Pricing for Tabular Workflows

When you train a model using a Tabular Workflow, you are charged based on the cost of the infrastructure and the dependent services. When you make predictions with this model, you are charged based on the cost of the infrastructure.

The cost of the infrastructure depends on the following factors:

  • The number of machines that you use. You can set associated parameters during model training, batch prediction, or online prediction.
  • The type of machines that you use. You can set this parameter during model training, batch prediction, or online prediction.
  • The length of time for which the machines are in use.
    • If you are training a model or making batch predictions, this is a measure of the total processing time of the operation.
    • If you are making online predictions, this is a measure of the time that your model is deployed to an endpoint.

Tabular Workflows runs multiple dependent services in your project on your behalf: Dataflow, BigQuery, Cloud Storage, Vertex AI Pipelines, Vertex AI Training. You will be charged by these services directly.

Examples of training cost calculation

Example 1: 110MB dataset in CSV format, trained for one hour with default hardware configuration.

The cost breakdown for the default workflow with Architecture Search and Training is as follows:

Service Cost
Dataflow example and stats generation $2 (Dataflow ran 7 min)
Dataflow data and feature transformations $3 (Dataflow ran 10 min)
Vertex AI Training 0.8hr x $20 + 0.2hr x $20 + $3.3 SSD cost + pipeline container cost = $24 (48min tuning, 12min training)
Vertex AI Pipelines 1 run x $0.03 = $0.03
Total excluding model distillation $27.03

Optionally, you can enable model distillation to reduce resulting model size. The cost breakdown is as follows:

Service Cost
Total excluding model distillation $27.03
Vertex AI Training for model distillation $1
Dataflow data, feature transformations for model distillation $3 (Dataflow ran 10 min)
Batch prediction for model distillation $7
Total including model distillation $38.03

Example 2: 1.84TB dataset in BigQuery, trained for 20 hours with hardware override.

The hardware configuration for this example is as follows:

Hardware Configuration Name Value
stats_and_example_gen_dataflow_machine_type n1-standard-16
stats_and_example_gen_dataflow_max_num_workers 100
stats_and_example_gen_dataflow_disk_size_gb 40
transform_dataflow_machine_type n1-standard-16
transform_dataflow_max_num_workers 100
transform_dataflow_disk_size_gb 200
distill_batch_predict_machine_type n1-standard-2
distill_batch_predict_starting_replica_count 200
distill_batch_predict_max_replica_count 200

The cost breakdown for the default workflow with Architecture Search and Training is as follows:

Service Cost
Dataflow example and stats generation $518 (Dataflow ran 6 hours)
Dataflow data, feature transformations $471 (Dataflow ran 6 hours)
Vertex AI Training 17hr x $20 + 3hr x $20 + $41.5 SSD cost + pipeline container cost = $555 (17 hours tuning, 3 hours training)
Vertex AI Pipelines 1 run x $0.03 = $0.03
Total $1544.03