Managed Service for Apache Spark (formerly Dataproc) pricing

Pricing for Managed Service for Apache Spark depends on your chosen deployment mode: serverless or clusters.

Both models offer access to Lightning Engine, our next-generation vectorized execution engine that delivers up to 4.9x faster performance than open source Spark and up to 2x the price-performance over the leading high-speed Spark alternative.

Pricing Overview

  1. Serverless: Pay only for the exact resources consumed during job execution (per-second billing). You can choose between standard and premium performance tiers. Compare serverless performance tiers in greater detail.
  2. Clusters: Pay a flat management fee based on cluster uptime, Lightning Engine add-on (if selected), plus the cost of the underlying Compute Engine VMs and Persistent Disks.

For both deployment types, you pay for incurred Google Cloud StorageBigQuery storage, and network egress charges.

Serverless pricing

With the serverless deployment model, you pay only for the resources you use. You are billed per second based on Data Compute Units (DCUs), shuffle storage, and any attached accelerators. Scale-to-zero ensures you never pay for idle capacity.

Serverless deployments offer two performance tiers to match your workload requirements:

  • Standard: Ideal for cost-effective batch processing, data transformations, and general-purpose Spark ETL.
  • Premium: Powered by Lightning Engine. Ideal for demanding workloads, interactive data science, and GPU-accelerated AI/ML.

Compare serverless performance tiers in greater detail

Note: The following rates are examples for the us-central1 region.

Data Compute Unit (DCU) pricing

The DCU rate shown below is an hourly / monthly rate. It is prorated and billed per second, with a 1-minute minimum charge.

  • Johannesburg (africa-south1)
  • Taiwan (asia-east1)
  • Hong Kong (asia-east2)
  • Tokyo (asia-northeast1)
  • Osaka (asia-northeast2)
  • Seoul (asia-northeast3)
  • Mumbai (asia-south1)
  • Delhi (asia-south2)
  • Singapore (asia-southeast1)
  • Jakarta (asia-southeast2)
  • Bangkok (asia-southeast3)
  • Sydney (australia-southeast1)
  • Melbourne (australia-southeast2)
  • Warsaw (europe-central2)
  • Finland (europe-north1)
  • Stockholm (europe-north2)
  • Madrid (europe-southwest1)
  • Belgium (europe-west1)
  • Berlin (europe-west10)
  • Turin (europe-west12)
  • London (europe-west2)
  • Frankfurt (europe-west3)
  • Netherlands (europe-west4)
  • Zurich (europe-west6)
  • Milan (europe-west8)
  • Paris (europe-west9)
  • Doha (me-central1)
  • Dammam (me-central2)
  • Tel Aviv (me-west1)
  • Montreal (northamerica-northeast1)
  • Toronto (northamerica-northeast2)
  • Mexico (northamerica-south1)
  • Sao Paulo (southamerica-east1)
  • Santiago (southamerica-west1)
  • Iowa (us-central1)
  • South Carolina (us-east1)
  • Northern Virginia (us-east4)
  • Columbus (us-east5)
  • Alabama (us-east7)
  • Dallas (us-south1)
  • Oregon (us-west1)
  • Los Angeles (us-west2)
  • Salt Lake City (us-west3)
  • Las Vegas (us-west4)
Rabattoptionen einblenden

Type

Default* (USD)
BigQuery CUD - 1 Year* (USD)
BigQuery CUD - 3 Year* (USD)

Data Compute Unit (standard)

0,06 $0,054 $0,048 $

Data Compute Unit (premium)

0,089 $0,0801 $0,0712 $
* Jedes Nutzungsmodell hat eine eindeutige ID. Möglicherweise ist Ihre Zustimmung nötig, um Rabatte auf das Nutzungsmodell erhalten zu können. Hier finden Sie weitere Informationen.

If you pay in a currency other than USD, the prices listed in your currency on Cloud Platform SKUs apply.

Interactive workloads are charged at premium.

Shuffle storage pricing

The shuffle storage rate shown below is an hourly / monthly rate. It is prorated and billed per second, with a 1-minute minimum charge for standard shuffle storage and a 5-minute minimum charge for premium shuffle storage. Premium shuffle storage can only be used with premium compute unit.

  • Johannesburg (africa-south1)
  • Taiwan (asia-east1)
  • Hong Kong (asia-east2)
  • Tokyo (asia-northeast1)
  • Osaka (asia-northeast2)
  • Seoul (asia-northeast3)
  • Mumbai (asia-south1)
  • Delhi (asia-south2)
  • Singapore (asia-southeast1)
  • Jakarta (asia-southeast2)
  • Bangkok (asia-southeast3)
  • Sydney (australia-southeast1)
  • Melbourne (australia-southeast2)
  • Warsaw (europe-central2)
  • Finland (europe-north1)
  • Stockholm (europe-north2)
  • Madrid (europe-southwest1)
  • Belgium (europe-west1)
  • Berlin (europe-west10)
  • Turin (europe-west12)
  • London (europe-west2)
  • Frankfurt (europe-west3)
  • Netherlands (europe-west4)
  • Zurich (europe-west6)
  • Milan (europe-west8)
  • Paris (europe-west9)
  • Doha (me-central1)
  • Dammam (me-central2)
  • Tel Aviv (me-west1)
  • Montreal (northamerica-northeast1)
  • Toronto (northamerica-northeast2)
  • Mexico (northamerica-south1)
  • Sao Paulo (southamerica-east1)
  • Santiago (southamerica-west1)
  • Iowa (us-central1)
  • South Carolina (us-east1)
  • Northern Virginia (us-east4)
  • Columbus (us-east5)
  • Alabama (us-east7)
  • Dallas (us-south1)
  • Oregon (us-west1)
  • Los Angeles (us-west2)
  • Salt Lake City (us-west3)
  • Las Vegas (us-west4)

Type

Price (USD) per hour

Shuffle Storage (standard)

0,000054795 $ / 1 gibibyte hour

Shuffle Storage (premium)

0,000136986 $ / 1 gibibyte hour

If you pay in a currency other than USD, the prices listed in your currency on Cloud Platform SKUs apply.

Accelerator pricing

The accelerator rate shown below is an hourly / monthly rate. It is prorated and billed per second, with a 5-minute minimum charge.

  • Johannesburg (africa-south1)
  • Taiwan (asia-east1)
  • Hong Kong (asia-east2)
  • Tokyo (asia-northeast1)
  • Osaka (asia-northeast2)
  • Seoul (asia-northeast3)
  • Mumbai (asia-south1)
  • Delhi (asia-south2)
  • Singapore (asia-southeast1)
  • Jakarta (asia-southeast2)
  • Bangkok (asia-southeast3)
  • Sydney (australia-southeast1)
  • Melbourne (australia-southeast2)
  • Warsaw (europe-central2)
  • Finland (europe-north1)
  • Stockholm (europe-north2)
  • Madrid (europe-southwest1)
  • Belgium (europe-west1)
  • Berlin (europe-west10)
  • Turin (europe-west12)
  • London (europe-west2)
  • Frankfurt (europe-west3)
  • Netherlands (europe-west4)
  • Zurich (europe-west6)
  • Milan (europe-west8)
  • Paris (europe-west9)
  • Doha (me-central1)
  • Dammam (me-central2)
  • Tel Aviv (me-west1)
  • Montreal (northamerica-northeast1)
  • Toronto (northamerica-northeast2)
  • Mexico (northamerica-south1)
  • Sao Paulo (southamerica-east1)
  • Santiago (southamerica-west1)
  • Iowa (us-central1)
  • Oklahoma (us-central2)
  • South Carolina (us-east1)
  • Northern Virginia (us-east4)
  • Columbus (us-east5)
  • Alabama (us-east7)
  • Dallas (us-south1)
  • Oregon (us-west1)
  • Los Angeles (us-west2)
  • Salt Lake City (us-west3)
  • Las Vegas (us-west4)
  • Phoenix (us-west8)

Type

Price (USD) per hour

NVIDIA L4

0,672048287 $

NVIDIA a100 40GB

3,5206896 $

NVIDIA a100 80GB

4,713696 $

If you pay in a currency other than USD, the prices listed in your currency on Cloud Platform SKUs apply.

Serverless pricing examples

If the serverless batch workload runs with 12 DCUs for 24 hours in the us-central1 region and consumes 25GB of shuffle storage, the price calculation is as follows:

(spark.driver.cores=4,spark.executor.cores=4,spark.executor.instances=2)

Note: The example assumes a 30-day month. Since the batch workload duration is one day, the monthly shuffle storage rate is divided by 30.

Wird geladen...

If the serverless batch workload runs with 12 DCUs and 2 L4 GPUs for 24 hours in the us-central1 region and consumes 25GB of shuffle storage, the price calculation is as follows:

(spark.driver.cores=4,spark.executor.cores=4, spark.executor.instances=2,spark.dataproc.driver.compute.tier=premium, spark.dataproc.executor.compute.tier=premium, spark.dataproc.executor.disk.tier=premium, spark.dataproc.executor.resource.accelerator.type=l4)

Note: The example assumes a 30-day month. Since the batch workload duration is one day, the monthly shuffle storage rate is divided by 30.

Wird geladen...

If the serverless interactive workload runs with 12 DCUs for 24 hours in the us-central1 region and consumes 25GB of shuffle storage, the price calculation is as follows:

(spark.driver.cores=4,spark.executor.cores=4,spark.executor.instances=2)

Note: The example assumes a 30-day month. Since the batch workload duration is one day, the monthly shuffle storage rate is divided by 30.

Wird geladen...

Serverless pricing estimation

When a serverless batch workload completes, Managed Service for Apache Spark calculates UsageMetrics, which contain an approximation of the total DCU, accelerator, and shuffle storage resources consumed by the completed workload. After running a workload, you can run the gcloud dataproc batches describe BATCH_ID command to view workload usage metrics to help you estimate the cost of running the workload.

Example:

Managed Service for Apache Spark runs a serverless workload on an ephemeral cluster with one master and two workers. Each node consumes 4 DCUs (default is 4 DCUs per core—see spark.dataproc.driver.disk.size) and 400 GB shuffle storage (default is 100GB per core—see spark.driver.cores). Workload run time is 60 seconds. Also, each worker has 1 GPU for a total of 2 across the cluster.

The user runs gcloud dataproc batches describe BATCH_ID --region REGION to obtain usage metrics. The command output includes the following snippet (milliDcuSeconds: 4 DCUs x 3 VMs x 60 seconds x 1000 = 720000, milliAcceleratorSeconds: 1 GPU x 2 VMs x 60 seconds x 1000 = 120000, and shuffleStorageGbSeconds: 400GB x 3 VMs x 60 seconds = 72000):

Note: The example assumes a 30-day month. Since the batch workload duration is one day, the monthly shuffle storage rate is divided by 30.

Wird geladen...

Serverless use of other Google Cloud resources

Your serverless Spark workload can optionally utilize the following resources, each billed at its own pricing, including but not limited to:

Clusters pricing

With the clusters deployment model, you pay for cluster uptime billed by the second. All clusters are billed in one-second clock-time increments, subject to a 1-minute minimum billing plus the cost of the underlying Google Cloud infrastructure.

  • Clusters pricing is in addition to the Compute Engine per-instance price for each virtual machine
  • Underlying Compute Engine VMs are eligible for Committed Use Discounts (CUDs)
  • The Lightning Engine add-on incurs additional charges
  • Spot VMs can be used to lower your Compute Engine costs for clusters
  • See supported machine types for information on the predefined and custom machine types you can use with clusters

Management Fee

The management fee is billed by the second, subject to a 1-minute minimum. It applies to all vCPUs across your master, worker, and secondary worker nodes.

The pricing formula is: $0.010 * # of vCPUs * hourly duration.

Price per vCPU / hour

0,01 $ / 1 hour

Lightning Engine add-on

Bring breakthrough performance by adding Lightning Engine to your clusters. You can experience up to 4.9x faster execution than open source Spark and up to 2x the price-performance over the leading high-speed Spark alternative

Read the Lightning Engine user guide for clusters

Price per vCPU / hour

Through 5/31/26: $0.0000 / 1 hour

Starting 6/1/26: $0.0025 / 1 hour

Clusters accrued charges

The following Managed Service for Apache Spark cluster operations and scenarios cause charges to accrue:

  • Scaling and autoscaling: When VMs are added to the cluster, charges accrue while the VMs are active. These accrued charges continue until the VMs are removed.
  • Clusters in an error state: When a cluster is in an error state, cluster VMs remain active and charges continue to accrue. These accrued charges continue until the cluster is deleted.

Clusters pricing example

Persistent cluster (standard tier)

You run persistent clusters for 730 hours (one month) in us-central1. The cluster consists of 1 master node and 4 worker nodes, totaling 20 vCPUs.


Wird geladen...

Clusters use of other Google Cloud resources

As a managed and integrated solution, Managed Service for Apache Spark clusters are built on top of other Google Cloud technologies. Clusters consume the following resources, each billed at its own pricing:

Clusters can optionally utilize the following resources, each billed at its own pricing, including but not limited to:

Clusters on GKE pricing

This section explains the charges that apply only to the virtual clusters that runs on a user-managed Google Kubernetes Engine. See GKE pricing to learn about the added charges that apply to the user-managed GKE cluster.

The Managed clusters on GKE pricing formula, $0.010 * # of vCPUs * hourly duration, is the same as clusters on Compute Engine pricing formula, and is applied to the aggregate number of virtual CPUs running in VMs instances in Managed Service for Apache Spark-created node pools in the cluster. The duration of a virtual machine instance is the length of time from its creation to its deletion. Clusters on GKE are billed by the second, subject to a 1-minute minimum billing per virtual machine instance. Other Google Cloud charges are applied in addition to Managed Service for Apache Spark charges.

Managed Service for Apache Spark-created node pools continue to exist after deletion of the cluster since they may be shared by multiple clusters. If you delete the node pools or scale node pools down to zero instances, continued Managed Service for Apache Spark charges will not be incurred. Any remaining node pool VMs will continue to incur charges until you delete them.

What's next

Request a custom quote

With Google Cloud's pay-as-you-go pricing, you only pay for the services you use. Connect with our sales team to get a custom quote for your organization.
Google Cloud