Quickstart using the console

This page shows you how to use the Google Cloud Console to create a Dataproc cluster, run a simple Apache Spark job in the cluster, then modify the number of workers in the cluster.

You can find out how to do the same tasks with Quickstarts Using the API Explorer and Quickstart using the gcloud command-line tool.

Before you begin

  1. Sign in to your Google Cloud account. If you're new to Google Cloud, create an account to evaluate how our products perform in real-world scenarios. New customers also get $300 in free credits to run, test, and deploy workloads.
  2. In the Google Cloud Console, on the project selector page, select or create a Google Cloud project.

    Go to project selector

  3. Make sure that billing is enabled for your Cloud project. Learn how to confirm that billing is enabled for your project.

  4. Enable the Dataproc API.

    Enable the API

Create a cluster

  1. Go to the Cloud Console Cloud Dataproc Clusters page.
  2. Click Create cluster.
  3. Enter example-cluster in the Name field.
  4. Select a region and zone for the cluster from the Region and Zone drop-down menus. You can select a distinct region, such as us-east1 or europe-west1, to isolate resources (including VM instances and Cloud Storage) and metadata storage locations utilized by Dataproc within the specified region. You can also select a global region, which is a special multi-region endpoint that is capable of deploying instances into any user-specified Compute Engine zone. See Regional endpoints to learn more about the difference between global and regional endpoints. See Available regions & zones for information on selecting a region and zone.
  5. Use the provided defaults for all the other options.
  6. Click Create to create the cluster.

Your new cluster should appear in the Clusters list. Cluster status is listed as "Provisioning" until the cluster is ready to use, then changes to "Running."

Submit a job

To run a sample Spark job:

  1. Select Jobs in the left nav to switch to Dataproc's jobs view.
  2. Click Submit job.
  3. You can accept the Job ID or provide your own, which must be unique within the project.
  4. Select the Region of your new example-cluster.
  5. Select example-cluster from the Cluster drop-down menu.
  6. Select Spark from the Job type drop-down menu.
  7. Enter org.apache.spark.examples.SparkPi in the Main class or jar field.
  8. Enter file:///usr/lib/spark/examples/jars/spark-examples.jar in the Jar files field.
  9. Enter 1000 in the Arguments field to set the number of tasks.
  10. Click Submit.

Your job should appear in the Jobs list, which shows your project's jobs with their cluster, type, and current status. Job status displays as "Running," and then "Succeeded" after it completes. To see your completed job's output:

  1. Click the job ID in the Jobs list.
  2. Select Line Wrapping to avoid scrolling.

You should see that your job has successfully calculated a rough value for pi!

Update a cluster

To change the number of worker instances in your cluster:

  1. Select Clusters in the left navigation pane to return to the Cloud Dataproc Clusters view.
  2. Click example-cluster in the Clusters list. By default, the page displays an overview of your cluster's CPU usage.
  3. Click Configuration to display your cluster's current settings.
  4. Click Edit. The number of worker nodes is now editable.
  5. Enter 5 in the Worker nodes field.
  6. Click Save.

Your cluster is now updated. You can follow the same procedure to decrease the number of worker nodes to the original value.

Clean up

To avoid incurring charges to your Google Cloud account for the resources used in this page, follow these steps.

  1. On the example-cluster Cluster page, click Delete to delete the cluster. You are prompted to confirm that you want to delete the cluster. Click OK.

What's next