The Bigtable to Cloud Storage Avro template is a pipeline that reads data from a Bigtable table and writes it to a Cloud Storage bucket in Avro format. You can use the template to move data from Bigtable to Cloud Storage.
Pipeline requirements
- The Bigtable table must exist.
 - The output Cloud Storage bucket must exist before running the pipeline.
 
Template parameters
Required parameters
- bigtableProjectId: The ID of the Google Cloud project that contains the Bigtable instance that you want to read data from.
 - bigtableInstanceId: The ID of the Bigtable instance that contains the table.
 - bigtableTableId: The ID of the Bigtable table to export.
 - outputDirectory: The Cloud Storage path where data is written. For example, 
gs://mybucket/somefolder. - filenamePrefix: The prefix of the Avro filename. For example, 
output-. Defaults to: part. 
Optional parameters
- bigtableAppProfileId: The ID of the Bigtable application profile to use for the export. If you don't specify an app profile, Bigtable uses the instance's default app profile: https://cloud.google.com/bigtable/docs/app-profiles#default-app-profile.
 
Run the template
Console
- Go to the Dataflow Create job from template page. Go to Create job from template
 - In the Job name field, enter a unique job name.
 - Optional: For Regional endpoint, select a value from the drop-down menu. The default
    region is 
us-central1.For a list of regions where you can run a Dataflow job, see Dataflow locations.
 - From the Dataflow template drop-down menu, select the Cloud Bigtable to Avro Files on Cloud Storage template .
 - In the provided parameter fields, enter your parameter values.
 - Click Run job.
 
gcloud
In your shell or terminal, run the template:
gcloud dataflow jobs run JOB_NAME \ --gcs-location gs://dataflow-templates-REGION_NAME/VERSION/Cloud_Bigtable_to_GCS_Avro \ --region REGION_NAME \ --parameters \ bigtableProjectId=BIGTABLE_PROJECT_ID,\ bigtableInstanceId=INSTANCE_ID,\ bigtableTableId=TABLE_ID,\ outputDirectory=OUTPUT_DIRECTORY,\ filenamePrefix=FILENAME_PREFIX
Replace the following:
JOB_NAME: a unique job name of your choiceVERSION: the version of the template that you want to useYou can use the following values:
latestto use the latest version of the template, which is available in the non-dated parent folder in the bucket— gs://dataflow-templates-REGION_NAME/latest/- the version name, like 
2023-09-12-00_RC00, to use a specific version of the template, which can be found nested in the respective dated parent folder in the bucket— gs://dataflow-templates-REGION_NAME/ 
REGION_NAME: the region where you want to deploy your Dataflow job—for example,us-central1BIGTABLE_PROJECT_ID: the ID of the Google Cloud project of the Bigtable instance that you want to read data fromINSTANCE_ID: the ID of the Bigtable instance that contains the tableTABLE_ID: the ID of the Bigtable table to exportOUTPUT_DIRECTORY: the Cloud Storage path where data is written, for example,gs://mybucket/somefolderFILENAME_PREFIX: the prefix of the Avro filename, for example,output-
API
To run the template using the REST API, send an HTTP POST request. For more information on the
    API and its authorization scopes, see
  projects.templates.launch.
POST https://dataflow.googleapis.com/v1b3/projects/PROJECT_ID/locations/LOCATION/templates:launch?gcsPath=gs://dataflow-templates-LOCATION/VERSION/Cloud_Bigtable_to_GCS_Avro { "jobName": "JOB_NAME", "parameters": { "bigtableProjectId": "BIGTABLE_PROJECT_ID", "bigtableInstanceId": "INSTANCE_ID", "bigtableTableId": "TABLE_ID", "outputDirectory": "OUTPUT_DIRECTORY", "filenamePrefix": "FILENAME_PREFIX", }, "environment": { "zone": "us-central1-f" } }
Replace the following:
PROJECT_ID: the Google Cloud project ID where you want to run the Dataflow jobJOB_NAME: a unique job name of your choiceVERSION: the version of the template that you want to useYou can use the following values:
latestto use the latest version of the template, which is available in the non-dated parent folder in the bucket— gs://dataflow-templates-REGION_NAME/latest/- the version name, like 
2023-09-12-00_RC00, to use a specific version of the template, which can be found nested in the respective dated parent folder in the bucket— gs://dataflow-templates-REGION_NAME/ 
LOCATION: the region where you want to deploy your Dataflow job—for example,us-central1BIGTABLE_PROJECT_ID: the ID of the Google Cloud project of the Bigtable instance that you want to read data fromINSTANCE_ID: the ID of the Bigtable instance that contains the tableTABLE_ID: the ID of the Bigtable table to exportOUTPUT_DIRECTORY: the Cloud Storage path where data is written, for example,gs://mybucket/somefolderFILENAME_PREFIX: the prefix of the Avro filename, for example,output-
What's next
- Learn about Dataflow templates.
 - See the list of Google-provided templates.