gcloud beta dataflow jobs run

NAME
gcloud beta dataflow jobs run - runs a job from the specified path
SYNOPSIS
gcloud beta dataflow jobs run JOB_NAME --gcs-location=GCS_LOCATION [--additional-experiments=[ADDITIONAL_EXPERIMENTS,…]] [--dataflow-kms-key=DATAFLOW_KMS_KEY] [--disable-public-ips] [--enable-streaming-engine] [--max-workers=MAX_WORKERS] [--network=NETWORK] [--num-workers=NUM_WORKERS] [--parameters=[PARAMETERS,…]] [--region=REGION_ID] [--service-account-email=SERVICE_ACCOUNT_EMAIL] [--staging-location=STAGING_LOCATION] [--subnetwork=SUBNETWORK] [--worker-machine-type=WORKER_MACHINE_TYPE] [--worker-region=WORKER_REGION     | --worker-zone=WORKER_ZONE     | --zone=ZONE] [GCLOUD_WIDE_FLAG]
DESCRIPTION
(BETA) Runs a job from the specified path.
POSITIONAL ARGUMENTS
JOB_NAME
The unique name to assign to the job.
REQUIRED FLAGS
--gcs-location=GCS_LOCATION
The Google Cloud Storage location of the job template to run. (Must be a URL beginning with 'gs://'.)
OPTIONAL FLAGS
--additional-experiments=[ADDITIONAL_EXPERIMENTS,…]
Additional experiments to pass to the job. These experiments are appended to any experiments already set by the template.
--dataflow-kms-key=DATAFLOW_KMS_KEY
The Cloud KMS key to protect the job resources.
--disable-public-ips
The Cloud Dataflow workers must not use public IP addresses. Overrides the default dataflow/disable_public_ips property value for this command invocation.
--enable-streaming-engine
Enabling Streaming Engine for the streaming job. Overrides the default dataflow/enable_streaming_engine property value for this command invocation.
--max-workers=MAX_WORKERS
The maximum number of workers to run.
--network=NETWORK
The Compute Engine network for launching instances to run your pipeline.
--num-workers=NUM_WORKERS
The initial number of workers to use.
--parameters=[PARAMETERS,…]
The parameters to pass to the job.
--region=REGION_ID
Region ID of the job's regional endpoint. Defaults to 'us-central1'.
--service-account-email=SERVICE_ACCOUNT_EMAIL
The service account to run the workers as.
--staging-location=STAGING_LOCATION
The Google Cloud Storage location to stage temporary files. (Must be a URL beginning with 'gs://'.)
--subnetwork=SUBNETWORK
The Compute Engine subnetwork for launching instances to run your pipeline.
--worker-machine-type=WORKER_MACHINE_TYPE
The type of machine to use for workers. Defaults to server-specified.
Worker location options.

At most one of these can be specified:

--worker-region=WORKER_REGION
The region to run the workers in.
--worker-zone=WORKER_ZONE
The zone to run the workers in.
--zone=ZONE
(DEPRECATED) The zone to run the workers in.

The --zone option is deprecated; use --worker-region or --worker-zone instead.

GCLOUD WIDE FLAGS
These flags are available to all commands: --access-token-file, --account, --billing-project, --configuration, --flags-file, --flatten, --format, --help, --impersonate-service-account, --log-http, --project, --quiet, --trace-token, --user-output-enabled, --verbosity.

Run $ gcloud help for details.

NOTES
This command is currently in beta and might change without notice. This variant is also available:
gcloud dataflow jobs run