Module types (0.8.1)

API documentation for dataproc_v1beta2.types module.

Classes

AcceleratorConfig

Specifies the type and number of accelerator cards attached to the instances of an instance group (see GPUs on Compute Engine <https://cloud.google.com/compute/docs/gpus/>__). .. attribute:: accelerator_type_uri

Full URL, partial URI, or short name of the accelerator type resource to expose to this instance. See Compute Engine AcceleratorTypes <https://cloud.google.com/compute/docs/refere nce/beta/acceleratorTypes> Examples * https://www.googl eapis.com/compute/beta/projects/[project_id]/zones/us- east1-a/acceleratorTypes/nvidia-tesla-k80 * projects/[project_id]/zones/us- east1-a/acceleratorTypes/nvidia-tesla-k80 * nvidia- tesla-k80 Auto Zone Exception: If you are using the Dataproc Auto Zone Placement <https://cloud.google.com/dataproc/docs/concepts/configuring- clusters/auto-zone#using_auto_zone_placement> feature, you must use the short name of the accelerator type resource, for example, nvidia-tesla-k80.

Any

API documentation for dataproc_v1beta2.types.Any class.

AutoscalingConfig

Autoscaling Policy config associated with the cluster. .. attribute:: policy_uri

Optional. The autoscaling policy used by the cluster. Only resource names including projectid and location (region) are valid. Examples: - https://www.googleapis.com/compute/v1/p rojects/[project_id]/locations/[dataproc_region]/autoscalingPo licies/[policy_id] - projects/[project_id]/locations/[dat aproc_region]/autoscalingPolicies/[policy_id] Note that the policy must be in the same project and Dataproc region.

AutoscalingPolicy

Describes an autoscaling policy for Dataproc cluster autoscaler. .. attribute:: id

Required. The policy id. The id must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), and hyphens (-). Cannot begin or end with underscore or hyphen. Must consist of between 3 and 50 characters.

Required. Autoscaling algorithm for policy.

Optional. Describes how the autoscaler will operate for secondary workers.

BasicAutoscalingAlgorithm

Basic algorithm for autoscaling. .. attribute:: yarn_config

Required. YARN autoscaling configuration.

BasicYarnAutoscalingConfig

Basic autoscaling configurations for YARN. .. attribute:: graceful_decommission_timeout

Required. Timeout for YARN graceful decommissioning of Node Managers. Specifies the duration to wait for jobs to complete before forcefully removing workers (and potentially interrupting jobs). Only applicable to downscaling operations. Bounds: [0s, 1d].

Required. Fraction of average pending memory in the last cooldown period for which to remove workers. A scale-down factor of 1 will result in scaling down so that there is no available memory remaining after the update (more aggressive scaling). A scale-down factor of 0 disables removing workers, which can be beneficial for autoscaling a single job. Bounds: [0.0, 1.0].

Optional. Minimum scale-down threshold as a fraction of total cluster size before scaling occurs. For example, in a 20-worker cluster, a threshold of 0.1 means the autoscaler must recommend at least a 2 worker scale-down for the cluster to scale. A threshold of 0 means the autoscaler will scale down on any recommended change. Bounds: [0.0, 1.0]. Default: 0.0.

CancelJobRequest

A request to cancel a job. .. attribute:: project_id

Required. The ID of the Google Cloud Platform project that the job belongs to.

Required. The job ID.

CancelOperationRequest

API documentation for dataproc_v1beta2.types.CancelOperationRequest class.

Cluster

Describes the identifying information, config, and status of a cluster of Compute Engine instances. .. attribute:: project_id

Required. The Google Cloud Platform project ID that the cluster belongs to.

Required. The cluster config. Note that Dataproc may set default values, and values may change when clusters are updated.

Output only. Cluster status.

Output only. A cluster UUID (Unique Universal Identifier). Dataproc generates this value when it creates the cluster.

ClusterConfig

The cluster config. .. attribute:: config_bucket

Optional. A Cloud Storage bucket used to stage job dependencies, config files, and job driver console output. If you do not specify a staging bucket, Cloud Dataproc will determine a Cloud Storage location (US, ASIA, or EU) for your cluster’s staging bucket according to the Compute Engine zone where your cluster is deployed, and then create and manage this project-level, per-location bucket (see Dataproc staging bucket <https://cloud.google.com/dataproc/docs/concepts/configuring- clusters/staging-bucket>__).

Optional. The Compute Engine config settings for the master instance in a cluster.

Optional. The Compute Engine config settings for additional worker instances in a cluster.

Optional. The config setting for auto delete cluster schedule.

Optional. Encryption settings for the cluster.

Optional. Port/endpoint configuration for this cluster

Optional. The Kubernetes Engine config for Dataproc clusters deployed to Kubernetes. Setting this is considered mutually exclusive with Compute Engine-based options such as gce_cluster_config, master_config, worker_config, secondary_worker_config, and autoscaling_config.

ClusterMetrics

Contains cluster daemon metrics, such as HDFS and YARN stats. Beta Feature: This report is available for testing purposes only. It may be changed before final release. .. attribute:: hdfs_metrics

The HDFS metrics.

ClusterOperation

The cluster operation triggered by a workflow. .. attribute:: operation_id

Output only. The id of the cluster operation.

Output only. Indicates the operation is done.

ClusterOperationMetadata

Metadata describing the operation. .. attribute:: cluster_name

Output only. Name of the cluster for the operation.

Output only. Current operation status.

Output only. The operation type.

Output only. Labels associated with the operation

ClusterOperationStatus

The status of the operation. .. attribute:: state

Output only. A message containing the operation state.

Output only. A message containing any operation metadata details.

ClusterSelector

A selector that chooses target cluster for jobs based on metadata. .. attribute:: zone

Optional. The zone where workflow process executes. This parameter does not affect the selection of the cluster. If unspecified, the zone of the first cluster matching the selector is used.

ClusterStatus

The status of a cluster and its instances. .. attribute:: state

Output only. The cluster’s state.

Output only. Time when this state was entered (see JSON representation of Timestamp <https://developers.google.com/protocol- buffers/docs/proto3#json>__).

CreateAutoscalingPolicyRequest

A request to create an autoscaling policy. .. attribute:: parent

Required. The “resource name” of the region or location, as described in https://cloud.google.com/apis/design/resource_names. - For projects.regions.autoscalingPolicies.create, the resource name has the following format: projects/{project_id}/regions/{region} - For projects.locations.autoscalingPolicies.create, the resource name has the following format: projects/{project_id}/locations/{location}

CreateClusterRequest

A request to create a cluster. .. attribute:: project_id

Required. The ID of the Google Cloud Platform project that the cluster belongs to.

Required. The cluster to create.

CreateWorkflowTemplateRequest

A request to create a workflow template. .. attribute:: parent

Required. The resource name of the region or location, as described in https://cloud.google.com/apis/design/resource_names. - For projects.regions.workflowTemplates,create, the resource name of the region has the following format: projects/{project_id}/regions/{region} - For projects.locations.workflowTemplates.create, the resource name of the location has the following format: projects/{project_id}/locations/{location}

DeleteAutoscalingPolicyRequest

A request to delete an autoscaling policy. Autoscaling policies in use by one or more clusters will not be deleted. .. attribute:: name

Required. The “resource name” of the autoscaling policy, as described in https://cloud.google.com/apis/design/resource_names. - For projects.regions.autoscalingPolicies.delete, the resource name of the policy has the following format: projects/{pr oject_id}/regions/{region}/autoscalingPolicies/{policy_id}

  • For projects.locations.autoscalingPolicies.delete, the resource name of the policy has the following format: projects/{project_id}/locations/{location}/autoscalingPolicies /{policy_id}

DeleteClusterRequest

A request to delete a cluster. .. attribute:: project_id

Required. The ID of the Google Cloud Platform project that the cluster belongs to.

Required. The cluster name.

Optional. A unique id used to identify the request. If the server receives two [DeleteClusterRequest][google.cloud.datapr oc.v1beta2.DeleteClusterRequest] requests with the same id, then the second request will be ignored and the first google.longrunning.Operation][google.longrunning.Operation] created and stored in the backend is returned. It is recommended to always set this value to a UUID <https://en.wi kipedia.org/wiki/Universally_unique_identifier>_. The id must contain only letters (a-z, A-Z), numbers (0-9), underscores (), and hyphens (-). The maximum length is 40 characters.

DeleteJobRequest

A request to delete a job. .. attribute:: project_id

Required. The ID of the Google Cloud Platform project that the job belongs to.

Required. The job ID.

DeleteOperationRequest

API documentation for dataproc_v1beta2.types.DeleteOperationRequest class.

DeleteWorkflowTemplateRequest

A request to delete a workflow template. Currently started workflows will remain running. .. attribute:: name

Required. The resource name of the workflow template, as described in https://cloud.google.com/apis/design/resource_names. - For projects.regions.workflowTemplates.delete, the resource name of the template has the following format: project s/{project_id}/regions/{region}/workflowTemplates/{template_id } - For projects.locations.workflowTemplates.instantiate, the resource name of the template has the following format: p rojects/{project_id}/locations/{location}/workflowTemplates/{t emplate_id}

DiagnoseClusterRequest

A request to collect cluster diagnostic information. .. attribute:: project_id

Required. The ID of the Google Cloud Platform project that the cluster belongs to.

Required. The cluster name.

DiagnoseClusterResults

The location of diagnostic output. .. attribute:: output_uri

Output only. The Cloud Storage URI of the diagnostic output. The output report is a plain text file with a summary of collected diagnostics.

DiskConfig

Specifies the config of disk options for a group of VM instances. .. attribute:: boot_disk_type

Optional. Type of the boot disk (default is “pd-standard”). Valid values: “pd-ssd” (Persistent Disk Solid State Drive) or “pd-standard” (Persistent Disk Hard Disk Drive).

Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS <https://hadoop.apache.org/docs/r1.2.1/hdfs_user_gui de.html>__ data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.

Duration

API documentation for dataproc_v1beta2.types.Duration class.

Empty

API documentation for dataproc_v1beta2.types.Empty class.

EncryptionConfig

Encryption settings for the cluster. .. attribute:: gce_pd_kms_key_name

Optional. The Cloud KMS key name to use for PD disk encryption for all instances in the cluster.

EndpointConfig

Endpoint config for this cluster .. attribute:: http_ports

Output only. The map of port descriptions to URLs. Will only be populated if enable_http_port_access is true.

FieldMask

API documentation for dataproc_v1beta2.types.FieldMask class.

GceClusterConfig

Common config settings for resources of Compute Engine cluster instances, applicable to all instances in the cluster. .. attribute:: zone_uri

Optional. The zone where the Compute Engine cluster will be located. On a create request, it is required in the “global” region. If omitted in a non-global Dataproc region, the service will pick a zone in the corresponding Compute Engine region. On a get request, zone will always be present. A full URL, partial URI, or short name are valid. Examples: - htt ps://www.googleapis.com/compute/v1/projects/[project_id]/zones /[zone] - projects/[project_id]/zones/[zone] - us- central1-f

Optional. The Compute Engine subnetwork to be used for machine communications. Cannot be specified with network_uri. A full URL, partial URI, or short name are valid. Examples: - htt ps://www.googleapis.com/compute/v1/projects/[project_id]/regio ns/us-east1/subnetworks/sub0 - projects/[project_id]/regions/us-east1/subnetworks/sub0 - sub0

Optional. The Dataproc service account <https://cloud.google.com/dataproc/docs/concepts/configuring- clusters/service- accounts#service_accounts_in_cloud_dataproc> (also see VM Data Plane identity <https://cloud.google.com/dataproc/docs/concepts/iam/dataproc- principals#vm_service_account_data_plane_identity>) used by Dataproc cluster VM instances to access Google Cloud Platform services. If not specified, the Compute Engine default service account <https://cloud.google.com/compute/docs/access/service- accounts#default_service_account>__ is used.

The Compute Engine tags to add to all instances (see Tagging instances <https://cloud.google.com/compute/docs/label-or-tag- resources#tags>__).

Optional. Reservation Affinity for consuming Zonal reservation.

GetAutoscalingPolicyRequest

A request to fetch an autoscaling policy. .. attribute:: name

Required. The “resource name” of the autoscaling policy, as described in https://cloud.google.com/apis/design/resource_names. - For projects.regions.autoscalingPolicies.get, the resource name of the policy has the following format: projects/ {project_id}/regions/{region}/autoscalingPolicies/{policy_id}` ` - Forprojects.locations.autoscalingPolicies.get, the resource name of the policy has the following format: projects/{project_id}/locations/{location}/autoscalingPolicies /{policy_id}``

GetClusterRequest

Request to get the resource representation for a cluster in a project. .. attribute:: project_id

Required. The ID of the Google Cloud Platform project that the cluster belongs to.

Required. The cluster name.

GetJobRequest

A request to get the resource representation for a job in a project. .. attribute:: project_id

Required. The ID of the Google Cloud Platform project that the job belongs to.

Required. The job ID.

GetOperationRequest

API documentation for dataproc_v1beta2.types.GetOperationRequest class.

GetWorkflowTemplateRequest

A request to fetch a workflow template. .. attribute:: name

Required. The resource name of the workflow template, as described in https://cloud.google.com/apis/design/resource_names. - For projects.regions.workflowTemplates.get, the resource name of the template has the following format: projects/{pr oject_id}/regions/{region}/workflowTemplates/{template_id}

  • For projects.locations.workflowTemplates.get, the resource name of the template has the following format: projects/{project_id}/locations/{location}/workflowTemplates /{template_id}

GkeClusterConfig

The GKE config for this cluster. .. attribute:: namespaced_gke_deployment_target

Optional. A target for the deployment.

HadoopJob

A Dataproc job for running Apache Hadoop MapReduce <https://hadoop.apache.org/docs/current/hadoop-mapreduce- client/hadoop-mapreduce-client-core/MapReduceTutorial.html> jobs on Apache Hadoop YARN <https://hadoop.apache.org/docs/r2.7.1/hadoop- yarn/hadoop-yarn-site/YARN.html>. .. attribute:: driver

Required. Indicates the location of the driver’s main class. Specify either the jar file that contains the main class or the main class name. To specify both, add the jar file to jar_file_uris, and then specify the main class name in this property.

The name of the driver’s main class. The jar file containing the class must be in the default CLASSPATH or specified in jar_file_uris.

Optional. Jar file URIs to add to the CLASSPATHs of the Hadoop driver and tasks.

Optional. HCFS URIs of archives to be extracted in the working directory of Hadoop drivers and tasks. Supported file types: .jar, .tar, .tar.gz, .tgz, or .zip.

Optional. The runtime log config for job execution.

HiveJob

A Dataproc job for running Apache Hive <https://hive.apache.org/>__ queries on YARN. .. attribute:: queries

Required. The sequence of Hive queries to execute, specified as either an HCFS file URI or a list of queries.

A list of queries.

Optional. Mapping of query variable names to values (equivalent to the Hive command: SET name="value";).

Optional. HCFS URIs of jar files to add to the CLASSPATH of the Hive server and Hadoop MapReduce (MR) tasks. Can contain Hive SerDes and UDFs.

InstanceGroupAutoscalingPolicyConfig

Configuration for the size bounds of an instance group, including its proportional size to other groups. .. attribute:: min_instances

Optional. Minimum number of instances for this group. Primary workers - Bounds: [2, max_instances]. Default: 2. Secondary workers - Bounds: [0, max_instances]. Default: 0.

Optional. Weight for the instance group, which is used to determine the fraction of total workers in the cluster from this instance group. For example, if primary workers have weight 2, and secondary workers have weight 1, the cluster will have approximately 2 primary workers for each secondary worker. The cluster may not reach the specified balance if constrained by min/max bounds or other autoscaling settings. For example, if max_instances for secondary workers is 0, then only primary workers will be added. The cluster can also be out of balance when created. If weight is not set on any instance group, the cluster will default to equal weight for all groups: the cluster will attempt to maintain an equal number of workers in each group within the configured size bounds for each group. If weight is set for one group only, the cluster will default to zero weight on the unset group. For example if weight is set only on primary workers, the cluster will use primary workers only and no secondary workers.

InstanceGroupConfig

The config settings for Compute Engine resources in an instance group, such as a master or worker group. .. attribute:: num_instances

Optional. The number of VM instances in the instance group. For master instance groups, must be set to 1.

Optional. The Compute Engine image resource used for cluster instances. The URI can represent an image or image family. Image examples: - https://www.googleapis.com/compute/beta/ projects/[project_id]/global/images/[image-id] - projects/[project_id]/global/images/[image-id] - image- id Image family examples. Dataproc will use the most recent image from the family: - https://www.googleapis.com/comput e/beta/projects/[project_id]/global/images/family/[custom- image-family-name] - projects/[project_id]/global/images/family/[custom-image- family-name] If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.

Optional. Disk option config settings.

Output only. The config for Compute Engine Instance Group Manager that manages this group. This is only used for preemptible instance groups.

Specifies the minimum cpu platform for the Instance Group. See Dataproc -> Minimum CPU Platform <https://cloud.google.com/da taproc/docs/concepts/compute/dataproc-min-cpu>__.

InstantiateInlineWorkflowTemplateRequest

A request to instantiate an inline workflow template. .. attribute:: parent

Required. The resource name of the region or location, as described in https://cloud.google.com/apis/design/resource_names. - For projects.regions.workflowTemplates,instantiateinline, the resource name of the region has the following format: projects/{project_id}/regions/{region} - For projects.locations.workflowTemplates.instantiateinline, the resource name of the location has the following format: projects/{project_id}/locations/{location}

Deprecated. Please use request_id field instead.

InstantiateWorkflowTemplateRequest

A request to instantiate a workflow template. .. attribute:: name

Required. The resource name of the workflow template, as described in https://cloud.google.com/apis/design/resource_names. - For projects.regions.workflowTemplates.instantiate, the resource name of the template has the following format: projects/{project_id}/regions/{region}/workflowTemplates/{te mplate_id} - For projects.locations.workflowTemplates.instantiate, the resource name of the template has the following format: p rojects/{project_id}/locations/{location}/workflowTemplates/{t emplate_id}

Deprecated. Please use request_id field instead.

Optional. Map from parameter names to values that should be used for those parameters. Values may not exceed 100 characters.

Job

A Dataproc job resource. .. attribute:: reference

Optional. The fully qualified reference to the job, which can be used to obtain the equivalent REST path of the job resource. If this property is not specified when a job is created, the server generates a job_id.

Required. The application/framework-specific portion of the job.

Optional. Job is a Spark job.

Optional. Job is a Hive job.

Optional. Job is a SparkR job.

Optional. Job is a Presto job.

Output only. The previous job status.

Output only. The email address of the user submitting the job. For jobs submitted on the cluster, the address is username@hostname.

Output only. If present, the location of miscellaneous control files which may be used as part of job setup and handling. If not present, control files may be placed in the same location as driver_output_uri.

Optional. Job scheduling configuration.

Output only. Indicates whether the job is completed. If the value is false, the job is still in progress. If true, the job is completed, and status.state field will indicate if it was successful, failed, or cancelled.

JobMetadata

Job Operation metadata. .. attribute:: job_id

Output only. The job id.

Output only. Operation type.

JobPlacement

Dataproc job config. .. attribute:: cluster_name

Required. The name of the cluster where the job will be submitted.

JobReference

Encapsulates the full scoping used to reference a job. .. attribute:: project_id

Required. The ID of the Google Cloud Platform project that the job belongs to.

JobScheduling

Job scheduling options. .. attribute:: max_failures_per_hour

Optional. Maximum number of times per hour a driver may be restarted as a result of driver terminating with non-zero code before job is reported failed. A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window. Maximum value is 10.

JobStatus

Dataproc job status. .. attribute:: state

Output only. A state message specifying the overall job state.

Output only. The time when this state was entered.

KerberosConfig

Specifies Kerberos related configuration. .. attribute:: enable_kerberos

Optional. Flag to indicate whether to Kerberize the cluster (default: false). Set this field to true to enable Kerberos on a cluster.

Required. The uri of the KMS key used to encrypt various sensitive files.

Optional. The Cloud Storage URI of the truststore file used for SSL encryption. If not provided, Dataproc will provide a self-signed certificate.

Optional. The Cloud Storage URI of a KMS encrypted file containing the password to the user provided key. For the self-signed certificate, this password is generated by Dataproc.

Optional. The remote realm the Dataproc on-cluster KDC will trust, should the user enable cross realm trust.

Optional. The admin server (IP or hostname) for the remote trusted realm in a cross realm trust relationship.

Optional. The Cloud Storage URI of a KMS encrypted file containing the master key of the KDC database.

Optional. The name of the on-cluster Kerberos realm. If not specified, the uppercased domain of hostnames will be the realm.

LifecycleConfig

Specifies the cluster auto-delete schedule configuration. .. attribute:: idle_delete_ttl

Optional. The duration to keep the cluster alive while idling (when no jobs are running). Passing this threshold will cause the cluster to be deleted. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration <https://developers.google.com/protocol- buffers/docs/proto3#json>__.

Optional. The time when cluster will be auto-deleted. (see JSON representation of Timestamp <https://developers.google.com/protocol- buffers/docs/proto3#json>__).

Output only. The time when cluster became idle (most recent job finished) and became eligible for deletion due to idleness (see JSON representation of Timestamp <https://developers.google.com/protocol- buffers/docs/proto3#json>__).

ListAutoscalingPoliciesRequest

A request to list autoscaling policies in a project. .. attribute:: parent

Required. The “resource name” of the region or location, as described in https://cloud.google.com/apis/design/resource_names. - For projects.regions.autoscalingPolicies.list, the resource name of the region has the following format: projects/{project_id}/regions/{region} - For projects.locations.autoscalingPolicies.list, the resource name of the location has the following format: projects/{project_id}/locations/{location}

Optional. The page token, returned by a previous call, to request the next page of results.

ListAutoscalingPoliciesResponse

A response to a request to list autoscaling policies in a project. .. attribute:: policies

Output only. Autoscaling policies list.

ListClustersRequest

A request to list the clusters in a project. .. attribute:: project_id

Required. The ID of the Google Cloud Platform project that the cluster belongs to.

Optional. A filter constraining the clusters to list. Filters are case-sensitive and have the following syntax: field = value [AND [field = value]] … where field is one of status.state, clusterName, or labels.[KEY], and [KEY] is a label key. value can be * to match all values. status.state can be one of the following: ACTIVE, INACTIVE, CREATING, RUNNING, ERROR, DELETING, or UPDATING. ACTIVE contains the CREATING, UPDATING, and RUNNING states. INACTIVE contains the DELETING and ERROR states. clusterName is the name of the cluster provided at creation time. Only the logical AND operator is supported; space-separated items are treated as having an implicit AND operator. Example filter: status.state = ACTIVE AND clusterName = mycluster AND labels.env = staging AND labels.starred = *

Optional. The standard List page token.

ListClustersResponse

The list of all clusters in a project. .. attribute:: clusters

Output only. The clusters in the project.

ListJobsRequest

A request to list jobs in a project. .. attribute:: project_id

Required. The ID of the Google Cloud Platform project that the job belongs to.

Optional. The number of results to return in each response.

Optional. If set, the returned jobs list includes only jobs that were submitted to the named cluster.

Optional. A filter constraining the jobs to list. Filters are case-sensitive and have the following syntax: [field = value] AND [field [= value]] … where field is status.state or labels.[KEY], and [KEY] is a label key. value can be * to match all values. status.state can be either ACTIVE or NON_ACTIVE. Only the logical AND operator is supported; space-separated items are treated as having an implicit AND operator. Example filter: status.state = ACTIVE AND labels.env = staging AND labels.starred = *

ListJobsResponse

A list of jobs in a project. .. attribute:: jobs

Output only. Jobs list.

ListOperationsRequest

API documentation for dataproc_v1beta2.types.ListOperationsRequest class.

ListOperationsResponse

API documentation for dataproc_v1beta2.types.ListOperationsResponse class.

ListWorkflowTemplatesRequest

A request to list workflow templates in a project. .. attribute:: parent

Required. The resource name of the region or location, as described in https://cloud.google.com/apis/design/resource_names. - For projects.regions.workflowTemplates,list, the resource name of the region has the following format: projects/{project_id}/regions/{region} - For projects.locations.workflowTemplates.list, the resource name of the location has the following format: projects/{project_id}/locations/{location}

Optional. The page token, returned by a previous call, to request the next page of results.

ListWorkflowTemplatesResponse

A response to a request to list workflow templates in a project. .. attribute:: templates

Output only. WorkflowTemplates list.

LoggingConfig

The runtime logging config of the job. .. attribute:: driver_log_levels

The per-package log levels for the driver. This may include “root” package name to configure rootLogger. Examples: ‘com.google = FATAL’, ‘root = INFO’, ‘org.apache = DEBUG’

ManagedCluster

Cluster that is managed by the workflow. .. attribute:: cluster_name

Required. The cluster name prefix. A unique cluster name will be formed by appending a random suffix. The name must contain only lower-case letters (a-z), numbers (0-9), and hyphens (-). Must begin with a letter. Cannot begin or end with hyphen. Must consist of between 2 and 35 characters.

Optional. The labels to associate with this cluster. Label keys must be between 1 and 63 characters long. Label values must be between 1 and 63 characters long. No more than 32 labels can be associated with a given cluster.

ManagedGroupConfig

Specifies the resources used to actively manage an instance group. .. attribute:: instance_template_name

Output only. The name of the Instance Template used for the Managed Instance Group.

NodeInitializationAction

Specifies an executable to run on a fully configured node and a timeout period for executable completion. .. attribute:: executable_file

Required. Cloud Storage URI of executable file.

Operation

API documentation for dataproc_v1beta2.types.Operation class.

OperationInfo

API documentation for dataproc_v1beta2.types.OperationInfo class.

OrderedJob

A job executed by the workflow. .. attribute:: step_id

Required. The step id. The id must be unique among all jobs within the template. The step id is used as prefix for job id, as job goog-dataproc-workflow-step-id label, and in [p rerequisiteStepIds][google.cloud.dataproc.v1beta2.OrderedJob.p rerequisite_step_ids] field from other steps. The id must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), and hyphens (-). Cannot begin or end with underscore or hyphen. Must consist of between 3 and 50 characters.

Spark R job

Optional. The labels to associate with this job. Label keys must be between 1 and 63 characters long. Label values must be between 1 and 63 characters long. No more than 32 labels can be associated with a given job.

Optional. The optional list of prerequisite job step_ids. If not specified, the job will start at the beginning of workflow.

ParameterValidation

Configuration for parameter validation. .. attribute:: validation_type

Required. The type of validation to be performed.

Validation based on a list of allowed values.

PigJob

A Dataproc job for running Apache Pig <https://pig.apache.org/>__ queries on YARN. .. attribute:: queries

Required. The sequence of Pig queries to execute, specified as an HCFS file URI or a list of queries.

A list of queries.

Optional. Mapping of query variable names to values (equivalent to the Pig command: name=[value]).

Optional. HCFS URIs of jar files to add to the CLASSPATH of the Pig Client and Hadoop MapReduce (MR) tasks. Can contain Pig UDFs.

PrestoJob

A Dataproc job for running Presto <https://prestosql.io/> queries. IMPORTANT: The Dataproc Presto Optional Component <https://cloud.google.com/dataproc/docs/concepts/components/presto> must be enabled when the cluster is created to submit a Presto job to the cluster. .. attribute:: queries

Required. The sequence of Presto queries to execute, specified as either an HCFS file URI or as a list of queries.

A list of queries.

Optional. The format in which query output will be displayed. See the Presto documentation for supported output formats

Optional. A mapping of property names to values. Used to set Presto session properties <https://prestodb.io/docs/current/sql/set-session.html>__ Equivalent to using the –session flag in the Presto CLI

PySparkJob

A Dataproc job for running Apache PySpark <https://spark.apache.org/docs/0.9.0/python-programming-guide.html>__ applications on YARN. .. attribute:: main_python_file_uri

Required. The HCFS URI of the main Python file to use as the driver. Must be a .py file.

Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.

Optional. HCFS URIs of files to be copied to the working directory of Python drivers and distributed tasks. Useful for naively parallel tasks.

Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.

QueryList

A list of queries to run on a cluster. .. attribute:: queries

Required. The queries to execute. You do not need to terminate a query with a semicolon. Multiple queries can be specified in one string by separating each with a semicolon. Here is an example of an Cloud Dataproc API snippet that uses a QueryList to specify a HiveJob: :: "hiveJob": { "queryList": { "queries": [ "query1", "query2", "query3;query4", ] } }

RegexValidation

Validation based on regular expressions. .. attribute:: regexes

Required. RE2 regular expressions used to validate the parameter’s value. The value must match the regex in its entirety (substring matches are not sufficient).

ReservationAffinity

Reservation Affinity for consuming Zonal reservation. .. attribute:: consume_reservation_type

Optional. Type of reservation to consume

Optional. Corresponds to the label values of reservation resource.

SecurityConfig

Security related configuration, including encryption, Kerberos, etc. .. attribute:: kerberos_config

Kerberos related configuration.

SoftwareConfig

Specifies the selection and config of software inside the cluster. .. attribute:: image_version

Optional. The version of software inside the cluster. It must be one of the supported Dataproc Versions <https://cloud.goog le.com/dataproc/docs/concepts/versioning/dataproc- versions#supported_cloud_dataproc_versions>, such as “1.2” (including a subminor version, such as “1.2.29”), or the “preview” version <https://cloud.google.com/dataproc/docs/con cepts/versioning/dataproc-versions#other_versions>. If unspecified, it defaults to the latest Debian version.

The set of optional components to activate on the cluster.

SparkJob

A Dataproc job for running Apache Spark <http://spark.apache.org/>__ applications on YARN. The specification of the main method to call to drive the job. Specify either the jar file that contains the main class or the main class name. To pass both a main jar and a main class in that jar, add the jar to CommonJob.jar_file_uris, and then specify the main class name in main_class. .. attribute:: main_jar_file_uri

The HCFS URI of the jar file that contains the main class.

Optional. The arguments to pass to the driver. Do not include arguments, such as --conf, that can be set as job properties, since a collision may occur that causes an incorrect job submission.

Optional. HCFS URIs of files to be copied to the working directory of Spark drivers and distributed tasks. Useful for naively parallel tasks.

Optional. A mapping of property names to values, used to configure Spark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.

SparkRJob

A Dataproc job for running Apache SparkR <https://spark.apache.org/docs/latest/sparkr.html>__ applications on YARN. .. attribute:: main_r_file_uri

Required. The HCFS URI of the main R file to use as the driver. Must be a .R file.

Optional. HCFS URIs of files to be copied to the working directory of R drivers and distributed tasks. Useful for naively parallel tasks.

Optional. A mapping of property names to values, used to configure SparkR. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.

SparkSqlJob

A Dataproc job for running Apache Spark SQL <http://spark.apache.org/sql/>__ queries. .. attribute:: queries

Required. The sequence of Spark SQL queries to execute, specified as either an HCFS file URI or as a list of queries.

A list of queries.

Optional. A mapping of property names to values, used to configure Spark SQL’s SparkConf. Properties that conflict with values set by the Dataproc API may be overwritten.

Optional. The runtime log config for job execution.

Status

API documentation for dataproc_v1beta2.types.Status class.

SubmitJobRequest

A request to submit a job. .. attribute:: project_id

Required. The ID of the Google Cloud Platform project that the job belongs to.

Required. The job resource.

TemplateParameter

A configurable parameter that replaces one or more fields in the template. Parameterizable fields: - Labels - File uris - Job properties - Job arguments - Script variables - Main class (in HadoopJob and SparkJob) - Zone (in ClusterSelector) .. attribute:: name

Required. Parameter name. The parameter name is used as the key, and paired with the parameter value, which are passed to the template when the template is instantiated. The name must contain only capital letters (A-Z), numbers (0-9), and underscores (_), and must not start with a number. The maximum length is 40 characters.

Optional. Brief description of the parameter. Must not exceed 1024 characters.

Timestamp

API documentation for dataproc_v1beta2.types.Timestamp class.

UpdateAutoscalingPolicyRequest

A request to update an autoscaling policy. .. attribute:: policy

Required. The updated autoscaling policy.

UpdateClusterRequest

A request to update a cluster. .. attribute:: project_id

Required. The ID of the Google Cloud Platform project the cluster belongs to.

Required. The cluster name.

Optional. Timeout for graceful YARN decomissioning. Graceful decommissioning allows removing nodes from the cluster without interrupting jobs in progress. Timeout specifies how long to wait for jobs in progress to finish before forcefully removing nodes (and potentially interrupting jobs). Default timeout is 0 (for forceful decommission), and the maximum allowed timeout is 1 day (see JSON representation of Duration <https://developers.google.com/protocol- buffers/docs/proto3#json>__). Only supported on Dataproc image versions 1.2 and higher.

Optional. A unique id used to identify the request. If the server receives two [UpdateClusterRequest][google.cloud.datapr oc.v1beta2.UpdateClusterRequest] requests with the same id, then the second request will be ignored and the first google.longrunning.Operation][google.longrunning.Operation] created and stored in the backend is returned. It is recommended to always set this value to a UUID <https://en.wi kipedia.org/wiki/Universally_unique_identifier>_. The id must contain only letters (a-z, A-Z), numbers (0-9), underscores (), and hyphens (-). The maximum length is 40 characters.

UpdateJobRequest

A request to update a job. .. attribute:: project_id

Required. The ID of the Google Cloud Platform project that the job belongs to.

Required. The job ID.

Required. Specifies the path, relative to Job, of the field to update. For example, to update the labels of a Job the update_mask parameter would be specified as labels, and the PATCH request body would specify the new value. Note: Currently, labels is the only field that can be updated.

UpdateWorkflowTemplateRequest

A request to update a workflow template. .. attribute:: template

Required. The updated workflow template. The template.version field must match the current version.

ValueValidation

Validation based on a list of allowed values. .. attribute:: values

Required. List of allowed values for the parameter.

WaitOperationRequest

API documentation for dataproc_v1beta2.types.WaitOperationRequest class.

WorkflowGraph

The workflow graph. .. attribute:: nodes

Output only. The workflow nodes.

WorkflowMetadata

A Dataproc workflow template resource. .. attribute:: template

Output only. The resource name of the workflow template as described in https://cloud.google.com/apis/design/resource_names. - For projects.regions.workflowTemplates, the resource name of the template has the following format: projects/{proje ct_id}/regions/{region}/workflowTemplates/{template_id} - For projects.locations.workflowTemplates, the resource name of the template has the following format: project s/{project_id}/locations/{location}/workflowTemplates/{templat e_id}

Output only. The create cluster operation metadata.

Output only. The delete cluster operation metadata.

Output only. The name of the target cluster.

Output only. Workflow start time.

Output only. The UUID of target cluster.

WorkflowNode

The workflow node. .. attribute:: step_id

Output only. The name of the node.

Output only. The job id; populated after the node enters RUNNING state.

Output only. The error detail.

WorkflowTemplate

A Dataproc workflow template resource. .. attribute:: id

Required. The template id. The id must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), and hyphens (-). Cannot begin or end with underscore or hyphen. Must consist of between 3 and 50 characters. .

Optional. Used to perform a consistent read-modify-write. This field should be left blank for a CreateWorkflowTemplate request. It is required for an UpdateWorkflowTemplate request, and must match the current server version. A typical update template flow would fetch the current template with a GetWorkflowTemplate request, which will return the current template with the version field filled in with the current server version. The user updates other fields in the template, then returns it as part of the UpdateWorkflowTemplate request.

Output only. The time template was last updated.

Required. WorkflowTemplate scheduling information.

Optional. Template parameters whose values are substituted into the template. Values for parameters must be provided when the template is instantiated.

WorkflowTemplatePlacement

Specifies workflow execution target. Either managed_cluster or cluster_selector is required. .. attribute:: placement

Required. Specifies where workflow executes; either on a managed cluster or an existing cluster chosen by labels.

Optional. A selector that chooses target cluster for jobs based on metadata. The selector is evaluated at the time each job is submitted.

YarnApplication

A YARN application created by a job. Application information is a subset of org.apache.hadoop.yarn.proto.YarnProtos.ApplicationReportProto. Beta Feature: This report is available for testing purposes only. It may be changed before final release. .. attribute:: name

Output only. The application name.

Output only. The numerical progress of the application, from 1 to 100.