Index
BatchController
(interface)SessionController
(interface)SessionTemplateController
(interface)AnalyzeOperationMetadata
(message)AnalyzeOperationMetadata.WorkloadType
(enum)AutotuningConfig
(message)AutotuningConfig.Scenario
(enum)Batch
(message)Batch.State
(enum)Batch.StateHistory
(message)BatchOperationMetadata
(message)BatchOperationMetadata.BatchOperationType
(enum)CreateBatchRequest
(message)CreateSessionRequest
(message)CreateSessionTemplateRequest
(message)DeleteBatchRequest
(message)DeleteSessionRequest
(message)DeleteSessionTemplateRequest
(message)DiagnoseClusterResults
(message)EnvironmentConfig
(message)ExecutionConfig
(message)GetBatchRequest
(message)GetSessionRequest
(message)GetSessionTemplateRequest
(message)JupyterConfig
(message)JupyterConfig.Kernel
(enum)ListBatchesRequest
(message)ListBatchesResponse
(message)ListSessionTemplatesRequest
(message)ListSessionTemplatesResponse
(message)ListSessionsRequest
(message)ListSessionsResponse
(message)PeripheralsConfig
(message)PyPiRepositoryConfig
(message)PySparkBatch
(message)RepositoryConfig
(message)RuntimeConfig
(message)RuntimeInfo
(message)Session
(message)Session.SessionStateHistory
(message)Session.State
(enum)SessionOperationMetadata
(message)SessionOperationMetadata.SessionOperationType
(enum)SessionTemplate
(message)SparkBatch
(message)SparkConnectConfig
(message)SparkHistoryServerConfig
(message)SparkRBatch
(message)SparkSqlBatch
(message)TerminateSessionRequest
(message)UpdateSessionTemplateRequest
(message)UsageMetrics
(message)UsageSnapshot
(message)
BatchController
The BatchController provides methods to manage batch workloads.
CreateBatch |
---|
Creates a batch workload that executes asynchronously.
|
DeleteBatch |
---|
Deletes the batch workload resource. If the batch is not in a
|
GetBatch |
---|
Gets the batch workload resource representation.
|
ListBatches |
---|
Lists batch workloads.
|
SessionController
The SessionController
provides methods to manage interactive sessions.
CreateSession |
---|
Create an interactive session asynchronously.
|
DeleteSession |
---|
Deletes the interactive session resource. If the session is not in terminal state, it is terminated, and then deleted.
|
GetSession |
---|
Gets the resource representation for an interactive session.
|
ListSessions |
---|
Lists interactive sessions.
|
TerminateSession |
---|
Terminates the interactive session.
|
SessionTemplateController
The SessionTemplateController provides methods to manage session templates.
CreateSessionTemplate |
---|
Create a session template synchronously.
|
DeleteSessionTemplate |
---|
Deletes a session template.
|
GetSessionTemplate |
---|
Gets the resource representation for a session template.
|
ListSessionTemplates |
---|
Lists session templates.
|
UpdateSessionTemplate |
---|
Updates the session template synchronously.
|
AnalyzeOperationMetadata
Metadata describing the Analyze operation.
Fields | |
---|---|
analyzed_ |
Output only. name of the workload being analyzed. |
analyzed_ |
Output only. Type of the workload being analyzed. |
analyzed_ |
Output only. unique identifier of the workload typically generated by control plane. E.g. batch uuid. |
create_ |
Output only. The time when the operation was created. |
done_ |
Output only. The time when the operation finished. |
description |
Output only. Short description of the operation. |
labels |
Output only. Labels associated with the operation. |
warnings[] |
Output only. Warnings encountered during operation execution. |
WorkloadType
Workload type
Enums | |
---|---|
WORKLOAD_TYPE_UNSPECIFIED |
Undefined option |
BATCH |
Serverless batch job |
AutotuningConfig
Autotuning configuration of the workload.
Fields | |
---|---|
scenarios[] |
Optional. Scenarios for which tunings are applied. |
Scenario
Scenario represents a specific goal that autotuning will attempt to achieve by modifying workloads.
Enums | |
---|---|
SCENARIO_UNSPECIFIED |
Default value. |
SCALING |
Scaling recommendations such as initialExecutors. |
BROADCAST_HASH_JOIN |
Adding hints for potential relation broadcasts. |
MEMORY |
Memory management for workloads. |
Batch
A representation of a batch workload in the service.
Fields | |
---|---|
name |
Output only. The resource name of the batch. |
uuid |
Output only. A batch UUID (Unique Universal Identifier). The service generates this value when it creates the batch. |
create_ |
Output only. The time when the batch was created. |
runtime_ |
Output only. Runtime information about batch execution. |
state |
Output only. The state of the batch. |
state_ |
Output only. Batch state details, such as a failure description if the state is |
state_ |
Output only. The time when the batch entered a current state. |
creator |
Output only. The email address of the user who created the batch. |
labels |
Optional. The labels to associate with this batch. Label keys must contain 1 to 63 characters, and must conform to RFC 1035. Label values may be empty, but, if present, must contain 1 to 63 characters, and must conform to RFC 1035. No more than 32 labels can be associated with a batch. |
runtime_ |
Optional. Runtime configuration for the batch execution. |
environment_ |
Optional. Environment configuration for the batch execution. |
operation |
Output only. The resource name of the operation associated with this batch. |
state_ |
Output only. Historical state information for the batch. |
Union field batch_config . The application/framework-specific portion of the batch configuration. batch_config can be only one of the following: |
|
pyspark_ |
Optional. PySpark batch config. |
spark_ |
Optional. Spark batch config. |
spark_ |
Optional. SparkR batch config. |
spark_ |
Optional. SparkSql batch config. |
State
The batch state.
Enums | |
---|---|
STATE_UNSPECIFIED |
The batch state is unknown. |
PENDING |
The batch is created before running. |
RUNNING |
The batch is running. |
CANCELLING |
The batch is cancelling. |
CANCELLED |
The batch cancellation was successful. |
SUCCEEDED |
The batch completed successfully. |
FAILED |
The batch is no longer running due to an error. |
StateHistory
Historical state information.
Fields | |
---|---|
state |
Output only. The state of the batch at this point in history. |
state_ |
Output only. Details about the state at this point in history. |
state_ |
Output only. The time when the batch entered the historical state. |
BatchOperationMetadata
Metadata describing the Batch operation.
Fields | |
---|---|
batch |
Name of the batch for the operation. |
batch_ |
Batch UUID for the operation. |
create_ |
The time when the operation was created. |
done_ |
The time when the operation finished. |
operation_ |
The operation type. |
description |
Short description of the operation. |
labels |
Labels associated with the operation. |
warnings[] |
Warnings encountered during operation execution. |
BatchOperationType
Operation type for Batch resources
Enums | |
---|---|
BATCH_OPERATION_TYPE_UNSPECIFIED |
Batch operation type is unknown. |
BATCH |
Batch operation type. |
CreateBatchRequest
A request to create a batch workload.
Fields | |
---|---|
parent |
Required. The parent resource where this batch will be created. Authorization requires the following IAM permission on the specified resource
|
batch |
Required. The batch to create. |
batch_ |
Optional. The ID to use for the batch, which will become the final component of the batch's resource name. This value must be 4-63 characters. Valid characters are |
request_ |
Optional. A unique ID used to identify the request. If the service receives two CreateBatchRequests with the same request_id, the second request is ignored and the Operation that corresponds to the first Batch created and stored in the backend is returned. Recommendation: Set this value to a UUID. The value must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), and hyphens (-). The maximum length is 40 characters. |
CreateSessionRequest
A request to create a session.
Fields | |
---|---|
parent |
Required. The parent resource where this session will be created. Authorization requires the following IAM permission on the specified resource
|
session |
Required. The interactive session to create. |
session_ |
Required. The ID to use for the session, which becomes the final component of the session's resource name. This value must be 4-63 characters. Valid characters are /[a-z][0-9]-/. |
request_ |
Optional. A unique ID used to identify the request. If the service receives two CreateSessionRequestss with the same ID, the second request is ignored, and the first Recommendation: Set this value to a UUID. The value must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), and hyphens (-). The maximum length is 40 characters. |
CreateSessionTemplateRequest
A request to create a session template.
Fields | |
---|---|
parent |
Required. The parent resource where this session template will be created. Authorization requires the following IAM permission on the specified resource
|
session_ |
Required. The session template to create. |
DeleteBatchRequest
A request to delete a batch workload.
Fields | |
---|---|
name |
Required. The fully qualified name of the batch to retrieve in the format "projects/PROJECT_ID/locations/DATAPROC_REGION/batches/BATCH_ID" Authorization requires the following IAM permission on the specified resource
|
DeleteSessionRequest
A request to delete a session.
Fields | |
---|---|
name |
Required. The name of the session resource to delete. Authorization requires the following IAM permission on the specified resource
|
request_ |
Optional. A unique ID used to identify the request. If the service receives two DeleteSessionRequests with the same ID, the second request is ignored. Recommendation: Set this value to a UUID. The value must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), and hyphens (-). The maximum length is 40 characters. |
DeleteSessionTemplateRequest
A request to delete a session template.
Fields | |
---|---|
name |
Required. The name of the session template resource to delete. Authorization requires the following IAM permission on the specified resource
|
DiagnoseClusterResults
The location of diagnostic output.
Fields | |
---|---|
output_ |
Output only. The Cloud Storage URI of the diagnostic output. The output report is a plain text file with a summary of collected diagnostics. |
EnvironmentConfig
Environment configuration for a workload.
Fields | |
---|---|
execution_ |
Optional. Execution configuration for a workload. |
peripherals_ |
Optional. Peripherals configuration that workload has access to. |
ExecutionConfig
Execution configuration for a workload.
Fields | |
---|---|
service_ |
Optional. Service account that used to execute workload. |
network_ |
Optional. Tags used for network traffic control. |
kms_ |
Optional. The Cloud KMS key to use for encryption. |
idle_ |
Optional. Applies to sessions only. The duration to keep the session alive while it's idling. Exceeding this threshold causes the session to terminate. This field cannot be set on a batch workload. Minimum value is 10 minutes; maximum value is 14 days (see JSON representation of Duration). Defaults to 1 hour if not set. If both |
ttl |
Optional. The duration after which the workload will be terminated, specified as the JSON representation for Duration. When the workload exceeds this duration, it will be unconditionally terminated without waiting for ongoing work to finish. If |
staging_ |
Optional. A Cloud Storage bucket used to stage workload dependencies, config files, and store workload output and other ephemeral data, such as Spark history files. If you do not specify a staging bucket, Cloud Dataproc will determine a Cloud Storage location according to the region where your workload is running, and then create and manage project-level, per-location staging and temporary buckets. This field requires a Cloud Storage bucket name, not a |
Union field network . Network configuration for workload execution. network can be only one of the following: |
|
network_ |
Optional. Network URI to connect workload to. |
subnetwork_ |
Optional. Subnetwork URI to connect workload to. |
GetBatchRequest
A request to get the resource representation for a batch workload.
Fields | |
---|---|
name |
Required. The fully qualified name of the batch to retrieve in the format "projects/PROJECT_ID/locations/DATAPROC_REGION/batches/BATCH_ID" Authorization requires the following IAM permission on the specified resource
|
GetSessionRequest
A request to get the resource representation for a session.
Fields | |
---|---|
name |
Required. The name of the session to retrieve. Authorization requires the following IAM permission on the specified resource
|
GetSessionTemplateRequest
A request to get the resource representation for a session template.
Fields | |
---|---|
name |
Required. The name of the session template to retrieve. Authorization requires the following IAM permission on the specified resource
|
JupyterConfig
Jupyter configuration for an interactive session.
Fields | |
---|---|
kernel |
Optional. Kernel |
display_ |
Optional. Display name, shown in the Jupyter kernelspec card. |
Kernel
Jupyter kernel types.
Enums | |
---|---|
KERNEL_UNSPECIFIED |
The kernel is unknown. |
PYTHON |
Python kernel. |
SCALA |
Scala kernel. |
ListBatchesRequest
A request to list batch workloads in a project.
Fields | |
---|---|
parent |
Required. The parent, which owns this collection of batches. Authorization requires the following IAM permission on the specified resource
|
page_ |
Optional. The maximum number of batches to return in each response. The service may return fewer than this value. The default page size is 20; the maximum page size is 1000. |
page_ |
Optional. A page token received from a previous |
filter |
Optional. A filter for the batches to return in the response. A filter is a logical expression constraining the values of various fields in each batch resource. Filters are case sensitive, and may contain multiple clauses combined with logical operators (AND/OR). Supported fields are e.g. See https://google.aip.dev/assets/misc/ebnf-filtering.txt for a detailed description of the filter syntax and a list of supported comparisons. |
order_ |
Optional. Field(s) on which to sort the list of batches. Currently the only supported sort orders are unspecified (empty) and See https://google.aip.dev/132#ordering for more details. |
ListBatchesResponse
A list of batch workloads.
Fields | |
---|---|
batches[] |
Output only. The batches from the specified collection. |
next_ |
A token, which can be sent as |
unreachable[] |
Output only. List of Batches that could not be included in the response. Attempting to get one of these resources may indicate why it was not included in the list response. |
ListSessionTemplatesRequest
A request to list session templates in a project.
Fields | |
---|---|
parent |
Required. The parent that owns this collection of session templates. Authorization requires the following IAM permission on the specified resource
|
page_ |
Optional. The maximum number of sessions to return in each response. The service may return fewer than this value. |
page_ |
Optional. A page token received from a previous |
filter |
Optional. A filter for the session templates to return in the response. Filters are case sensitive and have the following syntax: [field = value] AND [field [= value]] ... |
ListSessionTemplatesResponse
A list of session templates.
Fields | |
---|---|
session_ |
Output only. Session template list |
next_ |
A token, which can be sent as |
ListSessionsRequest
A request to list sessions in a project.
Fields | |
---|---|
parent |
Required. The parent, which owns this collection of sessions. Authorization requires the following IAM permission on the specified resource
|
page_ |
Optional. The maximum number of sessions to return in each response. The service may return fewer than this value. |
page_ |
Optional. A page token received from a previous |
filter |
Optional. A filter for the sessions to return in the response. A filter is a logical expression constraining the values of various fields in each session resource. Filters are case sensitive, and may contain multiple clauses combined with logical operators (AND, OR). Supported fields are Example: See https://google.aip.dev/assets/misc/ebnf-filtering.txt for a detailed description of the filter syntax and a list of supported comparators. |
ListSessionsResponse
A list of interactive sessions.
Fields | |
---|---|
sessions[] |
Output only. The sessions from the specified collection. |
next_ |
A token, which can be sent as |
PeripheralsConfig
Auxiliary services configuration for a workload.
Fields | |
---|---|
metastore_ |
Optional. Resource name of an existing Dataproc Metastore service. Example:
|
spark_ |
Optional. The Spark History Server configuration for the workload. |
PyPiRepositoryConfig
Configuration for PyPi repository
Fields | |
---|---|
pypi_ |
Optional. PyPi repository address |
PySparkBatch
A configuration for running an Apache PySpark batch workload.
Fields | |
---|---|
main_ |
Required. The HCFS URI of the main Python file to use as the Spark driver. Must be a .py file. |
args[] |
Optional. The arguments to pass to the driver. Do not include arguments that can be set as batch properties, such as |
python_ |
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: |
jar_ |
Optional. HCFS URIs of jar files to add to the classpath of the Spark driver and tasks. |
file_ |
Optional. HCFS URIs of files to be placed in the working directory of each executor. |
archive_ |
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: |
RepositoryConfig
Configuration for dependency repositories
Fields | |
---|---|
pypi_ |
Optional. Configuration for PyPi repository. |
RuntimeConfig
Runtime configuration for a workload.
Fields | |
---|---|
version |
Optional. Version of the batch runtime. |
container_ |
Optional. Optional custom container image for the job runtime environment. If not specified, a default container image will be used. |
properties |
Optional. A mapping of property names to values, which are used to configure workload execution. |
repository_ |
Optional. Dependency repository configuration. |
autotuning_ |
Optional. Autotuning configuration of the workload. |
cohort |
Optional. Cohort identifier. Identifies families of the workloads having the same shape, e.g. daily ETL jobs. |
RuntimeInfo
Runtime information about workload execution.
Fields | |
---|---|
endpoints |
Output only. Map of remote access endpoints (such as web interfaces and APIs) to their URIs. |
output_ |
Output only. A URI pointing to the location of the stdout and stderr of the workload. |
diagnostic_ |
Output only. A URI pointing to the location of the diagnostics tarball. |
approximate_ |
Output only. Approximate workload resource usage, calculated when the workload completes (see Dataproc Serverless pricing). Note: This metric calculation may change in the future, for example, to capture cumulative workload resource consumption during workload execution (see the Dataproc Serverless release notes for announcements, changes, fixes and other Dataproc developments). |
current_ |
Output only. Snapshot of current workload resource usage. |
Session
A representation of a session.
Fields | |
---|---|
name |
Required. The resource name of the session. |
uuid |
Output only. A session UUID (Unique Universal Identifier). The service generates this value when it creates the session. |
create_ |
Output only. The time when the session was created. |
runtime_ |
Output only. Runtime information about session execution. |
state |
Output only. A state of the session. |
state_ |
Output only. Session state details, such as the failure description if the state is |
state_ |
Output only. The time when the session entered the current state. |
creator |
Output only. The email address of the user who created the session. |
labels |
Optional. The labels to associate with the session. Label keys must contain 1 to 63 characters, and must conform to RFC 1035. Label values may be empty, but, if present, must contain 1 to 63 characters, and must conform to RFC 1035. No more than 32 labels can be associated with a session. |
runtime_ |
Optional. Runtime configuration for the session execution. |
environment_ |
Optional. Environment configuration for the session execution. |
user |
Optional. The email address of the user who owns the session. |
state_ |
Output only. Historical state information for the session. |
session_ |
Optional. The session template used by the session. Only resource names, including project ID and location, are valid. Example: * The template must be in the same project and Dataproc region as the session. |
Union field session_config . The session configuration. session_config can be only one of the following: |
|
jupyter_ |
Optional. Jupyter session config. |
spark_ |
Optional. Spark connect session config. |
SessionStateHistory
Historical state information.
Fields | |
---|---|
state |
Output only. The state of the session at this point in the session history. |
state_ |
Output only. Details about the state at this point in the session history. |
state_ |
Output only. The time when the session entered the historical state. |
State
The session state.
Enums | |
---|---|
STATE_UNSPECIFIED |
The session state is unknown. |
CREATING |
The session is created prior to running. |
ACTIVE |
The session is running. |
TERMINATING |
The session is terminating. |
TERMINATED |
The session is terminated successfully. |
FAILED |
The session is no longer running due to an error. |
SessionOperationMetadata
Metadata describing the Session operation.
Fields | |
---|---|
session |
Name of the session for the operation. |
session_ |
Session UUID for the operation. |
create_ |
The time when the operation was created. |
done_ |
The time when the operation was finished. |
operation_ |
The operation type. |
description |
Short description of the operation. |
labels |
Labels associated with the operation. |
warnings[] |
Warnings encountered during operation execution. |
SessionOperationType
Operation type for Session resources
Enums | |
---|---|
SESSION_OPERATION_TYPE_UNSPECIFIED |
Session operation type is unknown. |
CREATE |
Create Session operation type. |
TERMINATE |
Terminate Session operation type. |
DELETE |
Delete Session operation type. |
SessionTemplate
A representation of a session template.
Fields | |
---|---|
name |
Required. The resource name of the session template. |
description |
Optional. Brief description of the template. |
create_ |
Output only. The time when the template was created. |
creator |
Output only. The email address of the user who created the template. |
labels |
Optional. Labels to associate with sessions created using this template. Label keys must contain 1 to 63 characters, and must conform to RFC 1035. Label values can be empty, but, if present, must contain 1 to 63 characters and conform to RFC 1035. No more than 32 labels can be associated with a session. |
runtime_ |
Optional. Runtime configuration for session execution. |
environment_ |
Optional. Environment configuration for session execution. |
update_ |
Output only. The time the template was last updated. |
uuid |
Output only. A session template UUID (Unique Universal Identifier). The service generates this value when it creates the session template. |
Union field session_config . The session configuration. session_config can be only one of the following: |
|
jupyter_ |
Optional. Jupyter session config. |
spark_ |
Optional. Spark connect session config. |
SparkBatch
A configuration for running an Apache Spark batch workload.
Fields | |
---|---|
args[] |
Optional. The arguments to pass to the driver. Do not include arguments that can be set as batch properties, such as |
jar_ |
Optional. HCFS URIs of jar files to add to the classpath of the Spark driver and tasks. |
file_ |
Optional. HCFS URIs of files to be placed in the working directory of each executor. |
archive_ |
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: |
Union field driver . The specification of the main method to call to drive the Spark workload. Specify either the jar file that contains the main class or the main class name. To pass both a main jar and a main class in that jar, add the jar to jar_file_uris , and then specify the main class name in main_class . driver can be only one of the following: |
|
main_ |
Optional. The HCFS URI of the jar file that contains the main class. |
main_ |
Optional. The name of the driver main class. The jar file that contains the class must be in the classpath or specified in |
SparkConnectConfig
This type has no fields.
Spark connect configuration for an interactive session.
SparkHistoryServerConfig
Spark History Server configuration for the workload.
Fields | |
---|---|
dataproc_ |
Optional. Resource name of an existing Dataproc Cluster to act as a Spark History Server for the workload. Example:
|
SparkRBatch
A configuration for running an Apache SparkR batch workload.
Fields | |
---|---|
main_ |
Required. The HCFS URI of the main R file to use as the driver. Must be a |
args[] |
Optional. The arguments to pass to the Spark driver. Do not include arguments that can be set as batch properties, such as |
file_ |
Optional. HCFS URIs of files to be placed in the working directory of each executor. |
archive_ |
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: |
SparkSqlBatch
A configuration for running Apache Spark SQL queries as a batch workload.
Fields | |
---|---|
query_ |
Required. The HCFS URI of the script that contains Spark SQL queries to execute. |
query_ |
Optional. Mapping of query variable names to values (equivalent to the Spark SQL command: |
jar_ |
Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH. |
TerminateSessionRequest
A request to terminate an interactive session.
Fields | |
---|---|
name |
Required. The name of the session resource to terminate. Authorization requires the following IAM permission on the specified resource
|
request_ |
Optional. A unique ID used to identify the request. If the service receives two TerminateSessionRequests with the same ID, the second request is ignored. Recommendation: Set this value to a UUID. The value must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), and hyphens (-). The maximum length is 40 characters. |
UpdateSessionTemplateRequest
A request to update a session template.
Fields | |
---|---|
session_ |
Required. The updated session template. Authorization requires the following IAM permission on the specified resource
|
UsageMetrics
Usage metrics represent approximate total resources consumed by a workload.
Fields | |
---|---|
milli_ |
Optional. DCU (Dataproc Compute Units) usage in ( |
shuffle_ |
Optional. Shuffle storage usage in ( |
milli_ |
Optional. Accelerator usage in ( |
accelerator_ |
Optional. Accelerator type being used, if any |
UsageSnapshot
The usage snapshot represents the resources consumed by a workload at a specified time.
Fields | |
---|---|
milli_ |
Optional. Milli (one-thousandth) Dataproc Compute Units (DCUs) (see Dataproc Serverless pricing). |
shuffle_ |
Optional. Shuffle Storage in gigabytes (GB). (see Dataproc Serverless pricing) |
milli_ |
Optional. Milli (one-thousandth) Dataproc Compute Units (DCUs) charged at premium tier (see Dataproc Serverless pricing). |
shuffle_ |
Optional. Shuffle Storage in gigabytes (GB) charged at premium tier. (see Dataproc Serverless pricing) |
milli_ |
Optional. Milli (one-thousandth) accelerator. (see Dataproc Serverless pricing) |
accelerator_ |
Optional. Accelerator type being used, if any |
snapshot_ |
Optional. The timestamp of the usage snapshot. |