PySparkJob(mapping=None, *, ignore_unknown_fields=False, **kwargs)
A Dataproc job for running
applications on YARN.
Required. The HCFS URI of the main Python file to use as the driver. Must be a .py file.
Optional. The arguments to pass to the driver. Do not include arguments, such as ``--conf``, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
Optional. The runtime log config for job execution.
Inheritancebuiltins.object > proto.message.Message > PySparkJob
PropertiesEntry(mapping=None, *, ignore_unknown_fields=False, **kwargs)
The abstract base class for a message.
Keys and values corresponding to the fields of the message.
A dictionary or message to be used to determine the values for this message.
If True, do not raise errors for unknown fields. Only applied if