Class PySparkJob (1.1.3)

A Dataproc job for running Apache PySpark <>__ applications on YARN.

Optional. The arguments to pass to the driver. Do not include arguments, such as --conf, that can be set as job properties, since a collision may occur that causes an incorrect job submission.

Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.

Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.

Optional. The runtime log config for job execution.



API documentation for dataproc_v1.types.PySparkJob.PropertiesEntry class.