Reference documentation and code samples for the Cloud Dataproc V1beta2 API class Google::Cloud::Dataproc::V1beta2::SparkJob.
A Dataproc job for running Apache Spark
applications on YARN.
The specification of the main method to call to drive the job.
Specify either the jar file that contains the main class or the main class
name. To pass both a main jar and a main class in that jar, add the jar to
CommonJob.jar_file_uris
, and then specify the main class name in
main_class
.
Inherits
- Object
Extended By
- Google::Protobuf::MessageExts::ClassMethods
Includes
- Google::Protobuf::MessageExts
Methods
#archive_uris
def archive_uris() -> ::Array<::String>
- (::Array<::String>) — Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
#archive_uris=
def archive_uris=(value) -> ::Array<::String>
- value (::Array<::String>) — Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
- (::Array<::String>) — Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
#args
def args() -> ::Array<::String>
-
(::Array<::String>) — Optional. The arguments to pass to the driver. Do not include arguments,
such as
--conf
, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
#args=
def args=(value) -> ::Array<::String>
-
value (::Array<::String>) — Optional. The arguments to pass to the driver. Do not include arguments,
such as
--conf
, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
-
(::Array<::String>) — Optional. The arguments to pass to the driver. Do not include arguments,
such as
--conf
, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
#file_uris
def file_uris() -> ::Array<::String>
- (::Array<::String>) — Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
#file_uris=
def file_uris=(value) -> ::Array<::String>
- value (::Array<::String>) — Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
- (::Array<::String>) — Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
#jar_file_uris
def jar_file_uris() -> ::Array<::String>
- (::Array<::String>) — Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Spark driver and tasks.
#jar_file_uris=
def jar_file_uris=(value) -> ::Array<::String>
- value (::Array<::String>) — Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Spark driver and tasks.
- (::Array<::String>) — Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Spark driver and tasks.
#logging_config
def logging_config() -> ::Google::Cloud::Dataproc::V1beta2::LoggingConfig
- (::Google::Cloud::Dataproc::V1beta2::LoggingConfig) — Optional. The runtime log config for job execution.
#logging_config=
def logging_config=(value) -> ::Google::Cloud::Dataproc::V1beta2::LoggingConfig
- value (::Google::Cloud::Dataproc::V1beta2::LoggingConfig) — Optional. The runtime log config for job execution.
- (::Google::Cloud::Dataproc::V1beta2::LoggingConfig) — Optional. The runtime log config for job execution.
#main_class
def main_class() -> ::String
-
(::String) — The name of the driver's main class. The jar file that contains the class
must be in the default CLASSPATH or specified in
jar_file_uris
.
#main_class=
def main_class=(value) -> ::String
-
value (::String) — The name of the driver's main class. The jar file that contains the class
must be in the default CLASSPATH or specified in
jar_file_uris
.
-
(::String) — The name of the driver's main class. The jar file that contains the class
must be in the default CLASSPATH or specified in
jar_file_uris
.
#main_jar_file_uri
def main_jar_file_uri() -> ::String
- (::String) — The HCFS URI of the jar file that contains the main class.
#main_jar_file_uri=
def main_jar_file_uri=(value) -> ::String
- value (::String) — The HCFS URI of the jar file that contains the main class.
- (::String) — The HCFS URI of the jar file that contains the main class.
#properties
def properties() -> ::Google::Protobuf::Map{::String => ::String}
- (::Google::Protobuf::Map{::String => ::String}) — Optional. A mapping of property names to values, used to configure Spark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
#properties=
def properties=(value) -> ::Google::Protobuf::Map{::String => ::String}
- value (::Google::Protobuf::Map{::String => ::String}) — Optional. A mapping of property names to values, used to configure Spark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
- (::Google::Protobuf::Map{::String => ::String}) — Optional. A mapping of property names to values, used to configure Spark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.