- 4.51.0 (latest)
- 4.50.0
- 4.49.0
- 4.48.0
- 4.46.0
- 4.45.0
- 4.44.0
- 4.43.0
- 4.42.0
- 4.41.0
- 4.40.0
- 4.39.0
- 4.38.0
- 4.37.0
- 4.36.0
- 4.34.0
- 4.33.0
- 4.32.0
- 4.31.0
- 4.30.0
- 4.29.0
- 4.28.0
- 4.27.0
- 4.26.0
- 4.25.0
- 4.24.0
- 4.21.0
- 4.20.0
- 4.19.0
- 4.18.0
- 4.17.0
- 4.16.0
- 4.15.0
- 4.14.0
- 4.13.0
- 4.12.0
- 4.11.0
- 4.10.0
- 4.9.0
- 4.8.0
- 4.6.0
- 4.5.0
- 4.4.0
- 4.3.0
- 4.2.0
- 4.1.0
- 4.0.8
- 3.1.2
- 3.0.3
- 2.3.1
public static final class PySparkJob.Builder extends GeneratedMessageV3.Builder<PySparkJob.Builder> implements PySparkJobOrBuilder
A Dataproc job for running Apache PySpark applications on YARN.
Protobuf type google.cloud.dataproc.v1.PySparkJob
Inheritance
Object > AbstractMessageLite.Builder<MessageType,BuilderType> > AbstractMessage.Builder<BuilderType> > GeneratedMessageV3.Builder > PySparkJob.BuilderImplements
PySparkJobOrBuilderStatic Methods
getDescriptor()
public static final Descriptors.Descriptor getDescriptor()
Type | Description |
Descriptor |
Methods
addAllArchiveUris(Iterable<String> values)
public PySparkJob.Builder addAllArchiveUris(Iterable<String> values)
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
values | Iterable<String> The archiveUris to add. |
Type | Description |
PySparkJob.Builder | This builder for chaining. |
addAllArgs(Iterable<String> values)
public PySparkJob.Builder addAllArgs(Iterable<String> values)
Optional. The arguments to pass to the driver. Do not include arguments,
such as --conf
, that can be set as job properties, since a collision may
occur that causes an incorrect job submission.
repeated string args = 2 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
values | Iterable<String> The args to add. |
Type | Description |
PySparkJob.Builder | This builder for chaining. |
addAllFileUris(Iterable<String> values)
public PySparkJob.Builder addAllFileUris(Iterable<String> values)
Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
repeated string file_uris = 5 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
values | Iterable<String> The fileUris to add. |
Type | Description |
PySparkJob.Builder | This builder for chaining. |
addAllJarFileUris(Iterable<String> values)
public PySparkJob.Builder addAllJarFileUris(Iterable<String> values)
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
values | Iterable<String> The jarFileUris to add. |
Type | Description |
PySparkJob.Builder | This builder for chaining. |
addAllPythonFileUris(Iterable<String> values)
public PySparkJob.Builder addAllPythonFileUris(Iterable<String> values)
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
values | Iterable<String> The pythonFileUris to add. |
Type | Description |
PySparkJob.Builder | This builder for chaining. |
addArchiveUris(String value)
public PySparkJob.Builder addArchiveUris(String value)
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
value | String The archiveUris to add. |
Type | Description |
PySparkJob.Builder | This builder for chaining. |
addArchiveUrisBytes(ByteString value)
public PySparkJob.Builder addArchiveUrisBytes(ByteString value)
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
value | ByteString The bytes of the archiveUris to add. |
Type | Description |
PySparkJob.Builder | This builder for chaining. |
addArgs(String value)
public PySparkJob.Builder addArgs(String value)
Optional. The arguments to pass to the driver. Do not include arguments,
such as --conf
, that can be set as job properties, since a collision may
occur that causes an incorrect job submission.
repeated string args = 2 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
value | String The args to add. |
Type | Description |
PySparkJob.Builder | This builder for chaining. |
addArgsBytes(ByteString value)
public PySparkJob.Builder addArgsBytes(ByteString value)
Optional. The arguments to pass to the driver. Do not include arguments,
such as --conf
, that can be set as job properties, since a collision may
occur that causes an incorrect job submission.
repeated string args = 2 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
value | ByteString The bytes of the args to add. |
Type | Description |
PySparkJob.Builder | This builder for chaining. |
addFileUris(String value)
public PySparkJob.Builder addFileUris(String value)
Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
repeated string file_uris = 5 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
value | String The fileUris to add. |
Type | Description |
PySparkJob.Builder | This builder for chaining. |
addFileUrisBytes(ByteString value)
public PySparkJob.Builder addFileUrisBytes(ByteString value)
Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
repeated string file_uris = 5 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
value | ByteString The bytes of the fileUris to add. |
Type | Description |
PySparkJob.Builder | This builder for chaining. |
addJarFileUris(String value)
public PySparkJob.Builder addJarFileUris(String value)
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
value | String The jarFileUris to add. |
Type | Description |
PySparkJob.Builder | This builder for chaining. |
addJarFileUrisBytes(ByteString value)
public PySparkJob.Builder addJarFileUrisBytes(ByteString value)
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
value | ByteString The bytes of the jarFileUris to add. |
Type | Description |
PySparkJob.Builder | This builder for chaining. |
addPythonFileUris(String value)
public PySparkJob.Builder addPythonFileUris(String value)
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
value | String The pythonFileUris to add. |
Type | Description |
PySparkJob.Builder | This builder for chaining. |
addPythonFileUrisBytes(ByteString value)
public PySparkJob.Builder addPythonFileUrisBytes(ByteString value)
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
value | ByteString The bytes of the pythonFileUris to add. |
Type | Description |
PySparkJob.Builder | This builder for chaining. |
addRepeatedField(Descriptors.FieldDescriptor field, Object value)
public PySparkJob.Builder addRepeatedField(Descriptors.FieldDescriptor field, Object value)
Name | Description |
field | FieldDescriptor |
value | Object |
Type | Description |
PySparkJob.Builder |
build()
public PySparkJob build()
Type | Description |
PySparkJob |
buildPartial()
public PySparkJob buildPartial()
Type | Description |
PySparkJob |
clear()
public PySparkJob.Builder clear()
Type | Description |
PySparkJob.Builder |
clearArchiveUris()
public PySparkJob.Builder clearArchiveUris()
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6 [(.google.api.field_behavior) = OPTIONAL];
Type | Description |
PySparkJob.Builder | This builder for chaining. |
clearArgs()
public PySparkJob.Builder clearArgs()
Optional. The arguments to pass to the driver. Do not include arguments,
such as --conf
, that can be set as job properties, since a collision may
occur that causes an incorrect job submission.
repeated string args = 2 [(.google.api.field_behavior) = OPTIONAL];
Type | Description |
PySparkJob.Builder | This builder for chaining. |
clearField(Descriptors.FieldDescriptor field)
public PySparkJob.Builder clearField(Descriptors.FieldDescriptor field)
Name | Description |
field | FieldDescriptor |
Type | Description |
PySparkJob.Builder |
clearFileUris()
public PySparkJob.Builder clearFileUris()
Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
repeated string file_uris = 5 [(.google.api.field_behavior) = OPTIONAL];
Type | Description |
PySparkJob.Builder | This builder for chaining. |
clearJarFileUris()
public PySparkJob.Builder clearJarFileUris()
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4 [(.google.api.field_behavior) = OPTIONAL];
Type | Description |
PySparkJob.Builder | This builder for chaining. |
clearLoggingConfig()
public PySparkJob.Builder clearLoggingConfig()
Optional. The runtime log config for job execution.
.google.cloud.dataproc.v1.LoggingConfig logging_config = 8 [(.google.api.field_behavior) = OPTIONAL];
Type | Description |
PySparkJob.Builder |
clearMainPythonFileUri()
public PySparkJob.Builder clearMainPythonFileUri()
Required. The HCFS URI of the main Python file to use as the driver. Must be a .py file.
string main_python_file_uri = 1 [(.google.api.field_behavior) = REQUIRED];
Type | Description |
PySparkJob.Builder | This builder for chaining. |
clearOneof(Descriptors.OneofDescriptor oneof)
public PySparkJob.Builder clearOneof(Descriptors.OneofDescriptor oneof)
Name | Description |
oneof | OneofDescriptor |
Type | Description |
PySparkJob.Builder |
clearProperties()
public PySparkJob.Builder clearProperties()
Type | Description |
PySparkJob.Builder |
clearPythonFileUris()
public PySparkJob.Builder clearPythonFileUris()
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3 [(.google.api.field_behavior) = OPTIONAL];
Type | Description |
PySparkJob.Builder | This builder for chaining. |
clone()
public PySparkJob.Builder clone()
Type | Description |
PySparkJob.Builder |
containsProperties(String key)
public boolean containsProperties(String key)
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
key | String |
Type | Description |
boolean |
getArchiveUris(int index)
public String getArchiveUris(int index)
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
index | int The index of the element to return. |
Type | Description |
String | The archiveUris at the given index. |
getArchiveUrisBytes(int index)
public ByteString getArchiveUrisBytes(int index)
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
index | int The index of the value to return. |
Type | Description |
ByteString | The bytes of the archiveUris at the given index. |
getArchiveUrisCount()
public int getArchiveUrisCount()
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6 [(.google.api.field_behavior) = OPTIONAL];
Type | Description |
int | The count of archiveUris. |
getArchiveUrisList()
public ProtocolStringList getArchiveUrisList()
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6 [(.google.api.field_behavior) = OPTIONAL];
Type | Description |
ProtocolStringList | A list containing the archiveUris. |
getArgs(int index)
public String getArgs(int index)
Optional. The arguments to pass to the driver. Do not include arguments,
such as --conf
, that can be set as job properties, since a collision may
occur that causes an incorrect job submission.
repeated string args = 2 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
index | int The index of the element to return. |
Type | Description |
String | The args at the given index. |
getArgsBytes(int index)
public ByteString getArgsBytes(int index)
Optional. The arguments to pass to the driver. Do not include arguments,
such as --conf
, that can be set as job properties, since a collision may
occur that causes an incorrect job submission.
repeated string args = 2 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
index | int The index of the value to return. |
Type | Description |
ByteString | The bytes of the args at the given index. |
getArgsCount()
public int getArgsCount()
Optional. The arguments to pass to the driver. Do not include arguments,
such as --conf
, that can be set as job properties, since a collision may
occur that causes an incorrect job submission.
repeated string args = 2 [(.google.api.field_behavior) = OPTIONAL];
Type | Description |
int | The count of args. |
getArgsList()
public ProtocolStringList getArgsList()
Optional. The arguments to pass to the driver. Do not include arguments,
such as --conf
, that can be set as job properties, since a collision may
occur that causes an incorrect job submission.
repeated string args = 2 [(.google.api.field_behavior) = OPTIONAL];
Type | Description |
ProtocolStringList | A list containing the args. |
getDefaultInstanceForType()
public PySparkJob getDefaultInstanceForType()
Type | Description |
PySparkJob |
getDescriptorForType()
public Descriptors.Descriptor getDescriptorForType()
Type | Description |
Descriptor |
getFileUris(int index)
public String getFileUris(int index)
Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
repeated string file_uris = 5 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
index | int The index of the element to return. |
Type | Description |
String | The fileUris at the given index. |
getFileUrisBytes(int index)
public ByteString getFileUrisBytes(int index)
Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
repeated string file_uris = 5 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
index | int The index of the value to return. |
Type | Description |
ByteString | The bytes of the fileUris at the given index. |
getFileUrisCount()
public int getFileUrisCount()
Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
repeated string file_uris = 5 [(.google.api.field_behavior) = OPTIONAL];
Type | Description |
int | The count of fileUris. |
getFileUrisList()
public ProtocolStringList getFileUrisList()
Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
repeated string file_uris = 5 [(.google.api.field_behavior) = OPTIONAL];
Type | Description |
ProtocolStringList | A list containing the fileUris. |
getJarFileUris(int index)
public String getJarFileUris(int index)
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
index | int The index of the element to return. |
Type | Description |
String | The jarFileUris at the given index. |
getJarFileUrisBytes(int index)
public ByteString getJarFileUrisBytes(int index)
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
index | int The index of the value to return. |
Type | Description |
ByteString | The bytes of the jarFileUris at the given index. |
getJarFileUrisCount()
public int getJarFileUrisCount()
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4 [(.google.api.field_behavior) = OPTIONAL];
Type | Description |
int | The count of jarFileUris. |
getJarFileUrisList()
public ProtocolStringList getJarFileUrisList()
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4 [(.google.api.field_behavior) = OPTIONAL];
Type | Description |
ProtocolStringList | A list containing the jarFileUris. |
getLoggingConfig()
public LoggingConfig getLoggingConfig()
Optional. The runtime log config for job execution.
.google.cloud.dataproc.v1.LoggingConfig logging_config = 8 [(.google.api.field_behavior) = OPTIONAL];
Type | Description |
LoggingConfig | The loggingConfig. |
getLoggingConfigBuilder()
public LoggingConfig.Builder getLoggingConfigBuilder()
Optional. The runtime log config for job execution.
.google.cloud.dataproc.v1.LoggingConfig logging_config = 8 [(.google.api.field_behavior) = OPTIONAL];
Type | Description |
LoggingConfig.Builder |
getLoggingConfigOrBuilder()
public LoggingConfigOrBuilder getLoggingConfigOrBuilder()
Optional. The runtime log config for job execution.
.google.cloud.dataproc.v1.LoggingConfig logging_config = 8 [(.google.api.field_behavior) = OPTIONAL];
Type | Description |
LoggingConfigOrBuilder |
getMainPythonFileUri()
public String getMainPythonFileUri()
Required. The HCFS URI of the main Python file to use as the driver. Must be a .py file.
string main_python_file_uri = 1 [(.google.api.field_behavior) = REQUIRED];
Type | Description |
String | The mainPythonFileUri. |
getMainPythonFileUriBytes()
public ByteString getMainPythonFileUriBytes()
Required. The HCFS URI of the main Python file to use as the driver. Must be a .py file.
string main_python_file_uri = 1 [(.google.api.field_behavior) = REQUIRED];
Type | Description |
ByteString | The bytes for mainPythonFileUri. |
getMutableProperties()
public Map<String,String> getMutableProperties()
Use alternate mutation accessors instead.
Type | Description |
Map<String,String> |
getProperties()
public Map<String,String> getProperties()
Use #getPropertiesMap() instead.
Type | Description |
Map<String,String> |
getPropertiesCount()
public int getPropertiesCount()
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7 [(.google.api.field_behavior) = OPTIONAL];
Type | Description |
int |
getPropertiesMap()
public Map<String,String> getPropertiesMap()
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7 [(.google.api.field_behavior) = OPTIONAL];
Type | Description |
Map<String,String> |
getPropertiesOrDefault(String key, String defaultValue)
public String getPropertiesOrDefault(String key, String defaultValue)
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
key | String |
defaultValue | String |
Type | Description |
String |
getPropertiesOrThrow(String key)
public String getPropertiesOrThrow(String key)
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
key | String |
Type | Description |
String |
getPythonFileUris(int index)
public String getPythonFileUris(int index)
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
index | int The index of the element to return. |
Type | Description |
String | The pythonFileUris at the given index. |
getPythonFileUrisBytes(int index)
public ByteString getPythonFileUrisBytes(int index)
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
index | int The index of the value to return. |
Type | Description |
ByteString | The bytes of the pythonFileUris at the given index. |
getPythonFileUrisCount()
public int getPythonFileUrisCount()
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3 [(.google.api.field_behavior) = OPTIONAL];
Type | Description |
int | The count of pythonFileUris. |
getPythonFileUrisList()
public ProtocolStringList getPythonFileUrisList()
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3 [(.google.api.field_behavior) = OPTIONAL];
Type | Description |
ProtocolStringList | A list containing the pythonFileUris. |
hasLoggingConfig()
public boolean hasLoggingConfig()
Optional. The runtime log config for job execution.
.google.cloud.dataproc.v1.LoggingConfig logging_config = 8 [(.google.api.field_behavior) = OPTIONAL];
Type | Description |
boolean | Whether the loggingConfig field is set. |
internalGetFieldAccessorTable()
protected GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable()
Type | Description |
FieldAccessorTable |
internalGetMapField(int number)
protected MapField internalGetMapField(int number)
Name | Description |
number | int |
Type | Description |
MapField |
internalGetMutableMapField(int number)
protected MapField internalGetMutableMapField(int number)
Name | Description |
number | int |
Type | Description |
MapField |
isInitialized()
public final boolean isInitialized()
Type | Description |
boolean |
mergeFrom(PySparkJob other)
public PySparkJob.Builder mergeFrom(PySparkJob other)
Name | Description |
other | PySparkJob |
Type | Description |
PySparkJob.Builder |
mergeFrom(CodedInputStream input, ExtensionRegistryLite extensionRegistry)
public PySparkJob.Builder mergeFrom(CodedInputStream input, ExtensionRegistryLite extensionRegistry)
Name | Description |
input | CodedInputStream |
extensionRegistry | ExtensionRegistryLite |
Type | Description |
PySparkJob.Builder |
Type | Description |
IOException |
mergeFrom(Message other)
public PySparkJob.Builder mergeFrom(Message other)
Name | Description |
other | Message |
Type | Description |
PySparkJob.Builder |
mergeLoggingConfig(LoggingConfig value)
public PySparkJob.Builder mergeLoggingConfig(LoggingConfig value)
Optional. The runtime log config for job execution.
.google.cloud.dataproc.v1.LoggingConfig logging_config = 8 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
value | LoggingConfig |
Type | Description |
PySparkJob.Builder |
mergeUnknownFields(UnknownFieldSet unknownFields)
public final PySparkJob.Builder mergeUnknownFields(UnknownFieldSet unknownFields)
Name | Description |
unknownFields | UnknownFieldSet |
Type | Description |
PySparkJob.Builder |
putAllProperties(Map<String,String> values)
public PySparkJob.Builder putAllProperties(Map<String,String> values)
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
values | Map<String,String> |
Type | Description |
PySparkJob.Builder |
putProperties(String key, String value)
public PySparkJob.Builder putProperties(String key, String value)
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
key | String |
value | String |
Type | Description |
PySparkJob.Builder |
removeProperties(String key)
public PySparkJob.Builder removeProperties(String key)
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
key | String |
Type | Description |
PySparkJob.Builder |
setArchiveUris(int index, String value)
public PySparkJob.Builder setArchiveUris(int index, String value)
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
index | int The index to set the value at. |
value | String The archiveUris to set. |
Type | Description |
PySparkJob.Builder | This builder for chaining. |
setArgs(int index, String value)
public PySparkJob.Builder setArgs(int index, String value)
Optional. The arguments to pass to the driver. Do not include arguments,
such as --conf
, that can be set as job properties, since a collision may
occur that causes an incorrect job submission.
repeated string args = 2 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
index | int The index to set the value at. |
value | String The args to set. |
Type | Description |
PySparkJob.Builder | This builder for chaining. |
setField(Descriptors.FieldDescriptor field, Object value)
public PySparkJob.Builder setField(Descriptors.FieldDescriptor field, Object value)
Name | Description |
field | FieldDescriptor |
value | Object |
Type | Description |
PySparkJob.Builder |
setFileUris(int index, String value)
public PySparkJob.Builder setFileUris(int index, String value)
Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
repeated string file_uris = 5 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
index | int The index to set the value at. |
value | String The fileUris to set. |
Type | Description |
PySparkJob.Builder | This builder for chaining. |
setJarFileUris(int index, String value)
public PySparkJob.Builder setJarFileUris(int index, String value)
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
index | int The index to set the value at. |
value | String The jarFileUris to set. |
Type | Description |
PySparkJob.Builder | This builder for chaining. |
setLoggingConfig(LoggingConfig value)
public PySparkJob.Builder setLoggingConfig(LoggingConfig value)
Optional. The runtime log config for job execution.
.google.cloud.dataproc.v1.LoggingConfig logging_config = 8 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
value | LoggingConfig |
Type | Description |
PySparkJob.Builder |
setLoggingConfig(LoggingConfig.Builder builderForValue)
public PySparkJob.Builder setLoggingConfig(LoggingConfig.Builder builderForValue)
Optional. The runtime log config for job execution.
.google.cloud.dataproc.v1.LoggingConfig logging_config = 8 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
builderForValue | LoggingConfig.Builder |
Type | Description |
PySparkJob.Builder |
setMainPythonFileUri(String value)
public PySparkJob.Builder setMainPythonFileUri(String value)
Required. The HCFS URI of the main Python file to use as the driver. Must be a .py file.
string main_python_file_uri = 1 [(.google.api.field_behavior) = REQUIRED];
Name | Description |
value | String The mainPythonFileUri to set. |
Type | Description |
PySparkJob.Builder | This builder for chaining. |
setMainPythonFileUriBytes(ByteString value)
public PySparkJob.Builder setMainPythonFileUriBytes(ByteString value)
Required. The HCFS URI of the main Python file to use as the driver. Must be a .py file.
string main_python_file_uri = 1 [(.google.api.field_behavior) = REQUIRED];
Name | Description |
value | ByteString The bytes for mainPythonFileUri to set. |
Type | Description |
PySparkJob.Builder | This builder for chaining. |
setPythonFileUris(int index, String value)
public PySparkJob.Builder setPythonFileUris(int index, String value)
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3 [(.google.api.field_behavior) = OPTIONAL];
Name | Description |
index | int The index to set the value at. |
value | String The pythonFileUris to set. |
Type | Description |
PySparkJob.Builder | This builder for chaining. |
setRepeatedField(Descriptors.FieldDescriptor field, int index, Object value)
public PySparkJob.Builder setRepeatedField(Descriptors.FieldDescriptor field, int index, Object value)
Name | Description |
field | FieldDescriptor |
index | int |
value | Object |
Type | Description |
PySparkJob.Builder |
setUnknownFields(UnknownFieldSet unknownFields)
public final PySparkJob.Builder setUnknownFields(UnknownFieldSet unknownFields)
Name | Description |
unknownFields | UnknownFieldSet |
Type | Description |
PySparkJob.Builder |