Package com.google.cloud.dataproc.v1
Class PySparkJob.Builder
- java.lang.Object
-
- com.google.protobuf.AbstractMessageLite.Builder
-
- com.google.protobuf.AbstractMessage.Builder<BuilderT>
-
- com.google.protobuf.GeneratedMessageV3.Builder<PySparkJob.Builder>
-
- com.google.cloud.dataproc.v1.PySparkJob.Builder
-
- All Implemented Interfaces:
PySparkJobOrBuilder,com.google.protobuf.Message.Builder,com.google.protobuf.MessageLite.Builder,com.google.protobuf.MessageLiteOrBuilder,com.google.protobuf.MessageOrBuilder,Cloneable
- Enclosing class:
- PySparkJob
public static final class PySparkJob.Builder extends com.google.protobuf.GeneratedMessageV3.Builder<PySparkJob.Builder> implements PySparkJobOrBuilder
A Dataproc job for running [Apache PySpark](https://spark.apache.org/docs/0.9.0/python-programming-guide.html) applications on YARN.
Protobuf typegoogle.cloud.dataproc.v1.PySparkJob
-
-
Method Summary
All Methods Static Methods Instance Methods Concrete Methods Deprecated Methods Modifier and Type Method Description PySparkJob.BuilderaddAllArchiveUris(Iterable<String> values)Optional.PySparkJob.BuilderaddAllArgs(Iterable<String> values)Optional.PySparkJob.BuilderaddAllFileUris(Iterable<String> values)Optional.PySparkJob.BuilderaddAllJarFileUris(Iterable<String> values)Optional.PySparkJob.BuilderaddAllPythonFileUris(Iterable<String> values)Optional.PySparkJob.BuilderaddArchiveUris(String value)Optional.PySparkJob.BuilderaddArchiveUrisBytes(com.google.protobuf.ByteString value)Optional.PySparkJob.BuilderaddArgs(String value)Optional.PySparkJob.BuilderaddArgsBytes(com.google.protobuf.ByteString value)Optional.PySparkJob.BuilderaddFileUris(String value)Optional.PySparkJob.BuilderaddFileUrisBytes(com.google.protobuf.ByteString value)Optional.PySparkJob.BuilderaddJarFileUris(String value)Optional.PySparkJob.BuilderaddJarFileUrisBytes(com.google.protobuf.ByteString value)Optional.PySparkJob.BuilderaddPythonFileUris(String value)Optional.PySparkJob.BuilderaddPythonFileUrisBytes(com.google.protobuf.ByteString value)Optional.PySparkJob.BuilderaddRepeatedField(com.google.protobuf.Descriptors.FieldDescriptor field, Object value)PySparkJobbuild()PySparkJobbuildPartial()PySparkJob.Builderclear()PySparkJob.BuilderclearArchiveUris()Optional.PySparkJob.BuilderclearArgs()Optional.PySparkJob.BuilderclearField(com.google.protobuf.Descriptors.FieldDescriptor field)PySparkJob.BuilderclearFileUris()Optional.PySparkJob.BuilderclearJarFileUris()Optional.PySparkJob.BuilderclearLoggingConfig()Optional.PySparkJob.BuilderclearMainPythonFileUri()Required.PySparkJob.BuilderclearOneof(com.google.protobuf.Descriptors.OneofDescriptor oneof)PySparkJob.BuilderclearProperties()PySparkJob.BuilderclearPythonFileUris()Optional.PySparkJob.Builderclone()booleancontainsProperties(String key)Optional.StringgetArchiveUris(int index)Optional.com.google.protobuf.ByteStringgetArchiveUrisBytes(int index)Optional.intgetArchiveUrisCount()Optional.com.google.protobuf.ProtocolStringListgetArchiveUrisList()Optional.StringgetArgs(int index)Optional.com.google.protobuf.ByteStringgetArgsBytes(int index)Optional.intgetArgsCount()Optional.com.google.protobuf.ProtocolStringListgetArgsList()Optional.PySparkJobgetDefaultInstanceForType()static com.google.protobuf.Descriptors.DescriptorgetDescriptor()com.google.protobuf.Descriptors.DescriptorgetDescriptorForType()StringgetFileUris(int index)Optional.com.google.protobuf.ByteStringgetFileUrisBytes(int index)Optional.intgetFileUrisCount()Optional.com.google.protobuf.ProtocolStringListgetFileUrisList()Optional.StringgetJarFileUris(int index)Optional.com.google.protobuf.ByteStringgetJarFileUrisBytes(int index)Optional.intgetJarFileUrisCount()Optional.com.google.protobuf.ProtocolStringListgetJarFileUrisList()Optional.LoggingConfiggetLoggingConfig()Optional.LoggingConfig.BuildergetLoggingConfigBuilder()Optional.LoggingConfigOrBuildergetLoggingConfigOrBuilder()Optional.StringgetMainPythonFileUri()Required.com.google.protobuf.ByteStringgetMainPythonFileUriBytes()Required.Map<String,String>getMutableProperties()Deprecated.Map<String,String>getProperties()Deprecated.intgetPropertiesCount()Optional.Map<String,String>getPropertiesMap()Optional.StringgetPropertiesOrDefault(String key, String defaultValue)Optional.StringgetPropertiesOrThrow(String key)Optional.StringgetPythonFileUris(int index)Optional.com.google.protobuf.ByteStringgetPythonFileUrisBytes(int index)Optional.intgetPythonFileUrisCount()Optional.com.google.protobuf.ProtocolStringListgetPythonFileUrisList()Optional.booleanhasLoggingConfig()Optional.protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTableinternalGetFieldAccessorTable()protected com.google.protobuf.MapFieldinternalGetMapField(int number)protected com.google.protobuf.MapFieldinternalGetMutableMapField(int number)booleanisInitialized()PySparkJob.BuildermergeFrom(PySparkJob other)PySparkJob.BuildermergeFrom(com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)PySparkJob.BuildermergeFrom(com.google.protobuf.Message other)PySparkJob.BuildermergeLoggingConfig(LoggingConfig value)Optional.PySparkJob.BuildermergeUnknownFields(com.google.protobuf.UnknownFieldSet unknownFields)PySparkJob.BuilderputAllProperties(Map<String,String> values)Optional.PySparkJob.BuilderputProperties(String key, String value)Optional.PySparkJob.BuilderremoveProperties(String key)Optional.PySparkJob.BuildersetArchiveUris(int index, String value)Optional.PySparkJob.BuildersetArgs(int index, String value)Optional.PySparkJob.BuildersetField(com.google.protobuf.Descriptors.FieldDescriptor field, Object value)PySparkJob.BuildersetFileUris(int index, String value)Optional.PySparkJob.BuildersetJarFileUris(int index, String value)Optional.PySparkJob.BuildersetLoggingConfig(LoggingConfig value)Optional.PySparkJob.BuildersetLoggingConfig(LoggingConfig.Builder builderForValue)Optional.PySparkJob.BuildersetMainPythonFileUri(String value)Required.PySparkJob.BuildersetMainPythonFileUriBytes(com.google.protobuf.ByteString value)Required.PySparkJob.BuildersetPythonFileUris(int index, String value)Optional.PySparkJob.BuildersetRepeatedField(com.google.protobuf.Descriptors.FieldDescriptor field, int index, Object value)PySparkJob.BuildersetUnknownFields(com.google.protobuf.UnknownFieldSet unknownFields)-
Methods inherited from class com.google.protobuf.GeneratedMessageV3.Builder
getAllFields, getField, getFieldBuilder, getOneofFieldDescriptor, getParentForChildren, getRepeatedField, getRepeatedFieldBuilder, getRepeatedFieldCount, getUnknownFields, getUnknownFieldSetBuilder, hasField, hasOneof, isClean, markClean, mergeUnknownLengthDelimitedField, mergeUnknownVarintField, newBuilderForField, onBuilt, onChanged, parseUnknownField, setUnknownFieldSetBuilder, setUnknownFieldsProto3
-
Methods inherited from class com.google.protobuf.AbstractMessage.Builder
findInitializationErrors, getInitializationErrorString, internalMergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, newUninitializedMessageException, toString
-
Methods inherited from class com.google.protobuf.AbstractMessageLite.Builder
addAll, addAll, mergeDelimitedFrom, mergeDelimitedFrom, mergeFrom, newUninitializedMessageException
-
Methods inherited from class java.lang.Object
equals, finalize, getClass, hashCode, notify, notifyAll, wait, wait, wait
-
-
-
-
Method Detail
-
getDescriptor
public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()
-
internalGetMapField
protected com.google.protobuf.MapField internalGetMapField(int number)
- Overrides:
internalGetMapFieldin classcom.google.protobuf.GeneratedMessageV3.Builder<PySparkJob.Builder>
-
internalGetMutableMapField
protected com.google.protobuf.MapField internalGetMutableMapField(int number)
- Overrides:
internalGetMutableMapFieldin classcom.google.protobuf.GeneratedMessageV3.Builder<PySparkJob.Builder>
-
internalGetFieldAccessorTable
protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable()
- Specified by:
internalGetFieldAccessorTablein classcom.google.protobuf.GeneratedMessageV3.Builder<PySparkJob.Builder>
-
clear
public PySparkJob.Builder clear()
- Specified by:
clearin interfacecom.google.protobuf.Message.Builder- Specified by:
clearin interfacecom.google.protobuf.MessageLite.Builder- Overrides:
clearin classcom.google.protobuf.GeneratedMessageV3.Builder<PySparkJob.Builder>
-
getDescriptorForType
public com.google.protobuf.Descriptors.Descriptor getDescriptorForType()
- Specified by:
getDescriptorForTypein interfacecom.google.protobuf.Message.Builder- Specified by:
getDescriptorForTypein interfacecom.google.protobuf.MessageOrBuilder- Overrides:
getDescriptorForTypein classcom.google.protobuf.GeneratedMessageV3.Builder<PySparkJob.Builder>
-
getDefaultInstanceForType
public PySparkJob getDefaultInstanceForType()
- Specified by:
getDefaultInstanceForTypein interfacecom.google.protobuf.MessageLiteOrBuilder- Specified by:
getDefaultInstanceForTypein interfacecom.google.protobuf.MessageOrBuilder
-
build
public PySparkJob build()
- Specified by:
buildin interfacecom.google.protobuf.Message.Builder- Specified by:
buildin interfacecom.google.protobuf.MessageLite.Builder
-
buildPartial
public PySparkJob buildPartial()
- Specified by:
buildPartialin interfacecom.google.protobuf.Message.Builder- Specified by:
buildPartialin interfacecom.google.protobuf.MessageLite.Builder
-
clone
public PySparkJob.Builder clone()
- Specified by:
clonein interfacecom.google.protobuf.Message.Builder- Specified by:
clonein interfacecom.google.protobuf.MessageLite.Builder- Overrides:
clonein classcom.google.protobuf.GeneratedMessageV3.Builder<PySparkJob.Builder>
-
setField
public PySparkJob.Builder setField(com.google.protobuf.Descriptors.FieldDescriptor field, Object value)
- Specified by:
setFieldin interfacecom.google.protobuf.Message.Builder- Overrides:
setFieldin classcom.google.protobuf.GeneratedMessageV3.Builder<PySparkJob.Builder>
-
clearField
public PySparkJob.Builder clearField(com.google.protobuf.Descriptors.FieldDescriptor field)
- Specified by:
clearFieldin interfacecom.google.protobuf.Message.Builder- Overrides:
clearFieldin classcom.google.protobuf.GeneratedMessageV3.Builder<PySparkJob.Builder>
-
clearOneof
public PySparkJob.Builder clearOneof(com.google.protobuf.Descriptors.OneofDescriptor oneof)
- Specified by:
clearOneofin interfacecom.google.protobuf.Message.Builder- Overrides:
clearOneofin classcom.google.protobuf.GeneratedMessageV3.Builder<PySparkJob.Builder>
-
setRepeatedField
public PySparkJob.Builder setRepeatedField(com.google.protobuf.Descriptors.FieldDescriptor field, int index, Object value)
- Specified by:
setRepeatedFieldin interfacecom.google.protobuf.Message.Builder- Overrides:
setRepeatedFieldin classcom.google.protobuf.GeneratedMessageV3.Builder<PySparkJob.Builder>
-
addRepeatedField
public PySparkJob.Builder addRepeatedField(com.google.protobuf.Descriptors.FieldDescriptor field, Object value)
- Specified by:
addRepeatedFieldin interfacecom.google.protobuf.Message.Builder- Overrides:
addRepeatedFieldin classcom.google.protobuf.GeneratedMessageV3.Builder<PySparkJob.Builder>
-
mergeFrom
public PySparkJob.Builder mergeFrom(com.google.protobuf.Message other)
- Specified by:
mergeFromin interfacecom.google.protobuf.Message.Builder- Overrides:
mergeFromin classcom.google.protobuf.AbstractMessage.Builder<PySparkJob.Builder>
-
mergeFrom
public PySparkJob.Builder mergeFrom(PySparkJob other)
-
isInitialized
public final boolean isInitialized()
- Specified by:
isInitializedin interfacecom.google.protobuf.MessageLiteOrBuilder- Overrides:
isInitializedin classcom.google.protobuf.GeneratedMessageV3.Builder<PySparkJob.Builder>
-
mergeFrom
public PySparkJob.Builder mergeFrom(com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws IOException
- Specified by:
mergeFromin interfacecom.google.protobuf.Message.Builder- Specified by:
mergeFromin interfacecom.google.protobuf.MessageLite.Builder- Overrides:
mergeFromin classcom.google.protobuf.AbstractMessage.Builder<PySparkJob.Builder>- Throws:
IOException
-
getMainPythonFileUri
public String getMainPythonFileUri()
Required. The HCFS URI of the main Python file to use as the driver. Must be a .py file.
string main_python_file_uri = 1 [(.google.api.field_behavior) = REQUIRED];- Specified by:
getMainPythonFileUriin interfacePySparkJobOrBuilder- Returns:
- The mainPythonFileUri.
-
getMainPythonFileUriBytes
public com.google.protobuf.ByteString getMainPythonFileUriBytes()
Required. The HCFS URI of the main Python file to use as the driver. Must be a .py file.
string main_python_file_uri = 1 [(.google.api.field_behavior) = REQUIRED];- Specified by:
getMainPythonFileUriBytesin interfacePySparkJobOrBuilder- Returns:
- The bytes for mainPythonFileUri.
-
setMainPythonFileUri
public PySparkJob.Builder setMainPythonFileUri(String value)
Required. The HCFS URI of the main Python file to use as the driver. Must be a .py file.
string main_python_file_uri = 1 [(.google.api.field_behavior) = REQUIRED];- Parameters:
value- The mainPythonFileUri to set.- Returns:
- This builder for chaining.
-
clearMainPythonFileUri
public PySparkJob.Builder clearMainPythonFileUri()
Required. The HCFS URI of the main Python file to use as the driver. Must be a .py file.
string main_python_file_uri = 1 [(.google.api.field_behavior) = REQUIRED];- Returns:
- This builder for chaining.
-
setMainPythonFileUriBytes
public PySparkJob.Builder setMainPythonFileUriBytes(com.google.protobuf.ByteString value)
Required. The HCFS URI of the main Python file to use as the driver. Must be a .py file.
string main_python_file_uri = 1 [(.google.api.field_behavior) = REQUIRED];- Parameters:
value- The bytes for mainPythonFileUri to set.- Returns:
- This builder for chaining.
-
getArgsList
public com.google.protobuf.ProtocolStringList getArgsList()
Optional. The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
repeated string args = 2 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getArgsListin interfacePySparkJobOrBuilder- Returns:
- A list containing the args.
-
getArgsCount
public int getArgsCount()
Optional. The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
repeated string args = 2 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getArgsCountin interfacePySparkJobOrBuilder- Returns:
- The count of args.
-
getArgs
public String getArgs(int index)
Optional. The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
repeated string args = 2 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getArgsin interfacePySparkJobOrBuilder- Parameters:
index- The index of the element to return.- Returns:
- The args at the given index.
-
getArgsBytes
public com.google.protobuf.ByteString getArgsBytes(int index)
Optional. The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
repeated string args = 2 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getArgsBytesin interfacePySparkJobOrBuilder- Parameters:
index- The index of the value to return.- Returns:
- The bytes of the args at the given index.
-
setArgs
public PySparkJob.Builder setArgs(int index, String value)
Optional. The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
repeated string args = 2 [(.google.api.field_behavior) = OPTIONAL];- Parameters:
index- The index to set the value at.value- The args to set.- Returns:
- This builder for chaining.
-
addArgs
public PySparkJob.Builder addArgs(String value)
Optional. The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
repeated string args = 2 [(.google.api.field_behavior) = OPTIONAL];- Parameters:
value- The args to add.- Returns:
- This builder for chaining.
-
addAllArgs
public PySparkJob.Builder addAllArgs(Iterable<String> values)
Optional. The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
repeated string args = 2 [(.google.api.field_behavior) = OPTIONAL];- Parameters:
values- The args to add.- Returns:
- This builder for chaining.
-
clearArgs
public PySparkJob.Builder clearArgs()
Optional. The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
repeated string args = 2 [(.google.api.field_behavior) = OPTIONAL];- Returns:
- This builder for chaining.
-
addArgsBytes
public PySparkJob.Builder addArgsBytes(com.google.protobuf.ByteString value)
Optional. The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
repeated string args = 2 [(.google.api.field_behavior) = OPTIONAL];- Parameters:
value- The bytes of the args to add.- Returns:
- This builder for chaining.
-
getPythonFileUrisList
public com.google.protobuf.ProtocolStringList getPythonFileUrisList()
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getPythonFileUrisListin interfacePySparkJobOrBuilder- Returns:
- A list containing the pythonFileUris.
-
getPythonFileUrisCount
public int getPythonFileUrisCount()
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getPythonFileUrisCountin interfacePySparkJobOrBuilder- Returns:
- The count of pythonFileUris.
-
getPythonFileUris
public String getPythonFileUris(int index)
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getPythonFileUrisin interfacePySparkJobOrBuilder- Parameters:
index- The index of the element to return.- Returns:
- The pythonFileUris at the given index.
-
getPythonFileUrisBytes
public com.google.protobuf.ByteString getPythonFileUrisBytes(int index)
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getPythonFileUrisBytesin interfacePySparkJobOrBuilder- Parameters:
index- The index of the value to return.- Returns:
- The bytes of the pythonFileUris at the given index.
-
setPythonFileUris
public PySparkJob.Builder setPythonFileUris(int index, String value)
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3 [(.google.api.field_behavior) = OPTIONAL];- Parameters:
index- The index to set the value at.value- The pythonFileUris to set.- Returns:
- This builder for chaining.
-
addPythonFileUris
public PySparkJob.Builder addPythonFileUris(String value)
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3 [(.google.api.field_behavior) = OPTIONAL];- Parameters:
value- The pythonFileUris to add.- Returns:
- This builder for chaining.
-
addAllPythonFileUris
public PySparkJob.Builder addAllPythonFileUris(Iterable<String> values)
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3 [(.google.api.field_behavior) = OPTIONAL];- Parameters:
values- The pythonFileUris to add.- Returns:
- This builder for chaining.
-
clearPythonFileUris
public PySparkJob.Builder clearPythonFileUris()
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3 [(.google.api.field_behavior) = OPTIONAL];- Returns:
- This builder for chaining.
-
addPythonFileUrisBytes
public PySparkJob.Builder addPythonFileUrisBytes(com.google.protobuf.ByteString value)
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3 [(.google.api.field_behavior) = OPTIONAL];- Parameters:
value- The bytes of the pythonFileUris to add.- Returns:
- This builder for chaining.
-
getJarFileUrisList
public com.google.protobuf.ProtocolStringList getJarFileUrisList()
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getJarFileUrisListin interfacePySparkJobOrBuilder- Returns:
- A list containing the jarFileUris.
-
getJarFileUrisCount
public int getJarFileUrisCount()
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getJarFileUrisCountin interfacePySparkJobOrBuilder- Returns:
- The count of jarFileUris.
-
getJarFileUris
public String getJarFileUris(int index)
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getJarFileUrisin interfacePySparkJobOrBuilder- Parameters:
index- The index of the element to return.- Returns:
- The jarFileUris at the given index.
-
getJarFileUrisBytes
public com.google.protobuf.ByteString getJarFileUrisBytes(int index)
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getJarFileUrisBytesin interfacePySparkJobOrBuilder- Parameters:
index- The index of the value to return.- Returns:
- The bytes of the jarFileUris at the given index.
-
setJarFileUris
public PySparkJob.Builder setJarFileUris(int index, String value)
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4 [(.google.api.field_behavior) = OPTIONAL];- Parameters:
index- The index to set the value at.value- The jarFileUris to set.- Returns:
- This builder for chaining.
-
addJarFileUris
public PySparkJob.Builder addJarFileUris(String value)
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4 [(.google.api.field_behavior) = OPTIONAL];- Parameters:
value- The jarFileUris to add.- Returns:
- This builder for chaining.
-
addAllJarFileUris
public PySparkJob.Builder addAllJarFileUris(Iterable<String> values)
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4 [(.google.api.field_behavior) = OPTIONAL];- Parameters:
values- The jarFileUris to add.- Returns:
- This builder for chaining.
-
clearJarFileUris
public PySparkJob.Builder clearJarFileUris()
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4 [(.google.api.field_behavior) = OPTIONAL];- Returns:
- This builder for chaining.
-
addJarFileUrisBytes
public PySparkJob.Builder addJarFileUrisBytes(com.google.protobuf.ByteString value)
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4 [(.google.api.field_behavior) = OPTIONAL];- Parameters:
value- The bytes of the jarFileUris to add.- Returns:
- This builder for chaining.
-
getFileUrisList
public com.google.protobuf.ProtocolStringList getFileUrisList()
Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
repeated string file_uris = 5 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getFileUrisListin interfacePySparkJobOrBuilder- Returns:
- A list containing the fileUris.
-
getFileUrisCount
public int getFileUrisCount()
Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
repeated string file_uris = 5 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getFileUrisCountin interfacePySparkJobOrBuilder- Returns:
- The count of fileUris.
-
getFileUris
public String getFileUris(int index)
Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
repeated string file_uris = 5 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getFileUrisin interfacePySparkJobOrBuilder- Parameters:
index- The index of the element to return.- Returns:
- The fileUris at the given index.
-
getFileUrisBytes
public com.google.protobuf.ByteString getFileUrisBytes(int index)
Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
repeated string file_uris = 5 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getFileUrisBytesin interfacePySparkJobOrBuilder- Parameters:
index- The index of the value to return.- Returns:
- The bytes of the fileUris at the given index.
-
setFileUris
public PySparkJob.Builder setFileUris(int index, String value)
Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
repeated string file_uris = 5 [(.google.api.field_behavior) = OPTIONAL];- Parameters:
index- The index to set the value at.value- The fileUris to set.- Returns:
- This builder for chaining.
-
addFileUris
public PySparkJob.Builder addFileUris(String value)
Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
repeated string file_uris = 5 [(.google.api.field_behavior) = OPTIONAL];- Parameters:
value- The fileUris to add.- Returns:
- This builder for chaining.
-
addAllFileUris
public PySparkJob.Builder addAllFileUris(Iterable<String> values)
Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
repeated string file_uris = 5 [(.google.api.field_behavior) = OPTIONAL];- Parameters:
values- The fileUris to add.- Returns:
- This builder for chaining.
-
clearFileUris
public PySparkJob.Builder clearFileUris()
Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
repeated string file_uris = 5 [(.google.api.field_behavior) = OPTIONAL];- Returns:
- This builder for chaining.
-
addFileUrisBytes
public PySparkJob.Builder addFileUrisBytes(com.google.protobuf.ByteString value)
Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
repeated string file_uris = 5 [(.google.api.field_behavior) = OPTIONAL];- Parameters:
value- The bytes of the fileUris to add.- Returns:
- This builder for chaining.
-
getArchiveUrisList
public com.google.protobuf.ProtocolStringList getArchiveUrisList()
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getArchiveUrisListin interfacePySparkJobOrBuilder- Returns:
- A list containing the archiveUris.
-
getArchiveUrisCount
public int getArchiveUrisCount()
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getArchiveUrisCountin interfacePySparkJobOrBuilder- Returns:
- The count of archiveUris.
-
getArchiveUris
public String getArchiveUris(int index)
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getArchiveUrisin interfacePySparkJobOrBuilder- Parameters:
index- The index of the element to return.- Returns:
- The archiveUris at the given index.
-
getArchiveUrisBytes
public com.google.protobuf.ByteString getArchiveUrisBytes(int index)
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getArchiveUrisBytesin interfacePySparkJobOrBuilder- Parameters:
index- The index of the value to return.- Returns:
- The bytes of the archiveUris at the given index.
-
setArchiveUris
public PySparkJob.Builder setArchiveUris(int index, String value)
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6 [(.google.api.field_behavior) = OPTIONAL];- Parameters:
index- The index to set the value at.value- The archiveUris to set.- Returns:
- This builder for chaining.
-
addArchiveUris
public PySparkJob.Builder addArchiveUris(String value)
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6 [(.google.api.field_behavior) = OPTIONAL];- Parameters:
value- The archiveUris to add.- Returns:
- This builder for chaining.
-
addAllArchiveUris
public PySparkJob.Builder addAllArchiveUris(Iterable<String> values)
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6 [(.google.api.field_behavior) = OPTIONAL];- Parameters:
values- The archiveUris to add.- Returns:
- This builder for chaining.
-
clearArchiveUris
public PySparkJob.Builder clearArchiveUris()
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6 [(.google.api.field_behavior) = OPTIONAL];- Returns:
- This builder for chaining.
-
addArchiveUrisBytes
public PySparkJob.Builder addArchiveUrisBytes(com.google.protobuf.ByteString value)
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6 [(.google.api.field_behavior) = OPTIONAL];- Parameters:
value- The bytes of the archiveUris to add.- Returns:
- This builder for chaining.
-
getPropertiesCount
public int getPropertiesCount()
Description copied from interface:PySparkJobOrBuilderOptional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getPropertiesCountin interfacePySparkJobOrBuilder
-
containsProperties
public boolean containsProperties(String key)
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
containsPropertiesin interfacePySparkJobOrBuilder
-
getProperties
@Deprecated public Map<String,String> getProperties()
Deprecated.UsegetPropertiesMap()instead.- Specified by:
getPropertiesin interfacePySparkJobOrBuilder
-
getPropertiesMap
public Map<String,String> getPropertiesMap()
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getPropertiesMapin interfacePySparkJobOrBuilder
-
getPropertiesOrDefault
public String getPropertiesOrDefault(String key, String defaultValue)
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getPropertiesOrDefaultin interfacePySparkJobOrBuilder
-
getPropertiesOrThrow
public String getPropertiesOrThrow(String key)
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getPropertiesOrThrowin interfacePySparkJobOrBuilder
-
clearProperties
public PySparkJob.Builder clearProperties()
-
removeProperties
public PySparkJob.Builder removeProperties(String key)
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7 [(.google.api.field_behavior) = OPTIONAL];
-
getMutableProperties
@Deprecated public Map<String,String> getMutableProperties()
Deprecated.Use alternate mutation accessors instead.
-
putProperties
public PySparkJob.Builder putProperties(String key, String value)
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7 [(.google.api.field_behavior) = OPTIONAL];
-
putAllProperties
public PySparkJob.Builder putAllProperties(Map<String,String> values)
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7 [(.google.api.field_behavior) = OPTIONAL];
-
hasLoggingConfig
public boolean hasLoggingConfig()
Optional. The runtime log config for job execution.
.google.cloud.dataproc.v1.LoggingConfig logging_config = 8 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
hasLoggingConfigin interfacePySparkJobOrBuilder- Returns:
- Whether the loggingConfig field is set.
-
getLoggingConfig
public LoggingConfig getLoggingConfig()
Optional. The runtime log config for job execution.
.google.cloud.dataproc.v1.LoggingConfig logging_config = 8 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getLoggingConfigin interfacePySparkJobOrBuilder- Returns:
- The loggingConfig.
-
setLoggingConfig
public PySparkJob.Builder setLoggingConfig(LoggingConfig value)
Optional. The runtime log config for job execution.
.google.cloud.dataproc.v1.LoggingConfig logging_config = 8 [(.google.api.field_behavior) = OPTIONAL];
-
setLoggingConfig
public PySparkJob.Builder setLoggingConfig(LoggingConfig.Builder builderForValue)
Optional. The runtime log config for job execution.
.google.cloud.dataproc.v1.LoggingConfig logging_config = 8 [(.google.api.field_behavior) = OPTIONAL];
-
mergeLoggingConfig
public PySparkJob.Builder mergeLoggingConfig(LoggingConfig value)
Optional. The runtime log config for job execution.
.google.cloud.dataproc.v1.LoggingConfig logging_config = 8 [(.google.api.field_behavior) = OPTIONAL];
-
clearLoggingConfig
public PySparkJob.Builder clearLoggingConfig()
Optional. The runtime log config for job execution.
.google.cloud.dataproc.v1.LoggingConfig logging_config = 8 [(.google.api.field_behavior) = OPTIONAL];
-
getLoggingConfigBuilder
public LoggingConfig.Builder getLoggingConfigBuilder()
Optional. The runtime log config for job execution.
.google.cloud.dataproc.v1.LoggingConfig logging_config = 8 [(.google.api.field_behavior) = OPTIONAL];
-
getLoggingConfigOrBuilder
public LoggingConfigOrBuilder getLoggingConfigOrBuilder()
Optional. The runtime log config for job execution.
.google.cloud.dataproc.v1.LoggingConfig logging_config = 8 [(.google.api.field_behavior) = OPTIONAL];- Specified by:
getLoggingConfigOrBuilderin interfacePySparkJobOrBuilder
-
setUnknownFields
public final PySparkJob.Builder setUnknownFields(com.google.protobuf.UnknownFieldSet unknownFields)
- Specified by:
setUnknownFieldsin interfacecom.google.protobuf.Message.Builder- Overrides:
setUnknownFieldsin classcom.google.protobuf.GeneratedMessageV3.Builder<PySparkJob.Builder>
-
mergeUnknownFields
public final PySparkJob.Builder mergeUnknownFields(com.google.protobuf.UnknownFieldSet unknownFields)
- Specified by:
mergeUnknownFieldsin interfacecom.google.protobuf.Message.Builder- Overrides:
mergeUnknownFieldsin classcom.google.protobuf.GeneratedMessageV3.Builder<PySparkJob.Builder>
-
-