Vertex AI V1 API - Class Google::Cloud::AIPlatform::V1::CustomJobSpec (v0.54.0)

Reference documentation and code samples for the Vertex AI V1 API class Google::Cloud::AIPlatform::V1::CustomJobSpec.

Represents the spec of a CustomJob.

Inherits

  • Object

Extended By

  • Google::Protobuf::MessageExts::ClassMethods

Includes

  • Google::Protobuf::MessageExts

Methods

#base_output_directory

def base_output_directory() -> ::Google::Cloud::AIPlatform::V1::GcsDestination
Returns
  • (::Google::Cloud::AIPlatform::V1::GcsDestination) —

    The Cloud Storage location to store the output of this CustomJob or HyperparameterTuningJob. For HyperparameterTuningJob, the baseOutputDirectory of each child CustomJob backing a Trial is set to a subdirectory of name id under its parent HyperparameterTuningJob's baseOutputDirectory.

    The following Vertex AI environment variables will be passed to containers or python modules when this field is set:

    For CustomJob:

    • AIP_MODEL_DIR = <base_output_directory>/model/
    • AIP_CHECKPOINT_DIR = <base_output_directory>/checkpoints/
    • AIP_TENSORBOARD_LOG_DIR = <base_output_directory>/logs/

    For CustomJob backing a Trial of HyperparameterTuningJob:

    • AIP_MODEL_DIR = <base_output_directory>/<trial_id>/model/
    • AIP_CHECKPOINT_DIR = <base_output_directory>/<trial_id>/checkpoints/
    • AIP_TENSORBOARD_LOG_DIR = <base_output_directory>/<trial_id>/logs/

#base_output_directory=

def base_output_directory=(value) -> ::Google::Cloud::AIPlatform::V1::GcsDestination
Parameter
  • value (::Google::Cloud::AIPlatform::V1::GcsDestination) —

    The Cloud Storage location to store the output of this CustomJob or HyperparameterTuningJob. For HyperparameterTuningJob, the baseOutputDirectory of each child CustomJob backing a Trial is set to a subdirectory of name id under its parent HyperparameterTuningJob's baseOutputDirectory.

    The following Vertex AI environment variables will be passed to containers or python modules when this field is set:

    For CustomJob:

    • AIP_MODEL_DIR = <base_output_directory>/model/
    • AIP_CHECKPOINT_DIR = <base_output_directory>/checkpoints/
    • AIP_TENSORBOARD_LOG_DIR = <base_output_directory>/logs/

    For CustomJob backing a Trial of HyperparameterTuningJob:

    • AIP_MODEL_DIR = <base_output_directory>/<trial_id>/model/
    • AIP_CHECKPOINT_DIR = <base_output_directory>/<trial_id>/checkpoints/
    • AIP_TENSORBOARD_LOG_DIR = <base_output_directory>/<trial_id>/logs/
Returns
  • (::Google::Cloud::AIPlatform::V1::GcsDestination) —

    The Cloud Storage location to store the output of this CustomJob or HyperparameterTuningJob. For HyperparameterTuningJob, the baseOutputDirectory of each child CustomJob backing a Trial is set to a subdirectory of name id under its parent HyperparameterTuningJob's baseOutputDirectory.

    The following Vertex AI environment variables will be passed to containers or python modules when this field is set:

    For CustomJob:

    • AIP_MODEL_DIR = <base_output_directory>/model/
    • AIP_CHECKPOINT_DIR = <base_output_directory>/checkpoints/
    • AIP_TENSORBOARD_LOG_DIR = <base_output_directory>/logs/

    For CustomJob backing a Trial of HyperparameterTuningJob:

    • AIP_MODEL_DIR = <base_output_directory>/<trial_id>/model/
    • AIP_CHECKPOINT_DIR = <base_output_directory>/<trial_id>/checkpoints/
    • AIP_TENSORBOARD_LOG_DIR = <base_output_directory>/<trial_id>/logs/

#enable_dashboard_access

def enable_dashboard_access() -> ::Boolean
Returns

#enable_dashboard_access=

def enable_dashboard_access=(value) -> ::Boolean
Parameter
Returns

#enable_web_access

def enable_web_access() -> ::Boolean
Returns

#enable_web_access=

def enable_web_access=(value) -> ::Boolean
Parameter
Returns

#experiment

def experiment() -> ::String
Returns
  • (::String) — Optional. The Experiment associated with this job. Format: projects/{project}/locations/{location}/metadataStores/{metadataStores}/contexts/{experiment-name}

#experiment=

def experiment=(value) -> ::String
Parameter
  • value (::String) — Optional. The Experiment associated with this job. Format: projects/{project}/locations/{location}/metadataStores/{metadataStores}/contexts/{experiment-name}
Returns
  • (::String) — Optional. The Experiment associated with this job. Format: projects/{project}/locations/{location}/metadataStores/{metadataStores}/contexts/{experiment-name}

#experiment_run

def experiment_run() -> ::String
Returns
  • (::String) — Optional. The Experiment Run associated with this job. Format: projects/{project}/locations/{location}/metadataStores/{metadataStores}/contexts/{experiment-name}-{experiment-run-name}

#experiment_run=

def experiment_run=(value) -> ::String
Parameter
  • value (::String) — Optional. The Experiment Run associated with this job. Format: projects/{project}/locations/{location}/metadataStores/{metadataStores}/contexts/{experiment-name}-{experiment-run-name}
Returns
  • (::String) — Optional. The Experiment Run associated with this job. Format: projects/{project}/locations/{location}/metadataStores/{metadataStores}/contexts/{experiment-name}-{experiment-run-name}

#models

def models() -> ::Array<::String>
Returns
  • (::Array<::String>) — Optional. The name of the Model resources for which to generate a mapping to artifact URIs. Applicable only to some of the Google-provided custom jobs. Format: projects/{project}/locations/{location}/models/{model}

    In order to retrieve a specific version of the model, also provide the version ID or version alias. Example: projects/{project}/locations/{location}/models/{model}@2 or projects/{project}/locations/{location}/models/{model}@golden If no version ID or alias is specified, the "default" version will be returned. The "default" version alias is created for the first version of the model, and can be moved to other versions later on. There will be exactly one default version.

#models=

def models=(value) -> ::Array<::String>
Parameter
  • value (::Array<::String>) — Optional. The name of the Model resources for which to generate a mapping to artifact URIs. Applicable only to some of the Google-provided custom jobs. Format: projects/{project}/locations/{location}/models/{model}

    In order to retrieve a specific version of the model, also provide the version ID or version alias. Example: projects/{project}/locations/{location}/models/{model}@2 or projects/{project}/locations/{location}/models/{model}@golden If no version ID or alias is specified, the "default" version will be returned. The "default" version alias is created for the first version of the model, and can be moved to other versions later on. There will be exactly one default version.

Returns
  • (::Array<::String>) — Optional. The name of the Model resources for which to generate a mapping to artifact URIs. Applicable only to some of the Google-provided custom jobs. Format: projects/{project}/locations/{location}/models/{model}

    In order to retrieve a specific version of the model, also provide the version ID or version alias. Example: projects/{project}/locations/{location}/models/{model}@2 or projects/{project}/locations/{location}/models/{model}@golden If no version ID or alias is specified, the "default" version will be returned. The "default" version alias is created for the first version of the model, and can be moved to other versions later on. There will be exactly one default version.

#network

def network() -> ::String
Returns
  • (::String) — Optional. The full name of the Compute Engine network to which the Job should be peered. For example, projects/12345/global/networks/myVPC. Format is of the form projects/{project}/global/networks/{network}. Where {project} is a project number, as in 12345, and {network} is a network name.

    To specify this field, you must have already configured VPC Network Peering for Vertex AI.

    If this field is left unspecified, the job is not peered with any network.

#network=

def network=(value) -> ::String
Parameter
  • value (::String) — Optional. The full name of the Compute Engine network to which the Job should be peered. For example, projects/12345/global/networks/myVPC. Format is of the form projects/{project}/global/networks/{network}. Where {project} is a project number, as in 12345, and {network} is a network name.

    To specify this field, you must have already configured VPC Network Peering for Vertex AI.

    If this field is left unspecified, the job is not peered with any network.

Returns
  • (::String) — Optional. The full name of the Compute Engine network to which the Job should be peered. For example, projects/12345/global/networks/myVPC. Format is of the form projects/{project}/global/networks/{network}. Where {project} is a project number, as in 12345, and {network} is a network name.

    To specify this field, you must have already configured VPC Network Peering for Vertex AI.

    If this field is left unspecified, the job is not peered with any network.

#persistent_resource_id

def persistent_resource_id() -> ::String
Returns
  • (::String) — Optional. The ID of the PersistentResource in the same Project and Location which to run

    If this is specified, the job will be run on existing machines held by the PersistentResource instead of on-demand short-live machines. The network and CMEK configs on the job should be consistent with those on the PersistentResource, otherwise, the job will be rejected.

#persistent_resource_id=

def persistent_resource_id=(value) -> ::String
Parameter
  • value (::String) — Optional. The ID of the PersistentResource in the same Project and Location which to run

    If this is specified, the job will be run on existing machines held by the PersistentResource instead of on-demand short-live machines. The network and CMEK configs on the job should be consistent with those on the PersistentResource, otherwise, the job will be rejected.

Returns
  • (::String) — Optional. The ID of the PersistentResource in the same Project and Location which to run

    If this is specified, the job will be run on existing machines held by the PersistentResource instead of on-demand short-live machines. The network and CMEK configs on the job should be consistent with those on the PersistentResource, otherwise, the job will be rejected.

#protected_artifact_location_id

def protected_artifact_location_id() -> ::String
Returns
  • (::String) — The ID of the location to store protected artifacts. e.g. us-central1. Populate only when the location is different than CustomJob location. List of supported locations: https://cloud.google.com/vertex-ai/docs/general/locations

#protected_artifact_location_id=

def protected_artifact_location_id=(value) -> ::String
Parameter
  • value (::String) — The ID of the location to store protected artifacts. e.g. us-central1. Populate only when the location is different than CustomJob location. List of supported locations: https://cloud.google.com/vertex-ai/docs/general/locations
Returns
  • (::String) — The ID of the location to store protected artifacts. e.g. us-central1. Populate only when the location is different than CustomJob location. List of supported locations: https://cloud.google.com/vertex-ai/docs/general/locations

#reserved_ip_ranges

def reserved_ip_ranges() -> ::Array<::String>
Returns
  • (::Array<::String>) — Optional. A list of names for the reserved ip ranges under the VPC network that can be used for this job.

    If set, we will deploy the job within the provided ip ranges. Otherwise, the job will be deployed to any ip ranges under the provided VPC network.

    Example: ['vertex-ai-ip-range'].

#reserved_ip_ranges=

def reserved_ip_ranges=(value) -> ::Array<::String>
Parameter
  • value (::Array<::String>) — Optional. A list of names for the reserved ip ranges under the VPC network that can be used for this job.

    If set, we will deploy the job within the provided ip ranges. Otherwise, the job will be deployed to any ip ranges under the provided VPC network.

    Example: ['vertex-ai-ip-range'].

Returns
  • (::Array<::String>) — Optional. A list of names for the reserved ip ranges under the VPC network that can be used for this job.

    If set, we will deploy the job within the provided ip ranges. Otherwise, the job will be deployed to any ip ranges under the provided VPC network.

    Example: ['vertex-ai-ip-range'].

#scheduling

def scheduling() -> ::Google::Cloud::AIPlatform::V1::Scheduling
Returns

#scheduling=

def scheduling=(value) -> ::Google::Cloud::AIPlatform::V1::Scheduling
Parameter
Returns

#service_account

def service_account() -> ::String
Returns
  • (::String) — Specifies the service account for workload run-as account. Users submitting jobs must have act-as permission on this run-as account. If unspecified, the Vertex AI Custom Code Service Agent for the CustomJob's project is used.

#service_account=

def service_account=(value) -> ::String
Parameter
  • value (::String) — Specifies the service account for workload run-as account. Users submitting jobs must have act-as permission on this run-as account. If unspecified, the Vertex AI Custom Code Service Agent for the CustomJob's project is used.
Returns
  • (::String) — Specifies the service account for workload run-as account. Users submitting jobs must have act-as permission on this run-as account. If unspecified, the Vertex AI Custom Code Service Agent for the CustomJob's project is used.

#tensorboard

def tensorboard() -> ::String
Returns
  • (::String) — Optional. The name of a Vertex AI Tensorboard resource to which this CustomJob will upload Tensorboard logs. Format: projects/{project}/locations/{location}/tensorboards/{tensorboard}

#tensorboard=

def tensorboard=(value) -> ::String
Parameter
  • value (::String) — Optional. The name of a Vertex AI Tensorboard resource to which this CustomJob will upload Tensorboard logs. Format: projects/{project}/locations/{location}/tensorboards/{tensorboard}
Returns
  • (::String) — Optional. The name of a Vertex AI Tensorboard resource to which this CustomJob will upload Tensorboard logs. Format: projects/{project}/locations/{location}/tensorboards/{tensorboard}

#worker_pool_specs

def worker_pool_specs() -> ::Array<::Google::Cloud::AIPlatform::V1::WorkerPoolSpec>
Returns
  • (::Array<::Google::Cloud::AIPlatform::V1::WorkerPoolSpec>) — Required. The spec of the worker pools including machine type and Docker image. All worker pools except the first one are optional and can be skipped by providing an empty value.

#worker_pool_specs=

def worker_pool_specs=(value) -> ::Array<::Google::Cloud::AIPlatform::V1::WorkerPoolSpec>
Parameter
  • value (::Array<::Google::Cloud::AIPlatform::V1::WorkerPoolSpec>) — Required. The spec of the worker pools including machine type and Docker image. All worker pools except the first one are optional and can be skipped by providing an empty value.
Returns
  • (::Array<::Google::Cloud::AIPlatform::V1::WorkerPoolSpec>) — Required. The spec of the worker pools including machine type and Docker image. All worker pools except the first one are optional and can be skipped by providing an empty value.