Class Google::Cloud::AIPlatform::V1::CustomJobSpec (v0.1.0)

Stay organized with collections Save and categorize content based on your preferences.

Represents the spec of a CustomJob. Next Id: 14

Inherits

  • Object

Extended By

  • Google::Protobuf::MessageExts::ClassMethods

Includes

  • Google::Protobuf::MessageExts

Methods

#base_output_directory

def base_output_directory() -> ::Google::Cloud::AIPlatform::V1::GcsDestination
Returns
  • (::Google::Cloud::AIPlatform::V1::GcsDestination) —

    The Cloud Storage location to store the output of this CustomJob or HyperparameterTuningJob. For HyperparameterTuningJob, the baseOutputDirectory of each child CustomJob backing a Trial is set to a subdirectory of name id under its parent HyperparameterTuningJob's baseOutputDirectory.

    The following Vertex AI environment variables will be passed to containers or python modules when this field is set:

    For CustomJob:

    • AIP_MODEL_DIR = <base_output_directory>/model/
    • AIP_CHECKPOINT_DIR = <base_output_directory>/checkpoints/
    • AIP_TENSORBOARD_LOG_DIR = <base_output_directory>/logs/

    For CustomJob backing a Trial of HyperparameterTuningJob:

    • AIP_MODEL_DIR = <base_output_directory>/<trial_id>/model/
    • AIP_CHECKPOINT_DIR = <base_output_directory>/<trial_id>/checkpoints/
    • AIP_TENSORBOARD_LOG_DIR = <base_output_directory>/<trial_id>/logs/

#base_output_directory=

def base_output_directory=(value) -> ::Google::Cloud::AIPlatform::V1::GcsDestination
Parameter
  • value (::Google::Cloud::AIPlatform::V1::GcsDestination) —

    The Cloud Storage location to store the output of this CustomJob or HyperparameterTuningJob. For HyperparameterTuningJob, the baseOutputDirectory of each child CustomJob backing a Trial is set to a subdirectory of name id under its parent HyperparameterTuningJob's baseOutputDirectory.

    The following Vertex AI environment variables will be passed to containers or python modules when this field is set:

    For CustomJob:

    • AIP_MODEL_DIR = <base_output_directory>/model/
    • AIP_CHECKPOINT_DIR = <base_output_directory>/checkpoints/
    • AIP_TENSORBOARD_LOG_DIR = <base_output_directory>/logs/

    For CustomJob backing a Trial of HyperparameterTuningJob:

    • AIP_MODEL_DIR = <base_output_directory>/<trial_id>/model/
    • AIP_CHECKPOINT_DIR = <base_output_directory>/<trial_id>/checkpoints/
    • AIP_TENSORBOARD_LOG_DIR = <base_output_directory>/<trial_id>/logs/
Returns
  • (::Google::Cloud::AIPlatform::V1::GcsDestination) —

    The Cloud Storage location to store the output of this CustomJob or HyperparameterTuningJob. For HyperparameterTuningJob, the baseOutputDirectory of each child CustomJob backing a Trial is set to a subdirectory of name id under its parent HyperparameterTuningJob's baseOutputDirectory.

    The following Vertex AI environment variables will be passed to containers or python modules when this field is set:

    For CustomJob:

    • AIP_MODEL_DIR = <base_output_directory>/model/
    • AIP_CHECKPOINT_DIR = <base_output_directory>/checkpoints/
    • AIP_TENSORBOARD_LOG_DIR = <base_output_directory>/logs/

    For CustomJob backing a Trial of HyperparameterTuningJob:

    • AIP_MODEL_DIR = <base_output_directory>/<trial_id>/model/
    • AIP_CHECKPOINT_DIR = <base_output_directory>/<trial_id>/checkpoints/
    • AIP_TENSORBOARD_LOG_DIR = <base_output_directory>/<trial_id>/logs/

#enable_web_access

def enable_web_access() -> ::Boolean
Returns

#enable_web_access=

def enable_web_access=(value) -> ::Boolean
Parameter
Returns

#network

def network() -> ::String
Returns
  • (::String) — The full name of the Compute Engine network to which the Job should be peered. For example, projects/12345/global/networks/myVPC. Format is of the form projects/{project}/global/networks/{network}. Where {project} is a project number, as in 12345, and {network} is a network name.

    To specify this field, you must have already configured VPC Network Peering for Vertex AI.

    If this field is left unspecified, the job is not peered with any network.

#network=

def network=(value) -> ::String
Parameter
  • value (::String) — The full name of the Compute Engine network to which the Job should be peered. For example, projects/12345/global/networks/myVPC. Format is of the form projects/{project}/global/networks/{network}. Where {project} is a project number, as in 12345, and {network} is a network name.

    To specify this field, you must have already configured VPC Network Peering for Vertex AI.

    If this field is left unspecified, the job is not peered with any network.

Returns
  • (::String) — The full name of the Compute Engine network to which the Job should be peered. For example, projects/12345/global/networks/myVPC. Format is of the form projects/{project}/global/networks/{network}. Where {project} is a project number, as in 12345, and {network} is a network name.

    To specify this field, you must have already configured VPC Network Peering for Vertex AI.

    If this field is left unspecified, the job is not peered with any network.

#scheduling

def scheduling() -> ::Google::Cloud::AIPlatform::V1::Scheduling
Returns

#scheduling=

def scheduling=(value) -> ::Google::Cloud::AIPlatform::V1::Scheduling
Parameter
Returns

#service_account

def service_account() -> ::String
Returns
  • (::String) — Specifies the service account for workload run-as account. Users submitting jobs must have act-as permission on this run-as account. If unspecified, the Vertex AI Custom Code Service Agent for the CustomJob's project is used.

#service_account=

def service_account=(value) -> ::String
Parameter
  • value (::String) — Specifies the service account for workload run-as account. Users submitting jobs must have act-as permission on this run-as account. If unspecified, the Vertex AI Custom Code Service Agent for the CustomJob's project is used.
Returns
  • (::String) — Specifies the service account for workload run-as account. Users submitting jobs must have act-as permission on this run-as account. If unspecified, the Vertex AI Custom Code Service Agent for the CustomJob's project is used.

#tensorboard

def tensorboard() -> ::String
Returns
  • (::String) — Optional. The name of a Vertex AI Tensorboard resource to which this CustomJob will upload Tensorboard logs. Format: projects/{project}/locations/{location}/tensorboards/{tensorboard}

#tensorboard=

def tensorboard=(value) -> ::String
Parameter
  • value (::String) — Optional. The name of a Vertex AI Tensorboard resource to which this CustomJob will upload Tensorboard logs. Format: projects/{project}/locations/{location}/tensorboards/{tensorboard}
Returns
  • (::String) — Optional. The name of a Vertex AI Tensorboard resource to which this CustomJob will upload Tensorboard logs. Format: projects/{project}/locations/{location}/tensorboards/{tensorboard}

#worker_pool_specs

def worker_pool_specs() -> ::Array<::Google::Cloud::AIPlatform::V1::WorkerPoolSpec>
Returns
  • (::Array<::Google::Cloud::AIPlatform::V1::WorkerPoolSpec>) — Required. The spec of the worker pools including machine type and Docker image. All worker pools except the first one are optional and can be skipped by providing an empty value.

#worker_pool_specs=

def worker_pool_specs=(value) -> ::Array<::Google::Cloud::AIPlatform::V1::WorkerPoolSpec>
Parameter
  • value (::Array<::Google::Cloud::AIPlatform::V1::WorkerPoolSpec>) — Required. The spec of the worker pools including machine type and Docker image. All worker pools except the first one are optional and can be skipped by providing an empty value.
Returns
  • (::Array<::Google::Cloud::AIPlatform::V1::WorkerPoolSpec>) — Required. The spec of the worker pools including machine type and Docker image. All worker pools except the first one are optional and can be skipped by providing an empty value.