public static final class BatchPredictionJob.OutputConfig.Builder extends GeneratedMessageV3.Builder<BatchPredictionJob.OutputConfig.Builder> implements BatchPredictionJob.OutputConfigOrBuilder
Configures the output of BatchPredictionJob.
See Model.supported_output_storage_formats for supported output
formats, and how predictions are expressed via any of them.
Protobuf type google.cloud.aiplatform.v1.BatchPredictionJob.OutputConfig
Methods
public BatchPredictionJob.OutputConfig.Builder addRepeatedField(Descriptors.FieldDescriptor field, Object value)
Parameters
Returns
Overrides
public BatchPredictionJob.OutputConfig build()
Returns
public BatchPredictionJob.OutputConfig buildPartial()
Returns
public BatchPredictionJob.OutputConfig.Builder clear()
Returns
Overrides
public BatchPredictionJob.OutputConfig.Builder clearBigqueryDestination()
The BigQuery project or dataset location where the output is to be
written to. If project is provided, a new dataset is created with name
prediction_<model-display-name>_<job-create-time>
where <model-display-name> is made
BigQuery-dataset-name compatible (for example, most special characters
become underscores), and timestamp is in
YYYY_MM_DDThh_mm_ss_sssZ "based on ISO-8601" format. In the dataset
two tables will be created, predictions
, and errors
.
If the Model has both instance
and prediction schemata
defined then the tables have columns as follows: The predictions
table contains instances for which the prediction succeeded, it
has columns as per a concatenation of the Model's instance and
prediction schemata. The errors
table contains rows for which the
prediction has failed, it has instance columns, as per the
instance schema, followed by a single "errors" column, which as values
has google.rpc.Status
represented as a STRUCT, and containing only code
and message
.
.google.cloud.aiplatform.v1.BigQueryDestination bigquery_destination = 3;
Returns
public BatchPredictionJob.OutputConfig.Builder clearDestination()
Returns
public BatchPredictionJob.OutputConfig.Builder clearField(Descriptors.FieldDescriptor field)
Parameter
Returns
Overrides
public BatchPredictionJob.OutputConfig.Builder clearGcsDestination()
The Cloud Storage location of the directory where the output is
to be written to. In the given directory a new directory is created.
Its name is prediction-<model-display-name>-<job-create-time>
,
where timestamp is in YYYY-MM-DDThh:mm:ss.sssZ ISO-8601 format.
Inside of it files predictions_0001.<extension>
,
predictions_0002.<extension>
, ..., predictions_N.<extension>
are created where <extension>
depends on chosen
predictions_format, and N may equal 0001 and depends on the total
number of successfully predicted instances.
If the Model has both instance
and prediction schemata
defined then each such file contains predictions as per the
predictions_format.
If prediction for any instance failed (partially or completely), then
an additional errors_0001.<extension>
, errors_0002.<extension>
,...,
errors_N.<extension>
files are created (N depends on total number
of failed predictions). These files contain the failed instances,
as per their schema, followed by an additional error
field which as
value has google.rpc.Status
containing only code
and message
fields.
.google.cloud.aiplatform.v1.GcsDestination gcs_destination = 2;
Returns
public BatchPredictionJob.OutputConfig.Builder clearOneof(Descriptors.OneofDescriptor oneof)
Parameter
Returns
Overrides
public BatchPredictionJob.OutputConfig.Builder clearPredictionsFormat()
Required. The format in which Vertex AI gives the predictions, must be one of the
Model's
supported_output_storage_formats.
string predictions_format = 1 [(.google.api.field_behavior) = REQUIRED];
Returns
public BatchPredictionJob.OutputConfig.Builder clone()
Returns
Overrides
public BigQueryDestination getBigqueryDestination()
The BigQuery project or dataset location where the output is to be
written to. If project is provided, a new dataset is created with name
prediction_<model-display-name>_<job-create-time>
where <model-display-name> is made
BigQuery-dataset-name compatible (for example, most special characters
become underscores), and timestamp is in
YYYY_MM_DDThh_mm_ss_sssZ "based on ISO-8601" format. In the dataset
two tables will be created, predictions
, and errors
.
If the Model has both instance
and prediction schemata
defined then the tables have columns as follows: The predictions
table contains instances for which the prediction succeeded, it
has columns as per a concatenation of the Model's instance and
prediction schemata. The errors
table contains rows for which the
prediction has failed, it has instance columns, as per the
instance schema, followed by a single "errors" column, which as values
has google.rpc.Status
represented as a STRUCT, and containing only code
and message
.
.google.cloud.aiplatform.v1.BigQueryDestination bigquery_destination = 3;
Returns
public BigQueryDestination.Builder getBigqueryDestinationBuilder()
The BigQuery project or dataset location where the output is to be
written to. If project is provided, a new dataset is created with name
prediction_<model-display-name>_<job-create-time>
where <model-display-name> is made
BigQuery-dataset-name compatible (for example, most special characters
become underscores), and timestamp is in
YYYY_MM_DDThh_mm_ss_sssZ "based on ISO-8601" format. In the dataset
two tables will be created, predictions
, and errors
.
If the Model has both instance
and prediction schemata
defined then the tables have columns as follows: The predictions
table contains instances for which the prediction succeeded, it
has columns as per a concatenation of the Model's instance and
prediction schemata. The errors
table contains rows for which the
prediction has failed, it has instance columns, as per the
instance schema, followed by a single "errors" column, which as values
has google.rpc.Status
represented as a STRUCT, and containing only code
and message
.
.google.cloud.aiplatform.v1.BigQueryDestination bigquery_destination = 3;
Returns
public BigQueryDestinationOrBuilder getBigqueryDestinationOrBuilder()
The BigQuery project or dataset location where the output is to be
written to. If project is provided, a new dataset is created with name
prediction_<model-display-name>_<job-create-time>
where <model-display-name> is made
BigQuery-dataset-name compatible (for example, most special characters
become underscores), and timestamp is in
YYYY_MM_DDThh_mm_ss_sssZ "based on ISO-8601" format. In the dataset
two tables will be created, predictions
, and errors
.
If the Model has both instance
and prediction schemata
defined then the tables have columns as follows: The predictions
table contains instances for which the prediction succeeded, it
has columns as per a concatenation of the Model's instance and
prediction schemata. The errors
table contains rows for which the
prediction has failed, it has instance columns, as per the
instance schema, followed by a single "errors" column, which as values
has google.rpc.Status
represented as a STRUCT, and containing only code
and message
.
.google.cloud.aiplatform.v1.BigQueryDestination bigquery_destination = 3;
Returns
public BatchPredictionJob.OutputConfig getDefaultInstanceForType()
Returns
public static final Descriptors.Descriptor getDescriptor()
Returns
public Descriptors.Descriptor getDescriptorForType()
Returns
Overrides
public BatchPredictionJob.OutputConfig.DestinationCase getDestinationCase()
Returns
public GcsDestination getGcsDestination()
The Cloud Storage location of the directory where the output is
to be written to. In the given directory a new directory is created.
Its name is prediction-<model-display-name>-<job-create-time>
,
where timestamp is in YYYY-MM-DDThh:mm:ss.sssZ ISO-8601 format.
Inside of it files predictions_0001.<extension>
,
predictions_0002.<extension>
, ..., predictions_N.<extension>
are created where <extension>
depends on chosen
predictions_format, and N may equal 0001 and depends on the total
number of successfully predicted instances.
If the Model has both instance
and prediction schemata
defined then each such file contains predictions as per the
predictions_format.
If prediction for any instance failed (partially or completely), then
an additional errors_0001.<extension>
, errors_0002.<extension>
,...,
errors_N.<extension>
files are created (N depends on total number
of failed predictions). These files contain the failed instances,
as per their schema, followed by an additional error
field which as
value has google.rpc.Status
containing only code
and message
fields.
.google.cloud.aiplatform.v1.GcsDestination gcs_destination = 2;
Returns
public GcsDestination.Builder getGcsDestinationBuilder()
The Cloud Storage location of the directory where the output is
to be written to. In the given directory a new directory is created.
Its name is prediction-<model-display-name>-<job-create-time>
,
where timestamp is in YYYY-MM-DDThh:mm:ss.sssZ ISO-8601 format.
Inside of it files predictions_0001.<extension>
,
predictions_0002.<extension>
, ..., predictions_N.<extension>
are created where <extension>
depends on chosen
predictions_format, and N may equal 0001 and depends on the total
number of successfully predicted instances.
If the Model has both instance
and prediction schemata
defined then each such file contains predictions as per the
predictions_format.
If prediction for any instance failed (partially or completely), then
an additional errors_0001.<extension>
, errors_0002.<extension>
,...,
errors_N.<extension>
files are created (N depends on total number
of failed predictions). These files contain the failed instances,
as per their schema, followed by an additional error
field which as
value has google.rpc.Status
containing only code
and message
fields.
.google.cloud.aiplatform.v1.GcsDestination gcs_destination = 2;
Returns
public GcsDestinationOrBuilder getGcsDestinationOrBuilder()
The Cloud Storage location of the directory where the output is
to be written to. In the given directory a new directory is created.
Its name is prediction-<model-display-name>-<job-create-time>
,
where timestamp is in YYYY-MM-DDThh:mm:ss.sssZ ISO-8601 format.
Inside of it files predictions_0001.<extension>
,
predictions_0002.<extension>
, ..., predictions_N.<extension>
are created where <extension>
depends on chosen
predictions_format, and N may equal 0001 and depends on the total
number of successfully predicted instances.
If the Model has both instance
and prediction schemata
defined then each such file contains predictions as per the
predictions_format.
If prediction for any instance failed (partially or completely), then
an additional errors_0001.<extension>
, errors_0002.<extension>
,...,
errors_N.<extension>
files are created (N depends on total number
of failed predictions). These files contain the failed instances,
as per their schema, followed by an additional error
field which as
value has google.rpc.Status
containing only code
and message
fields.
.google.cloud.aiplatform.v1.GcsDestination gcs_destination = 2;
Returns
public String getPredictionsFormat()
Required. The format in which Vertex AI gives the predictions, must be one of the
Model's
supported_output_storage_formats.
string predictions_format = 1 [(.google.api.field_behavior) = REQUIRED];
Returns
Type | Description |
String | The predictionsFormat.
|
public ByteString getPredictionsFormatBytes()
Required. The format in which Vertex AI gives the predictions, must be one of the
Model's
supported_output_storage_formats.
string predictions_format = 1 [(.google.api.field_behavior) = REQUIRED];
Returns
Type | Description |
ByteString | The bytes for predictionsFormat.
|
public boolean hasBigqueryDestination()
The BigQuery project or dataset location where the output is to be
written to. If project is provided, a new dataset is created with name
prediction_<model-display-name>_<job-create-time>
where <model-display-name> is made
BigQuery-dataset-name compatible (for example, most special characters
become underscores), and timestamp is in
YYYY_MM_DDThh_mm_ss_sssZ "based on ISO-8601" format. In the dataset
two tables will be created, predictions
, and errors
.
If the Model has both instance
and prediction schemata
defined then the tables have columns as follows: The predictions
table contains instances for which the prediction succeeded, it
has columns as per a concatenation of the Model's instance and
prediction schemata. The errors
table contains rows for which the
prediction has failed, it has instance columns, as per the
instance schema, followed by a single "errors" column, which as values
has google.rpc.Status
represented as a STRUCT, and containing only code
and message
.
.google.cloud.aiplatform.v1.BigQueryDestination bigquery_destination = 3;
Returns
Type | Description |
boolean | Whether the bigqueryDestination field is set.
|
public boolean hasGcsDestination()
The Cloud Storage location of the directory where the output is
to be written to. In the given directory a new directory is created.
Its name is prediction-<model-display-name>-<job-create-time>
,
where timestamp is in YYYY-MM-DDThh:mm:ss.sssZ ISO-8601 format.
Inside of it files predictions_0001.<extension>
,
predictions_0002.<extension>
, ..., predictions_N.<extension>
are created where <extension>
depends on chosen
predictions_format, and N may equal 0001 and depends on the total
number of successfully predicted instances.
If the Model has both instance
and prediction schemata
defined then each such file contains predictions as per the
predictions_format.
If prediction for any instance failed (partially or completely), then
an additional errors_0001.<extension>
, errors_0002.<extension>
,...,
errors_N.<extension>
files are created (N depends on total number
of failed predictions). These files contain the failed instances,
as per their schema, followed by an additional error
field which as
value has google.rpc.Status
containing only code
and message
fields.
.google.cloud.aiplatform.v1.GcsDestination gcs_destination = 2;
Returns
Type | Description |
boolean | Whether the gcsDestination field is set.
|
protected GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable()
Returns
Overrides
public final boolean isInitialized()
Returns
Overrides
public BatchPredictionJob.OutputConfig.Builder mergeBigqueryDestination(BigQueryDestination value)
The BigQuery project or dataset location where the output is to be
written to. If project is provided, a new dataset is created with name
prediction_<model-display-name>_<job-create-time>
where <model-display-name> is made
BigQuery-dataset-name compatible (for example, most special characters
become underscores), and timestamp is in
YYYY_MM_DDThh_mm_ss_sssZ "based on ISO-8601" format. In the dataset
two tables will be created, predictions
, and errors
.
If the Model has both instance
and prediction schemata
defined then the tables have columns as follows: The predictions
table contains instances for which the prediction succeeded, it
has columns as per a concatenation of the Model's instance and
prediction schemata. The errors
table contains rows for which the
prediction has failed, it has instance columns, as per the
instance schema, followed by a single "errors" column, which as values
has google.rpc.Status
represented as a STRUCT, and containing only code
and message
.
.google.cloud.aiplatform.v1.BigQueryDestination bigquery_destination = 3;
Parameter
Returns
public BatchPredictionJob.OutputConfig.Builder mergeFrom(BatchPredictionJob.OutputConfig other)
Parameter
Returns
public BatchPredictionJob.OutputConfig.Builder mergeFrom(CodedInputStream input, ExtensionRegistryLite extensionRegistry)
Parameters
Returns
Overrides
Exceptions
public BatchPredictionJob.OutputConfig.Builder mergeFrom(Message other)
Parameter
Returns
Overrides
public BatchPredictionJob.OutputConfig.Builder mergeGcsDestination(GcsDestination value)
The Cloud Storage location of the directory where the output is
to be written to. In the given directory a new directory is created.
Its name is prediction-<model-display-name>-<job-create-time>
,
where timestamp is in YYYY-MM-DDThh:mm:ss.sssZ ISO-8601 format.
Inside of it files predictions_0001.<extension>
,
predictions_0002.<extension>
, ..., predictions_N.<extension>
are created where <extension>
depends on chosen
predictions_format, and N may equal 0001 and depends on the total
number of successfully predicted instances.
If the Model has both instance
and prediction schemata
defined then each such file contains predictions as per the
predictions_format.
If prediction for any instance failed (partially or completely), then
an additional errors_0001.<extension>
, errors_0002.<extension>
,...,
errors_N.<extension>
files are created (N depends on total number
of failed predictions). These files contain the failed instances,
as per their schema, followed by an additional error
field which as
value has google.rpc.Status
containing only code
and message
fields.
.google.cloud.aiplatform.v1.GcsDestination gcs_destination = 2;
Parameter
Returns
public final BatchPredictionJob.OutputConfig.Builder mergeUnknownFields(UnknownFieldSet unknownFields)
Parameter
Returns
Overrides
public BatchPredictionJob.OutputConfig.Builder setBigqueryDestination(BigQueryDestination value)
The BigQuery project or dataset location where the output is to be
written to. If project is provided, a new dataset is created with name
prediction_<model-display-name>_<job-create-time>
where <model-display-name> is made
BigQuery-dataset-name compatible (for example, most special characters
become underscores), and timestamp is in
YYYY_MM_DDThh_mm_ss_sssZ "based on ISO-8601" format. In the dataset
two tables will be created, predictions
, and errors
.
If the Model has both instance
and prediction schemata
defined then the tables have columns as follows: The predictions
table contains instances for which the prediction succeeded, it
has columns as per a concatenation of the Model's instance and
prediction schemata. The errors
table contains rows for which the
prediction has failed, it has instance columns, as per the
instance schema, followed by a single "errors" column, which as values
has google.rpc.Status
represented as a STRUCT, and containing only code
and message
.
.google.cloud.aiplatform.v1.BigQueryDestination bigquery_destination = 3;
Parameter
Returns
public BatchPredictionJob.OutputConfig.Builder setBigqueryDestination(BigQueryDestination.Builder builderForValue)
The BigQuery project or dataset location where the output is to be
written to. If project is provided, a new dataset is created with name
prediction_<model-display-name>_<job-create-time>
where <model-display-name> is made
BigQuery-dataset-name compatible (for example, most special characters
become underscores), and timestamp is in
YYYY_MM_DDThh_mm_ss_sssZ "based on ISO-8601" format. In the dataset
two tables will be created, predictions
, and errors
.
If the Model has both instance
and prediction schemata
defined then the tables have columns as follows: The predictions
table contains instances for which the prediction succeeded, it
has columns as per a concatenation of the Model's instance and
prediction schemata. The errors
table contains rows for which the
prediction has failed, it has instance columns, as per the
instance schema, followed by a single "errors" column, which as values
has google.rpc.Status
represented as a STRUCT, and containing only code
and message
.
.google.cloud.aiplatform.v1.BigQueryDestination bigquery_destination = 3;
Parameter
Returns
public BatchPredictionJob.OutputConfig.Builder setField(Descriptors.FieldDescriptor field, Object value)
Parameters
Returns
Overrides
public BatchPredictionJob.OutputConfig.Builder setGcsDestination(GcsDestination value)
The Cloud Storage location of the directory where the output is
to be written to. In the given directory a new directory is created.
Its name is prediction-<model-display-name>-<job-create-time>
,
where timestamp is in YYYY-MM-DDThh:mm:ss.sssZ ISO-8601 format.
Inside of it files predictions_0001.<extension>
,
predictions_0002.<extension>
, ..., predictions_N.<extension>
are created where <extension>
depends on chosen
predictions_format, and N may equal 0001 and depends on the total
number of successfully predicted instances.
If the Model has both instance
and prediction schemata
defined then each such file contains predictions as per the
predictions_format.
If prediction for any instance failed (partially or completely), then
an additional errors_0001.<extension>
, errors_0002.<extension>
,...,
errors_N.<extension>
files are created (N depends on total number
of failed predictions). These files contain the failed instances,
as per their schema, followed by an additional error
field which as
value has google.rpc.Status
containing only code
and message
fields.
.google.cloud.aiplatform.v1.GcsDestination gcs_destination = 2;
Parameter
Returns
public BatchPredictionJob.OutputConfig.Builder setGcsDestination(GcsDestination.Builder builderForValue)
The Cloud Storage location of the directory where the output is
to be written to. In the given directory a new directory is created.
Its name is prediction-<model-display-name>-<job-create-time>
,
where timestamp is in YYYY-MM-DDThh:mm:ss.sssZ ISO-8601 format.
Inside of it files predictions_0001.<extension>
,
predictions_0002.<extension>
, ..., predictions_N.<extension>
are created where <extension>
depends on chosen
predictions_format, and N may equal 0001 and depends on the total
number of successfully predicted instances.
If the Model has both instance
and prediction schemata
defined then each such file contains predictions as per the
predictions_format.
If prediction for any instance failed (partially or completely), then
an additional errors_0001.<extension>
, errors_0002.<extension>
,...,
errors_N.<extension>
files are created (N depends on total number
of failed predictions). These files contain the failed instances,
as per their schema, followed by an additional error
field which as
value has google.rpc.Status
containing only code
and message
fields.
.google.cloud.aiplatform.v1.GcsDestination gcs_destination = 2;
Parameter
Returns
public BatchPredictionJob.OutputConfig.Builder setPredictionsFormat(String value)
Required. The format in which Vertex AI gives the predictions, must be one of the
Model's
supported_output_storage_formats.
string predictions_format = 1 [(.google.api.field_behavior) = REQUIRED];
Parameter
Name | Description |
value | String
The predictionsFormat to set.
|
Returns
public BatchPredictionJob.OutputConfig.Builder setPredictionsFormatBytes(ByteString value)
Required. The format in which Vertex AI gives the predictions, must be one of the
Model's
supported_output_storage_formats.
string predictions_format = 1 [(.google.api.field_behavior) = REQUIRED];
Parameter
Name | Description |
value | ByteString
The bytes for predictionsFormat to set.
|
Returns
public BatchPredictionJob.OutputConfig.Builder setRepeatedField(Descriptors.FieldDescriptor field, int index, Object value)
Parameters
Returns
Overrides
public final BatchPredictionJob.OutputConfig.Builder setUnknownFields(UnknownFieldSet unknownFields)
Parameter
Returns
Overrides