BatchPredictionJob component allows you to run an asynchronous
prediction request. You request batch predictions directly from the
resource; you do not need to deploy the model to an
endpoint. For data types
that support both batch and online predictions you can use batch predictions.
This is useful when you don't require an immediate response and want to process
accumulated data by using a single request.
To make a batch prediction, you specify an input source and an output location
for Vertex AI to store predictions results. The inputs and outputs
depend on the
model type that you're working with. For example, batch
predictions for the AutoML image model type require an input
file and the name of a Cloud Storage bucket to store the output.
For more information about batch prediction, see
Get batch predictions.
You can use the
ModelBatchPredictOp to access this functionality through
Vertex AI Pipelines.
- For component reference, see the
- For Vertex AI API reference, see the
Version history and changelog
|03/2022||GCPC v1.0||Version 1.0 release of the components.
|02/2022||GCPC v0.3||New Experimental version of the components.|
|11/2021||GCPC v0.2||Experimental release of the components.|
Technical Support Contacts
If you have any questions, please reach out to email@example.com.