Perform an online prediction.
HTTP request
POST https://{service-endpoint}/v1/{endpoint}:predict
Where {service-endpoint}
is one of the supported service endpoints.
Path parameters
Parameters | |
---|---|
endpoint |
Required. The name of the Endpoint requested to serve the prediction. Format: |
Request body
The request body contains data with the following structure:
JSON representation |
---|
{ "instances": [ value ], "parameters": value } |
Fields | |
---|---|
instances[] |
Required. The instances that are the input to the prediction call. A DeployedModel may have an upper limit on the number of instances it supports per request, and when it is exceeded the prediction call errors in case of AutoML Models, or, in case of customer created Models, the behaviour is as documented by that Model. The schema of any single instance may be specified via Endpoint's DeployedModels' |
parameters |
The parameters that govern the prediction. The schema of the parameters may be specified via Endpoint's DeployedModels' |
Response body
If successful, the response body contains an instance of PredictResponse
.
Authorization scopes
Requires one of the following OAuth scopes:
https://www.googleapis.com/auth/cloud-platform
https://www.googleapis.com/auth/cloud-platform.read-only
https://www.googleapis.com/auth/cloud-vertex-ai.firstparty.predict
For more information, see the Authentication Overview.
IAM Permissions
Requires the following IAM permission on the endpoint
resource:
aiplatform.endpoints.predict
For more information, see the IAM documentation.