gcloud ai endpoints explain

NAME
gcloud ai endpoints explain - request an online explanation from an Vertex AI endpoint
SYNOPSIS
gcloud ai endpoints explain (ENDPOINT : --region=REGION) --json-request=JSON_REQUEST [--deployed-model-id=DEPLOYED_MODEL_ID] [GCLOUD_WIDE_FLAG]
DESCRIPTION
gcloud ai endpoints explain sends an explanation request to the Vertex AI endpoint for the given instances. This command reads up to 100 instances, though the service itself accepts instances up to the payload limit size (currently, 1.5MB).
EXAMPLES
To send an explanation request to the endpoint for the json file, input.json, run:
gcloud ai endpoints explain ENDPOINT_ID --region=us-central1 --json-request=input.json
POSITIONAL ARGUMENTS
Endpoint resource - The endpoint to request an online explanation. The arguments in this group can be used to specify the attributes of this resource. (NOTE) Some attributes are not given arguments in this group but can be set in other ways.

To set the project attribute:

  • provide the argument endpoint on the command line with a fully specified name;
  • provide the argument --project on the command line;
  • set the property core/project.

This must be specified.

ENDPOINT
ID of the endpoint or fully qualified identifier for the endpoint.

To set the name attribute:

  • provide the argument endpoint on the command line.

This positional argument must be specified if any of the other arguments in this group are specified.

--region=REGION
Cloud region for the endpoint.

To set the region attribute:

  • provide the argument endpoint on the command line with a fully specified name;
  • provide the argument --region on the command line;
  • set the property ai/region;
  • choose one from the prompted list of available regions.
REQUIRED FLAGS
--json-request=JSON_REQUEST
Path to a local file containing the body of a JSON request.

An example of a JSON request:

{
  "instances": [
    {"x": [1, 2], "y": [3, 4]},
    {"x": [-1, -2], "y": [-3, -4]}
  ]
}

This flag accepts "-" for stdin.

OPTIONAL FLAGS
--deployed-model-id=DEPLOYED_MODEL_ID
Id of the deployed model.
GCLOUD WIDE FLAGS
These flags are available to all commands: --access-token-file, --account, --billing-project, --configuration, --flags-file, --flatten, --format, --help, --impersonate-service-account, --log-http, --project, --quiet, --trace-token, --user-output-enabled, --verbosity.

Run $ gcloud help for details.

NOTES
These variants are also available:
gcloud alpha ai endpoints explain
gcloud beta ai endpoints explain