Index
DatasetService
(interface)DeploymentResourcePoolService
(interface)EndpointService
(interface)EvaluationService
(interface)FeatureOnlineStoreAdminService
(interface)FeatureOnlineStoreService
(interface)FeatureRegistryService
(interface)FeaturestoreOnlineServingService
(interface)FeaturestoreService
(interface)GenAiTuningService
(interface)IndexEndpointService
(interface)IndexService
(interface)JobService
(interface)LlmUtilityService
(interface)MatchService
(interface)MetadataService
(interface)MigrationService
(interface)ModelGardenService
(interface)ModelService
(interface)NotebookService
(interface)PersistentResourceService
(interface)PipelineService
(interface)PredictionService
(interface)ScheduleService
(interface)SpecialistPoolService
(interface)TensorboardService
(interface)VertexRagDataService
(interface)VertexRagService
(interface)VizierService
(interface)AcceleratorType
(enum)AddContextArtifactsAndExecutionsRequest
(message)AddContextArtifactsAndExecutionsResponse
(message)AddContextChildrenRequest
(message)AddContextChildrenResponse
(message)AddExecutionEventsRequest
(message)AddExecutionEventsResponse
(message)AddTrialMeasurementRequest
(message)Annotation
(message)AnnotationSpec
(message)ApiAuth
(message)ApiAuth.ApiKeyConfig
(message)Artifact
(message)Artifact.State
(enum)AssignNotebookRuntimeOperationMetadata
(message)AssignNotebookRuntimeRequest
(message)Attribution
(message)AugmentPromptRequest
(message)AugmentPromptRequest.Model
(message)AugmentPromptResponse
(message)AutomaticResources
(message)AutoscalingMetricSpec
(message)AvroSource
(message)BatchCancelPipelineJobsOperationMetadata
(message)BatchCancelPipelineJobsRequest
(message)BatchCancelPipelineJobsResponse
(message)BatchCreateFeaturesOperationMetadata
(message)BatchCreateFeaturesRequest
(message)BatchCreateFeaturesResponse
(message)BatchCreateTensorboardRunsRequest
(message)BatchCreateTensorboardRunsResponse
(message)BatchCreateTensorboardTimeSeriesRequest
(message)BatchCreateTensorboardTimeSeriesResponse
(message)BatchDedicatedResources
(message)BatchDeletePipelineJobsRequest
(message)BatchDeletePipelineJobsResponse
(message)BatchImportEvaluatedAnnotationsRequest
(message)BatchImportEvaluatedAnnotationsResponse
(message)BatchImportModelEvaluationSlicesRequest
(message)BatchImportModelEvaluationSlicesResponse
(message)BatchMigrateResourcesOperationMetadata
(message)BatchMigrateResourcesOperationMetadata.PartialResult
(message)BatchMigrateResourcesRequest
(message)BatchMigrateResourcesResponse
(message)BatchPredictionJob
(message)BatchPredictionJob.InputConfig
(message)BatchPredictionJob.InstanceConfig
(message)BatchPredictionJob.OutputConfig
(message)BatchPredictionJob.OutputInfo
(message)BatchReadFeatureValuesOperationMetadata
(message)BatchReadFeatureValuesRequest
(message)BatchReadFeatureValuesRequest.EntityTypeSpec
(message)BatchReadFeatureValuesRequest.PassThroughField
(message)BatchReadFeatureValuesResponse
(message)BatchReadTensorboardTimeSeriesDataRequest
(message)BatchReadTensorboardTimeSeriesDataResponse
(message)BigQueryDestination
(message)BigQuerySource
(message)BleuInput
(message)BleuInstance
(message)BleuMetricValue
(message)BleuResults
(message)BleuSpec
(message)Blob
(message)BlurBaselineConfig
(message)BoolArray
(message)CancelBatchPredictionJobRequest
(message)CancelCustomJobRequest
(message)CancelHyperparameterTuningJobRequest
(message)CancelNasJobRequest
(message)CancelPipelineJobRequest
(message)CancelTrainingPipelineRequest
(message)CancelTuningJobRequest
(message)Candidate
(message)Candidate.FinishReason
(enum)CheckTrialEarlyStoppingStateMetatdata
(message)CheckTrialEarlyStoppingStateRequest
(message)CheckTrialEarlyStoppingStateResponse
(message)Citation
(message)CitationMetadata
(message)Claim
(message)ClientConnectionConfig
(message)CoherenceInput
(message)CoherenceInstance
(message)CoherenceResult
(message)CoherenceSpec
(message)CometInput
(message)CometInstance
(message)CometResult
(message)CometSpec
(message)CometSpec.CometVersion
(enum)CompleteTrialRequest
(message)CompletionStats
(message)ComputeTokensRequest
(message)ComputeTokensResponse
(message)ContainerRegistryDestination
(message)ContainerSpec
(message)Content
(message)Context
(message)CopyModelOperationMetadata
(message)CopyModelRequest
(message)CopyModelResponse
(message)CorpusStatus
(message)CorpusStatus.State
(enum)CorroborateContentRequest
(message)CorroborateContentRequest.Parameters
(message)CorroborateContentResponse
(message)CountTokensRequest
(message)CountTokensResponse
(message)CreateArtifactRequest
(message)CreateBatchPredictionJobRequest
(message)CreateContextRequest
(message)CreateCustomJobRequest
(message)CreateDatasetOperationMetadata
(message)CreateDatasetRequest
(message)CreateDatasetVersionOperationMetadata
(message)CreateDatasetVersionRequest
(message)CreateDeploymentResourcePoolOperationMetadata
(message)CreateDeploymentResourcePoolRequest
(message)CreateEndpointOperationMetadata
(message)CreateEndpointRequest
(message)CreateEntityTypeOperationMetadata
(message)CreateEntityTypeRequest
(message)CreateExecutionRequest
(message)CreateFeatureGroupOperationMetadata
(message)CreateFeatureGroupRequest
(message)CreateFeatureOnlineStoreOperationMetadata
(message)CreateFeatureOnlineStoreRequest
(message)CreateFeatureOperationMetadata
(message)CreateFeatureRequest
(message)CreateFeatureViewOperationMetadata
(message)CreateFeatureViewRequest
(message)CreateFeaturestoreOperationMetadata
(message)CreateFeaturestoreRequest
(message)CreateHyperparameterTuningJobRequest
(message)CreateIndexEndpointOperationMetadata
(message)CreateIndexEndpointRequest
(message)CreateIndexOperationMetadata
(message)CreateIndexRequest
(message)CreateMetadataSchemaRequest
(message)CreateMetadataStoreOperationMetadata
(message)CreateMetadataStoreRequest
(message)CreateModelDeploymentMonitoringJobRequest
(message)CreateNasJobRequest
(message)CreateNotebookExecutionJobOperationMetadata
(message)CreateNotebookExecutionJobRequest
(message)CreateNotebookRuntimeTemplateOperationMetadata
(message)CreateNotebookRuntimeTemplateRequest
(message)CreatePersistentResourceOperationMetadata
(message)CreatePersistentResourceRequest
(message)CreatePipelineJobRequest
(message)CreateRagCorpusOperationMetadata
(message)CreateRagCorpusRequest
(message)CreateRegistryFeatureOperationMetadata
(message)CreateScheduleRequest
(message)CreateSpecialistPoolOperationMetadata
(message)CreateSpecialistPoolRequest
(message)CreateStudyRequest
(message)CreateTensorboardExperimentRequest
(message)CreateTensorboardOperationMetadata
(message)CreateTensorboardRequest
(message)CreateTensorboardRunRequest
(message)CreateTensorboardTimeSeriesRequest
(message)CreateTrainingPipelineRequest
(message)CreateTrialRequest
(message)CreateTuningJobRequest
(message)CsvDestination
(message)CsvSource
(message)CustomJob
(message)CustomJobSpec
(message)DataItem
(message)DataItemView
(message)Dataset
(message)DatasetVersion
(message)DedicatedResources
(message)DeleteArtifactRequest
(message)DeleteBatchPredictionJobRequest
(message)DeleteContextRequest
(message)DeleteCustomJobRequest
(message)DeleteDatasetRequest
(message)DeleteDatasetVersionRequest
(message)DeleteDeploymentResourcePoolRequest
(message)DeleteEndpointRequest
(message)DeleteEntityTypeRequest
(message)DeleteExecutionRequest
(message)DeleteFeatureGroupRequest
(message)DeleteFeatureOnlineStoreRequest
(message)DeleteFeatureRequest
(message)DeleteFeatureValuesOperationMetadata
(message)DeleteFeatureValuesRequest
(message)DeleteFeatureValuesRequest.SelectEntity
(message)DeleteFeatureValuesRequest.SelectTimeRangeAndFeature
(message)DeleteFeatureValuesResponse
(message)DeleteFeatureValuesResponse.SelectEntity
(message)DeleteFeatureValuesResponse.SelectTimeRangeAndFeature
(message)DeleteFeatureViewRequest
(message)DeleteFeaturestoreRequest
(message)DeleteHyperparameterTuningJobRequest
(message)DeleteIndexEndpointRequest
(message)DeleteIndexRequest
(message)DeleteMetadataStoreOperationMetadata
(message)DeleteMetadataStoreRequest
(message)DeleteModelDeploymentMonitoringJobRequest
(message)DeleteModelRequest
(message)DeleteModelVersionRequest
(message)DeleteNasJobRequest
(message)DeleteNotebookExecutionJobRequest
(message)DeleteNotebookRuntimeRequest
(message)DeleteNotebookRuntimeTemplateRequest
(message)DeleteOperationMetadata
(message)DeletePersistentResourceRequest
(message)DeletePipelineJobRequest
(message)DeleteRagCorpusRequest
(message)DeleteRagFileRequest
(message)DeleteSavedQueryRequest
(message)DeleteScheduleRequest
(message)DeleteSpecialistPoolRequest
(message)DeleteStudyRequest
(message)DeleteTensorboardExperimentRequest
(message)DeleteTensorboardRequest
(message)DeleteTensorboardRunRequest
(message)DeleteTensorboardTimeSeriesRequest
(message)DeleteTrainingPipelineRequest
(message)DeleteTrialRequest
(message)DeployIndexOperationMetadata
(message)DeployIndexRequest
(message)DeployIndexResponse
(message)DeployModelOperationMetadata
(message)DeployModelRequest
(message)DeployModelResponse
(message)DeployedIndex
(message)DeployedIndexAuthConfig
(message)DeployedIndexAuthConfig.AuthProvider
(message)DeployedIndexRef
(message)DeployedModel
(message)DeployedModel.Status
(message)DeployedModelRef
(message)DeploymentResourcePool
(message)DestinationFeatureSetting
(message)DirectPredictRequest
(message)DirectPredictResponse
(message)DirectRawPredictRequest
(message)DirectRawPredictResponse
(message)DirectUploadSource
(message)DiskSpec
(message)DoubleArray
(message)DynamicRetrievalConfig
(message)DynamicRetrievalConfig.Mode
(enum)EncryptionSpec
(message)Endpoint
(message)EntityIdSelector
(message)EntityType
(message)EnvVar
(message)ErrorAnalysisAnnotation
(message)ErrorAnalysisAnnotation.AttributedItem
(message)ErrorAnalysisAnnotation.QueryType
(enum)EvaluateInstancesRequest
(message)EvaluateInstancesResponse
(message)EvaluatedAnnotation
(message)EvaluatedAnnotation.EvaluatedAnnotationType
(enum)EvaluatedAnnotationExplanation
(message)Event
(message)Event.Type
(enum)ExactMatchInput
(message)ExactMatchInstance
(message)ExactMatchMetricValue
(message)ExactMatchResults
(message)ExactMatchSpec
(message)Examples
(message)Examples.ExampleGcsSource
(message)Examples.ExampleGcsSource.DataFormat
(enum)ExamplesOverride
(message)ExamplesOverride.DataFormat
(enum)ExamplesRestrictionsNamespace
(message)Execution
(message)Execution.State
(enum)ExplainRequest
(message)ExplainResponse
(message)Explanation
(message)ExplanationMetadata
(message)ExplanationMetadata.InputMetadata
(message)ExplanationMetadata.InputMetadata.Encoding
(enum)ExplanationMetadata.InputMetadata.FeatureValueDomain
(message)ExplanationMetadata.InputMetadata.Visualization
(message)ExplanationMetadata.InputMetadata.Visualization.ColorMap
(enum)ExplanationMetadata.InputMetadata.Visualization.OverlayType
(enum)ExplanationMetadata.InputMetadata.Visualization.Polarity
(enum)ExplanationMetadata.InputMetadata.Visualization.Type
(enum)ExplanationMetadata.OutputMetadata
(message)ExplanationMetadataOverride
(message)ExplanationMetadataOverride.InputMetadataOverride
(message)ExplanationParameters
(message)ExplanationSpec
(message)ExplanationSpecOverride
(message)ExportDataConfig
(message)ExportDataConfig.ExportUse
(enum)ExportDataOperationMetadata
(message)ExportDataRequest
(message)ExportDataResponse
(message)ExportFeatureValuesOperationMetadata
(message)ExportFeatureValuesRequest
(message)ExportFeatureValuesRequest.FullExport
(message)ExportFeatureValuesRequest.SnapshotExport
(message)ExportFeatureValuesResponse
(message)ExportFilterSplit
(message)ExportFractionSplit
(message)ExportModelOperationMetadata
(message)ExportModelOperationMetadata.OutputInfo
(message)ExportModelRequest
(message)ExportModelRequest.OutputConfig
(message)ExportModelResponse
(message)ExportTensorboardTimeSeriesDataRequest
(message)ExportTensorboardTimeSeriesDataResponse
(message)Fact
(message)FasterDeploymentConfig
(message)Feature
(message)Feature.MonitoringStatsAnomaly
(message)Feature.MonitoringStatsAnomaly.Objective
(enum)Feature.ValueType
(enum)FeatureGroup
(message)FeatureGroup.BigQuery
(message)FeatureGroup.BigQuery.TimeSeries
(message)FeatureNoiseSigma
(message)FeatureNoiseSigma.NoiseSigmaForFeature
(message)FeatureOnlineStore
(message)FeatureOnlineStore.Bigtable
(message)FeatureOnlineStore.Bigtable.AutoScaling
(message)FeatureOnlineStore.DedicatedServingEndpoint
(message)FeatureOnlineStore.Optimized
(message)FeatureOnlineStore.State
(enum)FeatureSelector
(message)FeatureStatsAnomaly
(message)FeatureValue
(message)FeatureValue.Metadata
(message)FeatureValueDestination
(message)FeatureValueList
(message)FeatureView
(message)FeatureView.BigQuerySource
(message)FeatureView.FeatureRegistrySource
(message)FeatureView.FeatureRegistrySource.FeatureGroup
(message)FeatureView.IndexConfig
(message)FeatureView.IndexConfig.BruteForceConfig
(message)FeatureView.IndexConfig.DistanceMeasureType
(enum)FeatureView.IndexConfig.TreeAHConfig
(message)FeatureView.SyncConfig
(message)FeatureView.VertexRagSource
(message)FeatureViewDataFormat
(enum)FeatureViewDataKey
(message)FeatureViewDataKey.CompositeKey
(message)FeatureViewSync
(message)FeatureViewSync.SyncSummary
(message)Featurestore
(message)Featurestore.OnlineServingConfig
(message)Featurestore.OnlineServingConfig.Scaling
(message)Featurestore.State
(enum)FeaturestoreMonitoringConfig
(message)FeaturestoreMonitoringConfig.ImportFeaturesAnalysis
(message)FeaturestoreMonitoringConfig.ImportFeaturesAnalysis.Baseline
(enum)FeaturestoreMonitoringConfig.ImportFeaturesAnalysis.State
(enum)FeaturestoreMonitoringConfig.SnapshotAnalysis
(message)FeaturestoreMonitoringConfig.ThresholdConfig
(message)FetchFeatureValuesRequest
(message)FetchFeatureValuesResponse
(message)FetchFeatureValuesResponse.FeatureNameValuePairList
(message)FetchFeatureValuesResponse.FeatureNameValuePairList.FeatureNameValuePair
(message)FileData
(message)FileStatus
(message)FileStatus.State
(enum)FilterSplit
(message)FluencyInput
(message)FluencyInstance
(message)FluencyResult
(message)FluencySpec
(message)FractionSplit
(message)FulfillmentInput
(message)FulfillmentInstance
(message)FulfillmentResult
(message)FulfillmentSpec
(message)FunctionCall
(message)FunctionCallingConfig
(message)FunctionCallingConfig.Mode
(enum)FunctionDeclaration
(message)FunctionResponse
(message)GcsDestination
(message)GcsSource
(message)GenerateContentRequest
(message)GenerateContentResponse
(message)GenerateContentResponse.PromptFeedback
(message)GenerateContentResponse.PromptFeedback.BlockedReason
(enum)GenerateContentResponse.UsageMetadata
(message)GenerationConfig
(message)GenerationConfig.RoutingConfig
(message)GenerationConfig.RoutingConfig.AutoRoutingMode
(message)GenerationConfig.RoutingConfig.AutoRoutingMode.ModelRoutingPreference
(enum)GenerationConfig.RoutingConfig.ManualRoutingMode
(message)GenericOperationMetadata
(message)GenieSource
(message)GetAnnotationSpecRequest
(message)GetArtifactRequest
(message)GetBatchPredictionJobRequest
(message)GetContextRequest
(message)GetCustomJobRequest
(message)GetDatasetRequest
(message)GetDatasetVersionRequest
(message)GetDeploymentResourcePoolRequest
(message)GetEndpointRequest
(message)GetEntityTypeRequest
(message)GetExecutionRequest
(message)GetFeatureGroupRequest
(message)GetFeatureOnlineStoreRequest
(message)GetFeatureRequest
(message)GetFeatureViewRequest
(message)GetFeatureViewSyncRequest
(message)GetFeaturestoreRequest
(message)GetHyperparameterTuningJobRequest
(message)GetIndexEndpointRequest
(message)GetIndexRequest
(message)GetMetadataSchemaRequest
(message)GetMetadataStoreRequest
(message)GetModelDeploymentMonitoringJobRequest
(message)GetModelEvaluationRequest
(message)GetModelEvaluationSliceRequest
(message)GetModelRequest
(message)GetNasJobRequest
(message)GetNasTrialDetailRequest
(message)GetNotebookExecutionJobRequest
(message)GetNotebookRuntimeRequest
(message)GetNotebookRuntimeTemplateRequest
(message)GetPersistentResourceRequest
(message)GetPipelineJobRequest
(message)GetPublisherModelRequest
(message)GetRagCorpusRequest
(message)GetRagFileRequest
(message)GetScheduleRequest
(message)GetSpecialistPoolRequest
(message)GetStudyRequest
(message)GetTensorboardExperimentRequest
(message)GetTensorboardRequest
(message)GetTensorboardRunRequest
(message)GetTensorboardTimeSeriesRequest
(message)GetTrainingPipelineRequest
(message)GetTrialRequest
(message)GetTuningJobRequest
(message)GoogleDriveSource
(message)GoogleDriveSource.ResourceId
(message)GoogleDriveSource.ResourceId.ResourceType
(enum)GoogleSearchRetrieval
(message)GroundednessInput
(message)GroundednessInstance
(message)GroundednessResult
(message)GroundednessSpec
(message)GroundingChunk
(message)GroundingChunk.RetrievedContext
(message)GroundingChunk.Web
(message)GroundingMetadata
(message)GroundingSupport
(message)HarmCategory
(enum)HyperparameterTuningJob
(message)IdMatcher
(message)ImportDataConfig
(message)ImportDataOperationMetadata
(message)ImportDataRequest
(message)ImportDataResponse
(message)ImportFeatureValuesOperationMetadata
(message)ImportFeatureValuesRequest
(message)ImportFeatureValuesRequest.FeatureSpec
(message)ImportFeatureValuesResponse
(message)ImportModelEvaluationRequest
(message)ImportRagFilesConfig
(message)ImportRagFilesOperationMetadata
(message)ImportRagFilesRequest
(message)ImportRagFilesResponse
(message)Index
(message)Index.IndexUpdateMethod
(enum)IndexDatapoint
(message)IndexDatapoint.CrowdingTag
(message)IndexDatapoint.NumericRestriction
(message)IndexDatapoint.NumericRestriction.Operator
(enum)IndexDatapoint.Restriction
(message)IndexDatapoint.SparseEmbedding
(message)IndexEndpoint
(message)IndexPrivateEndpoints
(message)IndexStats
(message)InputDataConfig
(message)Int64Array
(message)IntegratedGradientsAttribution
(message)JiraSource
(message)JiraSource.JiraQueries
(message)JobState
(enum)LargeModelReference
(message)LineageSubgraph
(message)ListAnnotationsRequest
(message)ListAnnotationsResponse
(message)ListArtifactsRequest
(message)ListArtifactsResponse
(message)ListBatchPredictionJobsRequest
(message)ListBatchPredictionJobsResponse
(message)ListContextsRequest
(message)ListContextsResponse
(message)ListCustomJobsRequest
(message)ListCustomJobsResponse
(message)ListDataItemsRequest
(message)ListDataItemsResponse
(message)ListDatasetVersionsRequest
(message)ListDatasetVersionsResponse
(message)ListDatasetsRequest
(message)ListDatasetsResponse
(message)ListDeploymentResourcePoolsRequest
(message)ListDeploymentResourcePoolsResponse
(message)ListEndpointsRequest
(message)ListEndpointsResponse
(message)ListEntityTypesRequest
(message)ListEntityTypesResponse
(message)ListExecutionsRequest
(message)ListExecutionsResponse
(message)ListFeatureGroupsRequest
(message)ListFeatureGroupsResponse
(message)ListFeatureOnlineStoresRequest
(message)ListFeatureOnlineStoresResponse
(message)ListFeatureViewSyncsRequest
(message)ListFeatureViewSyncsResponse
(message)ListFeatureViewsRequest
(message)ListFeatureViewsResponse
(message)ListFeaturesRequest
(message)ListFeaturesResponse
(message)ListFeaturestoresRequest
(message)ListFeaturestoresResponse
(message)ListHyperparameterTuningJobsRequest
(message)ListHyperparameterTuningJobsResponse
(message)ListIndexEndpointsRequest
(message)ListIndexEndpointsResponse
(message)ListIndexesRequest
(message)ListIndexesResponse
(message)ListMetadataSchemasRequest
(message)ListMetadataSchemasResponse
(message)ListMetadataStoresRequest
(message)ListMetadataStoresResponse
(message)ListModelDeploymentMonitoringJobsRequest
(message)ListModelDeploymentMonitoringJobsResponse
(message)ListModelEvaluationSlicesRequest
(message)ListModelEvaluationSlicesResponse
(message)ListModelEvaluationsRequest
(message)ListModelEvaluationsResponse
(message)ListModelVersionsRequest
(message)ListModelVersionsResponse
(message)ListModelsRequest
(message)ListModelsResponse
(message)ListNasJobsRequest
(message)ListNasJobsResponse
(message)ListNasTrialDetailsRequest
(message)ListNasTrialDetailsResponse
(message)ListNotebookExecutionJobsRequest
(message)ListNotebookExecutionJobsResponse
(message)ListNotebookRuntimeTemplatesRequest
(message)ListNotebookRuntimeTemplatesResponse
(message)ListNotebookRuntimesRequest
(message)ListNotebookRuntimesResponse
(message)ListOptimalTrialsRequest
(message)ListOptimalTrialsResponse
(message)ListPersistentResourcesRequest
(message)ListPersistentResourcesResponse
(message)ListPipelineJobsRequest
(message)ListPipelineJobsResponse
(message)ListRagCorporaRequest
(message)ListRagCorporaResponse
(message)ListRagFilesRequest
(message)ListRagFilesResponse
(message)ListSavedQueriesRequest
(message)ListSavedQueriesResponse
(message)ListSchedulesRequest
(message)ListSchedulesResponse
(message)ListSpecialistPoolsRequest
(message)ListSpecialistPoolsResponse
(message)ListStudiesRequest
(message)ListStudiesResponse
(message)ListTensorboardExperimentsRequest
(message)ListTensorboardExperimentsResponse
(message)ListTensorboardRunsRequest
(message)ListTensorboardRunsResponse
(message)ListTensorboardTimeSeriesRequest
(message)ListTensorboardTimeSeriesResponse
(message)ListTensorboardsRequest
(message)ListTensorboardsResponse
(message)ListTrainingPipelinesRequest
(message)ListTrainingPipelinesResponse
(message)ListTrialsRequest
(message)ListTrialsResponse
(message)ListTuningJobsRequest
(message)ListTuningJobsResponse
(message)LogprobsResult
(message)LogprobsResult.Candidate
(message)LogprobsResult.TopCandidates
(message)LookupStudyRequest
(message)MachineSpec
(message)ManualBatchTuningParameters
(message)Measurement
(message)Measurement.Metric
(message)MergeVersionAliasesRequest
(message)MetadataSchema
(message)MetadataSchema.MetadataSchemaType
(enum)MetadataStore
(message)MetadataStore.DataplexConfig
(message)MetadataStore.MetadataStoreState
(message)MigratableResource
(message)MigratableResource.AutomlDataset
(message)MigratableResource.AutomlModel
(message)MigratableResource.DataLabelingDataset
(message)MigratableResource.DataLabelingDataset.DataLabelingAnnotatedDataset
(message)MigratableResource.MlEngineModelVersion
(message)MigrateResourceRequest
(message)MigrateResourceRequest.MigrateAutomlDatasetConfig
(message)MigrateResourceRequest.MigrateAutomlModelConfig
(message)MigrateResourceRequest.MigrateDataLabelingDatasetConfig
(message)MigrateResourceRequest.MigrateDataLabelingDatasetConfig.MigrateDataLabelingAnnotatedDatasetConfig
(message)MigrateResourceRequest.MigrateMlEngineModelVersionConfig
(message)MigrateResourceResponse
(message)Model
(message)Model.BaseModelSource
(message)Model.DataStats
(message)Model.DeploymentResourcesType
(enum)Model.ExportFormat
(message)Model.ExportFormat.ExportableContent
(enum)Model.OriginalModelInfo
(message)ModelContainerSpec
(message)ModelDeploymentMonitoringBigQueryTable
(message)ModelDeploymentMonitoringBigQueryTable.LogSource
(enum)ModelDeploymentMonitoringBigQueryTable.LogType
(enum)ModelDeploymentMonitoringJob
(message)ModelDeploymentMonitoringJob.LatestMonitoringPipelineMetadata
(message)ModelDeploymentMonitoringJob.MonitoringScheduleState
(enum)ModelDeploymentMonitoringObjectiveConfig
(message)ModelDeploymentMonitoringObjectiveType
(enum)ModelDeploymentMonitoringScheduleConfig
(message)ModelEvaluation
(message)ModelEvaluation.ModelEvaluationExplanationSpec
(message)ModelEvaluationSlice
(message)ModelEvaluationSlice.Slice
(message)ModelEvaluationSlice.Slice.SliceSpec
(message)ModelEvaluationSlice.Slice.SliceSpec.Range
(message)ModelEvaluationSlice.Slice.SliceSpec.SliceConfig
(message)ModelEvaluationSlice.Slice.SliceSpec.Value
(message)ModelExplanation
(message)ModelGardenSource
(message)ModelMonitoringAlertConfig
(message)ModelMonitoringAlertConfig.EmailAlertConfig
(message)ModelMonitoringObjectiveConfig
(message)ModelMonitoringObjectiveConfig.ExplanationConfig
(message)ModelMonitoringObjectiveConfig.ExplanationConfig.ExplanationBaseline
(message)ModelMonitoringObjectiveConfig.ExplanationConfig.ExplanationBaseline.PredictionFormat
(enum)ModelMonitoringObjectiveConfig.PredictionDriftDetectionConfig
(message)ModelMonitoringObjectiveConfig.TrainingDataset
(message)ModelMonitoringObjectiveConfig.TrainingPredictionSkewDetectionConfig
(message)ModelMonitoringStatsAnomalies
(message)ModelMonitoringStatsAnomalies.FeatureHistoricStatsAnomalies
(message)ModelSourceInfo
(message)ModelSourceInfo.ModelSourceType
(enum)MutateDeployedIndexOperationMetadata
(message)MutateDeployedIndexRequest
(message)MutateDeployedIndexResponse
(message)MutateDeployedModelOperationMetadata
(message)MutateDeployedModelRequest
(message)MutateDeployedModelResponse
(message)NasJob
(message)NasJobOutput
(message)NasJobOutput.MultiTrialJobOutput
(message)NasJobSpec
(message)NasJobSpec.MultiTrialAlgorithmSpec
(message)NasJobSpec.MultiTrialAlgorithmSpec.MetricSpec
(message)NasJobSpec.MultiTrialAlgorithmSpec.MetricSpec.GoalType
(enum)NasJobSpec.MultiTrialAlgorithmSpec.MultiTrialAlgorithm
(enum)NasJobSpec.MultiTrialAlgorithmSpec.SearchTrialSpec
(message)NasJobSpec.MultiTrialAlgorithmSpec.TrainTrialSpec
(message)NasTrial
(message)NasTrial.State
(enum)NasTrialDetail
(message)NearestNeighborQuery
(message)NearestNeighborQuery.Embedding
(message)NearestNeighborQuery.NumericFilter
(message)NearestNeighborQuery.NumericFilter.Operator
(enum)NearestNeighborQuery.Parameters
(message)NearestNeighborQuery.StringFilter
(message)NearestNeighborSearchOperationMetadata
(message)NearestNeighborSearchOperationMetadata.ContentValidationStats
(message)NearestNeighborSearchOperationMetadata.RecordError
(message)NearestNeighborSearchOperationMetadata.RecordError.RecordErrorType
(enum)NearestNeighbors
(message)NearestNeighbors.Neighbor
(message)Neighbor
(message)NetworkSpec
(message)NfsMount
(message)NotebookEucConfig
(message)NotebookExecutionJob
(message)NotebookExecutionJob.CustomEnvironmentSpec
(message)NotebookExecutionJob.DataformRepositorySource
(message)NotebookExecutionJob.DirectNotebookSource
(message)NotebookExecutionJob.GcsNotebookSource
(message)NotebookExecutionJob.WorkbenchRuntime
(message)NotebookExecutionJobView
(enum)NotebookIdleShutdownConfig
(message)NotebookRuntime
(message)NotebookRuntime.HealthState
(enum)NotebookRuntime.RuntimeState
(enum)NotebookRuntimeTemplate
(message)NotebookRuntimeTemplateRef
(message)NotebookRuntimeType
(enum)PSCAutomationConfig
(message)PairwiseChoice
(enum)PairwiseMetricInput
(message)PairwiseMetricInstance
(message)PairwiseMetricResult
(message)PairwiseMetricSpec
(message)PairwiseQuestionAnsweringQualityInput
(message)PairwiseQuestionAnsweringQualityInstance
(message)PairwiseQuestionAnsweringQualityResult
(message)PairwiseQuestionAnsweringQualitySpec
(message)PairwiseSummarizationQualityInput
(message)PairwiseSummarizationQualityInstance
(message)PairwiseSummarizationQualityResult
(message)PairwiseSummarizationQualitySpec
(message)Part
(message)PauseModelDeploymentMonitoringJobRequest
(message)PauseScheduleRequest
(message)PersistentDiskSpec
(message)PersistentResource
(message)PersistentResource.State
(enum)PipelineFailurePolicy
(enum)PipelineJob
(message)PipelineJob.RuntimeConfig
(message)PipelineJob.RuntimeConfig.InputArtifact
(message)PipelineJobDetail
(message)PipelineState
(enum)PipelineTaskDetail
(message)PipelineTaskDetail.ArtifactList
(message)PipelineTaskDetail.PipelineTaskStatus
(message)PipelineTaskDetail.State
(enum)PipelineTaskExecutorDetail
(message)PipelineTaskExecutorDetail.ContainerDetail
(message)PipelineTaskExecutorDetail.CustomJobDetail
(message)PipelineTemplateMetadata
(message)PointwiseMetricInput
(message)PointwiseMetricInstance
(message)PointwiseMetricResult
(message)PointwiseMetricSpec
(message)Port
(message)PredefinedSplit
(message)PredictLongRunningRequest
(message)PredictRequest
(message)PredictRequestResponseLoggingConfig
(message)PredictResponse
(message)PredictSchemata
(message)Presets
(message)Presets.Modality
(enum)Presets.Query
(enum)PrivateEndpoints
(message)PrivateServiceConnectConfig
(message)Probe
(message)Probe.ExecAction
(message)PscAutomatedEndpoints
(message)PublisherModel
(message)PublisherModel.CallToAction
(message)PublisherModel.CallToAction.Deploy
(message)PublisherModel.CallToAction.Deploy.DeployMetadata
(message)PublisherModel.CallToAction.DeployGke
(message)PublisherModel.CallToAction.OpenFineTuningPipelines
(message)PublisherModel.CallToAction.OpenNotebooks
(message)PublisherModel.CallToAction.RegionalResourceReferences
(message)PublisherModel.CallToAction.ViewRestApi
(message)PublisherModel.Documentation
(message)PublisherModel.LaunchStage
(enum)PublisherModel.OpenSourceCategory
(enum)PublisherModel.ResourceReference
(message)PublisherModel.VersionState
(enum)PublisherModelView
(enum)PurgeArtifactsMetadata
(message)PurgeArtifactsRequest
(message)PurgeArtifactsResponse
(message)PurgeContextsMetadata
(message)PurgeContextsRequest
(message)PurgeContextsResponse
(message)PurgeExecutionsMetadata
(message)PurgeExecutionsRequest
(message)PurgeExecutionsResponse
(message)PythonPackageSpec
(message)QueryArtifactLineageSubgraphRequest
(message)QueryContextLineageSubgraphRequest
(message)QueryDeployedModelsRequest
(message)QueryDeployedModelsResponse
(message)QueryExecutionInputsAndOutputsRequest
(message)QuestionAnsweringCorrectnessInput
(message)QuestionAnsweringCorrectnessInstance
(message)QuestionAnsweringCorrectnessResult
(message)QuestionAnsweringCorrectnessSpec
(message)QuestionAnsweringHelpfulnessInput
(message)QuestionAnsweringHelpfulnessInstance
(message)QuestionAnsweringHelpfulnessResult
(message)QuestionAnsweringHelpfulnessSpec
(message)QuestionAnsweringQualityInput
(message)QuestionAnsweringQualityInstance
(message)QuestionAnsweringQualityResult
(message)QuestionAnsweringQualitySpec
(message)QuestionAnsweringRelevanceInput
(message)QuestionAnsweringRelevanceInstance
(message)QuestionAnsweringRelevanceResult
(message)QuestionAnsweringRelevanceSpec
(message)RagContexts
(message)RagContexts.Context
(message)RagCorpus
(message)RagEmbeddingModelConfig
(message)RagEmbeddingModelConfig.VertexPredictionEndpoint
(message)RagFile
(message)RagFileChunkingConfig
(message)RagFileChunkingConfig.FixedLengthChunking
(message)RagFileTransformationConfig
(message)RagQuery
(message)RagRetrievalConfig
(message)RagRetrievalConfig.Filter
(message)RagVectorDbConfig
(message)RagVectorDbConfig.Pinecone
(message)RagVectorDbConfig.RagManagedDb
(message)RagVectorDbConfig.VertexVectorSearch
(message)RawPredictRequest
(message)RayLogsSpec
(message)RayMetricSpec
(message)RaySpec
(message)ReadFeatureValuesRequest
(message)ReadFeatureValuesResponse
(message)ReadFeatureValuesResponse.EntityView
(message)ReadFeatureValuesResponse.EntityView.Data
(message)ReadFeatureValuesResponse.FeatureDescriptor
(message)ReadFeatureValuesResponse.Header
(message)ReadTensorboardBlobDataRequest
(message)ReadTensorboardBlobDataResponse
(message)ReadTensorboardSizeRequest
(message)ReadTensorboardSizeResponse
(message)ReadTensorboardTimeSeriesDataRequest
(message)ReadTensorboardTimeSeriesDataResponse
(message)ReadTensorboardUsageRequest
(message)ReadTensorboardUsageResponse
(message)ReadTensorboardUsageResponse.PerMonthUsageData
(message)ReadTensorboardUsageResponse.PerUserUsageData
(message)RebaseTunedModelOperationMetadata
(message)RebaseTunedModelRequest
(message)RebootPersistentResourceOperationMetadata
(message)RebootPersistentResourceRequest
(message)RemoveContextChildrenRequest
(message)RemoveContextChildrenResponse
(message)RemoveDatapointsRequest
(message)RemoveDatapointsResponse
(message)ReservationAffinity
(message)ReservationAffinity.Type
(enum)ResourcePool
(message)ResourcePool.AutoscalingSpec
(message)ResourceRuntime
(message)ResourceRuntimeSpec
(message)ResourcesConsumed
(message)RestoreDatasetVersionOperationMetadata
(message)RestoreDatasetVersionRequest
(message)ResumeModelDeploymentMonitoringJobRequest
(message)ResumeScheduleRequest
(message)Retrieval
(message)RetrievalMetadata
(message)RetrieveContextsRequest
(message)RetrieveContextsRequest.VertexRagStore
(message)RetrieveContextsRequest.VertexRagStore.RagResource
(message)RetrieveContextsResponse
(message)RougeInput
(message)RougeInstance
(message)RougeMetricValue
(message)RougeResults
(message)RougeSpec
(message)SafetyInput
(message)SafetyInstance
(message)SafetyRating
(message)SafetyRating.HarmProbability
(enum)SafetyRating.HarmSeverity
(enum)SafetyResult
(message)SafetySetting
(message)SafetySetting.HarmBlockMethod
(enum)SafetySetting.HarmBlockThreshold
(enum)SafetySpec
(message)SampledShapleyAttribution
(message)SamplingStrategy
(message)SamplingStrategy.RandomSampleConfig
(message)SavedQuery
(message)Scalar
(message)Schedule
(message)Schedule.RunResponse
(message)Schedule.State
(enum)Scheduling
(message)Scheduling.Strategy
(enum)Schema
(message)SearchDataItemsRequest
(message)SearchDataItemsRequest.OrderByAnnotation
(message)SearchDataItemsResponse
(message)SearchEntryPoint
(message)SearchFeaturesRequest
(message)SearchFeaturesResponse
(message)SearchMigratableResourcesRequest
(message)SearchMigratableResourcesResponse
(message)SearchModelDeploymentMonitoringStatsAnomaliesRequest
(message)SearchModelDeploymentMonitoringStatsAnomaliesRequest.StatsAnomaliesObjective
(message)SearchModelDeploymentMonitoringStatsAnomaliesResponse
(message)SearchNearestEntitiesRequest
(message)SearchNearestEntitiesResponse
(message)Segment
(message)ServiceAccountSpec
(message)SharePointSources
(message)SharePointSources.SharePointSource
(message)ShieldedVmConfig
(message)SlackSource
(message)SlackSource.SlackChannels
(message)SlackSource.SlackChannels.SlackChannel
(message)SmoothGradConfig
(message)SpecialistPool
(message)StartNotebookRuntimeOperationMetadata
(message)StartNotebookRuntimeRequest
(message)StartNotebookRuntimeResponse
(message)StopNotebookRuntimeOperationMetadata
(message)StopNotebookRuntimeRequest
(message)StopNotebookRuntimeResponse
(message)StopTrialRequest
(message)StratifiedSplit
(message)StreamDirectPredictRequest
(message)StreamDirectPredictResponse
(message)StreamDirectRawPredictRequest
(message)StreamDirectRawPredictResponse
(message)StreamRawPredictRequest
(message)StreamingPredictRequest
(message)StreamingPredictResponse
(message)StreamingRawPredictRequest
(message)StreamingRawPredictResponse
(message)StreamingReadFeatureValuesRequest
(message)StringArray
(message)StructFieldValue
(message)StructValue
(message)Study
(message)Study.State
(enum)StudySpec
(message)StudySpec.Algorithm
(enum)StudySpec.ConvexAutomatedStoppingSpec
(message)StudySpec.DecayCurveAutomatedStoppingSpec
(message)StudySpec.MeasurementSelectionType
(enum)StudySpec.MedianAutomatedStoppingSpec
(message)StudySpec.MetricSpec
(message)StudySpec.MetricSpec.GoalType
(enum)StudySpec.MetricSpec.SafetyMetricConfig
(message)StudySpec.ObservationNoise
(enum)StudySpec.ParameterSpec
(message)StudySpec.ParameterSpec.CategoricalValueSpec
(message)StudySpec.ParameterSpec.ConditionalParameterSpec
(message)StudySpec.ParameterSpec.ConditionalParameterSpec.CategoricalValueCondition
(message)StudySpec.ParameterSpec.ConditionalParameterSpec.DiscreteValueCondition
(message)StudySpec.ParameterSpec.ConditionalParameterSpec.IntValueCondition
(message)StudySpec.ParameterSpec.DiscreteValueSpec
(message)StudySpec.ParameterSpec.DoubleValueSpec
(message)StudySpec.ParameterSpec.IntegerValueSpec
(message)StudySpec.ParameterSpec.ScaleType
(enum)StudySpec.StudyStoppingConfig
(message)StudyTimeConstraint
(message)SuggestTrialsMetadata
(message)SuggestTrialsRequest
(message)SuggestTrialsResponse
(message)SummarizationHelpfulnessInput
(message)SummarizationHelpfulnessInstance
(message)SummarizationHelpfulnessResult
(message)SummarizationHelpfulnessSpec
(message)SummarizationQualityInput
(message)SummarizationQualityInstance
(message)SummarizationQualityResult
(message)SummarizationQualitySpec
(message)SummarizationVerbosityInput
(message)SummarizationVerbosityInstance
(message)SummarizationVerbosityResult
(message)SummarizationVerbositySpec
(message)SupervisedHyperParameters
(message)SupervisedHyperParameters.AdapterSize
(enum)SupervisedTuningDataStats
(message)SupervisedTuningDatasetDistribution
(message)SupervisedTuningDatasetDistribution.DatasetBucket
(message)SupervisedTuningSpec
(message)SyncFeatureViewRequest
(message)SyncFeatureViewResponse
(message)TFRecordDestination
(message)Tensor
(message)Tensor.DataType
(enum)Tensorboard
(message)TensorboardBlob
(message)TensorboardBlobSequence
(message)TensorboardExperiment
(message)TensorboardRun
(message)TensorboardTensor
(message)TensorboardTimeSeries
(message)TensorboardTimeSeries.Metadata
(message)TensorboardTimeSeries.ValueType
(enum)ThresholdConfig
(message)TimeSeriesData
(message)TimeSeriesDataPoint
(message)TimestampSplit
(message)TokensInfo
(message)Tool
(message)ToolCall
(message)ToolCallValidInput
(message)ToolCallValidInstance
(message)ToolCallValidMetricValue
(message)ToolCallValidResults
(message)ToolCallValidSpec
(message)ToolConfig
(message)ToolNameMatchInput
(message)ToolNameMatchInstance
(message)ToolNameMatchMetricValue
(message)ToolNameMatchResults
(message)ToolNameMatchSpec
(message)ToolParameterKVMatchInput
(message)ToolParameterKVMatchInstance
(message)ToolParameterKVMatchMetricValue
(message)ToolParameterKVMatchResults
(message)ToolParameterKVMatchSpec
(message)ToolParameterKeyMatchInput
(message)ToolParameterKeyMatchInstance
(message)ToolParameterKeyMatchMetricValue
(message)ToolParameterKeyMatchResults
(message)ToolParameterKeyMatchSpec
(message)TrainingPipeline
(message)Trajectory
(message)TrajectoryAnyOrderMatchInput
(message)TrajectoryAnyOrderMatchInstance
(message)TrajectoryAnyOrderMatchMetricValue
(message)TrajectoryAnyOrderMatchResults
(message)TrajectoryAnyOrderMatchSpec
(message)TrajectoryExactMatchInput
(message)TrajectoryExactMatchInstance
(message)TrajectoryExactMatchMetricValue
(message)TrajectoryExactMatchResults
(message)TrajectoryExactMatchSpec
(message)TrajectoryInOrderMatchInput
(message)TrajectoryInOrderMatchInstance
(message)TrajectoryInOrderMatchMetricValue
(message)TrajectoryInOrderMatchResults
(message)TrajectoryInOrderMatchSpec
(message)TrajectoryPrecisionInput
(message)TrajectoryPrecisionInstance
(message)TrajectoryPrecisionMetricValue
(message)TrajectoryPrecisionResults
(message)TrajectoryPrecisionSpec
(message)TrajectoryRecallInput
(message)TrajectoryRecallInstance
(message)TrajectoryRecallMetricValue
(message)TrajectoryRecallResults
(message)TrajectoryRecallSpec
(message)TrajectorySingleToolUseInput
(message)TrajectorySingleToolUseInstance
(message)TrajectorySingleToolUseMetricValue
(message)TrajectorySingleToolUseResults
(message)TrajectorySingleToolUseSpec
(message)Trial
(message)Trial.Parameter
(message)Trial.State
(enum)TrialContext
(message)TunedModel
(message)TunedModelRef
(message)TuningDataStats
(message)TuningJob
(message)Type
(enum)UndeployIndexOperationMetadata
(message)UndeployIndexRequest
(message)UndeployIndexResponse
(message)UndeployModelOperationMetadata
(message)UndeployModelRequest
(message)UndeployModelResponse
(message)UnmanagedContainerModel
(message)UpdateArtifactRequest
(message)UpdateContextRequest
(message)UpdateDatasetRequest
(message)UpdateDatasetVersionRequest
(message)UpdateDeploymentResourcePoolOperationMetadata
(message)UpdateDeploymentResourcePoolRequest
(message)UpdateEndpointLongRunningRequest
(message)UpdateEndpointOperationMetadata
(message)UpdateEndpointRequest
(message)UpdateEntityTypeRequest
(message)UpdateExecutionRequest
(message)UpdateExplanationDatasetOperationMetadata
(message)UpdateExplanationDatasetRequest
(message)UpdateExplanationDatasetResponse
(message)UpdateFeatureGroupOperationMetadata
(message)UpdateFeatureGroupRequest
(message)UpdateFeatureOnlineStoreOperationMetadata
(message)UpdateFeatureOnlineStoreRequest
(message)UpdateFeatureOperationMetadata
(message)UpdateFeatureRequest
(message)UpdateFeatureViewOperationMetadata
(message)UpdateFeatureViewRequest
(message)UpdateFeaturestoreOperationMetadata
(message)UpdateFeaturestoreRequest
(message)UpdateIndexEndpointRequest
(message)UpdateIndexOperationMetadata
(message)UpdateIndexRequest
(message)UpdateModelDeploymentMonitoringJobOperationMetadata
(message)UpdateModelDeploymentMonitoringJobRequest
(message)UpdateModelRequest
(message)UpdateNotebookRuntimeTemplateRequest
(message)UpdatePersistentResourceOperationMetadata
(message)UpdatePersistentResourceRequest
(message)UpdateRagCorpusOperationMetadata
(message)UpdateRagCorpusRequest
(message)UpdateScheduleRequest
(message)UpdateSpecialistPoolOperationMetadata
(message)UpdateSpecialistPoolRequest
(message)UpdateTensorboardExperimentRequest
(message)UpdateTensorboardOperationMetadata
(message)UpdateTensorboardRequest
(message)UpdateTensorboardRunRequest
(message)UpdateTensorboardTimeSeriesRequest
(message)UpgradeNotebookRuntimeOperationMetadata
(message)UpgradeNotebookRuntimeRequest
(message)UpgradeNotebookRuntimeResponse
(message)UploadModelOperationMetadata
(message)UploadModelRequest
(message)UploadModelResponse
(message)UploadRagFileConfig
(message)UpsertDatapointsRequest
(message)UpsertDatapointsResponse
(message)UserActionReference
(message)Value
(message)VertexAISearch
(message)VertexRagStore
(message)VertexRagStore.RagResource
(message)VideoMetadata
(message)WorkerPoolSpec
(message)WriteFeatureValuesPayload
(message)WriteFeatureValuesRequest
(message)WriteFeatureValuesResponse
(message)WriteTensorboardExperimentDataRequest
(message)WriteTensorboardExperimentDataResponse
(message)WriteTensorboardRunDataRequest
(message)WriteTensorboardRunDataResponse
(message)XraiAttribution
(message)
DatasetService
The service that manages Vertex AI Dataset and its child resources.
CreateDataset |
---|
Creates a Dataset.
|
CreateDatasetVersion |
---|
Create a version from a Dataset.
|
DeleteDataset |
---|
Deletes a Dataset.
|
DeleteDatasetVersion |
---|
Deletes a Dataset version.
|
DeleteSavedQuery |
---|
Deletes a SavedQuery.
|
ExportData |
---|
Exports data from a Dataset.
|
GetAnnotationSpec |
---|
Gets an AnnotationSpec.
|
GetDataset |
---|
Gets a Dataset.
|
GetDatasetVersion |
---|
Gets a Dataset version.
|
ImportData |
---|
Imports data into a Dataset.
|
ListAnnotations |
---|
Lists Annotations belongs to a dataitem This RPC is only available in InternalDatasetService. It is only used for exporting conversation data to CCAI Insights.
|
ListDataItems |
---|
Lists DataItems in a Dataset.
|
ListDatasetVersions |
---|
Lists DatasetVersions in a Dataset.
|
ListDatasets |
---|
Lists Datasets in a Location.
|
ListSavedQueries |
---|
Lists SavedQueries in a Dataset.
|
RestoreDatasetVersion |
---|
Restores a dataset version.
|
SearchDataItems |
---|
Searches DataItems in a Dataset.
|
UpdateDataset |
---|
Updates a Dataset.
|
UpdateDatasetVersion |
---|
Updates a DatasetVersion.
|
DeploymentResourcePoolService
A service that manages the DeploymentResourcePool resource.
CreateDeploymentResourcePool |
---|
Create a DeploymentResourcePool.
|
DeleteDeploymentResourcePool |
---|
Delete a DeploymentResourcePool.
|
GetDeploymentResourcePool |
---|
Get a DeploymentResourcePool.
|
ListDeploymentResourcePools |
---|
List DeploymentResourcePools in a location.
|
QueryDeployedModels |
---|
List DeployedModels that have been deployed on this DeploymentResourcePool.
|
UpdateDeploymentResourcePool |
---|
Update a DeploymentResourcePool.
|
EndpointService
A service for managing Vertex AI's Endpoints.
CreateEndpoint |
---|
Creates an Endpoint.
|
DeleteEndpoint |
---|
Deletes an Endpoint.
|
DeployModel |
---|
Deploys a Model into this Endpoint, creating a DeployedModel within it.
|
GetEndpoint |
---|
Gets an Endpoint.
|
ListEndpoints |
---|
Lists Endpoints in a Location.
|
MutateDeployedModel |
---|
Updates an existing deployed model. Updatable fields include
|
UndeployModel |
---|
Undeploys a Model from an Endpoint, removing a DeployedModel from it, and freeing all resources it's using.
|
UpdateEndpoint |
---|
Updates an Endpoint.
|
UpdateEndpointLongRunning |
---|
Updates an Endpoint with a long running operation.
|
EvaluationService
Vertex AI Online Evaluation Service.
EvaluateInstances |
---|
Evaluates instances based on a given metric.
|
FeatureOnlineStoreAdminService
The service that handles CRUD and List for resources for FeatureOnlineStore.
CreateFeatureOnlineStore |
---|
Creates a new FeatureOnlineStore in a given project and location.
|
CreateFeatureView |
---|
Creates a new FeatureView in a given FeatureOnlineStore.
|
DeleteFeatureOnlineStore |
---|
Deletes a single FeatureOnlineStore. The FeatureOnlineStore must not contain any FeatureViews.
|
DeleteFeatureView |
---|
Deletes a single FeatureView.
|
GetFeatureOnlineStore |
---|
Gets details of a single FeatureOnlineStore.
|
GetFeatureView |
---|
Gets details of a single FeatureView.
|
GetFeatureViewSync |
---|
Gets details of a single FeatureViewSync.
|
ListFeatureOnlineStores |
---|
Lists FeatureOnlineStores in a given project and location.
|
ListFeatureViewSyncs |
---|
Lists FeatureViewSyncs in a given FeatureView.
|
ListFeatureViews |
---|
Lists FeatureViews in a given FeatureOnlineStore.
|
SyncFeatureView |
---|
Triggers on-demand sync for the FeatureView.
|
UpdateFeatureOnlineStore |
---|
Updates the parameters of a single FeatureOnlineStore.
|
UpdateFeatureView |
---|
Updates the parameters of a single FeatureView.
|
FeatureOnlineStoreService
A service for fetching feature values from the online store.
FetchFeatureValues |
---|
Fetch feature values under a FeatureView.
|
SearchNearestEntities |
---|
Search the nearest entities under a FeatureView. Search only works for indexable feature view; if a feature view isn't indexable, returns Invalid argument response.
|
FeatureRegistryService
The service that handles CRUD and List for resources for FeatureRegistry.
BatchCreateFeatures |
---|
Creates a batch of Features in a given FeatureGroup.
|
CreateFeature |
---|
Creates a new Feature in a given FeatureGroup.
|
CreateFeatureGroup |
---|
Creates a new FeatureGroup in a given project and location.
|
DeleteFeature |
---|
Deletes a single Feature.
|
DeleteFeatureGroup |
---|
Deletes a single FeatureGroup.
|
GetFeature |
---|
Gets details of a single Feature.
|
GetFeatureGroup |
---|
Gets details of a single FeatureGroup.
|
ListFeatureGroups |
---|
Lists FeatureGroups in a given project and location.
|
ListFeatures |
---|
Lists Features in a given FeatureGroup.
|
UpdateFeature |
---|
Updates the parameters of a single Feature.
|
UpdateFeatureGroup |
---|
Updates the parameters of a single FeatureGroup.
|
FeaturestoreOnlineServingService
A service for serving online feature values.
ReadFeatureValues |
---|
Reads Feature values of a specific entity of an EntityType. For reading feature values of multiple entities of an EntityType, please use StreamingReadFeatureValues.
|
StreamingReadFeatureValues |
---|
Reads Feature values for multiple entities. Depending on their size, data for different entities may be broken up across multiple responses.
|
WriteFeatureValues |
---|
Writes Feature values of one or more entities of an EntityType. The Feature values are merged into existing entities if any. The Feature values to be written must have timestamp within the online storage retention.
|
FeaturestoreService
The service that handles CRUD and List for resources for Featurestore.
BatchCreateFeatures |
---|
Creates a batch of Features in a given EntityType.
|
BatchReadFeatureValues |
---|
Batch reads Feature values from a Featurestore. This API enables batch reading Feature values, where each read instance in the batch may read Feature values of entities from one or more EntityTypes. Point-in-time correctness is guaranteed for Feature values of each read instance as of each instance's read timestamp.
|
CreateEntityType |
---|
Creates a new EntityType in a given Featurestore.
|
CreateFeature |
---|
Creates a new Feature in a given EntityType.
|
CreateFeaturestore |
---|
Creates a new Featurestore in a given project and location.
|
DeleteEntityType |
---|
Deletes a single EntityType. The EntityType must not have any Features or
|
DeleteFeature |
---|
Deletes a single Feature.
|
DeleteFeatureValues |
---|
Delete Feature values from Featurestore. The progress of the deletion is tracked by the returned operation. The deleted feature values are guaranteed to be invisible to subsequent read operations after the operation is marked as successfully done. If a delete feature values operation fails, the feature values returned from reads and exports may be inconsistent. If consistency is required, the caller must retry the same delete request again and wait till the new operation returned is marked as successfully done.
|
DeleteFeaturestore |
---|
Deletes a single Featurestore. The Featurestore must not contain any EntityTypes or
|
ExportFeatureValues |
---|
Exports Feature values from all the entities of a target EntityType.
|
GetEntityType |
---|
Gets details of a single EntityType.
|
GetFeature |
---|
Gets details of a single Feature.
|
GetFeaturestore |
---|
Gets details of a single Featurestore.
|
ImportFeatureValues |
---|
Imports Feature values into the Featurestore from a source storage. The progress of the import is tracked by the returned operation. The imported features are guaranteed to be visible to subsequent read operations after the operation is marked as successfully done. If an import operation fails, the Feature values returned from reads and exports may be inconsistent. If consistency is required, the caller must retry the same import request again and wait till the new operation returned is marked as successfully done. There are also scenarios where the caller can cause inconsistency.
|
ListEntityTypes |
---|
Lists EntityTypes in a given Featurestore.
|
ListFeatures |
---|
Lists Features in a given EntityType.
|
ListFeaturestores |
---|
Lists Featurestores in a given project and location.
|
SearchFeatures |
---|
Searches Features matching a query in a given project.
|
UpdateEntityType |
---|
Updates the parameters of a single EntityType.
|
UpdateFeature |
---|
Updates the parameters of a single Feature.
|
UpdateFeaturestore |
---|
Updates the parameters of a single Featurestore.
|
GenAiTuningService
A service for creating and managing GenAI Tuning Jobs.
CancelTuningJob |
---|
Cancels a TuningJob. Starts asynchronous cancellation on the TuningJob. The server makes a best effort to cancel the job, but success is not guaranteed. Clients can use
|
CreateTuningJob |
---|
Creates a TuningJob. A created TuningJob right away will be attempted to be run.
|
GetTuningJob |
---|
Gets a TuningJob.
|
ListTuningJobs |
---|
Lists TuningJobs in a Location.
|
RebaseTunedModel |
---|
Rebase a TunedModel.
|
IndexEndpointService
A service for managing Vertex AI's IndexEndpoints.
CreateIndexEndpoint |
---|
Creates an IndexEndpoint.
|
DeleteIndexEndpoint |
---|
Deletes an IndexEndpoint.
|
DeployIndex |
---|
Deploys an Index into this IndexEndpoint, creating a DeployedIndex within it. Only non-empty Indexes can be deployed.
|
GetIndexEndpoint |
---|
Gets an IndexEndpoint.
|
ListIndexEndpoints |
---|
Lists IndexEndpoints in a Location.
|
MutateDeployedIndex |
---|
Update an existing DeployedIndex under an IndexEndpoint.
|
UndeployIndex |
---|
Undeploys an Index from an IndexEndpoint, removing a DeployedIndex from it, and freeing all resources it's using.
|
UpdateIndexEndpoint |
---|
Updates an IndexEndpoint.
|
IndexService
A service for creating and managing Vertex AI's Index resources.
CreateIndex |
---|
Creates an Index.
|
DeleteIndex |
---|
Deletes an Index. An Index can only be deleted when all its
|
GetIndex |
---|
Gets an Index.
|
ListIndexes |
---|
Lists Indexes in a Location.
|
RemoveDatapoints |
---|
Remove Datapoints from an Index.
|
UpdateIndex |
---|
Updates an Index.
|
UpsertDatapoints |
---|
Add/update Datapoints into an Index.
|
JobService
A service for creating and managing Vertex AI's jobs.
CancelBatchPredictionJob |
---|
Cancels a BatchPredictionJob. Starts asynchronous cancellation on the BatchPredictionJob. The server makes the best effort to cancel the job, but success is not guaranteed. Clients can use
|
CancelCustomJob |
---|
Cancels a CustomJob. Starts asynchronous cancellation on the CustomJob. The server makes a best effort to cancel the job, but success is not guaranteed. Clients can use
|
CancelHyperparameterTuningJob |
---|
Cancels a HyperparameterTuningJob. Starts asynchronous cancellation on the HyperparameterTuningJob. The server makes a best effort to cancel the job, but success is not guaranteed. Clients can use
|
CancelNasJob |
---|
Cancels a NasJob. Starts asynchronous cancellation on the NasJob. The server makes a best effort to cancel the job, but success is not guaranteed. Clients can use
|
CreateBatchPredictionJob |
---|
Creates a BatchPredictionJob. A BatchPredictionJob once created will right away be attempted to start.
|
CreateCustomJob |
---|
Creates a CustomJob. A created CustomJob right away will be attempted to be run.
|
CreateHyperparameterTuningJob |
---|
Creates a HyperparameterTuningJob
|
CreateModelDeploymentMonitoringJob |
---|
Creates a ModelDeploymentMonitoringJob. It will run periodically on a configured interval.
|
CreateNasJob |
---|
Creates a NasJob
|
DeleteBatchPredictionJob |
---|
Deletes a BatchPredictionJob. Can only be called on jobs that already finished.
|
DeleteCustomJob |
---|
Deletes a CustomJob.
|
DeleteHyperparameterTuningJob |
---|
Deletes a HyperparameterTuningJob.
|
DeleteModelDeploymentMonitoringJob |
---|
Deletes a ModelDeploymentMonitoringJob.
|
DeleteNasJob |
---|
Deletes a NasJob.
|
GetBatchPredictionJob |
---|
Gets a BatchPredictionJob
|
GetCustomJob |
---|
Gets a CustomJob.
|
GetHyperparameterTuningJob |
---|
Gets a HyperparameterTuningJob
|
GetModelDeploymentMonitoringJob |
---|
Gets a ModelDeploymentMonitoringJob.
|
GetNasJob |
---|
Gets a NasJob
|
GetNasTrialDetail |
---|
Gets a NasTrialDetail.
|
ListBatchPredictionJobs |
---|
Lists BatchPredictionJobs in a Location.
|
ListCustomJobs |
---|
Lists CustomJobs in a Location.
|
ListHyperparameterTuningJobs |
---|
Lists HyperparameterTuningJobs in a Location.
|
ListModelDeploymentMonitoringJobs |
---|
Lists ModelDeploymentMonitoringJobs in a Location.
|
ListNasJobs |
---|
Lists NasJobs in a Location.
|
ListNasTrialDetails |
---|
List top NasTrialDetails of a NasJob.
|
PauseModelDeploymentMonitoringJob |
---|
Pauses a ModelDeploymentMonitoringJob. If the job is running, the server makes a best effort to cancel the job. Will mark
|
ResumeModelDeploymentMonitoringJob |
---|
Resumes a paused ModelDeploymentMonitoringJob. It will start to run from next scheduled time. A deleted ModelDeploymentMonitoringJob can't be resumed.
|
SearchModelDeploymentMonitoringStatsAnomalies |
---|
Searches Model Monitoring Statistics generated within a given time window.
|
UpdateModelDeploymentMonitoringJob |
---|
Updates a ModelDeploymentMonitoringJob.
|
LlmUtilityService
Service for LLM related utility functions.
ComputeTokens |
---|
Return a list of tokens based on the input text.
|
CountTokens |
---|
Perform a token counting.
|
MatchService
MatchService is a Google managed service for efficient vector similarity search at scale.
MetadataService
Service for reading and writing metadata entries.
AddContextArtifactsAndExecutions |
---|
Adds a set of Artifacts and Executions to a Context. If any of the Artifacts or Executions have already been added to a Context, they are simply skipped.
|
AddContextChildren |
---|
Adds a set of Contexts as children to a parent Context. If any of the child Contexts have already been added to the parent Context, they are simply skipped. If this call would create a cycle or cause any Context to have more than 10 parents, the request will fail with an INVALID_ARGUMENT error.
|
AddExecutionEvents |
---|
Adds Events to the specified Execution. An Event indicates whether an Artifact was used as an input or output for an Execution. If an Event already exists between the Execution and the Artifact, the Event is skipped.
|
CreateArtifact |
---|
Creates an Artifact associated with a MetadataStore.
|
CreateContext |
---|
Creates a Context associated with a MetadataStore.
|
CreateExecution |
---|
Creates an Execution associated with a MetadataStore.
|
CreateMetadataSchema |
---|
Creates a MetadataSchema.
|
CreateMetadataStore |
---|
Initializes a MetadataStore, including allocation of resources.
|
DeleteArtifact |
---|
Deletes an Artifact.
|
DeleteContext |
---|
Deletes a stored Context.
|
DeleteExecution |
---|
Deletes an Execution.
|
DeleteMetadataStore |
---|
Deletes a single MetadataStore and all its child resources (Artifacts, Executions, and Contexts).
|
GetArtifact |
---|
Retrieves a specific Artifact.
|
GetContext |
---|
Retrieves a specific Context.
|
GetExecution |
---|
Retrieves a specific Execution.
|
GetMetadataSchema |
---|
Retrieves a specific MetadataSchema.
|
GetMetadataStore |
---|
Retrieves a specific MetadataStore.
|
ListArtifacts |
---|
Lists Artifacts in the MetadataStore.
|
ListContexts |
---|
Lists Contexts on the MetadataStore.
|
ListExecutions |
---|
Lists Executions in the MetadataStore.
|
ListMetadataSchemas |
---|
Lists MetadataSchemas.
|
ListMetadataStores |
---|
Lists MetadataStores for a Location.
|
PurgeArtifacts |
---|
Purges Artifacts.
|
PurgeContexts |
---|
Purges Contexts.
|
PurgeExecutions |
---|
Purges Executions.
|
QueryArtifactLineageSubgraph |
---|
Retrieves lineage of an Artifact represented through Artifacts and Executions connected by Event edges and returned as a LineageSubgraph.
|
QueryContextLineageSubgraph |
---|
Retrieves Artifacts and Executions within the specified Context, connected by Event edges and returned as a LineageSubgraph.
|
QueryExecutionInputsAndOutputs |
---|
Obtains the set of input and output Artifacts for this Execution, in the form of LineageSubgraph that also contains the Execution and connecting Events.
|
RemoveContextChildren |
---|
Remove a set of children contexts from a parent Context. If any of the child Contexts were NOT added to the parent Context, they are simply skipped.
|
UpdateArtifact |
---|
Updates a stored Artifact.
|
UpdateContext |
---|
Updates a stored Context.
|
UpdateExecution |
---|
Updates a stored Execution.
|
MigrationService
A service that migrates resources from automl.googleapis.com, datalabeling.googleapis.com and ml.googleapis.com to Vertex AI.
BatchMigrateResources |
---|
Batch migrates resources from ml.googleapis.com, automl.googleapis.com, and datalabeling.googleapis.com to Vertex AI.
|
SearchMigratableResources |
---|
Searches all of the resources in automl.googleapis.com, datalabeling.googleapis.com and ml.googleapis.com that can be migrated to Vertex AI's given location.
|
ModelGardenService
The interface of Model Garden Service.
GetPublisherModel |
---|
Gets a Model Garden publisher model.
|
ModelService
A service for managing Vertex AI's machine learning Models.
BatchImportEvaluatedAnnotations |
---|
Imports a list of externally generated EvaluatedAnnotations.
|
BatchImportModelEvaluationSlices |
---|
Imports a list of externally generated ModelEvaluationSlice.
|
CopyModel |
---|
Copies an already existing Vertex AI Model into the specified Location. The source Model must exist in the same Project. When copying custom Models, the users themselves are responsible for
|
DeleteModel |
---|
Deletes a Model. A model cannot be deleted if any
|
DeleteModelVersion |
---|
Deletes a Model version. Model version can only be deleted if there are no
|
ExportModel |
---|
Exports a trained, exportable Model to a location specified by the user. A Model is considered to be exportable if it has at least one
|
GetModel |
---|
Gets a Model.
|
GetModelEvaluation |
---|
Gets a ModelEvaluation.
|
GetModelEvaluationSlice |
---|
Gets a ModelEvaluationSlice.
|
ImportModelEvaluation |
---|
Imports an externally generated ModelEvaluation.
|
ListModelEvaluationSlices |
---|
Lists ModelEvaluationSlices in a ModelEvaluation.
|
ListModelEvaluations |
---|
Lists ModelEvaluations in a Model.
|
ListModelVersions |
---|
Lists versions of the specified model.
|
ListModels |
---|
Lists Models in a Location.
|
MergeVersionAliases |
---|
Merges a set of aliases for a Model version.
|
UpdateExplanationDataset |
---|
Incrementally update the dataset used for an examples model.
|
UpdateModel |
---|
Updates a Model.
|
UploadModel |
---|
Uploads a Model artifact into Vertex AI.
|
NotebookService
The interface for Vertex Notebook service (a.k.a. Colab on Workbench).
AssignNotebookRuntime |
---|
Assigns a NotebookRuntime to a user for a particular Notebook file. This method will either returns an existing assignment or generates a new one.
|
CreateNotebookExecutionJob |
---|
Creates a NotebookExecutionJob.
|
CreateNotebookRuntimeTemplate |
---|
Creates a NotebookRuntimeTemplate.
|
DeleteNotebookExecutionJob |
---|
Deletes a NotebookExecutionJob.
|
DeleteNotebookRuntime |
---|
Deletes a NotebookRuntime.
|
DeleteNotebookRuntimeTemplate |
---|
Deletes a NotebookRuntimeTemplate.
|
GetNotebookExecutionJob |
---|
Gets a NotebookExecutionJob.
|
GetNotebookRuntime |
---|
Gets a NotebookRuntime.
|
GetNotebookRuntimeTemplate |
---|
Gets a NotebookRuntimeTemplate.
|
ListNotebookExecutionJobs |
---|
Lists NotebookExecutionJobs in a Location.
|
ListNotebookRuntimeTemplates |
---|
Lists NotebookRuntimeTemplates in a Location.
|
ListNotebookRuntimes |
---|
Lists NotebookRuntimes in a Location.
|
StartNotebookRuntime |
---|
Starts a NotebookRuntime.
|
StopNotebookRuntime |
---|
Stops a NotebookRuntime.
|
UpdateNotebookRuntimeTemplate |
---|
Updates a NotebookRuntimeTemplate.
|
UpgradeNotebookRuntime |
---|
Upgrades a NotebookRuntime.
|
PersistentResourceService
A service for managing Vertex AI's machine learning PersistentResource.
CreatePersistentResource |
---|
Creates a PersistentResource.
|
DeletePersistentResource |
---|
Deletes a PersistentResource.
|
GetPersistentResource |
---|
Gets a PersistentResource.
|
ListPersistentResources |
---|
Lists PersistentResources in a Location.
|
RebootPersistentResource |
---|
Reboots a PersistentResource.
|
UpdatePersistentResource |
---|
Updates a PersistentResource.
|
PipelineService
A service for creating and managing Vertex AI's pipelines. This includes both TrainingPipeline
resources (used for AutoML and custom training) and PipelineJob
resources (used for Vertex AI Pipelines).
BatchCancelPipelineJobs |
---|
Batch cancel PipelineJobs. Firstly the server will check if all the jobs are in non-terminal states, and skip the jobs that are already terminated. If the operation failed, none of the pipeline jobs are cancelled. The server will poll the states of all the pipeline jobs periodically to check the cancellation status. This operation will return an LRO.
|
BatchDeletePipelineJobs |
---|
Batch deletes PipelineJobs The Operation is atomic. If it fails, none of the PipelineJobs are deleted. If it succeeds, all of the PipelineJobs are deleted.
|
CancelPipelineJob |
---|
Cancels a PipelineJob. Starts asynchronous cancellation on the PipelineJob. The server makes a best effort to cancel the pipeline, but success is not guaranteed. Clients can use
|
CancelTrainingPipeline |
---|
Cancels a TrainingPipeline. Starts asynchronous cancellation on the TrainingPipeline. The server makes a best effort to cancel the pipeline, but success is not guaranteed. Clients can use
|
CreatePipelineJob |
---|
Creates a PipelineJob. A PipelineJob will run immediately when created.
|
CreateTrainingPipeline |
---|
Creates a TrainingPipeline. A created TrainingPipeline right away will be attempted to be run.
|
DeletePipelineJob |
---|
Deletes a PipelineJob.
|
DeleteTrainingPipeline |
---|
Deletes a TrainingPipeline.
|
GetPipelineJob |
---|
Gets a PipelineJob.
|
GetTrainingPipeline |
---|
Gets a TrainingPipeline.
|
ListPipelineJobs |
---|
Lists PipelineJobs in a Location.
|
ListTrainingPipelines |
---|
Lists TrainingPipelines in a Location.
|
PredictionService
A service for online predictions and explanations.
DirectPredict |
---|
Perform an unary online prediction request to a gRPC model server for Vertex first-party products and frameworks.
|
DirectRawPredict |
---|
Perform an unary online prediction request to a gRPC model server for custom containers.
|
Explain |
---|
Perform an online explanation. If
|
GenerateContent |
---|
Generate content with multimodal inputs.
|
Predict |
---|
Perform an online prediction.
|
PredictLongRunning |
---|
|
RawPredict |
---|
Perform an online prediction with an arbitrary HTTP payload. The response includes the following HTTP headers:
|
ServerStreamingPredict |
---|
Perform a server-side streaming online prediction request for Vertex LLM streaming.
|
StreamDirectPredict |
---|
Perform a streaming online prediction request to a gRPC model server for Vertex first-party products and frameworks.
|
StreamDirectRawPredict |
---|
Perform a streaming online prediction request to a gRPC model server for custom containers.
|
StreamGenerateContent |
---|
Generate content with multimodal inputs with streaming support.
|
StreamRawPredict |
---|
Perform a streaming online prediction with an arbitrary HTTP payload.
|
StreamingPredict |
---|
Perform a streaming online prediction request for Vertex first-party products and frameworks.
|
StreamingRawPredict |
---|
Perform a streaming online prediction request through gRPC.
|
ScheduleService
A service for creating and managing Vertex AI's Schedule resources to periodically launch shceudled runs to make API calls.
CreateSchedule |
---|
Creates a Schedule.
|
DeleteSchedule |
---|
Deletes a Schedule.
|
GetSchedule |
---|
Gets a Schedule.
|
ListSchedules |
---|
Lists Schedules in a Location.
|
PauseSchedule |
---|
Pauses a Schedule. Will mark
|
ResumeSchedule |
---|
Resumes a paused Schedule to start scheduling new runs. Will mark When the Schedule is resumed, new runs will be scheduled starting from the next execution time after the current time based on the time_specification in the Schedule. If
|
UpdateSchedule |
---|
Updates an active or paused Schedule. When the Schedule is updated, new runs will be scheduled starting from the updated next execution time after the update time based on the time_specification in the updated Schedule. All unstarted runs before the update time will be skipped while already created runs will NOT be paused or canceled.
|
SpecialistPoolService
A service for creating and managing Customer SpecialistPools. When customers start Data Labeling jobs, they can reuse/create Specialist Pools to bring their own Specialists to label the data. Customers can add/remove Managers for the Specialist Pool on Cloud console, then Managers will get email notifications to manage Specialists and tasks on CrowdCompute console.
CreateSpecialistPool |
---|
Creates a SpecialistPool.
|
DeleteSpecialistPool |
---|
Deletes a SpecialistPool as well as all Specialists in the pool.
|
GetSpecialistPool |
---|
Gets a SpecialistPool.
|
ListSpecialistPools |
---|
Lists SpecialistPools in a Location.
|
UpdateSpecialistPool |
---|
Updates a SpecialistPool.
|
TensorboardService
TensorboardService
BatchCreateTensorboardRuns |
---|
Batch create TensorboardRuns.
|
BatchCreateTensorboardTimeSeries |
---|
Batch create TensorboardTimeSeries that belong to a TensorboardExperiment.
|
BatchReadTensorboardTimeSeriesData |
---|
Reads multiple TensorboardTimeSeries' data. The data point number limit is 1000 for scalars, 100 for tensors and blob references. If the number of data points stored is less than the limit, all data is returned. Otherwise, the number limit of data points is randomly selected from this time series and returned.
|
CreateTensorboard |
---|
Creates a Tensorboard.
|
CreateTensorboardExperiment |
---|
Creates a TensorboardExperiment.
|
CreateTensorboardRun |
---|
Creates a TensorboardRun.
|
CreateTensorboardTimeSeries |
---|
Creates a TensorboardTimeSeries.
|
DeleteTensorboard |
---|
Deletes a Tensorboard.
|
DeleteTensorboardExperiment |
---|
Deletes a TensorboardExperiment.
|
DeleteTensorboardRun |
---|
Deletes a TensorboardRun.
|
DeleteTensorboardTimeSeries |
---|
Deletes a TensorboardTimeSeries.
|
ExportTensorboardTimeSeriesData |
---|
Exports a TensorboardTimeSeries' data. Data is returned in paginated responses.
|
GetTensorboard |
---|
Gets a Tensorboard.
|
GetTensorboardExperiment |
---|
Gets a TensorboardExperiment.
|
GetTensorboardRun |
---|
Gets a TensorboardRun.
|
GetTensorboardTimeSeries |
---|
Gets a TensorboardTimeSeries.
|
ListTensorboardExperiments |
---|
Lists TensorboardExperiments in a Location.
|
ListTensorboardRuns |
---|
Lists TensorboardRuns in a Location.
|
ListTensorboardTimeSeries |
---|
Lists TensorboardTimeSeries in a Location.
|
ListTensorboards |
---|
Lists Tensorboards in a Location.
|
ReadTensorboardBlobData |
---|
Gets bytes of TensorboardBlobs. This is to allow reading blob data stored in consumer project's Cloud Storage bucket without users having to obtain Cloud Storage access permission.
|
ReadTensorboardSize |
---|
Returns the storage size for a given TensorBoard instance.
|
ReadTensorboardTimeSeriesData |
---|
Reads a TensorboardTimeSeries' data. By default, if the number of data points stored is less than 1000, all data is returned. Otherwise, 1000 data points is randomly selected from this time series and returned. This value can be changed by changing max_data_points, which can't be greater than 10k.
|
ReadTensorboardUsage |
---|
Returns a list of monthly active users for a given TensorBoard instance.
|
UpdateTensorboard |
---|
Updates a Tensorboard.
|
UpdateTensorboardExperiment |
---|
Updates a TensorboardExperiment.
|
UpdateTensorboardRun |
---|
Updates a TensorboardRun.
|
UpdateTensorboardTimeSeries |
---|
Updates a TensorboardTimeSeries.
|
WriteTensorboardExperimentData |
---|
Write time series data points of multiple TensorboardTimeSeries in multiple TensorboardRun's. If any data fail to be ingested, an error is returned.
|
WriteTensorboardRunData |
---|
Write time series data points into multiple TensorboardTimeSeries under a TensorboardRun. If any data fail to be ingested, an error is returned.
|
VertexRagDataService
A service for managing user data for RAG.
CreateRagCorpus |
---|
Creates a RagCorpus.
|
DeleteRagCorpus |
---|
Deletes a RagCorpus.
|
DeleteRagFile |
---|
Deletes a RagFile.
|
GetRagCorpus |
---|
Gets a RagCorpus.
|
GetRagFile |
---|
Gets a RagFile.
|
ImportRagFiles |
---|
Import files from Google Cloud Storage or Google Drive into a RagCorpus.
|
ListRagCorpora |
---|
Lists RagCorpora in a Location.
|
ListRagFiles |
---|
Lists RagFiles in a RagCorpus.
|
UpdateRagCorpus |
---|
Updates a RagCorpus.
|
VertexRagService
A service for retrieving relevant contexts.
AugmentPrompt |
---|
Given an input prompt, it returns augmented prompt from vertex rag store to guide LLM towards generating grounded responses.
|
CorroborateContent |
---|
Given an input text, it returns a score that evaluates the factuality of the text. It also extracts and returns claims from the text and provides supporting facts.
|
RetrieveContexts |
---|
Retrieves relevant contexts for a query.
|
VizierService
Vertex AI Vizier API.
Vertex AI Vizier is a service to solve blackbox optimization problems, such as tuning machine learning hyperparameters and searching over deep learning architectures.
AddTrialMeasurement |
---|
Adds a measurement of the objective metrics to a Trial. This measurement is assumed to have been taken before the Trial is complete.
|
CheckTrialEarlyStoppingState |
---|
Checks whether a Trial should stop or not. Returns a long-running operation. When the operation is successful, it will contain a
|
CompleteTrial |
---|
Marks a Trial as complete.
|
CreateStudy |
---|
Creates a Study. A resource name will be generated after creation of the Study.
|
CreateTrial |
---|
Adds a user provided Trial to a Study.
|
DeleteStudy |
---|
Deletes a Study.
|
DeleteTrial |
---|
Deletes a Trial.
|
GetStudy |
---|
Gets a Study by name.
|
GetTrial |
---|
Gets a Trial.
|
ListOptimalTrials |
---|
Lists the pareto-optimal Trials for multi-objective Study or the optimal Trials for single-objective Study. The definition of pareto-optimal can be checked in wiki page. https://en.wikipedia.org/wiki/Pareto_efficiency
|
ListStudies |
---|
Lists all the studies in a region for an associated project.
|
ListTrials |
---|
Lists the Trials associated with a Study.
|
LookupStudy |
---|
Looks a study up using the user-defined display_name field instead of the fully qualified resource name.
|
StopTrial |
---|
Stops a Trial.
|
SuggestTrials |
---|
Adds one or more Trials to a Study, with parameter values suggested by Vertex AI Vizier. Returns a long-running operation associated with the generation of Trial suggestions. When this long-running operation succeeds, it will contain a
|
AcceleratorType
Represents a hardware accelerator type.
Enums | |
---|---|
ACCELERATOR_TYPE_UNSPECIFIED |
Unspecified accelerator type, which means no accelerator. |
NVIDIA_TESLA_K80 |
Deprecated: Nvidia Tesla K80 GPU has reached end of support, see https://cloud.google.com/compute/docs/eol/k80-eol. |
NVIDIA_TESLA_P100 |
Nvidia Tesla P100 GPU. |
NVIDIA_TESLA_V100 |
Nvidia Tesla V100 GPU. |
NVIDIA_TESLA_P4 |
Nvidia Tesla P4 GPU. |
NVIDIA_TESLA_T4 |
Nvidia Tesla T4 GPU. |
NVIDIA_TESLA_A100 |
Nvidia Tesla A100 GPU. |
NVIDIA_A100_80GB |
Nvidia A100 80GB GPU. |
NVIDIA_L4 |
Nvidia L4 GPU. |
NVIDIA_H100_80GB |
Nvidia H100 80Gb GPU. |
NVIDIA_H100_MEGA_80GB |
Nvidia H100 Mega 80Gb GPU. |
TPU_V2 |
TPU v2. |
TPU_V3 |
TPU v3. |
TPU_V4_POD |
TPU v4. |
TPU_V5_LITEPOD |
TPU v5. |
AddContextArtifactsAndExecutionsRequest
Request message for MetadataService.AddContextArtifactsAndExecutions
.
Fields | |
---|---|
context |
Required. The resource name of the Context that the Artifacts and Executions belong to. Format: |
artifacts[] |
The resource names of the Artifacts to attribute to the Context. Format: |
executions[] |
The resource names of the Executions to associate with the Context. Format: |
AddContextArtifactsAndExecutionsResponse
This type has no fields.
Response message for MetadataService.AddContextArtifactsAndExecutions
.
AddContextChildrenRequest
Request message for MetadataService.AddContextChildren
.
Fields | |
---|---|
context |
Required. The resource name of the parent Context. Format: |
child_ |
The resource names of the child Contexts. |
AddContextChildrenResponse
This type has no fields.
Response message for MetadataService.AddContextChildren
.
AddExecutionEventsRequest
Request message for MetadataService.AddExecutionEvents
.
Fields | |
---|---|
execution |
Required. The resource name of the Execution that the Events connect Artifacts with. Format: |
events[] |
The Events to create and add. |
AddExecutionEventsResponse
This type has no fields.
Response message for MetadataService.AddExecutionEvents
.
AddTrialMeasurementRequest
Request message for VizierService.AddTrialMeasurement
.
Fields | |
---|---|
trial_ |
Required. The name of the trial to add measurement. Format: |
measurement |
Required. The measurement to be added to a Trial. |
Annotation
Used to assign specific AnnotationSpec to a particular area of a DataItem or the whole part of the DataItem.
Fields | |
---|---|
name |
Output only. Resource name of the Annotation. |
payload_ |
Required. Google Cloud Storage URI points to a YAML file describing |
payload |
Required. The schema of the payload can be found in |
create_ |
Output only. Timestamp when this Annotation was created. |
update_ |
Output only. Timestamp when this Annotation was last updated. |
etag |
Optional. Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
annotation_ |
Output only. The source of the Annotation. |
labels |
Optional. The labels with user-defined metadata to organize your Annotations. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Annotation(System labels are excluded). See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. Following system labels exist for each Annotation:
|
AnnotationSpec
Identifies a concept with which DataItems may be annotated with.
Fields | |
---|---|
name |
Output only. Resource name of the AnnotationSpec. |
display_ |
Required. The user-defined name of the AnnotationSpec. The name can be up to 128 characters long and can consist of any UTF-8 characters. |
create_ |
Output only. Timestamp when this AnnotationSpec was created. |
update_ |
Output only. Timestamp when AnnotationSpec was last updated. |
etag |
Optional. Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
ApiAuth
The generic reusable api auth config.
Fields | |
---|---|
Union field auth_config . The auth config. auth_config can be only one of the following: |
|
api_ |
The API secret. |
ApiKeyConfig
The API secret.
Fields | |
---|---|
api_ |
Required. The SecretManager secret version resource name storing API key. e.g. projects/{project}/secrets/{secret}/versions/{version} |
Artifact
Instance of a general artifact.
Fields | |
---|---|
name |
Output only. The resource name of the Artifact. |
display_ |
User provided display name of the Artifact. May be up to 128 Unicode characters. |
uri |
The uniform resource identifier of the artifact file. May be empty if there is no actual artifact file. |
etag |
An eTag used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
labels |
The labels with user-defined metadata to organize your Artifacts. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Artifact (System labels are excluded). |
create_ |
Output only. Timestamp when this Artifact was created. |
update_ |
Output only. Timestamp when this Artifact was last updated. |
state |
The state of this Artifact. This is a property of the Artifact, and does not imply or capture any ongoing process. This property is managed by clients (such as Vertex AI Pipelines), and the system does not prescribe or check the validity of state transitions. |
schema_ |
The title of the schema describing the metadata. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store. |
schema_ |
The version of the schema in schema_name to use. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store. |
metadata |
Properties of the Artifact. Top level metadata keys' heading and trailing spaces will be trimmed. The size of this field should not exceed 200KB. |
description |
Description of the Artifact |
State
Describes the state of the Artifact.
Enums | |
---|---|
STATE_UNSPECIFIED |
Unspecified state for the Artifact. |
PENDING |
A state used by systems like Vertex AI Pipelines to indicate that the underlying data item represented by this Artifact is being created. |
LIVE |
A state indicating that the Artifact should exist, unless something external to the system deletes it. |
AssignNotebookRuntimeOperationMetadata
Metadata information for NotebookService.AssignNotebookRuntime
.
Fields | |
---|---|
generic_ |
The operation generic information. |
progress_ |
A human-readable message that shows the intermediate progress details of NotebookRuntime. |
AssignNotebookRuntimeRequest
Request message for NotebookService.AssignNotebookRuntime
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to get the NotebookRuntime assignment. Format: |
notebook_ |
Required. The resource name of the NotebookRuntimeTemplate based on which a NotebookRuntime will be assigned (reuse or create a new one). |
notebook_ |
Required. Provide runtime specific information (e.g. runtime owner, notebook id) used for NotebookRuntime assignment. |
notebook_ |
Optional. User specified ID for the notebook runtime. |
Attribution
Attribution that explains a particular prediction output.
Fields | |
---|---|
baseline_ |
Output only. Model predicted output if the input instance is constructed from the baselines of all the features defined in If the Model's predicted output has multiple dimensions (rank > 1), this is the value in the output located by If there are multiple baselines, their output values are averaged. |
instance_ |
Output only. Model predicted output on the corresponding [explanation instance][ExplainRequest.instances]. The field name of the output is determined by the key in If the Model predicted output has multiple dimensions, this is the value in the output located by |
feature_ |
Output only. Attributions of each explained feature. Features are extracted from the The value is a struct, whose keys are the name of the feature. The values are how much the feature in the The format of the value is determined by the feature's input format:
The |
output_ |
Output only. The index that locates the explained prediction output. If the prediction output is a scalar value, output_index is not populated. If the prediction output has multiple dimensions, the length of the output_index list is the same as the number of dimensions of the output. The i-th element in output_index is the element index of the i-th dimension of the output vector. Indices start from 0. |
output_ |
Output only. The display name of the output identified by This field is only populated iff the Model predicts display names as a separate field along with the explained output. The predicted display name must has the same shape of the explained output, and can be located using output_index. |
approximation_ |
Output only. Error of
See this introduction for more information. |
output_ |
Output only. Name of the explain output. Specified as the key in |
AugmentPromptRequest
Request message for AugmentPrompt.
Fields | |
---|---|
parent |
Required. The resource name of the Location from which to augment prompt. The users must have permission to make a call in the project. Format: |
contents[] |
Optional. Input content to augment, only text format is supported for now. |
model |
Optional. Metadata of the backend deployed model. |
Union field data_source . The data source for retrieving contexts. data_source can be only one of the following: |
|
vertex_ |
Optional. Retrieves contexts from the Vertex RagStore. |
Model
Metadata of the backend deployed model.
Fields | |
---|---|
model |
Optional. The model that the user will send the augmented prompt for content generation. |
model_ |
Optional. The model version of the backend deployed model. |
AugmentPromptResponse
Response message for AugmentPrompt.
Fields | |
---|---|
augmented_ |
Augmented prompt, only text format is supported for now. |
facts[] |
Retrieved facts from RAG data sources. |
AutomaticResources
A description of resources that to large degree are decided by Vertex AI, and require only a modest additional configuration. Each Model supporting these resources documents its specific guidelines.
Fields | |
---|---|
min_ |
Immutable. The minimum number of replicas this DeployedModel will be always deployed on. If traffic against it increases, it may dynamically be deployed onto more replicas up to |
max_ |
Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, a no upper bound for scaling under heavy traffic will be assume, though Vertex AI may be unable to scale beyond certain replica number. |
AutoscalingMetricSpec
The metric specification that defines the target resource utilization (CPU utilization, accelerator's duty cycle, and so on) for calculating the desired replica count.
Fields | |
---|---|
metric_ |
Required. The resource metric name. Supported metrics:
|
target |
The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided. |
AvroSource
The storage details for Avro input content.
Fields | |
---|---|
gcs_ |
Required. Google Cloud Storage location. |
BatchCancelPipelineJobsOperationMetadata
Runtime operation information for PipelineService.BatchCancelPipelineJobs
.
Fields | |
---|---|
generic_ |
The common part of the operation metadata. |
BatchCancelPipelineJobsRequest
Request message for PipelineService.BatchCancelPipelineJobs
.
Fields | |
---|---|
parent |
Required. The name of the PipelineJobs' parent resource. Format: |
names[] |
Required. The names of the PipelineJobs to cancel. A maximum of 32 PipelineJobs can be cancelled in a batch. Format: |
BatchCancelPipelineJobsResponse
Response message for PipelineService.BatchCancelPipelineJobs
.
Fields | |
---|---|
pipeline_ |
PipelineJobs cancelled. |
BatchCreateFeaturesOperationMetadata
Details of operations that perform batch create Features.
Fields | |
---|---|
generic_ |
Operation metadata for Feature. |
BatchCreateFeaturesRequest
Request message for FeaturestoreService.BatchCreateFeatures
. Request message for FeatureRegistryService.BatchCreateFeatures
.
Fields | |
---|---|
parent |
Required. The resource name of the EntityType/FeatureGroup to create the batch of Features under. Format: |
requests[] |
Required. The request message specifying the Features to create. All Features must be created under the same parent EntityType / FeatureGroup. The |
BatchCreateFeaturesResponse
Response message for FeaturestoreService.BatchCreateFeatures
.
Fields | |
---|---|
features[] |
The Features created. |
BatchCreateTensorboardRunsRequest
Request message for TensorboardService.BatchCreateTensorboardRuns
.
Fields | |
---|---|
parent |
Required. The resource name of the TensorboardExperiment to create the TensorboardRuns in. Format: |
requests[] |
Required. The request message specifying the TensorboardRuns to create. A maximum of 1000 TensorboardRuns can be created in a batch. |
BatchCreateTensorboardRunsResponse
Response message for TensorboardService.BatchCreateTensorboardRuns
.
Fields | |
---|---|
tensorboard_ |
The created TensorboardRuns. |
BatchCreateTensorboardTimeSeriesRequest
Request message for TensorboardService.BatchCreateTensorboardTimeSeries
.
Fields | |
---|---|
parent |
Required. The resource name of the TensorboardExperiment to create the TensorboardTimeSeries in. Format: |
requests[] |
Required. The request message specifying the TensorboardTimeSeries to create. A maximum of 1000 TensorboardTimeSeries can be created in a batch. |
BatchCreateTensorboardTimeSeriesResponse
Response message for TensorboardService.BatchCreateTensorboardTimeSeries
.
Fields | |
---|---|
tensorboard_ |
The created TensorboardTimeSeries. |
BatchDedicatedResources
A description of resources that are used for performing batch operations, are dedicated to a Model, and need manual configuration.
Fields | |
---|---|
machine_ |
Required. Immutable. The specification of a single machine. |
starting_ |
Immutable. The number of machine replicas used at the start of the batch operation. If not set, Vertex AI decides starting number, not greater than |
max_ |
Immutable. The maximum number of machine replicas the batch operation may be scaled to. The default value is 10. |
BatchDeletePipelineJobsRequest
Request message for PipelineService.BatchDeletePipelineJobs
.
Fields | |
---|---|
parent |
Required. The name of the PipelineJobs' parent resource. Format: |
names[] |
Required. The names of the PipelineJobs to delete. A maximum of 32 PipelineJobs can be deleted in a batch. Format: |
BatchDeletePipelineJobsResponse
Response message for PipelineService.BatchDeletePipelineJobs
.
Fields | |
---|---|
pipeline_ |
PipelineJobs deleted. |
BatchImportEvaluatedAnnotationsRequest
Request message for ModelService.BatchImportEvaluatedAnnotations
Fields | |
---|---|
parent |
Required. The name of the parent ModelEvaluationSlice resource. Format: |
evaluated_ |
Required. Evaluated annotations resource to be imported. |
BatchImportEvaluatedAnnotationsResponse
Response message for ModelService.BatchImportEvaluatedAnnotations
Fields | |
---|---|
imported_ |
Output only. Number of EvaluatedAnnotations imported. |
BatchImportModelEvaluationSlicesRequest
Request message for ModelService.BatchImportModelEvaluationSlices
Fields | |
---|---|
parent |
Required. The name of the parent ModelEvaluation resource. Format: |
model_ |
Required. Model evaluation slice resource to be imported. |
BatchImportModelEvaluationSlicesResponse
Response message for ModelService.BatchImportModelEvaluationSlices
Fields | |
---|---|
imported_ |
Output only. List of imported |
BatchMigrateResourcesOperationMetadata
Runtime operation information for MigrationService.BatchMigrateResources
.
Fields | |
---|---|
generic_ |
The common part of the operation metadata. |
partial_ |
Partial results that reflect the latest migration operation progress. |
PartialResult
Represents a partial result in batch migration operation for one MigrateResourceRequest
.
Fields | |
---|---|
request |
It's the same as the value in |
Union field result . If the resource's migration is ongoing, none of the result will be set. If the resource's migration is finished, either error or one of the migrated resource name will be filled. result can be only one of the following: |
|
error |
The error result of the migration request in case of failure. |
model |
Migrated model resource name. |
dataset |
Migrated dataset resource name. |
BatchMigrateResourcesRequest
Request message for MigrationService.BatchMigrateResources
.
Fields | |
---|---|
parent |
Required. The location of the migrated resource will live in. Format: |
migrate_ |
Required. The request messages specifying the resources to migrate. They must be in the same location as the destination. Up to 50 resources can be migrated in one batch. |
BatchMigrateResourcesResponse
Response message for MigrationService.BatchMigrateResources
.
Fields | |
---|---|
migrate_ |
Successfully migrated resources. |
BatchPredictionJob
A job that uses a Model
to produce predictions on multiple input instances
. If predictions for significant portion of the instances fail, the job may finish without attempting predictions for all remaining instances.
Fields | |
---|---|
name |
Output only. Resource name of the BatchPredictionJob. |
display_ |
Required. The user-defined name of this BatchPredictionJob. |
model |
The name of the Model resource that produces the predictions via this job, must share the same ancestor Location. Starting this job has no impact on any existing deployments of the Model and their resources. Exactly one of model and unmanaged_container_model must be set. The model resource name may contain version id or version alias to specify the version. Example: The model resource could also be a publisher model. Example: |
model_ |
Output only. The version ID of the Model that produces the predictions via this job. |
unmanaged_ |
Contains model information necessary to perform batch prediction without requiring uploading to model registry. Exactly one of model and unmanaged_container_model must be set. |
input_ |
Required. Input configuration of the instances on which predictions are performed. The schema of any single instance may be specified via the |
instance_ |
Configuration for how to convert batch prediction input instances to the prediction instances that are sent to the Model. |
model_ |
The parameters that govern the predictions. The schema of the parameters may be specified via the |
output_ |
Required. The Configuration specifying where output predictions should be written. The schema of any single prediction may be specified as a concatenation of |
dedicated_ |
The config of resources used by the Model during the batch prediction. If the Model |
service_ |
The service account that the DeployedModel's container runs as. If not specified, a system generated one will be used, which has minimal permissions and the custom container, if used, may not have enough permission to access other Google Cloud resources. Users deploying the Model must have the |
manual_ |
Immutable. Parameters configuring the batch behavior. Currently only applicable when |
generate_ |
Generate explanation with the batch prediction results. When set to
If this field is set to true, either the |
explanation_ |
Explanation configuration for this BatchPredictionJob. Can be specified only if This value overrides the value of |
output_ |
Output only. Information further describing the output of this job. |
state |
Output only. The detailed state of the job. |
error |
Output only. Only populated when the job's state is JOB_STATE_FAILED or JOB_STATE_CANCELLED. |
partial_ |
Output only. Partial failures encountered. For example, single files that can't be read. This field never exceeds 20 entries. Status details fields contain standard Google Cloud error details. |
resources_ |
Output only. Information about resources that had been consumed by this job. Provided in real time at best effort basis, as well as a final value once the job completes. Note: This field currently may be not populated for batch predictions that use AutoML Models. |
completion_ |
Output only. Statistics on completed and failed prediction instances. |
create_ |
Output only. Time when the BatchPredictionJob was created. |
start_ |
Output only. Time when the BatchPredictionJob for the first time entered the |
end_ |
Output only. Time when the BatchPredictionJob entered any of the following states: |
update_ |
Output only. Time when the BatchPredictionJob was most recently updated. |
labels |
The labels with user-defined metadata to organize BatchPredictionJobs. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. |
encryption_ |
Customer-managed encryption key options for a BatchPredictionJob. If this is set, then all resources created by the BatchPredictionJob will be encrypted with the provided encryption key. |
disable_ |
For custom-trained Models and AutoML Tabular Models, the container of the DeployedModel instances will send User can disable container logging by setting this flag to true. |
satisfies_ |
Output only. Reserved for future use. |
satisfies_ |
Output only. Reserved for future use. |
InputConfig
Configures the input to BatchPredictionJob
. See Model.supported_input_storage_formats
for Model's supported input formats, and how instances should be expressed via any of them.
Fields | |
---|---|
instances_ |
Required. The format in which instances are given, must be one of the |
Union field source . Required. The source of the input. source can be only one of the following: |
|
gcs_ |
The Cloud Storage location for the input instances. |
bigquery_ |
The BigQuery location of the input table. The schema of the table should be in the format described by the given context OpenAPI Schema, if one is provided. The table may contain additional columns that are not described by the schema, and they will be ignored. |
InstanceConfig
Configuration defining how to transform batch prediction input instances to the instances that the Model accepts.
Fields | |
---|---|
instance_ |
The format of the instance that the Model accepts. Vertex AI will convert compatible Supported values are:
If not specified, Vertex AI converts the batch prediction input as follows:
|
key_ |
The name of the field that is considered as a key. The values identified by the key field is not included in the transformed instances that is sent to the Model. This is similar to specifying this name of the field in
The input must be JSONL with objects at each line, CSV, BigQuery or TfRecord. |
included_ |
Fields that will be included in the prediction instance that is sent to the Model. If When included_fields is populated, The input must be JSONL with objects at each line, BigQuery or TfRecord. |
excluded_ |
Fields that will be excluded in the prediction instance that is sent to the Model. Excluded will be attached to the batch prediction output if When excluded_fields is populated, The input must be JSONL with objects at each line, BigQuery or TfRecord. |
OutputConfig
Configures the output of BatchPredictionJob
. See Model.supported_output_storage_formats
for supported output formats, and how predictions are expressed via any of them.
Fields | |
---|---|
predictions_ |
Required. The format in which Vertex AI gives the predictions, must be one of the |
Union field destination . Required. The destination of the output. destination can be only one of the following: |
|
gcs_ |
The Cloud Storage location of the directory where the output is to be written to. In the given directory a new directory is created. Its name is |
bigquery_ |
The BigQuery project or dataset location where the output is to be written to. If project is provided, a new dataset is created with name |
OutputInfo
Further describes this job's output. Supplements output_config
.
Fields | |
---|---|
bigquery_ |
Output only. The name of the BigQuery table created, in |
Union field output_location . The output location into which prediction output is written. output_location can be only one of the following: |
|
gcs_ |
Output only. The full path of the Cloud Storage directory created, into which the prediction output is written. |
bigquery_ |
Output only. The path of the BigQuery dataset created, in |
BatchReadFeatureValuesOperationMetadata
Details of operations that batch reads Feature values.
Fields | |
---|---|
generic_ |
Operation metadata for Featurestore batch read Features values. |
BatchReadFeatureValuesRequest
Request message for FeaturestoreService.BatchReadFeatureValues
.
Fields | |
---|---|
featurestore |
Required. The resource name of the Featurestore from which to query Feature values. Format: |
destination |
Required. Specifies output location and format. |
pass_ |
When not empty, the specified fields in the *_read_instances source will be joined as-is in the output, in addition to those fields from the Featurestore Entity. For BigQuery source, the type of the pass-through values will be automatically inferred. For CSV source, the pass-through values will be passed as opaque bytes. |
entity_ |
Required. Specifies EntityType grouping Features to read values of and settings. |
start_ |
Optional. Excludes Feature values with feature generation timestamp before this timestamp. If not set, retrieve oldest values kept in Feature Store. Timestamp, if present, must not have higher than millisecond precision. |
Union field
|
|
csv_ |
Each read instance consists of exactly one read timestamp and one or more entity IDs identifying entities of the corresponding EntityTypes whose Features are requested. Each output instance contains Feature values of requested entities concatenated together as of the read time. An example read instance may be An example output instance may be Timestamp in each read instance must be millisecond-aligned.
The columns can be in any order. Values in the timestamp column must use the RFC 3339 format, e.g. |
bigquery_ |
Similar to csv_read_instances, but from BigQuery source. |
EntityTypeSpec
Selects Features of an EntityType to read values of and specifies read settings.
Fields | |
---|---|
entity_ |
Required. ID of the EntityType to select Features. The EntityType id is the |
feature_ |
Required. Selectors choosing which Feature values to read from the EntityType. |
settings[] |
Per-Feature settings for the batch read. |
PassThroughField
Describe pass-through fields in read_instance source.
Fields | |
---|---|
field_ |
Required. The name of the field in the CSV header or the name of the column in BigQuery table. The naming restriction is the same as |
BatchReadFeatureValuesResponse
This type has no fields.
Response message for FeaturestoreService.BatchReadFeatureValues
.
BatchReadTensorboardTimeSeriesDataRequest
Request message for TensorboardService.BatchReadTensorboardTimeSeriesData
.
Fields | |
---|---|
tensorboard |
Required. The resource name of the Tensorboard containing TensorboardTimeSeries to read data from. Format: |
time_ |
Required. The resource names of the TensorboardTimeSeries to read data from. Format: |
BatchReadTensorboardTimeSeriesDataResponse
Response message for TensorboardService.BatchReadTensorboardTimeSeriesData
.
Fields | |
---|---|
time_ |
The returned time series data. |
BigQueryDestination
The BigQuery location for the output content.
Fields | |
---|---|
output_ |
Required. BigQuery URI to a project or table, up to 2000 characters long. When only the project is specified, the Dataset and Table is created. When the full table reference is specified, the Dataset must exist and table must not exist. Accepted forms:
|
BigQuerySource
The BigQuery location for the input content.
Fields | |
---|---|
input_ |
Required. BigQuery URI to a table, up to 2000 characters long. Accepted forms:
|
BleuInput
Input for bleu metric.
Fields | |
---|---|
metric_ |
Required. Spec for bleu score metric. |
instances[] |
Required. Repeated bleu instances. |
BleuInstance
Spec for bleu instance.
Fields | |
---|---|
prediction |
Required. Output of the evaluated model. |
reference |
Required. Ground truth used to compare against the prediction. |
BleuMetricValue
Bleu metric value for an instance.
Fields | |
---|---|
score |
Output only. Bleu score. |
BleuResults
Results for bleu metric.
Fields | |
---|---|
bleu_ |
Output only. Bleu metric values. |
BleuSpec
Spec for bleu score metric - calculates the precision of n-grams in the prediction as compared to reference - returns a score ranging between 0 to 1.
Fields | |
---|---|
use_ |
Optional. Whether to use_effective_order to compute bleu score. |
Blob
Content blob.
Fields | |
---|---|
mime_ |
Required. The IANA standard MIME type of the source data. |
data |
Required. Raw bytes. |
BlurBaselineConfig
Config for blur baseline.
When enabled, a linear path from the maximally blurred image to the input image is created. Using a blurred baseline instead of zero (black image) is motivated by the BlurIG approach explained here: https://arxiv.org/abs/2004.03383
Fields | |
---|---|
max_ |
The standard deviation of the blur kernel for the blurred baseline. The same blurring parameter is used for both the height and the width dimension. If not set, the method defaults to the zero (i.e. black for images) baseline. |
BoolArray
A list of boolean values.
Fields | |
---|---|
values[] |
A list of bool values. |
CancelBatchPredictionJobRequest
Request message for JobService.CancelBatchPredictionJob
.
Fields | |
---|---|
name |
Required. The name of the BatchPredictionJob to cancel. Format: |
CancelCustomJobRequest
Request message for JobService.CancelCustomJob
.
Fields | |
---|---|
name |
Required. The name of the CustomJob to cancel. Format: |
CancelHyperparameterTuningJobRequest
Request message for JobService.CancelHyperparameterTuningJob
.
Fields | |
---|---|
name |
Required. The name of the HyperparameterTuningJob to cancel. Format: |
CancelNasJobRequest
Request message for JobService.CancelNasJob
.
Fields | |
---|---|
name |
Required. The name of the NasJob to cancel. Format: |
CancelPipelineJobRequest
Request message for PipelineService.CancelPipelineJob
.
Fields | |
---|---|
name |
Required. The name of the PipelineJob to cancel. Format: |
CancelTrainingPipelineRequest
Request message for PipelineService.CancelTrainingPipeline
.
Fields | |
---|---|
name |
Required. The name of the TrainingPipeline to cancel. Format: |
CancelTuningJobRequest
Request message for GenAiTuningService.CancelTuningJob
.
Fields | |
---|---|
name |
Required. The name of the TuningJob to cancel. Format: |
Candidate
A response candidate generated from the model.
Fields | |
---|---|
index |
Output only. Index of the candidate. |
content |
Output only. Content parts of the candidate. |
avg_ |
Output only. Average log probability score of the candidate. |
logprobs_ |
Output only. Log-likelihood scores for the response tokens and top tokens |
finish_ |
Output only. The reason why the model stopped generating tokens. If empty, the model has not stopped generating the tokens. |
safety_ |
Output only. List of ratings for the safety of a response candidate. There is at most one rating per category. |
citation_ |
Output only. Source attribution of the generated content. |
grounding_ |
Output only. Metadata specifies sources used to ground generated content. |
finish_ |
Output only. Describes the reason the mode stopped generating tokens in more detail. This is only filled when |
FinishReason
The reason why the model stopped generating tokens. If empty, the model has not stopped generating the tokens.
Enums | |
---|---|
FINISH_REASON_UNSPECIFIED |
The finish reason is unspecified. |
STOP |
Token generation reached a natural stopping point or a configured stop sequence. |
MAX_TOKENS |
Token generation reached the configured maximum output tokens. |
SAFETY |
Token generation stopped because the content potentially contains safety violations. NOTE: When streaming, content is empty if content filters blocks the output. |
RECITATION |
The token generation stopped because of potential recitation. |
OTHER |
All other reasons that stopped the token generation. |
BLOCKLIST |
Token generation stopped because the content contains forbidden terms. |
PROHIBITED_CONTENT |
Token generation stopped for potentially containing prohibited content. |
SPII |
Token generation stopped because the content potentially contains Sensitive Personally Identifiable Information (SPII). |
MALFORMED_FUNCTION_CALL |
The function call generated by the model is invalid. |
CheckTrialEarlyStoppingStateMetatdata
This message will be placed in the metadata field of a google.longrunning.Operation associated with a CheckTrialEarlyStoppingState request.
Fields | |
---|---|
generic_ |
Operation metadata for suggesting Trials. |
study |
The name of the Study that the Trial belongs to. |
trial |
The Trial name. |
CheckTrialEarlyStoppingStateRequest
Request message for VizierService.CheckTrialEarlyStoppingState
.
Fields | |
---|---|
trial_ |
Required. The Trial's name. Format: |
CheckTrialEarlyStoppingStateResponse
Response message for VizierService.CheckTrialEarlyStoppingState
.
Fields | |
---|---|
should_ |
True if the Trial should stop. |
Citation
Source attributions for content.
Fields | |
---|---|
start_ |
Output only. Start index into the content. |
end_ |
Output only. End index into the content. |
uri |
Output only. Url reference of the attribution. |
title |
Output only. Title of the attribution. |
license |
Output only. License of the attribution. |
publication_ |
Output only. Publication date of the attribution. |
CitationMetadata
A collection of source attributions for a piece of content.
Fields | |
---|---|
citations[] |
Output only. List of citations. |
Claim
Claim that is extracted from the input text and facts that support it.
Fields | |
---|---|
fact_ |
Indexes of the facts supporting this claim. |
start_ |
Index in the input text where the claim starts (inclusive). |
end_ |
Index in the input text where the claim ends (exclusive). |
score |
Confidence score of this corroboration. |
ClientConnectionConfig
Configurations (e.g. inference timeout) that are applied on your endpoints.
Fields | |
---|---|
inference_ |
Customizable online prediction request timeout. |
CoherenceInput
Input for coherence metric.
Fields | |
---|---|
metric_ |
Required. Spec for coherence score metric. |
instance |
Required. Coherence instance. |
CoherenceInstance
Spec for coherence instance.
Fields | |
---|---|
prediction |
Required. Output of the evaluated model. |
CoherenceResult
Spec for coherence result.
Fields | |
---|---|
explanation |
Output only. Explanation for coherence score. |
score |
Output only. Coherence score. |
confidence |
Output only. Confidence for coherence score. |
CoherenceSpec
Spec for coherence score metric.
Fields | |
---|---|
version |
Optional. Which version to use for evaluation. |
CometInput
Input for Comet metric.
Fields | |
---|---|
metric_ |
Required. Spec for comet metric. |
instance |
Required. Comet instance. |
CometInstance
Spec for Comet instance - The fields used for evaluation are dependent on the comet version.
Fields | |
---|---|
prediction |
Required. Output of the evaluated model. |
reference |
Optional. Ground truth used to compare against the prediction. |
source |
Optional. Source text in original language. |
CometResult
Spec for Comet result - calculates the comet score for the given instance using the version specified in the spec.
Fields | |
---|---|
score |
Output only. Comet score. Range depends on version. |
CometSpec
Spec for Comet metric.
Fields | |
---|---|
source_ |
Optional. Source language in BCP-47 format. |
target_ |
Optional. Target language in BCP-47 format. Covers both prediction and reference. |
version |
Required. Which version to use for evaluation. |
CometVersion
Comet version options.
Enums | |
---|---|
COMET_VERSION_UNSPECIFIED |
Comet version unspecified. |
COMET_22_SRC_REF |
Comet 22 for translation + source + reference (source-reference-combined). |
CompleteTrialRequest
Request message for VizierService.CompleteTrial
.
Fields | |
---|---|
name |
Required. The Trial's name. Format: |
final_ |
Optional. If provided, it will be used as the completed Trial's final_measurement; Otherwise, the service will auto-select a previously reported measurement as the final-measurement |
trial_ |
Optional. True if the Trial cannot be run with the given Parameter, and final_measurement will be ignored. |
infeasible_ |
Optional. A human readable reason why the trial was infeasible. This should only be provided if |
CompletionStats
Success and error statistics of processing multiple entities (for example, DataItems or structured data rows) in batch.
Fields | |
---|---|
successful_ |
Output only. The number of entities that had been processed successfully. |
failed_ |
Output only. The number of entities for which any error was encountered. |
incomplete_ |
Output only. In cases when enough errors are encountered a job, pipeline, or operation may be failed as a whole. Below is the number of entities for which the processing had not been finished (either in successful or failed state). Set to -1 if the number is unknown (for example, the operation failed before the total entity number could be collected). |
successful_ |
Output only. The number of the successful forecast points that are generated by the forecasting model. This is ONLY used by the forecasting batch prediction. |
ComputeTokensRequest
Request message for ComputeTokens RPC call.
Fields | |
---|---|
endpoint |
Required. The name of the Endpoint requested to get lists of tokens and token ids. |
instances[] |
Optional. The instances that are the input to token computing API call. Schema is identical to the prediction schema of the text model, even for the non-text models, like chat models, or Codey models. |
model |
Optional. The name of the publisher model requested to serve the prediction. Format: projects/{project}/locations/{location}/publishers/*/models/* |
contents[] |
Optional. Input content. |
ComputeTokensResponse
Response message for ComputeTokens RPC call.
Fields | |
---|---|
tokens_ |
Lists of tokens info from the input. A ComputeTokensRequest could have multiple instances with a prompt in each instance. We also need to return lists of tokens info for the request with multiple instances. |
ContainerRegistryDestination
The Container Registry location for the container image.
Fields | |
---|---|
output_ |
Required. Container Registry URI of a container image. Only Google Container Registry and Artifact Registry are supported now. Accepted forms:
If a tag is not specified, "latest" will be used as the default tag. |
ContainerSpec
The spec of a Container.
Fields | |
---|---|
image_ |
Required. The URI of a container image in the Container Registry that is to be run on each worker replica. |
command[] |
The command to be invoked when the container is started. It overrides the entrypoint instruction in Dockerfile when provided. |
args[] |
The arguments to be passed when starting the container. |
env[] |
Environment variables to be passed to the container. Maximum limit is 100. |
Content
The base structured datatype containing multi-part content of a message.
A Content
includes a role
field designating the producer of the Content
and a parts
field containing multi-part data that contains the content of the message turn.
Fields | |
---|---|
role |
Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. |
parts[] |
Required. Ordered |
Context
Instance of a general context.
Fields | |
---|---|
name |
Immutable. The resource name of the Context. |
display_ |
User provided display name of the Context. May be up to 128 Unicode characters. |
etag |
An eTag used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
labels |
The labels with user-defined metadata to organize your Contexts. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Context (System labels are excluded). |
create_ |
Output only. Timestamp when this Context was created. |
update_ |
Output only. Timestamp when this Context was last updated. |
parent_ |
Output only. A list of resource names of Contexts that are parents of this Context. A Context may have at most 10 parent_contexts. |
schema_ |
The title of the schema describing the metadata. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store. |
schema_ |
The version of the schema in schema_name to use. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store. |
metadata |
Properties of the Context. Top level metadata keys' heading and trailing spaces will be trimmed. The size of this field should not exceed 200KB. |
description |
Description of the Context |
CopyModelOperationMetadata
Details of ModelService.CopyModel
operation.
Fields | |
---|---|
generic_ |
The common part of the operation metadata. |
CopyModelRequest
Request message for ModelService.CopyModel
.
Fields | |
---|---|
parent |
Required. The resource name of the Location into which to copy the Model. Format: |
source_ |
Required. The resource name of the Model to copy. That Model must be in the same Project. Format: |
encryption_ |
Customer-managed encryption key options. If this is set, then the Model copy will be encrypted with the provided encryption key. |
Union field destination_model . If both fields are unset, a new Model will be created with a generated ID. destination_model can be only one of the following: |
|
model_ |
Optional. Copy source_model into a new Model with this ID. The ID will become the final component of the model resource name. This value may be up to 63 characters, and valid characters are |
parent_ |
Optional. Specify this field to copy source_model into this existing Model as a new version. Format: |
CopyModelResponse
Response message of ModelService.CopyModel
operation.
Fields | |
---|---|
model |
The name of the copied Model resource. Format: |
model_ |
Output only. The version ID of the model that is copied. |
CorpusStatus
RagCorpus status.
Fields | |
---|---|
state |
Output only. RagCorpus life state. |
error_ |
Output only. Only when the |
State
RagCorpus life state.
Enums | |
---|---|
UNKNOWN |
This state is not supposed to happen. |
INITIALIZED |
RagCorpus resource entry is initialized, but hasn't done validation. |
ACTIVE |
RagCorpus is provisioned successfully and is ready to serve. |
ERROR |
RagCorpus is in a problematic situation. See error_message field for details. |
CorroborateContentRequest
Request message for CorroborateContent.
Fields | |
---|---|
parent |
Required. The resource name of the Location from which to corroborate text. The users must have permission to make a call in the project. Format: |
facts[] |
Optional. Facts used to generate the text can also be used to corroborate the text. |
parameters |
Optional. Parameters that can be set to override default settings per request. |
content |
Optional. Input content to corroborate, only text format is supported for now. |
Parameters
Parameters that can be overrided per request.
Fields | |
---|---|
citation_ |
Optional. Only return claims with citation score larger than the threshold. |
CorroborateContentResponse
Response message for CorroborateContent.
Fields | |
---|---|
claims[] |
Claims that are extracted from the input content and facts that support the claims. |
corroboration_ |
Confidence score of corroborating content. Value is [0,1] with 1 is the most confidence. |
CountTokensRequest
Request message for [PredictionService.CountTokens][].
Fields | |
---|---|
endpoint |
Required. The name of the Endpoint requested to perform token counting. Format: |
model |
Optional. The name of the publisher model requested to serve the prediction. Format: |
instances[] |
Optional. The instances that are the input to token counting call. Schema is identical to the prediction schema of the underlying model. |
contents[] |
Optional. Input content. |
tools[] |
Optional. A list of A |
system_ |
Optional. The user provided system instructions for the model. Note: only text should be used in parts and content in each part will be in a separate paragraph. |
generation_ |
Optional. Generation config that the model will use to generate the response. |
CountTokensResponse
Response message for [PredictionService.CountTokens][].
Fields | |
---|---|
total_ |
The total number of tokens counted across all instances from the request. |
total_ |
The total number of billable characters counted across all instances from the request. |
CreateArtifactRequest
Request message for MetadataService.CreateArtifact
.
Fields | |
---|---|
parent |
Required. The resource name of the MetadataStore where the Artifact should be created. Format: |
artifact |
Required. The Artifact to create. |
artifact_ |
The {artifact} portion of the resource name with the format: |
CreateBatchPredictionJobRequest
Request message for JobService.CreateBatchPredictionJob
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to create the BatchPredictionJob in. Format: |
batch_ |
Required. The BatchPredictionJob to create. |
CreateContextRequest
Request message for MetadataService.CreateContext
.
Fields | |
---|---|
parent |
Required. The resource name of the MetadataStore where the Context should be created. Format: |
context |
Required. The Context to create. |
context_ |
The {context} portion of the resource name with the format: |
CreateCustomJobRequest
Request message for JobService.CreateCustomJob
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to create the CustomJob in. Format: |
custom_ |
Required. The CustomJob to create. |
CreateDatasetOperationMetadata
Runtime operation information for DatasetService.CreateDataset
.
Fields | |
---|---|
generic_ |
The operation generic information. |
CreateDatasetRequest
Request message for DatasetService.CreateDataset
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to create the Dataset in. Format: |
dataset |
Required. The Dataset to create. |
CreateDatasetVersionOperationMetadata
Runtime operation information for DatasetService.CreateDatasetVersion
.
Fields | |
---|---|
generic_ |
The common part of the operation metadata. |
CreateDatasetVersionRequest
Request message for DatasetService.CreateDatasetVersion
.
Fields | |
---|---|
parent |
Required. The name of the Dataset resource. Format: |
dataset_ |
Required. The version to be created. The same CMEK policies with the original Dataset will be applied the dataset version. So here we don't need to specify the EncryptionSpecType here. |
CreateDeploymentResourcePoolOperationMetadata
Runtime operation information for CreateDeploymentResourcePool method.
Fields | |
---|---|
generic_ |
The operation generic information. |
CreateDeploymentResourcePoolRequest
Request message for CreateDeploymentResourcePool method.
Fields | |
---|---|
parent |
Required. The parent location resource where this DeploymentResourcePool will be created. Format: |
deployment_ |
Required. The DeploymentResourcePool to create. |
deployment_ |
Required. The ID to use for the DeploymentResourcePool, which will become the final component of the DeploymentResourcePool's resource name. The maximum length is 63 characters, and valid characters are |
CreateEndpointOperationMetadata
Runtime operation information for EndpointService.CreateEndpoint
.
Fields | |
---|---|
generic_ |
The operation generic information. |
CreateEndpointRequest
Request message for EndpointService.CreateEndpoint
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to create the Endpoint in. Format: |
endpoint |
Required. The Endpoint to create. |
endpoint_ |
Immutable. The ID to use for endpoint, which will become the final component of the endpoint resource name. If not provided, Vertex AI will generate a value for this ID. If the first character is a letter, this value may be up to 63 characters, and valid characters are If the first character is a number, this value may be up to 9 characters, and valid characters are When using HTTP/JSON, this field is populated based on a query string argument, such as |
CreateEntityTypeOperationMetadata
Details of operations that perform create EntityType.
Fields | |
---|---|
generic_ |
Operation metadata for EntityType. |
CreateEntityTypeRequest
Request message for FeaturestoreService.CreateEntityType
.
Fields | |
---|---|
parent |
Required. The resource name of the Featurestore to create EntityTypes. Format: |
entity_ |
The EntityType to create. |
entity_ |
Required. The ID to use for the EntityType, which will become the final component of the EntityType's resource name. This value may be up to 60 characters, and valid characters are The value must be unique within a featurestore. |
CreateExecutionRequest
Request message for MetadataService.CreateExecution
.
Fields | |
---|---|
parent |
Required. The resource name of the MetadataStore where the Execution should be created. Format: |
execution |
Required. The Execution to create. |
execution_ |
The {execution} portion of the resource name with the format: |
CreateFeatureGroupOperationMetadata
Details of operations that perform create FeatureGroup.
Fields | |
---|---|
generic_ |
Operation metadata for FeatureGroup. |
CreateFeatureGroupRequest
Request message for FeatureRegistryService.CreateFeatureGroup
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to create FeatureGroups. Format: |
feature_ |
Required. The FeatureGroup to create. |
feature_ |
Required. The ID to use for this FeatureGroup, which will become the final component of the FeatureGroup's resource name. This value may be up to 128 characters, and valid characters are The value must be unique within the project and location. |
CreateFeatureOnlineStoreOperationMetadata
Details of operations that perform create FeatureOnlineStore.
Fields | |
---|---|
generic_ |
Operation metadata for FeatureOnlineStore. |
CreateFeatureOnlineStoreRequest
Request message for FeatureOnlineStoreAdminService.CreateFeatureOnlineStore
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to create FeatureOnlineStores. Format: |
feature_ |
Required. The FeatureOnlineStore to create. |
feature_ |
Required. The ID to use for this FeatureOnlineStore, which will become the final component of the FeatureOnlineStore's resource name. This value may be up to 60 characters, and valid characters are The value must be unique within the project and location. |
CreateFeatureOperationMetadata
Details of operations that perform create Feature.
Fields | |
---|---|
generic_ |
Operation metadata for Feature. |
CreateFeatureRequest
Request message for FeaturestoreService.CreateFeature
. Request message for FeatureRegistryService.CreateFeature
.
Fields | |
---|---|
parent |
Required. The resource name of the EntityType or FeatureGroup to create a Feature. Format for entity_type as parent: |
feature |
Required. The Feature to create. |
feature_ |
Required. The ID to use for the Feature, which will become the final component of the Feature's resource name. This value may be up to 128 characters, and valid characters are The value must be unique within an EntityType/FeatureGroup. |
CreateFeatureViewOperationMetadata
Details of operations that perform create FeatureView.
Fields | |
---|---|
generic_ |
Operation metadata for FeatureView Create. |
CreateFeatureViewRequest
Request message for FeatureOnlineStoreAdminService.CreateFeatureView
.
Fields | |
---|---|
parent |
Required. The resource name of the FeatureOnlineStore to create FeatureViews. Format: |
feature_ |
Required. The FeatureView to create. |
feature_ |
Required. The ID to use for the FeatureView, which will become the final component of the FeatureView's resource name. This value may be up to 60 characters, and valid characters are The value must be unique within a FeatureOnlineStore. |
run_ |
Immutable. If set to true, one on demand sync will be run immediately, regardless whether the |
CreateFeaturestoreOperationMetadata
Details of operations that perform create Featurestore.
Fields | |
---|---|
generic_ |
Operation metadata for Featurestore. |
CreateFeaturestoreRequest
Request message for FeaturestoreService.CreateFeaturestore
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to create Featurestores. Format: |
featurestore |
Required. The Featurestore to create. |
featurestore_ |
Required. The ID to use for this Featurestore, which will become the final component of the Featurestore's resource name. This value may be up to 60 characters, and valid characters are The value must be unique within the project and location. |
CreateHyperparameterTuningJobRequest
Request message for JobService.CreateHyperparameterTuningJob
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to create the HyperparameterTuningJob in. Format: |
hyperparameter_ |
Required. The HyperparameterTuningJob to create. |
CreateIndexEndpointOperationMetadata
Runtime operation information for IndexEndpointService.CreateIndexEndpoint
.
Fields | |
---|---|
generic_ |
The operation generic information. |
CreateIndexEndpointRequest
Request message for IndexEndpointService.CreateIndexEndpoint
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to create the IndexEndpoint in. Format: |
index_ |
Required. The IndexEndpoint to create. |
CreateIndexOperationMetadata
Runtime operation information for IndexService.CreateIndex
.
Fields | |
---|---|
generic_ |
The operation generic information. |
nearest_ |
The operation metadata with regard to Matching Engine Index operation. |
CreateIndexRequest
Request message for IndexService.CreateIndex
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to create the Index in. Format: |
index |
Required. The Index to create. |
CreateMetadataSchemaRequest
Request message for MetadataService.CreateMetadataSchema
.
Fields | |
---|---|
parent |
Required. The resource name of the MetadataStore where the MetadataSchema should be created. Format: |
metadata_ |
Required. The MetadataSchema to create. |
metadata_ |
The {metadata_schema} portion of the resource name with the format: |
CreateMetadataStoreOperationMetadata
Details of operations that perform MetadataService.CreateMetadataStore
.
Fields | |
---|---|
generic_ |
Operation metadata for creating a MetadataStore. |
CreateMetadataStoreRequest
Request message for MetadataService.CreateMetadataStore
.
Fields | |
---|---|
parent |
Required. The resource name of the Location where the MetadataStore should be created. Format: |
metadata_ |
Required. The MetadataStore to create. |
metadata_ |
The {metadatastore} portion of the resource name with the format: |
CreateModelDeploymentMonitoringJobRequest
Request message for JobService.CreateModelDeploymentMonitoringJob
.
Fields | |
---|---|
parent |
Required. The parent of the ModelDeploymentMonitoringJob. Format: |
model_ |
Required. The ModelDeploymentMonitoringJob to create |
CreateNasJobRequest
Request message for JobService.CreateNasJob
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to create the NasJob in. Format: |
nas_ |
Required. The NasJob to create. |
CreateNotebookExecutionJobOperationMetadata
Metadata information for NotebookService.CreateNotebookExecutionJob
.
Fields | |
---|---|
generic_ |
The operation generic information. |
progress_ |
A human-readable message that shows the intermediate progress details of NotebookRuntime. |
CreateNotebookExecutionJobRequest
Request message for [NotebookService.CreateNotebookExecutionJob]
Fields | |
---|---|
parent |
Required. The resource name of the Location to create the NotebookExecutionJob. Format: |
notebook_ |
Required. The NotebookExecutionJob to create. |
notebook_ |
Optional. User specified ID for the NotebookExecutionJob. |
CreateNotebookRuntimeTemplateOperationMetadata
Metadata information for NotebookService.CreateNotebookRuntimeTemplate
.
Fields | |
---|---|
generic_ |
The operation generic information. |
CreateNotebookRuntimeTemplateRequest
Request message for NotebookService.CreateNotebookRuntimeTemplate
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to create the NotebookRuntimeTemplate. Format: |
notebook_ |
Required. The NotebookRuntimeTemplate to create. |
notebook_ |
Optional. User specified ID for the notebook runtime template. |
CreatePersistentResourceOperationMetadata
Details of operations that perform create PersistentResource.
Fields | |
---|---|
generic_ |
Operation metadata for PersistentResource. |
progress_ |
Progress Message for Create LRO |
CreatePersistentResourceRequest
Request message for PersistentResourceService.CreatePersistentResource
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to create the PersistentResource in. Format: |
persistent_ |
Required. The PersistentResource to create. |
persistent_ |
Required. The ID to use for the PersistentResource, which become the final component of the PersistentResource's resource name. The maximum length is 63 characters, and valid characters are |
CreatePipelineJobRequest
Request message for PipelineService.CreatePipelineJob
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to create the PipelineJob in. Format: |
pipeline_ |
Required. The PipelineJob to create. |
pipeline_ |
The ID to use for the PipelineJob, which will become the final component of the PipelineJob name. If not provided, an ID will be automatically generated. This value should be less than 128 characters, and valid characters are |
CreateRagCorpusOperationMetadata
Runtime operation information for VertexRagDataService.CreateRagCorpus
.
Fields | |
---|---|
generic_ |
The operation generic information. |
CreateRagCorpusRequest
Request message for VertexRagDataService.CreateRagCorpus
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to create the RagCorpus in. Format: |
rag_ |
Required. The RagCorpus to create. |
CreateRegistryFeatureOperationMetadata
Details of operations that perform create FeatureGroup.
Fields | |
---|---|
generic_ |
Operation metadata for Feature. |
CreateScheduleRequest
Request message for ScheduleService.CreateSchedule
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to create the Schedule in. Format: |
schedule |
Required. The Schedule to create. |
CreateSpecialistPoolOperationMetadata
Runtime operation information for SpecialistPoolService.CreateSpecialistPool
.
Fields | |
---|---|
generic_ |
The operation generic information. |
CreateSpecialistPoolRequest
Request message for SpecialistPoolService.CreateSpecialistPool
.
Fields | |
---|---|
parent |
Required. The parent Project name for the new SpecialistPool. The form is |
specialist_ |
Required. The SpecialistPool to create. |
CreateStudyRequest
Request message for VizierService.CreateStudy
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to create the CustomJob in. Format: |
study |
Required. The Study configuration used to create the Study. |
CreateTensorboardExperimentRequest
Request message for TensorboardService.CreateTensorboardExperiment
.
Fields | |
---|---|
parent |
Required. The resource name of the Tensorboard to create the TensorboardExperiment in. Format: |
tensorboard_ |
The TensorboardExperiment to create. |
tensorboard_ |
Required. The ID to use for the Tensorboard experiment, which becomes the final component of the Tensorboard experiment's resource name. This value should be 1-128 characters, and valid characters are |
CreateTensorboardOperationMetadata
Details of operations that perform create Tensorboard.
Fields | |
---|---|
generic_ |
Operation metadata for Tensorboard. |
CreateTensorboardRequest
Request message for TensorboardService.CreateTensorboard
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to create the Tensorboard in. Format: |
tensorboard |
Required. The Tensorboard to create. |
CreateTensorboardRunRequest
Request message for TensorboardService.CreateTensorboardRun
.
Fields | |
---|---|
parent |
Required. The resource name of the TensorboardExperiment to create the TensorboardRun in. Format: |
tensorboard_ |
Required. The TensorboardRun to create. |
tensorboard_ |
Required. The ID to use for the Tensorboard run, which becomes the final component of the Tensorboard run's resource name. This value should be 1-128 characters, and valid characters are |
CreateTensorboardTimeSeriesRequest
Request message for TensorboardService.CreateTensorboardTimeSeries
.
Fields | |
---|---|
parent |
Required. The resource name of the TensorboardRun to create the TensorboardTimeSeries in. Format: |
tensorboard_ |
Optional. The user specified unique ID to use for the TensorboardTimeSeries, which becomes the final component of the TensorboardTimeSeries's resource name. This value should match "[a-z0-9][a-z0-9-]{0, 127}" |
tensorboard_ |
Required. The TensorboardTimeSeries to create. |
CreateTrainingPipelineRequest
Request message for PipelineService.CreateTrainingPipeline
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to create the TrainingPipeline in. Format: |
training_ |
Required. The TrainingPipeline to create. |
CreateTrialRequest
Request message for VizierService.CreateTrial
.
Fields | |
---|---|
parent |
Required. The resource name of the Study to create the Trial in. Format: |
trial |
Required. The Trial to create. |
CreateTuningJobRequest
Request message for GenAiTuningService.CreateTuningJob
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to create the TuningJob in. Format: |
tuning_ |
Required. The TuningJob to create. |
CsvDestination
The storage details for CSV output content.
Fields | |
---|---|
gcs_ |
Required. Google Cloud Storage location. |
CsvSource
The storage details for CSV input content.
Fields | |
---|---|
gcs_ |
Required. Google Cloud Storage location. |
CustomJob
Represents a job that runs custom workloads such as a Docker container or a Python package. A CustomJob can have multiple worker pools and each worker pool can have its own machine and input spec. A CustomJob will be cleaned up once the job enters terminal state (failed or succeeded).
Fields | |
---|---|
name |
Output only. Resource name of a CustomJob. |
display_ |
Required. The display name of the CustomJob. The name can be up to 128 characters long and can consist of any UTF-8 characters. |
job_ |
Required. Job spec. |
state |
Output only. The detailed state of the job. |
create_ |
Output only. Time when the CustomJob was created. |
start_ |
Output only. Time when the CustomJob for the first time entered the |
end_ |
Output only. Time when the CustomJob entered any of the following states: |
update_ |
Output only. Time when the CustomJob was most recently updated. |
error |
Output only. Only populated when job's state is |
labels |
The labels with user-defined metadata to organize CustomJobs. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. |
encryption_ |
Customer-managed encryption key options for a CustomJob. If this is set, then all resources created by the CustomJob will be encrypted with the provided encryption key. |
web_ |
Output only. URIs for accessing interactive shells (one URI for each training node). Only available if The keys are names of each node in the training job; for example, The values are the URIs for each node's interactive shell. |
satisfies_ |
Output only. Reserved for future use. |
satisfies_ |
Output only. Reserved for future use. |
CustomJobSpec
Represents the spec of a CustomJob.
Fields | |
---|---|
persistent_ |
Optional. The ID of the PersistentResource in the same Project and Location which to run If this is specified, the job will be run on existing machines held by the PersistentResource instead of on-demand short-live machines. The network and CMEK configs on the job should be consistent with those on the PersistentResource, otherwise, the job will be rejected. |
worker_ |
Required. The spec of the worker pools including machine type and Docker image. All worker pools except the first one are optional and can be skipped by providing an empty value. |
scheduling |
Scheduling options for a CustomJob. |
service_ |
Specifies the service account for workload run-as account. Users submitting jobs must have act-as permission on this run-as account. If unspecified, the Vertex AI Custom Code Service Agent for the CustomJob's project is used. |
network |
Optional. The full name of the Compute Engine network to which the Job should be peered. For example, To specify this field, you must have already configured VPC Network Peering for Vertex AI. If this field is left unspecified, the job is not peered with any network. |
reserved_ |
Optional. A list of names for the reserved ip ranges under the VPC network that can be used for this job. If set, we will deploy the job within the provided ip ranges. Otherwise, the job will be deployed to any ip ranges under the provided VPC network. Example: ['vertex-ai-ip-range']. |
base_ |
The Cloud Storage location to store the output of this CustomJob or HyperparameterTuningJob. For HyperparameterTuningJob, the baseOutputDirectory of each child CustomJob backing a Trial is set to a subdirectory of name The following Vertex AI environment variables will be passed to containers or python modules when this field is set: For CustomJob:
For CustomJob backing a Trial of HyperparameterTuningJob:
|
protected_ |
The ID of the location to store protected artifacts. e.g. us-central1. Populate only when the location is different than CustomJob location. List of supported locations: https://cloud.google.com/vertex-ai/docs/general/locations |
tensorboard |
Optional. The name of a Vertex AI |
enable_ |
Optional. Whether you want Vertex AI to enable interactive shell access to training containers. If set to |
enable_ |
Optional. Whether you want Vertex AI to enable access to the customized dashboard in training chief container. If set to |
experiment |
Optional. The Experiment associated with this job. Format: |
experiment_ |
Optional. The Experiment Run associated with this job. Format: |
models[] |
Optional. The name of the Model resources for which to generate a mapping to artifact URIs. Applicable only to some of the Google-provided custom jobs. Format: In order to retrieve a specific version of the model, also provide the version ID or version alias. Example: |
DataItem
A piece of data in a Dataset. Could be an image, a video, a document or plain text.
Fields | |
---|---|
name |
Output only. The resource name of the DataItem. |
create_ |
Output only. Timestamp when this DataItem was created. |
update_ |
Output only. Timestamp when this DataItem was last updated. |
labels |
Optional. The labels with user-defined metadata to organize your DataItems. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one DataItem(System labels are excluded). See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. |
payload |
Required. The data that the DataItem represents (for example, an image or a text snippet). The schema of the payload is stored in the parent Dataset's |
etag |
Optional. Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
satisfies_ |
Output only. Reserved for future use. |
satisfies_ |
Output only. Reserved for future use. |
DataItemView
A container for a single DataItem and Annotations on it.
Fields | |
---|---|
data_ |
The DataItem. |
annotations[] |
The Annotations on the DataItem. If too many Annotations should be returned for the DataItem, this field will be truncated per annotations_limit in request. If it was, then the has_truncated_annotations will be set to true. |
has_ |
True if and only if the Annotations field has been truncated. It happens if more Annotations for this DataItem met the request's annotation_filter than are allowed to be returned by annotations_limit. Note that if Annotations field is not being returned due to field mask, then this field will not be set to true no matter how many Annotations are there. |
Dataset
A collection of DataItems and Annotations on them.
Fields | |
---|---|
name |
Output only. Identifier. The resource name of the Dataset. Format: |
display_ |
Required. The user-defined name of the Dataset. The name can be up to 128 characters long and can consist of any UTF-8 characters. |
description |
The description of the Dataset. |
metadata_ |
Required. Points to a YAML file stored on Google Cloud Storage describing additional information about the Dataset. The schema is defined as an OpenAPI 3.0.2 Schema Object. The schema files that can be used here are found in gs://google-cloud-aiplatform/schema/dataset/metadata/. |
metadata |
Required. Additional information about the Dataset. |
data_ |
Output only. The number of DataItems in this Dataset. Only apply for non-structured Dataset. |
create_ |
Output only. Timestamp when this Dataset was created. |
update_ |
Output only. Timestamp when this Dataset was last updated. |
etag |
Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
labels |
The labels with user-defined metadata to organize your Datasets. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Dataset (System labels are excluded). See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. Following system labels exist for each Dataset:
|
saved_ |
All SavedQueries belong to the Dataset will be returned in List/Get Dataset response. The annotation_specs field will not be populated except for UI cases which will only use |
encryption_ |
Customer-managed encryption key spec for a Dataset. If set, this Dataset and all sub-resources of this Dataset will be secured by this key. |
metadata_ |
Output only. The resource name of the Artifact that was created in MetadataStore when creating the Dataset. The Artifact resource name pattern is |
model_ |
Optional. Reference to the public base model last used by the dataset. Only set for prompt datasets. |
satisfies_ |
Output only. Reserved for future use. |
satisfies_ |
Output only. Reserved for future use. |
DatasetVersion
Describes the dataset version.
Fields | |
---|---|
name |
Output only. Identifier. The resource name of the DatasetVersion. |
create_ |
Output only. Timestamp when this DatasetVersion was created. |
update_ |
Output only. Timestamp when this DatasetVersion was last updated. |
etag |
Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
big_ |
Output only. Name of the associated BigQuery dataset. |
display_ |
The user-defined name of the DatasetVersion. The name can be up to 128 characters long and can consist of any UTF-8 characters. |
metadata |
Required. Output only. Additional information about the DatasetVersion. |
model_ |
Output only. Reference to the public base model last used by the dataset version. Only set for prompt dataset versions. |
satisfies_ |
Output only. Reserved for future use. |
satisfies_ |
Output only. Reserved for future use. |
DedicatedResources
A description of resources that are dedicated to a DeployedModel, and that need a higher degree of manual configuration.
Fields | |
---|---|
machine_ |
Required. Immutable. The specification of a single machine used by the prediction. |
min_ |
Required. Immutable. The minimum number of machine replicas this DeployedModel will be always deployed on. This value must be greater than or equal to 1. If traffic against the DeployedModel increases, it may dynamically be deployed onto more replicas, and as traffic decreases, some of these extra replicas may be freed. |
max_ |
Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, will use The value of this field impacts the charge against Vertex CPU and GPU quotas. Specifically, you will be charged for (max_replica_count * number of cores in the selected machine type) and (max_replica_count * number of GPUs per replica in the selected machine type). |
required_ |
Optional. Number of required available replicas for the deployment to succeed. This field is only needed when partial model deployment/mutation is desired. If set, the model deploy/mutate operation will succeed once available_replica_count reaches required_replica_count, and the rest of the replicas will be retried. If not set, the default required_replica_count will be min_replica_count. |
autoscaling_ |
Immutable. The metric specifications that overrides a resource utilization metric (CPU utilization, accelerator's duty cycle, and so on) target value (default to 60 if not set). At most one entry is allowed per metric. If If For example, in the case of Online Prediction, if you want to override target CPU utilization to 80, you should set |
spot |
Optional. If true, schedule the deployment workload on spot VMs. |
DeleteArtifactRequest
Request message for MetadataService.DeleteArtifact
.
Fields | |
---|---|
name |
Required. The resource name of the Artifact to delete. Format: |
etag |
Optional. The etag of the Artifact to delete. If this is provided, it must match the server's etag. Otherwise, the request will fail with a FAILED_PRECONDITION. |
DeleteBatchPredictionJobRequest
Request message for JobService.DeleteBatchPredictionJob
.
Fields | |
---|---|
name |
Required. The name of the BatchPredictionJob resource to be deleted. Format: |
DeleteContextRequest
Request message for MetadataService.DeleteContext
.
Fields | |
---|---|
name |
Required. The resource name of the Context to delete. Format: |
force |
The force deletion semantics is still undefined. Users should not use this field. |
etag |
Optional. The etag of the Context to delete. If this is provided, it must match the server's etag. Otherwise, the request will fail with a FAILED_PRECONDITION. |
DeleteCustomJobRequest
Request message for JobService.DeleteCustomJob
.
Fields | |
---|---|
name |
Required. The name of the CustomJob resource to be deleted. Format: |
DeleteDatasetRequest
Request message for DatasetService.DeleteDataset
.
Fields | |
---|---|
name |
Required. The resource name of the Dataset to delete. Format: |
DeleteDatasetVersionRequest
Request message for DatasetService.DeleteDatasetVersion
.
Fields | |
---|---|
name |
Required. The resource name of the Dataset version to delete. Format: |
DeleteDeploymentResourcePoolRequest
Request message for DeleteDeploymentResourcePool method.
Fields | |
---|---|
name |
Required. The name of the DeploymentResourcePool to delete. Format: |
DeleteEndpointRequest
Request message for EndpointService.DeleteEndpoint
.
Fields | |
---|---|
name |
Required. The name of the Endpoint resource to be deleted. Format: |
DeleteEntityTypeRequest
Request message for FeaturestoreService.DeleteEntityType
.
Fields | |
---|---|
name |
Required. The name of the EntityType to be deleted. Format: |
force |
If set to true, any Features for this EntityType will also be deleted. (Otherwise, the request will only work if the EntityType has no Features.) |
DeleteExecutionRequest
Request message for MetadataService.DeleteExecution
.
Fields | |
---|---|
name |
Required. The resource name of the Execution to delete. Format: |
etag |
Optional. The etag of the Execution to delete. If this is provided, it must match the server's etag. Otherwise, the request will fail with a FAILED_PRECONDITION. |
DeleteFeatureGroupRequest
Request message for FeatureRegistryService.DeleteFeatureGroup
.
Fields | |
---|---|
name |
Required. The name of the FeatureGroup to be deleted. Format: |
force |
If set to true, any Features under this FeatureGroup will also be deleted. (Otherwise, the request will only work if the FeatureGroup has no Features.) |
DeleteFeatureOnlineStoreRequest
Request message for FeatureOnlineStoreAdminService.DeleteFeatureOnlineStore
.
Fields | |
---|---|
name |
Required. The name of the FeatureOnlineStore to be deleted. Format: |
force |
If set to true, any FeatureViews and Features for this FeatureOnlineStore will also be deleted. (Otherwise, the request will only work if the FeatureOnlineStore has no FeatureViews.) |
DeleteFeatureRequest
Request message for FeaturestoreService.DeleteFeature
. Request message for FeatureRegistryService.DeleteFeature
.
Fields | |
---|---|
name |
Required. The name of the Features to be deleted. Format: |
DeleteFeatureValuesOperationMetadata
Details of operations that delete Feature values.
Fields | |
---|---|
generic_ |
Operation metadata for Featurestore delete Features values. |
DeleteFeatureValuesRequest
Request message for FeaturestoreService.DeleteFeatureValues
.
Fields | |
---|---|
entity_ |
Required. The resource name of the EntityType grouping the Features for which values are being deleted from. Format: |
Union field DeleteOption . Defines options to select feature values to be deleted. DeleteOption can be only one of the following: |
|
select_ |
Select feature values to be deleted by specifying entities. |
select_ |
Select feature values to be deleted by specifying time range and features. |
SelectEntity
Message to select entity. If an entity id is selected, all the feature values corresponding to the entity id will be deleted, including the entityId.
Fields | |
---|---|
entity_ |
Required. Selectors choosing feature values of which entity id to be deleted from the EntityType. |
SelectTimeRangeAndFeature
Message to select time range and feature. Values of the selected feature generated within an inclusive time range will be deleted. Using this option permanently deletes the feature values from the specified feature IDs within the specified time range. This might include data from the online storage. If you want to retain any deleted historical data in the online storage, you must re-ingest it.
Fields | |
---|---|
time_ |
Required. Select feature generated within a half-inclusive time range. The time range is lower inclusive and upper exclusive. |
feature_ |
Required. Selectors choosing which feature values to be deleted from the EntityType. |
skip_ |
If set, data will not be deleted from online storage. When time range is older than the data in online storage, setting this to be true will make the deletion have no impact on online serving. |
DeleteFeatureValuesResponse
Response message for FeaturestoreService.DeleteFeatureValues
.
Fields | |
---|---|
Union field response . Response based on which delete option is specified in the request response can be only one of the following: |
|
select_ |
Response for request specifying the entities to delete |
select_ |
Response for request specifying time range and feature |
SelectEntity
Response message if the request uses the SelectEntity option.
Fields | |
---|---|
offline_ |
The count of deleted entity rows in the offline storage. Each row corresponds to the combination of an entity ID and a timestamp. One entity ID can have multiple rows in the offline storage. |
online_ |
The count of deleted entities in the online storage. Each entity ID corresponds to one entity. |
SelectTimeRangeAndFeature
Response message if the request uses the SelectTimeRangeAndFeature option.
Fields | |
---|---|
impacted_ |
The count of the features or columns impacted. This is the same as the feature count in the request. |
offline_ |
The count of modified entity rows in the offline storage. Each row corresponds to the combination of an entity ID and a timestamp. One entity ID can have multiple rows in the offline storage. Within each row, only the features specified in the request are deleted. |
online_ |
The count of modified entities in the online storage. Each entity ID corresponds to one entity. Within each entity, only the features specified in the request are deleted. |
DeleteFeatureViewRequest
Request message for FeatureOnlineStoreAdminService.DeleteFeatureView
.
Fields | |
---|---|
name |
Required. The name of the FeatureView to be deleted. Format: |
DeleteFeaturestoreRequest
Request message for FeaturestoreService.DeleteFeaturestore
.
Fields | |
---|---|
name |
Required. The name of the Featurestore to be deleted. Format: |
force |
If set to true, any EntityTypes and Features for this Featurestore will also be deleted. (Otherwise, the request will only work if the Featurestore has no EntityTypes.) |
DeleteHyperparameterTuningJobRequest
Request message for JobService.DeleteHyperparameterTuningJob
.
Fields | |
---|---|
name |
Required. The name of the HyperparameterTuningJob resource to be deleted. Format: |
DeleteIndexEndpointRequest
Request message for IndexEndpointService.DeleteIndexEndpoint
.
Fields | |
---|---|
name |
Required. The name of the IndexEndpoint resource to be deleted. Format: |
DeleteIndexRequest
Request message for IndexService.DeleteIndex
.
Fields | |
---|---|
name |
Required. The name of the Index resource to be deleted. Format: |
DeleteMetadataStoreOperationMetadata
Details of operations that perform MetadataService.DeleteMetadataStore
.
Fields | |
---|---|
generic_ |
Operation metadata for deleting a MetadataStore. |
DeleteMetadataStoreRequest
Request message for MetadataService.DeleteMetadataStore
.
Fields | |
---|---|
name |
Required. The resource name of the MetadataStore to delete. Format: |
force |
Deprecated: Field is no longer supported. |
DeleteModelDeploymentMonitoringJobRequest
Request message for JobService.DeleteModelDeploymentMonitoringJob
.
Fields | |
---|---|
name |
Required. The resource name of the model monitoring job to delete. Format: |
DeleteModelRequest
Request message for ModelService.DeleteModel
.
Fields | |
---|---|
name |
Required. The name of the Model resource to be deleted. Format: |
DeleteModelVersionRequest
Request message for ModelService.DeleteModelVersion
.
Fields | |
---|---|
name |
Required. The name of the model version to be deleted, with a version ID explicitly included. Example: |
DeleteNasJobRequest
Request message for JobService.DeleteNasJob
.
Fields | |
---|---|
name |
Required. The name of the NasJob resource to be deleted. Format: |
DeleteNotebookExecutionJobRequest
Request message for [NotebookService.DeleteNotebookExecutionJob]
Fields | |
---|---|
name |
Required. The name of the NotebookExecutionJob resource to be deleted. |
DeleteNotebookRuntimeRequest
Request message for NotebookService.DeleteNotebookRuntime
.
Fields | |
---|---|
name |
Required. The name of the NotebookRuntime resource to be deleted. Instead of checking whether the name is in valid NotebookRuntime resource name format, directly throw NotFound exception if there is no such NotebookRuntime in spanner. |
DeleteNotebookRuntimeTemplateRequest
Request message for NotebookService.DeleteNotebookRuntimeTemplate
.
Fields | |
---|---|
name |
Required. The name of the NotebookRuntimeTemplate resource to be deleted. Format: |
DeleteOperationMetadata
Details of operations that perform deletes of any entities.
Fields | |
---|---|
generic_ |
The common part of the operation metadata. |
DeletePersistentResourceRequest
Request message for PersistentResourceService.DeletePersistentResource
.
Fields | |
---|---|
name |
Required. The name of the PersistentResource to be deleted. Format: |
DeletePipelineJobRequest
Request message for PipelineService.DeletePipelineJob
.
Fields | |
---|---|
name |
Required. The name of the PipelineJob resource to be deleted. Format: |
DeleteRagCorpusRequest
Request message for VertexRagDataService.DeleteRagCorpus
.
Fields | |
---|---|
name |
Required. The name of the RagCorpus resource to be deleted. Format: |
force |
Optional. If set to true, any RagFiles in this RagCorpus will also be deleted. Otherwise, the request will only work if the RagCorpus has no RagFiles. |
DeleteRagFileRequest
Request message for VertexRagDataService.DeleteRagFile
.
Fields | |
---|---|
name |
Required. The name of the RagFile resource to be deleted. Format: |
DeleteSavedQueryRequest
Request message for DatasetService.DeleteSavedQuery
.
Fields | |
---|---|
name |
Required. The resource name of the SavedQuery to delete. Format: |
DeleteScheduleRequest
Request message for ScheduleService.DeleteSchedule
.
Fields | |
---|---|
name |
Required. The name of the Schedule resource to be deleted. Format: |
DeleteSpecialistPoolRequest
Request message for SpecialistPoolService.DeleteSpecialistPool
.
Fields | |
---|---|
name |
Required. The resource name of the SpecialistPool to delete. Format: |
force |
If set to true, any specialist managers in this SpecialistPool will also be deleted. (Otherwise, the request will only work if the SpecialistPool has no specialist managers.) |
DeleteStudyRequest
Request message for VizierService.DeleteStudy
.
Fields | |
---|---|
name |
Required. The name of the Study resource to be deleted. Format: |
DeleteTensorboardExperimentRequest
Request message for TensorboardService.DeleteTensorboardExperiment
.
Fields | |
---|---|
name |
Required. The name of the TensorboardExperiment to be deleted. Format: |
DeleteTensorboardRequest
Request message for TensorboardService.DeleteTensorboard
.
Fields | |
---|---|
name |
Required. The name of the Tensorboard to be deleted. Format: |
DeleteTensorboardRunRequest
Request message for TensorboardService.DeleteTensorboardRun
.
Fields | |
---|---|
name |
Required. The name of the TensorboardRun to be deleted. Format: |
DeleteTensorboardTimeSeriesRequest
Request message for TensorboardService.DeleteTensorboardTimeSeries
.
Fields | |
---|---|
name |
Required. The name of the TensorboardTimeSeries to be deleted. Format: |
DeleteTrainingPipelineRequest
Request message for PipelineService.DeleteTrainingPipeline
.
Fields | |
---|---|
name |
Required. The name of the TrainingPipeline resource to be deleted. Format: |
DeleteTrialRequest
Request message for VizierService.DeleteTrial
.
Fields | |
---|---|
name |
Required. The Trial's name. Format: |
DeployIndexOperationMetadata
Runtime operation information for IndexEndpointService.DeployIndex
.
Fields | |
---|---|
generic_ |
The operation generic information. |
deployed_ |
The unique index id specified by user |
DeployIndexRequest
Request message for IndexEndpointService.DeployIndex
.
Fields | |
---|---|
index_ |
Required. The name of the IndexEndpoint resource into which to deploy an Index. Format: |
deployed_ |
Required. The DeployedIndex to be created within the IndexEndpoint. |
DeployIndexResponse
Response message for IndexEndpointService.DeployIndex
.
Fields | |
---|---|
deployed_ |
The DeployedIndex that had been deployed in the IndexEndpoint. |
DeployModelOperationMetadata
Runtime operation information for EndpointService.DeployModel
.
Fields | |
---|---|
generic_ |
The operation generic information. |
DeployModelRequest
Request message for EndpointService.DeployModel
.
Fields | |
---|---|
endpoint |
Required. The name of the Endpoint resource into which to deploy a Model. Format: |
deployed_ |
Required. The DeployedModel to be created within the Endpoint. Note that |
traffic_ |
A map from a DeployedModel's ID to the percentage of this Endpoint's traffic that should be forwarded to that DeployedModel. If this field is non-empty, then the Endpoint's If this field is empty, then the Endpoint's |
DeployModelResponse
Response message for EndpointService.DeployModel
.
Fields | |
---|---|
deployed_ |
The DeployedModel that had been deployed in the Endpoint. |
DeployedIndex
A deployment of an Index. IndexEndpoints contain one or more DeployedIndexes.
Fields | |
---|---|
id |
Required. The user specified ID of the DeployedIndex. The ID can be up to 128 characters long and must start with a letter and only contain letters, numbers, and underscores. The ID must be unique within the project it is created in. |
index |
Required. The name of the Index this is the deployment of. We may refer to this Index as the DeployedIndex's "original" Index. |
display_ |
The display name of the DeployedIndex. If not provided upon creation, the Index's display_name is used. |
create_ |
Output only. Timestamp when the DeployedIndex was created. |
private_ |
Output only. Provides paths for users to send requests directly to the deployed index services running on Cloud via private services access. This field is populated if |
index_ |
Output only. The DeployedIndex may depend on various data on its original Index. Additionally when certain changes to the original Index are being done (e.g. when what the Index contains is being changed) the DeployedIndex may be asynchronously updated in the background to reflect these changes. If this timestamp's value is at least the |
automatic_ |
Optional. A description of resources that the DeployedIndex uses, which to large degree are decided by Vertex AI, and optionally allows only a modest additional configuration. If min_replica_count is not set, the default value is 2 (we don't provide SLA when min_replica_count=1). If max_replica_count is not set, the default value is min_replica_count. The max allowed replica count is 1000. |
dedicated_ |
Optional. A description of resources that are dedicated to the DeployedIndex, and that need a higher degree of manual configuration. The field min_replica_count must be set to a value strictly greater than 0, or else validation will fail. We don't provide SLA when min_replica_count=1. If max_replica_count is not set, the default value is min_replica_count. The max allowed replica count is 1000. Available machine types for SMALL shard: e2-standard-2 and all machine types available for MEDIUM and LARGE shard. Available machine types for MEDIUM shard: e2-standard-16 and all machine types available for LARGE shard. Available machine types for LARGE shard: e2-highmem-16, n2d-standard-32. n1-standard-16 and n1-standard-32 are still available, but we recommend e2-standard-16 and e2-highmem-16 for cost efficiency. |
enable_ |
Optional. If true, private endpoint's access logs are sent to Cloud Logging. These logs are like standard server access logs, containing information like timestamp and latency for each MatchRequest. Note that logs may incur a cost, especially if the deployed index receives a high queries per second rate (QPS). Estimate your costs before enabling this option. |
deployed_ |
Optional. If set, the authentication is enabled for the private endpoint. |
reserved_ |
Optional. A list of reserved ip ranges under the VPC network that can be used for this DeployedIndex. If set, we will deploy the index within the provided ip ranges. Otherwise, the index might be deployed to any ip ranges under the provided VPC network. The value should be the name of the address (https://cloud.google.com/compute/docs/reference/rest/v1/addresses) Example: ['vertex-ai-ip-range']. For more information about subnets and network IP ranges, please see https://cloud.google.com/vpc/docs/subnets#manually_created_subnet_ip_ranges. |
deployment_ |
Optional. The deployment group can be no longer than 64 characters (eg: 'test', 'prod'). If not set, we will use the 'default' deployment group. Creating Note: we only support up to 5 deployment groups(not including 'default'). |
psc_ |
Optional. If set for PSC deployed index, PSC connection will be automatically created after deployment is done and the endpoint information is populated in private_endpoints.psc_automated_endpoints. |
DeployedIndexAuthConfig
Used to set up the auth on the DeployedIndex's private endpoint.
Fields | |
---|---|
auth_ |
Defines the authentication provider that the DeployedIndex uses. |
AuthProvider
Configuration for an authentication provider, including support for JSON Web Token (JWT).
Fields | |
---|---|
audiences[] |
The list of JWT audiences. that are allowed to access. A JWT containing any of these audiences will be accepted. |
allowed_ |
A list of allowed JWT issuers. Each entry must be a valid Google service account, in the following format:
|
DeployedIndexRef
Points to a DeployedIndex.
Fields | |
---|---|
index_ |
Immutable. A resource name of the IndexEndpoint. |
deployed_ |
Immutable. The ID of the DeployedIndex in the above IndexEndpoint. |
display_ |
Output only. The display name of the DeployedIndex. |
DeployedModel
A deployment of a Model. Endpoints contain one or more DeployedModels.
Fields | |
---|---|
id |
Immutable. The ID of the DeployedModel. If not provided upon deployment, Vertex AI will generate a value for this ID. This value should be 1-10 characters, and valid characters are |
model |
Required. The resource name of the Model that this is the deployment of. Note that the Model may be in a different location than the DeployedModel's Endpoint. The resource name may contain version id or version alias to specify the version. Example: |
model_ |
Output only. The version ID of the model that is deployed. |
display_ |
The display name of the DeployedModel. If not provided upon creation, the Model's display_name is used. |
create_ |
Output only. Timestamp when the DeployedModel was created. |
explanation_ |
Explanation configuration for this DeployedModel. When deploying a Model using |
disable_ |
If true, deploy the model without explainable feature, regardless the existence of |
service_ |
The service account that the DeployedModel's container runs as. Specify the email address of the service account. If this service account is not specified, the container runs as a service account that doesn't have access to the resource project. Users deploying the Model must have the |
disable_ |
For custom-trained Models and AutoML Tabular Models, the container of the DeployedModel instances will send User can disable container logging by setting this flag to true. |
enable_ |
If true, online prediction access logs are sent to Cloud Logging. These logs are like standard server access logs, containing information like timestamp and latency for each prediction request. Note that logs may incur a cost, especially if your project receives prediction requests at a high queries per second rate (QPS). Estimate your costs before enabling this option. |
private_ |
Output only. Provide paths for users to send predict/explain/health requests directly to the deployed model services running on Cloud via private services access. This field is populated if |
faster_ |
Configuration for faster model deployment. |
status |
Output only. Runtime status of the deployed model. |
system_ |
System labels to apply to Model Garden deployments. System labels are managed by Google for internal use only. |
Union field prediction_resources . The prediction (for example, the machine) resources that the DeployedModel uses. The user is billed for the resources (at least their minimal amount) even if the DeployedModel receives no traffic. Not all Models support all resources types. See Model.supported_deployment_resources_types . Required except for Large Model Deploy use cases. prediction_resources can be only one of the following: |
|
dedicated_ |
A description of resources that are dedicated to the DeployedModel, and that need a higher degree of manual configuration. |
automatic_ |
A description of resources that to large degree are decided by Vertex AI, and require only a modest additional configuration. |
shared_ |
The resource name of the shared DeploymentResourcePool to deploy on. Format: |
Status
Runtime status of the deployed model.
Fields | |
---|---|
message |
Output only. The latest deployed model's status message (if any). |
last_ |
Output only. The time at which the status was last updated. |
available_ |
Output only. The number of available replicas of the deployed model. |
DeployedModelRef
Points to a DeployedModel.
Fields | |
---|---|
endpoint |
Immutable. A resource name of an Endpoint. |
deployed_ |
Immutable. An ID of a DeployedModel in the above Endpoint. |
DeploymentResourcePool
A description of resources that can be shared by multiple DeployedModels, whose underlying specification consists of a DedicatedResources.
Fields | |
---|---|
name |
Immutable. The resource name of the DeploymentResourcePool. Format: |
dedicated_ |
Required. The underlying DedicatedResources that the DeploymentResourcePool uses. |
encryption_ |
Customer-managed encryption key spec for a DeploymentResourcePool. If set, this DeploymentResourcePool will be secured by this key. Endpoints and the DeploymentResourcePool they deploy in need to have the same EncryptionSpec. |
service_ |
The service account that the DeploymentResourcePool's container(s) run as. Specify the email address of the service account. If this service account is not specified, the container(s) run as a service account that doesn't have access to the resource project. Users deploying the Models to this DeploymentResourcePool must have the |
disable_ |
If the DeploymentResourcePool is deployed with custom-trained Models or AutoML Tabular Models, the container(s) of the DeploymentResourcePool will send User can disable container logging by setting this flag to true. |
create_ |
Output only. Timestamp when this DeploymentResourcePool was created. |
satisfies_ |
Output only. Reserved for future use. |
satisfies_ |
Output only. Reserved for future use. |
DestinationFeatureSetting
Fields | |
---|---|
feature_ |
Required. The ID of the Feature to apply the setting to. |
destination_ |
Specify the field name in the export destination. If not specified, Feature ID is used. |
DirectPredictRequest
Request message for PredictionService.DirectPredict
.
Fields | |
---|---|
endpoint |
Required. The name of the Endpoint requested to serve the prediction. Format: |
inputs[] |
The prediction input. |
parameters |
The parameters that govern the prediction. |
DirectPredictResponse
Response message for PredictionService.DirectPredict
.
Fields | |
---|---|
outputs[] |
The prediction output. |
parameters |
The parameters that govern the prediction. |
DirectRawPredictRequest
Request message for PredictionService.DirectRawPredict
.
Fields | |
---|---|
endpoint |
Required. The name of the Endpoint requested to serve the prediction. Format: |
method_ |
Fully qualified name of the API method being invoked to perform predictions. Format: |
input |
The prediction input. |
DirectRawPredictResponse
Response message for PredictionService.DirectRawPredict
.
Fields | |
---|---|
output |
The prediction output. |
DirectUploadSource
This type has no fields.
The input content is encapsulated and uploaded in the request.
DiskSpec
Represents the spec of disk options.
Fields | |
---|---|
boot_ |
Type of the boot disk (default is "pd-ssd"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) or "pd-standard" (Persistent Disk Hard Disk Drive). |
boot_ |
Size in GB of the boot disk (default is 100GB). |
DoubleArray
A list of double values.
Fields | |
---|---|
values[] |
A list of double values. |
DynamicRetrievalConfig
Describes the options to customize dynamic retrieval.
Fields | |
---|---|
mode |
The mode of the predictor to be used in dynamic retrieval. |
dynamic_ |
Optional. The threshold to be used in dynamic retrieval. If not set, a system default value is used. |
Mode
The mode of the predictor to be used in dynamic retrieval.
Enums | |
---|---|
MODE_UNSPECIFIED |
Always trigger retrieval. |
MODE_DYNAMIC |
Run retrieval only when system decides it is necessary. |
EncryptionSpec
Represents a customer-managed encryption key spec that can be applied to a top-level resource.
Fields | |
---|---|
kms_ |
Required. The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form: |
Endpoint
Models are deployed into it, and afterwards Endpoint is called to obtain predictions and explanations.
Fields | |
---|---|
name |
Output only. The resource name of the Endpoint. |
display_ |
Required. The display name of the Endpoint. The name can be up to 128 characters long and can consist of any UTF-8 characters. |
description |
The description of the Endpoint. |
deployed_ |
Output only. The models deployed in this Endpoint. To add or remove DeployedModels use |
traffic_ |
A map from a DeployedModel's ID to the percentage of this Endpoint's traffic that should be forwarded to that DeployedModel. If a DeployedModel's ID is not listed in this map, then it receives no traffic. The traffic percentage values must add up to 100, or map must be empty if the Endpoint is to not accept any traffic at a moment. |
etag |
Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
labels |
The labels with user-defined metadata to organize your Endpoints. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. |
create_ |
Output only. Timestamp when this Endpoint was created. |
update_ |
Output only. Timestamp when this Endpoint was last updated. |
encryption_ |
Customer-managed encryption key spec for an Endpoint. If set, this Endpoint and all sub-resources of this Endpoint will be secured by this key. |
network |
Optional. The full name of the Google Compute Engine network to which the Endpoint should be peered. Private services access must already be configured for the network. If left unspecified, the Endpoint is not peered with any network. Only one of the fields, Format: |
enable_private_service_connect |
Deprecated: If true, expose the Endpoint via private service connect. Only one of the fields, |
private_ |
Optional. Configuration for private service connect.
|
model_ |
Output only. Resource name of the Model Monitoring job associated with this Endpoint if monitoring is enabled by |
predict_ |
Configures the request-response logging for online prediction. |
dedicated_ |
If true, the endpoint will be exposed through a dedicated DNS [Endpoint.dedicated_endpoint_dns]. Your request to the dedicated DNS will be isolated from other users' traffic and will have better performance and reliability. Note: Once you enabled dedicated endpoint, you won't be able to send request to the shared DNS {region}-aiplatform.googleapis.com. The limitation will be removed soon. |
dedicated_ |
Output only. DNS of the dedicated endpoint. Will only be populated if dedicated_endpoint_enabled is true. Format: |
client_ |
Configurations that are applied to the endpoint for online prediction. |
satisfies_ |
Output only. Reserved for future use. |
satisfies_ |
Output only. Reserved for future use. |
EntityIdSelector
Selector for entityId. Getting ids from the given source.
Fields | |
---|---|
entity_ |
Source column that holds entity IDs. If not provided, entity IDs are extracted from the column named entity_id. |
Union field EntityIdsSource . Details about the source data, including the location of the storage and the format. EntityIdsSource can be only one of the following: |
|
csv_ |
Source of Csv |
EntityType
An entity type is a type of object in a system that needs to be modeled and have stored information about. For example, driver is an entity type, and driver0 is an instance of an entity type driver.
Fields | |
---|---|
name |
Immutable. Name of the EntityType. Format: The last part entity_type is assigned by the client. The entity_type can be up to 64 characters long and can consist only of ASCII Latin letters A-Z and a-z and underscore(_), and ASCII digits 0-9 starting with a letter. The value will be unique given a featurestore. |
description |
Optional. Description of the EntityType. |
create_ |
Output only. Timestamp when this EntityType was created. |
update_ |
Output only. Timestamp when this EntityType was most recently updated. |
labels |
Optional. The labels with user-defined metadata to organize your EntityTypes. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one EntityType (System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. |
etag |
Optional. Used to perform a consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
monitoring_ |
Optional. The default monitoring configuration for all Features with value type ( If this is populated with [FeaturestoreMonitoringConfig.monitoring_interval] specified, snapshot analysis monitoring is enabled. Otherwise, snapshot analysis monitoring is disabled. |
offline_ |
Optional. Config for data retention policy in offline storage. TTL in days for feature values that will be stored in offline storage. The Feature Store offline storage periodically removes obsolete feature values older than |
satisfies_ |
Output only. Reserved for future use. |
satisfies_ |
Output only. Reserved for future use. |
EnvVar
Represents an environment variable present in a Container or Python Module.
Fields | |
---|---|
name |
Required. Name of the environment variable. Must be a valid C identifier. |
value |
Required. Variables that reference a $(VAR_NAME) are expanded using the previous defined environment variables in the container and any service environment variables. If a variable cannot be resolved, the reference in the input string will be unchanged. The $(VAR_NAME) syntax can be escaped with a double $$, ie: $$(VAR_NAME). Escaped references will never be expanded, regardless of whether the variable exists or not. |
ErrorAnalysisAnnotation
Model error analysis for each annotation.
Fields | |
---|---|
attributed_ |
Attributed items for a given annotation, typically representing neighbors from the training sets constrained by the query type. |
query_ |
The query type used for finding the attributed items. |
outlier_ |
The outlier score of this annotated item. Usually defined as the min of all distances from attributed items. |
outlier_ |
The threshold used to determine if this annotation is an outlier or not. |
AttributedItem
Attributed items for a given annotation, typically representing neighbors from the training sets constrained by the query type.
Fields | |
---|---|
annotation_ |
The unique ID for each annotation. Used by FE to allocate the annotation in DB. |
distance |
The distance of this item to the annotation. |
QueryType
The query type used for finding the attributed items.
Enums | |
---|---|
QUERY_TYPE_UNSPECIFIED |
Unspecified query type for model error analysis. |
ALL_SIMILAR |
Query similar samples across all classes in the dataset. |
SAME_CLASS_SIMILAR |
Query similar samples from the same class of the input sample. |
SAME_CLASS_DISSIMILAR |
Query dissimilar samples from the same class of the input sample. |
EvaluateInstancesRequest
Request message for EvaluationService.EvaluateInstances.
Fields | |
---|---|
location |
Required. The resource name of the Location to evaluate the instances. Format: |
Union field metric_inputs . Instances and specs for evaluation metric_inputs can be only one of the following: |
|
exact_ |
Auto metric instances. Instances and metric spec for exact match metric. |
bleu_ |
Instances and metric spec for bleu metric. |
rouge_ |
Instances and metric spec for rouge metric. |
fluency_ |
LLM-based metric instance. General text generation metrics, applicable to other categories. Input for fluency metric. |
coherence_ |
Input for coherence metric. |
safety_ |
Input for safety metric. |
groundedness_ |
Input for groundedness metric. |
fulfillment_ |
Input for fulfillment metric. |
summarization_ |
Input for summarization quality metric. |
pairwise_ |
Input for pairwise summarization quality metric. |
summarization_ |
Input for summarization helpfulness metric. |
summarization_ |
Input for summarization verbosity metric. |
question_ |
Input for question answering quality metric. |
pairwise_ |
Input for pairwise question answering quality metric. |
question_ |
Input for question answering relevance metric. |
question_ |
Input for question answering helpfulness metric. |
question_ |
Input for question answering correctness metric. |
pointwise_ |
Input for pointwise metric. |
pairwise_ |
Input for pairwise metric. |
tool_ |
Tool call metric instances. Input for tool call valid metric. |
tool_ |
Input for tool name match metric. |
tool_ |
Input for tool parameter key match metric. |
tool_ |
Input for tool parameter key value match metric. |
comet_ |
Translation metrics. Input for Comet metric. |
trajectory_ |
Input for trajectory exact match metric. |
trajectory_ |
Input for trajectory in order match metric. |
trajectory_ |
Input for trajectory match any order metric. |
trajectory_ |
Input for trajectory precision metric. |
trajectory_ |
Input for trajectory recall metric. |
trajectory_ |
Input for trajectory single tool use metric. |
EvaluateInstancesResponse
Response message for EvaluationService.EvaluateInstances.
Fields | |
---|---|
Union field evaluation_results . Evaluation results will be served in the same order as presented in EvaluationRequest.instances. evaluation_results can be only one of the following: |
|
exact_ |
Auto metric evaluation results. Results for exact match metric. |
bleu_ |
Results for bleu metric. |
rouge_ |
Results for rouge metric. |
fluency_ |
LLM-based metric evaluation result. General text generation metrics, applicable to other categories. Result for fluency metric. |
coherence_ |
Result for coherence metric. |
safety_ |
Result for safety metric. |
groundedness_ |
Result for groundedness metric. |
fulfillment_ |
Result for fulfillment metric. |
summarization_ |
Summarization only metrics. Result for summarization quality metric. |
pairwise_ |
Result for pairwise summarization quality metric. |
summarization_ |
Result for summarization helpfulness metric. |
summarization_ |
Result for summarization verbosity metric. |
question_ |
Question answering only metrics. Result for question answering quality metric. |
pairwise_ |
Result for pairwise question answering quality metric. |
question_ |
Result for question answering relevance metric. |
question_ |
Result for question answering helpfulness metric. |
question_ |
Result for question answering correctness metric. |
pointwise_ |
Generic metrics. Result for pointwise metric. |
pairwise_ |
Result for pairwise metric. |
tool_ |
Tool call metrics. Results for tool call valid metric. |
tool_ |
Results for tool name match metric. |
tool_ |
Results for tool parameter key match metric. |
tool_ |
Results for tool parameter key value match metric. |
comet_ |
Translation metrics. Result for Comet metric. |
trajectory_ |
Result for trajectory exact match metric. |
trajectory_ |
Result for trajectory in order match metric. |
trajectory_ |
Result for trajectory any order match metric. |
trajectory_ |
Result for trajectory precision metric. |
trajectory_ |
Results for trajectory recall metric. |
trajectory_ |
Results for trajectory single tool use metric. |
EvaluatedAnnotation
True positive, false positive, or false negative.
EvaluatedAnnotation is only available under ModelEvaluationSlice with slice of annotationSpec
dimension.
Fields | |
---|---|
type |
Output only. Type of the EvaluatedAnnotation. |
predictions[] |
Output only. The model predicted annotations. For true positive, there is one and only one prediction, which matches the only one ground truth annotation in For false positive, there is one and only one prediction, which doesn't match any ground truth annotation of the corresponding For false negative, there are zero or more predictions which are similar to the only ground truth annotation in The schema of the prediction is stored in |
ground_ |
Output only. The ground truth Annotations, i.e. the Annotations that exist in the test data the Model is evaluated on. For true positive, there is one and only one ground truth annotation, which matches the only prediction in For false positive, there are zero or more ground truth annotations that are similar to the only prediction in For false negative, there is one and only one ground truth annotation, which doesn't match any predictions created by the model. The schema of the ground truth is stored in |
data_ |
Output only. The data item payload that the Model predicted this EvaluatedAnnotation on. |
evaluated_ |
Output only. ID of the EvaluatedDataItemView under the same ancestor ModelEvaluation. The EvaluatedDataItemView consists of all ground truths and predictions on |
explanations[] |
Explanations of The attributions list in the |
error_ |
Annotations of model error analysis results. |
EvaluatedAnnotationType
Describes the type of the EvaluatedAnnotation. The type is determined
Enums | |
---|---|
EVALUATED_ANNOTATION_TYPE_UNSPECIFIED |
Invalid value. |
TRUE_POSITIVE |
The EvaluatedAnnotation is a true positive. It has a prediction created by the Model and a ground truth Annotation which the prediction matches. |
FALSE_POSITIVE |
The EvaluatedAnnotation is false positive. It has a prediction created by the Model which does not match any ground truth annotation. |
FALSE_NEGATIVE |
The EvaluatedAnnotation is false negative. It has a ground truth annotation which is not matched by any of the model created predictions. |
EvaluatedAnnotationExplanation
Explanation result of the prediction produced by the Model.
Fields | |
---|---|
explanation_ |
Explanation type. For AutoML Image Classification models, possible values are:
|
explanation |
Explanation attribution response details. |
Event
An edge describing the relationship between an Artifact and an Execution in a lineage graph.
Fields | |
---|---|
artifact |
Required. The relative resource name of the Artifact in the Event. |
execution |
Output only. The relative resource name of the Execution in the Event. |
event_ |
Output only. Time the Event occurred. |
type |
Required. The type of the Event. |
labels |
The labels with user-defined metadata to annotate Events. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Event (System labels are excluded). See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. |
Type
Describes whether an Event's Artifact is the Execution's input or output.
Enums | |
---|---|
TYPE_UNSPECIFIED |
Unspecified whether input or output of the Execution. |
INPUT |
An input of the Execution. |
OUTPUT |
An output of the Execution. |
ExactMatchInput
Input for exact match metric.
Fields | |
---|---|
metric_ |
Required. Spec for exact match metric. |
instances[] |
Required. Repeated exact match instances. |
ExactMatchInstance
Spec for exact match instance.
Fields | |
---|---|
prediction |
Required. Output of the evaluated model. |
reference |
Required. Ground truth used to compare against the prediction. |
ExactMatchMetricValue
Exact match metric value for an instance.
Fields | |
---|---|
score |
Output only. Exact match score. |
ExactMatchResults
Results for exact match metric.
Fields | |
---|---|
exact_ |
Output only. Exact match metric values. |
ExactMatchSpec
This type has no fields.
Spec for exact match metric - returns 1 if prediction and reference exactly matches, otherwise 0.
Examples
Example-based explainability that returns the nearest neighbors from the provided dataset.
Fields | |
---|---|
neighbor_ |
The number of neighbors to return when querying for examples. |
Union field
|
|
example_ |
The Cloud Storage input instances. |
Union field
|
|
nearest_ |
The full configuration for the generated index, the semantics are the same as |
presets |
Simplified preset configuration, which automatically sets configuration values based on the desired query speed-precision trade-off and modality. |
ExampleGcsSource
The Cloud Storage input instances.
Fields | |
---|---|
data_ |
The format in which instances are given, if not specified, assume it's JSONL format. Currently only JSONL format is supported. |
gcs_ |
The Cloud Storage location for the input instances. |
DataFormat
The format of the input example instances.
Enums | |
---|---|
DATA_FORMAT_UNSPECIFIED |
Format unspecified, used when unset. |
JSONL |
Examples are stored in JSONL files. |
ExamplesOverride
Overrides for example-based explanations.
Fields | |
---|---|
neighbor_ |
The number of neighbors to return. |
crowding_ |
The number of neighbors to return that have the same crowding tag. |
restrictions[] |
Restrict the resulting nearest neighbors to respect these constraints. |
return_ |
If true, return the embeddings instead of neighbors. |
data_ |
The format of the data being provided with each call. |
DataFormat
Data format enum.
Enums | |
---|---|
DATA_FORMAT_UNSPECIFIED |
Unspecified format. Must not be used. |
INSTANCES |
Provided data is a set of model inputs. |
EMBEDDINGS |
Provided data is a set of embeddings. |
ExamplesRestrictionsNamespace
Restrictions namespace for example-based explanations overrides.
Fields | |
---|---|
namespace_ |
The namespace name. |
allow[] |
The list of allowed tags. |
deny[] |
The list of deny tags. |
Execution
Instance of a general execution.
Fields | |
---|---|
name |
Output only. The resource name of the Execution. |
display_ |
User provided display name of the Execution. May be up to 128 Unicode characters. |
state |
The state of this Execution. This is a property of the Execution, and does not imply or capture any ongoing process. This property is managed by clients (such as Vertex AI Pipelines) and the system does not prescribe or check the validity of state transitions. |
etag |
An eTag used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
labels |
The labels with user-defined metadata to organize your Executions. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Execution (System labels are excluded). |
create_ |
Output only. Timestamp when this Execution was created. |
update_ |
Output only. Timestamp when this Execution was last updated. |
schema_ |
The title of the schema describing the metadata. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store. |
schema_ |
The version of the schema in Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store. |
metadata |
Properties of the Execution. Top level metadata keys' heading and trailing spaces will be trimmed. The size of this field should not exceed 200KB. |
description |
Description of the Execution |
State
Describes the state of the Execution.
Enums | |
---|---|
STATE_UNSPECIFIED |
Unspecified Execution state |
NEW |
The Execution is new |
RUNNING |
The Execution is running |
COMPLETE |
The Execution has finished running |
FAILED |
The Execution has failed |
CACHED |
The Execution completed through Cache hit. |
CANCELLED |
The Execution was cancelled. |
ExplainRequest
Request message for PredictionService.Explain
.
Fields | |
---|---|
endpoint |
Required. The name of the Endpoint requested to serve the explanation. Format: |
instances[] |
Required. The instances that are the input to the explanation call. A DeployedModel may have an upper limit on the number of instances it supports per request, and when it is exceeded the explanation call errors in case of AutoML Models, or, in case of customer created Models, the behaviour is as documented by that Model. The schema of any single instance may be specified via Endpoint's DeployedModels' |
parameters |
The parameters that govern the prediction. The schema of the parameters may be specified via Endpoint's DeployedModels' |
explanation_ |
If specified, overrides the |
deployed_ |
If specified, this ExplainRequest will be served by the chosen DeployedModel, overriding |
ExplainResponse
Response message for PredictionService.Explain
.
Fields | |
---|---|
explanations[] |
The explanations of the Model's It has the same number of elements as |
deployed_ |
ID of the Endpoint's DeployedModel that served this explanation. |
predictions[] |
The predictions that are the output of the predictions call. Same as |
Explanation
Explanation of a prediction (provided in PredictResponse.predictions
) produced by the Model on a given instance
.
Fields | |
---|---|
attributions[] |
Output only. Feature attributions grouped by predicted outputs. For Models that predict only one output, such as regression Models that predict only one score, there is only one attibution that explains the predicted output. For Models that predict multiple outputs, such as multiclass Models that predict multiple classes, each element explains one specific item. By default, we provide Shapley values for the predicted class. However, you can configure the explanation request to generate Shapley values for any other classes too. For example, if a model predicts a probability of If users set |
neighbors[] |
Output only. List of the nearest neighbors for example-based explanations. For models deployed with the examples explanations feature enabled, the attributions field is empty and instead the neighbors field is populated. |
ExplanationMetadata
Metadata describing the Model's input and output for explanation.
Fields | |
---|---|
inputs |
Required. Map from feature names to feature input metadata. Keys are the name of the features. Values are the specification of the feature. An empty InputMetadata is valid. It describes a text feature which has the name specified as the key in For Vertex AI-provided Tensorflow images, the key can be any friendly name of the feature. Once specified, For custom images, the key must match with the key in |
outputs |
Required. Map from output names to output metadata. For Vertex AI-provided Tensorflow images, keys can be any user defined string that consists of any UTF-8 characters. For custom images, keys are the name of the output field in the prediction to be explained. Currently only one key is allowed. |
feature_ |
Points to a YAML file stored on Google Cloud Storage describing the format of the |
latent_ |
Name of the source to generate embeddings for example based explanations. |
InputMetadata
Metadata of the input of a feature.
Fields other than InputMetadata.input_baselines
are applicable only for Models that are using Vertex AI-provided images for Tensorflow.
Fields | |
---|---|
input_ |
Baseline inputs for this feature. If no baseline is specified, Vertex AI chooses the baseline for this feature. If multiple baselines are specified, Vertex AI returns the average attributions across them in For Vertex AI-provided Tensorflow images (both 1.x and 2.x), the shape of each baseline must match the shape of the input tensor. If a scalar is provided, we broadcast to the same shape as the input tensor. For custom images, the element of the baselines must be in the same format as the feature's input in the |
input_ |
Name of the input tensor for this feature. Required and is only applicable to Vertex AI-provided images for Tensorflow. |
encoding |
Defines how the feature is encoded into the input tensor. Defaults to IDENTITY. |
modality |
Modality of the feature. Valid values are: numeric, image. Defaults to numeric. |
feature_ |
The domain details of the input feature value. Like min/max, original mean or standard deviation if normalized. |
indices_ |
Specifies the index of the values of the input tensor. Required when the input tensor is a sparse representation. Refer to Tensorflow documentation for more details: https://www.tensorflow.org/api_docs/python/tf/sparse/SparseTensor. |
dense_ |
Specifies the shape of the values of the input if the input is a sparse representation. Refer to Tensorflow documentation for more details: https://www.tensorflow.org/api_docs/python/tf/sparse/SparseTensor. |
index_ |
A list of feature names for each index in the input tensor. Required when the input |
encoded_ |
Encoded tensor is a transformation of the input tensor. Must be provided if choosing An encoded tensor is generated if the input tensor is encoded by a lookup table. |
encoded_ |
A list of baselines for the encoded tensor. The shape of each baseline should match the shape of the encoded tensor. If a scalar is provided, Vertex AI broadcasts to the same shape as the encoded tensor. |
visualization |
Visualization configurations for image explanation. |
group_ |
Name of the group that the input belongs to. Features with the same group name will be treated as one feature when computing attributions. Features grouped together can have different shapes in value. If provided, there will be one single attribution generated in |
Encoding
Defines how a feature is encoded. Defaults to IDENTITY.
Enums | |
---|---|
ENCODING_UNSPECIFIED |
Default value. This is the same as IDENTITY. |
IDENTITY |
The tensor represents one feature. |
BAG_OF_FEATURES |
The tensor represents a bag of features where each index maps to a feature.
|
BAG_OF_FEATURES_SPARSE |
The tensor represents a bag of features where each index maps to a feature. Zero values in the tensor indicates feature being non-existent.
|
INDICATOR |
The tensor is a list of binaries representing whether a feature exists or not (1 indicates existence).
|
COMBINED_EMBEDDING |
The tensor is encoded into a 1-dimensional array represented by an encoded tensor.
|
CONCAT_EMBEDDING |
Select this encoding when the input tensor is encoded into a 2-dimensional array represented by an encoded tensor.
|
FeatureValueDomain
Domain details of the input feature value. Provides numeric information about the feature, such as its range (min, max). If the feature has been pre-processed, for example with z-scoring, then it provides information about how to recover the original feature. For example, if the input feature is an image and it has been pre-processed to obtain 0-mean and stddev = 1 values, then original_mean, and original_stddev refer to the mean and stddev of the original feature (e.g. image tensor) from which input feature (with mean = 0 and stddev = 1) was obtained.
Fields | |
---|---|
min_ |
The minimum permissible value for this feature. |
max_ |
The maximum permissible value for this feature. |
original_ |
If this input feature has been normalized to a mean value of 0, the original_mean specifies the mean value of the domain prior to normalization. |
original_ |
If this input feature has been normalized to a standard deviation of 1.0, the original_stddev specifies the standard deviation of the domain prior to normalization. |
Visualization
Visualization configurations for image explanation.
Fields | |
---|---|
type |
Type of the image visualization. Only applicable to |
polarity |
Whether to only highlight pixels with positive contributions, negative or both. Defaults to POSITIVE. |
color_ |
The color scheme used for the highlighted areas. Defaults to PINK_GREEN for Defaults to VIRIDIS for |
clip_ |
Excludes attributions above the specified percentile from the highlighted areas. Using the clip_percent_upperbound and clip_percent_lowerbound together can be useful for filtering out noise and making it easier to see areas of strong attribution. Defaults to 99.9. |
clip_ |
Excludes attributions below the specified percentile, from the highlighted areas. Defaults to 62. |
overlay_ |
How the original image is displayed in the visualization. Adjusting the overlay can help increase visual clarity if the original image makes it difficult to view the visualization. Defaults to NONE. |
ColorMap
The color scheme used for highlighting areas.
Enums | |
---|---|
COLOR_MAP_UNSPECIFIED |
Should not be used. |
PINK_GREEN |
Positive: green. Negative: pink. |
VIRIDIS |
Viridis color map: A perceptually uniform color mapping which is easier to see by those with colorblindness and progresses from yellow to green to blue. Positive: yellow. Negative: blue. |
RED |
Positive: red. Negative: red. |
GREEN |
Positive: green. Negative: green. |
RED_GREEN |
Positive: green. Negative: red. |
PINK_WHITE_GREEN |
PiYG palette. |
OverlayType
How the original image is displayed in the visualization.
Enums | |
---|---|
OVERLAY_TYPE_UNSPECIFIED |
Default value. This is the same as NONE. |
NONE |
No overlay. |
ORIGINAL |
The attributions are shown on top of the original image. |
GRAYSCALE |
The attributions are shown on top of grayscaled version of the original image. |
MASK_BLACK |
The attributions are used as a mask to reveal predictive parts of the image and hide the un-predictive parts. |
Polarity
Whether to only highlight pixels with positive contributions, negative or both. Defaults to POSITIVE.
Enums | |
---|---|
POLARITY_UNSPECIFIED |
Default value. This is the same as POSITIVE. |
POSITIVE |
Highlights the pixels/outlines that were most influential to the model's prediction. |
NEGATIVE |
Setting polarity to negative highlights areas that does not lead to the models's current prediction. |
BOTH |
Shows both positive and negative attributions. |
Type
Type of the image visualization. Only applicable to Integrated Gradients attribution
.
Enums | |
---|---|
TYPE_UNSPECIFIED |
Should not be used. |
PIXELS |
Shows which pixel contributed to the image prediction. |
OUTLINES |
Shows which region contributed to the image prediction by outlining the region. |
OutputMetadata
Metadata of the prediction output to be explained.
Fields | |
---|---|
output_ |
Name of the output tensor. Required and is only applicable to Vertex AI provided images for Tensorflow. |
Union field If neither of the fields are specified, |
|
index_ |
Static mapping between the index and display name. Use this if the outputs are a deterministic n-dimensional array, e.g. a list of scores of all the classes in a pre-defined order for a multi-classification Model. It's not feasible if the outputs are non-deterministic, e.g. the Model produces top-k classes or sort the outputs by their values. The shape of the value must be an n-dimensional array of strings. The number of dimensions must match that of the outputs to be explained. The |
display_ |
Specify a field name in the prediction to look for the display name. Use this if the prediction contains the display names for the outputs. The display names in the prediction must have the same shape of the outputs, so that it can be located by |
ExplanationMetadataOverride
The ExplanationMetadata
entries that can be overridden at online explanation
time.
Fields | |
---|---|
inputs |
Required. Overrides the |
InputMetadataOverride
The input metadata
entries to be overridden.
Fields | |
---|---|
input_ |
Baseline inputs for this feature. This overrides the |
ExplanationParameters
Parameters to configure explaining for Model's predictions.
Fields | |
---|---|
top_ |
If populated, returns attributions for top K indices of outputs (defaults to 1). Only applies to Models that predicts more than one outputs (e,g, multi-class Models). When set to -1, returns explanations for all outputs. |
output_ |
If populated, only returns attributions that have If not populated, returns attributions for Only applicable to Models that predict multiple outputs (e,g, multi-class Models that predict multiple classes). |
Union field
|
|
sampled_ |
An attribution method that approximates Shapley values for features that contribute to the label being predicted. A sampling strategy is used to approximate the value rather than considering all subsets of features. Refer to this paper for model details: https://arxiv.org/abs/1306.4265. |
integrated_ |
An attribution method that computes Aumann-Shapley values taking advantage of the model's fully differentiable structure. Refer to this paper for more details: https://arxiv.org/abs/1703.01365 |
xrai_ |
An attribution method that redistributes Integrated Gradients attribution to segmented regions, taking advantage of the model's fully differentiable structure. Refer to this paper for more details: https://arxiv.org/abs/1906.02825 XRAI currently performs better on natural images, like a picture of a house or an animal. If the images are taken in artificial environments, like a lab or manufacturing line, or from diagnostic equipment, like x-rays or quality-control cameras, use Integrated Gradients instead. |
examples |
Example-based explanations that returns the nearest neighbors from the provided dataset. |
ExplanationSpec
Specification of Model explanation.
Fields | |
---|---|
parameters |
Required. Parameters that configure explaining of the Model's predictions. |
metadata |
Optional. Metadata describing the Model's input and output for explanation. |
ExplanationSpecOverride
The ExplanationSpec
entries that can be overridden at online explanation
time.
Fields | |
---|---|
parameters |
The parameters to be overridden. Note that the attribution method cannot be changed. If not specified, no parameter is overridden. |
metadata |
The metadata to be overridden. If not specified, no metadata is overridden. |
examples_ |
The example-based explanations parameter overrides. |
ExportDataConfig
Describes what part of the Dataset is to be exported, the destination of the export and how to export.
Fields | |
---|---|
annotations_ |
An expression for filtering what part of the Dataset is to be exported. Only Annotations that match this filter will be exported. The filter syntax is the same as in |
saved_ |
The ID of a SavedQuery (annotation set) under the Dataset specified by Only used for custom training data export use cases. Only applicable to Datasets that have SavedQueries. Only Annotations that are associated with this SavedQuery are used in respectively training. When used in conjunction with Only one of |
annotation_ |
The Cloud Storage URI that points to a YAML file describing the annotation schema. The schema is defined as an OpenAPI 3.0.2 Schema Object. The schema files that can be used here are found in gs://google-cloud-aiplatform/schema/dataset/annotation/, note that the chosen schema must be consistent with Only used for custom training data export use cases. Only applicable to Datasets that have DataItems and Annotations. Only Annotations that both match this schema and belong to DataItems not ignored by the split method are used in respectively training, validation or test role, depending on the role of the DataItem they are on. When used in conjunction with |
export_ |
Indicates the usage of the exported files. |
Union field destination . The destination of the output. destination can be only one of the following: |
|
gcs_ |
The Google Cloud Storage location where the output is to be written to. In the given directory a new directory will be created with name: |
Union field split . The instructions how the export data should be split between the training, validation and test sets. split can be only one of the following: |
|
fraction_ |
Split based on fractions defining the size of each set. |
filter_ |
Split based on the provided filters for each set. |
ExportUse
ExportUse indicates the usage of the exported files. It restricts file destination, format, annotations to be exported, whether to allow unannotated data to be exported and whether to clone files to temp Cloud Storage bucket.
Enums | |
---|---|
EXPORT_USE_UNSPECIFIED |
Regular user export. |
CUSTOM_CODE_TRAINING |
Export for custom code training. |
ExportDataOperationMetadata
Runtime operation information for DatasetService.ExportData
.
Fields | |
---|---|
generic_ |
The common part of the operation metadata. |
gcs_ |
A Google Cloud Storage directory which path ends with '/'. The exported data is stored in the directory. |
ExportDataRequest
Request message for DatasetService.ExportData
.
Fields | |
---|---|
name |
Required. The name of the Dataset resource. Format: |
export_ |
Required. The desired output location. |
ExportDataResponse
Response message for DatasetService.ExportData
.
Fields | |
---|---|
exported_ |
All of the files that are exported in this export operation. For custom code training export, only three (training, validation and test) Cloud Storage paths in wildcard format are populated (for example, gs://.../training-*). |
data_ |
Only present for custom code training export use case. Records data stats, i.e., train/validation/test item/annotation counts calculated during the export operation. |
ExportFeatureValuesOperationMetadata
Details of operations that exports Features values.
Fields | |
---|---|
generic_ |
Operation metadata for Featurestore export Feature values. |
ExportFeatureValuesRequest
Request message for FeaturestoreService.ExportFeatureValues
.
Fields | |
---|---|
entity_ |
Required. The resource name of the EntityType from which to export Feature values. Format: |
destination |
Required. Specifies destination location and format. |
feature_ |
Required. Selects Features to export values of. |
settings[] |
Per-Feature export settings. |
Union field mode . Required. The mode in which Feature values are exported. mode can be only one of the following: |
|
snapshot_ |
Exports the latest Feature values of all entities of the EntityType within a time range. |
full_ |
Exports all historical values of all entities of the EntityType within a time range |
FullExport
Describes exporting all historical Feature values of all entities of the EntityType between [start_time, end_time].
Fields | |
---|---|
start_ |
Excludes Feature values with feature generation timestamp before this timestamp. If not set, retrieve oldest values kept in Feature Store. Timestamp, if present, must not have higher than millisecond precision. |
end_ |
Exports Feature values as of this timestamp. If not set, retrieve values as of now. Timestamp, if present, must not have higher than millisecond precision. |
SnapshotExport
Describes exporting the latest Feature values of all entities of the EntityType between [start_time, snapshot_time].
Fields | |
---|---|
snapshot_ |
Exports Feature values as of this timestamp. If not set, retrieve values as of now. Timestamp, if present, must not have higher than millisecond precision. |
start_ |
Excludes Feature values with feature generation timestamp before this timestamp. If not set, retrieve oldest values kept in Feature Store. Timestamp, if present, must not have higher than millisecond precision. |
ExportFeatureValuesResponse
This type has no fields.
Response message for FeaturestoreService.ExportFeatureValues
.
ExportFilterSplit
Assigns input data to training, validation, and test sets based on the given filters, data pieces not matched by any filter are ignored. Currently only supported for Datasets containing DataItems. If any of the filters in this message are to match nothing, then they can be set as '-' (the minus sign).
Supported only for unstructured Datasets.
Fields | |
---|---|
training_ |
Required. A filter on DataItems of the Dataset. DataItems that match this filter are used to train the Model. A filter with same syntax as the one used in |
validation_ |
Required. A filter on DataItems of the Dataset. DataItems that match this filter are used to validate the Model. A filter with same syntax as the one used in |
test_ |
Required. A filter on DataItems of the Dataset. DataItems that match this filter are used to test the Model. A filter with same syntax as the one used in |
ExportFractionSplit
Assigns the input data to training, validation, and test sets as per the given fractions. Any of training_fraction
, validation_fraction
and test_fraction
may optionally be provided, they must sum to up to 1. If the provided ones sum to less than 1, the remainder is assigned to sets as decided by Vertex AI. If none of the fractions are set, by default roughly 80% of data is used for training, 10% for validation, and 10% for test.
Fields | |
---|---|
training_ |
The fraction of the input data that is to be used to train the Model. |
validation_ |
The fraction of the input data that is to be used to validate the Model. |
test_ |
The fraction of the input data that is to be used to evaluate the Model. |
ExportModelOperationMetadata
Details of ModelService.ExportModel
operation.
Fields | |
---|---|
generic_ |
The common part of the operation metadata. |
output_ |
Output only. Information further describing the output of this Model export. |
OutputInfo
Further describes the output of the ExportModel. Supplements ExportModelRequest.OutputConfig
.
Fields | |
---|---|
artifact_ |
Output only. If the Model artifact is being exported to Google Cloud Storage this is the full path of the directory created, into which the Model files are being written to. |
image_ |
Output only. If the Model image is being exported to Google Container Registry or Artifact Registry this is the full path of the image created. |
ExportModelRequest
Request message for ModelService.ExportModel
.
Fields | |
---|---|
name |
Required. The resource name of the Model to export. The resource name may contain version id or version alias to specify the version, if no version is specified, the default version will be exported. |
output_ |
Required. The desired output location and configuration. |
OutputConfig
Output configuration for the Model export.
Fields | |
---|---|
export_ |
The ID of the format in which the Model must be exported. Each Model lists the |
artifact_ |
The Cloud Storage location where the Model artifact is to be written to. Under the directory given as the destination a new one with name " |
image_ |
The Google Container Registry or Artifact Registry uri where the Model container image will be copied to. This field should only be set when the |
ExportModelResponse
This type has no fields.
Response message of ModelService.ExportModel
operation.
ExportTensorboardTimeSeriesDataRequest
Request message for TensorboardService.ExportTensorboardTimeSeriesData
.
Fields | |
---|---|
tensorboard_ |
Required. The resource name of the TensorboardTimeSeries to export data from. Format: |
filter |
Exports the TensorboardTimeSeries' data that match the filter expression. |
page_ |
The maximum number of data points to return per page. The default page_size is 1000. Values must be between 1 and 10000. Values above 10000 are coerced to 10000. |
page_ |
A page token, received from a previous When paginating, all other parameters provided to |
order_ |
Field to use to sort the TensorboardTimeSeries' data. By default, TensorboardTimeSeries' data is returned in a pseudo random order. |
ExportTensorboardTimeSeriesDataResponse
Response message for TensorboardService.ExportTensorboardTimeSeriesData
.
Fields | |
---|---|
time_ |
The returned time series data points. |
next_ |
A token, which can be sent as |
Fact
The fact used in grounding.
Fields | |
---|---|
query |
Query that is used to retrieve this fact. |
title |
If present, it refers to the title of this fact. |
uri |
If present, this uri links to the source of the fact. |
summary |
If present, the summary/snippet of the fact. |
vector_distance |
If present, the distance between the query vector and this fact vector. |
score |
If present, according to the underlying Vector DB and the selected metric type, the score can be either the distance or the similarity between the query and the fact and its range depends on the metric type. For example, if the metric type is COSINE_DISTANCE, it represents the distance between the query and the fact. The larger the distance, the less relevant the fact is to the query. The range is [0, 2], while 0 means the most relevant and 2 means the least relevant. |
FasterDeploymentConfig
Configuration for faster model deployment.
Fields | |
---|---|
fast_ |
If true, enable fast tryout feature for this deployed model. |
Feature
Feature Metadata information. For example, color is a feature that describes an apple.
Fields | |
---|---|
name |
Immutable. Name of the Feature. Format: The last part feature is assigned by the client. The feature can be up to 64 characters long and can consist only of ASCII Latin letters A-Z and a-z, underscore(_), and ASCII digits 0-9 starting with a letter. The value will be unique given an entity type. |
description |
Description of the Feature. |
value_ |
Immutable. Only applicable for Vertex AI Feature Store (Legacy). Type of Feature value. |
create_ |
Output only. Only applicable for Vertex AI Feature Store (Legacy). Timestamp when this EntityType was created. |
update_ |
Output only. Only applicable for Vertex AI Feature Store (Legacy). Timestamp when this EntityType was most recently updated. |
labels |
Optional. The labels with user-defined metadata to organize your Features. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one Feature (System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. |
etag |
Used to perform a consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
disable_ |
Optional. Only applicable for Vertex AI Feature Store (Legacy). If not set, use the monitoring_config defined for the EntityType this Feature belongs to. Only Features with type ( If set to true, all types of data monitoring are disabled despite the config on EntityType. |
monitoring_ |
Output only. Only applicable for Vertex AI Feature Store (Legacy). The list of historical stats and anomalies with specified objectives. |
version_ |
Only applicable for Vertex AI Feature Store. The name of the BigQuery Table/View column hosting data for this version. If no value is provided, will use feature_id. |
point_ |
Entity responsible for maintaining this feature. Can be comma separated list of email addresses or URIs. |
MonitoringStatsAnomaly
A list of historical SnapshotAnalysis
or ImportFeaturesAnalysis
stats requested by user, sorted by FeatureStatsAnomaly.start_time
descending.
Fields | |
---|---|
objective |
Output only. The objective for each stats. |
feature_ |
Output only. The stats and anomalies generated at specific timestamp. |
Objective
If the objective in the request is both Import Feature Analysis and Snapshot Analysis, this objective could be one of them. Otherwise, this objective should be the same as the objective in the request.
Enums | |
---|---|
OBJECTIVE_UNSPECIFIED |
If it's OBJECTIVE_UNSPECIFIED, monitoring_stats will be empty. |
IMPORT_FEATURE_ANALYSIS |
Stats are generated by Import Feature Analysis. |
SNAPSHOT_ANALYSIS |
Stats are generated by Snapshot Analysis. |
ValueType
Only applicable for Vertex AI Legacy Feature Store. An enum representing the value type of a feature.
Enums | |
---|---|
VALUE_TYPE_UNSPECIFIED |
The value type is unspecified. |
BOOL |
Used for Feature that is a boolean. |
BOOL_ARRAY |
Used for Feature that is a list of boolean. |
DOUBLE |
Used for Feature that is double. |
DOUBLE_ARRAY |
Used for Feature that is a list of double. |
INT64 |
Used for Feature that is INT64. |
INT64_ARRAY |
Used for Feature that is a list of INT64. |
STRING |
Used for Feature that is string. |
STRING_ARRAY |
Used for Feature that is a list of String. |
BYTES |
Used for Feature that is bytes. |
STRUCT |
Used for Feature that is struct. |
FeatureGroup
Vertex AI Feature Group.
Fields | |
---|---|
name |
Identifier. Name of the FeatureGroup. Format: |
create_ |
Output only. Timestamp when this FeatureGroup was created. |
update_ |
Output only. Timestamp when this FeatureGroup was last updated. |
etag |
Optional. Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
labels |
Optional. The labels with user-defined metadata to organize your FeatureGroup. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one FeatureGroup(System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. |
description |
Optional. Description of the FeatureGroup. |
Union field
|
|
big_ |
Indicates that features for this group come from BigQuery Table/View. By default treats the source as a sparse time series source. The BigQuery source table or view must have at least one entity ID column and a column named |
BigQuery
Input source type for BigQuery Tables and Views.
Fields | |
---|---|
big_ |
Required. Immutable. The BigQuery source URI that points to either a BigQuery Table or View. |
entity_ |
Optional. Columns to construct entity_id / row keys. If not provided defaults to |
static_ |
Optional. Set if the data source is not a time-series. |
time_ |
Optional. If the source is a time-series source, this can be set to control how downstream sources (ex: |
dense |
Optional. If set, all feature values will be fetched from a single row per unique entityId including nulls. If not set, will collapse all rows for each unique entityId into a singe row with any non-null values if present, if no non-null values are present will sync null. ex: If source has schema |
TimeSeries
Fields | |
---|---|
timestamp_ |
Optional. Column hosting timestamp values for a time-series source. Will be used to determine the latest |
FeatureNoiseSigma
Noise sigma by features. Noise sigma represents the standard deviation of the gaussian kernel that will be used to add noise to interpolated inputs prior to computing gradients.
Fields | |
---|---|
noise_ |
Noise sigma per feature. No noise is added to features that are not set. |
NoiseSigmaForFeature
Noise sigma for a single feature.
Fields | |
---|---|
name |
The name of the input feature for which noise sigma is provided. The features are defined in |
sigma |
This represents the standard deviation of the Gaussian kernel that will be used to add noise to the feature prior to computing gradients. Similar to |
FeatureOnlineStore
Vertex AI Feature Online Store provides a centralized repository for serving ML features and embedding indexes at low latency. The Feature Online Store is a top-level container.
Fields | |
---|---|
name |
Identifier. Name of the FeatureOnlineStore. Format: |
create_ |
Output only. Timestamp when this FeatureOnlineStore was created. |
update_ |
Output only. Timestamp when this FeatureOnlineStore was last updated. |
etag |
Optional. Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
labels |
Optional. The labels with user-defined metadata to organize your FeatureOnlineStore. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one FeatureOnlineStore(System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. |
state |
Output only. State of the featureOnlineStore. |
dedicated_ |
Optional. The dedicated serving endpoint for this FeatureOnlineStore, which is different from common Vertex service endpoint. |
encryption_ |
Optional. Customer-managed encryption key spec for data storage. If set, online store will be secured by this key. |
satisfies_ |
Output only. Reserved for future use. |
satisfies_ |
Output only. Reserved for future use. |
Union field
|
|
bigtable |
Contains settings for the Cloud Bigtable instance that will be created to serve featureValues for all FeatureViews under this FeatureOnlineStore. |
optimized |
Contains settings for the Optimized store that will be created to serve featureValues for all FeatureViews under this FeatureOnlineStore. When choose Optimized storage type, need to set |
Bigtable
Fields | |
---|---|
auto_ |
Required. Autoscaling config applied to Bigtable Instance. |
AutoScaling
Fields | |
---|---|
min_ |
Required. The minimum number of nodes to scale down to. Must be greater than or equal to 1. |
max_ |
Required. The maximum number of nodes to scale up to. Must be greater than or equal to min_node_count, and less than or equal to 10 times of 'min_node_count'. |
cpu_ |
Optional. A percentage of the cluster's CPU capacity. Can be from 10% to 80%. When a cluster's CPU utilization exceeds the target that you have set, Bigtable immediately adds nodes to the cluster. When CPU utilization is substantially lower than the target, Bigtable removes nodes. If not set will default to 50%. |
DedicatedServingEndpoint
The dedicated serving endpoint for this FeatureOnlineStore. Only need to set when you choose Optimized storage type. Public endpoint is provisioned by default.
Fields | |
---|---|
public_ |
Output only. This field will be populated with the domain name to use for this FeatureOnlineStore |
private_ |
Optional. Private service connect config. The private service connection is available only for Optimized storage type, not for embedding management now. If |
service_ |
Output only. The name of the service attachment resource. Populated if private service connect is enabled and after FeatureViewSync is created. |
Optimized
This type has no fields.
Optimized storage type
State
Possible states a featureOnlineStore can have.
Enums | |
---|---|
STATE_UNSPECIFIED |
Default value. This value is unused. |
STABLE |
State when the featureOnlineStore configuration is not being updated and the fields reflect the current configuration of the featureOnlineStore. The featureOnlineStore is usable in this state. |
UPDATING |
The state of the featureOnlineStore configuration when it is being updated. During an update, the fields reflect either the original configuration or the updated configuration of the featureOnlineStore. The featureOnlineStore is still usable in this state. |
FeatureSelector
Selector for Features of an EntityType.
Fields | |
---|---|
id_ |
Required. Matches Features based on ID. |
FeatureStatsAnomaly
Stats and Anomaly generated at specific timestamp for specific Feature. The start_time and end_time are used to define the time range of the dataset that current stats belongs to, e.g. prediction traffic is bucketed into prediction datasets by time window. If the Dataset is not defined by time window, start_time = end_time. Timestamp of the stats and anomalies always refers to end_time. Raw stats and anomalies are stored in stats_uri or anomaly_uri in the tensorflow defined protos. Field data_stats contains almost identical information with the raw stats in Vertex AI defined proto, for UI to display.
Fields | |
---|---|
score |
Feature importance score, only populated when cross-feature monitoring is enabled. For now only used to represent feature attribution score within range [0, 1] for |
stats_ |
Path of the stats file for current feature values in Cloud Storage bucket. Format: gs:// |
anomaly_ |
Path of the anomaly file for current feature values in Cloud Storage bucket. Format: gs:// |
distribution_ |
Deviation from the current stats to baseline stats. 1. For categorical feature, the distribution distance is calculated by L-inifinity norm. 2. For numerical feature, the distribution distance is calculated by Jensen–Shannon divergence. |
anomaly_ |
This is the threshold used when detecting anomalies. The threshold can be changed by user, so this one might be different from |
start_ |
The start timestamp of window where stats were generated. For objectives where time window doesn't make sense (e.g. Featurestore Snapshot Monitoring), start_time is only used to indicate the monitoring intervals, so it always equals to (end_time - monitoring_interval). |
end_ |
The end timestamp of window where stats were generated. For objectives where time window doesn't make sense (e.g. Featurestore Snapshot Monitoring), end_time indicates the timestamp of the data used to generate stats (e.g. timestamp we take snapshots for feature values). |
FeatureValue
Value for a feature.
Fields | |
---|---|
metadata |
Metadata of feature value. |
Union field value . Value for the feature. value can be only one of the following: |
|
bool_ |
Bool type feature value. |
double_ |
Double type feature value. |
int64_ |
Int64 feature value. |
string_ |
String feature value. |
bool_ |
A list of bool type feature value. |
double_ |
A list of double type feature value. |
int64_ |
A list of int64 type feature value. |
string_ |
A list of string type feature value. |
bytes_ |
Bytes feature value. |
struct_ |
A struct type feature value. |
Metadata
Metadata of feature value.
Fields | |
---|---|
generate_ |
Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future. |
FeatureValueDestination
A destination location for Feature values and format.
Fields | |
---|---|
Union field
|
|
bigquery_ |
Output in BigQuery format. |
tfrecord_ |
Output in TFRecord format. Below are the mapping from Feature value type in Featurestore to Feature value type in TFRecord:
|
csv_ |
Output in CSV format. Array Feature value types are not allowed in CSV format. |
FeatureValueList
Container for list of values.
Fields | |
---|---|
values[] |
A list of feature values. All of them should be the same data type. |
FeatureView
FeatureView is representation of values that the FeatureOnlineStore will serve based on its syncConfig.
Fields | |
---|---|
name |
Identifier. Name of the FeatureView. Format: |
create_ |
Output only. Timestamp when this FeatureView was created. |
update_ |
Output only. Timestamp when this FeatureView was last updated. |
etag |
Optional. Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
labels |
Optional. The labels with user-defined metadata to organize your FeatureViews. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one FeatureOnlineStore(System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. |
sync_ |
Configures when data is to be synced/updated for this FeatureView. At the end of the sync the latest featureValues for each entityId of this FeatureView are made ready for online serving. |
index_ |
Optional. Configuration for index preparation for vector search. It contains the required configurations to create an index from source data, so that approximate nearest neighbor (a.k.a ANN) algorithms search can be performed during online serving. |
satisfies_ |
Output only. Reserved for future use. |
satisfies_ |
Output only. Reserved for future use. |
Union field
|
|
big_ |
Optional. Configures how data is supposed to be extracted from a BigQuery source to be loaded onto the FeatureOnlineStore. |
feature_ |
Optional. Configures the features from a Feature Registry source that need to be loaded onto the FeatureOnlineStore. |
vertex_ |
Optional. The Vertex RAG Source that the FeatureView is linked to. |
BigQuerySource
Fields | |
---|---|
uri |
Required. The BigQuery view URI that will be materialized on each sync trigger based on FeatureView.SyncConfig. |
entity_ |
Required. Columns to construct entity_id / row keys. |
FeatureRegistrySource
A Feature Registry source for features that need to be synced to Online Store.
Fields | |
---|---|
feature_ |
Required. List of features that need to be synced to Online Store. |
project_ |
Optional. The project number of the parent project of the Feature Groups. |
FeatureGroup
Features belonging to a single feature group that will be synced to Online Store.
Fields | |
---|---|
feature_ |
Required. Identifier of the feature group. |
feature_ |
Required. Identifiers of features under the feature group. |
IndexConfig
Configuration for vector indexing.
Fields | |
---|---|
embedding_ |
Optional. Column of embedding. This column contains the source data to create index for vector search. embedding_column must be set when using vector search. |
filter_ |
Optional. Columns of features that're used to filter vector search results. |
crowding_ |
Optional. Column of crowding. This column contains crowding attribute which is a constraint on a neighbor list produced by |
distance_ |
Optional. The distance measure used in nearest neighbor search. |
Union field algorithm_config . The configuration with regard to the algorithms used for efficient search. algorithm_config can be only one of the following: |
|
tree_ |
Optional. Configuration options for the tree-AH algorithm (Shallow tree + Asymmetric Hashing). Please refer to this paper for more details: https://arxiv.org/abs/1908.10396 |
brute_ |
Optional. Configuration options for using brute force search, which simply implements the standard linear search in the database for each query. It is primarily meant for benchmarking and to generate the ground truth for approximate search. |
embedding_ |
Optional. The number of dimensions of the input embedding. |
BruteForceConfig
This type has no fields.
Configuration options for using brute force search.
DistanceMeasureType
The distance measure used in nearest neighbor search.
Enums | |
---|---|
DISTANCE_MEASURE_TYPE_UNSPECIFIED |
Should not be set. |
SQUARED_L2_DISTANCE |
Euclidean (L_2) Distance. |
COSINE_DISTANCE |
Cosine Distance. Defined as 1 - cosine similarity. We strongly suggest using DOT_PRODUCT_DISTANCE + UNIT_L2_NORM instead of COSINE distance. Our algorithms have been more optimized for DOT_PRODUCT distance which, when combined with UNIT_L2_NORM, is mathematically equivalent to COSINE distance and results in the same ranking. |
DOT_PRODUCT_DISTANCE |
Dot Product Distance. Defined as a negative of the dot product. |
TreeAHConfig
Configuration options for the tree-AH algorithm.
Fields | |
---|---|
leaf_ |
Optional. Number of embeddings on each leaf node. The default value is 1000 if not set. |
SyncConfig
Configuration for Sync. Only one option is set.
Fields | |
---|---|
cron |
Cron schedule (https://en.wikipedia.org/wiki/Cron) to launch scheduled runs. To explicitly set a timezone to the cron tab, apply a prefix in the cron tab: "CRON_TZ=${IANA_TIME_ZONE}" or "TZ=${IANA_TIME_ZONE}". The ${IANA_TIME_ZONE} may only be a valid string from IANA time zone database. For example, "CRON_TZ=America/New_York 1 * * * *", or "TZ=America/New_York 1 * * * *". |
continuous |
Optional. If true, syncs the FeatureView in a continuous manner to Online Store. |
VertexRagSource
A Vertex Rag source for features that need to be synced to Online Store.
Fields | |
---|---|
uri |
Required. The BigQuery view/table URI that will be materialized on each manual sync trigger. The table/view is expected to have the following columns and types at least: - |
rag_ |
Optional. The RAG corpus id corresponding to this FeatureView. |
FeatureViewDataFormat
Format of the data in the Feature View.
Enums | |
---|---|
FEATURE_VIEW_DATA_FORMAT_UNSPECIFIED |
Not set. Will be treated as the KeyValue format. |
KEY_VALUE |
Return response data in key-value format. |
PROTO_STRUCT |
Return response data in proto Struct format. |
FeatureViewDataKey
Lookup key for a feature view.
Fields | |
---|---|
Union field
|
|
key |
String key to use for lookup. |
composite_ |
The actual Entity ID will be composed from this struct. This should match with the way ID is defined in the FeatureView spec. |
CompositeKey
ID that is comprised from several parts (columns).
Fields | |
---|---|
parts[] |
Parts to construct Entity ID. Should match with the same ID columns as defined in FeatureView in the same order. |
FeatureViewSync
FeatureViewSync is a representation of sync operation which copies data from data source to Feature View in Online Store.
Fields | |
---|---|
name |
Identifier. Name of the FeatureViewSync. Format: |
create_ |
Output only. Time when this FeatureViewSync is created. Creation of a FeatureViewSync means that the job is pending / waiting for sufficient resources but may not have started the actual data transfer yet. |
run_ |
Output only. Time when this FeatureViewSync is finished. |
final_ |
Output only. Final status of the FeatureViewSync. |
sync_ |
Output only. Summary of the sync job. |
satisfies_ |
Output only. Reserved for future use. |
satisfies_ |
Output only. Reserved for future use. |
SyncSummary
Summary from the Sync job. For continuous syncs, the summary is updated periodically. For batch syncs, it gets updated on completion of the sync.
Fields | |
---|---|
row_ |
Output only. Total number of rows synced. |
total_ |
Output only. BigQuery slot milliseconds consumed for the sync job. |
system_ |
Lower bound of the system time watermark for the sync job. This is only set for continuously syncing feature views. |
Featurestore
Vertex AI Feature Store provides a centralized repository for organizing, storing, and serving ML features. The Featurestore is a top-level container for your features and their values.
Fields | |
---|---|
name |
Output only. Name of the Featurestore. Format: |
create_ |
Output only. Timestamp when this Featurestore was created. |
update_ |
Output only. Timestamp when this Featurestore was last updated. |
etag |
Optional. Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
labels |
Optional. The labels with user-defined metadata to organize your Featurestore. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one Featurestore(System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. |
online_ |
Optional. Config for online storage resources. The field should not co-exist with the field of |
state |
Output only. State of the featurestore. |
online_ |
Optional. TTL in days for feature values that will be stored in online serving storage. The Feature Store online storage periodically removes obsolete feature values older than |
encryption_ |
Optional. Customer-managed encryption key spec for data storage. If set, both of the online and offline data storage will be secured by this key. |
satisfies_ |
Output only. Reserved for future use. |
satisfies_ |
Output only. Reserved for future use. |
OnlineServingConfig
OnlineServingConfig specifies the details for provisioning online serving resources.
Fields | |
---|---|
fixed_ |
The number of nodes for the online store. The number of nodes doesn't scale automatically, but you can manually update the number of nodes. If set to 0, the featurestore will not have an online store and cannot be used for online serving. |
scaling |
Online serving scaling configuration. Only one of |
Scaling
Online serving scaling configuration. If min_node_count and max_node_count are set to the same value, the cluster will be configured with the fixed number of node (no auto-scaling).
Fields | |
---|---|
min_ |
Required. The minimum number of nodes to scale down to. Must be greater than or equal to 1. |
max_ |
The maximum number of nodes to scale up to. Must be greater than min_node_count, and less than or equal to 10 times of 'min_node_count'. |
cpu_ |
Optional. The cpu utilization that the Autoscaler should be trying to achieve. This number is on a scale from 0 (no utilization) to 100 (total utilization), and is limited between 10 and 80. When a cluster's CPU utilization exceeds the target that you have set, Bigtable immediately adds nodes to the cluster. When CPU utilization is substantially lower than the target, Bigtable removes nodes. If not set or set to 0, default to 50. |
State
Possible states a featurestore can have.
Enums | |
---|---|
STATE_UNSPECIFIED |
Default value. This value is unused. |
STABLE |
State when the featurestore configuration is not being updated and the fields reflect the current configuration of the featurestore. The featurestore is usable in this state. |
UPDATING |
The state of the featurestore configuration when it is being updated. During an update, the fields reflect either the original configuration or the updated configuration of the featurestore. For example, online_serving_config.fixed_node_count can take minutes to update. While the update is in progress, the featurestore is in the UPDATING state, and the value of fixed_node_count can be the original value or the updated value, depending on the progress of the operation. Until the update completes, the actual number of nodes can still be the original value of fixed_node_count . The featurestore is still usable in this state. |
FeaturestoreMonitoringConfig
Configuration of how features in Featurestore are monitored.
Fields | |
---|---|
snapshot_ |
The config for Snapshot Analysis Based Feature Monitoring. |
import_ |
The config for ImportFeatures Analysis Based Feature Monitoring. |
numerical_ |
Threshold for numerical features of anomaly detection. This is shared by all objectives of Featurestore Monitoring for numerical features (i.e. Features with type ( |
categorical_ |
Threshold for categorical features of anomaly detection. This is shared by all types of Featurestore Monitoring for categorical features (i.e. Features with type ( |
ImportFeaturesAnalysis
Configuration of the Featurestore's ImportFeature Analysis Based Monitoring. This type of analysis generates statistics for values of each Feature imported by every ImportFeatureValues
operation.
Fields | |
---|---|
state |
Whether to enable / disable / inherite default hebavior for import features analysis. |
anomaly_ |
The baseline used to do anomaly detection for the statistics generated by import features analysis. |
Baseline
Defines the baseline to do anomaly detection for feature values imported by each ImportFeatureValues
operation.
Enums | |
---|---|
BASELINE_UNSPECIFIED |
Should not be used. |
LATEST_STATS |
Choose the later one statistics generated by either most recent snapshot analysis or previous import features analysis. If non of them exists, skip anomaly detection and only generate a statistics. |
MOST_RECENT_SNAPSHOT_STATS |
Use the statistics generated by the most recent snapshot analysis if exists. |
PREVIOUS_IMPORT_FEATURES_STATS |
Use the statistics generated by the previous import features analysis if exists. |
State
The state defines whether to enable ImportFeature analysis.
Enums | |
---|---|
STATE_UNSPECIFIED |
Should not be used. |
DEFAULT |
The default behavior of whether to enable the monitoring. EntityType-level config: disabled. Feature-level config: inherited from the configuration of EntityType this Feature belongs to. |
ENABLED |
Explicitly enables import features analysis. EntityType-level config: by default enables import features analysis for all Features under it. Feature-level config: enables import features analysis regardless of the EntityType-level config. |
DISABLED |
Explicitly disables import features analysis. EntityType-level config: by default disables import features analysis for all Features under it. Feature-level config: disables import features analysis regardless of the EntityType-level config. |
SnapshotAnalysis
Configuration of the Featurestore's Snapshot Analysis Based Monitoring. This type of analysis generates statistics for each Feature based on a snapshot of the latest feature value of each entities every monitoring_interval.
Fields | |
---|---|
disabled |
The monitoring schedule for snapshot analysis. For EntityType-level config: unset / disabled = true indicates disabled by default for Features under it; otherwise by default enable snapshot analysis monitoring with monitoring_interval for Features under it. Feature-level config: disabled = true indicates disabled regardless of the EntityType-level config; unset monitoring_interval indicates going with EntityType-level config; otherwise run snapshot analysis monitoring with monitoring_interval regardless of the EntityType-level config. Explicitly Disable the snapshot analysis based monitoring. |
monitoring_ |
Configuration of the snapshot analysis based monitoring pipeline running interval. The value indicates number of days. |
staleness_ |
Customized export features time window for snapshot analysis. Unit is one day. Default value is 3 weeks. Minimum value is 1 day. Maximum value is 4000 days. |
ThresholdConfig
The config for Featurestore Monitoring threshold.
Fields | |
---|---|
Union field
|
|
value |
Specify a threshold value that can trigger the alert. 1. For categorical feature, the distribution distance is calculated by L-inifinity norm. 2. For numerical feature, the distribution distance is calculated by Jensen–Shannon divergence. Each feature must have a non-zero threshold if they need to be monitored. Otherwise no alert will be triggered for that feature. |
FetchFeatureValuesRequest
Request message for FeatureOnlineStoreService.FetchFeatureValues
. All the features under the requested feature view will be returned.
Fields | |
---|---|
feature_ |
Required. FeatureView resource format |
data_ |
Optional. The request key to fetch feature values for. |
data_ |
Optional. Response data format. If not set, |
FetchFeatureValuesResponse
Response message for FeatureOnlineStoreService.FetchFeatureValues
Fields | |
---|---|
data_ |
The data key associated with this response. Will only be populated for [FeatureOnlineStoreService.StreamingFetchFeatureValues][] RPCs. |
Union field
|
|
key_ |
Feature values in KeyValue format. |
proto_ |
Feature values in proto Struct format. |
FeatureNameValuePairList
Response structure in the format of key (feature name) and (feature) value pair.
Fields | |
---|---|
features[] |
List of feature names and values. |
FeatureNameValuePair
Feature name & value pair.
Fields | |
---|---|
name |
Feature short name. |
Union field
|
|
value |
Feature value. |
FileData
URI based data.
Fields | |
---|---|
mime_ |
Required. The IANA standard MIME type of the source data. |
file_ |
Required. URI. |
FileStatus
RagFile status.
Fields | |
---|---|
state |
Output only. RagFile state. |
error_ |
Output only. Only when the |
State
RagFile state.
Enums | |
---|---|
STATE_UNSPECIFIED |
RagFile state is unspecified. |
ACTIVE |
RagFile resource has been created and indexed successfully. |
ERROR |
RagFile resource is in a problematic state. See error_message field for details. |
FilterSplit
Assigns input data to training, validation, and test sets based on the given filters, data pieces not matched by any filter are ignored. Currently only supported for Datasets containing DataItems. If any of the filters in this message are to match nothing, then they can be set as '-' (the minus sign).
Supported only for unstructured Datasets.
Fields | |
---|---|
training_ |
Required. A filter on DataItems of the Dataset. DataItems that match this filter are used to train the Model. A filter with same syntax as the one used in |
validation_ |
Required. A filter on DataItems of the Dataset. DataItems that match this filter are used to validate the Model. A filter with same syntax as the one used in |
test_ |
Required. A filter on DataItems of the Dataset. DataItems that match this filter are used to test the Model. A filter with same syntax as the one used in |
FluencyInput
Input for fluency metric.
Fields | |
---|---|
metric_ |
Required. Spec for fluency score metric. |
instance |
Required. Fluency instance. |
FluencyInstance
Spec for fluency instance.
Fields | |
---|---|
prediction |
Required. Output of the evaluated model. |
FluencyResult
Spec for fluency result.
Fields | |
---|---|
explanation |
Output only. Explanation for fluency score. |
score |
Output only. Fluency score. |
confidence |
Output only. Confidence for fluency score. |
FluencySpec
Spec for fluency score metric.
Fields | |
---|---|
version |
Optional. Which version to use for evaluation. |
FractionSplit
Assigns the input data to training, validation, and test sets as per the given fractions. Any of training_fraction
, validation_fraction
and test_fraction
may optionally be provided, they must sum to up to 1. If the provided ones sum to less than 1, the remainder is assigned to sets as decided by Vertex AI. If none of the fractions are set, by default roughly 80% of data is used for training, 10% for validation, and 10% for test.
Fields | |
---|---|
training_ |
The fraction of the input data that is to be used to train the Model. |
validation_ |
The fraction of the input data that is to be used to validate the Model. |
test_ |
The fraction of the input data that is to be used to evaluate the Model. |
FulfillmentInput
Input for fulfillment metric.
Fields | |
---|---|
metric_ |
Required. Spec for fulfillment score metric. |
instance |
Required. Fulfillment instance. |
FulfillmentInstance
Spec for fulfillment instance.
Fields | |
---|---|
prediction |
Required. Output of the evaluated model. |
instruction |
Required. Inference instruction prompt to compare prediction with. |
FulfillmentResult
Spec for fulfillment result.
Fields | |
---|---|
explanation |
Output only. Explanation for fulfillment score. |
score |
Output only. Fulfillment score. |
confidence |
Output only. Confidence for fulfillment score. |
FulfillmentSpec
Spec for fulfillment metric.
Fields | |
---|---|
version |
Optional. Which version to use for evaluation. |
FunctionCall
A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values.
Fields | |
---|---|
name |
Required. The name of the function to call. Matches [FunctionDeclaration.name]. |
args |
Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. |
FunctionCallingConfig
Function calling config.
Fields | |
---|---|
mode |
Optional. Function calling mode. |
allowed_ |
Optional. Function names to call. Only set when the Mode is ANY. Function names should match [FunctionDeclaration.name]. With mode set to ANY, model will predict a function call from the set of function names provided. |
Mode
Function calling mode.
Enums | |
---|---|
MODE_UNSPECIFIED |
Unspecified function calling mode. This value should not be used. |
AUTO |
Default model behavior, model decides to predict either function calls or natural language response. |
ANY |
Model is constrained to always predicting function calls only. If "allowed_function_names" are set, the predicted function calls will be limited to any one of "allowed_function_names", else the predicted function calls will be any one of the provided "function_declarations". |
NONE |
Model will not predict any function calls. Model behavior is same as when not passing any function declarations. |
FunctionDeclaration
Structured representation of a function declaration as defined by the OpenAPI 3.0 specification. Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a Tool
by the model and executed by the client.
Fields | |
---|---|
name |
Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. |
description |
Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. |
parameters |
Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 |
response |
Optional. Describes the output from this function in JSON Schema format. Reflects the Open API 3.03 Response Object. The Schema defines the type used for the response value of the function. |
FunctionResponse
The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction.
Fields | |
---|---|
name |
Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. |
response |
Required. The function response in JSON object format. Use "output" key to specify function output and "error" key to specify error details (if any). If "output" and "error" keys are not specified, then whole "response" is treated as function output. |
GcsDestination
The Google Cloud Storage location where the output is to be written to.
Fields | |
---|---|
output_ |
Required. Google Cloud Storage URI to output directory. If the uri doesn't end with '/', a '/' will be automatically appended. The directory is created if it doesn't exist. |
GcsSource
The Google Cloud Storage location for the input content.
Fields | |
---|---|
uris[] |
Required. Google Cloud Storage URI(-s) to the input file(s). May contain wildcards. For more information on wildcards, see https://cloud.google.com/storage/docs/gsutil/addlhelp/WildcardNames. |
GenerateContentRequest
Request message for [PredictionService.GenerateContent].
Fields | |
---|---|
model |
Required. The fully qualified name of the publisher model or tuned model endpoint to use. Publisher model format: Tuned model endpoint format: |
contents[] |
Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries, this is a repeated field that contains conversation history + latest request. |
cached_ |
Optional. The name of the cached content used as context to serve the prediction. Note: only used in explicit caching, where users can have control over caching (e.g. what content to cache) and enjoy guaranteed cost savings. Format: |
tools[] |
Optional. A list of A |
tool_ |
Optional. Tool config. This config is shared for all tools provided in the request. |
labels |
Optional. The labels with user-defined metadata for the request. It is used for billing and reporting only. Label keys and values can be no longer than 63 characters (Unicode codepoints) and can only contain lowercase letters, numeric characters, underscores, and dashes. International characters are allowed. Label values are optional. Label keys must start with a letter. |
safety_ |
Optional. Per request settings for blocking unsafe content. Enforced on GenerateContentResponse.candidates. |
generation_ |
Optional. Generation config. |
system_ |
Optional. The user provided system instructions for the model. Note: only text should be used in parts and content in each part will be in a separate paragraph. |
GenerateContentResponse
Response message for [PredictionService.GenerateContent].
Fields | |
---|---|
candidates[] |
Output only. Generated candidates. |
model_ |
Output only. The model version used to generate the response. |
prompt_ |
Output only. Content filter results for a prompt sent in the request. Note: Sent only in the first stream chunk. Only happens when no candidates were generated due to content violations. |
usage_ |
Usage metadata about the response(s). |
PromptFeedback
Content filter results for a prompt sent in the request.
Fields | |
---|---|
block_ |
Output only. Blocked reason. |
safety_ |
Output only. Safety ratings. |
block_ |
Output only. A readable block reason message. |
BlockedReason
Blocked reason enumeration.
Enums | |
---|---|
BLOCKED_REASON_UNSPECIFIED |
Unspecified blocked reason. |
SAFETY |
Candidates blocked due to safety. |
OTHER |
Candidates blocked due to other reason. |
BLOCKLIST |
Candidates blocked due to the terms which are included from the terminology blocklist. |
PROHIBITED_CONTENT |
Candidates blocked due to prohibited content. |
UsageMetadata
Usage metadata about response(s).
Fields | |
---|---|
prompt_ |
Number of tokens in the request. When |
candidates_ |
Number of tokens in the response(s). |
total_ |
Total token count for prompt and response candidates. |
cached_ |
Output only. Number of tokens in the cached part in the input (the cached content). |
GenerationConfig
Generation config.
Fields | |
---|---|
stop_ |
Optional. Stop sequences. |
response_ |
Optional. Output response mimetype of the generated candidate text. Supported mimetype: - |
temperature |
Optional. Controls the randomness of predictions. |
top_ |
Optional. If specified, nucleus sampling will be used. |
top_ |
Optional. If specified, top-k sampling will be used. |
candidate_ |
Optional. Number of candidates to generate. |
max_ |
Optional. The maximum number of output tokens to generate per message. |
response_ |
Optional. If true, export the logprobs results in response. |
logprobs |
Optional. Logit probabilities. |
presence_ |
Optional. Positive penalties. |
frequency_ |
Optional. Frequency penalties. |
seed |
Optional. Seed. |
response_ |
Optional. The |
routing_ |
Optional. Routing configuration. |
RoutingConfig
The configuration for routing the request to a specific model.
Fields | |
---|---|
Union field routing_config . Routing mode. routing_config can be only one of the following: |
|
auto_ |
Automated routing. |
manual_ |
Manual routing. |
AutoRoutingMode
When automated routing is specified, the routing will be determined by the pretrained routing model and customer provided model routing preference.
Fields | |
---|---|
model_ |
The model routing preference. |
ModelRoutingPreference
The model routing preference.
Enums | |
---|---|
UNKNOWN |
Unspecified model routing preference. |
PRIORITIZE_QUALITY |
Prefer higher quality over low cost. |
BALANCED |
Balanced model routing preference. |
PRIORITIZE_COST |
Prefer lower cost over higher quality. |
ManualRoutingMode
When manual routing is set, the specified model will be used directly.
Fields | |
---|---|
model_ |
The model name to use. Only the public LLM models are accepted. e.g. 'gemini-1.5-pro-001'. |
GenericOperationMetadata
Generic Metadata shared by all operations.
Fields | |
---|---|
partial_ |
Output only. Partial failures encountered. E.g. single files that couldn't be read. This field should never exceed 20 entries. Status details field will contain standard Google Cloud error details. |
create_ |
Output only. Time when the operation was created. |
update_ |
Output only. Time when the operation was updated for the last time. If the operation has finished (successfully or not), this is the finish time. |
GenieSource
Contains information about the source of the models generated from Generative AI Studio.
Fields | |
---|---|
base_ |
Required. The public base model URI. |
GetAnnotationSpecRequest
Request message for DatasetService.GetAnnotationSpec
.
Fields | |
---|---|
name |
Required. The name of the AnnotationSpec resource. Format: |
read_ |
Mask specifying which fields to read. |
GetArtifactRequest
Request message for MetadataService.GetArtifact
.
Fields | |
---|---|
name |
Required. The resource name of the Artifact to retrieve. Format: |
GetBatchPredictionJobRequest
Request message for JobService.GetBatchPredictionJob
.
Fields | |
---|---|
name |
Required. The name of the BatchPredictionJob resource. Format: |
GetContextRequest
Request message for MetadataService.GetContext
.
Fields | |
---|---|
name |
Required. The resource name of the Context to retrieve. Format: |
GetCustomJobRequest
Request message for JobService.GetCustomJob
.
Fields | |
---|---|
name |
Required. The name of the CustomJob resource. Format: |
GetDatasetRequest
Request message for DatasetService.GetDataset
. Next ID: 4
Fields | |
---|---|
name |
Required. The name of the Dataset resource. |
read_ |
Mask specifying which fields to read. |
GetDatasetVersionRequest
Request message for DatasetService.GetDatasetVersion
. Next ID: 4
Fields | |
---|---|
name |
Required. The resource name of the Dataset version to delete. Format: |
read_ |
Mask specifying which fields to read. |
GetDeploymentResourcePoolRequest
Request message for GetDeploymentResourcePool method.
Fields | |
---|---|
name |
Required. The name of the DeploymentResourcePool to retrieve. Format: |
GetEndpointRequest
Request message for EndpointService.GetEndpoint
Fields | |
---|---|
name |
Required. The name of the Endpoint resource. Format: |
GetEntityTypeRequest
Request message for FeaturestoreService.GetEntityType
.
Fields | |
---|---|
name |
Required. The name of the EntityType resource. Format: |
GetExecutionRequest
Request message for MetadataService.GetExecution
.
Fields | |
---|---|
name |
Required. The resource name of the Execution to retrieve. Format: |
GetFeatureGroupRequest
Request message for FeatureRegistryService.GetFeatureGroup
.
Fields | |
---|---|
name |
Required. The name of the FeatureGroup resource. |
GetFeatureOnlineStoreRequest
Request message for FeatureOnlineStoreAdminService.GetFeatureOnlineStore
.
Fields | |
---|---|
name |
Required. The name of the FeatureOnlineStore resource. |
GetFeatureRequest
Request message for FeaturestoreService.GetFeature
. Request message for FeatureRegistryService.GetFeature
.
Fields | |
---|---|
name |
Required. The name of the Feature resource. Format for entity_type as parent: |
GetFeatureViewRequest
Request message for FeatureOnlineStoreAdminService.GetFeatureView
.
Fields | |
---|---|
name |
Required. The name of the FeatureView resource. Format: |
GetFeatureViewSyncRequest
Request message for FeatureOnlineStoreAdminService.GetFeatureViewSync
.
Fields | |
---|---|
name |
Required. The name of the FeatureViewSync resource. Format: |
GetFeaturestoreRequest
Request message for FeaturestoreService.GetFeaturestore
.
Fields | |
---|---|
name |
Required. The name of the Featurestore resource. |
GetHyperparameterTuningJobRequest
Request message for JobService.GetHyperparameterTuningJob
.
Fields | |
---|---|
name |
Required. The name of the HyperparameterTuningJob resource. Format: |
GetIndexEndpointRequest
Request message for IndexEndpointService.GetIndexEndpoint
Fields | |
---|---|
name |
Required. The name of the IndexEndpoint resource. Format: |
GetIndexRequest
Request message for IndexService.GetIndex
Fields | |
---|---|
name |
Required. The name of the Index resource. Format: |
GetMetadataSchemaRequest
Request message for MetadataService.GetMetadataSchema
.
Fields | |
---|---|
name |
Required. The resource name of the MetadataSchema to retrieve. Format: |
GetMetadataStoreRequest
Request message for MetadataService.GetMetadataStore
.
Fields | |
---|---|
name |
Required. The resource name of the MetadataStore to retrieve. Format: |
GetModelDeploymentMonitoringJobRequest
Request message for JobService.GetModelDeploymentMonitoringJob
.
Fields | |
---|---|
name |
Required. The resource name of the ModelDeploymentMonitoringJob. Format: |
GetModelEvaluationRequest
Request message for ModelService.GetModelEvaluation
.
Fields | |
---|---|
name |
Required. The name of the ModelEvaluation resource. Format: |
GetModelEvaluationSliceRequest
Request message for ModelService.GetModelEvaluationSlice
.
Fields | |
---|---|
name |
Required. The name of the ModelEvaluationSlice resource. Format: |
GetModelRequest
Request message for ModelService.GetModel
.
Fields | |
---|---|
name |
Required. The name of the Model resource. Format: In order to retrieve a specific version of the model, also provide the version ID or version alias. Example: |
GetNasJobRequest
Request message for JobService.GetNasJob
.
Fields | |
---|---|
name |
Required. The name of the NasJob resource. Format: |
GetNasTrialDetailRequest
Request message for JobService.GetNasTrialDetail
.
Fields | |
---|---|
name |
Required. The name of the NasTrialDetail resource. Format: |
GetNotebookExecutionJobRequest
Request message for [NotebookService.GetNotebookExecutionJob]
Fields | |
---|---|
name |
Required. The name of the NotebookExecutionJob resource. |
view |
Optional. The NotebookExecutionJob view. Defaults to BASIC. |
GetNotebookRuntimeRequest
Request message for NotebookService.GetNotebookRuntime
Fields | |
---|---|
name |
Required. The name of the NotebookRuntime resource. Instead of checking whether the name is in valid NotebookRuntime resource name format, directly throw NotFound exception if there is no such NotebookRuntime in spanner. |
GetNotebookRuntimeTemplateRequest
Request message for NotebookService.GetNotebookRuntimeTemplate
Fields | |
---|---|
name |
Required. The name of the NotebookRuntimeTemplate resource. Format: |
GetPersistentResourceRequest
Request message for PersistentResourceService.GetPersistentResource
.
Fields | |
---|---|
name |
Required. The name of the PersistentResource resource. Format: |
GetPipelineJobRequest
Request message for PipelineService.GetPipelineJob
.
Fields | |
---|---|
name |
Required. The name of the PipelineJob resource. Format: |
GetPublisherModelRequest
Request message for ModelGardenService.GetPublisherModel
Fields | |
---|---|
name |
Required. The name of the PublisherModel resource. Format: |
language_ |
Optional. The IETF BCP-47 language code representing the language in which the publisher model's text information should be written in. |
view |
Optional. PublisherModel view specifying which fields to read. |
is_ |
Optional. Boolean indicates whether the requested model is a Hugging Face model. |
hugging_ |
Optional. Token used to access Hugging Face gated models. |
GetRagCorpusRequest
Request message for VertexRagDataService.GetRagCorpus
Fields | |
---|---|
name |
Required. The name of the RagCorpus resource. Format: |
GetRagFileRequest
Request message for VertexRagDataService.GetRagFile
Fields | |
---|---|
name |
Required. The name of the RagFile resource. Format: |
GetScheduleRequest
Request message for ScheduleService.GetSchedule
.
Fields | |
---|---|
name |
Required. The name of the Schedule resource. Format: |
GetSpecialistPoolRequest
Request message for SpecialistPoolService.GetSpecialistPool
.
Fields | |
---|---|
name |
Required. The name of the SpecialistPool resource. The form is |
GetStudyRequest
Request message for VizierService.GetStudy
.
Fields | |
---|---|
name |
Required. The name of the Study resource. Format: |
GetTensorboardExperimentRequest
Request message for TensorboardService.GetTensorboardExperiment
.
Fields | |
---|---|
name |
Required. The name of the TensorboardExperiment resource. Format: |
GetTensorboardRequest
Request message for TensorboardService.GetTensorboard
.
Fields | |
---|---|
name |
Required. The name of the Tensorboard resource. Format: |
GetTensorboardRunRequest
Request message for TensorboardService.GetTensorboardRun
.
Fields | |
---|---|
name |
Required. The name of the TensorboardRun resource. Format: |
GetTensorboardTimeSeriesRequest
Request message for TensorboardService.GetTensorboardTimeSeries
.
Fields | |
---|---|
name |
Required. The name of the TensorboardTimeSeries resource. Format: |
GetTrainingPipelineRequest
Request message for PipelineService.GetTrainingPipeline
.
Fields | |
---|---|
name |
Required. The name of the TrainingPipeline resource. Format: |
GetTrialRequest
Request message for VizierService.GetTrial
.
Fields | |
---|---|
name |
Required. The name of the Trial resource. Format: |
GetTuningJobRequest
Request message for GenAiTuningService.GetTuningJob
.
Fields | |
---|---|
name |
Required. The name of the TuningJob resource. Format: |
GoogleDriveSource
The Google Drive location for the input content.
Fields | |
---|---|
resource_ |
Required. Google Drive resource IDs. |
ResourceId
The type and ID of the Google Drive resource.
Fields | |
---|---|
resource_ |
Required. The type of the Google Drive resource. |
resource_ |
Required. The ID of the Google Drive resource. |
ResourceType
The type of the Google Drive resource.
Enums | |
---|---|
RESOURCE_TYPE_UNSPECIFIED |
Unspecified resource type. |
RESOURCE_TYPE_FILE |
File resource type. |
RESOURCE_TYPE_FOLDER |
Folder resource type. |
GoogleSearchRetrieval
Tool to retrieve public web data for grounding, powered by Google.
Fields | |
---|---|
dynamic_ |
Specifies the dynamic retrieval configuration for the given source. |
GroundednessInput
Input for groundedness metric.
Fields | |
---|---|
metric_ |
Required. Spec for groundedness metric. |
instance |
Required. Groundedness instance. |
GroundednessInstance
Spec for groundedness instance.
Fields | |
---|---|
prediction |
Required. Output of the evaluated model. |
context |
Required. Background information provided in context used to compare against the prediction. |
GroundednessResult
Spec for groundedness result.
Fields | |
---|---|
explanation |
Output only. Explanation for groundedness score. |
score |
Output only. Groundedness score. |
confidence |
Output only. Confidence for groundedness score. |
GroundednessSpec
Spec for groundedness metric.
Fields | |
---|---|
version |
Optional. Which version to use for evaluation. |
GroundingChunk
Grounding chunk.
Fields | |
---|---|
Union field chunk_type . Chunk type. chunk_type can be only one of the following: |
|
web |
Grounding chunk from the web. |
retrieved_ |
Grounding chunk from context retrieved by the retrieval tools. |
RetrievedContext
Chunk from context retrieved by the retrieval tools.
Fields | |
---|---|
uri |
URI reference of the attribution. |
title |
Title of the attribution. |
text |
Text of the attribution. |
Web
Chunk from the web.
Fields | |
---|---|
uri |
URI reference of the chunk. |
title |
Title of the chunk. |
GroundingMetadata
Metadata returned to client when grounding is enabled.
Fields | |
---|---|
web_ |
Optional. Web search queries for the following-up web search. |
grounding_ |
List of supporting references retrieved from specified grounding source. |
grounding_ |
Optional. List of grounding support. |
search_ |
Optional. Google search entry for the following-up web searches. |
retrieval_ |
Optional. Output only. Retrieval metadata. |
GroundingSupport
Grounding support.
Fields | |
---|---|
grounding_ |
A list of indices (into 'grounding_chunk') specifying the citations associated with the claim. For instance [1,3,4] means that grounding_chunk[1], grounding_chunk[3], grounding_chunk[4] are the retrieved content attributed to the claim. |
confidence_ |
Confidence score of the support references. Ranges from 0 to 1. 1 is the most confident. This list must have the same size as the grounding_chunk_indices. |
segment |
Segment of the content this support belongs to. |
HarmCategory
Harm categories that will block the content.
Enums | |
---|---|
HARM_CATEGORY_UNSPECIFIED |
The harm category is unspecified. |
HARM_CATEGORY_HATE_SPEECH |
The harm category is hate speech. |
HARM_CATEGORY_DANGEROUS_CONTENT |
The harm category is dangerous content. |
HARM_CATEGORY_HARASSMENT |
The harm category is harassment. |
HARM_CATEGORY_SEXUALLY_EXPLICIT |
The harm category is sexually explicit content. |
HARM_CATEGORY_CIVIC_INTEGRITY |
The harm category is civic integrity. |
HyperparameterTuningJob
Represents a HyperparameterTuningJob. A HyperparameterTuningJob has a Study specification and multiple CustomJobs with identical CustomJob specification.
Fields | |
---|---|
name |
Output only. Resource name of the HyperparameterTuningJob. |
display_ |
Required. The display name of the HyperparameterTuningJob. The name can be up to 128 characters long and can consist of any UTF-8 characters. |
study_ |
Required. Study configuration of the HyperparameterTuningJob. |
max_ |
Required. The desired total number of Trials. |
parallel_ |
Required. The desired number of Trials to run in parallel. |
max_ |
The number of failed Trials that need to be seen before failing the HyperparameterTuningJob. If set to 0, Vertex AI decides how many Trials must fail before the whole job fails. |
trial_ |
Required. The spec of a trial job. The same spec applies to the CustomJobs created in all the trials. |
trials[] |
Output only. Trials of the HyperparameterTuningJob. |
state |
Output only. The detailed state of the job. |
create_ |
Output only. Time when the HyperparameterTuningJob was created. |
start_ |
Output only. Time when the HyperparameterTuningJob for the first time entered the |
end_ |
Output only. Time when the HyperparameterTuningJob entered any of the following states: |
update_ |
Output only. Time when the HyperparameterTuningJob was most recently updated. |
error |
Output only. Only populated when job's state is JOB_STATE_FAILED or JOB_STATE_CANCELLED. |
labels |
The labels with user-defined metadata to organize HyperparameterTuningJobs. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. |
encryption_ |
Customer-managed encryption key options for a HyperparameterTuningJob. If this is set, then all resources created by the HyperparameterTuningJob will be encrypted with the provided encryption key. |
satisfies_ |
Output only. Reserved for future use. |
satisfies_ |
Output only. Reserved for future use. |
IdMatcher
Matcher for Features of an EntityType by Feature ID.
Fields | |
---|---|
ids[] |
Required. The following are accepted as
|
ImportDataConfig
Describes the location from where we import data into a Dataset, together with the labels that will be applied to the DataItems and the Annotations.
Fields | |
---|---|
data_ |
Labels that will be applied to newly imported DataItems. If an identical DataItem as one being imported already exists in the Dataset, then these labels will be appended to these of the already existing one, and if labels with identical key is imported before, the old label value will be overwritten. If two DataItems are identical in the same import data operation, the labels will be combined and if key collision happens in this case, one of the values will be picked randomly. Two DataItems are considered identical if their content bytes are identical (e.g. image bytes or pdf bytes). These labels will be overridden by Annotation labels specified inside index file referenced by |
annotation_ |
Labels that will be applied to newly imported Annotations. If two Annotations are identical, one of them will be deduped. Two Annotations are considered identical if their |
import_ |
Required. Points to a YAML file stored on Google Cloud Storage describing the import format. Validation will be done against the schema. The schema is defined as an OpenAPI 3.0.2 Schema Object. |
Union field source . The source of the input. source can be only one of the following: |
|
gcs_ |
The Google Cloud Storage location for the input content. |
ImportDataOperationMetadata
Runtime operation information for DatasetService.ImportData
.
Fields | |
---|---|
generic_ |
The common part of the operation metadata. |
ImportDataRequest
Request message for DatasetService.ImportData
.
Fields | |
---|---|
name |
Required. The name of the Dataset resource. Format: |
import_ |
Required. The desired input locations. The contents of all input locations will be imported in one batch. |
ImportDataResponse
This type has no fields.
Response message for DatasetService.ImportData
.
ImportFeatureValuesOperationMetadata
Details of operations that perform import Feature values.
Fields | |
---|---|
generic_ |
Operation metadata for Featurestore import Feature values. |
imported_ |
Number of entities that have been imported by the operation. |
imported_ |
Number of Feature values that have been imported by the operation. |
source_ |
The source URI from where Feature values are imported. |
invalid_ |
The number of rows in input source that weren't imported due to either * Not having any featureValues. * Having a null entityId. * Having a null timestamp. * Not being parsable (applicable for CSV sources). |
timestamp_ |
The number rows that weren't ingested due to having timestamps outside the retention boundary. |
blocking_ |
List of ImportFeatureValues operations running under a single EntityType that are blocking this operation. |
ImportFeatureValuesRequest
Request message for FeaturestoreService.ImportFeatureValues
.
Fields | |
---|---|
entity_ |
Required. The resource name of the EntityType grouping the Features for which values are being imported. Format: |
entity_ |
Source column that holds entity IDs. If not provided, entity IDs are extracted from the column named entity_id. |
feature_ |
Required. Specifications defining which Feature values to import from the entity. The request fails if no feature_specs are provided, and having multiple feature_specs for one Feature is not allowed. |
disable_ |
If set, data will not be imported for online serving. This is typically used for backfilling, where Feature generation timestamps are not in the timestamp range needed for online serving. |
worker_ |
Specifies the number of workers that are used to write data to the Featurestore. Consider the online serving capacity that you require to achieve the desired import throughput without interfering with online serving. The value must be positive, and less than or equal to 100. If not set, defaults to using 1 worker. The low count ensures minimal impact on online serving performance. |
disable_ |
If true, API doesn't start ingestion analysis pipeline. |
Union field source . Details about the source data, including the location of the storage and the format. source can be only one of the following: |
|
avro_ |
|
bigquery_ |
|
csv_ |
|
Union field feature_time_source . Source of Feature timestamp for all Feature values of each entity. Timestamps must be millisecond-aligned. feature_time_source can be only one of the following: |
|
feature_ |
Source column that holds the Feature timestamp for all Feature values in each entity. |
feature_ |
Single Feature timestamp for all entities being imported. The timestamp must not have higher than millisecond precision. |
FeatureSpec
Defines the Feature value(s) to import.
Fields | |
---|---|
id |
Required. ID of the Feature to import values of. This Feature must exist in the target EntityType, or the request will fail. |
source_ |
Source column to get the Feature values from. If not set, uses the column with the same name as the Feature ID. |
ImportFeatureValuesResponse
Response message for FeaturestoreService.ImportFeatureValues
.
Fields | |
---|---|
imported_ |
Number of entities that have been imported by the operation. |
imported_ |
Number of Feature values that have been imported by the operation. |
invalid_ |
The number of rows in input source that weren't imported due to either * Not having any featureValues. * Having a null entityId. * Having a null timestamp. * Not being parsable (applicable for CSV sources). |
timestamp_ |
The number rows that weren't ingested due to having feature timestamps outside the retention boundary. |
ImportModelEvaluationRequest
Request message for ModelService.ImportModelEvaluation
Fields | |
---|---|
parent |
Required. The name of the parent model resource. Format: |
model_ |
Required. Model evaluation resource to be imported. |
ImportRagFilesConfig
Config for importing RagFiles.
Fields | |
---|---|
rag_ |
Specifies the transformation config for RagFiles. |
max_ |
Optional. The max number of queries per minute that this job is allowed to make to the embedding model specified on the corpus. This value is specific to this job and not shared across other import jobs. Consult the Quotas page on the project to set an appropriate value here. If unspecified, a default value of 1,000 QPM would be used. |
Union field import_source . The source of the import. import_source can be only one of the following: |
|
gcs_ |
Google Cloud Storage location. Supports importing individual files as well as entire Google Cloud Storage directories. Sample formats: - |
google_ |
Google Drive location. Supports importing individual files as well as Google Drive folders. |
slack_ |
Slack channels with their corresponding access tokens. |
jira_ |
Jira queries with their corresponding authentication. |
share_ |
SharePoint sources. |
Union field partial_failure_sink . Optional. If provided, all partial failures are written to the sink. Deprecated. Prefer to use the import_result_sink . partial_failure_sink can be only one of the following: |
|
partial_failure_gcs_sink |
The Cloud Storage path to write partial failures to. Deprecated. Prefer to use |
partial_failure_bigquery_sink |
The BigQuery destination to write partial failures to. It should be a bigquery table resource name (e.g. "bq://projectId.bqDatasetId.bqTableId"). The dataset must exist. If the table does not exist, it will be created with the expected schema. If the table exists, the schema will be validated and data will be added to this existing table. Deprecated. Prefer to use |
ImportRagFilesOperationMetadata
Runtime operation information for VertexRagDataService.ImportRagFiles
.
Fields | |
---|---|
generic_ |
The operation generic information. |
rag_ |
The resource ID of RagCorpus that this operation is executed on. |
import_ |
Output only. The config that was passed in the ImportRagFilesRequest. |
progress_ |
The progress percentage of the operation. Value is in the range [0, 100]. This percentage is calculated as follows: progress_percentage = 100 * (successes + failures + skips) / total |
ImportRagFilesRequest
Request message for VertexRagDataService.ImportRagFiles
.
Fields | |
---|---|
parent |
Required. The name of the RagCorpus resource into which to import files. Format: |
import_ |
Required. The config for the RagFiles to be synced and imported into the RagCorpus. |
ImportRagFilesResponse
Response message for VertexRagDataService.ImportRagFiles
.
Fields | |
---|---|
imported_ |
The number of RagFiles that had been imported into the RagCorpus. |
failed_ |
The number of RagFiles that had failed while importing into the RagCorpus. |
skipped_ |
The number of RagFiles that was skipped while importing into the RagCorpus. |
Union field partial_failure_sink . The location into which the partial failures were written. partial_failure_sink can be only one of the following: |
|
partial_ |
The Google Cloud Storage path into which the partial failures were written. |
partial_ |
The BigQuery table into which the partial failures were written. |
Index
A representation of a collection of database items organized in a way that allows for approximate nearest neighbor (a.k.a ANN) algorithms search.
Fields | |
---|---|
name |
Output only. The resource name of the Index. |
display_ |
Required. The display name of the Index. The name can be up to 128 characters long and can consist of any UTF-8 characters. |
description |
The description of the Index. |
metadata_ |
Immutable. Points to a YAML file stored on Google Cloud Storage describing additional information about the Index, that is specific to it. Unset if the Index does not have any additional information. The schema is defined as an OpenAPI 3.0.2 Schema Object. Note: The URI given on output will be immutable and probably different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access. |
metadata |
An additional information about the Index; the schema of the metadata can be found in |
deployed_ |
Output only. The pointers to DeployedIndexes created from this Index. An Index can be only deleted if all its DeployedIndexes had been undeployed first. |
etag |
Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
labels |
The labels with user-defined metadata to organize your Indexes. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. |
create_ |
Output only. Timestamp when this Index was created. |
update_ |
Output only. Timestamp when this Index was most recently updated. This also includes any update to the contents of the Index. Note that Operations working on this Index may have their |
index_ |
Output only. Stats of the index resource. |
index_ |
Immutable. The update method to use with this Index. If not set, BATCH_UPDATE will be used by default. |
encryption_ |
Immutable. Customer-managed encryption key spec for an Index. If set, this Index and all sub-resources of this Index will be secured by this key. |
satisfies_ |
Output only. Reserved for future use. |
satisfies_ |
Output only. Reserved for future use. |
IndexUpdateMethod
The update method of an Index.
Enums | |
---|---|
INDEX_UPDATE_METHOD_UNSPECIFIED |
Should not be used. |
BATCH_UPDATE |
BatchUpdate: user can call UpdateIndex with files on Cloud Storage of Datapoints to update. |
STREAM_UPDATE |
StreamUpdate: user can call UpsertDatapoints/DeleteDatapoints to update the Index and the updates will be applied in corresponding DeployedIndexes in nearly real-time. |
IndexDatapoint
A datapoint of Index.
Fields | |
---|---|
datapoint_ |
Required. Unique identifier of the datapoint. |
feature_ |
Required. Feature embedding vector for dense index. An array of numbers with the length of [NearestNeighborSearchConfig.dimensions]. |
sparse_ |
Optional. Feature embedding vector for sparse index. |
restricts[] |
Optional. List of Restrict of the datapoint, used to perform "restricted searches" where boolean rule are used to filter the subset of the database eligible for matching. This uses categorical tokens. See: https://cloud.google.com/vertex-ai/docs/matching-engine/filtering |
numeric_ |
Optional. List of Restrict of the datapoint, used to perform "restricted searches" where boolean rule are used to filter the subset of the database eligible for matching. This uses numeric comparisons. |
crowding_ |
Optional. CrowdingTag of the datapoint, the number of neighbors to return in each crowding can be configured during query. |
CrowdingTag
Crowding tag is a constraint on a neighbor list produced by nearest neighbor search requiring that no more than some value k' of the k neighbors returned have the same value of crowding_attribute.
Fields | |
---|---|
crowding_ |
The attribute value used for crowding. The maximum number of neighbors to return per crowding attribute value (per_crowding_attribute_num_neighbors) is configured per-query. This field is ignored if per_crowding_attribute_num_neighbors is larger than the total number of neighbors to return for a given query. |
NumericRestriction
This field allows restricts to be based on numeric comparisons rather than categorical tokens.
Fields | |
---|---|
namespace |
The namespace of this restriction. e.g.: cost. |
op |
This MUST be specified for queries and must NOT be specified for datapoints. |
Union field Value . The type of Value must be consistent for all datapoints with a given namespace name. This is verified at runtime. Value can be only one of the following: |
|
value_ |
Represents 64 bit integer. |
value_ |
Represents 32 bit float. |
value_ |
Represents 64 bit float. |
Operator
Which comparison operator to use. Should be specified for queries only; specifying this for a datapoint is an error.
Datapoints for which Operator is true relative to the query's Value field will be allowlisted.
Enums | |
---|---|
OPERATOR_UNSPECIFIED |
Default value of the enum. |
LESS |
Datapoints are eligible iff their value is < the query's. |
LESS_EQUAL |
Datapoints are eligible iff their value is <= the query's. |
EQUAL |
Datapoints are eligible iff their value is == the query's. |
GREATER_EQUAL |
Datapoints are eligible iff their value is >= the query's. |
GREATER |
Datapoints are eligible iff their value is > the query's. |
NOT_EQUAL |
Datapoints are eligible iff their value is != the query's. |
Restriction
Restriction of a datapoint which describe its attributes(tokens) from each of several attribute categories(namespaces).
Fields | |
---|---|
namespace |
The namespace of this restriction. e.g.: color. |
allow_ |
The attributes to allow in this namespace. e.g.: 'red' |
deny_ |
The attributes to deny in this namespace. e.g.: 'blue' |
SparseEmbedding
Feature embedding vector for sparse index. An array of numbers whose values are located in the specified dimensions.
Fields | |
---|---|
values[] |
Required. The list of embedding values of the sparse vector. |
dimensions[] |
Required. The list of indexes for the embedding values of the sparse vector. |
IndexEndpoint
Indexes are deployed into it. An IndexEndpoint can have multiple DeployedIndexes.
Fields | |
---|---|
name |
Output only. The resource name of the IndexEndpoint. |
display_ |
Required. The display name of the IndexEndpoint. The name can be up to 128 characters long and can consist of any UTF-8 characters. |
description |
The description of the IndexEndpoint. |
deployed_ |
Output only. The indexes deployed in this endpoint. |
etag |
Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
labels |
The labels with user-defined metadata to organize your IndexEndpoints. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. |
create_ |
Output only. Timestamp when this IndexEndpoint was created. |
update_ |
Output only. Timestamp when this IndexEndpoint was last updated. This timestamp is not updated when the endpoint's DeployedIndexes are updated, e.g. due to updates of the original Indexes they are the deployments of. |
network |
Optional. The full name of the Google Compute Engine network to which the IndexEndpoint should be peered. Private services access must already be configured for the network. If left unspecified, the Endpoint is not peered with any network.
Format: |
enable_private_service_connect |
Optional. Deprecated: If true, expose the IndexEndpoint via private service connect. Only one of the fields, |
private_ |
Optional. Configuration for private service connect.
|
public_ |
Optional. If true, the deployed index will be accessible through public endpoint. |
public_ |
Output only. If |
encryption_ |
Immutable. Customer-managed encryption key spec for an IndexEndpoint. If set, this IndexEndpoint and all sub-resources of this IndexEndpoint will be secured by this key. |
satisfies_ |
Output only. Reserved for future use. |
satisfies_ |
Output only. Reserved for future use. |
IndexPrivateEndpoints
IndexPrivateEndpoints proto is used to provide paths for users to send requests via private endpoints (e.g. private service access, private service connect). To send request via private service access, use match_grpc_address. To send request via private service connect, use service_attachment.
Fields | |
---|---|
match_ |
Output only. The ip address used to send match gRPC requests. |
service_ |
Output only. The name of the service attachment resource. Populated if private service connect is enabled. |
psc_ |
Output only. PscAutomatedEndpoints is populated if private service connect is enabled if PscAutomatedConfig is set. |
IndexStats
Stats of the Index.
Fields | |
---|---|
vectors_ |
Output only. The number of dense vectors in the Index. |
sparse_ |
Output only. The number of sparse vectors in the Index. |
shards_ |
Output only. The number of shards in the Index. |
InputDataConfig
Specifies Vertex AI owned input data to be used for training, and possibly evaluating, the Model.
Fields | |
---|---|
dataset_ |
Required. The ID of the Dataset in the same Project and Location which data will be used to train the Model. The Dataset must use schema compatible with Model being trained, and what is compatible should be described in the used TrainingPipeline's |
annotations_ |
Applicable only to Datasets that have DataItems and Annotations. A filter on Annotations of the Dataset. Only Annotations that both match this filter and belong to DataItems not ignored by the split method are used in respectively training, validation or test role, depending on the role of the DataItem they are on (for the auto-assigned that role is decided by Vertex AI). A filter with same syntax as the one used in |
annotation_ |
Applicable only to custom training with Datasets that have DataItems and Annotations. Cloud Storage URI that points to a YAML file describing the annotation schema. The schema is defined as an OpenAPI 3.0.2 Schema Object. The schema files that can be used here are found in gs://google-cloud-aiplatform/schema/dataset/annotation/ , note that the chosen schema must be consistent with Only Annotations that both match this schema and belong to DataItems not ignored by the split method are used in respectively training, validation or test role, depending on the role of the DataItem they are on. When used in conjunction with |
saved_ |
Only applicable to Datasets that have SavedQueries. The ID of a SavedQuery (annotation set) under the Dataset specified by Only Annotations that are associated with this SavedQuery are used in respectively training. When used in conjunction with Only one of |
persist_ |
Whether to persist the ML use assignment to data item system labels. |
Union field split . The instructions how the input data should be split between the training, validation and test sets. If no split type is provided, the fraction_split is used by default. split can be only one of the following: |
|
fraction_ |
Split based on fractions defining the size of each set. |
filter_ |
Split based on the provided filters for each set. |
predefined_ |
Supported only for tabular Datasets. Split based on a predefined key. |
timestamp_ |
Supported only for tabular Datasets. Split based on the timestamp of the input data pieces. |
stratified_ |
Supported only for tabular Datasets. Split based on the distribution of the specified column. |
Union field The destination of the training data to be written to. Supported destination file formats: * For non-tabular data: "jsonl". * For tabular data: "csv" and "bigquery". The following Vertex AI environment variables are passed to containers or python modules of the training task when this field is set:
|
|
gcs_ |
The Cloud Storage location where the training data is to be written to. In the given directory a new directory is created with name: The Vertex AI environment variables representing Cloud Storage data URIs are represented in the Cloud Storage wildcard format to support sharded data. e.g.: "gs://.../training-*.jsonl"
|
bigquery_ |
Only applicable to custom training with tabular Dataset with BigQuery source. The BigQuery project location where the training data is to be written to. In the given project a new dataset is created with name
|
Int64Array
A list of int64 values.
Fields | |
---|---|
values[] |
A list of int64 values. |
IntegratedGradientsAttribution
An attribution method that computes the Aumann-Shapley value taking advantage of the model's fully differentiable structure. Refer to this paper for more details: https://arxiv.org/abs/1703.01365
Fields | |
---|---|
step_ |
Required. The number of steps for approximating the path integral. A good value to start is 50 and gradually increase until the sum to diff property is within the desired error range. Valid range of its value is [1, 100], inclusively. |
smooth_ |
Config for SmoothGrad approximation of gradients. When enabled, the gradients are approximated by averaging the gradients from noisy samples in the vicinity of the inputs. Adding noise can help improve the computed gradients. Refer to this paper for more details: https://arxiv.org/pdf/1706.03825.pdf |
blur_ |
Config for IG with blur baseline. When enabled, a linear path from the maximally blurred image to the input image is created. Using a blurred baseline instead of zero (black image) is motivated by the BlurIG approach explained here: https://arxiv.org/abs/2004.03383 |
JiraSource
The Jira source for the ImportRagFilesRequest.
Fields | |
---|---|
jira_ |
Required. The Jira queries. |
JiraQueries
JiraQueries contains the Jira queries and corresponding authentication.
Fields | |
---|---|
projects[] |
A list of Jira projects to import in their entirety. |
custom_ |
A list of custom Jira queries to import. For information about JQL (Jira Query Language), see https://support.atlassian.com/jira-service-management-cloud/docs/use-advanced-search-with-jira-query-language-jql/ |
email |
Required. The Jira email address. |
server_ |
Required. The Jira server URI. |
api_ |
Required. The SecretManager secret version resource name (e.g. projects/{project}/secrets/{secret}/versions/{version}) storing the Jira API key. See Manage API tokens for your Atlassian account. |
JobState
Describes the state of a job.
Enums | |
---|---|
JOB_STATE_UNSPECIFIED |
The job state is unspecified. |
JOB_STATE_QUEUED |
The job has been just created or resumed and processing has not yet begun. |
JOB_STATE_PENDING |
The service is preparing to run the job. |
JOB_STATE_RUNNING |
The job is in progress. |
JOB_STATE_SUCCEEDED |
The job completed successfully. |
JOB_STATE_FAILED |
The job failed. |
JOB_STATE_CANCELLING |
The job is being cancelled. From this state the job may only go to either JOB_STATE_SUCCEEDED , JOB_STATE_FAILED or JOB_STATE_CANCELLED . |
JOB_STATE_CANCELLED |
The job has been cancelled. |
JOB_STATE_PAUSED |
The job has been stopped, and can be resumed. |
JOB_STATE_EXPIRED |
The job has expired. |
JOB_STATE_UPDATING |
The job is being updated. Only jobs in the RUNNING state can be updated. After updating, the job goes back to the RUNNING state. |
JOB_STATE_PARTIALLY_SUCCEEDED |
The job is partially succeeded, some results may be missing due to errors. |
LargeModelReference
Contains information about the Large Model.
Fields | |
---|---|
name |
Required. The unique name of the large Foundation or pre-built model. Like "chat-bison", "text-bison". Or model name with version ID, like "chat-bison@001", "text-bison@005", etc. |
LineageSubgraph
A subgraph of the overall lineage graph. Event edges connect Artifact and Execution nodes.
Fields | |
---|---|
artifacts[] |
The Artifact nodes in the subgraph. |
executions[] |
The Execution nodes in the subgraph. |
events[] |
The Event edges between Artifacts and Executions in the subgraph. |
ListAnnotationsRequest
Request message for DatasetService.ListAnnotations
.
Fields | |
---|---|
parent |
Required. The resource name of the DataItem to list Annotations from. Format: |
filter |
The standard list filter. |
page_ |
The standard list page size. |
page_ |
The standard list page token. |
read_ |
Mask specifying which fields to read. |
order_ |
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. |
ListAnnotationsResponse
Response message for DatasetService.ListAnnotations
.
Fields | |
---|---|
annotations[] |
A list of Annotations that matches the specified filter in the request. |
next_ |
The standard List next-page token. |
ListArtifactsRequest
Request message for MetadataService.ListArtifacts
.
Fields | |
---|---|
parent |
Required. The MetadataStore whose Artifacts should be listed. Format: |
page_ |
The maximum number of Artifacts to return. The service may return fewer. Must be in range 1-1000, inclusive. Defaults to 100. |
page_ |
A page token, received from a previous When paginating, all other provided parameters must match the call that provided the page token. (Otherwise the request will fail with INVALID_ARGUMENT error.) |
filter |
Filter specifying the boolean condition for the Artifacts to satisfy in order to be part of the result set. The syntax to define filter query is based on https://google.aip.dev/160. The supported set of filters include the following:
Each of the above supported filter types can be combined together using logical operators ( For example: |
order_ |
How the list of messages is ordered. Specify the values to order by and an ordering operation. The default sorting order is ascending. To specify descending order for a field, users append a " desc" suffix; for example: "foo desc, bar". Subfields are specified with a |
ListArtifactsResponse
Response message for MetadataService.ListArtifacts
.
Fields | |
---|---|
artifacts[] |
The Artifacts retrieved from the MetadataStore. |
next_ |
A token, which can be sent as |
ListBatchPredictionJobsRequest
Request message for JobService.ListBatchPredictionJobs
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to list the BatchPredictionJobs from. Format: |
filter |
The standard list filter. Supported fields:
Some examples of using the filter are:
|
page_ |
The standard list page size. |
page_ |
The standard list page token. Typically obtained via |
read_ |
Mask specifying which fields to read. |
ListBatchPredictionJobsResponse
Response message for JobService.ListBatchPredictionJobs
Fields | |
---|---|
batch_ |
List of BatchPredictionJobs in the requested page. |
next_ |
A token to retrieve the next page of results. Pass to |
ListContextsRequest
Request message for MetadataService.ListContexts
Fields | |
---|---|
parent |
Required. The MetadataStore whose Contexts should be listed. Format: |
page_ |
The maximum number of Contexts to return. The service may return fewer. Must be in range 1-1000, inclusive. Defaults to 100. |
page_ |
A page token, received from a previous When paginating, all other provided parameters must match the call that provided the page token. (Otherwise the request will fail with INVALID_ARGUMENT error.) |
filter |
Filter specifying the boolean condition for the Contexts to satisfy in order to be part of the result set. The syntax to define filter query is based on https://google.aip.dev/160. Following are the supported set of filters:
Each of the above supported filters can be combined together using logical operators ( For example: |
order_ |
How the list of messages is ordered. Specify the values to order by and an ordering operation. The default sorting order is ascending. To specify descending order for a field, users append a " desc" suffix; for example: "foo desc, bar". Subfields are specified with a |
ListContextsResponse
Response message for MetadataService.ListContexts
.
Fields | |
---|---|
contexts[] |
The Contexts retrieved from the MetadataStore. |
next_ |
A token, which can be sent as |
ListCustomJobsRequest
Request message for JobService.ListCustomJobs
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to list the CustomJobs from. Format: |
filter |
The standard list filter. Supported fields:
Some examples of using the filter are:
|
page_ |
The standard list page size. |
page_ |
The standard list page token. Typically obtained via |
read_ |
Mask specifying which fields to read. |
ListCustomJobsResponse
Response message for JobService.ListCustomJobs
Fields | |
---|---|
custom_ |
List of CustomJobs in the requested page. |
next_ |
A token to retrieve the next page of results. Pass to |
ListDataItemsRequest
Request message for DatasetService.ListDataItems
.
Fields | |
---|---|
parent |
Required. The resource name of the Dataset to list DataItems from. Format: |
filter |
The standard list filter. |
page_ |
The standard list page size. |
page_ |
The standard list page token. |
read_ |
Mask specifying which fields to read. |
order_ |
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. |
ListDataItemsResponse
Response message for DatasetService.ListDataItems
.
Fields | |
---|---|
data_ |
A list of DataItems that matches the specified filter in the request. |
next_ |
The standard List next-page token. |
ListDatasetVersionsRequest
Request message for DatasetService.ListDatasetVersions
.
Fields | |
---|---|
parent |
Required. The resource name of the Dataset to list DatasetVersions from. Format: |
filter |
Optional. The standard list filter. |
page_ |
Optional. The standard list page size. |
page_ |
Optional. The standard list page token. |
read_ |
Optional. Mask specifying which fields to read. |
order_ |
Optional. A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. |
ListDatasetVersionsResponse
Response message for DatasetService.ListDatasetVersions
.
Fields | |
---|---|
dataset_ |
A list of DatasetVersions that matches the specified filter in the request. |
next_ |
The standard List next-page token. |
ListDatasetsRequest
Request message for DatasetService.ListDatasets
.
Fields | |
---|---|
parent |
Required. The name of the Dataset's parent resource. Format: |
filter |
An expression for filtering the results of the request. For field names both snake_case and camelCase are supported.
Some examples:
|
page_ |
The standard list page size. |
page_ |
The standard list page token. |
read_ |
Mask specifying which fields to read. |
order_ |
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields:
|
ListDatasetsResponse
Response message for DatasetService.ListDatasets
.
Fields | |
---|---|
datasets[] |
A list of Datasets that matches the specified filter in the request. |
next_ |
The standard List next-page token. |
ListDeploymentResourcePoolsRequest
Request message for ListDeploymentResourcePools method.
Fields | |
---|---|
parent |
Required. The parent Location which owns this collection of DeploymentResourcePools. Format: |
page_ |
The maximum number of DeploymentResourcePools to return. The service may return fewer than this value. |
page_ |
A page token, received from a previous When paginating, all other parameters provided to |
ListDeploymentResourcePoolsResponse
Response message for ListDeploymentResourcePools method.
Fields | |
---|---|
deployment_ |
The DeploymentResourcePools from the specified location. |
next_ |
A token, which can be sent as |
ListEndpointsRequest
Request message for EndpointService.ListEndpoints
.
Fields | |
---|---|
parent |
Required. The resource name of the Location from which to list the Endpoints. Format: |
filter |
Optional. An expression for filtering the results of the request. For field names both snake_case and camelCase are supported.
Some examples:
|
page_ |
Optional. The standard list page size. |
page_ |
Optional. The standard list page token. Typically obtained via |
read_ |
Optional. Mask specifying which fields to read. |
order_ |
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields:
Example: |
ListEndpointsResponse
Response message for EndpointService.ListEndpoints
.
Fields | |
---|---|
endpoints[] |
List of Endpoints in the requested page. |
next_ |
A token to retrieve the next page of results. Pass to |
ListEntityTypesRequest
Request message for FeaturestoreService.ListEntityTypes
.
Fields | |
---|---|
parent |
Required. The resource name of the Featurestore to list EntityTypes. Format: |
filter |
Lists the EntityTypes that match the filter expression. The following filters are supported:
Examples:
|
page_ |
The maximum number of EntityTypes to return. The service may return fewer than this value. If unspecified, at most 1000 EntityTypes will be returned. The maximum value is 1000; any value greater than 1000 will be coerced to 1000. |
page_ |
A page token, received from a previous When paginating, all other parameters provided to |
order_ |
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields:
|
read_ |
Mask specifying which fields to read. |
ListEntityTypesResponse
Response message for FeaturestoreService.ListEntityTypes
.
Fields | |
---|---|
entity_ |
The EntityTypes matching the request. |
next_ |
A token, which can be sent as |
ListExecutionsRequest
Request message for MetadataService.ListExecutions
.
Fields | |
---|---|
parent |
Required. The MetadataStore whose Executions should be listed. Format: |
page_ |
The maximum number of Executions to return. The service may return fewer. Must be in range 1-1000, inclusive. Defaults to 100. |
page_ |
A page token, received from a previous When paginating, all other provided parameters must match the call that provided the page token. (Otherwise the request will fail with an INVALID_ARGUMENT error.) |
filter |
Filter specifying the boolean condition for the Executions to satisfy in order to be part of the result set. The syntax to define filter query is based on https://google.aip.dev/160. Following are the supported set of filters:
Each of the above supported filters can be combined together using logical operators ( For example: |
order_ |
How the list of messages is ordered. Specify the values to order by and an ordering operation. The default sorting order is ascending. To specify descending order for a field, users append a " desc" suffix; for example: "foo desc, bar". Subfields are specified with a |
ListExecutionsResponse
Response message for MetadataService.ListExecutions
.
Fields | |
---|---|
executions[] |
The Executions retrieved from the MetadataStore. |
next_ |
A token, which can be sent as |
ListFeatureGroupsRequest
Request message for FeatureRegistryService.ListFeatureGroups
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to list FeatureGroups. Format: |
filter |
Lists the FeatureGroups that match the filter expression. The following fields are supported:
Examples:
|
page_ |
The maximum number of FeatureGroups to return. The service may return fewer than this value. If unspecified, at most 100 FeatureGroups will be returned. The maximum value is 100; any value greater than 100 will be coerced to 100. |
page_ |
A page token, received from a previous When paginating, all other parameters provided to |
order_ |
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported Fields:
|
ListFeatureGroupsResponse
Response message for FeatureRegistryService.ListFeatureGroups
.
Fields | |
---|---|
feature_ |
The FeatureGroups matching the request. |
next_ |
A token, which can be sent as |
ListFeatureOnlineStoresRequest
Request message for FeatureOnlineStoreAdminService.ListFeatureOnlineStores
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to list FeatureOnlineStores. Format: |
filter |
Lists the FeatureOnlineStores that match the filter expression. The following fields are supported:
Examples:
|
page_ |
The maximum number of FeatureOnlineStores to return. The service may return fewer than this value. If unspecified, at most 100 FeatureOnlineStores will be returned. The maximum value is 100; any value greater than 100 will be coerced to 100. |
page_ |
A page token, received from a previous When paginating, all other parameters provided to |
order_ |
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported Fields:
|
ListFeatureOnlineStoresResponse
Response message for FeatureOnlineStoreAdminService.ListFeatureOnlineStores
.
Fields | |
---|---|
feature_ |
The FeatureOnlineStores matching the request. |
next_ |
A token, which can be sent as |
ListFeatureViewSyncsRequest
Request message for FeatureOnlineStoreAdminService.ListFeatureViewSyncs
.
Fields | |
---|---|
parent |
Required. The resource name of the FeatureView to list FeatureViewSyncs. Format: |
filter |
Lists the FeatureViewSyncs that match the filter expression. The following filters are supported:
Examples:
|
page_ |
The maximum number of FeatureViewSyncs to return. The service may return fewer than this value. If unspecified, at most 1000 FeatureViewSyncs will be returned. The maximum value is 1000; any value greater than 1000 will be coerced to 1000. |
page_ |
A page token, received from a previous When paginating, all other parameters provided to |
order_ |
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields:
|
ListFeatureViewSyncsResponse
Response message for FeatureOnlineStoreAdminService.ListFeatureViewSyncs
.
Fields | |
---|---|
feature_ |
The FeatureViewSyncs matching the request. |
next_ |
A token, which can be sent as |
ListFeatureViewsRequest
Request message for FeatureOnlineStoreAdminService.ListFeatureViews
.
Fields | |
---|---|
parent |
Required. The resource name of the FeatureOnlineStore to list FeatureViews. Format: |
filter |
Lists the FeatureViews that match the filter expression. The following filters are supported:
Examples:
|
page_ |
The maximum number of FeatureViews to return. The service may return fewer than this value. If unspecified, at most 1000 FeatureViews will be returned. The maximum value is 1000; any value greater than 1000 will be coerced to 1000. |
page_ |
A page token, received from a previous When paginating, all other parameters provided to |
order_ |
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields:
|
ListFeatureViewsResponse
Response message for FeatureOnlineStoreAdminService.ListFeatureViews
.
Fields | |
---|---|
feature_ |
The FeatureViews matching the request. |
next_ |
A token, which can be sent as |
ListFeaturesRequest
Request message for FeaturestoreService.ListFeatures
. Request message for FeatureRegistryService.ListFeatures
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to list Features. Format for entity_type as parent: |
filter |
Lists the Features that match the filter expression. The following filters are supported:
Examples:
|
page_ |
The maximum number of Features to return. The service may return fewer than this value. If unspecified, at most 1000 Features will be returned. The maximum value is 1000; any value greater than 1000 will be coerced to 1000. |
page_ |
A page token, received from a previous When paginating, all other parameters provided to |
order_ |
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields:
|
read_ |
Mask specifying which fields to read. |
latest_ |
Only applicable for Vertex AI Feature Store (Legacy). If set, return the most recent |
ListFeaturesResponse
Response message for FeaturestoreService.ListFeatures
. Response message for FeatureRegistryService.ListFeatures
.
Fields | |
---|---|
features[] |
The Features matching the request. |
next_ |
A token, which can be sent as |
ListFeaturestoresRequest
Request message for FeaturestoreService.ListFeaturestores
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to list Featurestores. Format: |
filter |
Lists the featurestores that match the filter expression. The following fields are supported:
Examples:
|
page_ |
The maximum number of Featurestores to return. The service may return fewer than this value. If unspecified, at most 100 Featurestores will be returned. The maximum value is 100; any value greater than 100 will be coerced to 100. |
page_ |
A page token, received from a previous When paginating, all other parameters provided to |
order_ |
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported Fields:
|
read_ |
Mask specifying which fields to read. |
ListFeaturestoresResponse
Response message for FeaturestoreService.ListFeaturestores
.
Fields | |
---|---|
featurestores[] |
The Featurestores matching the request. |
next_ |
A token, which can be sent as |
ListHyperparameterTuningJobsRequest
Request message for JobService.ListHyperparameterTuningJobs
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to list the HyperparameterTuningJobs from. Format: |
filter |
The standard list filter. Supported fields:
Some examples of using the filter are:
|
page_ |
The standard list page size. |
page_ |
The standard list page token. Typically obtained via |
read_ |
Mask specifying which fields to read. |
ListHyperparameterTuningJobsResponse
Response message for JobService.ListHyperparameterTuningJobs
Fields | |
---|---|
hyperparameter_ |
List of HyperparameterTuningJobs in the requested page. |
next_ |
A token to retrieve the next page of results. Pass to |
ListIndexEndpointsRequest
Request message for IndexEndpointService.ListIndexEndpoints
.
Fields | |
---|---|
parent |
Required. The resource name of the Location from which to list the IndexEndpoints. Format: |
filter |
Optional. An expression for filtering the results of the request. For field names both snake_case and camelCase are supported.
Some examples: * |
page_ |
Optional. The standard list page size. |
page_ |
Optional. The standard list page token. Typically obtained via |
read_ |
Optional. Mask specifying which fields to read. |
ListIndexEndpointsResponse
Response message for IndexEndpointService.ListIndexEndpoints
.
Fields | |
---|---|
index_ |
List of IndexEndpoints in the requested page. |
next_ |
A token to retrieve next page of results. Pass to |
ListIndexesRequest
Request message for IndexService.ListIndexes
.
Fields | |
---|---|
parent |
Required. The resource name of the Location from which to list the Indexes. Format: |
filter |
The standard list filter. |
page_ |
The standard list page size. |
page_ |
The standard list page token. Typically obtained via |
read_ |
Mask specifying which fields to read. |
ListIndexesResponse
Response message for IndexService.ListIndexes
.
Fields | |
---|---|
indexes[] |
List of indexes in the requested page. |
next_ |
A token to retrieve next page of results. Pass to |
ListMetadataSchemasRequest
Request message for MetadataService.ListMetadataSchemas
.
Fields | |
---|---|
parent |
Required. The MetadataStore whose MetadataSchemas should be listed. Format: |
page_ |
The maximum number of MetadataSchemas to return. The service may return fewer. Must be in range 1-1000, inclusive. Defaults to 100. |
page_ |
A page token, received from a previous When paginating, all other provided parameters must match the call that provided the page token. (Otherwise the request will fail with INVALID_ARGUMENT error.) |
filter |
A query to filter available MetadataSchemas for matching results. |
ListMetadataSchemasResponse
Response message for MetadataService.ListMetadataSchemas
.
Fields | |
---|---|
metadata_ |
The MetadataSchemas found for the MetadataStore. |
next_ |
A token, which can be sent as |
ListMetadataStoresRequest
Request message for MetadataService.ListMetadataStores
.
Fields | |
---|---|
parent |
Required. The Location whose MetadataStores should be listed. Format: |
page_ |
The maximum number of Metadata Stores to return. The service may return fewer. Must be in range 1-1000, inclusive. Defaults to 100. |
page_ |
A page token, received from a previous When paginating, all other provided parameters must match the call that provided the page token. (Otherwise the request will fail with INVALID_ARGUMENT error.) |
ListMetadataStoresResponse
Response message for MetadataService.ListMetadataStores
.
Fields | |
---|---|
metadata_ |
The MetadataStores found for the Location. |
next_ |
A token, which can be sent as |
ListModelDeploymentMonitoringJobsRequest
Request message for JobService.ListModelDeploymentMonitoringJobs
.
Fields | |
---|---|
parent |
Required. The parent of the ModelDeploymentMonitoringJob. Format: |
filter |
The standard list filter. Supported fields:
Some examples of using the filter are:
|
page_ |
The standard list page size. |
page_ |
The standard list page token. |
read_ |
Mask specifying which fields to read |
ListModelDeploymentMonitoringJobsResponse
Response message for JobService.ListModelDeploymentMonitoringJobs
.
Fields | |
---|---|
model_ |
A list of ModelDeploymentMonitoringJobs that matches the specified filter in the request. |
next_ |
The standard List next-page token. |
ListModelEvaluationSlicesRequest
Request message for ModelService.ListModelEvaluationSlices
.
Fields | |
---|---|
parent |
Required. The resource name of the ModelEvaluation to list the ModelEvaluationSlices from. Format: |
filter |
The standard list filter.
|
page_ |
The standard list page size. |
page_ |
The standard list page token. Typically obtained via |
read_ |
Mask specifying which fields to read. |
ListModelEvaluationSlicesResponse
Response message for ModelService.ListModelEvaluationSlices
.
Fields | |
---|---|
model_ |
List of ModelEvaluations in the requested page. |
next_ |
A token to retrieve next page of results. Pass to |
ListModelEvaluationsRequest
Request message for ModelService.ListModelEvaluations
.
Fields | |
---|---|
parent |
Required. The resource name of the Model to list the ModelEvaluations from. Format: |
filter |
The standard list filter. |
page_ |
The standard list page size. |
page_ |
The standard list page token. Typically obtained via |
read_ |
Mask specifying which fields to read. |
ListModelEvaluationsResponse
Response message for ModelService.ListModelEvaluations
.
Fields | |
---|---|
model_ |
List of ModelEvaluations in the requested page. |
next_ |
A token to retrieve next page of results. Pass to |
ListModelVersionsRequest
Request message for ModelService.ListModelVersions
.
Fields | |
---|---|
name |
Required. The name of the model to list versions for. |
page_ |
The standard list page size. |
page_ |
The standard list page token. Typically obtained via |
filter |
An expression for filtering the results of the request. For field names both snake_case and camelCase are supported.
Some examples:
|
read_ |
Mask specifying which fields to read. |
order_ |
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields:
Example: |
ListModelVersionsResponse
Response message for ModelService.ListModelVersions
Fields | |
---|---|
models[] |
List of Model versions in the requested page. In the returned Model name field, version ID instead of regvision tag will be included. |
next_ |
A token to retrieve the next page of results. Pass to |
ListModelsRequest
Request message for ModelService.ListModels
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to list the Models from. Format: |
filter |
An expression for filtering the results of the request. For field names both snake_case and camelCase are supported.
Some examples:
|
page_ |
The standard list page size. |
page_ |
The standard list page token. Typically obtained via |
read_ |
Mask specifying which fields to read. |
order_ |
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields:
Example: |
ListModelsResponse
Response message for ModelService.ListModels
Fields | |
---|---|
models[] |
List of Models in the requested page. |
next_ |
A token to retrieve next page of results. Pass to |
ListNasJobsRequest
Request message for JobService.ListNasJobs
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to list the NasJobs from. Format: |
filter |
The standard list filter. Supported fields:
Some examples of using the filter are:
|
page_ |
The standard list page size. |
page_ |
The standard list page token. Typically obtained via |
read_ |
Mask specifying which fields to read. |
ListNasJobsResponse
Response message for JobService.ListNasJobs
Fields | |
---|---|
nas_ |
List of NasJobs in the requested page. |
next_ |
A token to retrieve the next page of results. Pass to |
ListNasTrialDetailsRequest
Request message for JobService.ListNasTrialDetails
.
Fields | |
---|---|
parent |
Required. The name of the NasJob resource. Format: |
page_ |
The standard list page size. |
page_ |
The standard list page token. Typically obtained via |
ListNasTrialDetailsResponse
Response message for JobService.ListNasTrialDetails
Fields | |
---|---|
nas_ |
List of top NasTrials in the requested page. |
next_ |
A token to retrieve the next page of results. Pass to |
ListNotebookExecutionJobsRequest
Request message for [NotebookService.ListNotebookExecutionJobs]
Fields | |
---|---|
parent |
Required. The resource name of the Location from which to list the NotebookExecutionJobs. Format: |
filter |
Optional. An expression for filtering the results of the request. For field names both snake_case and camelCase are supported.
Some examples: * |
page_ |
Optional. The standard list page size. |
page_ |
Optional. The standard list page token. Typically obtained via |
order_ |
Optional. A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields:
Example: |
view |
Optional. The NotebookExecutionJob view. Defaults to BASIC. |
ListNotebookExecutionJobsResponse
Response message for [NotebookService.CreateNotebookExecutionJob]
Fields | |
---|---|
notebook_ |
List of NotebookExecutionJobs in the requested page. |
next_ |
A token to retrieve next page of results. Pass to |
ListNotebookRuntimeTemplatesRequest
Request message for NotebookService.ListNotebookRuntimeTemplates
.
Fields | |
---|---|
parent |
Required. The resource name of the Location from which to list the NotebookRuntimeTemplates. Format: |
filter |
Optional. An expression for filtering the results of the request. For field names both snake_case and camelCase are supported.
Some examples:
|
page_ |
Optional. The standard list page size. |
page_ |
Optional. The standard list page token. Typically obtained via |
read_ |
Optional. Mask specifying which fields to read. |
order_ |
Optional. A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields:
Example: |
ListNotebookRuntimeTemplatesResponse
Response message for NotebookService.ListNotebookRuntimeTemplates
.
Fields | |
---|---|
notebook_ |
List of NotebookRuntimeTemplates in the requested page. |
next_ |
A token to retrieve next page of results. Pass to |
ListNotebookRuntimesRequest
Request message for NotebookService.ListNotebookRuntimes
.
Fields | |
---|---|
parent |
Required. The resource name of the Location from which to list the NotebookRuntimes. Format: |
filter |
Optional. An expression for filtering the results of the request. For field names both snake_case and camelCase are supported.
Some examples:
|
page_ |
Optional. The standard list page size. |
page_ |
Optional. The standard list page token. Typically obtained via |
read_ |
Optional. Mask specifying which fields to read. |
order_ |
Optional. A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields:
Example: |
ListNotebookRuntimesResponse
Response message for NotebookService.ListNotebookRuntimes
.
Fields | |
---|---|
notebook_ |
List of NotebookRuntimes in the requested page. |
next_ |
A token to retrieve next page of results. Pass to |
ListOptimalTrialsRequest
Request message for VizierService.ListOptimalTrials
.
Fields | |
---|---|
parent |
Required. The name of the Study that the optimal Trial belongs to. |
ListOptimalTrialsResponse
Response message for VizierService.ListOptimalTrials
.
Fields | |
---|---|
optimal_ |
The pareto-optimal Trials for multiple objective Study or the optimal trial for single objective Study. The definition of pareto-optimal can be checked in wiki page. https://en.wikipedia.org/wiki/Pareto_efficiency |
ListPersistentResourcesRequest
Request message for PersistentResourceService.ListPersistentResources
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to list the PersistentResources from. Format: |
page_ |
Optional. The standard list page size. |
page_ |
Optional. The standard list page token. Typically obtained via |
ListPersistentResourcesResponse
Response message for PersistentResourceService.ListPersistentResources
Fields | |
---|---|
persistent_ |
|
next_ |
A token to retrieve next page of results. Pass to |
ListPipelineJobsRequest
Request message for PipelineService.ListPipelineJobs
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to list the PipelineJobs from. Format: |
filter |
Lists the PipelineJobs that match the filter expression. The following fields are supported:
Filter expressions can be combined together using logical operators ( The syntax to define filter expression is based on https://google.aip.dev/160. Examples:
|
page_ |
The standard list page size. |
page_ |
The standard list page token. Typically obtained via |
order_ |
A comma-separated list of fields to order by. The default sort order is in ascending order. Use "desc" after a field name for descending. You can have multiple order_by fields provided e.g. "create_time desc, end_time", "end_time, start_time, update_time" For example, using "create_time desc, end_time" will order results by create time in descending order, and if there are multiple jobs having the same create time, order them by the end time in ascending order. if order_by is not specified, it will order by default order is create time in descending order. Supported fields:
|
read_ |
Mask specifying which fields to read. |
ListPipelineJobsResponse
Response message for PipelineService.ListPipelineJobs
Fields | |
---|---|
pipeline_ |
List of PipelineJobs in the requested page. |
next_ |
A token to retrieve the next page of results. Pass to |
ListRagCorporaRequest
Request message for VertexRagDataService.ListRagCorpora
.
Fields | |
---|---|
parent |
Required. The resource name of the Location from which to list the RagCorpora. Format: |
page_ |
Optional. The standard list page size. |
page_ |
Optional. The standard list page token. Typically obtained via |
ListRagCorporaResponse
Response message for VertexRagDataService.ListRagCorpora
.
Fields | |
---|---|
rag_ |
List of RagCorpora in the requested page. |
next_ |
A token to retrieve the next page of results. Pass to |
ListRagFilesRequest
Request message for VertexRagDataService.ListRagFiles
.
Fields | |
---|---|
parent |
Required. The resource name of the RagCorpus from which to list the RagFiles. Format: |
page_ |
Optional. The standard list page size. |
page_ |
Optional. The standard list page token. Typically obtained via |
ListRagFilesResponse
Response message for VertexRagDataService.ListRagFiles
.
Fields | |
---|---|
rag_ |
List of RagFiles in the requested page. |
next_ |
A token to retrieve the next page of results. Pass to |
ListSavedQueriesRequest
Request message for DatasetService.ListSavedQueries
.
Fields | |
---|---|
parent |
Required. The resource name of the Dataset to list SavedQueries from. Format: |
filter |
The standard list filter. |
page_ |
The standard list page size. |
page_ |
The standard list page token. |
read_ |
Mask specifying which fields to read. |
order_ |
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. |
ListSavedQueriesResponse
Response message for DatasetService.ListSavedQueries
.
Fields | |
---|---|
saved_ |
A list of SavedQueries that match the specified filter in the request. |
next_ |
The standard List next-page token. |
ListSchedulesRequest
Request message for ScheduleService.ListSchedules
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to list the Schedules from. Format: |
filter |
Lists the Schedules that match the filter expression. The following fields are supported:
Filter expressions can be combined together using logical operators ( Examples:
|
page_ |
The standard list page size. Default to 100 if not specified. |
page_ |
The standard list page token. Typically obtained via |
order_ |
A comma-separated list of fields to order by. The default sort order is in ascending order. Use "desc" after a field name for descending. You can have multiple order_by fields provided. For example, using "create_time desc, end_time" will order results by create time in descending order, and if there are multiple schedules having the same create time, order them by the end time in ascending order. If order_by is not specified, it will order by default with create_time in descending order. Supported fields:
|
ListSchedulesResponse
Response message for ScheduleService.ListSchedules
Fields | |
---|---|
schedules[] |
List of Schedules in the requested page. |
next_ |
A token to retrieve the next page of results. Pass to |
ListSpecialistPoolsRequest
Request message for SpecialistPoolService.ListSpecialistPools
.
Fields | |
---|---|
parent |
Required. The name of the SpecialistPool's parent resource. Format: |
page_ |
The standard list page size. |
page_ |
The standard list page token. Typically obtained by |
read_ |
Mask specifying which fields to read. FieldMask represents a set of |
ListSpecialistPoolsResponse
Response message for SpecialistPoolService.ListSpecialistPools
.
Fields | |
---|---|
specialist_ |
A list of SpecialistPools that matches the specified filter in the request. |
next_ |
The standard List next-page token. |
ListStudiesRequest
Request message for VizierService.ListStudies
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to list the Study from. Format: |
page_ |
Optional. A page token to request the next page of results. If unspecified, there are no subsequent pages. |
page_ |
Optional. The maximum number of studies to return per "page" of results. If unspecified, service will pick an appropriate default. |
ListStudiesResponse
Response message for VizierService.ListStudies
.
Fields | |
---|---|
studies[] |
The studies associated with the project. |
next_ |
Passes this token as the |
ListTensorboardExperimentsRequest
Request message for TensorboardService.ListTensorboardExperiments
.
Fields | |
---|---|
parent |
Required. The resource name of the Tensorboard to list TensorboardExperiments. Format: |
filter |
Lists the TensorboardExperiments that match the filter expression. |
page_ |
The maximum number of TensorboardExperiments to return. The service may return fewer than this value. If unspecified, at most 50 TensorboardExperiments are returned. The maximum value is 1000; values above 1000 are coerced to 1000. |
page_ |
A page token, received from a previous When paginating, all other parameters provided to |
order_ |
Field to use to sort the list. |
read_ |
Mask specifying which fields to read. |
ListTensorboardExperimentsResponse
Response message for TensorboardService.ListTensorboardExperiments
.
Fields | |
---|---|
tensorboard_ |
The TensorboardExperiments mathching the request. |
next_ |
A token, which can be sent as |
ListTensorboardRunsRequest
Request message for TensorboardService.ListTensorboardRuns
.
Fields | |
---|---|
parent |
Required. The resource name of the TensorboardExperiment to list TensorboardRuns. Format: |
filter |
Lists the TensorboardRuns that match the filter expression. |
page_ |
The maximum number of TensorboardRuns to return. The service may return fewer than this value. If unspecified, at most 50 TensorboardRuns are returned. The maximum value is 1000; values above 1000 are coerced to 1000. |
page_ |
A page token, received from a previous When paginating, all other parameters provided to |
order_ |
Field to use to sort the list. |
read_ |
Mask specifying which fields to read. |
ListTensorboardRunsResponse
Response message for TensorboardService.ListTensorboardRuns
.
Fields | |
---|---|
tensorboard_ |
The TensorboardRuns mathching the request. |
next_ |
A token, which can be sent as |
ListTensorboardTimeSeriesRequest
Request message for TensorboardService.ListTensorboardTimeSeries
.
Fields | |
---|---|
parent |
Required. The resource name of the TensorboardRun to list TensorboardTimeSeries. Format: |
filter |
Lists the TensorboardTimeSeries that match the filter expression. |
page_ |
The maximum number of TensorboardTimeSeries to return. The service may return fewer than this value. If unspecified, at most 50 TensorboardTimeSeries are returned. The maximum value is 1000; values above 1000 are coerced to 1000. |
page_ |
A page token, received from a previous When paginating, all other parameters provided to |
order_ |
Field to use to sort the list. |
read_ |
Mask specifying which fields to read. |
ListTensorboardTimeSeriesResponse
Response message for TensorboardService.ListTensorboardTimeSeries
.
Fields | |
---|---|
tensorboard_ |
The TensorboardTimeSeries mathching the request. |
next_ |
A token, which can be sent as |
ListTensorboardsRequest
Request message for TensorboardService.ListTensorboards
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to list Tensorboards. Format: |
filter |
Lists the Tensorboards that match the filter expression. |
page_ |
The maximum number of Tensorboards to return. The service may return fewer than this value. If unspecified, at most 100 Tensorboards are returned. The maximum value is 100; values above 100 are coerced to 100. |
page_ |
A page token, received from a previous When paginating, all other parameters provided to |
order_ |
Field to use to sort the list. |
read_ |
Mask specifying which fields to read. |
ListTensorboardsResponse
Response message for TensorboardService.ListTensorboards
.
Fields | |
---|---|
tensorboards[] |
The Tensorboards mathching the request. |
next_ |
A token, which can be sent as |
ListTrainingPipelinesRequest
Request message for PipelineService.ListTrainingPipelines
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to list the TrainingPipelines from. Format: |
filter |
The standard list filter. Supported fields:
Some examples of using the filter are:
|
page_ |
The standard list page size. |
page_ |
The standard list page token. Typically obtained via |
read_ |
Mask specifying which fields to read. |
ListTrainingPipelinesResponse
Response message for PipelineService.ListTrainingPipelines
Fields | |
---|---|
training_ |
List of TrainingPipelines in the requested page. |
next_ |
A token to retrieve the next page of results. Pass to |
ListTrialsRequest
Request message for VizierService.ListTrials
.
Fields | |
---|---|
parent |
Required. The resource name of the Study to list the Trial from. Format: |
page_ |
Optional. A page token to request the next page of results. If unspecified, there are no subsequent pages. |
page_ |
Optional. The number of Trials to retrieve per "page" of results. If unspecified, the service will pick an appropriate default. |
ListTrialsResponse
Response message for VizierService.ListTrials
.
Fields | |
---|---|
trials[] |
The Trials associated with the Study. |
next_ |
Pass this token as the |
ListTuningJobsRequest
Request message for GenAiTuningService.ListTuningJobs
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to list the TuningJobs from. Format: |
filter |
Optional. The standard list filter. |
page_ |
Optional. The standard list page size. |
page_ |
Optional. The standard list page token. Typically obtained via |
ListTuningJobsResponse
Response message for GenAiTuningService.ListTuningJobs
Fields | |
---|---|
tuning_ |
List of TuningJobs in the requested page. |
next_ |
A token to retrieve the next page of results. Pass to |
LogprobsResult
Logprobs Result
Fields | |
---|---|
top_ |
Length = total number of decoding steps. |
chosen_ |
Length = total number of decoding steps. The chosen candidates may or may not be in top_candidates. |
Candidate
Candidate for the logprobs token and score.
Fields | |
---|---|
token |
The candidate's token string value. |
token_ |
The candidate's token id value. |
log_ |
The candidate's log probability. |
TopCandidates
Candidates with top log probabilities at each decoding step.
Fields | |
---|---|
candidates[] |
Sorted by log probability in descending order. |
LookupStudyRequest
Request message for VizierService.LookupStudy
.
Fields | |
---|---|
parent |
Required. The resource name of the Location to get the Study from. Format: |
display_ |
Required. The user-defined display name of the Study |
MachineSpec
Specification of a single machine.
Fields | |
---|---|
machine_ |
Immutable. The type of the machine. See the list of machine types supported for prediction See the list of machine types supported for custom training. For |
accelerator_ |
Immutable. The type of accelerator(s) that may be attached to the machine as per |
accelerator_ |
The number of accelerators to attach to the machine. |
tpu_ |
Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). |
reservation_ |
Optional. Immutable. Configuration controlling how this resource pool consumes reservation. |
ManualBatchTuningParameters
Manual batch tuning parameters.
Fields | |
---|---|
batch_ |
Immutable. The number of the records (e.g. instances) of the operation given in each batch to a machine replica. Machine type, and size of a single record should be considered when setting this parameter, higher value speeds up the batch operation's execution, but too high value will result in a whole batch not fitting in a machine's memory, and the whole operation will fail. The default value is 64. |
Measurement
A message representing a Measurement of a Trial. A Measurement contains the Metrics got by executing a Trial using suggested hyperparameter values.
Fields | |
---|---|
elapsed_ |
Output only. Time that the Trial has been running at the point of this Measurement. |
step_ |
Output only. The number of steps the machine learning model has been trained for. Must be non-negative. |
metrics[] |
Output only. A list of metrics got by evaluating the objective functions using suggested Parameter values. |
Metric
A message representing a metric in the measurement.
Fields | |
---|---|
metric_ |
Output only. The ID of the Metric. The Metric should be defined in |
value |
Output only. The value for this metric. |
MergeVersionAliasesRequest
Request message for ModelService.MergeVersionAliases
.
Fields | |
---|---|
name |
Required. The name of the model version to merge aliases, with a version ID explicitly included. Example: |
version_ |
Required. The set of version aliases to merge. The alias should be at most 128 characters, and match There is NO ordering in aliases, which means 1) The aliases returned from GetModel API might not have the exactly same order from this MergeVersionAliases API. 2) Adding and deleting the same alias in the request is not recommended, and the 2 operations will be cancelled out. |
MetadataSchema
Instance of a general MetadataSchema.
Fields | |
---|---|
name |
Output only. The resource name of the MetadataSchema. |
schema_ |
The version of the MetadataSchema. The version's format must match the following regular expression: |
schema |
Required. The raw YAML string representation of the MetadataSchema. The combination of [MetadataSchema.version] and the schema name given by The schema is defined as an OpenAPI 3.0.2 MetadataSchema Object |
schema_ |
The type of the MetadataSchema. This is a property that identifies which metadata types will use the MetadataSchema. |
create_ |
Output only. Timestamp when this MetadataSchema was created. |
description |
Description of the Metadata Schema |
MetadataSchemaType
Describes the type of the MetadataSchema.
Enums | |
---|---|
METADATA_SCHEMA_TYPE_UNSPECIFIED |
Unspecified type for the MetadataSchema. |
ARTIFACT_TYPE |
A type indicating that the MetadataSchema will be used by Artifacts. |
EXECUTION_TYPE |
A typee indicating that the MetadataSchema will be used by Executions. |
CONTEXT_TYPE |
A state indicating that the MetadataSchema will be used by Contexts. |
MetadataStore
Instance of a metadata store. Contains a set of metadata that can be queried.
Fields | |
---|---|
name |
Output only. The resource name of the MetadataStore instance. |
create_ |
Output only. Timestamp when this MetadataStore was created. |
update_ |
Output only. Timestamp when this MetadataStore was last updated. |
encryption_ |
Customer-managed encryption key spec for a Metadata Store. If set, this Metadata Store and all sub-resources of this Metadata Store are secured using this key. |
description |
Description of the MetadataStore. |
state |
Output only. State information of the MetadataStore. |
dataplex_ |
Optional. Dataplex integration settings. |
DataplexConfig
Represents Dataplex integration settings.
Fields | |
---|---|
enabled_ |
Optional. Whether or not Data Lineage synchronization is enabled for Vertex Pipelines. |
MetadataStoreState
Represents state information for a MetadataStore.
Fields | |
---|---|
disk_ |
The disk utilization of the MetadataStore in bytes. |
MigratableResource
Represents one resource that exists in automl.googleapis.com, datalabeling.googleapis.com or ml.googleapis.com.
Fields | |
---|---|
last_ |
Output only. Timestamp when the last migration attempt on this MigratableResource started. Will not be set if there's no migration attempt on this MigratableResource. |
last_ |
Output only. Timestamp when this MigratableResource was last updated. |
Union field
|
|
ml_ |
Output only. Represents one Version in ml.googleapis.com. |
automl_ |
Output only. Represents one Model in automl.googleapis.com. |
automl_ |
Output only. Represents one Dataset in automl.googleapis.com. |
data_ |
Output only. Represents one Dataset in datalabeling.googleapis.com. |
AutomlDataset
Represents one Dataset in automl.googleapis.com.
Fields | |
---|---|
dataset |
Full resource name of automl Dataset. Format: |
dataset_ |
The Dataset's display name in automl.googleapis.com. |
AutomlModel
Represents one Model in automl.googleapis.com.
Fields | |
---|---|
model |
Full resource name of automl Model. Format: |
model_ |
The Model's display name in automl.googleapis.com. |
DataLabelingDataset
Represents one Dataset in datalabeling.googleapis.com.
Fields | |
---|---|
dataset |
Full resource name of data labeling Dataset. Format: |
dataset_ |
The Dataset's display name in datalabeling.googleapis.com. |
data_ |
The migratable AnnotatedDataset in datalabeling.googleapis.com belongs to the data labeling Dataset. |
DataLabelingAnnotatedDataset
Represents one AnnotatedDataset in datalabeling.googleapis.com.
Fields | |
---|---|
annotated_ |
Full resource name of data labeling AnnotatedDataset. Format: |
annotated_ |
The AnnotatedDataset's display name in datalabeling.googleapis.com. |
MlEngineModelVersion
Represents one model Version in ml.googleapis.com.
Fields | |
---|---|
endpoint |
The ml.googleapis.com endpoint that this model Version currently lives in. Example values:
|
version |
Full resource name of ml engine model Version. Format: |
MigrateResourceRequest
Config of migrating one resource from automl.googleapis.com, datalabeling.googleapis.com and ml.googleapis.com to Vertex AI.
Fields | |
---|---|
Union field
|
|
migrate_ |
Config for migrating Version in ml.googleapis.com to Vertex AI's Model. |
migrate_ |
Config for migrating Model in automl.googleapis.com to Vertex AI's Model. |
migrate_ |
Config for migrating Dataset in automl.googleapis.com to Vertex AI's Dataset. |
migrate_ |
Config for migrating Dataset in datalabeling.googleapis.com to Vertex AI's Dataset. |
MigrateAutomlDatasetConfig
Config for migrating Dataset in automl.googleapis.com to Vertex AI's Dataset.
Fields | |
---|---|
dataset |
Required. Full resource name of automl Dataset. Format: |
dataset_ |
Required. Display name of the Dataset in Vertex AI. System will pick a display name if unspecified. |
MigrateAutomlModelConfig
Config for migrating Model in automl.googleapis.com to Vertex AI's Model.
Fields | |
---|---|
model |
Required. Full resource name of automl Model. Format: |
model_ |
Optional. Display name of the model in Vertex AI. System will pick a display name if unspecified. |
MigrateDataLabelingDatasetConfig
Config for migrating Dataset in datalabeling.googleapis.com to Vertex AI's Dataset.
Fields | |
---|---|
dataset |
Required. Full resource name of data labeling Dataset. Format: |
dataset_ |
Optional. Display name of the Dataset in Vertex AI. System will pick a display name if unspecified. |
migrate_ |
Optional. Configs for migrating AnnotatedDataset in datalabeling.googleapis.com to Vertex AI's SavedQuery. The specified AnnotatedDatasets have to belong to the datalabeling Dataset. |
MigrateDataLabelingAnnotatedDatasetConfig
Config for migrating AnnotatedDataset in datalabeling.googleapis.com to Vertex AI's SavedQuery.
Fields | |
---|---|
annotated_ |
Required. Full resource name of data labeling AnnotatedDataset. Format: |
MigrateMlEngineModelVersionConfig
Config for migrating version in ml.googleapis.com to Vertex AI's Model.
Fields | |
---|---|
endpoint |
Required. The ml.googleapis.com endpoint that this model version should be migrated from. Example values:
|
model_ |
Required. Full resource name of ml engine model version. Format: |
model_ |
Required. Display name of the model in Vertex AI. System will pick a display name if unspecified. |
MigrateResourceResponse
Describes a successfully migrated resource.
Fields | |
---|---|
migratable_ |
Before migration, the identifier in ml.googleapis.com, automl.googleapis.com or datalabeling.googleapis.com. |
Union field migrated_resource . After migration, the resource name in Vertex AI. migrated_resource can be only one of the following: |
|
dataset |
Migrated Dataset's resource name. |
model |
Migrated Model's resource name. |
Model
A trained machine learning Model.
Fields | |
---|---|
name |
The resource name of the Model. |
version_ |
Output only. Immutable. The version ID of the model. A new version is committed when a new model version is uploaded or trained under an existing model id. It is an auto-incrementing decimal number in string representation. |
version_ |
User provided version aliases so that a model version can be referenced via alias (i.e. |
version_ |
Output only. Timestamp when this version was created. |
version_ |
Output only. Timestamp when this version was most recently updated. |
display_ |
Required. The display name of the Model. The name can be up to 128 characters long and can consist of any UTF-8 characters. |
description |
The description of the Model. |
version_ |
The description of this version. |
predict_ |
The schemata that describe formats of the Model's predictions and explanations as given and returned via |
metadata_ |
Immutable. Points to a YAML file stored on Google Cloud Storage describing additional information about the Model, that is specific to it. Unset if the Model does not have any additional information. The schema is defined as an OpenAPI 3.0.2 Schema Object. AutoML Models always have this field populated by Vertex AI, if no additional metadata is needed, this field is set to an empty string. Note: The URI given on output will be immutable and probably different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access. |
metadata |
Immutable. An additional information about the Model; the schema of the metadata can be found in |
supported_ |
Output only. The formats in which this Model may be exported. If empty, this Model is not available for export. |
training_ |
Output only. The resource name of the TrainingPipeline that uploaded this Model, if any. |
pipeline_ |
Optional. This field is populated if the model is produced by a pipeline job. |
container_ |
Input only. The specification of the container that is to be used when deploying this Model. The specification is ingested upon |
artifact_ |
Immutable. The path to the directory containing the Model artifact and any of its supporting files. Not required for AutoML Models. |
supported_ |
Output only. When this Model is deployed, its prediction resources are described by the |
supported_ |
Output only. The formats this Model supports in The possible formats are:
If this Model doesn't support any of these formats it means it cannot be used with a |
supported_ |
Output only. The formats this Model supports in The possible formats are:
If this Model doesn't support any of these formats it means it cannot be used with a |
create_ |
Output only. Timestamp when this Model was uploaded into Vertex AI. |
update_ |
Output only. Timestamp when this Model was most recently updated. |
deployed_ |
Output only. The pointers to DeployedModels created from this Model. Note that Model could have been deployed to Endpoints in different Locations. |
explanation_ |
The default explanation specification for this Model. The Model can be used for All fields of the explanation_spec can be overridden by If the default explanation specification is not set for this Model, this Model can still be used for |
etag |
Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
labels |
The labels with user-defined metadata to organize your Models. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. |
data_ |
Stats of data used for training or evaluating the Model. Only populated when the Model is trained by a TrainingPipeline with |
encryption_ |
Customer-managed encryption key spec for a Model. If set, this Model and all sub-resources of this Model will be secured by this key. |
model_ |
Output only. Source of a model. It can either be automl training pipeline, custom training pipeline, BigQuery ML, or saved and tuned from Genie or Model Garden. |
original_ |
Output only. If this Model is a copy of another Model, this contains info about the original. |
metadata_ |
Output only. The resource name of the Artifact that was created in MetadataStore when creating the Model. The Artifact resource name pattern is |
base_ |
Optional. User input field to specify the base model source. Currently it only supports specifing the Model Garden models and Genie models. |
satisfies_ |
Output only. Reserved for future use. |
satisfies_ |
Output only. Reserved for future use. |
BaseModelSource
User input field to specify the base model source. Currently it only supports specifing the Model Garden models and Genie models.
Fields | |
---|---|
Union field
|
|
model_ |
Source information of Model Garden models. |
genie_ |
Information about the base model of Genie models. |
DataStats
Stats of data used for train or evaluate the Model.
Fields | |
---|---|
training_ |
Number of DataItems that were used for training this Model. |
validation_ |
Number of DataItems that were used for validating this Model during training. |
test_ |
Number of DataItems that were used for evaluating this Model. If the Model is evaluated multiple times, this will be the number of test DataItems used by the first evaluation. If the Model is not evaluated, the number is 0. |
training_ |
Number of Annotations that are used for training this Model. |
validation_ |
Number of Annotations that are used for validating this Model during training. |
test_ |
Number of Annotations that are used for evaluating this Model. If the Model is evaluated multiple times, this will be the number of test Annotations used by the first evaluation. If the Model is not evaluated, the number is 0. |
DeploymentResourcesType
Identifies a type of Model's prediction resources.
Enums | |
---|---|
DEPLOYMENT_RESOURCES_TYPE_UNSPECIFIED |
Should not be used. |
DEDICATED_RESOURCES |
Resources that are dedicated to the DeployedModel , and that need a higher degree of manual configuration. |
AUTOMATIC_RESOURCES |
Resources that to large degree are decided by Vertex AI, and require only a modest additional configuration. |
SHARED_RESOURCES |
Resources that can be shared by multiple DeployedModels . A pre-configured DeploymentResourcePool is required. |
ExportFormat
Represents export format supported by the Model. All formats export to Google Cloud Storage.
Fields | |
---|---|
id |
Output only. The ID of the export format. The possible format IDs are:
|
exportable_ |
Output only. The content of this Model that may be exported. |
ExportableContent
The Model content that can be exported.
Enums | |
---|---|
EXPORTABLE_CONTENT_UNSPECIFIED |
Should not be used. |
ARTIFACT |
Model artifact and any of its supported files. Will be exported to the location specified by the artifactDestination field of the ExportModelRequest.output_config object. |
IMAGE |
The container image that is to be used when deploying this Model. Will be exported to the location specified by the imageDestination field of the ExportModelRequest.output_config object. |
OriginalModelInfo
Contains information about the original Model if this Model is a copy.
Fields | |
---|---|
model |
Output only. The resource name of the Model this Model is a copy of, including the revision. Format: |
ModelContainerSpec
Specification of a container for serving predictions. Some fields in this message correspond to fields in the Kubernetes Container v1 core specification.
Fields | |
---|---|
image_ |
Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the container publishing requirements, including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon To learn about the requirements for the Docker image itself, see Custom container requirements. You can use the URI to one of Vertex AI's pre-built container images for prediction in this field. |
command[] |
Immutable. Specifies the command that runs when the container starts. This overrides the container's ENTRYPOINT. Specify this field as an array of executable and arguments, similar to a Docker If you do not specify this field, then the container's If you specify this field, then you can also specify the In this field, you can reference environment variables set by Vertex AI and environment variables set in the
Note that this differs from Bash variable expansion, which does not use parentheses. If a variable cannot be resolved, the reference in the input string is used unchanged. To avoid variable expansion, you can escape this syntax with
This field corresponds to the |
args[] |
Immutable. Specifies arguments for the command that runs when the container starts. This overrides the container's If you don't specify this field but do specify the If you don't specify this field and don't specify the In this field, you can reference environment variables set by Vertex AI and environment variables set in the
Note that this differs from Bash variable expansion, which does not use parentheses. If a variable cannot be resolved, the reference in the input string is used unchanged. To avoid variable expansion, you can escape this syntax with
This field corresponds to the |
env[] |
Immutable. List of environment variables to set in the container. After the container starts running, code running in the container can read these environment variables. Additionally, the
If you switch the order of the variables in the example, then the expansion does not occur. This field corresponds to the |
ports[] |
Immutable. List of ports to expose from the container. Vertex AI sends any prediction requests that it receives to the first port on this list. Vertex AI also sends liveness and health checks to this port. If you do not specify this field, it defaults to following value:
Vertex AI does not use ports other than the first one listed. This field corresponds to the |
predict_ |
Immutable. HTTP path on the container to send prediction requests to. Vertex AI forwards requests sent using For example, if you set this field to If you don't specify this field, it defaults to the following value when you
The placeholders in this value are replaced as follows:
|
health_ |
Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about health checks. For example, if you set this field to If you don't specify this field, it defaults to the following value when you
The placeholders in this value are replaced as follows:
|
grpc_ |
Immutable. List of ports to expose from the container. Vertex AI sends gRPC prediction requests that it receives to the first port on this list. Vertex AI also sends liveness and health checks to this port. If you do not specify this field, gRPC requests to the container will be disabled. Vertex AI does not use ports other than the first one listed. This field corresponds to the |
deployment_ |
Immutable. Deployment timeout. Limit for deployment timeout is 2 hours. |
shared_ |
Immutable. The amount of the VM memory to reserve as the shared memory for the model in megabytes. |
startup_ |
Immutable. Specification for Kubernetes startup probe. |
health_ |
Immutable. Specification for Kubernetes readiness probe. |
ModelDeploymentMonitoringBigQueryTable
ModelDeploymentMonitoringBigQueryTable specifies the BigQuery table name as well as some information of the logs stored in this table.
Fields | |
---|---|
log_ |
The source of log. |
log_ |
The type of log. |
bigquery_ |
The created BigQuery table to store logs. Customer could do their own query & analysis. Format: |
request_ |
Output only. The schema version of the request/response logging BigQuery table. Default to v1 if unset. |
LogSource
Indicates where does the log come from.
Enums | |
---|---|
LOG_SOURCE_UNSPECIFIED |
Unspecified source. |
TRAINING |
Logs coming from Training dataset. |
SERVING |
Logs coming from Serving traffic. |
LogType
Indicates what type of traffic does the log belong to.
Enums | |
---|---|
LOG_TYPE_UNSPECIFIED |
Unspecified type. |
PREDICT |
Predict logs. |
EXPLAIN |
Explain logs. |
ModelDeploymentMonitoringJob
Represents a job that runs periodically to monitor the deployed models in an endpoint. It will analyze the logged training & prediction data to detect any abnormal behaviors.
Fields | |
---|---|
name |
Output only. Resource name of a ModelDeploymentMonitoringJob. |
display_ |
Required. The user-defined name of the ModelDeploymentMonitoringJob. The name can be up to 128 characters long and can consist of any UTF-8 characters. Display name of a ModelDeploymentMonitoringJob. |
endpoint |
Required. Endpoint resource name. Format: |
state |
Output only. The detailed state of the monitoring job. When the job is still creating, the state will be 'PENDING'. Once the job is successfully created, the state will be 'RUNNING'. Pause the job, the state will be 'PAUSED'. Resume the job, the state will return to 'RUNNING'. |
schedule_ |
Output only. Schedule state when the monitoring job is in Running state. |
latest_ |
Output only. Latest triggered monitoring pipeline metadata. |
model_ |
Required. The config for monitoring objectives. This is a per DeployedModel config. Each DeployedModel needs to be configured separately. |
model_ |
Required. Schedule config for running the monitoring job. |
logging_ |
Required. Sample Strategy for logging. |
model_ |
Alert config for model monitoring. |
predict_ |
YAML schema file uri describing the format of a single instance, which are given to format this Endpoint's prediction (and explanation). If not set, we will generate predict schema from collected predict requests. |
sample_ |
Sample Predict instance, same format as |
analysis_ |
YAML schema file uri describing the format of a single instance that you want Tensorflow Data Validation (TFDV) to analyze. If this field is empty, all the feature data types are inferred from |
bigquery_ |
Output only. The created bigquery tables for the job under customer project. Customer could do their own query & analysis. There could be 4 log tables in maximum: 1. Training data logging predict request/response 2. Serving data logging predict request/response |
log_ |
The TTL of BigQuery tables in user projects which stores logs. A day is the basic unit of the TTL and we take the ceil of TTL/86400(a day). e.g. { second: 3600} indicates ttl = 1 day. |
labels |
The labels with user-defined metadata to organize your ModelDeploymentMonitoringJob. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. |
create_ |
Output only. Timestamp when this ModelDeploymentMonitoringJob was created. |
update_ |
Output only. Timestamp when this ModelDeploymentMonitoringJob was updated most recently. |
next_ |
Output only. Timestamp when this monitoring pipeline will be scheduled to run for the next round. |
stats_ |
Stats anomalies base folder path. |
encryption_ |
Customer-managed encryption key spec for a ModelDeploymentMonitoringJob. If set, this ModelDeploymentMonitoringJob and all sub-resources of this ModelDeploymentMonitoringJob will be secured by this key. |
enable_ |
If true, the scheduled monitoring pipeline logs are sent to Google Cloud Logging, including pipeline status and anomalies detected. Please note the logs incur cost, which are subject to Cloud Logging pricing. |
error |
Output only. Only populated when the job's state is |
satisfies_ |
Output only. Reserved for future use. |
satisfies_ |
Output only. Reserved for future use. |
LatestMonitoringPipelineMetadata
All metadata of most recent monitoring pipelines.
Fields | |
---|---|
run_ |
The time that most recent monitoring pipelines that is related to this run. |
status |
The status of the most recent monitoring pipeline. |
MonitoringScheduleState
The state to Specify the monitoring pipeline.
Enums | |
---|---|
MONITORING_SCHEDULE_STATE_UNSPECIFIED |
Unspecified state. |
PENDING |
The pipeline is picked up and wait to run. |
OFFLINE |
The pipeline is offline and will be scheduled for next run. |
RUNNING |
The pipeline is running. |
ModelDeploymentMonitoringObjectiveConfig
ModelDeploymentMonitoringObjectiveConfig contains the pair of deployed_model_id to ModelMonitoringObjectiveConfig.
Fields | |
---|---|
deployed_ |
The DeployedModel ID of the objective config. |
objective_ |
The objective config of for the modelmonitoring job of this deployed model. |
ModelDeploymentMonitoringObjectiveType
The Model Monitoring Objective types.
Enums | |
---|---|
MODEL_DEPLOYMENT_MONITORING_OBJECTIVE_TYPE_UNSPECIFIED |
Default value, should not be set. |
RAW_FEATURE_SKEW |
Raw feature values' stats to detect skew between Training-Prediction datasets. |
RAW_FEATURE_DRIFT |
Raw feature values' stats to detect drift between Serving-Prediction datasets. |
FEATURE_ATTRIBUTION_SKEW |
Feature attribution scores to detect skew between Training-Prediction datasets. |
FEATURE_ATTRIBUTION_DRIFT |
Feature attribution scores to detect skew between Prediction datasets collected within different time windows. |
ModelDeploymentMonitoringScheduleConfig
The config for scheduling monitoring job.
Fields | |
---|---|
monitor_ |
Required. The model monitoring job scheduling interval. It will be rounded up to next full hour. This defines how often the monitoring jobs are triggered. |
monitor_ |
The time window of the prediction data being included in each prediction dataset. This window specifies how long the data should be collected from historical model results for each run. If not set, |
ModelEvaluation
A collection of metrics calculated by comparing Model's predictions on all of the test data against annotations from the test data.
Fields | |
---|---|
name |
Output only. The resource name of the ModelEvaluation. |
display_ |
The display name of the ModelEvaluation. |
metrics_ |
Points to a YAML file stored on Google Cloud Storage describing the |
metrics |
Evaluation metrics of the Model. The schema of the metrics is stored in |
create_ |
Output only. Timestamp when this ModelEvaluation was created. |
slice_ |
All possible |
data_ |
Points to a YAML file stored on Google Cloud Storage describing [EvaluatedDataItemView.data_item_payload][] and This field is not populated if there are neither EvaluatedDataItemViews nor EvaluatedAnnotations under this ModelEvaluation. |
annotation_ |
Points to a YAML file stored on Google Cloud Storage describing [EvaluatedDataItemView.predictions][], [EvaluatedDataItemView.ground_truths][], This field is not populated if there are neither EvaluatedDataItemViews nor EvaluatedAnnotations under this ModelEvaluation. |
model_ |
Aggregated explanation metrics for the Model's prediction output over the data this ModelEvaluation uses. This field is populated only if the Model is evaluated with explanations, and only for AutoML tabular Models. |
explanation_ |
Describes the values of |
metadata |
The metadata of the ModelEvaluation. For the ModelEvaluation uploaded from Managed Pipeline, metadata contains a structured value with keys of "pipeline_job_id", "evaluation_dataset_type", "evaluation_dataset_path", "row_based_metrics_path". |
ModelEvaluationExplanationSpec
Fields | |
---|---|
explanation_ |
Explanation type. For AutoML Image Classification models, possible values are:
|
explanation_ |
Explanation spec details. |
ModelEvaluationSlice
A collection of metrics calculated by comparing Model's predictions on a slice of the test data against ground truth annotations.
Fields | |
---|---|
name |
Output only. The resource name of the ModelEvaluationSlice. |
slice |
Output only. The slice of the test data that is used to evaluate the Model. |
metrics_ |
Output only. Points to a YAML file stored on Google Cloud Storage describing the |
metrics |
Output only. Sliced evaluation metrics of the Model. The schema of the metrics is stored in |
create_ |
Output only. Timestamp when this ModelEvaluationSlice was created. |
model_ |
Output only. Aggregated explanation metrics for the Model's prediction output over the data this ModelEvaluation uses. This field is populated only if the Model is evaluated with explanations, and only for tabular Models. |
Slice
Definition of a slice.
Fields | |
---|---|
dimension |
Output only. The dimension of the slice. Well-known dimensions are: * |
value |
Output only. The value of the dimension in this slice. |
slice_ |
Output only. Specification for how the data was sliced. |
SliceSpec
Specification for how the data should be sliced.
Fields | |
---|---|
configs |
Mapping configuration for this SliceSpec. The key is the name of the feature. By default, the key will be prefixed by "instance" as a dictionary prefix for Vertex Batch Predictions output format. |
Range
A range of values for slice(s). low
is inclusive, high
is exclusive.
Fields | |
---|---|
low |
Inclusive low value for the range. |
high |
Exclusive high value for the range. |
SliceConfig
Specification message containing the config for this SliceSpec. When kind
is selected as value
and/or range
, only a single slice will be computed. When all_values
is present, a separate slice will be computed for each possible label/value for the corresponding key in config
. Examples, with feature zip_code with values 12345, 23334, 88888 and feature country with values "US", "Canada", "Mexico" in the dataset:
Example 1:
{
"zip_code": { "value": { "float_value": 12345.0 } }
}
A single slice for any data with zip_code 12345 in the dataset.
Example 2:
{
"zip_code": { "range": { "low": 12345, "high": 20000 } }
}
A single slice containing data where the zip_codes between 12345 and 20000 For this example, data with the zip_code of 12345 will be in this slice.
Example 3:
{
"zip_code": { "range": { "low": 10000, "high": 20000 } },
"country": { "value": { "string_value": "US" } }
}
A single slice containing data where the zip_codes between 10000 and 20000 has the country "US". For this example, data with the zip_code of 12345 and country "US" will be in this slice.
Example 4:
{ "country": {"all_values": { "value": true } } }
Three slices are computed, one for each unique country in the dataset.
Example 5:
{
"country": { "all_values": { "value": true } },
"zip_code": { "value": { "float_value": 12345.0 } }
}
Three slices are computed, one for each unique country in the dataset where the zip_code is also 12345. For this example, data with zip_code 12345 and country "US" will be in one slice, zip_code 12345 and country "Canada" in another slice, and zip_code 12345 and country "Mexico" in another slice, totaling 3 slices.
Fields | |
---|---|
Union field
|
|
value |
A unique specific value for a given feature. Example: |
range |
A range of values for a numerical feature. Example: |
all_ |
If all_values is set to true, then all possible labels of the keyed feature will have another slice computed. Example: |
Value
Single value that supports strings and floats.
Fields | |
---|---|
Union field
|
|
string_ |
String type. |
float_ |
Float type. |
ModelExplanation
Aggregated explanation metrics for a Model over a set of instances.
Fields | |
---|---|
mean_ |
Output only. Aggregated attributions explaining the Model's prediction outputs over the set of instances. The attributions are grouped by outputs. For Models that predict only one output, such as regression Models that predict only one score, there is only one attibution that explains the predicted output. For Models that predict multiple outputs, such as multiclass Models that predict multiple classes, each element explains one specific item. The NOTE: Currently AutoML tabular classification Models produce only one attribution, which averages attributions over all the classes it predicts. |
ModelGardenSource
Contains information about the source of the models generated from Model Garden.
Fields | |
---|---|
public_ |
Required. The model garden source model resource name. |
ModelMonitoringAlertConfig
The alert config for model monitoring.
Fields | |
---|---|
enable_ |
Dump the anomalies to Cloud Logging. The anomalies will be put to json payload encoded from proto |
notification_ |
Resource names of the NotificationChannels to send alert. Must be of the format |
Union field
|
|
email_ |
Email alert config. |
EmailAlertConfig
The config for email alert.
Fields | |
---|---|
user_ |
The email addresses to send the alert. |
ModelMonitoringObjectiveConfig
The objective configuration for model monitoring, including the information needed to detect anomalies for one particular model.
Fields | |
---|---|
training_ |
Training dataset for models. This field has to be set only if TrainingPredictionSkewDetectionConfig is specified. |
training_ |
The config for skew between training data and prediction data. |
prediction_ |
The config for drift of prediction data. |
explanation_ |
The config for integrating with Vertex Explainable AI. |
ExplanationConfig
The config for integrating with Vertex Explainable AI. Only applicable if the Model has explanation_spec populated.
Fields | |
---|---|
enable_ |
If want to analyze the Vertex Explainable AI feature attribute scores or not. If set to true, Vertex AI will log the feature attributions from explain response and do the skew/drift detection for them. |
explanation_ |
Predictions generated by the BatchPredictionJob using baseline dataset. |
ExplanationBaseline
Output from BatchPredictionJob
for Model Monitoring baseline dataset, which can be used to generate baseline attribution scores.
Fields | |
---|---|
prediction_ |
The storage format of the predictions generated BatchPrediction job. |
Union field destination . The configuration specifying of BatchExplain job output. This can be used to generate the baseline of feature attribution scores. destination can be only one of the following: |
|
gcs |
Cloud Storage location for BatchExplain output. |
bigquery |
BigQuery location for BatchExplain output. |
PredictionFormat
The storage format of the predictions generated BatchPrediction job.
Enums | |
---|---|
PREDICTION_FORMAT_UNSPECIFIED |
Should not be set. |
JSONL |
Predictions are in JSONL files. |
BIGQUERY |
Predictions are in BigQuery. |
PredictionDriftDetectionConfig
The config for Prediction data drift detection.
Fields | |
---|---|
drift_ |
Key is the feature name and value is the threshold. If a feature needs to be monitored for drift, a value threshold must be configured for that feature. The threshold here is against feature distribution distance between different time windws. |
attribution_ |
Key is the feature name and value is the threshold. The threshold here is against attribution score distance between different time windows. |
default_ |
Drift anomaly detection threshold used by all features. When the per-feature thresholds are not set, this field can be used to specify a threshold for all features. |
TrainingDataset
Training Dataset information.
Fields | |
---|---|
data_ |
Data format of the dataset, only applicable if the input is from Google Cloud Storage. The possible formats are: "tf-record" The source file is a TFRecord file. "csv" The source file is a CSV file. "jsonl" The source file is a JSONL file. |
target_ |
The target field name the model is to predict. This field will be excluded when doing Predict and (or) Explain for the training data. |
logging_ |
Strategy to sample data from Training Dataset. If not set, we process the whole dataset. |
Union field
|
|
dataset |
The resource name of the Dataset used to train this Model. |
gcs_ |
The Google Cloud Storage uri of the unmanaged Dataset used to train this Model. |
bigquery_ |
The BigQuery table of the unmanaged Dataset used to train this Model. |
TrainingPredictionSkewDetectionConfig
The config for Training & Prediction data skew detection. It specifies the training dataset sources and the skew detection parameters.
Fields | |
---|---|
skew_ |
Key is the feature name and value is the threshold. If a feature needs to be monitored for skew, a value threshold must be configured for that feature. The threshold here is against feature distribution distance between the training and prediction feature. |
attribution_ |
Key is the feature name and value is the threshold. The threshold here is against attribution score distance between the training and prediction feature. |
default_ |
Skew anomaly detection threshold used by all features. When the per-feature thresholds are not set, this field can be used to specify a threshold for all features. |
ModelMonitoringStatsAnomalies
Statistics and anomalies generated by Model Monitoring.
Fields | |
---|---|
objective |
Model Monitoring Objective those stats and anomalies belonging to. |
deployed_ |
Deployed Model ID. |
anomaly_ |
Number of anomalies within all stats. |
feature_ |
A list of historical Stats and Anomalies generated for all Features. |
FeatureHistoricStatsAnomalies
Historical Stats (and Anomalies) for a specific Feature.
Fields | |
---|---|
feature_ |
Display Name of the Feature. |
threshold |
Threshold for anomaly detection. |
training_ |
Stats calculated for the Training Dataset. |
prediction_ |
A list of historical stats generated by different time window's Prediction Dataset. |
ModelSourceInfo
Detail description of the source information of the model.
Fields | |
---|---|
source_ |
Type of the model source. |
ModelSourceType
Source of the model. Different from objective
field, this ModelSourceType
enum indicates the source from which the model was accessed or obtained, whereas the objective
indicates the overall aim or function of this model.
Enums | |
---|---|
MODEL_SOURCE_TYPE_UNSPECIFIED |
Should not be used. |
AUTOML |
The Model is uploaded by automl training pipeline. |
CUSTOM |
The Model is uploaded by user or custom training pipeline. |
BQML |
The Model is registered and sync'ed from BigQuery ML. |
MODEL_GARDEN |
The Model is saved or tuned from Model Garden. |
GENIE |
The Model is saved or tuned from Genie. |
CUSTOM_TEXT_EMBEDDING |
The Model is uploaded by text embedding finetuning pipeline. |
MARKETPLACE |
The Model is saved or tuned from Marketplace. |
MutateDeployedIndexOperationMetadata
Runtime operation information for IndexEndpointService.MutateDeployedIndex
.
Fields | |
---|---|
generic_ |
The operation generic information. |
deployed_ |
The unique index id specified by user |
MutateDeployedIndexRequest
Request message for IndexEndpointService.MutateDeployedIndex
.
Fields | |
---|---|
index_ |
Required. The name of the IndexEndpoint resource into which to deploy an Index. Format: |
deployed_ |
Required. The DeployedIndex to be updated within the IndexEndpoint. Currently, the updatable fields are |
MutateDeployedIndexResponse
Response message for IndexEndpointService.MutateDeployedIndex
.
Fields | |
---|---|
deployed_ |
The DeployedIndex that had been updated in the IndexEndpoint. |
MutateDeployedModelOperationMetadata
Runtime operation information for EndpointService.MutateDeployedModel
.
Fields | |
---|---|
generic_ |
The operation generic information. |
MutateDeployedModelRequest
Request message for EndpointService.MutateDeployedModel
.
Fields | |
---|---|
endpoint |
Required. The name of the Endpoint resource into which to mutate a DeployedModel. Format: |
deployed_ |
Required. The DeployedModel to be mutated within the Endpoint. Only the following fields can be mutated:
|
update_ |
Required. The update mask applies to the resource. See |
MutateDeployedModelResponse
Response message for EndpointService.MutateDeployedModel
.
Fields | |
---|---|
deployed_ |
The DeployedModel that's being mutated. |
NasJob
Represents a Neural Architecture Search (NAS) job.
Fields | |
---|---|
name |
Output only. Resource name of the NasJob. |
display_ |
Required. The display name of the NasJob. The name can be up to 128 characters long and can consist of any UTF-8 characters. |
nas_ |
Required. The specification of a NasJob. |
nas_ |
Output only. Output of the NasJob. |
state |
Output only. The detailed state of the job. |
create_ |
Output only. Time when the NasJob was created. |
start_ |
Output only. Time when the NasJob for the first time entered the |
end_ |
Output only. Time when the NasJob entered any of the following states: |
update_ |
Output only. Time when the NasJob was most recently updated. |
error |
Output only. Only populated when job's state is JOB_STATE_FAILED or JOB_STATE_CANCELLED. |
labels |
The labels with user-defined metadata to organize NasJobs. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. |
encryption_ |
Customer-managed encryption key options for a NasJob. If this is set, then all resources created by the NasJob will be encrypted with the provided encryption key. |
enable_restricted_image_training |
Optional. Enable a separation of Custom model training and restricted image training for tenant project. |
satisfies_ |
Output only. Reserved for future use. |
satisfies_ |
Output only. Reserved for future use. |
NasJobOutput
Represents a uCAIP NasJob output.
Fields | |
---|---|
Union field output . The output of this Neural Architecture Search (NAS) job. output can be only one of the following: |
|
multi_ |
Output only. The output of this multi-trial Neural Architecture Search (NAS) job. |
MultiTrialJobOutput
The output of a multi-trial Neural Architecture Search (NAS) jobs.
Fields | |
---|---|
search_ |
Output only. List of NasTrials that were started as part of search stage. |
train_ |
Output only. List of NasTrials that were started as part of train stage. |
NasJobSpec
Represents the spec of a NasJob.
Fields | |
---|---|
resume_ |
The ID of the existing NasJob in the same Project and Location which will be used to resume search. search_space_spec and nas_algorithm_spec are obtained from previous NasJob hence should not provide them again for this NasJob. |
search_ |
It defines the search space for Neural Architecture Search (NAS). |
Union field nas_algorithm_spec . The Neural Architecture Search (NAS) algorithm specification. nas_algorithm_spec can be only one of the following: |
|
multi_ |
The spec of multi-trial algorithms. |
MultiTrialAlgorithmSpec
The spec of multi-trial Neural Architecture Search (NAS).
Fields | |
---|---|
multi_ |
The multi-trial Neural Architecture Search (NAS) algorithm type. Defaults to |
metric |
Metric specs for the NAS job. Validation for this field is done at |
search_ |
Required. Spec for search trials. |
train_ |
Spec for train trials. Top N [TrainTrialSpec.max_parallel_trial_count] search trials will be trained for every M [TrainTrialSpec.frequency] trials searched. |
MetricSpec
Represents a metric to optimize.
Fields | |
---|---|
metric_ |
Required. The ID of the metric. Must not contain whitespaces. |
goal |
Required. The optimization goal of the metric. |
GoalType
The available types of optimization goals.
Enums | |
---|---|
GOAL_TYPE_UNSPECIFIED |
Goal Type will default to maximize. |
MAXIMIZE |
Maximize the goal metric. |
MINIMIZE |
Minimize the goal metric. |
MultiTrialAlgorithm
The available types of multi-trial algorithms.
Enums | |
---|---|
MULTI_TRIAL_ALGORITHM_UNSPECIFIED |
Defaults to REINFORCEMENT_LEARNING . |
REINFORCEMENT_LEARNING |
The Reinforcement Learning Algorithm for Multi-trial Neural Architecture Search (NAS). |
GRID_SEARCH |
The Grid Search Algorithm for Multi-trial Neural Architecture Search (NAS). |
SearchTrialSpec
Represent spec for search trials.
Fields | |
---|---|
search_ |
Required. The spec of a search trial job. The same spec applies to all search trials. |
max_ |
Required. The maximum number of Neural Architecture Search (NAS) trials to run. |
max_ |
Required. The maximum number of trials to run in parallel. |
max_ |
The number of failed trials that need to be seen before failing the NasJob. If set to 0, Vertex AI decides how many trials must fail before the whole job fails. |
TrainTrialSpec
Represent spec for train trials.
Fields | |
---|---|
train_ |
Required. The spec of a train trial job. The same spec applies to all train trials. |
max_ |
Required. The maximum number of trials to run in parallel. |
frequency |
Required. Frequency of search trials to start train stage. Top N [TrainTrialSpec.max_parallel_trial_count] search trials will be trained for every M [TrainTrialSpec.frequency] trials searched. |
NasTrial
Represents a uCAIP NasJob trial.
Fields | |
---|---|
id |
Output only. The identifier of the NasTrial assigned by the service. |
state |
Output only. The detailed state of the NasTrial. |
final_ |
Output only. The final measurement containing the objective value. |
start_ |
Output only. Time when the NasTrial was started. |
end_ |
Output only. Time when the NasTrial's status changed to |
State
Describes a NasTrial state.
Enums | |
---|---|
STATE_UNSPECIFIED |
The NasTrial state is unspecified. |
REQUESTED |
Indicates that a specific NasTrial has been requested, but it has not yet been suggested by the service. |
ACTIVE |
Indicates that the NasTrial has been suggested. |
STOPPING |
Indicates that the NasTrial should stop according to the service. |
SUCCEEDED |
Indicates that the NasTrial is completed successfully. |
INFEASIBLE |
Indicates that the NasTrial should not be attempted again. The service will set a NasTrial to INFEASIBLE when it's done but missing the final_measurement. |
NasTrialDetail
Represents a NasTrial details along with its parameters. If there is a corresponding train NasTrial, the train NasTrial is also returned.
Fields | |
---|---|
name |
Output only. Resource name of the NasTrialDetail. |
parameters |
The parameters for the NasJob NasTrial. |
search_ |
The requested search NasTrial. |
train_ |
The train NasTrial corresponding to |
NearestNeighborQuery
A query to find a number of similar entities.
Fields | |
---|---|
neighbor_ |
Optional. The number of similar entities to be retrieved from feature view for each query. |
string_ |
Optional. The list of string filters. |
numeric_ |
Optional. The list of numeric filters. |
per_ |
Optional. Crowding is a constraint on a neighbor list produced by nearest neighbor search requiring that no more than sper_crowding_attribute_neighbor_count of the k neighbors returned have the same value of crowding_attribute. It's used for improving result diversity. |
parameters |
Optional. Parameters that can be set to tune query on the fly. |
Union field
|
|
entity_ |
Optional. The entity id whose similar entities should be searched for. If embedding is set, search will use embedding instead of entity_id. |
embedding |
Optional. The embedding vector that be used for similar search. |
Embedding
The embedding vector.
Fields | |
---|---|
value[] |
Optional. Individual value in the embedding. |
NumericFilter
Numeric filter is used to search a subset of the entities by using boolean rules on numeric columns. For example: Database Point 0: {name: "a" value_int: 42} {name: "b" value_float: 1.0} Database Point 1: {name: "a" value_int: 10} {name: "b" value_float: 2.0} Database Point 2: {name: "a" value_int: -1} {name: "b" value_float: 3.0} Query: {name: "a" value_int: 12 operator: LESS} // Matches Point 1, 2 {name: "b" value_float: 2.0 operator: EQUAL} // Matches Point 1
Fields | |
---|---|
name |
Required. Column name in BigQuery that used as filters. |
Union field Value . The type of Value must be consistent for all datapoints with a given name. This is verified at runtime. Value can be only one of the following: |
|
value_ |
int value type. |
value_ |
float value type. |
value_ |
double value type. |
op |
Optional. This MUST be specified for queries and must NOT be specified for database points. |
Operator
Datapoints for which Operator is true relative to the query's Value field will be allowlisted.
Enums | |
---|---|
OPERATOR_UNSPECIFIED |
Unspecified operator. |
LESS |
Entities are eligible if their value is < the query's. |
LESS_EQUAL |
Entities are eligible if their value is <= the query's. |
EQUAL |
Entities are eligible if their value is == the query's. |
GREATER_EQUAL |
Entities are eligible if their value is >= the query's. |
GREATER |
Entities are eligible if their value is > the query's. |
NOT_EQUAL |
Entities are eligible if their value is != the query's. |
Parameters
Parameters that can be overrided in each query to tune query latency and recall.
Fields | |
---|---|
approximate_ |
Optional. The number of neighbors to find via approximate search before exact reordering is performed; if set, this value must be > neighbor_count. |
leaf_ |
Optional. The fraction of the number of leaves to search, set at query time allows user to tune search performance. This value increase result in both search accuracy and latency increase. The value should be between 0.0 and 1.0. |
StringFilter
String filter is used to search a subset of the entities by using boolean rules on string columns. For example: if a query specifies string filter with 'name = color, allow_tokens = {red, blue}, deny_tokens = {purple}',' then that query will match entities that are red or blue, but if those points are also purple, then they will be excluded even if they are red/blue. Only string filter is supported for now, numeric filter will be supported in the near future.
Fields | |
---|---|
name |
Required. Column names in BigQuery that used as filters. |
allow_ |
Optional. The allowed tokens. |
deny_ |
Optional. The denied tokens. |
NearestNeighborSearchOperationMetadata
Runtime operation metadata with regard to Matching Engine Index.
Fields | |
---|---|
content_ |
The validation stats of the content (per file) to be inserted or updated on the Matching Engine Index resource. Populated if contentsDeltaUri is provided as part of |
data_ |
The ingested data size in bytes. |
ContentValidationStats
Fields | |
---|---|
source_ |
Cloud Storage URI pointing to the original file in user's bucket. |
valid_ |
Number of records in this file that were successfully processed. |
invalid_ |
Number of records in this file we skipped due to validate errors. |
partial_ |
The detail information of the partial failures encountered for those invalid records that couldn't be parsed. Up to 50 partial errors will be reported. |
valid_ |
Number of sparse records in this file that were successfully processed. |
invalid_ |
Number of sparse records in this file we skipped due to validate errors. |
RecordError
Fields | |
---|---|
error_ |
The error type of this record. |
error_ |
A human-readable message that is shown to the user to help them fix the error. Note that this message may change from time to time, your code should check against error_type as the source of truth. |
source_ |
Cloud Storage URI pointing to the original file in user's bucket. |
embedding_ |
Empty if the embedding id is failed to parse. |
raw_ |
The original content of this record. |
RecordErrorType
Enums | |
---|---|
ERROR_TYPE_UNSPECIFIED |
Default, shall not be used. |
EMPTY_LINE |
The record is empty. |
INVALID_JSON_SYNTAX |
Invalid json format. |
INVALID_CSV_SYNTAX |
Invalid csv format. |
INVALID_AVRO_SYNTAX |
Invalid avro format. |
INVALID_EMBEDDING_ID |
The embedding id is not valid. |
EMBEDDING_SIZE_MISMATCH |
The size of the dense embedding vectors does not match with the specified dimension. |
NAMESPACE_MISSING |
The namespace field is missing. |
PARSING_ERROR |
Generic catch-all error. Only used for validation failure where the root cause cannot be easily retrieved programmatically. |
DUPLICATE_NAMESPACE |
There are multiple restricts with the same namespace value. |
OP_IN_DATAPOINT |
Numeric restrict has operator specified in datapoint. |
MULTIPLE_VALUES |
Numeric restrict has multiple values specified. |
INVALID_NUMERIC_VALUE |
Numeric restrict has invalid numeric value specified. |
INVALID_ENCODING |
File is not in UTF_8 format. |
INVALID_SPARSE_DIMENSIONS |
Error parsing sparse dimensions field. |
INVALID_TOKEN_VALUE |
Token restrict value is invalid. |
INVALID_SPARSE_EMBEDDING |
Invalid sparse embedding. |
INVALID_EMBEDDING |
Invalid dense embedding. |
NearestNeighbors
Nearest neighbors for one query.
Fields | |
---|---|
neighbors[] |
All its neighbors. |
Neighbor
A neighbor of the query vector.
Fields | |
---|---|
entity_ |
The id of the similar entity. |
distance |
The distance between the neighbor and the query vector. |
entity_ |
The attributes of the neighbor, e.g. filters, crowding and metadata Note that full entities are returned only when "return_full_entity" is set to true. Otherwise, only the "entity_id" and "distance" fields are populated. |
Neighbor
Neighbors for example-based explanations.
Fields | |
---|---|
neighbor_ |
Output only. The neighbor id. |
neighbor_ |
Output only. The neighbor distance. |
NetworkSpec
Network spec.
Fields | |
---|---|
enable_ |
Whether to enable public internet access. Default false. |
network |
The full name of the Google Compute Engine network |
subnetwork |
The name of the subnet that this instance is in. Format: |
NfsMount
Represents a mount configuration for Network File System (NFS) to mount.
Fields | |
---|---|
server |
Required. IP address of the NFS server. |
path |
Required. Source path exported from NFS server. Has to start with '/', and combined with the ip address, it indicates the source mount path in the form of |
mount_ |
Required. Destination mount path. The NFS will be mounted for the user under /mnt/nfs/ |
NotebookEucConfig
The euc configuration of NotebookRuntimeTemplate.
Fields | |
---|---|
euc_ |
Input only. Whether EUC is disabled in this NotebookRuntimeTemplate. In proto3, the default value of a boolean is false. In this way, by default EUC will be enabled for NotebookRuntimeTemplate. |
bypass_ |
Output only. Whether ActAs check is bypassed for service account attached to the VM. If false, we need ActAs check for the default Compute Engine Service account. When a Runtime is created, a VM is allocated using Default Compute Engine Service Account. Any user requesting to use this Runtime requires Service Account User (ActAs) permission over this SA. If true, Runtime owner is using EUC and does not require the above permission as VM no longer use default Compute Engine SA, but a P4SA. |
NotebookExecutionJob
NotebookExecutionJob represents an instance of a notebook execution.
Fields | |
---|---|
name |
Output only. The resource name of this NotebookExecutionJob. Format: |
display_ |
The display name of the NotebookExecutionJob. The name can be up to 128 characters long and can consist of any UTF-8 characters. |
execution_ |
Max running time of the execution job in seconds (default 86400s / 24 hrs). |
schedule_ |
Output only. The Schedule resource name if this job is triggered by one. Format: |
job_ |
Output only. The state of the NotebookExecutionJob. |
status |
Output only. Populated when the NotebookExecutionJob is completed. When there is an error during notebook execution, the error details are populated. |
create_ |
Output only. Timestamp when this NotebookExecutionJob was created. |
update_ |
Output only. Timestamp when this NotebookExecutionJob was most recently updated. |
labels |
The labels with user-defined metadata to organize NotebookExecutionJobs. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. |
kernel_ |
The name of the kernel to use during notebook execution. If unset, the default kernel is used. |
encryption_ |
Customer-managed encryption key spec for the notebook execution job. This field is auto-populated if the |
Union field notebook_source . The input notebook. notebook_source can be only one of the following: |
|
dataform_ |
The Dataform Repository pointing to a single file notebook repository. |
gcs_ |
The Cloud Storage url pointing to the ipynb file. Format: |
direct_ |
The contents of an input notebook file. |
Union field environment_spec . The compute config to use for an execution job. environment_spec can be only one of the following: |
|
notebook_ |
The NotebookRuntimeTemplate to source compute configuration from. |
custom_ |
The custom compute configuration for an execution job. |
Union field execution_sink . The location to store the notebook execution result. execution_sink can be only one of the following: |
|
gcs_ |
The Cloud Storage location to upload the result to. Format: |
Union field execution_identity . The identity to run the execution as. execution_identity can be only one of the following: |
|
execution_ |
The user email to run the execution as. Only supported by Colab runtimes. |
service_ |
The service account to run the execution as. |
Union field runtime_environment . Runtime environment for the notebook execution job. If unspecified, the default runtime of Colab is used. runtime_environment can be only one of the following: |
|
workbench_ |
The Workbench runtime configuration to use for the notebook execution. |
CustomEnvironmentSpec
Compute configuration to use for an execution job.
Fields | |
---|---|
machine_ |
The specification of a single machine for the execution job. |
persistent_ |
The specification of a persistent disk to attach for the execution job. |
network_ |
The network configuration to use for the execution job. |
DataformRepositorySource
The Dataform Repository containing the input notebook.
Fields | |
---|---|
dataform_ |
The resource name of the Dataform Repository. Format: |
commit_ |
The commit SHA to read repository with. If unset, the file will be read at HEAD. |
DirectNotebookSource
The content of the input notebook in ipynb format.
Fields | |
---|---|
content |
The base64-encoded contents of the input notebook file. |
GcsNotebookSource
The Cloud Storage uri for the input notebook.
Fields | |
---|---|
uri |
The Cloud Storage uri pointing to the ipynb file. Format: |
generation |
The version of the Cloud Storage object to read. If unset, the current version of the object is read. See https://cloud.google.com/storage/docs/metadata#generation-number. |
WorkbenchRuntime
This type has no fields.
Configuration for a Workbench Instances-based environment.
NotebookExecutionJobView
Views for Get/List NotebookExecutionJob
Enums | |
---|---|
NOTEBOOK_EXECUTION_JOB_VIEW_UNSPECIFIED |
When unspecified, the API defaults to the BASIC view. |
NOTEBOOK_EXECUTION_JOB_VIEW_BASIC |
Includes all fields except for direct notebook inputs. |
NOTEBOOK_EXECUTION_JOB_VIEW_FULL |
Includes all fields. |
NotebookIdleShutdownConfig
The idle shutdown configuration of NotebookRuntimeTemplate, which contains the idle_timeout as required field.
Fields | |
---|---|
idle_ |
Required. Duration is accurate to the second. In Notebook, Idle Timeout is accurate to minute so the range of idle_timeout (second) is: 10 * 60 ~ 1440 * 60. |
idle_ |
Whether Idle Shutdown is disabled in this NotebookRuntimeTemplate. |
NotebookRuntime
A runtime is a virtual machine allocated to a particular user for a particular Notebook file on temporary basis with lifetime limited to 24 hours.
Fields | |
---|---|
name |
Output only. The resource name of the NotebookRuntime. |
runtime_ |
Required. The user email of the NotebookRuntime. |
notebook_ |
Output only. The pointer to NotebookRuntimeTemplate this NotebookRuntime is created from. |
proxy_ |
Output only. The proxy endpoint used to access the NotebookRuntime. |
create_ |
Output only. Timestamp when this NotebookRuntime was created. |
update_ |
Output only. Timestamp when this NotebookRuntime was most recently updated. |
health_ |
Output only. The health state of the NotebookRuntime. |
display_ |
Required. The display name of the NotebookRuntime. The name can be up to 128 characters long and can consist of any UTF-8 characters. |
description |
The description of the NotebookRuntime. |
service_ |
Output only. Deprecated: This field is no longer used and the "Vertex AI Notebook Service Account" (service-PROJECT_NUMBER@gcp-sa-aiplatform-vm.iam.gserviceaccount.com) is used for the runtime workload identity. See https://cloud.google.com/iam/docs/service-agents#vertex-ai-notebook-service-account for more details. The service account that the NotebookRuntime workload runs as. |
runtime_ |
Output only. The runtime (instance) state of the NotebookRuntime. |
is_ |
Output only. Whether NotebookRuntime is upgradable. |
labels |
The labels with user-defined metadata to organize your NotebookRuntime. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one NotebookRuntime (System labels are excluded). See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. Following system labels exist for NotebookRuntime:
|
expiration_ |
Output only. Timestamp when this NotebookRuntime will be expired: 1. System Predefined NotebookRuntime: 24 hours after creation. After expiration, system predifined runtime will be deleted. 2. User created NotebookRuntime: 6 months after last upgrade. After expiration, user created runtime will be stopped and allowed for upgrade. |
version |
Output only. The VM os image version of NotebookRuntime. |
notebook_ |
Output only. The type of the notebook runtime. |
idle_ |
Output only. The idle shutdown configuration of the notebook runtime. |
euc_ |
Output only. EUC configuration of the notebook runtime. |
shielded_ |
Output only. Runtime Shielded VM spec. |
network_ |
Optional. The Compute Engine tags to add to runtime (see Tagging instances). |
encryption_ |
Output only. Customer-managed encryption key spec for the notebook runtime. |
satisfies_ |
Output only. Reserved for future use. |
satisfies_ |
Output only. Reserved for future use. |
HealthState
The substate of the NotebookRuntime to display health information.
Enums | |
---|---|
HEALTH_STATE_UNSPECIFIED |
Unspecified health state. |
HEALTHY |
NotebookRuntime is in healthy state. Applies to ACTIVE state. |
UNHEALTHY |
NotebookRuntime is in unhealthy state. Applies to ACTIVE state. |
RuntimeState
The substate of the NotebookRuntime to display state of runtime. The resource of NotebookRuntime is in ACTIVE state for these sub state.
Enums | |
---|---|
RUNTIME_STATE_UNSPECIFIED |
Unspecified runtime state. |
RUNNING |
NotebookRuntime is in running state. |
BEING_STARTED |
NotebookRuntime is in starting state. |
BEING_STOPPED |
NotebookRuntime is in stopping state. |
STOPPED |
NotebookRuntime is in stopped state. |
BEING_UPGRADED |
NotebookRuntime is in upgrading state. It is in the middle of upgrading process. |
ERROR |
NotebookRuntime was unable to start/stop properly. |
INVALID |
NotebookRuntime is in invalid state. Cannot be recovered. |
NotebookRuntimeTemplate
A template that specifies runtime configurations such as machine type, runtime version, network configurations, etc. Multiple runtimes can be created from a runtime template.
Fields | |
---|---|
name |
The resource name of the NotebookRuntimeTemplate. |
display_ |
Required. The display name of the NotebookRuntimeTemplate. The name can be up to 128 characters long and can consist of any UTF-8 characters. |
description |
The description of the NotebookRuntimeTemplate. |
is_default |
Output only. Deprecated: This field has no behavior. Use notebook_runtime_type = 'ONE_CLICK' instead. The default template to use if not specified. |
machine_ |
Optional. Immutable. The specification of a single machine for the template. |
data_ |
Optional. The specification of [persistent disk][https://cloud.google.com/compute/docs/disks/persistent-disks] attached to the runtime as data disk storage. |
network_ |
Optional. Network spec. |
service_account |
Deprecated: This field is ignored and the "Vertex AI Notebook Service Account" (service-PROJECT_NUMBER@gcp-sa-aiplatform-vm.iam.gserviceaccount.com) is used for the runtime workload identity. See https://cloud.google.com/iam/docs/service-agents#vertex-ai-notebook-service-account for more details. For NotebookExecutionJob, use NotebookExecutionJob.service_account instead. The service account that the runtime workload runs as. You can use any service account within the same project, but you must have the service account user permission to use the instance. If not specified, the Compute Engine default service account is used. |
etag |
Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
labels |
The labels with user-defined metadata to organize the NotebookRuntimeTemplates. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. |
idle_ |
The idle shutdown configuration of NotebookRuntimeTemplate. This config will only be set when idle shutdown is enabled. |
euc_ |
EUC configuration of the NotebookRuntimeTemplate. |
create_ |
Output only. Timestamp when this NotebookRuntimeTemplate was created. |
update_ |
Output only. Timestamp when this NotebookRuntimeTemplate was most recently updated. |
notebook_ |
Optional. Immutable. The type of the notebook runtime template. |
shielded_ |
Optional. Immutable. Runtime Shielded VM spec. |
network_ |
Optional. The Compute Engine tags to add to runtime (see Tagging instances). |
encryption_ |
Customer-managed encryption key spec for the notebook runtime. |
NotebookRuntimeTemplateRef
Points to a NotebookRuntimeTemplateRef.
Fields | |
---|---|
notebook_ |
Immutable. A resource name of the NotebookRuntimeTemplate. |
NotebookRuntimeType
Represents a notebook runtime type.
Enums | |
---|---|
NOTEBOOK_RUNTIME_TYPE_UNSPECIFIED |
Unspecified notebook runtime type, NotebookRuntimeType will default to USER_DEFINED. |
USER_DEFINED |
runtime or template with coustomized configurations from user. |
ONE_CLICK |
runtime or template with system defined configurations. |
PSCAutomationConfig
PSC config that is used to automatically create forwarding rule via ServiceConnectionMap.
Fields | |
---|---|
project_ |
Required. Project id used to create forwarding rule. |
network |
Required. The full name of the Google Compute Engine network. Format: |
PairwiseChoice
Pairwise prediction autorater preference.
Enums | |
---|---|
PAIRWISE_CHOICE_UNSPECIFIED |
Unspecified prediction choice. |
BASELINE |
Baseline prediction wins |
CANDIDATE |
Candidate prediction wins |
TIE |
Winner cannot be determined |
PairwiseMetricInput
Input for pairwise metric.
Fields | |
---|---|
metric_ |
Required. Spec for pairwise metric. |
instance |
Required. Pairwise metric instance. |
PairwiseMetricInstance
Pairwise metric instance. Usually one instance corresponds to one row in an evaluation dataset.
Fields | |
---|---|
Union field instance . Instance for pairwise metric. instance can be only one of the following: |
|
json_ |
Instance specified as a json string. String key-value pairs are expected in the json_instance to render PairwiseMetricSpec.instance_prompt_template. |
PairwiseMetricResult
Spec for pairwise metric result.
Fields | |
---|---|
pairwise_ |
Output only. Pairwise metric choice. |
explanation |
Output only. Explanation for pairwise metric score. |
PairwiseMetricSpec
Spec for pairwise metric.
Fields | |
---|---|
metric_ |
Required. Metric prompt template for pairwise metric. |
PairwiseQuestionAnsweringQualityInput
Input for pairwise question answering quality metric.
Fields | |
---|---|
metric_ |
Required. Spec for pairwise question answering quality score metric. |
instance |
Required. Pairwise question answering quality instance. |
PairwiseQuestionAnsweringQualityInstance
Spec for pairwise question answering quality instance.
Fields | |
---|---|
prediction |
Required. Output of the candidate model. |
baseline_ |
Required. Output of the baseline model. |
reference |
Optional. Ground truth used to compare against the prediction. |
context |
Required. Text to answer the question. |
instruction |
Required. Question Answering prompt for LLM. |
PairwiseQuestionAnsweringQualityResult
Spec for pairwise question answering quality result.
Fields | |
---|---|
pairwise_ |
Output only. Pairwise question answering prediction choice. |
explanation |
Output only. Explanation for question answering quality score. |
confidence |
Output only. Confidence for question answering quality score. |
PairwiseQuestionAnsweringQualitySpec
Spec for pairwise question answering quality score metric.
Fields | |
---|---|
use_ |
Optional. Whether to use instance.reference to compute question answering quality. |
version |
Optional. Which version to use for evaluation. |
PairwiseSummarizationQualityInput
Input for pairwise summarization quality metric.
Fields | |
---|---|
metric_ |
Required. Spec for pairwise summarization quality score metric. |
instance |
Required. Pairwise summarization quality instance. |
PairwiseSummarizationQualityInstance
Spec for pairwise summarization quality instance.
Fields | |
---|---|
prediction |
Required. Output of the candidate model. |
baseline_ |
Required. Output of the baseline model. |
reference |
Optional. Ground truth used to compare against the prediction. |
context |
Required. Text to be summarized. |
instruction |
Required. Summarization prompt for LLM. |
PairwiseSummarizationQualityResult
Spec for pairwise summarization quality result.
Fields | |
---|---|
pairwise_ |
Output only. Pairwise summarization prediction choice. |
explanation |
Output only. Explanation for summarization quality score. |
confidence |
Output only. Confidence for summarization quality score. |
PairwiseSummarizationQualitySpec
Spec for pairwise summarization quality score metric.
Fields | |
---|---|
use_ |
Optional. Whether to use instance.reference to compute pairwise summarization quality. |
version |
Optional. Which version to use for evaluation. |
Part
A datatype containing media that is part of a multi-part Content
message.
A Part
consists of data which has an associated datatype. A Part
can only contain one of the accepted types in Part.data
.
A Part
must have a fixed IANA MIME type identifying the type and subtype of the media if inline_data
or file_data
field is filled with raw bytes.
Fields | |
---|---|
Union field
|
|
text |
Optional. Text part (can be code). |
inline_ |
Optional. Inlined bytes data. |
file_ |
Optional. URI based data. |
function_ |
Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. |
function_ |
Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. |
Union field
|
|
video_ |
Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. |
PauseModelDeploymentMonitoringJobRequest
Request message for JobService.PauseModelDeploymentMonitoringJob
.
Fields | |
---|---|
name |
Required. The resource name of the ModelDeploymentMonitoringJob to pause. Format: |
PauseScheduleRequest
Request message for ScheduleService.PauseSchedule
.
Fields | |
---|---|
name |
Required. The name of the Schedule resource to be paused. Format: |
PersistentDiskSpec
Represents the spec of [persistent disk][https://cloud.google.com/compute/docs/disks/persistent-disks] options.
Fields | |
---|---|
disk_ |
Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) |
disk_ |
Size in GB of the disk (default is 100GB). |
PersistentResource
Represents long-lasting resources that are dedicated to users to runs custom workloads. A PersistentResource can have multiple node pools and each node pool can have its own machine spec.
Fields | |
---|---|
name |
Immutable. Resource name of a PersistentResource. |
display_ |
Optional. The display name of the PersistentResource. The name can be up to 128 characters long and can consist of any UTF-8 characters. |
resource_ |
Required. The spec of the pools of different resources. |
state |
Output only. The detailed state of a Study. |
error |
Output only. Only populated when persistent resource's state is |
create_ |
Output only. Time when the PersistentResource was created. |
start_ |
Output only. Time when the PersistentResource for the first time entered the |
update_ |
Output only. Time when the PersistentResource was most recently updated. |
labels |
Optional. The labels with user-defined metadata to organize PersistentResource. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. |
network |
Optional. The full name of the Compute Engine network to peered with Vertex AI to host the persistent resources. For example, To specify this field, you must have already configured VPC Network Peering for Vertex AI. If this field is left unspecified, the resources aren't peered with any network. |
encryption_ |
Optional. Customer-managed encryption key spec for a PersistentResource. If set, this PersistentResource and all sub-resources of this PersistentResource will be secured by this key. |
resource_ |
Optional. Persistent Resource runtime spec. For example, used for Ray cluster configuration. |
resource_ |
Output only. Runtime information of the Persistent Resource. |
reserved_ |
Optional. A list of names for the reserved IP ranges under the VPC network that can be used for this persistent resource. If set, we will deploy the persistent resource within the provided IP ranges. Otherwise, the persistent resource is deployed to any IP ranges under the provided VPC network. Example: ['vertex-ai-ip-range']. |
satisfies_ |
Output only. Reserved for future use. |
satisfies_ |
Output only. Reserved for future use. |
State
Describes the PersistentResource state.
Enums | |
---|---|
STATE_UNSPECIFIED |
Not set. |
PROVISIONING |
The PROVISIONING state indicates the persistent resources is being created. |
RUNNING |
The RUNNING state indicates the persistent resource is healthy and fully usable. |
STOPPING |
The STOPPING state indicates the persistent resource is being deleted. |
ERROR |
The ERROR state indicates the persistent resource may be unusable. Details can be found in the error field. |
REBOOTING |
The REBOOTING state indicates the persistent resource is being rebooted (PR is not available right now but is expected to be ready again later). |
UPDATING |
The UPDATING state indicates the persistent resource is being updated. |
PipelineFailurePolicy
Represents the failure policy of a pipeline. Currently, the default of a pipeline is that the pipeline will continue to run until no more tasks can be executed, also known as PIPELINE_FAILURE_POLICY_FAIL_SLOW. However, if a pipeline is set to PIPELINE_FAILURE_POLICY_FAIL_FAST, it will stop scheduling any new tasks when a task has failed. Any scheduled tasks will continue to completion.
Enums | |
---|---|
PIPELINE_FAILURE_POLICY_UNSPECIFIED |
Default value, and follows fail slow behavior. |
PIPELINE_FAILURE_POLICY_FAIL_SLOW |
Indicates that the pipeline should continue to run until all possible tasks have been scheduled and completed. |
PIPELINE_FAILURE_POLICY_FAIL_FAST |
Indicates that the pipeline should stop scheduling new tasks after a task has failed. |
PipelineJob
An instance of a machine learning PipelineJob.
Fields | |
---|---|
name |
Output only. The resource name of the PipelineJob. |
display_ |
The display name of the Pipeline. The name can be up to 128 characters long and can consist of any UTF-8 characters. |
create_ |
Output only. Pipeline creation time. |
start_ |
Output only. Pipeline start time. |
end_ |
Output only. Pipeline end time. |
update_ |
Output only. Timestamp when this PipelineJob was most recently updated. |
pipeline_ |
The spec of the pipeline. |
state |
Output only. The detailed state of the job. |
job_ |
Output only. The details of pipeline run. Not available in the list view. |
error |
Output only. The error that occurred during pipeline execution. Only populated when the pipeline's state is FAILED or CANCELLED. |
labels |
The labels with user-defined metadata to organize PipelineJob. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. Note there is some reserved label key for Vertex AI Pipelines. - |
runtime_ |
Runtime config of the pipeline. |
encryption_ |
Customer-managed encryption key spec for a pipelineJob. If set, this PipelineJob and all of its sub-resources will be secured by this key. |
service_ |
The service account that the pipeline workload runs as. If not specified, the Compute Engine default service account in the project will be used. See https://cloud.google.com/compute/docs/access/service-accounts#default_service_account Users starting the pipeline must have the |
network |
The full name of the Compute Engine network to which the Pipeline Job's workload should be peered. For example, Private services access must already be configured for the network. Pipeline job will apply the network configuration to the Google Cloud resources being launched, if applied, such as Vertex AI Training or Dataflow job. If left unspecified, the workload is not peered with any network. |
reserved_ |
A list of names for the reserved ip ranges under the VPC network that can be used for this Pipeline Job's workload. If set, we will deploy the Pipeline Job's workload within the provided ip ranges. Otherwise, the job will be deployed to any ip ranges under the provided VPC network. Example: ['vertex-ai-ip-range']. |
template_ |
A template uri from where the |
template_ |
Output only. Pipeline template metadata. Will fill up fields if |
schedule_ |
Output only. The schedule resource name. Only returned if the Pipeline is created by Schedule API. |
preflight_ |
Optional. Whether to do component level validations before job creation. |
RuntimeConfig
The runtime config of a PipelineJob.
Fields | |
---|---|
parameters |
Deprecated. Use |
gcs_ |
Required. A path in a Cloud Storage bucket, which will be treated as the root output directory of the pipeline. It is used by the system to generate the paths of output artifacts. The artifact paths are generated with a sub-path pattern |
parameter_ |
The runtime parameters of the PipelineJob. The parameters will be passed into |
failure_ |
Represents the failure policy of a pipeline. Currently, the default of a pipeline is that the pipeline will continue to run until no more tasks can be executed, also known as PIPELINE_FAILURE_POLICY_FAIL_SLOW. However, if a pipeline is set to PIPELINE_FAILURE_POLICY_FAIL_FAST, it will stop scheduling any new tasks when a task has failed. Any scheduled tasks will continue to completion. |
input_ |
The runtime artifacts of the PipelineJob. The key will be the input artifact name and the value would be one of the InputArtifact. |
InputArtifact
The type of an input artifact.
Fields | |
---|---|
Union field
|
|
artifact_ |
Artifact resource id from MLMD. Which is the last portion of an artifact resource name: |
PipelineJobDetail
The runtime detail of PipelineJob.
Fields | |
---|---|
pipeline_ |
Output only. The context of the pipeline. |
pipeline_ |
Output only. The context of the current pipeline run. |
task_ |
Output only. The runtime details of the tasks under the pipeline. |
PipelineState
Describes the state of a pipeline.
Enums | |
---|---|
PIPELINE_STATE_UNSPECIFIED |
The pipeline state is unspecified. |
PIPELINE_STATE_QUEUED |
The pipeline has been created or resumed, and processing has not yet begun. |
PIPELINE_STATE_PENDING |
The service is preparing to run the pipeline. |
PIPELINE_STATE_RUNNING |
The pipeline is in progress. |
PIPELINE_STATE_SUCCEEDED |
The pipeline completed successfully. |
PIPELINE_STATE_FAILED |
The pipeline failed. |
PIPELINE_STATE_CANCELLING |
The pipeline is being cancelled. From this state, the pipeline may only go to either PIPELINE_STATE_SUCCEEDED, PIPELINE_STATE_FAILED or PIPELINE_STATE_CANCELLED. |
PIPELINE_STATE_CANCELLED |
The pipeline has been cancelled. |
PIPELINE_STATE_PAUSED |
The pipeline has been stopped, and can be resumed. |
PipelineTaskDetail
The runtime detail of a task execution.
Fields | |
---|---|
task_ |
Output only. The system generated ID of the task. |
parent_ |
Output only. The id of the parent task if the task is within a component scope. Empty if the task is at the root level. |
task_ |
Output only. The user specified name of the task that is defined in |
create_ |
Output only. Task create time. |
start_ |
Output only. Task start time. |
end_ |
Output only. Task end time. |
executor_ |
Output only. The detailed execution info. |
state |
Output only. State of the task. |
execution |
Output only. The execution metadata of the task. |
error |
Output only. The error that occurred during task execution. Only populated when the task's state is FAILED or CANCELLED. |
pipeline_ |
Output only. A list of task status. This field keeps a record of task status evolving over time. |
inputs |
Output only. The runtime input artifacts of the task. |
outputs |
Output only. The runtime output artifacts of the task. |
ArtifactList
A list of artifact metadata.
Fields | |
---|---|
artifacts[] |
Output only. A list of artifact metadata. |
PipelineTaskStatus
A single record of the task status.
Fields | |
---|---|
update_ |
Output only. Update time of this status. |
state |
Output only. The state of the task. |
error |
Output only. The error that occurred during the state. May be set when the state is any of the non-final state (PENDING/RUNNING/CANCELLING) or FAILED state. If the state is FAILED, the error here is final and not going to be retried. If the state is a non-final state, the error indicates a system-error being retried. |
State
Specifies state of TaskExecution
Enums | |
---|---|
STATE_UNSPECIFIED |
Unspecified. |
PENDING |
Specifies pending state for the task. |
RUNNING |
Specifies task is being executed. |
SUCCEEDED |
Specifies task completed successfully. |
CANCEL_PENDING |
Specifies Task cancel is in pending state. |
CANCELLING |
Specifies task is being cancelled. |
CANCELLED |
Specifies task was cancelled. |
FAILED |
Specifies task failed. |
SKIPPED |
Specifies task was skipped due to cache hit. |
NOT_TRIGGERED |
Specifies that the task was not triggered because the task's trigger policy is not satisfied. The trigger policy is specified in the condition field of PipelineJob.pipeline_spec . |
PipelineTaskExecutorDetail
The runtime detail of a pipeline executor.
Fields | |
---|---|
Union field
|
|
container_ |
Output only. The detailed info for a container executor. |
custom_ |
Output only. The detailed info for a custom job executor. |
ContainerDetail
The detail of a container execution. It contains the job names of the lifecycle of a container execution.
Fields | |
---|---|
main_ |
Output only. The name of the |
pre_ |
Output only. The name of the |
failed_ |
Output only. The names of the previously failed |
failed_ |
Output only. The names of the previously failed |
CustomJobDetail
The detailed info for a custom job executor.
Fields | |
---|---|
job |
Output only. The name of the |
failed_ |
Output only. The names of the previously failed |
PipelineTemplateMetadata
Pipeline template metadata if PipelineJob.template_uri
is from supported template registry. Currently, the only supported registry is Artifact Registry.
Fields | |
---|---|
version |
The version_name in artifact registry. Will always be presented in output if the Format is "sha256:abcdef123456...". |
PointwiseMetricInput
Input for pointwise metric.
Fields | |
---|---|
metric_ |
Required. Spec for pointwise metric. |
instance |
Required. Pointwise metric instance. |
PointwiseMetricInstance
Pointwise metric instance. Usually one instance corresponds to one row in an evaluation dataset.
Fields | |
---|---|
Union field instance . Instance for pointwise metric. instance can be only one of the following: |
|
json_ |
Instance specified as a json string. String key-value pairs are expected in the json_instance to render PointwiseMetricSpec.instance_prompt_template. |
PointwiseMetricResult
Spec for pointwise metric result.
Fields | |
---|---|
explanation |
Output only. Explanation for pointwise metric score. |
score |
Output only. Pointwise metric score. |
PointwiseMetricSpec
Spec for pointwise metric.
Fields | |
---|---|
metric_ |
Required. Metric prompt template for pointwise metric. |
Port
Represents a network port in a container.
Fields | |
---|---|
container_ |
The number of the port to expose on the pod's IP address. Must be a valid port number, between 1 and 65535 inclusive. |
PredefinedSplit
Assigns input data to training, validation, and test sets based on the value of a provided key.
Supported only for tabular Datasets.
Fields | |
---|---|
key |
Required. The key is a name of one of the Dataset's data columns. The value of the key (either the label's value or value in the column) must be one of { |
PredictLongRunningRequest
Request message for PredictionService.PredictLongRunning
.
Fields | |
---|---|
endpoint |
Required. The name of the Endpoint requested to serve the prediction. Format: |
instances[] |
Required. The instances that are the input to the prediction call. A DeployedModel may have an upper limit on the number of instances it supports per request, and when it is exceeded the prediction call errors in case of AutoML Models, or, in case of customer created Models, the behaviour is as documented by that Model. The schema of any single instance may be specified via Endpoint's DeployedModels' |
parameters |
Optional. The parameters that govern the prediction. The schema of the parameters may be specified via Endpoint's DeployedModels' |
PredictRequest
Request message for PredictionService.Predict
.
Fields | |
---|---|
endpoint |
Required. The name of the Endpoint requested to serve the prediction. Format: |
instances[] |
Required. The instances that are the input to the prediction call. A DeployedModel may have an upper limit on the number of instances it supports per request, and when it is exceeded the prediction call errors in case of AutoML Models, or, in case of customer created Models, the behaviour is as documented by that Model. The schema of any single instance may be specified via Endpoint's DeployedModels' |
parameters |
The parameters that govern the prediction. The schema of the parameters may be specified via Endpoint's DeployedModels' |
PredictRequestResponseLoggingConfig
Configuration for logging request-response to a BigQuery table.
Fields | |
---|---|
enabled |
If logging is enabled or not. |
sampling_ |
Percentage of requests to be logged, expressed as a fraction in range(0,1]. |
bigquery_ |
BigQuery table for logging. If only given a project, a new dataset will be created with name |
PredictResponse
Response message for PredictionService.Predict
.
Fields | |
---|---|
predictions[] |
The predictions that are the output of the predictions call. The schema of any single prediction may be specified via Endpoint's DeployedModels' |
deployed_ |
ID of the Endpoint's DeployedModel that served this prediction. |
model |
Output only. The resource name of the Model which is deployed as the DeployedModel that this prediction hits. |
model_ |
Output only. The version ID of the Model which is deployed as the DeployedModel that this prediction hits. |
model_ |
Output only. The |
metadata |
Output only. Request-level metadata returned by the model. The metadata type will be dependent upon the model implementation. |
PredictSchemata
Contains the schemata used in Model's predictions and explanations via PredictionService.Predict
, PredictionService.Explain
and BatchPredictionJob
.
Fields | |
---|---|
instance_ |
Immutable. Points to a YAML file stored on Google Cloud Storage describing the format of a single instance, which are used in |
parameters_ |
Immutable. Points to a YAML file stored on Google Cloud Storage describing the parameters of prediction and explanation via |
prediction_ |
Immutable. Points to a YAML file stored on Google Cloud Storage describing the format of a single prediction produced by this Model, which are returned via |
Presets
Preset configuration for example-based explanations
Fields | |
---|---|
modality |
The modality of the uploaded model, which automatically configures the distance measurement and feature normalization for the underlying example index and queries. If your model does not precisely fit one of these types, it is okay to choose the closest type. |
query |
Preset option controlling parameters for speed-precision trade-off when querying for examples. If omitted, defaults to |
Modality
Preset option controlling parameters for different modalities
Enums | |
---|---|
MODALITY_UNSPECIFIED |
Should not be set. Added as a recommended best practice for enums |
IMAGE |
IMAGE modality |
TEXT |
TEXT modality |
TABULAR |
TABULAR modality |
Query
Preset option controlling parameters for query speed-precision trade-off
Enums | |
---|---|
PRECISE |
More precise neighbors as a trade-off against slower response. |
FAST |
Faster response as a trade-off against less precise neighbors. |
PrivateEndpoints
PrivateEndpoints proto is used to provide paths for users to send requests privately. To send request via private service access, use predict_http_uri, explain_http_uri or health_http_uri. To send request via private service connect, use service_attachment.
Fields | |
---|---|
predict_ |
Output only. Http(s) path to send prediction requests. |
explain_ |
Output only. Http(s) path to send explain requests. |
health_ |
Output only. Http(s) path to send health check requests. |
service_ |
Output only. The name of the service attachment resource. Populated if private service connect is enabled. |
PrivateServiceConnectConfig
Represents configuration for private service connect.
Fields | |
---|---|
enable_ |
Required. If true, expose the IndexEndpoint via private service connect. |
project_ |
A list of Projects from which the forwarding rule will target the service attachment. |
service_ |
Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect. |
Probe
Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic.
Fields | |
---|---|
period_ |
How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1. Must be less than timeout_seconds. Maps to Kubernetes probe argument 'periodSeconds'. |
timeout_ |
Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. Must be greater or equal to period_seconds. Maps to Kubernetes probe argument 'timeoutSeconds'. |
Union field
|
|
exec |
ExecAction probes the health of a container by executing a command. |
ExecAction
ExecAction specifies a command to execute.
Fields | |
---|---|
command[] |
Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. |
PscAutomatedEndpoints
PscAutomatedEndpoints defines the output of the forwarding rule automatically created by each PscAutomationConfig.
Fields | |
---|---|
project_ |
Corresponding project_id in pscAutomationConfigs |
network |
Corresponding network in pscAutomationConfigs. |
match_ |
Ip Address created by the automated forwarding rule. |
PublisherModel
A Model Garden Publisher Model.
Fields | |
---|---|
name |
Output only. The resource name of the PublisherModel. |
version_ |
Output only. Immutable. The version ID of the PublisherModel. A new version is committed when a new model version is uploaded under an existing model id. It is an auto-incrementing decimal number in string representation. |
open_ |
Required. Indicates the open source category of the publisher model. |
supported_ |
Optional. Supported call-to-action options. |
frameworks[] |
Optional. Additional information about the model's Frameworks. |
launch_ |
Optional. Indicates the launch stage of the model. |
version_ |
Optional. Indicates the state of the model version. |
publisher_ |
Optional. Output only. Immutable. Used to indicate this model has a publisher model and provide the template of the publisher model resource name. |
predict_ |
Optional. The schemata that describes formats of the PublisherModel's predictions and explanations as given and returned via |
CallToAction
Actions could take on this Publisher Model.
Fields | |
---|---|
view_ |
Optional. To view Rest API docs. |
open_ |
Optional. Open notebook of the PublisherModel. |
create_ |
Optional. Create application using the PublisherModel. |
open_ |
Optional. Open fine-tuning pipeline of the PublisherModel. |
open_ |
Optional. Open prompt-tuning pipeline of the PublisherModel. |
open_ |
Optional. Open Genie / Playground. |
deploy |
Optional. Deploy the PublisherModel to Vertex Endpoint. |
deploy_ |
Optional. Deploy PublisherModel to Google Kubernetes Engine. |
open_ |
Optional. Open in Generation AI Studio. |
request_ |
Optional. Request for access. |
open_ |
Optional. Open evaluation pipeline of the PublisherModel. |
open_ |
Optional. Open notebooks of the PublisherModel. |
open_ |
Optional. Open fine-tuning pipelines of the PublisherModel. |
Deploy
Model metadata that is needed for UploadModel or DeployModel/CreateEndpoint requests.
Fields | |
---|---|
model_ |
Optional. Default model display name. |
large_ |
Optional. Large model reference. When this is set, model_artifact_spec is not needed. |
container_ |
Optional. The specification of the container that is to be used when deploying this Model in Vertex AI. Not present for Large Models. |
artifact_ |
Optional. The path to the directory containing the Model artifact and any of its supporting files. |
title |
Required. The title of the regional resource reference. |
public_ |
Optional. The signed URI for ephemeral Cloud Storage access to model artifact. |
Union field prediction_resources . The prediction (for example, the machine) resources that the DeployedModel uses. prediction_resources can be only one of the following: |
|
dedicated_ |
A description of resources that are dedicated to the DeployedModel, and that need a higher degree of manual configuration. |
automatic_ |
A description of resources that to large degree are decided by Vertex AI, and require only a modest additional configuration. |
shared_ |
The resource name of the shared DeploymentResourcePool to deploy on. Format: |
deploy_ |
Optional. The name of the deploy task (e.g., "text to image generation"). |
deploy_ |
Optional. Metadata information about this deployment config. |
DeployMetadata
Metadata information about the deployment for managing deployment config.
Fields | |
---|---|
labels |
Optional. Labels for the deployment config. For managing deployment config like verifying, source of deployment config, etc. |
sample_ |
Optional. Sample request for deployed endpoint. |
DeployGke
Configurations for PublisherModel GKE deployment
Fields | |
---|---|
gke_ |
Optional. GKE deployment configuration in yaml format. |
OpenFineTuningPipelines
Open fine tuning pipelines.
Fields | |
---|---|
fine_ |
Required. Regional resource references to fine tuning pipelines. |
OpenNotebooks
Open notebooks.
Fields | |
---|---|
notebooks[] |
Required. Regional resource references to notebooks. |
RegionalResourceReferences
The regional resource name or the URI. Key is region, e.g., us-central1, europe-west2, global, etc..
Fields | |
---|---|
references |
Required. |
title |
Required. |
resource_ |
Optional. Title of the resource. |
resource_ |
Optional. Use case (CUJ) of the resource. |
resource_ |
Optional. Description of the resource. |
ViewRestApi
Rest API docs.
Fields | |
---|---|
documentations[] |
Required. |
title |
Required. The title of the view rest API. |
Documentation
A named piece of documentation.
Fields | |
---|---|
title |
Required. E.g., OVERVIEW, USE CASES, DOCUMENTATION, SDK & SAMPLES, JAVA, NODE.JS, etc.. |
content |
Required. Content of this piece of document (in Markdown format). |
LaunchStage
An enum representing the launch stage of a PublisherModel.
Enums | |
---|---|
LAUNCH_STAGE_UNSPECIFIED |
The model launch stage is unspecified. |
EXPERIMENTAL |
Used to indicate the PublisherModel is at Experimental launch stage, available to a small set of customers. |
PRIVATE_PREVIEW |
Used to indicate the PublisherModel is at Private Preview launch stage, only available to a small set of customers, although a larger set of customers than an Experimental launch. Previews are the first launch stage used to get feedback from customers. |
PUBLIC_PREVIEW |
Used to indicate the PublisherModel is at Public Preview launch stage, available to all customers, although not supported for production workloads. |
GA |
Used to indicate the PublisherModel is at GA launch stage, available to all customers and ready for production workload. |
OpenSourceCategory
An enum representing the open source category of a PublisherModel.
Enums | |
---|---|
OPEN_SOURCE_CATEGORY_UNSPECIFIED |
The open source category is unspecified, which should not be used. |
PROPRIETARY |
Used to indicate the PublisherModel is not open sourced. |
GOOGLE_OWNED_OSS_WITH_GOOGLE_CHECKPOINT |
Used to indicate the PublisherModel is a Google-owned open source model w/ Google checkpoint. |
THIRD_PARTY_OWNED_OSS_WITH_GOOGLE_CHECKPOINT |
Used to indicate the PublisherModel is a 3p-owned open source model w/ Google checkpoint. |
GOOGLE_OWNED_OSS |
Used to indicate the PublisherModel is a Google-owned pure open source model. |
THIRD_PARTY_OWNED_OSS |
Used to indicate the PublisherModel is a 3p-owned pure open source model. |
ResourceReference
Reference to a resource.
Fields | |
---|---|
Union field
|
|
uri |
The URI of the resource. |
resource_ |
The resource name of the Google Cloud resource. |
use_case |
Use case (CUJ) of the resource. |
description |
Description of the resource. |
VersionState
An enum representing the state of the PublicModelVersion.
Enums | |
---|---|
VERSION_STATE_UNSPECIFIED |
The version state is unspecified. |
VERSION_STATE_STABLE |
Used to indicate the version is stable. |
VERSION_STATE_UNSTABLE |
Used to indicate the version is unstable. |
PublisherModelView
View enumeration of PublisherModel.
Enums | |
---|---|
PUBLISHER_MODEL_VIEW_UNSPECIFIED |
The default / unset value. The API will default to the BASIC view. |
PUBLISHER_MODEL_VIEW_BASIC |
Include basic metadata about the publisher model, but not the full contents. |
PUBLISHER_MODEL_VIEW_FULL |
Include everything. |
PUBLISHER_MODEL_VERSION_VIEW_BASIC |
Include: VersionId, ModelVersionExternalName, and SupportedActions. |
PurgeArtifactsMetadata
Details of operations that perform MetadataService.PurgeArtifacts
.
Fields | |
---|---|
generic_ |
Operation metadata for purging Artifacts. |
PurgeArtifactsRequest
Request message for MetadataService.PurgeArtifacts
.
Fields | |
---|---|
parent |
Required. The metadata store to purge Artifacts from. Format: |
filter |
Required. A required filter matching the Artifacts to be purged. E.g., |
force |
Optional. Flag to indicate to actually perform the purge. If |
PurgeArtifactsResponse
Response message for MetadataService.PurgeArtifacts
.
Fields | |
---|---|
purge_ |
The number of Artifacts that this request deleted (or, if |
purge_ |
A sample of the Artifact names that will be deleted. Only populated if |
PurgeContextsMetadata
Details of operations that perform MetadataService.PurgeContexts
.
Fields | |
---|---|
generic_ |
Operation metadata for purging Contexts. |
PurgeContextsRequest
Request message for MetadataService.PurgeContexts
.
Fields | |
---|---|
parent |
Required. The metadata store to purge Contexts from. Format: |
filter |
Required. A required filter matching the Contexts to be purged. E.g., |
force |
Optional. Flag to indicate to actually perform the purge. If |
PurgeContextsResponse
Response message for MetadataService.PurgeContexts
.
Fields | |
---|---|
purge_ |
The number of Contexts that this request deleted (or, if |
purge_ |
A sample of the Context names that will be deleted. Only populated if |
PurgeExecutionsMetadata
Details of operations that perform MetadataService.PurgeExecutions
.
Fields | |
---|---|
generic_ |
Operation metadata for purging Executions. |
PurgeExecutionsRequest
Request message for MetadataService.PurgeExecutions
.
Fields | |
---|---|
parent |
Required. The metadata store to purge Executions from. Format: |
filter |
Required. A required filter matching the Executions to be purged. E.g., |
force |
Optional. Flag to indicate to actually perform the purge. If |
PurgeExecutionsResponse
Response message for MetadataService.PurgeExecutions
.
Fields | |
---|---|
purge_ |
The number of Executions that this request deleted (or, if |
purge_ |
A sample of the Execution names that will be deleted. Only populated if |
PythonPackageSpec
The spec of a Python packaged code.
Fields | |
---|---|
executor_ |
Required. The URI of a container image in Artifact Registry that will run the provided Python package. Vertex AI provides a wide range of executor images with pre-installed packages to meet users' various use cases. See the list of pre-built containers for training. You must use an image from this list. |
package_ |
Required. The Google Cloud Storage location of the Python package files which are the training program and its dependent packages. The maximum number of package URIs is 100. Authorization requires the following IAM permission on the specified resource
|
python_ |
Required. The Python module name to run after installing the packages. |
args[] |
Command line arguments to be passed to the Python task. |
env[] |
Environment variables to be passed to the python module. Maximum limit is 100. |
QueryArtifactLineageSubgraphRequest
Request message for MetadataService.QueryArtifactLineageSubgraph
.
Fields | |
---|---|
artifact |
Required. The resource name of the Artifact whose Lineage needs to be retrieved as a LineageSubgraph. Format: The request may error with FAILED_PRECONDITION if the number of Artifacts, the number of Executions, or the number of Events that would be returned for the Context exceeds 1000. |
max_ |
Specifies the size of the lineage graph in terms of number of hops from the specified artifact. Negative Value: INVALID_ARGUMENT error is returned 0: Only input artifact is returned. No value: Transitive closure is performed to return the complete graph. |
filter |
Filter specifying the boolean condition for the Artifacts to satisfy in order to be part of the Lineage Subgraph. The syntax to define filter query is based on https://google.aip.dev/160. The supported set of filters include the following:
Each of the above supported filter types can be combined together using logical operators ( For example: |
QueryContextLineageSubgraphRequest
Request message for MetadataService.QueryContextLineageSubgraph
.
Fields | |
---|---|
context |
Required. The resource name of the Context whose Artifacts and Executions should be retrieved as a LineageSubgraph. Format: The request may error with FAILED_PRECONDITION if the number of Artifacts, the number of Executions, or the number of Events that would be returned for the Context exceeds 1000. |
QueryDeployedModelsRequest
Request message for QueryDeployedModels method.
Fields | |
---|---|
deployment_ |
Required. The name of the target DeploymentResourcePool to query. Format: |
page_ |
The maximum number of DeployedModels to return. The service may return fewer than this value. |
page_ |
A page token, received from a previous When paginating, all other parameters provided to |
QueryDeployedModelsResponse
Response message for QueryDeployedModels method.
Fields | |
---|---|
deployed_models[] |
DEPRECATED Use deployed_model_refs instead. |
next_ |
A token, which can be sent as |
deployed_ |
References to the DeployedModels that share the specified deploymentResourcePool. |
total_ |
The total number of DeployedModels on this DeploymentResourcePool. |
total_ |
The total number of Endpoints that have DeployedModels on this DeploymentResourcePool. |
QueryExecutionInputsAndOutputsRequest
Request message for MetadataService.QueryExecutionInputsAndOutputs
.
Fields | |
---|---|
execution |
Required. The resource name of the Execution whose input and output Artifacts should be retrieved as a LineageSubgraph. Format: |
QuestionAnsweringCorrectnessInput
Input for question answering correctness metric.
Fields | |
---|---|
metric_ |
Required. Spec for question answering correctness score metric. |
instance |
Required. Question answering correctness instance. |
QuestionAnsweringCorrectnessInstance
Spec for question answering correctness instance.
Fields | |
---|---|
prediction |
Required. Output of the evaluated model. |
reference |
Optional. Ground truth used to compare against the prediction. |
context |
Optional. Text provided as context to answer the question. |
instruction |
Required. The question asked and other instruction in the inference prompt. |
QuestionAnsweringCorrectnessResult
Spec for question answering correctness result.
Fields | |
---|---|
explanation |
Output only. Explanation for question answering correctness score. |
score |
Output only. Question Answering Correctness score. |
confidence |
Output only. Confidence for question answering correctness score. |
QuestionAnsweringCorrectnessSpec
Spec for question answering correctness metric.
Fields | |
---|---|
use_ |
Optional. Whether to use instance.reference to compute question answering correctness. |
version |
Optional. Which version to use for evaluation. |
QuestionAnsweringHelpfulnessInput
Input for question answering helpfulness metric.
Fields | |
---|---|
metric_ |
Required. Spec for question answering helpfulness score metric. |
instance |
Required. Question answering helpfulness instance. |
QuestionAnsweringHelpfulnessInstance
Spec for question answering helpfulness instance.
Fields | |
---|---|
prediction |
Required. Output of the evaluated model. |
reference |
Optional. Ground truth used to compare against the prediction. |
context |
Optional. Text provided as context to answer the question. |
instruction |
Required. The question asked and other instruction in the inference prompt. |
QuestionAnsweringHelpfulnessResult
Spec for question answering helpfulness result.
Fields | |
---|---|
explanation |
Output only. Explanation for question answering helpfulness score. |
score |
Output only. Question Answering Helpfulness score. |
confidence |
Output only. Confidence for question answering helpfulness score. |
QuestionAnsweringHelpfulnessSpec
Spec for question answering helpfulness metric.
Fields | |
---|---|
use_ |
Optional. Whether to use instance.reference to compute question answering helpfulness. |
version |
Optional. Which version to use for evaluation. |
QuestionAnsweringQualityInput
Input for question answering quality metric.
Fields | |
---|---|
metric_ |
Required. Spec for question answering quality score metric. |
instance |
Required. Question answering quality instance. |
QuestionAnsweringQualityInstance
Spec for question answering quality instance.
Fields | |
---|---|
prediction |
Required. Output of the evaluated model. |
reference |
Optional. Ground truth used to compare against the prediction. |
context |
Required. Text to answer the question. |
instruction |
Required. Question Answering prompt for LLM. |
QuestionAnsweringQualityResult
Spec for question answering quality result.
Fields | |
---|---|
explanation |
Output only. Explanation for question answering quality score. |
score |
Output only. Question Answering Quality score. |
confidence |
Output only. Confidence for question answering quality score. |
QuestionAnsweringQualitySpec
Spec for question answering quality score metric.
Fields | |
---|---|
use_ |
Optional. Whether to use instance.reference to compute question answering quality. |
version |
Optional. Which version to use for evaluation. |
QuestionAnsweringRelevanceInput
Input for question answering relevance metric.
Fields | |
---|---|
metric_ |
Required. Spec for question answering relevance score metric. |
instance |
Required. Question answering relevance instance. |
QuestionAnsweringRelevanceInstance
Spec for question answering relevance instance.
Fields | |
---|---|
prediction |
Required. Output of the evaluated model. |
reference |
Optional. Ground truth used to compare against the prediction. |
context |
Optional. Text provided as context to answer the question. |
instruction |
Required. The question asked and other instruction in the inference prompt. |
QuestionAnsweringRelevanceResult
Spec for question answering relevance result.
Fields | |
---|---|
explanation |
Output only. Explanation for question answering relevance score. |
score |
Output only. Question Answering Relevance score. |
confidence |
Output only. Confidence for question answering relevance score. |
QuestionAnsweringRelevanceSpec
Spec for question answering relevance metric.
Fields | |
---|---|
use_ |
Optional. Whether to use instance.reference to compute question answering relevance. |
version |
Optional. Which version to use for evaluation. |
RagContexts
Relevant contexts for one query.
Fields | |
---|---|
contexts[] |
All its contexts. |
Context
A context of the query.
Fields | |
---|---|
source_ |
If the file is imported from Cloud Storage or Google Drive, source_uri will be original file URI in Cloud Storage or Google Drive; if file is uploaded, source_uri will be file display name. |
source_ |
The file display name. |
text |
The text chunk. |
score |
According to the underlying Vector DB and the selected metric type, the score can be either the distance or the similarity between the query and the context and its range depends on the metric type. For example, if the metric type is COSINE_DISTANCE, it represents the distance between the query and the context. The larger the distance, the less relevant the context is to the query. The range is [0, 2], while 0 means the most relevant and 2 means the least relevant. |
RagCorpus
A RagCorpus is a RagFile container and a project can have multiple RagCorpora.
Fields | |
---|---|
name |
Output only. The resource name of the RagCorpus. |
display_ |
Required. The display name of the RagCorpus. The name can be up to 128 characters long and can consist of any UTF-8 characters. |
description |
Optional. The description of the RagCorpus. |
create_ |
Output only. Timestamp when this RagCorpus was created. |
update_ |
Output only. Timestamp when this RagCorpus was last updated. |
corpus_ |
Output only. RagCorpus state. |
Union field backend_config . The backend config of the RagCorpus. It can be data store and/or retrieval engine. backend_config can be only one of the following: |
|
vector_ |
Optional. Immutable. The config for the Vector DBs. |
RagEmbeddingModelConfig
Config for the embedding model to use for RAG.
Fields | |
---|---|
Union field model_config . The model config to use. model_config can be only one of the following: |
|
vertex_ |
The Vertex AI Prediction Endpoint that either refers to a publisher model or an endpoint that is hosting a 1P fine-tuned text embedding model. Endpoints hosting non-1P fine-tuned text embedding models are currently not supported. This is used for dense vector search. |
VertexPredictionEndpoint
Config representing a model hosted on Vertex Prediction Endpoint.
Fields | |
---|---|
endpoint |
Required. The endpoint resource name. Format: |
model |
Output only. The resource name of the model that is deployed on the endpoint. Present only when the endpoint is not a publisher model. Pattern: |
model_ |
Output only. Version ID of the model that is deployed on the endpoint. Present only when the endpoint is not a publisher model. |
RagFile
A RagFile contains user data for chunking, embedding and indexing.
Fields | |
---|---|
name |
Output only. The resource name of the RagFile. |
display_ |
Required. The display name of the RagFile. The name can be up to 128 characters long and can consist of any UTF-8 characters. |
description |
Optional. The description of the RagFile. |
create_ |
Output only. Timestamp when this RagFile was created. |
update_ |
Output only. Timestamp when this RagFile was last updated. |
file_ |
Output only. State of the RagFile. |
Union field rag_file_source . The origin location of the RagFile if it is imported from Google Cloud Storage or Google Drive. rag_file_source can be only one of the following: |
|
gcs_ |
Output only. Google Cloud Storage location of the RagFile. It does not support wildcards in the Cloud Storage uri for now. |
google_ |
Output only. Google Drive location. Supports importing individual files as well as Google Drive folders. |
direct_ |
Output only. The RagFile is encapsulated and uploaded in the UploadRagFile request. |
slack_ |
The RagFile is imported from a Slack channel. |
jira_ |
The RagFile is imported from a Jira query. |
share_ |
The RagFile is imported from a SharePoint source. |
RagFileChunkingConfig
Specifies the size and overlap of chunks for RagFiles.
Fields | |
---|---|
Union field chunking_config . Specifies the chunking config for RagFiles. chunking_config can be only one of the following: |
|
fixed_ |
Specifies the fixed length chunking config. |
FixedLengthChunking
Specifies the fixed length chunking config.
Fields | |
---|---|
chunk_ |
The size of the chunks. |
chunk_ |
The overlap between chunks. |
RagFileTransformationConfig
Specifies the transformation config for RagFiles.
Fields | |
---|---|
rag_ |
Specifies the chunking config for RagFiles. |
RagQuery
A query to retrieve relevant contexts.
Fields | |
---|---|
rag_ |
Optional. The retrieval config for the query. |
Union field query . The query to retrieve contexts. Currently only text query is supported. query can be only one of the following: |
|
text |
Optional. The query in text format to get relevant contexts. |
RagRetrievalConfig
Specifies the context retrieval config.
Fields | |
---|---|
top_ |
Optional. The number of contexts to retrieve. |
filter |
Optional. Config for filters. |
Filter
Config for filters.
Fields | |
---|---|
metadata_ |
Optional. String for metadata filtering. |
Union field vector_db_threshold . Filter contexts retrieved from the vector DB based on either vector distance or vector similarity. vector_db_threshold can be only one of the following: |
|
vector_ |
Optional. Only returns contexts with vector distance smaller than the threshold. |
vector_ |
Optional. Only returns contexts with vector similarity larger than the threshold. |
RagVectorDbConfig
Config for the Vector DB to use for RAG.
Fields | |
---|---|
api_ |
Authentication config for the chosen Vector DB. |
rag_ |
Optional. Immutable. The embedding model config of the Vector DB. |
Union field vector_db . The config for the Vector DB. vector_db can be only one of the following: |
|
rag_ |
The config for the RAG-managed Vector DB. |
pinecone |
The config for the Pinecone. |
vertex_ |
The config for the Vertex Vector Search. |
Pinecone
The config for the Pinecone.
Fields | |
---|---|
index_ |
Pinecone index name. This value cannot be changed after it's set. |
RagManagedDb
This type has no fields.
The config for the default RAG-managed Vector DB.
VertexVectorSearch
The config for the Vertex Vector Search.
Fields | |
---|---|
index_ |
The resource name of the Index Endpoint. Format: |
index |
The resource name of the Index. Format: |
RawPredictRequest
Request message for PredictionService.RawPredict
.
Fields | |
---|---|
endpoint |
Required. The name of the Endpoint requested to serve the prediction. Format: |
http_ |
The prediction input. Supports HTTP headers and arbitrary data payload. A You can specify the schema for each instance in the |
RayLogsSpec
Configuration for the Ray OSS Logs.
Fields | |
---|---|
disabled |
Optional. Flag to disable the export of Ray OSS logs to Cloud Logging. |
RayMetricSpec
Configuration for the Ray metrics.
Fields | |
---|---|
disabled |
Optional. Flag to disable the Ray metrics collection. |
RaySpec
Configuration information for the Ray cluster. For experimental launch, Ray cluster creation and Persistent cluster creation are 1:1 mapping: We will provision all the nodes within the Persistent cluster as Ray nodes.
Fields | |
---|---|
image_ |
Optional. Default image for user to choose a preferred ML framework (for example, TensorFlow or Pytorch) by choosing from Vertex prebuilt images. Either this or the resource_pool_images is required. Use this field if you need all the resource pools to have the same Ray image. Otherwise, use the {@code resource_pool_images} field. |
resource_ |
Optional. Required if image_uri isn't set. A map of resource_pool_id to prebuild Ray image if user need to use different images for different head/worker pools. This map needs to cover all the resource pool ids. Example: { "ray_head_node_pool": "head image" "ray_worker_node_pool1": "worker image" "ray_worker_node_pool2": "another worker image" } |
head_ |
Optional. This will be used to indicate which resource pool will serve as the Ray head node(the first node within that pool). Will use the machine from the first workerpool as the head node by default if this field isn't set. |
ray_ |
Optional. Ray metrics configurations. |
ray_ |
Optional. OSS Ray logging configurations. |
ReadFeatureValuesRequest
Request message for FeaturestoreOnlineServingService.ReadFeatureValues
.
Fields | |
---|---|
entity_ |
Required. The resource name of the EntityType for the entity being read. Value format: |
entity_ |
Required. ID for a specific entity. For example, for a machine learning model predicting user clicks on a website, an entity ID could be |
feature_ |
Required. Selector choosing Features of the target EntityType. |
ReadFeatureValuesResponse
Response message for FeaturestoreOnlineServingService.ReadFeatureValues
.
Fields | |
---|---|
header |
Response header. |
entity_ |
Entity view with Feature values. This may be the entity in the Featurestore if values for all Features were requested, or a projection of the entity in the Featurestore if values for only some Features were requested. |
EntityView
Entity view with Feature values.
Fields | |
---|---|
entity_ |
ID of the requested entity. |
data[] |
Each piece of data holds the k requested values for one requested Feature. If no values for the requested Feature exist, the corresponding cell will be empty. This has the same size and is in the same order as the features from the header |
Data
Container to hold value(s), successive in time, for one Feature from the request.
Fields | |
---|---|
Union field
|
|
value |
Feature value if a single value is requested. |
values |
Feature values list if values, successive in time, are requested. If the requested number of values is greater than the number of existing Feature values, nonexistent values are omitted instead of being returned as empty. |
FeatureDescriptor
Metadata for requested Features.
Fields | |
---|---|
id |
Feature ID. |
Header
Response header with metadata for the requested ReadFeatureValuesRequest.entity_type
and Features.
Fields | |
---|---|
entity_ |
The resource name of the EntityType from the |
feature_ |
List of Feature metadata corresponding to each piece of |
ReadTensorboardBlobDataRequest
Request message for TensorboardService.ReadTensorboardBlobData
.
Fields | |
---|---|
time_ |
Required. The resource name of the TensorboardTimeSeries to list Blobs. Format: |
blob_ |
IDs of the blobs to read. |
ReadTensorboardBlobDataResponse
Response message for TensorboardService.ReadTensorboardBlobData
.
Fields | |
---|---|
blobs[] |
Blob messages containing blob bytes. |
ReadTensorboardSizeRequest
Request message for TensorboardService.ReadTensorboardSize
.
Fields | |
---|---|
tensorboard |
Required. The name of the Tensorboard resource. Format: |
ReadTensorboardSizeResponse
Response message for TensorboardService.ReadTensorboardSize
.
Fields | |
---|---|
storage_ |
Payload storage size for the TensorBoard |
ReadTensorboardTimeSeriesDataRequest
Request message for TensorboardService.ReadTensorboardTimeSeriesData
.
Fields | |
---|---|
tensorboard_ |
Required. The resource name of the TensorboardTimeSeries to read data from. Format: |
max_ |
The maximum number of TensorboardTimeSeries' data to return. This value should be a positive integer. This value can be set to -1 to return all data. |
filter |
Reads the TensorboardTimeSeries' data that match the filter expression. |
ReadTensorboardTimeSeriesDataResponse
Response message for TensorboardService.ReadTensorboardTimeSeriesData
.
Fields | |
---|---|
time_ |
The returned time series data. |
ReadTensorboardUsageRequest
Request message for TensorboardService.ReadTensorboardUsage
.
Fields | |
---|---|
tensorboard |
Required. The name of the Tensorboard resource. Format: |
ReadTensorboardUsageResponse
Response message for TensorboardService.ReadTensorboardUsage
.
Fields | |
---|---|
monthly_ |
Maps year-month (YYYYMM) string to per month usage data. |
PerMonthUsageData
Per month usage data
Fields | |
---|---|
user_ |
Usage data for each user in the given month. |
PerUserUsageData
Per user usage data.
Fields | |
---|---|
username |
User's username |
view_ |
Number of times the user has read data within the Tensorboard. |
RebaseTunedModelOperationMetadata
Runtime operation information for GenAiTuningService.RebaseTunedModel
.
Fields | |
---|---|
generic_ |
The common part of the operation generic information. |
RebaseTunedModelRequest
Request message for GenAiTuningService.RebaseTunedModel
.
Fields | |
---|---|
parent |
Required. The resource name of the Location into which to rebase the Model. Format: |
tuned_ |
Required. TunedModel reference to retrieve the legacy model information. |
tuning_ |
Optional. The TuningJob to be updated. Users can use this TuningJob field to overwrite tuning configs. |
artifact_ |
Optional. The Google Cloud Storage location to write the artifacts. |
deploy_ |
Optional. By default, bison to gemini migration will always create new model/endpoint, but for gemini-1.0 to gemini-1.5 migration, we default deploy to the same endpoint. See details in this Section. |
RebootPersistentResourceOperationMetadata
Details of operations that perform reboot PersistentResource.
Fields | |
---|---|
generic_ |
Operation metadata for PersistentResource. |
progress_ |
Progress Message for Reboot LRO |
RebootPersistentResourceRequest
Request message for PersistentResourceService.RebootPersistentResource
.
Fields | |
---|---|
name |
Required. The name of the PersistentResource resource. Format: |
RemoveContextChildrenRequest
Request message for [MetadataService.DeleteContextChildrenRequest][].
Fields | |
---|---|
context |
Required. The resource name of the parent Context. Format: |
child_ |
The resource names of the child Contexts. |
RemoveContextChildrenResponse
This type has no fields.
Response message for MetadataService.RemoveContextChildren
.
RemoveDatapointsRequest
Request message for IndexService.RemoveDatapoints
Fields | |
---|---|
index |
Required. The name of the Index resource to be updated. Format: |
datapoint_ |
A list of datapoint ids to be deleted. |
RemoveDatapointsResponse
This type has no fields.
Response message for IndexService.RemoveDatapoints
ReservationAffinity
A ReservationAffinity can be used to configure a Vertex AI resource (e.g., a DeployedModel) to draw its Compute Engine resources from a Shared Reservation, or exclusively from on-demand capacity.
Fields | |
---|---|
reservation_ |
Required. Specifies the reservation affinity type. |
key |
Optional. Corresponds to the label key of a reservation resource. To target a SPECIFIC_RESERVATION by name, use |
values[] |
Optional. Corresponds to the label values of a reservation resource. This must be the full resource name of the reservation. |
Type
Identifies a type of reservation affinity.
Enums | |
---|---|
TYPE_UNSPECIFIED |
Default value. This should not be used. |
NO_RESERVATION |
Do not consume from any reserved capacity, only use on-demand. |
ANY_RESERVATION |
Consume any reservation available, falling back to on-demand. |
SPECIFIC_RESERVATION |
Consume from a specific reservation. When chosen, the reservation must be identified via the key and values fields. |
ResourcePool
Represents the spec of a group of resources of the same type, for example machine type, disk, and accelerators, in a PersistentResource.
Fields | |
---|---|
id |
Immutable. The unique ID in a PersistentResource for referring to this resource pool. User can specify it if necessary. Otherwise, it's generated automatically. |
machine_ |
Required. Immutable. The specification of a single machine. |
disk_ |
Optional. Disk spec for the machine in this node pool. |
used_ |
Output only. The number of machines currently in use by training jobs for this resource pool. Will replace idle_replica_count. |
autoscaling_ |
Optional. Optional spec to configure GKE or Ray-on-Vertex autoscaling |
replica_ |
Optional. The total number of machines to use for this resource pool. |
AutoscalingSpec
The min/max number of replicas allowed if enabling autoscaling
Fields | |
---|---|
min_ |
Optional. min replicas in the node pool, must be ≤ replica_count and < max_replica_count or will throw error. For autoscaling enabled Ray-on-Vertex, we allow min_replica_count of a resource_pool to be 0 to match the OSS Ray behavior(https://docs.ray.io/en/latest/cluster/vms/user-guides/configuring-autoscaling.html#cluster-config-parameters). As for Persistent Resource, the min_replica_count must be > 0, we added a corresponding validation inside CreatePersistentResourceRequestValidator.java. |
max_ |
Optional. max replicas in the node pool, must be ≥ replica_count and > min_replica_count or will throw error |
ResourceRuntime
Persistent Cluster runtime information as output
Fields | |
---|---|
access_ |
Output only. URIs for user to connect to the Cluster. Example: { "RAY_HEAD_NODE_INTERNAL_IP": "head-node-IP:10001" "RAY_DASHBOARD_URI": "ray-dashboard-address:8888" } |
ResourceRuntimeSpec
Configuration for the runtime on a PersistentResource instance, including but not limited to:
- Service accounts used to run the workloads.
- Whether to make it a dedicated Ray Cluster.
Fields | |
---|---|
service_ |
Optional. Configure the use of workload identity on the PersistentResource |
ray_ |
Optional. Ray cluster configuration. Required when creating a dedicated RayCluster on the PersistentResource. |
ResourcesConsumed
Statistics information about resource consumption.
Fields | |
---|---|
replica_ |
Output only. The number of replica hours used. Note that many replicas may run in parallel, and additionally any given work may be queued for some time. Therefore this value is not strictly related to wall time. |
RestoreDatasetVersionOperationMetadata
Runtime operation information for DatasetService.RestoreDatasetVersion
.
Fields | |
---|---|
generic_ |
The common part of the operation metadata. |
RestoreDatasetVersionRequest
Request message for DatasetService.RestoreDatasetVersion
.
Fields | |
---|---|
name |
Required. The name of the DatasetVersion resource. Format: |
ResumeModelDeploymentMonitoringJobRequest
Request message for JobService.ResumeModelDeploymentMonitoringJob
.
Fields | |
---|---|
name |
Required. The resource name of the ModelDeploymentMonitoringJob to resume. Format: |
ResumeScheduleRequest
Request message for ScheduleService.ResumeSchedule
.
Fields | |
---|---|
name |
Required. The name of the Schedule resource to be resumed. Format: |
catch_ |
Optional. Whether to backfill missed runs when the schedule is resumed from PAUSED state. If set to true, all missed runs will be scheduled. New runs will be scheduled after the backfill is complete. This will also update |
Retrieval
Defines a retrieval tool that model can call to access external knowledge.
Fields | |
---|---|
disable_attribution |
Optional. Deprecated. This option is no longer supported. |
Union field source . The source of the retrieval. source can be only one of the following: |
|
vertex_ |
Set to use data source powered by Vertex AI Search. |
vertex_ |
Set to use data source powered by Vertex RAG store. User data is uploaded via the VertexRagDataService. |
RetrievalMetadata
Metadata related to retrieval in the grounding flow.
Fields | |
---|---|
google_ |
Optional. Score indicating how likely information from Google Search could help answer the prompt. The score is in the range |
RetrieveContextsRequest
Request message for VertexRagService.RetrieveContexts
.
Fields | |
---|---|
parent |
Required. The resource name of the Location from which to retrieve RagContexts. The users must have permission to make a call in the project. Format: |
query |
Required. Single RAG retrieve query. |
Union field data_source . Data Source to retrieve contexts. data_source can be only one of the following: |
|
vertex_ |
The data source for Vertex RagStore. |
VertexRagStore
The data source for Vertex RagStore.
Fields | |
---|---|
rag_ |
Optional. The representation of the rag source. It can be used to specify corpus only or ragfiles. Currently only support one corpus or multiple files from one corpus. In the future we may open up multiple corpora support. |
vector_distance_threshold |
Optional. Only return contexts with vector distance smaller than the threshold. |
RagResource
The definition of the Rag resource.
Fields | |
---|---|
rag_ |
Optional. RagCorpora resource name. Format: |
rag_ |
Optional. rag_file_id. The files should be in the same rag_corpus set in rag_corpus field. |
RetrieveContextsResponse
Response message for VertexRagService.RetrieveContexts
.
Fields | |
---|---|
contexts |
The contexts of the query. |
RougeInput
Input for rouge metric.
Fields | |
---|---|
metric_ |
Required. Spec for rouge score metric. |
instances[] |
Required. Repeated rouge instances. |
RougeInstance
Spec for rouge instance.
Fields | |
---|---|
prediction |
Required. Output of the evaluated model. |
reference |
Required. Ground truth used to compare against the prediction. |
RougeMetricValue
Rouge metric value for an instance.
Fields | |
---|---|
score |
Output only. Rouge score. |
RougeResults
Results for rouge metric.
Fields | |
---|---|
rouge_ |
Output only. Rouge metric values. |
RougeSpec
Spec for rouge score metric - calculates the recall of n-grams in prediction as compared to reference - returns a score ranging between 0 and 1.
Fields | |
---|---|
rouge_ |
Optional. Supported rouge types are rougen[1-9], rougeL, and rougeLsum. |
use_ |
Optional. Whether to use stemmer to compute rouge score. |
split_ |
Optional. Whether to split summaries while using rougeLsum. |
SafetyInput
Input for safety metric.
Fields | |
---|---|
metric_ |
Required. Spec for safety metric. |
instance |
Required. Safety instance. |
SafetyInstance
Spec for safety instance.
Fields | |
---|---|
prediction |
Required. Output of the evaluated model. |
SafetyRating
Safety rating corresponding to the generated content.
Fields | |
---|---|
category |
Output only. Harm category. |
probability |
Output only. Harm probability levels in the content. |
probability_ |
Output only. Harm probability score. |
severity |
Output only. Harm severity levels in the content. |
severity_ |
Output only. Harm severity score. |
blocked |
Output only. Indicates whether the content was filtered out because of this rating. |
HarmProbability
Harm probability levels in the content.
Enums | |
---|---|
HARM_PROBABILITY_UNSPECIFIED |
Harm probability unspecified. |
NEGLIGIBLE |
Negligible level of harm. |
LOW |
Low level of harm. |
MEDIUM |
Medium level of harm. |
HIGH |
High level of harm. |
HarmSeverity
Harm severity levels.
Enums | |
---|---|
HARM_SEVERITY_UNSPECIFIED |
Harm severity unspecified. |
HARM_SEVERITY_NEGLIGIBLE |
Negligible level of harm severity. |
HARM_SEVERITY_LOW |
Low level of harm severity. |
HARM_SEVERITY_MEDIUM |
Medium level of harm severity. |
HARM_SEVERITY_HIGH |
High level of harm severity. |
SafetyResult
Spec for safety result.
Fields | |
---|---|
explanation |
Output only. Explanation for safety score. |
score |
Output only. Safety score. |
confidence |
Output only. Confidence for safety score. |
SafetySetting
Safety settings.
Fields | |
---|---|
category |
Required. Harm category. |
threshold |
Required. The harm block threshold. |
method |
Optional. Specify if the threshold is used for probability or severity score. If not specified, the threshold is used for probability score. |
HarmBlockMethod
Probability vs severity.
Enums | |
---|---|
HARM_BLOCK_METHOD_UNSPECIFIED |
The harm block method is unspecified. |
SEVERITY |
The harm block method uses both probability and severity scores. |
PROBABILITY |
The harm block method uses the probability score. |
HarmBlockThreshold
Probability based thresholds levels for blocking.
Enums | |
---|---|
HARM_BLOCK_THRESHOLD_UNSPECIFIED |
Unspecified harm block threshold. |
BLOCK_LOW_AND_ABOVE |
Block low threshold and above (i.e. block more). |
BLOCK_MEDIUM_AND_ABOVE |
Block medium threshold and above. |
BLOCK_ONLY_HIGH |
Block only high threshold (i.e. block less). |
BLOCK_NONE |
Block none. |
OFF |
Turn off the safety filter. |
SafetySpec
Spec for safety metric.
Fields | |
---|---|
version |
Optional. Which version to use for evaluation. |
SampledShapleyAttribution
An attribution method that approximates Shapley values for features that contribute to the label being predicted. A sampling strategy is used to approximate the value rather than considering all subsets of features.
Fields | |
---|---|
path_ |
Required. The number of feature permutations to consider when approximating the Shapley values. Valid range of its value is [1, 50], inclusively. |
SamplingStrategy
Sampling Strategy for logging, can be for both training and prediction dataset.
Fields | |
---|---|
random_ |
Random sample config. Will support more sampling strategies later. |
RandomSampleConfig
Requests are randomly selected.
Fields | |
---|---|
sample_ |
Sample rate (0, 1] |
SavedQuery
A SavedQuery is a view of the dataset. It references a subset of annotations by problem type and filters.
Fields | |
---|---|
name |
Output only. Resource name of the SavedQuery. |
display_ |
Required. The user-defined name of the SavedQuery. The name can be up to 128 characters long and can consist of any UTF-8 characters. |
metadata |
Some additional information about the SavedQuery. |
create_ |
Output only. Timestamp when this SavedQuery was created. |
update_ |
Output only. Timestamp when SavedQuery was last updated. |
annotation_ |
Output only. Filters on the Annotations in the dataset. |
problem_ |
Required. Problem type of the SavedQuery. Allowed values:
|
annotation_ |
Output only. Number of AnnotationSpecs in the context of the SavedQuery. |
etag |
Used to perform a consistent read-modify-write update. If not set, a blind "overwrite" update happens. |
support_ |
Output only. If the Annotations belonging to the SavedQuery can be used for AutoML training. |
Scalar
One point viewable on a scalar metric plot.
Fields | |
---|---|
value |
Value of the point at this step / timestamp. |
Schedule
An instance of a Schedule periodically schedules runs to make API calls based on user specified time specification and API request type.
Fields | |
---|---|
name |
Immutable. The resource name of the Schedule. |
display_ |
Required. User provided name of the Schedule. The name can be up to 128 characters long and can consist of any UTF-8 characters. |
start_ |
Optional. Timestamp after which the first run can be scheduled. Default to Schedule create time if not specified. |
end_ |
Optional. Timestamp after which no new runs can be scheduled. If specified, The schedule will be completed when either end_time is reached or when scheduled_run_count >= max_run_count. If not specified, new runs will keep getting scheduled until this Schedule is paused or deleted. Already scheduled runs will be allowed to complete. Unset if not specified. |
max_ |
Optional. Maximum run count of the schedule. If specified, The schedule will be completed when either started_run_count >= max_run_count or when end_time is reached. If not specified, new runs will keep getting scheduled until this Schedule is paused or deleted. Already scheduled runs will be allowed to complete. Unset if not specified. |
started_ |
Output only. The number of runs started by this schedule. |
state |
Output only. The state of this Schedule. |
create_ |
Output only. Timestamp when this Schedule was created. |
update_ |
Output only. Timestamp when this Schedule was updated. |
next_ |
Output only. Timestamp when this Schedule should schedule the next run. Having a next_run_time in the past means the runs are being started behind schedule. |
last_ |
Output only. Timestamp when this Schedule was last paused. Unset if never paused. |
last_ |
Output only. Timestamp when this Schedule was last resumed. Unset if never resumed from pause. |
max_ |
Required. Maximum number of runs that can be started concurrently for this Schedule. This is the limit for starting the scheduled requests and not the execution of the operations/jobs created by the requests (if applicable). |
allow_ |
Optional. Whether new scheduled runs can be queued when max_concurrent_runs limit is reached. If set to true, new runs will be queued instead of skipped. Default to false. |
catch_ |
Output only. Whether to backfill missed runs when the schedule is resumed from PAUSED state. If set to true, all missed runs will be scheduled. New runs will be scheduled after the backfill is complete. Default to false. |
last_ |
Output only. Response of the last scheduled run. This is the response for starting the scheduled requests and not the execution of the operations/jobs created by the requests (if applicable). Unset if no run has been scheduled yet. |
Union field time_specification . Required. The time specification to launch scheduled runs. time_specification can be only one of the following: |
|
cron |
Cron schedule (https://en.wikipedia.org/wiki/Cron) to launch scheduled runs. To explicitly set a timezone to the cron tab, apply a prefix in the cron tab: "CRON_TZ=${IANA_TIME_ZONE}" or "TZ=${IANA_TIME_ZONE}". The ${IANA_TIME_ZONE} may only be a valid string from IANA time zone database. For example, "CRON_TZ=America/New_York 1 * * * *", or "TZ=America/New_York 1 * * * *". |
Union field request . Required. The API request template to launch the scheduled runs. User-specified ID is not supported in the request template. request can be only one of the following: |
|
create_ |
Request for |
create_ |
Request for |
RunResponse
Status of a scheduled run.
Fields | |
---|---|
scheduled_ |
The scheduled run time based on the user-specified schedule. |
run_ |
The response of the scheduled run. |
State
Possible state of the schedule.
Enums | |
---|---|
STATE_UNSPECIFIED |
Unspecified. |
ACTIVE |
The Schedule is active. Runs are being scheduled on the user-specified timespec. |
PAUSED |
The schedule is paused. No new runs will be created until the schedule is resumed. Already started runs will be allowed to complete. |
COMPLETED |
The Schedule is completed. No new runs will be scheduled. Already started runs will be allowed to complete. Schedules in completed state cannot be paused or resumed. |
Scheduling
All parameters related to queuing and scheduling of custom jobs.
Fields | |
---|---|
timeout |
Optional. The maximum job running time. The default is 7 days. |
restart_ |
Optional. Restarts the entire CustomJob if a worker gets restarted. This feature can be used by distributed training jobs that are not resilient to workers leaving and joining a job. |
strategy |
Optional. This determines which type of scheduling strategy to use. |
disable_ |
Optional. Indicates if the job should retry for internal errors after the job starts running. If true, overrides |
max_ |
Optional. This is the maximum duration that a job will wait for the requested resources to be provisioned if the scheduling strategy is set to [Strategy.DWS_FLEX_START]. If set to 0, the job will wait indefinitely. The default is 24 hours. |
Strategy
Optional. This determines which type of scheduling strategy to use. Right now users have two options such as STANDARD which will use regular on demand resources to schedule the job, the other is SPOT which would leverage spot resources alongwith regular resources to schedule the job.
Enums | |
---|---|
STRATEGY_UNSPECIFIED |
Strategy will default to STANDARD. |
ON_DEMAND |
Deprecated. Regular on-demand provisioning strategy. |
LOW_COST |
Deprecated. Low cost by making potential use of spot resources. |
STANDARD |
Standard provisioning strategy uses regular on-demand resources. |
SPOT |
Spot provisioning strategy uses spot resources. |
FLEX_START |
Flex Start strategy uses DWS to queue for resources. |
Schema
Schema is used to define the format of input/output data. Represents a select subset of an OpenAPI 3.0 schema object. More fields may be added in the future as needed.
Fields | |
---|---|
type |
Optional. The type of the data. |
format |
Optional. The format of the data. Supported formats: for NUMBER type: "float", "double" for INTEGER type: "int32", "int64" for STRING type: "email", "byte", etc |
title |
Optional. The title of the Schema. |
description |
Optional. The description of the data. |
nullable |
Optional. Indicates if the value may be null. |
default |
Optional. Default value of the data. |
items |
Optional. SCHEMA FIELDS FOR TYPE ARRAY Schema of the elements of Type.ARRAY. |
min_ |
Optional. Minimum number of the elements for Type.ARRAY. |
max_ |
Optional. Maximum number of the elements for Type.ARRAY. |
enum[] |
Optional. Possible values of the element of primitive type with enum format. Examples: 1. We can define direction as : {type:STRING, format:enum, enum:["EAST", NORTH", "SOUTH", "WEST"]} 2. We can define apartment number as : {type:INTEGER, format:enum, enum:["101", "201", "301"]} |
properties |
Optional. SCHEMA FIELDS FOR TYPE OBJECT Properties of Type.OBJECT. |
property_ |
Optional. The order of the properties. Not a standard field in open api spec. Only used to support the order of the properties. |
required[] |
Optional. Required properties of Type.OBJECT. |
min_ |
Optional. Minimum number of the properties for Type.OBJECT. |
max_ |
Optional. Maximum number of the properties for Type.OBJECT. |
minimum |
Optional. SCHEMA FIELDS FOR TYPE INTEGER and NUMBER Minimum value of the Type.INTEGER and Type.NUMBER |
maximum |
Optional. Maximum value of the Type.INTEGER and Type.NUMBER |
min_ |
Optional. SCHEMA FIELDS FOR TYPE STRING Minimum length of the Type.STRING |
max_ |
Optional. Maximum length of the Type.STRING |
pattern |
Optional. Pattern of the Type.STRING to restrict a string to a regular expression. |
example |
Optional. Example of the object. Will only populated when the object is the root. |
any_ |
Optional. The value should be validated against any (one or more) of the subschemas in the list. |
SearchDataItemsRequest
Request message for DatasetService.SearchDataItems
.
Fields | |
---|---|
dataset |
Required. The resource name of the Dataset from which to search DataItems. Format: |
saved_query |
The resource name of a SavedQuery(annotation set in UI). Format: |
data_ |
The resource name of a DataLabelingJob. Format: |
data_ |
An expression for filtering the DataItem that will be returned.
For example:
|
annotations_filter |
An expression for filtering the Annotations that will be returned per DataItem. * |
annotation_ |
An expression that specifies what Annotations will be returned per DataItem. Annotations satisfied either of the conditions will be returned. * |
field_ |
Mask specifying which fields of |
annotations_ |
If set, only up to this many of Annotations will be returned per DataItemView. The maximum value is 1000. If not set, the maximum value will be used. |
page_ |
Requested page size. Server may return fewer results than requested. Default and maximum page size is 100. |
order_by |
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. |
page_ |
A token identifying a page of results for the server to return Typically obtained via |
Union field
|
|
order_ |
A comma-separated list of data item fields to order by, sorted in ascending order. Use "desc" after a field name for descending. |
order_ |
Expression that allows ranking results based on annotation's property. |
OrderByAnnotation
Expression that allows ranking results based on annotation's property.
Fields | |
---|---|
saved_ |
Required. Saved query of the Annotation. Only Annotations belong to this saved query will be considered for ordering. |
order_ |
A comma-separated list of annotation fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Must also specify saved_query. |
SearchDataItemsResponse
Response message for DatasetService.SearchDataItems
.
Fields | |
---|---|
data_ |
The DataItemViews read. |
next_ |
A token to retrieve next page of results. Pass to |
SearchEntryPoint
Google search entry point.
Fields | |
---|---|
rendered_ |
Optional. Web content snippet that can be embedded in a web page or an app webview. |
sdk_ |
Optional. Base64 encoded JSON representing array of <search term, search url> tuple. |
SearchFeaturesRequest
Request message for FeaturestoreService.SearchFeatures
.
Fields | |
---|---|
location |
Required. The resource name of the Location to search Features. Format: |
query |
Query string that is a conjunction of field-restricted queries and/or field-restricted filters. Field-restricted queries and filters can be combined using A field query is in the form FIELD:QUERY. This implicitly checks if QUERY exists as a substring within Feature's FIELD. The QUERY and the FIELD are converted to a sequence of words (i.e. tokens) for comparison. This is done by:
A QUERY must be either a singular token or a phrase. A phrase is one or multiple words enclosed in double quotation marks ("). With phrases, the order of the words is important. Words in the phrase must be matching in order and consecutively. Supported FIELDs for field-restricted queries:
Examples:
Besides field queries, the following exact-match filters are supported. The exact-match filters do not support wildcards. Unlike field-restricted queries, exact-match filters are case-sensitive.
Examples:
|
page_ |
The maximum number of Features to return. The service may return fewer than this value. If unspecified, at most 100 Features will be returned. The maximum value is 100; any value greater than 100 will be coerced to 100. |
page_ |
A page token, received from a previous When paginating, all other parameters provided to |
SearchFeaturesResponse
Response message for FeaturestoreService.SearchFeatures
.
Fields | |
---|---|
features[] |
The Features matching the request. Fields returned:
|
next_ |
A token, which can be sent as |
SearchMigratableResourcesRequest
Request message for MigrationService.SearchMigratableResources
.
Fields | |
---|---|
parent |
Required. The location that the migratable resources should be searched from. It's the Vertex AI location that the resources can be migrated to, not the resources' original location. Format: |
page_ |
The standard page size. The default and maximum value is 100. |
page_ |
The standard page token. |
filter |
A filter for your search. You can use the following types of filters:
|
SearchMigratableResourcesResponse
Response message for MigrationService.SearchMigratableResources
.
Fields | |
---|---|
migratable_ |
All migratable resources that can be migrated to the location specified in the request. |
next_ |
The standard next-page token. The migratable_resources may not fill page_size in SearchMigratableResourcesRequest even when there are subsequent pages. |
SearchModelDeploymentMonitoringStatsAnomaliesRequest
Request message for JobService.SearchModelDeploymentMonitoringStatsAnomalies
.
Fields | |
---|---|
model_ |
Required. ModelDeploymentMonitoring Job resource name. Format: |
deployed_ |
Required. The DeployedModel ID of the [ModelDeploymentMonitoringObjectiveConfig.deployed_model_id]. |
feature_ |
The feature display name. If specified, only return the stats belonging to this feature. Format: |
objectives[] |
Required. Objectives of the stats to retrieve. |
page_ |
The standard list page size. |
page_ |
A page token received from a previous |
start_ |
The earliest timestamp of stats being generated. If not set, indicates fetching stats till the earliest possible one. |
end_ |
The latest timestamp of stats being generated. If not set, indicates feching stats till the latest possible one. |
StatsAnomaliesObjective
Stats requested for specific objective.
Fields | |
---|---|
type |
|
top_ |
If set, all attribution scores between |
SearchModelDeploymentMonitoringStatsAnomaliesResponse
Response message for JobService.SearchModelDeploymentMonitoringStatsAnomalies
.
Fields | |
---|---|
monitoring_ |
Stats retrieved for requested objectives. There are at most 1000 |
next_ |
The page token that can be used by the next |
SearchNearestEntitiesRequest
The request message for FeatureOnlineStoreService.SearchNearestEntities
.
Fields | |
---|---|
feature_ |
Required. FeatureView resource format |
query |
Required. The query. |
return_ |
Optional. If set to true, the full entities (including all vector values and metadata) of the nearest neighbors are returned; otherwise only entity id of the nearest neighbors will be returned. Note that returning full entities will significantly increase the latency and cost of the query. |
SearchNearestEntitiesResponse
Response message for FeatureOnlineStoreService.SearchNearestEntities
Fields | |
---|---|
nearest_ |
The nearest neighbors of the query entity. |
Segment
Segment of the content.
Fields | |
---|---|
part_ |
Output only. The index of a Part object within its parent Content object. |
start_ |
Output only. Start index in the given Part, measured in bytes. Offset from the start of the Part, inclusive, starting at zero. |
end_ |
Output only. End index in the given Part, measured in bytes. Offset from the start of the Part, exclusive, starting at zero. |
text |
Output only. The text corresponding to the segment from the response. |
ServiceAccountSpec
Configuration for the use of custom service account to run the workloads.
Fields | |
---|---|
enable_ |
Required. If true, custom user-managed service account is enforced to run any workloads (for example, Vertex Jobs) on the resource. Otherwise, uses the Vertex AI Custom Code Service Agent. |
service_ |
Optional. Required when all below conditions are met * The users must have Do not set this field if you want to submit jobs using custom service account to this PersistentResource after creation, but only specify the |
ShieldedVmConfig
A set of Shielded Instance options. See Images using supported Shielded VM features.
Fields | |
---|---|
enable_ |
Defines whether the instance has Secure Boot enabled. Secure Boot helps ensure that the system only runs authentic software by verifying the digital signature of all boot components, and halting the boot process if signature verification fails. |
SlackSource
The Slack source for the ImportRagFilesRequest.
Fields | |
---|---|
channels[] |
Required. The Slack channels. |
SlackChannels
SlackChannels contains the Slack channels and corresponding access token.
Fields | |
---|---|
channels[] |
Required. The Slack channel IDs. |
api_ |
Required. The SecretManager secret version resource name (e.g. projects/{project}/secrets/{secret}/versions/{version}) storing the Slack channel access token that has access to the slack channel IDs. See: https://api.slack.com/tutorials/tracks/getting-a-token. |
SlackChannel
SlackChannel contains the Slack channel ID and the time range to import.
Fields | |
---|---|
channel_ |
Required. The Slack channel ID. |
start_ |
Optional. The starting timestamp for messages to import. |
end_ |
Optional. The ending timestamp for messages to import. |
SmoothGradConfig
Config for SmoothGrad approximation of gradients.
When enabled, the gradients are approximated by averaging the gradients from noisy samples in the vicinity of the inputs. Adding noise can help improve the computed gradients. Refer to this paper for more details: https://arxiv.org/pdf/1706.03825.pdf
Fields | |
---|---|
noisy_ |
The number of gradient samples to use for approximation. The higher this number, the more accurate the gradient is, but the runtime complexity increases by this factor as well. Valid range of its value is [1, 50]. Defaults to 3. |
Union field GradientNoiseSigma . Represents the standard deviation of the gaussian kernel that will be used to add noise to the interpolated inputs prior to computing gradients. GradientNoiseSigma can be only one of the following: |
|
noise_ |
This is a single float value and will be used to add noise to all the features. Use this field when all features are normalized to have the same distribution: scale to range [0, 1], [-1, 1] or z-scoring, where features are normalized to have 0-mean and 1-variance. Learn more about normalization. For best results the recommended value is about 10% - 20% of the standard deviation of the input feature. Refer to section 3.2 of the SmoothGrad paper: https://arxiv.org/pdf/1706.03825.pdf. Defaults to 0.1. If the distribution is different per feature, set |
feature_ |
This is similar to |
SpecialistPool
SpecialistPool represents customers' own workforce to work on their data labeling jobs. It includes a group of specialist managers and workers. Managers are responsible for managing the workers in this pool as well as customers' data labeling jobs associated with this pool. Customers create specialist pool as well as start data labeling jobs on Cloud, managers and workers handle the jobs using CrowdCompute console.
Fields | |
---|---|
name |
Required. The resource name of the SpecialistPool. |
display_ |
Required. The user-defined name of the SpecialistPool. The name can be up to 128 characters long and can consist of any UTF-8 characters. This field should be unique on project-level. |
specialist_ |
Output only. The number of managers in this SpecialistPool. |
specialist_ |
The email addresses of the managers in the SpecialistPool. |
pending_ |
Output only. The resource name of the pending data labeling jobs. |
specialist_ |
The email addresses of workers in the SpecialistPool. |
StartNotebookRuntimeOperationMetadata
Metadata information for NotebookService.StartNotebookRuntime
.
Fields | |
---|---|
generic_ |
The operation generic information. |
progress_ |
A human-readable message that shows the intermediate progress details of NotebookRuntime. |
StartNotebookRuntimeRequest
Request message for NotebookService.StartNotebookRuntime
.
Fields | |
---|---|
name |
Required. The name of the NotebookRuntime resource to be started. Instead of checking whether the name is in valid NotebookRuntime resource name format, directly throw NotFound exception if there is no such NotebookRuntime in spanner. |
StartNotebookRuntimeResponse
This type has no fields.
Response message for NotebookService.StartNotebookRuntime
.
StopNotebookRuntimeOperationMetadata
Metadata information for NotebookService.StopNotebookRuntime
.
Fields | |
---|---|
generic_ |
The operation generic information. |
StopNotebookRuntimeRequest
Request message for NotebookService.StopNotebookRuntime
.
Fields | |
---|---|
name |
Required. The name of the NotebookRuntime resource to be stopped. Instead of checking whether the name is in valid NotebookRuntime resource name format, directly throw NotFound exception if there is no such NotebookRuntime in spanner. |
StopNotebookRuntimeResponse
This type has no fields.
Response message for NotebookService.StopNotebookRuntime
.
StopTrialRequest
Request message for VizierService.StopTrial
.
Fields | |
---|---|
name |
Required. The Trial's name. Format: |
StratifiedSplit
Assigns input data to the training, validation, and test sets so that the distribution of values found in the categorical column (as specified by the key
field) is mirrored within each split. The fraction values determine the relative sizes of the splits.
For example, if the specified column has three values, with 50% of the rows having value "A", 25% value "B", and 25% value "C", and the split fractions are specified as 80/10/10, then the training set will constitute 80% of the training data, with about 50% of the training set rows having the value "A" for the specified column, about 25% having the value "B", and about 25% having the value "C".
Only the top 500 occurring values are used; any values not in the top 500 values are randomly assigned to a split. If less than three rows contain a specific value, those rows are randomly assigned.
Supported only for tabular Datasets.
Fields | |
---|---|
training_ |
The fraction of the input data that is to be used to train the Model. |
validation_ |
The fraction of the input data that is to be used to validate the Model. |
test_ |
The fraction of the input data that is to be used to evaluate the Model. |
key |
Required. The key is a name of one of the Dataset's data columns. The key provided must be for a categorical column. |
StreamDirectPredictRequest
Request message for PredictionService.StreamDirectPredict
.
The first message must contain endpoint
field and optionally [input][]. The subsequent messages must contain [input][].
Fields | |
---|---|
endpoint |
Required. The name of the Endpoint requested to serve the prediction. Format: |
inputs[] |
Optional. The prediction input. |
parameters |
Optional. The parameters that govern the prediction. |
StreamDirectPredictResponse
Response message for PredictionService.StreamDirectPredict
.
Fields | |
---|---|
outputs[] |
The prediction output. |
parameters |
The parameters that govern the prediction. |
StreamDirectRawPredictRequest
Request message for PredictionService.StreamDirectRawPredict
.
The first message must contain endpoint
and method_name
fields and optionally input
. The subsequent messages must contain input
. method_name
in the subsequent messages have no effect.
Fields | |
---|---|
endpoint |
Required. The name of the Endpoint requested to serve the prediction. Format: |
method_ |
Optional. Fully qualified name of the API method being invoked to perform predictions. Format: |
input |
Optional. The prediction input. |
StreamDirectRawPredictResponse
Response message for PredictionService.StreamDirectRawPredict
.
Fields | |
---|---|
output |
The prediction output. |
StreamRawPredictRequest
Request message for PredictionService.StreamRawPredict
.
Fields | |
---|---|
endpoint |
Required. The name of the Endpoint requested to serve the prediction. Format: |
http_ |
The prediction input. Supports HTTP headers and arbitrary data payload. |
StreamingPredictRequest
Request message for PredictionService.StreamingPredict
.
The first message must contain endpoint
field and optionally [input][]. The subsequent messages must contain [input][].
Fields | |
---|---|
endpoint |
Required. The name of the Endpoint requested to serve the prediction. Format: |
inputs[] |
The prediction input. |
parameters |
The parameters that govern the prediction. |
StreamingPredictResponse
Response message for PredictionService.StreamingPredict
.
Fields | |
---|---|
outputs[] |
The prediction output. |
parameters |
The parameters that govern the prediction. |
StreamingRawPredictRequest
Request message for PredictionService.StreamingRawPredict
.
The first message must contain endpoint
and method_name
fields and optionally input
. The subsequent messages must contain input
. method_name
in the subsequent messages have no effect.
Fields | |
---|---|
endpoint |
Required. The name of the Endpoint requested to serve the prediction. Format: |
method_ |
Fully qualified name of the API method being invoked to perform predictions. Format: |
input |
The prediction input. |
StreamingRawPredictResponse
Response message for PredictionService.StreamingRawPredict
.
Fields | |
---|---|
output |
The prediction output. |
StreamingReadFeatureValuesRequest
Request message for FeaturestoreOnlineServingService.StreamingReadFeatureValues
.
Fields | |
---|---|
entity_ |
Required. The resource name of the entities' type. Value format: |
entity_ |
Required. IDs of entities to read Feature values of. The maximum number of IDs is 100. For example, for a machine learning model predicting user clicks on a website, an entity ID could be |
feature_ |
Required. Selector choosing Features of the target EntityType. Feature IDs will be deduplicated. |
StringArray
A list of string values.
Fields | |
---|---|
values[] |
A list of string values. |
StructFieldValue
One field of a Struct (or object) type feature value.
Fields | |
---|---|
name |
Name of the field in the struct feature. |
value |
The value for this field. |
StructValue
Struct (or object) type feature value.
Fields | |
---|---|
values[] |
A list of field values. |
Study
A message representing a Study.
Fields | |
---|---|
name |
Output only. The name of a study. The study's globally unique identifier. Format: |
display_ |
Required. Describes the Study, default value is empty string. |
study_ |
Required. Configuration of the Study. |
state |
Output only. The detailed state of a Study. |
create_ |
Output only. Time at which the study was created. |
inactive_ |
Output only. A human readable reason why the Study is inactive. This should be empty if a study is ACTIVE or COMPLETED. |
State
Describes the Study state.
Enums | |
---|---|
STATE_UNSPECIFIED |
The study state is unspecified. |
ACTIVE |
The study is active. |
INACTIVE |
The study is stopped due to an internal error. |
COMPLETED |
The study is done when the service exhausts the parameter search space or max_trial_count is reached. |
StudySpec
Represents specification of a Study.
Fields | |
---|---|
metrics[] |
Required. Metric specs for the Study. |
parameters[] |
Required. The set of parameters to tune. |
algorithm |
The search algorithm specified for the Study. |
observation_ |
The observation noise level of the study. Currently only supported by the Vertex AI Vizier service. Not supported by HyperparameterTuningJob or TrainingPipeline. |
measurement_ |
Describe which measurement selection type will be used |
Union field
|
|
decay_ |
The automated early stopping spec using decay curve rule. |
median_ |
The automated early stopping spec using median rule. |
convex_ |
The automated early stopping spec using convex stopping rule. |
study_ |
Conditions for automated stopping of a Study. Enable automated stopping by configuring at least one condition. |
Algorithm
The available search algorithms for the Study.
Enums | |
---|---|
ALGORITHM_UNSPECIFIED |
The default algorithm used by Vertex AI for hyperparameter tuning and Vertex AI Vizier. |
GRID_SEARCH |
Simple grid search within the feasible space. To use grid search, all parameters must be INTEGER , CATEGORICAL , or DISCRETE . |
RANDOM_SEARCH |
Simple random search within the feasible space. |
ConvexAutomatedStoppingSpec
Configuration for ConvexAutomatedStoppingSpec. When there are enough completed trials (configured by min_measurement_count), for pending trials with enough measurements and steps, the policy first computes an overestimate of the objective value at max_num_steps according to the slope of the incomplete objective value curve. No prediction can be made if the curve is completely flat. If the overestimation is worse than the best objective value of the completed trials, this pending trial will be early-stopped, but a last measurement will be added to the pending trial with max_num_steps and predicted objective value from the autoregression model.
Fields | |
---|---|
max_ |
Steps used in predicting the final objective for early stopped trials. In general, it's set to be the same as the defined steps in training / tuning. If not defined, it will learn it from the completed trials. When use_steps is false, this field is set to the maximum elapsed seconds. |
min_ |
Minimum number of steps for a trial to complete. Trials which do not have a measurement with step_count > min_step_count won't be considered for early stopping. It's ok to set it to 0, and a trial can be early stopped at any stage. By default, min_step_count is set to be one-tenth of the max_step_count. When use_elapsed_duration is true, this field is set to the minimum elapsed seconds. |
min_ |
The minimal number of measurements in a Trial. Early-stopping checks will not trigger if less than min_measurement_count+1 completed trials or pending trials with less than min_measurement_count measurements. If not defined, the default value is 5. |
learning_ |
The hyper-parameter name used in the tuning job that stands for learning rate. Leave it blank if learning rate is not in a parameter in tuning. The learning_rate is used to estimate the objective value of the ongoing trial. |
use_ |
This bool determines whether or not the rule is applied based on elapsed_secs or steps. If use_elapsed_duration==false, the early stopping decision is made according to the predicted objective values according to the target steps. If use_elapsed_duration==true, elapsed_secs is used instead of steps. Also, in this case, the parameters max_num_steps and min_num_steps are overloaded to contain max_elapsed_seconds and min_elapsed_seconds. |
update_ |
ConvexAutomatedStoppingSpec by default only updates the trials that needs to be early stopped using a newly trained auto-regressive model. When this flag is set to True, all stopped trials from the beginning are potentially updated in terms of their |
DecayCurveAutomatedStoppingSpec
The decay curve automated stopping rule builds a Gaussian Process Regressor to predict the final objective value of a Trial based on the already completed Trials and the intermediate measurements of the current Trial. Early stopping is requested for the current Trial if there is very low probability to exceed the optimal value found so far.
Fields | |
---|---|
use_ |
True if |
MeasurementSelectionType
This indicates which measurement to use if/when the service automatically selects the final measurement from previously reported intermediate measurements. Choose this based on two considerations: A) Do you expect your measurements to monotonically improve? If so, choose LAST_MEASUREMENT. On the other hand, if you're in a situation where your system can "over-train" and you expect the performance to get better for a while but then start declining, choose BEST_MEASUREMENT. B) Are your measurements significantly noisy and/or irreproducible? If so, BEST_MEASUREMENT will tend to be over-optimistic, and it may be better to choose LAST_MEASUREMENT. If both or neither of (A) and (B) apply, it doesn't matter which selection type is chosen.
Enums | |
---|---|
MEASUREMENT_SELECTION_TYPE_UNSPECIFIED |
Will be treated as LAST_MEASUREMENT. |
LAST_MEASUREMENT |
Use the last measurement reported. |
BEST_MEASUREMENT |
Use the best measurement reported. |
MedianAutomatedStoppingSpec
The median automated stopping rule stops a pending Trial if the Trial's best objective_value is strictly below the median 'performance' of all completed Trials reported up to the Trial's last measurement. Currently, 'performance' refers to the running average of the objective values reported by the Trial in each measurement.
Fields | |
---|---|
use_ |
True if median automated stopping rule applies on |
MetricSpec
Represents a metric to optimize.
Fields | |
---|---|
metric_ |
Required. The ID of the metric. Must not contain whitespaces and must be unique amongst all MetricSpecs. |
goal |
Required. The optimization goal of the metric. |
safety_ |
Used for safe search. In the case, the metric will be a safety metric. You must provide a separate metric for objective metric. |
GoalType
The available types of optimization goals.
Enums | |
---|---|
GOAL_TYPE_UNSPECIFIED |
Goal Type will default to maximize. |
MAXIMIZE |
Maximize the goal metric. |
MINIMIZE |
Minimize the goal metric. |
SafetyMetricConfig
Used in safe optimization to specify threshold levels and risk tolerance.
Fields | |
---|---|
safety_ |
Safety threshold (boundary value between safe and unsafe). NOTE that if you leave SafetyMetricConfig unset, a default value of 0 will be used. |
desired_ |
Desired minimum fraction of safe trials (over total number of trials) that should be targeted by the algorithm at any time during the study (best effort). This should be between 0.0 and 1.0 and a value of 0.0 means that there is no minimum and an algorithm proceeds without targeting any specific fraction. A value of 1.0 means that the algorithm attempts to only Suggest safe Trials. |
ObservationNoise
Describes the noise level of the repeated observations.
"Noisy" means that the repeated observations with the same Trial parameters may lead to different metric evaluations.
Enums | |
---|---|
OBSERVATION_NOISE_UNSPECIFIED |
The default noise level chosen by Vertex AI. |
LOW |
Vertex AI assumes that the objective function is (nearly) perfectly reproducible, and will never repeat the same Trial parameters. |
HIGH |
Vertex AI will estimate the amount of noise in metric evaluations, it may repeat the same Trial parameters more than once. |
ParameterSpec
Represents a single parameter to optimize.
Fields | |
---|---|
parameter_ |
Required. The ID of the parameter. Must not contain whitespaces and must be unique amongst all ParameterSpecs. |
scale_ |
How the parameter should be scaled. Leave unset for |
conditional_ |
A conditional parameter node is active if the parameter's value matches the conditional node's parent_value_condition. If two items in conditional_parameter_specs have the same name, they must have disjoint parent_value_condition. |
Union field
|
|
double_ |
The value spec for a 'DOUBLE' parameter. |
integer_ |
The value spec for an 'INTEGER' parameter. |
categorical_ |
The value spec for a 'CATEGORICAL' parameter. |
discrete_ |
The value spec for a 'DISCRETE' parameter. |
CategoricalValueSpec
Value specification for a parameter in CATEGORICAL
type.
Fields | |
---|---|
values[] |
Required. The list of possible categories. |
default_ |
A default value for a Currently only supported by the Vertex AI Vizier service. Not supported by HyperparameterTuningJob or TrainingPipeline. |
ConditionalParameterSpec
Represents a parameter spec with condition from its parent parameter.
Fields | |
---|---|
parameter_ |
Required. The spec for a conditional parameter. |
Union field parent_value_condition . A set of parameter values from the parent ParameterSpec's feasible space. parent_value_condition can be only one of the following: |
|
parent_ |
The spec for matching values from a parent parameter of |
parent_ |
The spec for matching values from a parent parameter of |
parent_ |
The spec for matching values from a parent parameter of |
CategoricalValueCondition
Represents the spec to match categorical values from parent parameter.
Fields | |
---|---|
values[] |
Required. Matches values of the parent parameter of 'CATEGORICAL' type. All values must exist in |
DiscreteValueCondition
Represents the spec to match discrete values from parent parameter.
Fields | |
---|---|
values[] |
Required. Matches values of the parent parameter of 'DISCRETE' type. All values must exist in The Epsilon of the value matching is 1e-10. |
IntValueCondition
Represents the spec to match integer values from parent parameter.
Fields | |
---|---|
values[] |
Required. Matches values of the parent parameter of 'INTEGER' type. All values must lie in |
DiscreteValueSpec
Value specification for a parameter in DISCRETE
type.
Fields | |
---|---|
values[] |
Required. A list of possible values. The list should be in increasing order and at least 1e-10 apart. For instance, this parameter might have possible settings of 1.5, 2.5, and 4.0. This list should not contain more than 1,000 values. |
default_ |
A default value for a Currently only supported by the Vertex AI Vizier service. Not supported by HyperparameterTuningJob or TrainingPipeline. |
DoubleValueSpec
Value specification for a parameter in DOUBLE
type.
Fields | |
---|---|
min_ |
Required. Inclusive minimum value of the parameter. |
max_ |
Required. Inclusive maximum value of the parameter. |
default_ |
A default value for a Currently only supported by the Vertex AI Vizier service. Not supported by HyperparameterTuningJob or TrainingPipeline. |
IntegerValueSpec
Value specification for a parameter in INTEGER
type.
Fields | |
---|---|
min_ |
Required. Inclusive minimum value of the parameter. |
max_ |
Required. Inclusive maximum value of the parameter. |
default_ |
A default value for an Currently only supported by the Vertex AI Vizier service. Not supported by HyperparameterTuningJob or TrainingPipeline. |
ScaleType
The type of scaling that should be applied to this parameter.
Enums | |
---|---|
SCALE_TYPE_UNSPECIFIED |
By default, no scaling is applied. |
UNIT_LINEAR_SCALE |
Scales the feasible space to (0, 1) linearly. |
UNIT_LOG_SCALE |
Scales the feasible space logarithmically to (0, 1). The entire feasible space must be strictly positive. |
UNIT_REVERSE_LOG_SCALE |
Scales the feasible space "reverse" logarithmically to (0, 1). The result is that values close to the top of the feasible space are spread out more than points near the bottom. The entire feasible space must be strictly positive. |
StudyStoppingConfig
The configuration (stopping conditions) for automated stopping of a Study. Conditions include trial budgets, time budgets, and convergence detection.
Fields | |
---|---|
should_ |
If true, a Study enters STOPPING_ASAP whenever it would normally enters STOPPING state. The bottom line is: set to true if you want to interrupt on-going evaluations of Trials as soon as the study stopping condition is met. (Please see Study.State documentation for the source of truth). |
minimum_ |
Each "stopping rule" in this proto specifies an "if" condition. Before Vizier would generate a new suggestion, it first checks each specified stopping rule, from top to bottom in this list. Note that the first few rules (e.g. minimum_runtime_constraint, min_num_trials) will prevent other stopping rules from being evaluated until they are met. For example, setting |
maximum_ |
If the specified time or duration has passed, stop the study. |
min_ |
If there are fewer than this many COMPLETED trials, do not stop the study. |
max_ |
If there are more than this many trials, stop the study. |
max_ |
If the objective value has not improved for this many consecutive trials, stop the study. WARNING: Effective only for single-objective studies. |
max_ |
If the objective value has not improved for this much time, stop the study. WARNING: Effective only for single-objective studies. |
StudyTimeConstraint
Time-based Constraint for Study
Fields | |
---|---|
Union field
|
|
max_ |
Counts the wallclock time passed since the creation of this Study. |
end_ |
Compares the wallclock time to this time. Must use UTC timezone. |
SuggestTrialsMetadata
Details of operations that perform Trials suggestion.
Fields | |
---|---|
generic_ |
Operation metadata for suggesting Trials. |
client_ |
The identifier of the client that is requesting the suggestion. If multiple SuggestTrialsRequests have the same |
SuggestTrialsRequest
Request message for VizierService.SuggestTrials
.
Fields | |
---|---|
parent |
Required. The project and location that the Study belongs to. Format: |
suggestion_ |
Required. The number of suggestions requested. It must be positive. |
client_ |
Required. The identifier of the client that is requesting the suggestion. If multiple SuggestTrialsRequests have the same |
contexts[] |
Optional. This allows you to specify the "context" for a Trial; a context is a slice (a subspace) of the search space. Typical uses for contexts: 1) You are using Vizier to tune a server for best performance, but there's a strong weekly cycle. The context specifies the day-of-week. This allows Tuesday to generalize from Wednesday without assuming that everything is identical. 2) Imagine you're optimizing some medical treatment for people. As they walk in the door, you know certain facts about them (e.g. sex, weight, height, blood-pressure). Put that information in the context, and Vizier will adapt its suggestions to the patient. 3) You want to do a fair A/B test efficiently. Specify the "A" and "B" conditions as contexts, and Vizier will generalize between "A" and "B" conditions. If they are similar, this will allow Vizier to converge to the optimum faster than if "A" and "B" were separate Studies. NOTE: You can also enter contexts as REQUESTED Trials, e.g. via the CreateTrial() RPC; that's the asynchronous option where you don't need a close association between contexts and suggestions. NOTE: All the Parameters you set in a context MUST be defined in the Study. NOTE: You must supply 0 or $suggestion_count contexts. If you don't supply any contexts, Vizier will make suggestions from the full search space specified in the StudySpec; if you supply a full set of context, each suggestion will match the corresponding context. NOTE: A Context with no features set matches anything, and allows suggestions from the full search space. NOTE: Contexts MUST lie within the search space specified in the StudySpec. It's an error if they don't. NOTE: Contexts preferentially match ACTIVE then REQUESTED trials before new suggestions are generated. NOTE: Generation of suggestions involves a match between a Context and (optionally) a REQUESTED trial; if that match is not fully specified, a suggestion will be geneated in the merged subspace. |
SuggestTrialsResponse
Response message for VizierService.SuggestTrials
.
Fields | |
---|---|
trials[] |
A list of Trials. |
study_ |
The state of the Study. |
start_ |
The time at which the operation was started. |
end_ |
The time at which operation processing completed. |
SummarizationHelpfulnessInput
Input for summarization helpfulness metric.
Fields | |
---|---|
metric_ |
Required. Spec for summarization helpfulness score metric. |
instance |
Required. Summarization helpfulness instance. |
SummarizationHelpfulnessInstance
Spec for summarization helpfulness instance.
Fields | |
---|---|
prediction |
Required. Output of the evaluated model. |
reference |
Optional. Ground truth used to compare against the prediction. |
context |
Required. Text to be summarized. |
instruction |
Optional. Summarization prompt for LLM. |
SummarizationHelpfulnessResult
Spec for summarization helpfulness result.
Fields | |
---|---|
explanation |
Output only. Explanation for summarization helpfulness score. |
score |
Output only. Summarization Helpfulness score. |
confidence |
Output only. Confidence for summarization helpfulness score. |
SummarizationHelpfulnessSpec
Spec for summarization helpfulness score metric.
Fields | |
---|---|
use_ |
Optional. Whether to use instance.reference to compute summarization helpfulness. |
version |
Optional. Which version to use for evaluation. |
SummarizationQualityInput
Input for summarization quality metric.
Fields | |
---|---|
metric_ |
Required. Spec for summarization quality score metric. |
instance |
Required. Summarization quality instance. |
SummarizationQualityInstance
Spec for summarization quality instance.
Fields | |
---|---|
prediction |
Required. Output of the evaluated model. |
reference |
Optional. Ground truth used to compare against the prediction. |
context |
Required. Text to be summarized. |
instruction |
Required. Summarization prompt for LLM. |
SummarizationQualityResult
Spec for summarization quality result.
Fields | |
---|---|
explanation |
Output only. Explanation for summarization quality score. |
score |
Output only. Summarization Quality score. |
confidence |
Output only. Confidence for summarization quality score. |
SummarizationQualitySpec
Spec for summarization quality score metric.
Fields | |
---|---|
use_ |
Optional. Whether to use instance.reference to compute summarization quality. |
version |
Optional. Which version to use for evaluation. |
SummarizationVerbosityInput
Input for summarization verbosity metric.
Fields | |
---|---|
metric_ |
Required. Spec for summarization verbosity score metric. |
instance |
Required. Summarization verbosity instance. |
SummarizationVerbosityInstance
Spec for summarization verbosity instance.
Fields | |
---|---|
prediction |
Required. Output of the evaluated model. |
reference |
Optional. Ground truth used to compare against the prediction. |
context |
Required. Text to be summarized. |
instruction |
Optional. Summarization prompt for LLM. |
SummarizationVerbosityResult
Spec for summarization verbosity result.
Fields | |
---|---|
explanation |
Output only. Explanation for summarization verbosity score. |
score |
Output only. Summarization Verbosity score. |
confidence |
Output only. Confidence for summarization verbosity score. |
SummarizationVerbositySpec
Spec for summarization verbosity score metric.
Fields | |
---|---|
use_ |
Optional. Whether to use instance.reference to compute summarization verbosity. |
version |
Optional. Which version to use for evaluation. |
SupervisedHyperParameters
Hyperparameters for SFT.
Fields | |
---|---|
epoch_ |
Optional. Number of complete passes the model makes over the entire training dataset during training. |
learning_ |
Optional. Multiplier for adjusting the default learning rate. |
adapter_ |
Optional. Adapter size for tuning. |
AdapterSize
Supported adapter sizes for tuning.
Enums | |
---|---|
ADAPTER_SIZE_UNSPECIFIED |
Adapter size is unspecified. |
ADAPTER_SIZE_ONE |
Adapter size 1. |
ADAPTER_SIZE_FOUR |
Adapter size 4. |
ADAPTER_SIZE_EIGHT |
Adapter size 8. |
ADAPTER_SIZE_SIXTEEN |
Adapter size 16. |
ADAPTER_SIZE_THIRTY_TWO |
Adapter size 32. |
SupervisedTuningDataStats
Tuning data statistics for Supervised Tuning.
Fields | |
---|---|
tuning_ |
Output only. Number of examples in the tuning dataset. |
total_ |
Output only. Number of tuning characters in the tuning dataset. |
total_billable_character_count |
Output only. Number of billable characters in the tuning dataset. |
total_ |
Output only. Number of billable tokens in the tuning dataset. |
tuning_ |
Output only. Number of tuning steps for this Tuning Job. |
user_ |
Output only. Dataset distributions for the user input tokens. |
user_ |
Output only. Dataset distributions for the user output tokens. |
user_ |
Output only. Dataset distributions for the messages per example. |
user_ |
Output only. Sample user messages in the training dataset uri. |
total_ |
The number of examples in the dataset that have been truncated by any amount. |
truncated_ |
A partial sample of the indices (starting from 1) of the truncated examples. |
SupervisedTuningDatasetDistribution
Dataset distribution for Supervised Tuning.
Fields | |
---|---|
sum |
Output only. Sum of a given population of values. |
billable_ |
Output only. Sum of a given population of values that are billable. |
min |
Output only. The minimum of the population values. |
max |
Output only. The maximum of the population values. |
mean |
Output only. The arithmetic mean of the values in the population. |
median |
Output only. The median of the values in the population. |
p5 |
Output only. The 5th percentile of the values in the population. |
p95 |
Output only. The 95th percentile of the values in the population. |
buckets[] |
Output only. Defines the histogram bucket. |
DatasetBucket
Dataset bucket used to create a histogram for the distribution given a population of values.
Fields | |
---|---|
count |
Output only. Number of values in the bucket. |
left |
Output only. Left bound of the bucket. |
right |
Output only. Right bound of the bucket. |
SupervisedTuningSpec
Tuning Spec for Supervised Tuning for first party models.
Fields | |
---|---|
training_ |
Required. Cloud Storage path to file containing training dataset for tuning. The dataset must be formatted as a JSONL file. |
validation_ |
Optional. Cloud Storage path to file containing validation dataset for tuning. The dataset must be formatted as a JSONL file. |
hyper_ |
Optional. Hyperparameters for SFT. |
SyncFeatureViewRequest
Request message for FeatureOnlineStoreAdminService.SyncFeatureView
.
Fields | |
---|---|
feature_ |
Required. Format: |
SyncFeatureViewResponse
Response message for FeatureOnlineStoreAdminService.SyncFeatureView
.
Fields | |
---|---|
feature_ |
Format: |
TFRecordDestination
The storage details for TFRecord output content.
Fields | |
---|---|
gcs_ |
Required. Google Cloud Storage location. |
Tensor
A tensor value type.
Fields | |
---|---|
dtype |
The data type of tensor. |
shape[] |
Shape of the tensor. |
bool_ |
Type specific representations that make it easy to create tensor protos in all languages. Only the representation corresponding to "dtype" can be set. The values hold the flattened representation of the tensor in row major order. |
string_ |
|
bytes_ |
|
float_ |
|
double_ |
|
int_ |
|
int64_ |
|
uint_ |
|
uint64_ |
|
list_ |
A list of tensor values. |
struct_ |
A map of string to tensor. |
tensor_ |
Serialized raw tensor content. |
DataType
Data type of the tensor.
Enums | |
---|---|
DATA_TYPE_UNSPECIFIED |
Not a legal value for DataType. Used to indicate a DataType field has not been set. |
BOOL |
Data types that all computation devices are expected to be capable to support. |
STRING |
|
FLOAT |
|
DOUBLE |
|
INT8 |
|
INT16 |
|
INT32 |
|
INT64 |
|
UINT8 |
|
UINT16 |
|
UINT32 |
|
UINT64 |
Tensorboard
Tensorboard is a physical database that stores users' training metrics. A default Tensorboard is provided in each region of a Google Cloud project. If needed users can also create extra Tensorboards in their projects.
Fields | |
---|---|
name |
Output only. Name of the Tensorboard. Format: |
display_ |
Required. User provided name of this Tensorboard. |
description |
Description of this Tensorboard. |
encryption_ |
Customer-managed encryption key spec for a Tensorboard. If set, this Tensorboard and all sub-resources of this Tensorboard will be secured by this key. |
blob_ |
Output only. Consumer project Cloud Storage path prefix used to store blob data, which can either be a bucket or directory. Does not end with a '/'. |
run_ |
Output only. The number of Runs stored in this Tensorboard. |
create_ |
Output only. Timestamp when this Tensorboard was created. |
update_ |
Output only. Timestamp when this Tensorboard was last updated. |
labels |
The labels with user-defined metadata to organize your Tensorboards. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Tensorboard (System labels are excluded). See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. |
etag |
Used to perform a consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
is_ |
Used to indicate if the TensorBoard instance is the default one. Each project & region can have at most one default TensorBoard instance. Creation of a default TensorBoard instance and updating an existing TensorBoard instance to be default will mark all other TensorBoard instances (if any) as non default. |
satisfies_ |
Output only. Reserved for future use. |
satisfies_ |
Output only. Reserved for future use. |
TensorboardBlob
One blob (e.g, image, graph) viewable on a blob metric plot.
Fields | |
---|---|
id |
Output only. A URI safe key uniquely identifying a blob. Can be used to locate the blob stored in the Cloud Storage bucket of the consumer project. |
data |
Optional. The bytes of the blob is not present unless it's returned by the ReadTensorboardBlobData endpoint. |
TensorboardBlobSequence
One point viewable on a blob metric plot, but mostly just a wrapper message to work around repeated fields can't be used directly within oneof
fields.
Fields | |
---|---|
values[] |
List of blobs contained within the sequence. |
TensorboardExperiment
A TensorboardExperiment is a group of TensorboardRuns, that are typically the results of a training job run, in a Tensorboard.
Fields | |
---|---|
name |
Output only. Name of the TensorboardExperiment. Format: |
display_ |
User provided name of this TensorboardExperiment. |
description |
Description of this TensorboardExperiment. |
create_ |
Output only. Timestamp when this TensorboardExperiment was created. |
update_ |
Output only. Timestamp when this TensorboardExperiment was last updated. |
labels |
The labels with user-defined metadata to organize your TensorboardExperiment. Label keys and values cannot be longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Dataset (System labels are excluded). See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with
|
etag |
Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
source |
Immutable. Source of the TensorboardExperiment. Example: a custom training job. |
TensorboardRun
TensorboardRun maps to a specific execution of a training job with a given set of hyperparameter values, model definition, dataset, etc
Fields | |
---|---|
name |
Output only. Name of the TensorboardRun. Format: |
display_ |
Required. User provided name of this TensorboardRun. This value must be unique among all TensorboardRuns belonging to the same parent TensorboardExperiment. |
description |
Description of this TensorboardRun. |
create_ |
Output only. Timestamp when this TensorboardRun was created. |
update_ |
Output only. Timestamp when this TensorboardRun was last updated. |
labels |
The labels with user-defined metadata to organize your TensorboardRuns. This field will be used to filter and visualize Runs in the Tensorboard UI. For example, a Vertex AI training job can set a label aiplatform.googleapis.com/training_job_id=xxxxx to all the runs created within that job. An end user can set a label experiment_id=xxxxx for all the runs produced in a Jupyter notebook. These runs can be grouped by a label value and visualized together in the Tensorboard UI. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one TensorboardRun (System labels are excluded). See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. |
etag |
Used to perform a consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
TensorboardTensor
One point viewable on a tensor metric plot.
Fields | |
---|---|
value |
Required. Serialized form of https://github.com/tensorflow/tensorflow/blob/master/tensorflow/core/framework/tensor.proto |
version_ |
Optional. Version number of TensorProto used to serialize |
TensorboardTimeSeries
TensorboardTimeSeries maps to times series produced in training runs
Fields | |
---|---|
name |
Output only. Name of the TensorboardTimeSeries. |
display_ |
Required. User provided name of this TensorboardTimeSeries. This value should be unique among all TensorboardTimeSeries resources belonging to the same TensorboardRun resource (parent resource). |
description |
Description of this TensorboardTimeSeries. |
value_ |
Required. Immutable. Type of TensorboardTimeSeries value. |
create_ |
Output only. Timestamp when this TensorboardTimeSeries was created. |
update_ |
Output only. Timestamp when this TensorboardTimeSeries was last updated. |
etag |
Used to perform a consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
plugin_ |
Immutable. Name of the plugin this time series pertain to. Such as Scalar, Tensor, Blob |
plugin_ |
Data of the current plugin, with the size limited to 65KB. |
metadata |
Output only. Scalar, Tensor, or Blob metadata for this TensorboardTimeSeries. |
Metadata
Describes metadata for a TensorboardTimeSeries.
Fields | |
---|---|
max_ |
Output only. Max step index of all data points within a TensorboardTimeSeries. |
max_ |
Output only. Max wall clock timestamp of all data points within a TensorboardTimeSeries. |
max_ |
Output only. The largest blob sequence length (number of blobs) of all data points in this time series, if its ValueType is BLOB_SEQUENCE. |
ValueType
An enum representing the value type of a TensorboardTimeSeries.
Enums | |
---|---|
VALUE_TYPE_UNSPECIFIED |
The value type is unspecified. |
SCALAR |
Used for TensorboardTimeSeries that is a list of scalars. E.g. accuracy of a model over epochs/time. |
TENSOR |
Used for TensorboardTimeSeries that is a list of tensors. E.g. histograms of weights of layer in a model over epoch/time. |
BLOB_SEQUENCE |
Used for TensorboardTimeSeries that is a list of blob sequences. E.g. set of sample images with labels over epochs/time. |
ThresholdConfig
The config for feature monitoring threshold.
Fields | |
---|---|
Union field
|
|
value |
Specify a threshold value that can trigger the alert. If this threshold config is for feature distribution distance: 1. For categorical feature, the distribution distance is calculated by L-inifinity norm. 2. For numerical feature, the distribution distance is calculated by Jensen–Shannon divergence. Each feature must have a non-zero threshold if they need to be monitored. Otherwise no alert will be triggered for that feature. |
TimeSeriesData
All the data stored in a TensorboardTimeSeries.
Fields | |
---|---|
tensorboard_ |
Required. The ID of the TensorboardTimeSeries, which will become the final component of the TensorboardTimeSeries' resource name |
value_ |
Required. Immutable. The value type of this time series. All the values in this time series data must match this value type. |
values[] |
Required. Data points in this time series. |
TimeSeriesDataPoint
A TensorboardTimeSeries data point.
Fields | |
---|---|
wall_ |
Wall clock timestamp when this data point is generated by the end user. |
step |
Step index of this data point within the run. |
Union field value . Value of this time series data point. value can be only one of the following: |
|
scalar |
A scalar value. |
tensor |
A tensor value. |
blobs |
A blob sequence value. |
TimestampSplit
Assigns input data to training, validation, and test sets based on a provided timestamps. The youngest data pieces are assigned to training set, next to validation set, and the oldest to the test set.
Supported only for tabular Datasets.
Fields | |
---|---|
training_ |
The fraction of the input data that is to be used to train the Model. |
validation_ |
The fraction of the input data that is to be used to validate the Model. |
test_ |
The fraction of the input data that is to be used to evaluate the Model. |
key |
Required. The key is a name of one of the Dataset's data columns. The values of the key (the values in the column) must be in RFC 3339 |
TokensInfo
Tokens info with a list of tokens and the corresponding list of token ids.
Fields | |
---|---|
tokens[] |
A list of tokens from the input. |
token_ |
A list of token ids from the input. |
role |
Optional. Optional fields for the role from the corresponding Content. |
Tool
Tool details that the model may use to generate response.
A Tool
is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. A Tool object should contain exactly one type of Tool (e.g FunctionDeclaration, Retrieval or GoogleSearchRetrieval).
Fields | |
---|---|
function_ |
Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating |
retrieval |
Optional. Retrieval tool type. System will always execute the provided retrieval tool(s) to get external knowledge to answer the prompt. Retrieval results are presented to the model for generation. |
google_ |
Optional. GoogleSearchRetrieval tool type. Specialized retrieval tool that is powered by Google search. |
ToolCall
Spec for tool call.
Fields | |
---|---|
tool_ |
Required. Spec for tool name |
tool_ |
Optional. Spec for tool input |
ToolCallValidInput
Input for tool call valid metric.
Fields | |
---|---|
metric_ |
Required. Spec for tool call valid metric. |
instances[] |
Required. Repeated tool call valid instances. |
ToolCallValidInstance
Spec for tool call valid instance.
Fields | |
---|---|
prediction |
Required. Output of the evaluated model. |
reference |
Required. Ground truth used to compare against the prediction. |
ToolCallValidMetricValue
Tool call valid metric value for an instance.
Fields | |
---|---|
score |
Output only. Tool call valid score. |
ToolCallValidResults
Results for tool call valid metric.
Fields | |
---|---|
tool_ |
Output only. Tool call valid metric values. |
ToolCallValidSpec
This type has no fields.
Spec for tool call valid metric.
ToolConfig
Tool config. This config is shared for all tools provided in the request.
Fields | |
---|---|
function_ |
Optional. Function calling config. |
ToolNameMatchInput
Input for tool name match metric.
Fields | |
---|---|
metric_ |
Required. Spec for tool name match metric. |
instances[] |
Required. Repeated tool name match instances. |
ToolNameMatchInstance
Spec for tool name match instance.
Fields | |
---|---|
prediction |
Required. Output of the evaluated model. |
reference |
Required. Ground truth used to compare against the prediction. |
ToolNameMatchMetricValue
Tool name match metric value for an instance.
Fields | |
---|---|
score |
Output only. Tool name match score. |
ToolNameMatchResults
Results for tool name match metric.
Fields | |
---|---|
tool_ |
Output only. Tool name match metric values. |
ToolNameMatchSpec
This type has no fields.
Spec for tool name match metric.
ToolParameterKVMatchInput
Input for tool parameter key value match metric.
Fields | |
---|---|
metric_ |
Required. Spec for tool parameter key value match metric. |
instances[] |
Required. Repeated tool parameter key value match instances. |
ToolParameterKVMatchInstance
Spec for tool parameter key value match instance.
Fields | |
---|---|
prediction |
Required. Output of the evaluated model. |
reference |
Required. Ground truth used to compare against the prediction. |
ToolParameterKVMatchMetricValue
Tool parameter key value match metric value for an instance.
Fields | |
---|---|
score |
Output only. Tool parameter key value match score. |
ToolParameterKVMatchResults
Results for tool parameter key value match metric.
Fields | |
---|---|
tool_ |
Output only. Tool parameter key value match metric values. |
ToolParameterKVMatchSpec
Spec for tool parameter key value match metric.
Fields | |
---|---|
use_ |
Optional. Whether to use STRICT string match on parameter values. |
ToolParameterKeyMatchInput
Input for tool parameter key match metric.
Fields | |
---|---|
metric_ |
Required. Spec for tool parameter key match metric. |
instances[] |
Required. Repeated tool parameter key match instances. |
ToolParameterKeyMatchInstance
Spec for tool parameter key match instance.
Fields | |
---|---|
prediction |
Required. Output of the evaluated model. |
reference |
Required. Ground truth used to compare against the prediction. |
ToolParameterKeyMatchMetricValue
Tool parameter key match metric value for an instance.
Fields | |
---|---|
score |
Output only. Tool parameter key match score. |
ToolParameterKeyMatchResults
Results for tool parameter key match metric.
Fields | |
---|---|
tool_ |
Output only. Tool parameter key match metric values. |
ToolParameterKeyMatchSpec
This type has no fields.
Spec for tool parameter key match metric.
TrainingPipeline
The TrainingPipeline orchestrates tasks associated with training a Model. It always executes the training task, and optionally may also export data from Vertex AI's Dataset which becomes the training input, upload
the Model to Vertex AI, and evaluate the Model.
Fields | |
---|---|
name |
Output only. Resource name of the TrainingPipeline. |
display_ |
Required. The user-defined name of this TrainingPipeline. |
input_ |
Specifies Vertex AI owned input data that may be used for training the Model. The TrainingPipeline's |
training_ |
Required. A Google Cloud Storage path to the YAML file that defines the training task which is responsible for producing the model artifact, and may also include additional auxiliary work. The definition files that can be used here are found in gs://google-cloud-aiplatform/schema/trainingjob/definition/. Note: The URI given on output will be immutable and probably different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access. |
training_ |
Required. The training task's parameter(s), as specified in the |
training_ |
Output only. The metadata information as specified in the |
model_ |
Describes the Model that may be uploaded (via |
model_ |
Optional. The ID to use for the uploaded Model, which will become the final component of the model resource name. This value may be up to 63 characters, and valid characters are |
parent_ |
Optional. When specify this field, the |
state |
Output only. The detailed state of the pipeline. |
error |
Output only. Only populated when the pipeline's state is |
create_ |
Output only. Time when the TrainingPipeline was created. |
start_ |
Output only. Time when the TrainingPipeline for the first time entered the |
end_ |
Output only. Time when the TrainingPipeline entered any of the following states: |
update_ |
Output only. Time when the TrainingPipeline was most recently updated. |
labels |
The labels with user-defined metadata to organize TrainingPipelines. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. |
encryption_ |
Customer-managed encryption key spec for a TrainingPipeline. If set, this TrainingPipeline will be secured by this key. Note: Model trained by this TrainingPipeline is also secured by this key if |
Trajectory
Spec for trajectory.
Fields | |
---|---|
tool_ |
Required. Tool calls in the trajectory. |
TrajectoryAnyOrderMatchInput
Instances and metric spec for TrajectoryAnyOrderMatch metric.
Fields | |
---|---|
metric_ |
Required. Spec for TrajectoryAnyOrderMatch metric. |
instances[] |
Required. Repeated TrajectoryAnyOrderMatch instance. |
TrajectoryAnyOrderMatchInstance
Spec for TrajectoryAnyOrderMatch instance.
Fields | |
---|---|
predicted_ |
Required. Spec for predicted tool call trajectory. |
reference_ |
Required. Spec for reference tool call trajectory. |
TrajectoryAnyOrderMatchMetricValue
TrajectoryAnyOrderMatch metric value for an instance.
Fields | |
---|---|
score |
Output only. TrajectoryAnyOrderMatch score. |
TrajectoryAnyOrderMatchResults
Results for TrajectoryAnyOrderMatch metric.
Fields | |
---|---|
trajectory_ |
Output only. TrajectoryAnyOrderMatch metric values. |
TrajectoryAnyOrderMatchSpec
This type has no fields.
Spec for TrajectoryAnyOrderMatch metric - returns 1 if all tool calls in the reference trajectory appear in the predicted trajectory in any order, else 0.
TrajectoryExactMatchInput
Instances and metric spec for TrajectoryExactMatch metric.
Fields | |
---|---|
metric_ |
Required. Spec for TrajectoryExactMatch metric. |
instances[] |
Required. Repeated TrajectoryExactMatch instance. |
TrajectoryExactMatchInstance
Spec for TrajectoryExactMatch instance.
Fields | |
---|---|
predicted_ |
Required. Spec for predicted tool call trajectory. |
reference_ |
Required. Spec for reference tool call trajectory. |
TrajectoryExactMatchMetricValue
TrajectoryExactMatch metric value for an instance.
Fields | |
---|---|
score |
Output only. TrajectoryExactMatch score. |
TrajectoryExactMatchResults
Results for TrajectoryExactMatch metric.
Fields | |
---|---|
trajectory_ |
Output only. TrajectoryExactMatch metric values. |
TrajectoryExactMatchSpec
This type has no fields.
Spec for TrajectoryExactMatch metric - returns 1 if tool calls in the reference trajectory exactly match the predicted trajectory, else 0.
TrajectoryInOrderMatchInput
Instances and metric spec for TrajectoryInOrderMatch metric.
Fields | |
---|---|
metric_ |
Required. Spec for TrajectoryInOrderMatch metric. |
instances[] |
Required. Repeated TrajectoryInOrderMatch instance. |
TrajectoryInOrderMatchInstance
Spec for TrajectoryInOrderMatch instance.
Fields | |
---|---|
predicted_ |
Required. Spec for predicted tool call trajectory. |
reference_ |
Required. Spec for reference tool call trajectory. |
TrajectoryInOrderMatchMetricValue
TrajectoryInOrderMatch metric value for an instance.
Fields | |
---|---|
score |
Output only. TrajectoryInOrderMatch score. |
TrajectoryInOrderMatchResults
Results for TrajectoryInOrderMatch metric.
Fields | |
---|---|
trajectory_ |
Output only. TrajectoryInOrderMatch metric values. |
TrajectoryInOrderMatchSpec
This type has no fields.
Spec for TrajectoryInOrderMatch metric - returns 1 if tool calls in the reference trajectory appear in the predicted trajectory in the same order, else 0.
TrajectoryPrecisionInput
Instances and metric spec for TrajectoryPrecision metric.
Fields | |
---|---|
metric_ |
Required. Spec for TrajectoryPrecision metric. |
instances[] |
Required. Repeated TrajectoryPrecision instance. |
TrajectoryPrecisionInstance
Spec for TrajectoryPrecision instance.
Fields | |
---|---|
predicted_ |
Required. Spec for predicted tool call trajectory. |
reference_ |
Required. Spec for reference tool call trajectory. |
TrajectoryPrecisionMetricValue
TrajectoryPrecision metric value for an instance.
Fields | |
---|---|
score |
Output only. TrajectoryPrecision score. |
TrajectoryPrecisionResults
Results for TrajectoryPrecision metric.
Fields | |
---|---|
trajectory_ |
Output only. TrajectoryPrecision metric values. |
TrajectoryPrecisionSpec
This type has no fields.
Spec for TrajectoryPrecision metric - returns a float score based on average precision of individual tool calls.
TrajectoryRecallInput
Instances and metric spec for TrajectoryRecall metric.
Fields | |
---|---|
metric_ |
Required. Spec for TrajectoryRecall metric. |
instances[] |
Required. Repeated TrajectoryRecall instance. |
TrajectoryRecallInstance
Spec for TrajectoryRecall instance.
Fields | |
---|---|
predicted_ |
Required. Spec for predicted tool call trajectory. |
reference_ |
Required. Spec for reference tool call trajectory. |
TrajectoryRecallMetricValue
TrajectoryRecall metric value for an instance.
Fields | |
---|---|
score |
Output only. TrajectoryRecall score. |
TrajectoryRecallResults
Results for TrajectoryRecall metric.
Fields | |
---|---|
trajectory_ |
Output only. TrajectoryRecall metric values. |
TrajectoryRecallSpec
This type has no fields.
Spec for TrajectoryRecall metric - returns a float score based on average recall of individual tool calls.
TrajectorySingleToolUseInput
Instances and metric spec for TrajectorySingleToolUse metric.
Fields | |
---|---|
metric_ |
Required. Spec for TrajectorySingleToolUse metric. |
instances[] |
Required. Repeated TrajectorySingleToolUse instance. |
TrajectorySingleToolUseInstance
Spec for TrajectorySingleToolUse instance.
Fields | |
---|---|
predicted_ |
Required. Spec for predicted tool call trajectory. |
TrajectorySingleToolUseMetricValue
TrajectorySingleToolUse metric value for an instance.
Fields | |
---|---|
score |
Output only. TrajectorySingleToolUse score. |
TrajectorySingleToolUseResults
Results for TrajectorySingleToolUse metric.
Fields | |
---|---|
trajectory_ |
Output only. TrajectorySingleToolUse metric values. |
TrajectorySingleToolUseSpec
Spec for TrajectorySingleToolUse metric - returns 1 if tool is present in the predicted trajectory, else 0.
Fields | |
---|---|
tool_ |
Required. Spec for tool name to be checked for in the predicted trajectory. |
Trial
A message representing a Trial. A Trial contains a unique set of Parameters that has been or will be evaluated, along with the objective metrics got by running the Trial.
Fields | |
---|---|
name |
Output only. Resource name of the Trial assigned by the service. |
id |
Output only. The identifier of the Trial assigned by the service. |
state |
Output only. The detailed state of the Trial. |
parameters[] |
Output only. The parameters of the Trial. |
final_ |
Output only. The final measurement containing the objective value. |
measurements[] |
Output only. A list of measurements that are strictly lexicographically ordered by their induced tuples (steps, elapsed_duration). These are used for early stopping computations. |
start_ |
Output only. Time when the Trial was started. |
end_ |
Output only. Time when the Trial's status changed to |
client_ |
Output only. The identifier of the client that originally requested this Trial. Each client is identified by a unique client_id. When a client asks for a suggestion, Vertex AI Vizier will assign it a Trial. The client should evaluate the Trial, complete it, and report back to Vertex AI Vizier. If suggestion is asked again by same client_id before the Trial is completed, the same Trial will be returned. Multiple clients with different client_ids can ask for suggestions simultaneously, each of them will get their own Trial. |
infeasible_ |
Output only. A human readable string describing why the Trial is infeasible. This is set only if Trial state is |
custom_ |
Output only. The CustomJob name linked to the Trial. It's set for a HyperparameterTuningJob's Trial. |
web_ |
Output only. URIs for accessing interactive shells (one URI for each training node). Only available if this trial is part of a The keys are names of each node used for the trial; for example, The values are the URIs for each node's interactive shell. |
Parameter
A message representing a parameter to be tuned.
Fields | |
---|---|
parameter_ |
Output only. The ID of the parameter. The parameter should be defined in |
value |
Output only. The value of the parameter. |
State
Describes a Trial state.
Enums | |
---|---|
STATE_UNSPECIFIED |
The Trial state is unspecified. |
REQUESTED |
Indicates that a specific Trial has been requested, but it has not yet been suggested by the service. |
ACTIVE |
Indicates that the Trial has been suggested. |
STOPPING |
Indicates that the Trial should stop according to the service. |
SUCCEEDED |
Indicates that the Trial is completed successfully. |
INFEASIBLE |
Indicates that the Trial should not be attempted again. The service will set a Trial to INFEASIBLE when it's done but missing the final_measurement. |
TrialContext
Fields | |
---|---|
description |
A human-readable field which can store a description of this context. This will become part of the resulting Trial's description field. |
parameters[] |
If/when a Trial is generated or selected from this Context, its Parameters will match any parameters specified here. (I.e. if this context specifies parameter name:'a' int_value:3, then a resulting Trial will have int_value:3 for its parameter named 'a'.) Note that we first attempt to match existing REQUESTED Trials with contexts, and if there are no matches, we generate suggestions in the subspace defined by the parameters specified here. NOTE: a Context without any Parameters matches the entire feasible search space. |
TunedModel
The Model Registry Model and Online Prediction Endpoint assiociated with this TuningJob
.
Fields | |
---|---|
model |
Output only. The resource name of the TunedModel. Format: |
endpoint |
Output only. A resource name of an Endpoint. Format: |
TunedModelRef
TunedModel Reference for legacy model migration.
Fields | |
---|---|
Union field tuned_model_ref . The Tuned Model Reference for the model. tuned_model_ref can be only one of the following: |
|
tuned_ |
Support migration from model registry. |
tuning_ |
Support migration from tuning job list page, from gemini-1.0-pro-002 to 1.5 and above. |
pipeline_ |
Support migration from tuning job list page, from bison model to gemini model. |
TuningDataStats
The tuning data statistic values for TuningJob
.
Fields | |
---|---|
Union field
|
|
supervised_ |
The SFT Tuning data stats. |
TuningJob
Represents a TuningJob that runs with Google owned models.
Fields | |
---|---|
name |
Output only. Identifier. Resource name of a TuningJob. Format: |
tuned_ |
Optional. The display name of the |
description |
Optional. The description of the |
state |
Output only. The detailed state of the job. |
create_ |
Output only. Time when the |
start_ |
Output only. Time when the |
end_ |
Output only. Time when the TuningJob entered any of the following |
update_ |
Output only. Time when the |
error |
Output only. Only populated when job's state is |
labels |
Optional. The labels with user-defined metadata to organize Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. |
experiment |
Output only. The Experiment associated with this |
tuned_ |
Output only. The tuned model resources assiociated with this |
tuning_ |
Output only. The tuning data statistics associated with this |
encryption_ |
Customer-managed encryption key options for a TuningJob. If this is set, then all resources created by the TuningJob will be encrypted with the provided encryption key. |
service_ |
The service account that the tuningJob workload runs as. If not specified, the Vertex AI Secure Fine-Tuned Service Agent in the project will be used. See https://cloud.google.com/iam/docs/service-agents#vertex-ai-secure-fine-tuning-service-agent Users starting the pipeline must have the |
Union field
|
|
base_ |
The base model that is being tuned, e.g., "gemini-1.0-pro-002". . |
Union field
|
|
supervised_ |
Tuning Spec for Supervised Fine Tuning. |
Type
Type contains the list of OpenAPI data types as defined by https://swagger.io/docs/specification/data-models/data-types/
Enums | |
---|---|
TYPE_UNSPECIFIED |
Not specified, should not be used. |
STRING |
OpenAPI string type |
NUMBER |
OpenAPI number type |
INTEGER |
OpenAPI integer type |
BOOLEAN |
OpenAPI boolean type |
ARRAY |
OpenAPI array type |
OBJECT |
OpenAPI object type |
UndeployIndexOperationMetadata
Runtime operation information for IndexEndpointService.UndeployIndex
.
Fields | |
---|---|
generic_ |
The operation generic information. |
UndeployIndexRequest
Request message for IndexEndpointService.UndeployIndex
.
Fields | |
---|---|
index_ |
Required. The name of the IndexEndpoint resource from which to undeploy an Index. Format: |
deployed_ |
Required. The ID of the DeployedIndex to be undeployed from the IndexEndpoint. |
UndeployIndexResponse
This type has no fields.
Response message for IndexEndpointService.UndeployIndex
.
UndeployModelOperationMetadata
Runtime operation information for EndpointService.UndeployModel
.
Fields | |
---|---|
generic_ |
The operation generic information. |
UndeployModelRequest
Request message for EndpointService.UndeployModel
.
Fields | |
---|---|
endpoint |
Required. The name of the Endpoint resource from which to undeploy a Model. Format: |
deployed_ |
Required. The ID of the DeployedModel to be undeployed from the Endpoint. |
traffic_ |
If this field is provided, then the Endpoint's |
UndeployModelResponse
This type has no fields.
Response message for EndpointService.UndeployModel
.
UnmanagedContainerModel
Contains model information necessary to perform batch prediction without requiring a full model import.
Fields | |
---|---|
artifact_ |
The path to the directory containing the Model artifact and any of its supporting files. |
predict_ |
Contains the schemata used in Model's predictions and explanations |
container_ |
Input only. The specification of the container that is to be used when deploying this Model. |
UpdateArtifactRequest
Request message for MetadataService.UpdateArtifact
.
Fields | |
---|---|
artifact |
Required. The Artifact containing updates. The Artifact's |
update_ |
Optional. A FieldMask indicating which fields should be updated. |
allow_ |
If set to true, and the |
UpdateContextRequest
Request message for MetadataService.UpdateContext
.
Fields | |
---|---|
context |
Required. The Context containing updates. The Context's |
update_ |
Optional. A FieldMask indicating which fields should be updated. |
allow_ |
If set to true, and the |
UpdateDatasetRequest
Request message for DatasetService.UpdateDataset
.
Fields | |
---|---|
dataset |
Required. The Dataset which replaces the resource on the server. |
update_ |
Required. The update mask applies to the resource. For the
|
UpdateDatasetVersionRequest
Request message for DatasetService.UpdateDatasetVersion
.
Fields | |
---|---|
dataset_ |
Required. The DatasetVersion which replaces the resource on the server. |
update_ |
Required. The update mask applies to the resource. For the
|
UpdateDeploymentResourcePoolOperationMetadata
Runtime operation information for UpdateDeploymentResourcePool method.
Fields | |
---|---|
generic_ |
The operation generic information. |
UpdateDeploymentResourcePoolRequest
Request message for UpdateDeploymentResourcePool method.
Fields | |
---|---|
deployment_ |
Required. The DeploymentResourcePool to update. The DeploymentResourcePool's |
update_ |
Required. The list of fields to update. |
UpdateEndpointLongRunningRequest
Request message for EndpointService.UpdateEndpointLongRunning
.
Fields | |
---|---|
endpoint |
Required. The Endpoint which replaces the resource on the server. Currently we only support updating the |
UpdateEndpointOperationMetadata
Runtime operation information for EndpointService.UpdateEndpointLongRunning
.
Fields | |
---|---|
generic_ |
The operation generic information. |
UpdateEndpointRequest
Request message for EndpointService.UpdateEndpoint
.
Fields | |
---|---|
endpoint |
Required. The Endpoint which replaces the resource on the server. |
update_ |
Required. The update mask applies to the resource. See |
UpdateEntityTypeRequest
Request message for FeaturestoreService.UpdateEntityType
.
Fields | |
---|---|
entity_ |
Required. The EntityType's |
update_ |
Field mask is used to specify the fields to be overwritten in the EntityType resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field will be overwritten if it is in the mask. If the user does not provide a mask then only the non-empty fields present in the request will be overwritten. Set the update_mask to Updatable fields:
|
UpdateExecutionRequest
Request message for MetadataService.UpdateExecution
.
Fields | |
---|---|
execution |
Required. The Execution containing updates. The Execution's |
update_ |
Optional. A FieldMask indicating which fields should be updated. |
allow_ |
If set to true, and the |
UpdateExplanationDatasetOperationMetadata
Runtime operation information for ModelService.UpdateExplanationDataset
.
Fields | |
---|---|
generic_ |
The common part of the operation metadata. |
UpdateExplanationDatasetRequest
Request message for ModelService.UpdateExplanationDataset
.
Fields | |
---|---|
model |
Required. The resource name of the Model to update. Format: |
examples |
The example config containing the location of the dataset. |
UpdateExplanationDatasetResponse
This type has no fields.
Response message of ModelService.UpdateExplanationDataset
operation.
UpdateFeatureGroupOperationMetadata
Details of operations that perform update FeatureGroup.
Fields | |
---|---|
generic_ |
Operation metadata for FeatureGroup. |
UpdateFeatureGroupRequest
Request message for FeatureRegistryService.UpdateFeatureGroup
.
Fields | |
---|---|
feature_ |
Required. The FeatureGroup's |
update_ |
Field mask is used to specify the fields to be overwritten in the FeatureGroup resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field will be overwritten if it is in the mask. If the user does not provide a mask then only the non-empty fields present in the request will be overwritten. Set the update_mask to Updatable fields:
|
UpdateFeatureOnlineStoreOperationMetadata
Details of operations that perform update FeatureOnlineStore.
Fields | |
---|---|
generic_ |
Operation metadata for FeatureOnlineStore. |
UpdateFeatureOnlineStoreRequest
Request message for FeatureOnlineStoreAdminService.UpdateFeatureOnlineStore
.
Fields | |
---|---|
feature_ |
Required. The FeatureOnlineStore's |
update_ |
Field mask is used to specify the fields to be overwritten in the FeatureOnlineStore resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field will be overwritten if it is in the mask. If the user does not provide a mask then only the non-empty fields present in the request will be overwritten. Set the update_mask to Updatable fields:
|
UpdateFeatureOperationMetadata
Details of operations that perform update Feature.
Fields | |
---|---|
generic_ |
Operation metadata for Feature Update. |
UpdateFeatureRequest
Request message for FeaturestoreService.UpdateFeature
. Request message for FeatureRegistryService.UpdateFeature
.
Fields | |
---|---|
feature |
Required. The Feature's |
update_ |
Field mask is used to specify the fields to be overwritten in the Features resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field will be overwritten if it is in the mask. If the user does not provide a mask then only the non-empty fields present in the request will be overwritten. Set the update_mask to Updatable fields:
|
UpdateFeatureViewOperationMetadata
Details of operations that perform update FeatureView.
Fields | |
---|---|
generic_ |
Operation metadata for FeatureView Update. |
UpdateFeatureViewRequest
Request message for FeatureOnlineStoreAdminService.UpdateFeatureView
.
Fields | |
---|---|
feature_ |
Required. The FeatureView's |
update_ |
Field mask is used to specify the fields to be overwritten in the FeatureView resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field will be overwritten if it is in the mask. If the user does not provide a mask then only the non-empty fields present in the request will be overwritten. Set the update_mask to Updatable fields:
|
UpdateFeaturestoreOperationMetadata
Details of operations that perform update Featurestore.
Fields | |
---|---|
generic_ |
Operation metadata for Featurestore. |
UpdateFeaturestoreRequest
Request message for FeaturestoreService.UpdateFeaturestore
.
Fields | |
---|---|
featurestore |
Required. The Featurestore's |
update_ |
Field mask is used to specify the fields to be overwritten in the Featurestore resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field will be overwritten if it is in the mask. If the user does not provide a mask then only the non-empty fields present in the request will be overwritten. Set the update_mask to Updatable fields:
|
UpdateIndexEndpointRequest
Request message for IndexEndpointService.UpdateIndexEndpoint
.
Fields | |
---|---|
index_ |
Required. The IndexEndpoint which replaces the resource on the server. |
update_ |
Required. The update mask applies to the resource. See |
UpdateIndexOperationMetadata
Runtime operation information for IndexService.UpdateIndex
.
Fields | |
---|---|
generic_ |
The operation generic information. |
nearest_ |
The operation metadata with regard to Matching Engine Index operation. |
UpdateIndexRequest
Request message for IndexService.UpdateIndex
.
Fields | |
---|---|
index |
Required. The Index which updates the resource on the server. |
update_ |
The update mask applies to the resource. For the |
UpdateModelDeploymentMonitoringJobOperationMetadata
Runtime operation information for JobService.UpdateModelDeploymentMonitoringJob
.
Fields | |
---|---|
generic_ |
The operation generic information. |
UpdateModelDeploymentMonitoringJobRequest
Request message for JobService.UpdateModelDeploymentMonitoringJob
.
Fields | |
---|---|
model_ |
Required. The model monitoring configuration which replaces the resource on the server. |
update_ |
Required. The update mask is used to specify the fields to be overwritten in the ModelDeploymentMonitoringJob resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field will be overwritten if it is in the mask. If the user does not provide a mask then only the non-empty fields present in the request will be overwritten. Set the update_mask to Updatable fields:
|
UpdateModelRequest
Request message for ModelService.UpdateModel
.
Fields | |
---|---|
model |
Required. The Model which replaces the resource on the server. When Model Versioning is enabled, the model.name will be used to determine whether to update the model or model version. 1. model.name with the @ value, e.g. models/123@1, refers to a version specific update. 2. model.name without the @ value, e.g. models/123, refers to a model update. 3. model.name with @-, e.g. models/123@-, refers to a model update. 4. Supported model fields: display_name, description; supported version-specific fields: version_description. Labels are supported in both scenarios. Both the model labels and the version labels are merged when a model is returned. When updating labels, if the request is for model-specific update, model label gets updated. Otherwise, version labels get updated. 5. A model name or model version name fields update mismatch will cause a precondition error. 6. One request cannot update both the model and the version fields. You must update them separately. |
update_ |
Required. The update mask applies to the resource. For the |
UpdateNotebookRuntimeTemplateRequest
Request message for NotebookService.UpdateNotebookRuntimeTemplate
.
Fields | |
---|---|
notebook_ |
Required. The NotebookRuntimeTemplate to update. |
update_ |
Required. The update mask applies to the resource. For the
|
UpdatePersistentResourceOperationMetadata
Details of operations that perform update PersistentResource.
Fields | |
---|---|
generic_ |
Operation metadata for PersistentResource. |
progress_ |
Progress Message for Update LRO |
UpdatePersistentResourceRequest
Request message for UpdatePersistentResource method.
Fields | |
---|---|
persistent_ |
Required. The PersistentResource to update. The PersistentResource's |
update_ |
Required. Specify the fields to be overwritten in the PersistentResource by the update method. |
UpdateRagCorpusOperationMetadata
Runtime operation information for VertexRagDataService.UpdateRagCorpus
.
Fields | |
---|---|
generic_ |
The operation generic information. |
UpdateRagCorpusRequest
Request message for VertexRagDataService.UpdateRagCorpus
.
Fields | |
---|---|
rag_ |
Required. The RagCorpus which replaces the resource on the server. |
UpdateScheduleRequest
Request message for ScheduleService.UpdateSchedule
.
Fields | |
---|---|
schedule |
Required. The Schedule which replaces the resource on the server. The following restrictions will be applied:
|
update_ |
Required. The update mask applies to the resource. See |
UpdateSpecialistPoolOperationMetadata
Runtime operation metadata for SpecialistPoolService.UpdateSpecialistPool
.
Fields | |
---|---|
specialist_ |
Output only. The name of the SpecialistPool to which the specialists are being added. Format: |
generic_ |
The operation generic information. |
UpdateSpecialistPoolRequest
Request message for SpecialistPoolService.UpdateSpecialistPool
.
Fields | |
---|---|
specialist_ |
Required. The SpecialistPool which replaces the resource on the server. |
update_ |
Required. The update mask applies to the resource. |
UpdateTensorboardExperimentRequest
Request message for TensorboardService.UpdateTensorboardExperiment
.
Fields | |
---|---|
update_ |
Required. Field mask is used to specify the fields to be overwritten in the TensorboardExperiment resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field is overwritten if it's in the mask. If the user does not provide a mask then all fields are overwritten if new values are specified. |
tensorboard_ |
Required. The TensorboardExperiment's |
UpdateTensorboardOperationMetadata
Details of operations that perform update Tensorboard.
Fields | |
---|---|
generic_ |
Operation metadata for Tensorboard. |
UpdateTensorboardRequest
Request message for TensorboardService.UpdateTensorboard
.
Fields | |
---|---|
update_ |
Required. Field mask is used to specify the fields to be overwritten in the Tensorboard resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field is overwritten if it's in the mask. If the user does not provide a mask then all fields are overwritten if new values are specified. |
tensorboard |
Required. The Tensorboard's |
UpdateTensorboardRunRequest
Request message for TensorboardService.UpdateTensorboardRun
.
Fields | |
---|---|
update_ |
Required. Field mask is used to specify the fields to be overwritten in the TensorboardRun resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field is overwritten if it's in the mask. If the user does not provide a mask then all fields are overwritten if new values are specified. |
tensorboard_ |
Required. The TensorboardRun's |
UpdateTensorboardTimeSeriesRequest
Request message for TensorboardService.UpdateTensorboardTimeSeries
.
Fields | |
---|---|
update_ |
Required. Field mask is used to specify the fields to be overwritten in the TensorboardTimeSeries resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field is overwritten if it's in the mask. If the user does not provide a mask then all fields are overwritten if new values are specified. |
tensorboard_ |
Required. The TensorboardTimeSeries' |
UpgradeNotebookRuntimeOperationMetadata
Metadata information for NotebookService.UpgradeNotebookRuntime
.
Fields | |
---|---|
generic_ |
The operation generic information. |
progress_ |
A human-readable message that shows the intermediate progress details of NotebookRuntime. |
UpgradeNotebookRuntimeRequest
Request message for NotebookService.UpgradeNotebookRuntime
.
Fields | |
---|---|
name |
Required. The name of the NotebookRuntime resource to be upgrade. Instead of checking whether the name is in valid NotebookRuntime resource name format, directly throw NotFound exception if there is no such NotebookRuntime in spanner. |
UpgradeNotebookRuntimeResponse
This type has no fields.
Response message for NotebookService.UpgradeNotebookRuntime
.
UploadModelOperationMetadata
Details of ModelService.UploadModel
operation.
Fields | |
---|---|
generic_ |
The common part of the operation metadata. |
UploadModelRequest
Request message for ModelService.UploadModel
.
Fields | |
---|---|
parent |
Required. The resource name of the Location into which to upload the Model. Format: |
parent_ |
Optional. The resource name of the model into which to upload the version. Only specify this field when uploading a new version. |
model_ |
Optional. The ID to use for the uploaded Model, which will become the final component of the model resource name. This value may be up to 63 characters, and valid characters are |
model |
Required. The Model to create. |
service_ |
Optional. The user-provided custom service account to use to do the model upload. If empty, Vertex AI Service Agent will be used to access resources needed to upload the model. This account must belong to the target project where the model is uploaded to, i.e., the project specified in the |
UploadModelResponse
Response message of ModelService.UploadModel
operation.
Fields | |
---|---|
model |
The name of the uploaded Model resource. Format: |
model_ |
Output only. The version ID of the model that is uploaded. |
UploadRagFileConfig
Config for uploading RagFile.
Fields | |
---|---|
rag_ |
Specifies the transformation config for RagFiles. |
UpsertDatapointsRequest
Request message for IndexService.UpsertDatapoints
Fields | |
---|---|
index |
Required. The name of the Index resource to be updated. Format: |
datapoints[] |
A list of datapoints to be created/updated. |
update_ |
Optional. Update mask is used to specify the fields to be overwritten in the datapoints by the update. The fields specified in the update_mask are relative to each IndexDatapoint inside datapoints, not the full request. Updatable fields:
|
UpsertDatapointsResponse
This type has no fields.
Response message for IndexService.UpsertDatapoints
UserActionReference
References an API call. It contains more information about long running operation and Jobs that are triggered by the API call.
Fields | |
---|---|
method |
The method name of the API RPC call. For example, "/google.cloud.aiplatform.{apiVersion}.DatasetService.CreateDataset" |
Union field
|
|
operation |
For API calls that return a long running operation. Resource name of the long running operation. Format: |
data_ |
For API calls that start a LabelingJob. Resource name of the LabelingJob. Format: |
Value
Value is the value of the field.
Fields | |
---|---|
Union field
|
|
int_ |
An integer value. |
double_ |
A double value. |
string_ |
A string value. |
VertexAISearch
Retrieve from Vertex AI Search datastore for grounding. See https://cloud.google.com/products/agent-builder
Fields | |
---|---|
datastore |
Required. Fully-qualified Vertex AI Search data store resource ID. Format: |
VertexRagStore
Retrieve from Vertex RAG Store for grounding.
Fields | |
---|---|
rag_ |
Optional. The representation of the rag source. It can be used to specify corpus only or ragfiles. Currently only support one corpus or multiple files from one corpus. In the future we may open up multiple corpora support. |
rag_ |
Optional. The retrieval config for the Rag query. |
similarity_top_k |
Optional. Number of top k results to return from the selected corpora. |
vector_distance_threshold |
Optional. Only return results with vector distance smaller than the threshold. |
RagResource
The definition of the Rag resource.
Fields | |
---|---|
rag_ |
Optional. RagCorpora resource name. Format: |
rag_ |
Optional. rag_file_id. The files should be in the same rag_corpus set in rag_corpus field. |
VideoMetadata
Metadata describes the input video content.
Fields | |
---|---|
start_ |
Optional. The start offset of the video. |
end_ |
Optional. The end offset of the video. |
WorkerPoolSpec
Represents the spec of a worker pool in a job.
Fields | |
---|---|
machine_ |
Optional. Immutable. The specification of a single machine. |
replica_ |
Optional. The number of worker replicas to use for this worker pool. |
nfs_ |
Optional. List of NFS mount spec. |
disk_ |
Disk spec. |
Union field task . The custom task to be executed in this worker pool. task can be only one of the following: |
|
container_ |
The custom container task. |
python_ |
The Python packaged task. |
WriteFeatureValuesPayload
Contains Feature values to be written for a specific entity.
Fields | |
---|---|
entity_ |
Required. The ID of the entity. |
feature_ |
Required. Feature values to be written, mapping from Feature ID to value. Up to 100,000 |
WriteFeatureValuesRequest
Request message for FeaturestoreOnlineServingService.WriteFeatureValues
.
Fields | |
---|---|
entity_ |
Required. The resource name of the EntityType for the entities being written. Value format: |
payloads[] |
Required. The entities to be written. Up to 100,000 feature values can be written across all |
WriteFeatureValuesResponse
This type has no fields.
Response message for FeaturestoreOnlineServingService.WriteFeatureValues
.
WriteTensorboardExperimentDataRequest
Request message for TensorboardService.WriteTensorboardExperimentData
.
Fields | |
---|---|
tensorboard_ |
Required. The resource name of the TensorboardExperiment to write data to. Format: |
write_ |
Required. Requests containing per-run TensorboardTimeSeries data to write. |
WriteTensorboardExperimentDataResponse
This type has no fields.
Response message for TensorboardService.WriteTensorboardExperimentData
.
WriteTensorboardRunDataRequest
Request message for TensorboardService.WriteTensorboardRunData
.
Fields | |
---|---|
tensorboard_ |
Required. The resource name of the TensorboardRun to write data to. Format: |
time_ |
Required. The TensorboardTimeSeries data to write. Values with in a time series are indexed by their step value. Repeated writes to the same step will overwrite the existing value for that step. The upper limit of data points per write request is 5000. |
WriteTensorboardRunDataResponse
This type has no fields.
Response message for TensorboardService.WriteTensorboardRunData
.
XraiAttribution
An explanation method that redistributes Integrated Gradients attributions to segmented regions, taking advantage of the model's fully differentiable structure. Refer to this paper for more details: https://arxiv.org/abs/1906.02825
Supported only by image Models.
Fields | |
---|---|
step_ |
Required. The number of steps for approximating the path integral. A good value to start is 50 and gradually increase until the sum to diff property is met within the desired error range. Valid range of its value is [1, 100], inclusively. |
smooth_ |
Config for SmoothGrad approximation of gradients. When enabled, the gradients are approximated by averaging the gradients from noisy samples in the vicinity of the inputs. Adding noise can help improve the computed gradients. Refer to this paper for more details: https://arxiv.org/pdf/1706.03825.pdf |
blur_ |
Config for XRAI with blur baseline. When enabled, a linear path from the maximally blurred image to the input image is created. Using a blurred baseline instead of zero (black image) is motivated by the BlurIG approach explained here: https://arxiv.org/abs/2004.03383 |