- 1.75.0 (latest)
- 1.74.0
- 1.73.0
- 1.72.0
- 1.71.1
- 1.70.0
- 1.69.0
- 1.68.0
- 1.67.1
- 1.66.0
- 1.65.0
- 1.63.0
- 1.62.0
- 1.60.0
- 1.59.0
- 1.58.0
- 1.57.0
- 1.56.0
- 1.55.0
- 1.54.1
- 1.53.0
- 1.52.0
- 1.51.0
- 1.50.0
- 1.49.0
- 1.48.0
- 1.47.0
- 1.46.0
- 1.45.0
- 1.44.0
- 1.43.0
- 1.39.0
- 1.38.1
- 1.37.0
- 1.36.4
- 1.35.0
- 1.34.0
- 1.33.1
- 1.32.0
- 1.31.1
- 1.30.1
- 1.29.0
- 1.28.1
- 1.27.1
- 1.26.1
- 1.25.0
- 1.24.1
- 1.23.0
- 1.22.1
- 1.21.0
- 1.20.0
- 1.19.1
- 1.18.3
- 1.17.1
- 1.16.1
- 1.15.1
- 1.14.0
- 1.13.1
- 1.12.1
- 1.11.0
- 1.10.0
- 1.9.0
- 1.8.1
- 1.7.1
- 1.6.2
- 1.5.0
- 1.4.3
- 1.3.0
- 1.2.0
- 1.1.1
- 1.0.1
- 0.9.0
- 0.8.0
- 0.7.1
- 0.6.0
- 0.5.1
- 0.4.0
- 0.3.1
ModelMonitoringJob(
model_monitoring_job_name: str,
model_monitor_id: typing.Optional[str] = None,
project: typing.Optional[str] = None,
location: typing.Optional[str] = None,
credentials: typing.Optional[google.auth.credentials.Credentials] = None,
)
Initializer for ModelMonitoringJob.
Example Usage:
my_monitoring_job = aiplatform.ModelMonitoringJob(
model_monitoring_job_name='projects/123/locations/us-central1/modelMonitors/\
my_model_monitor_id/modelMonitoringJobs/my_monitoring_job_id'
)
or
my_monitoring_job = aiplatform.aiplatform.ModelMonitoringJob(
model_monitoring_job_name='my_monitoring_job_id',
model_monitor_id='my_model_monitor_id',
)
Parameters |
|
---|---|
Name | Description |
model_monitoring_job_name |
str
Required. The resource name for the Model Monitoring Job if provided alone, or the model monitoring job id if provided with model_monitor_id. |
model_monitor_id |
str
Optional. The model monitor id depends on the way of initializing ModelMonitoringJob. |
project |
str
Required. Project to retrieve endpoint from. If not set, project set in aiplatform.init will be used. |
location |
str
Required. Location to retrieve endpoint from. If not set, location set in aiplatform.init will be used. |
credentials |
auth_credentials.Credentials
Optional. Custom credentials to use to init model monitoring job. Overrides credentials set in aiplatform.init. |
Properties
create_time
Time this resource was created.
display_name
Display name of this resource.
encryption_spec
Customer-managed encryption key options for this Vertex AI resource.
If this is set, then all resources created by this Vertex AI resource will be encrypted with the provided encryption key.
gca_resource
The underlying resource proto representation.
labels
User-defined labels containing metadata about this resource.
Read more about labels at https://goo.gl/xmQnxf
name
Name of this resource.
resource_name
Full qualified resource name.
state
Fetch Job again and return the current JobState.
Returns | |
---|---|
Type | Description |
state (job_state.JobState) |
Enum that describes the state of a Model Monitoring Job. |
update_time
Time this resource was last updated.
Methods
ModelMonitoringJob
ModelMonitoringJob(
model_monitoring_job_name: str,
model_monitor_id: typing.Optional[str] = None,
project: typing.Optional[str] = None,
location: typing.Optional[str] = None,
credentials: typing.Optional[google.auth.credentials.Credentials] = None,
)
Initializes class with project, location, and api_client.
Parameters | |
---|---|
Name | Description |
project |
str
Optional. Project of the resource noun. |
location |
str
Optional. The location of the resource noun. |
credentials |
google.auth.credentials.Credentials
Optional. custom credentials to use when accessing interacting with resource noun. |
resource_name |
str
A fully-qualified resource name or ID. |
create
create(
model_monitor_name: typing.Optional[str] = None,
target_dataset: typing.Optional[
vertexai.resources.preview.ml_monitoring.spec.objective.MonitoringInput
] = None,
display_name: typing.Optional[str] = None,
model_monitoring_job_id: typing.Optional[str] = None,
project: typing.Optional[str] = None,
location: typing.Optional[str] = None,
credentials: typing.Optional[google.auth.credentials.Credentials] = None,
baseline_dataset: typing.Optional[
vertexai.resources.preview.ml_monitoring.spec.objective.MonitoringInput
] = None,
tabular_objective_spec: typing.Optional[
vertexai.resources.preview.ml_monitoring.spec.objective.TabularObjective
] = None,
output_spec: typing.Optional[
vertexai.resources.preview.ml_monitoring.spec.output.OutputSpec
] = None,
notification_spec: typing.Optional[
vertexai.resources.preview.ml_monitoring.spec.notification.NotificationSpec
] = None,
explanation_spec: typing.Optional[
google.cloud.aiplatform_v1beta1.types.explanation.ExplanationSpec
] = None,
) -> vertexai.resources.preview.ml_monitoring.model_monitors.ModelMonitoringJob
Creates a new ModelMonitoringJob.
Parameters | |
---|---|
Name | Description |
model_monitor_name |
str
Required. The parent model monitor resource name. Format: |
target_dataset |
objective.MonitoringInput
Required. The target dataset for analysis. |
display_name |
str
Optional. The user-defined name of the ModelMonitoringJob. The name can be up to 128 characters long and can comprise any UTF-8 character. |
model_monitoring_job_id |
str
Optional. The unique ID of the model monitoring job run, which will become the final component of the model monitoring job resource name. The maximum length is 63 characters, and valid characters are /^a-z?$/. If not specified, it will be generated by Vertex AI. |
project |
str
Required. Project to retrieve endpoint from. If not set, project set in aiplatform.init will be used. |
location |
str
Required. Location to retrieve endpoint from. If not set, location set in aiplatform.init will be used. |
credentials |
auth_credentials.Credentials
Optional. Custom credentials to use to create model monitoring job. Overrides credentials set in aiplatform.init. |
baseline_dataset |
objective.MonitoringInput
Optional. The baseline dataset for monitoring job. If not set, the training dataset in ModelMonitor will be used as baseline dataset. |
output_spec |
output.OutputSpec
Optional. The monitoring metrics/logs export spec. If not set, will use the default output_spec defined in ModelMonitor. |
notification_spec |
notification.NotificationSpec
Optional. The notification spec for monitoring result. If not set, will use the default notification_spec defined in ModelMonitor. |
explanation_spec |
explanation.ExplanationSpec
Optional. The explanation spec for feature attribution monitoring. If not set, will use the default explanation_spec defined in ModelMonitor. |
Returns | |
---|---|
Type | Description |
ModelMonitoringJob |
The model monitoring job that was created. |
delete
delete() -> None
Deletes an Model Monitoring Job.
done
done() -> bool
Method indicating whether a job has completed.
list
list(
filter: typing.Optional[str] = None,
order_by: typing.Optional[str] = None,
project: typing.Optional[str] = None,
location: typing.Optional[str] = None,
credentials: typing.Optional[google.auth.credentials.Credentials] = None,
parent: typing.Optional[str] = None,
) -> typing.List[google.cloud.aiplatform.base.VertexAiResourceNoun]
List all instances of this Vertex AI Resource.
Example Usage:
aiplatform.BatchPredictionJobs.list( filter='state="JOB_STATE_SUCCEEDED" AND display_name="my_job"', )
aiplatform.Model.list(order_by="create_time desc, display_name")
Parameters | |
---|---|
Name | Description |
filter |
str
Optional. An expression for filtering the results of the request. For field names both snake_case and camelCase are supported. |
order_by |
str
Optional. A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields: |
project |
str
Optional. Project to retrieve list from. If not set, project set in aiplatform.init will be used. |
location |
str
Optional. Location to retrieve list from. If not set, location set in aiplatform.init will be used. |
credentials |
auth_credentials.Credentials
Optional. Custom credentials to use to retrieve list. Overrides credentials set in aiplatform.init. |
parent |
str
Optional. The parent resource name if any to retrieve list from. |
to_dict
to_dict() -> typing.Dict[str, typing.Any]
Returns the resource proto as a dictionary.
wait
wait()
Helper method that blocks until all futures are complete.