The goal when developing a model for a problem is to identify the best model for that particular use case. To this end, Vertex AI Experiments enables you to track, analyze, compare, and search across different ML Frameworks (for example, TensorFlow, PyTorch, scikit-learn) and training environments.
Experiment runs do not incur additional charges. You're only charged for resources that you use during your experiment as described in Vertex AI pricing.
What do you want to do? | Check out notebook sample |
---|---|
track metrics and parameters | Compare models |
track experiment lineage | Model training |
track pipeline runs | Compare pipeline runs |
Track steps, inputs, and outputs
Vertex AI Experiments enables you to track
- steps of an experiment run, for example, preprocessing, training,
- inputs, for example, algorithm, parameters, datasets,
- outputs of those steps, for example, models, checkpoints, metrics.
You can then figure out what worked, what did not, and identify further avenues for experimentation.
For user journey examples, check out:
Analyze model performance
Vertex AI Experiments enables you to track and evaluate how their model performed in aggregate, against test datasets, and during the training run. This ability helps to understand the performance characteristics of their models -- how well a particular model works overall, where it fails, and where the model excels.
For user journey examples, check out:
Compare model performance
Vertex AI Experiments enables you to group and compare multiple models across experiment runs. Each model has its own specified parameters, modeling techniques, architectures, and input. This approach helps select the best model.
For user journey examples, check out:
Search experiments
The console provides a centralized view of experiments, a cross-sectional view of the experiment runs, and the details for each run. The Vertex AI SDK for Python provides APIs to consume experiments, experiment runs, experiment run parameters, metrics, and artifacts.
Vertex AI Experiments, along with Vertex ML Metadata, provides a way to find the artifacts tracked in an experiment so you can quickly view the artifact's lineage, and the artifacts consumed and produced by steps in a run.
Scope of support
Vertex AI Experiments supports development of models using Vertex AI custom training, Vertex AI Workbench notebooks, Notebooks, and all Python ML Frameworks across most ML Frameworks. For some ML frameworks, such as TensorFlow, Vertex AI Experiments provides deep integrations into the framework that makes the user experience automagical. For other ML frameworks, Vertex AI Experiments provides a framework neutral Vertex AI SDK for Python that you can use. (see: Pre-built containers for TensorFlow, sckikit-learn, PyTorch, XGBoost).
Data models and concepts
Vertex AI Experiments is a context in Vertex ML Metadata where an experiment can contain n experiment runs in addition to n pipeline runs. An experiment run consists of parameters, summary metrics, time series metrics, Vertex resources (PipelineJob), artifacts, and executions. Vertex AI TensorBoard, a managed version of open source TensorBoard, is used for time-series metrics storage. Executions and artifacts of a pipeline run are viewable in the console.
Vertex AI Experiments terms
Experiment, experiment run, and pipeline run
See Associate a pipeline with an experiment.
Parameters and metrics
See Log parameters.
See Log summary metrics.
Resource types
Vertex AI Experiments lets you to define the type of artifact, for
example, supported types include system.Dataset
, system.Model
,
system.Artifact
.