Autologging: Notebook

As part of the data science team, you want to try different modeling approaches during experimentation phase.To guarantee reproducibility, each approach has different parameters that you need to manually track. Vertex AI SDK autologging, which is a one-line code SDK capability leveraging MLflow, provides automatic metrics and parameters tracking associated with your Vertex AI Experiments and experiment runs.

Notebook: Vertex AI Experiments Autologging

In the "Vertex AI Experiments: Autologging" notebook, you'll learn how to use Vertex AI Experiments to:

  • Enable autologging in the Vertex AI SDK.
  • Train scikit-learn model and see the resulting experiment run with metrics and parameters autologged to Vertex AI Experiments without setting an experiment run.
  • Train TensorFlow model, check autologged metrics and parameters to Vertex AI Experiments by manually setting an experiment run with aiplatform.start_run() and aiplatform.end_run().
  • Disable autologging in Vertex AI SDK, train a PyTorch model and check that none of the parameters or metrics are logged.

Relevant content