Class ModelExportOutputConfig (0.4.0)

Output configuration for ModelExport Action.

The Google Cloud Storage location where the model is to be written to. This location may only be set for the following model formats: "tflite", "edgetpu_tflite", "core_ml", "docker". Under the directory given as the destination a new one with name "model-export--", where timestamp is in YYYY-MM- DDThh:mm:ss.sssZ ISO-8601 format, will be created. Inside the model and any of its supporting files will be written.

The format in which the model must be exported. The available, and default, formats depend on the problem and model type (if given problem and type combination doesn't have a format listed, it means its models are not exportable): - For Image Classification mobile-low-latency-1, mobile-versatile-1, mobile-high-accuracy-1: "tflite" (default), "edgetpu_tflite", "tf_saved_model", "docker". - For Image Classification mobile-core-ml-low-latency-1, mobile-core-ml-versatile-1, mobile-core-ml-high-accuracy-1: "core_ml" (default). Formats description: - tflite - Used for Android mobile devices. - edgetpu_tflite - Used for Edge TPU <https://cloud.google.com/edge-tpu/> devices. - tf_saved_model - A tensorflow model in SavedModel format. - docker - Used for Docker containers. Use the params field to customize the container. The container is verified to work correctly on ubuntu 16.04 operating system. See more at containers quickstart <https://cloud.google.com/vision/automl/docs/containers-gcs- quickstart> - core_ml - Used for iOS mobile devices.

Classes

ParamsEntry

API documentation for automl_v1beta1.types.ModelExportOutputConfig.ParamsEntry class.