- 3.27.0 (latest)
- 3.26.0
- 3.25.0
- 3.24.0
- 3.23.1
- 3.22.0
- 3.21.0
- 3.20.1
- 3.19.0
- 3.18.0
- 3.17.2
- 3.16.0
- 3.15.0
- 3.14.1
- 3.13.0
- 3.12.0
- 3.11.4
- 3.4.0
- 3.3.6
- 3.2.0
- 3.1.0
- 3.0.1
- 2.34.4
- 2.33.0
- 2.32.0
- 2.31.0
- 2.30.1
- 2.29.0
- 2.28.1
- 2.27.1
- 2.26.0
- 2.25.2
- 2.24.1
- 2.23.3
- 2.22.1
- 2.21.0
- 2.20.0
- 2.19.0
- 2.18.0
- 2.17.0
- 2.16.1
- 2.15.0
- 2.14.0
- 2.13.1
- 2.12.0
- 2.11.0
- 2.10.0
- 2.9.0
- 2.8.0
- 2.7.0
- 2.6.2
- 2.5.0
- 2.4.0
- 2.3.1
- 2.2.0
- 2.1.0
- 2.0.0
- 1.28.2
- 1.27.2
- 1.26.1
- 1.25.0
- 1.24.0
- 1.23.1
- 1.22.0
- 1.21.0
- 1.20.0
- 1.19.0
- 1.18.0
- 1.17.0
- 1.16.0
Protocol buffer.
Required. Unique identifier for this model.
Output only. The time when this model was last modified, in millisecs since the epoch.
[Optional] A descriptive name for this model.
[Optional] The time when this model expires, in milliseconds since the epoch. If not present, the model will persist indefinitely. Expired models will be deleted and their storage reclaimed. The defaultTableExpirationMs property of the encapsulating dataset can be used to set a default expirationTime on newly created models.
Output only. Type of the model resource.
Output only. Input feature columns that were used to train this model.
Inheritance
builtins.object > google.protobuf.pyext._message.CMessage > builtins.object > google.protobuf.message.Message > ModelClasses
AggregateClassificationMetrics
Aggregate metrics for classification/classifier models. For multi-class models, the metrics are either macro-averaged or micro-averaged. When macro-averaged, the metrics are calculated for each label and then an unweighted average is taken of those values. When micro-averaged, the metric is calculated globally by counting the total number of correctly predicted rows.
.. attribute:: precision
Precision is the fraction of actual positive predictions that had positive actual labels. For multiclass this is a macro- averaged metric treating each class as a binary classifier.
Accuracy is the fraction of predictions given the correct label. For multiclass this is a micro-averaged metric.
The F1 score is an average of recall and precision. For multiclass this is a macro-averaged metric.
Area Under a ROC Curve. For multiclass this is a macro- averaged metric.
BinaryClassificationMetrics
Evaluation metrics for binary classification/classifier models.
.. attribute:: aggregate_classification_metrics
Aggregate classification metrics.
Label representing the positive class.
ClusteringMetrics
Evaluation metrics for clustering models.
.. attribute:: davies_bouldin_index
Davies-Bouldin index.
EvaluationMetrics
Evaluation metrics of a model. These are either computed on all training data or just the eval data based on whether eval data was used during training. These are not present for imported models.
.. attribute:: regression_metrics
Populated for regression models.
Populated for multi-class classification/classifier models.
LabelsEntry
API documentation for bigquery_v2.types.Model.LabelsEntry
class.
MultiClassClassificationMetrics
Evaluation metrics for multi-class classification/classifier models.
.. attribute:: aggregate_classification_metrics
Aggregate classification metrics.
RegressionMetrics
Evaluation metrics for regression models.
.. attribute:: mean_absolute_error
Mean absolute error.
Mean squared log error.
R^2 score.
TrainingRun
Information about a single training query run for the model.
.. attribute:: training_options
Options that were used for this training run, includes user specified and default options that were used.
Output of each iteration run, results.size() <= max_iterations.