Interpret prediction results from text classification models
Stay organized with collections
Save and categorize content based on your preferences.
After requesting a prediction, Vertex AI returns results based on your
model's objective. Predictions from multi-label classification models return one
or more labels for each document and a confidence score for each label. For
single-label classification models, predictions return only one label and
confidence score per document.
The confidence score communicates how strongly your model associates each
class or label with a test item. The higher the number, the higher the model's
confidence that the label should be applied to that item. You decide how high
the confidence score must be for you to accept the model's results.
Score threshold slider
In the Google Cloud console, Vertex AI provides a slider that's
used to adjust the confidence threshold for all classes or labels, or an
individual class or label. The slider is available on a model's detail page in
the Evaluate tab. The confidence threshold is the confidence level that
the model must have for it to assign a class or label to a test item. As you
adjust the threshold, you can see how your model's precision and recall
changes. Higher thresholds typically increase precision and lower recall.
Example batch prediction output
The following sample is the predicted result for a multi-label classification
model. The model applied the GreatService, Suggestion, and InfoRequest
labels to the submitted document. The confidence values apply to each of the
labels in order. In this example, the model predicted GreatService as the most
relevant label.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Hard to understand","hardToUnderstand","thumb-down"],["Incorrect information or sample code","incorrectInformationOrSampleCode","thumb-down"],["Missing the information/samples I need","missingTheInformationSamplesINeed","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2025-09-04 UTC."],[],[],null,["# Interpret prediction results from text classification models\n\n| Starting on September 15, 2024, you can only customize classification, entity extraction, and sentiment analysis objectives by moving to Vertex AI Gemini prompts and tuning. Training or updating models for Vertex AI AutoML for Text classification, entity extraction, and sentiment analysis objectives will no longer be available. You can continue using existing Vertex AI AutoML Text models until June 15, 2025. For a comparison of AutoML text and Gemini, see [Gemini for AutoML text users](/vertex-ai/docs/start/automl-gemini-comparison). For more information about how Gemini offers enhanced user experience through improved prompting capabilities, see [Introduction to tuning](/vertex-ai/generative-ai/docs/models/tune-gemini-overview). To get started with tuning, see [Model tuning for Gemini text models](/vertex-ai/generative-ai/docs/models/tune_gemini/tune-gemini-learn)\n\nAfter requesting a prediction, Vertex AI returns results based on your\nmodel's objective. Predictions from multi-label classification models return one\nor more labels for each document and a confidence score for each label. For\nsingle-label classification models, predictions return only one label and\nconfidence score per document.\n\n\nThe confidence score communicates how strongly your model associates each\nclass or label with a test item. The higher the number, the higher the model's\nconfidence that the label should be applied to that item. You decide how high\nthe confidence score must be for you to accept the model's results.\n\n\u003cbr /\u003e\n\nScore threshold slider\n----------------------\n\n\nIn the Google Cloud console, Vertex AI provides a slider that's\nused to adjust the confidence threshold for all classes or labels, or an\nindividual class or label. The slider is available on a model's detail page in\nthe **Evaluate** tab. The confidence threshold is the confidence level that\nthe model must have for it to assign a class or label to a test item. As you\nadjust the threshold, you can see how your model's precision and recall\nchanges. Higher thresholds typically increase precision and lower recall.\n\n\u003cbr /\u003e\n\nExample batch prediction output\n-------------------------------\n\nThe following sample is the predicted result for a multi-label classification\nmodel. The model applied the `GreatService`, `Suggestion`, and `InfoRequest`\nlabels to the submitted document. The confidence values apply to each of the\nlabels in order. In this example, the model predicted `GreatService` as the most\nrelevant label.\n\n\n| **Note**: The following JSON Lines example includes line breaks for\n| readability. In your JSON Lines files, line breaks are included only after each\n| each JSON object.\n\n\u003cbr /\u003e\n\n\n```\n{\n \"instance\": {\"content\": \"gs://bucket/text.txt\", \"mimeType\": \"text/plain\"},\n \"predictions\": [\n {\n \"ids\": [\n \"1234567890123456789\",\n \"2234567890123456789\",\n \"3234567890123456789\"\n ],\n \"displayNames\": [\n \"GreatService\",\n \"Suggestion\",\n \"InfoRequest\"\n ],\n \"confidences\": [\n 0.8986392080783844,\n 0.81984345316886902,\n 0.7722353458404541\n ]\n }\n ]\n}\n```"]]