Settings for LLM models.
| JSON representation | 
|---|
{
  "model": string,
  "promptText": string,
  "parameters": {
    object ( | 
            
| Fields | |
|---|---|
model | 
              
                 
 The selected LLM model.  | 
            
promptText | 
              
                 
 The custom prompt to use.  | 
            
parameters | 
              
                 
 Generative model parameters.  | 
            
Parameters
Generative model parameters to control the model behavior.
| JSON representation | 
|---|
{ "temperature": number, "inputTokenLimit": enum (  | 
              
| Fields | |
|---|---|
temperature | 
                
                   
 The temperature used for sampling during response generation. Value ranges from 0 to 1. Temperature controls the degree of randomness in token selection. Lower temperature means less randomness, while higher temperature means more randomness. Valid range: [0.0, 1.0]  | 
              
inputTokenLimit | 
                
                   
 The input token limit. This setting is currently only supported by playbooks.  | 
              
outputTokenLimit | 
                
                   
 The output token limit. This setting is currently only supported by playbooks. Only one of outputTokenLimit and max_output_tokens is allowed to be set.  | 
              
InputTokenLimit
The input token limits for 1 LLM call. For the limit of each model, see https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models for more information.
| Enums | |
|---|---|
INPUT_TOKEN_LIMIT_UNSPECIFIED | 
                Limit not specified. Treated as 'INPUT_TOKEN_LIMIT_SHORT'. | 
INPUT_TOKEN_LIMIT_SHORT | 
                Input token limit up to 8k. | 
INPUT_TOKEN_LIMIT_MEDIUM | 
                Input token limit up to 32k. | 
INPUT_TOKEN_LIMIT_LONG | 
                Input token limit up to 100k. | 
OutputTokenLimit
The output token limits for 1 LLM call. The limits are subject to change. For the limit of each model, see https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models for more information.
| Enums | |
|---|---|
OUTPUT_TOKEN_LIMIT_UNSPECIFIED | 
                Limit not specified. | 
OUTPUT_TOKEN_LIMIT_SHORT | 
                Input token limit up to 512 tokens. | 
OUTPUT_TOKEN_LIMIT_MEDIUM | 
                Input token limit up to 1k. | 
OUTPUT_TOKEN_LIMIT_LONG | 
                Input token limit up to 2k. |