CodeChatSession(
model: vertexai.language_models.CodeChatModel,
context: typing.Optional[str] = None,
max_output_tokens: typing.Optional[int] = None,
temperature: typing.Optional[float] = None,
message_history: typing.Optional[
typing.List[vertexai.language_models.ChatMessage]
] = None,
stop_sequences: typing.Optional[typing.List[str]] = None,
)
CodeChatSession represents a chat session with code chat language model.
Within a code chat session, the model keeps context and remembers the previous converstion.
Properties
message_history
List of previous messages.
Methods
send_message
send_message(
message: str,
*,
max_output_tokens: typing.Optional[int] = None,
temperature: typing.Optional[float] = None,
stop_sequences: typing.Optional[typing.List[str]] = None,
candidate_count: typing.Optional[int] = None
) -> vertexai.language_models.MultiCandidateTextGenerationResponse
Sends message to the code chat model and gets a response.
Parameter | |
---|---|
Name | Description |
message |
str
Message to send to the model |
send_message_async
send_message_async(
message: str,
*,
max_output_tokens: typing.Optional[int] = None,
temperature: typing.Optional[float] = None,
candidate_count: typing.Optional[int] = None
) -> vertexai.language_models.MultiCandidateTextGenerationResponse
Asynchronously sends message to the code chat model and gets a response.
Parameter | |
---|---|
Name | Description |
message |
str
Message to send to the model |
send_message_streaming
send_message_streaming(
message: str,
*,
max_output_tokens: typing.Optional[int] = None,
temperature: typing.Optional[float] = None,
stop_sequences: typing.Optional[typing.List[str]] = None
) -> typing.Iterator[vertexai.language_models.TextGenerationResponse]
Sends message to the language model and gets a streamed response.
The response is only added to the history once it's fully read.
Parameter | |
---|---|
Name | Description |
message |
str
Message to send to the model |
send_message_streaming_async
send_message_streaming_async(
message: str,
*,
max_output_tokens: typing.Optional[int] = None,
temperature: typing.Optional[float] = None,
stop_sequences: typing.Optional[typing.List[str]] = None
) -> typing.AsyncIterator[vertexai.language_models.TextGenerationResponse]
Asynchronously sends message to the language model and gets a streamed response.
The response is only added to the history once it's fully read.
Parameter | |
---|---|
Name | Description |
message |
str
Message to send to the model |