The top-level message sent by the client to the
Sessions.StreamingDetectIntent
method.
Multiple request messages should be sent in order:
The first message must contain
session,
query_input
plus optionally
query_params.
If the client wants to receive an audio response, it should also contain
output_audio_config.
The message must not contain
input_audio.
If
query_input
was set to
query_input.audio_config,
all subsequent messages must contain
input_audio
to continue with Speech recognition. If you decide to rather detect an
intent from text input after you already started Speech recognition,
please send a message with
query_input.text.
However, note that:
Dialogflow will bill you for the audio duration so far.
Dialogflow discards all Speech recognition results in favor of the
input text.
Dialogflow will use the language code from the first message.
After you sent all input, you must half-close or abort the request stream.
Protobuf type google.cloud.dialogflow.v2beta1.StreamingDetectIntentRequest
The input audio content to be recognized. Must be sent if
query_input was set to a streaming input audio config. The complete audio
over all streaming messages must not exceed 1 minute.
Instructs the speech synthesizer how to generate the output
audio. If this field is not set and agent-level speech synthesizer is not
configured, no output audio is generated.
Mask for
output_audio_config
indicating which settings in this request-level config should override
speech synthesizer settings defined at agent-level.
If unspecified or empty,
output_audio_config
replaces the agent-level config in its entirety.
Mask for
output_audio_config
indicating which settings in this request-level config should override
speech synthesizer settings defined at agent-level.
If unspecified or empty,
output_audio_config
replaces the agent-level config in its entirety.
Instructs the speech synthesizer how to generate the output
audio. If this field is not set and agent-level speech synthesizer is not
configured, no output audio is generated.
projects/<Project ID>/locations/<Location
ID>/agent/environments/<Environment ID>/users/<User ID>/sessions/<Session
ID>,
If Location ID is not specified we assume default 'us' location. If
Environment ID is not specified, we assume default 'draft' environment.
If User ID is not specified, we are using "-". It's up to the API caller
to choose an appropriate Session ID and User Id. They can be a random
number or some type of user and session identifiers (preferably hashed).
The length of the Session ID and User ID` must not exceed 36 characters.
For more information, see the API interactions
guide.
Note: Always use agent versions for production traffic.
See Versions and
environments.
projects/<Project ID>/locations/<Location
ID>/agent/environments/<Environment ID>/users/<User ID>/sessions/<Session
ID>,
If Location ID is not specified we assume default 'us' location. If
Environment ID is not specified, we assume default 'draft' environment.
If User ID is not specified, we are using "-". It's up to the API caller
to choose an appropriate Session ID and User Id. They can be a random
number or some type of user and session identifiers (preferably hashed).
The length of the Session ID and User ID` must not exceed 36 characters.
For more information, see the API interactions
guide.
Note: Always use agent versions for production traffic.
See Versions and
environments.
Deprecated.google.cloud.dialogflow.v2beta1.StreamingDetectIntentRequest.single_utterance is
deprecated. See google/cloud/dialogflow/v2beta1/session.proto;l=564
DEPRECATED. Please use
InputAudioConfig.single_utterance
instead. If false (default), recognition does not cease until the client
closes the stream. If true, the recognizer will detect a single spoken
utterance in input audio. Recognition ceases when it detects the audio's
voice has stopped or paused. In this case, once a detected intent is
received, the client should close the stream and start a new request with a
new stream as needed. This setting is ignored when query_input is a piece
of text or an event.
Instructs the speech synthesizer how to generate the output
audio. If this field is not set and agent-level speech synthesizer is not
configured, no output audio is generated.
Mask for
output_audio_config
indicating which settings in this request-level config should override
speech synthesizer settings defined at agent-level.
If unspecified or empty,
output_audio_config
replaces the agent-level config in its entirety.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Hard to understand","hardToUnderstand","thumb-down"],["Incorrect information or sample code","incorrectInformationOrSampleCode","thumb-down"],["Missing the information/samples I need","missingTheInformationSamplesINeed","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2025-01-28 UTC."],[],[]]