StreamingDetectIntentRequest(
mapping=None, *, ignore_unknown_fields=False, **kwargs
)
The top-level message sent by the client to the Sessions.StreamingDetectIntent method.
Multiple request messages should be sent in order:
The first message must contain session, query_input plus optionally query_params. If the client wants to receive an audio response, it should also contain output_audio_config. The message must not contain input_audio.
If query_input was set to query_input.audio_config, all subsequent messages must contain input_audio to continue with Speech recognition. If you decide to rather detect an intent from text input after you already started Speech recognition, please send a message with query_input.text.
However, note that:
- Dialogflow will bill you for the audio duration so far.
- Dialogflow discards all Speech recognition results in favor of the input text.
- Dialogflow will use the language code from the first message.
After you sent all input, you must half-close or abort the request stream.
Attributes
Name | Description |
session |
str
Required. The name of the session the query is sent to. Supported formats: - \`projects//agent/sessions/, - ``projects/ |
query_params |
google.cloud.dialogflow_v2beta1.types.QueryParameters
The parameters of this query. |
query_input |
google.cloud.dialogflow_v2beta1.types.QueryInput
Required. The input specification. It can be set to: 1. an audio config which instructs the speech recognizer how to process the speech audio, 2. a conversational query in the form of text, or 3. an event that specifies which intent to trigger. |
single_utterance |
bool
DEPRECATED. Please use InputAudioConfig.single_utterance instead. If ``false`` (default), recognition does not cease until the client closes the stream. If ``true``, the recognizer will detect a single spoken utterance in input audio. Recognition ceases when it detects the audio's voice has stopped or paused. In this case, once a detected intent is received, the client should close the stream and start a new request with a new stream as needed. This setting is ignored when ``query_input`` is a piece of text or an event. |
output_audio_config |
google.cloud.dialogflow_v2beta1.types.OutputAudioConfig
Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated. |
output_audio_config_mask |
google.protobuf.field_mask_pb2.FieldMask
Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level. If unspecified or empty, output_audio_config replaces the agent-level config in its entirety. |
input_audio |
bytes
The input audio content to be recognized. Must be sent if ``query_input`` was set to a streaming input audio config. The complete audio over all streaming messages must not exceed 1 minute. |