Reference documentation and code samples for the Google Cloud Dialogflow V2 Client class StreamingDetectIntentRequest.
The top-level message sent by the client to the Sessions.StreamingDetectIntent method.
Multiple request messages should be sent in order:
- The first message must contain session, query_input plus optionally query_params. If the client wants to receive an audio response, it should also contain output_audio_config. The message must not contain input_audio.
- If query_input was set to
query_input.audio_config, all subsequent
messages must contain
input_audio to continue with
Speech recognition.
If you decide to rather detect an intent from text input after you
already started Speech recognition, please send a message with
query_input.text.
However, note that:
- Dialogflow will bill you for the audio duration so far.
- Dialogflow discards all Speech recognition results in favor of the input text.
- Dialogflow will use the language code from the first message. After you sent all input, you must half-close or abort the request stream.
Generated from protobuf message google.cloud.dialogflow.v2.StreamingDetectIntentRequest
Methods
__construct
Constructor.
Parameters | |
---|---|
Name | Description |
data |
array
Optional. Data for populating the Message object. |
↳ session |
string
Required. The name of the session the query is sent to. Format of the session name: |
↳ query_params |
Google\Cloud\Dialogflow\V2\QueryParameters
The parameters of this query. |
↳ query_input |
Google\Cloud\Dialogflow\V2\QueryInput
Required. The input specification. It can be set to: 1. an audio config which instructs the speech recognizer how to process the speech audio, 2. a conversational query in the form of text, or 3. an event that specifies which intent to trigger. |
↳ single_utterance |
bool
Please use InputAudioConfig.single_utterance instead. If |
↳ output_audio_config |
Google\Cloud\Dialogflow\V2\OutputAudioConfig
Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated. |
↳ output_audio_config_mask |
Google\Protobuf\FieldMask
Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level. If unspecified or empty, output_audio_config replaces the agent-level config in its entirety. |
↳ input_audio |
string
The input audio content to be recognized. Must be sent if |
getSession
Required. The name of the session the query is sent to.
Format of the session name:
projects/<Project ID>/agent/sessions/<Session ID>
, or
projects/<Project ID>/agent/environments/<Environment ID>/users/<User
ID>/sessions/<Session ID>
. If Environment ID
is not specified, we assume
default 'draft' environment. If User ID
is not specified, we are using
"-". It's up to the API caller to choose an appropriate Session ID
and
User Id
. They can be a random number or some type of user and session
identifiers (preferably hashed). The length of the Session ID
and
User ID
must not exceed 36 characters.
For more information, see the API interactions
guide.
Note: Always use agent versions for production traffic.
See Versions and
environments.
Generated from protobuf field string session = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = {
Returns | |
---|---|
Type | Description |
string |
setSession
Required. The name of the session the query is sent to.
Format of the session name:
projects/<Project ID>/agent/sessions/<Session ID>
, or
projects/<Project ID>/agent/environments/<Environment ID>/users/<User
ID>/sessions/<Session ID>
. If Environment ID
is not specified, we assume
default 'draft' environment. If User ID
is not specified, we are using
"-". It's up to the API caller to choose an appropriate Session ID
and
User Id
. They can be a random number or some type of user and session
identifiers (preferably hashed). The length of the Session ID
and
User ID
must not exceed 36 characters.
For more information, see the API interactions
guide.
Note: Always use agent versions for production traffic.
See Versions and
environments.
Generated from protobuf field string session = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = {
Parameter | |
---|---|
Name | Description |
var |
string
|
Returns | |
---|---|
Type | Description |
$this |
getQueryParams
The parameters of this query.
Generated from protobuf field .google.cloud.dialogflow.v2.QueryParameters query_params = 2;
Returns | |
---|---|
Type | Description |
Google\Cloud\Dialogflow\V2\QueryParameters|null |
hasQueryParams
clearQueryParams
setQueryParams
The parameters of this query.
Generated from protobuf field .google.cloud.dialogflow.v2.QueryParameters query_params = 2;
Parameter | |
---|---|
Name | Description |
var |
Google\Cloud\Dialogflow\V2\QueryParameters
|
Returns | |
---|---|
Type | Description |
$this |
getQueryInput
Required. The input specification. It can be set to:
- an audio config which instructs the speech recognizer how to process the speech audio,
- a conversational query in the form of text, or
- an event that specifies which intent to trigger.
Generated from protobuf field .google.cloud.dialogflow.v2.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
Returns | |
---|---|
Type | Description |
Google\Cloud\Dialogflow\V2\QueryInput|null |
hasQueryInput
clearQueryInput
setQueryInput
Required. The input specification. It can be set to:
- an audio config which instructs the speech recognizer how to process the speech audio,
- a conversational query in the form of text, or
- an event that specifies which intent to trigger.
Generated from protobuf field .google.cloud.dialogflow.v2.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
Parameter | |
---|---|
Name | Description |
var |
Google\Cloud\Dialogflow\V2\QueryInput
|
Returns | |
---|---|
Type | Description |
$this |
getSingleUtterance
Please use InputAudioConfig.single_utterance instead.
If false
(default), recognition does not cease until
the client closes the stream. If true
, the recognizer will detect a
single spoken utterance in input audio. Recognition ceases when it detects
the audio's voice has stopped or paused. In this case, once a detected
intent is received, the client should close the stream and start a new
request with a new stream as needed.
This setting is ignored when query_input
is a piece of text or an event.
Generated from protobuf field bool single_utterance = 4 [deprecated = true];
Returns | |
---|---|
Type | Description |
bool |
setSingleUtterance
Please use InputAudioConfig.single_utterance instead.
If false
(default), recognition does not cease until
the client closes the stream. If true
, the recognizer will detect a
single spoken utterance in input audio. Recognition ceases when it detects
the audio's voice has stopped or paused. In this case, once a detected
intent is received, the client should close the stream and start a new
request with a new stream as needed.
This setting is ignored when query_input
is a piece of text or an event.
Generated from protobuf field bool single_utterance = 4 [deprecated = true];
Parameter | |
---|---|
Name | Description |
var |
bool
|
Returns | |
---|---|
Type | Description |
$this |
getOutputAudioConfig
Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
Generated from protobuf field .google.cloud.dialogflow.v2.OutputAudioConfig output_audio_config = 5;
Returns | |
---|---|
Type | Description |
Google\Cloud\Dialogflow\V2\OutputAudioConfig|null |
hasOutputAudioConfig
clearOutputAudioConfig
setOutputAudioConfig
Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
Generated from protobuf field .google.cloud.dialogflow.v2.OutputAudioConfig output_audio_config = 5;
Parameter | |
---|---|
Name | Description |
var |
Google\Cloud\Dialogflow\V2\OutputAudioConfig
|
Returns | |
---|---|
Type | Description |
$this |
getOutputAudioConfigMask
Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level.
If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
Generated from protobuf field .google.protobuf.FieldMask output_audio_config_mask = 7;
Returns | |
---|---|
Type | Description |
Google\Protobuf\FieldMask|null |
hasOutputAudioConfigMask
clearOutputAudioConfigMask
setOutputAudioConfigMask
Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level.
If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
Generated from protobuf field .google.protobuf.FieldMask output_audio_config_mask = 7;
Parameter | |
---|---|
Name | Description |
var |
Google\Protobuf\FieldMask
|
Returns | |
---|---|
Type | Description |
$this |
getInputAudio
The input audio content to be recognized. Must be sent if
query_input
was set to a streaming input audio config. The complete audio
over all streaming messages must not exceed 1 minute.
Generated from protobuf field bytes input_audio = 6;
Returns | |
---|---|
Type | Description |
string |
setInputAudio
The input audio content to be recognized. Must be sent if
query_input
was set to a streaming input audio config. The complete audio
over all streaming messages must not exceed 1 minute.
Generated from protobuf field bytes input_audio = 6;
Parameter | |
---|---|
Name | Description |
var |
string
|
Returns | |
---|---|
Type | Description |
$this |