- 4.58.0 (latest)
- 4.57.0
- 4.55.0
- 4.54.0
- 4.53.0
- 4.52.0
- 4.51.0
- 4.50.0
- 4.49.0
- 4.48.0
- 4.47.0
- 4.46.0
- 4.45.0
- 4.43.0
- 4.42.0
- 4.41.0
- 4.40.0
- 4.39.0
- 4.38.0
- 4.37.0
- 4.36.0
- 4.35.0
- 4.34.0
- 4.33.0
- 4.30.0
- 4.29.0
- 4.28.0
- 4.27.0
- 4.26.0
- 4.25.0
- 4.24.0
- 4.23.0
- 4.22.0
- 4.21.0
- 4.20.0
- 4.19.0
- 4.18.0
- 4.17.0
- 4.15.0
- 4.14.0
- 4.13.0
- 4.12.0
- 4.11.0
- 4.10.0
- 4.9.1
- 4.8.6
- 4.7.5
- 4.6.0
- 4.5.11
- 4.4.0
- 4.3.1
public static final class StreamingDetectIntentRequest.Builder extends GeneratedMessageV3.Builder<StreamingDetectIntentRequest.Builder> implements StreamingDetectIntentRequestOrBuilder
The top-level message sent by the client to the Sessions.StreamingDetectIntent method. Multiple request messages should be sent in order:
- The first message must contain session, query_input plus optionally query_params. If the client wants to receive an audio response, it should also contain output_audio_config. The message must not contain input_audio.
- If query_input was set to
query_input.audio_config, all subsequent
messages must contain
input_audio to continue with
Speech recognition.
If you decide to rather detect an intent from text input after you
already started Speech recognition, please send a message with
query_input.text.
However, note that:
- Dialogflow will bill you for the audio duration so far.
- Dialogflow discards all Speech recognition results in favor of the input text.
- Dialogflow will use the language code from the first message. After you sent all input, you must half-close or abort the request stream.
Protobuf type google.cloud.dialogflow.v2.StreamingDetectIntentRequest
Inheritance
Object > AbstractMessageLite.Builder<MessageType,BuilderType> > AbstractMessage.Builder<BuilderType> > GeneratedMessageV3.Builder > StreamingDetectIntentRequest.BuilderImplements
StreamingDetectIntentRequestOrBuilderStatic Methods
getDescriptor()
public static final Descriptors.Descriptor getDescriptor()
Type | Description |
Descriptor |
Methods
addRepeatedField(Descriptors.FieldDescriptor field, Object value)
public StreamingDetectIntentRequest.Builder addRepeatedField(Descriptors.FieldDescriptor field, Object value)
Name | Description |
field | FieldDescriptor |
value | Object |
Type | Description |
StreamingDetectIntentRequest.Builder |
build()
public StreamingDetectIntentRequest build()
Type | Description |
StreamingDetectIntentRequest |
buildPartial()
public StreamingDetectIntentRequest buildPartial()
Type | Description |
StreamingDetectIntentRequest |
clear()
public StreamingDetectIntentRequest.Builder clear()
Type | Description |
StreamingDetectIntentRequest.Builder |
clearField(Descriptors.FieldDescriptor field)
public StreamingDetectIntentRequest.Builder clearField(Descriptors.FieldDescriptor field)
Name | Description |
field | FieldDescriptor |
Type | Description |
StreamingDetectIntentRequest.Builder |
clearInputAudio()
public StreamingDetectIntentRequest.Builder clearInputAudio()
The input audio content to be recognized. Must be sent if
query_input
was set to a streaming input audio config. The complete audio
over all streaming messages must not exceed 1 minute.
bytes input_audio = 6;
Type | Description |
StreamingDetectIntentRequest.Builder | This builder for chaining. |
clearOneof(Descriptors.OneofDescriptor oneof)
public StreamingDetectIntentRequest.Builder clearOneof(Descriptors.OneofDescriptor oneof)
Name | Description |
oneof | OneofDescriptor |
Type | Description |
StreamingDetectIntentRequest.Builder |
clearOutputAudioConfig()
public StreamingDetectIntentRequest.Builder clearOutputAudioConfig()
Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
.google.cloud.dialogflow.v2.OutputAudioConfig output_audio_config = 5;
Type | Description |
StreamingDetectIntentRequest.Builder |
clearOutputAudioConfigMask()
public StreamingDetectIntentRequest.Builder clearOutputAudioConfigMask()
Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level. If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
.google.protobuf.FieldMask output_audio_config_mask = 7;
Type | Description |
StreamingDetectIntentRequest.Builder |
clearQueryInput()
public StreamingDetectIntentRequest.Builder clearQueryInput()
Required. The input specification. It can be set to:
- an audio config which instructs the speech recognizer how to process the speech audio,
- a conversational query in the form of text, or
- an event that specifies which intent to trigger.
.google.cloud.dialogflow.v2.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
Type | Description |
StreamingDetectIntentRequest.Builder |
clearQueryParams()
public StreamingDetectIntentRequest.Builder clearQueryParams()
The parameters of this query.
.google.cloud.dialogflow.v2.QueryParameters query_params = 2;
Type | Description |
StreamingDetectIntentRequest.Builder |
clearSession()
public StreamingDetectIntentRequest.Builder clearSession()
Required. The name of the session the query is sent to.
Format of the session name:
projects/<Project ID>/agent/sessions/<Session ID>
, or
projects/<Project ID>/agent/environments/<Environment ID>/users/<User
ID>/sessions/<Session ID>
. If Environment ID
is not specified, we assume
default 'draft' environment. If User ID
is not specified, we are using
"-". It's up to the API caller to choose an appropriate Session ID
and
User Id
. They can be a random number or some type of user and session
identifiers (preferably hashed). The length of the Session ID
and
User ID
must not exceed 36 characters.
For more information, see the API interactions
guide.
Note: Always use agent versions for production traffic.
See Versions and
environments.
string session = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... }
Type | Description |
StreamingDetectIntentRequest.Builder | This builder for chaining. |
clearSingleUtterance() (deprecated)
public StreamingDetectIntentRequest.Builder clearSingleUtterance()
Deprecated. google.cloud.dialogflow.v2.StreamingDetectIntentRequest.single_utterance is deprecated. See google/cloud/dialogflow/v2/session.proto;l=457
Please use InputAudioConfig.single_utterance instead.
If false
(default), recognition does not cease until
the client closes the stream. If true
, the recognizer will detect a
single spoken utterance in input audio. Recognition ceases when it detects
the audio's voice has stopped or paused. In this case, once a detected
intent is received, the client should close the stream and start a new
request with a new stream as needed.
This setting is ignored when query_input
is a piece of text or an event.
bool single_utterance = 4 [deprecated = true];
Type | Description |
StreamingDetectIntentRequest.Builder | This builder for chaining. |
clone()
public StreamingDetectIntentRequest.Builder clone()
Type | Description |
StreamingDetectIntentRequest.Builder |
getDefaultInstanceForType()
public StreamingDetectIntentRequest getDefaultInstanceForType()
Type | Description |
StreamingDetectIntentRequest |
getDescriptorForType()
public Descriptors.Descriptor getDescriptorForType()
Type | Description |
Descriptor |
getInputAudio()
public ByteString getInputAudio()
The input audio content to be recognized. Must be sent if
query_input
was set to a streaming input audio config. The complete audio
over all streaming messages must not exceed 1 minute.
bytes input_audio = 6;
Type | Description |
ByteString | The inputAudio. |
getOutputAudioConfig()
public OutputAudioConfig getOutputAudioConfig()
Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
.google.cloud.dialogflow.v2.OutputAudioConfig output_audio_config = 5;
Type | Description |
OutputAudioConfig | The outputAudioConfig. |
getOutputAudioConfigBuilder()
public OutputAudioConfig.Builder getOutputAudioConfigBuilder()
Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
.google.cloud.dialogflow.v2.OutputAudioConfig output_audio_config = 5;
Type | Description |
OutputAudioConfig.Builder |
getOutputAudioConfigMask()
public FieldMask getOutputAudioConfigMask()
Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level. If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
.google.protobuf.FieldMask output_audio_config_mask = 7;
Type | Description |
FieldMask | The outputAudioConfigMask. |
getOutputAudioConfigMaskBuilder()
public FieldMask.Builder getOutputAudioConfigMaskBuilder()
Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level. If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
.google.protobuf.FieldMask output_audio_config_mask = 7;
Type | Description |
Builder |
getOutputAudioConfigMaskOrBuilder()
public FieldMaskOrBuilder getOutputAudioConfigMaskOrBuilder()
Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level. If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
.google.protobuf.FieldMask output_audio_config_mask = 7;
Type | Description |
FieldMaskOrBuilder |
getOutputAudioConfigOrBuilder()
public OutputAudioConfigOrBuilder getOutputAudioConfigOrBuilder()
Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
.google.cloud.dialogflow.v2.OutputAudioConfig output_audio_config = 5;
Type | Description |
OutputAudioConfigOrBuilder |
getQueryInput()
public QueryInput getQueryInput()
Required. The input specification. It can be set to:
- an audio config which instructs the speech recognizer how to process the speech audio,
- a conversational query in the form of text, or
- an event that specifies which intent to trigger.
.google.cloud.dialogflow.v2.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
Type | Description |
QueryInput | The queryInput. |
getQueryInputBuilder()
public QueryInput.Builder getQueryInputBuilder()
Required. The input specification. It can be set to:
- an audio config which instructs the speech recognizer how to process the speech audio,
- a conversational query in the form of text, or
- an event that specifies which intent to trigger.
.google.cloud.dialogflow.v2.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
Type | Description |
QueryInput.Builder |
getQueryInputOrBuilder()
public QueryInputOrBuilder getQueryInputOrBuilder()
Required. The input specification. It can be set to:
- an audio config which instructs the speech recognizer how to process the speech audio,
- a conversational query in the form of text, or
- an event that specifies which intent to trigger.
.google.cloud.dialogflow.v2.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
Type | Description |
QueryInputOrBuilder |
getQueryParams()
public QueryParameters getQueryParams()
The parameters of this query.
.google.cloud.dialogflow.v2.QueryParameters query_params = 2;
Type | Description |
QueryParameters | The queryParams. |
getQueryParamsBuilder()
public QueryParameters.Builder getQueryParamsBuilder()
The parameters of this query.
.google.cloud.dialogflow.v2.QueryParameters query_params = 2;
Type | Description |
QueryParameters.Builder |
getQueryParamsOrBuilder()
public QueryParametersOrBuilder getQueryParamsOrBuilder()
The parameters of this query.
.google.cloud.dialogflow.v2.QueryParameters query_params = 2;
Type | Description |
QueryParametersOrBuilder |
getSession()
public String getSession()
Required. The name of the session the query is sent to.
Format of the session name:
projects/<Project ID>/agent/sessions/<Session ID>
, or
projects/<Project ID>/agent/environments/<Environment ID>/users/<User
ID>/sessions/<Session ID>
. If Environment ID
is not specified, we assume
default 'draft' environment. If User ID
is not specified, we are using
"-". It's up to the API caller to choose an appropriate Session ID
and
User Id
. They can be a random number or some type of user and session
identifiers (preferably hashed). The length of the Session ID
and
User ID
must not exceed 36 characters.
For more information, see the API interactions
guide.
Note: Always use agent versions for production traffic.
See Versions and
environments.
string session = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... }
Type | Description |
String | The session. |
getSessionBytes()
public ByteString getSessionBytes()
Required. The name of the session the query is sent to.
Format of the session name:
projects/<Project ID>/agent/sessions/<Session ID>
, or
projects/<Project ID>/agent/environments/<Environment ID>/users/<User
ID>/sessions/<Session ID>
. If Environment ID
is not specified, we assume
default 'draft' environment. If User ID
is not specified, we are using
"-". It's up to the API caller to choose an appropriate Session ID
and
User Id
. They can be a random number or some type of user and session
identifiers (preferably hashed). The length of the Session ID
and
User ID
must not exceed 36 characters.
For more information, see the API interactions
guide.
Note: Always use agent versions for production traffic.
See Versions and
environments.
string session = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... }
Type | Description |
ByteString | The bytes for session. |
getSingleUtterance() (deprecated)
public boolean getSingleUtterance()
Deprecated. google.cloud.dialogflow.v2.StreamingDetectIntentRequest.single_utterance is deprecated. See google/cloud/dialogflow/v2/session.proto;l=457
Please use InputAudioConfig.single_utterance instead.
If false
(default), recognition does not cease until
the client closes the stream. If true
, the recognizer will detect a
single spoken utterance in input audio. Recognition ceases when it detects
the audio's voice has stopped or paused. In this case, once a detected
intent is received, the client should close the stream and start a new
request with a new stream as needed.
This setting is ignored when query_input
is a piece of text or an event.
bool single_utterance = 4 [deprecated = true];
Type | Description |
boolean | The singleUtterance. |
hasOutputAudioConfig()
public boolean hasOutputAudioConfig()
Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
.google.cloud.dialogflow.v2.OutputAudioConfig output_audio_config = 5;
Type | Description |
boolean | Whether the outputAudioConfig field is set. |
hasOutputAudioConfigMask()
public boolean hasOutputAudioConfigMask()
Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level. If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
.google.protobuf.FieldMask output_audio_config_mask = 7;
Type | Description |
boolean | Whether the outputAudioConfigMask field is set. |
hasQueryInput()
public boolean hasQueryInput()
Required. The input specification. It can be set to:
- an audio config which instructs the speech recognizer how to process the speech audio,
- a conversational query in the form of text, or
- an event that specifies which intent to trigger.
.google.cloud.dialogflow.v2.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
Type | Description |
boolean | Whether the queryInput field is set. |
hasQueryParams()
public boolean hasQueryParams()
The parameters of this query.
.google.cloud.dialogflow.v2.QueryParameters query_params = 2;
Type | Description |
boolean | Whether the queryParams field is set. |
internalGetFieldAccessorTable()
protected GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable()
Type | Description |
FieldAccessorTable |
isInitialized()
public final boolean isInitialized()
Type | Description |
boolean |
mergeFrom(StreamingDetectIntentRequest other)
public StreamingDetectIntentRequest.Builder mergeFrom(StreamingDetectIntentRequest other)
Name | Description |
other | StreamingDetectIntentRequest |
Type | Description |
StreamingDetectIntentRequest.Builder |
mergeFrom(CodedInputStream input, ExtensionRegistryLite extensionRegistry)
public StreamingDetectIntentRequest.Builder mergeFrom(CodedInputStream input, ExtensionRegistryLite extensionRegistry)
Name | Description |
input | CodedInputStream |
extensionRegistry | ExtensionRegistryLite |
Type | Description |
StreamingDetectIntentRequest.Builder |
Type | Description |
IOException |
mergeFrom(Message other)
public StreamingDetectIntentRequest.Builder mergeFrom(Message other)
Name | Description |
other | Message |
Type | Description |
StreamingDetectIntentRequest.Builder |
mergeOutputAudioConfig(OutputAudioConfig value)
public StreamingDetectIntentRequest.Builder mergeOutputAudioConfig(OutputAudioConfig value)
Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
.google.cloud.dialogflow.v2.OutputAudioConfig output_audio_config = 5;
Name | Description |
value | OutputAudioConfig |
Type | Description |
StreamingDetectIntentRequest.Builder |
mergeOutputAudioConfigMask(FieldMask value)
public StreamingDetectIntentRequest.Builder mergeOutputAudioConfigMask(FieldMask value)
Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level. If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
.google.protobuf.FieldMask output_audio_config_mask = 7;
Name | Description |
value | FieldMask |
Type | Description |
StreamingDetectIntentRequest.Builder |
mergeQueryInput(QueryInput value)
public StreamingDetectIntentRequest.Builder mergeQueryInput(QueryInput value)
Required. The input specification. It can be set to:
- an audio config which instructs the speech recognizer how to process the speech audio,
- a conversational query in the form of text, or
- an event that specifies which intent to trigger.
.google.cloud.dialogflow.v2.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
Name | Description |
value | QueryInput |
Type | Description |
StreamingDetectIntentRequest.Builder |
mergeQueryParams(QueryParameters value)
public StreamingDetectIntentRequest.Builder mergeQueryParams(QueryParameters value)
The parameters of this query.
.google.cloud.dialogflow.v2.QueryParameters query_params = 2;
Name | Description |
value | QueryParameters |
Type | Description |
StreamingDetectIntentRequest.Builder |
mergeUnknownFields(UnknownFieldSet unknownFields)
public final StreamingDetectIntentRequest.Builder mergeUnknownFields(UnknownFieldSet unknownFields)
Name | Description |
unknownFields | UnknownFieldSet |
Type | Description |
StreamingDetectIntentRequest.Builder |
setField(Descriptors.FieldDescriptor field, Object value)
public StreamingDetectIntentRequest.Builder setField(Descriptors.FieldDescriptor field, Object value)
Name | Description |
field | FieldDescriptor |
value | Object |
Type | Description |
StreamingDetectIntentRequest.Builder |
setInputAudio(ByteString value)
public StreamingDetectIntentRequest.Builder setInputAudio(ByteString value)
The input audio content to be recognized. Must be sent if
query_input
was set to a streaming input audio config. The complete audio
over all streaming messages must not exceed 1 minute.
bytes input_audio = 6;
Name | Description |
value | ByteString The inputAudio to set. |
Type | Description |
StreamingDetectIntentRequest.Builder | This builder for chaining. |
setOutputAudioConfig(OutputAudioConfig value)
public StreamingDetectIntentRequest.Builder setOutputAudioConfig(OutputAudioConfig value)
Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
.google.cloud.dialogflow.v2.OutputAudioConfig output_audio_config = 5;
Name | Description |
value | OutputAudioConfig |
Type | Description |
StreamingDetectIntentRequest.Builder |
setOutputAudioConfig(OutputAudioConfig.Builder builderForValue)
public StreamingDetectIntentRequest.Builder setOutputAudioConfig(OutputAudioConfig.Builder builderForValue)
Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
.google.cloud.dialogflow.v2.OutputAudioConfig output_audio_config = 5;
Name | Description |
builderForValue | OutputAudioConfig.Builder |
Type | Description |
StreamingDetectIntentRequest.Builder |
setOutputAudioConfigMask(FieldMask value)
public StreamingDetectIntentRequest.Builder setOutputAudioConfigMask(FieldMask value)
Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level. If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
.google.protobuf.FieldMask output_audio_config_mask = 7;
Name | Description |
value | FieldMask |
Type | Description |
StreamingDetectIntentRequest.Builder |
setOutputAudioConfigMask(FieldMask.Builder builderForValue)
public StreamingDetectIntentRequest.Builder setOutputAudioConfigMask(FieldMask.Builder builderForValue)
Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level. If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
.google.protobuf.FieldMask output_audio_config_mask = 7;
Name | Description |
builderForValue | Builder |
Type | Description |
StreamingDetectIntentRequest.Builder |
setQueryInput(QueryInput value)
public StreamingDetectIntentRequest.Builder setQueryInput(QueryInput value)
Required. The input specification. It can be set to:
- an audio config which instructs the speech recognizer how to process the speech audio,
- a conversational query in the form of text, or
- an event that specifies which intent to trigger.
.google.cloud.dialogflow.v2.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
Name | Description |
value | QueryInput |
Type | Description |
StreamingDetectIntentRequest.Builder |
setQueryInput(QueryInput.Builder builderForValue)
public StreamingDetectIntentRequest.Builder setQueryInput(QueryInput.Builder builderForValue)
Required. The input specification. It can be set to:
- an audio config which instructs the speech recognizer how to process the speech audio,
- a conversational query in the form of text, or
- an event that specifies which intent to trigger.
.google.cloud.dialogflow.v2.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
Name | Description |
builderForValue | QueryInput.Builder |
Type | Description |
StreamingDetectIntentRequest.Builder |
setQueryParams(QueryParameters value)
public StreamingDetectIntentRequest.Builder setQueryParams(QueryParameters value)
The parameters of this query.
.google.cloud.dialogflow.v2.QueryParameters query_params = 2;
Name | Description |
value | QueryParameters |
Type | Description |
StreamingDetectIntentRequest.Builder |
setQueryParams(QueryParameters.Builder builderForValue)
public StreamingDetectIntentRequest.Builder setQueryParams(QueryParameters.Builder builderForValue)
The parameters of this query.
.google.cloud.dialogflow.v2.QueryParameters query_params = 2;
Name | Description |
builderForValue | QueryParameters.Builder |
Type | Description |
StreamingDetectIntentRequest.Builder |
setRepeatedField(Descriptors.FieldDescriptor field, int index, Object value)
public StreamingDetectIntentRequest.Builder setRepeatedField(Descriptors.FieldDescriptor field, int index, Object value)
Name | Description |
field | FieldDescriptor |
index | int |
value | Object |
Type | Description |
StreamingDetectIntentRequest.Builder |
setSession(String value)
public StreamingDetectIntentRequest.Builder setSession(String value)
Required. The name of the session the query is sent to.
Format of the session name:
projects/<Project ID>/agent/sessions/<Session ID>
, or
projects/<Project ID>/agent/environments/<Environment ID>/users/<User
ID>/sessions/<Session ID>
. If Environment ID
is not specified, we assume
default 'draft' environment. If User ID
is not specified, we are using
"-". It's up to the API caller to choose an appropriate Session ID
and
User Id
. They can be a random number or some type of user and session
identifiers (preferably hashed). The length of the Session ID
and
User ID
must not exceed 36 characters.
For more information, see the API interactions
guide.
Note: Always use agent versions for production traffic.
See Versions and
environments.
string session = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... }
Name | Description |
value | String The session to set. |
Type | Description |
StreamingDetectIntentRequest.Builder | This builder for chaining. |
setSessionBytes(ByteString value)
public StreamingDetectIntentRequest.Builder setSessionBytes(ByteString value)
Required. The name of the session the query is sent to.
Format of the session name:
projects/<Project ID>/agent/sessions/<Session ID>
, or
projects/<Project ID>/agent/environments/<Environment ID>/users/<User
ID>/sessions/<Session ID>
. If Environment ID
is not specified, we assume
default 'draft' environment. If User ID
is not specified, we are using
"-". It's up to the API caller to choose an appropriate Session ID
and
User Id
. They can be a random number or some type of user and session
identifiers (preferably hashed). The length of the Session ID
and
User ID
must not exceed 36 characters.
For more information, see the API interactions
guide.
Note: Always use agent versions for production traffic.
See Versions and
environments.
string session = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... }
Name | Description |
value | ByteString The bytes for session to set. |
Type | Description |
StreamingDetectIntentRequest.Builder | This builder for chaining. |
setSingleUtterance(boolean value) (deprecated)
public StreamingDetectIntentRequest.Builder setSingleUtterance(boolean value)
Deprecated. google.cloud.dialogflow.v2.StreamingDetectIntentRequest.single_utterance is deprecated. See google/cloud/dialogflow/v2/session.proto;l=457
Please use InputAudioConfig.single_utterance instead.
If false
(default), recognition does not cease until
the client closes the stream. If true
, the recognizer will detect a
single spoken utterance in input audio. Recognition ceases when it detects
the audio's voice has stopped or paused. In this case, once a detected
intent is received, the client should close the stream and start a new
request with a new stream as needed.
This setting is ignored when query_input
is a piece of text or an event.
bool single_utterance = 4 [deprecated = true];
Name | Description |
value | boolean The singleUtterance to set. |
Type | Description |
StreamingDetectIntentRequest.Builder | This builder for chaining. |
setUnknownFields(UnknownFieldSet unknownFields)
public final StreamingDetectIntentRequest.Builder setUnknownFields(UnknownFieldSet unknownFields)
Name | Description |
unknownFields | UnknownFieldSet |
Type | Description |
StreamingDetectIntentRequest.Builder |