- 4.61.0 (latest)
- 4.60.0
- 4.59.0
- 4.58.0
- 4.57.0
- 4.55.0
- 4.54.0
- 4.53.0
- 4.52.0
- 4.51.0
- 4.50.0
- 4.49.0
- 4.48.0
- 4.47.0
- 4.46.0
- 4.45.0
- 4.43.0
- 4.42.0
- 4.41.0
- 4.40.0
- 4.39.0
- 4.38.0
- 4.37.0
- 4.36.0
- 4.35.0
- 4.34.0
- 4.33.0
- 4.30.0
- 4.29.0
- 4.28.0
- 4.27.0
- 4.26.0
- 4.25.0
- 4.24.0
- 4.23.0
- 4.22.0
- 4.21.0
- 4.20.0
- 4.19.0
- 4.18.0
- 4.17.0
- 4.15.0
- 4.14.0
- 4.13.0
- 4.12.0
- 4.11.0
- 4.10.0
- 4.9.1
- 4.8.6
- 4.7.5
- 4.6.0
- 4.5.11
- 4.4.0
- 4.3.1
public static final class StreamingDetectIntentRequest.Builder extends GeneratedMessageV3.Builder<StreamingDetectIntentRequest.Builder> implements StreamingDetectIntentRequestOrBuilder
The top-level message sent by the client to the Sessions.StreamingDetectIntent method. Multiple request messages should be sent in order:
- The first message must contain session, query_input plus optionally query_params. If the client wants to receive an audio response, it should also contain output_audio_config. The message must not contain input_audio.
- If
query_input
was set to
query_input.audio_config,
all subsequent messages must contain
input_audio
to continue with Speech recognition. If you decide to rather detect an
intent from text input after you already started Speech recognition,
please send a message with
query_input.text.
However, note that:
- Dialogflow will bill you for the audio duration so far.
- Dialogflow discards all Speech recognition results in favor of the input text.
- Dialogflow will use the language code from the first message. After you sent all input, you must half-close or abort the request stream.
Protobuf type google.cloud.dialogflow.v2beta1.StreamingDetectIntentRequest
Inheritance
Object > AbstractMessageLite.Builder<MessageType,BuilderType> > AbstractMessage.Builder<BuilderType> > GeneratedMessageV3.Builder > StreamingDetectIntentRequest.BuilderImplements
StreamingDetectIntentRequestOrBuilderStatic Methods
getDescriptor()
public static final Descriptors.Descriptor getDescriptor()
Type | Description |
Descriptor |
Methods
addRepeatedField(Descriptors.FieldDescriptor field, Object value)
public StreamingDetectIntentRequest.Builder addRepeatedField(Descriptors.FieldDescriptor field, Object value)
Name | Description |
field |
FieldDescriptor |
value |
Object |
Type | Description |
StreamingDetectIntentRequest.Builder |
build()
public StreamingDetectIntentRequest build()
Type | Description |
StreamingDetectIntentRequest |
buildPartial()
public StreamingDetectIntentRequest buildPartial()
Type | Description |
StreamingDetectIntentRequest |
clear()
public StreamingDetectIntentRequest.Builder clear()
Type | Description |
StreamingDetectIntentRequest.Builder |
clearField(Descriptors.FieldDescriptor field)
public StreamingDetectIntentRequest.Builder clearField(Descriptors.FieldDescriptor field)
Name | Description |
field |
FieldDescriptor |
Type | Description |
StreamingDetectIntentRequest.Builder |
clearInputAudio()
public StreamingDetectIntentRequest.Builder clearInputAudio()
The input audio content to be recognized. Must be sent if
query_input
was set to a streaming input audio config. The complete audio
over all streaming messages must not exceed 1 minute.
bytes input_audio = 6;
Type | Description |
StreamingDetectIntentRequest.Builder |
This builder for chaining. |
clearOneof(Descriptors.OneofDescriptor oneof)
public StreamingDetectIntentRequest.Builder clearOneof(Descriptors.OneofDescriptor oneof)
Name | Description |
oneof |
OneofDescriptor |
Type | Description |
StreamingDetectIntentRequest.Builder |
clearOutputAudioConfig()
public StreamingDetectIntentRequest.Builder clearOutputAudioConfig()
Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
.google.cloud.dialogflow.v2beta1.OutputAudioConfig output_audio_config = 5;
Type | Description |
StreamingDetectIntentRequest.Builder |
clearOutputAudioConfigMask()
public StreamingDetectIntentRequest.Builder clearOutputAudioConfigMask()
Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level. If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
.google.protobuf.FieldMask output_audio_config_mask = 7;
Type | Description |
StreamingDetectIntentRequest.Builder |
clearQueryInput()
public StreamingDetectIntentRequest.Builder clearQueryInput()
Required. The input specification. It can be set to:
- an audio config which instructs the speech recognizer how to process the speech audio,
- a conversational query in the form of text, or
- an event that specifies which intent to trigger.
.google.cloud.dialogflow.v2beta1.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
Type | Description |
StreamingDetectIntentRequest.Builder |
clearQueryParams()
public StreamingDetectIntentRequest.Builder clearQueryParams()
The parameters of this query.
.google.cloud.dialogflow.v2beta1.QueryParameters query_params = 2;
Type | Description |
StreamingDetectIntentRequest.Builder |
clearSession()
public StreamingDetectIntentRequest.Builder clearSession()
Required. The name of the session the query is sent to. Supported formats:
projects/<Project ID>/agent/sessions/<Session ID>,
- projects/<Project ID>/locations/<Location ID>/agent/sessions/<Session
ID>
,
- projects/<Project ID>/agent/environments/<Environment ID>/users/<User
ID>/sessions/<Session ID>
,
- projects/<Project ID>/locations/<Location
ID>/agent/environments/<Environment ID>/users/<User ID>/sessions/<Session
ID>
, If
Location IDis not specified we assume default 'us' location. If
Environment IDis not specified, we assume default 'draft' environment. If
User IDis not specified, we are using "-". It's up to the API caller to choose an appropriate
Session IDand
User Id. They can be a random number or some type of user and session identifiers (preferably hashed). The length of the
Session IDand
User ID` must not exceed 36 characters. For more information, see the API interactions guide. Note: Always use agent versions for production traffic. See Versions and environments.
string session = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... }
Type | Description |
StreamingDetectIntentRequest.Builder |
This builder for chaining. |
clearSingleUtterance() (deprecated)
public StreamingDetectIntentRequest.Builder clearSingleUtterance()
Deprecated. google.cloud.dialogflow.v2beta1.StreamingDetectIntentRequest.single_utterance is deprecated. See google/cloud/dialogflow/v2beta1/session.proto;l=564
DEPRECATED. Please use
InputAudioConfig.single_utterance
instead. If false
(default), recognition does not cease until the client
closes the stream. If true
, the recognizer will detect a single spoken
utterance in input audio. Recognition ceases when it detects the audio's
voice has stopped or paused. In this case, once a detected intent is
received, the client should close the stream and start a new request with a
new stream as needed. This setting is ignored when query_input
is a piece
of text or an event.
bool single_utterance = 4 [deprecated = true];
Type | Description |
StreamingDetectIntentRequest.Builder |
This builder for chaining. |
clone()
public StreamingDetectIntentRequest.Builder clone()
Type | Description |
StreamingDetectIntentRequest.Builder |
getDefaultInstanceForType()
public StreamingDetectIntentRequest getDefaultInstanceForType()
Type | Description |
StreamingDetectIntentRequest |
getDescriptorForType()
public Descriptors.Descriptor getDescriptorForType()
Type | Description |
Descriptor |
getInputAudio()
public ByteString getInputAudio()
The input audio content to be recognized. Must be sent if
query_input
was set to a streaming input audio config. The complete audio
over all streaming messages must not exceed 1 minute.
bytes input_audio = 6;
Type | Description |
ByteString |
The inputAudio. |
getOutputAudioConfig()
public OutputAudioConfig getOutputAudioConfig()
Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
.google.cloud.dialogflow.v2beta1.OutputAudioConfig output_audio_config = 5;
Type | Description |
OutputAudioConfig |
The outputAudioConfig. |
getOutputAudioConfigBuilder()
public OutputAudioConfig.Builder getOutputAudioConfigBuilder()
Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
.google.cloud.dialogflow.v2beta1.OutputAudioConfig output_audio_config = 5;
Type | Description |
OutputAudioConfig.Builder |
getOutputAudioConfigMask()
public FieldMask getOutputAudioConfigMask()
Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level. If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
.google.protobuf.FieldMask output_audio_config_mask = 7;
Type | Description |
FieldMask |
The outputAudioConfigMask. |
getOutputAudioConfigMaskBuilder()
public FieldMask.Builder getOutputAudioConfigMaskBuilder()
Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level. If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
.google.protobuf.FieldMask output_audio_config_mask = 7;
Type | Description |
Builder |
getOutputAudioConfigMaskOrBuilder()
public FieldMaskOrBuilder getOutputAudioConfigMaskOrBuilder()
Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level. If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
.google.protobuf.FieldMask output_audio_config_mask = 7;
Type | Description |
FieldMaskOrBuilder |
getOutputAudioConfigOrBuilder()
public OutputAudioConfigOrBuilder getOutputAudioConfigOrBuilder()
Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
.google.cloud.dialogflow.v2beta1.OutputAudioConfig output_audio_config = 5;
Type | Description |
OutputAudioConfigOrBuilder |
getQueryInput()
public QueryInput getQueryInput()
Required. The input specification. It can be set to:
- an audio config which instructs the speech recognizer how to process the speech audio,
- a conversational query in the form of text, or
- an event that specifies which intent to trigger.
.google.cloud.dialogflow.v2beta1.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
Type | Description |
QueryInput |
The queryInput. |
getQueryInputBuilder()
public QueryInput.Builder getQueryInputBuilder()
Required. The input specification. It can be set to:
- an audio config which instructs the speech recognizer how to process the speech audio,
- a conversational query in the form of text, or
- an event that specifies which intent to trigger.
.google.cloud.dialogflow.v2beta1.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
Type | Description |
QueryInput.Builder |
getQueryInputOrBuilder()
public QueryInputOrBuilder getQueryInputOrBuilder()
Required. The input specification. It can be set to:
- an audio config which instructs the speech recognizer how to process the speech audio,
- a conversational query in the form of text, or
- an event that specifies which intent to trigger.
.google.cloud.dialogflow.v2beta1.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
Type | Description |
QueryInputOrBuilder |
getQueryParams()
public QueryParameters getQueryParams()
The parameters of this query.
.google.cloud.dialogflow.v2beta1.QueryParameters query_params = 2;
Type | Description |
QueryParameters |
The queryParams. |
getQueryParamsBuilder()
public QueryParameters.Builder getQueryParamsBuilder()
The parameters of this query.
.google.cloud.dialogflow.v2beta1.QueryParameters query_params = 2;
Type | Description |
QueryParameters.Builder |
getQueryParamsOrBuilder()
public QueryParametersOrBuilder getQueryParamsOrBuilder()
The parameters of this query.
.google.cloud.dialogflow.v2beta1.QueryParameters query_params = 2;
Type | Description |
QueryParametersOrBuilder |
getSession()
public String getSession()
Required. The name of the session the query is sent to. Supported formats:
projects/<Project ID>/agent/sessions/<Session ID>,
- projects/<Project ID>/locations/<Location ID>/agent/sessions/<Session
ID>
,
- projects/<Project ID>/agent/environments/<Environment ID>/users/<User
ID>/sessions/<Session ID>
,
- projects/<Project ID>/locations/<Location
ID>/agent/environments/<Environment ID>/users/<User ID>/sessions/<Session
ID>
, If
Location IDis not specified we assume default 'us' location. If
Environment IDis not specified, we assume default 'draft' environment. If
User IDis not specified, we are using "-". It's up to the API caller to choose an appropriate
Session IDand
User Id. They can be a random number or some type of user and session identifiers (preferably hashed). The length of the
Session IDand
User ID` must not exceed 36 characters. For more information, see the API interactions guide. Note: Always use agent versions for production traffic. See Versions and environments.
string session = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... }
Type | Description |
String |
The session. |
getSessionBytes()
public ByteString getSessionBytes()
Required. The name of the session the query is sent to. Supported formats:
projects/<Project ID>/agent/sessions/<Session ID>,
- projects/<Project ID>/locations/<Location ID>/agent/sessions/<Session
ID>
,
- projects/<Project ID>/agent/environments/<Environment ID>/users/<User
ID>/sessions/<Session ID>
,
- projects/<Project ID>/locations/<Location
ID>/agent/environments/<Environment ID>/users/<User ID>/sessions/<Session
ID>
, If
Location IDis not specified we assume default 'us' location. If
Environment IDis not specified, we assume default 'draft' environment. If
User IDis not specified, we are using "-". It's up to the API caller to choose an appropriate
Session IDand
User Id. They can be a random number or some type of user and session identifiers (preferably hashed). The length of the
Session IDand
User ID` must not exceed 36 characters. For more information, see the API interactions guide. Note: Always use agent versions for production traffic. See Versions and environments.
string session = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... }
Type | Description |
ByteString |
The bytes for session. |
getSingleUtterance() (deprecated)
public boolean getSingleUtterance()
Deprecated. google.cloud.dialogflow.v2beta1.StreamingDetectIntentRequest.single_utterance is deprecated. See google/cloud/dialogflow/v2beta1/session.proto;l=564
DEPRECATED. Please use
InputAudioConfig.single_utterance
instead. If false
(default), recognition does not cease until the client
closes the stream. If true
, the recognizer will detect a single spoken
utterance in input audio. Recognition ceases when it detects the audio's
voice has stopped or paused. In this case, once a detected intent is
received, the client should close the stream and start a new request with a
new stream as needed. This setting is ignored when query_input
is a piece
of text or an event.
bool single_utterance = 4 [deprecated = true];
Type | Description |
boolean |
The singleUtterance. |
hasOutputAudioConfig()
public boolean hasOutputAudioConfig()
Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
.google.cloud.dialogflow.v2beta1.OutputAudioConfig output_audio_config = 5;
Type | Description |
boolean |
Whether the outputAudioConfig field is set. |
hasOutputAudioConfigMask()
public boolean hasOutputAudioConfigMask()
Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level. If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
.google.protobuf.FieldMask output_audio_config_mask = 7;
Type | Description |
boolean |
Whether the outputAudioConfigMask field is set. |
hasQueryInput()
public boolean hasQueryInput()
Required. The input specification. It can be set to:
- an audio config which instructs the speech recognizer how to process the speech audio,
- a conversational query in the form of text, or
- an event that specifies which intent to trigger.
.google.cloud.dialogflow.v2beta1.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
Type | Description |
boolean |
Whether the queryInput field is set. |
hasQueryParams()
public boolean hasQueryParams()
The parameters of this query.
.google.cloud.dialogflow.v2beta1.QueryParameters query_params = 2;
Type | Description |
boolean |
Whether the queryParams field is set. |
internalGetFieldAccessorTable()
protected GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable()
Type | Description |
FieldAccessorTable |
isInitialized()
public final boolean isInitialized()
Type | Description |
boolean |
mergeFrom(StreamingDetectIntentRequest other)
public StreamingDetectIntentRequest.Builder mergeFrom(StreamingDetectIntentRequest other)
Name | Description |
other |
StreamingDetectIntentRequest |
Type | Description |
StreamingDetectIntentRequest.Builder |
mergeFrom(CodedInputStream input, ExtensionRegistryLite extensionRegistry)
public StreamingDetectIntentRequest.Builder mergeFrom(CodedInputStream input, ExtensionRegistryLite extensionRegistry)
Name | Description |
input |
CodedInputStream |
extensionRegistry |
ExtensionRegistryLite |
Type | Description |
StreamingDetectIntentRequest.Builder |
Type | Description |
IOException |
mergeFrom(Message other)
public StreamingDetectIntentRequest.Builder mergeFrom(Message other)
Name | Description |
other |
Message |
Type | Description |
StreamingDetectIntentRequest.Builder |
mergeOutputAudioConfig(OutputAudioConfig value)
public StreamingDetectIntentRequest.Builder mergeOutputAudioConfig(OutputAudioConfig value)
Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
.google.cloud.dialogflow.v2beta1.OutputAudioConfig output_audio_config = 5;
Name | Description |
value |
OutputAudioConfig |
Type | Description |
StreamingDetectIntentRequest.Builder |
mergeOutputAudioConfigMask(FieldMask value)
public StreamingDetectIntentRequest.Builder mergeOutputAudioConfigMask(FieldMask value)
Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level. If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
.google.protobuf.FieldMask output_audio_config_mask = 7;
Name | Description |
value |
FieldMask |
Type | Description |
StreamingDetectIntentRequest.Builder |
mergeQueryInput(QueryInput value)
public StreamingDetectIntentRequest.Builder mergeQueryInput(QueryInput value)
Required. The input specification. It can be set to:
- an audio config which instructs the speech recognizer how to process the speech audio,
- a conversational query in the form of text, or
- an event that specifies which intent to trigger.
.google.cloud.dialogflow.v2beta1.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
Name | Description |
value |
QueryInput |
Type | Description |
StreamingDetectIntentRequest.Builder |
mergeQueryParams(QueryParameters value)
public StreamingDetectIntentRequest.Builder mergeQueryParams(QueryParameters value)
The parameters of this query.
.google.cloud.dialogflow.v2beta1.QueryParameters query_params = 2;
Name | Description |
value |
QueryParameters |
Type | Description |
StreamingDetectIntentRequest.Builder |
mergeUnknownFields(UnknownFieldSet unknownFields)
public final StreamingDetectIntentRequest.Builder mergeUnknownFields(UnknownFieldSet unknownFields)
Name | Description |
unknownFields |
UnknownFieldSet |
Type | Description |
StreamingDetectIntentRequest.Builder |
setField(Descriptors.FieldDescriptor field, Object value)
public StreamingDetectIntentRequest.Builder setField(Descriptors.FieldDescriptor field, Object value)
Name | Description |
field |
FieldDescriptor |
value |
Object |
Type | Description |
StreamingDetectIntentRequest.Builder |
setInputAudio(ByteString value)
public StreamingDetectIntentRequest.Builder setInputAudio(ByteString value)
The input audio content to be recognized. Must be sent if
query_input
was set to a streaming input audio config. The complete audio
over all streaming messages must not exceed 1 minute.
bytes input_audio = 6;
Name | Description |
value |
ByteString The inputAudio to set. |
Type | Description |
StreamingDetectIntentRequest.Builder |
This builder for chaining. |
setOutputAudioConfig(OutputAudioConfig value)
public StreamingDetectIntentRequest.Builder setOutputAudioConfig(OutputAudioConfig value)
Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
.google.cloud.dialogflow.v2beta1.OutputAudioConfig output_audio_config = 5;
Name | Description |
value |
OutputAudioConfig |
Type | Description |
StreamingDetectIntentRequest.Builder |
setOutputAudioConfig(OutputAudioConfig.Builder builderForValue)
public StreamingDetectIntentRequest.Builder setOutputAudioConfig(OutputAudioConfig.Builder builderForValue)
Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
.google.cloud.dialogflow.v2beta1.OutputAudioConfig output_audio_config = 5;
Name | Description |
builderForValue |
OutputAudioConfig.Builder |
Type | Description |
StreamingDetectIntentRequest.Builder |
setOutputAudioConfigMask(FieldMask value)
public StreamingDetectIntentRequest.Builder setOutputAudioConfigMask(FieldMask value)
Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level. If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
.google.protobuf.FieldMask output_audio_config_mask = 7;
Name | Description |
value |
FieldMask |
Type | Description |
StreamingDetectIntentRequest.Builder |
setOutputAudioConfigMask(FieldMask.Builder builderForValue)
public StreamingDetectIntentRequest.Builder setOutputAudioConfigMask(FieldMask.Builder builderForValue)
Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level. If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
.google.protobuf.FieldMask output_audio_config_mask = 7;
Name | Description |
builderForValue |
Builder |
Type | Description |
StreamingDetectIntentRequest.Builder |
setQueryInput(QueryInput value)
public StreamingDetectIntentRequest.Builder setQueryInput(QueryInput value)
Required. The input specification. It can be set to:
- an audio config which instructs the speech recognizer how to process the speech audio,
- a conversational query in the form of text, or
- an event that specifies which intent to trigger.
.google.cloud.dialogflow.v2beta1.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
Name | Description |
value |
QueryInput |
Type | Description |
StreamingDetectIntentRequest.Builder |
setQueryInput(QueryInput.Builder builderForValue)
public StreamingDetectIntentRequest.Builder setQueryInput(QueryInput.Builder builderForValue)
Required. The input specification. It can be set to:
- an audio config which instructs the speech recognizer how to process the speech audio,
- a conversational query in the form of text, or
- an event that specifies which intent to trigger.
.google.cloud.dialogflow.v2beta1.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
Name | Description |
builderForValue |
QueryInput.Builder |
Type | Description |
StreamingDetectIntentRequest.Builder |
setQueryParams(QueryParameters value)
public StreamingDetectIntentRequest.Builder setQueryParams(QueryParameters value)
The parameters of this query.
.google.cloud.dialogflow.v2beta1.QueryParameters query_params = 2;
Name | Description |
value |
QueryParameters |
Type | Description |
StreamingDetectIntentRequest.Builder |
setQueryParams(QueryParameters.Builder builderForValue)
public StreamingDetectIntentRequest.Builder setQueryParams(QueryParameters.Builder builderForValue)
The parameters of this query.
.google.cloud.dialogflow.v2beta1.QueryParameters query_params = 2;
Name | Description |
builderForValue |
QueryParameters.Builder |
Type | Description |
StreamingDetectIntentRequest.Builder |
setRepeatedField(Descriptors.FieldDescriptor field, int index, Object value)
public StreamingDetectIntentRequest.Builder setRepeatedField(Descriptors.FieldDescriptor field, int index, Object value)
Name | Description |
field |
FieldDescriptor |
index |
int |
value |
Object |
Type | Description |
StreamingDetectIntentRequest.Builder |
setSession(String value)
public StreamingDetectIntentRequest.Builder setSession(String value)
Required. The name of the session the query is sent to. Supported formats:
projects/<Project ID>/agent/sessions/<Session ID>,
- projects/<Project ID>/locations/<Location ID>/agent/sessions/<Session
ID>
,
- projects/<Project ID>/agent/environments/<Environment ID>/users/<User
ID>/sessions/<Session ID>
,
- projects/<Project ID>/locations/<Location
ID>/agent/environments/<Environment ID>/users/<User ID>/sessions/<Session
ID>
, If
Location IDis not specified we assume default 'us' location. If
Environment IDis not specified, we assume default 'draft' environment. If
User IDis not specified, we are using "-". It's up to the API caller to choose an appropriate
Session IDand
User Id. They can be a random number or some type of user and session identifiers (preferably hashed). The length of the
Session IDand
User ID` must not exceed 36 characters. For more information, see the API interactions guide. Note: Always use agent versions for production traffic. See Versions and environments.
string session = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... }
Name | Description |
value |
String The session to set. |
Type | Description |
StreamingDetectIntentRequest.Builder |
This builder for chaining. |
setSessionBytes(ByteString value)
public StreamingDetectIntentRequest.Builder setSessionBytes(ByteString value)
Required. The name of the session the query is sent to. Supported formats:
projects/<Project ID>/agent/sessions/<Session ID>,
- projects/<Project ID>/locations/<Location ID>/agent/sessions/<Session
ID>
,
- projects/<Project ID>/agent/environments/<Environment ID>/users/<User
ID>/sessions/<Session ID>
,
- projects/<Project ID>/locations/<Location
ID>/agent/environments/<Environment ID>/users/<User ID>/sessions/<Session
ID>
, If
Location IDis not specified we assume default 'us' location. If
Environment IDis not specified, we assume default 'draft' environment. If
User IDis not specified, we are using "-". It's up to the API caller to choose an appropriate
Session IDand
User Id. They can be a random number or some type of user and session identifiers (preferably hashed). The length of the
Session IDand
User ID` must not exceed 36 characters. For more information, see the API interactions guide. Note: Always use agent versions for production traffic. See Versions and environments.
string session = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... }
Name | Description |
value |
ByteString The bytes for session to set. |
Type | Description |
StreamingDetectIntentRequest.Builder |
This builder for chaining. |
setSingleUtterance(boolean value) (deprecated)
public StreamingDetectIntentRequest.Builder setSingleUtterance(boolean value)
Deprecated. google.cloud.dialogflow.v2beta1.StreamingDetectIntentRequest.single_utterance is deprecated. See google/cloud/dialogflow/v2beta1/session.proto;l=564
DEPRECATED. Please use
InputAudioConfig.single_utterance
instead. If false
(default), recognition does not cease until the client
closes the stream. If true
, the recognizer will detect a single spoken
utterance in input audio. Recognition ceases when it detects the audio's
voice has stopped or paused. In this case, once a detected intent is
received, the client should close the stream and start a new request with a
new stream as needed. This setting is ignored when query_input
is a piece
of text or an event.
bool single_utterance = 4 [deprecated = true];
Name | Description |
value |
boolean The singleUtterance to set. |
Type | Description |
StreamingDetectIntentRequest.Builder |
This builder for chaining. |
setUnknownFields(UnknownFieldSet unknownFields)
public final StreamingDetectIntentRequest.Builder setUnknownFields(UnknownFieldSet unknownFields)
Name | Description |
unknownFields |
UnknownFieldSet |
Type | Description |
StreamingDetectIntentRequest.Builder |