Interface StreamingDetectIntentRequestOrBuilder (4.4.0)

public interface StreamingDetectIntentRequestOrBuilder extends MessageOrBuilder

Implements

MessageOrBuilder

Methods

getInputAudio()

public abstract ByteString getInputAudio()

The input audio content to be recognized. Must be sent if query_input was set to a streaming input audio config. The complete audio over all streaming messages must not exceed 1 minute.

bytes input_audio = 6;

Returns
TypeDescription
ByteString

The inputAudio.

getOutputAudioConfig()

public abstract OutputAudioConfig getOutputAudioConfig()

Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.

.google.cloud.dialogflow.v2.OutputAudioConfig output_audio_config = 5;

Returns
TypeDescription
OutputAudioConfig

The outputAudioConfig.

getOutputAudioConfigMask()

public abstract FieldMask getOutputAudioConfigMask()

Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level. If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.

.google.protobuf.FieldMask output_audio_config_mask = 7;

Returns
TypeDescription
FieldMask

The outputAudioConfigMask.

getOutputAudioConfigMaskOrBuilder()

public abstract FieldMaskOrBuilder getOutputAudioConfigMaskOrBuilder()

Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level. If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.

.google.protobuf.FieldMask output_audio_config_mask = 7;

Returns
TypeDescription
FieldMaskOrBuilder

getOutputAudioConfigOrBuilder()

public abstract OutputAudioConfigOrBuilder getOutputAudioConfigOrBuilder()

Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.

.google.cloud.dialogflow.v2.OutputAudioConfig output_audio_config = 5;

Returns
TypeDescription
OutputAudioConfigOrBuilder

getQueryInput()

public abstract QueryInput getQueryInput()

Required. The input specification. It can be set to:

  1. an audio config which instructs the speech recognizer how to process the speech audio,
  2. a conversational query in the form of text, or
  3. an event that specifies which intent to trigger.

.google.cloud.dialogflow.v2.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];

Returns
TypeDescription
QueryInput

The queryInput.

getQueryInputOrBuilder()

public abstract QueryInputOrBuilder getQueryInputOrBuilder()

Required. The input specification. It can be set to:

  1. an audio config which instructs the speech recognizer how to process the speech audio,
  2. a conversational query in the form of text, or
  3. an event that specifies which intent to trigger.

.google.cloud.dialogflow.v2.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];

Returns
TypeDescription
QueryInputOrBuilder

getQueryParams()

public abstract QueryParameters getQueryParams()

The parameters of this query.

.google.cloud.dialogflow.v2.QueryParameters query_params = 2;

Returns
TypeDescription
QueryParameters

The queryParams.

getQueryParamsOrBuilder()

public abstract QueryParametersOrBuilder getQueryParamsOrBuilder()

The parameters of this query.

.google.cloud.dialogflow.v2.QueryParameters query_params = 2;

Returns
TypeDescription
QueryParametersOrBuilder

getSession()

public abstract String getSession()

Required. The name of the session the query is sent to. Format of the session name: projects/<Project ID>/agent/sessions/<Session ID>, or projects/<Project ID>/agent/environments/<Environment ID>/users/<User ID>/sessions/<Session ID>. If Environment ID is not specified, we assume default 'draft' environment. If User ID is not specified, we are using "-". It's up to the API caller to choose an appropriate Session ID and User Id. They can be a random number or some type of user and session identifiers (preferably hashed). The length of the Session ID and User ID must not exceed 36 characters. For more information, see the API interactions guide. Note: Always use agent versions for production traffic. See Versions and environments.

string session = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... }

Returns
TypeDescription
String

The session.

getSessionBytes()

public abstract ByteString getSessionBytes()

Required. The name of the session the query is sent to. Format of the session name: projects/<Project ID>/agent/sessions/<Session ID>, or projects/<Project ID>/agent/environments/<Environment ID>/users/<User ID>/sessions/<Session ID>. If Environment ID is not specified, we assume default 'draft' environment. If User ID is not specified, we are using "-". It's up to the API caller to choose an appropriate Session ID and User Id. They can be a random number or some type of user and session identifiers (preferably hashed). The length of the Session ID and User ID must not exceed 36 characters. For more information, see the API interactions guide. Note: Always use agent versions for production traffic. See Versions and environments.

string session = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... }

Returns
TypeDescription
ByteString

The bytes for session.

getSingleUtterance()

public abstract boolean getSingleUtterance()

Please use InputAudioConfig.single_utterance instead. If false (default), recognition does not cease until the client closes the stream. If true, the recognizer will detect a single spoken utterance in input audio. Recognition ceases when it detects the audio's voice has stopped or paused. In this case, once a detected intent is received, the client should close the stream and start a new request with a new stream as needed. This setting is ignored when query_input is a piece of text or an event.

bool single_utterance = 4 [deprecated = true];

Returns
TypeDescription
boolean

The singleUtterance.

hasOutputAudioConfig()

public abstract boolean hasOutputAudioConfig()

Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.

.google.cloud.dialogflow.v2.OutputAudioConfig output_audio_config = 5;

Returns
TypeDescription
boolean

Whether the outputAudioConfig field is set.

hasOutputAudioConfigMask()

public abstract boolean hasOutputAudioConfigMask()

Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level. If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.

.google.protobuf.FieldMask output_audio_config_mask = 7;

Returns
TypeDescription
boolean

Whether the outputAudioConfigMask field is set.

hasQueryInput()

public abstract boolean hasQueryInput()

Required. The input specification. It can be set to:

  1. an audio config which instructs the speech recognizer how to process the speech audio,
  2. a conversational query in the form of text, or
  3. an event that specifies which intent to trigger.

.google.cloud.dialogflow.v2.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];

Returns
TypeDescription
boolean

Whether the queryInput field is set.

hasQueryParams()

public abstract boolean hasQueryParams()

The parameters of this query.

.google.cloud.dialogflow.v2.QueryParameters query_params = 2;

Returns
TypeDescription
boolean

Whether the queryParams field is set.