- 0.64.0 (latest)
- 0.63.0
- 0.62.0
- 0.60.0
- 0.59.0
- 0.58.0
- 0.57.0
- 0.56.0
- 0.55.0
- 0.54.0
- 0.53.0
- 0.52.0
- 0.51.0
- 0.50.0
- 0.48.0
- 0.47.0
- 0.46.0
- 0.45.0
- 0.44.0
- 0.43.0
- 0.42.0
- 0.41.0
- 0.40.0
- 0.39.0
- 0.38.0
- 0.35.0
- 0.34.0
- 0.33.0
- 0.32.0
- 0.31.0
- 0.30.0
- 0.29.0
- 0.28.0
- 0.27.0
- 0.26.0
- 0.25.0
- 0.24.0
- 0.23.0
- 0.22.0
- 0.20.0
- 0.19.0
- 0.18.0
- 0.17.0
- 0.16.0
- 0.15.0
- 0.14.7
- 0.13.1
- 0.12.1
- 0.11.5
public static final class InputAudioConfig.Builder extends GeneratedMessageV3.Builder<InputAudioConfig.Builder> implements InputAudioConfigOrBuilder
Instructs the speech recognizer on how to process the audio content.
Protobuf type google.cloud.dialogflow.cx.v3beta1.InputAudioConfig
Inheritance
Object > AbstractMessageLite.Builder<MessageType,BuilderType> > AbstractMessage.Builder<BuilderType> > GeneratedMessageV3.Builder > InputAudioConfig.BuilderImplements
InputAudioConfigOrBuilderStatic Methods
getDescriptor()
public static final Descriptors.Descriptor getDescriptor()
Type | Description |
Descriptor |
Methods
addAllPhraseHints(Iterable<String> values)
public InputAudioConfig.Builder addAllPhraseHints(Iterable<String> values)
Optional. A list of strings containing words and phrases that the speech recognizer should recognize with higher likelihood. See the Cloud Speech documentation for more details.
repeated string phrase_hints = 4;
Name | Description |
values |
Iterable<String> The phraseHints to add. |
Type | Description |
InputAudioConfig.Builder |
This builder for chaining. |
addPhraseHints(String value)
public InputAudioConfig.Builder addPhraseHints(String value)
Optional. A list of strings containing words and phrases that the speech recognizer should recognize with higher likelihood. See the Cloud Speech documentation for more details.
repeated string phrase_hints = 4;
Name | Description |
value |
String The phraseHints to add. |
Type | Description |
InputAudioConfig.Builder |
This builder for chaining. |
addPhraseHintsBytes(ByteString value)
public InputAudioConfig.Builder addPhraseHintsBytes(ByteString value)
Optional. A list of strings containing words and phrases that the speech recognizer should recognize with higher likelihood. See the Cloud Speech documentation for more details.
repeated string phrase_hints = 4;
Name | Description |
value |
ByteString The bytes of the phraseHints to add. |
Type | Description |
InputAudioConfig.Builder |
This builder for chaining. |
addRepeatedField(Descriptors.FieldDescriptor field, Object value)
public InputAudioConfig.Builder addRepeatedField(Descriptors.FieldDescriptor field, Object value)
Name | Description |
field |
FieldDescriptor |
value |
Object |
Type | Description |
InputAudioConfig.Builder |
build()
public InputAudioConfig build()
Type | Description |
InputAudioConfig |
buildPartial()
public InputAudioConfig buildPartial()
Type | Description |
InputAudioConfig |
clear()
public InputAudioConfig.Builder clear()
Type | Description |
InputAudioConfig.Builder |
clearAudioEncoding()
public InputAudioConfig.Builder clearAudioEncoding()
Required. Audio encoding of the audio content to process.
.google.cloud.dialogflow.cx.v3beta1.AudioEncoding audio_encoding = 1 [(.google.api.field_behavior) = REQUIRED];
Type | Description |
InputAudioConfig.Builder |
This builder for chaining. |
clearEnableWordInfo()
public InputAudioConfig.Builder clearEnableWordInfo()
Optional. If true
, Dialogflow returns
SpeechWordInfo in
StreamingRecognitionResult
with information about the recognized speech words, e.g. start and end time
offsets. If false or unspecified, Speech doesn't return any word-level
information.
bool enable_word_info = 13;
Type | Description |
InputAudioConfig.Builder |
This builder for chaining. |
clearField(Descriptors.FieldDescriptor field)
public InputAudioConfig.Builder clearField(Descriptors.FieldDescriptor field)
Name | Description |
field |
FieldDescriptor |
Type | Description |
InputAudioConfig.Builder |
clearModel()
public InputAudioConfig.Builder clearModel()
Optional. Which Speech model to select for the given request. Select the model best suited to your domain to get best results. If a model is not explicitly specified, then we auto-select a model based on the parameters in the InputAudioConfig. If enhanced speech model is enabled for the agent and an enhanced version of the specified model for the language does not exist, then the speech is recognized using the standard version of the specified model. Refer to Cloud Speech API documentation for more details.
string model = 7;
Type | Description |
InputAudioConfig.Builder |
This builder for chaining. |
clearModelVariant()
public InputAudioConfig.Builder clearModelVariant()
Optional. Which variant of the Speech model to use.
.google.cloud.dialogflow.cx.v3beta1.SpeechModelVariant model_variant = 10;
Type | Description |
InputAudioConfig.Builder |
This builder for chaining. |
clearOneof(Descriptors.OneofDescriptor oneof)
public InputAudioConfig.Builder clearOneof(Descriptors.OneofDescriptor oneof)
Name | Description |
oneof |
OneofDescriptor |
Type | Description |
InputAudioConfig.Builder |
clearPhraseHints()
public InputAudioConfig.Builder clearPhraseHints()
Optional. A list of strings containing words and phrases that the speech recognizer should recognize with higher likelihood. See the Cloud Speech documentation for more details.
repeated string phrase_hints = 4;
Type | Description |
InputAudioConfig.Builder |
This builder for chaining. |
clearSampleRateHertz()
public InputAudioConfig.Builder clearSampleRateHertz()
Sample rate (in Hertz) of the audio content sent in the query. Refer to Cloud Speech API documentation for more details.
int32 sample_rate_hertz = 2;
Type | Description |
InputAudioConfig.Builder |
This builder for chaining. |
clearSingleUtterance()
public InputAudioConfig.Builder clearSingleUtterance()
Optional. If false
(default), recognition does not cease until the
client closes the stream.
If true
, the recognizer will detect a single spoken utterance in input
audio. Recognition ceases when it detects the audio's voice has
stopped or paused. In this case, once a detected intent is received, the
client should close the stream and start a new request with a new stream as
needed.
Note: This setting is relevant only for streaming methods.
bool single_utterance = 8;
Type | Description |
InputAudioConfig.Builder |
This builder for chaining. |
clone()
public InputAudioConfig.Builder clone()
Type | Description |
InputAudioConfig.Builder |
getAudioEncoding()
public AudioEncoding getAudioEncoding()
Required. Audio encoding of the audio content to process.
.google.cloud.dialogflow.cx.v3beta1.AudioEncoding audio_encoding = 1 [(.google.api.field_behavior) = REQUIRED];
Type | Description |
AudioEncoding |
The audioEncoding. |
getAudioEncodingValue()
public int getAudioEncodingValue()
Required. Audio encoding of the audio content to process.
.google.cloud.dialogflow.cx.v3beta1.AudioEncoding audio_encoding = 1 [(.google.api.field_behavior) = REQUIRED];
Type | Description |
int |
The enum numeric value on the wire for audioEncoding. |
getDefaultInstanceForType()
public InputAudioConfig getDefaultInstanceForType()
Type | Description |
InputAudioConfig |
getDescriptorForType()
public Descriptors.Descriptor getDescriptorForType()
Type | Description |
Descriptor |
getEnableWordInfo()
public boolean getEnableWordInfo()
Optional. If true
, Dialogflow returns
SpeechWordInfo in
StreamingRecognitionResult
with information about the recognized speech words, e.g. start and end time
offsets. If false or unspecified, Speech doesn't return any word-level
information.
bool enable_word_info = 13;
Type | Description |
boolean |
The enableWordInfo. |
getModel()
public String getModel()
Optional. Which Speech model to select for the given request. Select the model best suited to your domain to get best results. If a model is not explicitly specified, then we auto-select a model based on the parameters in the InputAudioConfig. If enhanced speech model is enabled for the agent and an enhanced version of the specified model for the language does not exist, then the speech is recognized using the standard version of the specified model. Refer to Cloud Speech API documentation for more details.
string model = 7;
Type | Description |
String |
The model. |
getModelBytes()
public ByteString getModelBytes()
Optional. Which Speech model to select for the given request. Select the model best suited to your domain to get best results. If a model is not explicitly specified, then we auto-select a model based on the parameters in the InputAudioConfig. If enhanced speech model is enabled for the agent and an enhanced version of the specified model for the language does not exist, then the speech is recognized using the standard version of the specified model. Refer to Cloud Speech API documentation for more details.
string model = 7;
Type | Description |
ByteString |
The bytes for model. |
getModelVariant()
public SpeechModelVariant getModelVariant()
Optional. Which variant of the Speech model to use.
.google.cloud.dialogflow.cx.v3beta1.SpeechModelVariant model_variant = 10;
Type | Description |
SpeechModelVariant |
The modelVariant. |
getModelVariantValue()
public int getModelVariantValue()
Optional. Which variant of the Speech model to use.
.google.cloud.dialogflow.cx.v3beta1.SpeechModelVariant model_variant = 10;
Type | Description |
int |
The enum numeric value on the wire for modelVariant. |
getPhraseHints(int index)
public String getPhraseHints(int index)
Optional. A list of strings containing words and phrases that the speech recognizer should recognize with higher likelihood. See the Cloud Speech documentation for more details.
repeated string phrase_hints = 4;
Name | Description |
index |
int The index of the element to return. |
Type | Description |
String |
The phraseHints at the given index. |
getPhraseHintsBytes(int index)
public ByteString getPhraseHintsBytes(int index)
Optional. A list of strings containing words and phrases that the speech recognizer should recognize with higher likelihood. See the Cloud Speech documentation for more details.
repeated string phrase_hints = 4;
Name | Description |
index |
int The index of the value to return. |
Type | Description |
ByteString |
The bytes of the phraseHints at the given index. |
getPhraseHintsCount()
public int getPhraseHintsCount()
Optional. A list of strings containing words and phrases that the speech recognizer should recognize with higher likelihood. See the Cloud Speech documentation for more details.
repeated string phrase_hints = 4;
Type | Description |
int |
The count of phraseHints. |
getPhraseHintsList()
public ProtocolStringList getPhraseHintsList()
Optional. A list of strings containing words and phrases that the speech recognizer should recognize with higher likelihood. See the Cloud Speech documentation for more details.
repeated string phrase_hints = 4;
Type | Description |
ProtocolStringList |
A list containing the phraseHints. |
getSampleRateHertz()
public int getSampleRateHertz()
Sample rate (in Hertz) of the audio content sent in the query. Refer to Cloud Speech API documentation for more details.
int32 sample_rate_hertz = 2;
Type | Description |
int |
The sampleRateHertz. |
getSingleUtterance()
public boolean getSingleUtterance()
Optional. If false
(default), recognition does not cease until the
client closes the stream.
If true
, the recognizer will detect a single spoken utterance in input
audio. Recognition ceases when it detects the audio's voice has
stopped or paused. In this case, once a detected intent is received, the
client should close the stream and start a new request with a new stream as
needed.
Note: This setting is relevant only for streaming methods.
bool single_utterance = 8;
Type | Description |
boolean |
The singleUtterance. |
internalGetFieldAccessorTable()
protected GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable()
Type | Description |
FieldAccessorTable |
isInitialized()
public final boolean isInitialized()
Type | Description |
boolean |
mergeFrom(InputAudioConfig other)
public InputAudioConfig.Builder mergeFrom(InputAudioConfig other)
Name | Description |
other |
InputAudioConfig |
Type | Description |
InputAudioConfig.Builder |
mergeFrom(CodedInputStream input, ExtensionRegistryLite extensionRegistry)
public InputAudioConfig.Builder mergeFrom(CodedInputStream input, ExtensionRegistryLite extensionRegistry)
Name | Description |
input |
CodedInputStream |
extensionRegistry |
ExtensionRegistryLite |
Type | Description |
InputAudioConfig.Builder |
Type | Description |
IOException |
mergeFrom(Message other)
public InputAudioConfig.Builder mergeFrom(Message other)
Name | Description |
other |
Message |
Type | Description |
InputAudioConfig.Builder |
mergeUnknownFields(UnknownFieldSet unknownFields)
public final InputAudioConfig.Builder mergeUnknownFields(UnknownFieldSet unknownFields)
Name | Description |
unknownFields |
UnknownFieldSet |
Type | Description |
InputAudioConfig.Builder |
setAudioEncoding(AudioEncoding value)
public InputAudioConfig.Builder setAudioEncoding(AudioEncoding value)
Required. Audio encoding of the audio content to process.
.google.cloud.dialogflow.cx.v3beta1.AudioEncoding audio_encoding = 1 [(.google.api.field_behavior) = REQUIRED];
Name | Description |
value |
AudioEncoding The audioEncoding to set. |
Type | Description |
InputAudioConfig.Builder |
This builder for chaining. |
setAudioEncodingValue(int value)
public InputAudioConfig.Builder setAudioEncodingValue(int value)
Required. Audio encoding of the audio content to process.
.google.cloud.dialogflow.cx.v3beta1.AudioEncoding audio_encoding = 1 [(.google.api.field_behavior) = REQUIRED];
Name | Description |
value |
int The enum numeric value on the wire for audioEncoding to set. |
Type | Description |
InputAudioConfig.Builder |
This builder for chaining. |
setEnableWordInfo(boolean value)
public InputAudioConfig.Builder setEnableWordInfo(boolean value)
Optional. If true
, Dialogflow returns
SpeechWordInfo in
StreamingRecognitionResult
with information about the recognized speech words, e.g. start and end time
offsets. If false or unspecified, Speech doesn't return any word-level
information.
bool enable_word_info = 13;
Name | Description |
value |
boolean The enableWordInfo to set. |
Type | Description |
InputAudioConfig.Builder |
This builder for chaining. |
setField(Descriptors.FieldDescriptor field, Object value)
public InputAudioConfig.Builder setField(Descriptors.FieldDescriptor field, Object value)
Name | Description |
field |
FieldDescriptor |
value |
Object |
Type | Description |
InputAudioConfig.Builder |
setModel(String value)
public InputAudioConfig.Builder setModel(String value)
Optional. Which Speech model to select for the given request. Select the model best suited to your domain to get best results. If a model is not explicitly specified, then we auto-select a model based on the parameters in the InputAudioConfig. If enhanced speech model is enabled for the agent and an enhanced version of the specified model for the language does not exist, then the speech is recognized using the standard version of the specified model. Refer to Cloud Speech API documentation for more details.
string model = 7;
Name | Description |
value |
String The model to set. |
Type | Description |
InputAudioConfig.Builder |
This builder for chaining. |
setModelBytes(ByteString value)
public InputAudioConfig.Builder setModelBytes(ByteString value)
Optional. Which Speech model to select for the given request. Select the model best suited to your domain to get best results. If a model is not explicitly specified, then we auto-select a model based on the parameters in the InputAudioConfig. If enhanced speech model is enabled for the agent and an enhanced version of the specified model for the language does not exist, then the speech is recognized using the standard version of the specified model. Refer to Cloud Speech API documentation for more details.
string model = 7;
Name | Description |
value |
ByteString The bytes for model to set. |
Type | Description |
InputAudioConfig.Builder |
This builder for chaining. |
setModelVariant(SpeechModelVariant value)
public InputAudioConfig.Builder setModelVariant(SpeechModelVariant value)
Optional. Which variant of the Speech model to use.
.google.cloud.dialogflow.cx.v3beta1.SpeechModelVariant model_variant = 10;
Name | Description |
value |
SpeechModelVariant The modelVariant to set. |
Type | Description |
InputAudioConfig.Builder |
This builder for chaining. |
setModelVariantValue(int value)
public InputAudioConfig.Builder setModelVariantValue(int value)
Optional. Which variant of the Speech model to use.
.google.cloud.dialogflow.cx.v3beta1.SpeechModelVariant model_variant = 10;
Name | Description |
value |
int The enum numeric value on the wire for modelVariant to set. |
Type | Description |
InputAudioConfig.Builder |
This builder for chaining. |
setPhraseHints(int index, String value)
public InputAudioConfig.Builder setPhraseHints(int index, String value)
Optional. A list of strings containing words and phrases that the speech recognizer should recognize with higher likelihood. See the Cloud Speech documentation for more details.
repeated string phrase_hints = 4;
Name | Description |
index |
int The index to set the value at. |
value |
String The phraseHints to set. |
Type | Description |
InputAudioConfig.Builder |
This builder for chaining. |
setRepeatedField(Descriptors.FieldDescriptor field, int index, Object value)
public InputAudioConfig.Builder setRepeatedField(Descriptors.FieldDescriptor field, int index, Object value)
Name | Description |
field |
FieldDescriptor |
index |
int |
value |
Object |
Type | Description |
InputAudioConfig.Builder |
setSampleRateHertz(int value)
public InputAudioConfig.Builder setSampleRateHertz(int value)
Sample rate (in Hertz) of the audio content sent in the query. Refer to Cloud Speech API documentation for more details.
int32 sample_rate_hertz = 2;
Name | Description |
value |
int The sampleRateHertz to set. |
Type | Description |
InputAudioConfig.Builder |
This builder for chaining. |
setSingleUtterance(boolean value)
public InputAudioConfig.Builder setSingleUtterance(boolean value)
Optional. If false
(default), recognition does not cease until the
client closes the stream.
If true
, the recognizer will detect a single spoken utterance in input
audio. Recognition ceases when it detects the audio's voice has
stopped or paused. In this case, once a detected intent is received, the
client should close the stream and start a new request with a new stream as
needed.
Note: This setting is relevant only for streaming methods.
bool single_utterance = 8;
Name | Description |
value |
boolean The singleUtterance to set. |
Type | Description |
InputAudioConfig.Builder |
This builder for chaining. |
setUnknownFields(UnknownFieldSet unknownFields)
public final InputAudioConfig.Builder setUnknownFields(UnknownFieldSet unknownFields)
Name | Description |
unknownFields |
UnknownFieldSet |
Type | Description |
InputAudioConfig.Builder |