Cloud Speech V1p1beta1 Client - Class StreamingRecognitionConfig (1.15.0)

Reference documentation and code samples for the Cloud Speech V1p1beta1 Client class StreamingRecognitionConfig.

Provides information to the recognizer that specifies how to process the request.

Generated from protobuf message google.cloud.speech.v1p1beta1.StreamingRecognitionConfig

Namespace

Google \ Cloud \ Speech \ V1p1beta1

Methods

__construct

Constructor.

Parameters
NameDescription
data array

Optional. Data for populating the Message object.

↳ config Google\Cloud\Speech\V1p1beta1\RecognitionConfig

Required. Provides information to the recognizer that specifies how to process the request.

↳ single_utterance bool

If false or omitted, the recognizer will perform continuous recognition (continuing to wait for and process audio even if the user pauses speaking) until the client closes the input stream (gRPC API) or until the maximum time limit has been reached. May return multiple StreamingRecognitionResults with the is_final flag set to true. If true, the recognizer will detect a single spoken utterance. When it detects that the user has paused or stopped speaking, it will return an END_OF_SINGLE_UTTERANCE event and cease recognition. It will return no more than one StreamingRecognitionResult with the is_final flag set to true. The single_utterance field can only be used with specified models, otherwise an error is thrown. The model field in RecognitionConfig must be set to: * command_and_search * phone_call AND additional field useEnhanced=true * The model field is left undefined. In this case the API auto-selects a model based on any other parameters that you set in RecognitionConfig.

↳ interim_results bool

If true, interim results (tentative hypotheses) may be returned as they become available (these interim results are indicated with the is_final=false flag). If false or omitted, only is_final=true result(s) are returned.

↳ enable_voice_activity_events bool

If true, responses with voice activity speech events will be returned as they are detected.

↳ voice_activity_timeout Google\Cloud\Speech\V1p1beta1\StreamingRecognitionConfig\VoiceActivityTimeout

If set, the server will automatically close the stream after the specified duration has elapsed after the last VOICE_ACTIVITY speech event has been sent. The field voice_activity_events must also be set to true.

getConfig

Required. Provides information to the recognizer that specifies how to process the request.

Returns
TypeDescription
Google\Cloud\Speech\V1p1beta1\RecognitionConfig|null

hasConfig

clearConfig

setConfig

Required. Provides information to the recognizer that specifies how to process the request.

Parameter
NameDescription
var Google\Cloud\Speech\V1p1beta1\RecognitionConfig
Returns
TypeDescription
$this

getSingleUtterance

If false or omitted, the recognizer will perform continuous recognition (continuing to wait for and process audio even if the user pauses speaking) until the client closes the input stream (gRPC API) or until the maximum time limit has been reached. May return multiple StreamingRecognitionResults with the is_final flag set to true.

If true, the recognizer will detect a single spoken utterance. When it detects that the user has paused or stopped speaking, it will return an END_OF_SINGLE_UTTERANCE event and cease recognition. It will return no more than one StreamingRecognitionResult with the is_final flag set to true. The single_utterance field can only be used with specified models, otherwise an error is thrown. The model field in RecognitionConfig must be set to:

  • command_and_search
  • phone_call AND additional field useEnhanced=true
  • The model field is left undefined. In this case the API auto-selects a model based on any other parameters that you set in RecognitionConfig.
Returns
TypeDescription
bool

setSingleUtterance

If false or omitted, the recognizer will perform continuous recognition (continuing to wait for and process audio even if the user pauses speaking) until the client closes the input stream (gRPC API) or until the maximum time limit has been reached. May return multiple StreamingRecognitionResults with the is_final flag set to true.

If true, the recognizer will detect a single spoken utterance. When it detects that the user has paused or stopped speaking, it will return an END_OF_SINGLE_UTTERANCE event and cease recognition. It will return no more than one StreamingRecognitionResult with the is_final flag set to true. The single_utterance field can only be used with specified models, otherwise an error is thrown. The model field in RecognitionConfig must be set to:

  • command_and_search
  • phone_call AND additional field useEnhanced=true
  • The model field is left undefined. In this case the API auto-selects a model based on any other parameters that you set in RecognitionConfig.
Parameter
NameDescription
var bool
Returns
TypeDescription
$this

getInterimResults

If true, interim results (tentative hypotheses) may be returned as they become available (these interim results are indicated with the is_final=false flag).

If false or omitted, only is_final=true result(s) are returned.

Returns
TypeDescription
bool

setInterimResults

If true, interim results (tentative hypotheses) may be returned as they become available (these interim results are indicated with the is_final=false flag).

If false or omitted, only is_final=true result(s) are returned.

Parameter
NameDescription
var bool
Returns
TypeDescription
$this

getEnableVoiceActivityEvents

If true, responses with voice activity speech events will be returned as they are detected.

Returns
TypeDescription
bool

setEnableVoiceActivityEvents

If true, responses with voice activity speech events will be returned as they are detected.

Parameter
NameDescription
var bool
Returns
TypeDescription
$this

getVoiceActivityTimeout

If set, the server will automatically close the stream after the specified duration has elapsed after the last VOICE_ACTIVITY speech event has been sent. The field voice_activity_events must also be set to true.

Returns
TypeDescription
Google\Cloud\Speech\V1p1beta1\StreamingRecognitionConfig\VoiceActivityTimeout|null

hasVoiceActivityTimeout

clearVoiceActivityTimeout

setVoiceActivityTimeout

If set, the server will automatically close the stream after the specified duration has elapsed after the last VOICE_ACTIVITY speech event has been sent. The field voice_activity_events must also be set to true.

Parameter
NameDescription
var Google\Cloud\Speech\V1p1beta1\StreamingRecognitionConfig\VoiceActivityTimeout
Returns
TypeDescription
$this