StreamingRecognitionConfig(mapping=None, *, ignore_unknown_fields=False, **kwargs)
Provides information to the recognizer that specifies how to process the request.
Attributes
Name | Description |
config |
google.cloud.speech_v1.types.RecognitionConfig
Required. Provides information to the recognizer that specifies how to process the request. |
single_utterance |
bool
If false or omitted, the recognizer will perform
continuous recognition (continuing to wait for and process
audio even if the user pauses speaking) until the client
closes the input stream (gRPC API) or until the maximum time
limit has been reached. May return multiple
StreamingRecognitionResult \ s with the is_final flag
set to true .
If true , the recognizer will detect a single spoken
utterance. When it detects that the user has paused or
stopped speaking, it will return an
END_OF_SINGLE_UTTERANCE event and cease recognition. It
will return no more than one StreamingRecognitionResult
with the is_final flag set to true .
The single_utterance field can only be used with
specified models, otherwise an error is thrown. The
model field in [RecognitionConfig ][] must be set to:
- command_and_search
- phone_call AND additional field
useEnhanced \ =\ true
- The model field is left undefined. In this case the
API auto-selects a model based on any other parameters
that you set in RecognitionConfig .
|
interim_results |
bool
If true , interim results (tentative hypotheses) may be
returned as they become available (these interim results are
indicated with the is_final=false flag). If false or
omitted, only is_final=true result(s) are returned.
|