Cloud Speech-to-Text V1p1beta1 API - Class Google::Cloud::Speech::V1p1beta1::StreamingRecognitionConfig (v0.13.0)

Reference documentation and code samples for the Cloud Speech-to-Text V1p1beta1 API class Google::Cloud::Speech::V1p1beta1::StreamingRecognitionConfig.

Provides information to the recognizer that specifies how to process the request.

Inherits

  • Object

Extended By

  • Google::Protobuf::MessageExts::ClassMethods

Includes

  • Google::Protobuf::MessageExts

Methods

#config

def config() -> ::Google::Cloud::Speech::V1p1beta1::RecognitionConfig
Returns

#config=

def config=(value) -> ::Google::Cloud::Speech::V1p1beta1::RecognitionConfig
Parameter
Returns

#interim_results

def interim_results() -> ::Boolean
Returns
  • (::Boolean) — If true, interim results (tentative hypotheses) may be returned as they become available (these interim results are indicated with the is_final=false flag). If false or omitted, only is_final=true result(s) are returned.

#interim_results=

def interim_results=(value) -> ::Boolean
Parameter
  • value (::Boolean) — If true, interim results (tentative hypotheses) may be returned as they become available (these interim results are indicated with the is_final=false flag). If false or omitted, only is_final=true result(s) are returned.
Returns
  • (::Boolean) — If true, interim results (tentative hypotheses) may be returned as they become available (these interim results are indicated with the is_final=false flag). If false or omitted, only is_final=true result(s) are returned.

#single_utterance

def single_utterance() -> ::Boolean
Returns
  • (::Boolean) —

    If false or omitted, the recognizer will perform continuous recognition (continuing to wait for and process audio even if the user pauses speaking) until the client closes the input stream (gRPC API) or until the maximum time limit has been reached. May return multiple StreamingRecognitionResults with the is_final flag set to true.

    If true, the recognizer will detect a single spoken utterance. When it detects that the user has paused or stopped speaking, it will return an END_OF_SINGLE_UTTERANCE event and cease recognition. It will return no more than one StreamingRecognitionResult with the is_final flag set to true.

    The single_utterance field can only be used with specified models, otherwise an error is thrown. The model field in [RecognitionConfig][] must be set to:

    • command_and_search
    • phone_call AND additional field useEnhanced=true
    • The model field is left undefined. In this case the API auto-selects a model based on any other parameters that you set in RecognitionConfig.

#single_utterance=

def single_utterance=(value) -> ::Boolean
Parameter
  • value (::Boolean) —

    If false or omitted, the recognizer will perform continuous recognition (continuing to wait for and process audio even if the user pauses speaking) until the client closes the input stream (gRPC API) or until the maximum time limit has been reached. May return multiple StreamingRecognitionResults with the is_final flag set to true.

    If true, the recognizer will detect a single spoken utterance. When it detects that the user has paused or stopped speaking, it will return an END_OF_SINGLE_UTTERANCE event and cease recognition. It will return no more than one StreamingRecognitionResult with the is_final flag set to true.

    The single_utterance field can only be used with specified models, otherwise an error is thrown. The model field in [RecognitionConfig][] must be set to:

    • command_and_search
    • phone_call AND additional field useEnhanced=true
    • The model field is left undefined. In this case the API auto-selects a model based on any other parameters that you set in RecognitionConfig.
Returns
  • (::Boolean) —

    If false or omitted, the recognizer will perform continuous recognition (continuing to wait for and process audio even if the user pauses speaking) until the client closes the input stream (gRPC API) or until the maximum time limit has been reached. May return multiple StreamingRecognitionResults with the is_final flag set to true.

    If true, the recognizer will detect a single spoken utterance. When it detects that the user has paused or stopped speaking, it will return an END_OF_SINGLE_UTTERANCE event and cease recognition. It will return no more than one StreamingRecognitionResult with the is_final flag set to true.

    The single_utterance field can only be used with specified models, otherwise an error is thrown. The model field in [RecognitionConfig][] must be set to:

    • command_and_search
    • phone_call AND additional field useEnhanced=true
    • The model field is left undefined. In this case the API auto-selects a model based on any other parameters that you set in RecognitionConfig.