- 1.36.0 (latest)
- 1.35.0
- 1.34.1
- 1.33.0
- 1.32.1
- 1.31.1
- 1.30.1
- 1.29.0
- 1.28.0
- 1.27.0
- 1.26.0
- 1.25.0
- 1.24.0
- 1.23.1
- 1.22.0
- 1.21.0
- 1.20.1
- 1.19.0
- 1.18.0
- 1.17.1
- 1.16.0
- 1.15.0
- 1.14.1
- 1.13.5
- 1.12.1
- 1.11.0
- 1.10.0
- 1.9.1
- 1.8.0
- 1.7.0
- 1.6.0
- 1.5.0
- 1.4.0
- 1.3.2
- 1.2.0
- 1.1.1
- 1.0.0
- 0.8.2
- 0.7.1
- 0.6.0
- 0.5.0
- 0.4.1
- 0.3.0
- 0.2.0
- 0.1.1
StreamingRecognitionResult(mapping=None, *, ignore_unknown_fields=False, **kwargs)
Contains a speech recognition result corresponding to a portion of the audio that is currently being processed or an indication that this is the end of the single requested utterance.
Example:
transcript: "tube"
transcript: "to be a"
transcript: "to be"
transcript: "to be or not to be" is_final: true
transcript: " that's"
transcript: " that is"
message_type:
END_OF_SINGLE_UTTERANCE
transcript: " that is the question" is_final: true
Only two of the responses contain final results (#4 and #8 indicated
by is_final: true
). Concatenating these generates the full
transcript: "to be or not to be that is the question".
In each response we populate:
for
TRANSCRIPT
:transcript
and possiblyis_final
.for
END_OF_SINGLE_UTTERANCE
: onlymessage_type
.
Attributes | |
---|---|
Name | Description |
message_type |
google.cloud.dialogflowcx_v3.types.StreamingRecognitionResult.MessageType
Type of the result message. |
transcript |
str
Transcript text representing the words that the user spoke. Populated if and only if message_type = TRANSCRIPT .
|
is_final |
bool
If false , the StreamingRecognitionResult represents
an interim result that may change. If true , the
recognizer will not return any further hypotheses about this
piece of the audio. May only be populated for
message_type = TRANSCRIPT .
|
confidence |
float
The Speech confidence between 0.0 and 1.0 for the current portion of audio. A higher number indicates an estimated greater likelihood that the recognized words are correct. The default of 0.0 is a sentinel value indicating that confidence was not set. This field is typically only provided if is_final is
true and you should not rely on it being accurate or even
set.
|
stability |
float
An estimate of the likelihood that the speech recognizer will not change its guess about this interim recognition result: - If the value is unspecified or 0.0, Dialogflow didn't compute the stability. In particular, Dialogflow will only provide stability for TRANSCRIPT results with
is_final = false .
- Otherwise, the value is in (0.0, 1.0] where 0.0 means
completely unstable and 1.0 means completely stable.
|
speech_word_info |
Sequence[google.cloud.dialogflowcx_v3.types.SpeechWordInfo]
Word-specific information for the words recognized by Speech in transcript. Populated if and only if message_type = TRANSCRIPT
and [InputAudioConfig.enable_word_info] is set.
|
speech_end_offset |
google.protobuf.duration_pb2.Duration
Time offset of the end of this Speech recognition result relative to the beginning of the audio. Only populated for message_type = TRANSCRIPT .
|
Classes
MessageType
MessageType(value)
Type of the response message.