- 1.37.0 (latest)
- 1.36.0
- 1.35.0
- 1.34.1
- 1.33.0
- 1.32.1
- 1.31.1
- 1.30.1
- 1.29.0
- 1.28.0
- 1.27.0
- 1.26.0
- 1.25.0
- 1.24.0
- 1.23.1
- 1.22.0
- 1.21.0
- 1.20.1
- 1.19.0
- 1.18.0
- 1.17.1
- 1.16.0
- 1.15.0
- 1.14.1
- 1.13.5
- 1.12.1
- 1.11.0
- 1.10.0
- 1.9.1
- 1.8.0
- 1.7.0
- 1.6.0
- 1.5.0
- 1.4.0
- 1.3.2
- 1.2.0
- 1.1.1
- 1.0.0
- 0.8.2
- 0.7.1
- 0.6.0
- 0.5.0
- 0.4.1
- 0.3.0
- 0.2.0
- 0.1.1
StreamingRecognitionResult(mapping=None, *, ignore_unknown_fields=False, **kwargs)
Contains a speech recognition result corresponding to a portion of the audio that is currently being processed or an indication that this is the end of the single requested utterance.
Example:
transcript: "tube"
transcript: "to be a"
transcript: "to be"
transcript: "to be or not to be" is_final: true
transcript: " that's"
transcript: " that is"
message_type:
END_OF_SINGLE_UTTERANCE
transcript: " that is the question" is_final: true
Only two of the responses contain final results (#4 and #8 indicated
by is_final: true
). Concatenating these generates the full
transcript: "to be or not to be that is the question".
In each response we populate:
for
TRANSCRIPT
:transcript
and possiblyis_final
.for
END_OF_SINGLE_UTTERANCE
: onlymessage_type
.
Attributes:
message_type (google.cloud.dialogflowcx_v3.types.StreamingRecognitionResult.MessageType):
Type of the result message.
transcript (str):
Transcript text representing the words that the user spoke.
Populated if and only if message_type
= TRANSCRIPT
.
is_final (bool):
If false
, the StreamingRecognitionResult
represents
an interim result that may change. If true
, the
recognizer will not return any further hypotheses about this
piece of the audio. May only be populated for
message_type
= TRANSCRIPT
.
confidence (float):
The Speech confidence between 0.0 and 1.0 for the current
portion of audio. A higher number indicates an estimated
greater likelihood that the recognized words are correct.
The default of 0.0 is a sentinel value indicating that
confidence was not set.
This field is typically only provided if ``is_final`` is
true and you should not rely on it being accurate or even
set.
stability (float):
An estimate of the likelihood that the speech recognizer
will not change its guess about this interim recognition
result:
- If the value is unspecified or 0.0, Dialogflow didn't
compute the stability. In particular, Dialogflow will
only provide stability for ``TRANSCRIPT`` results with
``is_final = false``.
- Otherwise, the value is in (0.0, 1.0] where 0.0 means
completely unstable and 1.0 means completely stable.
speech_word_info (Sequence[google.cloud.dialogflowcx_v3.types.SpeechWordInfo]):
Word-specific information for the words recognized by Speech
in
[transcript][google.cloud.dialogflow.cx.v3.StreamingRecognitionResult.transcript].
Populated if and only if ``message_type`` = ``TRANSCRIPT``
and [InputAudioConfig.enable_word_info] is set.
speech_end_offset (google.protobuf.duration_pb2.Duration):
Time offset of the end of this Speech recognition result
relative to the beginning of the audio. Only populated for
``message_type`` = ``TRANSCRIPT``.