|
| 1 | +# Copyright 2019 Google LLC |
| 2 | +# |
| 3 | +# Licensed under the Apache License, Version 2.0 (the "License"); |
| 4 | +# you may not use this file except in compliance with the License. |
| 5 | +# You may obtain a copy of the License at |
| 6 | +# |
| 7 | +# http://www.apache.org/licenses/LICENSE-2.0 |
| 8 | +# |
| 9 | +# Unless required by applicable law or agreed to in writing, software |
| 10 | +# distributed under the License is distributed on an "AS IS" BASIS, |
| 11 | +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| 12 | +# See the License for the specific language governing permissions and |
| 13 | +# limitations under the License. |
| 14 | + |
| 15 | +# NOTE: This file is auto generated by the elixir code generator program. |
| 16 | +# Do not edit this file manually. |
| 17 | + |
| 18 | +defmodule GoogleApi.Dialogflow.V2.Model.GoogleCloudDialogflowV2StreamingRecognitionResult do |
| 19 | + @moduledoc """ |
| 20 | + Contains a speech recognition result corresponding to a portion of the audio that is currently being processed or an indication that this is the end of the single requested utterance. While end-user audio is being processed, Dialogflow sends a series of results. Each result may contain a `transcript` value. A transcript represents a portion of the utterance. While the recognizer is processing audio, transcript values may be interim values or finalized values. Once a transcript is finalized, the `is_final` value is set to true and processing continues for the next transcript. If `StreamingDetectIntentRequest.query_input.audio_config.single_utterance` was true, and the recognizer has completed processing audio, the `message_type` value is set to `END_OF_SINGLE_UTTERANCE and the following (last) result contains the last finalized transcript. The complete end-user utterance is determined by concatenating the finalized transcript values received for the series of results. In the following example, single utterance is enabled. In the case where single utterance is not enabled, result 7 would not occur. ``` Num | transcript | message_type | is_final --- | ----------------------- | ----------------------- | -------- 1 | "tube" | TRANSCRIPT | false 2 | "to be a" | TRANSCRIPT | false 3 | "to be" | TRANSCRIPT | false 4 | "to be or not to be" | TRANSCRIPT | true 5 | "that's" | TRANSCRIPT | false 6 | "that is | TRANSCRIPT | false 7 | unset | END_OF_SINGLE_UTTERANCE | unset 8 | " that is the question" | TRANSCRIPT | true ``` Concatenating the finalized transcripts with `is_final` set to true, the complete utterance becomes "to be or not to be that is the question". |
| 21 | +
|
| 22 | + ## Attributes |
| 23 | +
|
| 24 | + * `confidence` (*type:* `number()`, *default:* `nil`) - The Speech confidence between 0.0 and 1.0 for the current portion of audio. A higher number indicates an estimated greater likelihood that the recognized words are correct. The default of 0.0 is a sentinel value indicating that confidence was not set. This field is typically only provided if `is_final` is true and you should not rely on it being accurate or even set. |
| 25 | + * `isFinal` (*type:* `boolean()`, *default:* `nil`) - If `false`, the `StreamingRecognitionResult` represents an interim result that may change. If `true`, the recognizer will not return any further hypotheses about this piece of the audio. May only be populated for `message_type` = `TRANSCRIPT`. |
| 26 | + * `languageCode` (*type:* `String.t`, *default:* `nil`) - Detected language code for the transcript. |
| 27 | + * `messageType` (*type:* `String.t`, *default:* `nil`) - Type of the result message. |
| 28 | + * `speechEndOffset` (*type:* `String.t`, *default:* `nil`) - Time offset of the end of this Speech recognition result relative to the beginning of the audio. Only populated for `message_type` = `TRANSCRIPT`. |
| 29 | + * `speechWordInfo` (*type:* `list(GoogleApi.Dialogflow.V2.Model.GoogleCloudDialogflowV2SpeechWordInfo.t)`, *default:* `nil`) - Word-specific information for the words recognized by Speech in transcript. Populated if and only if `message_type` = `TRANSCRIPT` and [InputAudioConfig.enable_word_info] is set. |
| 30 | + * `transcript` (*type:* `String.t`, *default:* `nil`) - Transcript text representing the words that the user spoke. Populated if and only if `message_type` = `TRANSCRIPT`. |
| 31 | + """ |
| 32 | + |
| 33 | + use GoogleApi.Gax.ModelBase |
| 34 | + |
| 35 | + @type t :: %__MODULE__{ |
| 36 | + :confidence => number() | nil, |
| 37 | + :isFinal => boolean() | nil, |
| 38 | + :languageCode => String.t() | nil, |
| 39 | + :messageType => String.t() | nil, |
| 40 | + :speechEndOffset => String.t() | nil, |
| 41 | + :speechWordInfo => |
| 42 | + list(GoogleApi.Dialogflow.V2.Model.GoogleCloudDialogflowV2SpeechWordInfo.t()) | nil, |
| 43 | + :transcript => String.t() | nil |
| 44 | + } |
| 45 | + |
| 46 | + field(:confidence) |
| 47 | + field(:isFinal) |
| 48 | + field(:languageCode) |
| 49 | + field(:messageType) |
| 50 | + field(:speechEndOffset) |
| 51 | + |
| 52 | + field(:speechWordInfo, |
| 53 | + as: GoogleApi.Dialogflow.V2.Model.GoogleCloudDialogflowV2SpeechWordInfo, |
| 54 | + type: :list |
| 55 | + ) |
| 56 | + |
| 57 | + field(:transcript) |
| 58 | +end |
| 59 | + |
| 60 | +defimpl Poison.Decoder, |
| 61 | + for: GoogleApi.Dialogflow.V2.Model.GoogleCloudDialogflowV2StreamingRecognitionResult do |
| 62 | + def decode(value, options) do |
| 63 | + GoogleApi.Dialogflow.V2.Model.GoogleCloudDialogflowV2StreamingRecognitionResult.decode( |
| 64 | + value, |
| 65 | + options |
| 66 | + ) |
| 67 | + end |
| 68 | +end |
| 69 | + |
| 70 | +defimpl Poison.Encoder, |
| 71 | + for: GoogleApi.Dialogflow.V2.Model.GoogleCloudDialogflowV2StreamingRecognitionResult do |
| 72 | + def encode(value, options) do |
| 73 | + GoogleApi.Gax.ModelBase.encode(value, options) |
| 74 | + end |
| 75 | +end |
0 commit comments