lib/google/cloud/dialogflow/v2/doc/google/cloud/dialogflow/v2/session.rb in google-cloud-dialogflow-0.3.0 vs lib/google/cloud/dialogflow/v2/doc/google/cloud/dialogflow/v2/session.rb in google-cloud-dialogflow-0.4.0
- old
+ new
@@ -129,16 +129,18 @@
# Represents the result of conversational query or event processing.
# @!attribute [rw] query_text
# @return [String]
# The original conversational query text:
+ #
# * If natural language text was provided as input, `query_text` contains
# a copy of the input.
# * If natural language speech audio was provided as input, `query_text`
# contains the speech recognition result. If speech recognizer produced
# multiple alternatives, a particular one is picked.
- # * If an event was provided as input, `query_text` is not set.
+ # * If automatic spell correction is enabled, `query_text` will contain the
+ # corrected user input.
# @!attribute [rw] language_code
# @return [String]
# The language that was triggered during intent detection.
# See [Language
# Support](https://cloud.google.com/dialogflow-enterprise/docs/reference/language)
@@ -161,10 +163,11 @@
# @return [Google::Protobuf::Struct]
# The collection of extracted parameters.
# @!attribute [rw] all_required_params_present
# @return [true, false]
# This field is set to:
+ #
# * `false` if the matched intent has required parameters and not all of
# the required parameter values have been collected.
# * `true` if all required parameter values have been collected, or if the
# matched intent doesn't contain any required parameters.
# @!attribute [rw] fulfillment_text
@@ -372,38 +375,10 @@
# `single_utterance` was set to `true`, and is not used otherwise.
END_OF_SINGLE_UTTERANCE = 2
end
end
- # Instructs the speech recognizer how to process the audio content.
- # @!attribute [rw] audio_encoding
- # @return [Google::Cloud::Dialogflow::V2::AudioEncoding]
- # Required. Audio encoding of the audio content to process.
- # @!attribute [rw] sample_rate_hertz
- # @return [Integer]
- # Required. Sample rate (in Hertz) of the audio content sent in the query.
- # Refer to
- # [Cloud Speech API
- # documentation](https://cloud.google.com/speech-to-text/docs/basics) for
- # more details.
- # @!attribute [rw] language_code
- # @return [String]
- # Required. The language of the supplied audio. Dialogflow does not do
- # translations. See [Language
- # Support](https://cloud.google.com/dialogflow-enterprise/docs/reference/language)
- # for a list of the currently supported language codes. Note that queries in
- # the same session do not necessarily need to specify the same language.
- # @!attribute [rw] phrase_hints
- # @return [Array<String>]
- # Optional. The collection of phrase hints which are used to boost accuracy
- # of speech recognition.
- # Refer to
- # [Cloud Speech API
- # documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints)
- # for more details.
- class InputAudioConfig; end
-
# Represents the natural language text to be processed.
# @!attribute [rw] text
# @return [String]
# Required. The UTF-8 encoded natural language text to be processed.
# Text length must not exceed 256 characters.
@@ -458,58 +433,9 @@
# @!attribute [rw] magnitude
# @return [Float]
# A non-negative number in the [0, +inf) range, which represents the absolute
# magnitude of sentiment, regardless of score (positive or negative).
class Sentiment; end
-
- # Audio encoding of the audio content sent in the conversational query request.
- # Refer to the
- # [Cloud Speech API
- # documentation](https://cloud.google.com/speech-to-text/docs/basics) for more
- # details.
- module AudioEncoding
- # Not specified.
- AUDIO_ENCODING_UNSPECIFIED = 0
-
- # Uncompressed 16-bit signed little-endian samples (Linear PCM).
- AUDIO_ENCODING_LINEAR_16 = 1
-
- # [`FLAC`](https://xiph.org/flac/documentation.html) (Free Lossless Audio
- # Codec) is the recommended encoding because it is lossless (therefore
- # recognition is not compromised) and requires only about half the
- # bandwidth of `LINEAR16`. `FLAC` stream encoding supports 16-bit and
- # 24-bit samples, however, not all fields in `STREAMINFO` are supported.
- AUDIO_ENCODING_FLAC = 2
-
- # 8-bit samples that compand 14-bit audio samples using G.711 PCMU/mu-law.
- AUDIO_ENCODING_MULAW = 3
-
- # Adaptive Multi-Rate Narrowband codec. `sample_rate_hertz` must be 8000.
- AUDIO_ENCODING_AMR = 4
-
- # Adaptive Multi-Rate Wideband codec. `sample_rate_hertz` must be 16000.
- AUDIO_ENCODING_AMR_WB = 5
-
- # Opus encoded audio frames in Ogg container
- # ([OggOpus](https://wiki.xiph.org/OggOpus)).
- # `sample_rate_hertz` must be 16000.
- AUDIO_ENCODING_OGG_OPUS = 6
-
- # Although the use of lossy encodings is not recommended, if a very low
- # bitrate encoding is required, `OGG_OPUS` is highly preferred over
- # Speex encoding. The [Speex](https://speex.org/) encoding supported by
- # Dialogflow API has a header byte in each block, as in MIME type
- # `audio/x-speex-with-header-byte`.
- # It is a variant of the RTP Speex encoding defined in
- # [RFC 5574](https://tools.ietf.org/html/rfc5574).
- # The stream is a sequence of blocks, one block per RTP packet. Each block
- # starts with a byte containing the length of the block, in bytes, followed
- # by one or more frames of Speex data, padded to an integral number of
- # bytes (octets) as specified in RFC 5574. In other words, each RTP header
- # is replaced with a single byte containing the block length. Only Speex
- # wideband is supported. `sample_rate_hertz` must be 16000.
- AUDIO_ENCODING_SPEEX_WITH_HEADER_BYTE = 7
- end
end
end
end
end
\ No newline at end of file