lib/google/cloud/dialogflow/v2/doc/google/cloud/dialogflow/v2/session.rb in google-cloud-dialogflow-0.3.0 vs lib/google/cloud/dialogflow/v2/doc/google/cloud/dialogflow/v2/session.rb in google-cloud-dialogflow-0.4.0

- old
+ new

@@ -129,16 +129,18 @@ # Represents the result of conversational query or event processing. # @!attribute [rw] query_text # @return [String] # The original conversational query text: + # # * If natural language text was provided as input, `query_text` contains # a copy of the input. # * If natural language speech audio was provided as input, `query_text` # contains the speech recognition result. If speech recognizer produced # multiple alternatives, a particular one is picked. - # * If an event was provided as input, `query_text` is not set. + # * If automatic spell correction is enabled, `query_text` will contain the + # corrected user input. # @!attribute [rw] language_code # @return [String] # The language that was triggered during intent detection. # See [Language # Support](https://cloud.google.com/dialogflow-enterprise/docs/reference/language) @@ -161,10 +163,11 @@ # @return [Google::Protobuf::Struct] # The collection of extracted parameters. # @!attribute [rw] all_required_params_present # @return [true, false] # This field is set to: + # # * `false` if the matched intent has required parameters and not all of # the required parameter values have been collected. # * `true` if all required parameter values have been collected, or if the # matched intent doesn't contain any required parameters. # @!attribute [rw] fulfillment_text @@ -372,38 +375,10 @@ # `single_utterance` was set to `true`, and is not used otherwise. END_OF_SINGLE_UTTERANCE = 2 end end - # Instructs the speech recognizer how to process the audio content. - # @!attribute [rw] audio_encoding - # @return [Google::Cloud::Dialogflow::V2::AudioEncoding] - # Required. Audio encoding of the audio content to process. - # @!attribute [rw] sample_rate_hertz - # @return [Integer] - # Required. Sample rate (in Hertz) of the audio content sent in the query. - # Refer to - # [Cloud Speech API - # documentation](https://cloud.google.com/speech-to-text/docs/basics) for - # more details. - # @!attribute [rw] language_code - # @return [String] - # Required. The language of the supplied audio. Dialogflow does not do - # translations. See [Language - # Support](https://cloud.google.com/dialogflow-enterprise/docs/reference/language) - # for a list of the currently supported language codes. Note that queries in - # the same session do not necessarily need to specify the same language. - # @!attribute [rw] phrase_hints - # @return [Array<String>] - # Optional. The collection of phrase hints which are used to boost accuracy - # of speech recognition. - # Refer to - # [Cloud Speech API - # documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints) - # for more details. - class InputAudioConfig; end - # Represents the natural language text to be processed. # @!attribute [rw] text # @return [String] # Required. The UTF-8 encoded natural language text to be processed. # Text length must not exceed 256 characters. @@ -458,58 +433,9 @@ # @!attribute [rw] magnitude # @return [Float] # A non-negative number in the [0, +inf) range, which represents the absolute # magnitude of sentiment, regardless of score (positive or negative). class Sentiment; end - - # Audio encoding of the audio content sent in the conversational query request. - # Refer to the - # [Cloud Speech API - # documentation](https://cloud.google.com/speech-to-text/docs/basics) for more - # details. - module AudioEncoding - # Not specified. - AUDIO_ENCODING_UNSPECIFIED = 0 - - # Uncompressed 16-bit signed little-endian samples (Linear PCM). - AUDIO_ENCODING_LINEAR_16 = 1 - - # [`FLAC`](https://xiph.org/flac/documentation.html) (Free Lossless Audio - # Codec) is the recommended encoding because it is lossless (therefore - # recognition is not compromised) and requires only about half the - # bandwidth of `LINEAR16`. `FLAC` stream encoding supports 16-bit and - # 24-bit samples, however, not all fields in `STREAMINFO` are supported. - AUDIO_ENCODING_FLAC = 2 - - # 8-bit samples that compand 14-bit audio samples using G.711 PCMU/mu-law. - AUDIO_ENCODING_MULAW = 3 - - # Adaptive Multi-Rate Narrowband codec. `sample_rate_hertz` must be 8000. - AUDIO_ENCODING_AMR = 4 - - # Adaptive Multi-Rate Wideband codec. `sample_rate_hertz` must be 16000. - AUDIO_ENCODING_AMR_WB = 5 - - # Opus encoded audio frames in Ogg container - # ([OggOpus](https://wiki.xiph.org/OggOpus)). - # `sample_rate_hertz` must be 16000. - AUDIO_ENCODING_OGG_OPUS = 6 - - # Although the use of lossy encodings is not recommended, if a very low - # bitrate encoding is required, `OGG_OPUS` is highly preferred over - # Speex encoding. The [Speex](https://speex.org/) encoding supported by - # Dialogflow API has a header byte in each block, as in MIME type - # `audio/x-speex-with-header-byte`. - # It is a variant of the RTP Speex encoding defined in - # [RFC 5574](https://tools.ietf.org/html/rfc5574). - # The stream is a sequence of blocks, one block per RTP packet. Each block - # starts with a byte containing the length of the block, in bytes, followed - # by one or more frames of Speex data, padded to an integral number of - # bytes (octets) as specified in RFC 5574. In other words, each RTP header - # is replaced with a single byte containing the block length. Only Speex - # wideband is supported. `sample_rate_hertz` must be 16000. - AUDIO_ENCODING_SPEEX_WITH_HEADER_BYTE = 7 - end end end end end \ No newline at end of file