lib/aws-sdk-bedrockagent/client.rb in aws-sdk-bedrockagent-1.21.0 vs lib/aws-sdk-bedrockagent/client.rb in aws-sdk-bedrockagent-1.22.0
- old
+ new
@@ -521,16 +521,20 @@
#
# * To override the default prompt behavior for agent orchestration and
# to use advanced prompts, include a `promptOverrideConfiguration`
# object. For more information, see [Advanced prompts][2].
#
- # * If you agent fails to be created, the response returns a list of
+ # * If your agent fails to be created, the response returns a list of
# `failureReasons` alongside a list of `recommendedActions` for you to
# troubleshoot.
#
+ # * The agent instructions will not be honored if your agent has only
+ # one knowledge base, uses default prompts, has no action group, and
+ # user input is disabled.
#
#
+ #
# [1]: https://docs.aws.amazon.com/bedrock/latest/userguide/agents-configure-memory.html
# [2]: https://docs.aws.amazon.com/bedrock/latest/userguide/advanced-prompts.html
#
# @option params [required, String] :agent_name
# A name for the agent that you create.
@@ -2015,10 +2019,16 @@
# temperature: 1.0,
# top_k: 1,
# top_p: 1.0,
# },
# },
+ # metadata: [
+ # {
+ # key: "PromptMetadataKey", # required
+ # value: "PromptMetadataValue", # required
+ # },
+ # ],
# model_id: "PromptModelIdentifier",
# name: "PromptVariantName", # required
# template_configuration: {
# text: {
# input_variables: [
@@ -2049,10 +2059,13 @@
# resp.variants[0].inference_configuration.text.stop_sequences #=> Array
# resp.variants[0].inference_configuration.text.stop_sequences[0] #=> String
# resp.variants[0].inference_configuration.text.temperature #=> Float
# resp.variants[0].inference_configuration.text.top_k #=> Integer
# resp.variants[0].inference_configuration.text.top_p #=> Float
+ # resp.variants[0].metadata #=> Array
+ # resp.variants[0].metadata[0].key #=> String
+ # resp.variants[0].metadata[0].value #=> String
# resp.variants[0].model_id #=> String
# resp.variants[0].name #=> String
# resp.variants[0].template_configuration.text.input_variables #=> Array
# resp.variants[0].template_configuration.text.input_variables[0].name #=> String
# resp.variants[0].template_configuration.text.text #=> String
@@ -2143,10 +2156,13 @@
# resp.variants[0].inference_configuration.text.stop_sequences #=> Array
# resp.variants[0].inference_configuration.text.stop_sequences[0] #=> String
# resp.variants[0].inference_configuration.text.temperature #=> Float
# resp.variants[0].inference_configuration.text.top_k #=> Integer
# resp.variants[0].inference_configuration.text.top_p #=> Float
+ # resp.variants[0].metadata #=> Array
+ # resp.variants[0].metadata[0].key #=> String
+ # resp.variants[0].metadata[0].value #=> String
# resp.variants[0].model_id #=> String
# resp.variants[0].name #=> String
# resp.variants[0].template_configuration.text.input_variables #=> Array
# resp.variants[0].template_configuration.text.input_variables[0].name #=> String
# resp.variants[0].template_configuration.text.text #=> String
@@ -3389,10 +3405,13 @@
# resp.variants[0].inference_configuration.text.stop_sequences #=> Array
# resp.variants[0].inference_configuration.text.stop_sequences[0] #=> String
# resp.variants[0].inference_configuration.text.temperature #=> Float
# resp.variants[0].inference_configuration.text.top_k #=> Integer
# resp.variants[0].inference_configuration.text.top_p #=> Float
+ # resp.variants[0].metadata #=> Array
+ # resp.variants[0].metadata[0].key #=> String
+ # resp.variants[0].metadata[0].value #=> String
# resp.variants[0].model_id #=> String
# resp.variants[0].name #=> String
# resp.variants[0].template_configuration.text.input_variables #=> Array
# resp.variants[0].template_configuration.text.input_variables[0].name #=> String
# resp.variants[0].template_configuration.text.text #=> String
@@ -5564,10 +5583,16 @@
# temperature: 1.0,
# top_k: 1,
# top_p: 1.0,
# },
# },
+ # metadata: [
+ # {
+ # key: "PromptMetadataKey", # required
+ # value: "PromptMetadataValue", # required
+ # },
+ # ],
# model_id: "PromptModelIdentifier",
# name: "PromptVariantName", # required
# template_configuration: {
# text: {
# input_variables: [
@@ -5598,10 +5623,13 @@
# resp.variants[0].inference_configuration.text.stop_sequences #=> Array
# resp.variants[0].inference_configuration.text.stop_sequences[0] #=> String
# resp.variants[0].inference_configuration.text.temperature #=> Float
# resp.variants[0].inference_configuration.text.top_k #=> Integer
# resp.variants[0].inference_configuration.text.top_p #=> Float
+ # resp.variants[0].metadata #=> Array
+ # resp.variants[0].metadata[0].key #=> String
+ # resp.variants[0].metadata[0].value #=> String
# resp.variants[0].model_id #=> String
# resp.variants[0].name #=> String
# resp.variants[0].template_configuration.text.input_variables #=> Array
# resp.variants[0].template_configuration.text.input_variables[0].name #=> String
# resp.variants[0].template_configuration.text.text #=> String
@@ -5633,10 +5661,10 @@
params: params,
config: config,
tracer: tracer
)
context[:gem_name] = 'aws-sdk-bedrockagent'
- context[:gem_version] = '1.21.0'
+ context[:gem_version] = '1.22.0'
Seahorse::Client::Request.new(handlers, context)
end
# @api private
# @deprecated