lib/aws-sdk-bedrockagent/client.rb in aws-sdk-bedrockagent-1.21.0 vs lib/aws-sdk-bedrockagent/client.rb in aws-sdk-bedrockagent-1.22.0

- old
+ new

@@ -521,16 +521,20 @@ # # * To override the default prompt behavior for agent orchestration and # to use advanced prompts, include a `promptOverrideConfiguration` # object. For more information, see [Advanced prompts][2]. # - # * If you agent fails to be created, the response returns a list of + # * If your agent fails to be created, the response returns a list of # `failureReasons` alongside a list of `recommendedActions` for you to # troubleshoot. # + # * The agent instructions will not be honored if your agent has only + # one knowledge base, uses default prompts, has no action group, and + # user input is disabled. # # + # # [1]: https://docs.aws.amazon.com/bedrock/latest/userguide/agents-configure-memory.html # [2]: https://docs.aws.amazon.com/bedrock/latest/userguide/advanced-prompts.html # # @option params [required, String] :agent_name # A name for the agent that you create. @@ -2015,10 +2019,16 @@ # temperature: 1.0, # top_k: 1, # top_p: 1.0, # }, # }, + # metadata: [ + # { + # key: "PromptMetadataKey", # required + # value: "PromptMetadataValue", # required + # }, + # ], # model_id: "PromptModelIdentifier", # name: "PromptVariantName", # required # template_configuration: { # text: { # input_variables: [ @@ -2049,10 +2059,13 @@ # resp.variants[0].inference_configuration.text.stop_sequences #=> Array # resp.variants[0].inference_configuration.text.stop_sequences[0] #=> String # resp.variants[0].inference_configuration.text.temperature #=> Float # resp.variants[0].inference_configuration.text.top_k #=> Integer # resp.variants[0].inference_configuration.text.top_p #=> Float + # resp.variants[0].metadata #=> Array + # resp.variants[0].metadata[0].key #=> String + # resp.variants[0].metadata[0].value #=> String # resp.variants[0].model_id #=> String # resp.variants[0].name #=> String # resp.variants[0].template_configuration.text.input_variables #=> Array # resp.variants[0].template_configuration.text.input_variables[0].name #=> String # resp.variants[0].template_configuration.text.text #=> String @@ -2143,10 +2156,13 @@ # resp.variants[0].inference_configuration.text.stop_sequences #=> Array # resp.variants[0].inference_configuration.text.stop_sequences[0] #=> String # resp.variants[0].inference_configuration.text.temperature #=> Float # resp.variants[0].inference_configuration.text.top_k #=> Integer # resp.variants[0].inference_configuration.text.top_p #=> Float + # resp.variants[0].metadata #=> Array + # resp.variants[0].metadata[0].key #=> String + # resp.variants[0].metadata[0].value #=> String # resp.variants[0].model_id #=> String # resp.variants[0].name #=> String # resp.variants[0].template_configuration.text.input_variables #=> Array # resp.variants[0].template_configuration.text.input_variables[0].name #=> String # resp.variants[0].template_configuration.text.text #=> String @@ -3389,10 +3405,13 @@ # resp.variants[0].inference_configuration.text.stop_sequences #=> Array # resp.variants[0].inference_configuration.text.stop_sequences[0] #=> String # resp.variants[0].inference_configuration.text.temperature #=> Float # resp.variants[0].inference_configuration.text.top_k #=> Integer # resp.variants[0].inference_configuration.text.top_p #=> Float + # resp.variants[0].metadata #=> Array + # resp.variants[0].metadata[0].key #=> String + # resp.variants[0].metadata[0].value #=> String # resp.variants[0].model_id #=> String # resp.variants[0].name #=> String # resp.variants[0].template_configuration.text.input_variables #=> Array # resp.variants[0].template_configuration.text.input_variables[0].name #=> String # resp.variants[0].template_configuration.text.text #=> String @@ -5564,10 +5583,16 @@ # temperature: 1.0, # top_k: 1, # top_p: 1.0, # }, # }, + # metadata: [ + # { + # key: "PromptMetadataKey", # required + # value: "PromptMetadataValue", # required + # }, + # ], # model_id: "PromptModelIdentifier", # name: "PromptVariantName", # required # template_configuration: { # text: { # input_variables: [ @@ -5598,10 +5623,13 @@ # resp.variants[0].inference_configuration.text.stop_sequences #=> Array # resp.variants[0].inference_configuration.text.stop_sequences[0] #=> String # resp.variants[0].inference_configuration.text.temperature #=> Float # resp.variants[0].inference_configuration.text.top_k #=> Integer # resp.variants[0].inference_configuration.text.top_p #=> Float + # resp.variants[0].metadata #=> Array + # resp.variants[0].metadata[0].key #=> String + # resp.variants[0].metadata[0].value #=> String # resp.variants[0].model_id #=> String # resp.variants[0].name #=> String # resp.variants[0].template_configuration.text.input_variables #=> Array # resp.variants[0].template_configuration.text.input_variables[0].name #=> String # resp.variants[0].template_configuration.text.text #=> String @@ -5633,10 +5661,10 @@ params: params, config: config, tracer: tracer ) context[:gem_name] = 'aws-sdk-bedrockagent' - context[:gem_version] = '1.21.0' + context[:gem_version] = '1.22.0' Seahorse::Client::Request.new(handlers, context) end # @api private # @deprecated