# WARNING ABOUT GENERATED CODE # # This file is generated. See the contributing guide for more information: # https://github.com/aws/aws-sdk-ruby/blob/master/CONTRIBUTING.md # # WARNING ABOUT GENERATED CODE module Aws::Rekognition module Types # You are not authorized to perform the action. # class AccessDeniedException < Aws::EmptyStructure; end # Structure containing the estimated age range, in years, for a face. # # Amazon Rekognition estimates an age range for faces detected in the # input image. Estimated age ranges can overlap. A face of a 5-year-old # might have an estimated range of 4-6, while the face of a 6-year-old # might have an estimated range of 4-8. # # @!attribute [rw] low # The lowest estimated age. # @return [Integer] # # @!attribute [rw] high # The highest estimated age. # @return [Integer] # class AgeRange < Struct.new( :low, :high) include Aws::Structure end # Assets are the images that you use to train and evaluate a model # version. Assets are referenced by Sagemaker GroundTruth manifest # files. # # @note When making an API call, you may pass Asset # data as a hash: # # { # ground_truth_manifest: { # s3_object: { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # }, # }, # } # # @!attribute [rw] ground_truth_manifest # The S3 bucket that contains the Ground Truth manifest file. # @return [Types::GroundTruthManifest] # class Asset < Struct.new( :ground_truth_manifest) include Aws::Structure end # Indicates whether or not the face has a beard, and the confidence # level in the determination. # # @!attribute [rw] value # Boolean value that indicates whether the face has beard or not. # @return [Boolean] # # @!attribute [rw] confidence # Level of confidence in the determination. # @return [Float] # class Beard < Struct.new( :value, :confidence) include Aws::Structure end # Identifies the bounding box around the label, face, or text. The # `left` (x-coordinate) and `top` (y-coordinate) are coordinates # representing the top and left sides of the bounding box. Note that the # upper-left corner of the image is the origin (0,0). # # The `top` and `left` values returned are ratios of the overall image # size. For example, if the input image is 700x200 pixels, and the # top-left coordinate of the bounding box is 350x50 pixels, the API # returns a `left` value of 0.5 (350/700) and a `top` value of 0.25 # (50/200). # # The `width` and `height` values represent the dimensions of the # bounding box as a ratio of the overall image dimension. For example, # if the input image is 700x200 pixels, and the bounding box width is 70 # pixels, the width returned is 0.1. # # The bounding box coordinates can have negative values. For example, if # Amazon Rekognition is able to detect a face that is at the image edge # and is only partially visible, the service can return coordinates that # are outside the image bounds and, depending on the image edge, you # might get negative values or values greater than 1 for the `left` or # `top` values. # # # # @note When making an API call, you may pass BoundingBox # data as a hash: # # { # width: 1.0, # height: 1.0, # left: 1.0, # top: 1.0, # } # # @!attribute [rw] width # Width of the bounding box as a ratio of the overall image width. # @return [Float] # # @!attribute [rw] height # Height of the bounding box as a ratio of the overall image height. # @return [Float] # # @!attribute [rw] left # Left coordinate of the bounding box as a ratio of overall image # width. # @return [Float] # # @!attribute [rw] top # Top coordinate of the bounding box as a ratio of overall image # height. # @return [Float] # class BoundingBox < Struct.new( :width, :height, :left, :top) include Aws::Structure end # Provides information about a celebrity recognized by the # RecognizeCelebrities operation. # # @!attribute [rw] urls # An array of URLs pointing to additional information about the # celebrity. If there is no additional information about the # celebrity, this list is empty. # @return [Array] # # @!attribute [rw] name # The name of the celebrity. # @return [String] # # @!attribute [rw] id # A unique identifier for the celebrity. # @return [String] # # @!attribute [rw] face # Provides information about the celebrity's face, such as its # location on the image. # @return [Types::ComparedFace] # # @!attribute [rw] match_confidence # The confidence, in percentage, that Amazon Rekognition has that the # recognized face is the celebrity. # @return [Float] # class Celebrity < Struct.new( :urls, :name, :id, :face, :match_confidence) include Aws::Structure end # Information about a recognized celebrity. # # @!attribute [rw] urls # An array of URLs pointing to additional celebrity information. # @return [Array] # # @!attribute [rw] name # The name of the celebrity. # @return [String] # # @!attribute [rw] id # The unique identifier for the celebrity. # @return [String] # # @!attribute [rw] confidence # The confidence, in percentage, that Amazon Rekognition has that the # recognized face is the celebrity. # @return [Float] # # @!attribute [rw] bounding_box # Bounding box around the body of a celebrity. # @return [Types::BoundingBox] # # @!attribute [rw] face # Face details for the recognized celebrity. # @return [Types::FaceDetail] # class CelebrityDetail < Struct.new( :urls, :name, :id, :confidence, :bounding_box, :face) include Aws::Structure end # Information about a detected celebrity and the time the celebrity was # detected in a stored video. For more information, see # GetCelebrityRecognition in the Amazon Rekognition Developer Guide. # # @!attribute [rw] timestamp # The time, in milliseconds from the start of the video, that the # celebrity was recognized. # @return [Integer] # # @!attribute [rw] celebrity # Information about a recognized celebrity. # @return [Types::CelebrityDetail] # class CelebrityRecognition < Struct.new( :timestamp, :celebrity) include Aws::Structure end # Provides information about a face in a target image that matches the # source image face analyzed by `CompareFaces`. The `Face` property # contains the bounding box of the face in the target image. The # `Similarity` property is the confidence that the source image face # matches the face in the bounding box. # # @!attribute [rw] similarity # Level of confidence that the faces match. # @return [Float] # # @!attribute [rw] face # Provides face metadata (bounding box and confidence that the # bounding box actually contains a face). # @return [Types::ComparedFace] # class CompareFacesMatch < Struct.new( :similarity, :face) include Aws::Structure end # @note When making an API call, you may pass CompareFacesRequest # data as a hash: # # { # source_image: { # required # bytes: "data", # s3_object: { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # }, # }, # target_image: { # required # bytes: "data", # s3_object: { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # }, # }, # similarity_threshold: 1.0, # quality_filter: "NONE", # accepts NONE, AUTO, LOW, MEDIUM, HIGH # } # # @!attribute [rw] source_image # The input image as base64-encoded bytes or an S3 object. If you use # the AWS CLI to call Amazon Rekognition operations, passing # base64-encoded image bytes is not supported. # # If you are using an AWS SDK to call Amazon Rekognition, you might # not need to base64-encode image bytes passed using the `Bytes` # field. For more information, see Images in the Amazon Rekognition # developer guide. # @return [Types::Image] # # @!attribute [rw] target_image # The target image as base64-encoded bytes or an S3 object. If you use # the AWS CLI to call Amazon Rekognition operations, passing # base64-encoded image bytes is not supported. # # If you are using an AWS SDK to call Amazon Rekognition, you might # not need to base64-encode image bytes passed using the `Bytes` # field. For more information, see Images in the Amazon Rekognition # developer guide. # @return [Types::Image] # # @!attribute [rw] similarity_threshold # The minimum level of confidence in the face matches that a match # must meet to be included in the `FaceMatches` array. # @return [Float] # # @!attribute [rw] quality_filter # A filter that specifies a quality bar for how much filtering is done # to identify faces. Filtered faces aren't compared. If you specify # `AUTO`, Amazon Rekognition chooses the quality bar. If you specify # `LOW`, `MEDIUM`, or `HIGH`, filtering removes all faces that don’t # meet the chosen quality bar. The quality bar is based on a variety # of common use cases. Low-quality detections can occur for a number # of reasons. Some examples are an object that's misidentified as a # face, a face that's too blurry, or a face with a pose that's too # extreme to use. If you specify `NONE`, no filtering is performed. # The default value is `NONE`. # # To use quality filtering, the collection you are using must be # associated with version 3 of the face model or higher. # @return [String] # class CompareFacesRequest < Struct.new( :source_image, :target_image, :similarity_threshold, :quality_filter) include Aws::Structure end # @!attribute [rw] source_image_face # The face in the source image that was used for comparison. # @return [Types::ComparedSourceImageFace] # # @!attribute [rw] face_matches # An array of faces in the target image that match the source image # face. Each `CompareFacesMatch` object provides the bounding box, the # confidence level that the bounding box contains a face, and the # similarity score for the face in the bounding box and the face in # the source image. # @return [Array] # # @!attribute [rw] unmatched_faces # An array of faces in the target image that did not match the source # image face. # @return [Array] # # @!attribute [rw] source_image_orientation_correction # The value of `SourceImageOrientationCorrection` is always null. # # If the input image is in .jpeg format, it might contain exchangeable # image file format (Exif) metadata that includes the image's # orientation. Amazon Rekognition uses this orientation information to # perform image correction. The bounding box coordinates are # translated to represent object locations after the orientation # information in the Exif metadata is used to correct the image # orientation. Images in .png format don't contain Exif metadata. # # Amazon Rekognition doesn’t perform image correction for images in # .png format and .jpeg images without orientation information in the # image Exif metadata. The bounding box coordinates aren't translated # and represent the object locations before the image is rotated. # @return [String] # # @!attribute [rw] target_image_orientation_correction # The value of `TargetImageOrientationCorrection` is always null. # # If the input image is in .jpeg format, it might contain exchangeable # image file format (Exif) metadata that includes the image's # orientation. Amazon Rekognition uses this orientation information to # perform image correction. The bounding box coordinates are # translated to represent object locations after the orientation # information in the Exif metadata is used to correct the image # orientation. Images in .png format don't contain Exif metadata. # # Amazon Rekognition doesn’t perform image correction for images in # .png format and .jpeg images without orientation information in the # image Exif metadata. The bounding box coordinates aren't translated # and represent the object locations before the image is rotated. # @return [String] # class CompareFacesResponse < Struct.new( :source_image_face, :face_matches, :unmatched_faces, :source_image_orientation_correction, :target_image_orientation_correction) include Aws::Structure end # Provides face metadata for target image faces that are analyzed by # `CompareFaces` and `RecognizeCelebrities`. # # @!attribute [rw] bounding_box # Bounding box of the face. # @return [Types::BoundingBox] # # @!attribute [rw] confidence # Level of confidence that what the bounding box contains is a face. # @return [Float] # # @!attribute [rw] landmarks # An array of facial landmarks. # @return [Array] # # @!attribute [rw] pose # Indicates the pose of the face as determined by its pitch, roll, and # yaw. # @return [Types::Pose] # # @!attribute [rw] quality # Identifies face image brightness and sharpness. # @return [Types::ImageQuality] # class ComparedFace < Struct.new( :bounding_box, :confidence, :landmarks, :pose, :quality) include Aws::Structure end # Type that describes the face Amazon Rekognition chose to compare with # the faces in the target. This contains a bounding box for the selected # face and confidence level that the bounding box contains a face. Note # that Amazon Rekognition selects the largest face in the source image # for this comparison. # # @!attribute [rw] bounding_box # Bounding box of the face. # @return [Types::BoundingBox] # # @!attribute [rw] confidence # Confidence level that the selected bounding box contains a face. # @return [Float] # class ComparedSourceImageFace < Struct.new( :bounding_box, :confidence) include Aws::Structure end # Information about an unsafe content label detection in a stored video. # # @!attribute [rw] timestamp # Time, in milliseconds from the beginning of the video, that the # unsafe content label was detected. # @return [Integer] # # @!attribute [rw] moderation_label # The unsafe content label detected by in the stored video. # @return [Types::ModerationLabel] # class ContentModerationDetection < Struct.new( :timestamp, :moderation_label) include Aws::Structure end # @note When making an API call, you may pass CreateCollectionRequest # data as a hash: # # { # collection_id: "CollectionId", # required # } # # @!attribute [rw] collection_id # ID for the collection that you are creating. # @return [String] # class CreateCollectionRequest < Struct.new( :collection_id) include Aws::Structure end # @!attribute [rw] status_code # HTTP status code indicating the result of the operation. # @return [Integer] # # @!attribute [rw] collection_arn # Amazon Resource Name (ARN) of the collection. You can use this to # manage permissions on your resources. # @return [String] # # @!attribute [rw] face_model_version # Version number of the face detection model associated with the # collection you are creating. # @return [String] # class CreateCollectionResponse < Struct.new( :status_code, :collection_arn, :face_model_version) include Aws::Structure end # @note When making an API call, you may pass CreateProjectRequest # data as a hash: # # { # project_name: "ProjectName", # required # } # # @!attribute [rw] project_name # The name of the project to create. # @return [String] # class CreateProjectRequest < Struct.new( :project_name) include Aws::Structure end # @!attribute [rw] project_arn # The Amazon Resource Name (ARN) of the new project. You can use the # ARN to configure IAM access to the project. # @return [String] # class CreateProjectResponse < Struct.new( :project_arn) include Aws::Structure end # @note When making an API call, you may pass CreateProjectVersionRequest # data as a hash: # # { # project_arn: "ProjectArn", # required # version_name: "VersionName", # required # output_config: { # required # s3_bucket: "S3Bucket", # s3_key_prefix: "S3KeyPrefix", # }, # training_data: { # required # assets: [ # { # ground_truth_manifest: { # s3_object: { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # }, # }, # }, # ], # }, # testing_data: { # required # assets: [ # { # ground_truth_manifest: { # s3_object: { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # }, # }, # }, # ], # auto_create: false, # }, # } # # @!attribute [rw] project_arn # The ARN of the Amazon Rekognition Custom Labels project that manages # the model that you want to train. # @return [String] # # @!attribute [rw] version_name # A name for the version of the model. This value must be unique. # @return [String] # # @!attribute [rw] output_config # The Amazon S3 location to store the results of training. # @return [Types::OutputConfig] # # @!attribute [rw] training_data # The dataset to use for training. # @return [Types::TrainingData] # # @!attribute [rw] testing_data # The dataset to use for testing. # @return [Types::TestingData] # class CreateProjectVersionRequest < Struct.new( :project_arn, :version_name, :output_config, :training_data, :testing_data) include Aws::Structure end # @!attribute [rw] project_version_arn # The ARN of the model version that was created. Use # `DescribeProjectVersion` to get the current status of the training # operation. # @return [String] # class CreateProjectVersionResponse < Struct.new( :project_version_arn) include Aws::Structure end # @note When making an API call, you may pass CreateStreamProcessorRequest # data as a hash: # # { # input: { # required # kinesis_video_stream: { # arn: "KinesisVideoArn", # }, # }, # output: { # required # kinesis_data_stream: { # arn: "KinesisDataArn", # }, # }, # name: "StreamProcessorName", # required # settings: { # required # face_search: { # collection_id: "CollectionId", # face_match_threshold: 1.0, # }, # }, # role_arn: "RoleArn", # required # } # # @!attribute [rw] input # Kinesis video stream stream that provides the source streaming # video. If you are using the AWS CLI, the parameter name is # `StreamProcessorInput`. # @return [Types::StreamProcessorInput] # # @!attribute [rw] output # Kinesis data stream stream to which Amazon Rekognition Video puts # the analysis results. If you are using the AWS CLI, the parameter # name is `StreamProcessorOutput`. # @return [Types::StreamProcessorOutput] # # @!attribute [rw] name # An identifier you assign to the stream processor. You can use `Name` # to manage the stream processor. For example, you can get the current # status of the stream processor by calling DescribeStreamProcessor. # `Name` is idempotent. # @return [String] # # @!attribute [rw] settings # Face recognition input parameters to be used by the stream # processor. Includes the collection to use for face recognition and # the face attributes to detect. # @return [Types::StreamProcessorSettings] # # @!attribute [rw] role_arn # ARN of the IAM role that allows access to the stream processor. # @return [String] # class CreateStreamProcessorRequest < Struct.new( :input, :output, :name, :settings, :role_arn) include Aws::Structure end # @!attribute [rw] stream_processor_arn # ARN for the newly create stream processor. # @return [String] # class CreateStreamProcessorResponse < Struct.new( :stream_processor_arn) include Aws::Structure end # A custom label detected in an image by a call to DetectCustomLabels. # # @!attribute [rw] name # The name of the custom label. # @return [String] # # @!attribute [rw] confidence # The confidence that the model has in the detection of the custom # label. The range is 0-100. A higher value indicates a higher # confidence. # @return [Float] # # @!attribute [rw] geometry # The location of the detected object on the image that corresponds to # the custom label. Includes an axis aligned coarse bounding box # surrounding the object and a finer grain polygon for more accurate # spatial information. # @return [Types::Geometry] # class CustomLabel < Struct.new( :name, :confidence, :geometry) include Aws::Structure end # @note When making an API call, you may pass DeleteCollectionRequest # data as a hash: # # { # collection_id: "CollectionId", # required # } # # @!attribute [rw] collection_id # ID of the collection to delete. # @return [String] # class DeleteCollectionRequest < Struct.new( :collection_id) include Aws::Structure end # @!attribute [rw] status_code # HTTP status code that indicates the result of the operation. # @return [Integer] # class DeleteCollectionResponse < Struct.new( :status_code) include Aws::Structure end # @note When making an API call, you may pass DeleteFacesRequest # data as a hash: # # { # collection_id: "CollectionId", # required # face_ids: ["FaceId"], # required # } # # @!attribute [rw] collection_id # Collection from which to remove the specific faces. # @return [String] # # @!attribute [rw] face_ids # An array of face IDs to delete. # @return [Array] # class DeleteFacesRequest < Struct.new( :collection_id, :face_ids) include Aws::Structure end # @!attribute [rw] deleted_faces # An array of strings (face IDs) of the faces that were deleted. # @return [Array] # class DeleteFacesResponse < Struct.new( :deleted_faces) include Aws::Structure end # @note When making an API call, you may pass DeleteProjectRequest # data as a hash: # # { # project_arn: "ProjectArn", # required # } # # @!attribute [rw] project_arn # The Amazon Resource Name (ARN) of the project that you want to # delete. # @return [String] # class DeleteProjectRequest < Struct.new( :project_arn) include Aws::Structure end # @!attribute [rw] status # The current status of the delete project operation. # @return [String] # class DeleteProjectResponse < Struct.new( :status) include Aws::Structure end # @note When making an API call, you may pass DeleteProjectVersionRequest # data as a hash: # # { # project_version_arn: "ProjectVersionArn", # required # } # # @!attribute [rw] project_version_arn # The Amazon Resource Name (ARN) of the model version that you want to # delete. # @return [String] # class DeleteProjectVersionRequest < Struct.new( :project_version_arn) include Aws::Structure end # @!attribute [rw] status # The status of the deletion operation. # @return [String] # class DeleteProjectVersionResponse < Struct.new( :status) include Aws::Structure end # @note When making an API call, you may pass DeleteStreamProcessorRequest # data as a hash: # # { # name: "StreamProcessorName", # required # } # # @!attribute [rw] name # The name of the stream processor you want to delete. # @return [String] # class DeleteStreamProcessorRequest < Struct.new( :name) include Aws::Structure end class DeleteStreamProcessorResponse < Aws::EmptyStructure; end # @note When making an API call, you may pass DescribeCollectionRequest # data as a hash: # # { # collection_id: "CollectionId", # required # } # # @!attribute [rw] collection_id # The ID of the collection to describe. # @return [String] # class DescribeCollectionRequest < Struct.new( :collection_id) include Aws::Structure end # @!attribute [rw] face_count # The number of faces that are indexed into the collection. To index # faces into a collection, use IndexFaces. # @return [Integer] # # @!attribute [rw] face_model_version # The version of the face model that's used by the collection for # face detection. # # For more information, see Model Versioning in the Amazon Rekognition # Developer Guide. # @return [String] # # @!attribute [rw] collection_arn # The Amazon Resource Name (ARN) of the collection. # @return [String] # # @!attribute [rw] creation_timestamp # The number of milliseconds since the Unix epoch time until the # creation of the collection. The Unix epoch time is 00:00:00 # Coordinated Universal Time (UTC), Thursday, 1 January 1970. # @return [Time] # class DescribeCollectionResponse < Struct.new( :face_count, :face_model_version, :collection_arn, :creation_timestamp) include Aws::Structure end # @note When making an API call, you may pass DescribeProjectVersionsRequest # data as a hash: # # { # project_arn: "ProjectArn", # required # version_names: ["VersionName"], # next_token: "ExtendedPaginationToken", # max_results: 1, # } # # @!attribute [rw] project_arn # The Amazon Resource Name (ARN) of the project that contains the # models you want to describe. # @return [String] # # @!attribute [rw] version_names # A list of model version names that you want to describe. You can add # up to 10 model version names to the list. If you don't specify a # value, all model descriptions are returned. # @return [Array] # # @!attribute [rw] next_token # If the previous response was incomplete (because there is more # results to retrieve), Amazon Rekognition Custom Labels returns a # pagination token in the response. You can use this pagination token # to retrieve the next set of results. # @return [String] # # @!attribute [rw] max_results # The maximum number of results to return per paginated call. The # largest value you can specify is 100. If you specify a value greater # than 100, a ValidationException error occurs. The default value is # 100. # @return [Integer] # class DescribeProjectVersionsRequest < Struct.new( :project_arn, :version_names, :next_token, :max_results) include Aws::Structure end # @!attribute [rw] project_version_descriptions # A list of model descriptions. The list is sorted by the creation # date and time of the model versions, latest to earliest. # @return [Array] # # @!attribute [rw] next_token # If the previous response was incomplete (because there is more # results to retrieve), Amazon Rekognition Custom Labels returns a # pagination token in the response. You can use this pagination token # to retrieve the next set of results. # @return [String] # class DescribeProjectVersionsResponse < Struct.new( :project_version_descriptions, :next_token) include Aws::Structure end # @note When making an API call, you may pass DescribeProjectsRequest # data as a hash: # # { # next_token: "ExtendedPaginationToken", # max_results: 1, # } # # @!attribute [rw] next_token # If the previous response was incomplete (because there is more # results to retrieve), Amazon Rekognition Custom Labels returns a # pagination token in the response. You can use this pagination token # to retrieve the next set of results. # @return [String] # # @!attribute [rw] max_results # The maximum number of results to return per paginated call. The # largest value you can specify is 100. If you specify a value greater # than 100, a ValidationException error occurs. The default value is # 100. # @return [Integer] # class DescribeProjectsRequest < Struct.new( :next_token, :max_results) include Aws::Structure end # @!attribute [rw] project_descriptions # A list of project descriptions. The list is sorted by the date and # time the projects are created. # @return [Array] # # @!attribute [rw] next_token # If the previous response was incomplete (because there is more # results to retrieve), Amazon Rekognition Custom Labels returns a # pagination token in the response. You can use this pagination token # to retrieve the next set of results. # @return [String] # class DescribeProjectsResponse < Struct.new( :project_descriptions, :next_token) include Aws::Structure end # @note When making an API call, you may pass DescribeStreamProcessorRequest # data as a hash: # # { # name: "StreamProcessorName", # required # } # # @!attribute [rw] name # Name of the stream processor for which you want information. # @return [String] # class DescribeStreamProcessorRequest < Struct.new( :name) include Aws::Structure end # @!attribute [rw] name # Name of the stream processor. # @return [String] # # @!attribute [rw] stream_processor_arn # ARN of the stream processor. # @return [String] # # @!attribute [rw] status # Current status of the stream processor. # @return [String] # # @!attribute [rw] status_message # Detailed status message about the stream processor. # @return [String] # # @!attribute [rw] creation_timestamp # Date and time the stream processor was created # @return [Time] # # @!attribute [rw] last_update_timestamp # The time, in Unix format, the stream processor was last updated. For # example, when the stream processor moves from a running state to a # failed state, or when the user starts or stops the stream processor. # @return [Time] # # @!attribute [rw] input # Kinesis video stream that provides the source streaming video. # @return [Types::StreamProcessorInput] # # @!attribute [rw] output # Kinesis data stream to which Amazon Rekognition Video puts the # analysis results. # @return [Types::StreamProcessorOutput] # # @!attribute [rw] role_arn # ARN of the IAM role that allows access to the stream processor. # @return [String] # # @!attribute [rw] settings # Face recognition input parameters that are being used by the stream # processor. Includes the collection to use for face recognition and # the face attributes to detect. # @return [Types::StreamProcessorSettings] # class DescribeStreamProcessorResponse < Struct.new( :name, :stream_processor_arn, :status, :status_message, :creation_timestamp, :last_update_timestamp, :input, :output, :role_arn, :settings) include Aws::Structure end # @note When making an API call, you may pass DetectCustomLabelsRequest # data as a hash: # # { # project_version_arn: "ProjectVersionArn", # required # image: { # required # bytes: "data", # s3_object: { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # }, # }, # max_results: 1, # min_confidence: 1.0, # } # # @!attribute [rw] project_version_arn # The ARN of the model version that you want to use. # @return [String] # # @!attribute [rw] image # Provides the input image either as bytes or an S3 object. # # You pass image bytes to an Amazon Rekognition API operation by using # the `Bytes` property. For example, you would use the `Bytes` # property to pass an image loaded from a local file system. Image # bytes passed by using the `Bytes` property must be base64-encoded. # Your code may not need to encode image bytes if you are using an AWS # SDK to call Amazon Rekognition API operations. # # For more information, see Analyzing an Image Loaded from a Local # File System in the Amazon Rekognition Developer Guide. # # You pass images stored in an S3 bucket to an Amazon Rekognition API # operation by using the `S3Object` property. Images stored in an S3 # bucket do not need to be base64-encoded. # # The region for the S3 bucket containing the S3 object must match the # region you use for Amazon Rekognition operations. # # If you use the AWS CLI to call Amazon Rekognition operations, # passing image bytes using the Bytes property is not supported. You # must first upload the image to an Amazon S3 bucket and then call the # operation using the S3Object property. # # For Amazon Rekognition to process an S3 object, the user must have # permission to access the S3 object. For more information, see # Resource Based Policies in the Amazon Rekognition Developer Guide. # @return [Types::Image] # # @!attribute [rw] max_results # Maximum number of results you want the service to return in the # response. The service returns the specified number of highest # confidence labels ranked from highest confidence to lowest. # @return [Integer] # # @!attribute [rw] min_confidence # Specifies the minimum confidence level for the labels to return. # Amazon Rekognition doesn't return any labels with a confidence # lower than this specified value. If you specify a value of 0, all # labels are return, regardless of the default thresholds that the # model version applies. # @return [Float] # class DetectCustomLabelsRequest < Struct.new( :project_version_arn, :image, :max_results, :min_confidence) include Aws::Structure end # @!attribute [rw] custom_labels # An array of custom labels detected in the input image. # @return [Array] # class DetectCustomLabelsResponse < Struct.new( :custom_labels) include Aws::Structure end # @note When making an API call, you may pass DetectFacesRequest # data as a hash: # # { # image: { # required # bytes: "data", # s3_object: { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # }, # }, # attributes: ["DEFAULT"], # accepts DEFAULT, ALL # } # # @!attribute [rw] image # The input image as base64-encoded bytes or an S3 object. If you use # the AWS CLI to call Amazon Rekognition operations, passing # base64-encoded image bytes is not supported. # # If you are using an AWS SDK to call Amazon Rekognition, you might # not need to base64-encode image bytes passed using the `Bytes` # field. For more information, see Images in the Amazon Rekognition # developer guide. # @return [Types::Image] # # @!attribute [rw] attributes # An array of facial attributes you want to be returned. This can be # the default list of attributes or all attributes. If you don't # specify a value for `Attributes` or if you specify `["DEFAULT"]`, # the API returns the following subset of facial attributes: # `BoundingBox`, `Confidence`, `Pose`, `Quality`, and `Landmarks`. If # you provide `["ALL"]`, all facial attributes are returned, but the # operation takes longer to complete. # # If you provide both, `["ALL", "DEFAULT"]`, the service uses a # logical AND operator to determine which attributes to return (in # this case, all attributes). # @return [Array] # class DetectFacesRequest < Struct.new( :image, :attributes) include Aws::Structure end # @!attribute [rw] face_details # Details of each face found in the image. # @return [Array] # # @!attribute [rw] orientation_correction # The value of `OrientationCorrection` is always null. # # If the input image is in .jpeg format, it might contain exchangeable # image file format (Exif) metadata that includes the image's # orientation. Amazon Rekognition uses this orientation information to # perform image correction. The bounding box coordinates are # translated to represent object locations after the orientation # information in the Exif metadata is used to correct the image # orientation. Images in .png format don't contain Exif metadata. # # Amazon Rekognition doesn’t perform image correction for images in # .png format and .jpeg images without orientation information in the # image Exif metadata. The bounding box coordinates aren't translated # and represent the object locations before the image is rotated. # @return [String] # class DetectFacesResponse < Struct.new( :face_details, :orientation_correction) include Aws::Structure end # @note When making an API call, you may pass DetectLabelsRequest # data as a hash: # # { # image: { # required # bytes: "data", # s3_object: { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # }, # }, # max_labels: 1, # min_confidence: 1.0, # } # # @!attribute [rw] image # The input image as base64-encoded bytes or an S3 object. If you use # the AWS CLI to call Amazon Rekognition operations, passing image # bytes is not supported. Images stored in an S3 Bucket do not need to # be base64-encoded. # # If you are using an AWS SDK to call Amazon Rekognition, you might # not need to base64-encode image bytes passed using the `Bytes` # field. For more information, see Images in the Amazon Rekognition # developer guide. # @return [Types::Image] # # @!attribute [rw] max_labels # Maximum number of labels you want the service to return in the # response. The service returns the specified number of highest # confidence labels. # @return [Integer] # # @!attribute [rw] min_confidence # Specifies the minimum confidence level for the labels to return. # Amazon Rekognition doesn't return any labels with confidence lower # than this specified value. # # If `MinConfidence` is not specified, the operation returns labels # with a confidence values greater than or equal to 55 percent. # @return [Float] # class DetectLabelsRequest < Struct.new( :image, :max_labels, :min_confidence) include Aws::Structure end # @!attribute [rw] labels # An array of labels for the real-world objects detected. # @return [Array] # # @!attribute [rw] orientation_correction # The value of `OrientationCorrection` is always null. # # If the input image is in .jpeg format, it might contain exchangeable # image file format (Exif) metadata that includes the image's # orientation. Amazon Rekognition uses this orientation information to # perform image correction. The bounding box coordinates are # translated to represent object locations after the orientation # information in the Exif metadata is used to correct the image # orientation. Images in .png format don't contain Exif metadata. # # Amazon Rekognition doesn’t perform image correction for images in # .png format and .jpeg images without orientation information in the # image Exif metadata. The bounding box coordinates aren't translated # and represent the object locations before the image is rotated. # @return [String] # # @!attribute [rw] label_model_version # Version number of the label detection model that was used to detect # labels. # @return [String] # class DetectLabelsResponse < Struct.new( :labels, :orientation_correction, :label_model_version) include Aws::Structure end # @note When making an API call, you may pass DetectModerationLabelsRequest # data as a hash: # # { # image: { # required # bytes: "data", # s3_object: { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # }, # }, # min_confidence: 1.0, # human_loop_config: { # human_loop_name: "HumanLoopName", # required # flow_definition_arn: "FlowDefinitionArn", # required # data_attributes: { # content_classifiers: ["FreeOfPersonallyIdentifiableInformation"], # accepts FreeOfPersonallyIdentifiableInformation, FreeOfAdultContent # }, # }, # } # # @!attribute [rw] image # The input image as base64-encoded bytes or an S3 object. If you use # the AWS CLI to call Amazon Rekognition operations, passing # base64-encoded image bytes is not supported. # # If you are using an AWS SDK to call Amazon Rekognition, you might # not need to base64-encode image bytes passed using the `Bytes` # field. For more information, see Images in the Amazon Rekognition # developer guide. # @return [Types::Image] # # @!attribute [rw] min_confidence # Specifies the minimum confidence level for the labels to return. # Amazon Rekognition doesn't return any labels with a confidence # level lower than this specified value. # # If you don't specify `MinConfidence`, the operation returns labels # with confidence values greater than or equal to 50 percent. # @return [Float] # # @!attribute [rw] human_loop_config # Sets up the configuration for human evaluation, including the # FlowDefinition the image will be sent to. # @return [Types::HumanLoopConfig] # class DetectModerationLabelsRequest < Struct.new( :image, :min_confidence, :human_loop_config) include Aws::Structure end # @!attribute [rw] moderation_labels # Array of detected Moderation labels and the time, in milliseconds # from the start of the video, they were detected. # @return [Array] # # @!attribute [rw] moderation_model_version # Version number of the moderation detection model that was used to # detect unsafe content. # @return [String] # # @!attribute [rw] human_loop_activation_output # Shows the results of the human in the loop evaluation. # @return [Types::HumanLoopActivationOutput] # class DetectModerationLabelsResponse < Struct.new( :moderation_labels, :moderation_model_version, :human_loop_activation_output) include Aws::Structure end # A set of optional parameters that you can use to set the criteria that # the text must meet to be included in your response. `WordFilter` looks # at a word’s height, width, and minimum confidence. `RegionOfInterest` # lets you set a specific region of the image to look for text in. # # @note When making an API call, you may pass DetectTextFilters # data as a hash: # # { # word_filter: { # min_confidence: 1.0, # min_bounding_box_height: 1.0, # min_bounding_box_width: 1.0, # }, # regions_of_interest: [ # { # bounding_box: { # width: 1.0, # height: 1.0, # left: 1.0, # top: 1.0, # }, # }, # ], # } # # @!attribute [rw] word_filter # A set of parameters that allow you to filter out certain results # from your returned results. # @return [Types::DetectionFilter] # # @!attribute [rw] regions_of_interest # A Filter focusing on a certain area of the image. Uses a # `BoundingBox` object to set the region of the image. # @return [Array] # class DetectTextFilters < Struct.new( :word_filter, :regions_of_interest) include Aws::Structure end # @note When making an API call, you may pass DetectTextRequest # data as a hash: # # { # image: { # required # bytes: "data", # s3_object: { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # }, # }, # filters: { # word_filter: { # min_confidence: 1.0, # min_bounding_box_height: 1.0, # min_bounding_box_width: 1.0, # }, # regions_of_interest: [ # { # bounding_box: { # width: 1.0, # height: 1.0, # left: 1.0, # top: 1.0, # }, # }, # ], # }, # } # # @!attribute [rw] image # The input image as base64-encoded bytes or an Amazon S3 object. If # you use the AWS CLI to call Amazon Rekognition operations, you # can't pass image bytes. # # If you are using an AWS SDK to call Amazon Rekognition, you might # not need to base64-encode image bytes passed using the `Bytes` # field. For more information, see Images in the Amazon Rekognition # developer guide. # @return [Types::Image] # # @!attribute [rw] filters # Optional parameters that let you set the criteria that the text must # meet to be included in your response. # @return [Types::DetectTextFilters] # class DetectTextRequest < Struct.new( :image, :filters) include Aws::Structure end # @!attribute [rw] text_detections # An array of text that was detected in the input image. # @return [Array] # # @!attribute [rw] text_model_version # The model version used to detect text. # @return [String] # class DetectTextResponse < Struct.new( :text_detections, :text_model_version) include Aws::Structure end # A set of parameters that allow you to filter out certain results from # your returned results. # # @note When making an API call, you may pass DetectionFilter # data as a hash: # # { # min_confidence: 1.0, # min_bounding_box_height: 1.0, # min_bounding_box_width: 1.0, # } # # @!attribute [rw] min_confidence # Sets confidence of word detection. Words with detection confidence # below this will be excluded from the result. Values should be # between 0.5 and 1 as Text in Video will not return any result below # 0.5. # @return [Float] # # @!attribute [rw] min_bounding_box_height # Sets the minimum height of the word bounding box. Words with # bounding box heights lesser than this value will be excluded from # the result. Value is relative to the video frame height. # @return [Float] # # @!attribute [rw] min_bounding_box_width # Sets the minimum width of the word bounding box. Words with bounding # boxes widths lesser than this value will be excluded from the # result. Value is relative to the video frame width. # @return [Float] # class DetectionFilter < Struct.new( :min_confidence, :min_bounding_box_height, :min_bounding_box_width) include Aws::Structure end # The emotions that appear to be expressed on the face, and the # confidence level in the determination. The API is only making a # determination of the physical appearance of a person's face. It is # not a determination of the person’s internal emotional state and # should not be used in such a way. For example, a person pretending to # have a sad face might not be sad emotionally. # # @!attribute [rw] type # Type of emotion detected. # @return [String] # # @!attribute [rw] confidence # Level of confidence in the determination. # @return [Float] # class Emotion < Struct.new( :type, :confidence) include Aws::Structure end # The evaluation results for the training of a model. # # @!attribute [rw] f1_score # The F1 score for the evaluation of all labels. The F1 score metric # evaluates the overall precision and recall performance of the model # as a single value. A higher value indicates better precision and # recall performance. A lower score indicates that precision, recall, # or both are performing poorly. # @return [Float] # # @!attribute [rw] summary # The S3 bucket that contains the training summary. # @return [Types::Summary] # class EvaluationResult < Struct.new( :f1_score, :summary) include Aws::Structure end # Indicates whether or not the eyes on the face are open, and the # confidence level in the determination. # # @!attribute [rw] value # Boolean value that indicates whether the eyes on the face are open. # @return [Boolean] # # @!attribute [rw] confidence # Level of confidence in the determination. # @return [Float] # class EyeOpen < Struct.new( :value, :confidence) include Aws::Structure end # Indicates whether or not the face is wearing eye glasses, and the # confidence level in the determination. # # @!attribute [rw] value # Boolean value that indicates whether the face is wearing eye glasses # or not. # @return [Boolean] # # @!attribute [rw] confidence # Level of confidence in the determination. # @return [Float] # class Eyeglasses < Struct.new( :value, :confidence) include Aws::Structure end # Describes the face properties such as the bounding box, face ID, image # ID of the input image, and external image ID that you assigned. # # @!attribute [rw] face_id # Unique identifier that Amazon Rekognition assigns to the face. # @return [String] # # @!attribute [rw] bounding_box # Bounding box of the face. # @return [Types::BoundingBox] # # @!attribute [rw] image_id # Unique identifier that Amazon Rekognition assigns to the input # image. # @return [String] # # @!attribute [rw] external_image_id # Identifier that you assign to all the faces in the input image. # @return [String] # # @!attribute [rw] confidence # Confidence level that the bounding box contains a face (and not a # different object such as a tree). # @return [Float] # class Face < Struct.new( :face_id, :bounding_box, :image_id, :external_image_id, :confidence) include Aws::Structure end # Structure containing attributes of the face that the algorithm # detected. # # A `FaceDetail` object contains either the default facial attributes or # all facial attributes. The default attributes are `BoundingBox`, # `Confidence`, `Landmarks`, `Pose`, and `Quality`. # # GetFaceDetection is the only Amazon Rekognition Video stored video # operation that can return a `FaceDetail` object with all attributes. # To specify which attributes to return, use the `FaceAttributes` input # parameter for StartFaceDetection. The following Amazon Rekognition # Video operations return only the default attributes. The corresponding # Start operations don't have a `FaceAttributes` input parameter. # # * GetCelebrityRecognition # # * GetPersonTracking # # * GetFaceSearch # # The Amazon Rekognition Image DetectFaces and IndexFaces operations can # return all facial attributes. To specify which attributes to return, # use the `Attributes` input parameter for `DetectFaces`. For # `IndexFaces`, use the `DetectAttributes` input parameter. # # @!attribute [rw] bounding_box # Bounding box of the face. Default attribute. # @return [Types::BoundingBox] # # @!attribute [rw] age_range # The estimated age range, in years, for the face. Low represents the # lowest estimated age and High represents the highest estimated age. # @return [Types::AgeRange] # # @!attribute [rw] smile # Indicates whether or not the face is smiling, and the confidence # level in the determination. # @return [Types::Smile] # # @!attribute [rw] eyeglasses # Indicates whether or not the face is wearing eye glasses, and the # confidence level in the determination. # @return [Types::Eyeglasses] # # @!attribute [rw] sunglasses # Indicates whether or not the face is wearing sunglasses, and the # confidence level in the determination. # @return [Types::Sunglasses] # # @!attribute [rw] gender # The predicted gender of a detected face. # @return [Types::Gender] # # @!attribute [rw] beard # Indicates whether or not the face has a beard, and the confidence # level in the determination. # @return [Types::Beard] # # @!attribute [rw] mustache # Indicates whether or not the face has a mustache, and the confidence # level in the determination. # @return [Types::Mustache] # # @!attribute [rw] eyes_open # Indicates whether or not the eyes on the face are open, and the # confidence level in the determination. # @return [Types::EyeOpen] # # @!attribute [rw] mouth_open # Indicates whether or not the mouth on the face is open, and the # confidence level in the determination. # @return [Types::MouthOpen] # # @!attribute [rw] emotions # The emotions that appear to be expressed on the face, and the # confidence level in the determination. The API is only making a # determination of the physical appearance of a person's face. It is # not a determination of the person’s internal emotional state and # should not be used in such a way. For example, a person pretending # to have a sad face might not be sad emotionally. # @return [Array] # # @!attribute [rw] landmarks # Indicates the location of landmarks on the face. Default attribute. # @return [Array] # # @!attribute [rw] pose # Indicates the pose of the face as determined by its pitch, roll, and # yaw. Default attribute. # @return [Types::Pose] # # @!attribute [rw] quality # Identifies image brightness and sharpness. Default attribute. # @return [Types::ImageQuality] # # @!attribute [rw] confidence # Confidence level that the bounding box contains a face (and not a # different object such as a tree). Default attribute. # @return [Float] # class FaceDetail < Struct.new( :bounding_box, :age_range, :smile, :eyeglasses, :sunglasses, :gender, :beard, :mustache, :eyes_open, :mouth_open, :emotions, :landmarks, :pose, :quality, :confidence) include Aws::Structure end # Information about a face detected in a video analysis request and the # time the face was detected in the video. # # @!attribute [rw] timestamp # Time, in milliseconds from the start of the video, that the face was # detected. # @return [Integer] # # @!attribute [rw] face # The face properties for the detected face. # @return [Types::FaceDetail] # class FaceDetection < Struct.new( :timestamp, :face) include Aws::Structure end # Provides face metadata. In addition, it also provides the confidence # in the match of this face with the input face. # # @!attribute [rw] similarity # Confidence in the match of this face with the input face. # @return [Float] # # @!attribute [rw] face # Describes the face properties such as the bounding box, face ID, # image ID of the source image, and external image ID that you # assigned. # @return [Types::Face] # class FaceMatch < Struct.new( :similarity, :face) include Aws::Structure end # Object containing both the face metadata (stored in the backend # database), and facial attributes that are detected but aren't stored # in the database. # # @!attribute [rw] face # Describes the face properties such as the bounding box, face ID, # image ID of the input image, and external image ID that you # assigned. # @return [Types::Face] # # @!attribute [rw] face_detail # Structure containing attributes of the face that the algorithm # detected. # @return [Types::FaceDetail] # class FaceRecord < Struct.new( :face, :face_detail) include Aws::Structure end # Input face recognition parameters for an Amazon Rekognition stream # processor. `FaceRecognitionSettings` is a request parameter for # CreateStreamProcessor. # # @note When making an API call, you may pass FaceSearchSettings # data as a hash: # # { # collection_id: "CollectionId", # face_match_threshold: 1.0, # } # # @!attribute [rw] collection_id # The ID of a collection that contains faces that you want to search # for. # @return [String] # # @!attribute [rw] face_match_threshold # Minimum face match confidence score that must be met to return a # result for a recognized face. Default is 80. 0 is the lowest # confidence. 100 is the highest confidence. # @return [Float] # class FaceSearchSettings < Struct.new( :collection_id, :face_match_threshold) include Aws::Structure end # The predicted gender of a detected face. # # Amazon Rekognition makes gender binary (male/female) predictions based # on the physical appearance of a face in a particular image. This kind # of prediction is not designed to categorize a person’s gender # identity, and you shouldn't use Amazon Rekognition to make such a # determination. For example, a male actor wearing a long-haired wig and # earrings for a role might be predicted as female. # # Using Amazon Rekognition to make gender binary predictions is best # suited for use cases where aggregate gender distribution statistics # need to be analyzed without identifying specific users. For example, # the percentage of female users compared to male users on a social # media platform. # # We don't recommend using gender binary predictions to make decisions # that impact
 an individual's rights, privacy, or access to services. # # @!attribute [rw] value # The predicted gender of the face. # @return [String] # # @!attribute [rw] confidence # Level of confidence in the prediction. # @return [Float] # class Gender < Struct.new( :value, :confidence) include Aws::Structure end # Information about where an object (DetectCustomLabels) or text # (DetectText) is located on an image. # # @!attribute [rw] bounding_box # An axis-aligned coarse representation of the detected item's # location on the image. # @return [Types::BoundingBox] # # @!attribute [rw] polygon # Within the bounding box, a fine-grained polygon around the detected # item. # @return [Array] # class Geometry < Struct.new( :bounding_box, :polygon) include Aws::Structure end # @note When making an API call, you may pass GetCelebrityInfoRequest # data as a hash: # # { # id: "RekognitionUniqueId", # required # } # # @!attribute [rw] id # The ID for the celebrity. You get the celebrity ID from a call to # the RecognizeCelebrities operation, which recognizes celebrities in # an image. # @return [String] # class GetCelebrityInfoRequest < Struct.new( :id) include Aws::Structure end # @!attribute [rw] urls # An array of URLs pointing to additional celebrity information. # @return [Array] # # @!attribute [rw] name # The name of the celebrity. # @return [String] # class GetCelebrityInfoResponse < Struct.new( :urls, :name) include Aws::Structure end # @note When making an API call, you may pass GetCelebrityRecognitionRequest # data as a hash: # # { # job_id: "JobId", # required # max_results: 1, # next_token: "PaginationToken", # sort_by: "ID", # accepts ID, TIMESTAMP # } # # @!attribute [rw] job_id # Job identifier for the required celebrity recognition analysis. You # can get the job identifer from a call to # `StartCelebrityRecognition`. # @return [String] # # @!attribute [rw] max_results # Maximum number of results to return per paginated call. The largest # value you can specify is 1000. If you specify a value greater than # 1000, a maximum of 1000 results is returned. The default value is # 1000. # @return [Integer] # # @!attribute [rw] next_token # If the previous response was incomplete (because there is more # recognized celebrities to retrieve), Amazon Rekognition Video # returns a pagination token in the response. You can use this # pagination token to retrieve the next set of celebrities. # @return [String] # # @!attribute [rw] sort_by # Sort to use for celebrities returned in `Celebrities` field. Specify # `ID` to sort by the celebrity identifier, specify `TIMESTAMP` to # sort by the time the celebrity was recognized. # @return [String] # class GetCelebrityRecognitionRequest < Struct.new( :job_id, :max_results, :next_token, :sort_by) include Aws::Structure end # @!attribute [rw] job_status # The current status of the celebrity recognition job. # @return [String] # # @!attribute [rw] status_message # If the job fails, `StatusMessage` provides a descriptive error # message. # @return [String] # # @!attribute [rw] video_metadata # Information about a video that Amazon Rekognition Video analyzed. # `Videometadata` is returned in every page of paginated responses # from a Amazon Rekognition Video operation. # @return [Types::VideoMetadata] # # @!attribute [rw] next_token # If the response is truncated, Amazon Rekognition Video returns this # token that you can use in the subsequent request to retrieve the # next set of celebrities. # @return [String] # # @!attribute [rw] celebrities # Array of celebrities recognized in the video. # @return [Array] # class GetCelebrityRecognitionResponse < Struct.new( :job_status, :status_message, :video_metadata, :next_token, :celebrities) include Aws::Structure end # @note When making an API call, you may pass GetContentModerationRequest # data as a hash: # # { # job_id: "JobId", # required # max_results: 1, # next_token: "PaginationToken", # sort_by: "NAME", # accepts NAME, TIMESTAMP # } # # @!attribute [rw] job_id # The identifier for the unsafe content job. Use `JobId` to identify # the job in a subsequent call to `GetContentModeration`. # @return [String] # # @!attribute [rw] max_results # Maximum number of results to return per paginated call. The largest # value you can specify is 1000. If you specify a value greater than # 1000, a maximum of 1000 results is returned. The default value is # 1000. # @return [Integer] # # @!attribute [rw] next_token # If the previous response was incomplete (because there is more data # to retrieve), Amazon Rekognition returns a pagination token in the # response. You can use this pagination token to retrieve the next set # of unsafe content labels. # @return [String] # # @!attribute [rw] sort_by # Sort to use for elements in the `ModerationLabelDetections` array. # Use `TIMESTAMP` to sort array elements by the time labels are # detected. Use `NAME` to alphabetically group elements for a label # together. Within each label group, the array element are sorted by # detection confidence. The default sort is by `TIMESTAMP`. # @return [String] # class GetContentModerationRequest < Struct.new( :job_id, :max_results, :next_token, :sort_by) include Aws::Structure end # @!attribute [rw] job_status # The current status of the unsafe content analysis job. # @return [String] # # @!attribute [rw] status_message # If the job fails, `StatusMessage` provides a descriptive error # message. # @return [String] # # @!attribute [rw] video_metadata # Information about a video that Amazon Rekognition analyzed. # `Videometadata` is returned in every page of paginated responses # from `GetContentModeration`. # @return [Types::VideoMetadata] # # @!attribute [rw] moderation_labels # The detected unsafe content labels and the time(s) they were # detected. # @return [Array] # # @!attribute [rw] next_token # If the response is truncated, Amazon Rekognition Video returns this # token that you can use in the subsequent request to retrieve the # next set of unsafe content labels. # @return [String] # # @!attribute [rw] moderation_model_version # Version number of the moderation detection model that was used to # detect unsafe content. # @return [String] # class GetContentModerationResponse < Struct.new( :job_status, :status_message, :video_metadata, :moderation_labels, :next_token, :moderation_model_version) include Aws::Structure end # @note When making an API call, you may pass GetFaceDetectionRequest # data as a hash: # # { # job_id: "JobId", # required # max_results: 1, # next_token: "PaginationToken", # } # # @!attribute [rw] job_id # Unique identifier for the face detection job. The `JobId` is # returned from `StartFaceDetection`. # @return [String] # # @!attribute [rw] max_results # Maximum number of results to return per paginated call. The largest # value you can specify is 1000. If you specify a value greater than # 1000, a maximum of 1000 results is returned. The default value is # 1000. # @return [Integer] # # @!attribute [rw] next_token # If the previous response was incomplete (because there are more # faces to retrieve), Amazon Rekognition Video returns a pagination # token in the response. You can use this pagination token to retrieve # the next set of faces. # @return [String] # class GetFaceDetectionRequest < Struct.new( :job_id, :max_results, :next_token) include Aws::Structure end # @!attribute [rw] job_status # The current status of the face detection job. # @return [String] # # @!attribute [rw] status_message # If the job fails, `StatusMessage` provides a descriptive error # message. # @return [String] # # @!attribute [rw] video_metadata # Information about a video that Amazon Rekognition Video analyzed. # `Videometadata` is returned in every page of paginated responses # from a Amazon Rekognition video operation. # @return [Types::VideoMetadata] # # @!attribute [rw] next_token # If the response is truncated, Amazon Rekognition returns this token # that you can use in the subsequent request to retrieve the next set # of faces. # @return [String] # # @!attribute [rw] faces # An array of faces detected in the video. Each element contains a # detected face's details and the time, in milliseconds from the # start of the video, the face was detected. # @return [Array] # class GetFaceDetectionResponse < Struct.new( :job_status, :status_message, :video_metadata, :next_token, :faces) include Aws::Structure end # @note When making an API call, you may pass GetFaceSearchRequest # data as a hash: # # { # job_id: "JobId", # required # max_results: 1, # next_token: "PaginationToken", # sort_by: "INDEX", # accepts INDEX, TIMESTAMP # } # # @!attribute [rw] job_id # The job identifer for the search request. You get the job identifier # from an initial call to `StartFaceSearch`. # @return [String] # # @!attribute [rw] max_results # Maximum number of results to return per paginated call. The largest # value you can specify is 1000. If you specify a value greater than # 1000, a maximum of 1000 results is returned. The default value is # 1000. # @return [Integer] # # @!attribute [rw] next_token # If the previous response was incomplete (because there is more # search results to retrieve), Amazon Rekognition Video returns a # pagination token in the response. You can use this pagination token # to retrieve the next set of search results. # @return [String] # # @!attribute [rw] sort_by # Sort to use for grouping faces in the response. Use `TIMESTAMP` to # group faces by the time that they are recognized. Use `INDEX` to # sort by recognized faces. # @return [String] # class GetFaceSearchRequest < Struct.new( :job_id, :max_results, :next_token, :sort_by) include Aws::Structure end # @!attribute [rw] job_status # The current status of the face search job. # @return [String] # # @!attribute [rw] status_message # If the job fails, `StatusMessage` provides a descriptive error # message. # @return [String] # # @!attribute [rw] next_token # If the response is truncated, Amazon Rekognition Video returns this # token that you can use in the subsequent request to retrieve the # next set of search results. # @return [String] # # @!attribute [rw] video_metadata # Information about a video that Amazon Rekognition analyzed. # `Videometadata` is returned in every page of paginated responses # from a Amazon Rekognition Video operation. # @return [Types::VideoMetadata] # # @!attribute [rw] persons # An array of persons, PersonMatch, in the video whose face(s) match # the face(s) in an Amazon Rekognition collection. It also includes # time information for when persons are matched in the video. You # specify the input collection in an initial call to # `StartFaceSearch`. Each `Persons` element includes a time the person # was matched, face match details (`FaceMatches`) for matching faces # in the collection, and person information (`Person`) for the matched # person. # @return [Array] # class GetFaceSearchResponse < Struct.new( :job_status, :status_message, :next_token, :video_metadata, :persons) include Aws::Structure end # @note When making an API call, you may pass GetLabelDetectionRequest # data as a hash: # # { # job_id: "JobId", # required # max_results: 1, # next_token: "PaginationToken", # sort_by: "NAME", # accepts NAME, TIMESTAMP # } # # @!attribute [rw] job_id # Job identifier for the label detection operation for which you want # results returned. You get the job identifer from an initial call to # `StartlabelDetection`. # @return [String] # # @!attribute [rw] max_results # Maximum number of results to return per paginated call. The largest # value you can specify is 1000. If you specify a value greater than # 1000, a maximum of 1000 results is returned. The default value is # 1000. # @return [Integer] # # @!attribute [rw] next_token # If the previous response was incomplete (because there are more # labels to retrieve), Amazon Rekognition Video returns a pagination # token in the response. You can use this pagination token to retrieve # the next set of labels. # @return [String] # # @!attribute [rw] sort_by # Sort to use for elements in the `Labels` array. Use `TIMESTAMP` to # sort array elements by the time labels are detected. Use `NAME` to # alphabetically group elements for a label together. Within each # label group, the array element are sorted by detection confidence. # The default sort is by `TIMESTAMP`. # @return [String] # class GetLabelDetectionRequest < Struct.new( :job_id, :max_results, :next_token, :sort_by) include Aws::Structure end # @!attribute [rw] job_status # The current status of the label detection job. # @return [String] # # @!attribute [rw] status_message # If the job fails, `StatusMessage` provides a descriptive error # message. # @return [String] # # @!attribute [rw] video_metadata # Information about a video that Amazon Rekognition Video analyzed. # `Videometadata` is returned in every page of paginated responses # from a Amazon Rekognition video operation. # @return [Types::VideoMetadata] # # @!attribute [rw] next_token # If the response is truncated, Amazon Rekognition Video returns this # token that you can use in the subsequent request to retrieve the # next set of labels. # @return [String] # # @!attribute [rw] labels # An array of labels detected in the video. Each element contains the # detected label and the time, in milliseconds from the start of the # video, that the label was detected. # @return [Array] # # @!attribute [rw] label_model_version # Version number of the label detection model that was used to detect # labels. # @return [String] # class GetLabelDetectionResponse < Struct.new( :job_status, :status_message, :video_metadata, :next_token, :labels, :label_model_version) include Aws::Structure end # @note When making an API call, you may pass GetPersonTrackingRequest # data as a hash: # # { # job_id: "JobId", # required # max_results: 1, # next_token: "PaginationToken", # sort_by: "INDEX", # accepts INDEX, TIMESTAMP # } # # @!attribute [rw] job_id # The identifier for a job that tracks persons in a video. You get the # `JobId` from a call to `StartPersonTracking`. # @return [String] # # @!attribute [rw] max_results # Maximum number of results to return per paginated call. The largest # value you can specify is 1000. If you specify a value greater than # 1000, a maximum of 1000 results is returned. The default value is # 1000. # @return [Integer] # # @!attribute [rw] next_token # If the previous response was incomplete (because there are more # persons to retrieve), Amazon Rekognition Video returns a pagination # token in the response. You can use this pagination token to retrieve # the next set of persons. # @return [String] # # @!attribute [rw] sort_by # Sort to use for elements in the `Persons` array. Use `TIMESTAMP` to # sort array elements by the time persons are detected. Use `INDEX` to # sort by the tracked persons. If you sort by `INDEX`, the array # elements for each person are sorted by detection confidence. The # default sort is by `TIMESTAMP`. # @return [String] # class GetPersonTrackingRequest < Struct.new( :job_id, :max_results, :next_token, :sort_by) include Aws::Structure end # @!attribute [rw] job_status # The current status of the person tracking job. # @return [String] # # @!attribute [rw] status_message # If the job fails, `StatusMessage` provides a descriptive error # message. # @return [String] # # @!attribute [rw] video_metadata # Information about a video that Amazon Rekognition Video analyzed. # `Videometadata` is returned in every page of paginated responses # from a Amazon Rekognition Video operation. # @return [Types::VideoMetadata] # # @!attribute [rw] next_token # If the response is truncated, Amazon Rekognition Video returns this # token that you can use in the subsequent request to retrieve the # next set of persons. # @return [String] # # @!attribute [rw] persons # An array of the persons detected in the video and the time(s) their # path was tracked throughout the video. An array element will exist # for each time a person's path is tracked. # @return [Array] # class GetPersonTrackingResponse < Struct.new( :job_status, :status_message, :video_metadata, :next_token, :persons) include Aws::Structure end # @note When making an API call, you may pass GetTextDetectionRequest # data as a hash: # # { # job_id: "JobId", # required # max_results: 1, # next_token: "PaginationToken", # } # # @!attribute [rw] job_id # Job identifier for the label detection operation for which you want # results returned. You get the job identifer from an initial call to # `StartTextDetection`. # @return [String] # # @!attribute [rw] max_results # Maximum number of results to return per paginated call. The largest # value you can specify is 1000. # @return [Integer] # # @!attribute [rw] next_token # If the previous response was incomplete (because there are more # labels to retrieve), Amazon Rekognition Video returns a pagination # token in the response. You can use this pagination token to retrieve # the next set of text. # @return [String] # class GetTextDetectionRequest < Struct.new( :job_id, :max_results, :next_token) include Aws::Structure end # @!attribute [rw] job_status # Current status of the text detection job. # @return [String] # # @!attribute [rw] status_message # If the job fails, `StatusMessage` provides a descriptive error # message. # @return [String] # # @!attribute [rw] video_metadata # Information about a video that Amazon Rekognition analyzed. # `Videometadata` is returned in every page of paginated responses # from a Amazon Rekognition video operation. # @return [Types::VideoMetadata] # # @!attribute [rw] text_detections # An array of text detected in the video. Each element contains the # detected text, the time in milliseconds from the start of the video # that the text was detected, and where it was detected on the screen. # @return [Array] # # @!attribute [rw] next_token # If the response is truncated, Amazon Rekognition Video returns this # token that you can use in the subsequent request to retrieve the # next set of text. # @return [String] # # @!attribute [rw] text_model_version # Version number of the text detection model that was used to detect # text. # @return [String] # class GetTextDetectionResponse < Struct.new( :job_status, :status_message, :video_metadata, :text_detections, :next_token, :text_model_version) include Aws::Structure end # The S3 bucket that contains the Ground Truth manifest file. # # @note When making an API call, you may pass GroundTruthManifest # data as a hash: # # { # s3_object: { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # }, # } # # @!attribute [rw] s3_object # Provides the S3 bucket name and object name. # # The region for the S3 bucket containing the S3 object must match the # region you use for Amazon Rekognition operations. # # For Amazon Rekognition to process an S3 object, the user must have # permission to access the S3 object. For more information, see # Resource-Based Policies in the Amazon Rekognition Developer Guide. # @return [Types::S3Object] # class GroundTruthManifest < Struct.new( :s3_object) include Aws::Structure end # Shows the results of the human in the loop evaluation. If there is no # HumanLoopArn, the input did not trigger human review. # # @!attribute [rw] human_loop_arn # The Amazon Resource Name (ARN) of the HumanLoop created. # @return [String] # # @!attribute [rw] human_loop_activation_reasons # Shows if and why human review was needed. # @return [Array] # # @!attribute [rw] human_loop_activation_conditions_evaluation_results # Shows the result of condition evaluations, including those # conditions which activated a human review. # @return [String] # class HumanLoopActivationOutput < Struct.new( :human_loop_arn, :human_loop_activation_reasons, :human_loop_activation_conditions_evaluation_results) include Aws::Structure end # Sets up the flow definition the image will be sent to if one of the # conditions is met. You can also set certain attributes of the image # before review. # # @note When making an API call, you may pass HumanLoopConfig # data as a hash: # # { # human_loop_name: "HumanLoopName", # required # flow_definition_arn: "FlowDefinitionArn", # required # data_attributes: { # content_classifiers: ["FreeOfPersonallyIdentifiableInformation"], # accepts FreeOfPersonallyIdentifiableInformation, FreeOfAdultContent # }, # } # # @!attribute [rw] human_loop_name # The name of the human review used for this image. This should be # kept unique within a region. # @return [String] # # @!attribute [rw] flow_definition_arn # The Amazon Resource Name (ARN) of the flow definition. # @return [String] # # @!attribute [rw] data_attributes # Sets attributes of the input data. # @return [Types::HumanLoopDataAttributes] # class HumanLoopConfig < Struct.new( :human_loop_name, :flow_definition_arn, :data_attributes) include Aws::Structure end # Allows you to set attributes of the image. Currently, you can declare # an image as free of personally identifiable information. # # @note When making an API call, you may pass HumanLoopDataAttributes # data as a hash: # # { # content_classifiers: ["FreeOfPersonallyIdentifiableInformation"], # accepts FreeOfPersonallyIdentifiableInformation, FreeOfAdultContent # } # # @!attribute [rw] content_classifiers # Sets whether the input image is free of personally identifiable # information. # @return [Array] # class HumanLoopDataAttributes < Struct.new( :content_classifiers) include Aws::Structure end # The number of in-progress human reviews you have has exceeded the # number allowed. # # @!attribute [rw] resource_type # @return [String] # # @!attribute [rw] quota_code # @return [String] # # @!attribute [rw] service_code # @return [String] # class HumanLoopQuotaExceededException < Struct.new( :resource_type, :quota_code, :service_code) include Aws::Structure end # A `ClientRequestToken` input parameter was reused with an operation, # but at least one of the other input parameters is different from the # previous call to the operation. # class IdempotentParameterMismatchException < Aws::EmptyStructure; end # Provides the input image either as bytes or an S3 object. # # You pass image bytes to an Amazon Rekognition API operation by using # the `Bytes` property. For example, you would use the `Bytes` property # to pass an image loaded from a local file system. Image bytes passed # by using the `Bytes` property must be base64-encoded. Your code may # not need to encode image bytes if you are using an AWS SDK to call # Amazon Rekognition API operations. # # For more information, see Analyzing an Image Loaded from a Local File # System in the Amazon Rekognition Developer Guide. # # You pass images stored in an S3 bucket to an Amazon Rekognition API # operation by using the `S3Object` property. Images stored in an S3 # bucket do not need to be base64-encoded. # # The region for the S3 bucket containing the S3 object must match the # region you use for Amazon Rekognition operations. # # If you use the AWS CLI to call Amazon Rekognition operations, passing # image bytes using the Bytes property is not supported. You must first # upload the image to an Amazon S3 bucket and then call the operation # using the S3Object property. # # For Amazon Rekognition to process an S3 object, the user must have # permission to access the S3 object. For more information, see Resource # Based Policies in the Amazon Rekognition Developer Guide. # # @note When making an API call, you may pass Image # data as a hash: # # { # bytes: "data", # s3_object: { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # }, # } # # @!attribute [rw] bytes # Blob of image bytes up to 5 MBs. # @return [String] # # @!attribute [rw] s3_object # Identifies an S3 object as the image source. # @return [Types::S3Object] # class Image < Struct.new( :bytes, :s3_object) include Aws::Structure end # Identifies face image brightness and sharpness. # # @!attribute [rw] brightness # Value representing brightness of the face. The service returns a # value between 0 and 100 (inclusive). A higher value indicates a # brighter face image. # @return [Float] # # @!attribute [rw] sharpness # Value representing sharpness of the face. The service returns a # value between 0 and 100 (inclusive). A higher value indicates a # sharper face image. # @return [Float] # class ImageQuality < Struct.new( :brightness, :sharpness) include Aws::Structure end # The input image size exceeds the allowed limit. For more information, # see Limits in Amazon Rekognition in the Amazon Rekognition Developer # Guide. # class ImageTooLargeException < Aws::EmptyStructure; end # @note When making an API call, you may pass IndexFacesRequest # data as a hash: # # { # collection_id: "CollectionId", # required # image: { # required # bytes: "data", # s3_object: { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # }, # }, # external_image_id: "ExternalImageId", # detection_attributes: ["DEFAULT"], # accepts DEFAULT, ALL # max_faces: 1, # quality_filter: "NONE", # accepts NONE, AUTO, LOW, MEDIUM, HIGH # } # # @!attribute [rw] collection_id # The ID of an existing collection to which you want to add the faces # that are detected in the input images. # @return [String] # # @!attribute [rw] image # The input image as base64-encoded bytes or an S3 object. If you use # the AWS CLI to call Amazon Rekognition operations, passing # base64-encoded image bytes isn't supported. # # If you are using an AWS SDK to call Amazon Rekognition, you might # not need to base64-encode image bytes passed using the `Bytes` # field. For more information, see Images in the Amazon Rekognition # developer guide. # @return [Types::Image] # # @!attribute [rw] external_image_id # The ID you want to assign to all the faces detected in the image. # @return [String] # # @!attribute [rw] detection_attributes # An array of facial attributes that you want to be returned. This can # be the default list of attributes or all attributes. If you don't # specify a value for `Attributes` or if you specify `["DEFAULT"]`, # the API returns the following subset of facial attributes: # `BoundingBox`, `Confidence`, `Pose`, `Quality`, and `Landmarks`. If # you provide `["ALL"]`, all facial attributes are returned, but the # operation takes longer to complete. # # If you provide both, `["ALL", "DEFAULT"]`, the service uses a # logical AND operator to determine which attributes to return (in # this case, all attributes). # @return [Array] # # @!attribute [rw] max_faces # The maximum number of faces to index. The value of `MaxFaces` must # be greater than or equal to 1. `IndexFaces` returns no more than 100 # detected faces in an image, even if you specify a larger value for # `MaxFaces`. # # If `IndexFaces` detects more faces than the value of `MaxFaces`, the # faces with the lowest quality are filtered out first. If there are # still more faces than the value of `MaxFaces`, the faces with the # smallest bounding boxes are filtered out (up to the number that's # needed to satisfy the value of `MaxFaces`). Information about the # unindexed faces is available in the `UnindexedFaces` array. # # The faces that are returned by `IndexFaces` are sorted by the # largest face bounding box size to the smallest size, in descending # order. # # `MaxFaces` can be used with a collection associated with any version # of the face model. # @return [Integer] # # @!attribute [rw] quality_filter # A filter that specifies a quality bar for how much filtering is done # to identify faces. Filtered faces aren't indexed. If you specify # `AUTO`, Amazon Rekognition chooses the quality bar. If you specify # `LOW`, `MEDIUM`, or `HIGH`, filtering removes all faces that don’t # meet the chosen quality bar. The default value is `AUTO`. The # quality bar is based on a variety of common use cases. Low-quality # detections can occur for a number of reasons. Some examples are an # object that's misidentified as a face, a face that's too blurry, # or a face with a pose that's too extreme to use. If you specify # `NONE`, no filtering is performed. # # To use quality filtering, the collection you are using must be # associated with version 3 of the face model or higher. # @return [String] # class IndexFacesRequest < Struct.new( :collection_id, :image, :external_image_id, :detection_attributes, :max_faces, :quality_filter) include Aws::Structure end # @!attribute [rw] face_records # An array of faces detected and added to the collection. For more # information, see Searching Faces in a Collection in the Amazon # Rekognition Developer Guide. # @return [Array] # # @!attribute [rw] orientation_correction # If your collection is associated with a face detection model that's # later than version 3.0, the value of `OrientationCorrection` is # always null and no orientation information is returned. # # If your collection is associated with a face detection model that's # version 3.0 or earlier, the following applies: # # * If the input image is in .jpeg format, it might contain # exchangeable image file format (Exif) metadata that includes the # image's orientation. Amazon Rekognition uses this orientation # information to perform image correction - the bounding box # coordinates are translated to represent object locations after the # orientation information in the Exif metadata is used to correct # the image orientation. Images in .png format don't contain Exif # metadata. The value of `OrientationCorrection` is null. # # * If the image doesn't contain orientation information in its Exif # metadata, Amazon Rekognition returns an estimated orientation # (ROTATE\_0, ROTATE\_90, ROTATE\_180, ROTATE\_270). Amazon # Rekognition doesn’t perform image correction for images. The # bounding box coordinates aren't translated and represent the # object locations before the image is rotated. # # Bounding box information is returned in the `FaceRecords` array. You # can get the version of the face detection model by calling # DescribeCollection. # @return [String] # # @!attribute [rw] face_model_version # The version number of the face detection model that's associated # with the input collection (`CollectionId`). # @return [String] # # @!attribute [rw] unindexed_faces # An array of faces that were detected in the image but weren't # indexed. They weren't indexed because the quality filter identified # them as low quality, or the `MaxFaces` request parameter filtered # them out. To use the quality filter, you specify the `QualityFilter` # request parameter. # @return [Array] # class IndexFacesResponse < Struct.new( :face_records, :orientation_correction, :face_model_version, :unindexed_faces) include Aws::Structure end # An instance of a label returned by Amazon Rekognition Image # (DetectLabels) or by Amazon Rekognition Video (GetLabelDetection). # # @!attribute [rw] bounding_box # The position of the label instance on the image. # @return [Types::BoundingBox] # # @!attribute [rw] confidence # The confidence that Amazon Rekognition has in the accuracy of the # bounding box. # @return [Float] # class Instance < Struct.new( :bounding_box, :confidence) include Aws::Structure end # Amazon Rekognition experienced a service issue. Try your call again. # class InternalServerError < Aws::EmptyStructure; end # The provided image format is not supported. # class InvalidImageFormatException < Aws::EmptyStructure; end # Pagination token in the request is not valid. # class InvalidPaginationTokenException < Aws::EmptyStructure; end # Input parameter violated a constraint. Validate your parameter before # calling the API operation again. # class InvalidParameterException < Aws::EmptyStructure; end # Amazon Rekognition is unable to access the S3 object specified in the # request. # class InvalidS3ObjectException < Aws::EmptyStructure; end # The Kinesis data stream Amazon Rekognition to which the analysis # results of a Amazon Rekognition stream processor are streamed. For # more information, see CreateStreamProcessor in the Amazon Rekognition # Developer Guide. # # @note When making an API call, you may pass KinesisDataStream # data as a hash: # # { # arn: "KinesisDataArn", # } # # @!attribute [rw] arn # ARN of the output Amazon Kinesis Data Streams stream. # @return [String] # class KinesisDataStream < Struct.new( :arn) include Aws::Structure end # Kinesis video stream stream that provides the source streaming video # for a Amazon Rekognition Video stream processor. For more information, # see CreateStreamProcessor in the Amazon Rekognition Developer Guide. # # @note When making an API call, you may pass KinesisVideoStream # data as a hash: # # { # arn: "KinesisVideoArn", # } # # @!attribute [rw] arn # ARN of the Kinesis video stream stream that streams the source # video. # @return [String] # class KinesisVideoStream < Struct.new( :arn) include Aws::Structure end # Structure containing details about the detected label, including the # name, detected instances, parent labels, and level of confidence. # # @!attribute [rw] name # The name (label) of the object or scene. # @return [String] # # @!attribute [rw] confidence # Level of confidence. # @return [Float] # # @!attribute [rw] instances # If `Label` represents an object, `Instances` contains the bounding # boxes for each instance of the detected object. Bounding boxes are # returned for common object labels such as people, cars, furniture, # apparel or pets. # @return [Array] # # @!attribute [rw] parents # The parent labels for a label. The response includes all ancestor # labels. # @return [Array] # class Label < Struct.new( :name, :confidence, :instances, :parents) include Aws::Structure end # Information about a label detected in a video analysis request and the # time the label was detected in the video. # # @!attribute [rw] timestamp # Time, in milliseconds from the start of the video, that the label # was detected. # @return [Integer] # # @!attribute [rw] label # Details about the detected label. # @return [Types::Label] # class LabelDetection < Struct.new( :timestamp, :label) include Aws::Structure end # Indicates the location of the landmark on the face. # # @!attribute [rw] type # Type of landmark. # @return [String] # # @!attribute [rw] x # The x-coordinate from the top left of the landmark expressed as the # ratio of the width of the image. For example, if the image is 700 x # 200 and the x-coordinate of the landmark is at 350 pixels, this # value is 0.5. # @return [Float] # # @!attribute [rw] y # The y-coordinate from the top left of the landmark expressed as the # ratio of the height of the image. For example, if the image is 700 x # 200 and the y-coordinate of the landmark is at 100 pixels, this # value is 0.5. # @return [Float] # class Landmark < Struct.new( :type, :x, :y) include Aws::Structure end # An Amazon Rekognition service limit was exceeded. For example, if you # start too many Amazon Rekognition Video jobs concurrently, calls to # start operations (`StartLabelDetection`, for example) will raise a # `LimitExceededException` exception (HTTP status code: 400) until the # number of concurrently running jobs is below the Amazon Rekognition # service limit. # class LimitExceededException < Aws::EmptyStructure; end # @note When making an API call, you may pass ListCollectionsRequest # data as a hash: # # { # next_token: "PaginationToken", # max_results: 1, # } # # @!attribute [rw] next_token # Pagination token from the previous response. # @return [String] # # @!attribute [rw] max_results # Maximum number of collection IDs to return. # @return [Integer] # class ListCollectionsRequest < Struct.new( :next_token, :max_results) include Aws::Structure end # @!attribute [rw] collection_ids # An array of collection IDs. # @return [Array] # # @!attribute [rw] next_token # If the result is truncated, the response provides a `NextToken` that # you can use in the subsequent request to fetch the next set of # collection IDs. # @return [String] # # @!attribute [rw] face_model_versions # Version numbers of the face detection models associated with the # collections in the array `CollectionIds`. For example, the value of # `FaceModelVersions[2]` is the version number for the face detection # model used by the collection in `CollectionId[2]`. # @return [Array] # class ListCollectionsResponse < Struct.new( :collection_ids, :next_token, :face_model_versions) include Aws::Structure end # @note When making an API call, you may pass ListFacesRequest # data as a hash: # # { # collection_id: "CollectionId", # required # next_token: "PaginationToken", # max_results: 1, # } # # @!attribute [rw] collection_id # ID of the collection from which to list the faces. # @return [String] # # @!attribute [rw] next_token # If the previous response was incomplete (because there is more data # to retrieve), Amazon Rekognition returns a pagination token in the # response. You can use this pagination token to retrieve the next set # of faces. # @return [String] # # @!attribute [rw] max_results # Maximum number of faces to return. # @return [Integer] # class ListFacesRequest < Struct.new( :collection_id, :next_token, :max_results) include Aws::Structure end # @!attribute [rw] faces # An array of `Face` objects. # @return [Array] # # @!attribute [rw] next_token # If the response is truncated, Amazon Rekognition returns this token # that you can use in the subsequent request to retrieve the next set # of faces. # @return [String] # # @!attribute [rw] face_model_version # Version number of the face detection model associated with the input # collection (`CollectionId`). # @return [String] # class ListFacesResponse < Struct.new( :faces, :next_token, :face_model_version) include Aws::Structure end # @note When making an API call, you may pass ListStreamProcessorsRequest # data as a hash: # # { # next_token: "PaginationToken", # max_results: 1, # } # # @!attribute [rw] next_token # If the previous response was incomplete (because there are more # stream processors to retrieve), Amazon Rekognition Video returns a # pagination token in the response. You can use this pagination token # to retrieve the next set of stream processors. # @return [String] # # @!attribute [rw] max_results # Maximum number of stream processors you want Amazon Rekognition # Video to return in the response. The default is 1000. # @return [Integer] # class ListStreamProcessorsRequest < Struct.new( :next_token, :max_results) include Aws::Structure end # @!attribute [rw] next_token # If the response is truncated, Amazon Rekognition Video returns this # token that you can use in the subsequent request to retrieve the # next set of stream processors. # @return [String] # # @!attribute [rw] stream_processors # List of stream processors that you have created. # @return [Array] # class ListStreamProcessorsResponse < Struct.new( :next_token, :stream_processors) include Aws::Structure end # Provides information about a single type of unsafe content found in an # image or video. Each type of moderated content has a label within a # hierarchical taxonomy. For more information, see Detecting Unsafe # Content in the Amazon Rekognition Developer Guide. # # @!attribute [rw] confidence # Specifies the confidence that Amazon Rekognition has that the label # has been correctly identified. # # If you don't specify the `MinConfidence` parameter in the call to # `DetectModerationLabels`, the operation returns labels with a # confidence value greater than or equal to 50 percent. # @return [Float] # # @!attribute [rw] name # The label name for the type of unsafe content detected in the image. # @return [String] # # @!attribute [rw] parent_name # The name for the parent label. Labels at the top level of the # hierarchy have the parent label `""`. # @return [String] # class ModerationLabel < Struct.new( :confidence, :name, :parent_name) include Aws::Structure end # Indicates whether or not the mouth on the face is open, and the # confidence level in the determination. # # @!attribute [rw] value # Boolean value that indicates whether the mouth on the face is open # or not. # @return [Boolean] # # @!attribute [rw] confidence # Level of confidence in the determination. # @return [Float] # class MouthOpen < Struct.new( :value, :confidence) include Aws::Structure end # Indicates whether or not the face has a mustache, and the confidence # level in the determination. # # @!attribute [rw] value # Boolean value that indicates whether the face has mustache or not. # @return [Boolean] # # @!attribute [rw] confidence # Level of confidence in the determination. # @return [Float] # class Mustache < Struct.new( :value, :confidence) include Aws::Structure end # The Amazon Simple Notification Service topic to which Amazon # Rekognition publishes the completion status of a video analysis # operation. For more information, see api-video. # # @note When making an API call, you may pass NotificationChannel # data as a hash: # # { # sns_topic_arn: "SNSTopicArn", # required # role_arn: "RoleArn", # required # } # # @!attribute [rw] sns_topic_arn # The Amazon SNS topic to which Amazon Rekognition to posts the # completion status. # @return [String] # # @!attribute [rw] role_arn # The ARN of an IAM role that gives Amazon Rekognition publishing # permissions to the Amazon SNS topic. # @return [String] # class NotificationChannel < Struct.new( :sns_topic_arn, :role_arn) include Aws::Structure end # The S3 bucket and folder location where training output is placed. # # @note When making an API call, you may pass OutputConfig # data as a hash: # # { # s3_bucket: "S3Bucket", # s3_key_prefix: "S3KeyPrefix", # } # # @!attribute [rw] s3_bucket # The S3 bucket where training output is placed. # @return [String] # # @!attribute [rw] s3_key_prefix # The prefix applied to the training output files. # @return [String] # class OutputConfig < Struct.new( :s3_bucket, :s3_key_prefix) include Aws::Structure end # A parent label for a label. A label can have 0, 1, or more parents. # # @!attribute [rw] name # The name of the parent label. # @return [String] # class Parent < Struct.new( :name) include Aws::Structure end # Details about a person detected in a video analysis request. # # @!attribute [rw] index # Identifier for the person detected person within a video. Use to # keep track of the person throughout the video. The identifier is not # stored by Amazon Rekognition. # @return [Integer] # # @!attribute [rw] bounding_box # Bounding box around the detected person. # @return [Types::BoundingBox] # # @!attribute [rw] face # Face details for the detected person. # @return [Types::FaceDetail] # class PersonDetail < Struct.new( :index, :bounding_box, :face) include Aws::Structure end # Details and path tracking information for a single time a person's # path is tracked in a video. Amazon Rekognition operations that track # people's paths return an array of `PersonDetection` objects with # elements for each time a person's path is tracked in a video. # # For more information, see GetPersonTracking in the Amazon Rekognition # Developer Guide. # # @!attribute [rw] timestamp # The time, in milliseconds from the start of the video, that the # person's path was tracked. # @return [Integer] # # @!attribute [rw] person # Details about a person whose path was tracked in a video. # @return [Types::PersonDetail] # class PersonDetection < Struct.new( :timestamp, :person) include Aws::Structure end # Information about a person whose face matches a face(s) in an Amazon # Rekognition collection. Includes information about the faces in the # Amazon Rekognition collection (FaceMatch), information about the # person (PersonDetail), and the time stamp for when the person was # detected in a video. An array of `PersonMatch` objects is returned by # GetFaceSearch. # # @!attribute [rw] timestamp # The time, in milliseconds from the beginning of the video, that the # person was matched in the video. # @return [Integer] # # @!attribute [rw] person # Information about the matched person. # @return [Types::PersonDetail] # # @!attribute [rw] face_matches # Information about the faces in the input collection that match the # face of a person in the video. # @return [Array] # class PersonMatch < Struct.new( :timestamp, :person, :face_matches) include Aws::Structure end # The X and Y coordinates of a point on an image. The X and Y values # returned are ratios of the overall image size. For example, if the # input image is 700x200 and the operation returns X=0.5 and Y=0.25, # then the point is at the (350,50) pixel coordinate on the image. # # An array of `Point` objects, `Polygon`, is returned by DetectText and # by DetectCustomLabels. `Polygon` represents a fine-grained polygon # around a detected item. For more information, see Geometry in the # Amazon Rekognition Developer Guide. # # @!attribute [rw] x # The value of the X coordinate for a point on a `Polygon`. # @return [Float] # # @!attribute [rw] y # The value of the Y coordinate for a point on a `Polygon`. # @return [Float] # class Point < Struct.new( :x, :y) include Aws::Structure end # Indicates the pose of the face as determined by its pitch, roll, and # yaw. # # @!attribute [rw] roll # Value representing the face rotation on the roll axis. # @return [Float] # # @!attribute [rw] yaw # Value representing the face rotation on the yaw axis. # @return [Float] # # @!attribute [rw] pitch # Value representing the face rotation on the pitch axis. # @return [Float] # class Pose < Struct.new( :roll, :yaw, :pitch) include Aws::Structure end # A description of a Amazon Rekognition Custom Labels project. # # @!attribute [rw] project_arn # The Amazon Resource Name (ARN) of the project. # @return [String] # # @!attribute [rw] creation_timestamp # The Unix timestamp for the date and time that the project was # created. # @return [Time] # # @!attribute [rw] status # The current status of the project. # @return [String] # class ProjectDescription < Struct.new( :project_arn, :creation_timestamp, :status) include Aws::Structure end # The description of a version of a model. # # @!attribute [rw] project_version_arn # The Amazon Resource Name (ARN) of the model version. # @return [String] # # @!attribute [rw] creation_timestamp # The Unix datetime for the date and time that training started. # @return [Time] # # @!attribute [rw] min_inference_units # The minimum number of inference units used by the model. For more # information, see StartProjectVersion. # @return [Integer] # # @!attribute [rw] status # The current status of the model version. # @return [String] # # @!attribute [rw] status_message # A descriptive message for an error or warning that occurred. # @return [String] # # @!attribute [rw] billable_training_time_in_seconds # The duration, in seconds, that the model version has been billed for # training. This value is only returned if the model version has been # successfully trained. # @return [Integer] # # @!attribute [rw] training_end_timestamp # The Unix date and time that training of the model ended. # @return [Time] # # @!attribute [rw] output_config # The location where training results are saved. # @return [Types::OutputConfig] # # @!attribute [rw] training_data_result # The manifest file that represents the training results. # @return [Types::TrainingDataResult] # # @!attribute [rw] testing_data_result # The manifest file that represents the testing results. # @return [Types::TestingDataResult] # # @!attribute [rw] evaluation_result # The training results. `EvaluationResult` is only returned if # training is successful. # @return [Types::EvaluationResult] # class ProjectVersionDescription < Struct.new( :project_version_arn, :creation_timestamp, :min_inference_units, :status, :status_message, :billable_training_time_in_seconds, :training_end_timestamp, :output_config, :training_data_result, :testing_data_result, :evaluation_result) include Aws::Structure end # The number of requests exceeded your throughput limit. If you want to # increase this limit, contact Amazon Rekognition. # class ProvisionedThroughputExceededException < Aws::EmptyStructure; end # @note When making an API call, you may pass RecognizeCelebritiesRequest # data as a hash: # # { # image: { # required # bytes: "data", # s3_object: { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # }, # }, # } # # @!attribute [rw] image # The input image as base64-encoded bytes or an S3 object. If you use # the AWS CLI to call Amazon Rekognition operations, passing # base64-encoded image bytes is not supported. # # If you are using an AWS SDK to call Amazon Rekognition, you might # not need to base64-encode image bytes passed using the `Bytes` # field. For more information, see Images in the Amazon Rekognition # developer guide. # @return [Types::Image] # class RecognizeCelebritiesRequest < Struct.new( :image) include Aws::Structure end # @!attribute [rw] celebrity_faces # Details about each celebrity found in the image. Amazon Rekognition # can detect a maximum of 15 celebrities in an image. # @return [Array] # # @!attribute [rw] unrecognized_faces # Details about each unrecognized face in the image. # @return [Array] # # @!attribute [rw] orientation_correction # The orientation of the input image (counterclockwise direction). If # your application displays the image, you can use this value to # correct the orientation. The bounding box coordinates returned in # `CelebrityFaces` and `UnrecognizedFaces` represent face locations # before the image orientation is corrected. # # If the input image is in .jpeg format, it might contain exchangeable # image (Exif) metadata that includes the image's orientation. If so, # and the Exif metadata for the input image populates the orientation # field, the value of `OrientationCorrection` is null. The # `CelebrityFaces` and `UnrecognizedFaces` bounding box coordinates # represent face locations after Exif metadata is used to correct the # image orientation. Images in .png format don't contain Exif # metadata. # # # @return [String] # class RecognizeCelebritiesResponse < Struct.new( :celebrity_faces, :unrecognized_faces, :orientation_correction) include Aws::Structure end # Specifies a location within the frame that Rekognition checks for # text. Uses a `BoundingBox` object to set a region of the screen. # # A word is included in the region if the word is more than half in that # region. If there is more than one region, the word will be compared # with all regions of the screen. Any word more than half in a region is # kept in the results. # # @note When making an API call, you may pass RegionOfInterest # data as a hash: # # { # bounding_box: { # width: 1.0, # height: 1.0, # left: 1.0, # top: 1.0, # }, # } # # @!attribute [rw] bounding_box # The box representing a region of interest on screen. # @return [Types::BoundingBox] # class RegionOfInterest < Struct.new( :bounding_box) include Aws::Structure end # A collection with the specified ID already exists. # class ResourceAlreadyExistsException < Aws::EmptyStructure; end class ResourceInUseException < Aws::EmptyStructure; end # The collection specified in the request cannot be found. # class ResourceNotFoundException < Aws::EmptyStructure; end # The requested resource isn't ready. For example, this exception # occurs when you call `DetectCustomLabels` with a model version that # isn't deployed. # class ResourceNotReadyException < Aws::EmptyStructure; end # Provides the S3 bucket name and object name. # # The region for the S3 bucket containing the S3 object must match the # region you use for Amazon Rekognition operations. # # For Amazon Rekognition to process an S3 object, the user must have # permission to access the S3 object. For more information, see # Resource-Based Policies in the Amazon Rekognition Developer Guide. # # @note When making an API call, you may pass S3Object # data as a hash: # # { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # } # # @!attribute [rw] bucket # Name of the S3 bucket. # @return [String] # # @!attribute [rw] name # S3 object key name. # @return [String] # # @!attribute [rw] version # If the bucket is versioning enabled, you can specify the object # version. # @return [String] # class S3Object < Struct.new( :bucket, :name, :version) include Aws::Structure end # @note When making an API call, you may pass SearchFacesByImageRequest # data as a hash: # # { # collection_id: "CollectionId", # required # image: { # required # bytes: "data", # s3_object: { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # }, # }, # max_faces: 1, # face_match_threshold: 1.0, # quality_filter: "NONE", # accepts NONE, AUTO, LOW, MEDIUM, HIGH # } # # @!attribute [rw] collection_id # ID of the collection to search. # @return [String] # # @!attribute [rw] image # The input image as base64-encoded bytes or an S3 object. If you use # the AWS CLI to call Amazon Rekognition operations, passing # base64-encoded image bytes is not supported. # # If you are using an AWS SDK to call Amazon Rekognition, you might # not need to base64-encode image bytes passed using the `Bytes` # field. For more information, see Images in the Amazon Rekognition # developer guide. # @return [Types::Image] # # @!attribute [rw] max_faces # Maximum number of faces to return. The operation returns the maximum # number of faces with the highest confidence in the match. # @return [Integer] # # @!attribute [rw] face_match_threshold # (Optional) Specifies the minimum confidence in the face match to # return. For example, don't return any matches where confidence in # matches is less than 70%. The default value is 80%. # @return [Float] # # @!attribute [rw] quality_filter # A filter that specifies a quality bar for how much filtering is done # to identify faces. Filtered faces aren't searched for in the # collection. If you specify `AUTO`, Amazon Rekognition chooses the # quality bar. If you specify `LOW`, `MEDIUM`, or `HIGH`, filtering # removes all faces that don’t meet the chosen quality bar. The # quality bar is based on a variety of common use cases. Low-quality # detections can occur for a number of reasons. Some examples are an # object that's misidentified as a face, a face that's too blurry, # or a face with a pose that's too extreme to use. If you specify # `NONE`, no filtering is performed. The default value is `NONE`. # # To use quality filtering, the collection you are using must be # associated with version 3 of the face model or higher. # @return [String] # class SearchFacesByImageRequest < Struct.new( :collection_id, :image, :max_faces, :face_match_threshold, :quality_filter) include Aws::Structure end # @!attribute [rw] searched_face_bounding_box # The bounding box around the face in the input image that Amazon # Rekognition used for the search. # @return [Types::BoundingBox] # # @!attribute [rw] searched_face_confidence # The level of confidence that the `searchedFaceBoundingBox`, contains # a face. # @return [Float] # # @!attribute [rw] face_matches # An array of faces that match the input face, along with the # confidence in the match. # @return [Array] # # @!attribute [rw] face_model_version # Version number of the face detection model associated with the input # collection (`CollectionId`). # @return [String] # class SearchFacesByImageResponse < Struct.new( :searched_face_bounding_box, :searched_face_confidence, :face_matches, :face_model_version) include Aws::Structure end # @note When making an API call, you may pass SearchFacesRequest # data as a hash: # # { # collection_id: "CollectionId", # required # face_id: "FaceId", # required # max_faces: 1, # face_match_threshold: 1.0, # } # # @!attribute [rw] collection_id # ID of the collection the face belongs to. # @return [String] # # @!attribute [rw] face_id # ID of a face to find matches for in the collection. # @return [String] # # @!attribute [rw] max_faces # Maximum number of faces to return. The operation returns the maximum # number of faces with the highest confidence in the match. # @return [Integer] # # @!attribute [rw] face_match_threshold # Optional value specifying the minimum confidence in the face match # to return. For example, don't return any matches where confidence # in matches is less than 70%. The default value is 80%. # @return [Float] # class SearchFacesRequest < Struct.new( :collection_id, :face_id, :max_faces, :face_match_threshold) include Aws::Structure end # @!attribute [rw] searched_face_id # ID of the face that was searched for matches in a collection. # @return [String] # # @!attribute [rw] face_matches # An array of faces that matched the input face, along with the # confidence in the match. # @return [Array] # # @!attribute [rw] face_model_version # Version number of the face detection model associated with the input # collection (`CollectionId`). # @return [String] # class SearchFacesResponse < Struct.new( :searched_face_id, :face_matches, :face_model_version) include Aws::Structure end # Indicates whether or not the face is smiling, and the confidence level # in the determination. # # @!attribute [rw] value # Boolean value that indicates whether the face is smiling or not. # @return [Boolean] # # @!attribute [rw] confidence # Level of confidence in the determination. # @return [Float] # class Smile < Struct.new( :value, :confidence) include Aws::Structure end # @note When making an API call, you may pass StartCelebrityRecognitionRequest # data as a hash: # # { # video: { # required # s3_object: { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # }, # }, # client_request_token: "ClientRequestToken", # notification_channel: { # sns_topic_arn: "SNSTopicArn", # required # role_arn: "RoleArn", # required # }, # job_tag: "JobTag", # } # # @!attribute [rw] video # The video in which you want to recognize celebrities. The video must # be stored in an Amazon S3 bucket. # @return [Types::Video] # # @!attribute [rw] client_request_token # Idempotent token used to identify the start request. If you use the # same token with multiple `StartCelebrityRecognition` requests, the # same `JobId` is returned. Use `ClientRequestToken` to prevent the # same job from being accidently started more than once. # @return [String] # # @!attribute [rw] notification_channel # The Amazon SNS topic ARN that you want Amazon Rekognition Video to # publish the completion status of the celebrity recognition analysis # to. # @return [Types::NotificationChannel] # # @!attribute [rw] job_tag # An identifier you specify that's returned in the completion # notification that's published to your Amazon Simple Notification # Service topic. For example, you can use `JobTag` to group related # jobs and identify them in the completion notification. # @return [String] # class StartCelebrityRecognitionRequest < Struct.new( :video, :client_request_token, :notification_channel, :job_tag) include Aws::Structure end # @!attribute [rw] job_id # The identifier for the celebrity recognition analysis job. Use # `JobId` to identify the job in a subsequent call to # `GetCelebrityRecognition`. # @return [String] # class StartCelebrityRecognitionResponse < Struct.new( :job_id) include Aws::Structure end # @note When making an API call, you may pass StartContentModerationRequest # data as a hash: # # { # video: { # required # s3_object: { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # }, # }, # min_confidence: 1.0, # client_request_token: "ClientRequestToken", # notification_channel: { # sns_topic_arn: "SNSTopicArn", # required # role_arn: "RoleArn", # required # }, # job_tag: "JobTag", # } # # @!attribute [rw] video # The video in which you want to detect unsafe content. The video must # be stored in an Amazon S3 bucket. # @return [Types::Video] # # @!attribute [rw] min_confidence # Specifies the minimum confidence that Amazon Rekognition must have # in order to return a moderated content label. Confidence represents # how certain Amazon Rekognition is that the moderated content is # correctly identified. 0 is the lowest confidence. 100 is the highest # confidence. Amazon Rekognition doesn't return any moderated content # labels with a confidence level lower than this specified value. If # you don't specify `MinConfidence`, `GetContentModeration` returns # labels with confidence values greater than or equal to 50 percent. # @return [Float] # # @!attribute [rw] client_request_token # Idempotent token used to identify the start request. If you use the # same token with multiple `StartContentModeration` requests, the same # `JobId` is returned. Use `ClientRequestToken` to prevent the same # job from being accidently started more than once. # @return [String] # # @!attribute [rw] notification_channel # The Amazon SNS topic ARN that you want Amazon Rekognition Video to # publish the completion status of the unsafe content analysis to. # @return [Types::NotificationChannel] # # @!attribute [rw] job_tag # An identifier you specify that's returned in the completion # notification that's published to your Amazon Simple Notification # Service topic. For example, you can use `JobTag` to group related # jobs and identify them in the completion notification. # @return [String] # class StartContentModerationRequest < Struct.new( :video, :min_confidence, :client_request_token, :notification_channel, :job_tag) include Aws::Structure end # @!attribute [rw] job_id # The identifier for the unsafe content analysis job. Use `JobId` to # identify the job in a subsequent call to `GetContentModeration`. # @return [String] # class StartContentModerationResponse < Struct.new( :job_id) include Aws::Structure end # @note When making an API call, you may pass StartFaceDetectionRequest # data as a hash: # # { # video: { # required # s3_object: { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # }, # }, # client_request_token: "ClientRequestToken", # notification_channel: { # sns_topic_arn: "SNSTopicArn", # required # role_arn: "RoleArn", # required # }, # face_attributes: "DEFAULT", # accepts DEFAULT, ALL # job_tag: "JobTag", # } # # @!attribute [rw] video # The video in which you want to detect faces. The video must be # stored in an Amazon S3 bucket. # @return [Types::Video] # # @!attribute [rw] client_request_token # Idempotent token used to identify the start request. If you use the # same token with multiple `StartFaceDetection` requests, the same # `JobId` is returned. Use `ClientRequestToken` to prevent the same # job from being accidently started more than once. # @return [String] # # @!attribute [rw] notification_channel # The ARN of the Amazon SNS topic to which you want Amazon Rekognition # Video to publish the completion status of the face detection # operation. # @return [Types::NotificationChannel] # # @!attribute [rw] face_attributes # The face attributes you want returned. # # `DEFAULT` - The following subset of facial attributes are returned: # BoundingBox, Confidence, Pose, Quality and Landmarks. # # `ALL` - All facial attributes are returned. # @return [String] # # @!attribute [rw] job_tag # An identifier you specify that's returned in the completion # notification that's published to your Amazon Simple Notification # Service topic. For example, you can use `JobTag` to group related # jobs and identify them in the completion notification. # @return [String] # class StartFaceDetectionRequest < Struct.new( :video, :client_request_token, :notification_channel, :face_attributes, :job_tag) include Aws::Structure end # @!attribute [rw] job_id # The identifier for the face detection job. Use `JobId` to identify # the job in a subsequent call to `GetFaceDetection`. # @return [String] # class StartFaceDetectionResponse < Struct.new( :job_id) include Aws::Structure end # @note When making an API call, you may pass StartFaceSearchRequest # data as a hash: # # { # video: { # required # s3_object: { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # }, # }, # client_request_token: "ClientRequestToken", # face_match_threshold: 1.0, # collection_id: "CollectionId", # required # notification_channel: { # sns_topic_arn: "SNSTopicArn", # required # role_arn: "RoleArn", # required # }, # job_tag: "JobTag", # } # # @!attribute [rw] video # The video you want to search. The video must be stored in an Amazon # S3 bucket. # @return [Types::Video] # # @!attribute [rw] client_request_token # Idempotent token used to identify the start request. If you use the # same token with multiple `StartFaceSearch` requests, the same # `JobId` is returned. Use `ClientRequestToken` to prevent the same # job from being accidently started more than once. # @return [String] # # @!attribute [rw] face_match_threshold # The minimum confidence in the person match to return. For example, # don't return any matches where confidence in matches is less than # 70%. The default value is 80%. # @return [Float] # # @!attribute [rw] collection_id # ID of the collection that contains the faces you want to search for. # @return [String] # # @!attribute [rw] notification_channel # The ARN of the Amazon SNS topic to which you want Amazon Rekognition # Video to publish the completion status of the search. # @return [Types::NotificationChannel] # # @!attribute [rw] job_tag # An identifier you specify that's returned in the completion # notification that's published to your Amazon Simple Notification # Service topic. For example, you can use `JobTag` to group related # jobs and identify them in the completion notification. # @return [String] # class StartFaceSearchRequest < Struct.new( :video, :client_request_token, :face_match_threshold, :collection_id, :notification_channel, :job_tag) include Aws::Structure end # @!attribute [rw] job_id # The identifier for the search job. Use `JobId` to identify the job # in a subsequent call to `GetFaceSearch`. # @return [String] # class StartFaceSearchResponse < Struct.new( :job_id) include Aws::Structure end # @note When making an API call, you may pass StartLabelDetectionRequest # data as a hash: # # { # video: { # required # s3_object: { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # }, # }, # client_request_token: "ClientRequestToken", # min_confidence: 1.0, # notification_channel: { # sns_topic_arn: "SNSTopicArn", # required # role_arn: "RoleArn", # required # }, # job_tag: "JobTag", # } # # @!attribute [rw] video # The video in which you want to detect labels. The video must be # stored in an Amazon S3 bucket. # @return [Types::Video] # # @!attribute [rw] client_request_token # Idempotent token used to identify the start request. If you use the # same token with multiple `StartLabelDetection` requests, the same # `JobId` is returned. Use `ClientRequestToken` to prevent the same # job from being accidently started more than once. # @return [String] # # @!attribute [rw] min_confidence # Specifies the minimum confidence that Amazon Rekognition Video must # have in order to return a detected label. Confidence represents how # certain Amazon Rekognition is that a label is correctly identified.0 # is the lowest confidence. 100 is the highest confidence. Amazon # Rekognition Video doesn't return any labels with a confidence level # lower than this specified value. # # If you don't specify `MinConfidence`, the operation returns labels # with confidence values greater than or equal to 50 percent. # @return [Float] # # @!attribute [rw] notification_channel # The Amazon SNS topic ARN you want Amazon Rekognition Video to # publish the completion status of the label detection operation to. # @return [Types::NotificationChannel] # # @!attribute [rw] job_tag # An identifier you specify that's returned in the completion # notification that's published to your Amazon Simple Notification # Service topic. For example, you can use `JobTag` to group related # jobs and identify them in the completion notification. # @return [String] # class StartLabelDetectionRequest < Struct.new( :video, :client_request_token, :min_confidence, :notification_channel, :job_tag) include Aws::Structure end # @!attribute [rw] job_id # The identifier for the label detection job. Use `JobId` to identify # the job in a subsequent call to `GetLabelDetection`. # @return [String] # class StartLabelDetectionResponse < Struct.new( :job_id) include Aws::Structure end # @note When making an API call, you may pass StartPersonTrackingRequest # data as a hash: # # { # video: { # required # s3_object: { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # }, # }, # client_request_token: "ClientRequestToken", # notification_channel: { # sns_topic_arn: "SNSTopicArn", # required # role_arn: "RoleArn", # required # }, # job_tag: "JobTag", # } # # @!attribute [rw] video # The video in which you want to detect people. The video must be # stored in an Amazon S3 bucket. # @return [Types::Video] # # @!attribute [rw] client_request_token # Idempotent token used to identify the start request. If you use the # same token with multiple `StartPersonTracking` requests, the same # `JobId` is returned. Use `ClientRequestToken` to prevent the same # job from being accidently started more than once. # @return [String] # # @!attribute [rw] notification_channel # The Amazon SNS topic ARN you want Amazon Rekognition Video to # publish the completion status of the people detection operation to. # @return [Types::NotificationChannel] # # @!attribute [rw] job_tag # An identifier you specify that's returned in the completion # notification that's published to your Amazon Simple Notification # Service topic. For example, you can use `JobTag` to group related # jobs and identify them in the completion notification. # @return [String] # class StartPersonTrackingRequest < Struct.new( :video, :client_request_token, :notification_channel, :job_tag) include Aws::Structure end # @!attribute [rw] job_id # The identifier for the person detection job. Use `JobId` to identify # the job in a subsequent call to `GetPersonTracking`. # @return [String] # class StartPersonTrackingResponse < Struct.new( :job_id) include Aws::Structure end # @note When making an API call, you may pass StartProjectVersionRequest # data as a hash: # # { # project_version_arn: "ProjectVersionArn", # required # min_inference_units: 1, # required # } # # @!attribute [rw] project_version_arn # The Amazon Resource Name(ARN) of the model version that you want to # start. # @return [String] # # @!attribute [rw] min_inference_units # The minimum number of inference units to use. A single inference # unit represents 1 hour of processing and can support up to 5 # Transaction Pers Second (TPS). Use a higher number to increase the # TPS throughput of your model. You are charged for the number of # inference units that you use. # @return [Integer] # class StartProjectVersionRequest < Struct.new( :project_version_arn, :min_inference_units) include Aws::Structure end # @!attribute [rw] status # The current running status of the model. # @return [String] # class StartProjectVersionResponse < Struct.new( :status) include Aws::Structure end # @note When making an API call, you may pass StartStreamProcessorRequest # data as a hash: # # { # name: "StreamProcessorName", # required # } # # @!attribute [rw] name # The name of the stream processor to start processing. # @return [String] # class StartStreamProcessorRequest < Struct.new( :name) include Aws::Structure end class StartStreamProcessorResponse < Aws::EmptyStructure; end # Set of optional parameters that let you set the criteria text must # meet to be included in your response. `WordFilter` looks at a word's # height, width and minimum confidence. `RegionOfInterest` lets you set # a specific region of the screen to look for text in. # # @note When making an API call, you may pass StartTextDetectionFilters # data as a hash: # # { # word_filter: { # min_confidence: 1.0, # min_bounding_box_height: 1.0, # min_bounding_box_width: 1.0, # }, # regions_of_interest: [ # { # bounding_box: { # width: 1.0, # height: 1.0, # left: 1.0, # top: 1.0, # }, # }, # ], # } # # @!attribute [rw] word_filter # Filters focusing on qualities of the text, such as confidence or # size. # @return [Types::DetectionFilter] # # @!attribute [rw] regions_of_interest # Filter focusing on a certain area of the frame. Uses a `BoundingBox` # object to set the region of the screen. # @return [Array] # class StartTextDetectionFilters < Struct.new( :word_filter, :regions_of_interest) include Aws::Structure end # @note When making an API call, you may pass StartTextDetectionRequest # data as a hash: # # { # video: { # required # s3_object: { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # }, # }, # client_request_token: "ClientRequestToken", # notification_channel: { # sns_topic_arn: "SNSTopicArn", # required # role_arn: "RoleArn", # required # }, # job_tag: "JobTag", # filters: { # word_filter: { # min_confidence: 1.0, # min_bounding_box_height: 1.0, # min_bounding_box_width: 1.0, # }, # regions_of_interest: [ # { # bounding_box: { # width: 1.0, # height: 1.0, # left: 1.0, # top: 1.0, # }, # }, # ], # }, # } # # @!attribute [rw] video # Video file stored in an Amazon S3 bucket. Amazon Rekognition video # start operations such as StartLabelDetection use `Video` to specify # a video for analysis. The supported file formats are .mp4, .mov and # .avi. # @return [Types::Video] # # @!attribute [rw] client_request_token # Idempotent token used to identify the start request. If you use the # same token with multiple `StartTextDetection` requests, the same # `JobId` is returned. Use `ClientRequestToken` to prevent the same # job from being accidentaly started more than once. # @return [String] # # @!attribute [rw] notification_channel # The Amazon Simple Notification Service topic to which Amazon # Rekognition publishes the completion status of a video analysis # operation. For more information, see api-video. # @return [Types::NotificationChannel] # # @!attribute [rw] job_tag # An identifier returned in the completion status published by your # Amazon Simple Notification Service topic. For example, you can use # `JobTag` to group related jobs and identify them in the completion # notification. # @return [String] # # @!attribute [rw] filters # Optional parameters that let you set criteria the text must meet to # be included in your response. # @return [Types::StartTextDetectionFilters] # class StartTextDetectionRequest < Struct.new( :video, :client_request_token, :notification_channel, :job_tag, :filters) include Aws::Structure end # @!attribute [rw] job_id # Identifier for the text detection job. Use `JobId` to identify the # job in a subsequent call to `GetTextDetection`. # @return [String] # class StartTextDetectionResponse < Struct.new( :job_id) include Aws::Structure end # @note When making an API call, you may pass StopProjectVersionRequest # data as a hash: # # { # project_version_arn: "ProjectVersionArn", # required # } # # @!attribute [rw] project_version_arn # The Amazon Resource Name (ARN) of the model version that you want to # delete. # # This operation requires permissions to perform the # `rekognition:StopProjectVersion` action. # @return [String] # class StopProjectVersionRequest < Struct.new( :project_version_arn) include Aws::Structure end # @!attribute [rw] status # The current status of the stop operation. # @return [String] # class StopProjectVersionResponse < Struct.new( :status) include Aws::Structure end # @note When making an API call, you may pass StopStreamProcessorRequest # data as a hash: # # { # name: "StreamProcessorName", # required # } # # @!attribute [rw] name # The name of a stream processor created by CreateStreamProcessor. # @return [String] # class StopStreamProcessorRequest < Struct.new( :name) include Aws::Structure end class StopStreamProcessorResponse < Aws::EmptyStructure; end # An object that recognizes faces in a streaming video. An Amazon # Rekognition stream processor is created by a call to # CreateStreamProcessor. The request parameters for # `CreateStreamProcessor` describe the Kinesis video stream source for # the streaming video, face recognition parameters, and where to stream # the analysis resullts. # # @!attribute [rw] name # Name of the Amazon Rekognition stream processor. # @return [String] # # @!attribute [rw] status # Current status of the Amazon Rekognition stream processor. # @return [String] # class StreamProcessor < Struct.new( :name, :status) include Aws::Structure end # Information about the source streaming video. # # @note When making an API call, you may pass StreamProcessorInput # data as a hash: # # { # kinesis_video_stream: { # arn: "KinesisVideoArn", # }, # } # # @!attribute [rw] kinesis_video_stream # The Kinesis video stream input stream for the source streaming # video. # @return [Types::KinesisVideoStream] # class StreamProcessorInput < Struct.new( :kinesis_video_stream) include Aws::Structure end # Information about the Amazon Kinesis Data Streams stream to which a # Amazon Rekognition Video stream processor streams the results of a # video analysis. For more information, see CreateStreamProcessor in the # Amazon Rekognition Developer Guide. # # @note When making an API call, you may pass StreamProcessorOutput # data as a hash: # # { # kinesis_data_stream: { # arn: "KinesisDataArn", # }, # } # # @!attribute [rw] kinesis_data_stream # The Amazon Kinesis Data Streams stream to which the Amazon # Rekognition stream processor streams the analysis results. # @return [Types::KinesisDataStream] # class StreamProcessorOutput < Struct.new( :kinesis_data_stream) include Aws::Structure end # Input parameters used to recognize faces in a streaming video analyzed # by a Amazon Rekognition stream processor. # # @note When making an API call, you may pass StreamProcessorSettings # data as a hash: # # { # face_search: { # collection_id: "CollectionId", # face_match_threshold: 1.0, # }, # } # # @!attribute [rw] face_search # Face search settings to use on a streaming video. # @return [Types::FaceSearchSettings] # class StreamProcessorSettings < Struct.new( :face_search) include Aws::Structure end # The S3 bucket that contains the training summary. The training summary # includes aggregated evaluation metrics for the entire testing dataset # and metrics for each individual label. # # You get the training summary S3 bucket location by calling # DescribeProjectVersions. # # @!attribute [rw] s3_object # Provides the S3 bucket name and object name. # # The region for the S3 bucket containing the S3 object must match the # region you use for Amazon Rekognition operations. # # For Amazon Rekognition to process an S3 object, the user must have # permission to access the S3 object. For more information, see # Resource-Based Policies in the Amazon Rekognition Developer Guide. # @return [Types::S3Object] # class Summary < Struct.new( :s3_object) include Aws::Structure end # Indicates whether or not the face is wearing sunglasses, and the # confidence level in the determination. # # @!attribute [rw] value # Boolean value that indicates whether the face is wearing sunglasses # or not. # @return [Boolean] # # @!attribute [rw] confidence # Level of confidence in the determination. # @return [Float] # class Sunglasses < Struct.new( :value, :confidence) include Aws::Structure end # The dataset used for testing. Optionally, if `AutoCreate` is set, # Amazon Rekognition Custom Labels creates a testing dataset using an # 80/20 split of the training dataset. # # @note When making an API call, you may pass TestingData # data as a hash: # # { # assets: [ # { # ground_truth_manifest: { # s3_object: { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # }, # }, # }, # ], # auto_create: false, # } # # @!attribute [rw] assets # The assets used for testing. # @return [Array] # # @!attribute [rw] auto_create # If specified, Amazon Rekognition Custom Labels creates a testing # dataset with an 80/20 split of the training dataset. # @return [Boolean] # class TestingData < Struct.new( :assets, :auto_create) include Aws::Structure end # A Sagemaker Groundtruth format manifest file representing the dataset # used for testing. # # @!attribute [rw] input # The testing dataset that was supplied for training. # @return [Types::TestingData] # # @!attribute [rw] output # The subset of the dataset that was actually tested. Some images # (assets) might not be tested due to file formatting and other # issues. # @return [Types::TestingData] # class TestingDataResult < Struct.new( :input, :output) include Aws::Structure end # Information about a word or line of text detected by DetectText. # # The `DetectedText` field contains the text that Amazon Rekognition # detected in the image. # # Every word and line has an identifier (`Id`). Each word belongs to a # line and has a parent identifier (`ParentId`) that identifies the line # of text in which the word appears. The word `Id` is also an index for # the word within a line of words. # # For more information, see Detecting Text in the Amazon Rekognition # Developer Guide. # # @!attribute [rw] detected_text # The word or line of text recognized by Amazon Rekognition. # @return [String] # # @!attribute [rw] type # The type of text that was detected. # @return [String] # # @!attribute [rw] id # The identifier for the detected text. The identifier is only unique # for a single call to `DetectText`. # @return [Integer] # # @!attribute [rw] parent_id # The Parent identifier for the detected text identified by the value # of `ID`. If the type of detected text is `LINE`, the value of # `ParentId` is `Null`. # @return [Integer] # # @!attribute [rw] confidence # The confidence that Amazon Rekognition has in the accuracy of the # detected text and the accuracy of the geometry points around the # detected text. # @return [Float] # # @!attribute [rw] geometry # The location of the detected text on the image. Includes an axis # aligned coarse bounding box surrounding the text and a finer grain # polygon for more accurate spatial information. # @return [Types::Geometry] # class TextDetection < Struct.new( :detected_text, :type, :id, :parent_id, :confidence, :geometry) include Aws::Structure end # Information about text detected in a video. Incudes the detected text, # the time in milliseconds from the start of the video that the text was # detected, and where it was detected on the screen. # # @!attribute [rw] timestamp # The time, in milliseconds from the start of the video, that the text # was detected. # @return [Integer] # # @!attribute [rw] text_detection # Details about text detected in a video. # @return [Types::TextDetection] # class TextDetectionResult < Struct.new( :timestamp, :text_detection) include Aws::Structure end # Amazon Rekognition is temporarily unable to process the request. Try # your call again. # class ThrottlingException < Aws::EmptyStructure; end # The dataset used for training. # # @note When making an API call, you may pass TrainingData # data as a hash: # # { # assets: [ # { # ground_truth_manifest: { # s3_object: { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # }, # }, # }, # ], # } # # @!attribute [rw] assets # A Sagemaker GroundTruth manifest file that contains the training # images (assets). # @return [Array] # class TrainingData < Struct.new( :assets) include Aws::Structure end # A Sagemaker Groundtruth format manifest file that represents the # dataset used for training. # # @!attribute [rw] input # The training assets that you supplied for training. # @return [Types::TrainingData] # # @!attribute [rw] output # The images (assets) that were actually trained by Amazon Rekognition # Custom Labels. # @return [Types::TrainingData] # class TrainingDataResult < Struct.new( :input, :output) include Aws::Structure end # A face that IndexFaces detected, but didn't index. Use the `Reasons` # response attribute to determine why a face wasn't indexed. # # @!attribute [rw] reasons # An array of reasons that specify why a face wasn't indexed. # # * EXTREME\_POSE - The face is at a pose that can't be detected. For # example, the head is turned too far away from the camera. # # * EXCEEDS\_MAX\_FACES - The number of faces detected is already # higher than that specified by the `MaxFaces` input parameter for # `IndexFaces`. # # * LOW\_BRIGHTNESS - The image is too dark. # # * LOW\_SHARPNESS - The image is too blurry. # # * LOW\_CONFIDENCE - The face was detected with a low confidence. # # * SMALL\_BOUNDING\_BOX - The bounding box around the face is too # small. # @return [Array] # # @!attribute [rw] face_detail # The structure that contains attributes of a face that # `IndexFaces`detected, but didn't index. # @return [Types::FaceDetail] # class UnindexedFace < Struct.new( :reasons, :face_detail) include Aws::Structure end # Video file stored in an Amazon S3 bucket. Amazon Rekognition video # start operations such as StartLabelDetection use `Video` to specify a # video for analysis. The supported file formats are .mp4, .mov and # .avi. # # @note When making an API call, you may pass Video # data as a hash: # # { # s3_object: { # bucket: "S3Bucket", # name: "S3ObjectName", # version: "S3ObjectVersion", # }, # } # # @!attribute [rw] s3_object # The Amazon S3 bucket name and file name for the video. # @return [Types::S3Object] # class Video < Struct.new( :s3_object) include Aws::Structure end # Information about a video that Amazon Rekognition analyzed. # `Videometadata` is returned in every page of paginated responses from # a Amazon Rekognition video operation. # # @!attribute [rw] codec # Type of compression used in the analyzed video. # @return [String] # # @!attribute [rw] duration_millis # Length of the video in milliseconds. # @return [Integer] # # @!attribute [rw] format # Format of the analyzed video. Possible values are MP4, MOV and AVI. # @return [String] # # @!attribute [rw] frame_rate # Number of frames per second in the video. # @return [Float] # # @!attribute [rw] frame_height # Vertical pixel dimension of the video. # @return [Integer] # # @!attribute [rw] frame_width # Horizontal pixel dimension of the video. # @return [Integer] # class VideoMetadata < Struct.new( :codec, :duration_millis, :format, :frame_rate, :frame_height, :frame_width) include Aws::Structure end # The file size or duration of the supplied media is too large. The # maximum file size is 10GB. The maximum duration is 6 hours. # class VideoTooLargeException < Aws::EmptyStructure; end end end