# frozen_string_literal: true # WARNING ABOUT GENERATED CODE # # This file is generated. See the contributing guide for more information: # https://github.com/aws/aws-sdk-ruby/blob/version-3/CONTRIBUTING.md # # WARNING ABOUT GENERATED CODE module Aws::S3 class Object extend Aws::Deprecations # @overload def initialize(bucket_name, key, options = {}) # @param [String] bucket_name # @param [String] key # @option options [Client] :client # @overload def initialize(options = {}) # @option options [required, String] :bucket_name # @option options [required, String] :key # @option options [Client] :client def initialize(*args) options = Hash === args.last ? args.pop.dup : {} @bucket_name = extract_bucket_name(args, options) @key = extract_key(args, options) @data = options.delete(:data) @client = options.delete(:client) || Client.new(options) @waiter_block_warned = false end # @!group Read-Only Attributes # @return [String] def bucket_name @bucket_name end # @return [String] def key @key end # Specifies whether the object retrieved was (true) or was not (false) a # Delete Marker. If false, this response header does not appear in the # response. # # This functionality is not supported for directory buckets. # # # @return [Boolean] def delete_marker data[:delete_marker] end # Indicates that a range of bytes was specified. # @return [String] def accept_ranges data[:accept_ranges] end # If the object expiration is configured (see [ # `PutBucketLifecycleConfiguration` ][1]), the response includes this # header. It includes the `expiry-date` and `rule-id` key-value pairs # providing object expiration information. The value of the `rule-id` is # URL-encoded. # # Object expiration information is not returned in directory buckets and # this header returns the value "`NotImplemented`" in all responses # for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/API/API_PutBucketLifecycleConfiguration.html # @return [String] def expiration data[:expiration] end # If the object is an archived object (an object whose storage class is # GLACIER), the response includes this header if either the archive # restoration is in progress (see [RestoreObject][1] or an archive copy # is already restored. # # If an archive copy is already restored, the header value indicates # when Amazon S3 is scheduled to delete the object copy. For example: # # `x-amz-restore: ongoing-request="false", expiry-date="Fri, 21 Dec 2012 # 00:00:00 GMT"` # # If the object restoration is in progress, the header returns the value # `ongoing-request="true"`. # # For more information about archiving objects, see [Transitioning # Objects: General Considerations][2]. # # This functionality is not supported for directory buckets. Only the S3 # Express One Zone storage class is supported by directory buckets to # store objects. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/API/API_RestoreObject.html # [2]: https://docs.aws.amazon.com/AmazonS3/latest/dev/object-lifecycle-mgmt.html#lifecycle-transition-general-considerations # @return [String] def restore data[:restore] end # The archive state of the head object. # # This functionality is not supported for directory buckets. # # # @return [String] def archive_status data[:archive_status] end # Date and time when the object was last modified. # @return [Time] def last_modified data[:last_modified] end # Size of the body in bytes. # @return [Integer] def content_length data[:content_length] end # The Base64 encoded, 32-bit `CRC-32 checksum` of the object. This # checksum is only be present if the checksum was uploaded with the # object. When you use an API operation on an object that was uploaded # using multipart uploads, this value may not be a direct checksum value # of the full object. Instead, it's a calculation based on the checksum # values of each individual part. For more information about how # checksums are calculated with multipart uploads, see [ Checking object # integrity][1] in the *Amazon S3 User Guide*. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/checking-object-integrity.html#large-object-checksums # @return [String] def checksum_crc32 data[:checksum_crc32] end # The Base64 encoded, 32-bit `CRC-32C` checksum of the object. This # checksum is only present if the checksum was uploaded with the object. # When you use an API operation on an object that was uploaded using # multipart uploads, this value may not be a direct checksum value of # the full object. Instead, it's a calculation based on the checksum # values of each individual part. For more information about how # checksums are calculated with multipart uploads, see [ Checking object # integrity][1] in the *Amazon S3 User Guide*. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/checking-object-integrity.html#large-object-checksums # @return [String] def checksum_crc32c data[:checksum_crc32c] end # The Base64 encoded, 64-bit `CRC-64NVME` checksum of the object. For # more information, see [Checking object integrity in the Amazon S3 User # Guide][1]. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/checking-object-integrity.html # @return [String] def checksum_crc64nvme data[:checksum_crc64nvme] end # The Base64 encoded, 160-bit `SHA-1` digest of the object. This will # only be present if the object was uploaded with the object. When you # use the API operation on an object that was uploaded using multipart # uploads, this value may not be a direct checksum value of the full # object. Instead, it's a calculation based on the checksum values of # each individual part. For more information about how checksums are # calculated with multipart uploads, see [ Checking object integrity][1] # in the *Amazon S3 User Guide*. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/checking-object-integrity.html#large-object-checksums # @return [String] def checksum_sha1 data[:checksum_sha1] end # The Base64 encoded, 256-bit `SHA-256` digest of the object. This will # only be present if the object was uploaded with the object. When you # use an API operation on an object that was uploaded using multipart # uploads, this value may not be a direct checksum value of the full # object. Instead, it's a calculation based on the checksum values of # each individual part. For more information about how checksums are # calculated with multipart uploads, see [ Checking object integrity][1] # in the *Amazon S3 User Guide*. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/checking-object-integrity.html#large-object-checksums # @return [String] def checksum_sha256 data[:checksum_sha256] end # The checksum type, which determines how part-level checksums are # combined to create an object-level checksum for multipart objects. You # can use this header response to verify that the checksum type that is # received is the same checksum type that was specified in # `CreateMultipartUpload` request. For more information, see [Checking # object integrity in the Amazon S3 User Guide][1]. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/checking-object-integrity.html # @return [String] def checksum_type data[:checksum_type] end # An entity tag (ETag) is an opaque identifier assigned by a web server # to a specific version of a resource found at a URL. # @return [String] def etag data[:etag] end # This is set to the number of metadata entries not returned in # `x-amz-meta` headers. This can happen if you create metadata using an # API like SOAP that supports more flexible metadata than the REST API. # For example, using SOAP, you can create metadata whose values are not # legal HTTP headers. # # This functionality is not supported for directory buckets. # # # @return [Integer] def missing_meta data[:missing_meta] end # Version ID of the object. # # This functionality is not supported for directory buckets. # # # @return [String] def version_id data[:version_id] end # Specifies caching behavior along the request/reply chain. # @return [String] def cache_control data[:cache_control] end # Specifies presentational information for the object. # @return [String] def content_disposition data[:content_disposition] end # Indicates what content encodings have been applied to the object and # thus what decoding mechanisms must be applied to obtain the media-type # referenced by the Content-Type header field. # @return [String] def content_encoding data[:content_encoding] end # The language the content is in. # @return [String] def content_language data[:content_language] end # A standard MIME type describing the format of the object data. # @return [String] def content_type data[:content_type] end # The date and time at which the object is no longer cacheable. # @return [Time] def expires data[:expires] end # @return [String] def expires_string data[:expires_string] end # If the bucket is configured as a website, redirects requests for this # object to another object in the same bucket or to an external URL. # Amazon S3 stores the value of this header in the object metadata. # # This functionality is not supported for directory buckets. # # # @return [String] def website_redirect_location data[:website_redirect_location] end # The server-side encryption algorithm used when you store this object # in Amazon S3 (for example, `AES256`, `aws:kms`, `aws:kms:dsse`). # @return [String] def server_side_encryption data[:server_side_encryption] end # A map of metadata to store with the object in S3. # @return [Hash] def metadata data[:metadata] end # If server-side encryption with a customer-provided encryption key was # requested, the response will include this header to confirm the # encryption algorithm that's used. # # This functionality is not supported for directory buckets. # # # @return [String] def sse_customer_algorithm data[:sse_customer_algorithm] end # If server-side encryption with a customer-provided encryption key was # requested, the response will include this header to provide the # round-trip message integrity verification of the customer-provided # encryption key. # # This functionality is not supported for directory buckets. # # # @return [String] def sse_customer_key_md5 data[:sse_customer_key_md5] end # If present, indicates the ID of the KMS key that was used for object # encryption. # @return [String] def ssekms_key_id data[:ssekms_key_id] end # Indicates whether the object uses an S3 Bucket Key for server-side # encryption with Key Management Service (KMS) keys (SSE-KMS). # @return [Boolean] def bucket_key_enabled data[:bucket_key_enabled] end # Provides storage class information of the object. Amazon S3 returns # this header for all objects except for S3 Standard storage class # objects. # # For more information, see [Storage Classes][1]. # # Directory buckets - Only the S3 Express One Zone storage class # is supported by directory buckets to store objects. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/storage-class-intro.html # @return [String] def storage_class data[:storage_class] end # If present, indicates that the requester was successfully charged for # the request. # # This functionality is not supported for directory buckets. # # # @return [String] def request_charged data[:request_charged] end # Amazon S3 can return this header if your request involves a bucket # that is either a source or a destination in a replication rule. # # In replication, you have a source bucket on which you configure # replication and destination bucket or buckets where Amazon S3 stores # object replicas. When you request an object (`GetObject`) or object # metadata (`HeadObject`) from these buckets, Amazon S3 will return the # `x-amz-replication-status` header in the response as follows: # # * **If requesting an object from the source bucket**, Amazon S3 will # return the `x-amz-replication-status` header if the object in your # request is eligible for replication. # # For example, suppose that in your replication configuration, you # specify object prefix `TaxDocs` requesting Amazon S3 to replicate # objects with key prefix `TaxDocs`. Any objects you upload with this # key name prefix, for example `TaxDocs/document1.pdf`, are eligible # for replication. For any object request with this key name prefix, # Amazon S3 will return the `x-amz-replication-status` header with # value PENDING, COMPLETED or FAILED indicating object replication # status. # # * **If requesting an object from a destination bucket**, Amazon S3 # will return the `x-amz-replication-status` header with value REPLICA # if the object in your request is a replica that Amazon S3 created # and there is no replica modification replication in progress. # # * **When replicating objects to multiple destination buckets**, the # `x-amz-replication-status` header acts differently. The header of # the source object will only return a value of COMPLETED when # replication is successful to all destinations. The header will # remain at value PENDING until replication has completed for all # destinations. If one or more destinations fails replication the # header will return FAILED. # # For more information, see [Replication][1]. # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/NotificationHowTo.html # @return [String] def replication_status data[:replication_status] end # The count of parts this object has. This value is only returned if you # specify `partNumber` in your request and the object was uploaded as a # multipart upload. # @return [Integer] def parts_count data[:parts_count] end # The Object Lock mode, if any, that's in effect for this object. This # header is only returned if the requester has the # `s3:GetObjectRetention` permission. For more information about S3 # Object Lock, see [Object Lock][1]. # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/object-lock.html # @return [String] def object_lock_mode data[:object_lock_mode] end # The date and time when the Object Lock retention period expires. This # header is only returned if the requester has the # `s3:GetObjectRetention` permission. # # This functionality is not supported for directory buckets. # # # @return [Time] def object_lock_retain_until_date data[:object_lock_retain_until_date] end # Specifies whether a legal hold is in effect for this object. This # header is only returned if the requester has the # `s3:GetObjectLegalHold` permission. This header is not returned if the # specified version of this object has never had a legal hold applied. # For more information about S3 Object Lock, see [Object Lock][1]. # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/object-lock.html # @return [String] def object_lock_legal_hold_status data[:object_lock_legal_hold_status] end # @!endgroup # @return [Client] def client @client end # Loads, or reloads {#data} for the current {Object}. # Returns `self` making it possible to chain methods. # # object.reload.data # # @return [self] def load resp = Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do @client.head_object( bucket: @bucket_name, key: @key ) end @data = resp.data self end alias :reload :load # @return [Types::HeadObjectOutput] # Returns the data for this {Object}. Calls # {Client#head_object} if {#data_loaded?} is `false`. def data load unless @data @data end # @return [Boolean] # Returns `true` if this resource is loaded. Accessing attributes or # {#data} on an unloaded resource will trigger a call to {#load}. def data_loaded? !!@data end # @param [Hash] options ({}) # @return [Boolean] # Returns `true` if the Object exists. def exists?(options = {}) begin wait_until_exists(options.merge(max_attempts: 1)) true rescue Aws::Waiters::Errors::UnexpectedError => e raise e.error rescue Aws::Waiters::Errors::WaiterFailed false end end # @param [Hash] options ({}) # @option options [Integer] :max_attempts (20) # @option options [Float] :delay (5) # @option options [Proc] :before_attempt # @option options [Proc] :before_wait # @return [Object] def wait_until_exists(options = {}, &block) options, params = separate_params_and_options(options) waiter = Waiters::ObjectExists.new(options) yield_waiter_and_warn(waiter, &block) if block_given? Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do waiter.wait(params.merge(bucket: @bucket_name, key: @key)) end Object.new({ bucket_name: @bucket_name, key: @key, client: @client }) end # @param [Hash] options ({}) # @option options [Integer] :max_attempts (20) # @option options [Float] :delay (5) # @option options [Proc] :before_attempt # @option options [Proc] :before_wait # @return [Object] def wait_until_not_exists(options = {}, &block) options, params = separate_params_and_options(options) waiter = Waiters::ObjectNotExists.new(options) yield_waiter_and_warn(waiter, &block) if block_given? Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do waiter.wait(params.merge(bucket: @bucket_name, key: @key)) end Object.new({ bucket_name: @bucket_name, key: @key, client: @client }) end # @deprecated Use [Aws::S3::Client] #wait_until instead # # Waiter polls an API operation until a resource enters a desired # state. # # @note The waiting operation is performed on a copy. The original resource # remains unchanged. # # ## Basic Usage # # Waiter will polls until it is successful, it fails by # entering a terminal state, or until a maximum number of attempts # are made. # # # polls in a loop until condition is true # resource.wait_until(options) {|resource| condition} # # ## Example # # instance.wait_until(max_attempts:10, delay:5) do |instance| # instance.state.name == 'running' # end # # ## Configuration # # You can configure the maximum number of polling attempts, and the # delay (in seconds) between each polling attempt. The waiting condition is # set by passing a block to {#wait_until}: # # # poll for ~25 seconds # resource.wait_until(max_attempts:5,delay:5) {|resource|...} # # ## Callbacks # # You can be notified before each polling attempt and before each # delay. If you throw `:success` or `:failure` from these callbacks, # it will terminate the waiter. # # started_at = Time.now # # poll for 1 hour, instead of a number of attempts # proc = Proc.new do |attempts, response| # throw :failure if Time.now - started_at > 3600 # end # # # disable max attempts # instance.wait_until(before_wait:proc, max_attempts:nil) {...} # # ## Handling Errors # # When a waiter is successful, it returns the Resource. When a waiter # fails, it raises an error. # # begin # resource.wait_until(...) # rescue Aws::Waiters::Errors::WaiterFailed # # resource did not enter the desired state in time # end # # @yieldparam [Resource] resource to be used in the waiting condition. # # @raise [Aws::Waiters::Errors::FailureStateError] Raised when the waiter # terminates because the waiter has entered a state that it will not # transition out of, preventing success. # # yet successful. # # @raise [Aws::Waiters::Errors::UnexpectedError] Raised when an error is # encountered while polling for a resource that is not expected. # # @raise [NotImplementedError] Raised when the resource does not # # @option options [Integer] :max_attempts (10) Maximum number of # attempts # @option options [Integer] :delay (10) Delay between each # attempt in seconds # @option options [Proc] :before_attempt (nil) Callback # invoked before each attempt # @option options [Proc] :before_wait (nil) Callback # invoked before each wait # @return [Resource] if the waiter was successful def wait_until(options = {}, &block) self_copy = self.dup attempts = 0 options[:max_attempts] = 10 unless options.key?(:max_attempts) options[:delay] ||= 10 options[:poller] = Proc.new do attempts += 1 if block.call(self_copy) [:success, self_copy] else self_copy.reload unless attempts == options[:max_attempts] :retry end end Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do Aws::Waiters::Waiter.new(options).wait({}) end end # @!group Actions # @example Request syntax with placeholder values # # object.copy_from({ # acl: "private", # accepts private, public-read, public-read-write, authenticated-read, aws-exec-read, bucket-owner-read, bucket-owner-full-control # cache_control: "CacheControl", # checksum_algorithm: "CRC32", # accepts CRC32, CRC32C, SHA1, SHA256, CRC64NVME # content_disposition: "ContentDisposition", # content_encoding: "ContentEncoding", # content_language: "ContentLanguage", # content_type: "ContentType", # copy_source: "CopySource", # required # copy_source_if_match: "CopySourceIfMatch", # copy_source_if_modified_since: Time.now, # copy_source_if_none_match: "CopySourceIfNoneMatch", # copy_source_if_unmodified_since: Time.now, # expires: Time.now, # grant_full_control: "GrantFullControl", # grant_read: "GrantRead", # grant_read_acp: "GrantReadACP", # grant_write_acp: "GrantWriteACP", # metadata: { # "MetadataKey" => "MetadataValue", # }, # metadata_directive: "COPY", # accepts COPY, REPLACE # tagging_directive: "COPY", # accepts COPY, REPLACE # server_side_encryption: "AES256", # accepts AES256, aws:kms, aws:kms:dsse # storage_class: "STANDARD", # accepts STANDARD, REDUCED_REDUNDANCY, STANDARD_IA, ONEZONE_IA, INTELLIGENT_TIERING, GLACIER, DEEP_ARCHIVE, OUTPOSTS, GLACIER_IR, SNOW, EXPRESS_ONEZONE # website_redirect_location: "WebsiteRedirectLocation", # sse_customer_algorithm: "SSECustomerAlgorithm", # sse_customer_key: "SSECustomerKey", # sse_customer_key_md5: "SSECustomerKeyMD5", # ssekms_key_id: "SSEKMSKeyId", # ssekms_encryption_context: "SSEKMSEncryptionContext", # bucket_key_enabled: false, # copy_source_sse_customer_algorithm: "CopySourceSSECustomerAlgorithm", # copy_source_sse_customer_key: "CopySourceSSECustomerKey", # copy_source_sse_customer_key_md5: "CopySourceSSECustomerKeyMD5", # request_payer: "requester", # accepts requester # tagging: "TaggingHeader", # object_lock_mode: "GOVERNANCE", # accepts GOVERNANCE, COMPLIANCE # object_lock_retain_until_date: Time.now, # object_lock_legal_hold_status: "ON", # accepts ON, OFF # expected_bucket_owner: "AccountId", # expected_source_bucket_owner: "AccountId", # }) # @param [Hash] options ({}) # @option options [String] :acl # The canned access control list (ACL) to apply to the object. # # When you copy an object, the ACL metadata is not preserved and is set # to `private` by default. Only the owner has full access control. To # override the default ACL setting, specify a new ACL when you generate # a copy request. For more information, see [Using ACLs][1]. # # If the destination bucket that you're copying objects to uses the # bucket owner enforced setting for S3 Object Ownership, ACLs are # disabled and no longer affect permissions. Buckets that use this # setting only accept `PUT` requests that don't specify an ACL or `PUT` # requests that specify bucket owner full control ACLs, such as the # `bucket-owner-full-control` canned ACL or an equivalent form of this # ACL expressed in the XML format. For more information, see # [Controlling ownership of objects and disabling ACLs][2] in the # *Amazon S3 User Guide*. # # * If your destination bucket uses the bucket owner enforced setting # for Object Ownership, all objects written to the bucket by any # account will be owned by the bucket owner. # # * This functionality is not supported for directory buckets. # # * This functionality is not supported for Amazon S3 on Outposts. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/S3_ACLs_UsingACLs.html # [2]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/about-object-ownership.html # @option options [String] :cache_control # Specifies the caching behavior along the request/reply chain. # @option options [String] :checksum_algorithm # Indicates the algorithm that you want Amazon S3 to use to create the # checksum for the object. For more information, see [Checking object # integrity][1] in the *Amazon S3 User Guide*. # # When you copy an object, if the source object has a checksum, that # checksum value will be copied to the new object by default. If the # `CopyObject` request does not include this `x-amz-checksum-algorithm` # header, the checksum algorithm will be copied from the source object # to the destination object (if it's present on the source object). You # can optionally specify a different checksum algorithm to use with the # `x-amz-checksum-algorithm` header. Unrecognized or unsupported values # will respond with the HTTP status code `400 Bad Request`. # # For directory buckets, when you use Amazon Web Services SDKs, `CRC32` # is the default checksum algorithm that's used for performance. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/checking-object-integrity.html # @option options [String] :content_disposition # Specifies presentational information for the object. Indicates whether # an object should be displayed in a web browser or downloaded as a # file. It allows specifying the desired filename for the downloaded # file. # @option options [String] :content_encoding # Specifies what content encodings have been applied to the object and # thus what decoding mechanisms must be applied to obtain the media-type # referenced by the Content-Type header field. # # For directory buckets, only the `aws-chunked` value is supported in # this header field. # # # @option options [String] :content_language # The language the content is in. # @option options [String] :content_type # A standard MIME type that describes the format of the object data. # @option options [required, String] :copy_source # Specifies the source object for the copy operation. The source object # can be up to 5 GB. If the source object is an object that was uploaded # by using a multipart upload, the object copy will be a single part # object after the source object is copied to the destination bucket. # # You specify the value of the copy source in one of two formats, # depending on whether you want to access the source object through an # [access point][1]: # # * For objects not accessed through an access point, specify the name # of the source bucket and the key of the source object, separated by # a slash (/). For example, to copy the object `reports/january.pdf` # from the general purpose bucket `awsexamplebucket`, use # `awsexamplebucket/reports/january.pdf`. The value must be # URL-encoded. To copy the object `reports/january.pdf` from the # directory bucket `awsexamplebucket--use1-az5--x-s3`, use # `awsexamplebucket--use1-az5--x-s3/reports/january.pdf`. The value # must be URL-encoded. # # * For objects accessed through access points, specify the Amazon # Resource Name (ARN) of the object as accessed through the access # point, in the format # `arn:aws:s3:::accesspoint//object/`. # For example, to copy the object `reports/january.pdf` through access # point `my-access-point` owned by account `123456789012` in Region # `us-west-2`, use the URL encoding of # `arn:aws:s3:us-west-2:123456789012:accesspoint/my-access-point/object/reports/january.pdf`. # The value must be URL encoded. # # * Amazon S3 supports copy operations using Access points only when # the source and destination buckets are in the same Amazon Web # Services Region. # # * Access points are not supported by directory buckets. # # # # Alternatively, for objects accessed through Amazon S3 on Outposts, # specify the ARN of the object as accessed in the format # `arn:aws:s3-outposts:::outpost//object/`. # For example, to copy the object `reports/january.pdf` through # outpost `my-outpost` owned by account `123456789012` in Region # `us-west-2`, use the URL encoding of # `arn:aws:s3-outposts:us-west-2:123456789012:outpost/my-outpost/object/reports/january.pdf`. # The value must be URL-encoded. # # If your source bucket versioning is enabled, the `x-amz-copy-source` # header by default identifies the current version of an object to copy. # If the current version is a delete marker, Amazon S3 behaves as if the # object was deleted. To copy a different version, use the `versionId` # query parameter. Specifically, append `?versionId=` to the # value (for example, # `awsexamplebucket/reports/january.pdf?versionId=QUpfdndhfd8438MNFDN93jdnJFkdmqnh893`). # If you don't specify a version ID, Amazon S3 copies the latest # version of the source object. # # If you enable versioning on the destination bucket, Amazon S3 # generates a unique version ID for the copied object. This version ID # is different from the version ID of the source object. Amazon S3 # returns the version ID of the copied object in the `x-amz-version-id` # response header in the response. # # If you do not enable versioning or suspend it on the destination # bucket, the version ID that Amazon S3 generates in the # `x-amz-version-id` response header is always null. # # **Directory buckets** - S3 Versioning isn't enabled and supported for # directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/access-points.html # @option options [String] :copy_source_if_match # Copies the object if its entity tag (ETag) matches the specified tag. # # If both the `x-amz-copy-source-if-match` and # `x-amz-copy-source-if-unmodified-since` headers are present in the # request and evaluate as follows, Amazon S3 returns `200 OK` and copies # the data: # # * `x-amz-copy-source-if-match` condition evaluates to true # # * `x-amz-copy-source-if-unmodified-since` condition evaluates to false # @option options [Time,DateTime,Date,Integer,String] :copy_source_if_modified_since # Copies the object if it has been modified since the specified time. # # If both the `x-amz-copy-source-if-none-match` and # `x-amz-copy-source-if-modified-since` headers are present in the # request and evaluate as follows, Amazon S3 returns the `412 # Precondition Failed` response code: # # * `x-amz-copy-source-if-none-match` condition evaluates to false # # * `x-amz-copy-source-if-modified-since` condition evaluates to true # @option options [String] :copy_source_if_none_match # Copies the object if its entity tag (ETag) is different than the # specified ETag. # # If both the `x-amz-copy-source-if-none-match` and # `x-amz-copy-source-if-modified-since` headers are present in the # request and evaluate as follows, Amazon S3 returns the `412 # Precondition Failed` response code: # # * `x-amz-copy-source-if-none-match` condition evaluates to false # # * `x-amz-copy-source-if-modified-since` condition evaluates to true # @option options [Time,DateTime,Date,Integer,String] :copy_source_if_unmodified_since # Copies the object if it hasn't been modified since the specified # time. # # If both the `x-amz-copy-source-if-match` and # `x-amz-copy-source-if-unmodified-since` headers are present in the # request and evaluate as follows, Amazon S3 returns `200 OK` and copies # the data: # # * `x-amz-copy-source-if-match` condition evaluates to true # # * `x-amz-copy-source-if-unmodified-since` condition evaluates to false # @option options [Time,DateTime,Date,Integer,String] :expires # The date and time at which the object is no longer cacheable. # @option options [String] :grant_full_control # Gives the grantee READ, READ\_ACP, and WRITE\_ACP permissions on the # object. # # * This functionality is not supported for directory buckets. # # * This functionality is not supported for Amazon S3 on Outposts. # # # @option options [String] :grant_read # Allows grantee to read the object data and its metadata. # # * This functionality is not supported for directory buckets. # # * This functionality is not supported for Amazon S3 on Outposts. # # # @option options [String] :grant_read_acp # Allows grantee to read the object ACL. # # * This functionality is not supported for directory buckets. # # * This functionality is not supported for Amazon S3 on Outposts. # # # @option options [String] :grant_write_acp # Allows grantee to write the ACL for the applicable object. # # * This functionality is not supported for directory buckets. # # * This functionality is not supported for Amazon S3 on Outposts. # # # @option options [Hash] :metadata # A map of metadata to store with the object in S3. # @option options [String] :metadata_directive # Specifies whether the metadata is copied from the source object or # replaced with metadata that's provided in the request. When copying # an object, you can preserve all metadata (the default) or specify new # metadata. If this header isn’t specified, `COPY` is the default # behavior. # # **General purpose bucket** - For general purpose buckets, when you # grant permissions, you can use the `s3:x-amz-metadata-directive` # condition key to enforce certain metadata behavior when objects are # uploaded. For more information, see [Amazon S3 condition key # examples][1] in the *Amazon S3 User Guide*. # # `x-amz-website-redirect-location` is unique to each object and is not # copied when using the `x-amz-metadata-directive` header. To copy the # value, you must specify `x-amz-website-redirect-location` in the # request header. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/amazon-s3-policy-keys.html # @option options [String] :tagging_directive # Specifies whether the object tag-set is copied from the source object # or replaced with the tag-set that's provided in the request. # # The default value is `COPY`. # # **Directory buckets** - For directory buckets in a `CopyObject` # operation, only the empty tag-set is supported. Any requests that # attempt to write non-empty tags into directory buckets will receive a # `501 Not Implemented` status code. When the destination bucket is a # directory bucket, you will receive a `501 Not Implemented` response in # any of the following situations: # # * When you attempt to `COPY` the tag-set from an S3 source object that # has non-empty tags. # # * When you attempt to `REPLACE` the tag-set of a source object and set # a non-empty value to `x-amz-tagging`. # # * When you don't set the `x-amz-tagging-directive` header and the # source object has non-empty tags. This is because the default value # of `x-amz-tagging-directive` is `COPY`. # # Because only the empty tag-set is supported for directory buckets in a # `CopyObject` operation, the following situations are allowed: # # * When you attempt to `COPY` the tag-set from a directory bucket # source object that has no tags to a general purpose bucket. It # copies an empty tag-set to the destination object. # # * When you attempt to `REPLACE` the tag-set of a directory bucket # source object and set the `x-amz-tagging` value of the directory # bucket destination object to empty. # # * When you attempt to `REPLACE` the tag-set of a general purpose # bucket source object that has non-empty tags and set the # `x-amz-tagging` value of the directory bucket destination object to # empty. # # * When you attempt to `REPLACE` the tag-set of a directory bucket # source object and don't set the `x-amz-tagging` value of the # directory bucket destination object. This is because the default # value of `x-amz-tagging` is the empty value. # # # @option options [String] :server_side_encryption # The server-side encryption algorithm used when storing this object in # Amazon S3. Unrecognized or unsupported values won’t write a # destination object and will receive a `400 Bad Request` response. # # Amazon S3 automatically encrypts all new objects that are copied to an # S3 bucket. When copying an object, if you don't specify encryption # information in your copy request, the encryption setting of the target # object is set to the default encryption configuration of the # destination bucket. By default, all buckets have a base level of # encryption configuration that uses server-side encryption with Amazon # S3 managed keys (SSE-S3). If the destination bucket has a different # default encryption configuration, Amazon S3 uses the corresponding # encryption key to encrypt the target object copy. # # With server-side encryption, Amazon S3 encrypts your data as it writes # your data to disks in its data centers and decrypts the data when you # access it. For more information about server-side encryption, see # [Using Server-Side Encryption][1] in the *Amazon S3 User Guide*. # # General purpose buckets # # * For general purpose buckets, there are the following supported # options for server-side encryption: server-side encryption with Key # Management Service (KMS) keys (SSE-KMS), dual-layer server-side # encryption with Amazon Web Services KMS keys (DSSE-KMS), and # server-side encryption with customer-provided encryption keys # (SSE-C). Amazon S3 uses the corresponding KMS key, or a # customer-provided key to encrypt the target object copy. # # * When you perform a `CopyObject` operation, if you want to use a # different type of encryption setting for the target object, you can # specify appropriate encryption-related headers to encrypt the target # object with an Amazon S3 managed key, a KMS key, or a # customer-provided key. If the encryption setting in your request is # different from the default encryption configuration of the # destination bucket, the encryption setting in your request takes # precedence. # # Directory buckets # # * For directory buckets, there are only two supported options for # server-side encryption: server-side encryption with Amazon S3 # managed keys (SSE-S3) (`AES256`) and server-side encryption with KMS # keys (SSE-KMS) (`aws:kms`). We recommend that the bucket's default # encryption uses the desired encryption configuration and you don't # override the bucket default encryption in your `CreateSession` # requests or `PUT` object requests. Then, new objects are # automatically encrypted with the desired encryption settings. For # more information, see [Protecting data with server-side # encryption][2] in the *Amazon S3 User Guide*. For more information # about the encryption overriding behaviors in directory buckets, see # [Specifying server-side encryption with KMS for new object # uploads][3]. # # * To encrypt new object copies to a directory bucket with SSE-KMS, we # recommend you specify SSE-KMS as the directory bucket's default # encryption configuration with a KMS key (specifically, a [customer # managed key][4]). The [Amazon Web Services managed key][5] # (`aws/s3`) isn't supported. Your SSE-KMS configuration can only # support 1 [customer managed key][4] per directory bucket for the # lifetime of the bucket. After you specify a customer managed key for # SSE-KMS, you can't override the customer managed key for the # bucket's SSE-KMS configuration. Then, when you perform a # `CopyObject` operation and want to specify server-side encryption # settings for new object copies with SSE-KMS in the # encryption-related request headers, you must ensure the encryption # key is the same customer managed key that you specified for the # directory bucket's default encryption configuration. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/serv-side-encryption.html # [2]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/s3-express-serv-side-encryption.html # [3]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/s3-express-specifying-kms-encryption.html # [4]: https://docs.aws.amazon.com/kms/latest/developerguide/concepts.html#customer-cmk # [5]: https://docs.aws.amazon.com/kms/latest/developerguide/concepts.html#aws-managed-cmk # @option options [String] :storage_class # If the `x-amz-storage-class` header is not used, the copied object # will be stored in the `STANDARD` Storage Class by default. The # `STANDARD` storage class provides high durability and high # availability. Depending on performance needs, you can specify a # different Storage Class. # # * Directory buckets - For directory buckets, only the S3 # Express One Zone storage class is supported to store newly created # objects. Unsupported storage class values won't write a destination # object and will respond with the HTTP status code `400 Bad Request`. # # * Amazon S3 on Outposts - S3 on Outposts only uses the # `OUTPOSTS` Storage Class. # # # # You can use the `CopyObject` action to change the storage class of an # object that is already stored in Amazon S3 by using the # `x-amz-storage-class` header. For more information, see [Storage # Classes][1] in the *Amazon S3 User Guide*. # # Before using an object as a source object for the copy operation, you # must restore a copy of it if it meets any of the following conditions: # # * The storage class of the source object is `GLACIER` or # `DEEP_ARCHIVE`. # # * The storage class of the source object is `INTELLIGENT_TIERING` and # it's [S3 Intelligent-Tiering access tier][2] is `Archive Access` or # `Deep Archive Access`. # # For more information, see [RestoreObject][3] and [Copying Objects][4] # in the *Amazon S3 User Guide*. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/storage-class-intro.html # [2]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/intelligent-tiering-overview.html#intel-tiering-tier-definition # [3]: https://docs.aws.amazon.com/AmazonS3/latest/API/API_RestoreObject.html # [4]: https://docs.aws.amazon.com/AmazonS3/latest/dev/CopyingObjectsExamples.html # @option options [String] :website_redirect_location # If the destination bucket is configured as a website, redirects # requests for this object copy to another object in the same bucket or # to an external URL. Amazon S3 stores the value of this header in the # object metadata. This value is unique to each object and is not copied # when using the `x-amz-metadata-directive` header. Instead, you may opt # to provide this header in combination with the # `x-amz-metadata-directive` header. # # This functionality is not supported for directory buckets. # # # @option options [String] :sse_customer_algorithm # Specifies the algorithm to use when encrypting the object (for # example, `AES256`). # # When you perform a `CopyObject` operation, if you want to use a # different type of encryption setting for the target object, you can # specify appropriate encryption-related headers to encrypt the target # object with an Amazon S3 managed key, a KMS key, or a # customer-provided key. If the encryption setting in your request is # different from the default encryption configuration of the destination # bucket, the encryption setting in your request takes precedence. # # This functionality is not supported when the destination bucket is a # directory bucket. # # # @option options [String] :sse_customer_key # Specifies the customer-provided encryption key for Amazon S3 to use in # encrypting data. This value is used to store the object and then it is # discarded. Amazon S3 does not store the encryption key. The key must # be appropriate for use with the algorithm specified in the # `x-amz-server-side-encryption-customer-algorithm` header. # # This functionality is not supported when the destination bucket is a # directory bucket. # # # @option options [String] :sse_customer_key_md5 # Specifies the 128-bit MD5 digest of the encryption key according to # RFC 1321. Amazon S3 uses this header for a message integrity check to # ensure that the encryption key was transmitted without error. # # This functionality is not supported when the destination bucket is a # directory bucket. # # # @option options [String] :ssekms_key_id # Specifies the KMS key ID (Key ID, Key ARN, or Key Alias) to use for # object encryption. All GET and PUT requests for an object protected by # KMS will fail if they're not made via SSL or using SigV4. For # information about configuring any of the officially supported Amazon # Web Services SDKs and Amazon Web Services CLI, see [Specifying the # Signature Version in Request Authentication][1] in the *Amazon S3 User # Guide*. # # **Directory buckets** - If you specify `x-amz-server-side-encryption` # with `aws:kms`, the ` x-amz-server-side-encryption-aws-kms-key-id` # header is implicitly assigned the ID of the KMS symmetric encryption # customer managed key that's configured for your directory bucket's # default encryption setting. If you want to specify the ` # x-amz-server-side-encryption-aws-kms-key-id` header explicitly, you # can only specify it with the ID (Key ID or Key ARN) of the KMS # customer managed key that's configured for your directory bucket's # default encryption setting. Otherwise, you get an HTTP `400 Bad # Request` error. Only use the key ID or key ARN. The key alias format # of the KMS key isn't supported. Your SSE-KMS configuration can only # support 1 [customer managed key][2] per directory bucket for the # lifetime of the bucket. The [Amazon Web Services managed key][3] # (`aws/s3`) isn't supported. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/UsingAWSSDK.html#specify-signature-version # [2]: https://docs.aws.amazon.com/kms/latest/developerguide/concepts.html#customer-cmk # [3]: https://docs.aws.amazon.com/kms/latest/developerguide/concepts.html#aws-managed-cmk # @option options [String] :ssekms_encryption_context # Specifies the Amazon Web Services KMS Encryption Context as an # additional encryption context to use for the destination object # encryption. The value of this header is a base64-encoded UTF-8 string # holding JSON with the encryption context key-value pairs. # # **General purpose buckets** - This value must be explicitly added to # specify encryption context for `CopyObject` requests if you want an # additional encryption context for your destination object. The # additional encryption context of the source object won't be copied to # the destination object. For more information, see [Encryption # context][1] in the *Amazon S3 User Guide*. # # **Directory buckets** - You can optionally provide an explicit # encryption context value. The value must match the default encryption # context - the bucket Amazon Resource Name (ARN). An additional # encryption context value is not supported. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/UsingKMSEncryption.html#encryption-context # @option options [Boolean] :bucket_key_enabled # Specifies whether Amazon S3 should use an S3 Bucket Key for object # encryption with server-side encryption using Key Management Service # (KMS) keys (SSE-KMS). If a target object uses SSE-KMS, you can enable # an S3 Bucket Key for the object. # # Setting this header to `true` causes Amazon S3 to use an S3 Bucket Key # for object encryption with SSE-KMS. Specifying this header with a COPY # action doesn’t affect bucket-level settings for S3 Bucket Key. # # For more information, see [Amazon S3 Bucket Keys][1] in the *Amazon S3 # User Guide*. # # **Directory buckets** - S3 Bucket Keys aren't supported, when you # copy SSE-KMS encrypted objects from general purpose buckets to # directory buckets, from directory buckets to general purpose buckets, # or between directory buckets, through [CopyObject][2]. In this case, # Amazon S3 makes a call to KMS every time a copy request is made for a # KMS-encrypted object. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/bucket-key.html # [2]: https://docs.aws.amazon.com/AmazonS3/latest/API/API_CopyObject.html # @option options [String] :copy_source_sse_customer_algorithm # Specifies the algorithm to use when decrypting the source object (for # example, `AES256`). # # If the source object for the copy is stored in Amazon S3 using SSE-C, # you must provide the necessary encryption information in your request # so that Amazon S3 can decrypt the object for copying. # # This functionality is not supported when the source object is in a # directory bucket. # # # @option options [String] :copy_source_sse_customer_key # Specifies the customer-provided encryption key for Amazon S3 to use to # decrypt the source object. The encryption key provided in this header # must be the same one that was used when the source object was created. # # If the source object for the copy is stored in Amazon S3 using SSE-C, # you must provide the necessary encryption information in your request # so that Amazon S3 can decrypt the object for copying. # # This functionality is not supported when the source object is in a # directory bucket. # # # @option options [String] :copy_source_sse_customer_key_md5 # Specifies the 128-bit MD5 digest of the encryption key according to # RFC 1321. Amazon S3 uses this header for a message integrity check to # ensure that the encryption key was transmitted without error. # # If the source object for the copy is stored in Amazon S3 using SSE-C, # you must provide the necessary encryption information in your request # so that Amazon S3 can decrypt the object for copying. # # This functionality is not supported when the source object is in a # directory bucket. # # # @option options [String] :request_payer # Confirms that the requester knows that they will be charged for the # request. Bucket owners need not specify this parameter in their # requests. If either the source or destination S3 bucket has Requester # Pays enabled, the requester will pay for corresponding charges to copy # the object. For information about downloading objects from Requester # Pays buckets, see [Downloading Objects in Requester Pays Buckets][1] # in the *Amazon S3 User Guide*. # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/ObjectsinRequesterPaysBuckets.html # @option options [String] :tagging # The tag-set for the object copy in the destination bucket. This value # must be used in conjunction with the `x-amz-tagging-directive` if you # choose `REPLACE` for the `x-amz-tagging-directive`. If you choose # `COPY` for the `x-amz-tagging-directive`, you don't need to set the # `x-amz-tagging` header, because the tag-set will be copied from the # source object directly. The tag-set must be encoded as URL Query # parameters. # # The default value is the empty value. # # **Directory buckets** - For directory buckets in a `CopyObject` # operation, only the empty tag-set is supported. Any requests that # attempt to write non-empty tags into directory buckets will receive a # `501 Not Implemented` status code. When the destination bucket is a # directory bucket, you will receive a `501 Not Implemented` response in # any of the following situations: # # * When you attempt to `COPY` the tag-set from an S3 source object that # has non-empty tags. # # * When you attempt to `REPLACE` the tag-set of a source object and set # a non-empty value to `x-amz-tagging`. # # * When you don't set the `x-amz-tagging-directive` header and the # source object has non-empty tags. This is because the default value # of `x-amz-tagging-directive` is `COPY`. # # Because only the empty tag-set is supported for directory buckets in a # `CopyObject` operation, the following situations are allowed: # # * When you attempt to `COPY` the tag-set from a directory bucket # source object that has no tags to a general purpose bucket. It # copies an empty tag-set to the destination object. # # * When you attempt to `REPLACE` the tag-set of a directory bucket # source object and set the `x-amz-tagging` value of the directory # bucket destination object to empty. # # * When you attempt to `REPLACE` the tag-set of a general purpose # bucket source object that has non-empty tags and set the # `x-amz-tagging` value of the directory bucket destination object to # empty. # # * When you attempt to `REPLACE` the tag-set of a directory bucket # source object and don't set the `x-amz-tagging` value of the # directory bucket destination object. This is because the default # value of `x-amz-tagging` is the empty value. # # # @option options [String] :object_lock_mode # The Object Lock mode that you want to apply to the object copy. # # This functionality is not supported for directory buckets. # # # @option options [Time,DateTime,Date,Integer,String] :object_lock_retain_until_date # The date and time when you want the Object Lock of the object copy to # expire. # # This functionality is not supported for directory buckets. # # # @option options [String] :object_lock_legal_hold_status # Specifies whether you want to apply a legal hold to the object copy. # # This functionality is not supported for directory buckets. # # # @option options [String] :expected_bucket_owner # The account ID of the expected destination bucket owner. If the # account ID that you provide does not match the actual owner of the # destination bucket, the request fails with the HTTP status code `403 # Forbidden` (access denied). # @option options [String] :expected_source_bucket_owner # The account ID of the expected source bucket owner. If the account ID # that you provide does not match the actual owner of the source bucket, # the request fails with the HTTP status code `403 Forbidden` (access # denied). # @return [Types::CopyObjectOutput] def copy_from(options = {}) options = options.merge( bucket: @bucket_name, key: @key ) resp = Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do @client.copy_object(options) end resp.data end # @example Request syntax with placeholder values # # object.delete({ # mfa: "MFA", # version_id: "ObjectVersionId", # request_payer: "requester", # accepts requester # bypass_governance_retention: false, # expected_bucket_owner: "AccountId", # if_match: "IfMatch", # if_match_last_modified_time: Time.now, # if_match_size: 1, # }) # @param [Hash] options ({}) # @option options [String] :mfa # The concatenation of the authentication device's serial number, a # space, and the value that is displayed on your authentication device. # Required to permanently delete a versioned object if versioning is # configured with MFA delete enabled. # # This functionality is not supported for directory buckets. # # # @option options [String] :version_id # Version ID used to reference a specific version of the object. # # For directory buckets in this API operation, only the `null` value of # the version ID is supported. # # # @option options [String] :request_payer # Confirms that the requester knows that they will be charged for the # request. Bucket owners need not specify this parameter in their # requests. If either the source or destination S3 bucket has Requester # Pays enabled, the requester will pay for corresponding charges to copy # the object. For information about downloading objects from Requester # Pays buckets, see [Downloading Objects in Requester Pays Buckets][1] # in the *Amazon S3 User Guide*. # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/ObjectsinRequesterPaysBuckets.html # @option options [Boolean] :bypass_governance_retention # Indicates whether S3 Object Lock should bypass Governance-mode # restrictions to process this operation. To use this header, you must # have the `s3:BypassGovernanceRetention` permission. # # This functionality is not supported for directory buckets. # # # @option options [String] :expected_bucket_owner # The account ID of the expected bucket owner. If the account ID that # you provide does not match the actual owner of the bucket, the request # fails with the HTTP status code `403 Forbidden` (access denied). # @option options [String] :if_match # The `If-Match` header field makes the request method conditional on # ETags. If the ETag value does not match, the operation returns a `412 # Precondition Failed` error. If the ETag matches or if the object # doesn't exist, the operation will return a `204 Success (No Content) # response`. # # For more information about conditional requests, see [RFC 7232][1]. # # This functionality is only supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/https:/tools.ietf.org/html/rfc7232 # @option options [Time,DateTime,Date,Integer,String] :if_match_last_modified_time # If present, the object is deleted only if its modification times # matches the provided `Timestamp`. If the `Timestamp` values do not # match, the operation returns a `412 Precondition Failed` error. If the # `Timestamp` matches or if the object doesn’t exist, the operation # returns a `204 Success (No Content)` response. # # This functionality is only supported for directory buckets. # # # @option options [Integer] :if_match_size # If present, the object is deleted only if its size matches the # provided size in bytes. If the `Size` value does not match, the # operation returns a `412 Precondition Failed` error. If the `Size` # matches or if the object doesn’t exist, the operation returns a `204 # Success (No Content)` response. # # This functionality is only supported for directory buckets. # # # # You can use the `If-Match`, `x-amz-if-match-last-modified-time` and # `x-amz-if-match-size` conditional headers in conjunction with # each-other or individually. # @return [Types::DeleteObjectOutput] def delete(options = {}) options = options.merge( bucket: @bucket_name, key: @key ) resp = Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do @client.delete_object(options) end resp.data end # @example Request syntax with placeholder values # # object.get({ # if_match: "IfMatch", # if_modified_since: Time.now, # if_none_match: "IfNoneMatch", # if_unmodified_since: Time.now, # range: "Range", # response_cache_control: "ResponseCacheControl", # response_content_disposition: "ResponseContentDisposition", # response_content_encoding: "ResponseContentEncoding", # response_content_language: "ResponseContentLanguage", # response_content_type: "ResponseContentType", # response_expires: Time.now, # version_id: "ObjectVersionId", # sse_customer_algorithm: "SSECustomerAlgorithm", # sse_customer_key: "SSECustomerKey", # sse_customer_key_md5: "SSECustomerKeyMD5", # request_payer: "requester", # accepts requester # part_number: 1, # expected_bucket_owner: "AccountId", # checksum_mode: "ENABLED", # accepts ENABLED # }) # @param [Hash] options ({}) # @option options [String] :if_match # Return the object only if its entity tag (ETag) is the same as the one # specified in this header; otherwise, return a `412 Precondition # Failed` error. # # If both of the `If-Match` and `If-Unmodified-Since` headers are # present in the request as follows: `If-Match` condition evaluates to # `true`, and; `If-Unmodified-Since` condition evaluates to `false`; # then, S3 returns `200 OK` and the data requested. # # For more information about conditional requests, see [RFC 7232][1]. # # # # [1]: https://tools.ietf.org/html/rfc7232 # @option options [Time,DateTime,Date,Integer,String] :if_modified_since # Return the object only if it has been modified since the specified # time; otherwise, return a `304 Not Modified` error. # # If both of the `If-None-Match` and `If-Modified-Since` headers are # present in the request as follows:` If-None-Match` condition evaluates # to `false`, and; `If-Modified-Since` condition evaluates to `true`; # then, S3 returns `304 Not Modified` status code. # # For more information about conditional requests, see [RFC 7232][1]. # # # # [1]: https://tools.ietf.org/html/rfc7232 # @option options [String] :if_none_match # Return the object only if its entity tag (ETag) is different from the # one specified in this header; otherwise, return a `304 Not Modified` # error. # # If both of the `If-None-Match` and `If-Modified-Since` headers are # present in the request as follows:` If-None-Match` condition evaluates # to `false`, and; `If-Modified-Since` condition evaluates to `true`; # then, S3 returns `304 Not Modified` HTTP status code. # # For more information about conditional requests, see [RFC 7232][1]. # # # # [1]: https://tools.ietf.org/html/rfc7232 # @option options [Time,DateTime,Date,Integer,String] :if_unmodified_since # Return the object only if it has not been modified since the specified # time; otherwise, return a `412 Precondition Failed` error. # # If both of the `If-Match` and `If-Unmodified-Since` headers are # present in the request as follows: `If-Match` condition evaluates to # `true`, and; `If-Unmodified-Since` condition evaluates to `false`; # then, S3 returns `200 OK` and the data requested. # # For more information about conditional requests, see [RFC 7232][1]. # # # # [1]: https://tools.ietf.org/html/rfc7232 # @option options [String] :range # Downloads the specified byte range of an object. For more information # about the HTTP Range header, see # [https://www.rfc-editor.org/rfc/rfc9110.html#name-range][1]. # # Amazon S3 doesn't support retrieving multiple ranges of data per # `GET` request. # # # # # # [1]: https://www.rfc-editor.org/rfc/rfc9110.html#name-range # @option options [String] :response_cache_control # Sets the `Cache-Control` header of the response. # @option options [String] :response_content_disposition # Sets the `Content-Disposition` header of the response. # @option options [String] :response_content_encoding # Sets the `Content-Encoding` header of the response. # @option options [String] :response_content_language # Sets the `Content-Language` header of the response. # @option options [String] :response_content_type # Sets the `Content-Type` header of the response. # @option options [Time,DateTime,Date,Integer,String] :response_expires # Sets the `Expires` header of the response. # @option options [String] :version_id # Version ID used to reference a specific version of the object. # # By default, the `GetObject` operation returns the current version of # an object. To return a different version, use the `versionId` # subresource. # # * If you include a `versionId` in your request header, you must have # the `s3:GetObjectVersion` permission to access a specific version of # an object. The `s3:GetObject` permission is not required in this # scenario. # # * If you request the current version of an object without a specific # `versionId` in the request header, only the `s3:GetObject` # permission is required. The `s3:GetObjectVersion` permission is not # required in this scenario. # # * **Directory buckets** - S3 Versioning isn't enabled and supported # for directory buckets. For this API operation, only the `null` value # of the version ID is supported by directory buckets. You can only # specify `null` to the `versionId` query parameter in the request. # # # # For more information about versioning, see [PutBucketVersioning][1]. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/API/API_PutBucketVersioning.html # @option options [String] :sse_customer_algorithm # Specifies the algorithm to use when decrypting the object (for # example, `AES256`). # # If you encrypt an object by using server-side encryption with # customer-provided encryption keys (SSE-C) when you store the object in # Amazon S3, then when you GET the object, you must use the following # headers: # # * `x-amz-server-side-encryption-customer-algorithm` # # * `x-amz-server-side-encryption-customer-key` # # * `x-amz-server-side-encryption-customer-key-MD5` # # For more information about SSE-C, see [Server-Side Encryption (Using # Customer-Provided Encryption Keys)][1] in the *Amazon S3 User Guide*. # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/ServerSideEncryptionCustomerKeys.html # @option options [String] :sse_customer_key # Specifies the customer-provided encryption key that you originally # provided for Amazon S3 to encrypt the data before storing it. This # value is used to decrypt the object when recovering it and must match # the one used when storing the data. The key must be appropriate for # use with the algorithm specified in the # `x-amz-server-side-encryption-customer-algorithm` header. # # If you encrypt an object by using server-side encryption with # customer-provided encryption keys (SSE-C) when you store the object in # Amazon S3, then when you GET the object, you must use the following # headers: # # * `x-amz-server-side-encryption-customer-algorithm` # # * `x-amz-server-side-encryption-customer-key` # # * `x-amz-server-side-encryption-customer-key-MD5` # # For more information about SSE-C, see [Server-Side Encryption (Using # Customer-Provided Encryption Keys)][1] in the *Amazon S3 User Guide*. # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/ServerSideEncryptionCustomerKeys.html # @option options [String] :sse_customer_key_md5 # Specifies the 128-bit MD5 digest of the customer-provided encryption # key according to RFC 1321. Amazon S3 uses this header for a message # integrity check to ensure that the encryption key was transmitted # without error. # # If you encrypt an object by using server-side encryption with # customer-provided encryption keys (SSE-C) when you store the object in # Amazon S3, then when you GET the object, you must use the following # headers: # # * `x-amz-server-side-encryption-customer-algorithm` # # * `x-amz-server-side-encryption-customer-key` # # * `x-amz-server-side-encryption-customer-key-MD5` # # For more information about SSE-C, see [Server-Side Encryption (Using # Customer-Provided Encryption Keys)][1] in the *Amazon S3 User Guide*. # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/ServerSideEncryptionCustomerKeys.html # @option options [String] :request_payer # Confirms that the requester knows that they will be charged for the # request. Bucket owners need not specify this parameter in their # requests. If either the source or destination S3 bucket has Requester # Pays enabled, the requester will pay for corresponding charges to copy # the object. For information about downloading objects from Requester # Pays buckets, see [Downloading Objects in Requester Pays Buckets][1] # in the *Amazon S3 User Guide*. # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/ObjectsinRequesterPaysBuckets.html # @option options [Integer] :part_number # Part number of the object being read. This is a positive integer # between 1 and 10,000. Effectively performs a 'ranged' GET request # for the part specified. Useful for downloading just a part of an # object. # @option options [String] :expected_bucket_owner # The account ID of the expected bucket owner. If the account ID that # you provide does not match the actual owner of the bucket, the request # fails with the HTTP status code `403 Forbidden` (access denied). # @option options [String] :checksum_mode # To retrieve the checksum, this mode must be enabled. # # **General purpose buckets** - In addition, if you enable checksum mode # and the object is uploaded with a [checksum][1] and encrypted with an # Key Management Service (KMS) key, you must have permission to use the # `kms:Decrypt` action to retrieve the checksum. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/API/API_Checksum.html # @return [Types::GetObjectOutput] def get(options = {}, &block) options = options.merge( bucket: @bucket_name, key: @key ) resp = Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do @client.get_object(options, &block) end resp.data end # @example Request syntax with placeholder values # # multipartupload = object.initiate_multipart_upload({ # acl: "private", # accepts private, public-read, public-read-write, authenticated-read, aws-exec-read, bucket-owner-read, bucket-owner-full-control # cache_control: "CacheControl", # content_disposition: "ContentDisposition", # content_encoding: "ContentEncoding", # content_language: "ContentLanguage", # content_type: "ContentType", # expires: Time.now, # grant_full_control: "GrantFullControl", # grant_read: "GrantRead", # grant_read_acp: "GrantReadACP", # grant_write_acp: "GrantWriteACP", # metadata: { # "MetadataKey" => "MetadataValue", # }, # server_side_encryption: "AES256", # accepts AES256, aws:kms, aws:kms:dsse # storage_class: "STANDARD", # accepts STANDARD, REDUCED_REDUNDANCY, STANDARD_IA, ONEZONE_IA, INTELLIGENT_TIERING, GLACIER, DEEP_ARCHIVE, OUTPOSTS, GLACIER_IR, SNOW, EXPRESS_ONEZONE # website_redirect_location: "WebsiteRedirectLocation", # sse_customer_algorithm: "SSECustomerAlgorithm", # sse_customer_key: "SSECustomerKey", # sse_customer_key_md5: "SSECustomerKeyMD5", # ssekms_key_id: "SSEKMSKeyId", # ssekms_encryption_context: "SSEKMSEncryptionContext", # bucket_key_enabled: false, # request_payer: "requester", # accepts requester # tagging: "TaggingHeader", # object_lock_mode: "GOVERNANCE", # accepts GOVERNANCE, COMPLIANCE # object_lock_retain_until_date: Time.now, # object_lock_legal_hold_status: "ON", # accepts ON, OFF # expected_bucket_owner: "AccountId", # checksum_algorithm: "CRC32", # accepts CRC32, CRC32C, SHA1, SHA256, CRC64NVME # checksum_type: "COMPOSITE", # accepts COMPOSITE, FULL_OBJECT # }) # @param [Hash] options ({}) # @option options [String] :acl # The canned ACL to apply to the object. Amazon S3 supports a set of # predefined ACLs, known as *canned ACLs*. Each canned ACL has a # predefined set of grantees and permissions. For more information, see # [Canned ACL][1] in the *Amazon S3 User Guide*. # # By default, all objects are private. Only the owner has full access # control. When uploading an object, you can grant access permissions to # individual Amazon Web Services accounts or to predefined groups # defined by Amazon S3. These permissions are then added to the access # control list (ACL) on the new object. For more information, see [Using # ACLs][2]. One way to grant the permissions using the request headers # is to specify a canned ACL with the `x-amz-acl` request header. # # * This functionality is not supported for directory buckets. # # * This functionality is not supported for Amazon S3 on Outposts. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/acl-overview.html#CannedACL # [2]: https://docs.aws.amazon.com/AmazonS3/latest/dev/S3_ACLs_UsingACLs.html # @option options [String] :cache_control # Specifies caching behavior along the request/reply chain. # @option options [String] :content_disposition # Specifies presentational information for the object. # @option options [String] :content_encoding # Specifies what content encodings have been applied to the object and # thus what decoding mechanisms must be applied to obtain the media-type # referenced by the Content-Type header field. # # For directory buckets, only the `aws-chunked` value is supported in # this header field. # # # @option options [String] :content_language # The language that the content is in. # @option options [String] :content_type # A standard MIME type describing the format of the object data. # @option options [Time,DateTime,Date,Integer,String] :expires # The date and time at which the object is no longer cacheable. # @option options [String] :grant_full_control # Specify access permissions explicitly to give the grantee READ, # READ\_ACP, and WRITE\_ACP permissions on the object. # # By default, all objects are private. Only the owner has full access # control. When uploading an object, you can use this header to # explicitly grant access permissions to specific Amazon Web Services # accounts or groups. This header maps to specific permissions that # Amazon S3 supports in an ACL. For more information, see [Access # Control List (ACL) Overview][1] in the *Amazon S3 User Guide*. # # You specify each grantee as a type=value pair, where the type is one # of the following: # # * `id` – if the value specified is the canonical user ID of an Amazon # Web Services account # # * `uri` – if you are granting permissions to a predefined group # # * `emailAddress` – if the value specified is the email address of an # Amazon Web Services account # # Using email addresses to specify a grantee is only supported in the # following Amazon Web Services Regions: # # * US East (N. Virginia) # # * US West (N. California) # # * US West (Oregon) # # * Asia Pacific (Singapore) # # * Asia Pacific (Sydney) # # * Asia Pacific (Tokyo) # # * Europe (Ireland) # # * South America (São Paulo) # # For a list of all the Amazon S3 supported Regions and endpoints, see # [Regions and Endpoints][2] in the Amazon Web Services General # Reference. # # # # For example, the following `x-amz-grant-read` header grants the Amazon # Web Services accounts identified by account IDs permissions to read # object data and its metadata: # # `x-amz-grant-read: id="11112222333", id="444455556666" ` # # * This functionality is not supported for directory buckets. # # * This functionality is not supported for Amazon S3 on Outposts. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/acl-overview.html # [2]: https://docs.aws.amazon.com/general/latest/gr/rande.html#s3_region # @option options [String] :grant_read # Specify access permissions explicitly to allow grantee to read the # object data and its metadata. # # By default, all objects are private. Only the owner has full access # control. When uploading an object, you can use this header to # explicitly grant access permissions to specific Amazon Web Services # accounts or groups. This header maps to specific permissions that # Amazon S3 supports in an ACL. For more information, see [Access # Control List (ACL) Overview][1] in the *Amazon S3 User Guide*. # # You specify each grantee as a type=value pair, where the type is one # of the following: # # * `id` – if the value specified is the canonical user ID of an Amazon # Web Services account # # * `uri` – if you are granting permissions to a predefined group # # * `emailAddress` – if the value specified is the email address of an # Amazon Web Services account # # Using email addresses to specify a grantee is only supported in the # following Amazon Web Services Regions: # # * US East (N. Virginia) # # * US West (N. California) # # * US West (Oregon) # # * Asia Pacific (Singapore) # # * Asia Pacific (Sydney) # # * Asia Pacific (Tokyo) # # * Europe (Ireland) # # * South America (São Paulo) # # For a list of all the Amazon S3 supported Regions and endpoints, see # [Regions and Endpoints][2] in the Amazon Web Services General # Reference. # # # # For example, the following `x-amz-grant-read` header grants the Amazon # Web Services accounts identified by account IDs permissions to read # object data and its metadata: # # `x-amz-grant-read: id="11112222333", id="444455556666" ` # # * This functionality is not supported for directory buckets. # # * This functionality is not supported for Amazon S3 on Outposts. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/acl-overview.html # [2]: https://docs.aws.amazon.com/general/latest/gr/rande.html#s3_region # @option options [String] :grant_read_acp # Specify access permissions explicitly to allows grantee to read the # object ACL. # # By default, all objects are private. Only the owner has full access # control. When uploading an object, you can use this header to # explicitly grant access permissions to specific Amazon Web Services # accounts or groups. This header maps to specific permissions that # Amazon S3 supports in an ACL. For more information, see [Access # Control List (ACL) Overview][1] in the *Amazon S3 User Guide*. # # You specify each grantee as a type=value pair, where the type is one # of the following: # # * `id` – if the value specified is the canonical user ID of an Amazon # Web Services account # # * `uri` – if you are granting permissions to a predefined group # # * `emailAddress` – if the value specified is the email address of an # Amazon Web Services account # # Using email addresses to specify a grantee is only supported in the # following Amazon Web Services Regions: # # * US East (N. Virginia) # # * US West (N. California) # # * US West (Oregon) # # * Asia Pacific (Singapore) # # * Asia Pacific (Sydney) # # * Asia Pacific (Tokyo) # # * Europe (Ireland) # # * South America (São Paulo) # # For a list of all the Amazon S3 supported Regions and endpoints, see # [Regions and Endpoints][2] in the Amazon Web Services General # Reference. # # # # For example, the following `x-amz-grant-read` header grants the Amazon # Web Services accounts identified by account IDs permissions to read # object data and its metadata: # # `x-amz-grant-read: id="11112222333", id="444455556666" ` # # * This functionality is not supported for directory buckets. # # * This functionality is not supported for Amazon S3 on Outposts. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/acl-overview.html # [2]: https://docs.aws.amazon.com/general/latest/gr/rande.html#s3_region # @option options [String] :grant_write_acp # Specify access permissions explicitly to allows grantee to allow # grantee to write the ACL for the applicable object. # # By default, all objects are private. Only the owner has full access # control. When uploading an object, you can use this header to # explicitly grant access permissions to specific Amazon Web Services # accounts or groups. This header maps to specific permissions that # Amazon S3 supports in an ACL. For more information, see [Access # Control List (ACL) Overview][1] in the *Amazon S3 User Guide*. # # You specify each grantee as a type=value pair, where the type is one # of the following: # # * `id` – if the value specified is the canonical user ID of an Amazon # Web Services account # # * `uri` – if you are granting permissions to a predefined group # # * `emailAddress` – if the value specified is the email address of an # Amazon Web Services account # # Using email addresses to specify a grantee is only supported in the # following Amazon Web Services Regions: # # * US East (N. Virginia) # # * US West (N. California) # # * US West (Oregon) # # * Asia Pacific (Singapore) # # * Asia Pacific (Sydney) # # * Asia Pacific (Tokyo) # # * Europe (Ireland) # # * South America (São Paulo) # # For a list of all the Amazon S3 supported Regions and endpoints, see # [Regions and Endpoints][2] in the Amazon Web Services General # Reference. # # # # For example, the following `x-amz-grant-read` header grants the Amazon # Web Services accounts identified by account IDs permissions to read # object data and its metadata: # # `x-amz-grant-read: id="11112222333", id="444455556666" ` # # * This functionality is not supported for directory buckets. # # * This functionality is not supported for Amazon S3 on Outposts. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/acl-overview.html # [2]: https://docs.aws.amazon.com/general/latest/gr/rande.html#s3_region # @option options [Hash] :metadata # A map of metadata to store with the object in S3. # @option options [String] :server_side_encryption # The server-side encryption algorithm used when you store this object # in Amazon S3 (for example, `AES256`, `aws:kms`). # # * Directory buckets - For directory buckets, there are only # two supported options for server-side encryption: server-side # encryption with Amazon S3 managed keys (SSE-S3) (`AES256`) and # server-side encryption with KMS keys (SSE-KMS) (`aws:kms`). We # recommend that the bucket's default encryption uses the desired # encryption configuration and you don't override the bucket default # encryption in your `CreateSession` requests or `PUT` object # requests. Then, new objects are automatically encrypted with the # desired encryption settings. For more information, see [Protecting # data with server-side encryption][1] in the *Amazon S3 User Guide*. # For more information about the encryption overriding behaviors in # directory buckets, see [Specifying server-side encryption with KMS # for new object uploads][2]. # # In the Zonal endpoint API calls (except [CopyObject][3] and # [UploadPartCopy][4]) using the REST API, the encryption request # headers must match the encryption settings that are specified in the # `CreateSession` request. You can't override the values of the # encryption settings (`x-amz-server-side-encryption`, # `x-amz-server-side-encryption-aws-kms-key-id`, # `x-amz-server-side-encryption-context`, and # `x-amz-server-side-encryption-bucket-key-enabled`) that are # specified in the `CreateSession` request. You don't need to # explicitly specify these encryption settings values in Zonal # endpoint API calls, and Amazon S3 will use the encryption settings # values from the `CreateSession` request to protect new objects in # the directory bucket. # # When you use the CLI or the Amazon Web Services SDKs, for # `CreateSession`, the session token refreshes automatically to avoid # service interruptions when a session expires. The CLI or the Amazon # Web Services SDKs use the bucket's default encryption configuration # for the `CreateSession` request. It's not supported to override the # encryption settings values in the `CreateSession` request. So in the # Zonal endpoint API calls (except [CopyObject][3] and # [UploadPartCopy][4]), the encryption request headers must match the # default encryption configuration of the directory bucket. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/s3-express-serv-side-encryption.html # [2]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/s3-express-specifying-kms-encryption.html # [3]: https://docs.aws.amazon.com/AmazonS3/latest/API/API_CopyObject.html # [4]: https://docs.aws.amazon.com/AmazonS3/latest/API/API_UploadPartCopy.html # @option options [String] :storage_class # By default, Amazon S3 uses the STANDARD Storage Class to store newly # created objects. The STANDARD storage class provides high durability # and high availability. Depending on performance needs, you can specify # a different Storage Class. For more information, see [Storage # Classes][1] in the *Amazon S3 User Guide*. # # * For directory buckets, only the S3 Express One Zone storage class is # supported to store newly created objects. # # * Amazon S3 on Outposts only uses the OUTPOSTS Storage Class. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/storage-class-intro.html # @option options [String] :website_redirect_location # If the bucket is configured as a website, redirects requests for this # object to another object in the same bucket or to an external URL. # Amazon S3 stores the value of this header in the object metadata. # # This functionality is not supported for directory buckets. # # # @option options [String] :sse_customer_algorithm # Specifies the algorithm to use when encrypting the object (for # example, AES256). # # This functionality is not supported for directory buckets. # # # @option options [String] :sse_customer_key # Specifies the customer-provided encryption key for Amazon S3 to use in # encrypting data. This value is used to store the object and then it is # discarded; Amazon S3 does not store the encryption key. The key must # be appropriate for use with the algorithm specified in the # `x-amz-server-side-encryption-customer-algorithm` header. # # This functionality is not supported for directory buckets. # # # @option options [String] :sse_customer_key_md5 # Specifies the 128-bit MD5 digest of the customer-provided encryption # key according to RFC 1321. Amazon S3 uses this header for a message # integrity check to ensure that the encryption key was transmitted # without error. # # This functionality is not supported for directory buckets. # # # @option options [String] :ssekms_key_id # Specifies the KMS key ID (Key ID, Key ARN, or Key Alias) to use for # object encryption. If the KMS key doesn't exist in the same account # that's issuing the command, you must use the full Key ARN not the Key # ID. # # **General purpose buckets** - If you specify # `x-amz-server-side-encryption` with `aws:kms` or `aws:kms:dsse`, this # header specifies the ID (Key ID, Key ARN, or Key Alias) of the KMS key # to use. If you specify `x-amz-server-side-encryption:aws:kms` or # `x-amz-server-side-encryption:aws:kms:dsse`, but do not provide # `x-amz-server-side-encryption-aws-kms-key-id`, Amazon S3 uses the # Amazon Web Services managed key (`aws/s3`) to protect the data. # # **Directory buckets** - If you specify `x-amz-server-side-encryption` # with `aws:kms`, the ` x-amz-server-side-encryption-aws-kms-key-id` # header is implicitly assigned the ID of the KMS symmetric encryption # customer managed key that's configured for your directory bucket's # default encryption setting. If you want to specify the ` # x-amz-server-side-encryption-aws-kms-key-id` header explicitly, you # can only specify it with the ID (Key ID or Key ARN) of the KMS # customer managed key that's configured for your directory bucket's # default encryption setting. Otherwise, you get an HTTP `400 Bad # Request` error. Only use the key ID or key ARN. The key alias format # of the KMS key isn't supported. Your SSE-KMS configuration can only # support 1 [customer managed key][1] per directory bucket for the # lifetime of the bucket. The [Amazon Web Services managed key][2] # (`aws/s3`) isn't supported. # # # # [1]: https://docs.aws.amazon.com/kms/latest/developerguide/concepts.html#customer-cmk # [2]: https://docs.aws.amazon.com/kms/latest/developerguide/concepts.html#aws-managed-cmk # @option options [String] :ssekms_encryption_context # Specifies the Amazon Web Services KMS Encryption Context to use for # object encryption. The value of this header is a Base64 encoded string # of a UTF-8 encoded JSON, which contains the encryption context as # key-value pairs. # # **Directory buckets** - You can optionally provide an explicit # encryption context value. The value must match the default encryption # context - the bucket Amazon Resource Name (ARN). An additional # encryption context value is not supported. # @option options [Boolean] :bucket_key_enabled # Specifies whether Amazon S3 should use an S3 Bucket Key for object # encryption with server-side encryption using Key Management Service # (KMS) keys (SSE-KMS). # # **General purpose buckets** - Setting this header to `true` causes # Amazon S3 to use an S3 Bucket Key for object encryption with SSE-KMS. # Also, specifying this header with a PUT action doesn't affect # bucket-level settings for S3 Bucket Key. # # **Directory buckets** - S3 Bucket Keys are always enabled for `GET` # and `PUT` operations in a directory bucket and can’t be disabled. S3 # Bucket Keys aren't supported, when you copy SSE-KMS encrypted objects # from general purpose buckets to directory buckets, from directory # buckets to general purpose buckets, or between directory buckets, # through [CopyObject][1], [UploadPartCopy][2], [the Copy operation in # Batch Operations][3], or [the import jobs][4]. In this case, Amazon S3 # makes a call to KMS every time a copy request is made for a # KMS-encrypted object. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/API/API_CopyObject.html # [2]: https://docs.aws.amazon.com/AmazonS3/latest/API/API_UploadPartCopy.html # [3]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/directory-buckets-objects-Batch-Ops # [4]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/create-import-job # @option options [String] :request_payer # Confirms that the requester knows that they will be charged for the # request. Bucket owners need not specify this parameter in their # requests. If either the source or destination S3 bucket has Requester # Pays enabled, the requester will pay for corresponding charges to copy # the object. For information about downloading objects from Requester # Pays buckets, see [Downloading Objects in Requester Pays Buckets][1] # in the *Amazon S3 User Guide*. # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/ObjectsinRequesterPaysBuckets.html # @option options [String] :tagging # The tag-set for the object. The tag-set must be encoded as URL Query # parameters. # # This functionality is not supported for directory buckets. # # # @option options [String] :object_lock_mode # Specifies the Object Lock mode that you want to apply to the uploaded # object. # # This functionality is not supported for directory buckets. # # # @option options [Time,DateTime,Date,Integer,String] :object_lock_retain_until_date # Specifies the date and time when you want the Object Lock to expire. # # This functionality is not supported for directory buckets. # # # @option options [String] :object_lock_legal_hold_status # Specifies whether you want to apply a legal hold to the uploaded # object. # # This functionality is not supported for directory buckets. # # # @option options [String] :expected_bucket_owner # The account ID of the expected bucket owner. If the account ID that # you provide does not match the actual owner of the bucket, the request # fails with the HTTP status code `403 Forbidden` (access denied). # @option options [String] :checksum_algorithm # Indicates the algorithm that you want Amazon S3 to use to create the # checksum for the object. For more information, see [Checking object # integrity][1] in the *Amazon S3 User Guide*. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/checking-object-integrity.html # @option options [String] :checksum_type # Indicates the checksum type that you want Amazon S3 to use to # calculate the object’s checksum value. For more information, see # [Checking object integrity in the Amazon S3 User Guide][1]. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/checking-object-integrity.html # @return [MultipartUpload] def initiate_multipart_upload(options = {}) options = options.merge( bucket: @bucket_name, key: @key ) resp = Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do @client.create_multipart_upload(options) end MultipartUpload.new( bucket_name: @bucket_name, object_key: @key, id: resp.data.upload_id, client: @client ) end # @example Request syntax with placeholder values # # object.put({ # acl: "private", # accepts private, public-read, public-read-write, authenticated-read, aws-exec-read, bucket-owner-read, bucket-owner-full-control # body: source_file, # cache_control: "CacheControl", # content_disposition: "ContentDisposition", # content_encoding: "ContentEncoding", # content_language: "ContentLanguage", # content_length: 1, # content_md5: "ContentMD5", # content_type: "ContentType", # checksum_algorithm: "CRC32", # accepts CRC32, CRC32C, SHA1, SHA256, CRC64NVME # checksum_crc32: "ChecksumCRC32", # checksum_crc32c: "ChecksumCRC32C", # checksum_crc64nvme: "ChecksumCRC64NVME", # checksum_sha1: "ChecksumSHA1", # checksum_sha256: "ChecksumSHA256", # expires: Time.now, # if_match: "IfMatch", # if_none_match: "IfNoneMatch", # grant_full_control: "GrantFullControl", # grant_read: "GrantRead", # grant_read_acp: "GrantReadACP", # grant_write_acp: "GrantWriteACP", # write_offset_bytes: 1, # metadata: { # "MetadataKey" => "MetadataValue", # }, # server_side_encryption: "AES256", # accepts AES256, aws:kms, aws:kms:dsse # storage_class: "STANDARD", # accepts STANDARD, REDUCED_REDUNDANCY, STANDARD_IA, ONEZONE_IA, INTELLIGENT_TIERING, GLACIER, DEEP_ARCHIVE, OUTPOSTS, GLACIER_IR, SNOW, EXPRESS_ONEZONE # website_redirect_location: "WebsiteRedirectLocation", # sse_customer_algorithm: "SSECustomerAlgorithm", # sse_customer_key: "SSECustomerKey", # sse_customer_key_md5: "SSECustomerKeyMD5", # ssekms_key_id: "SSEKMSKeyId", # ssekms_encryption_context: "SSEKMSEncryptionContext", # bucket_key_enabled: false, # request_payer: "requester", # accepts requester # tagging: "TaggingHeader", # object_lock_mode: "GOVERNANCE", # accepts GOVERNANCE, COMPLIANCE # object_lock_retain_until_date: Time.now, # object_lock_legal_hold_status: "ON", # accepts ON, OFF # expected_bucket_owner: "AccountId", # }) # @param [Hash] options ({}) # @option options [String] :acl # The canned ACL to apply to the object. For more information, see # [Canned ACL][1] in the *Amazon S3 User Guide*. # # When adding a new object, you can use headers to grant ACL-based # permissions to individual Amazon Web Services accounts or to # predefined groups defined by Amazon S3. These permissions are then # added to the ACL on the object. By default, all objects are private. # Only the owner has full access control. For more information, see # [Access Control List (ACL) Overview][2] and [Managing ACLs Using the # REST API][3] in the *Amazon S3 User Guide*. # # If the bucket that you're uploading objects to uses the bucket owner # enforced setting for S3 Object Ownership, ACLs are disabled and no # longer affect permissions. Buckets that use this setting only accept # PUT requests that don't specify an ACL or PUT requests that specify # bucket owner full control ACLs, such as the # `bucket-owner-full-control` canned ACL or an equivalent form of this # ACL expressed in the XML format. PUT requests that contain other ACLs # (for example, custom grants to certain Amazon Web Services accounts) # fail and return a `400` error with the error code # `AccessControlListNotSupported`. For more information, see [ # Controlling ownership of objects and disabling ACLs][4] in the *Amazon # S3 User Guide*. # # * This functionality is not supported for directory buckets. # # * This functionality is not supported for Amazon S3 on Outposts. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/acl-overview.html#CannedACL # [2]: https://docs.aws.amazon.com/AmazonS3/latest/dev/acl-overview.html # [3]: https://docs.aws.amazon.com/AmazonS3/latest/dev/acl-using-rest-api.html # [4]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/about-object-ownership.html # @option options [String, StringIO, File] :body # Object data. # @option options [String] :cache_control # Can be used to specify caching behavior along the request/reply chain. # For more information, see # [http://www.w3.org/Protocols/rfc2616/rfc2616-sec14.html#sec14.9][1]. # # # # [1]: http://www.w3.org/Protocols/rfc2616/rfc2616-sec14.html#sec14.9 # @option options [String] :content_disposition # Specifies presentational information for the object. For more # information, see # [https://www.rfc-editor.org/rfc/rfc6266#section-4][1]. # # # # [1]: https://www.rfc-editor.org/rfc/rfc6266#section-4 # @option options [String] :content_encoding # Specifies what content encodings have been applied to the object and # thus what decoding mechanisms must be applied to obtain the media-type # referenced by the Content-Type header field. For more information, see # [https://www.rfc-editor.org/rfc/rfc9110.html#field.content-encoding][1]. # # # # [1]: https://www.rfc-editor.org/rfc/rfc9110.html#field.content-encoding # @option options [String] :content_language # The language the content is in. # @option options [Integer] :content_length # Size of the body in bytes. This parameter is useful when the size of # the body cannot be determined automatically. For more information, see # [https://www.rfc-editor.org/rfc/rfc9110.html#name-content-length][1]. # # # # [1]: https://www.rfc-editor.org/rfc/rfc9110.html#name-content-length # @option options [String] :content_md5 # The Base64 encoded 128-bit `MD5` digest of the message (without the # headers) according to RFC 1864. This header can be used as a message # integrity check to verify that the data is the same data that was # originally sent. Although it is optional, we recommend using the # Content-MD5 mechanism as an end-to-end integrity check. For more # information about REST request authentication, see [REST # Authentication][1]. # # The `Content-MD5` or `x-amz-sdk-checksum-algorithm` header is required # for any request to upload an object with a retention period configured # using Amazon S3 Object Lock. For more information, see [Uploading # objects to an Object Lock enabled bucket ][2] in the *Amazon S3 User # Guide*. # # # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/RESTAuthentication.html # [2]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/object-lock-managing.html#object-lock-put-object # @option options [String] :content_type # A standard MIME type describing the format of the contents. For more # information, see # [https://www.rfc-editor.org/rfc/rfc9110.html#name-content-type][1]. # # # # [1]: https://www.rfc-editor.org/rfc/rfc9110.html#name-content-type # @option options [String] :checksum_algorithm # Indicates the algorithm used to create the checksum for the object # when you use the SDK. This header will not provide any additional # functionality if you don't use the SDK. When you send this header, # there must be a corresponding `x-amz-checksum-algorithm ` or # `x-amz-trailer` header sent. Otherwise, Amazon S3 fails the request # with the HTTP status code `400 Bad Request`. # # For the `x-amz-checksum-algorithm ` header, replace ` algorithm ` with # the supported algorithm from the following list: # # * `CRC-32` # # * `CRC-32C` # # * `CRC-64NVME` # # * `SHA-1` # # * `SHA-256` # # For more information, see [Checking object integrity][1] in the # *Amazon S3 User Guide*. # # If the individual checksum value you provide through # `x-amz-checksum-algorithm ` doesn't match the checksum algorithm you # set through `x-amz-sdk-checksum-algorithm`, Amazon S3 fails the # request with a `BadDigest` error. # # The `Content-MD5` or `x-amz-sdk-checksum-algorithm` header is required # for any request to upload an object with a retention period configured # using Amazon S3 Object Lock. For more information, see [Uploading # objects to an Object Lock enabled bucket ][2] in the *Amazon S3 User # Guide*. # # # # For directory buckets, when you use Amazon Web Services SDKs, `CRC32` # is the default checksum algorithm that's used for performance. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/checking-object-integrity.html # [2]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/object-lock-managing.html#object-lock-put-object # @option options [String] :checksum_crc32 # This header can be used as a data integrity check to verify that the # data received is the same data that was originally sent. This header # specifies the Base64 encoded, 32-bit `CRC-32` checksum of the object. # For more information, see [Checking object integrity][1] in the # *Amazon S3 User Guide*. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/checking-object-integrity.html # @option options [String] :checksum_crc32c # This header can be used as a data integrity check to verify that the # data received is the same data that was originally sent. This header # specifies the Base64 encoded, 32-bit `CRC-32C` checksum of the object. # For more information, see [Checking object integrity][1] in the # *Amazon S3 User Guide*. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/checking-object-integrity.html # @option options [String] :checksum_crc64nvme # This header can be used as a data integrity check to verify that the # data received is the same data that was originally sent. This header # specifies the Base64 encoded, 64-bit `CRC-64NVME` checksum of the # object. The `CRC-64NVME` checksum is always a full object checksum. # For more information, see [Checking object integrity in the Amazon S3 # User Guide][1]. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/checking-object-integrity.html # @option options [String] :checksum_sha1 # This header can be used as a data integrity check to verify that the # data received is the same data that was originally sent. This header # specifies the Base64 encoded, 160-bit `SHA-1` digest of the object. # For more information, see [Checking object integrity][1] in the # *Amazon S3 User Guide*. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/checking-object-integrity.html # @option options [String] :checksum_sha256 # This header can be used as a data integrity check to verify that the # data received is the same data that was originally sent. This header # specifies the Base64 encoded, 256-bit `SHA-256` digest of the object. # For more information, see [Checking object integrity][1] in the # *Amazon S3 User Guide*. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/checking-object-integrity.html # @option options [Time,DateTime,Date,Integer,String] :expires # The date and time at which the object is no longer cacheable. For more # information, see # [https://www.rfc-editor.org/rfc/rfc7234#section-5.3][1]. # # # # [1]: https://www.rfc-editor.org/rfc/rfc7234#section-5.3 # @option options [String] :if_match # Uploads the object only if the ETag (entity tag) value provided during # the WRITE operation matches the ETag of the object in S3. If the ETag # values do not match, the operation returns a `412 Precondition Failed` # error. # # If a conflicting operation occurs during the upload S3 returns a `409 # ConditionalRequestConflict` response. On a 409 failure you should # fetch the object's ETag and retry the upload. # # Expects the ETag value as a string. # # For more information about conditional requests, see [RFC 7232][1], or # [Conditional requests][2] in the *Amazon S3 User Guide*. # # # # [1]: https://tools.ietf.org/html/rfc7232 # [2]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/conditional-requests.html # @option options [String] :if_none_match # Uploads the object only if the object key name does not already exist # in the bucket specified. Otherwise, Amazon S3 returns a `412 # Precondition Failed` error. # # If a conflicting operation occurs during the upload S3 returns a `409 # ConditionalRequestConflict` response. On a 409 failure you should # retry the upload. # # Expects the '*' (asterisk) character. # # For more information about conditional requests, see [RFC 7232][1], or # [Conditional requests][2] in the *Amazon S3 User Guide*. # # # # [1]: https://tools.ietf.org/html/rfc7232 # [2]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/conditional-requests.html # @option options [String] :grant_full_control # Gives the grantee READ, READ\_ACP, and WRITE\_ACP permissions on the # object. # # * This functionality is not supported for directory buckets. # # * This functionality is not supported for Amazon S3 on Outposts. # # # @option options [String] :grant_read # Allows grantee to read the object data and its metadata. # # * This functionality is not supported for directory buckets. # # * This functionality is not supported for Amazon S3 on Outposts. # # # @option options [String] :grant_read_acp # Allows grantee to read the object ACL. # # * This functionality is not supported for directory buckets. # # * This functionality is not supported for Amazon S3 on Outposts. # # # @option options [String] :grant_write_acp # Allows grantee to write the ACL for the applicable object. # # * This functionality is not supported for directory buckets. # # * This functionality is not supported for Amazon S3 on Outposts. # # # @option options [Integer] :write_offset_bytes # Specifies the offset for appending data to existing objects in bytes. # The offset must be equal to the size of the existing object being # appended to. If no object exists, setting this header to 0 will create # a new object. # # This functionality is only supported for objects in the Amazon S3 # Express One Zone storage class in directory buckets. # # # @option options [Hash] :metadata # A map of metadata to store with the object in S3. # @option options [String] :server_side_encryption # The server-side encryption algorithm that was used when you store this # object in Amazon S3 (for example, `AES256`, `aws:kms`, # `aws:kms:dsse`). # # * General purpose buckets - You have four mutually exclusive # options to protect data using server-side encryption in Amazon S3, # depending on how you choose to manage the encryption keys. # Specifically, the encryption key options are Amazon S3 managed keys # (SSE-S3), Amazon Web Services KMS keys (SSE-KMS or DSSE-KMS), and # customer-provided keys (SSE-C). Amazon S3 encrypts data with # server-side encryption by using Amazon S3 managed keys (SSE-S3) by # default. You can optionally tell Amazon S3 to encrypt data at rest # by using server-side encryption with other key options. For more # information, see [Using Server-Side Encryption][1] in the *Amazon S3 # User Guide*. # # * Directory buckets - For directory buckets, there are only # two supported options for server-side encryption: server-side # encryption with Amazon S3 managed keys (SSE-S3) (`AES256`) and # server-side encryption with KMS keys (SSE-KMS) (`aws:kms`). We # recommend that the bucket's default encryption uses the desired # encryption configuration and you don't override the bucket default # encryption in your `CreateSession` requests or `PUT` object # requests. Then, new objects are automatically encrypted with the # desired encryption settings. For more information, see [Protecting # data with server-side encryption][2] in the *Amazon S3 User Guide*. # For more information about the encryption overriding behaviors in # directory buckets, see [Specifying server-side encryption with KMS # for new object uploads][3]. # # In the Zonal endpoint API calls (except [CopyObject][4] and # [UploadPartCopy][5]) using the REST API, the encryption request # headers must match the encryption settings that are specified in the # `CreateSession` request. You can't override the values of the # encryption settings (`x-amz-server-side-encryption`, # `x-amz-server-side-encryption-aws-kms-key-id`, # `x-amz-server-side-encryption-context`, and # `x-amz-server-side-encryption-bucket-key-enabled`) that are # specified in the `CreateSession` request. You don't need to # explicitly specify these encryption settings values in Zonal # endpoint API calls, and Amazon S3 will use the encryption settings # values from the `CreateSession` request to protect new objects in # the directory bucket. # # When you use the CLI or the Amazon Web Services SDKs, for # `CreateSession`, the session token refreshes automatically to avoid # service interruptions when a session expires. The CLI or the Amazon # Web Services SDKs use the bucket's default encryption configuration # for the `CreateSession` request. It's not supported to override the # encryption settings values in the `CreateSession` request. So in the # Zonal endpoint API calls (except [CopyObject][4] and # [UploadPartCopy][5]), the encryption request headers must match the # default encryption configuration of the directory bucket. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/UsingServerSideEncryption.html # [2]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/s3-express-serv-side-encryption.html # [3]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/s3-express-specifying-kms-encryption.html # [4]: https://docs.aws.amazon.com/AmazonS3/latest/API/API_CopyObject.html # [5]: https://docs.aws.amazon.com/AmazonS3/latest/API/API_UploadPartCopy.html # @option options [String] :storage_class # By default, Amazon S3 uses the STANDARD Storage Class to store newly # created objects. The STANDARD storage class provides high durability # and high availability. Depending on performance needs, you can specify # a different Storage Class. For more information, see [Storage # Classes][1] in the *Amazon S3 User Guide*. # # * For directory buckets, only the S3 Express One Zone storage class is # supported to store newly created objects. # # * Amazon S3 on Outposts only uses the OUTPOSTS Storage Class. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/storage-class-intro.html # @option options [String] :website_redirect_location # If the bucket is configured as a website, redirects requests for this # object to another object in the same bucket or to an external URL. # Amazon S3 stores the value of this header in the object metadata. For # information about object metadata, see [Object Key and Metadata][1] in # the *Amazon S3 User Guide*. # # In the following example, the request header sets the redirect to an # object (anotherPage.html) in the same bucket: # # `x-amz-website-redirect-location: /anotherPage.html` # # In the following example, the request header sets the object redirect # to another website: # # `x-amz-website-redirect-location: http://www.example.com/` # # For more information about website hosting in Amazon S3, see [Hosting # Websites on Amazon S3][2] and [How to Configure Website Page # Redirects][3] in the *Amazon S3 User Guide*. # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/UsingMetadata.html # [2]: https://docs.aws.amazon.com/AmazonS3/latest/dev/WebsiteHosting.html # [3]: https://docs.aws.amazon.com/AmazonS3/latest/dev/how-to-page-redirect.html # @option options [String] :sse_customer_algorithm # Specifies the algorithm to use when encrypting the object (for # example, `AES256`). # # This functionality is not supported for directory buckets. # # # @option options [String] :sse_customer_key # Specifies the customer-provided encryption key for Amazon S3 to use in # encrypting data. This value is used to store the object and then it is # discarded; Amazon S3 does not store the encryption key. The key must # be appropriate for use with the algorithm specified in the # `x-amz-server-side-encryption-customer-algorithm` header. # # This functionality is not supported for directory buckets. # # # @option options [String] :sse_customer_key_md5 # Specifies the 128-bit MD5 digest of the encryption key according to # RFC 1321. Amazon S3 uses this header for a message integrity check to # ensure that the encryption key was transmitted without error. # # This functionality is not supported for directory buckets. # # # @option options [String] :ssekms_key_id # Specifies the KMS key ID (Key ID, Key ARN, or Key Alias) to use for # object encryption. If the KMS key doesn't exist in the same account # that's issuing the command, you must use the full Key ARN not the Key # ID. # # **General purpose buckets** - If you specify # `x-amz-server-side-encryption` with `aws:kms` or `aws:kms:dsse`, this # header specifies the ID (Key ID, Key ARN, or Key Alias) of the KMS key # to use. If you specify `x-amz-server-side-encryption:aws:kms` or # `x-amz-server-side-encryption:aws:kms:dsse`, but do not provide # `x-amz-server-side-encryption-aws-kms-key-id`, Amazon S3 uses the # Amazon Web Services managed key (`aws/s3`) to protect the data. # # **Directory buckets** - If you specify `x-amz-server-side-encryption` # with `aws:kms`, the ` x-amz-server-side-encryption-aws-kms-key-id` # header is implicitly assigned the ID of the KMS symmetric encryption # customer managed key that's configured for your directory bucket's # default encryption setting. If you want to specify the ` # x-amz-server-side-encryption-aws-kms-key-id` header explicitly, you # can only specify it with the ID (Key ID or Key ARN) of the KMS # customer managed key that's configured for your directory bucket's # default encryption setting. Otherwise, you get an HTTP `400 Bad # Request` error. Only use the key ID or key ARN. The key alias format # of the KMS key isn't supported. Your SSE-KMS configuration can only # support 1 [customer managed key][1] per directory bucket for the # lifetime of the bucket. The [Amazon Web Services managed key][2] # (`aws/s3`) isn't supported. # # # # [1]: https://docs.aws.amazon.com/kms/latest/developerguide/concepts.html#customer-cmk # [2]: https://docs.aws.amazon.com/kms/latest/developerguide/concepts.html#aws-managed-cmk # @option options [String] :ssekms_encryption_context # Specifies the Amazon Web Services KMS Encryption Context as an # additional encryption context to use for object encryption. The value # of this header is a Base64 encoded string of a UTF-8 encoded JSON, # which contains the encryption context as key-value pairs. This value # is stored as object metadata and automatically gets passed on to # Amazon Web Services KMS for future `GetObject` operations on this # object. # # **General purpose buckets** - This value must be explicitly added # during `CopyObject` operations if you want an additional encryption # context for your object. For more information, see [Encryption # context][1] in the *Amazon S3 User Guide*. # # **Directory buckets** - You can optionally provide an explicit # encryption context value. The value must match the default encryption # context - the bucket Amazon Resource Name (ARN). An additional # encryption context value is not supported. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/UsingKMSEncryption.html#encryption-context # @option options [Boolean] :bucket_key_enabled # Specifies whether Amazon S3 should use an S3 Bucket Key for object # encryption with server-side encryption using Key Management Service # (KMS) keys (SSE-KMS). # # **General purpose buckets** - Setting this header to `true` causes # Amazon S3 to use an S3 Bucket Key for object encryption with SSE-KMS. # Also, specifying this header with a PUT action doesn't affect # bucket-level settings for S3 Bucket Key. # # **Directory buckets** - S3 Bucket Keys are always enabled for `GET` # and `PUT` operations in a directory bucket and can’t be disabled. S3 # Bucket Keys aren't supported, when you copy SSE-KMS encrypted objects # from general purpose buckets to directory buckets, from directory # buckets to general purpose buckets, or between directory buckets, # through [CopyObject][1], [UploadPartCopy][2], [the Copy operation in # Batch Operations][3], or [the import jobs][4]. In this case, Amazon S3 # makes a call to KMS every time a copy request is made for a # KMS-encrypted object. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/API/API_CopyObject.html # [2]: https://docs.aws.amazon.com/AmazonS3/latest/API/API_UploadPartCopy.html # [3]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/directory-buckets-objects-Batch-Ops # [4]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/create-import-job # @option options [String] :request_payer # Confirms that the requester knows that they will be charged for the # request. Bucket owners need not specify this parameter in their # requests. If either the source or destination S3 bucket has Requester # Pays enabled, the requester will pay for corresponding charges to copy # the object. For information about downloading objects from Requester # Pays buckets, see [Downloading Objects in Requester Pays Buckets][1] # in the *Amazon S3 User Guide*. # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/ObjectsinRequesterPaysBuckets.html # @option options [String] :tagging # The tag-set for the object. The tag-set must be encoded as URL Query # parameters. (For example, "Key1=Value1") # # This functionality is not supported for directory buckets. # # # @option options [String] :object_lock_mode # The Object Lock mode that you want to apply to this object. # # This functionality is not supported for directory buckets. # # # @option options [Time,DateTime,Date,Integer,String] :object_lock_retain_until_date # The date and time when you want this object's Object Lock to expire. # Must be formatted as a timestamp parameter. # # This functionality is not supported for directory buckets. # # # @option options [String] :object_lock_legal_hold_status # Specifies whether a legal hold will be applied to this object. For # more information about S3 Object Lock, see [Object Lock][1] in the # *Amazon S3 User Guide*. # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/object-lock.html # @option options [String] :expected_bucket_owner # The account ID of the expected bucket owner. If the account ID that # you provide does not match the actual owner of the bucket, the request # fails with the HTTP status code `403 Forbidden` (access denied). # @return [Types::PutObjectOutput] def put(options = {}) options = options.merge( bucket: @bucket_name, key: @key ) resp = Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do @client.put_object(options) end resp.data end # @example Request syntax with placeholder values # # object.restore_object({ # version_id: "ObjectVersionId", # restore_request: { # days: 1, # glacier_job_parameters: { # tier: "Standard", # required, accepts Standard, Bulk, Expedited # }, # type: "SELECT", # accepts SELECT # tier: "Standard", # accepts Standard, Bulk, Expedited # description: "Description", # select_parameters: { # input_serialization: { # required # csv: { # file_header_info: "USE", # accepts USE, IGNORE, NONE # comments: "Comments", # quote_escape_character: "QuoteEscapeCharacter", # record_delimiter: "RecordDelimiter", # field_delimiter: "FieldDelimiter", # quote_character: "QuoteCharacter", # allow_quoted_record_delimiter: false, # }, # compression_type: "NONE", # accepts NONE, GZIP, BZIP2 # json: { # type: "DOCUMENT", # accepts DOCUMENT, LINES # }, # parquet: { # }, # }, # expression_type: "SQL", # required, accepts SQL # expression: "Expression", # required # output_serialization: { # required # csv: { # quote_fields: "ALWAYS", # accepts ALWAYS, ASNEEDED # quote_escape_character: "QuoteEscapeCharacter", # record_delimiter: "RecordDelimiter", # field_delimiter: "FieldDelimiter", # quote_character: "QuoteCharacter", # }, # json: { # record_delimiter: "RecordDelimiter", # }, # }, # }, # output_location: { # s3: { # bucket_name: "BucketName", # required # prefix: "LocationPrefix", # required # encryption: { # encryption_type: "AES256", # required, accepts AES256, aws:kms, aws:kms:dsse # kms_key_id: "SSEKMSKeyId", # kms_context: "KMSContext", # }, # canned_acl: "private", # accepts private, public-read, public-read-write, authenticated-read, aws-exec-read, bucket-owner-read, bucket-owner-full-control # access_control_list: [ # { # grantee: { # display_name: "DisplayName", # email_address: "EmailAddress", # id: "ID", # type: "CanonicalUser", # required, accepts CanonicalUser, AmazonCustomerByEmail, Group # uri: "URI", # }, # permission: "FULL_CONTROL", # accepts FULL_CONTROL, WRITE, WRITE_ACP, READ, READ_ACP # }, # ], # tagging: { # tag_set: [ # required # { # key: "ObjectKey", # required # value: "Value", # required # }, # ], # }, # user_metadata: [ # { # name: "MetadataKey", # value: "MetadataValue", # }, # ], # storage_class: "STANDARD", # accepts STANDARD, REDUCED_REDUNDANCY, STANDARD_IA, ONEZONE_IA, INTELLIGENT_TIERING, GLACIER, DEEP_ARCHIVE, OUTPOSTS, GLACIER_IR, SNOW, EXPRESS_ONEZONE # }, # }, # }, # request_payer: "requester", # accepts requester # checksum_algorithm: "CRC32", # accepts CRC32, CRC32C, SHA1, SHA256, CRC64NVME # expected_bucket_owner: "AccountId", # }) # @param [Hash] options ({}) # @option options [String] :version_id # VersionId used to reference a specific version of the object. # @option options [Types::RestoreRequest] :restore_request # Container for restore job parameters. # @option options [String] :request_payer # Confirms that the requester knows that they will be charged for the # request. Bucket owners need not specify this parameter in their # requests. If either the source or destination S3 bucket has Requester # Pays enabled, the requester will pay for corresponding charges to copy # the object. For information about downloading objects from Requester # Pays buckets, see [Downloading Objects in Requester Pays Buckets][1] # in the *Amazon S3 User Guide*. # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/ObjectsinRequesterPaysBuckets.html # @option options [String] :checksum_algorithm # Indicates the algorithm used to create the checksum for the object # when you use the SDK. This header will not provide any additional # functionality if you don't use the SDK. When you send this header, # there must be a corresponding `x-amz-checksum` or `x-amz-trailer` # header sent. Otherwise, Amazon S3 fails the request with the HTTP # status code `400 Bad Request`. For more information, see [Checking # object integrity][1] in the *Amazon S3 User Guide*. # # If you provide an individual checksum, Amazon S3 ignores any provided # `ChecksumAlgorithm` parameter. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/checking-object-integrity.html # @option options [String] :expected_bucket_owner # The account ID of the expected bucket owner. If the account ID that # you provide does not match the actual owner of the bucket, the request # fails with the HTTP status code `403 Forbidden` (access denied). # @return [Types::RestoreObjectOutput] def restore_object(options = {}) options = options.merge( bucket: @bucket_name, key: @key ) resp = Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do @client.restore_object(options) end resp.data end # @example Request syntax with placeholder values # # object.head({ # if_match: "IfMatch", # if_modified_since: Time.now, # if_none_match: "IfNoneMatch", # if_unmodified_since: Time.now, # range: "Range", # response_cache_control: "ResponseCacheControl", # response_content_disposition: "ResponseContentDisposition", # response_content_encoding: "ResponseContentEncoding", # response_content_language: "ResponseContentLanguage", # response_content_type: "ResponseContentType", # response_expires: Time.now, # version_id: "ObjectVersionId", # sse_customer_algorithm: "SSECustomerAlgorithm", # sse_customer_key: "SSECustomerKey", # sse_customer_key_md5: "SSECustomerKeyMD5", # request_payer: "requester", # accepts requester # part_number: 1, # expected_bucket_owner: "AccountId", # checksum_mode: "ENABLED", # accepts ENABLED # }) # @param [Hash] options ({}) # @option options [String] :if_match # Return the object only if its entity tag (ETag) is the same as the one # specified; otherwise, return a 412 (precondition failed) error. # # If both of the `If-Match` and `If-Unmodified-Since` headers are # present in the request as follows: # # * `If-Match` condition evaluates to `true`, and; # # * `If-Unmodified-Since` condition evaluates to `false`; # # Then Amazon S3 returns `200 OK` and the data requested. # # For more information about conditional requests, see [RFC 7232][1]. # # # # [1]: https://tools.ietf.org/html/rfc7232 # @option options [Time,DateTime,Date,Integer,String] :if_modified_since # Return the object only if it has been modified since the specified # time; otherwise, return a 304 (not modified) error. # # If both of the `If-None-Match` and `If-Modified-Since` headers are # present in the request as follows: # # * `If-None-Match` condition evaluates to `false`, and; # # * `If-Modified-Since` condition evaluates to `true`; # # Then Amazon S3 returns the `304 Not Modified` response code. # # For more information about conditional requests, see [RFC 7232][1]. # # # # [1]: https://tools.ietf.org/html/rfc7232 # @option options [String] :if_none_match # Return the object only if its entity tag (ETag) is different from the # one specified; otherwise, return a 304 (not modified) error. # # If both of the `If-None-Match` and `If-Modified-Since` headers are # present in the request as follows: # # * `If-None-Match` condition evaluates to `false`, and; # # * `If-Modified-Since` condition evaluates to `true`; # # Then Amazon S3 returns the `304 Not Modified` response code. # # For more information about conditional requests, see [RFC 7232][1]. # # # # [1]: https://tools.ietf.org/html/rfc7232 # @option options [Time,DateTime,Date,Integer,String] :if_unmodified_since # Return the object only if it has not been modified since the specified # time; otherwise, return a 412 (precondition failed) error. # # If both of the `If-Match` and `If-Unmodified-Since` headers are # present in the request as follows: # # * `If-Match` condition evaluates to `true`, and; # # * `If-Unmodified-Since` condition evaluates to `false`; # # Then Amazon S3 returns `200 OK` and the data requested. # # For more information about conditional requests, see [RFC 7232][1]. # # # # [1]: https://tools.ietf.org/html/rfc7232 # @option options [String] :range # HeadObject returns only the metadata for an object. If the Range is # satisfiable, only the `ContentLength` is affected in the response. If # the Range is not satisfiable, S3 returns a `416 - Requested Range Not # Satisfiable` error. # @option options [String] :response_cache_control # Sets the `Cache-Control` header of the response. # @option options [String] :response_content_disposition # Sets the `Content-Disposition` header of the response. # @option options [String] :response_content_encoding # Sets the `Content-Encoding` header of the response. # @option options [String] :response_content_language # Sets the `Content-Language` header of the response. # @option options [String] :response_content_type # Sets the `Content-Type` header of the response. # @option options [Time,DateTime,Date,Integer,String] :response_expires # Sets the `Expires` header of the response. # @option options [String] :version_id # Version ID used to reference a specific version of the object. # # For directory buckets in this API operation, only the `null` value of # the version ID is supported. # # # @option options [String] :sse_customer_algorithm # Specifies the algorithm to use when encrypting the object (for # example, AES256). # # This functionality is not supported for directory buckets. # # # @option options [String] :sse_customer_key # Specifies the customer-provided encryption key for Amazon S3 to use in # encrypting data. This value is used to store the object and then it is # discarded; Amazon S3 does not store the encryption key. The key must # be appropriate for use with the algorithm specified in the # `x-amz-server-side-encryption-customer-algorithm` header. # # This functionality is not supported for directory buckets. # # # @option options [String] :sse_customer_key_md5 # Specifies the 128-bit MD5 digest of the encryption key according to # RFC 1321. Amazon S3 uses this header for a message integrity check to # ensure that the encryption key was transmitted without error. # # This functionality is not supported for directory buckets. # # # @option options [String] :request_payer # Confirms that the requester knows that they will be charged for the # request. Bucket owners need not specify this parameter in their # requests. If either the source or destination S3 bucket has Requester # Pays enabled, the requester will pay for corresponding charges to copy # the object. For information about downloading objects from Requester # Pays buckets, see [Downloading Objects in Requester Pays Buckets][1] # in the *Amazon S3 User Guide*. # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/ObjectsinRequesterPaysBuckets.html # @option options [Integer] :part_number # Part number of the object being read. This is a positive integer # between 1 and 10,000. Effectively performs a 'ranged' HEAD request # for the part specified. Useful querying about the size of the part and # the number of parts in this object. # @option options [String] :expected_bucket_owner # The account ID of the expected bucket owner. If the account ID that # you provide does not match the actual owner of the bucket, the request # fails with the HTTP status code `403 Forbidden` (access denied). # @option options [String] :checksum_mode # To retrieve the checksum, this parameter must be enabled. # # **General purpose buckets** - If you enable checksum mode and the # object is uploaded with a [checksum][1] and encrypted with an Key # Management Service (KMS) key, you must have permission to use the # `kms:Decrypt` action to retrieve the checksum. # # **Directory buckets** - If you enable `ChecksumMode` and the object is # encrypted with Amazon Web Services Key Management Service (Amazon Web # Services KMS), you must also have the `kms:GenerateDataKey` and # `kms:Decrypt` permissions in IAM identity-based policies and KMS key # policies for the KMS key to retrieve the checksum of the object. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/API/API_Checksum.html # @return [Types::HeadObjectOutput] def head(options = {}) options = options.merge( bucket: @bucket_name, key: @key ) resp = Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do @client.head_object(options) end resp.data end # @!group Associations # @return [ObjectAcl] def acl ObjectAcl.new( bucket_name: @bucket_name, object_key: @key, client: @client ) end # @return [Bucket] def bucket Bucket.new( name: @bucket_name, client: @client ) end # @param [String] id # @return [MultipartUpload] def multipart_upload(id) MultipartUpload.new( bucket_name: @bucket_name, object_key: @key, id: id, client: @client ) end # @param [String] id # @return [ObjectVersion] def version(id) ObjectVersion.new( bucket_name: @bucket_name, object_key: @key, id: id, client: @client ) end # @deprecated # @api private def identifiers { bucket_name: @bucket_name, key: @key } end deprecated(:identifiers) private def extract_bucket_name(args, options) value = args[0] || options.delete(:bucket_name) case value when String then value when nil then raise ArgumentError, "missing required option :bucket_name" else msg = "expected :bucket_name to be a String, got #{value.class}" raise ArgumentError, msg end end def extract_key(args, options) value = args[1] || options.delete(:key) case value when String then value when nil then raise ArgumentError, "missing required option :key" else msg = "expected :key to be a String, got #{value.class}" raise ArgumentError, msg end end def yield_waiter_and_warn(waiter, &block) if !@waiter_block_warned msg = "pass options to configure the waiter; "\ "yielding the waiter is deprecated" warn(msg) @waiter_block_warned = true end yield(waiter.waiter) end def separate_params_and_options(options) opts = Set.new( [:client, :max_attempts, :delay, :before_attempt, :before_wait] ) waiter_opts = {} waiter_params = {} options.each_pair do |key, value| if opts.include?(key) waiter_opts[key] = value else waiter_params[key] = value end end waiter_opts[:client] ||= @client [waiter_opts, waiter_params] end class Collection < Aws::Resources::Collection # @!group Batch Actions # @example Request syntax with placeholder values # # object.batch_delete!({ # mfa: "MFA", # request_payer: "requester", # accepts requester # bypass_governance_retention: false, # expected_bucket_owner: "AccountId", # checksum_algorithm: "CRC32", # accepts CRC32, CRC32C, SHA1, SHA256, CRC64NVME # }) # @param options ({}) # @option options [String] :mfa # The concatenation of the authentication device's serial number, a # space, and the value that is displayed on your authentication device. # Required to permanently delete a versioned object if versioning is # configured with MFA delete enabled. # # When performing the `DeleteObjects` operation on an MFA delete enabled # bucket, which attempts to delete the specified versioned objects, you # must include an MFA token. If you don't provide an MFA token, the # entire request will fail, even if there are non-versioned objects that # you are trying to delete. If you provide an invalid token, whether # there are versioned object keys in the request or not, the entire # Multi-Object Delete request will fail. For information about MFA # Delete, see [ MFA Delete][1] in the *Amazon S3 User Guide*. # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/Versioning.html#MultiFactorAuthenticationDelete # @option options [String] :request_payer # Confirms that the requester knows that they will be charged for the # request. Bucket owners need not specify this parameter in their # requests. If either the source or destination S3 bucket has Requester # Pays enabled, the requester will pay for corresponding charges to copy # the object. For information about downloading objects from Requester # Pays buckets, see [Downloading Objects in Requester Pays Buckets][1] # in the *Amazon S3 User Guide*. # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/ObjectsinRequesterPaysBuckets.html # @option options [Boolean] :bypass_governance_retention # Specifies whether you want to delete this object even if it has a # Governance-type Object Lock in place. To use this header, you must # have the `s3:BypassGovernanceRetention` permission. # # This functionality is not supported for directory buckets. # # # @option options [String] :expected_bucket_owner # The account ID of the expected bucket owner. If the account ID that # you provide does not match the actual owner of the bucket, the request # fails with the HTTP status code `403 Forbidden` (access denied). # @option options [String] :checksum_algorithm # Indicates the algorithm used to create the checksum for the object # when you use the SDK. This header will not provide any additional # functionality if you don't use the SDK. When you send this header, # there must be a corresponding `x-amz-checksum-algorithm ` or # `x-amz-trailer` header sent. Otherwise, Amazon S3 fails the request # with the HTTP status code `400 Bad Request`. # # For the `x-amz-checksum-algorithm ` header, replace ` algorithm ` with # the supported algorithm from the following list: # # * `CRC-32` # # * `CRC-32C` # # * `CRC-64NVME` # # * `SHA-1` # # * `SHA-256` # # For more information, see [Checking object integrity][1] in the # *Amazon S3 User Guide*. # # If the individual checksum value you provide through # `x-amz-checksum-algorithm ` doesn't match the checksum algorithm you # set through `x-amz-sdk-checksum-algorithm`, Amazon S3 fails the # request with a `BadDigest` error. # # If you provide an individual checksum, Amazon S3 ignores any provided # `ChecksumAlgorithm` parameter. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/checking-object-integrity.html # @return [void] def batch_delete!(options = {}) batch_enum.each do |batch| params = Aws::Util.copy_hash(options) params[:bucket] = batch[0].bucket_name params[:delete] ||= {} params[:delete][:objects] ||= [] batch.each do |item| params[:delete][:objects] << { key: item.key } end Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do batch[0].client.delete_objects(params) end end nil end # @!endgroup end end end # Load customizations if they exist require 'aws-sdk-s3/customizations/object'