# frozen_string_literal: true # WARNING ABOUT GENERATED CODE # # This file is generated. See the contributing guide for more information: # https://github.com/aws/aws-sdk-ruby/blob/version-3/CONTRIBUTING.md # # WARNING ABOUT GENERATED CODE module Aws::S3 class Bucket extend Aws::Deprecations # @overload def initialize(name, options = {}) # @param [String] name # @option options [Client] :client # @overload def initialize(options = {}) # @option options [required, String] :name # @option options [Client] :client def initialize(*args) options = Hash === args.last ? args.pop.dup : {} @name = extract_name(args, options) @data = options.delete(:data) @client = options.delete(:client) || Client.new(options) @waiter_block_warned = false end # @!group Read-Only Attributes # @return [String] def name @name end # Date the bucket was created. This date can change when making changes # to your bucket, such as editing its bucket policy. # @return [Time] def creation_date data[:creation_date] end # `BucketRegion` indicates the Amazon Web Services region where the # bucket is located. If the request contains at least one valid # parameter, it is included in the response. # @return [String] def bucket_region data[:bucket_region] end # @!endgroup # @return [Client] def client @client end # @raise [NotImplementedError] # @api private def load msg = "#load is not implemented, data only available via enumeration" raise NotImplementedError, msg end alias :reload :load # @raise [NotImplementedError] Raises when {#data_loaded?} is `false`. # @return [Types::Bucket] # Returns the data for this {Bucket}. def data load unless @data @data end # @return [Boolean] # Returns `true` if this resource is loaded. Accessing attributes or # {#data} on an unloaded resource will trigger a call to {#load}. def data_loaded? !!@data end # @param [Hash] options ({}) # @return [Boolean] # Returns `true` if the Bucket exists. def exists?(options = {}) begin wait_until_exists(options.merge(max_attempts: 1)) true rescue Aws::Waiters::Errors::UnexpectedError => e raise e.error rescue Aws::Waiters::Errors::WaiterFailed false end end # @param [Hash] options ({}) # @option options [Integer] :max_attempts (20) # @option options [Float] :delay (5) # @option options [Proc] :before_attempt # @option options [Proc] :before_wait # @return [Bucket] def wait_until_exists(options = {}, &block) options, params = separate_params_and_options(options) waiter = Waiters::BucketExists.new(options) yield_waiter_and_warn(waiter, &block) if block_given? Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do waiter.wait(params.merge(bucket: @name)) end Bucket.new({ name: @name, client: @client }) end # @param [Hash] options ({}) # @option options [Integer] :max_attempts (20) # @option options [Float] :delay (5) # @option options [Proc] :before_attempt # @option options [Proc] :before_wait # @return [Bucket] def wait_until_not_exists(options = {}, &block) options, params = separate_params_and_options(options) waiter = Waiters::BucketNotExists.new(options) yield_waiter_and_warn(waiter, &block) if block_given? Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do waiter.wait(params.merge(bucket: @name)) end Bucket.new({ name: @name, client: @client }) end # @deprecated Use [Aws::S3::Client] #wait_until instead # # Waiter polls an API operation until a resource enters a desired # state. # # @note The waiting operation is performed on a copy. The original resource # remains unchanged. # # ## Basic Usage # # Waiter will polls until it is successful, it fails by # entering a terminal state, or until a maximum number of attempts # are made. # # # polls in a loop until condition is true # resource.wait_until(options) {|resource| condition} # # ## Example # # instance.wait_until(max_attempts:10, delay:5) do |instance| # instance.state.name == 'running' # end # # ## Configuration # # You can configure the maximum number of polling attempts, and the # delay (in seconds) between each polling attempt. The waiting condition is # set by passing a block to {#wait_until}: # # # poll for ~25 seconds # resource.wait_until(max_attempts:5,delay:5) {|resource|...} # # ## Callbacks # # You can be notified before each polling attempt and before each # delay. If you throw `:success` or `:failure` from these callbacks, # it will terminate the waiter. # # started_at = Time.now # # poll for 1 hour, instead of a number of attempts # proc = Proc.new do |attempts, response| # throw :failure if Time.now - started_at > 3600 # end # # # disable max attempts # instance.wait_until(before_wait:proc, max_attempts:nil) {...} # # ## Handling Errors # # When a waiter is successful, it returns the Resource. When a waiter # fails, it raises an error. # # begin # resource.wait_until(...) # rescue Aws::Waiters::Errors::WaiterFailed # # resource did not enter the desired state in time # end # # @yieldparam [Resource] resource to be used in the waiting condition. # # @raise [Aws::Waiters::Errors::FailureStateError] Raised when the waiter # terminates because the waiter has entered a state that it will not # transition out of, preventing success. # # yet successful. # # @raise [Aws::Waiters::Errors::UnexpectedError] Raised when an error is # encountered while polling for a resource that is not expected. # # @raise [NotImplementedError] Raised when the resource does not # # @option options [Integer] :max_attempts (10) Maximum number of # attempts # @option options [Integer] :delay (10) Delay between each # attempt in seconds # @option options [Proc] :before_attempt (nil) Callback # invoked before each attempt # @option options [Proc] :before_wait (nil) Callback # invoked before each wait # @return [Resource] if the waiter was successful def wait_until(options = {}, &block) self_copy = self.dup attempts = 0 options[:max_attempts] = 10 unless options.key?(:max_attempts) options[:delay] ||= 10 options[:poller] = Proc.new do attempts += 1 if block.call(self_copy) [:success, self_copy] else self_copy.reload unless attempts == options[:max_attempts] :retry end end Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do Aws::Waiters::Waiter.new(options).wait({}) end end # @!group Actions # @example Request syntax with placeholder values # # bucket.create({ # acl: "private", # accepts private, public-read, public-read-write, authenticated-read # create_bucket_configuration: { # location_constraint: "af-south-1", # accepts af-south-1, ap-east-1, ap-northeast-1, ap-northeast-2, ap-northeast-3, ap-south-1, ap-south-2, ap-southeast-1, ap-southeast-2, ap-southeast-3, ca-central-1, cn-north-1, cn-northwest-1, EU, eu-central-1, eu-north-1, eu-south-1, eu-south-2, eu-west-1, eu-west-2, eu-west-3, me-south-1, sa-east-1, us-east-2, us-gov-east-1, us-gov-west-1, us-west-1, us-west-2 # location: { # type: "AvailabilityZone", # accepts AvailabilityZone # name: "LocationNameAsString", # }, # bucket: { # data_redundancy: "SingleAvailabilityZone", # accepts SingleAvailabilityZone # type: "Directory", # accepts Directory # }, # }, # grant_full_control: "GrantFullControl", # grant_read: "GrantRead", # grant_read_acp: "GrantReadACP", # grant_write: "GrantWrite", # grant_write_acp: "GrantWriteACP", # object_lock_enabled_for_bucket: false, # object_ownership: "BucketOwnerPreferred", # accepts BucketOwnerPreferred, ObjectWriter, BucketOwnerEnforced # }) # @param [Hash] options ({}) # @option options [String] :acl # The canned ACL to apply to the bucket. # # This functionality is not supported for directory buckets. # # # @option options [Types::CreateBucketConfiguration] :create_bucket_configuration # The configuration information for the bucket. # @option options [String] :grant_full_control # Allows grantee the read, write, read ACP, and write ACP permissions on # the bucket. # # This functionality is not supported for directory buckets. # # # @option options [String] :grant_read # Allows grantee to list the objects in the bucket. # # This functionality is not supported for directory buckets. # # # @option options [String] :grant_read_acp # Allows grantee to read the bucket ACL. # # This functionality is not supported for directory buckets. # # # @option options [String] :grant_write # Allows grantee to create new objects in the bucket. # # For the bucket and object owners of existing objects, also allows # deletions and overwrites of those objects. # # This functionality is not supported for directory buckets. # # # @option options [String] :grant_write_acp # Allows grantee to write the ACL for the applicable bucket. # # This functionality is not supported for directory buckets. # # # @option options [Boolean] :object_lock_enabled_for_bucket # Specifies whether you want S3 Object Lock to be enabled for the new # bucket. # # This functionality is not supported for directory buckets. # # # @option options [String] :object_ownership # The container element for object ownership for a bucket's ownership # controls. # # `BucketOwnerPreferred` - Objects uploaded to the bucket change # ownership to the bucket owner if the objects are uploaded with the # `bucket-owner-full-control` canned ACL. # # `ObjectWriter` - The uploading account will own the object if the # object is uploaded with the `bucket-owner-full-control` canned ACL. # # `BucketOwnerEnforced` - Access control lists (ACLs) are disabled and # no longer affect permissions. The bucket owner automatically owns and # has full control over every object in the bucket. The bucket only # accepts PUT requests that don't specify an ACL or specify bucket # owner full control ACLs (such as the predefined # `bucket-owner-full-control` canned ACL or a custom ACL in XML format # that grants the same permissions). # # By default, `ObjectOwnership` is set to `BucketOwnerEnforced` and ACLs # are disabled. We recommend keeping ACLs disabled, except in uncommon # use cases where you must control access for each object individually. # For more information about S3 Object Ownership, see [Controlling # ownership of objects and disabling ACLs for your bucket][1] in the # *Amazon S3 User Guide*. # # This functionality is not supported for directory buckets. Directory # buckets use the bucket owner enforced setting for S3 Object Ownership. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/about-object-ownership.html # @return [Types::CreateBucketOutput] def create(options = {}) options = options.merge(bucket: @name) resp = Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do @client.create_bucket(options) end resp.data end # @example Request syntax with placeholder values # # bucket.delete({ # expected_bucket_owner: "AccountId", # }) # @param [Hash] options ({}) # @option options [String] :expected_bucket_owner # The account ID of the expected bucket owner. If the account ID that # you provide does not match the actual owner of the bucket, the request # fails with the HTTP status code `403 Forbidden` (access denied). # # For directory buckets, this header is not supported in this API # operation. If you specify this header, the request fails with the HTTP # status code `501 Not Implemented`. # # # @return [EmptyStructure] def delete(options = {}) options = options.merge(bucket: @name) resp = Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do @client.delete_bucket(options) end resp.data end # @example Request syntax with placeholder values # # bucket.delete_objects({ # delete: { # required # objects: [ # required # { # key: "ObjectKey", # required # version_id: "ObjectVersionId", # }, # ], # quiet: false, # }, # mfa: "MFA", # request_payer: "requester", # accepts requester # bypass_governance_retention: false, # expected_bucket_owner: "AccountId", # checksum_algorithm: "CRC32", # accepts CRC32, CRC32C, SHA1, SHA256 # }) # @param [Hash] options ({}) # @option options [required, Types::Delete] :delete # Container for the request. # @option options [String] :mfa # The concatenation of the authentication device's serial number, a # space, and the value that is displayed on your authentication device. # Required to permanently delete a versioned object if versioning is # configured with MFA delete enabled. # # When performing the `DeleteObjects` operation on an MFA delete enabled # bucket, which attempts to delete the specified versioned objects, you # must include an MFA token. If you don't provide an MFA token, the # entire request will fail, even if there are non-versioned objects that # you are trying to delete. If you provide an invalid token, whether # there are versioned object keys in the request or not, the entire # Multi-Object Delete request will fail. For information about MFA # Delete, see [ MFA Delete][1] in the *Amazon S3 User Guide*. # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/Versioning.html#MultiFactorAuthenticationDelete # @option options [String] :request_payer # Confirms that the requester knows that they will be charged for the # request. Bucket owners need not specify this parameter in their # requests. If either the source or destination S3 bucket has Requester # Pays enabled, the requester will pay for corresponding charges to copy # the object. For information about downloading objects from Requester # Pays buckets, see [Downloading Objects in Requester Pays Buckets][1] # in the *Amazon S3 User Guide*. # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/ObjectsinRequesterPaysBuckets.html # @option options [Boolean] :bypass_governance_retention # Specifies whether you want to delete this object even if it has a # Governance-type Object Lock in place. To use this header, you must # have the `s3:BypassGovernanceRetention` permission. # # This functionality is not supported for directory buckets. # # # @option options [String] :expected_bucket_owner # The account ID of the expected bucket owner. If the account ID that # you provide does not match the actual owner of the bucket, the request # fails with the HTTP status code `403 Forbidden` (access denied). # @option options [String] :checksum_algorithm # Indicates the algorithm used to create the checksum for the object # when you use the SDK. This header will not provide any additional # functionality if you don't use the SDK. When you send this header, # there must be a corresponding `x-amz-checksum-algorithm ` or # `x-amz-trailer` header sent. Otherwise, Amazon S3 fails the request # with the HTTP status code `400 Bad Request`. # # For the `x-amz-checksum-algorithm ` header, replace ` algorithm ` with # the supported algorithm from the following list: # # * `CRC32` # # * `CRC32C` # # * `SHA1` # # * `SHA256` # # For more information, see [Checking object integrity][1] in the # *Amazon S3 User Guide*. # # If the individual checksum value you provide through # `x-amz-checksum-algorithm ` doesn't match the checksum algorithm you # set through `x-amz-sdk-checksum-algorithm`, Amazon S3 ignores any # provided `ChecksumAlgorithm` parameter and uses the checksum algorithm # that matches the provided value in `x-amz-checksum-algorithm `. # # If you provide an individual checksum, Amazon S3 ignores any provided # `ChecksumAlgorithm` parameter. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/checking-object-integrity.html # @return [Types::DeleteObjectsOutput] def delete_objects(options = {}) options = options.merge(bucket: @name) resp = Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do @client.delete_objects(options) end resp.data end # @example Request syntax with placeholder values # # object = bucket.put_object({ # acl: "private", # accepts private, public-read, public-read-write, authenticated-read, aws-exec-read, bucket-owner-read, bucket-owner-full-control # body: source_file, # cache_control: "CacheControl", # content_disposition: "ContentDisposition", # content_encoding: "ContentEncoding", # content_language: "ContentLanguage", # content_length: 1, # content_md5: "ContentMD5", # content_type: "ContentType", # checksum_algorithm: "CRC32", # accepts CRC32, CRC32C, SHA1, SHA256 # checksum_crc32: "ChecksumCRC32", # checksum_crc32c: "ChecksumCRC32C", # checksum_sha1: "ChecksumSHA1", # checksum_sha256: "ChecksumSHA256", # expires: Time.now, # if_none_match: "IfNoneMatch", # grant_full_control: "GrantFullControl", # grant_read: "GrantRead", # grant_read_acp: "GrantReadACP", # grant_write_acp: "GrantWriteACP", # key: "ObjectKey", # required # metadata: { # "MetadataKey" => "MetadataValue", # }, # server_side_encryption: "AES256", # accepts AES256, aws:kms, aws:kms:dsse # storage_class: "STANDARD", # accepts STANDARD, REDUCED_REDUNDANCY, STANDARD_IA, ONEZONE_IA, INTELLIGENT_TIERING, GLACIER, DEEP_ARCHIVE, OUTPOSTS, GLACIER_IR, SNOW, EXPRESS_ONEZONE # website_redirect_location: "WebsiteRedirectLocation", # sse_customer_algorithm: "SSECustomerAlgorithm", # sse_customer_key: "SSECustomerKey", # sse_customer_key_md5: "SSECustomerKeyMD5", # ssekms_key_id: "SSEKMSKeyId", # ssekms_encryption_context: "SSEKMSEncryptionContext", # bucket_key_enabled: false, # request_payer: "requester", # accepts requester # tagging: "TaggingHeader", # object_lock_mode: "GOVERNANCE", # accepts GOVERNANCE, COMPLIANCE # object_lock_retain_until_date: Time.now, # object_lock_legal_hold_status: "ON", # accepts ON, OFF # expected_bucket_owner: "AccountId", # }) # @param [Hash] options ({}) # @option options [String] :acl # The canned ACL to apply to the object. For more information, see # [Canned ACL][1] in the *Amazon S3 User Guide*. # # When adding a new object, you can use headers to grant ACL-based # permissions to individual Amazon Web Services accounts or to # predefined groups defined by Amazon S3. These permissions are then # added to the ACL on the object. By default, all objects are private. # Only the owner has full access control. For more information, see # [Access Control List (ACL) Overview][2] and [Managing ACLs Using the # REST API][3] in the *Amazon S3 User Guide*. # # If the bucket that you're uploading objects to uses the bucket owner # enforced setting for S3 Object Ownership, ACLs are disabled and no # longer affect permissions. Buckets that use this setting only accept # PUT requests that don't specify an ACL or PUT requests that specify # bucket owner full control ACLs, such as the # `bucket-owner-full-control` canned ACL or an equivalent form of this # ACL expressed in the XML format. PUT requests that contain other ACLs # (for example, custom grants to certain Amazon Web Services accounts) # fail and return a `400` error with the error code # `AccessControlListNotSupported`. For more information, see [ # Controlling ownership of objects and disabling ACLs][4] in the *Amazon # S3 User Guide*. # # * This functionality is not supported for directory buckets. # # * This functionality is not supported for Amazon S3 on Outposts. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/acl-overview.html#CannedACL # [2]: https://docs.aws.amazon.com/AmazonS3/latest/dev/acl-overview.html # [3]: https://docs.aws.amazon.com/AmazonS3/latest/dev/acl-using-rest-api.html # [4]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/about-object-ownership.html # @option options [String, StringIO, File] :body # Object data. # @option options [String] :cache_control # Can be used to specify caching behavior along the request/reply chain. # For more information, see # [http://www.w3.org/Protocols/rfc2616/rfc2616-sec14.html#sec14.9][1]. # # # # [1]: http://www.w3.org/Protocols/rfc2616/rfc2616-sec14.html#sec14.9 # @option options [String] :content_disposition # Specifies presentational information for the object. For more # information, see # [https://www.rfc-editor.org/rfc/rfc6266#section-4][1]. # # # # [1]: https://www.rfc-editor.org/rfc/rfc6266#section-4 # @option options [String] :content_encoding # Specifies what content encodings have been applied to the object and # thus what decoding mechanisms must be applied to obtain the media-type # referenced by the Content-Type header field. For more information, see # [https://www.rfc-editor.org/rfc/rfc9110.html#field.content-encoding][1]. # # # # [1]: https://www.rfc-editor.org/rfc/rfc9110.html#field.content-encoding # @option options [String] :content_language # The language the content is in. # @option options [Integer] :content_length # Size of the body in bytes. This parameter is useful when the size of # the body cannot be determined automatically. For more information, see # [https://www.rfc-editor.org/rfc/rfc9110.html#name-content-length][1]. # # # # [1]: https://www.rfc-editor.org/rfc/rfc9110.html#name-content-length # @option options [String] :content_md5 # The base64-encoded 128-bit MD5 digest of the message (without the # headers) according to RFC 1864. This header can be used as a message # integrity check to verify that the data is the same data that was # originally sent. Although it is optional, we recommend using the # Content-MD5 mechanism as an end-to-end integrity check. For more # information about REST request authentication, see [REST # Authentication][1]. # # The `Content-MD5` or `x-amz-sdk-checksum-algorithm` header is required # for any request to upload an object with a retention period configured # using Amazon S3 Object Lock. For more information, see [Uploading # objects to an Object Lock enabled bucket ][2] in the *Amazon S3 User # Guide*. # # # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/RESTAuthentication.html # [2]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/object-lock-managing.html#object-lock-put-object # @option options [String] :content_type # A standard MIME type describing the format of the contents. For more # information, see # [https://www.rfc-editor.org/rfc/rfc9110.html#name-content-type][1]. # # # # [1]: https://www.rfc-editor.org/rfc/rfc9110.html#name-content-type # @option options [String] :checksum_algorithm # Indicates the algorithm used to create the checksum for the object # when you use the SDK. This header will not provide any additional # functionality if you don't use the SDK. When you send this header, # there must be a corresponding `x-amz-checksum-algorithm ` or # `x-amz-trailer` header sent. Otherwise, Amazon S3 fails the request # with the HTTP status code `400 Bad Request`. # # For the `x-amz-checksum-algorithm ` header, replace ` algorithm ` with # the supported algorithm from the following list: # # * `CRC32` # # * `CRC32C` # # * `SHA1` # # * `SHA256` # # For more information, see [Checking object integrity][1] in the # *Amazon S3 User Guide*. # # If the individual checksum value you provide through # `x-amz-checksum-algorithm ` doesn't match the checksum algorithm you # set through `x-amz-sdk-checksum-algorithm`, Amazon S3 ignores any # provided `ChecksumAlgorithm` parameter and uses the checksum algorithm # that matches the provided value in `x-amz-checksum-algorithm `. # # The `Content-MD5` or `x-amz-sdk-checksum-algorithm` header is required # for any request to upload an object with a retention period configured # using Amazon S3 Object Lock. For more information, see [Uploading # objects to an Object Lock enabled bucket ][2] in the *Amazon S3 User # Guide*. # # # # For directory buckets, when you use Amazon Web Services SDKs, `CRC32` # is the default checksum algorithm that's used for performance. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/checking-object-integrity.html # [2]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/object-lock-managing.html#object-lock-put-object # @option options [String] :checksum_crc32 # This header can be used as a data integrity check to verify that the # data received is the same data that was originally sent. This header # specifies the base64-encoded, 32-bit CRC-32 checksum of the object. # For more information, see [Checking object integrity][1] in the # *Amazon S3 User Guide*. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/checking-object-integrity.html # @option options [String] :checksum_crc32c # This header can be used as a data integrity check to verify that the # data received is the same data that was originally sent. This header # specifies the base64-encoded, 32-bit CRC-32C checksum of the object. # For more information, see [Checking object integrity][1] in the # *Amazon S3 User Guide*. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/checking-object-integrity.html # @option options [String] :checksum_sha1 # This header can be used as a data integrity check to verify that the # data received is the same data that was originally sent. This header # specifies the base64-encoded, 160-bit SHA-1 digest of the object. For # more information, see [Checking object integrity][1] in the *Amazon S3 # User Guide*. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/checking-object-integrity.html # @option options [String] :checksum_sha256 # This header can be used as a data integrity check to verify that the # data received is the same data that was originally sent. This header # specifies the base64-encoded, 256-bit SHA-256 digest of the object. # For more information, see [Checking object integrity][1] in the # *Amazon S3 User Guide*. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/checking-object-integrity.html # @option options [Time,DateTime,Date,Integer,String] :expires # The date and time at which the object is no longer cacheable. For more # information, see # [https://www.rfc-editor.org/rfc/rfc7234#section-5.3][1]. # # # # [1]: https://www.rfc-editor.org/rfc/rfc7234#section-5.3 # @option options [String] :if_none_match # Uploads the object only if the object key name does not already exist # in the bucket specified. Otherwise, Amazon S3 returns a `412 # Precondition Failed` error. # # If a conflicting operation occurs during the upload S3 returns a `409 # ConditionalRequestConflict` response. On a 409 failure you should # retry the upload. # # Expects the '*' (asterisk) character. # # For more information about conditional requests, see [RFC 7232][1], or # [Conditional requests][2] in the *Amazon S3 User Guide*. # # # # [1]: https://tools.ietf.org/html/rfc7232 # [2]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/conditional-requests.html # @option options [String] :grant_full_control # Gives the grantee READ, READ\_ACP, and WRITE\_ACP permissions on the # object. # # * This functionality is not supported for directory buckets. # # * This functionality is not supported for Amazon S3 on Outposts. # # # @option options [String] :grant_read # Allows grantee to read the object data and its metadata. # # * This functionality is not supported for directory buckets. # # * This functionality is not supported for Amazon S3 on Outposts. # # # @option options [String] :grant_read_acp # Allows grantee to read the object ACL. # # * This functionality is not supported for directory buckets. # # * This functionality is not supported for Amazon S3 on Outposts. # # # @option options [String] :grant_write_acp # Allows grantee to write the ACL for the applicable object. # # * This functionality is not supported for directory buckets. # # * This functionality is not supported for Amazon S3 on Outposts. # # # @option options [required, String] :key # Object key for which the PUT action was initiated. # @option options [Hash] :metadata # A map of metadata to store with the object in S3. # @option options [String] :server_side_encryption # The server-side encryption algorithm that was used when you store this # object in Amazon S3 (for example, `AES256`, `aws:kms`, # `aws:kms:dsse`). # # * General purpose buckets - You have four mutually exclusive # options to protect data using server-side encryption in Amazon S3, # depending on how you choose to manage the encryption keys. # Specifically, the encryption key options are Amazon S3 managed keys # (SSE-S3), Amazon Web Services KMS keys (SSE-KMS or DSSE-KMS), and # customer-provided keys (SSE-C). Amazon S3 encrypts data with # server-side encryption by using Amazon S3 managed keys (SSE-S3) by # default. You can optionally tell Amazon S3 to encrypt data at rest # by using server-side encryption with other key options. For more # information, see [Using Server-Side Encryption][1] in the *Amazon S3 # User Guide*. # # * Directory buckets - For directory buckets, there are only # two supported options for server-side encryption: server-side # encryption with Amazon S3 managed keys (SSE-S3) (`AES256`) and # server-side encryption with KMS keys (SSE-KMS) (`aws:kms`). We # recommend that the bucket's default encryption uses the desired # encryption configuration and you don't override the bucket default # encryption in your `CreateSession` requests or `PUT` object # requests. Then, new objects are automatically encrypted with the # desired encryption settings. For more information, see [Protecting # data with server-side encryption][2] in the *Amazon S3 User Guide*. # For more information about the encryption overriding behaviors in # directory buckets, see [Specifying server-side encryption with KMS # for new object uploads][3]. # # In the Zonal endpoint API calls (except [CopyObject][4] and # [UploadPartCopy][5]) using the REST API, the encryption request # headers must match the encryption settings that are specified in the # `CreateSession` request. You can't override the values of the # encryption settings (`x-amz-server-side-encryption`, # `x-amz-server-side-encryption-aws-kms-key-id`, # `x-amz-server-side-encryption-context`, and # `x-amz-server-side-encryption-bucket-key-enabled`) that are # specified in the `CreateSession` request. You don't need to # explicitly specify these encryption settings values in Zonal # endpoint API calls, and Amazon S3 will use the encryption settings # values from the `CreateSession` request to protect new objects in # the directory bucket. # # When you use the CLI or the Amazon Web Services SDKs, for # `CreateSession`, the session token refreshes automatically to avoid # service interruptions when a session expires. The CLI or the Amazon # Web Services SDKs use the bucket's default encryption configuration # for the `CreateSession` request. It's not supported to override the # encryption settings values in the `CreateSession` request. So in the # Zonal endpoint API calls (except [CopyObject][4] and # [UploadPartCopy][5]), the encryption request headers must match the # default encryption configuration of the directory bucket. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/UsingServerSideEncryption.html # [2]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/s3-express-serv-side-encryption.html # [3]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/s3-express-specifying-kms-encryption.html # [4]: https://docs.aws.amazon.com/AmazonS3/latest/API/API_CopyObject.html # [5]: https://docs.aws.amazon.com/AmazonS3/latest/API/API_UploadPartCopy.html # @option options [String] :storage_class # By default, Amazon S3 uses the STANDARD Storage Class to store newly # created objects. The STANDARD storage class provides high durability # and high availability. Depending on performance needs, you can specify # a different Storage Class. For more information, see [Storage # Classes][1] in the *Amazon S3 User Guide*. # # * For directory buckets, only the S3 Express One Zone storage class is # supported to store newly created objects. # # * Amazon S3 on Outposts only uses the OUTPOSTS Storage Class. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/storage-class-intro.html # @option options [String] :website_redirect_location # If the bucket is configured as a website, redirects requests for this # object to another object in the same bucket or to an external URL. # Amazon S3 stores the value of this header in the object metadata. For # information about object metadata, see [Object Key and Metadata][1] in # the *Amazon S3 User Guide*. # # In the following example, the request header sets the redirect to an # object (anotherPage.html) in the same bucket: # # `x-amz-website-redirect-location: /anotherPage.html` # # In the following example, the request header sets the object redirect # to another website: # # `x-amz-website-redirect-location: http://www.example.com/` # # For more information about website hosting in Amazon S3, see [Hosting # Websites on Amazon S3][2] and [How to Configure Website Page # Redirects][3] in the *Amazon S3 User Guide*. # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/UsingMetadata.html # [2]: https://docs.aws.amazon.com/AmazonS3/latest/dev/WebsiteHosting.html # [3]: https://docs.aws.amazon.com/AmazonS3/latest/dev/how-to-page-redirect.html # @option options [String] :sse_customer_algorithm # Specifies the algorithm to use when encrypting the object (for # example, `AES256`). # # This functionality is not supported for directory buckets. # # # @option options [String] :sse_customer_key # Specifies the customer-provided encryption key for Amazon S3 to use in # encrypting data. This value is used to store the object and then it is # discarded; Amazon S3 does not store the encryption key. The key must # be appropriate for use with the algorithm specified in the # `x-amz-server-side-encryption-customer-algorithm` header. # # This functionality is not supported for directory buckets. # # # @option options [String] :sse_customer_key_md5 # Specifies the 128-bit MD5 digest of the encryption key according to # RFC 1321. Amazon S3 uses this header for a message integrity check to # ensure that the encryption key was transmitted without error. # # This functionality is not supported for directory buckets. # # # @option options [String] :ssekms_key_id # Specifies the KMS key ID (Key ID, Key ARN, or Key Alias) to use for # object encryption. If the KMS key doesn't exist in the same account # that's issuing the command, you must use the full Key ARN not the Key # ID. # # **General purpose buckets** - If you specify # `x-amz-server-side-encryption` with `aws:kms` or `aws:kms:dsse`, this # header specifies the ID (Key ID, Key ARN, or Key Alias) of the KMS key # to use. If you specify `x-amz-server-side-encryption:aws:kms` or # `x-amz-server-side-encryption:aws:kms:dsse`, but do not provide # `x-amz-server-side-encryption-aws-kms-key-id`, Amazon S3 uses the # Amazon Web Services managed key (`aws/s3`) to protect the data. # # **Directory buckets** - If you specify `x-amz-server-side-encryption` # with `aws:kms`, the ` x-amz-server-side-encryption-aws-kms-key-id` # header is implicitly assigned the ID of the KMS symmetric encryption # customer managed key that's configured for your directory bucket's # default encryption setting. If you want to specify the ` # x-amz-server-side-encryption-aws-kms-key-id` header explicitly, you # can only specify it with the ID (Key ID or Key ARN) of the KMS # customer managed key that's configured for your directory bucket's # default encryption setting. Otherwise, you get an HTTP `400 Bad # Request` error. Only use the key ID or key ARN. The key alias format # of the KMS key isn't supported. Your SSE-KMS configuration can only # support 1 [customer managed key][1] per directory bucket for the # lifetime of the bucket. The [Amazon Web Services managed key][2] # (`aws/s3`) isn't supported. # # # # [1]: https://docs.aws.amazon.com/kms/latest/developerguide/concepts.html#customer-cmk # [2]: https://docs.aws.amazon.com/kms/latest/developerguide/concepts.html#aws-managed-cmk # @option options [String] :ssekms_encryption_context # Specifies the Amazon Web Services KMS Encryption Context as an # additional encryption context to use for object encryption. The value # of this header is a Base64-encoded string of a UTF-8 encoded JSON, # which contains the encryption context as key-value pairs. This value # is stored as object metadata and automatically gets passed on to # Amazon Web Services KMS for future `GetObject` operations on this # object. # # **General purpose buckets** - This value must be explicitly added # during `CopyObject` operations if you want an additional encryption # context for your object. For more information, see [Encryption # context][1] in the *Amazon S3 User Guide*. # # **Directory buckets** - You can optionally provide an explicit # encryption context value. The value must match the default encryption # context - the bucket Amazon Resource Name (ARN). An additional # encryption context value is not supported. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/UsingKMSEncryption.html#encryption-context # @option options [Boolean] :bucket_key_enabled # Specifies whether Amazon S3 should use an S3 Bucket Key for object # encryption with server-side encryption using Key Management Service # (KMS) keys (SSE-KMS). # # **General purpose buckets** - Setting this header to `true` causes # Amazon S3 to use an S3 Bucket Key for object encryption with SSE-KMS. # Also, specifying this header with a PUT action doesn't affect # bucket-level settings for S3 Bucket Key. # # **Directory buckets** - S3 Bucket Keys are always enabled for `GET` # and `PUT` operations in a directory bucket and can’t be disabled. S3 # Bucket Keys aren't supported, when you copy SSE-KMS encrypted objects # from general purpose buckets to directory buckets, from directory # buckets to general purpose buckets, or between directory buckets, # through [CopyObject][1], [UploadPartCopy][2], [the Copy operation in # Batch Operations][3], or [the import jobs][4]. In this case, Amazon S3 # makes a call to KMS every time a copy request is made for a # KMS-encrypted object. # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/API/API_CopyObject.html # [2]: https://docs.aws.amazon.com/AmazonS3/latest/API/API_UploadPartCopy.html # [3]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/directory-buckets-objects-Batch-Ops # [4]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/create-import-job # @option options [String] :request_payer # Confirms that the requester knows that they will be charged for the # request. Bucket owners need not specify this parameter in their # requests. If either the source or destination S3 bucket has Requester # Pays enabled, the requester will pay for corresponding charges to copy # the object. For information about downloading objects from Requester # Pays buckets, see [Downloading Objects in Requester Pays Buckets][1] # in the *Amazon S3 User Guide*. # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/ObjectsinRequesterPaysBuckets.html # @option options [String] :tagging # The tag-set for the object. The tag-set must be encoded as URL Query # parameters. (For example, "Key1=Value1") # # This functionality is not supported for directory buckets. # # # @option options [String] :object_lock_mode # The Object Lock mode that you want to apply to this object. # # This functionality is not supported for directory buckets. # # # @option options [Time,DateTime,Date,Integer,String] :object_lock_retain_until_date # The date and time when you want this object's Object Lock to expire. # Must be formatted as a timestamp parameter. # # This functionality is not supported for directory buckets. # # # @option options [String] :object_lock_legal_hold_status # Specifies whether a legal hold will be applied to this object. For # more information about S3 Object Lock, see [Object Lock][1] in the # *Amazon S3 User Guide*. # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/object-lock.html # @option options [String] :expected_bucket_owner # The account ID of the expected bucket owner. If the account ID that # you provide does not match the actual owner of the bucket, the request # fails with the HTTP status code `403 Forbidden` (access denied). # @return [Object] def put_object(options = {}) options = options.merge(bucket: @name) Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do @client.put_object(options) end Object.new( bucket_name: @name, key: options[:key], client: @client ) end # @!group Associations # @return [BucketAcl] def acl BucketAcl.new( bucket_name: @name, client: @client ) end # @return [BucketCors] def cors BucketCors.new( bucket_name: @name, client: @client ) end # @return [BucketLifecycle] def lifecycle BucketLifecycle.new( bucket_name: @name, client: @client ) end # @return [BucketLifecycleConfiguration] def lifecycle_configuration BucketLifecycleConfiguration.new( bucket_name: @name, client: @client ) end # @return [BucketLogging] def logging BucketLogging.new( bucket_name: @name, client: @client ) end # @example Request syntax with placeholder values # # multipart_uploads = bucket.multipart_uploads({ # delimiter: "Delimiter", # encoding_type: "url", # accepts url # key_marker: "KeyMarker", # prefix: "Prefix", # upload_id_marker: "UploadIdMarker", # expected_bucket_owner: "AccountId", # request_payer: "requester", # accepts requester # }) # @param [Hash] options ({}) # @option options [String] :delimiter # Character you use to group keys. # # All keys that contain the same string between the prefix, if # specified, and the first occurrence of the delimiter after the prefix # are grouped under a single result element, `CommonPrefixes`. If you # don't specify the prefix parameter, then the substring starts at the # beginning of the key. The keys that are grouped under `CommonPrefixes` # result element are not returned elsewhere in the response. # # **Directory buckets** - For directory buckets, `/` is the only # supported delimiter. # # # @option options [String] :encoding_type # Encoding type used by Amazon S3 to encode the [object keys][1] in the # response. Responses are encoded only in UTF-8. An object key can # contain any Unicode character. However, the XML 1.0 parser can't # parse certain characters, such as characters with an ASCII value from # 0 to 10. For characters that aren't supported in XML 1.0, you can add # this parameter to request that Amazon S3 encode the keys in the # response. For more information about characters to avoid in object key # names, see [Object key naming guidelines][2]. # # When using the URL encoding type, non-ASCII characters that are used # in an object's key name will be percent-encoded according to UTF-8 # code values. For example, the object `test_file(3).png` will appear as # `test_file%283%29.png`. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/object-keys.html # [2]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/object-keys.html#object-key-guidelines # @option options [String] :key_marker # Specifies the multipart upload after which listing should begin. # # * **General purpose buckets** - For general purpose buckets, # `key-marker` is an object key. Together with `upload-id-marker`, # this parameter specifies the multipart upload after which listing # should begin. # # If `upload-id-marker` is not specified, only the keys # lexicographically greater than the specified `key-marker` will be # included in the list. # # If `upload-id-marker` is specified, any multipart uploads for a key # equal to the `key-marker` might also be included, provided those # multipart uploads have upload IDs lexicographically greater than the # specified `upload-id-marker`. # # * **Directory buckets** - For directory buckets, `key-marker` is # obfuscated and isn't a real object key. The `upload-id-marker` # parameter isn't supported by directory buckets. To list the # additional multipart uploads, you only need to set the value of # `key-marker` to the `NextKeyMarker` value from the previous # response. # # In the `ListMultipartUploads` response, the multipart uploads # aren't sorted lexicographically based on the object keys. # # # @option options [String] :prefix # Lists in-progress uploads only for those keys that begin with the # specified prefix. You can use prefixes to separate a bucket into # different grouping of keys. (You can think of using `prefix` to make # groups in the same way that you'd use a folder in a file system.) # # **Directory buckets** - For directory buckets, only prefixes that end # in a delimiter (`/`) are supported. # # # @option options [String] :upload_id_marker # Together with key-marker, specifies the multipart upload after which # listing should begin. If key-marker is not specified, the # upload-id-marker parameter is ignored. Otherwise, any multipart # uploads for a key equal to the key-marker might be included in the # list only if they have an upload ID lexicographically greater than the # specified `upload-id-marker`. # # This functionality is not supported for directory buckets. # # # @option options [String] :expected_bucket_owner # The account ID of the expected bucket owner. If the account ID that # you provide does not match the actual owner of the bucket, the request # fails with the HTTP status code `403 Forbidden` (access denied). # @option options [String] :request_payer # Confirms that the requester knows that they will be charged for the # request. Bucket owners need not specify this parameter in their # requests. If either the source or destination S3 bucket has Requester # Pays enabled, the requester will pay for corresponding charges to copy # the object. For information about downloading objects from Requester # Pays buckets, see [Downloading Objects in Requester Pays Buckets][1] # in the *Amazon S3 User Guide*. # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/ObjectsinRequesterPaysBuckets.html # @return [MultipartUpload::Collection] def multipart_uploads(options = {}) batches = Enumerator.new do |y| options = options.merge(bucket: @name) resp = Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do @client.list_multipart_uploads(options) end resp.each_page do |page| batch = [] page.data.uploads.each do |u| batch << MultipartUpload.new( bucket_name: @name, object_key: u.key, id: u.upload_id, data: u, client: @client ) end y.yield(batch) end end MultipartUpload::Collection.new(batches) end # @return [BucketNotification] def notification BucketNotification.new( bucket_name: @name, client: @client ) end # @param [String] key # @return [Object] def object(key) Object.new( bucket_name: @name, key: key, client: @client ) end # @example Request syntax with placeholder values # # object_versions = bucket.object_versions({ # delimiter: "Delimiter", # encoding_type: "url", # accepts url # key_marker: "KeyMarker", # prefix: "Prefix", # version_id_marker: "VersionIdMarker", # expected_bucket_owner: "AccountId", # request_payer: "requester", # accepts requester # optional_object_attributes: ["RestoreStatus"], # accepts RestoreStatus # }) # @param [Hash] options ({}) # @option options [String] :delimiter # A delimiter is a character that you specify to group keys. All keys # that contain the same string between the `prefix` and the first # occurrence of the delimiter are grouped under a single result element # in `CommonPrefixes`. These groups are counted as one result against # the `max-keys` limitation. These keys are not returned elsewhere in # the response. # @option options [String] :encoding_type # Encoding type used by Amazon S3 to encode the [object keys][1] in the # response. Responses are encoded only in UTF-8. An object key can # contain any Unicode character. However, the XML 1.0 parser can't # parse certain characters, such as characters with an ASCII value from # 0 to 10. For characters that aren't supported in XML 1.0, you can add # this parameter to request that Amazon S3 encode the keys in the # response. For more information about characters to avoid in object key # names, see [Object key naming guidelines][2]. # # When using the URL encoding type, non-ASCII characters that are used # in an object's key name will be percent-encoded according to UTF-8 # code values. For example, the object `test_file(3).png` will appear as # `test_file%283%29.png`. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/object-keys.html # [2]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/object-keys.html#object-key-guidelines # @option options [String] :key_marker # Specifies the key to start with when listing objects in a bucket. # @option options [String] :prefix # Use this parameter to select only those keys that begin with the # specified prefix. You can use prefixes to separate a bucket into # different groupings of keys. (You can think of using `prefix` to make # groups in the same way that you'd use a folder in a file system.) You # can use `prefix` with `delimiter` to roll up numerous objects into a # single result under `CommonPrefixes`. # @option options [String] :version_id_marker # Specifies the object version you want to start listing from. # @option options [String] :expected_bucket_owner # The account ID of the expected bucket owner. If the account ID that # you provide does not match the actual owner of the bucket, the request # fails with the HTTP status code `403 Forbidden` (access denied). # @option options [String] :request_payer # Confirms that the requester knows that they will be charged for the # request. Bucket owners need not specify this parameter in their # requests. If either the source or destination S3 bucket has Requester # Pays enabled, the requester will pay for corresponding charges to copy # the object. For information about downloading objects from Requester # Pays buckets, see [Downloading Objects in Requester Pays Buckets][1] # in the *Amazon S3 User Guide*. # # This functionality is not supported for directory buckets. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/ObjectsinRequesterPaysBuckets.html # @option options [Array] :optional_object_attributes # Specifies the optional fields that you want returned in the response. # Fields that you do not specify are not returned. # @return [ObjectVersion::Collection] def object_versions(options = {}) batches = Enumerator.new do |y| options = options.merge(bucket: @name) resp = Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do @client.list_object_versions(options) end resp.each_page do |page| batch = [] page.data.versions_delete_markers.each do |v| batch << ObjectVersion.new( bucket_name: @name, object_key: v.key, id: v.version_id, data: v, client: @client ) end y.yield(batch) end end ObjectVersion::Collection.new(batches) end # @example Request syntax with placeholder values # # objects = bucket.objects({ # delimiter: "Delimiter", # encoding_type: "url", # accepts url # prefix: "Prefix", # fetch_owner: false, # start_after: "StartAfter", # request_payer: "requester", # accepts requester # expected_bucket_owner: "AccountId", # optional_object_attributes: ["RestoreStatus"], # accepts RestoreStatus # }) # @param [Hash] options ({}) # @option options [String] :delimiter # A delimiter is a character that you use to group keys. # # * **Directory buckets** - For directory buckets, `/` is the only # supported delimiter. # # * Directory buckets - When you query `ListObjectsV2` with a # delimiter during in-progress multipart uploads, the `CommonPrefixes` # response parameter contains the prefixes that are associated with # the in-progress multipart uploads. For more information about # multipart uploads, see [Multipart Upload Overview][1] in the *Amazon # S3 User Guide*. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/dev/mpuoverview.html # @option options [String] :encoding_type # Encoding type used by Amazon S3 to encode the [object keys][1] in the # response. Responses are encoded only in UTF-8. An object key can # contain any Unicode character. However, the XML 1.0 parser can't # parse certain characters, such as characters with an ASCII value from # 0 to 10. For characters that aren't supported in XML 1.0, you can add # this parameter to request that Amazon S3 encode the keys in the # response. For more information about characters to avoid in object key # names, see [Object key naming guidelines][2]. # # When using the URL encoding type, non-ASCII characters that are used # in an object's key name will be percent-encoded according to UTF-8 # code values. For example, the object `test_file(3).png` will appear as # `test_file%283%29.png`. # # # # # # [1]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/object-keys.html # [2]: https://docs.aws.amazon.com/AmazonS3/latest/userguide/object-keys.html#object-key-guidelines # @option options [String] :prefix # Limits the response to keys that begin with the specified prefix. # # **Directory buckets** - For directory buckets, only prefixes that end # in a delimiter (`/`) are supported. # # # @option options [Boolean] :fetch_owner # The owner field is not present in `ListObjectsV2` by default. If you # want to return the owner field with each key in the result, then set # the `FetchOwner` field to `true`. # # **Directory buckets** - For directory buckets, the bucket owner is # returned as the object owner for all objects. # # # @option options [String] :start_after # StartAfter is where you want Amazon S3 to start listing from. Amazon # S3 starts listing after this specified key. StartAfter can be any key # in the bucket. # # This functionality is not supported for directory buckets. # # # @option options [String] :request_payer # Confirms that the requester knows that she or he will be charged for # the list objects request in V2 style. Bucket owners need not specify # this parameter in their requests. # # This functionality is not supported for directory buckets. # # # @option options [String] :expected_bucket_owner # The account ID of the expected bucket owner. If the account ID that # you provide does not match the actual owner of the bucket, the request # fails with the HTTP status code `403 Forbidden` (access denied). # @option options [Array] :optional_object_attributes # Specifies the optional fields that you want returned in the response. # Fields that you do not specify are not returned. # # This functionality is not supported for directory buckets. # # # @return [ObjectSummary::Collection] def objects(options = {}) batches = Enumerator.new do |y| options = options.merge(bucket: @name) resp = Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do @client.list_objects_v2(options) end resp.each_page do |page| batch = [] page.data.contents.each do |c| batch << ObjectSummary.new( bucket_name: @name, key: c.key, data: c, client: @client ) end y.yield(batch) end end ObjectSummary::Collection.new(batches) end # @return [BucketPolicy] def policy BucketPolicy.new( bucket_name: @name, client: @client ) end # @return [BucketRequestPayment] def request_payment BucketRequestPayment.new( bucket_name: @name, client: @client ) end # @return [BucketTagging] def tagging BucketTagging.new( bucket_name: @name, client: @client ) end # @return [BucketVersioning] def versioning BucketVersioning.new( bucket_name: @name, client: @client ) end # @return [BucketWebsite] def website BucketWebsite.new( bucket_name: @name, client: @client ) end # @deprecated # @api private def identifiers { name: @name } end deprecated(:identifiers) private def extract_name(args, options) value = args[0] || options.delete(:name) case value when String then value when nil then raise ArgumentError, "missing required option :name" else msg = "expected :name to be a String, got #{value.class}" raise ArgumentError, msg end end def yield_waiter_and_warn(waiter, &block) if !@waiter_block_warned msg = "pass options to configure the waiter; "\ "yielding the waiter is deprecated" warn(msg) @waiter_block_warned = true end yield(waiter.waiter) end def separate_params_and_options(options) opts = Set.new( [:client, :max_attempts, :delay, :before_attempt, :before_wait] ) waiter_opts = {} waiter_params = {} options.each_pair do |key, value| if opts.include?(key) waiter_opts[key] = value else waiter_params[key] = value end end waiter_opts[:client] ||= @client [waiter_opts, waiter_params] end class Collection < Aws::Resources::Collection; end end end # Load customizations if they exist require 'aws-sdk-s3/customizations/bucket'