# frozen_string_literal: true module Aws module S3 class Object alias size content_length # Make the method redefinable alias_method :copy_from, :copy_from # Copies another object to this object. Use `multipart_copy: true` # for large objects. This is required for objects that exceed 5GB. # # @param [S3::Object, S3::ObjectVersion, S3::ObjectSummary, String, Hash] # source Where to copy object data from. `source` must be one of the # following: # # * {Aws::S3::Object} # * {Aws::S3::ObjectSummary} # * {Aws::S3::ObjectVersion} # * Hash - with `:bucket` and `:key` and optional `:version_id` # * String - formatted like `"source-bucket-name/uri-escaped-key"` # or `"source-bucket-name/uri-escaped-key?versionId=version-id"` # # @option options [Boolean] :multipart_copy (false) When `true`, # the object will be copied using the multipart APIs. This is # necessary for objects larger than 5GB and can provide # performance improvements on large objects. Amazon S3 does # not accept multipart copies for objects smaller than 5MB. # Object metadata such as Content-Type will be copied, however, # Checksums are not copied. # # @option options [Integer] :content_length Only used when # `:multipart_copy` is `true`. Passing this options avoids a HEAD # request to query the source object size but prevents object metadata # from being copied. Raises an `ArgumentError` if # this option is provided when `:multipart_copy` is `false` or not set. # # @option options [S3::Client] :copy_source_client Only used when # `:multipart_copy` is `true` and the source object is in a # different region. You do not need to specify this option # if you have provided `:content_length`. # # @option options [String] :copy_source_region Only used when # `:multipart_copy` is `true` and the source object is in a # different region. You do not need to specify this option # if you have provided a `:source_client` or a `:content_length`. # # @option options [Boolean] :use_source_parts (false) Only used when # `:multipart_copy` is `true`. Use part sizes defined on the source # object if any exist. If copying or moving an object that # is already multipart, this does not re-part the object, instead # re-using the part definitions on the original. That means the etag # and any checksums will not change. This is especially useful if the # source object has parts with varied sizes. # # @example Basic object copy # # bucket = Aws::S3::Bucket.new('target-bucket') # object = bucket.object('target-key') # # # source as String # object.copy_from('source-bucket/source-key') # # # source as Hash # object.copy_from(bucket:'source-bucket', key:'source-key') # # # source as Aws::S3::Object # object.copy_from(bucket.object('source-key')) # # @example Managed copy of large objects # # # uses multipart upload APIs to copy object # object.copy_from('src-bucket/src-key', multipart_copy: true) # # @see #copy_to # def copy_from(source, options = {}) Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do if Hash === source && source[:copy_source] # for backwards compatibility @client.copy_object(source.merge(bucket: bucket_name, key: key)) else ObjectCopier.new(self, options).copy_from(source, options) end end end # Copies this object to another object. Use `multipart_copy: true` # for large objects. This is required for objects that exceed 5GB. # # @note If you need to copy to a bucket in a different region, use # {#copy_from}. # # @param [S3::Object, String, Hash] target Where to copy the object # data to. `target` must be one of the following: # # * {Aws::S3::Object} # * Hash - with `:bucket` and `:key` # * String - formatted like `"target-bucket-name/target-key"` # # @example Basic object copy # # bucket = Aws::S3::Bucket.new('source-bucket') # object = bucket.object('source-key') # # # target as String # object.copy_to('target-bucket/target-key') # # # target as Hash # object.copy_to(bucket: 'target-bucket', key: 'target-key') # # # target as Aws::S3::Object # object.copy_to(bucket.object('target-key')) # # @example Managed copy of large objects # # # uses multipart upload APIs to copy object # object.copy_to('src-bucket/src-key', multipart_copy: true) # def copy_to(target, options = {}) Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do ObjectCopier.new(self, options).copy_to(target, options) end end # Copies and deletes the current object. The object will only be deleted # if the copy operation succeeds. # # @param (see Object#copy_to) # @option (see Object#copy_to) # @return [void] # @see Object#copy_to # @see Object#delete def move_to(target, options = {}) copy_to(target, options) delete end # Creates a {PresignedPost} that makes it easy to upload a file from # a web browser direct to Amazon S3 using an HTML post form with # a file field. # # See the {PresignedPost} documentation for more information. # # @option (see PresignedPost#initialize) # @return [PresignedPost] # @see PresignedPost def presigned_post(options = {}) PresignedPost.new( client.config.credentials, client.config.region, bucket_name, { key: key, url: bucket.url }.merge(options) ) end # Generates a pre-signed URL for this object. # # @example Pre-signed GET URL, valid for one hour # # obj.presigned_url(:get, expires_in: 3600) # #=> "https://bucket-name.s3.amazonaws.com/object-key?..." # # @example Pre-signed PUT with a canned ACL # # # the object uploaded using this URL will be publicly accessible # obj.presigned_url(:put, acl: 'public-read') # #=> "https://bucket-name.s3.amazonaws.com/object-key?..." # # @example Pre-signed UploadPart PUT # # # the object uploaded using this URL will be publicly accessible # obj.presigned_url(:upload_part, part_number: 1, upload_id: 'uploadIdToken') # #=> "https://bucket-name.s3.amazonaws.com/object-key?..." # # @param [Symbol] method # The S3 operation to generate a presigned URL for. Valid values # are `:get`, `:put`, `:head`, `:delete`, `:create_multipart_upload`, # `:list_multipart_uploads`, `:complete_multipart_upload`, # `:abort_multipart_upload`, `:list_parts`, and `:upload_part`. # # @param [Hash] params # Additional request parameters to use when generating the pre-signed # URL. See the related documentation in {Client} for accepted # params. # # | Method | Client Method | # |------------------------------|------------------------------------| # | `:get` | {Client#get_object} | # | `:put` | {Client#put_object} | # | `:head` | {Client#head_object} | # | `:delete` | {Client#delete_object} | # | `:create_multipart_upload` | {Client#create_multipart_upload} | # | `:list_multipart_uploads` | {Client#list_multipart_uploads} | # | `:complete_multipart_upload` | {Client#complete_multipart_upload} | # | `:abort_multipart_upload` | {Client#abort_multipart_upload} | # | `:list_parts` | {Client#list_parts} | # | `:upload_part` | {Client#upload_part} | # # @option params [Boolean] :virtual_host (false) When `true` the # presigned URL will use the bucket name as a virtual host. # # bucket = Aws::S3::Bucket.new('my.bucket.com') # bucket.object('key').presigned_url(virtual_host: true) # #=> "http://my.bucket.com/key?..." # # @option params [Integer] :expires_in (900) Number of seconds before # the pre-signed URL expires. This may not exceed one week (604800 # seconds). Note that the pre-signed URL is also only valid as long as # credentials used to sign it are. For example, when using IAM roles, # temporary tokens generated for signing also have a default expiration # which will affect the effective expiration of the pre-signed URL. # # @raise [ArgumentError] Raised if `:expires_in` exceeds one week # (604800 seconds). # # @return [String] # def presigned_url(method, params = {}) presigner = Presigner.new(client: client) if %w(delete head get put).include?(method.to_s) method = "#{method}_object".to_sym end presigner.presigned_url( method.downcase, params.merge(bucket: bucket_name, key: key) ) end # Allows you to create presigned URL requests for S3 operations. This # method returns a tuple containing the URL and the signed X-amz-* headers # to be used with the presigned url. # # @example Pre-signed GET URL, valid for one hour # # obj.presigned_request(:get, expires_in: 3600) # #=> ["https://bucket-name.s3.amazonaws.com/object-key?...", {}] # # @example Pre-signed PUT with a canned ACL # # # the object uploaded using this URL will be publicly accessible # obj.presigned_request(:put, acl: 'public-read') # #=> ["https://bucket-name.s3.amazonaws.com/object-key?...", # {"x-amz-acl"=>"public-read"}] # # @param [Symbol] method # The S3 operation to generate a presigned request for. Valid values # are `:get`, `:put`, `:head`, `:delete`, `:create_multipart_upload`, # `:list_multipart_uploads`, `:complete_multipart_upload`, # `:abort_multipart_upload`, `:list_parts`, and `:upload_part`. # # @param [Hash] params # Additional request parameters to use when generating the pre-signed # request. See the related documentation in {Client} for accepted # params. # # | Method | Client Method | # |------------------------------|------------------------------------| # | `:get` | {Client#get_object} | # | `:put` | {Client#put_object} | # | `:head` | {Client#head_object} | # | `:delete` | {Client#delete_object} | # | `:create_multipart_upload` | {Client#create_multipart_upload} | # | `:list_multipart_uploads` | {Client#list_multipart_uploads} | # | `:complete_multipart_upload` | {Client#complete_multipart_upload} | # | `:abort_multipart_upload` | {Client#abort_multipart_upload} | # | `:list_parts` | {Client#list_parts} | # | `:upload_part` | {Client#upload_part} | # # @option params [Boolean] :virtual_host (false) When `true` the # presigned URL will use the bucket name as a virtual host. # # bucket = Aws::S3::Bucket.new('my.bucket.com') # bucket.object('key').presigned_request(virtual_host: true) # #=> ["http://my.bucket.com/key?...", {}] # # @option params [Integer] :expires_in (900) Number of seconds before # the pre-signed URL expires. This may not exceed one week (604800 # seconds). Note that the pre-signed URL is also only valid as long as # credentials used to sign it are. For example, when using IAM roles, # temporary tokens generated for signing also have a default expiration # which will affect the effective expiration of the pre-signed URL. # # @raise [ArgumentError] Raised if `:expires_in` exceeds one week # (604800 seconds). # # @return [String, Hash] A tuple with a presigned URL and headers that # should be included with the request. # def presigned_request(method, params = {}) presigner = Presigner.new(client: client) if %w(delete head get put).include?(method.to_s) method = "#{method}_object".to_sym end presigner.presigned_request( method.downcase, params.merge(bucket: bucket_name, key: key) ) end # Returns the public (un-signed) URL for this object. # # s3.bucket('bucket-name').object('obj-key').public_url # #=> "https://bucket-name.s3.amazonaws.com/obj-key" # # To use virtual hosted bucket url. # Uses https unless secure: false is set. If the bucket # name contains dots (.) then you will need to set secure: false. # # s3.bucket('my-bucket.com').object('key') # .public_url(virtual_host: true) # #=> "https://my-bucket.com/key" # # @option options [Boolean] :virtual_host (false) When `true`, the bucket # name will be used as the host name. This is useful when you have # a CNAME configured for the bucket. # # @option options [Boolean] :secure (true) When `false`, http # will be used with virtual_host. This is required when # the bucket name has a dot (.) in it. # # @return [String] def public_url(options = {}) url = URI.parse(bucket.url(options)) url.path += '/' unless url.path[-1] == '/' url.path += key.gsub(/[^\/]+/) { |s| Seahorse::Util.uri_escape(s) } url.to_s end # Uploads a stream in a streaming fashion to the current object in S3. # # Passed chunks automatically split into multipart upload parts and the # parts are uploaded in parallel. This allows for streaming uploads that # never touch the disk. # # Note that this is known to have issues in JRuby until jruby-9.1.15.0, # so avoid using this with older versions of JRuby. # # @example Streaming chunks of data # obj.upload_stream do |write_stream| # 10.times { write_stream << 'foo' } # end # @example Streaming chunks of data # obj.upload_stream do |write_stream| # IO.copy_stream(IO.popen('ls'), write_stream) # end # @example Streaming chunks of data # obj.upload_stream do |write_stream| # IO.copy_stream(STDIN, write_stream) # end # @param [Hash] options # Additional options for {Client#create_multipart_upload}, # {Client#complete_multipart_upload}, # and {Client#upload_part} can be provided. # # @option options [Integer] :thread_count (10) The number of parallel # multipart uploads # # @option options [Boolean] :tempfile (false) Normally read data is stored # in memory when building the parts in order to complete the underlying # multipart upload. By passing `:tempfile => true` data read will be # temporarily stored on disk reducing the memory footprint vastly. # # @option options [Integer] :part_size (5242880) # Define how big each part size but the last should be. # Default `:part_size` is `5 * 1024 * 1024`. # # @raise [MultipartUploadError] If an object is being uploaded in # parts, and the upload can not be completed, then the upload is # aborted and this error is raised. The raised error has a `#errors` # method that returns the failures that caused the upload to be # aborted. # # @return [Boolean] Returns `true` when the object is uploaded # without any errors. # # @see Client#create_multipart_upload # @see Client#complete_multipart_upload # @see Client#upload_part def upload_stream(options = {}, &block) uploading_options = options.dup uploader = MultipartStreamUploader.new( client: client, thread_count: uploading_options.delete(:thread_count), tempfile: uploading_options.delete(:tempfile), part_size: uploading_options.delete(:part_size) ) Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do uploader.upload( uploading_options.merge(bucket: bucket_name, key: key), &block ) end true end # Uploads a file from disk to the current object in S3. # # # small files are uploaded in a single API call # obj.upload_file('/path/to/file') # # Files larger than or equal to `:multipart_threshold` are uploaded # using the Amazon S3 multipart upload APIs. # # # large files are automatically split into parts # # and the parts are uploaded in parallel # obj.upload_file('/path/to/very_large_file') # # The response of the S3 upload API is yielded if a block given. # # # API response will have etag value of the file # obj.upload_file('/path/to/file') do |response| # etag = response.etag # end # # You can provide a callback to monitor progress of the upload: # # # bytes and totals are each an array with 1 entry per part # progress = Proc.new do |bytes, totals| # puts bytes.map.with_index { |b, i| "Part #{i+1}: #{b} / #{totals[i]}"}.join(' ') + "Total: #{100.0 * bytes.sum / totals.sum }%" } # end # obj.upload_file('/path/to/file', progress_callback: progress) # # @param [String, Pathname, File, Tempfile] source A file on the local # file system that will be uploaded as this object. This can either be # a String or Pathname to the file, an open File object, or an open # Tempfile object. If you pass an open File or Tempfile object, then # you are responsible for closing it after the upload completes. When # using an open Tempfile, rewind it before uploading or else the object # will be empty. # # @param [Hash] options # Additional options for {Client#put_object} # when file sizes below the multipart threshold. For files larger than # the multipart threshold, options for {Client#create_multipart_upload}, # {Client#complete_multipart_upload}, # and {Client#upload_part} can be provided. # # @option options [Integer] :multipart_threshold (104857600) Files larger # than or equal to `:multipart_threshold` are uploaded using the S3 # multipart APIs. # Default threshold is 100MB. # # @option options [Integer] :thread_count (10) The number of parallel # multipart uploads. This option is not used if the file is smaller than # `:multipart_threshold`. # # @option options [Proc] :progress_callback # A Proc that will be called when each chunk of the upload is sent. # It will be invoked with [bytes_read], [total_sizes] # # @raise [MultipartUploadError] If an object is being uploaded in # parts, and the upload can not be completed, then the upload is # aborted and this error is raised. The raised error has a `#errors` # method that returns the failures that caused the upload to be # aborted. # # @return [Boolean] Returns `true` when the object is uploaded # without any errors. # # @see Client#put_object # @see Client#create_multipart_upload # @see Client#complete_multipart_upload # @see Client#upload_part def upload_file(source, options = {}) uploading_options = options.dup uploader = FileUploader.new( multipart_threshold: uploading_options.delete(:multipart_threshold), client: client ) response = Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do uploader.upload( source, uploading_options.merge(bucket: bucket_name, key: key) ) end yield response if block_given? true end # Downloads a file in S3 to a path on disk. # # # small files (< 5MB) are downloaded in a single API call # obj.download_file('/path/to/file') # # Files larger than 5MB are downloaded using multipart method # # # large files are split into parts # # and the parts are downloaded in parallel # obj.download_file('/path/to/very_large_file') # # You can provide a callback to monitor progress of the download: # # # bytes and part_sizes are each an array with 1 entry per part # # part_sizes may not be known until the first bytes are retrieved # progress = Proc.new do |bytes, part_sizes, file_size| # puts bytes.map.with_index { |b, i| "Part #{i+1}: #{b} / #{part_sizes[i]}"}.join(' ') + "Total: #{100.0 * bytes.sum / file_size}%" } # end # obj.download_file('/path/to/file', progress_callback: progress) # # @param [String] destination Where to download the file to. # # @param [Hash] options # Additional options for {Client#get_object} and #{Client#head_object} # may be provided. # # @option options [String] mode `auto`, `single_request`, `get_range` # `single_request` mode forces only 1 GET request is made in download, # `get_range` mode allows `chunk_size` parameter to configured in # customizing each range size in multipart_download, # By default, `auto` mode is enabled, which performs multipart_download # # @option options [Integer] chunk_size required in get_range mode. # # @option options [Integer] thread_count (10) Customize threads used in # the multipart download. # # @option options [String] version_id The object version id used to # retrieve the object. For more about object versioning, see: # https://docs.aws.amazon.com/AmazonS3/latest/dev/ObjectVersioning.html # # @option options [String] checksum_mode (ENABLED) When `ENABLED` and # the object has a stored checksum, it will be used to validate the # download and will raise an `Aws::Errors::ChecksumError` if # checksum validation fails. You may provide a `on_checksum_validated` # callback if you need to verify that validation occurred and which # algorithm was used. To disable checksum validation, set # `checksum_mode` to "DISABLED". # # @option options [Callable] on_checksum_validated Called each time a # request's checksum is validated with the checksum algorithm and the # response. For multipart downloads, this will be called for each # part that is downloaded and validated. # # @option options [Proc] :progress_callback # A Proc that will be called when each chunk of the download is received. # It will be invoked with [bytes_read], [part_sizes], file_size. # When the object is downloaded as parts (rather than by ranges), the # part_sizes will not be known ahead of time and will be nil in the # callback until the first bytes in the part are received. # # @return [Boolean] Returns `true` when the file is downloaded without # any errors. # # @see Client#get_object # @see Client#head_object def download_file(destination, options = {}) downloader = FileDownloader.new(client: client) Aws::Plugins::UserAgent.metric('RESOURCE_MODEL') do downloader.download( destination, options.merge(bucket: bucket_name, key: key) ) end true end class Collection < Aws::Resources::Collection alias_method :delete, :batch_delete! extend Aws::Deprecations deprecated :delete, use: :batch_delete! end end end end