module CarrierWave module Storage ## # Stores things using the "fog" gem. # # fog supports storing files with AWS, Google, Local and Rackspace # # You need to setup some options to configure your usage: # # [:fog_credentials] host info and credentials for service # [:fog_directory] specifies name of directory to store data in, assumed to already exist # # [:fog_attributes] (optional) additional attributes to set on files # [:fog_public] (optional) public readability, defaults to true # [:fog_authenticated_url_expiration] (optional) time (in seconds) that authenticated urls # will be valid, when fog_public is false and provider is AWS or Google, defaults to 600 # [:fog_use_ssl_for_aws] (optional) #public_url will use https for the AWS generated URL] # [:fog_aws_accelerate] (optional) #public_url will use s3-accelerate subdomain # instead of s3, defaults to false # # # AWS credentials contain the following keys: # # [:aws_access_key_id] # [:aws_secret_access_key] # [:region] (optional) defaults to 'us-east-1' # :region should be one of ['eu-west-1', 'us-east-1', 'ap-southeast-1', 'us-west-1', 'ap-northeast-1', 'eu-central-1'] # # # Google credentials contain the following keys: # [:google_storage_access_key_id] # [:google_storage_secret_access_key] # # # Local credentials contain the following keys: # # [:local_root] local path to files # # # Rackspace credentials contain the following keys: # # [:rackspace_username] # [:rackspace_api_key] # # # A full example with AWS credentials: # CarrierWave.configure do |config| # config.fog_credentials = { # :aws_access_key_id => 'xxxxxx', # :aws_secret_access_key => 'yyyyyy', # :provider => 'AWS' # } # config.fog_directory = 'directoryname' # config.fog_public = true # end # class Fog < Abstract class << self def connection_cache @connection_cache ||= {} end def eager_load # see #1198. This will hopefully no longer be necessary in future release of fog fog_credentials = CarrierWave::Uploader::Base.fog_credentials if fog_credentials.present? CarrierWave::Storage::Fog.connection_cache[fog_credentials] ||= ::Fog::Storage.new(fog_credentials) end end end ## # Store a file # # === Parameters # # [file (CarrierWave::SanitizedFile)] the file to store # # === Returns # # [CarrierWave::Storage::Fog::File] the stored file # def store!(file) f = CarrierWave::Storage::Fog::File.new(uploader, self, uploader.store_path) f.store(file) f end ## # Retrieve a file # # === Parameters # # [identifier (String)] unique identifier for file # # === Returns # # [CarrierWave::Storage::Fog::File] the stored file # def retrieve!(identifier) CarrierWave::Storage::Fog::File.new(uploader, self, uploader.store_path(identifier)) end ## # Stores given file to cache directory. # # === Parameters # # [new_file (File, IOString, Tempfile)] any kind of file object # # === Returns # # [CarrierWave::SanitizedFile] a sanitized file # def cache!(new_file) f = CarrierWave::Storage::Fog::File.new(uploader, self, uploader.cache_path) f.store(new_file) f end ## # Retrieves the file with the given cache_name from the cache. # # === Parameters # # [cache_name (String)] uniquely identifies a cache file # # === Raises # # [CarrierWave::InvalidParameter] if the cache_name is incorrectly formatted. # def retrieve_from_cache!(identifier) CarrierWave::Storage::Fog::File.new(uploader, self, uploader.cache_path(identifier)) end ## # Deletes a cache dir # def delete_dir!(path) # do nothing, because there's no such things as 'empty directory' end def clean_cache!(seconds) connection.directories.new( :key => uploader.fog_directory, :public => uploader.fog_public ).files.all(:prefix => uploader.cache_dir).each do |file| # generate_cache_id returns key formatted TIMEINT-PID(-COUNTER)-RND matched = file.key.match(/(\d+)-\d+-\d+(?:-\d+)?/) next unless matched time = Time.at(matched[1].to_i) file.destroy if time < (Time.now.utc - seconds) end end def connection @connection ||= begin options = credentials = uploader.fog_credentials self.class.connection_cache[credentials] ||= ::Fog::Storage.new(options) end end class File DEFAULT_S3_REGION = 'us-east-1' include CarrierWave::Utilities::Uri include CarrierWave::Utilities::FileName ## # Current local path to file # # === Returns # # [String] a path to file # attr_reader :path ## # Return all attributes from file # # === Returns # # [Hash] attributes from file # def attributes file.attributes end ## # Return a temporary authenticated url to a private file, if available # Only supported for AWS, Rackspace, Google, AzureRM and Aliyun providers # # === Returns # # [String] temporary authenticated url # or # [NilClass] no authenticated url available # def authenticated_url(options = {}) if ['AWS', 'Google', 'Rackspace', 'OpenStack', 'AzureRM', 'Aliyun', 'backblaze'].include?(@uploader.fog_credentials[:provider]) # avoid a get by using local references local_directory = connection.directories.new(:key => @uploader.fog_directory) local_file = local_directory.files.new(:key => path) expire_at = options[:expire_at] || ::Fog::Time.now.since(@uploader.fog_authenticated_url_expiration.to_i) case @uploader.fog_credentials[:provider] when 'AWS', 'Google' # Older versions of fog-google do not support options as a parameter if url_options_supported?(local_file) local_file.url(expire_at, options) else warn "Options hash not supported in #{local_file.class}. You may need to upgrade your Fog provider." local_file.url(expire_at) end when 'Rackspace', 'OpenStack' connection.get_object_https_url(@uploader.fog_directory, path, expire_at, options) when 'Aliyun' expire_at = expire_at - Time.now local_file.url(expire_at) else local_file.url(expire_at) end end end ## # Lookup value for file content-type header # # === Returns # # [String] value of content-type # def content_type @content_type || file.try(:content_type) end ## # Set non-default content-type header (default is file.content_type) # # === Returns # # [String] returns new content type value # def content_type=(new_content_type) @content_type = new_content_type end ## # Remove the file from service # # === Returns # # [Boolean] true for success or raises error # def delete # avoid a get by just using local reference directory.files.new(:key => path).destroy.tap do |result| @file = nil if result end end ## # deprecated: All attributes from file (includes headers) # # === Returns # # [Hash] attributes from file # def headers location = caller.first warning = "[yellow][WARN] headers is deprecated, use attributes instead[/]" warning << " [light_black](#{location})[/]" Formatador.display_line(warning) attributes end def initialize(uploader, base, path) @uploader, @base, @path, @content_type = uploader, base, path, nil end ## # Read content of file from service # # === Returns # # [String] contents of file def read file_body = file.body return if file_body.nil? return file_body unless file_body.is_a?(::File) # Fog::Storage::XXX::File#body could return the source file which was upoloaded to the remote server. read_source_file(file_body) if ::File.exist?(file_body.path) # If the source file doesn't exist, the remote content is read @file = nil # rubocop:disable Gitlab/ModuleWithInstanceVariables file.body end ## # Return size of file body # # === Returns # # [Integer] size of file body # def size file.nil? ? 0 : file.content_length end ## # Check if the file exists on the remote service # # === Returns # # [Boolean] true if file exists or false def exists? !!file end ## # Write file to service # # === Returns # # [Boolean] true on success or raises error def store(new_file) if new_file.is_a?(self.class) new_file.copy_to(path) else fog_file = new_file.to_file @content_type ||= new_file.content_type @file = directory.files.create({ :body => fog_file ? fog_file : new_file.read, :content_type => @content_type, :key => path, :public => @uploader.fog_public }.merge(@uploader.fog_attributes)) fog_file.close if fog_file && !fog_file.closed? end true end ## # Return a url to a public file, if available # # === Returns # # [String] public url # or # [NilClass] no public url available # def public_url encoded_path = encode_path(path) if host = @uploader.asset_host if host.respond_to? :call "#{host.call(self)}/#{encoded_path}" else "#{host}/#{encoded_path}" end else # AWS/Google optimized for speed over correctness case fog_provider when 'AWS' # check if some endpoint is set in fog_credentials if @uploader.fog_credentials.has_key?(:endpoint) "#{@uploader.fog_credentials[:endpoint]}/#{@uploader.fog_directory}/#{encoded_path}" else protocol = @uploader.fog_use_ssl_for_aws ? "https" : "http" subdomain_regex = /^(?:[a-z]|\d(?!\d{0,2}(?:\d{1,3}){3}$))(?:[a-z0-9\.]|(?![\-])|\-(?![\.])){1,61}[a-z0-9]$/ valid_subdomain = @uploader.fog_directory.to_s =~ subdomain_regex && !(protocol == 'https' && @uploader.fog_directory =~ /\./) # if directory is a valid subdomain, use that style for access if valid_subdomain s3_subdomain = @uploader.fog_aws_accelerate ? "s3-accelerate" : "s3" "#{protocol}://#{@uploader.fog_directory}.#{s3_subdomain}.amazonaws.com/#{encoded_path}" else # directory is not a valid subdomain, so use path style for access region = @uploader.fog_credentials[:region].to_s host = case region when DEFAULT_S3_REGION, '' 's3.amazonaws.com' else "s3.#{region}.amazonaws.com" end "#{protocol}://#{host}/#{@uploader.fog_directory}/#{encoded_path}" end end when 'Google' # https://cloud.google.com/storage/docs/access-public-data "https://storage.googleapis.com/#{@uploader.fog_directory}/#{encoded_path}" else # avoid a get by just using local reference directory.files.new(:key => path).public_url end end end ## # Return url to file, if available # # === Returns # # [String] url # or # [NilClass] no url available # def url(options = {}) if !@uploader.fog_public authenticated_url(options) else public_url end end ## # Return file name, if available # # === Returns # # [String] file name # or # [NilClass] no file name available # def filename(options = {}) return unless file_url = url(options) CGI.unescape(file_url.split('?').first).gsub(/.*\/(.*?$)/, '\1') end ## # Creates a copy of this file and returns it. # # === Parameters # # [new_path (String)] The path where the file should be copied to. # # === Returns # # @return [CarrierWave::Storage::Fog::File] the location where the file will be stored. # def copy_to(new_path) file.copy(@uploader.fog_directory, new_path, copy_options) CarrierWave::Storage::Fog::File.new(@uploader, @base, new_path) end private ## # connection to service # # === Returns # # [Fog::#{provider}::Storage] connection to service # def connection @base.connection end ## # local reference to directory containing file # # === Returns # # [Fog::#{provider}::Directory] containing directory # def directory @directory ||= begin connection.directories.new( :key => @uploader.fog_directory, :public => @uploader.fog_public ) end end ## # lookup file # # === Returns # # [Fog::#{provider}::File] file data from remote service # def file @file ||= directory.files.head(path) end def copy_options options = {} options.merge!(acl_header) if acl_header.present? options[fog_provider == "Google" ? :content_type : 'Content-Type'] ||= content_type if content_type options.merge(@uploader.fog_attributes) end def acl_header if fog_provider == 'AWS' { 'x-amz-acl' => @uploader.fog_public ? 'public-read' : 'private' } elsif fog_provider == "Google" @uploader.fog_public ? { destination_predefined_acl: "publicRead" } : {} else {} end end def fog_provider @uploader.fog_credentials[:provider].to_s end def read_source_file(file_body) return unless ::File.exist?(file_body.path) begin file_body = ::File.open(file_body.path) if file_body.closed? # Reopen if it's already closed file_body.read ensure file_body.close end end def url_options_supported?(local_file) parameters = local_file.method(:url).parameters parameters.count == 2 && parameters[1].include?(:options) end end end # Fog end # Storage end # CarrierWave