# frozen_string_literal: true require "zlib" require "active_support/core_ext/array/extract_options" require "active_support/core_ext/enumerable" require "active_support/core_ext/module/attribute_accessors" require "active_support/core_ext/numeric/bytes" require "active_support/core_ext/object/to_param" require "active_support/core_ext/object/try" require "active_support/core_ext/string/inflections" require_relative "cache/coder" require_relative "cache/entry" require_relative "cache/serializer_with_fallback" module ActiveSupport # See ActiveSupport::Cache::Store for documentation. module Cache autoload :FileStore, "active_support/cache/file_store" autoload :MemoryStore, "active_support/cache/memory_store" autoload :MemCacheStore, "active_support/cache/mem_cache_store" autoload :NullStore, "active_support/cache/null_store" autoload :RedisCacheStore, "active_support/cache/redis_cache_store" # These options mean something to all cache implementations. Individual cache # implementations may support additional options. UNIVERSAL_OPTIONS = [ :coder, :compress, :compress_threshold, :compressor, :expire_in, :expired_in, :expires_in, :namespace, :race_condition_ttl, :serializer, :skip_nil, ] # Mapping of canonical option names to aliases that a store will recognize. OPTION_ALIASES = { expires_in: [:expire_in, :expired_in] }.freeze DEFAULT_COMPRESS_LIMIT = 1.kilobyte # Raised by coders when the cache entry can't be deserialized. # This error is treated as a cache miss. DeserializationError = Class.new(StandardError) module Strategy autoload :LocalCache, "active_support/cache/strategy/local_cache" end @format_version = 7.0 class << self attr_accessor :format_version # Creates a new Store object according to the given options. # # If no arguments are passed to this method, then a new # ActiveSupport::Cache::MemoryStore object will be returned. # # If you pass a Symbol as the first argument, then a corresponding cache # store class under the ActiveSupport::Cache namespace will be created. # For example: # # ActiveSupport::Cache.lookup_store(:memory_store) # # => returns a new ActiveSupport::Cache::MemoryStore object # # ActiveSupport::Cache.lookup_store(:mem_cache_store) # # => returns a new ActiveSupport::Cache::MemCacheStore object # # Any additional arguments will be passed to the corresponding cache store # class's constructor: # # ActiveSupport::Cache.lookup_store(:file_store, '/tmp/cache') # # => same as: ActiveSupport::Cache::FileStore.new('/tmp/cache') # # If the first argument is not a Symbol, then it will simply be returned: # # ActiveSupport::Cache.lookup_store(MyOwnCacheStore.new) # # => returns MyOwnCacheStore.new def lookup_store(store = nil, *parameters) case store when Symbol options = parameters.extract_options! retrieve_store_class(store).new(*parameters, **options) when Array lookup_store(*store) when nil ActiveSupport::Cache::MemoryStore.new else store end end # Expands out the +key+ argument into a key that can be used for the # cache store. Optionally accepts a namespace, and all keys will be # scoped within that namespace. # # If the +key+ argument provided is an array, or responds to +to_a+, then # each of elements in the array will be turned into parameters/keys and # concatenated into a single key. For example: # # ActiveSupport::Cache.expand_cache_key([:foo, :bar]) # => "foo/bar" # ActiveSupport::Cache.expand_cache_key([:foo, :bar], "namespace") # => "namespace/foo/bar" # # The +key+ argument can also respond to +cache_key+ or +to_param+. def expand_cache_key(key, namespace = nil) expanded_cache_key = namespace ? +"#{namespace}/" : +"" if prefix = ENV["RAILS_CACHE_ID"] || ENV["RAILS_APP_VERSION"] expanded_cache_key << "#{prefix}/" end expanded_cache_key << retrieve_cache_key(key) expanded_cache_key end private def retrieve_cache_key(key) case when key.respond_to?(:cache_key_with_version) then key.cache_key_with_version when key.respond_to?(:cache_key) then key.cache_key when key.is_a?(Array) then key.map { |element| retrieve_cache_key(element) }.to_param when key.respond_to?(:to_a) then retrieve_cache_key(key.to_a) else key.to_param end.to_s end # Obtains the specified cache store class, given the name of the +store+. # Raises an error when the store class cannot be found. def retrieve_store_class(store) # require_relative cannot be used here because the class might be # provided by another gem, like redis-activesupport for example. require "active_support/cache/#{store}" rescue LoadError => e raise "Could not find cache store adapter for #{store} (#{e})" else ActiveSupport::Cache.const_get(store.to_s.camelize) end end # = Active Support \Cache \Store # # An abstract cache store class. There are multiple cache store # implementations, each having its own additional features. See the classes # under the ActiveSupport::Cache module, e.g. # ActiveSupport::Cache::MemCacheStore. MemCacheStore is currently the most # popular cache store for large production websites. # # Some implementations may not support all methods beyond the basic cache # methods of #fetch, #write, #read, #exist?, and #delete. # # +ActiveSupport::Cache::Store+ can store any Ruby object that is supported # by its +coder+'s +dump+ and +load+ methods. # # cache = ActiveSupport::Cache::MemoryStore.new # # cache.read('city') # => nil # cache.write('city', "Duckburgh") # => true # cache.read('city') # => "Duckburgh" # # cache.write('not serializable', Proc.new {}) # => TypeError # # Keys are always translated into Strings and are case sensitive. When an # object is specified as a key and has a +cache_key+ method defined, this # method will be called to define the key. Otherwise, the +to_param+ # method will be called. Hashes and Arrays can also be used as keys. The # elements will be delimited by slashes, and the elements within a Hash # will be sorted by key so they are consistent. # # cache.read('city') == cache.read(:city) # => true # # Nil values can be cached. # # If your cache is on a shared infrastructure, you can define a namespace # for your cache entries. If a namespace is defined, it will be prefixed on # to every key. The namespace can be either a static value or a Proc. If it # is a Proc, it will be invoked when each key is evaluated so that you can # use application logic to invalidate keys. # # cache.namespace = -> { @last_mod_time } # Set the namespace to a variable # @last_mod_time = Time.now # Invalidate the entire cache by changing namespace # class Store cattr_accessor :logger, instance_writer: true cattr_accessor :raise_on_invalid_cache_expiration_time, default: false attr_reader :silence, :options alias :silence? :silence class << self private DEFAULT_POOL_OPTIONS = { size: 5, timeout: 5 }.freeze private_constant :DEFAULT_POOL_OPTIONS def retrieve_pool_options(options) if options.key?(:pool) pool_options = options.delete(:pool) else pool_options = true end case pool_options when false, nil return false when true pool_options = DEFAULT_POOL_OPTIONS when Hash pool_options[:size] = Integer(pool_options[:size]) if pool_options.key?(:size) pool_options[:timeout] = Float(pool_options[:timeout]) if pool_options.key?(:timeout) pool_options = DEFAULT_POOL_OPTIONS.merge(pool_options) else raise TypeError, "Invalid :pool argument, expected Hash, got: #{pool_options.inspect}" end pool_options unless pool_options.empty? end end # Creates a new cache. # # ==== Options # # [+:namespace+] # Sets the namespace for the cache. This option is especially useful if # your application shares a cache with other applications. # # [+:serializer+] # The serializer for cached values. Must respond to +dump+ and +load+. # # The default serializer depends on the cache format version (set via # +config.active_support.cache_format_version+ when using Rails). The # default serializer for each format version includes a fallback # mechanism to deserialize values from any format version. This behavior # makes it easy to migrate between format versions without invalidating # the entire cache. # # You can also specify serializer: :message_pack to use a # preconfigured serializer based on ActiveSupport::MessagePack. The # +:message_pack+ serializer includes the same deserialization fallback # mechanism, allowing easy migration from (or to) the default # serializer. The +:message_pack+ serializer may improve performance, # but it requires the +msgpack+ gem. # # [+:compressor+] # The compressor for serialized cache values. Must respond to +deflate+ # and +inflate+. # # The default compressor is +Zlib+. To define a new custom compressor # that also decompresses old cache entries, you can check compressed # values for Zlib's "\x78" signature: # # module MyCompressor # def self.deflate(dumped) # # compression logic... (make sure result does not start with "\x78"!) # end # # def self.inflate(compressed) # if compressed.start_with?("\x78") # Zlib.inflate(compressed) # else # # decompression logic... # end # end # end # # ActiveSupport::Cache.lookup_store(:redis_cache_store, compressor: MyCompressor) # # [+:coder+] # The coder for serializing and (optionally) compressing cache entries. # Must respond to +dump+ and +load+. # # The default coder composes the serializer and compressor, and includes # some performance optimizations. If you only need to override the # serializer or compressor, you should specify the +:serializer+ or # +:compressor+ options instead. # # If the store can handle cache entries directly, you may also specify # coder: nil to omit the serializer, compressor, and coder. For # example, if you are using ActiveSupport::Cache::MemoryStore and can # guarantee that cache values will not be mutated, you can specify # coder: nil to avoid the overhead of safeguarding against # mutation. # # The +:coder+ option is mutually exclusive with the +:serializer+ and # +:compressor+ options. Specifying them together will raise an # +ArgumentError+. # # Any other specified options are treated as default options for the # relevant cache operations, such as #read, #write, and #fetch. def initialize(options = nil) @options = options ? validate_options(normalize_options(options)) : {} @options[:compress] = true unless @options.key?(:compress) @options[:compress_threshold] ||= DEFAULT_COMPRESS_LIMIT @coder = @options.delete(:coder) do legacy_serializer = Cache.format_version < 7.1 && !@options[:serializer] serializer = @options.delete(:serializer) || default_serializer serializer = Cache::SerializerWithFallback[serializer] if serializer.is_a?(Symbol) compressor = @options.delete(:compressor) { Zlib } Cache::Coder.new(serializer, compressor, legacy_serializer: legacy_serializer) end @coder ||= Cache::SerializerWithFallback[:passthrough] @coder_supports_compression = @coder.respond_to?(:dump_compressed) end # Silences the logger. def silence! @silence = true self end # Silences the logger within a block. def mute previous_silence, @silence = @silence, true yield ensure @silence = previous_silence end # Fetches data from the cache, using the given key. If there is data in # the cache with the given key, then that data is returned. # # If there is no such data in the cache (a cache miss), then +nil+ will be # returned. However, if a block has been passed, that block will be passed # the key and executed in the event of a cache miss. The return value of the # block will be written to the cache under the given cache key, and that # return value will be returned. # # cache.write('today', 'Monday') # cache.fetch('today') # => "Monday" # # cache.fetch('city') # => nil # cache.fetch('city') do # 'Duckburgh' # end # cache.fetch('city') # => "Duckburgh" # # ==== Options # # Internally, +fetch+ calls +read_entry+, and calls +write_entry+ on a # cache miss. Thus, +fetch+ supports the same options as #read and #write. # Additionally, +fetch+ supports the following options: # # * force: true - Forces a cache "miss," meaning we treat the # cache value as missing even if it's present. Passing a block is # required when +force+ is true so this always results in a cache write. # # cache.write('today', 'Monday') # cache.fetch('today', force: true) { 'Tuesday' } # => 'Tuesday' # cache.fetch('today', force: true) # => ArgumentError # # The +:force+ option is useful when you're calling some other method to # ask whether you should force a cache write. Otherwise, it's clearer to # just call +write+. # # * skip_nil: true - Prevents caching a nil result: # # cache.fetch('foo') { nil } # cache.fetch('bar', skip_nil: true) { nil } # cache.exist?('foo') # => true # cache.exist?('bar') # => false # # * +:race_condition_ttl+ - Specifies the number of seconds during which # an expired value can be reused while a new value is being generated. # This can be used to prevent race conditions when cache entries expire, # by preventing multiple processes from simultaneously regenerating the # same entry (also known as the dog pile effect). # # When a process encounters a cache entry that has expired less than # +:race_condition_ttl+ seconds ago, it will bump the expiration time by # +:race_condition_ttl+ seconds before generating a new value. During # this extended time window, while the process generates a new value, # other processes will continue to use the old value. After the first # process writes the new value, other processes will then use it. # # If the first process errors out while generating a new value, another # process can try to generate a new value after the extended time window # has elapsed. # # # Set all values to expire after one minute. # cache = ActiveSupport::Cache::MemoryStore.new(expires_in: 1) # # cache.write("foo", "original value") # val_1 = nil # val_2 = nil # p cache.read("foo") # => "original value" # # sleep 1 # wait until the cache expires # # t1 = Thread.new do # # fetch does the following: # # 1. gets an recent expired entry # # 2. extends the expiry by 2 seconds (race_condition_ttl) # # 3. regenerates the new value # val_1 = cache.fetch("foo", race_condition_ttl: 2) do # sleep 1 # "new value 1" # end # end # # # Wait until t1 extends the expiry of the entry # # but before generating the new value # sleep 0.1 # # val_2 = cache.fetch("foo", race_condition_ttl: 2) do # # This block won't be executed because t1 extended the expiry # "new value 2" # end # # t1.join # # p val_1 # => "new value 1" # p val_2 # => "original value" # p cache.fetch("foo") # => "new value 1" # # # The entry requires 3 seconds to expire (expires_in + race_condition_ttl) # # We have waited 2 seconds already (sleep(1) + t1.join) thus we need to wait 1 # # more second to see the entry expire. # sleep 1 # # p cache.fetch("foo") # => nil # # ==== Dynamic Options # # In some cases it may be necessary to dynamically compute options based # on the cached value. To support this, an ActiveSupport::Cache::WriteOptions # instance is passed as the second argument to the block. For example: # # cache.fetch("authentication-token:#{user.id}") do |key, options| # token = authenticate_to_service # options.expires_at = token.expires_at # token # end # def fetch(name, options = nil, &block) if block_given? options = merged_options(options) key = normalize_key(name, options) entry = nil unless options[:force] instrument(:read, key, options) do |payload| cached_entry = read_entry(key, **options, event: payload) entry = handle_expired_entry(cached_entry, key, options) if entry if entry.mismatched?(normalize_version(name, options)) entry = nil else begin entry.value rescue DeserializationError entry = nil end end end payload[:super_operation] = :fetch if payload payload[:hit] = !!entry if payload end end if entry get_entry_value(entry, name, options) else save_block_result_to_cache(name, key, options, &block) end elsif options && options[:force] raise ArgumentError, "Missing block: Calling `Cache#fetch` with `force: true` requires a block." else read(name, options) end end # Reads data from the cache, using the given key. If there is data in # the cache with the given key, then that data is returned. Otherwise, # +nil+ is returned. # # Note, if data was written with the :expires_in or # :version options, both of these conditions are applied before # the data is returned. # # ==== Options # # * +:namespace+ - Replace the store namespace for this call. # * +:version+ - Specifies a version for the cache entry. If the cached # version does not match the requested version, the read will be treated # as a cache miss. This feature is used to support recyclable cache keys. # # Other options will be handled by the specific cache store implementation. def read(name, options = nil) options = merged_options(options) key = normalize_key(name, options) version = normalize_version(name, options) instrument(:read, key, options) do |payload| entry = read_entry(key, **options, event: payload) if entry if entry.expired? delete_entry(key, **options) payload[:hit] = false if payload nil elsif entry.mismatched?(version) payload[:hit] = false if payload nil else payload[:hit] = true if payload begin entry.value rescue DeserializationError payload[:hit] = false nil end end else payload[:hit] = false if payload nil end end end # Reads multiple values at once from the cache. Options can be passed # in the last argument. # # Some cache implementation may optimize this method. # # Returns a hash mapping the names provided to the values found. def read_multi(*names) return {} if names.empty? options = names.extract_options! options = merged_options(options) keys = names.map { |name| normalize_key(name, options) } instrument_multi :read_multi, keys, options do |payload| read_multi_entries(names, **options, event: payload).tap do |results| payload[:hits] = results.keys.map { |name| normalize_key(name, options) } end end end # Cache Storage API to write multiple values at once. def write_multi(hash, options = nil) return hash if hash.empty? options = merged_options(options) normalized_hash = hash.transform_keys { |key| normalize_key(key, options) } instrument_multi :write_multi, normalized_hash, options do |payload| entries = hash.each_with_object({}) do |(name, value), memo| memo[normalize_key(name, options)] = Entry.new(value, **options.merge(version: normalize_version(name, options))) end write_multi_entries entries, **options end end # Fetches data from the cache, using the given keys. If there is data in # the cache with the given keys, then that data is returned. Otherwise, # the supplied block is called for each key for which there was no data, # and the result will be written to the cache and returned. # Therefore, you need to pass a block that returns the data to be written # to the cache. If you do not want to write the cache when the cache is # not found, use #read_multi. # # Returns a hash with the data for each of the names. For example: # # cache.write("bim", "bam") # cache.fetch_multi("bim", "unknown_key") do |key| # "Fallback value for key: #{key}" # end # # => { "bim" => "bam", # # "unknown_key" => "Fallback value for key: unknown_key" } # # You may also specify additional options via the +options+ argument. See #fetch for details. # Other options are passed to the underlying cache implementation. For example: # # cache.fetch_multi("fizz", expires_in: 5.seconds) do |key| # "buzz" # end # # => {"fizz"=>"buzz"} # cache.read("fizz") # # => "buzz" # sleep(6) # cache.read("fizz") # # => nil def fetch_multi(*names) raise ArgumentError, "Missing block: `Cache#fetch_multi` requires a block." unless block_given? return {} if names.empty? options = names.extract_options! options = merged_options(options) keys = names.map { |name| normalize_key(name, options) } writes = {} ordered = instrument_multi :read_multi, keys, options do |payload| if options[:force] reads = {} else reads = read_multi_entries(names, **options) end ordered = names.index_with do |name| reads.fetch(name) { writes[name] = yield(name) } end writes.compact! if options[:skip_nil] payload[:hits] = reads.keys.map { |name| normalize_key(name, options) } payload[:super_operation] = :fetch_multi ordered end write_multi(writes, options) ordered end # Writes the value to the cache with the key. The value must be supported # by the +coder+'s +dump+ and +load+ methods. # # Returns +true+ if the write succeeded, +nil+ if there was an error talking # to the cache backend, or +false+ if the write failed for another reason. # # By default, cache entries larger than 1kB are compressed. Compression # allows more data to be stored in the same memory footprint, leading to # fewer cache evictions and higher hit rates. # # ==== Options # # * compress: false - Disables compression of the cache entry. # # * +:compress_threshold+ - The compression threshold, specified in bytes. # \Cache entries larger than this threshold will be compressed. Defaults # to +1.kilobyte+. # # * +:expires_in+ - Sets a relative expiration time for the cache entry, # specified in seconds. +:expire_in+ and +:expired_in+ are aliases for # +:expires_in+. # # cache = ActiveSupport::Cache::MemoryStore.new(expires_in: 5.minutes) # cache.write(key, value, expires_in: 1.minute) # Set a lower value for one entry # # * +:expires_at+ - Sets an absolute expiration time for the cache entry. # # cache = ActiveSupport::Cache::MemoryStore.new # cache.write(key, value, expires_at: Time.now.at_end_of_hour) # # * +:version+ - Specifies a version for the cache entry. When reading # from the cache, if the cached version does not match the requested # version, the read will be treated as a cache miss. This feature is # used to support recyclable cache keys. # # Other options will be handled by the specific cache store implementation. def write(name, value, options = nil) options = merged_options(options) key = normalize_key(name, options) instrument(:write, key, options) do entry = Entry.new(value, **options.merge(version: normalize_version(name, options))) write_entry(key, entry, **options) end end # Deletes an entry in the cache. Returns +true+ if an entry is deleted # and +false+ otherwise. # # Options are passed to the underlying cache implementation. def delete(name, options = nil) options = merged_options(options) key = normalize_key(name, options) instrument(:delete, key, options) do delete_entry(key, **options) end end # Deletes multiple entries in the cache. Returns the number of deleted # entries. # # Options are passed to the underlying cache implementation. def delete_multi(names, options = nil) return 0 if names.empty? options = merged_options(options) names.map! { |key| normalize_key(key, options) } instrument_multi(:delete_multi, names, options) do delete_multi_entries(names, **options) end end # Returns +true+ if the cache contains an entry for the given key. # # Options are passed to the underlying cache implementation. def exist?(name, options = nil) options = merged_options(options) key = normalize_key(name, options) instrument(:exist?, key) do |payload| entry = read_entry(key, **options, event: payload) (entry && !entry.expired? && !entry.mismatched?(normalize_version(name, options))) || false end end def new_entry(value, options = nil) # :nodoc: Entry.new(value, **merged_options(options)) end # Deletes all entries with keys matching the pattern. # # Options are passed to the underlying cache implementation. # # Some implementations may not support this method. def delete_matched(matcher, options = nil) raise NotImplementedError.new("#{self.class.name} does not support delete_matched") end # Increments an integer value in the cache. # # Options are passed to the underlying cache implementation. # # Some implementations may not support this method. def increment(name, amount = 1, options = nil) raise NotImplementedError.new("#{self.class.name} does not support increment") end # Decrements an integer value in the cache. # # Options are passed to the underlying cache implementation. # # Some implementations may not support this method. def decrement(name, amount = 1, options = nil) raise NotImplementedError.new("#{self.class.name} does not support decrement") end # Cleans up the cache by removing expired entries. # # Options are passed to the underlying cache implementation. # # Some implementations may not support this method. def cleanup(options = nil) raise NotImplementedError.new("#{self.class.name} does not support cleanup") end # Clears the entire cache. Be careful with this method since it could # affect other processes if shared cache is being used. # # The options hash is passed to the underlying cache implementation. # # Some implementations may not support this method. def clear(options = nil) raise NotImplementedError.new("#{self.class.name} does not support clear") end private def default_serializer case Cache.format_version when 7.0 Cache::SerializerWithFallback[:marshal_7_0] when 7.1 Cache::SerializerWithFallback[:marshal_7_1] else raise ArgumentError, "Unrecognized ActiveSupport::Cache.format_version: #{Cache.format_version.inspect}" end end # Adds the namespace defined in the options to a pattern designed to # match keys. Implementations that support delete_matched should call # this method to translate a pattern that matches names into one that # matches namespaced keys. def key_matcher(pattern, options) # :doc: prefix = options[:namespace].is_a?(Proc) ? options[:namespace].call : options[:namespace] if prefix source = pattern.source if source.start_with?("^") source = source[1, source.length] else source = ".*#{source[0, source.length]}" end Regexp.new("^#{Regexp.escape(prefix)}:#{source}", pattern.options) else pattern end end # Reads an entry from the cache implementation. Subclasses must implement # this method. def read_entry(key, **options) raise NotImplementedError.new end # Writes an entry to the cache implementation. Subclasses must implement # this method. def write_entry(key, entry, **options) raise NotImplementedError.new end def serialize_entry(entry, **options) options = merged_options(options) if @coder_supports_compression && options[:compress] @coder.dump_compressed(entry, options[:compress_threshold]) else @coder.dump(entry) end end def deserialize_entry(payload, **) payload.nil? ? nil : @coder.load(payload) rescue DeserializationError nil end # Reads multiple entries from the cache implementation. Subclasses MAY # implement this method. def read_multi_entries(names, **options) names.each_with_object({}) do |name, results| key = normalize_key(name, options) entry = read_entry(key, **options) next unless entry version = normalize_version(name, options) if entry.expired? delete_entry(key, **options) elsif !entry.mismatched?(version) results[name] = entry.value end end end # Writes multiple entries to the cache implementation. Subclasses MAY # implement this method. def write_multi_entries(hash, **options) hash.each do |key, entry| write_entry key, entry, **options end end # Deletes an entry from the cache implementation. Subclasses must # implement this method. def delete_entry(key, **options) raise NotImplementedError.new end # Deletes multiples entries in the cache implementation. Subclasses MAY # implement this method. def delete_multi_entries(entries, **options) entries.count { |key| delete_entry(key, **options) } end # Merges the default options with ones specific to a method call. def merged_options(call_options) if call_options call_options = normalize_options(call_options) if call_options.key?(:expires_in) && call_options.key?(:expires_at) raise ArgumentError, "Either :expires_in or :expires_at can be supplied, but not both" end expires_at = call_options.delete(:expires_at) call_options[:expires_in] = (expires_at - Time.now) if expires_at if call_options[:expires_in].is_a?(Time) expires_in = call_options[:expires_in] raise ArgumentError.new("expires_in parameter should not be a Time. Did you mean to use expires_at? Got: #{expires_in}") end if call_options[:expires_in]&.negative? expires_in = call_options.delete(:expires_in) handle_invalid_expires_in("Cache expiration time is invalid, cannot be negative: #{expires_in}") end if options.empty? call_options else options.merge(call_options) end else options end end def handle_invalid_expires_in(message) error = ArgumentError.new(message) if ActiveSupport::Cache::Store.raise_on_invalid_cache_expiration_time raise error else ActiveSupport.error_reporter&.report(error, handled: true, severity: :warning) logger.error("#{error.class}: #{error.message}") if logger end end # Normalize aliased options to their canonical form def normalize_options(options) options = options.dup OPTION_ALIASES.each do |canonical_name, aliases| alias_key = aliases.detect { |key| options.key?(key) } options[canonical_name] ||= options[alias_key] if alias_key options.except!(*aliases) end options end def validate_options(options) if options.key?(:coder) && options[:serializer] raise ArgumentError, "Cannot specify :serializer and :coder options together" end if options.key?(:coder) && options[:compressor] raise ArgumentError, "Cannot specify :compressor and :coder options together" end if Cache.format_version < 7.1 && !options[:serializer] && options[:compressor] raise ArgumentError, "Cannot specify :compressor option when using" \ " default serializer and cache format version is < 7.1" end options end # Expands and namespaces the cache key. # Raises an exception when the key is +nil+ or an empty string. # May be overridden by cache stores to do additional normalization. def normalize_key(key, options = nil) str_key = expanded_key(key) raise(ArgumentError, "key cannot be blank") if !str_key || str_key.empty? namespace_key str_key, options end # Prefix the key with a namespace string: # # namespace_key 'foo', namespace: 'cache' # # => 'cache:foo' # # With a namespace block: # # namespace_key 'foo', namespace: -> { 'cache' } # # => 'cache:foo' def namespace_key(key, call_options = nil) namespace = if call_options&.key?(:namespace) call_options[:namespace] else options[:namespace] end if namespace.respond_to?(:call) namespace = namespace.call end if key && key.encoding != Encoding::UTF_8 key = key.dup.force_encoding(Encoding::UTF_8) end if namespace "#{namespace}:#{key}" else key end end # Expands key to be a consistent string value. Invokes +cache_key+ if # object responds to +cache_key+. Otherwise, +to_param+ method will be # called. If the key is a Hash, then keys will be sorted alphabetically. def expanded_key(key) return key.cache_key.to_s if key.respond_to?(:cache_key) case key when Array if key.size > 1 key.collect { |element| expanded_key(element) } else expanded_key(key.first) end when Hash key.collect { |k, v| "#{k}=#{v}" }.sort! else key end.to_param end def normalize_version(key, options = nil) (options && options[:version].try(:to_param)) || expanded_version(key) end def expanded_version(key) case when key.respond_to?(:cache_version) then key.cache_version.to_param when key.is_a?(Array) then key.map { |element| expanded_version(element) }.tap(&:compact!).to_param when key.respond_to?(:to_a) then expanded_version(key.to_a) end end def instrument(operation, key, options = nil, &block) _instrument(operation, key: key, options: options, &block) end def instrument_multi(operation, keys, options = nil, &block) _instrument(operation, multi: true, key: keys, options: options, &block) end def _instrument(operation, multi: false, options: nil, **payload, &block) if logger && logger.debug? && !silence? debug_key = if multi ": #{payload[:key].size} key(s) specified" elsif payload[:key] ": #{payload[:key]}" end debug_options = " (#{options.inspect})" unless options.blank? logger.debug "Cache #{operation}#{debug_key}#{debug_options}" end payload[:store] = self.class.name payload.merge!(options) if options.is_a?(Hash) ActiveSupport::Notifications.instrument("cache_#{operation}.active_support", payload) do block&.call(payload) end end def handle_expired_entry(entry, key, options) if entry && entry.expired? race_ttl = options[:race_condition_ttl].to_i if (race_ttl > 0) && (Time.now.to_f - entry.expires_at <= race_ttl) # When an entry has a positive :race_condition_ttl defined, put the stale entry back into the cache # for a brief period while the entry is being recalculated. entry.expires_at = Time.now.to_f + race_ttl options[:expires_in] = race_ttl * 2 write_entry(key, entry, **options) else delete_entry(key, **options) end entry = nil end entry end def get_entry_value(entry, name, options) instrument(:fetch_hit, name, options) entry.value end def save_block_result_to_cache(name, key, options) options = options.dup result = instrument(:generate, key, options) do yield(name, WriteOptions.new(options)) end write(name, result, options) unless result.nil? && options[:skip_nil] result end end # Enables the dynamic configuration of Cache entry options while ensuring # that conflicting options are not both set. When a block is given to # ActiveSupport::Cache::Store#fetch, the second argument will be an # instance of +WriteOptions+. class WriteOptions def initialize(options) # :nodoc: @options = options end def version @options[:version] end def version=(version) @options[:version] = version end def expires_in @options[:expires_in] end # Sets the Cache entry's +expires_in+ value. If an +expires_at+ option was # previously set, this will unset it since +expires_in+ and +expires_at+ # cannot both be set. def expires_in=(expires_in) @options.delete(:expires_at) @options[:expires_in] = expires_in end def expires_at @options[:expires_at] end # Sets the Cache entry's +expires_at+ value. If an +expires_in+ option was # previously set, this will unset it since +expires_at+ and +expires_in+ # cannot both be set. def expires_at=(expires_at) @options.delete(:expires_in) @options[:expires_at] = expires_at end end end end