# frozen_string_literal: true require 'deimos/avro_data_encoder' require 'deimos/message' require 'deimos/shared_config' require 'deimos/schema_coercer' require 'phobos/producer' require 'active_support/notifications' # :nodoc: module Deimos class << self # Run a block without allowing any messages to be produced to Kafka. # Optionally add a list of producer classes to limit the disabling to those # classes. # @param producer_classes [Array|Class] def disable_producers(*producer_classes, &block) if producer_classes.any? _disable_producer_classes(producer_classes, &block) return end if Thread.current[:frk_disable_all_producers] # nested disable block yield return end begin Thread.current[:frk_disable_all_producers] = true yield ensure Thread.current[:frk_disable_all_producers] = false end end # :nodoc: def _disable_producer_classes(producer_classes) Thread.current[:frk_disabled_producers] ||= Set.new producers_to_disable = producer_classes - Thread.current[:frk_disabled_producers].to_a Thread.current[:frk_disabled_producers] += producers_to_disable yield Thread.current[:frk_disabled_producers] -= producers_to_disable end # Are producers disabled? If a class is passed in, check only that class. # Otherwise check if the global disable flag is set. # @return [Boolean] def producers_disabled?(producer_class=nil) Thread.current[:frk_disable_all_producers] || Thread.current[:frk_disabled_producers]&.include?(producer_class) end end # Producer to publish messages to a given kafka topic. class Producer include SharedConfig MAX_BATCH_SIZE = 500 class << self # @return [Hash] def config @config ||= { encode_key: true, namespace: Deimos.config.producer_schema_namespace } end # Set the topic. # @param topic [String] # @return [String] the current topic if no argument given. def topic(topic=nil) if topic config[:topic] = topic return end # accessor "#{Deimos.config.producer_topic_prefix}#{config[:topic]}" end # Override the default partition key (which is the payload key). # @param _payload [Hash] the payload being passed into the produce method. # Will include `payload_key` if it is part of the original payload. # @return [String] def partition_key(_payload) nil end # Publish the payload to the topic. # @param payload [Hash] with an optional payload_key hash key. def publish(payload) publish_list([payload]) end # Publish a list of messages. # @param payloads [Hash|Array] with optional payload_key hash key. # @param sync [Boolean] if given, override the default setting of # whether to publish synchronously. # @param force_send [Boolean] if true, ignore the configured backend # and send immediately to Kafka. def publish_list(payloads, sync: nil, force_send: false) return if Deimos.config.seed_broker.blank? || Deimos.config.disable_producers || Deimos.producers_disabled?(self) backend_class = determine_backend_class(sync, force_send) Deimos.instrument( 'encode_messages', producer: self, topic: topic, payloads: payloads ) do messages = Array(payloads).map { |p| Deimos::Message.new(p, self) } messages.each(&method(:_process_message)) messages.in_groups_of(MAX_BATCH_SIZE, false) do |batch| self.produce_batch(backend_class, batch) end end end # @param sync [Boolean] # @param force_send [Boolean] # @return [Class < Deimos::Backend] def determine_backend_class(sync, force_send) backend = if force_send :kafka else Deimos.config.publish_backend end if backend == :kafka_async && sync backend = :kafka elsif backend == :kafka && sync == false backend = :kafka_async end "Deimos::Backends::#{backend.to_s.classify}".constantize end # Send a batch to the backend. # @param backend [Class < Deimos::Backend] # @param batch [Array] def produce_batch(backend, batch) backend.publish(producer_class: self, messages: batch) end # @return [AvroDataEncoder] def encoder @encoder ||= AvroDataEncoder.new(schema: config[:schema], namespace: config[:namespace]) end # @return [AvroDataEncoder] def key_encoder @key_encoder ||= AvroDataEncoder.new(schema: config[:key_schema], namespace: config[:namespace]) end # Override this in active record producers to add # non-schema fields to check for updates # @return [Array] fields to check for updates def watched_attributes self.encoder.avro_schema.fields.map(&:name) end private # @param message [Message] def _process_message(message) # this violates the Law of Demeter but it has to happen in a very # specific order and requires a bunch of methods on the producer # to work correctly. message.add_fields(encoder.avro_schema) message.partition_key = self.partition_key(message.payload) message.key = _retrieve_key(message.payload) # need to do this before _coerce_fields because that might result # in an empty payload which is an *error* whereas this is intended. message.payload = nil if message.payload.blank? message.coerce_fields(encoder.avro_schema) message.encoded_key = _encode_key(message.key) message.topic = self.topic message.encoded_payload = if message.payload.nil? nil else encoder.encode(message.payload, topic: "#{config[:topic]}-value") end end # @param key [Object] # @return [String|Object] def _encode_key(key) if key.nil? return nil if config[:no_keys] # no key is fine, otherwise it's a problem raise 'No key given but a key is required! Use `key_config none: true` to avoid using keys.' end if config[:encode_key] && config[:key_field].nil? && config[:key_schema].nil? raise 'No key config given - if you are not encoding keys, please use `key_config plain: true`' end if config[:key_field] encoder.encode_key(config[:key_field], key, "#{config[:topic]}-key") elsif config[:key_schema] key_encoder.encode(key, topic: "#{config[:topic]}-key") else key end end # @param payload [Hash] # @return [String] def _retrieve_key(payload) key = payload.delete(:payload_key) return key if key config[:key_field] ? payload[config[:key_field]] : nil end end end end