# frozen_string_literal: true module Karafka module Connection # Class used as a wrapper around Ruby-Kafka client to simplify additional # features that we provide/might provide in future and to hide the internal implementation class Client extend Forwardable %i[ seek trigger_heartbeat trigger_heartbeat! ].each do |delegated_method| def_delegator :kafka_consumer, delegated_method end # Creates a queue consumer client that will pull the data from Kafka # @param consumer_group [Karafka::Routing::ConsumerGroup] consumer group for which # we create a client # @return [Karafka::Connection::Client] group consumer that can subscribe to # multiple topics def initialize(consumer_group) @consumer_group = consumer_group Persistence::Client.write(self) end # Opens connection, gets messages and calls a block for each of the incoming messages # @yieldparam [Array, Symbol] kafka response with an info about # the type of the fetcher that is being used # @note This will yield with raw messages - no preprocessing or reformatting. def fetch_loop settings = ApiAdapter.consumption(consumer_group) if consumer_group.batch_fetching kafka_consumer.each_batch(**settings) { |batch| yield(batch, :batch) } else kafka_consumer.each_message(**settings) { |message| yield(message, :message) } end # @note We catch only the processing errors as any other are considered critical (exceptions) # and should require a client restart with a backoff rescue Kafka::ProcessingError => e # If there was an error during consumption, we have to log it, pause current partition # and process other things Karafka.monitor.instrument( 'connection.client.fetch_loop.error', caller: self, error: e.cause ) pause(e.topic, e.partition) retry end # Gracefully stops topic consumption # @note Stopping running consumers without a really important reason is not recommended # as until all the consumers are stopped, the server will keep running serving only # part of the messages def stop @kafka_consumer&.stop @kafka_consumer = nil end # Pauses fetching and consumption of a given topic partition # @param topic [String] topic that we want to pause # @param partition [Integer] number partition that we want to pause def pause(topic, partition) args, kwargs = ApiAdapter.pause(topic, partition, consumer_group).values_at(:args, :kwargs) kafka_consumer.pause(*args, **kwargs) end # Marks given message as consumed # @param [Karafka::Params::Params] params message that we want to mark as processed # @note This method won't trigger automatic offsets commits, rather relying on the ruby-kafka # offsets time-interval based committing def mark_as_consumed(params) kafka_consumer.mark_message_as_processed( *ApiAdapter.mark_message_as_processed(params) ) end # Marks a given message as consumed and commit the offsets in a blocking way # @param [Karafka::Params::Params] params message that we want to mark as processed # @note This method commits the offset for each manual marking to be sure # that offset commit happen asap in case of a crash def mark_as_consumed!(params) mark_as_consumed(params) # Trigger an immediate, blocking offset commit in order to minimize the risk of crashing # before the automatic triggers have kicked in. kafka_consumer.commit_offsets end private attr_reader :consumer_group # @return [Kafka::Consumer] returns a ready to consume Kafka consumer # that is set up to consume from topics of a given consumer group def kafka_consumer # @note We don't cache the connection internally because we cache kafka_consumer that uses # kafka client object instance @kafka_consumer ||= Builder.call(consumer_group).consumer( **ApiAdapter.consumer(consumer_group) ).tap do |consumer| consumer_group.topics.each do |topic| settings = ApiAdapter.subscribe(topic) consumer.subscribe(settings[0], **settings[1]) end end rescue Kafka::ConnectionError # If we would not wait it will spam log file with failed # attempts if Kafka is down sleep(consumer_group.reconnect_timeout) # We don't log and just re-raise - this will be logged # down the road raise end end end end