# -*- coding: utf-8 -*- require 'tempfile' require 'securerandom' require 'fluent/mixin/plaintextformatter' class Fluent::WebHDFSOutput < Fluent::TimeSlicedOutput Fluent::Plugin.register_output('webhdfs', self) config_set_default :buffer_type, 'memory' config_set_default :time_slice_format, '%Y%m%d' # For fluentd v0.12.16 or earlier class << self unless method_defined?(:desc) def desc(description) end end end desc 'WebHDFS/HttpFs host' config_param :host, :string, :default => nil desc 'WebHDFS/HttpFs port' config_param :port, :integer, :default => 50070 desc 'Namenode (host:port)' config_param :namenode, :string, :default => nil # host:port desc 'Standby namenode for Namenode HA (host:port)' config_param :standby_namenode, :string, :default => nil # host:port desc 'Ignore errors on start up' config_param :ignore_start_check_error, :bool, :default => false desc 'Output file path on HDFS' config_param :path, :string desc 'User name for pseudo authentication' config_param :username, :string, :default => nil desc 'Store data over HttpFs instead of WebHDFS' config_param :httpfs, :bool, :default => false desc 'Number of seconds to wait for the connection to open' config_param :open_timeout, :integer, :default => 30 # from ruby net/http default desc 'Number of seconds to wait for one block to be read' config_param :read_timeout, :integer, :default => 60 # from ruby net/http default desc 'Retry automatically when known errors of HDFS are occurred' config_param :retry_known_errors, :bool, :default => false desc 'Retry interval' config_param :retry_interval, :integer, :default => nil desc 'The number of retries' config_param :retry_times, :integer, :default => nil # how many times of write failure before switch to standby namenode # by default it's 11 times that costs 1023 seconds inside fluentd, # which is considered enough to exclude the scenes that caused by temporary network fail or single datanode fail desc 'How many times of write failure before switch to standby namenode' config_param :failures_before_use_standby, :integer, :default => 11 include Fluent::Mixin::PlainTextFormatter config_param :default_tag, :string, :default => 'tag_missing' desc 'Append data or not' config_param :append, :bool, :default => true desc 'Use SSL or not' config_param :ssl, :bool, :default => false desc 'OpenSSL certificate authority file' config_param :ssl_ca_file, :string, :default => nil desc 'OpenSSL verify mode (none,peer)' config_param :ssl_verify_mode, :default => nil do |val| case val when 'none' :none when 'peer' :peer else raise Fluent::ConfigError, "unexpected parameter on ssl_verify_mode: #{val}" end end desc 'Use kerberos authentication or not' config_param :kerberos, :bool, :default => false desc 'kerberos keytab file' config_param :kerberos_keytab, :string, :default => nil SUPPORTED_COMPRESS = ['gzip', 'bzip2', 'snappy', 'lzo_command', 'text'] desc "Compress method (#{SUPPORTED_COMPRESS.join(',')})" config_param :compress, :default => nil do |val| unless SUPPORTED_COMPRESS.include? val raise Fluent::ConfigError, "unsupported compress: #{val}" end val end CHUNK_ID_PLACE_HOLDER = '${chunk_id}' attr_reader :compressor def initialize super require 'net/http' require 'time' require 'webhdfs' @compressor = nil end # Define `log` method for v0.10.42 or earlier unless method_defined?(:log) define_method("log") { $log } end def configure(conf) if conf['path'] if conf['path'].index('%S') conf['time_slice_format'] = '%Y%m%d%H%M%S' elsif conf['path'].index('%M') conf['time_slice_format'] = '%Y%m%d%H%M' elsif conf['path'].index('%H') conf['time_slice_format'] = '%Y%m%d%H' end end verify_config_placeholders_in_path!(conf) super @replace_random_uuid = @path.include?('%{uuid}') || @path.include?('%{uuid_flush}') if @replace_random_uuid # to check SecureRandom.uuid is available or not (NotImplementedError raised in such environment) begin SecureRandom.uuid rescue raise Fluent::ConfigError, "uuid feature (SecureRandom) is unavailable in this environment" end end begin @compressor = COMPRESSOR_REGISTRY.lookup(@compress || 'text').new rescue Fluent::ConfigError raise rescue $log.warn "#{@comress} not found. Use 'text' instead" @compressor = COMPRESSOR_REGISTRY.lookup('text').new end if @host @namenode_host = @host @namenode_port = @port elsif @namenode unless /\A([a-zA-Z0-9][-a-zA-Z0-9.]*):(\d+)\Z/ =~ @namenode raise Fluent::ConfigError, "Invalid config value about namenode: '#{@namenode}', needs NAMENODE_HOST:PORT" end @namenode_host = $1 @namenode_port = $2.to_i else raise Fluent::ConfigError, "WebHDFS host or namenode missing" end if @standby_namenode unless /\A([a-zA-Z0-9][-a-zA-Z0-9.]*):(\d+)\Z/ =~ @standby_namenode raise Fluent::ConfigError, "Invalid config value about standby namenode: '#{@standby_namenode}', needs STANDBY_NAMENODE_HOST:PORT" end if @httpfs raise Fluent::ConfigError, "Invalid configuration: specified to use both of standby_namenode and httpfs." end @standby_namenode_host = $1 @standby_namenode_port = $2.to_i end unless @path.index('/') == 0 raise Fluent::ConfigError, "Path on hdfs MUST starts with '/', but '#{@path}'" end @client = prepare_client(@namenode_host, @namenode_port, @username) if @standby_namenode_host @client_standby = prepare_client(@standby_namenode_host, @standby_namenode_port, @username) else @client_standby = nil end if not @append if @path.index(CHUNK_ID_PLACE_HOLDER).nil? raise Fluent::ConfigError, "path must contain ${chunk_id}, which is the placeholder for chunk_id, when append is set to false." end end end def prepare_client(host, port, username) client = WebHDFS::Client.new(host, port, username) if @httpfs client.httpfs_mode = true end client.open_timeout = @open_timeout client.read_timeout = @read_timeout if @retry_known_errors client.retry_known_errors = true client.retry_interval = @retry_interval if @retry_interval client.retry_times = @retry_times if @retry_times end if @ssl client.ssl = true client.ssl_ca_file = @ssl_ca_file if @ssl_ca_file client.ssl_verify_mode = @ssl_verify_mode if @ssl_verify_mode end if @kerberos client.kerberos = true client.kerberos_keytab = @kerberos_keytab if @kerberos_keytab end client end def namenode_available(client) if client available = true begin client.list('/') rescue => e log.warn "webhdfs check request failed. (namenode: #{client.host}:#{client.port}, error: #{e.message})" available = false end available else false end end def start super if namenode_available(@client) log.info "webhdfs connection confirmed: #{@namenode_host}:#{@namenode_port}" return end if @client_standby && namenode_available(@client_standby) log.info "webhdfs connection confirmed: #{@standby_namenode_host}:#{@standby_namenode_port}" return end unless @ignore_start_check_error raise RuntimeError, "webhdfs is not available now." end end def shutdown super end def path_format(chunk_key) Time.strptime(chunk_key, @time_slice_format).strftime(@path) end def is_standby_exception(e) e.is_a?(WebHDFS::IOError) && e.message.match(/org\.apache\.hadoop\.ipc\.StandbyException/) end def namenode_failover if @standby_namenode @client, @client_standby = @client_standby, @client log.warn "Namenode failovered, now using #{@client.host}:#{@client.port}." end end def chunk_unique_id_to_str(unique_id) unique_id.unpack('C*').map{|x| x.to_s(16).rjust(2,'0')}.join('') end # TODO check conflictions def send_data(path, data) if @append begin @client.append(path, data) rescue WebHDFS::FileNotFoundError @client.create(path, data) end else @client.create(path, data, {'overwrite' => 'true'}) end end HOSTNAME_PLACEHOLDERS_DEPRECATED = ['${hostname}', '%{hostname}', '__HOSTNAME__'] UUID_RANDOM_PLACEHOLDERS_DEPRECATED = ['${uuid}', '${uuid:random}', '__UUID__', '__UUID_RANDOM__'] UUID_OTHER_PLACEHOLDERS_OBSOLETED = ['${uuid:hostname}', '%{uuid:hostname}', '__UUID_HOSTNAME__', '${uuid:timestamp}', '%{uuid:timestamp}', '__UUID_TIMESTAMP__'] def verify_config_placeholders_in_path!(conf) return unless conf.has_key?('path') path = conf['path'] # check @path for ${hostname}, %{hostname} and __HOSTNAME__ to warn to use #{Socket.gethostbyname} if HOSTNAME_PLACEHOLDERS_DEPRECATED.any?{|ph| path.include?(ph) } log.warn "hostname placeholder is now deprecated. use '\#\{Socket.gethostname\}' instead." hostname = conf['hostname'] || Socket.gethostname HOSTNAME_PLACEHOLDERS_DEPRECATED.each do |ph| path.gsub!(ph, hostname) end end if UUID_RANDOM_PLACEHOLDERS_DEPRECATED.any?{|ph| path.include?(ph) } log.warn "random uuid placeholders are now deprecated. use %{uuid} (or %{uuid_flush}) instead." UUID_RANDOM_PLACEHOLDERS_DEPRECATED.each do |ph| path.gsub!(ph, '%{uuid}') end end if UUID_OTHER_PLACEHOLDERS_OBSOLETED.any?{|ph| path.include?(ph) } UUID_OTHER_PLACEHOLDERS_OBSOLETED.each do |ph| if path.include?(ph) log.error "configuration placeholder #{ph} is now unsupported by webhdfs output plugin." end end raise Fluent::ConfigError, "there are unsupported placeholders in path." end end def generate_path(chunk) hdfs_path = if @append path_format(chunk.key) else path_format(chunk.key).gsub(CHUNK_ID_PLACE_HOLDER, chunk_unique_id_to_str(chunk.unique_id)) end hdfs_path = "#{hdfs_path}#{@compressor.ext}" if @replace_random_uuid uuid_random = SecureRandom.uuid hdfs_path = hdfs_path.gsub('%{uuid}', uuid_random).gsub('%{uuid_flush}', uuid_random) end hdfs_path end def compress_context(chunk, &block) begin tmp = Tempfile.new("webhdfs-") @compressor.compress(chunk, tmp) tmp.rewind yield tmp ensure tmp.close(true) rescue nil end end def write(chunk) hdfs_path = generate_path(chunk) failovered = false begin compress_context(chunk) do |data| send_data(hdfs_path, data) end rescue => e log.warn "failed to communicate hdfs cluster, path: #{hdfs_path}" raise e if !@client_standby || failovered if is_standby_exception(e) && namenode_available(@client_standby) log.warn "Seems the connected host status is not active (maybe due to failovers). Gonna try another namenode immediately." namenode_failover failovered = true retry end if @num_errors && ((@num_errors + 1) >= @failures_before_use_standby) && namenode_available(@client_standby) log.warn "Too many failures. Try to use the standby namenode instead." namenode_failover failovered = true retry end raise e end hdfs_path end class Compressor include Fluent::Configurable def initialise(options = {}) super() end def configure(conf) super end def ext end def compress(chunk) end private def check_command(command, algo = nil) require 'open3' algo = command if algo.nil? begin Open3.capture3("#{command} -V") rescue Errno::ENOENT raise Fluent::ConfigError, "'#{command}' utility must be in PATH for #{algo} compression" end end end COMPRESSOR_REGISTRY = Fluent::Registry.new(:webhdfs_compressor_type, 'fluent/plugin/webhdfs_compressor_') def self.register_compressor(name, compressor) COMPRESSOR_REGISTRY.register(name, compressor) end end require 'fluent/plugin/webhdfs_compressor_text' require 'fluent/plugin/webhdfs_compressor_gzip' require 'fluent/plugin/webhdfs_compressor_bzip2' require 'fluent/plugin/webhdfs_compressor_snappy' require 'fluent/plugin/webhdfs_compressor_lzo_command'