README.rdoc in fluent-plugin-s3-0.2.0 vs README.rdoc in fluent-plugin-s3-0.2.1

- old
+ new

@@ -1,2 +1,55 @@ = Amazon S3 output plugin for Fluent event collector +== Overview + +*s3* output plugin buffers event logs in local file and upload it to S3 periodically. + +This plugin splits files exactly by using the time of event logs (not the time when the logs are received). For example, a log '2011-01-02 message B' is reached, and then another log '2011-01-03 message B' is reached in this order, the former one is stored in "20110102.gz" file, and latter one in "20110103.gz" file. + + +== Installation + +Simply use RubyGems: + + gem install fluent-plugin-s3 + +== Configuration + + <match pattern> + type s3 + + aws_key_id YOUR_AWS_KEY_ID + aws_sec_key YOUR_AWS_SECRET/KEY + s3_bucket YOUR_S3_BUCKET_NAME + path logs/ + buffer_path /var/log/fluent/s3 + + time_slice_format %Y%m%d-%H + time_slice_wait 10m + utc + </match> + +[aws_key_id (required)] AWS access key id. + +[aws_sec_key (required)] AWS secret key. + +[s3_bucket (required)] S3 bucket name. + +[path] path prefix of the files on S3. Default is "" (no prefix). + +[buffer_path (required)] path prefix of the files to buffer logs. + +[time_slice_format] Format of the time used as the file name. Default is '%Y%m%d'. Use '%Y%m%d%H' to split files hourly. + +[time_slice_wait] The time to wait old logs. Default is 10 minutes. Specify larger value if old logs may reache. + +[utc] Use UTC instead of local time. + +The actual path on S3 will be: "{path}{time_slice_format}_{sequential_number}.gz" + + +== Copyright + +Copyright:: Copyright (c) 2011 Sadayuki Furuhashi +License:: Apache License, Version 2.0 +