wukong.gemspec in wukong-2.0.0 vs wukong.gemspec in wukong-2.0.1
- old
+ new
@@ -3,15 +3,15 @@
# Instead, edit Jeweler::Tasks in Rakefile, and run 'rake gemspec'
# -*- encoding: utf-8 -*-
Gem::Specification.new do |s|
s.name = %q{wukong}
- s.version = "2.0.0"
+ s.version = "2.0.1"
s.required_rubygems_version = Gem::Requirement.new(">= 0") if s.respond_to? :required_rubygems_version=
s.authors = ["Philip (flip) Kromer"]
- s.date = %q{2011-01-29}
+ s.date = %q{2011-07-01}
s.description = %q{ Treat your dataset like a:
* stream of lines when it's efficient to process by lines
* stream of field arrays when it's efficient to deal directly with fields
* stream of lightweight objects when it's efficient to deal with objects
@@ -137,10 +137,11 @@
"examples/contrib/jeans/data/sizes",
"examples/contrib/jeans/normalize.rb",
"examples/contrib/jeans/sizes.rb",
"examples/corpus/bucket_counter.rb",
"examples/corpus/dbpedia_abstract_to_sentences.rb",
+ "examples/corpus/sentence_bigrams.rb",
"examples/corpus/sentence_coocurrence.rb",
"examples/corpus/words_to_bigrams.rb",
"examples/emr/README.textile",
"examples/emr/dot_wukong_dir/credentials.json",
"examples/emr/dot_wukong_dir/emr.yaml",
@@ -160,11 +161,14 @@
"examples/pagerank/run_pagerank.sh",
"examples/sample_records.rb",
"examples/server_logs/apache_log_parser.rb",
"examples/server_logs/breadcrumbs.rb",
"examples/server_logs/logline.rb",
+ "examples/server_logs/nook.rb",
+ "examples/server_logs/nook/faraday_dummy_adapter.rb",
"examples/server_logs/user_agent.rb",
+ "examples/simple_word_count.rb",
"examples/size.rb",
"examples/stats/avg_value_frequency.rb",
"examples/stats/binning_percentile_estimator.rb",
"examples/stats/data/avg_value_frequency.tsv",
"examples/stats/rank_and_bin.rb",
@@ -250,17 +254,18 @@
"spec/wukong/script_spec.rb",
"wukong.gemspec"
]
s.homepage = %q{http://mrflip.github.com/wukong}
s.require_paths = ["lib"]
- s.rubygems_version = %q{1.4.2}
+ s.rubygems_version = %q{1.5.0}
s.summary = %q{Hadoop Streaming for Ruby. Wukong makes Hadoop so easy a chimpanzee can use it, yet handles terabyte-scale computation with ease.}
s.test_files = [
"examples/contrib/jeans/normalize.rb",
"examples/contrib/jeans/sizes.rb",
"examples/corpus/bucket_counter.rb",
"examples/corpus/dbpedia_abstract_to_sentences.rb",
+ "examples/corpus/sentence_bigrams.rb",
"examples/corpus/sentence_coocurrence.rb",
"examples/corpus/words_to_bigrams.rb",
"examples/emr/elastic_mapreduce_example.rb",
"examples/ignore_me/counting.rb",
"examples/ignore_me/grouper.rb",
@@ -273,10 +278,13 @@
"examples/pagerank/pagerank_initialize.rb",
"examples/sample_records.rb",
"examples/server_logs/apache_log_parser.rb",
"examples/server_logs/breadcrumbs.rb",
"examples/server_logs/logline.rb",
+ "examples/server_logs/nook.rb",
+ "examples/server_logs/nook/faraday_dummy_adapter.rb",
"examples/server_logs/user_agent.rb",
+ "examples/simple_word_count.rb",
"examples/size.rb",
"examples/stats/avg_value_frequency.rb",
"examples/stats/binning_percentile_estimator.rb",
"examples/stats/rank_and_bin.rb",
"examples/stupidly_simple_filter.rb",