lib/fts_lite/tokenizer.rb in fts_lite-0.0.2 vs lib/fts_lite/tokenizer.rb in fts_lite-0.1.0

- old
+ new

@@ -68,12 +68,12 @@ class WakachiBigram def vector(text) split(text).join(" ") end def split(text) - words = BimyouSegmenter.segment(Tokenizer.normalize(text), - :white_space => false, - :symbol => false).map {|word| + BimyouSegmenter.segment(Tokenizer.normalize(text), + :white_space => false, + :symbol => false).map {|word| if (word.size == 1) word else 0.upto(word.size - 2).map {|i| word[i, 2] } end