lib/fts_lite/tokenizer.rb in fts_lite-0.0.2 vs lib/fts_lite/tokenizer.rb in fts_lite-0.1.0
- old
+ new
@@ -68,12 +68,12 @@
class WakachiBigram
def vector(text)
split(text).join(" ")
end
def split(text)
- words = BimyouSegmenter.segment(Tokenizer.normalize(text),
- :white_space => false,
- :symbol => false).map {|word|
+ BimyouSegmenter.segment(Tokenizer.normalize(text),
+ :white_space => false,
+ :symbol => false).map {|word|
if (word.size == 1)
word
else
0.upto(word.size - 2).map {|i| word[i, 2] }
end