Sha256: 236484526f73cb3d4b1c9cf147f622e732e80bed5701e62b0a0db108a828de6e
Contents?: true
Size: 375 Bytes
Versions: 10
Compression:
Stored size: 375 Bytes
Contents
require 'spec_helper' describe PragmaticTokenizer do context 'Language: French (fr)' do it 'tokenizes a string #001' do text = "L'art de l'univers, c'est un art" pt = PragmaticTokenizer::Tokenizer.new( language: 'fr' ) expect(pt.tokenize(text)).to eq(["l'", "art", "de", "l'", "univers", ",", "c'est", "un", "art"]) end end end
Version data entries
10 entries across 10 versions & 1 rubygems