lib/dentaku/tokenizer.rb in dentaku-1.2.1 vs lib/dentaku/tokenizer.rb in dentaku-1.2.2
- old
+ new
@@ -8,10 +8,10 @@
RPAREN = TokenMatcher.new(:grouping, :close)
def tokenize(string)
@nesting = 0
@tokens = []
- input = string.dup
+ input = string.to_s.dup
until input.empty?
raise "parse error at: '#{ input }'" unless TokenScanner.scanners.any? do |scanner|
scanned, input = scan(input, scanner)
scanned