require 'dentaku/tokenizer' describe Dentaku::Tokenizer do let(:tokenizer) { described_class.new } it 'handles an empty expression' do expect(tokenizer.tokenize('')).to be_empty end it 'tokenizes addition' do tokens = tokenizer.tokenize('1+1') expect(tokens.map(&:category)).to eq([:numeric, :operator, :numeric]) expect(tokens.map(&:value)).to eq([1, :add, 1]) end it 'tokenizes comparison with =' do tokens = tokenizer.tokenize('number = 5') expect(tokens.map(&:category)).to eq([:identifier, :comparator, :numeric]) expect(tokens.map(&:value)).to eq(['number', :eq, 5]) end it 'tokenizes comparison with =' do tokens = tokenizer.tokenize('number = 5') expect(tokens.map(&:category)).to eq([:identifier, :comparator, :numeric]) expect(tokens.map(&:value)).to eq(['number', :eq, 5]) end it 'tokenizes comparison with alternate ==' do tokens = tokenizer.tokenize('number == 5') expect(tokens.map(&:category)).to eq([:identifier, :comparator, :numeric]) expect(tokens.map(&:value)).to eq(['number', :eq, 5]) end it 'ignores whitespace' do tokens = tokenizer.tokenize('1 / 1 ') expect(tokens.map(&:category)).to eq([:numeric, :operator, :numeric]) expect(tokens.map(&:value)).to eq([1, :divide, 1]) end it 'tokenizes power operations' do tokens = tokenizer.tokenize('10 ^ 2') expect(tokens.map(&:category)).to eq([:numeric, :operator, :numeric]) expect(tokens.map(&:value)).to eq([10, :pow, 2]) end it 'tokenizes power operations' do tokens = tokenizer.tokenize('0 * 10 ^ -5') expect(tokens.map(&:category)).to eq([:numeric, :operator, :numeric, :operator, :operator, :numeric]) expect(tokens.map(&:value)).to eq([0, :multiply, 10, :pow, :subtract, 5]) end it 'handles floating point' do tokens = tokenizer.tokenize('1.5 * 3.7') expect(tokens.map(&:category)).to eq([:numeric, :operator, :numeric]) expect(tokens.map(&:value)).to eq([1.5, :multiply, 3.7]) end it 'does not require leading zero' do tokens = tokenizer.tokenize('.5 * 3.7') expect(tokens.map(&:category)).to eq([:numeric, :operator, :numeric]) expect(tokens.map(&:value)).to eq([0.5, :multiply, 3.7]) end it 'accepts arbitrary identifiers' do tokens = tokenizer.tokenize('monkeys > 1500') expect(tokens.map(&:category)).to eq([:identifier, :comparator, :numeric]) expect(tokens.map(&:value)).to eq(['monkeys', :gt, 1500]) end it 'recognizes double-quoted strings' do tokens = tokenizer.tokenize('animal = "giraffe"') expect(tokens.map(&:category)).to eq([:identifier, :comparator, :string]) expect(tokens.map(&:value)).to eq(['animal', :eq, 'giraffe']) end it 'recognizes single-quoted strings' do tokens = tokenizer.tokenize("animal = 'giraffe'") expect(tokens.map(&:category)).to eq([:identifier, :comparator, :string]) expect(tokens.map(&:value)).to eq(['animal', :eq, 'giraffe']) end it 'recognizes binary minus operator' do tokens = tokenizer.tokenize('2 - 3') expect(tokens.map(&:category)).to eq([:numeric, :operator, :numeric]) expect(tokens.map(&:value)).to eq([2, :subtract, 3]) end it 'recognizes unary minus operator' do tokens = tokenizer.tokenize('-2 + 3') expect(tokens.map(&:category)).to eq([:operator, :numeric, :operator, :numeric]) expect(tokens.map(&:value)).to eq([:subtract, 2, :add, 3]) end it 'recognizes unary minus operator' do tokens = tokenizer.tokenize('2 - -3') expect(tokens.map(&:category)).to eq([:numeric, :operator, :operator, :numeric]) expect(tokens.map(&:value)).to eq([2, :subtract, :subtract, 3]) end it 'matches "<=" before "<"' do tokens = tokenizer.tokenize('perimeter <= 7500') expect(tokens.map(&:category)).to eq([:identifier, :comparator, :numeric]) expect(tokens.map(&:value)).to eq(['perimeter', :le, 7500]) end it 'matches "and" for logical expressions' do tokens = tokenizer.tokenize('octopi <= 7500 AND sharks > 1500') expect(tokens.map(&:category)).to eq([:identifier, :comparator, :numeric, :combinator, :identifier, :comparator, :numeric]) expect(tokens.map(&:value)).to eq(['octopi', :le, 7500, :and, 'sharks', :gt, 1500]) end it 'matches "or" for logical expressions' do tokens = tokenizer.tokenize('size < 3 or admin = 1') expect(tokens.map(&:category)).to eq([:identifier, :comparator, :numeric, :combinator, :identifier, :comparator, :numeric]) expect(tokens.map(&:value)).to eq(['size', :lt, 3, :or, 'admin', :eq, 1]) end it 'detects unbalanced parentheses' do expect { tokenizer.tokenize('(5+3') }.to raise_error expect { tokenizer.tokenize(')') }.to raise_error end it 'recognizes identifiers that share initial substrings with combinators' do tokens = tokenizer.tokenize('andover < 10') expect(tokens.length).to eq(3) expect(tokens.map(&:category)).to eq([:identifier, :comparator, :numeric]) expect(tokens.map(&:value)).to eq(['andover', :lt, 10]) end describe 'functions' do it 'include IF' do tokens = tokenizer.tokenize('if(x < 10, y, z)') expect(tokens.length).to eq(10) expect(tokens.map(&:category)).to eq([:function, :grouping, :identifier, :comparator, :numeric, :grouping, :identifier, :grouping, :identifier, :grouping]) expect(tokens.map(&:value)).to eq([:if, :fopen, 'x', :lt, 10, :comma, 'y', :comma, 'z', :close]) end it 'include ROUND/UP/DOWN' do tokens = tokenizer.tokenize('round(8.2)') expect(tokens.length).to eq(4) expect(tokens.map(&:category)).to eq([:function, :grouping, :numeric, :grouping]) expect(tokens.map(&:value)).to eq([:round, :fopen, BigDecimal.new('8.2'), :close]) tokens = tokenizer.tokenize('round(8.75, 1)') expect(tokens.length).to eq(6) expect(tokens.map(&:category)).to eq([:function, :grouping, :numeric, :grouping, :numeric, :grouping]) expect(tokens.map(&:value)).to eq([:round, :fopen, BigDecimal.new('8.75'), :comma, 1, :close]) tokens = tokenizer.tokenize('ROUNDUP(8.2)') expect(tokens.length).to eq(4) expect(tokens.map(&:category)).to eq([:function, :grouping, :numeric, :grouping]) expect(tokens.map(&:value)).to eq([:roundup, :fopen, BigDecimal.new('8.2'), :close]) tokens = tokenizer.tokenize('RoundDown(8.2)') expect(tokens.length).to eq(4) expect(tokens.map(&:category)).to eq([:function, :grouping, :numeric, :grouping]) expect(tokens.map(&:value)).to eq([:rounddown, :fopen, BigDecimal.new('8.2'), :close]) end it 'include NOT' do tokens = tokenizer.tokenize('not(8 < 5)') expect(tokens.length).to eq(6) expect(tokens.map(&:category)).to eq([:function, :grouping, :numeric, :comparator, :numeric, :grouping]) expect(tokens.map(&:value)).to eq([:not, :fopen, 8, :lt, 5, :close]) end it 'handles whitespace after function name' do tokens = tokenizer.tokenize('not (8 < 5)') expect(tokens.length).to eq(6) expect(tokens.map(&:category)).to eq([:function, :grouping, :numeric, :comparator, :numeric, :grouping]) expect(tokens.map(&:value)).to eq([:not, :fopen, 8, :lt, 5, :close]) end end end