|
|
|
package main
|
|
|
|
|
|
|
|
import (
|
|
|
|
"testing"
|
|
|
|
)
|
|
|
|
|
|
|
|
var primitivesTests = []struct {
|
|
|
|
in string
|
|
|
|
out []token
|
|
|
|
}{
|
|
|
|
{`"x"`, []token{{t_string, "x"}}},
|
|
|
|
{`"yes"`, []token{{t_string, "yes"}}},
|
|
|
|
{`"this one has spaces"`, []token{{t_string, "this one has spaces"}}},
|
|
|
|
{`"this one has \"quotes\" in it"`, []token{{t_string, `this one has "quotes" in it`}}},
|
|
|
|
{"`this one is delimited by backticks`", []token{{t_string, "this one is delimited by backticks"}}},
|
|
|
|
{` "this one has white space on either end" `, []token{{t_string, "this one has white space on either end"}}},
|
|
|
|
{`name`, []token{{t_name, "name"}}},
|
|
|
|
{`name_with_underscore`, []token{{t_name, "name_with_underscore"}}},
|
|
|
|
{` name_surrounded_by_whitespace `, []token{{t_name, "name_surrounded_by_whitespace"}}},
|
|
|
|
{`name1`, []token{{t_name, "name1"}}},
|
|
|
|
{`camelName`, []token{{t_name, "camelName"}}},
|
|
|
|
{`Type`, []token{{t_type, "Type"}}},
|
|
|
|
{`CamelType`, []token{{t_type, "CamelType"}}},
|
|
|
|
{`Type_1_2`, []token{{t_type, "Type_1_2"}}},
|
|
|
|
{`:`, []token{{t_object_separator, ":"}}},
|
|
|
|
{` : `, []token{{t_object_separator, ":"}}},
|
|
|
|
{`"x" "y"`, []token{{t_string, "x"}, {t_string, "y"}}},
|
|
|
|
{`x: "sam"`, []token{
|
|
|
|
{t_name, "x"},
|
|
|
|
{t_object_separator, ":"},
|
|
|
|
{t_string, "sam"},
|
|
|
|
}},
|
|
|
|
{`# this is a comment`, []token{{t_comment, " this is a comment"}}},
|
|
|
|
{`
|
|
|
|
# comment line one
|
|
|
|
# comment line two
|
|
|
|
`, []token{{t_comment, " comment line one"}, {t_comment, " comment line two"}}},
|
|
|
|
{`[]`, []token{{t_list_start, "["}, {t_list_end, "]"}}},
|
|
|
|
{`["item"]`, []token{{t_list_start, "["}, {t_string, "item"}, {t_list_end, "]"}}},
|
|
|
|
{`{}`, []token{{t_object_start, "{"}, {t_object_end, "}"}}},
|
|
|
|
{`{first_name: "jordan", last_name: "orelli"}`, []token{
|
|
|
|
{t_object_start, "{"},
|
|
|
|
{t_name, "first_name"},
|
|
|
|
{t_object_separator, ":"},
|
|
|
|
{t_string, "jordan"},
|
|
|
|
{t_list_separator, ","},
|
|
|
|
{t_name, "last_name"},
|
|
|
|
{t_object_separator, ":"},
|
|
|
|
{t_string, "orelli"},
|
|
|
|
{t_object_end, "}"},
|
|
|
|
}},
|
|
|
|
{`{
|
|
|
|
first_name: "jordan",
|
|
|
|
last_name: "orelli",
|
|
|
|
}`, []token{
|
|
|
|
{t_object_start, "{"},
|
|
|
|
{t_name, "first_name"},
|
|
|
|
{t_object_separator, ":"},
|
|
|
|
{t_string, "jordan"},
|
|
|
|
{t_list_separator, ","},
|
|
|
|
{t_name, "last_name"},
|
|
|
|
{t_object_separator, ":"},
|
|
|
|
{t_string, "orelli"},
|
|
|
|
{t_list_separator, ","},
|
|
|
|
{t_object_end, "}"},
|
|
|
|
}},
|
|
|
|
{`0`, []token{{t_real_number, "0"}}},
|
|
|
|
{`-0`, []token{{t_real_number, "-0"}}},
|
|
|
|
{`+0`, []token{{t_real_number, "+0"}}},
|
|
|
|
{`+125`, []token{{t_real_number, "+125"}}},
|
|
|
|
{`-125`, []token{{t_real_number, "-125"}}},
|
|
|
|
{`.0`, []token{{t_real_number, ".0"}}},
|
|
|
|
{`15`, []token{{t_real_number, "15"}}},
|
|
|
|
{`0x0`, []token{{t_real_number, "0x0"}}},
|
|
|
|
{`0xa`, []token{{t_real_number, "0xa"}}},
|
|
|
|
{`0xc0dea5cf`, []token{{t_real_number, "0xc0dea5cf"}}},
|
|
|
|
{`12.345`, []token{{t_real_number, "12.345"}}},
|
|
|
|
{`12.345 name`, []token{{t_real_number, "12.345"}, {t_name, "name"}}},
|
|
|
|
{`[12.345]`, []token{
|
|
|
|
{t_list_start, "["},
|
|
|
|
{t_real_number, "12.345"},
|
|
|
|
{t_list_end, "]"},
|
|
|
|
}},
|
|
|
|
{`[1, 2, 3]`, []token{
|
|
|
|
{t_list_start, "["},
|
|
|
|
{t_real_number, "1"},
|
|
|
|
{t_list_separator, ","},
|
|
|
|
{t_real_number, "2"},
|
|
|
|
{t_list_separator, ","},
|
|
|
|
{t_real_number, "3"},
|
|
|
|
{t_list_end, "]"},
|
|
|
|
}},
|
|
|
|
{`1i`, []token{{t_imaginary_number, "1i"}}},
|
|
|
|
// a complex number generates two lexemes; one for its real component,
|
|
|
|
// and one for its imaginary component.
|
|
|
|
{`1+2i`, []token{{t_real_number, "1"}, {t_imaginary_number, "+2i"}}},
|
|
|
|
{`1e9`, []token{{t_real_number, "1e9"}}},
|
|
|
|
{`1e+9`, []token{{t_real_number, "1e+9"}}},
|
|
|
|
{`1E-9`, []token{{t_real_number, "1E-9"}}},
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestLexPrimities(t *testing.T) {
|
|
|
|
for _, test := range primitivesTests {
|
|
|
|
tokens, err := fullTokens(lexString(test.in))
|
|
|
|
if err != nil {
|
|
|
|
t.Error(err)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
// tokens = tokens[:len(tokens)-1]
|
|
|
|
if len(tokens) != len(test.out) {
|
|
|
|
t.Errorf("expected %d token, saw %d: %v", len(test.out), len(tokens), tokens)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
for i := range tokens {
|
|
|
|
if tokens[i] != test.out[i] {
|
|
|
|
t.Errorf("token %d is %v, expected %v", i, tokens[i], test.out[i])
|
|
|
|
}
|
|
|
|
}
|
|
|
|
t.Logf("OK: %s", test.in)
|
|
|
|
}
|
|
|
|
}
|