mirror of
https://github.com/Luzifer/ansible-role-version.git
synced 2024-12-23 19:11:20 +00:00
418 lines
11 KiB
Go
418 lines
11 KiB
Go
|
// Copyright 2009 The Go Authors. All rights reserved.
|
|||
|
// Use of this source code is governed by a BSD-style
|
|||
|
// license that can be found in the LICENSE file.
|
|||
|
|
|||
|
package scanner
|
|||
|
|
|||
|
import (
|
|||
|
"os"
|
|||
|
"strings"
|
|||
|
"testing"
|
|||
|
)
|
|||
|
|
|||
|
import (
|
|||
|
"github.com/src-d/gcfg/token"
|
|||
|
)
|
|||
|
|
|||
|
var fset = token.NewFileSet()
|
|||
|
|
|||
|
const /* class */ (
|
|||
|
special = iota
|
|||
|
literal
|
|||
|
operator
|
|||
|
)
|
|||
|
|
|||
|
func tokenclass(tok token.Token) int {
|
|||
|
switch {
|
|||
|
case tok.IsLiteral():
|
|||
|
return literal
|
|||
|
case tok.IsOperator():
|
|||
|
return operator
|
|||
|
}
|
|||
|
return special
|
|||
|
}
|
|||
|
|
|||
|
type elt struct {
|
|||
|
tok token.Token
|
|||
|
lit string
|
|||
|
class int
|
|||
|
pre string
|
|||
|
suf string
|
|||
|
}
|
|||
|
|
|||
|
var tokens = [...]elt{
|
|||
|
// Special tokens
|
|||
|
{token.COMMENT, "; a comment", special, "", "\n"},
|
|||
|
{token.COMMENT, "# a comment", special, "", "\n"},
|
|||
|
|
|||
|
// Operators and delimiters
|
|||
|
{token.ASSIGN, "=", operator, "", "value"},
|
|||
|
{token.LBRACK, "[", operator, "", ""},
|
|||
|
{token.RBRACK, "]", operator, "", ""},
|
|||
|
{token.EOL, "\n", operator, "", ""},
|
|||
|
|
|||
|
// Identifiers
|
|||
|
{token.IDENT, "foobar", literal, "", ""},
|
|||
|
{token.IDENT, "a۰۱۸", literal, "", ""},
|
|||
|
{token.IDENT, "foo६४", literal, "", ""},
|
|||
|
{token.IDENT, "bar9876", literal, "", ""},
|
|||
|
{token.IDENT, "foo-bar", literal, "", ""},
|
|||
|
{token.IDENT, "foo", literal, ";\n", ""},
|
|||
|
// String literals (subsection names)
|
|||
|
{token.STRING, `"foobar"`, literal, "", ""},
|
|||
|
{token.STRING, `"\""`, literal, "", ""},
|
|||
|
// String literals (values)
|
|||
|
{token.STRING, `"\n"`, literal, "=", ""},
|
|||
|
{token.STRING, `"foobar"`, literal, "=", ""},
|
|||
|
{token.STRING, `"foo\nbar"`, literal, "=", ""},
|
|||
|
{token.STRING, `"foo\"bar"`, literal, "=", ""},
|
|||
|
{token.STRING, `"foo\\bar"`, literal, "=", ""},
|
|||
|
{token.STRING, `"foobar"`, literal, "=", ""},
|
|||
|
{token.STRING, `"foobar"`, literal, "= ", ""},
|
|||
|
{token.STRING, `"foobar"`, literal, "=", "\n"},
|
|||
|
{token.STRING, `"foobar"`, literal, "=", ";"},
|
|||
|
{token.STRING, `"foobar"`, literal, "=", " ;"},
|
|||
|
{token.STRING, `"foobar"`, literal, "=", "#"},
|
|||
|
{token.STRING, `"foobar"`, literal, "=", " #"},
|
|||
|
{token.STRING, "foobar", literal, "=", ""},
|
|||
|
{token.STRING, "foobar", literal, "= ", ""},
|
|||
|
{token.STRING, "foobar", literal, "=", " "},
|
|||
|
{token.STRING, `"foo" "bar"`, literal, "=", " "},
|
|||
|
{token.STRING, "foo\\\nbar", literal, "=", ""},
|
|||
|
{token.STRING, "foo\\\r\nbar", literal, "=", ""},
|
|||
|
}
|
|||
|
|
|||
|
const whitespace = " \t \n\n\n" // to separate tokens
|
|||
|
|
|||
|
var source = func() []byte {
|
|||
|
var src []byte
|
|||
|
for _, t := range tokens {
|
|||
|
src = append(src, t.pre...)
|
|||
|
src = append(src, t.lit...)
|
|||
|
src = append(src, t.suf...)
|
|||
|
src = append(src, whitespace...)
|
|||
|
}
|
|||
|
return src
|
|||
|
}()
|
|||
|
|
|||
|
func newlineCount(s string) int {
|
|||
|
n := 0
|
|||
|
for i := 0; i < len(s); i++ {
|
|||
|
if s[i] == '\n' {
|
|||
|
n++
|
|||
|
}
|
|||
|
}
|
|||
|
return n
|
|||
|
}
|
|||
|
|
|||
|
func checkPos(t *testing.T, lit string, p token.Pos, expected token.Position) {
|
|||
|
pos := fset.Position(p)
|
|||
|
if pos.Filename != expected.Filename {
|
|||
|
t.Errorf("bad filename for %q: got %s, expected %s", lit, pos.Filename, expected.Filename)
|
|||
|
}
|
|||
|
if pos.Offset != expected.Offset {
|
|||
|
t.Errorf("bad position for %q: got %d, expected %d", lit, pos.Offset, expected.Offset)
|
|||
|
}
|
|||
|
if pos.Line != expected.Line {
|
|||
|
t.Errorf("bad line for %q: got %d, expected %d", lit, pos.Line, expected.Line)
|
|||
|
}
|
|||
|
if pos.Column != expected.Column {
|
|||
|
t.Errorf("bad column for %q: got %d, expected %d", lit, pos.Column, expected.Column)
|
|||
|
}
|
|||
|
}
|
|||
|
|
|||
|
// Verify that calling Scan() provides the correct results.
|
|||
|
func TestScan(t *testing.T) {
|
|||
|
// make source
|
|||
|
src_linecount := newlineCount(string(source))
|
|||
|
whitespace_linecount := newlineCount(whitespace)
|
|||
|
|
|||
|
index := 0
|
|||
|
|
|||
|
// error handler
|
|||
|
eh := func(_ token.Position, msg string) {
|
|||
|
t.Errorf("%d: error handler called (msg = %s)", index, msg)
|
|||
|
}
|
|||
|
|
|||
|
// verify scan
|
|||
|
var s Scanner
|
|||
|
s.Init(fset.AddFile("", fset.Base(), len(source)), source, eh, ScanComments)
|
|||
|
// epos is the expected position
|
|||
|
epos := token.Position{
|
|||
|
Filename: "",
|
|||
|
Offset: 0,
|
|||
|
Line: 1,
|
|||
|
Column: 1,
|
|||
|
}
|
|||
|
for {
|
|||
|
pos, tok, lit := s.Scan()
|
|||
|
if lit == "" {
|
|||
|
// no literal value for non-literal tokens
|
|||
|
lit = tok.String()
|
|||
|
}
|
|||
|
e := elt{token.EOF, "", special, "", ""}
|
|||
|
if index < len(tokens) {
|
|||
|
e = tokens[index]
|
|||
|
}
|
|||
|
if tok == token.EOF {
|
|||
|
lit = "<EOF>"
|
|||
|
epos.Line = src_linecount
|
|||
|
epos.Column = 2
|
|||
|
}
|
|||
|
if e.pre != "" && strings.ContainsRune("=;#", rune(e.pre[0])) {
|
|||
|
epos.Column = 1
|
|||
|
checkPos(t, lit, pos, epos)
|
|||
|
var etok token.Token
|
|||
|
if e.pre[0] == '=' {
|
|||
|
etok = token.ASSIGN
|
|||
|
} else {
|
|||
|
etok = token.COMMENT
|
|||
|
}
|
|||
|
if tok != etok {
|
|||
|
t.Errorf("bad token for %q: got %q, expected %q", lit, tok, etok)
|
|||
|
}
|
|||
|
pos, tok, lit = s.Scan()
|
|||
|
}
|
|||
|
epos.Offset += len(e.pre)
|
|||
|
if tok != token.EOF {
|
|||
|
epos.Column = 1 + len(e.pre)
|
|||
|
}
|
|||
|
if e.pre != "" && e.pre[len(e.pre)-1] == '\n' {
|
|||
|
epos.Offset--
|
|||
|
epos.Column--
|
|||
|
checkPos(t, lit, pos, epos)
|
|||
|
if tok != token.EOL {
|
|||
|
t.Errorf("bad token for %q: got %q, expected %q", lit, tok, token.EOL)
|
|||
|
}
|
|||
|
epos.Line++
|
|||
|
epos.Offset++
|
|||
|
epos.Column = 1
|
|||
|
pos, tok, lit = s.Scan()
|
|||
|
}
|
|||
|
checkPos(t, lit, pos, epos)
|
|||
|
if tok != e.tok {
|
|||
|
t.Errorf("bad token for %q: got %q, expected %q", lit, tok, e.tok)
|
|||
|
}
|
|||
|
if e.tok.IsLiteral() {
|
|||
|
// no CRs in value string literals
|
|||
|
elit := e.lit
|
|||
|
if strings.ContainsRune(e.pre, '=') {
|
|||
|
elit = string(stripCR([]byte(elit)))
|
|||
|
epos.Offset += len(e.lit) - len(lit) // correct position
|
|||
|
}
|
|||
|
if lit != elit {
|
|||
|
t.Errorf("bad literal for %q: got %q, expected %q", lit, lit, elit)
|
|||
|
}
|
|||
|
}
|
|||
|
if tokenclass(tok) != e.class {
|
|||
|
t.Errorf("bad class for %q: got %d, expected %d", lit, tokenclass(tok), e.class)
|
|||
|
}
|
|||
|
epos.Offset += len(lit) + len(e.suf) + len(whitespace)
|
|||
|
epos.Line += newlineCount(lit) + newlineCount(e.suf) + whitespace_linecount
|
|||
|
index++
|
|||
|
if tok == token.EOF {
|
|||
|
break
|
|||
|
}
|
|||
|
if e.suf == "value" {
|
|||
|
pos, tok, lit = s.Scan()
|
|||
|
if tok != token.STRING {
|
|||
|
t.Errorf("bad token for %q: got %q, expected %q", lit, tok, token.STRING)
|
|||
|
}
|
|||
|
} else if strings.ContainsRune(e.suf, ';') || strings.ContainsRune(e.suf, '#') {
|
|||
|
pos, tok, lit = s.Scan()
|
|||
|
if tok != token.COMMENT {
|
|||
|
t.Errorf("bad token for %q: got %q, expected %q", lit, tok, token.COMMENT)
|
|||
|
}
|
|||
|
}
|
|||
|
// skip EOLs
|
|||
|
for i := 0; i < whitespace_linecount+newlineCount(e.suf); i++ {
|
|||
|
pos, tok, lit = s.Scan()
|
|||
|
if tok != token.EOL {
|
|||
|
t.Errorf("bad token for %q: got %q, expected %q", lit, tok, token.EOL)
|
|||
|
}
|
|||
|
}
|
|||
|
}
|
|||
|
if s.ErrorCount != 0 {
|
|||
|
t.Errorf("found %d errors", s.ErrorCount)
|
|||
|
}
|
|||
|
}
|
|||
|
|
|||
|
func TestScanValStringEOF(t *testing.T) {
|
|||
|
var s Scanner
|
|||
|
src := "= value"
|
|||
|
f := fset.AddFile("src", fset.Base(), len(src))
|
|||
|
s.Init(f, []byte(src), nil, 0)
|
|||
|
s.Scan() // =
|
|||
|
s.Scan() // value
|
|||
|
_, tok, _ := s.Scan() // EOF
|
|||
|
if tok != token.EOF {
|
|||
|
t.Errorf("bad token: got %s, expected %s", tok, token.EOF)
|
|||
|
}
|
|||
|
if s.ErrorCount > 0 {
|
|||
|
t.Error("scanning error")
|
|||
|
}
|
|||
|
}
|
|||
|
|
|||
|
// Verify that initializing the same scanner more then once works correctly.
|
|||
|
func TestInit(t *testing.T) {
|
|||
|
var s Scanner
|
|||
|
|
|||
|
// 1st init
|
|||
|
src1 := "\nname = value"
|
|||
|
f1 := fset.AddFile("src1", fset.Base(), len(src1))
|
|||
|
s.Init(f1, []byte(src1), nil, 0)
|
|||
|
if f1.Size() != len(src1) {
|
|||
|
t.Errorf("bad file size: got %d, expected %d", f1.Size(), len(src1))
|
|||
|
}
|
|||
|
s.Scan() // \n
|
|||
|
s.Scan() // name
|
|||
|
_, tok, _ := s.Scan() // =
|
|||
|
if tok != token.ASSIGN {
|
|||
|
t.Errorf("bad token: got %s, expected %s", tok, token.ASSIGN)
|
|||
|
}
|
|||
|
|
|||
|
// 2nd init
|
|||
|
src2 := "[section]"
|
|||
|
f2 := fset.AddFile("src2", fset.Base(), len(src2))
|
|||
|
s.Init(f2, []byte(src2), nil, 0)
|
|||
|
if f2.Size() != len(src2) {
|
|||
|
t.Errorf("bad file size: got %d, expected %d", f2.Size(), len(src2))
|
|||
|
}
|
|||
|
_, tok, _ = s.Scan() // [
|
|||
|
if tok != token.LBRACK {
|
|||
|
t.Errorf("bad token: got %s, expected %s", tok, token.LBRACK)
|
|||
|
}
|
|||
|
|
|||
|
if s.ErrorCount != 0 {
|
|||
|
t.Errorf("found %d errors", s.ErrorCount)
|
|||
|
}
|
|||
|
}
|
|||
|
|
|||
|
func TestStdErrorHandler(t *testing.T) {
|
|||
|
const src = "@\n" + // illegal character, cause an error
|
|||
|
"@ @\n" // two errors on the same line
|
|||
|
|
|||
|
var list ErrorList
|
|||
|
eh := func(pos token.Position, msg string) { list.Add(pos, msg) }
|
|||
|
|
|||
|
var s Scanner
|
|||
|
s.Init(fset.AddFile("File1", fset.Base(), len(src)), []byte(src), eh, 0)
|
|||
|
for {
|
|||
|
if _, tok, _ := s.Scan(); tok == token.EOF {
|
|||
|
break
|
|||
|
}
|
|||
|
}
|
|||
|
|
|||
|
if len(list) != s.ErrorCount {
|
|||
|
t.Errorf("found %d errors, expected %d", len(list), s.ErrorCount)
|
|||
|
}
|
|||
|
|
|||
|
if len(list) != 3 {
|
|||
|
t.Errorf("found %d raw errors, expected 3", len(list))
|
|||
|
PrintError(os.Stderr, list)
|
|||
|
}
|
|||
|
|
|||
|
list.Sort()
|
|||
|
if len(list) != 3 {
|
|||
|
t.Errorf("found %d sorted errors, expected 3", len(list))
|
|||
|
PrintError(os.Stderr, list)
|
|||
|
}
|
|||
|
|
|||
|
list.RemoveMultiples()
|
|||
|
if len(list) != 2 {
|
|||
|
t.Errorf("found %d one-per-line errors, expected 2", len(list))
|
|||
|
PrintError(os.Stderr, list)
|
|||
|
}
|
|||
|
}
|
|||
|
|
|||
|
type errorCollector struct {
|
|||
|
cnt int // number of errors encountered
|
|||
|
msg string // last error message encountered
|
|||
|
pos token.Position // last error position encountered
|
|||
|
}
|
|||
|
|
|||
|
func checkError(t *testing.T, src string, tok token.Token, pos int, err string) {
|
|||
|
var s Scanner
|
|||
|
var h errorCollector
|
|||
|
eh := func(pos token.Position, msg string) {
|
|||
|
h.cnt++
|
|||
|
h.msg = msg
|
|||
|
h.pos = pos
|
|||
|
}
|
|||
|
s.Init(fset.AddFile("", fset.Base(), len(src)), []byte(src), eh, ScanComments)
|
|||
|
if src[0] == '=' {
|
|||
|
_, _, _ = s.Scan()
|
|||
|
}
|
|||
|
_, tok0, _ := s.Scan()
|
|||
|
_, tok1, _ := s.Scan()
|
|||
|
if tok0 != tok {
|
|||
|
t.Errorf("%q: got %s, expected %s", src, tok0, tok)
|
|||
|
}
|
|||
|
if tok1 != token.EOF {
|
|||
|
t.Errorf("%q: got %s, expected EOF", src, tok1)
|
|||
|
}
|
|||
|
cnt := 0
|
|||
|
if err != "" {
|
|||
|
cnt = 1
|
|||
|
}
|
|||
|
if h.cnt != cnt {
|
|||
|
t.Errorf("%q: got cnt %d, expected %d", src, h.cnt, cnt)
|
|||
|
}
|
|||
|
if h.msg != err {
|
|||
|
t.Errorf("%q: got msg %q, expected %q", src, h.msg, err)
|
|||
|
}
|
|||
|
if h.pos.Offset != pos {
|
|||
|
t.Errorf("%q: got offset %d, expected %d", src, h.pos.Offset, pos)
|
|||
|
}
|
|||
|
}
|
|||
|
|
|||
|
var errors = []struct {
|
|||
|
src string
|
|||
|
tok token.Token
|
|||
|
pos int
|
|||
|
err string
|
|||
|
}{
|
|||
|
{"\a", token.ILLEGAL, 0, "illegal character U+0007"},
|
|||
|
{"/", token.ILLEGAL, 0, "illegal character U+002F '/'"},
|
|||
|
{"_", token.ILLEGAL, 0, "illegal character U+005F '_'"},
|
|||
|
{`…`, token.ILLEGAL, 0, "illegal character U+2026 '…'"},
|
|||
|
{`""`, token.STRING, 0, ""},
|
|||
|
{`"`, token.STRING, 0, "string not terminated"},
|
|||
|
{"\"\n", token.STRING, 0, "string not terminated"},
|
|||
|
{`="`, token.STRING, 1, "string not terminated"},
|
|||
|
{"=\"\n", token.STRING, 1, "string not terminated"},
|
|||
|
{"=\\", token.STRING, 1, "unquoted '\\' must be followed by new line"},
|
|||
|
{"=\\\r", token.STRING, 1, "unquoted '\\' must be followed by new line"},
|
|||
|
{`"\z"`, token.STRING, 2, "unknown escape sequence"},
|
|||
|
{`"\a"`, token.STRING, 2, "unknown escape sequence"},
|
|||
|
{`"\b"`, token.STRING, 2, "unknown escape sequence"},
|
|||
|
{`"\f"`, token.STRING, 2, "unknown escape sequence"},
|
|||
|
{`"\r"`, token.STRING, 2, "unknown escape sequence"},
|
|||
|
{`"\t"`, token.STRING, 2, "unknown escape sequence"},
|
|||
|
{`"\v"`, token.STRING, 2, "unknown escape sequence"},
|
|||
|
{`"\0"`, token.STRING, 2, "unknown escape sequence"},
|
|||
|
}
|
|||
|
|
|||
|
func TestScanErrors(t *testing.T) {
|
|||
|
for _, e := range errors {
|
|||
|
checkError(t, e.src, e.tok, e.pos, e.err)
|
|||
|
}
|
|||
|
}
|
|||
|
|
|||
|
func BenchmarkScan(b *testing.B) {
|
|||
|
b.StopTimer()
|
|||
|
fset := token.NewFileSet()
|
|||
|
file := fset.AddFile("", fset.Base(), len(source))
|
|||
|
var s Scanner
|
|||
|
b.StartTimer()
|
|||
|
for i := b.N - 1; i >= 0; i-- {
|
|||
|
s.Init(file, source, nil, ScanComments)
|
|||
|
for {
|
|||
|
_, tok, _ := s.Scan()
|
|||
|
if tok == token.EOF {
|
|||
|
break
|
|||
|
}
|
|||
|
}
|
|||
|
}
|
|||
|
}
|