refactor peek functionality, begin whole feature parse tests
Этот коммит содержится в:
родитель
df957153a8
коммит
951b1357b4
5 изменённых файлов: 144 добавлений и 73 удалений
|
@ -9,7 +9,6 @@ import (
|
|||
|
||||
type Lexer struct {
|
||||
reader *bufio.Reader
|
||||
peek *Token
|
||||
lines int
|
||||
}
|
||||
|
||||
|
@ -19,30 +18,7 @@ func New(r io.Reader) *Lexer {
|
|||
}
|
||||
}
|
||||
|
||||
func (l *Lexer) Next(skip ...TokenType) (t *Token) {
|
||||
if l.peek != nil {
|
||||
t = l.peek
|
||||
l.peek = nil
|
||||
} else {
|
||||
t = l.read()
|
||||
}
|
||||
|
||||
for _, typ := range skip {
|
||||
if t.Type == typ {
|
||||
return l.Next(skip...)
|
||||
}
|
||||
}
|
||||
return
|
||||
}
|
||||
|
||||
func (l *Lexer) Peek() *Token {
|
||||
if l.peek == nil {
|
||||
l.peek = l.read()
|
||||
}
|
||||
return l.peek
|
||||
}
|
||||
|
||||
func (l *Lexer) read() *Token {
|
||||
func (l *Lexer) Next() *Token {
|
||||
line, err := l.reader.ReadString(byte('\n'))
|
||||
if err != nil && len(line) == 0 {
|
||||
return &Token{
|
||||
|
|
|
@ -12,10 +12,21 @@ import (
|
|||
|
||||
type Tag string
|
||||
|
||||
type Tags []Tag
|
||||
|
||||
func (t Tags) Has(tag Tag) bool {
|
||||
for _, tg := range t {
|
||||
if tg == tag {
|
||||
return true
|
||||
}
|
||||
}
|
||||
return false
|
||||
}
|
||||
|
||||
type Scenario struct {
|
||||
Title string
|
||||
Steps []*Step
|
||||
Tags []Tag
|
||||
Tags Tags
|
||||
}
|
||||
|
||||
type Background struct {
|
||||
|
@ -38,7 +49,8 @@ type Step struct {
|
|||
}
|
||||
|
||||
type Feature struct {
|
||||
Tags []Tag
|
||||
Path string
|
||||
Tags Tags
|
||||
Description string
|
||||
Title string
|
||||
Background *Background
|
||||
|
@ -65,9 +77,10 @@ var allSteps = []lexer.TokenType{
|
|||
var ErrEmpty = errors.New("the feature file is empty")
|
||||
|
||||
type parser struct {
|
||||
lx *lexer.Lexer
|
||||
path string
|
||||
ast *AST
|
||||
lx *lexer.Lexer
|
||||
path string
|
||||
ast *AST
|
||||
peeked *lexer.Token
|
||||
}
|
||||
|
||||
func Parse(path string) (*Feature, error) {
|
||||
|
@ -89,21 +102,23 @@ func (p *parser) next() *lexer.Token {
|
|||
if p.ast.tail != nil && p.ast.tail.value.Type == lexer.EOF {
|
||||
return p.ast.tail.value // has reached EOF, do not record it more than once
|
||||
}
|
||||
tok := p.lx.Next()
|
||||
tok := p.peek()
|
||||
p.ast.addTail(tok)
|
||||
if tok.OfType(lexer.COMMENT, lexer.NEW_LINE) {
|
||||
return p.next()
|
||||
}
|
||||
p.peeked = nil
|
||||
return tok
|
||||
}
|
||||
|
||||
// peaks into next token, skips comments or new lines
|
||||
func (p *parser) peek() *lexer.Token {
|
||||
if tok := p.lx.Peek(); !tok.OfType(lexer.COMMENT, lexer.NEW_LINE) {
|
||||
return tok
|
||||
if p.peeked != nil {
|
||||
return p.peeked
|
||||
}
|
||||
p.next()
|
||||
return p.peek()
|
||||
|
||||
for p.peeked = p.lx.Next(); p.peeked.OfType(lexer.COMMENT, lexer.NEW_LINE); p.peeked = p.lx.Next() {
|
||||
p.ast.addTail(p.peeked) // record comments and newlines
|
||||
}
|
||||
|
||||
return p.peeked
|
||||
}
|
||||
|
||||
func (p *parser) err(s string, l int) error {
|
||||
|
@ -111,48 +126,51 @@ func (p *parser) err(s string, l int) error {
|
|||
}
|
||||
|
||||
func (p *parser) parseFeature() (ft *Feature, err error) {
|
||||
var tok *lexer.Token = p.next()
|
||||
if tok.Type == lexer.EOF {
|
||||
return nil, ErrEmpty
|
||||
}
|
||||
|
||||
ft = &Feature{}
|
||||
if tok.Type == lexer.TAGS {
|
||||
if p.peek().Type != lexer.FEATURE {
|
||||
return ft, p.err("tags must be a single line next to a feature definition", tok.Line)
|
||||
}
|
||||
ft.Tags = p.parseTags(tok.Value)
|
||||
tok = p.next()
|
||||
|
||||
ft = &Feature{Path: p.path, AST: p.ast}
|
||||
switch p.peek().Type {
|
||||
case lexer.EOF:
|
||||
// if p.ast.tail != nil {
|
||||
// log.Println("peeked at:", p.peek().Type, p.ast.tail.prev.value.Type)
|
||||
// }
|
||||
return ft, ErrEmpty
|
||||
case lexer.TAGS:
|
||||
ft.Tags = p.parseTags()
|
||||
}
|
||||
|
||||
tok := p.next()
|
||||
if tok.Type != lexer.FEATURE {
|
||||
return ft, p.err("expected a file to begin with a feature definition, but got '"+tok.Type.String()+"' instead", tok.Line)
|
||||
}
|
||||
|
||||
ft.Title = tok.Value
|
||||
|
||||
var desc []string
|
||||
for ; p.peek().Type == lexer.TEXT; tok = p.next() {
|
||||
desc = append(desc, tok.Value)
|
||||
}
|
||||
ft.Description = strings.Join(desc, "\n")
|
||||
|
||||
tok = p.next()
|
||||
for tok = p.next(); tok.Type != lexer.EOF; p.next() {
|
||||
for tok = p.peek(); tok.Type != lexer.EOF; tok = p.peek() {
|
||||
// log.Println("loop peeked:", tok.Type)
|
||||
// there may be a background
|
||||
if tok.Type == lexer.BACKGROUND {
|
||||
if ft.Background != nil {
|
||||
return ft, p.err("there can only be a single background section, but found another", tok.Line)
|
||||
}
|
||||
|
||||
ft.Background = &Background{}
|
||||
p.next() // jump to background steps
|
||||
if ft.Background.Steps, err = p.parseSteps(); err != nil {
|
||||
return ft, err
|
||||
}
|
||||
continue
|
||||
tok = p.peek() // peek to scenario or tags
|
||||
}
|
||||
|
||||
// there may be tags before scenario
|
||||
sc := &Scenario{}
|
||||
if tok.Type == lexer.TAGS {
|
||||
sc.Tags, tok = p.parseTags(tok.Value), p.next()
|
||||
sc.Tags = p.parseTags()
|
||||
tok = p.peek()
|
||||
}
|
||||
|
||||
// there must be a scenario otherwise
|
||||
|
@ -161,19 +179,18 @@ func (p *parser) parseFeature() (ft *Feature, err error) {
|
|||
}
|
||||
|
||||
sc.Title = tok.Value
|
||||
p.next() // jump to scenario steps
|
||||
if sc.Steps, err = p.parseSteps(); err != nil {
|
||||
return ft, err
|
||||
}
|
||||
ft.Scenarios = append(ft.Scenarios, sc)
|
||||
}
|
||||
|
||||
ft.AST = p.ast
|
||||
return ft, nil
|
||||
}
|
||||
|
||||
func (p *parser) parseSteps() (steps []*Step, err error) {
|
||||
for tok := p.peek(); tok.OfType(allSteps...); tok = p.peek() {
|
||||
p.next() // move over the step
|
||||
step := &Step{Text: tok.Value}
|
||||
switch tok.Type {
|
||||
case lexer.GIVEN:
|
||||
|
@ -190,9 +207,10 @@ func (p *parser) parseSteps() (steps []*Step, err error) {
|
|||
}
|
||||
}
|
||||
|
||||
p.next() // have read a peeked step
|
||||
if step.Text[len(step.Text)-1] == ':' {
|
||||
next := p.peek()
|
||||
switch next.Type {
|
||||
tok = p.peek()
|
||||
switch tok.Type {
|
||||
case lexer.PYSTRING:
|
||||
if err := p.parsePystring(step); err != nil {
|
||||
return steps, err
|
||||
|
@ -202,12 +220,13 @@ func (p *parser) parseSteps() (steps []*Step, err error) {
|
|||
return steps, err
|
||||
}
|
||||
default:
|
||||
return steps, p.err("pystring or table row was expected, but got: '"+next.Type.String()+"' instead", next.Line)
|
||||
return steps, p.err("pystring or table row was expected, but got: '"+tok.Type.String()+"' instead", tok.Line)
|
||||
}
|
||||
}
|
||||
|
||||
steps = append(steps, step)
|
||||
}
|
||||
|
||||
return steps, nil
|
||||
}
|
||||
|
||||
|
@ -242,11 +261,11 @@ func (p *parser) parseTable(s *Step) error {
|
|||
return nil
|
||||
}
|
||||
|
||||
func (p *parser) parseTags(s string) (tags []Tag) {
|
||||
for _, tag := range strings.Split(s, " ") {
|
||||
t := strings.Trim(tag, "@ ")
|
||||
if len(t) > 0 {
|
||||
tags = append(tags, Tag(t))
|
||||
func (p *parser) parseTags() (tags Tags) {
|
||||
for _, tag := range strings.Split(p.next().Value, " ") {
|
||||
t := Tag(strings.Trim(tag, "@ "))
|
||||
if len(t) > 0 && !tags.Has(t) {
|
||||
tags = append(tags, t)
|
||||
}
|
||||
}
|
||||
return
|
||||
|
|
|
@ -8,7 +8,7 @@ import (
|
|||
)
|
||||
|
||||
var testFeatureSamples = map[string]string{
|
||||
"full": `Feature: gherkin parser
|
||||
"feature": `Feature: gherkin parser
|
||||
in order to run features
|
||||
as gherkin lexer
|
||||
I need to be able to parse a feature`,
|
||||
|
@ -20,9 +20,15 @@ var testFeatureSamples = map[string]string{
|
|||
Feature: gherkin`,
|
||||
}
|
||||
|
||||
func (f *Feature) assertTitle(title string, t *testing.T) {
|
||||
if f.Title != title {
|
||||
t.Fatalf("expected feature title to be '%s', but got '%s'", title, f.Title)
|
||||
}
|
||||
}
|
||||
|
||||
func Test_parse_normal_feature(t *testing.T) {
|
||||
p := &parser{
|
||||
lx: lexer.New(strings.NewReader(testFeatureSamples["full"])),
|
||||
lx: lexer.New(strings.NewReader(testFeatureSamples["feature"])),
|
||||
path: "some.feature",
|
||||
ast: newAST(),
|
||||
}
|
||||
|
@ -42,7 +48,6 @@ func Test_parse_normal_feature(t *testing.T) {
|
|||
lexer.TEXT,
|
||||
lexer.TEXT,
|
||||
lexer.TEXT,
|
||||
lexer.EOF,
|
||||
}, t)
|
||||
}
|
||||
|
||||
|
@ -65,7 +70,6 @@ func Test_parse_feature_without_description(t *testing.T) {
|
|||
|
||||
ft.AST.assertMatchesTypes([]lexer.TokenType{
|
||||
lexer.FEATURE,
|
||||
lexer.EOF,
|
||||
}, t)
|
||||
}
|
||||
|
||||
|
@ -79,9 +83,6 @@ func Test_parse_empty_feature_file(t *testing.T) {
|
|||
if err != ErrEmpty {
|
||||
t.Fatalf("expected an empty file error, but got none")
|
||||
}
|
||||
p.ast.assertMatchesTypes([]lexer.TokenType{
|
||||
lexer.EOF,
|
||||
}, t)
|
||||
}
|
||||
|
||||
func Test_parse_invalid_feature_with_random_text(t *testing.T) {
|
||||
|
@ -120,6 +121,5 @@ func Test_parse_feature_with_newlines(t *testing.T) {
|
|||
lexer.NEW_LINE,
|
||||
lexer.NEW_LINE,
|
||||
lexer.FEATURE,
|
||||
lexer.EOF,
|
||||
}, t)
|
||||
}
|
||||
|
|
73
gherkin/parse_test.go
Обычный файл
73
gherkin/parse_test.go
Обычный файл
|
@ -0,0 +1,73 @@
|
|||
package gherkin
|
||||
|
||||
import (
|
||||
"strings"
|
||||
"testing"
|
||||
|
||||
"github.com/l3pp4rd/go-behat/gherkin/lexer"
|
||||
)
|
||||
|
||||
func Test_parse_feature_file(t *testing.T) {
|
||||
|
||||
content := strings.Join([]string{
|
||||
// feature
|
||||
"@global-one @cust",
|
||||
testFeatureSamples["feature"] + "\n",
|
||||
// background
|
||||
indent(2, "Background:"),
|
||||
testStepSamples["given_table_hash"] + "\n",
|
||||
// scenario - normal without tags
|
||||
indent(2, "Scenario: user is able to register"),
|
||||
testStepSamples["step_group"] + "\n",
|
||||
// scenario - repeated tag, one extra
|
||||
indent(2, "@user @cust"),
|
||||
indent(2, "Scenario: password is required to login"),
|
||||
testStepSamples["step_group_another"] + "\n",
|
||||
// scenario - no steps yet
|
||||
indent(2, "@todo"), // cust - tag is repeated
|
||||
indent(2, "Scenario: user is able to reset his password"),
|
||||
}, "\n")
|
||||
|
||||
p := &parser{
|
||||
lx: lexer.New(strings.NewReader(content)),
|
||||
path: "usual.feature",
|
||||
ast: newAST(),
|
||||
}
|
||||
ft, err := p.parseFeature()
|
||||
if err != nil {
|
||||
t.Fatalf("unexpected error: %s", err)
|
||||
}
|
||||
ft.assertTitle("gherkin parser", t)
|
||||
|
||||
ft.AST.assertMatchesTypes([]lexer.TokenType{
|
||||
lexer.TAGS,
|
||||
lexer.FEATURE,
|
||||
lexer.TEXT,
|
||||
lexer.TEXT,
|
||||
lexer.TEXT,
|
||||
lexer.NEW_LINE,
|
||||
|
||||
lexer.BACKGROUND,
|
||||
lexer.GIVEN,
|
||||
lexer.TABLE_ROW,
|
||||
lexer.NEW_LINE,
|
||||
|
||||
lexer.SCENARIO,
|
||||
lexer.GIVEN,
|
||||
lexer.AND,
|
||||
lexer.WHEN,
|
||||
lexer.THEN,
|
||||
lexer.NEW_LINE,
|
||||
|
||||
lexer.TAGS,
|
||||
lexer.SCENARIO,
|
||||
lexer.GIVEN,
|
||||
lexer.AND,
|
||||
lexer.WHEN,
|
||||
lexer.THEN,
|
||||
lexer.NEW_LINE,
|
||||
|
||||
lexer.TAGS,
|
||||
lexer.SCENARIO,
|
||||
}, t)
|
||||
}
|
|
@ -11,6 +11,9 @@ func (a *AST) assertMatchesTypes(expected []lexer.TokenType, t *testing.T) {
|
|||
key := -1
|
||||
for item := a.head; item != nil; item = item.next {
|
||||
key += 1
|
||||
if len(expected) <= key {
|
||||
t.Fatalf("there are more tokens in AST then expected, next is '%s'", item.value.Type)
|
||||
}
|
||||
if expected[key] != item.value.Type {
|
||||
t.Fatalf("expected ast token '%s', but got '%s' at position: %d", expected[key], item.value.Type, key)
|
||||
}
|
||||
|
|
Загрузка…
Создание таблицы
Сослаться в новой задаче