[refactoring] rename struct field scanner.Token.Hidden

This commit is contained in:
Vadym Slizov 2020-05-18 00:04:04 +03:00
parent e98607f5d0
commit 5620697cab
11 changed files with 1179 additions and 1179 deletions

View File

@ -50,7 +50,7 @@ func (l *Parser) Error(msg string) {
} }
func (l *Parser) WithTokens() { func (l *Parser) WithTokens() {
l.Lexer.SetWithTokens(true) l.Lexer.SetWithHiddenTokens(true)
} }
// Parse the php7 Parser entrypoint // Parse the php7 Parser entrypoint
@ -93,7 +93,7 @@ func isDollar(r rune) bool {
} }
func (l *Parser) MoveFreeFloating(src ast.Vertex, dst ast.Vertex) { func (l *Parser) MoveFreeFloating(src ast.Vertex, dst ast.Vertex) {
if l.Lexer.GetWithFreeFloating() == false { if l.Lexer.GetWithHiddenTokens() == false {
return return
} }
@ -106,7 +106,7 @@ func (l *Parser) MoveFreeFloating(src ast.Vertex, dst ast.Vertex) {
} }
func (l *Parser) setFreeFloating(dst ast.Vertex, p token.Position, strings []token.Token) { func (l *Parser) setFreeFloating(dst ast.Vertex, p token.Position, strings []token.Token) {
if l.Lexer.GetWithFreeFloating() == false { if l.Lexer.GetWithHiddenTokens() == false {
return return
} }
@ -123,7 +123,7 @@ func (l *Parser) setFreeFloating(dst ast.Vertex, p token.Position, strings []tok
} }
func (l *Parser) GetFreeFloatingToken(t *scanner.Token) []token.Token { func (l *Parser) GetFreeFloatingToken(t *scanner.Token) []token.Token {
if l.Lexer.GetWithFreeFloating() == false { if l.Lexer.GetWithHiddenTokens() == false {
return []token.Token{} return []token.Token{}
} }
@ -136,7 +136,7 @@ func (l *Parser) GetFreeFloatingToken(t *scanner.Token) []token.Token {
} }
func (l *Parser) addDollarToken(v ast.Vertex) { func (l *Parser) addDollarToken(v ast.Vertex) {
if l.Lexer.GetWithFreeFloating() == false { if l.Lexer.GetWithHiddenTokens() == false {
return return
} }
@ -149,7 +149,7 @@ func (l *Parser) addDollarToken(v ast.Vertex) {
} }
func (l *Parser) splitSemiColonAndPhpCloseTag(htmlNode ast.Vertex, prevNode ast.Vertex) { func (l *Parser) splitSemiColonAndPhpCloseTag(htmlNode ast.Vertex, prevNode ast.Vertex) {
if l.Lexer.GetWithFreeFloating() == false { if l.Lexer.GetWithHiddenTokens() == false {
return return
} }

Binary file not shown.

File diff suppressed because it is too large Load Diff

View File

@ -48,7 +48,7 @@ func (l *Parser) Error(msg string) {
} }
func (l *Parser) WithTokens() { func (l *Parser) WithTokens() {
l.Lexer.SetWithTokens(true) l.Lexer.SetWithHiddenTokens(true)
} }
// Parse the php7 Parser entrypoint // Parse the php7 Parser entrypoint
@ -91,7 +91,7 @@ func isDollar(r rune) bool {
} }
func (l *Parser) MoveFreeFloating(src ast.Vertex, dst ast.Vertex) { func (l *Parser) MoveFreeFloating(src ast.Vertex, dst ast.Vertex) {
if l.Lexer.GetWithFreeFloating() == false { if l.Lexer.GetWithHiddenTokens() == false {
return return
} }
@ -104,7 +104,7 @@ func (l *Parser) MoveFreeFloating(src ast.Vertex, dst ast.Vertex) {
} }
func (l *Parser) setFreeFloating(dst ast.Vertex, p token.Position, strings []token.Token) { func (l *Parser) setFreeFloating(dst ast.Vertex, p token.Position, strings []token.Token) {
if l.Lexer.GetWithFreeFloating() == false { if l.Lexer.GetWithHiddenTokens() == false {
return return
} }
@ -121,7 +121,7 @@ func (l *Parser) setFreeFloating(dst ast.Vertex, p token.Position, strings []tok
} }
func (l *Parser) GetFreeFloatingToken(t *scanner.Token) []token.Token { func (l *Parser) GetFreeFloatingToken(t *scanner.Token) []token.Token {
if l.Lexer.GetWithFreeFloating() == false { if l.Lexer.GetWithHiddenTokens() == false {
return []token.Token{} return []token.Token{}
} }
@ -134,7 +134,7 @@ func (l *Parser) GetFreeFloatingToken(t *scanner.Token) []token.Token {
} }
func (l *Parser) addDollarToken(v ast.Vertex) { func (l *Parser) addDollarToken(v ast.Vertex) {
if l.Lexer.GetWithFreeFloating() == false { if l.Lexer.GetWithHiddenTokens() == false {
return return
} }
@ -147,7 +147,7 @@ func (l *Parser) addDollarToken(v ast.Vertex) {
} }
func (l *Parser) splitSemiColonAndPhpCloseTag(htmlNode ast.Vertex, prevNode ast.Vertex) { func (l *Parser) splitSemiColonAndPhpCloseTag(htmlNode ast.Vertex, prevNode ast.Vertex) {
if l.Lexer.GetWithFreeFloating() == false { if l.Lexer.GetWithHiddenTokens() == false {
return return
} }

Binary file not shown.

File diff suppressed because it is too large Load Diff

View File

@ -16,8 +16,8 @@ type Scanner interface {
GetPhpDocComment() string GetPhpDocComment() string
SetPhpDocComment(string) SetPhpDocComment(string)
GetErrors() []*errors.Error GetErrors() []*errors.Error
GetWithFreeFloating() bool GetWithHiddenTokens() bool
SetWithTokens(bool) SetWithHiddenTokens(bool)
AddError(e *errors.Error) AddError(e *errors.Error)
SetErrors(e []*errors.Error) SetErrors(e []*errors.Error)
} }
@ -35,14 +35,14 @@ type Lexer struct {
top int top int
heredocLabel []byte heredocLabel []byte
TokenPool *TokenPool TokenPool *TokenPool
Tokens []token.Token HiddenTokens []token.Token
WithTokens bool WithHiddenTokens bool
PhpDocComment string PhpDocComment string
lastToken *Token lastToken *Token
Errors []*errors.Error Errors []*errors.Error
NewLines NewLines NewLines NewLines
PHPVersion string PHPVersion string
} }
func (l *Lexer) ReturnTokenToPool(t *Token) { func (l *Lexer) ReturnTokenToPool(t *Token) {
@ -61,12 +61,12 @@ func (l *Lexer) GetErrors() []*errors.Error {
return l.Errors return l.Errors
} }
func (l *Lexer) GetWithFreeFloating() bool { func (l *Lexer) GetWithHiddenTokens() bool {
return l.WithTokens return l.WithHiddenTokens
} }
func (l *Lexer) SetWithTokens(b bool) { func (l *Lexer) SetWithHiddenTokens(b bool) {
l.WithTokens = b l.WithHiddenTokens = b
} }
func (l *Lexer) AddError(e *errors.Error) { func (l *Lexer) AddError(e *errors.Error) {
@ -85,11 +85,11 @@ func (lex *Lexer) setTokenPosition(token *Token) {
} }
func (lex *Lexer) addToken(id TokenID, ps, pe int) { func (lex *Lexer) addToken(id TokenID, ps, pe int) {
if !lex.WithTokens { if !lex.WithHiddenTokens {
return return
} }
lex.Tokens = append(lex.Tokens, token.Token{ lex.HiddenTokens = append(lex.HiddenTokens, token.Token{
ID: token.ID(id), ID: token.ID(id),
Value: lex.data[ps:pe], Value: lex.data[ps:pe],
}) })

Binary file not shown.

View File

@ -28,12 +28,12 @@ func NewLexer(data []byte) *Lexer {
} }
func (lex *Lexer) Lex(lval Lval) int { func (lex *Lexer) Lex(lval Lval) int {
lex.Tokens = nil lex.HiddenTokens = nil
eof := lex.pe eof := lex.pe
var tok TokenID var tok TokenID
token := lex.TokenPool.Get() token := lex.TokenPool.Get()
token.Tokens = lex.Tokens token.Hidden = nil
token.Value = lex.data[0:0] token.Value = lex.data[0:0]
lblStart := 0 lblStart := 0
@ -508,7 +508,7 @@ func (lex *Lexer) Lex(lval Lval) int {
write exec; write exec;
}%% }%%
token.Tokens = lex.Tokens token.Hidden = lex.HiddenTokens
token.Value = lex.data[lex.ts:lex.te] token.Value = lex.data[lex.ts:lex.te]
token.ID = tok token.ID = tok

View File

@ -360,7 +360,7 @@ func TestTokens(t *testing.T) {
} }
lexer := NewLexer([]byte(src)) lexer := NewLexer([]byte(src))
lexer.WithTokens = true lexer.WithHiddenTokens = true
lv := &lval{} lv := &lval{}
actual := []string{} actual := []string{}
@ -389,14 +389,14 @@ func TestShebang(t *testing.T) {
} }
lexer := NewLexer([]byte(src)) lexer := NewLexer([]byte(src))
lexer.WithTokens = true lexer.WithHiddenTokens = true
lv := &lval{} lv := &lval{}
actual := []string{} actual := []string{}
token := lexer.Lex(lv) token := lexer.Lex(lv)
assert.Equal(t, token, int(T_DNUMBER)) assert.Equal(t, token, int(T_DNUMBER))
for _, tt := range lv.Tkn.Tokens { for _, tt := range lv.Tkn.Hidden {
actual = append(actual, string(tt.Value)) actual = append(actual, string(tt.Value))
} }
@ -410,12 +410,12 @@ func TestShebangHtml(t *testing.T) {
` `
lexer := NewLexer([]byte(src)) lexer := NewLexer([]byte(src))
lexer.WithTokens = true lexer.WithHiddenTokens = true
lv := &lval{} lv := &lval{}
token := lexer.Lex(lv) token := lexer.Lex(lv)
assert.Equal(t, token, int(T_INLINE_HTML)) assert.Equal(t, token, int(T_INLINE_HTML))
assert.Equal(t, string(lv.Tkn.Tokens[0].Value), "#!/usr/bin/env php\n") assert.Equal(t, string(lv.Tkn.Hidden[0].Value), "#!/usr/bin/env php\n")
token = lexer.Lex(lv) token = lexer.Lex(lv)
assert.Equal(t, token, int(T_DNUMBER)) assert.Equal(t, token, int(T_DNUMBER))
@ -461,7 +461,7 @@ func TestNumberTokens(t *testing.T) {
} }
lexer := NewLexer([]byte(src)) lexer := NewLexer([]byte(src))
lexer.WithTokens = true lexer.WithHiddenTokens = true
lv := &lval{} lv := &lval{}
actual := []string{} actual := []string{}
@ -519,7 +519,7 @@ func TestConstantStrings(t *testing.T) {
} }
lexer := NewLexer([]byte(src)) lexer := NewLexer([]byte(src))
lexer.WithTokens = true lexer.WithHiddenTokens = true
lv := &lval{} lv := &lval{}
actual := []string{} actual := []string{}
@ -655,7 +655,7 @@ func TestTeplateStringTokens(t *testing.T) {
} }
lexer := NewLexer([]byte(src)) lexer := NewLexer([]byte(src))
lexer.WithTokens = true lexer.WithHiddenTokens = true
lv := &lval{} lv := &lval{}
actual := []string{} actual := []string{}
@ -741,7 +741,7 @@ func TestBackquoteStringTokens(t *testing.T) {
} }
lexer := NewLexer([]byte(src)) lexer := NewLexer([]byte(src))
lexer.WithTokens = true lexer.WithHiddenTokens = true
lv := &lval{} lv := &lval{}
actual := []string{} actual := []string{}
@ -836,7 +836,7 @@ CAT;
} }
lexer := NewLexer([]byte(src)) lexer := NewLexer([]byte(src))
lexer.WithTokens = true lexer.WithHiddenTokens = true
lv := &lval{} lv := &lval{}
actual := []string{} actual := []string{}
@ -910,7 +910,7 @@ CAT
} }
lexer := NewLexer([]byte(src)) lexer := NewLexer([]byte(src))
lexer.WithTokens = true lexer.WithHiddenTokens = true
lv := &lval{} lv := &lval{}
actual := []string{} actual := []string{}
@ -950,7 +950,7 @@ CAT;
} }
lexer := NewLexer([]byte(src)) lexer := NewLexer([]byte(src))
lexer.WithTokens = true lexer.WithHiddenTokens = true
lv := &lval{} lv := &lval{}
actual := []string{} actual := []string{}
@ -982,7 +982,7 @@ func TestHereDocTokens73(t *testing.T) {
} }
lexer := NewLexer([]byte(src)) lexer := NewLexer([]byte(src))
lexer.WithTokens = true lexer.WithHiddenTokens = true
lv := &lval{} lv := &lval{}
actual := []string{} actual := []string{}
@ -1014,7 +1014,7 @@ CAT;`
lexer := NewLexer([]byte(src)) lexer := NewLexer([]byte(src))
lexer.PHPVersion = "7.2" lexer.PHPVersion = "7.2"
lexer.WithTokens = true lexer.WithHiddenTokens = true
lv := &lval{} lv := &lval{}
actual := []string{} actual := []string{}
@ -1047,7 +1047,7 @@ func TestInlineHtmlNopTokens(t *testing.T) {
} }
lexer := NewLexer([]byte(src)) lexer := NewLexer([]byte(src))
lexer.WithTokens = true lexer.WithHiddenTokens = true
lv := &lval{} lv := &lval{}
actual := []string{} actual := []string{}
@ -1155,12 +1155,12 @@ func TestCommentEnd(t *testing.T) {
} }
lexer := NewLexer([]byte(src)) lexer := NewLexer([]byte(src))
lexer.WithTokens = true lexer.WithHiddenTokens = true
lv := &lval{} lv := &lval{}
lexer.Lex(lv) lexer.Lex(lv)
actual := lexer.Tokens actual := lexer.HiddenTokens
assert.DeepEqual(t, expected, actual) assert.DeepEqual(t, expected, actual)
} }
@ -1185,12 +1185,12 @@ func TestCommentNewLine(t *testing.T) {
} }
lexer := NewLexer([]byte(src)) lexer := NewLexer([]byte(src))
lexer.WithTokens = true lexer.WithHiddenTokens = true
lv := &lval{} lv := &lval{}
lexer.Lex(lv) lexer.Lex(lv)
actual := lv.Tkn.Tokens actual := lv.Tkn.Hidden
assert.DeepEqual(t, expected, actual) assert.DeepEqual(t, expected, actual)
} }
@ -1215,12 +1215,12 @@ func TestCommentNewLine1(t *testing.T) {
} }
lexer := NewLexer([]byte(src)) lexer := NewLexer([]byte(src))
lexer.WithTokens = true lexer.WithHiddenTokens = true
lv := &lval{} lv := &lval{}
lexer.Lex(lv) lexer.Lex(lv)
actual := lv.Tkn.Tokens actual := lv.Tkn.Hidden
assert.DeepEqual(t, expected, actual) assert.DeepEqual(t, expected, actual)
} }
@ -1245,12 +1245,12 @@ func TestCommentNewLine2(t *testing.T) {
} }
lexer := NewLexer([]byte(src)) lexer := NewLexer([]byte(src))
lexer.WithTokens = true lexer.WithHiddenTokens = true
lv := &lval{} lv := &lval{}
lexer.Lex(lv) lexer.Lex(lv)
actual := lv.Tkn.Tokens actual := lv.Tkn.Hidden
assert.DeepEqual(t, expected, actual) assert.DeepEqual(t, expected, actual)
} }
@ -1276,12 +1276,12 @@ func TestCommentWithPhpEndTag(t *testing.T) {
} }
lexer := NewLexer([]byte(src)) lexer := NewLexer([]byte(src))
lexer.WithTokens = true lexer.WithHiddenTokens = true
lv := &lval{} lv := &lval{}
lexer.Lex(lv) lexer.Lex(lv)
actual := lv.Tkn.Tokens actual := lv.Tkn.Hidden
assert.DeepEqual(t, expected, actual) assert.DeepEqual(t, expected, actual)
} }
@ -1307,12 +1307,12 @@ func TestInlineComment(t *testing.T) {
} }
lexer := NewLexer([]byte(src)) lexer := NewLexer([]byte(src))
lexer.WithTokens = true lexer.WithHiddenTokens = true
lv := &lval{} lv := &lval{}
lexer.Lex(lv) lexer.Lex(lv)
actual := lv.Tkn.Tokens actual := lv.Tkn.Hidden
assert.DeepEqual(t, expected, actual) assert.DeepEqual(t, expected, actual)
} }
@ -1338,12 +1338,12 @@ func TestInlineComment2(t *testing.T) {
} }
lexer := NewLexer([]byte(src)) lexer := NewLexer([]byte(src))
lexer.WithTokens = true lexer.WithHiddenTokens = true
lv := &lval{} lv := &lval{}
lexer.Lex(lv) lexer.Lex(lv)
actual := lexer.Tokens actual := lexer.HiddenTokens
assert.DeepEqual(t, expected, actual) assert.DeepEqual(t, expected, actual)
} }
@ -1373,12 +1373,12 @@ func TestEmptyInlineComment(t *testing.T) {
} }
lexer := NewLexer([]byte(src)) lexer := NewLexer([]byte(src))
lexer.WithTokens = true lexer.WithHiddenTokens = true
lv := &lval{} lv := &lval{}
lexer.Lex(lv) lexer.Lex(lv)
actual := lexer.Tokens actual := lexer.HiddenTokens
assert.DeepEqual(t, expected, actual) assert.DeepEqual(t, expected, actual)
} }
@ -1404,12 +1404,12 @@ func TestEmptyInlineComment2(t *testing.T) {
} }
lexer := NewLexer([]byte(src)) lexer := NewLexer([]byte(src))
lexer.WithTokens = true lexer.WithHiddenTokens = true
lv := &lval{} lv := &lval{}
lexer.Lex(lv) lexer.Lex(lv)
actual := lv.Tkn.Tokens actual := lv.Tkn.Hidden
assert.DeepEqual(t, expected, actual) assert.DeepEqual(t, expected, actual)
} }
@ -1419,7 +1419,7 @@ func TestMethodCallTokens(t *testing.T) {
$a -> bar ( '' ) ;` $a -> bar ( '' ) ;`
lexer := NewLexer([]byte(src)) lexer := NewLexer([]byte(src))
lexer.WithTokens = true lexer.WithHiddenTokens = true
lv := &lval{} lv := &lval{}
expected := []token.Token{ expected := []token.Token{
@ -1433,7 +1433,7 @@ func TestMethodCallTokens(t *testing.T) {
}, },
} }
lexer.Lex(lv) lexer.Lex(lv)
actual := lv.Tkn.Tokens actual := lv.Tkn.Hidden
assert.DeepEqual(t, expected, actual) assert.DeepEqual(t, expected, actual)
expected = []token.Token{ expected = []token.Token{
@ -1443,7 +1443,7 @@ func TestMethodCallTokens(t *testing.T) {
}, },
} }
lexer.Lex(lv) lexer.Lex(lv)
actual = lv.Tkn.Tokens actual = lv.Tkn.Hidden
assert.DeepEqual(t, expected, actual) assert.DeepEqual(t, expected, actual)
expected = []token.Token{ expected = []token.Token{
@ -1453,7 +1453,7 @@ func TestMethodCallTokens(t *testing.T) {
}, },
} }
lexer.Lex(lv) lexer.Lex(lv)
actual = lv.Tkn.Tokens actual = lv.Tkn.Hidden
assert.DeepEqual(t, expected, actual) assert.DeepEqual(t, expected, actual)
expected = []token.Token{ expected = []token.Token{
@ -1463,7 +1463,7 @@ func TestMethodCallTokens(t *testing.T) {
}, },
} }
lexer.Lex(lv) lexer.Lex(lv)
actual = lv.Tkn.Tokens actual = lv.Tkn.Hidden
assert.DeepEqual(t, expected, actual) assert.DeepEqual(t, expected, actual)
expected = []token.Token{ expected = []token.Token{
@ -1473,7 +1473,7 @@ func TestMethodCallTokens(t *testing.T) {
}, },
} }
lexer.Lex(lv) lexer.Lex(lv)
actual = lv.Tkn.Tokens actual = lv.Tkn.Hidden
assert.DeepEqual(t, expected, actual) assert.DeepEqual(t, expected, actual)
expected = []token.Token{ expected = []token.Token{
@ -1483,7 +1483,7 @@ func TestMethodCallTokens(t *testing.T) {
}, },
} }
lexer.Lex(lv) lexer.Lex(lv)
actual = lv.Tkn.Tokens actual = lv.Tkn.Hidden
assert.DeepEqual(t, expected, actual) assert.DeepEqual(t, expected, actual)
expected = []token.Token{ expected = []token.Token{
@ -1493,7 +1493,7 @@ func TestMethodCallTokens(t *testing.T) {
}, },
} }
lexer.Lex(lv) lexer.Lex(lv)
actual = lv.Tkn.Tokens actual = lv.Tkn.Hidden
assert.DeepEqual(t, expected, actual) assert.DeepEqual(t, expected, actual)
} }
@ -1502,7 +1502,7 @@ func TestYieldFromTokens(t *testing.T) {
yield from $a` yield from $a`
lexer := NewLexer([]byte(src)) lexer := NewLexer([]byte(src))
lexer.WithTokens = true lexer.WithHiddenTokens = true
lv := &lval{} lv := &lval{}
expected := []token.Token{ expected := []token.Token{
@ -1516,7 +1516,7 @@ func TestYieldFromTokens(t *testing.T) {
}, },
} }
lexer.Lex(lv) lexer.Lex(lv)
actual := lv.Tkn.Tokens actual := lv.Tkn.Hidden
assert.DeepEqual(t, expected, actual) assert.DeepEqual(t, expected, actual)
expected = []token.Token{ expected = []token.Token{
@ -1526,7 +1526,7 @@ func TestYieldFromTokens(t *testing.T) {
}, },
} }
lexer.Lex(lv) lexer.Lex(lv)
actual = lv.Tkn.Tokens actual = lv.Tkn.Hidden
assert.DeepEqual(t, expected, actual) assert.DeepEqual(t, expected, actual)
} }

View File

@ -9,6 +9,6 @@ import (
type Token struct { type Token struct {
ID TokenID ID TokenID
Value []byte Value []byte
Tokens []token.Token Hidden []token.Token
Position position.Position Position position.Position
} }