Started working on ast
This commit is contained in:
parent
9aee1f2272
commit
1d4dc219db
1344 changed files with 118659 additions and 107610 deletions
|
|
@ -14,52 +14,44 @@
|
|||
|
||||
bool lex_normal_s285(t_lexer *lexer, t_lexer_state *s)
|
||||
{
|
||||
lex_accept_token(anon_sym_EQ, lexer, s);
|
||||
if (s->lookahead == '\\')
|
||||
return (lex_advance(234, lexer, s));
|
||||
if ((!s->eof && \
|
||||
set_contains(sym__comment_word_character_set_1(), 10, \
|
||||
s->lookahead)))
|
||||
return (lex_advance(533, lexer, s));
|
||||
lex_accept_token(anon_sym_BANG, lexer, s);
|
||||
return (lex_end_state(lexer, s));
|
||||
}
|
||||
|
||||
bool lex_normal_s286(t_lexer *lexer, t_lexer_state *s)
|
||||
{
|
||||
lex_accept_token(anon_sym_LT, lexer, s);
|
||||
if (s->lookahead == '&')
|
||||
return (lex_advance(297, lexer, s));
|
||||
lex_accept_token(anon_sym_BANG, lexer, s);
|
||||
if (s->lookahead == '\\')
|
||||
return (lex_advance(238, lexer, s));
|
||||
if ((!s->eof && \
|
||||
set_contains(sym__comment_word_character_set_1(), 10, \
|
||||
s->lookahead)))
|
||||
return (lex_advance(528, lexer, s));
|
||||
return (lex_end_state(lexer, s));
|
||||
}
|
||||
|
||||
bool lex_normal_s287(t_lexer *lexer, t_lexer_state *s)
|
||||
{
|
||||
lex_accept_token(anon_sym_LT, lexer, s);
|
||||
if (s->lookahead == '&')
|
||||
return (lex_advance(297, lexer, s));
|
||||
if (s->lookahead == '<')
|
||||
return (lex_advance(303, lexer, s));
|
||||
if (s->lookahead == '=')
|
||||
return (lex_advance(358, lexer, s));
|
||||
lex_accept_token(anon_sym_EQ, lexer, s);
|
||||
return (lex_end_state(lexer, s));
|
||||
}
|
||||
|
||||
bool lex_normal_s288(t_lexer *lexer, t_lexer_state *s)
|
||||
{
|
||||
lex_accept_token(anon_sym_LT, lexer, s);
|
||||
if (s->lookahead == '&')
|
||||
return (lex_advance(297, lexer, s));
|
||||
if (s->lookahead == '<')
|
||||
return (lex_advance(302, lexer, s));
|
||||
lex_accept_token(anon_sym_EQ, lexer, s);
|
||||
if (s->lookahead == '=')
|
||||
return (lex_advance(360, lexer, s));
|
||||
return (lex_end_state(lexer, s));
|
||||
}
|
||||
|
||||
bool lex_normal_s289(t_lexer *lexer, t_lexer_state *s)
|
||||
{
|
||||
lex_accept_token(anon_sym_LT, lexer, s);
|
||||
if (s->lookahead == '<')
|
||||
return (lex_advance(304, lexer, s));
|
||||
if (s->lookahead == '=')
|
||||
return (lex_advance(358, lexer, s));
|
||||
lex_accept_token(anon_sym_EQ, lexer, s);
|
||||
if (s->lookahead == '\\')
|
||||
return (lex_advance(238, lexer, s));
|
||||
if ((!s->eof && \
|
||||
set_contains(sym__comment_word_character_set_1(), 10, \
|
||||
s->lookahead)))
|
||||
return (lex_advance(528, lexer, s));
|
||||
return (lex_end_state(lexer, s));
|
||||
}
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue