Remove _TOKEN_NONE and use _TOKEN_COUNT instead
This commit is contained in:
parent
ced5a56424
commit
6bd9d4a09b
@ -19,7 +19,6 @@ class <%= @classname %>
|
|||||||
_TOKEN_COUNT = <%= @grammar.tokens.size %>,
|
_TOKEN_COUNT = <%= @grammar.tokens.size %>,
|
||||||
_TOKEN_DECODE_ERROR = <%= TOKEN_DECODE_ERROR %>,
|
_TOKEN_DECODE_ERROR = <%= TOKEN_DECODE_ERROR %>,
|
||||||
_TOKEN_DROP = <%= TOKEN_DROP %>,
|
_TOKEN_DROP = <%= TOKEN_DROP %>,
|
||||||
_TOKEN_NONE = <%= TOKEN_NONE %>,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static immutable string token_names[] = [
|
static immutable string token_names[] = [
|
||||||
@ -166,7 +165,7 @@ class <%= @classname %>
|
|||||||
for (;;)
|
for (;;)
|
||||||
{
|
{
|
||||||
LexedToken lt = attempt_lex_token();
|
LexedToken lt = attempt_lex_token();
|
||||||
if ((lt.token != _TOKEN_DROP) && (lt.token != _TOKEN_NONE))
|
if (lt.token < _TOKEN_COUNT)
|
||||||
{
|
{
|
||||||
return lt;
|
return lt;
|
||||||
}
|
}
|
||||||
@ -178,7 +177,7 @@ class <%= @classname %>
|
|||||||
*
|
*
|
||||||
* @param code_id The ID of the user code block to execute.
|
* @param code_id The ID of the user code block to execute.
|
||||||
*
|
*
|
||||||
* @return Token ID to accept, or _TOKEN_NONE if the user code does
|
* @return Token ID to accept, or _TOKEN_COUNT if the user code does
|
||||||
* not explicitly return a token.
|
* not explicitly return a token.
|
||||||
*/
|
*/
|
||||||
private uint user_code(uint code_id)
|
private uint user_code(uint code_id)
|
||||||
@ -195,12 +194,12 @@ class <%= @classname %>
|
|||||||
default: break;
|
default: break;
|
||||||
}
|
}
|
||||||
|
|
||||||
return _TOKEN_NONE;
|
return _TOKEN_COUNT;
|
||||||
}
|
}
|
||||||
|
|
||||||
private LexedToken attempt_lex_token()
|
private LexedToken attempt_lex_token()
|
||||||
{
|
{
|
||||||
LexedToken lt = LexedToken(m_input_row, m_input_col, 0, _TOKEN_NONE);
|
LexedToken lt = LexedToken(m_input_row, m_input_col, 0, _TOKEN_COUNT);
|
||||||
struct MatchInfo
|
struct MatchInfo
|
||||||
{
|
{
|
||||||
size_t length;
|
size_t length;
|
||||||
@ -210,7 +209,7 @@ class <%= @classname %>
|
|||||||
uint code_id;
|
uint code_id;
|
||||||
}
|
}
|
||||||
MatchInfo longest_match_info;
|
MatchInfo longest_match_info;
|
||||||
longest_match_info.token = _TOKEN_NONE;
|
longest_match_info.token = _TOKEN_COUNT;
|
||||||
MatchInfo attempt_match_info;
|
MatchInfo attempt_match_info;
|
||||||
uint current_state;
|
uint current_state;
|
||||||
for (;;)
|
for (;;)
|
||||||
@ -239,7 +238,7 @@ class <%= @classname %>
|
|||||||
attempt_match_info.delta_col++;
|
attempt_match_info.delta_col++;
|
||||||
}
|
}
|
||||||
current_state = dest;
|
current_state = dest;
|
||||||
if ((states[current_state].token != _TOKEN_NONE) ||
|
if ((states[current_state].token != _TOKEN_COUNT) ||
|
||||||
(states[current_state].code_id != 0xFFFF_FFFFu))
|
(states[current_state].code_id != 0xFFFF_FFFFu))
|
||||||
{
|
{
|
||||||
attempt_match_info.token = states[current_state].token;
|
attempt_match_info.token = states[current_state].token;
|
||||||
@ -260,17 +259,17 @@ class <%= @classname %>
|
|||||||
if (longest_match_info.code_id != 0xFFFF_FFFFu)
|
if (longest_match_info.code_id != 0xFFFF_FFFFu)
|
||||||
{
|
{
|
||||||
uint user_code_token = user_code(longest_match_info.code_id);
|
uint user_code_token = user_code(longest_match_info.code_id);
|
||||||
/* A return of _TOKEN_NONE from user_code() means
|
/* A return of _TOKEN_COUNT from user_code() means
|
||||||
* that the user code did not explicitly return a
|
* that the user code did not explicitly return a
|
||||||
* token. So only override the token to return if the
|
* token. So only override the token to return if the
|
||||||
* user code does explicitly return a token. */
|
* user code does explicitly return a token. */
|
||||||
if (user_code_token != _TOKEN_NONE)
|
if (user_code_token != _TOKEN_COUNT)
|
||||||
{
|
{
|
||||||
token_to_accept = user_code_token;
|
token_to_accept = user_code_token;
|
||||||
}
|
}
|
||||||
pattern_accepted = true;
|
pattern_accepted = true;
|
||||||
}
|
}
|
||||||
if (pattern_accepted || (token_to_accept != _TOKEN_NONE))
|
if (pattern_accepted || (token_to_accept != _TOKEN_COUNT))
|
||||||
{
|
{
|
||||||
/* Update the input position tracking. */
|
/* Update the input position tracking. */
|
||||||
m_input_position += longest_match_info.length;
|
m_input_position += longest_match_info.length;
|
||||||
@ -360,12 +359,12 @@ class <%= @classname %>
|
|||||||
bool parse()
|
bool parse()
|
||||||
{
|
{
|
||||||
Lexer.LexedToken lexed_token;
|
Lexer.LexedToken lexed_token;
|
||||||
uint token = _TOKEN_NONE;
|
uint token = _TOKEN_COUNT;
|
||||||
uint[] states = new uint[](1);
|
uint[] states = new uint[](1);
|
||||||
uint reduced_rule_set = 0xFFFFFFFFu;
|
uint reduced_rule_set = 0xFFFFFFFFu;
|
||||||
for (;;)
|
for (;;)
|
||||||
{
|
{
|
||||||
if (token == _TOKEN_NONE)
|
if (token == _TOKEN_COUNT)
|
||||||
{
|
{
|
||||||
lexed_token = m_lexer.lex_token();
|
lexed_token = m_lexer.lex_token();
|
||||||
token = lexed_token.token;
|
token = lexed_token.token;
|
||||||
@ -389,7 +388,7 @@ class <%= @classname %>
|
|||||||
states ~= shift_state;
|
states ~= shift_state;
|
||||||
if (reduced_rule_set == 0xFFFFFFFFu)
|
if (reduced_rule_set == 0xFFFFFFFFu)
|
||||||
{
|
{
|
||||||
token = _TOKEN_NONE;
|
token = _TOKEN_COUNT;
|
||||||
}
|
}
|
||||||
else
|
else
|
||||||
{
|
{
|
||||||
@ -449,7 +448,7 @@ class <%= @classname %>
|
|||||||
for (uint i = start; i < end; i++)
|
for (uint i = start; i < end; i++)
|
||||||
{
|
{
|
||||||
if ((reduces[i].token == token) ||
|
if ((reduces[i].token == token) ||
|
||||||
(reduces[i].token == _TOKEN_NONE))
|
(reduces[i].token == _TOKEN_COUNT))
|
||||||
{
|
{
|
||||||
// write("Reducing rule ", reduces[i].rule, ", rule set ", reduces[i].rule_set, " lookahead ");
|
// write("Reducing rule ", reduces[i].rule, ", rule set ", reduces[i].rule_set, " lookahead ");
|
||||||
// if (token < _TOKEN_COUNT)
|
// if (token < _TOKEN_COUNT)
|
||||||
|
@ -31,9 +31,6 @@ class Propane
|
|||||||
# Token ID for a "dropped" token.
|
# Token ID for a "dropped" token.
|
||||||
TOKEN_DROP = 0xFFFFFFFE
|
TOKEN_DROP = 0xFFFFFFFE
|
||||||
|
|
||||||
# Invalid token ID.
|
|
||||||
TOKEN_NONE = 0xFFFFFFFF
|
|
||||||
|
|
||||||
class Error < RuntimeError
|
class Error < RuntimeError
|
||||||
end
|
end
|
||||||
|
|
||||||
|
@ -81,7 +81,7 @@ class Propane
|
|||||||
end
|
end
|
||||||
determine_possibly_empty_rulesets!(rule_sets)
|
determine_possibly_empty_rulesets!(rule_sets)
|
||||||
# Generate the lexer.
|
# Generate the lexer.
|
||||||
@lexer = Lexer.new(@grammar.patterns)
|
@lexer = Lexer.new(@grammar)
|
||||||
# Generate the parser.
|
# Generate the parser.
|
||||||
@parser = Parser.new(@grammar, rule_sets, @log)
|
@parser = Parser.new(@grammar, rule_sets, @log)
|
||||||
end
|
end
|
||||||
|
@ -5,8 +5,9 @@ class Propane
|
|||||||
# Lexer DFA.
|
# Lexer DFA.
|
||||||
attr_accessor :dfa
|
attr_accessor :dfa
|
||||||
|
|
||||||
def initialize(patterns)
|
def initialize(grammar)
|
||||||
@dfa = DFA.new(patterns)
|
@grammar = grammar
|
||||||
|
@dfa = DFA.new(grammar.patterns)
|
||||||
end
|
end
|
||||||
|
|
||||||
def build_tables
|
def build_tables
|
||||||
@ -16,13 +17,13 @@ class Propane
|
|||||||
states.each do |state, id|
|
states.each do |state, id|
|
||||||
token =
|
token =
|
||||||
if state.accepts.nil?
|
if state.accepts.nil?
|
||||||
TOKEN_NONE
|
@grammar.tokens.size
|
||||||
elsif state.accepts.drop?
|
elsif state.accepts.drop?
|
||||||
TOKEN_DROP
|
TOKEN_DROP
|
||||||
elsif state.accepts.token
|
elsif state.accepts.token
|
||||||
state.accepts.token.id
|
state.accepts.token.id
|
||||||
else
|
else
|
||||||
TOKEN_NONE
|
@grammar.tokens.size
|
||||||
end
|
end
|
||||||
code_id =
|
code_id =
|
||||||
if state.accepts && state.accepts.code_id
|
if state.accepts && state.accepts.code_id
|
||||||
|
@ -55,7 +55,7 @@ class Propane
|
|||||||
reduce_entries =
|
reduce_entries =
|
||||||
case ra = item_set.reduce_actions
|
case ra = item_set.reduce_actions
|
||||||
when Rule
|
when Rule
|
||||||
[{token_id: TOKEN_NONE, rule_id: ra.id,
|
[{token_id: @grammar.tokens.size, rule_id: ra.id,
|
||||||
rule_set_id: ra.rule_set.id, n_states: ra.components.size}]
|
rule_set_id: ra.rule_set.id, n_states: ra.components.size}]
|
||||||
when Hash
|
when Hash
|
||||||
ra.map do |token, rule|
|
ra.map do |token, rule|
|
||||||
|
Loading…
x
Reference in New Issue
Block a user