Compare commits

...

7 Commits

Author SHA1 Message Date
36213d9e9c wip 2023-08-20 17:25:48 -04:00
7b1d903b00 wip 2023-08-20 16:58:02 -04:00
59e8e0a095 wip 2023-08-20 16:56:00 -04:00
dace12310a wip 2023-08-20 16:53:54 -04:00
c7185edef0 Add spec support for C tests 2023-08-20 16:38:35 -04:00
9d2b3be20b Determine output language by output file extension 2023-08-20 16:27:58 -04:00
2b515e1a7a wip 2023-08-20 16:13:18 -04:00
5 changed files with 452 additions and 239 deletions

View File

@ -1,4 +1,4 @@
#include "<TBD>.h" #include "<%= File.basename(output_file).sub(%r{\.[a-z]+$}, "") %>.h"
#include <stdbool.h> #include <stdbool.h>
#include <stdlib.h> #include <stdlib.h>
#include <string.h> #include <string.h>
@ -326,15 +326,15 @@ static lexer_state_id_t check_lexer_transition(uint32_t current_state, uint32_t
static size_t find_longest_match(<%= @grammar.prefix %>context_t * context, static size_t find_longest_match(<%= @grammar.prefix %>context_t * context,
lexer_match_info_t * out_match_info, size_t * out_unexpected_input_length) lexer_match_info_t * out_match_info, size_t * out_unexpected_input_length)
{ {
lexer_match_info_t longest_match; lexer_match_info_t longest_match = {0};
lexer_match_info_t attempt_match; lexer_match_info_t attempt_match = {0};
*out_match_info = longest_match; *out_match_info = longest_match;
uint32_t current_state = lexer_mode_table[context.mode].state_table_offset; uint32_t current_state = lexer_mode_table[context->mode].state_table_offset;
for (;;) for (;;)
{ {
size_t const input_index = context.input_index + attempt_match.length; size_t const input_index = context->input_index + attempt_match.length;
uint8_t const * input = &context.input[input_index]; uint8_t const * input = &context->input[input_index];
size_t input_length = context.input_length - input_index; size_t input_length = context->input_length - input_index;
<%= @grammar.prefix %>code_point_t code_point; <%= @grammar.prefix %>code_point_t code_point;
uint8_t code_point_length; uint8_t code_point_length;
size_t result = <%= @grammar.prefix %>decode_code_point(input, input_length, &code_point, &code_point_length); size_t result = <%= @grammar.prefix %>decode_code_point(input, input_length, &code_point, &code_point_length);
@ -427,8 +427,8 @@ static size_t find_longest_match(<%= @grammar.prefix %>context_t * context,
*/ */
static size_t attempt_lex_token(<%= @grammar.prefix %>context_t * context, <%= @grammar.prefix %>token_info_t * out_token_info) static size_t attempt_lex_token(<%= @grammar.prefix %>context_t * context, <%= @grammar.prefix %>token_info_t * out_token_info)
{ {
<%= @grammar.prefix %>token_info_t token_info; <%= @grammar.prefix %>token_info_t token_info = {0};
token_info.position = context.text_position; token_info.position = context->text_position;
token_info.token = INVALID_TOKEN_ID; token_info.token = INVALID_TOKEN_ID;
*out_token_info = token_info; // TODO: remove *out_token_info = token_info; // TODO: remove
lexer_match_info_t match_info; lexer_match_info_t match_info;
@ -437,12 +437,12 @@ static size_t attempt_lex_token(<%= @grammar.prefix %>context_t * context, <%= @
switch (result) switch (result)
{ {
case P_SUCCESS: case P_SUCCESS:
<%= @grammar.prefix %>token_t token_to_accept = match_info.accepting_state.token; <%= @grammar.prefix %>token_t token_to_accept = match_info.accepting_state->token;
if (match_info.accepting_state.code_id != INVALID_USER_CODE_ID) if (match_info.accepting_state->code_id != INVALID_USER_CODE_ID)
{ {
uint8_t const * match = &context.input[context.input_index]; uint8_t const * match = &context->input[context->input_index];
<%= @grammar.prefix %>token_t user_code_token = lexer_user_code(context, <%= @grammar.prefix %>token_t user_code_token = lexer_user_code(context,
match_info.accepting_state.code_id, match, match_info.length, &token_info); match_info.accepting_state->code_id, match, match_info.length, &token_info);
/* An invalid token returned from lexer_user_code() means that the /* An invalid token returned from lexer_user_code() means that the
* user code did not explicitly return a token. So only override * user code did not explicitly return a token. So only override
* the token to return if the user code does explicitly return a * the token to return if the user code does explicitly return a
@ -454,15 +454,15 @@ static size_t attempt_lex_token(<%= @grammar.prefix %>context_t * context, <%= @
} }
/* Update the input position tracking. */ /* Update the input position tracking. */
context.input_index += match_info.length; context->input_index += match_info.length;
context.text_position.row += match_info.delta_position.row; context->text_position.row += match_info.delta_position.row;
if (match_info.delta_position.row != 0u) if (match_info.delta_position.row != 0u)
{ {
context.text_position.col = match_info.delta_position.col; context->text_position.col = match_info.delta_position.col;
} }
else else
{ {
context.text_position.col += match_info.delta_position.col; context->text_position.col += match_info.delta_position.col;
} }
if (token_to_accept == INVALID_TOKEN_ID) if (token_to_accept == INVALID_TOKEN_ID)
@ -481,15 +481,15 @@ static size_t attempt_lex_token(<%= @grammar.prefix %>context_t * context, <%= @
case P_DECODE_ERROR: case P_DECODE_ERROR:
/* Update the input position tracking. */ /* Update the input position tracking. */
context.input_index += match_info.length; context->input_index += match_info.length;
context.text_position.row += match_info.delta_position.row; context->text_position.row += match_info.delta_position.row;
if (match_info.delta_position.row != 0u) if (match_info.delta_position.row != 0u)
{ {
context.text_position.col = match_info.delta_position.col; context->text_position.col = match_info.delta_position.col;
} }
else else
{ {
context.text_position.col += match_info.delta_position.col; context->text_position.col += match_info.delta_position.col;
} }
return result; return result;
@ -531,7 +531,7 @@ size_t <%= @grammar.prefix %>lex(<%= @grammar.prefix %>context_t * context, <%=
*************************************************************************/ *************************************************************************/
/** Reduce ID type. */ /** Reduce ID type. */
alias reduce_id_t = <%= get_type_for(@parser.reduce_table.size) %>; typedef <%= get_type_for(@parser.reduce_table.size) %> reduce_id_t;
/** /**
* A symbol ID can hold either a token ID or a rule set ID. * A symbol ID can hold either a token ID or a rule set ID.
@ -539,16 +539,16 @@ alias reduce_id_t = <%= get_type_for(@parser.reduce_table.size) %>;
* Token IDs and rule set IDs share the same namespace, with rule set IDs * Token IDs and rule set IDs share the same namespace, with rule set IDs
* beginning after token IDs end. * beginning after token IDs end.
*/ */
alias symbol_id_t = <%= get_type_for(@parser.rule_sets.map(&:last).map(&:id).max) %>; typedef <%= get_type_for(@parser.rule_sets.map(&:last).map(&:id).max) %> symbol_id_t;
/** Parser state ID type. */ /** Parser state ID type. */
alias parser_state_id_t = <%= get_type_for(@parser.state_table.size) %>; typedef <%= get_type_for(@parser.state_table.size) %> parser_state_id_t;
/** Parser rule ID type. */ /** Parser rule ID type. */
alias rule_id_t = <%= get_type_for(@grammar.rules.size) %>; typedef <%= get_type_for(@grammar.rules.size) %> rule_id_t;
/** Parser shift ID type. */ /** Parser shift ID type. */
alias shift_id_t = <%= get_type_for(@parser.shift_table.size) %>; typedef <%= get_type_for(@parser.shift_table.size) %> shift_id_t;
/** Shift table entry. */ /** Shift table entry. */
typedef struct typedef struct
@ -622,25 +622,25 @@ typedef struct
} state_value_t; } state_value_t;
/** Parser shift table. */ /** Parser shift table. */
static immutable shift_t[] parser_shift_table = [ static const shift_t parser_shift_table[] = {
<% @parser.shift_table.each do |shift| %> <% @parser.shift_table.each do |shift| %>
shift_t(<%= shift[:symbol_id] %>u, <%= shift[:state_id] %>u), {<%= shift[:symbol_id] %>u, <%= shift[:state_id] %>u},
<% end %> <% end %>
]; };
/** Parser reduce table. */ /** Parser reduce table. */
static immutable reduce_t[] parser_reduce_table = [ static const reduce_t parser_reduce_table[] = {
<% @parser.reduce_table.each do |reduce| %> <% @parser.reduce_table.each do |reduce| %>
reduce_t(<%= reduce[:token_id] %>u, <%= reduce[:rule_id] %>u, <%= reduce[:rule_set_id] %>u, <%= reduce[:n_states] %>u), {<%= reduce[:token_id] %>u, <%= reduce[:rule_id] %>u, <%= reduce[:rule_set_id] %>u, <%= reduce[:n_states] %>u},
<% end %> <% end %>
]; };
/** Parser state table. */ /** Parser state table. */
static immutable parser_state_t[] parser_state_table = [ static const parser_state_t parser_state_table[] = {
<% @parser.state_table.each do |state| %> <% @parser.state_table.each do |state| %>
parser_state_t(<%= state[:shift_index] %>u, <%= state[:n_shifts] %>u, <%= state[:reduce_index] %>u, <%= state[:n_reduces] %>u), {<%= state[:shift_index] %>u, <%= state[:n_shifts] %>u, <%= state[:reduce_index] %>u, <%= state[:n_reduces] %>u},
<% end %> <% end %>
]; };
/* state_values stack functionality */ /* state_values stack functionality */
@ -658,7 +658,7 @@ typedef struct
* @param stack * @param stack
* state_values stack structure. * state_values stack structure.
*/ */
void state_values_stack_init(state_values_stack_t * stack) static void state_values_stack_init(state_values_stack_t * stack)
{ {
const size_t initial_capacity = 10u; const size_t initial_capacity = 10u;
stack->length = 0u; stack->length = 0u;
@ -666,13 +666,35 @@ void state_values_stack_init(state_values_stack_t * stack)
stack->entries = (state_value_t *)malloc(initial_capacity * sizeof(state_value_t)); stack->entries = (state_value_t *)malloc(initial_capacity * sizeof(state_value_t));
} }
/**
* Index a state_values stack.
*
* @param stack
* state_values stack structure.
* @param index
* Index to the stack.
*
* @return Pointer to the state value structure at the given index.
*/
static state_value_t * state_values_stack_index(state_values_stack_t * stack, int index)
{
if (index >= 0)
{
return &stack->entries[index];
}
else
{
return &stack->entries[stack->length - (size_t)(unsigned int)(-index)];
}
}
/** /**
* Push a new state_value to the state_values stack. * Push a new state_value to the state_values stack.
* *
* @param stack * @param stack
* state_values stack structure. * state_values stack structure.
*/ */
void state_values_stack_push(state_values_stack_t * stack) static void state_values_stack_push(state_values_stack_t * stack)
{ {
size_t const current_capacity = stack->capacity; size_t const current_capacity = stack->capacity;
size_t const current_length = stack->length; size_t const current_length = stack->length;
@ -680,7 +702,7 @@ void state_values_stack_push(state_values_stack_t * stack)
{ {
size_t const new_capacity = current_capacity * 2u; size_t const new_capacity = current_capacity * 2u;
state_value_t * new_entries = malloc(new_capacity * sizeof(state_value_t)); state_value_t * new_entries = malloc(new_capacity * sizeof(state_value_t));
memcpy(new_entries, stack->entries, current_length * sizeof(state_value_t); memcpy(new_entries, stack->entries, current_length * sizeof(state_value_t));
free(stack->entries); free(stack->entries);
stack->capacity = new_capacity; stack->capacity = new_capacity;
stack->entries = new_entries; stack->entries = new_entries;
@ -689,6 +711,30 @@ void state_values_stack_push(state_values_stack_t * stack)
stack->length = current_length + 1u; stack->length = current_length + 1u;
} }
/**
* Pop entries from a state_values stack.
*
* @param stack
* state_values stack structure.
* @param n
* Number of states to pop.
*/
static void state_values_stack_pop(state_values_stack_t * stack, size_t n)
{
stack->length -= n;
}
/**
* Free memory for a state_values stack structure.
*
* @param stack
* state_values stack structure.
*/
static void state_values_stack_free(state_values_stack_t * stack)
{
free(stack->entries);
}
/** /**
* Execute user code associated with a parser rule. * Execute user code associated with a parser rule.
* *
@ -696,9 +742,9 @@ void state_values_stack_push(state_values_stack_t * stack)
* *
* @return Parse value. * @return Parse value.
*/ */
static <%= @grammar.prefix %>value_t parser_user_code(uint32_t rule, state_value_t[] statevalues, uint32_t n_states) static <%= @grammar.prefix %>value_t parser_user_code(uint32_t rule, state_values_stack_t * statevalues, uint32_t n_states)
{ {
<%= @grammar.prefix %>value_t _pvalue; <%= @grammar.prefix %>value_t _pvalue = {0};
switch (rule) switch (rule)
{ {
@ -775,7 +821,7 @@ static size_t check_reduce(size_t state_id, <%= @grammar.prefix %>token_t token)
* can be accessed with <%= @grammar.prefix %>result(). * can be accessed with <%= @grammar.prefix %>result().
* @retval P_UNEXPECTED_TOKEN * @retval P_UNEXPECTED_TOKEN
* An unexpected token was encountered that does not match any grammar rule. * An unexpected token was encountered that does not match any grammar rule.
* The value context.token holds the unexpected token. * The value context->token holds the unexpected token.
* @reval P_DECODE_ERROR * @reval P_DECODE_ERROR
* The decoder encountered invalid text encoding. * The decoder encountered invalid text encoding.
* @reval P_UNEXPECTED_INPUT * @reval P_UNEXPECTED_INPUT
@ -785,9 +831,12 @@ size_t <%= @grammar.prefix %>parse(<%= @grammar.prefix %>context_t * context)
{ {
<%= @grammar.prefix %>token_info_t token_info; <%= @grammar.prefix %>token_info_t token_info;
<%= @grammar.prefix %>token_t token = INVALID_TOKEN_ID; <%= @grammar.prefix %>token_t token = INVALID_TOKEN_ID;
state_value_t[] statevalues = new state_value_t[](1); state_values_stack_t statevalues;
size_t reduced_rule_set = INVALID_ID; size_t reduced_rule_set = INVALID_ID;
<%= @grammar.prefix %>value_t reduced_parser_value; <%= @grammar.prefix %>value_t reduced_parser_value;
state_values_stack_init(&statevalues);
state_values_stack_push(&statevalues);
size_t result;
for (;;) for (;;)
{ {
if (token == INVALID_TOKEN_ID) if (token == INVALID_TOKEN_ID)
@ -795,53 +844,56 @@ size_t <%= @grammar.prefix %>parse(<%= @grammar.prefix %>context_t * context)
size_t lexer_result = <%= @grammar.prefix %>lex(context, &token_info); size_t lexer_result = <%= @grammar.prefix %>lex(context, &token_info);
if (lexer_result != P_SUCCESS) if (lexer_result != P_SUCCESS)
{ {
return lexer_result; result = lexer_result;
break;
} }
token = token_info.token; token = token_info.token;
} }
size_t shift_state = INVALID_ID; size_t shift_state = INVALID_ID;
if (reduced_rule_set != INVALID_ID) if (reduced_rule_set != INVALID_ID)
{ {
shift_state = check_shift(statevalues[$-1].state_id, reduced_rule_set); shift_state = check_shift(state_values_stack_index(&statevalues, -1)->state_id, reduced_rule_set);
} }
if (shift_state == INVALID_ID) if (shift_state == INVALID_ID)
{ {
shift_state = check_shift(statevalues[$-1].state_id, token); shift_state = check_shift(state_values_stack_index(&statevalues, -1)->state_id, token);
if ((shift_state != INVALID_ID) && (token == TOKEN___EOF)) if ((shift_state != INVALID_ID) && (token == TOKEN___EOF))
{ {
/* Successful parse. */ /* Successful parse. */
context.parse_result = statevalues[$-1].pvalue; context->parse_result = state_values_stack_index(&statevalues, -1)->pvalue;
return P_SUCCESS; result = P_SUCCESS;
break;
} }
} }
if (shift_state != INVALID_ID) if (shift_state != INVALID_ID)
{ {
/* We have something to shift. */ /* We have something to shift. */
statevalues ~= state_value_t(shift_state); state_values_stack_push(&statevalues);
state_values_stack_index(&statevalues, -1)->state_id = shift_state;
if (reduced_rule_set == INVALID_ID) if (reduced_rule_set == INVALID_ID)
{ {
/* We shifted a token, mark it consumed. */ /* We shifted a token, mark it consumed. */
token = INVALID_TOKEN_ID; token = INVALID_TOKEN_ID;
statevalues[$-1].pvalue = token_info.pvalue; state_values_stack_index(&statevalues, -1)->pvalue = token_info.pvalue;
} }
else else
{ {
/* We shifted a RuleSet. */ /* We shifted a RuleSet. */
statevalues[$-1].pvalue = reduced_parser_value; state_values_stack_index(&statevalues, -1)->pvalue = reduced_parser_value;
<%= @grammar.prefix %>value_t new_parse_result; <%= @grammar.prefix %>value_t new_parse_result = {0};
reduced_parser_value = new_parse_result; reduced_parser_value = new_parse_result;
reduced_rule_set = INVALID_ID; reduced_rule_set = INVALID_ID;
} }
continue; continue;
} }
size_t reduce_index = check_reduce(statevalues[$-1].state_id, token); size_t reduce_index = check_reduce(state_values_stack_index(&statevalues, -1)->state_id, token);
if (reduce_index != INVALID_ID) if (reduce_index != INVALID_ID)
{ {
/* We have something to reduce. */ /* We have something to reduce. */
reduced_parser_value = parser_user_code(parser_reduce_table[reduce_index].rule, statevalues, parser_reduce_table[reduce_index].n_states); reduced_parser_value = parser_user_code(parser_reduce_table[reduce_index].rule, &statevalues, parser_reduce_table[reduce_index].n_states);
reduced_rule_set = parser_reduce_table[reduce_index].rule_set; reduced_rule_set = parser_reduce_table[reduce_index].rule_set;
statevalues.length -= parser_reduce_table[reduce_index].n_states; state_values_stack_pop(&statevalues, parser_reduce_table[reduce_index].n_states);
continue; continue;
} }
@ -850,10 +902,13 @@ size_t <%= @grammar.prefix %>parse(<%= @grammar.prefix %>context_t * context)
* the context text position to point to the token rather than the text * the context text position to point to the token rather than the text
* after it, so that if the caller wants to report the error position, * after it, so that if the caller wants to report the error position,
* it will point to the correct position of the unexpected token. */ * it will point to the correct position of the unexpected token. */
context.text_position = token_info.position; context->text_position = token_info.position;
context.token = token; context->token = token;
return P_UNEXPECTED_TOKEN; result = P_UNEXPECTED_TOKEN;
break;
} }
state_values_stack_free(&statevalues);
return result;
} }
/** /**
@ -866,7 +921,7 @@ size_t <%= @grammar.prefix %>parse(<%= @grammar.prefix %>context_t * context)
*/ */
<%= start_rule_type[1] %> <%= @grammar.prefix %>result(<%= @grammar.prefix %>context_t * context) <%= start_rule_type[1] %> <%= @grammar.prefix %>result(<%= @grammar.prefix %>context_t * context)
{ {
return context.parse_result.v_<%= start_rule_type[0] %>; return context->parse_result.v_<%= start_rule_type[0] %>;
} }
/** /**
@ -879,5 +934,5 @@ size_t <%= @grammar.prefix %>parse(<%= @grammar.prefix %>context_t * context)
*/ */
<%= @grammar.prefix %>position_t <%= @grammar.prefix %>position(<%= @grammar.prefix %>context_t * context) <%= @grammar.prefix %>position_t <%= @grammar.prefix %>position(<%= @grammar.prefix %>context_t * context)
{ {
return context.text_position; return context->text_position;
} }

View File

@ -8,6 +8,7 @@
#define PROPANE_PARSER_H #define PROPANE_PARSER_H
#include <stdint.h> #include <stdint.h>
#include <stddef.h>
/************************************************************************** /**************************************************************************
* Public types * Public types
@ -87,6 +88,9 @@ typedef struct
/** Input text. */ /** Input text. */
uint8_t const * input; uint8_t const * input;
/** Input text length. */
size_t input_length;
/** Input text index (byte offset). */ /** Input text index (byte offset). */
size_t input_index; size_t input_index;
@ -105,4 +109,17 @@ typedef struct
<%= @grammar.prefix %>token_t token; <%= @grammar.prefix %>token_t token;
} <%= @grammar.prefix %>context_t; } <%= @grammar.prefix %>context_t;
void <%= @grammar.prefix %>context_init(<%= @grammar.prefix %>context_t * context, uint8_t const * input, size_t input_length);
size_t <%= @grammar.prefix %>decode_code_point(uint8_t const * input, size_t input_length,
<%= @grammar.prefix %>code_point_t * out_code_point, uint8_t * out_code_point_length);
size_t <%= @grammar.prefix %>lex(<%= @grammar.prefix %>context_t * context, <%= @grammar.prefix %>token_info_t * out_token_info);
size_t <%= @grammar.prefix %>parse(<%= @grammar.prefix %>context_t * context);
<%= start_rule_type[1] %> <%= @grammar.prefix %>result(<%= @grammar.prefix %>context_t * context);
<%= @grammar.prefix %>position_t <%= @grammar.prefix %>position(<%= @grammar.prefix %>context_t * context);
#endif #endif

View File

@ -11,15 +11,28 @@ class Propane
@log = StringIO.new @log = StringIO.new
end end
@classname = @grammar.classname || File.basename(output_file).sub(%r{[^a-zA-Z0-9].*}, "").capitalize @classname = @grammar.classname || File.basename(output_file).sub(%r{[^a-zA-Z0-9].*}, "").capitalize
@language =
if output_file =~ /\.([a-z]+)$/
$1
else
"d"
end
process_grammar! process_grammar!
end end
def generate def generate
erb = ERB.new(File.read(File.join(File.dirname(File.expand_path(__FILE__)), "../../assets/parser.d.erb")), trim_mode: "<>") extensions = [@language]
if @language == "c"
extensions += %w[h]
end
extensions.each do |extension|
erb = ERB.new(File.read(File.join(File.dirname(File.expand_path(__FILE__)), "../../assets/parser.#{extension}.erb")), trim_mode: "<>")
output_file = @output_file.sub(%r{\.[a-z]+$}, ".#{extension}")
result = erb.result(binding.clone) result = erb.result(binding.clone)
File.open(@output_file, "wb") do |fh| File.open(output_file, "wb") do |fh|
fh.write(result) fh.write(result)
end end
end
@log.close @log.close
end end
@ -191,8 +204,13 @@ class Propane
end end
code = code.gsub(/\$(\d+)/) do |match| code = code.gsub(/\$(\d+)/) do |match|
index = $1.to_i index = $1.to_i
case @language
when "c"
"state_value_stack_index(statevalues, -1 - (int)n_states + #{index})->pvalue.v_#{rule.components[index - 1].ptypename}"
when "d"
"statevalues[$-1-n_states+#{index}].pvalue.v_#{rule.components[index - 1].ptypename}" "statevalues[$-1-n_states+#{index}].pvalue.v_#{rule.components[index - 1].ptypename}"
end end
end
else else
code = code.gsub(/\$\$/) do |match| code = code.gsub(/\$\$/) do |match|
"out_token_info.pvalue.v_#{pattern.ptypename}" "out_token_info.pvalue.v_#{pattern.ptypename}"
@ -203,9 +221,14 @@ class Propane
unless mode_id unless mode_id
raise Error.new("Lexer mode '#{mode_name}' not found") raise Error.new("Lexer mode '#{mode_name}' not found")
end end
case @language
when "c"
"context->mode = #{mode_id}u"
when "d"
"context.mode = #{mode_id}u" "context.mode = #{mode_id}u"
end end
end end
end
code code
end end
@ -229,13 +252,28 @@ class Propane
# Type. # Type.
def get_type_for(max) def get_type_for(max)
if max <= 0xFF if max <= 0xFF
case @language
when "c"
"uint8_t"
when "d"
"ubyte" "ubyte"
end
elsif max <= 0xFFFF elsif max <= 0xFFFF
case @language
when "c"
"uint16_t"
when "d"
"ushort" "ushort"
end
else
case @language
when "c"
"uint32_t"
else else
"uint" "uint"
end end
end end
end
end end

View File

@ -11,7 +11,7 @@ describe Propane do
def build_parser(options = {}) def build_parser(options = {})
options[:name] ||= "" options[:name] ||= ""
command = %W[./propane.sh spec/run/testparser#{options[:name]}.propane spec/run/testparser#{options[:name]}.d --log spec/run/testparser#{options[:name]}.log] command = %W[./propane.sh spec/run/testparser#{options[:name]}.propane spec/run/testparser#{options[:name]}.#{options[:language]} --log spec/run/testparser#{options[:name]}.log]
if (options[:capture]) if (options[:capture])
stdout, stderr, status = Open3.capture3(*command) stdout, stderr, status = Open3.capture3(*command)
Results.new(stdout, stderr, status) Results.new(stdout, stderr, status)
@ -25,9 +25,14 @@ describe Propane do
test_files = Array(test_files) test_files = Array(test_files)
options[:parsers] ||= [""] options[:parsers] ||= [""]
parsers = options[:parsers].map do |name| parsers = options[:parsers].map do |name|
"spec/run/testparser#{name}.d" "spec/run/testparser#{name}.#{options[:language]}"
end end
case options[:language]
when "c"
result = system(*%w[gcc -Wall -o spec/run/testparser -Ispec -Ispec/run], *parsers, *test_files)
when "d"
result = system(*%w[ldc2 --unittest -of spec/run/testparser -Ispec], *parsers, *test_files, "spec/testutils.d") result = system(*%w[ldc2 --unittest -of spec/run/testparser -Ispec], *parsers, *test_files, "spec/testutils.d")
end
expect(result).to be_truthy expect(result).to be_truthy
end end
@ -69,6 +74,10 @@ describe Propane do
FileUtils.mkdir_p("spec/run") FileUtils.mkdir_p("spec/run")
end end
%w[d c].each do |language|
context "#{language.upcase} language" do
it "generates a lexer" do it "generates a lexer" do
write_grammar <<EOF write_grammar <<EOF
token int /\\d+/; token int /\\d+/;
@ -81,8 +90,8 @@ Foo -> int <<
Foo -> plus << Foo -> plus <<
>> >>
EOF EOF
build_parser build_parser(language: language)
compile("spec/test_lexer.d") compile("spec/test_lexer.#{language}", language: language)
results = run results = run
expect(results.stderr).to eq "" expect(results.stderr).to eq ""
expect(results.status).to eq 0 expect(results.status).to eq 0
@ -104,8 +113,8 @@ Start -> int <<
$$ = $1; $$ = $1;
>> >>
EOF EOF
build_parser build_parser(language: language)
compile("spec/test_lexer_unknown_character.d") compile("spec/test_lexer_unknown_character.#{language}", language: language)
results = run results = run
expect(results.stderr).to eq "" expect(results.stderr).to eq ""
expect(results.status).to eq 0 expect(results.status).to eq 0
@ -124,7 +133,7 @@ E -> B;
B -> zero; B -> zero;
B -> one; B -> one;
EOF EOF
build_parser build_parser(language: language)
end end
it "generates a parser that does basic math - user guide example" do it "generates a parser that does basic math - user guide example" do
@ -179,8 +188,8 @@ E4 -> lparen E1 rparen <<
$$ = $2; $$ = $2;
>> >>
EOF EOF
build_parser build_parser(language: language)
compile("spec/test_basic_math_grammar.d") compile("spec/test_basic_math_grammar.#{language}", language: language)
results = run results = run
expect(results.stderr).to eq "" expect(results.stderr).to eq ""
expect(results.status).to eq 0 expect(results.status).to eq 0
@ -193,7 +202,7 @@ Start -> E;
E -> one E; E -> one E;
E -> one; E -> one;
EOF EOF
build_parser build_parser(language: language)
end end
it "distinguishes between multiple identical rules with lookahead symbol" do it "distinguishes between multiple identical rules with lookahead symbol" do
@ -205,8 +214,8 @@ Start -> R2 b;
R1 -> a b; R1 -> a b;
R2 -> a b; R2 -> a b;
EOF EOF
build_parser build_parser(language: language)
compile("spec/test_parser_identical_rules_lookahead.d") compile("spec/test_parser_identical_rules_lookahead.#{language}", language: language)
results = run results = run
expect(results.status).to eq 0 expect(results.status).to eq 0
end end
@ -220,8 +229,8 @@ Start -> a R1;
Start -> b R1; Start -> b R1;
R1 -> b; R1 -> b;
EOF EOF
build_parser build_parser(language: language)
compile("spec/test_parser_rule_from_multiple_states.d") compile("spec/test_parser_rule_from_multiple_states.#{language}", language: language)
results = run results = run
expect(results.status).to eq 0 expect(results.status).to eq 0
end end
@ -239,8 +248,8 @@ Start -> Abcs def;
Abcs -> ; Abcs -> ;
Abcs -> abc Abcs; Abcs -> abc Abcs;
EOF EOF
build_parser build_parser(language: language)
compile("spec/test_user_code.d") compile("spec/test_user_code.#{language}", language: language)
results = run results = run
expect(results.status).to eq 0 expect(results.status).to eq 0
verify_lines(results.stdout, [ verify_lines(results.stdout, [
@ -263,8 +272,8 @@ token abc;
>> >>
Start -> abc; Start -> abc;
EOF EOF
build_parser build_parser(language: language)
compile("spec/test_pattern.d") compile("spec/test_pattern.#{language}", language: language)
results = run results = run
expect(results.status).to eq 0 expect(results.status).to eq 0
verify_lines(results.stdout, [ verify_lines(results.stdout, [
@ -291,8 +300,8 @@ token abc;
>> >>
Start -> abc; Start -> abc;
EOF EOF
build_parser build_parser(language: language)
compile("spec/test_return_token_from_pattern.d") compile("spec/test_return_token_from_pattern.#{language}", language: language)
results = run results = run
expect(results.status).to eq 0 expect(results.status).to eq 0
verify_lines(results.stdout, [ verify_lines(results.stdout, [
@ -324,8 +333,8 @@ string: /"/ <<
>> >>
Start -> abc string def; Start -> abc string def;
EOF EOF
build_parser build_parser(language: language)
compile("spec/test_lexer_modes.d") compile("spec/test_lexer_modes.#{language}", language: language)
results = run results = run
expect(results.status).to eq 0 expect(results.status).to eq 0
verify_lines(results.stdout, [ verify_lines(results.stdout, [
@ -355,8 +364,8 @@ B -> b <<
writeln("B!"); writeln("B!");
>> >>
EOF EOF
build_parser build_parser(language: language)
compile("spec/test_parser_rule_user_code.d") compile("spec/test_parser_rule_user_code.#{language}", language: language)
results = run results = run
expect(results.status).to eq 0 expect(results.status).to eq 0
verify_lines(results.stdout, [ verify_lines(results.stdout, [
@ -380,8 +389,8 @@ As -> As a <<
$$ = $1 + 1u; $$ = $1 + 1u;
>> >>
EOF EOF
build_parser build_parser(language: language)
compile("spec/test_parsing_lists.d") compile("spec/test_parsing_lists.#{language}", language: language)
results = run results = run
expect(results.status).to eq 0 expect(results.status).to eq 0
expect(results.stderr).to eq "" expect(results.stderr).to eq ""
@ -401,7 +410,7 @@ Start -> b E d;
E -> e; E -> e;
F -> e; F -> e;
EOF EOF
results = build_parser(capture: true) results = build_parser(capture: true, language: language)
expect(results.status).to_not eq 0 expect(results.status).to_not eq 0
expect(results.stderr).to match %r{reduce/reduce conflict.*\(E\).*\(F\)} expect(results.stderr).to match %r{reduce/reduce conflict.*\(E\).*\(F\)}
end end
@ -416,8 +425,8 @@ token id /[a-zA-Z_][a-zA-Z0-9_]*/ <<
>> >>
Start -> id; Start -> id;
EOF EOF
build_parser build_parser(language: language)
compile("spec/test_lexer_match_text.d") compile("spec/test_lexer_match_text.#{language}", language: language)
results = run results = run
expect(results.status).to eq 0 expect(results.status).to eq 0
verify_lines(results.stdout, [ verify_lines(results.stdout, [
@ -436,8 +445,8 @@ Start -> word <<
$$ = $1; $$ = $1;
>> >>
EOF EOF
build_parser build_parser(language: language)
compile("spec/test_lexer_result_value.d") compile("spec/test_lexer_result_value.#{language}", language: language)
results = run results = run
expect(results.stderr).to eq "" expect(results.stderr).to eq ""
expect(results.status).to eq 0 expect(results.status).to eq 0
@ -451,8 +460,8 @@ drop /\\s+/;
Start -> a num Start; Start -> a num Start;
Start -> a num; Start -> a num;
EOF EOF
build_parser build_parser(language: language)
compile("spec/test_error_positions.d") compile("spec/test_error_positions.#{language}", language: language)
results = run results = run
expect(results.stderr).to eq "" expect(results.stderr).to eq ""
expect(results.status).to eq 0 expect(results.status).to eq 0
@ -460,8 +469,8 @@ EOF
it "allows creating a JSON parser" do it "allows creating a JSON parser" do
write_grammar(File.read("spec/json_parser.propane")) write_grammar(File.read("spec/json_parser.propane"))
build_parser build_parser(language: language)
compile(["spec/test_parsing_json.d", "spec/json_types.d"]) compile(["spec/test_parsing_json.#{language}", "spec/json_types.#{language}"], language: language)
end end
it "allows generating multiple parsers in the same program" do it "allows generating multiple parsers in the same program" do
@ -472,17 +481,19 @@ token num /\\d+/;
drop /\\s+/; drop /\\s+/;
Start -> a num; Start -> a num;
EOF EOF
build_parser(name: "myp1") build_parser(name: "myp1", language: language)
write_grammar(<<EOF, name: "myp2") write_grammar(<<EOF, name: "myp2")
prefix myp2_; prefix myp2_;
token b; token b;
token c; token c;
Start -> b c b; Start -> b c b;
EOF EOF
build_parser(name: "myp2") build_parser(name: "myp2", language: language)
compile("spec/test_multiple_parsers.d", parsers: %w[myp1 myp2]) compile("spec/test_multiple_parsers.#{language}", parsers: %w[myp1 myp2], language: language)
results = run results = run
expect(results.stderr).to eq "" expect(results.stderr).to eq ""
expect(results.status).to eq 0 expect(results.status).to eq 0
end end
end end
end
end

92
spec/test_lexer.c Normal file
View File

@ -0,0 +1,92 @@
#include "testparser.h"
#include <assert.h>
#include <string.h>
int main()
{
size_t result;
p_code_point_t code_point;
uint8_t code_point_length;
result = p_decode_code_point((uint8_t const *)"5", 1u, &code_point, &code_point_length);
assert(result == P_SUCCESS);
assert(code_point == '5');
assert(code_point_length == 1u);
result = p_decode_code_point((uint8_t const *)"", 0u, &code_point, &code_point_length);
assert(result == P_EOF);
result = p_decode_code_point((uint8_t const *)"\xC2\xA9", 2u, &code_point, &code_point_length);
assert(result == P_SUCCESS);
assert(code_point == 0xA9u);
assert(code_point_length == 2u);
result = p_decode_code_point((uint8_t const *)"\xf0\x9f\xa7\xa1", 4u, &code_point, &code_point_length);
assert(result == P_SUCCESS);
assert(code_point == 0x1F9E1u);
assert(code_point_length == 4u);
result = p_decode_code_point((uint8_t const *)"\xf0\x9f\x27", 3u, &code_point, &code_point_length);
assert(result == P_DECODE_ERROR);
result = p_decode_code_point((uint8_t const *)"\xf0\x9f\xa7\xFF", 4u, &code_point, &code_point_length);
assert(result == P_DECODE_ERROR);
result = p_decode_code_point((uint8_t const *)"\xfe", 1u, &code_point, &code_point_length);
assert(result == P_DECODE_ERROR);
p_token_info_t token_info;
char const * input = "5 + 4 * \n677 + 567";
p_context_t context;
p_context_init(&context, (uint8_t const *)input, strlen(input));
assert(p_lex(&context, &token_info) == P_SUCCESS);
assert(token_info.position.row == 0u);
assert(token_info.position.col == 0u);
assert(token_info.length == 1u);
assert(token_info.token == TOKEN_int);
assert(p_lex(&context, &token_info) == P_SUCCESS);
assert(token_info.position.row == 0u);
assert(token_info.position.col == 2u);
assert(token_info.length == 1u);
assert(token_info.token == TOKEN_plus);
assert(p_lex(&context, &token_info) == P_SUCCESS);
assert(token_info.position.row == 0u);
assert(token_info.position.col == 4u);
assert(token_info.length == 1u);
assert(token_info.token == TOKEN_int);
assert(p_lex(&context, &token_info) == P_SUCCESS);
assert(token_info.position.row == 0u);
assert(token_info.position.col == 6u);
assert(token_info.length == 1u);
assert(token_info.token == TOKEN_times);
assert(p_lex(&context, &token_info) == P_SUCCESS);
assert(token_info.position.row == 1u);
assert(token_info.position.col == 0u);
assert(token_info.length == 3u);
assert(token_info.token == TOKEN_int);
assert(p_lex(&context, &token_info) == P_SUCCESS);
assert(token_info.position.row == 1u);
assert(token_info.position.col == 4u);
assert(token_info.length == 1u);
assert(token_info.token == TOKEN_plus);
assert(p_lex(&context, &token_info) == P_SUCCESS);
assert(token_info.position.row == 1u);
assert(token_info.position.col == 6u);
assert(token_info.length == 3u);
assert(token_info.token == TOKEN_int);
assert(p_lex(&context, &token_info) == P_SUCCESS);
assert(token_info.position.row == 1u);
assert(token_info.position.col == 9u);
assert(token_info.length == 0u);
assert(token_info.token == TOKEN___EOF);
p_context_init(&context, (uint8_t const *)"", 0u);
assert(p_lex(&context, &token_info) == P_SUCCESS);
assert(token_info.position.row == 0u);
assert(token_info.position.col == 0u);
assert(token_info.length == 0u);
assert(token_info.token == TOKEN___EOF);
return 0;
}