Write parser log file

Fix bug of skipping rule set IDs.
Remove unneeded out_sets from ItemSet class.
This commit is contained in:
Josh Holtrop 2022-06-26 11:06:55 -04:00
parent 2fbe13e071
commit 30f4cfcc99
7 changed files with 95 additions and 43 deletions

View File

@ -1,5 +1,6 @@
require "erb" require "erb"
require "set" require "set"
require "stringio"
require_relative "propane/cli" require_relative "propane/cli"
require_relative "propane/code_point_range" require_relative "propane/code_point_range"
require_relative "propane/fa" require_relative "propane/fa"
@ -18,6 +19,7 @@ require_relative "propane/regex/unit"
require_relative "propane/rule_set" require_relative "propane/rule_set"
require_relative "propane/rule" require_relative "propane/rule"
require_relative "propane/token" require_relative "propane/token"
require_relative "propane/util"
require_relative "propane/version" require_relative "propane/version"
class Propane class Propane

View File

@ -5,7 +5,11 @@ class Propane
def initialize(grammar, output_file, log_file) def initialize(grammar, output_file, log_file)
@grammar = grammar @grammar = grammar
@output_file = output_file @output_file = output_file
@log_file = log_file if log_file
@log = File.open(log_file, "wb")
else
@log = StringIO.new
end
@classname = @grammar.classname || File.basename(output_file).sub(%r{[^a-zA-Z0-9].*}, "").capitalize @classname = @grammar.classname || File.basename(output_file).sub(%r{[^a-zA-Z0-9].*}, "").capitalize
process_grammar! process_grammar!
end end
@ -16,6 +20,7 @@ class Propane
File.open(@output_file, "wb") do |fh| File.open(@output_file, "wb") do |fh|
fh.write(result) fh.write(result)
end end
@log.close
end end
private private
@ -34,8 +39,10 @@ class Propane
raise Error.new("Rule name collides with token name #{rule.name.inspect}") raise Error.new("Rule name collides with token name #{rule.name.inspect}")
end end
@_rule_set_id ||= @grammar.tokens.size @_rule_set_id ||= @grammar.tokens.size
rule_sets[rule.name] ||= RuleSet.new(rule.name, @_rule_set_id) unless rule_sets[rule.name]
rule_sets[rule.name] = RuleSet.new(rule.name, @_rule_set_id)
@_rule_set_id += 1 @_rule_set_id += 1
end
rule.rule_set = rule_sets[rule.name] rule.rule_set = rule_sets[rule.name]
rule_sets[rule.name] << rule rule_sets[rule.name] << rule
end end
@ -54,13 +61,8 @@ class Propane
end end
end end
determine_possibly_empty_rulesets!(rule_sets) determine_possibly_empty_rulesets!(rule_sets)
puts "Start token set"
rule_sets.each do |rule_set_name, rule_set|
puts "RuleSet #{rule_set_name}:"
puts " " + rule_set.start_token_set.map(&:name).join(", ")
end
@lexer = Lexer.new(@grammar.tokens, @grammar.drop_tokens) @lexer = Lexer.new(@grammar.tokens, @grammar.drop_tokens)
@parser = Parser.new(rule_sets["Start"]) @parser = Parser.new(@grammar, rule_sets, rule_sets["Start"], @log)
end end
# Determine which grammar rules could expand to empty sequences. # Determine which grammar rules could expand to empty sequences.

View File

@ -2,12 +2,15 @@ class Propane
class Parser class Parser
def initialize(start_rule_set) def initialize(grammar, rule_sets, start_rule_set, log)
@grammar = grammar
@rule_sets = rule_sets
@log = log
@eof_token = Token.new(name: "$", id: TOKEN_EOF) @eof_token = Token.new(name: "$", id: TOKEN_EOF)
start_rule = Rule.new("$$", [start_rule_set, @eof_token], nil, nil, 0) @start_rule = Rule.new("$$", [start_rule_set, @eof_token], nil, nil, 0)
@item_sets = [] @item_sets = []
@item_sets_set = {} @item_sets_set = {}
start_item = Item.new(start_rule, 0) start_item = Item.new(@start_rule, 0)
eval_item_sets = Set[ItemSet.new([start_item])] eval_item_sets = Set[ItemSet.new([start_item])]
while eval_item_sets.size > 0 while eval_item_sets.size > 0
@ -31,30 +34,7 @@ class Propane
end end
build_reduce_actions! build_reduce_actions!
write_log!
@item_sets.each do |item_set|
puts "Item set #{item_set.id}:"
ids = item_set.in_sets.map(&:id)
puts " In sets: #{ids.join(", ")}"
puts " Out sets:"
item_set.out_sets.each do |symbol, out_set|
puts " #{symbol.name} => #{out_set.id}"
end
puts item_set
item_set.following_item_set.each do |following_symbol, following_item_set|
puts " #{following_symbol.name} => #{following_item_set.id}"
end
puts " Reduce actions:"
case item_set.reduce_actions
when Rule
puts " * => #{item_set.reduce_actions.id} (#{item_set.reduce_actions.name})"
when Hash
item_set.reduce_actions.each do |token, rule|
puts " #{token.name} => #{rule.id} (#{rule.name})"
end
end
puts
end
end end
def build_tables def build_tables
@ -107,7 +87,6 @@ class Propane
following_set = @item_sets_set[item_set.build_following_item_set(following_symbol)] following_set = @item_sets_set[item_set.build_following_item_set(following_symbol)]
item_set.following_item_set[following_symbol] = following_set item_set.following_item_set[following_symbol] = following_set
following_set.in_sets << item_set following_set.in_sets << item_set
item_set.out_sets[following_symbol] = following_set
end end
end end
end end
@ -225,6 +204,55 @@ class Propane
lookahead_tokens lookahead_tokens
end end
def write_log!
@log.puts Util.banner("Parser Rules")
([@start_rule] + @grammar.rules).each do |rule|
@log.puts
@log.puts "Rule #{rule.id}:"
@log.puts " #{rule}"
end
@log.puts
@log.puts Util.banner("Parser Tokens")
@log.puts
@grammar.tokens.each do |token|
@log.puts "Token #{token.id}: #{token.name}"
end
@log.puts
@log.puts Util.banner("Parser Rule Sets")
@rule_sets.each do |rule_set_name, rule_set|
@log.puts
@log.puts "Rule Set #{rule_set.id}: #{rule_set_name}"
@log.puts " Start token set: #{rule_set.start_token_set.map(&:name).join(", ")}"
end
@log.puts
@log.puts Util.banner("Parser States")
@item_sets.each do |item_set|
@log.puts
@log.puts "State #{item_set.id}:"
@log.puts item_set.to_s.gsub(/^/, " ")
incoming_ids = item_set.in_sets.map(&:id)
@log.puts
@log.puts " Incoming states: #{incoming_ids.join(", ")}"
@log.puts " Outgoing states:"
item_set.following_item_set.each do |following_symbol, following_item_set|
@log.puts " #{following_symbol.name} => #{following_item_set.id}"
end
@log.puts
@log.puts " Reduce actions:"
case item_set.reduce_actions
when Rule
@log.puts " * => #{item_set.reduce_actions.id} (#{item_set.reduce_actions.name})"
when Hash
item_set.reduce_actions.each do |token, rule|
@log.puts " lookahead #{token.name} => #{rule.name} (#{rule.id}), rule set ##{rule.rule_set.id}"
end
end
end
end
end end
end end

View File

@ -21,10 +21,6 @@ class Propane
# ItemSets leading to this item set. # ItemSets leading to this item set.
attr_reader :in_sets attr_reader :in_sets
# @return [Hash]
# ItemSets reached from this item set. Key is a Token or Rule.
attr_reader :out_sets
# @return [nil, Rule, Hash] # @return [nil, Rule, Hash]
# Reduce actions, mapping lookahead tokens to rules. # Reduce actions, mapping lookahead tokens to rules.
attr_accessor :reduce_actions attr_accessor :reduce_actions
@ -37,7 +33,6 @@ class Propane
@items = Set.new(items) @items = Set.new(items)
@following_item_set = {} @following_item_set = {}
@in_sets = Set.new @in_sets = Set.new
@out_sets = {}
close! close!
end end

View File

@ -56,6 +56,14 @@ class Propane
@components.empty? @components.empty?
end end
# Represent the Rule as a String.
#
# @return [String]
# Rule represented as a String.
def to_s
"#{@name} -> #{@components.map(&:name).join(" ")}"
end
end end
end end

17
lib/propane/util.rb Normal file
View File

@ -0,0 +1,17 @@
class Propane
# Utility methods.
module Util
class << self
def banner(message)
s = "*" * (message.size + 4)
"#{s}\n* #{message} *\n#{s}\n"
end
end
end
end

View File

@ -6,7 +6,7 @@ describe Propane do
end end
def build_parser def build_parser
result = system(*%w[./propane.sh spec/run/testparser.propane spec/run/testparser.d]) result = system(*%w[./propane.sh spec/run/testparser.propane spec/run/testparser.d --log spec/run/testparser.log])
expect(result).to be_truthy expect(result).to be_truthy
end end