Write parser log file
Fix bug of skipping rule set IDs. Remove unneeded out_sets from ItemSet class.
This commit is contained in:
parent
2fbe13e071
commit
30f4cfcc99
@ -1,5 +1,6 @@
|
||||
require "erb"
|
||||
require "set"
|
||||
require "stringio"
|
||||
require_relative "propane/cli"
|
||||
require_relative "propane/code_point_range"
|
||||
require_relative "propane/fa"
|
||||
@ -18,6 +19,7 @@ require_relative "propane/regex/unit"
|
||||
require_relative "propane/rule_set"
|
||||
require_relative "propane/rule"
|
||||
require_relative "propane/token"
|
||||
require_relative "propane/util"
|
||||
require_relative "propane/version"
|
||||
|
||||
class Propane
|
||||
|
@ -5,7 +5,11 @@ class Propane
|
||||
def initialize(grammar, output_file, log_file)
|
||||
@grammar = grammar
|
||||
@output_file = output_file
|
||||
@log_file = log_file
|
||||
if log_file
|
||||
@log = File.open(log_file, "wb")
|
||||
else
|
||||
@log = StringIO.new
|
||||
end
|
||||
@classname = @grammar.classname || File.basename(output_file).sub(%r{[^a-zA-Z0-9].*}, "").capitalize
|
||||
process_grammar!
|
||||
end
|
||||
@ -16,6 +20,7 @@ class Propane
|
||||
File.open(@output_file, "wb") do |fh|
|
||||
fh.write(result)
|
||||
end
|
||||
@log.close
|
||||
end
|
||||
|
||||
private
|
||||
@ -34,8 +39,10 @@ class Propane
|
||||
raise Error.new("Rule name collides with token name #{rule.name.inspect}")
|
||||
end
|
||||
@_rule_set_id ||= @grammar.tokens.size
|
||||
rule_sets[rule.name] ||= RuleSet.new(rule.name, @_rule_set_id)
|
||||
@_rule_set_id += 1
|
||||
unless rule_sets[rule.name]
|
||||
rule_sets[rule.name] = RuleSet.new(rule.name, @_rule_set_id)
|
||||
@_rule_set_id += 1
|
||||
end
|
||||
rule.rule_set = rule_sets[rule.name]
|
||||
rule_sets[rule.name] << rule
|
||||
end
|
||||
@ -54,13 +61,8 @@ class Propane
|
||||
end
|
||||
end
|
||||
determine_possibly_empty_rulesets!(rule_sets)
|
||||
puts "Start token set"
|
||||
rule_sets.each do |rule_set_name, rule_set|
|
||||
puts "RuleSet #{rule_set_name}:"
|
||||
puts " " + rule_set.start_token_set.map(&:name).join(", ")
|
||||
end
|
||||
@lexer = Lexer.new(@grammar.tokens, @grammar.drop_tokens)
|
||||
@parser = Parser.new(rule_sets["Start"])
|
||||
@parser = Parser.new(@grammar, rule_sets, rule_sets["Start"], @log)
|
||||
end
|
||||
|
||||
# Determine which grammar rules could expand to empty sequences.
|
||||
|
@ -2,12 +2,15 @@ class Propane
|
||||
|
||||
class Parser
|
||||
|
||||
def initialize(start_rule_set)
|
||||
def initialize(grammar, rule_sets, start_rule_set, log)
|
||||
@grammar = grammar
|
||||
@rule_sets = rule_sets
|
||||
@log = log
|
||||
@eof_token = Token.new(name: "$", id: TOKEN_EOF)
|
||||
start_rule = Rule.new("$$", [start_rule_set, @eof_token], nil, nil, 0)
|
||||
@start_rule = Rule.new("$$", [start_rule_set, @eof_token], nil, nil, 0)
|
||||
@item_sets = []
|
||||
@item_sets_set = {}
|
||||
start_item = Item.new(start_rule, 0)
|
||||
start_item = Item.new(@start_rule, 0)
|
||||
eval_item_sets = Set[ItemSet.new([start_item])]
|
||||
|
||||
while eval_item_sets.size > 0
|
||||
@ -31,30 +34,7 @@ class Propane
|
||||
end
|
||||
|
||||
build_reduce_actions!
|
||||
|
||||
@item_sets.each do |item_set|
|
||||
puts "Item set #{item_set.id}:"
|
||||
ids = item_set.in_sets.map(&:id)
|
||||
puts " In sets: #{ids.join(", ")}"
|
||||
puts " Out sets:"
|
||||
item_set.out_sets.each do |symbol, out_set|
|
||||
puts " #{symbol.name} => #{out_set.id}"
|
||||
end
|
||||
puts item_set
|
||||
item_set.following_item_set.each do |following_symbol, following_item_set|
|
||||
puts " #{following_symbol.name} => #{following_item_set.id}"
|
||||
end
|
||||
puts " Reduce actions:"
|
||||
case item_set.reduce_actions
|
||||
when Rule
|
||||
puts " * => #{item_set.reduce_actions.id} (#{item_set.reduce_actions.name})"
|
||||
when Hash
|
||||
item_set.reduce_actions.each do |token, rule|
|
||||
puts " #{token.name} => #{rule.id} (#{rule.name})"
|
||||
end
|
||||
end
|
||||
puts
|
||||
end
|
||||
write_log!
|
||||
end
|
||||
|
||||
def build_tables
|
||||
@ -107,7 +87,6 @@ class Propane
|
||||
following_set = @item_sets_set[item_set.build_following_item_set(following_symbol)]
|
||||
item_set.following_item_set[following_symbol] = following_set
|
||||
following_set.in_sets << item_set
|
||||
item_set.out_sets[following_symbol] = following_set
|
||||
end
|
||||
end
|
||||
end
|
||||
@ -225,6 +204,55 @@ class Propane
|
||||
lookahead_tokens
|
||||
end
|
||||
|
||||
def write_log!
|
||||
@log.puts Util.banner("Parser Rules")
|
||||
([@start_rule] + @grammar.rules).each do |rule|
|
||||
@log.puts
|
||||
@log.puts "Rule #{rule.id}:"
|
||||
@log.puts " #{rule}"
|
||||
end
|
||||
|
||||
@log.puts
|
||||
@log.puts Util.banner("Parser Tokens")
|
||||
@log.puts
|
||||
@grammar.tokens.each do |token|
|
||||
@log.puts "Token #{token.id}: #{token.name}"
|
||||
end
|
||||
|
||||
@log.puts
|
||||
@log.puts Util.banner("Parser Rule Sets")
|
||||
@rule_sets.each do |rule_set_name, rule_set|
|
||||
@log.puts
|
||||
@log.puts "Rule Set #{rule_set.id}: #{rule_set_name}"
|
||||
@log.puts " Start token set: #{rule_set.start_token_set.map(&:name).join(", ")}"
|
||||
end
|
||||
|
||||
@log.puts
|
||||
@log.puts Util.banner("Parser States")
|
||||
@item_sets.each do |item_set|
|
||||
@log.puts
|
||||
@log.puts "State #{item_set.id}:"
|
||||
@log.puts item_set.to_s.gsub(/^/, " ")
|
||||
incoming_ids = item_set.in_sets.map(&:id)
|
||||
@log.puts
|
||||
@log.puts " Incoming states: #{incoming_ids.join(", ")}"
|
||||
@log.puts " Outgoing states:"
|
||||
item_set.following_item_set.each do |following_symbol, following_item_set|
|
||||
@log.puts " #{following_symbol.name} => #{following_item_set.id}"
|
||||
end
|
||||
@log.puts
|
||||
@log.puts " Reduce actions:"
|
||||
case item_set.reduce_actions
|
||||
when Rule
|
||||
@log.puts " * => #{item_set.reduce_actions.id} (#{item_set.reduce_actions.name})"
|
||||
when Hash
|
||||
item_set.reduce_actions.each do |token, rule|
|
||||
@log.puts " lookahead #{token.name} => #{rule.name} (#{rule.id}), rule set ##{rule.rule_set.id}"
|
||||
end
|
||||
end
|
||||
end
|
||||
end
|
||||
|
||||
end
|
||||
|
||||
end
|
||||
|
@ -21,10 +21,6 @@ class Propane
|
||||
# ItemSets leading to this item set.
|
||||
attr_reader :in_sets
|
||||
|
||||
# @return [Hash]
|
||||
# ItemSets reached from this item set. Key is a Token or Rule.
|
||||
attr_reader :out_sets
|
||||
|
||||
# @return [nil, Rule, Hash]
|
||||
# Reduce actions, mapping lookahead tokens to rules.
|
||||
attr_accessor :reduce_actions
|
||||
@ -37,7 +33,6 @@ class Propane
|
||||
@items = Set.new(items)
|
||||
@following_item_set = {}
|
||||
@in_sets = Set.new
|
||||
@out_sets = {}
|
||||
close!
|
||||
end
|
||||
|
||||
|
@ -56,6 +56,14 @@ class Propane
|
||||
@components.empty?
|
||||
end
|
||||
|
||||
# Represent the Rule as a String.
|
||||
#
|
||||
# @return [String]
|
||||
# Rule represented as a String.
|
||||
def to_s
|
||||
"#{@name} -> #{@components.map(&:name).join(" ")}"
|
||||
end
|
||||
|
||||
end
|
||||
|
||||
end
|
||||
|
17
lib/propane/util.rb
Normal file
17
lib/propane/util.rb
Normal file
@ -0,0 +1,17 @@
|
||||
class Propane
|
||||
|
||||
# Utility methods.
|
||||
module Util
|
||||
|
||||
class << self
|
||||
|
||||
def banner(message)
|
||||
s = "*" * (message.size + 4)
|
||||
"#{s}\n* #{message} *\n#{s}\n"
|
||||
end
|
||||
|
||||
end
|
||||
|
||||
end
|
||||
|
||||
end
|
@ -6,7 +6,7 @@ describe Propane do
|
||||
end
|
||||
|
||||
def build_parser
|
||||
result = system(*%w[./propane.sh spec/run/testparser.propane spec/run/testparser.d])
|
||||
result = system(*%w[./propane.sh spec/run/testparser.propane spec/run/testparser.d --log spec/run/testparser.log])
|
||||
expect(result).to be_truthy
|
||||
end
|
||||
|
||||
|
Loading…
x
Reference in New Issue
Block a user