2010-02-11 01:57:33 -05:00
|
|
|
// Jison, an LR(0), SLR(1), LARL(1), LR(1) Parser Generator
|
|
|
|
// Zachary Carter <zach@carter.name>
|
|
|
|
// MIT X Licensed
|
|
|
|
|
|
|
|
if (typeof exports === 'undefined') {
|
|
|
|
exports = {};
|
|
|
|
} else {
|
|
|
|
// assume we're in commonjs land
|
|
|
|
//var system = require("system");
|
|
|
|
var typal = require('./jison/util/typal').typal;
|
|
|
|
var Set = require('./jison/util/set').Set;
|
|
|
|
var RegExpLexer = require('./jison/lexer').RegExpLexer;
|
|
|
|
}
|
|
|
|
|
|
|
|
var Jison = exports.Jison = exports;
|
|
|
|
|
|
|
|
// detect print
|
|
|
|
if (typeof puts !== 'undefined') {
|
|
|
|
Jison.print = function print () { puts([].join.call(arguments, ' ')); };
|
|
|
|
} else if (typeof print !== 'undefined') {
|
|
|
|
Jison.print = print;
|
|
|
|
} else {
|
|
|
|
Jison.print = function print () {};
|
|
|
|
}
|
|
|
|
|
|
|
|
Jison.Parser = (function () {
|
|
|
|
|
|
|
|
// iterator utility
|
|
|
|
function each (obj, func) {
|
|
|
|
if (obj.forEach) {
|
|
|
|
obj.forEach(func);
|
|
|
|
} else {
|
|
|
|
var p;
|
|
|
|
for (p in obj) {
|
|
|
|
if (obj.hasOwnProperty(p)) {
|
|
|
|
func.call(obj, obj[p], p, obj);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
var Nonterminal = typal.construct({
|
|
|
|
constructor: function Nonterminal (symbol) {
|
|
|
|
this.symbol = symbol;
|
|
|
|
this.productions = new Set();
|
|
|
|
this.first = [];
|
|
|
|
this.follows = [];
|
|
|
|
this.nullable = false;
|
|
|
|
},
|
|
|
|
toString: function Nonterminal_toString () {
|
|
|
|
var str = this.symbol+"\n";
|
|
|
|
str += (this.nullable ? 'nullable' : 'not nullable');
|
|
|
|
str += "\nFirsts: "+this.first.join(', ');
|
|
|
|
str += "\nFollows: "+this.first.join(', ');
|
|
|
|
str += "\nProductions:\n "+this.productions.join('\n ');
|
|
|
|
|
|
|
|
return str;
|
|
|
|
}
|
|
|
|
});
|
|
|
|
|
|
|
|
var Production = typal.construct({
|
|
|
|
constructor: function Production (symbol, handle, id) {
|
|
|
|
this.symbol = symbol;
|
|
|
|
this.handle = handle;
|
|
|
|
this.nullable = false;
|
|
|
|
this.id = id;
|
|
|
|
this.first = [];
|
|
|
|
this.precedence = 0;
|
|
|
|
},
|
|
|
|
toString: function Production_toString () {
|
|
|
|
return this.symbol+" -> "+this.handle.join(' ');
|
|
|
|
}
|
|
|
|
});
|
|
|
|
|
|
|
|
var generator = typal.beget();
|
|
|
|
|
|
|
|
generator.constructor = function Jison_Generator (grammar, opt) {
|
|
|
|
if (typeof grammar === 'string') {
|
|
|
|
grammar = require("jison/bnf").parse(grammar);
|
|
|
|
}
|
|
|
|
|
|
|
|
var options = typal.mix.call({}, grammar.options, opt);
|
|
|
|
this.terms = {};
|
|
|
|
this.operators = {};
|
|
|
|
this.productions = [];
|
|
|
|
this.conflicts = 0;
|
|
|
|
this.resolutions = [];
|
|
|
|
this.options = options;
|
|
|
|
this.yy = {}; // accessed as yy free variable in the parser/lexer actions
|
|
|
|
|
|
|
|
// source included in semantic action execution scope
|
|
|
|
if (grammar.actionInclude) {
|
|
|
|
if (typeof grammar.actionInclude === 'function') {
|
|
|
|
grammar.actionInclude = String(grammar.actionInclude).replace(/^\s*function \(\) \{/, '').replace(/\}\s*$/, '');
|
|
|
|
}
|
|
|
|
this.actionInclude = grammar.actionInclude;
|
|
|
|
}
|
|
|
|
|
|
|
|
this.DEBUG = options.debug || false;
|
|
|
|
if (this.DEBUG) this.mix(generatorDebug); // mixin debug methods
|
|
|
|
|
|
|
|
this.processGrammar(grammar);
|
|
|
|
|
|
|
|
if (grammar.lex) {
|
|
|
|
this.lexer = new RegExpLexer(grammar.lex, null, this.terminals_);
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
generator.processGrammar = function processGrammarDef (grammar) {
|
|
|
|
var bnf = grammar.bnf,
|
|
|
|
tokens = grammar.tokens,
|
|
|
|
nonterminals = this.nonterminals = {},
|
|
|
|
productions = this.productions,
|
|
|
|
self = this;
|
|
|
|
|
|
|
|
if (tokens) {
|
|
|
|
if (typeof tokens === 'string') {
|
|
|
|
tokens = tokens.trim().split(' ');
|
|
|
|
} else {
|
|
|
|
tokens = tokens.slice(0);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
var symbols = this.symbols = [];
|
|
|
|
|
|
|
|
// calculate precedence of operators
|
|
|
|
var operators = this.operators = processOperators(grammar.operators);
|
|
|
|
|
|
|
|
// build productions from cfg
|
|
|
|
this.buildProductions(grammar.bnf, productions, nonterminals, symbols, operators);
|
|
|
|
|
|
|
|
if (tokens && this.terminals.length !== tokens.length) {
|
|
|
|
self.trace("Warning: declared tokens differ from tokens found in rules.");
|
|
|
|
self.trace(this.terminals);
|
|
|
|
self.trace(tokens);
|
|
|
|
}
|
|
|
|
|
|
|
|
// augment the grammar
|
|
|
|
this.augmentGrammar(grammar);
|
|
|
|
};
|
|
|
|
|
|
|
|
generator.augmentGrammar = function augmentGrammar (grammar) {
|
|
|
|
// use specified start symbol, or default to first user defined production
|
|
|
|
this.startSymbol = grammar.start || grammar.startSymbol || this.productions[0].symbol;
|
|
|
|
if (!this.nonterminals[this.startSymbol]) {
|
|
|
|
throw new Error("Grammar error: startSymbol must be a non-terminal found in your grammar.");
|
|
|
|
}
|
|
|
|
this.EOF = "$end";
|
|
|
|
|
|
|
|
// augment the grammar
|
|
|
|
var acceptProduction = new Production('$accept', [this.startSymbol, '$end'], 0);
|
|
|
|
this.productions.unshift(acceptProduction);
|
|
|
|
|
|
|
|
// prepend parser tokens
|
|
|
|
this.symbols.unshift("$accept","$end");
|
|
|
|
this.symbols_["$accept"] = 0;
|
|
|
|
this.symbols_["$end"] = 1;
|
|
|
|
this.terminals.unshift("$end");
|
|
|
|
|
|
|
|
this.nonterminals["$accept"] = new Nonterminal("$accept");
|
|
|
|
this.nonterminals["$accept"].productions.push(acceptProduction);
|
|
|
|
|
|
|
|
// add follow $ to start symbol
|
|
|
|
this.nonterminals[this.startSymbol].follows.push(this.EOF);
|
|
|
|
};
|
|
|
|
|
|
|
|
// set precedence and associativity of operators
|
|
|
|
function processOperators (ops) {
|
|
|
|
if (!ops) return {};
|
|
|
|
var operators = {};
|
|
|
|
for (var i=0,k,prec;prec=ops[i]; i++) {
|
|
|
|
for (k=1;k < prec.length;k++) {
|
|
|
|
operators[prec[k]] = {precedence: i+1, assoc: prec[0]};
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return operators;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
generator.buildProductions = function buildProductions(bnf, productions, nonterminals, symbols, operators) {
|
|
|
|
var actions = [this.actionInclude || "", "var $$ = arguments[5],$0=arguments[5].length;",'switch(arguments[4]) {'],
|
|
|
|
prods, symbol;
|
|
|
|
var productions_ = [0];
|
|
|
|
var symbolId = 1;
|
|
|
|
var symbols_ = {};
|
|
|
|
|
|
|
|
function addSymbol (s) {
|
|
|
|
if (s && !symbols_[s]) {
|
|
|
|
symbols_[s] = ++symbolId;
|
|
|
|
symbols.push(s);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
for (symbol in bnf) {
|
|
|
|
if (!bnf.hasOwnProperty(symbol)) continue;
|
|
|
|
|
|
|
|
addSymbol(symbol);
|
|
|
|
nonterminals[symbol] = new Nonterminal(symbol);
|
|
|
|
|
|
|
|
if (typeof bnf[symbol] === 'string') {
|
|
|
|
prods = bnf[symbol].split(/\s*\|\s*/g);
|
|
|
|
} else {
|
|
|
|
prods = bnf[symbol].slice(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
prods.forEach(function buildProds_forEach (handle) {
|
|
|
|
var r, rhs, i;
|
|
|
|
if (handle.constructor === Array) {
|
|
|
|
if (typeof handle[0] === 'string')
|
|
|
|
rhs = handle[0].trim().split(' ');
|
2010-02-11 23:34:45 -05:00
|
|
|
else
|
2010-02-11 01:57:33 -05:00
|
|
|
rhs = handle[0].slice(0);
|
|
|
|
|
|
|
|
for (i=0; i<rhs.length; i++) if (!symbols_[rhs[i]]) {
|
|
|
|
addSymbol(rhs[i]);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (typeof handle[1] === 'string' || handle.length == 3) {
|
|
|
|
// semantic action specified
|
|
|
|
var action = 'case '+(productions.length+1)+':'+handle[1]+'\nbreak;';
|
|
|
|
|
|
|
|
// replace named semantic values ($nonterminal)
|
|
|
|
if (action.match(/\$[a-zA-Z][a-zA-Z0-9_]*/)) {
|
|
|
|
var count = {},
|
|
|
|
names = {};
|
|
|
|
for (i=0;i<rhs.length;i++) {
|
|
|
|
if (names[rhs[i]]) {
|
|
|
|
names[rhs[i]+(++count[rhs[i]])] = i+1;
|
|
|
|
} else {
|
|
|
|
names[rhs[i]] = i+1;
|
|
|
|
names[rhs[i]+"1"] = i+1;
|
|
|
|
count[rhs[i]] = 1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
action = action.replace(/\$([a-zA-Z][a-zA-Z0-9_]*)/g, function (str, pl) {
|
|
|
|
return names[pl] ? '$'+names[pl] : pl;
|
|
|
|
});
|
|
|
|
}
|
|
|
|
action = action.replace(/\$(?:0|\$)/g, "this.$")
|
|
|
|
.replace(/\$(\d+)/g, "$$$[\$0-"+rhs.length+"+$1-1]");
|
|
|
|
actions.push(action);
|
|
|
|
|
|
|
|
r = new Production(symbol, rhs, productions.length+1);
|
|
|
|
// precedence specified also
|
|
|
|
if (handle[2]) {
|
|
|
|
r.precedence = operators[handle[2].prec].precedence;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
// only precedence specified
|
|
|
|
r = new Production(symbol, rhs, productions.length+1);
|
|
|
|
r.precedence = operators[handle[1].prec].precedence;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
rhs = handle.trim().split(' ');
|
|
|
|
for (i=0; i<rhs.length; i++) if (!symbols_[rhs[i]]) {
|
|
|
|
addSymbol(rhs[i]);
|
|
|
|
}
|
|
|
|
r = new Production(symbol, rhs, productions.length+1);
|
|
|
|
}
|
|
|
|
if (r.precedence === 0) {
|
|
|
|
// set precedence
|
|
|
|
for (i=r.handle.length-1; i>=0; i--) {
|
|
|
|
if (!(r.handle[i] in nonterminals) && r.handle[i] in operators) {
|
|
|
|
r.precedence = operators[r.handle[i]].precedence;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
productions.push(r);
|
|
|
|
productions_.push([symbols_[r.symbol], r.handle[0] === '' ? 0 : r.handle.length]);
|
|
|
|
nonterminals[symbol].productions.push(r);
|
|
|
|
});
|
|
|
|
}
|
|
|
|
|
|
|
|
var sym, terms = [], terms_ = {};
|
|
|
|
each(symbols_, function (id, sym) {
|
|
|
|
if (!nonterminals[sym]) {
|
|
|
|
terms.push(sym);
|
|
|
|
terms_[id] = sym;
|
|
|
|
}
|
|
|
|
});
|
|
|
|
|
|
|
|
this.terminals = terms;
|
|
|
|
this.terminals_ = terms_;
|
|
|
|
this.symbols_ = symbols_;
|
|
|
|
|
|
|
|
this.productions_ = productions_;
|
|
|
|
actions.push('}');
|
|
|
|
this.performAction = Function("yytext","yyleng","yylineno","yy", actions.join("\n"));
|
|
|
|
};
|
|
|
|
|
|
|
|
generator.createParser = function createParser () {
|
|
|
|
throw 'Calling abstract method.';
|
|
|
|
};
|
|
|
|
|
|
|
|
// noop. implemented in debug mixin
|
|
|
|
generator.trace = function trace () { };
|
|
|
|
|
|
|
|
generator.warn = function warn () {
|
|
|
|
Jison.print.apply(null,arguments);
|
|
|
|
};
|
|
|
|
|
|
|
|
generator.error = function error (msg) {
|
|
|
|
throw msg;
|
|
|
|
};
|
|
|
|
|
|
|
|
// Generator debug mixin
|
|
|
|
|
|
|
|
var generatorDebug = {
|
|
|
|
trace: function trace () {
|
|
|
|
Jison.print.apply(null, arguments);
|
|
|
|
},
|
|
|
|
beforeprocessGrammar: function () {
|
|
|
|
this.trace("Processing grammar.");
|
|
|
|
},
|
|
|
|
afteraugmentGrammar: function () {
|
|
|
|
var trace = this.trace;
|
|
|
|
each(this.symbols, function (sym, i) {
|
|
|
|
trace(sym+"("+i+")");
|
|
|
|
});
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Mixin for common behaviors of lookahead parsers
|
|
|
|
* */
|
|
|
|
var lookaheadMixin = {};
|
|
|
|
|
|
|
|
lookaheadMixin.computeLookaheads = function computeLookaheads () {
|
|
|
|
if (this.DEBUG) this.mix(lookaheadDebug); // mixin debug methods
|
|
|
|
|
|
|
|
this.computeLookaheads = function () {};
|
|
|
|
this.nullableSets();
|
|
|
|
this.firstSets();
|
|
|
|
this.followSets();
|
|
|
|
};
|
|
|
|
|
|
|
|
// calculate follow sets typald on first and nullable
|
|
|
|
lookaheadMixin.followSets = function followSets () {
|
|
|
|
var productions = this.productions,
|
|
|
|
nonterminals = this.nonterminals,
|
|
|
|
self = this,
|
|
|
|
cont = true;
|
|
|
|
|
|
|
|
// loop until no further changes have been made
|
|
|
|
while(cont) {
|
|
|
|
cont = false;
|
|
|
|
|
|
|
|
productions.forEach(function Follow_prod_forEach (production, k) {
|
|
|
|
//self.trace(production.symbol,nonterminals[production.symbol].follows);
|
|
|
|
// q is used in Simple LALR algorithm determine follows in context
|
|
|
|
var q;
|
|
|
|
var ctx = !!self.go_;
|
|
|
|
|
|
|
|
var set = [],oldcount;
|
2010-02-11 23:34:45 -05:00
|
|
|
for (var i=0,t;t=production.handle[i];++i) {
|
2010-02-11 01:57:33 -05:00
|
|
|
if (!nonterminals[t]) continue;
|
|
|
|
|
2010-02-11 23:34:45 -05:00
|
|
|
// for Simple LALR algorithm, self.go_ checks if
|
2010-02-11 01:57:33 -05:00
|
|
|
if (ctx)
|
|
|
|
q = self.go_(production.symbol, production.handle.slice(0, i));
|
2010-02-11 23:34:45 -05:00
|
|
|
var bool = !ctx || q === parseInt(self.nterms_[t]);
|
2010-02-11 01:57:33 -05:00
|
|
|
|
|
|
|
if (i === production.handle.length+1 && bool) {
|
|
|
|
set = nonterminals[production.symbol].follows
|
|
|
|
} else {
|
|
|
|
var part = production.handle.slice(i+1);
|
|
|
|
|
|
|
|
set = self.first(part);
|
2010-02-23 21:03:05 -05:00
|
|
|
if (self.nullable(part) && bool) {
|
|
|
|
set.push.apply(set, nonterminals[production.symbol].follows);
|
2010-02-11 01:57:33 -05:00
|
|
|
}
|
|
|
|
}
|
|
|
|
oldcount = nonterminals[t].follows.length;
|
|
|
|
Set.union(nonterminals[t].follows, set);
|
|
|
|
if (oldcount !== nonterminals[t].follows.length) {
|
|
|
|
cont = true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
});
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
// return the FIRST set of a symbol or series of symbols
|
|
|
|
lookaheadMixin.first = function first (symbol) {
|
|
|
|
// epsilon
|
|
|
|
if (symbol === '') {
|
|
|
|
return [];
|
|
|
|
// RHS
|
|
|
|
} else if (symbol instanceof Array) {
|
|
|
|
var firsts = [];
|
2010-02-11 23:34:45 -05:00
|
|
|
for (var i=0,t;t=symbol[i];++i) {
|
2010-03-07 18:48:15 -05:00
|
|
|
if (!this.nonterminals[t]) {
|
|
|
|
if (firsts.indexOf(t) === -1)
|
|
|
|
firsts.push(t);
|
|
|
|
} else {
|
|
|
|
Set.union(firsts, this.nonterminals[t].first);
|
|
|
|
}
|
2010-02-11 01:57:33 -05:00
|
|
|
if (!this.nullable(t))
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
return firsts;
|
|
|
|
// terminal
|
|
|
|
} else if (!this.nonterminals[symbol]) {
|
|
|
|
return [symbol];
|
|
|
|
// nonterminal
|
|
|
|
} else {
|
|
|
|
return this.nonterminals[symbol].first;
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
// fixed-point calculation of FIRST sets
|
|
|
|
lookaheadMixin.firstSets = function firstSets () {
|
|
|
|
var productions = this.productions,
|
|
|
|
nonterminals = this.nonterminals,
|
|
|
|
self = this,
|
|
|
|
cont = true,
|
|
|
|
symbol,firsts;
|
|
|
|
|
|
|
|
// loop until no further changes have been made
|
|
|
|
while(cont) {
|
|
|
|
cont = false;
|
|
|
|
|
|
|
|
productions.forEach(function FirstSets_forEach (production, k) {
|
|
|
|
var firsts = self.first(production.handle);
|
|
|
|
if (firsts.length !== production.first.length) {
|
|
|
|
production.first = firsts;
|
|
|
|
cont=true;
|
|
|
|
}
|
|
|
|
});
|
|
|
|
|
|
|
|
for (symbol in nonterminals) {
|
|
|
|
firsts = [];
|
|
|
|
nonterminals[symbol].productions.forEach(function (production) {
|
|
|
|
Set.union(firsts, production.first);
|
|
|
|
});
|
|
|
|
if (firsts.length !== nonterminals[symbol].first.length) {
|
|
|
|
nonterminals[symbol].first = firsts;
|
|
|
|
cont=true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
// fixed-point calculation of NULLABLE
|
|
|
|
lookaheadMixin.nullableSets = function nullableSets () {
|
|
|
|
var firsts = this.firsts = {},
|
|
|
|
nonterminals = this.nonterminals,
|
|
|
|
self = this,
|
|
|
|
cont = true;
|
|
|
|
|
|
|
|
// loop until no further changes have been made
|
|
|
|
while(cont) {
|
|
|
|
cont = false;
|
|
|
|
|
|
|
|
// check if each production is nullable
|
|
|
|
this.productions.forEach(function (production, k) {
|
|
|
|
if (!production.nullable) {
|
|
|
|
for (var i=0,n=0,t;t=production.handle[i];++i) {
|
|
|
|
if (self.nullable(t)) n++;
|
|
|
|
}
|
|
|
|
if (n===i) { // production is nullable if all tokens are nullable
|
|
|
|
production.nullable = cont = true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
});
|
|
|
|
|
|
|
|
//check if each symbol is nullable
|
|
|
|
for (var symbol in nonterminals) {
|
|
|
|
if (!this.nullable(symbol)) {
|
|
|
|
for (var i=0,production;production=nonterminals[symbol].productions.item(i);i++) {
|
|
|
|
if (production.nullable)
|
|
|
|
nonterminals[symbol].nullable = cont = true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
// check if a token or series of tokens is nullable
|
|
|
|
lookaheadMixin.nullable = function nullable (symbol) {
|
|
|
|
// epsilon
|
|
|
|
if (symbol === '') {
|
|
|
|
return true
|
|
|
|
// RHS
|
|
|
|
} else if (symbol instanceof Array) {
|
|
|
|
for (var i=0,t;t=symbol[i];++i) {
|
|
|
|
if (!this.nullable(t))
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
// terminal
|
|
|
|
} else if (!this.nonterminals[symbol]) {
|
|
|
|
return false;
|
|
|
|
// nonterminal
|
|
|
|
} else {
|
|
|
|
return this.nonterminals[symbol].nullable;
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
|
|
// lookahead debug mixin
|
|
|
|
var lookaheadDebug = {
|
|
|
|
beforenullableSets: function () {
|
|
|
|
this.trace("Computing Nullable sets.");
|
|
|
|
},
|
|
|
|
beforefirstSets: function () {
|
|
|
|
this.trace("Computing First sets.");
|
|
|
|
},
|
|
|
|
beforefollowSets: function () {
|
|
|
|
this.trace("Computing Follow sets.");
|
|
|
|
},
|
|
|
|
afterfollowSets: function () {
|
|
|
|
var trace = this.trace;
|
|
|
|
each(this.nonterminals, function (nt, t) {
|
|
|
|
trace(nt, '\n');
|
|
|
|
});
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Mixin for common LR parser behavior
|
|
|
|
* */
|
|
|
|
var lrGeneratorMixin = {};
|
|
|
|
|
|
|
|
lrGeneratorMixin.buildTable = function buildTable () {
|
|
|
|
if (this.DEBUG) this.mix(lrGeneratorDebug); // mixin debug methods
|
|
|
|
|
|
|
|
this.states = this.canonicalCollection();
|
|
|
|
this.table = this.parseTable(this.states);
|
|
|
|
};
|
|
|
|
|
|
|
|
lrGeneratorMixin.Item = typal.construct({
|
|
|
|
constructor: function Item(production, dot, f, predecessor) {
|
|
|
|
this.production = production;
|
|
|
|
this.dotPosition = dot || 0;
|
2010-02-11 23:34:45 -05:00
|
|
|
this.follows = f || [];
|
2010-02-11 01:57:33 -05:00
|
|
|
this.predecessor = predecessor;
|
|
|
|
this.id = parseInt(production.id+'a'+this.dotPosition, 36);
|
|
|
|
this.markedSymbol = this.production.handle[this.dotPosition];
|
|
|
|
},
|
|
|
|
remainingHandle: function () {
|
|
|
|
return this.production.handle.slice(this.dotPosition+1);
|
|
|
|
},
|
|
|
|
eq: function (e) {
|
|
|
|
return e.id === this.id;
|
|
|
|
},
|
|
|
|
toString: function () {
|
|
|
|
var temp = this.production.handle.slice(0);
|
|
|
|
temp[this.dotPosition] = '.'+(temp[this.dotPosition]||'');
|
|
|
|
return '['+this.production.symbol+" -> "+temp.join(' ')
|
|
|
|
+(this.follows.length === 0 ? "" : ", "+this.follows.join('/'))
|
|
|
|
+']';
|
|
|
|
}
|
|
|
|
});
|
|
|
|
|
|
|
|
lrGeneratorMixin.ItemSet = Set.prototype.construct({
|
|
|
|
afterconstructor: function () {
|
|
|
|
this.reductions = [];
|
|
|
|
this.goes = {};
|
|
|
|
this.edges = {};
|
|
|
|
this.shifts = false;
|
|
|
|
this.inadequate = false;
|
2010-02-11 23:34:45 -05:00
|
|
|
this.hash_ = {};
|
|
|
|
for (var i=this._items.length-1;i >=0;i--) {
|
|
|
|
this.hash_[this._items[i].id] = true; //i;
|
|
|
|
}
|
|
|
|
},
|
|
|
|
concat: function concat (set) {
|
|
|
|
var a = set._items || set;
|
|
|
|
for (var i=a.length-1;i >=0;i--) {
|
|
|
|
this.hash_[a[i].id] = true; //i;
|
|
|
|
}
|
|
|
|
this._items.push.apply(this._items, a);
|
|
|
|
return this;
|
|
|
|
},
|
|
|
|
push: function (item) {
|
|
|
|
this.hash_[item.id] = true;
|
|
|
|
return this._items.push(item);
|
|
|
|
},
|
|
|
|
contains: function (item) {
|
|
|
|
return this.hash_[item.id];
|
2010-02-11 01:57:33 -05:00
|
|
|
},
|
2010-03-07 18:48:15 -05:00
|
|
|
valueOf: function toValue () {
|
|
|
|
var v = this._items.map(function (a) {return a.id}).sort().join('|');
|
|
|
|
return (this.valueOf = function toValue_inner() {return v;})();
|
2010-02-11 01:57:33 -05:00
|
|
|
}
|
|
|
|
});
|
|
|
|
|
|
|
|
lrGeneratorMixin.closureOperation = function closureOperation (itemSet /*, closureSet*/) {
|
|
|
|
var closureSet = new this.ItemSet();
|
|
|
|
var self = this;
|
|
|
|
|
|
|
|
var set = itemSet,
|
|
|
|
itemQueue, syms = {};
|
|
|
|
|
|
|
|
do {
|
|
|
|
itemQueue = new Set();
|
|
|
|
closureSet.concat(set);
|
|
|
|
set.forEach(function CO_set_forEach (item) {
|
|
|
|
var symbol = item.markedSymbol;
|
|
|
|
|
|
|
|
// if token is a non-terminal, recursively add closures
|
|
|
|
if (symbol && self.nonterminals[symbol]) {
|
|
|
|
if(!syms[symbol]) {
|
2010-02-11 23:34:45 -05:00
|
|
|
self.nonterminals[symbol].productions.forEach(function CO_nt_forEach (production) {
|
2010-02-11 01:57:33 -05:00
|
|
|
var newItem = new self.Item(production, 0);
|
|
|
|
if(!closureSet.contains(newItem))
|
|
|
|
itemQueue.push(newItem);
|
|
|
|
});
|
|
|
|
syms[symbol] = true;
|
|
|
|
}
|
|
|
|
} else if (!symbol) {
|
|
|
|
// reduction
|
|
|
|
closureSet.reductions.push(item);
|
|
|
|
closureSet.inadequate = closureSet.reductions.length > 1 || closureSet.shifts;
|
|
|
|
} else {
|
|
|
|
// shift
|
|
|
|
closureSet.shifts = true;
|
|
|
|
closureSet.inadequate = closureSet.reductions.length > 0;
|
|
|
|
}
|
|
|
|
});
|
|
|
|
|
|
|
|
set = itemQueue;
|
|
|
|
|
|
|
|
} while (!itemQueue.isEmpty());
|
|
|
|
|
|
|
|
return closureSet;
|
|
|
|
};
|
|
|
|
|
|
|
|
lrGeneratorMixin.gotoOperation = function gotoOperation (itemSet, symbol) {
|
|
|
|
var gotoSet = new this.ItemSet(),
|
|
|
|
self = this;
|
|
|
|
|
|
|
|
itemSet.forEach(function goto_forEach(item, n) {
|
2010-03-07 18:48:15 -05:00
|
|
|
if (item.markedSymbol === symbol) {
|
2010-02-11 01:57:33 -05:00
|
|
|
gotoSet.push(new self.Item(item.production, item.dotPosition+1, item.follows, n));
|
|
|
|
}
|
|
|
|
});
|
|
|
|
|
|
|
|
return gotoSet.isEmpty() ? gotoSet : this.closureOperation(gotoSet);
|
|
|
|
};
|
|
|
|
|
|
|
|
/* Create unique set of item sets
|
|
|
|
* */
|
|
|
|
lrGeneratorMixin.canonicalCollection = function canonicalCollection () {
|
|
|
|
var item1 = new this.Item(this.productions[0], 0, new Set(this.EOF));
|
|
|
|
var firstState = this.closureOperation(new this.ItemSet(item1)),
|
|
|
|
states = new Set(firstState),
|
|
|
|
marked = 0,
|
|
|
|
self = this,
|
|
|
|
itemSet;
|
|
|
|
|
2010-03-07 18:48:15 -05:00
|
|
|
states.has = {};
|
|
|
|
states.has[firstState] = 0;
|
|
|
|
|
2010-02-11 01:57:33 -05:00
|
|
|
while (marked !== states.size()) {
|
|
|
|
itemSet = states.item(marked); marked++;
|
2010-03-07 18:48:15 -05:00
|
|
|
itemSet.forEach(function CC_itemSet_forEach (item) {
|
|
|
|
if (item.markedSymbol && item.markedSymbol !== self.EOF)
|
2010-02-11 01:57:33 -05:00
|
|
|
self.canonicalCollectionInsert(item.markedSymbol, itemSet, states, marked-1);
|
|
|
|
});
|
|
|
|
}
|
|
|
|
|
|
|
|
return states;
|
|
|
|
};
|
|
|
|
|
|
|
|
// Pushes a unique state into the que. Some parsing algorithms may perform additional operations
|
|
|
|
lrGeneratorMixin.canonicalCollectionInsert = function canonicalCollectionInsert (symbol, itemSet, states, stateNum) {
|
|
|
|
var g = this.gotoOperation(itemSet, symbol);
|
|
|
|
if (!g.predecessors)
|
|
|
|
g.predecessors = {};
|
|
|
|
// add g to que if not empty or duplicate
|
|
|
|
if (!g.isEmpty()) {
|
2010-03-07 18:48:15 -05:00
|
|
|
var gv = g.valueOf(),
|
|
|
|
i = states.has[gv];
|
|
|
|
if (i === -1 || typeof i === 'undefined') {
|
|
|
|
states.has[gv] = states.size();
|
2010-02-11 01:57:33 -05:00
|
|
|
itemSet.edges[symbol] = states.size(); // store goto transition for table
|
2010-02-11 23:34:45 -05:00
|
|
|
states.push(g);
|
2010-02-11 01:57:33 -05:00
|
|
|
g.predecessors[symbol] = [stateNum];
|
|
|
|
} else {
|
|
|
|
itemSet.edges[symbol] = i; // store goto transition for table
|
|
|
|
states.item(i).predecessors[symbol].push(stateNum);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
2010-03-07 18:48:15 -05:00
|
|
|
var NONASSOC = 0;
|
2010-02-11 01:57:33 -05:00
|
|
|
lrGeneratorMixin.parseTable = function parseTable (itemSets) {
|
|
|
|
var states = [],
|
|
|
|
nonterminals = this.nonterminals,
|
|
|
|
operators = this.operators,
|
|
|
|
self = this,
|
|
|
|
s = 1, // shift
|
|
|
|
r = 2, // reduce
|
|
|
|
a = 3; // accept
|
|
|
|
|
|
|
|
// for each item set
|
|
|
|
itemSets.forEach(function (itemSet, k) {
|
|
|
|
var state = states[k] = {};
|
|
|
|
var action, stackSymbol;
|
|
|
|
|
|
|
|
// set shift and goto actions
|
|
|
|
for (stackSymbol in itemSet.edges) {
|
|
|
|
itemSet.forEach(function (item, j) {
|
|
|
|
// find shift and goto actions
|
|
|
|
if (item.markedSymbol == stackSymbol) {
|
|
|
|
var gotoState = itemSet.edges[stackSymbol];
|
|
|
|
if (nonterminals[stackSymbol]) {
|
|
|
|
// store state to go to after a reduce
|
|
|
|
//self.trace(k, stackSymbol, 'g'+gotoState);
|
2010-02-11 23:34:45 -05:00
|
|
|
state[self.symbols_[stackSymbol]] = gotoState;
|
2010-02-11 01:57:33 -05:00
|
|
|
} else {
|
|
|
|
//self.trace(k, stackSymbol, 's'+gotoState);
|
2010-03-07 18:48:15 -05:00
|
|
|
state[self.symbols_[stackSymbol]] = [s,gotoState];
|
2010-02-11 01:57:33 -05:00
|
|
|
}
|
|
|
|
}
|
|
|
|
});
|
|
|
|
}
|
|
|
|
|
|
|
|
// set accept action
|
|
|
|
itemSet.forEach(function (item, j) {
|
|
|
|
if (item.markedSymbol == self.EOF) {
|
|
|
|
// accept
|
2010-03-07 18:48:15 -05:00
|
|
|
state[self.symbols_[self.EOF]] = [a];
|
2010-02-11 01:57:33 -05:00
|
|
|
//self.trace(k, self.EOF, state[self.EOF]);
|
|
|
|
}
|
|
|
|
});
|
|
|
|
|
|
|
|
var allterms = self.lookAheads ? false : self.terminals;
|
|
|
|
|
|
|
|
// set reductions and resolve potential conflicts
|
|
|
|
itemSet.reductions.forEach(function (item, j) {
|
|
|
|
// if parser uses lookahead, only enumerate those terminals
|
|
|
|
var terminals = allterms || self.lookAheads(itemSet, item);
|
|
|
|
|
|
|
|
terminals.forEach(function (stackSymbol) {
|
2010-03-07 18:48:15 -05:00
|
|
|
action = state[self.symbols_[stackSymbol]];
|
2010-02-11 01:57:33 -05:00
|
|
|
var op = operators[stackSymbol];
|
2010-03-07 18:48:15 -05:00
|
|
|
|
2010-02-11 01:57:33 -05:00
|
|
|
// Reading a terminal and current position is at the end of a production, try to reduce
|
2010-03-07 18:48:15 -05:00
|
|
|
if (action || action && action.length) {
|
|
|
|
var sol = resolveConflict(item.production, op, [r,item.production.id], action[0] instanceof Array ? action[0] : action);
|
2010-02-11 01:57:33 -05:00
|
|
|
self.resolutions.push([k,stackSymbol,sol]);
|
|
|
|
if (sol.bydefault) {
|
|
|
|
self.conflicts++;
|
|
|
|
if (!self.DEBUG) {
|
|
|
|
self.warn('Conflict in grammar (state:',k, ', token:',stackSymbol, ")\n ", printAction(sol.r, self), "\n ", printAction(sol.s, self));
|
|
|
|
}
|
|
|
|
if (self.options.noDefaultResolve) {
|
2010-03-07 18:48:15 -05:00
|
|
|
if (!(action[0] instanceof Array))
|
|
|
|
action = [action];
|
2010-02-11 01:57:33 -05:00
|
|
|
action.push(sol.r);
|
|
|
|
}
|
|
|
|
} else {
|
2010-03-07 18:48:15 -05:00
|
|
|
action = sol.action;
|
2010-02-11 01:57:33 -05:00
|
|
|
}
|
|
|
|
} else {
|
2010-03-07 18:48:15 -05:00
|
|
|
action = [r,item.production.id];
|
2010-02-11 01:57:33 -05:00
|
|
|
}
|
|
|
|
if (action && action.length) {
|
|
|
|
state[self.symbols_[stackSymbol]] = action;
|
2010-03-07 18:48:15 -05:00
|
|
|
} else if (action === NONASSOC) {
|
|
|
|
state[self.symbols_[stackSymbol]] = undefined;
|
2010-02-11 01:57:33 -05:00
|
|
|
}
|
|
|
|
});
|
|
|
|
});
|
|
|
|
|
|
|
|
});
|
|
|
|
|
|
|
|
return states;
|
|
|
|
};
|
|
|
|
|
|
|
|
// resolves shift-reduce and reduce-reduce conflicts
|
|
|
|
function resolveConflict (production, op, reduce, shift) {
|
|
|
|
var sln = {production: production, operator: op, r: reduce, s: shift},
|
|
|
|
s = 1, // shift
|
|
|
|
r = 2, // reduce
|
|
|
|
a = 3; // accept
|
|
|
|
|
|
|
|
if (shift[0] === r) {
|
|
|
|
sln.msg = "Resolve R/R conflict (use first production declared in grammar.)";
|
|
|
|
sln.action = shift[1] < reduce[1] ? shift : reduce;
|
|
|
|
sln.bydefault = true;
|
|
|
|
return sln;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (production.precedence === 0 || !op) {
|
|
|
|
sln.msg = "Resolve S/R conflict (shift by default.)";
|
|
|
|
sln.bydefault = true;
|
|
|
|
sln.action = shift;
|
|
|
|
} else if (production.precedence < op.precedence ) {
|
|
|
|
sln.msg = "Resolve S/R conflict (shift for higher precedent operator.)";
|
|
|
|
sln.action = shift;
|
|
|
|
} else if (production.precedence === op.precedence) {
|
|
|
|
if (op.assoc === "right" ) {
|
|
|
|
sln.msg = "Resolve S/R conflict (shift for right associative operator.)";
|
|
|
|
sln.action = shift;
|
|
|
|
} else if (op.assoc === "left" ) {
|
|
|
|
sln.msg = "Resolve S/R conflict (reduce for left associative operator.)";
|
|
|
|
sln.action = reduce;
|
|
|
|
} else if (op.assoc === "nonassoc" ) {
|
|
|
|
sln.msg = "Resolve S/R conflict (no action for non-associative operator.)";
|
2010-03-07 18:48:15 -05:00
|
|
|
sln.action = NONASSOC;
|
2010-02-11 01:57:33 -05:00
|
|
|
}
|
|
|
|
} else {
|
|
|
|
sln.msg = "Resolve conflict (reduce for higher precedent production.)";
|
|
|
|
sln.action = reduce;
|
|
|
|
}
|
|
|
|
|
|
|
|
return sln;
|
|
|
|
}
|
|
|
|
|
|
|
|
lrGeneratorMixin.generate = function parser_generate (opt) {
|
|
|
|
opt = typal.mix.call({}, this.options, opt);
|
|
|
|
var code = "";
|
|
|
|
switch (opt.moduleType) {
|
|
|
|
case "js":
|
|
|
|
code = this.generateModule(opt);
|
|
|
|
break;
|
|
|
|
case "commonjs":
|
|
|
|
default:
|
|
|
|
code = this.generateCommonJSModule(opt);
|
|
|
|
}
|
|
|
|
|
|
|
|
return code;
|
|
|
|
};
|
|
|
|
|
|
|
|
lrGeneratorMixin.generateCommonJSModule = function generateCommonJSModule (opt) {
|
|
|
|
opt = typal.mix.call({}, this.options, opt);
|
|
|
|
var moduleName = opt.moduleName || "parser";
|
|
|
|
var out = this.generateModule(opt);
|
|
|
|
out += "\nif (typeof require !== 'undefined') {";
|
|
|
|
out += "\nexports.parser = "+moduleName+";";
|
|
|
|
out += "\nexports.parse = function () { return "+moduleName+".parse.apply("+moduleName+", arguments); }";
|
|
|
|
out += "\nexports.main = "+ String(opt.moduleMain || commonjsMain);
|
|
|
|
out += "\nif (require.main === module) {\n\texports.main(require(\"system\").args);\n}";
|
|
|
|
out += "\n}";
|
|
|
|
|
|
|
|
return out;
|
|
|
|
};
|
|
|
|
|
|
|
|
lrGeneratorMixin.generateModule = function generateModule (opt) {
|
|
|
|
opt = typal.mix.call({}, this.options, opt);
|
|
|
|
var moduleName = opt.moduleName || "parser";
|
|
|
|
var out = "/* Jison generated parser */\n";
|
|
|
|
out += (moduleName.match(/\./) ? moduleName : "var "+moduleName)+" = (function(){";
|
|
|
|
out += "\nvar parser = "+this.generateModule_();
|
2010-02-23 21:03:05 -05:00
|
|
|
if (this.lexer && this.lexer.generateModule) {
|
2010-02-11 01:57:33 -05:00
|
|
|
out += this.lexer.generateModule();
|
|
|
|
out += "\nparser.lexer = lexer;";
|
|
|
|
}
|
|
|
|
out += "\nreturn parser;\n})();";
|
|
|
|
|
|
|
|
return out;
|
|
|
|
};
|
|
|
|
|
|
|
|
lrGeneratorMixin.generateModule_ = function generateModule_ () {
|
|
|
|
var out = "{";
|
|
|
|
out += [
|
|
|
|
"trace: " + String(this.trace),
|
|
|
|
"yy: {}",
|
|
|
|
"symbols_: " + JSON.stringify(this.symbols_),
|
|
|
|
"terminals_: " + JSON.stringify(this.terminals_),
|
|
|
|
"productions_: " + JSON.stringify(this.productions_),
|
|
|
|
"performAction: " + String(this.performAction),
|
|
|
|
"table: " + JSON.stringify(this.table),
|
|
|
|
"parseError: " + String(this.parseError || parser.parseError),
|
|
|
|
"parse: " + String(parser.parse)
|
|
|
|
].join(",\n");
|
|
|
|
out += "};";
|
|
|
|
|
|
|
|
return out;
|
|
|
|
};
|
|
|
|
|
|
|
|
// default main method for generated commonjs modules
|
|
|
|
function commonjsMain (args) {
|
|
|
|
var cwd = require("file").path(require("file").cwd());
|
|
|
|
if (!args[1])
|
|
|
|
throw new Error('Usage: '+args[0]+' FILE');
|
|
|
|
var source = cwd.join(args[1]).read({charset: "utf-8"});
|
|
|
|
this.parse(source);
|
|
|
|
}
|
|
|
|
|
|
|
|
// debug mixin for LR parser generators
|
|
|
|
|
|
|
|
function printAction (a, gen) {
|
|
|
|
var s = a[0] == 1 ? 'shift '+gen.symbols[a[1]] :
|
|
|
|
a[0] == 2 ? 'reduce by '+gen.productions[a[1]] :
|
|
|
|
'accept' ;
|
|
|
|
|
|
|
|
return s;
|
|
|
|
}
|
|
|
|
|
|
|
|
var lrGeneratorDebug = {
|
|
|
|
beforeparseTable: function () {
|
|
|
|
this.trace("Building parse table.");
|
|
|
|
},
|
|
|
|
afterparseTable: function () {
|
|
|
|
var self = this;
|
|
|
|
if (this.conflicts > 0) {
|
|
|
|
this.resolutions.forEach(function (r, i) {
|
|
|
|
if (r[2].bydefault) {
|
2010-02-12 13:35:34 -05:00
|
|
|
self.warn('Conflict at state:',r[0], ', Token:',r[1], "\n ", printAction(r[2].r, self), "\n ", printAction(r[2].s, self));
|
2010-02-11 01:57:33 -05:00
|
|
|
}
|
|
|
|
});
|
2010-02-12 13:35:34 -05:00
|
|
|
this.trace("\n"+this.conflicts+" Conflict(s) found in grammar.");
|
2010-02-11 01:57:33 -05:00
|
|
|
}
|
|
|
|
this.trace("Done.");
|
|
|
|
},
|
|
|
|
aftercanonicalCollection: function (states) {
|
|
|
|
var trace = this.trace;
|
|
|
|
trace("\nItem sets\n------");
|
|
|
|
|
|
|
|
states.forEach(function (state, i) {
|
|
|
|
trace("\nitem set",i,"\n"+state.join("\n"), '\ntransitions -> ', JSON.stringify(state.edges));
|
|
|
|
});
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
var parser = typal.beget();
|
|
|
|
|
|
|
|
lrGeneratorMixin.createParser = function createParser () {
|
|
|
|
var p = parser.beget();
|
|
|
|
|
|
|
|
p.init({
|
2010-02-11 23:34:45 -05:00
|
|
|
table: this.table,
|
2010-02-11 01:57:33 -05:00
|
|
|
productions_: this.productions_,
|
|
|
|
symbols_: this.symbols_,
|
2010-02-13 18:19:59 -05:00
|
|
|
terminals_: this.terminals_,
|
2010-02-11 01:57:33 -05:00
|
|
|
performAction: this.performAction
|
|
|
|
});
|
|
|
|
|
|
|
|
// for debugging
|
|
|
|
p.productions = this.productions;
|
|
|
|
|
|
|
|
// backwards compatability
|
|
|
|
p.generate = this.generate;
|
|
|
|
p.lexer = this.lexer;
|
|
|
|
p.generateModule = this.generateModule;
|
|
|
|
p.generateCommonJSModule = this.generateCommonJSModule;
|
|
|
|
p.generateModule_ = this.generateModule_;
|
|
|
|
|
|
|
|
return p;
|
|
|
|
};
|
|
|
|
|
|
|
|
parser.yy = {};
|
|
|
|
parser.trace = generator.trace;
|
|
|
|
parser.warn = generator.warn;
|
|
|
|
parser.error = generator.error;
|
|
|
|
|
|
|
|
parser.parseError = lrGeneratorMixin.parseError = function parseError (str, hash) {
|
|
|
|
throw new Error(str);
|
|
|
|
};
|
|
|
|
|
|
|
|
parser.parse = function parse (input) {
|
|
|
|
var self = this,
|
|
|
|
stack = [0],
|
|
|
|
vstack = [null], // semantic value stack
|
|
|
|
table = this.table,
|
|
|
|
yytext = '',
|
|
|
|
yylineno = 0,
|
|
|
|
yyleng = 0,
|
|
|
|
shifts = 0,
|
|
|
|
reductions = 0;
|
|
|
|
|
|
|
|
this.lexer.setInput(input);
|
|
|
|
this.lexer.yy = this.yy;
|
|
|
|
|
|
|
|
var parseError = this.yy.parseError = this.yy.parseError || this.parseError;
|
|
|
|
|
|
|
|
function lex() {
|
|
|
|
var token;
|
|
|
|
token = self.lexer.lex() || 1; // $end = 1
|
|
|
|
// if token isn't its numeric value, convert
|
|
|
|
if (typeof token !== 'number') {
|
|
|
|
token = self.symbols_[token];
|
|
|
|
}
|
|
|
|
return token;
|
|
|
|
};
|
|
|
|
|
|
|
|
var symbol, state, action, a, r, yyval={},p,len,ip=0,newState, expected;
|
2010-02-11 23:34:45 -05:00
|
|
|
symbol = lex();
|
2010-02-11 01:57:33 -05:00
|
|
|
while (true) {
|
|
|
|
// set first input
|
|
|
|
state = stack[stack.length-1];
|
|
|
|
// read action for current state and first input
|
|
|
|
action = table[state] && table[state][symbol];
|
|
|
|
|
2010-03-07 18:48:15 -05:00
|
|
|
if (typeof action === 'undefined' || !action.length || !action[0]) {
|
2010-02-11 01:57:33 -05:00
|
|
|
expected = [];
|
|
|
|
for (p in table[state]) if (this.terminals_[p] && p != 1) {
|
|
|
|
expected.push("'"+this.terminals_[p]+"'");
|
|
|
|
}
|
2010-02-13 18:19:59 -05:00
|
|
|
if (this.lexer.showPosition) {
|
|
|
|
parseError('Parse error on line '+(yylineno+1)+":\n"+this.lexer.showPosition()+'\nExpecting '+expected.join(', '),
|
|
|
|
{text: this.lexer.match, token: this.terminals_[symbol], line: this.lexer.yylineno, expected: expected});
|
|
|
|
} else {
|
|
|
|
parseError('Parse error on line '+(yylineno+1)+": Unexpected '"+this.terminals_[symbol]+"'",
|
|
|
|
{text: this.lexer.match, token: this.terminals_[symbol], line: this.lexer.yylineno, expected: expected});
|
|
|
|
}
|
2010-02-11 01:57:33 -05:00
|
|
|
}
|
|
|
|
|
|
|
|
// this shouldn't happen, unless resolve defaults are off
|
2010-03-07 18:48:15 -05:00
|
|
|
if (action[0] instanceof Array && action.length > 1) {
|
2010-02-11 01:57:33 -05:00
|
|
|
throw new Error('Parse Error: multiple actions possible at state: '+state+', token: '+symbol);
|
|
|
|
}
|
|
|
|
|
2010-03-07 18:48:15 -05:00
|
|
|
a = action;
|
2010-02-11 01:57:33 -05:00
|
|
|
|
|
|
|
switch (a[0]) {
|
|
|
|
|
|
|
|
case 1: // shift
|
|
|
|
shifts++;
|
|
|
|
|
|
|
|
stack.push(symbol);++ip;
|
|
|
|
yyleng = this.lexer.yyleng;
|
|
|
|
yytext = this.lexer.yytext;
|
|
|
|
yylineno = this.lexer.yylineno;
|
2010-02-11 23:34:45 -05:00
|
|
|
symbol = lex();
|
2010-02-11 01:57:33 -05:00
|
|
|
vstack.push(null); // semantic values or junk only, no terminals
|
|
|
|
stack.push(a[1]); // push state
|
|
|
|
break;
|
|
|
|
|
|
|
|
case 2: // reduce
|
|
|
|
reductions++;
|
|
|
|
|
|
|
|
len = this.productions_[a[1]][1];
|
|
|
|
|
|
|
|
// perform semantic action
|
|
|
|
yyval.$ = vstack[vstack.length-len]; // default to $$ = $1
|
|
|
|
r = this.performAction.call(yyval, yytext, yyleng, yylineno, this.yy, a[1], vstack);
|
|
|
|
|
|
|
|
if (typeof r !== 'undefined') {
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
|
|
|
|
// pop off stack
|
|
|
|
if (len) {
|
|
|
|
stack = stack.slice(0,-1*len*2);
|
|
|
|
vstack = vstack.slice(0, -1*len);
|
|
|
|
}
|
|
|
|
|
|
|
|
stack.push(this.productions_[a[1]][0]); // push nonterminal (reduce)
|
|
|
|
vstack.push(yyval.$);
|
|
|
|
// goto new state = table[STATE][NONTERMINAL]
|
|
|
|
newState = table[stack[stack.length-2]][stack[stack.length-1]];
|
|
|
|
stack.push(newState);
|
|
|
|
break;
|
|
|
|
|
|
|
|
case 3: // accept
|
|
|
|
|
2010-02-12 13:35:34 -05:00
|
|
|
this.reductionCount = reductions;
|
|
|
|
this.shiftCount = shifts;
|
2010-02-11 01:57:33 -05:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
return true;
|
|
|
|
};
|
|
|
|
|
|
|
|
parser.init = function parser_init (dict) {
|
|
|
|
this.table = dict.table;
|
|
|
|
this.performAction = dict.performAction;
|
|
|
|
this.productions_ = dict.productions_;
|
|
|
|
this.symbols_ = dict.symbols_;
|
|
|
|
this.terminals_ = dict.terminals_;
|
|
|
|
};
|
|
|
|
|
|
|
|
/*
|
|
|
|
* LR(0) Parser
|
|
|
|
* */
|
|
|
|
|
|
|
|
var lr0 = generator.beget(lookaheadMixin, lrGeneratorMixin, {
|
|
|
|
type: "LR(0)",
|
|
|
|
afterconstructor: function lr0_afterconstructor () {
|
|
|
|
this.buildTable();
|
|
|
|
}
|
|
|
|
});
|
|
|
|
|
|
|
|
var LR0Generator = exports.LR0Generator = lr0.construct();
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Simple LALR(1)
|
|
|
|
* */
|
|
|
|
|
|
|
|
var lalr = generator.beget(lookaheadMixin, lrGeneratorMixin, {
|
|
|
|
type: "LALR(1)",
|
|
|
|
|
|
|
|
afterconstructor: function (grammar, options) {
|
2010-02-11 23:34:45 -05:00
|
|
|
if (this.DEBUG) this.mix(lrGeneratorDebug, lalrGeneratorDebug); // mixin debug methods
|
2010-02-11 01:57:33 -05:00
|
|
|
|
|
|
|
options = options || {};
|
|
|
|
this.states = this.canonicalCollection();
|
2010-02-11 23:34:45 -05:00
|
|
|
this.terms_ = {};
|
2010-02-11 01:57:33 -05:00
|
|
|
|
|
|
|
var newg = this.newg = typal.beget(lookaheadMixin,{
|
|
|
|
oldg: this,
|
|
|
|
trace: this.trace,
|
2010-02-11 23:34:45 -05:00
|
|
|
nterms_: {},
|
2010-02-11 01:57:33 -05:00
|
|
|
DEBUG: false,
|
|
|
|
go_: function (r, B) {
|
|
|
|
r = r.split(":")[0]; // grab state #
|
|
|
|
B = B.map(function (b) { return b.slice(b.indexOf(":")+1)});
|
|
|
|
return this.oldg.go(r, B);
|
|
|
|
}
|
|
|
|
});
|
|
|
|
newg.nonterminals = {};
|
|
|
|
newg.productions = [];
|
|
|
|
|
|
|
|
this.inadequateStates = [];
|
|
|
|
|
|
|
|
// if true, only lookaheads in inadequate states are computed (faster, larger table)
|
|
|
|
// if false, lookaheads for all reductions will be computed (slower, smaller table)
|
|
|
|
this.onDemandLookahead = options.onDemandLookahead || false;
|
|
|
|
|
|
|
|
this.buildNewGrammar();
|
|
|
|
newg.computeLookaheads();
|
|
|
|
this.unionLookaheads();
|
|
|
|
|
|
|
|
this.table = this.parseTable(this.states);
|
|
|
|
},
|
|
|
|
|
|
|
|
lookAheads: function LALR_lookaheads (state, item) {
|
|
|
|
return (!!this.onDemandLookahead && !state.inadequate) ? this.terminals : item.follows;
|
|
|
|
},
|
|
|
|
go: function LALR_go (p, w) {
|
|
|
|
var q = parseInt(p);
|
|
|
|
for (var i=0;i<w.length;i++) {
|
|
|
|
q = this.states.item(q).edges[w[i]] || q;
|
|
|
|
}
|
|
|
|
return q;
|
|
|
|
},
|
|
|
|
goPath: function LALR_goPath (p, w) {
|
2010-02-11 23:34:45 -05:00
|
|
|
var q = parseInt(p),t,
|
2010-02-11 01:57:33 -05:00
|
|
|
path = [];
|
|
|
|
for (var i=0;i<w.length;i++) {
|
2010-02-11 23:34:45 -05:00
|
|
|
t = w[i] ? q+":"+w[i] : '';
|
|
|
|
if (t) this.newg.nterms_[t] = q;
|
|
|
|
path.push(t);
|
2010-02-11 01:57:33 -05:00
|
|
|
q = this.states.item(q).edges[w[i]] || q;
|
2010-02-11 23:34:45 -05:00
|
|
|
this.terms_[t] = w[i];
|
2010-02-11 01:57:33 -05:00
|
|
|
}
|
|
|
|
return {path: path, endState: q};
|
|
|
|
},
|
|
|
|
// every disjoint reduction of a nonterminal becomes a produciton in G'
|
|
|
|
buildNewGrammar: function LALR_buildNewGrammar () {
|
|
|
|
var self = this,
|
|
|
|
newg = this.newg;
|
|
|
|
|
|
|
|
this.states.forEach(function (state, i) {
|
|
|
|
state.forEach(function (item) {
|
|
|
|
if (item.dotPosition === 0) {
|
|
|
|
// new symbols are a combination of state and transition symbol
|
|
|
|
var symbol = i+":"+item.production.symbol;
|
2010-02-11 23:34:45 -05:00
|
|
|
self.terms_[symbol] = item.production.symbol;
|
|
|
|
newg.nterms_[symbol] = i;
|
2010-02-11 01:57:33 -05:00
|
|
|
if (!newg.nonterminals[symbol])
|
|
|
|
newg.nonterminals[symbol] = new Nonterminal(symbol);
|
|
|
|
var pathInfo = self.goPath(i, item.production.handle);
|
|
|
|
var p = new Production(symbol, pathInfo.path, newg.productions.length);
|
|
|
|
newg.productions.push(p);
|
|
|
|
newg.nonterminals[symbol].productions.push(p);
|
|
|
|
|
|
|
|
// store the transition that get's 'backed up to' after reduction on path
|
|
|
|
var handle = item.production.handle.join(' ');
|
2010-02-11 23:34:45 -05:00
|
|
|
var goes = self.states.item(pathInfo.endState).goes;
|
|
|
|
if (!goes[handle])
|
|
|
|
goes[handle] = [];
|
|
|
|
goes[handle].push(symbol);
|
2010-02-11 01:57:33 -05:00
|
|
|
|
|
|
|
//self.trace('new production:',p);
|
|
|
|
}
|
|
|
|
});
|
|
|
|
if (state.inadequate)
|
|
|
|
self.inadequateStates.push(i);
|
|
|
|
});
|
|
|
|
},
|
|
|
|
unionLookaheads: function LALR_unionLookaheads () {
|
|
|
|
var self = this,
|
|
|
|
newg = this.newg,
|
|
|
|
states = !!this.onDemandLookahead ? this.inadequateStates : this.states;
|
|
|
|
|
|
|
|
states.forEach(function union_states_forEach (i) {
|
|
|
|
var state = typeof i === 'number' ? self.states.item(i) : i,
|
|
|
|
follows = [];
|
2010-02-11 23:34:45 -05:00
|
|
|
if (state.reductions.length)
|
2010-02-11 01:57:33 -05:00
|
|
|
state.reductions.forEach(function union_reduction_forEach (item) {
|
2010-02-11 23:34:45 -05:00
|
|
|
var follows = {};
|
|
|
|
for (var k=0;k<item.follows.length;k++) {
|
|
|
|
follows[item.follows[k]] = true;
|
|
|
|
}
|
2010-02-11 01:57:33 -05:00
|
|
|
state.goes[item.production.handle.join(' ')].forEach(function reduction_goes_forEach (symbol) {
|
|
|
|
newg.nonterminals[symbol].follows.forEach(function goes_follows_forEach (symbol) {
|
2010-02-11 23:34:45 -05:00
|
|
|
var terminal = self.terms_[symbol];
|
|
|
|
if (!follows[terminal]) {
|
|
|
|
follows[terminal]=true;
|
2010-02-11 01:57:33 -05:00
|
|
|
item.follows.push(terminal);
|
2010-02-11 23:34:45 -05:00
|
|
|
}
|
2010-02-11 01:57:33 -05:00
|
|
|
});
|
|
|
|
});
|
|
|
|
//self.trace('unioned item', item);
|
|
|
|
});
|
|
|
|
});
|
|
|
|
}
|
|
|
|
});
|
|
|
|
|
|
|
|
var LALRGenerator = exports.LALRGenerator = lalr.construct();
|
|
|
|
|
|
|
|
// LALR generator debug mixin
|
|
|
|
|
|
|
|
var lalrGeneratorDebug = {
|
|
|
|
trace: function trace () {
|
|
|
|
Jison.print.apply(null, arguments);
|
|
|
|
},
|
|
|
|
beforebuildNewGrammar: function () {
|
|
|
|
this.trace(this.states.size()+" states.");
|
|
|
|
this.trace("Building lookahead grammar.");
|
|
|
|
},
|
|
|
|
beforeunionLookaheads: function () {
|
|
|
|
this.trace("Computing lookaheads.");
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
2010-02-11 23:34:45 -05:00
|
|
|
/*
|
2010-02-11 01:57:33 -05:00
|
|
|
* Lookahead parser definitions
|
|
|
|
*
|
|
|
|
* Define base type
|
|
|
|
* */
|
|
|
|
var lrLookaheadGenerator = generator.beget(lookaheadMixin, lrGeneratorMixin, {
|
|
|
|
afterconstructor: function lr_aftercontructor () {
|
|
|
|
this.computeLookaheads();
|
|
|
|
this.buildTable();
|
|
|
|
}
|
|
|
|
});
|
|
|
|
|
|
|
|
/*
|
|
|
|
* SLR Parser
|
|
|
|
* */
|
|
|
|
var SLRGenerator = exports.SLRGenerator = lrLookaheadGenerator.construct({
|
|
|
|
type: "SLR(1)",
|
|
|
|
|
|
|
|
lookAheads: function SLR_lookAhead (state, item) {
|
|
|
|
return this.nonterminals[item.production.symbol].follows;
|
|
|
|
}
|
|
|
|
});
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
* LR(1) Parser
|
|
|
|
* */
|
|
|
|
var lr1 = lrLookaheadGenerator.beget({
|
|
|
|
type: "Canonical LR(1)",
|
|
|
|
|
|
|
|
lookAheads: function LR_lookAheads (state, item) {
|
|
|
|
return item.follows;
|
|
|
|
},
|
|
|
|
Item: lrGeneratorMixin.Item.prototype.construct({
|
2010-02-11 23:34:45 -05:00
|
|
|
afterconstructor: function () {
|
|
|
|
this.id = this.production.id+'a'+this.dotPosition+'a'+this.follows.sort().join(',');
|
|
|
|
},
|
2010-02-11 01:57:33 -05:00
|
|
|
eq: function (e) {
|
2010-02-11 23:34:45 -05:00
|
|
|
return e.id === this.id;
|
2010-02-11 01:57:33 -05:00
|
|
|
}
|
|
|
|
}),
|
|
|
|
|
|
|
|
closureOperation: function LR_ClosureOperation (itemSet /*, closureSet*/) {
|
|
|
|
var closureSet = new this.ItemSet();
|
|
|
|
var self = this;
|
|
|
|
|
|
|
|
var set = itemSet,
|
|
|
|
itemQueue, syms = {};
|
|
|
|
|
|
|
|
do {
|
|
|
|
itemQueue = new Set();
|
|
|
|
closureSet.concat(set);
|
|
|
|
set.forEach(function (item) {
|
|
|
|
var symbol = item.markedSymbol;
|
|
|
|
var b;
|
|
|
|
|
|
|
|
// if token is a nonterminal, recursively add closures
|
|
|
|
if (symbol && self.nonterminals[symbol]) {
|
|
|
|
b = self.first(item.remainingHandle());
|
|
|
|
if (b.length === 0) b = item.follows;
|
|
|
|
self.nonterminals[symbol].productions.forEach(function (production) {
|
|
|
|
var newItem = new self.Item(production, 0, b);
|
|
|
|
if(!closureSet.contains(newItem) && !itemQueue.contains(newItem)) {
|
|
|
|
itemQueue.push(newItem);
|
|
|
|
}
|
|
|
|
});
|
|
|
|
} else if (!symbol) {
|
|
|
|
// reduction
|
|
|
|
closureSet.reductions.push(item);
|
|
|
|
}
|
|
|
|
});
|
|
|
|
|
|
|
|
set = itemQueue;
|
|
|
|
} while (!itemQueue.isEmpty());
|
|
|
|
|
|
|
|
return closureSet;
|
|
|
|
}
|
|
|
|
});
|
|
|
|
|
|
|
|
var LR1Generator = exports.LR1Generator = lr1.construct();
|
|
|
|
|
|
|
|
/*
|
|
|
|
* LL Parser
|
|
|
|
* */
|
|
|
|
var ll = generator.beget(lookaheadMixin, {
|
|
|
|
type: "LL(1)",
|
|
|
|
|
|
|
|
afterconstructor: function ll_aftercontructor () {
|
|
|
|
this.computeLookaheads();
|
|
|
|
this.table = this.parseTable(this.productions);
|
|
|
|
},
|
|
|
|
parseTable: function llParseTable (productions) {
|
|
|
|
var table = {},
|
|
|
|
self = this;
|
|
|
|
productions.forEach(function (production, i) {
|
|
|
|
var row = table[production.symbol] || {};
|
|
|
|
var tokens = production.first;
|
|
|
|
if (self.nullable(production.handle)) {
|
|
|
|
Set.union(tokens, self.nonterminals[production.symbol].follows);
|
|
|
|
}
|
|
|
|
tokens.forEach(function (token) {
|
|
|
|
if (row[token]) {
|
|
|
|
row[token].push(i);
|
|
|
|
self.conflicts++;
|
|
|
|
} else {
|
|
|
|
row[token] = [i];
|
|
|
|
}
|
|
|
|
});
|
|
|
|
table[production.symbol] = row;
|
|
|
|
});
|
|
|
|
|
|
|
|
return table;
|
|
|
|
}
|
|
|
|
});
|
|
|
|
|
|
|
|
var LLGenerator = exports.LLGenerator = ll.construct();
|
|
|
|
|
|
|
|
Jison.Generator = function Jison_Generator (g, options) {
|
|
|
|
var opt = typal.mix.call({}, g.options, options);
|
|
|
|
switch (opt.type) {
|
|
|
|
case 'lr0':
|
|
|
|
return new LR0Generator(g, opt);
|
|
|
|
case 'slr':
|
|
|
|
return new SLRGenerator(g, opt);
|
|
|
|
case 'lr':
|
|
|
|
return new LR1Generator(g, opt);
|
|
|
|
case 'll':
|
|
|
|
return new LLGenerator(g, opt);
|
|
|
|
case 'lalr':
|
|
|
|
default:
|
|
|
|
return new LALRGenerator(g, opt);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return function Parser (g, options) {
|
|
|
|
var opt = typal.mix.call({}, g.options, options);
|
|
|
|
var gen;
|
|
|
|
switch (opt.type) {
|
|
|
|
case 'lr0':
|
|
|
|
gen = new LR0Generator(g, opt);
|
|
|
|
case 'slr':
|
|
|
|
gen = new SLRGenerator(g, opt);
|
|
|
|
case 'lr':
|
|
|
|
gen = new LR1Generator(g, opt);
|
|
|
|
case 'll':
|
|
|
|
gen = new LLGenerator(g, opt);
|
|
|
|
case 'lalr':
|
|
|
|
default:
|
|
|
|
gen = new LALRGenerator(g, opt);
|
|
|
|
}
|
|
|
|
return gen.createParser();
|
|
|
|
}
|
|
|
|
|
|
|
|
})();
|
|
|
|
|
|
|
|
exports.main = function main (args) {
|
|
|
|
//var parser = new require("args").Parser();
|
|
|
|
var fs = require("file");
|
|
|
|
gfile = fs.path(fs.cwd()).join(args[1]);
|
|
|
|
|
|
|
|
// try to parse as JSON, else use BNF parser
|
|
|
|
if (gfile.extension() === '.json') {
|
|
|
|
var grammar = JSON.parse(gfile.read({charset: "utf-8"}));
|
|
|
|
} else if (gfile.extension() === '.jison') {
|
|
|
|
var grammar = require("jison/bnf").parse(gfile.read({charset: "utf-8"}));
|
|
|
|
}
|
|
|
|
|
|
|
|
var opt = grammar.options || {};
|
|
|
|
|
|
|
|
// lexer file
|
|
|
|
if (args[2]) {
|
|
|
|
var lfile = fs.path(fs.cwd()).join(args[2]);
|
|
|
|
|
|
|
|
// try to parse as JSON, else use BNF parser
|
|
|
|
if (lfile.extension() === '.json') {
|
|
|
|
grammar.lex = JSON.parse(lfile.read({charset: "utf-8"}));
|
|
|
|
} else if (lfile.extension() === '.jisonlex') {
|
|
|
|
grammar.lex = require("jison/jisonlex").parse(lfile.read({charset: "utf-8"}));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!opt.moduleName)
|
|
|
|
opt.moduleName = gfile.basename().replace(new RegExp(gfile.extension()+"$"), "");
|
|
|
|
if (!opt.moduleType)
|
|
|
|
opt.moduleType = "commonjs";
|
|
|
|
|
|
|
|
var generator = new Jison.Generator(grammar, opt);
|
|
|
|
fname = fs.path(fs.cwd()).join(opt.moduleName + ".js"),
|
|
|
|
source = generator.generate(opt),
|
|
|
|
stream = fname.open("w");
|
|
|
|
|
|
|
|
stream.print(source);
|
|
|
|
stream.close();
|
|
|
|
};
|
|
|
|
|