Difference between revisions of "Constrained Dynamic Parser"

This is being worked at and not in a usable state for now...

This page describes the requirements for a generic parser for the Event-B Mathematical Language.

A first design proposal is also drafted.

Requirements

In order to be usable mathematical extensions require that the event-b mathematical language syntax can be extended by the final user.

The lexical analyser and the syntaxic parser thus have to be extensible in a simple enough way (from a user point of vue).

Requirements Exported by the Current Language Design

Operator Priority

• operator are defined by group,
• group of operator have a defined precedences,
• there may be precedences defined inside groups.

Operator Compatibility

• a compatibility table defines allowed associativities inside a group,
• a compatibility table defines allowed associativities between groups (it allows to forbid a syntaxic construction like $f(x)^{-1}\;$)
nota: this requirement was added afterwards with consistency in mind.

Expected Extension Schemes

We do want to at least define operators of the following form :

• infix : $a + b\;$ or $a \vdash b : c\;$
• prefix : $\neg a\;$
• postfix : $R^*\;$
• closed : $\|a\|\;$
• parentheses sugar : $(a +b) * c\;$
• functional postfix : $a \langle b \rangle\;$
• functional prefix : $\langle b \rangle f\;$
• lambda like : $\lambda x\mapsto y . P | E\;$
• Union like : $\Union\{ e \mid P\}$ or $\Union\{ x,y . P \mid e\}$
• set like : $\{a, b, c + e\}\;$ or $\{ e \mid P\}\;$ or $\{x,y . P \mid e\}\;$

We also want to define precisely precedence and associativity between existing and new operators.

Requirements exported by the dynamic feature

• the precedence should not be enumerated, which would hinder extensibility,

but defined by a relation, like: '+' < '*' and '*' < '**', ...

Design Alternatives

Three design alternatives where identified:

Make Existing Parser Extensible

The existing parser is a LL recursive descent parser generated by the Coco/R Compiler Generator, which makes extensive use of pre-computed lookahead, which makes it very difficult to be transformed in a parser witgh enough extensibility.

Parser Combinator

This paper is interesting in its proposal of using an acyclic graph to define operator precedence.

Pratt Parser

Note that a pratt parser is essentially a parser for expressions (where

Limitations

In a first implementation, we do not want to support several alternative for one symbol (using backtracking -- see below). As a side effect, it won't be possible to define new lambda like, union like and sets like terms, or terms that use one of the following symbol: λ, ⋂, ⋃ or ∣ expression parts are themselves expressions)... But it may be tweaked to limit the accepted sub-expressions or to enforce some non-expression parts inside an expression. More on that hereafter.

Design Proposal

The chosen design is based on Pratt parser and is explained hereafter.

Main Concepts

symbol
a symbol is a lexem (finite, $\Union$, ...)or a lexem class (number, identifier) known by the parser.
group
each symbol belongs to one and only one group (binary relation, assignment substitution,...).
Groups are used to ease the definition of a hierarchy between operators in order to ease the definition of precedence and compatibility.
symbol compatibility
a relation telling if inside a group two symbol are compatibles.
Two symbol are compatibles is they can be parsed without parentheses (see Event-B Mathematical Language for more on compatibility).
group compatibility
a relation telling if two groups are compatibles.
Two symbol of different groups are said compatibles if their group are compatibles.
symbol associativity
a relation which gives the associative precedence of symbols inside a given group.
group associativity
a relation which gives the associative precedence of groups.
Two symbol of different groups have a relative precedence given by the associative precedence of their groups.
generic parser
the parser which is defined.
This parser may be called recursively (an infix expression being build upon a left and right sub-expression) and is often the default parser used when building the AST associated with a symbol.
specific parser
some specific parser, predefined as helper functions to build specific AST.
nonterminal
A nonterminal is a set of symbol. Nonterminals are used for acceptance or rejection of the AST produced by a parser on a sub-formula.
nonterminals may be defined by mean of symbols, groups or other nonterminals.

Proposed User Interface

The hereafter proposition does not preclude the final UI from taќing other (more user friendly) form). But the concepts should probably stay the same

We propose that the user will be able to create several predefined type of symbols (new types may certainly be further added by mean of eclipse extension points). When defining a symbol, the user can choose the parser used to parse the sub-expressions (default to generic parser, but a specific parser may be used) and can enforce the form of the resulting parsed sub-expression in terms of their AST root type.

Predefined Symbol Builders

The proposed predefined symbol builders are:

infix
For example predicate conjunction which expects two predicates for left and right parts may be defined by infix("∧", gid="(logic pred)", expected_nonterm=["predicate", "predicate"]), saying that '∧' is an operator belonging to the logic predicate group which accept a predicate on left and right.
As another example, substitution :∣ of group '(infix subst)' which expect one ident list for left term and and a predicate for the right one may be defined by infix(":∣", "(infix subst)", expected_nonterm=["ident_list", "predicate"])
prefix
prefix("¬", "(not pred)", expected_nonterm=["predicate"])
atomic
atomic("⊤", "(atomic pred)")
prefix closed
used to build prefix symbol which must be followed by enclosing symbol(brackets, parentheses,..). For example finite may be defined by: prefix_closed("finite", "(", ")","(atomic pred)", expected_nonterm=["expression"])
postfix; postfix("∼", gid="(unary relation)", expected_nonterm=["expression"])
quantified
quantified("∀", "·", expected_nonterm=["predicate"]
lambda like
lambda_like("λ", "·", u"∣", gid="(quantification)", expected_nonterm=["ident list", "predicate", "expression"])
quantified union like
Union_like("⋃", "·", "∣", gid="(quantification)")
closed sugar
use for parentheses added to enforce associativity: closed_sugar("(", ")")
functional postfix
functional_postfix("(", ")", gid="(functional)", expected_nonterm=["expression", "expression"])

Some right associative versions of those symbol builders may also be supplied (infix_right, or a specific parameter right associative parameter may also be provided to the infix symbol builder).

TODO: Definition and use of nonterminals'

Predefined Specific Parser

• expression lists
• ident lists
• maplet tree (for lambda)

Defining Compatibility and Precedences

Inside a group, compatibility may be defined like:

g = group["(binop)"]
[   ("∪", "∪"),
("∩", "∩"),
("∩", "∖"),
("∩", "▷"),
("∩", "⩥"),
("∘", "∘"),
(";", ";"),
(";", "▷"),
(";", "⩥"),
("", ""),
("◁", "∩"),
("◁", "∖"),
("◁", ";"),
("◁", "⊗"),
("◁", "▷"),
("◁", "⩥"),
("⩥", "∩"),
("⩥", "∖"),
("⩥", ";"),
("⩥", "⊗"),
("⩥", "▷"),
("⩥", "⩥"),
("×", "×"), ]

which defines the compatibility table for expression binary operator as defined in Event-B Mathematical Language.

Inside a group, precedence may be defined like:

g = group["(arithmetic)"]
[   ("^", "÷"),
("^", "∗"),
("^", "+"),
("^", "−"),
("^", "mod"),
("+", "∗"),
("−", "∗"),
("+", "÷"),
("−", "÷"),
("+", "mod"),
("−", "mod"), ] )

which defines ^ as having the least precedence amongst the operators of the expression arithmetic group.

Similarly, precedences and compatibilities between groups may be defined :

group_compatibility.add_all( [ "(binop)", "(interval)", (arithmetic)", "(unary relation)", "(unary relation)", "(bound unary)",....])

Some Difficulties

Some difficulties appear when trying to implement an Event-B parser with the preceding scheme:

• A symbol may be declared by several different builders (think of unary and infix minus for example).
Pratt parsers are able to cope with a symbol appearing either at the beginning of an expression or inside an expression (see led and nud hereafter). For example, unary minus and infix minus can be both defined by the proposed scheme. However, in their default form, pratt parser failed in parsing symbol like '{' which appears at the beginning of extension sets, comprehension sets and quantified sets.
SOLUTION #1: The existing Event-B parser uses an infinite lookahead to determine which one of those three sets is actually being parsed.
This solution hinders the extensibility of the parser, because the addition of a symbol requires to modify the lookahead.
However, such an issue may be bypassed, at least in a first version, in enforcing some usage rules. Thus, a new symbol contributed by an extension but used in the lookahead may be rejected as in invalid symbol.
SOLUTION #2: Implementing a backtracking parser.
More precisely, it means that the parsing is attempted on every alternative (for example, on the three types of sets parser: extension, comprehension, and quantified). The parsing is successful if and only if one of these attempts lead to a result (i.e., it is a failure if several attempts or none lead to a result). All the error are collected and returned to the user in case of no match.
It is a good solution provided that we use memorisation to alleviate the cost of parsing formulas that require heavy backtracking.
• Some binary infix operators are in fact parsed as n-ary operators (with more than 2 childrens if possible).
An attribute is required to specify if an operator shall be flattened (i.e., parsed as an n-ary operator) or not. It is indeed impossible to "deduce" this information (e.g., the minus operator, which is left associative and self-compatible, shall not be flattened).
The flattening should then be done by the parser associated to every symbol.
• The returned parsing errors shall be meaningful for users (e.g., the "Missing parenthesis" message may help to solve the problem when the "Invalid expression" message is too imprecise)
TODO: how does it impact actual design?.
• Defining a grammar or a grammar extension is not an easy job, even with those simple mechanisms.
It would certainly be a good idea to implement a user interface to test the grammar, so that an extension designer has a way to ensure that he agrees with the parser on the semantic of his extension (or at least on the AST obtained by using it). For example, the representation of the syntax tree returned by Formula.getSyntaxTree could be compared to an expected entered string.

Limitations

In a first implementation, we do not want to support several alternative for one symbol (using backtracking -- see here before). As a side effect, it won't be possible to define new lambda like, union like and sets like terms, or terms that use one of the following symbol: λ, ⋂, ⋃ or ∣

Core Algorithm

Rough description

The algorithm is based on a table, associating a SymbolParser for each known operator (or lexem class). The MainParser then constructs the AST by using the led or nud parser instance attached to each symbol, taking into account symbol information such as group and precedence for constructing the resulting AST.

Each led or nud parser instance may in turn make recursive calls to either the standard MainParser or to ad-hoc parsers. The choice of the subparsers used is given when the SymbolParser is created.

Parsing an Example

We will try to explain how the parser manage to build an AST from the input 2

• 3 + a.

MainParser will try to apply the nud function of the parser associated to the literal symbol valued with '2', which simply return an AST node (literal 2). MainParser will then passed this AST to the led function of the parser associated to the '*' operator, which will use it for its left child and will then call MainParser to recognize its right term.

This new call to MainParser will in turn build the AST node (literal 3) and then will return, because precedences won't allow to build a ('+' (literal 3) (name a)) sub-expression.

The initial MainParser call now get the following left term: ('*' (literal 2) (literal 3)) and then call the led function associated to the '+' operator, which will then call MainParser which will return the simple AST (ident a) and build the resulting AST: ('+' ('*' (literal 2) (literal 3)) (name a) ).

Main Objects

KnownSymbols
a table associating a SymbolParser for each known symbol.
Tokenizer
tokenizer
which provides the SymbolParser associated to the next lexem of the formula being parsed.
TODO: shall be a singleton ? Define how is it use by Parser object
SymbolParser
a Parser for a specific symbol. It should at least have the following attributes:
• Array<LedParser>
led_parsers
for parsing the expression associated to the symbol
id
when it appears in the middle of an expression (left denotation)
• Array<NudParser>
nud_parsers
for parsing the expression associated to the symbol
id
when it appears at the beginning of an expression (null denotation)
• String
id
which is the lexem concerned by those parsers.
• String
gid
designing the group to which the
id
symbol belongs to.
• bool flat_ast
true if the associated binary AST shall be flatten in an n-ary AST.
and the following members:
• AST led(AST left): which should iterate with backtracking over led_parsers.led()
• AST nud(AST left): which should iterate with backtracking over nud_parsers.nud()
Note that parsers in led (respectively nud) will be tried one after another (with backtracking to the original state in between). If several of them succeed, an error is returned to the user, else the AST generated by the only one succeeding is returned. If none succeeded, an error is raised.
LedOrNudParser
An abstract parser, that may be used by the MainParser, which should have the following attributes
• Array<Parser>
subparsers
: an array of sub-parsers used by the
parse
method. They defaults to
MainParser
, but may be set to specific parsers.
• Array<NonterminalSet>
nonterminals
: set of nonterminals that are expected to be return by the parse function of the associated sub-parser (with the same index). If the root of the AST returned by the parse function is not one of the expected nonterminals, then an error is raised stating which non-terminals were expected, and which one was found.
LedParser
An abstract parser which inherits from LedOrNudParser and provides the following method:
• AST led(AST left): which given an already parsed left AST will construct the AST associated to the lexem being parsed, by analysing the part of the formula on its right. This method is responsible for verifying that the returned AST has one of the expected type (see LedOrNudParser#
nonterminals
)
NudParser
An abstract parser which inherits from LedOrNudParser and provides the following method:
• AST nud(): which build an AST associated to the lexem being parsed, by analysing the part of the formula on its right. This method is responsible for verifying that the returned AST has one of the expected type (see LedOrNudParser#
nonterminals
)
Parser
An abstract class of a recursive descent like parser which is used by the led or nud functions to build parts from AST. It may be instantiated by the MainParser or by any specific parser (used to parse list of identifier, or other structures that do not fit well in the general Pratt parser schema ...).
and the following method:
• AST parse(): this method should consume tokens provided by the tokenizer and returns an AST. It may use several
subparsers
.

MainParser
This class inherits from Parser and its
parse
function implements the core of the pratt algorithm.
Without taking into account backtracking, the algorithm reads:
def parse(reckognized_token, has_right_associativity=False):
global token
t = token                    # t is the processed symbol
token = tokenizer.next()         # token is the next symbol
left = t.nud()               # a new expression is being parsed (null denotation : nothing on the left)
while has_less_priority(reckognized_token, token, has_right_associativity):
t = token
token = next_token()     # if precedence allows it, we parse next expression
left = t.led(left)       # has being left expression of current symbol (left denotation)
return left

The has_less_priority function used the compatibility and precedence relations to determine if the next token should be parsed as a part of the current expression of should be left over for the calling parser. I may looks like:

• when the recognize symbol is left associative:
def has_less_priority_than(self, other , is_right_associative=False):
# every operator of group 0 is not associative
if (self.gid == u"0" and other.gid == u"0"):
return False
if (self.gid, other.gid) in group_associativity:
return True
elif (other.gid, self.gid) in group_associativity:
return False
elif (other.gid, self.gid) not in group_compatibility:
raise SyntaxError
elif self.gid == other.gid:
if (self.id, other.id) in group[self.gid].operator_associativity:
return True # right associativity
if (other.id, self.id) in group[self.gid].operator_associativity:
return False # left associativity
if  group[self.gid].is_compatible(self.id, other.id): #raise SyntaxError if not compatible
return False # left associativity is chosen for
# compatible operators
raise AssertionError("Should not be reached")
else:
return False
• when the recognize symbol is right associative:
def has_less_priority_than(self, other , is_right_associative=False):
if (self.gid == u"0" and other.gid == u"0"):
return False
if (self.gid, other.gid) in group_associativity:
return True
elif (other.gid, self.gid) in group_associativity:
return False
elif (other.gid, self.gid) not in group_compatibility:
raise SyntaxError
elif self.gid == other.gid:
if (self.id, other.id) in group[self.gid].operator_associativity:
return False # lef associativity
if (other.id, self.id) in group[self.gid].operator_associativity:
return True # right associativity
if group[self.gid].is_compatible(self.id, other.id) :
return True
raise AssertionError("Should not be reached")
else:
return False
The parsing of a new expression is then initiated by:
token = tokenizer.next()
ast = MainParser.parse(KnownSymbols["(eof)"])

identListParser
inherits from Parser. Its
parse
function implements the parsing of an ident list (like $a,b,c\;$).
identMapletListParser
inherits from Parser . Its
parse
function implements the parsing of a maplet list as used in lambda expression (like $a \mapsto (b \mapsto c)\;$).
expressionListParser
inherits from Parser . Its
parse
function implements a list of expression as used in the extension set expression.
infixParser
inherits from LedParser. Its
led

function implements something like (pseudocode):

led(AST left, Symbol value):
ast = new AST(this.id, this.gid, value)
ast.first = left
verify_expected_group(ast.first, this.expected_non_terminals[0])
ast.second = this.subparsers[0].parse(ast, is_right_associative=False)
# is_right_associative parameter should be set to True for a right associative infix parser
verify_expected_group(ast.second, this.expected_non_terminals[1])
if this.flat:
# flatten the AST if ast.first or ast.second have the same root operator tha the AST being built.
return ast

TODO: each parser should be passed a context to allow rewriting of bound identifiers on the AST being constructed. A special rewriting should be done for comprehension sets and alike, because the bouns variables are not known before the parsing of the actual expression

atomicParser
inherits from NudParser. Its
nud

function implements something like (pseudocode):

nud(Symbol value):
ast = new AST(this.id, this.gid, value)
return ast
postfixParser
inherits from LedParser. Its
led

function implements something like (pseudocode):

led(AST left, Symbol value):
ast = new AST(this.id, this.gid, value)
ast.first = left
verify_expected_group(ast.first, this.expected_non_terminals[0])
if this.flat:
# flatten the AST if ast.first or ast.second have the same root operator tha the AST being built.
return ast
prefixParser
inherits from NudParser. Its
nud

function implements something like (pseudocode):

nud(Symbol value):
ast = new AST(this.id, this.gid, value)
ast.first = this.subparsers[0].parse(ast, is_right_associative=False)
verify_expected_group(ast.first, this.expected_non_terminals[0])
if this.flat:
# flatten the AST if ast.first or ast.second have the same root operator tha the AST being built.
return ast
functionalPostfixParser
inherits from LedParser. It has a supplementary
close_symbol
attribute, and its
led
function implements something like (pseudocode):
led(AST left, Symbol value):
ast = new AST(this.id, this.gid, value)
ast.first = left
verify_expected_group(ast.first, this.expected_non_terminals[0])
ast.second = this.subparsers[0].parse(ast, is_right_associative=False)
# is_right_associative parameter should be set to True for a right associative infix parser
verify_expected_group(ast.second, this.expected_non_terminals[1])
this.tokenizer.advance(this.close_symbol) #consume the close symbol or raise a SyntaxError
return ast

Open Questions

How does the user specify the expected AST?

Some examples in the existing parser:
• Bound identifiers are converted to de Bruijn indexes while constructing the AST. Is this desirable ? How does the user specify that identifier are bound or not ?
In fact, the rewriting of bound identifier is done in the pre-existing parsers, so there is no need for the user to specify if an identifier will be bound or not. The user has just to select the parser he want to use (lambda like, set like, ...)

Sample Implementation

A prototype as been developed in Python to quickly try different solutions.

The development tree of the python prototype is available at http://bitbucket.org/matclab/eventb_pratt_parser/ with the default branch implementing a rather functional approach and the object branch being a first try to a more object based implementation.

Tests

• Regression tests. The tests of the existing parser shall succeed when applied to the generic parser.
• New tests.