pmacs3/lex2.py

672 lines
28 KiB
Python
Executable File

import re
valid_name_re = re.compile('^[a-zA-Z_][a-zA-Z0-9_]*$')
reserved_names = ['start', 'middle', 'end', 'null']
class RuleContext:
# to be clear:
# x, y: where the rule processing began
# rule: the rule which began
# flag: a signal to be used to resume the rule correctly
# context: the previous rule namespace(s)
# matchd: the dictionary returned by the rule's matching
def __init__(self, y, x, rule, flag, context, matchd):
self.y = y
self.x = x
self.rule = rule
self.flag = flag
self.context = context
self.matchd = matchd
class Token(object):
def __init__(self, name, rule, y, x, s, **vargs):
self.name = name
self.rule = rule
self.y = y
self.x = x
self.string = s
self.vargs = vargs
def copy(self):
return Token(self.name, None, self.y, self.x, self.string, **self.vargs)
def add_to_string(self, s):
self.string += s
def end_x(self):
return self.x + len(self.string)
def __eq__(self, other):
return (self.y == other.y and
self.x == other.x and
self.string == other.string and
self.name == other.name and
self.vargs == other.vargs)
def __repr__(self):
if len(self.string) < 10:
s = self.string
else:
s = self.string[:10] + '...'
return "<Token(%r, %r, %d, %d, %r)>" % (self.name, self.rule, self.y, self.x, s)
def render(self):
return (self,)
class Rule:
name = 'abstract'
def match(self, lexer, context=[], d={}):
raise Exception, "%s rule cannot match!" % self.name
def make_token(self, lexer, s, name, **vargs):
return Token(name, self, lexer.y, lexer.x, s, **vargs)
class ConstantRule(Rule):
def __init__(self, name, constant):
assert valid_name_re.match(name), 'invalid name %r' % name
assert name not in reserved_names, "reserved rule name: %r" % name
self.name = name
self.constant = constant
def match(self, lexer, context=[], d={}):
if lexer.lines[lexer.y][lexer.x:].startswith(self.constant):
name = '.'.join(context + [self.name])
lexer.add_token(self.make_token(lexer, self.constant, name, grammar=lexer.grammar))
lexer.x += len(self.constant)
return True
else:
return False
class PatternRule(Rule):
def __init__(self, name, pattern):
assert valid_name_re.match(name), 'invalid name %r' % name
assert name not in reserved_names, "reserved rule name: %r" % name
self.name = name
self.pattern = pattern
self.re = re.compile(pattern)
def match(self, lexer, context=[], d={}):
m = self.re.match(lexer.lines[lexer.y], lexer.x)
if m:
name = '.'.join(context + [self.name])
lexer.add_token(self.make_token(lexer, m.group(0), name, grammar=lexer.grammar))
lexer.x += len(m.group(0))
return True
else:
return False
class ContextPatternRule(Rule):
def __init__(self, name, pattern, fallback):
assert valid_name_re.match(name), 'invalid name %r' % name
assert name not in reserved_names, "reserved rule name: %r" % name
self.name = name
self.pattern = pattern
self.fallback = fallback
self.fallback_re = re.compile(fallback)
def match(self, lexer, context=[], d={}):
try:
r = re.compile(self.pattern % d)
except KeyError:
r = self.fallback_re
m = r.match(lexer.lines[lexer.y], lexer.x)
if m:
name = '.'.join(context + [self.name])
lexer.add_token(self.make_token(lexer, m.group(0), name, grammar=lexer.grammar))
lexer.x += len(m.group(0))
return True
else:
return False
class RegionRule(Rule):
def __init__(self, name, start, grammar, end):
assert valid_name_re.match(name), 'invalid name %r' % name
assert name not in reserved_names, "reserved rule name: %r" % name
self.name = name
self.start = start
self.grammar = grammar
self.end = end
self.start_re = re.compile(start)
def _add_from_regex(self, context, name, lexer, m, grammar):
t_name = '.'.join(context + [self.name, name])
t = self.make_token(lexer, m.group(0), t_name, grammar=grammar)
lexer.add_token(t)
lexer.x += len(m.group(0))
def resume(self, lexer, context, flag, d, rulecontexts):
assert rulecontexts, "can't resume without rulecontexts!"
self._match(lexer, context, d, None, rulecontexts)
return True
def match(self, lexer, context=[], d={}):
# see if we can match our start token
m = self.start_re.match(lexer.lines[lexer.y], lexer.x)
if m:
# region was match, so let's do this
return self._match(lexer, context, m.groupdict(), m, [])
else:
# region was not matched; we never started. so return false
return False
def _match(self, lexer, context, d, m, rulecontext=[]):
# if we have been given rulecontext, then we are going to "resume" a
# parse that can already be assumed to have started
reenter = len(rulecontext) > 0
assert m or reenter
# first let's do some bookkeeping
lexer.context.append(RuleContext(lexer.y, lexer.x, self, 'start',
list(context), dict(d)))
if m is not None:
self._add_from_regex(context, 'start', lexer, m, lexer.grammar)
null_t_name = '.'.join(context + [self.name, 'null'])
null_t = None
# if we have an end token, then build it here. notice that it can
# reference named groups from the start token. if we have no end,
# well, then, we're never getting out of here alive!
if self.end:
end_re = re.compile(self.end % d)
# ok, so as long as we aren't done (we haven't found an end token),
# keep reading input
done = False
while not done and lexer.y < len(lexer.lines):
old_y = lexer.y
# if this line is empty, then we skip it, but here we insert
# an empty null token just so we have something
if not reenter and len(lexer.lines[lexer.y]) == 0:
null_t = Token(null_t_name, None, lexer.y, lexer.x, '')
lexer.add_token(null_t)
null_t = None
# ok, as long as we haven't found the end token, and have more
# data on the current line to read, we will process tokens
while not done and lexer.y == old_y and lexer.x < len(lexer.lines[lexer.y]):
# if we are reentering mid-parse, then that takes precedence
if reenter:
reenter = False
rule2 = rulecontext[0].rule
context2 = rulecontext[0].context
d2 = rulecontext[0].matchd
assert rule2.resume(lexer, context2, d2, rulecontext[1:]), \
"%r %r %r %r" % (lexer, context2, d2, rulecontext[1:])
found = True
null_t = None
break
# if we are looking for an end token, then see if we've
# found it. if so, then we are done!
if self.end:
m = end_re.match(lexer.lines[lexer.y], lexer.x)
if m:
self._add_from_regex(context, 'end', lexer, m, None)
done = True
break
# ok, we need to check all our rules now, in order. if we
# find a token, note that we found one and exit the loop
found = False
for rule in self.grammar.rules:
if rule.match(lexer, context + [self.name], d):
found = True
null_t = None
break
# if we never found a token, then we need to add another
# character to the current null token (which we should
# create if it isn't set).
if not found:
if null_t is None:
null_t = Token(null_t_name, None, lexer.y, lexer.x, '')
lexer.add_token(null_t)
null_t.add_to_string(lexer.lines[lexer.y][lexer.x])
lexer.x += 1
# ok, since we're soon going to be on a different line (or
# already are), we want a new null token. so forget about the
# current one (i.e. stop adding to it).
null_t = None
# if we're still on the same line at this point (and not done)
# then that means we're finished with the line and should move
# on to the next one here
if not done and old_y == lexer.y:
lexer.save_context()
lexer.y += 1
lexer.x = 0
# alright, we're finally done procesing the region, so return true
lexer.context.pop(-1)
return True
class DualRegionRule(Rule):
def __init__(self, name, start, grammar1, middle, grammar2, end):
assert valid_name_re.match(name), 'invalid name %r' % name
assert name not in reserved_names, "reserved rule name: %r" % name
self.name = name
self.start = start
self.grammar1 = grammar1
self.middle = middle
self.grammar2 = grammar2
self.end = end
self.start_re = re.compile(start)
def _add_from_regex(self, context, name, lexer, m, grammar=None):
t_name = '.'.join(context + [self.name, name])
t = self.make_token(lexer, m.group(0), t_name, grammar=grammar)
lexer.add_token(t)
lexer.x += len(m.group(0))
def resume(self, lexer, context, flag, d, rulecontexts):
if flag == 'start':
d2 = self._match_first(lexer, context, d, None, rulecontexts)
d3 = dict(d.items() + d2.items())
self._match_second(lexer, context, d3, None, rulecontexts)
return True
elif flag == 'middle':
self._match_second(lexer, context, flag, d, None, rulecontexts)
return True
else:
raise Exception, "invalid flag %r" % flag
def match(self, lexer, context=[], d={}):
# see if we can match our start token
m = self.start_re.match(lexer.lines[lexer.y], lexer.x)
if m:
# region was match, so let's do this
d1 = m.groupdict()
d2 = self._match_first(lexer, context, d1, m, [])
d3 = dict(d1.items() + d2.items())
self._match_second(lexer, context, d3, None, [])
return True
else:
# region was not matched; we never started. so return false
return False
def _match_first(self, lexer, context, d1, m1, rulecontext=[]):
# if we have been given rulecontext, then we are going to "resume" a
# parse that can already be assumed to have started
reenter = len(rulecontext) > 0
assert m1 or reenter
# first let's do some bookkeeping
lexer.context.append(RuleContext(lexer.y, lexer.x, self, 'start',
list(context), dict(d1)))
# ok, so create our start token, and get ready to start reading data
if m1 is not None:
self._add_from_regex(context, 'start', lexer, m1, lexer.grammar)
null_t_name = '.'.join(context + [self.name, 'null'])
null_t = None
middle_re = re.compile(self.middle % d1)
d2 = {}
# ok, so as long as we aren't done (we haven't found an end token),
# keep reading input
done = False
while not done and lexer.y < len(lexer.lines):
old_y = lexer.y
# if this line is empty, then we will skip it, but here weinsert
# an empty null token just so we have something
if len(lexer.lines[lexer.y]) == 0:
null_t = Token(null_t_name, None, lexer.y, lexer.x, '')
lexer.add_token(null_t)
null_t = None
# ok, as long as we haven't found the end token, and have more
# data on the current line to read, we will process tokens
while not done and lexer.y == old_y and lexer.x < len(lexer.lines[lexer.y]):
# if we are reentering mid-parse, then that takes precedence
if reenter:
reenter = False
xrule = rulecontext[0].rule
xcontext = rulecontext[0].context
xd = rulecontext[0].matchd
assert rule2.resume(lexer, xcontext, xd, rulecontext[1:]), \
"%r %r %r %r" % (lexer, xcontext, xd, rulecontext[1:])
found = True
null_t = None
break
# see if we have found the middle token. if so, we can then
# proceed to "stage 2"
m2 = middle_re.match(lexer.lines[lexer.y], lexer.x)
if m2:
d2 = m2.groupdict()
self._add_from_regex(context, 'middle', lexer, m2, None)
done = True
break
# ok, we need to check all our rules now, in order. if we
# find a token, note that we found one and exit the loop
found = False
for rule in self.grammar1.rules:
if rule.match(lexer, context + [self.name], d1):
found = True
null_t = None
break
# if we never found a token, then we need to add another
# character to the current null token (which we should
# create if it isn't set).
if not found:
if null_t is None:
null_t = Token(null_t_name, None, lexer.y, lexer.x, '')
lexer.add_token(null_t)
null_t.add_to_string(lexer.lines[lexer.y][lexer.x])
lexer.x += 1
# ok, since we're soon going to be on a different line (or
# already are), we want a new null token. so forget about the
# current one.
null_t = None
# if we're still on the same line at this point (and not done)
# then that means we're finished with the line and should move
# on to the next one here
if not done and old_y == lexer.y:
lexer.save_context()
lexer.y += 1
lexer.x = 0
lexer.context.pop(-1)
return d2
def _match_second(self, lexer, context, d3, m, rulecontext=[]):
# if we have been given rulecontext, then we are going to "resume" a
# parse that can already be assumed to have started
reenter = len(rulecontext) > 0
# ok stage 2 is like stage 1, only we are looking for end tokens
# instead of middle tokens
null_t_name = '.'.join(context + [self.name, 'null'])
null_t = None
end_re = re.compile(self.end % d3)
lexer.context.append(RuleContext(lexer.y, lexer.x, self, 'middle',
list(context), dict(d3)))
# ok, so as long as we aren't done (we haven't found an end token),
# keep reading input
done = False
while not done and lexer.y < len(lexer.lines):
old_y = lexer.y
# if we are reentering mid-parse, then that takes precedence
if reenter:
reenter = False
xrule = rulecontext[0].rule
xcontext = rulecontext[0].context
xd = rulecontext[0].matchd
assert rule2.resume(lexer, xcontext, xd, rulecontext[1:]), \
"%r %r %r %r" % (lexer, xcontext, xd, rulecontext[1:])
found = True
null_t = None
break
# if this line is empty, then we will skip it, but here weinsert
# an empty null token just so we have something
if len(lexer.lines[lexer.y]) == 0:
null_t = Token(null_t_name, None, lexer.y, lexer.x, '')
lexer.add_token(null_t)
null_t = None
# ok, as long as we haven't found the end token, and have more
# data on the current line to read, we will process tokens
while not done and lexer.y == old_y and lexer.x < len(lexer.lines[lexer.y]):
# see if we have found the middle token. if so, we can then
# proceed to "stage 2"
m3 = end_re.match(lexer.lines[lexer.y], lexer.x)
if m3:
self._add_from_regex(context, 'end', lexer, m3, None)
done = True
break
# ok, we need to check all our rules now, in order. if we
# find a token, note that we found one and exit the loop
found = False
for rule in self.grammar2.rules:
if rule.match(lexer, context + [self.name], d3):
found = True
null_t = None
break
# if we never found a token, then we need to add another
# character to the current null token (which we should
# create if it isn't set).
if not found:
if null_t is None:
null_t = Token(null_t_name, None, lexer.y, lexer.x, '')
lexer.add_token(null_t)
null_t.add_to_string(lexer.lines[lexer.y][lexer.x])
lexer.x += 1
# ok, since we're soon going to be on a different line (or
# already are), we want a new null token. so forget about the
# current one.
null_t = None
# if we're still on the same line at this point (and not done)
# then that means we're finished with the line and should move
# on to the next one here
if not done and old_y == lexer.y:
lexer.save_context()
lexer.y += 1
lexer.x = 0
# alright, we're finally done processing; return true
lexer.context.pop(-1)
return True
# def matchOLD(self, lexer, context=[], d={}):
# m1 = self.start_re.match(lexer.lines[lexer.y], lexer.x)
# # see if we can match out start token
# if m1:
# # ok, so create our start token, and get ready to start reading data
# self._add_from_regex(context, 'start', lexer, m1, lexer.grammar)
# null_t_name = '.'.join(context + [self.name, 'null'])
# null_t = None
#
# d1 = m1.groupdict()
# lexer.context.append(RuleContext(lexer.y, lexer.x, self, 'start',
# list(context), dict(d1)))
# d2 = {}
# middle_re = re.compile(self.middle % d1)
#
# # ok, so as long as we aren't done (we haven't found an end token),
# # keep reading input
# done = False
# while not done and lexer.y < len(lexer.lines):
# old_y = lexer.y
# # if this line is empty, then we will skip it, but here weinsert
# # an empty null token just so we have something
# if len(lexer.lines[lexer.y]) == 0:
# null_t = Token(null_t_name, None, lexer.y, lexer.x, '')
# lexer.add_token(null_t)
# null_t = None
#
# # ok, as long as we haven't found the end token, and have more
# # data on the current line to read, we will process tokens
# while not done and lexer.y == old_y and lexer.x < len(lexer.lines[lexer.y]):
# # see if we have found the middle token. if so, we can then
# # proceed to "stage 2"
# m2 = middle_re.match(lexer.lines[lexer.y], lexer.x)
# if m2:
# d2 = m2.groupdict()
# self._add_from_regex(context, 'middle', lexer, m2, None)
# done = True
# break
#
# # ok, we need to check all our rules now, in order. if we
# # find a token, note that we found one and exit the loop
# found = False
# for rule in self.grammar1.rules:
# if rule.match(lexer, context + [self.name], d1):
# found = True
# null_t = None
# break
#
# # if we never found a token, then we need to add another
# # character to the current null token (which we should
# # create if it isn't set).
# if not found:
# if null_t is None:
# null_t = Token(null_t_name, None, lexer.y, lexer.x, '')
# lexer.add_token(null_t)
# null_t.add_to_string(lexer.lines[lexer.y][lexer.x])
# lexer.x += 1
#
# # ok, since we're soon going to be on a different line (or
# # already are), we want a new null token. so forget about the
# # current one.
# null_t = None
#
# # if we're still on the same line at this point (and not done)
# # then that means we're finished with the line and should move
# # on to the next one here
# if not done and old_y == lexer.y:
# lexer.save_context()
# lexer.y += 1
# lexer.x = 0
#
# # ok stage 2 is like stage 1, only we are looking for end tokens
# # instead of middle tokens
# d3 = dict(d1.items() + d2.items())
# end_re = re.compile(self.end % d3)
# lexer.context.pop(-1)
# lexer.context.append(RuleContext(lexer.y, lexer.x, self, 'middle',
# list(context), dict(d3)))
#
# # ok, so as long as we aren't done (we haven't found an end token),
# # keep reading input
# done = False
# while not done and lexer.y < len(lexer.lines):
# old_y = lexer.y
# # if this line is empty, then we will skip it, but here weinsert
# # an empty null token just so we have something
# if len(lexer.lines[lexer.y]) == 0:
# null_t = Token(null_t_name, None, lexer.y, lexer.x, '')
# lexer.add_token(null_t)
# null_t = None
#
# # ok, as long as we haven't found the end token, and have more
# # data on the current line to read, we will process tokens
# while not done and lexer.y == old_y and lexer.x < len(lexer.lines[lexer.y]):
# # see if we have found the middle token. if so, we can then
# # proceed to "stage 2"
# m3 = end_re.match(lexer.lines[lexer.y], lexer.x)
# if m3:
# self._add_from_regex(context, 'end', lexer, m3, None)
# done = True
# break
#
# # ok, we need to check all our rules now, in order. if we
# # find a token, note that we found one and exit the loop
# found = False
# for rule in self.grammar2.rules:
# if rule.match(lexer, context + [self.name], d3):
# found = True
# null_t = None
# break
#
# # if we never found a token, then we need to add another
# # character to the current null token (which we should
# # create if it isn't set).
# if not found:
# if null_t is None:
# null_t = Token(null_t_name, None, lexer.y, lexer.x, '')
# lexer.add_token(null_t)
# null_t.add_to_string(lexer.lines[lexer.y][lexer.x])
# lexer.x += 1
#
# # ok, since we're soon going to be on a different line (or
# # already are), we want a new null token. so forget about the
# # current one.
# null_t = None
#
# # if we're still on the same line at this point (and not done)
# # then that means we're finished with the line and should move
# # on to the next one here
# if not done and old_y == lexer.y:
# lexer.save_context()
# lexer.y += 1
# lexer.x = 0
#
# # alright, we're finally done processing; return true
# lexer.context.pop(-1)
# return True
# else:
# # dual region was not matched; we never started. so return false
# return False
class Grammar:
rules = []
def __init__(self):
for rule in self.rules:
if hasattr(rule, 'grammar') and rule.grammar is None:
rule.grammar = self
class Lexer:
def __init__(self, name, grammar):
self.name = name
self.grammar = grammar
self.y = 0
self.x = 0
self.lines = None
self.tokens = []
self.context = []
self.line_contexts = {}
def add_token(self, t):
self.tokens.append(t)
def lex(self, lines, y=0, x=0):
self.y = y
self.x = x
self.lines = lines
self.tokens = []
self.context = []
self.line_contexts = {}
def resume(self, lines, y=0, x=0, rulecontexts=[]):
if len(rulecontexts) == 0:
self.lex(lines, y, x)
else:
self.y = y
self.x = x
self.lines = lines
self.tokens = []
rc = rulecontexts[0]
rc.rule.resume(self, rc.context, rc.flag, rc.matchd, rulecontexts[1:])
def __iter__(self):
if self.lines is None:
raise Exception, "no lines to lex"
return self
def save_context(self):
self.line_contexts[self.y] = list(self.context)
def next(self):
null_t_name = 'null'
null_t = None
if self.tokens:
return self.tokens.pop(0)
while self.y < len(self.lines):
line = self.lines[self.y]
while self.x < len(line):
curr_t = None
for rule in self.grammar.rules:
if rule.match(self):
assert self.tokens, "match rendered no tokens?"
return self.tokens.pop(0)
if null_t is None:
null_t = Token(null_t_name, None, self.y, self.x, '')
self.add_token(null_t)
null_t.add_to_string(line[self.x])
self.x += 1
self.save_context()
null_t = None
self.y += 1
self.x = 0
if self.tokens:
return self.tokens.pop(0)
else:
raise StopIteration