parent
72e7eed05a
commit
e47d74aa40
151
mode/awk.py
151
mode/awk.py
|
@ -2,7 +2,7 @@ import commands
|
|||
import color, mode, tab
|
||||
from lex import Grammar, PatternRule, RegionRule
|
||||
from mode.python import StringGrammar2
|
||||
from tab import Marker
|
||||
from tab import StackTabber2
|
||||
|
||||
class RegexGrammar(Grammar):
|
||||
rules = [
|
||||
|
@ -37,37 +37,17 @@ class AwkGrammar(Grammar):
|
|||
PatternRule(r'eol', r'\n'),
|
||||
]
|
||||
|
||||
class AwkTabber(tab.StackTabber):
|
||||
open_tokens = {'{': '}', '(': ')', '[': ']'}
|
||||
close_tokens = {'}': '{', ')': '(', ']': '['}
|
||||
def __init__(self, m):
|
||||
self.mode = m
|
||||
self.name = m.name()
|
||||
self.lines = {}
|
||||
self._reset()
|
||||
def region_added(self, p, newlines):
|
||||
self.lines = {}
|
||||
def region_removed(self, p1, p2):
|
||||
self.lines = {}
|
||||
def get_level(self, y):
|
||||
if y not in self.lines:
|
||||
self._calc_level(y)
|
||||
return self.lines.get(y)
|
||||
|
||||
def _calc_level(self, y):
|
||||
target = y
|
||||
while not self._is_base(y) and y > 0:
|
||||
y -= 1
|
||||
self._reset()
|
||||
while y <= target:
|
||||
self._save_curr_level()
|
||||
self._handle_tokens(y)
|
||||
y += 1
|
||||
|
||||
class AwkTabber(StackTabber2):
|
||||
open_tokens = {'delimiter': {'{': '}', '(': ')', '[': ']'}}
|
||||
close_tokens = {'delimiter': {'}': '{', ')': '(', ']': '['}}
|
||||
control_tokens = {
|
||||
'keyword': {'if': 1, 'else': 1, 'while': 1, 'do': 1, 'for': 1},
|
||||
}
|
||||
end_at_eof = True
|
||||
end_at_tokens = {}
|
||||
def _is_base(self, y):
|
||||
if y == 0:
|
||||
return True
|
||||
|
||||
t = self._get_tokens(y)[0]
|
||||
if t.fqname() == 'awk_regex.start':
|
||||
return True
|
||||
|
@ -77,116 +57,11 @@ class AwkTabber(tab.StackTabber):
|
|||
return True
|
||||
else:
|
||||
return False
|
||||
def _is_indent(self, t):
|
||||
return t.name == 'spaces'
|
||||
def _is_ignored(self, t):
|
||||
return t.name in ('spaces', 'eol', 'comment')
|
||||
|
||||
def _reset(self):
|
||||
self.record = {}
|
||||
self.stack = []
|
||||
self.markers = self.stack
|
||||
self.curr_level = 0
|
||||
def _get_curr_level(self):
|
||||
if self.stack:
|
||||
return self.stack[-1].level
|
||||
else:
|
||||
return 0
|
||||
def _get_next_level(self):
|
||||
return self._get_curr_level() + self.mode.tabwidth
|
||||
def _save_curr_level(self):
|
||||
self.curr_level = self._get_curr_level()
|
||||
|
||||
def _match(self, *names):
|
||||
return self.stack and self.stack[-1].name in names
|
||||
def _nomatch(self, *names):
|
||||
return self.stack and self.stack[-1].name not in names
|
||||
def _pop(self, *names):
|
||||
if self._match(*names):
|
||||
self.stack.pop()
|
||||
def _pop_while(self, *names):
|
||||
while self._match(*names):
|
||||
self.stack.pop()
|
||||
def _pop_until(self, *names):
|
||||
while self._nomatch(*names):
|
||||
self.stack.pop()
|
||||
|
||||
def _append(self, name, level):
|
||||
self.stack.append(Marker(name, level))
|
||||
def _append_unless(self, name, level):
|
||||
if self._nomatch(name):
|
||||
self.stack.append(Marker(name, level))
|
||||
|
||||
def _get_tokens(self, y):
|
||||
return self.mode.window.buffer.highlights[self.name].tokens[y]
|
||||
def _handle_tokens(self, y):
|
||||
tokens = self._get_tokens(y)
|
||||
assert tokens
|
||||
start = int(self._is_indent(tokens[0]))
|
||||
end = len(tokens) - 1
|
||||
|
||||
while end > 0 and self._is_ignored(tokens[end]):
|
||||
end -= 1
|
||||
|
||||
for i in range(0, end + 1 - start):
|
||||
token = tokens[start + i]
|
||||
if self._is_ignored(token):
|
||||
pass
|
||||
elif self._is_close_token(token):
|
||||
self._handle_close_token(y, tokens, start, end, i, token)
|
||||
elif self._is_open_token(token):
|
||||
self._handle_open_token(y, tokens, start, end, i, token)
|
||||
else:
|
||||
self._handle_other_token(y, tokens, start, end, i, token)
|
||||
self.lines[y] = self.curr_level
|
||||
self.record[y] = tuple(self.stack)
|
||||
|
||||
def _is_indent(self, token):
|
||||
return token.name == 'spaces'
|
||||
def _is_ignored(self, token):
|
||||
return token.name in ('spaces', 'eol', 'comment')
|
||||
|
||||
def _is_close_token(self, token):
|
||||
return token.name == 'delimiter' and token.string in self.close_tokens
|
||||
def _handle_close_token(self, y, tokens, start, end, i, token):
|
||||
if not self.stack:
|
||||
raise Exception, "unmatched %r, line %d" % (token.string, y)
|
||||
while True:
|
||||
marker = self.stack[-1]
|
||||
if marker.name in ('control', 'continue'):
|
||||
self.stack.pop()
|
||||
elif marker.name in self.open_tokens:
|
||||
if self.open_tokens[marker.name] == token.string:
|
||||
self.stack.pop()
|
||||
break
|
||||
else:
|
||||
raise Exception, "mismatched %r, line %d (expected %r)" % \
|
||||
(token.string, y, d[marker.name])
|
||||
else:
|
||||
raise Exception, "what? %r" % marker.name
|
||||
if i == 0:
|
||||
self._save_curr_level()
|
||||
|
||||
def _is_open_token(self, token):
|
||||
return token.name == 'delimiter' and token.string in self.open_tokens
|
||||
def _handle_open_token(self, y, tokens, start, end, i, token):
|
||||
if i == 0 and self.stack and self.stack[-1].name == 'continue':
|
||||
self.stack.pop()
|
||||
if token.string == '{':
|
||||
self._pop_while('continue', 'control')
|
||||
if i == end - start:
|
||||
level = self._get_next_level()
|
||||
else:
|
||||
level = tokens[i + 1].x
|
||||
self._append(token.string, level)
|
||||
|
||||
def _handle_other_token(self, y, tokens, start, end, i, token):
|
||||
name, s = token.name, token.string
|
||||
if i + start == end:
|
||||
self._pop_while('continue', 'control')
|
||||
|
||||
if name == 'continuation':
|
||||
self._append_unless('continue', self._get_next_level())
|
||||
elif name == 'keyword' and s in ('if', 'else', 'while', 'do', 'for'):
|
||||
if i == start:
|
||||
self._save_curr_level()
|
||||
self._append_unless('control', self._get_next_level())
|
||||
|
||||
class Awk(mode.Fundamental):
|
||||
tabbercls = AwkTabber
|
||||
|
|
51
mode/c.py
51
mode/c.py
|
@ -36,7 +36,6 @@ class CGrammar(Grammar):
|
|||
|
||||
PatternRule(r"delimiter", r"\.|\(|\)|\[|\]|{|}|@|,|:|`|;|=(?!=)|\?|->"),
|
||||
PatternRule(r'eol', r"\n$"),
|
||||
PatternRule(r'function', r'[a-zA-Z_][a-zA-Z0-9_]*(?= *\()'),
|
||||
|
||||
PatternGroupRule(r'structgroup', r'keyword', r'struct', r'spaces',
|
||||
r' +', r'structname', r'[a-zA-Z_][a-zA-Z0-9_]*'),
|
||||
|
@ -44,6 +43,8 @@ class CGrammar(Grammar):
|
|||
r' +', r'enumname', r'[a-zA-Z_][a-zA-Z0-9_]*'),
|
||||
PatternRule(r'keyword', r"(?:auto|break|case|char|const|continue|default|double|do|else|enum|extern|float|for|goto|if|int|long|register|return|short|signed|sizeof|static|struct|switch|typedef|union|unsigned|void|volatile|while)(?![a-zA-Z_])"),
|
||||
|
||||
PatternRule(r'function', r'[a-zA-Z_][a-zA-Z0-9_]*(?= *\()'),
|
||||
|
||||
PatternRule(r'builtin', r"(?:NULL|TRUE|FALSE)"),
|
||||
PatternRule(r'label', r'[a-zA-Z_][a-zA-Z0-9_]*(?=:)'),
|
||||
|
||||
|
@ -72,6 +73,51 @@ class CGrammar(Grammar):
|
|||
OverridePatternRule(r'comment', r'// *@@:(?P<token>[.a-zA-Z0-9_]+):(?P<mode>[.a-zA-Z0-9_]+) *$'),
|
||||
]
|
||||
|
||||
class CTabber2(tab.StackTabber2):
|
||||
open_tokens = {'delimiter': {'{': '}', '(': ')', '[': ']'}}
|
||||
close_tokens = {'delimiter': {'}': '{', ')': '(', ']': '['}}
|
||||
control_tokens = {'keyword': {'if': 1, 'else': 1, 'while': 1, 'do': 1, 'for': 1}}
|
||||
end_at_eof = False
|
||||
end_at_tokens = {'delimiter': {';': 1}}
|
||||
nocontinue_tokens = {'delimiter': {';': 1}}
|
||||
start_free_tokens = {'string.start': 'string.end'}
|
||||
end_free_tokens = {'string.end': 'string.start'}
|
||||
def is_base(self, y):
|
||||
if y == 0:
|
||||
return True
|
||||
tokens = self._get_tokens(y)
|
||||
|
||||
# this assumes that people aren't gonna use these macros inside of
|
||||
# blocks, which is probably ok.
|
||||
t0 = tokens[0]
|
||||
if t0.fqname() == 'macro.start' and t0.string in ('#define', '#include'):
|
||||
return True
|
||||
|
||||
# detecting function declarations is annoying; this assumes that people
|
||||
# won't put a variable type and name on different lines, but that they
|
||||
# might do that for function return type and name.
|
||||
#
|
||||
# unfortunately, valid function return types might include any of the
|
||||
# four types of tokens below
|
||||
decl = False
|
||||
for t in tokens:
|
||||
if t.name in ('keyword', 'identifier', 'structname', 'enumname'):
|
||||
decl = True
|
||||
continue
|
||||
if decl and t.name == 'function':
|
||||
break
|
||||
else:
|
||||
decl = False
|
||||
break
|
||||
return decl
|
||||
|
||||
def _is_indent(self, t):
|
||||
return t.name == 'spaces'
|
||||
def _is_ignored(self, t):
|
||||
if t.name in ('spaces', 'eol', 'comment'): return True
|
||||
elif t.fqname() in ('comment.start', 'comment.null', 'comment.end'): return True
|
||||
else: return False
|
||||
|
||||
class CTabber(tab.StackTabber):
|
||||
wst = ('spaces', 'eol', 'comment', 'comment.start', 'comment.null', 'comment.end')
|
||||
def token_is_whitespace(self, y, i):
|
||||
|
@ -222,7 +268,8 @@ class CMake(method.shell.Exec):
|
|||
class C(mode.Fundamental):
|
||||
modename = 'C'
|
||||
extensions = ['.c', '.h', '.cpp']
|
||||
tabbercls = CTabber
|
||||
#tabbercls = CTabber
|
||||
tabbercls = CTabber2
|
||||
grammar = CGrammar
|
||||
opentokens = ('delimiter',)
|
||||
opentags = {'(': ')', '[': ']', '{': '}'}
|
||||
|
|
182
tab.py
182
tab.py
|
@ -1,5 +1,6 @@
|
|||
import regex, util
|
||||
from point import Point
|
||||
from sets import Set
|
||||
|
||||
class Marker(object):
|
||||
def __init__(self, name, level):
|
||||
|
@ -221,3 +222,184 @@ class StackTabber(Tabber):
|
|||
def _opt_pop(self, *names):
|
||||
if self.markers and self.markers[-1].name in names:
|
||||
self.markers.pop(-1)
|
||||
|
||||
class Marker2(object):
|
||||
def __init__(self, name, type_, level):
|
||||
self.name = name
|
||||
self.type_ = type_
|
||||
self.level = level
|
||||
def __repr__(self):
|
||||
return '<Marker2(%r, %r, %r)>' % (self.name, self.type_, self.level)
|
||||
|
||||
#class StackTabber2(tab.StackTabber):
|
||||
class StackTabber2(Tabber):
|
||||
open_tokens = {'delimiter': {'{': '}', '(': ')', '[': ']'}}
|
||||
close_tokens = {'delimiter': {'}': '{', ')': '(', ']': '['}}
|
||||
scope_tokens = {'delimiter': Set(['{'])}
|
||||
control_tokens = {'keyword': Set(['if', 'else', 'while', 'do', 'for'])}
|
||||
end_at_eof = True
|
||||
end_at_tokens = {}
|
||||
continue_tokens = {}
|
||||
nocontinue_tokens = {}
|
||||
start_free_tokens = {'string.start': 'string.end'}
|
||||
end_free_tokens = {'string.end': 'string.start'}
|
||||
def __init__(self, m):
|
||||
self.mode = m
|
||||
self.name = m.name()
|
||||
self.lines = {}
|
||||
self._reset()
|
||||
def region_added(self, p, newlines):
|
||||
self.lines = {}
|
||||
def region_removed(self, p1, p2):
|
||||
self.lines = {}
|
||||
def get_level(self, y):
|
||||
if y not in self.lines:
|
||||
self._calc_level(y)
|
||||
return self.lines.get(y)
|
||||
|
||||
def _calc_level(self, y):
|
||||
target = y
|
||||
while not self._is_base(y) and y > 0:
|
||||
y -= 1
|
||||
self._reset()
|
||||
while y <= target:
|
||||
self._save_curr_level()
|
||||
self._handle_tokens(y)
|
||||
y += 1
|
||||
|
||||
def _is_base(self, y):
|
||||
return y == 0
|
||||
|
||||
def _reset(self):
|
||||
self.record = {}
|
||||
self.stack = []
|
||||
self.markers = self.stack
|
||||
self.curr_level = 0
|
||||
def _get_curr_level(self):
|
||||
if self.stack:
|
||||
return self.stack[-1].level
|
||||
else:
|
||||
return 0
|
||||
def _get_next_level(self):
|
||||
return self._get_curr_level() + self.mode.tabwidth
|
||||
def _save_curr_level(self):
|
||||
self.curr_level = self._get_curr_level()
|
||||
|
||||
def _match(self, *names):
|
||||
return self.stack and self.stack[-1].name in names
|
||||
def _nomatch(self, *names):
|
||||
return self.stack and self.stack[-1].name not in names
|
||||
def _pop(self, *names):
|
||||
if self._match(*names):
|
||||
self.stack.pop()
|
||||
def _pop_while(self, *names):
|
||||
while self._match(*names):
|
||||
self.stack.pop()
|
||||
def _pop_until(self, *names):
|
||||
while self._nomatch(*names):
|
||||
self.stack.pop()
|
||||
|
||||
def _append(self, name, type_, level):
|
||||
self.stack.append(Marker2(name, type_, level))
|
||||
def _append_unless(self, name, type_, level):
|
||||
if self._nomatch(name):
|
||||
self.stack.append(Marker2(name, type_, level))
|
||||
def _peek(self):
|
||||
if self.stack:
|
||||
return self.stack[-1]
|
||||
else:
|
||||
return None
|
||||
|
||||
def _get_tokens(self, y):
|
||||
return self.mode.window.buffer.highlights[self.name].tokens[y]
|
||||
def _handle_tokens(self, y):
|
||||
tokens = self._get_tokens(y)
|
||||
assert tokens
|
||||
start = int(self._is_indent(tokens[0]))
|
||||
end = len(tokens) - 1
|
||||
|
||||
while end > 0 and self._is_ignored(tokens[end]):
|
||||
end -= 1
|
||||
|
||||
for i in range(0, end + 1 - start):
|
||||
t = tokens[start + i]
|
||||
if self._is_ignored(t):
|
||||
pass
|
||||
elif self._is_close_token(t):
|
||||
self._handle_close_token(y, tokens, start, end, i, t)
|
||||
elif self._is_open_token(t):
|
||||
self._handle_open_token(y, tokens, start, end, i, t)
|
||||
else:
|
||||
self._handle_other_token(y, tokens, start, end, i, t)
|
||||
self.lines[y] = self.curr_level
|
||||
self.record[y] = tuple(self.stack)
|
||||
|
||||
def _is_indent(self, t):
|
||||
return t.name == 'spaces'
|
||||
def _is_ignored(self, t):
|
||||
return t.name in ('spaces', 'eol', 'comment')
|
||||
|
||||
def _is_close_token(self, t):
|
||||
return (t.name == 'delimiter' and
|
||||
t.string in self.close_tokens['delimiter'])
|
||||
def _handle_close_token(self, y, tokens, start, end, i, t):
|
||||
if not self.stack:
|
||||
raise Exception, "unmatched %r, line %d" % (t.string, y)
|
||||
while True:
|
||||
marker = self.stack[-1]
|
||||
if marker.name in ('control', 'continue'):
|
||||
self.stack.pop()
|
||||
elif marker.name in self.open_tokens[marker.type_]:
|
||||
if self.open_tokens[marker.type_][marker.name] == t.string:
|
||||
self.stack.pop()
|
||||
break
|
||||
else:
|
||||
raise Exception, "mismatched %r, line %d (expected %r)" % \
|
||||
(t.string, y, d[marker.name])
|
||||
else:
|
||||
raise Exception, "what? %r" % marker.name
|
||||
if i == 0:
|
||||
self._save_curr_level()
|
||||
|
||||
def _is_open_token(self, t):
|
||||
return (t.name == 'delimiter' and
|
||||
t.string in self.open_tokens['delimiter'])
|
||||
def _handle_open_token(self, y, tokens, start, end, i, t):
|
||||
if i == 0 and self.stack and self.stack[-1].name == 'continue':
|
||||
self.stack.pop()
|
||||
if t.string in self.scope_tokens.get(t.name, {}):
|
||||
self._pop_while('continue', 'control')
|
||||
if i == end - start:
|
||||
level = self._get_next_level()
|
||||
else:
|
||||
level = tokens[i + 1].x + 1
|
||||
self._append(t.string, t.name, level)
|
||||
|
||||
def _handle_other_token(self, y, tokens, start, end, i, t):
|
||||
name, s = t.fqname(), t.string
|
||||
if name in self.start_free_tokens:
|
||||
self._append('free', name, None)
|
||||
return
|
||||
elif name in self.end_free_tokens:
|
||||
self._pop('free')
|
||||
|
||||
if self.end_at_eof and i + start == end:
|
||||
self._pop_while('continue', 'control')
|
||||
elif self.end_at_tokens.get(name, {}).get(s):
|
||||
self._pop_while('continue', 'control')
|
||||
|
||||
top = self._peek()
|
||||
if top and top.name in self.scope_tokens.get(top.type_, {}):
|
||||
if self.continue_tokens:
|
||||
if s in self.continue_tokens.get(name, {}):
|
||||
self._append_unless('continue', name, self._get_next_level())
|
||||
elif self.nocontinue_tokens:
|
||||
if s not in self.nocontinue_tokens.get(name, {}):
|
||||
self._append_unless('continue', name, self._get_next_level())
|
||||
|
||||
if name == 'continuation':
|
||||
self._append_unless('continue', name, self._get_next_level())
|
||||
elif s in self.control_tokens.get(name, {}):
|
||||
if i == start:
|
||||
self._save_curr_level()
|
||||
self._append_unless('control', name, self._get_next_level())
|
||||
|
|
Loading…
Reference in New Issue