return token_matcher
_TOKEN_MATCHERS = [
- ('keyword', r'(def|end)(?![a-z_])'),
+ ('keyword', r'(def|do|end)(?![a-z_])'),
('open_parenthese', r'\('),
('close_parenthese', r'\)'),
('comma', r','),
index += 1
continue
+ if source[index] == '#':
+ while index < len(source) and source[index] != '\n':
+ index += 1
+
+ continue
+
success = False
for matcher in _TOKEN_MATCHERS: