Closures (and garbage collection for environments, which is required by closures...
[fur] / tokenization.py
index 1a2653e..3c4dc6f 100644 (file)
@@ -58,6 +58,12 @@ def tokenize(source):
             index += 1
             continue
 
+        if source[index] == '#':
+            while index < len(source) and source[index] != '\n':
+                index += 1
+
+            continue
+
         success = False
 
         for matcher in _TOKEN_MATCHERS: