@@ -32,13 +32,57 @@ def tokenizer_tests():
32
32
{ "type" : "TABS" , "content" : "\t \t " }, { "type" : "WORD" , "content" : "Mass" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "0.0" }, { "type" : "NEWLINES" , "content" : "\n " },
33
33
{ "type" : "TABS" , "content" : "\t \t " }, { "type" : "WORD" , "content" : "Xd" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "42" },
34
34
])
35
+ test ("deindentation_1" , [
36
+ { "type" : "WORD" , "content" : "PresetName" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "Foo" }, { "type" : "NEWLINES" , "content" : "\n " },
37
+
38
+ { "type" : "TABS" , "content" : "\t " }, { "type" : "WORD" , "content" : "A1" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "X" }, { "type" : "NEWLINES" , "content" : "\n \n " },
39
+ { "type" : "TABS" , "content" : "\t " }, { "type" : "WORD" , "content" : "A2" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "X" }, { "type" : "NEWLINES" , "content" : "\n " },
40
+
41
+ { "type" : "TABS" , "content" : "\t " }, { "type" : "WORD" , "content" : "B1" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "X" }, { "type" : "NEWLINES" , "content" : "\n " },
42
+ { "type" : "EXTRA" , "content" : " " }, { "type" : "NEWLINES" , "content" : "\n " },
43
+ { "type" : "TABS" , "content" : "\t " }, { "type" : "WORD" , "content" : "B2" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "X" }, { "type" : "NEWLINES" , "content" : "\n " },
44
+
45
+ { "type" : "TABS" , "content" : "\t " }, { "type" : "WORD" , "content" : "C1" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "X" }, { "type" : "NEWLINES" , "content" : "\n " },
46
+ { "type" : "EXTRA" , "content" : "//foo" }, { "type" : "NEWLINES" , "content" : "\n " },
47
+ { "type" : "TABS" , "content" : "\t " }, { "type" : "WORD" , "content" : "C2" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "X" },
48
+ ])
49
+ test ("deindentation_2" , [
50
+ { "type" : "WORD" , "content" : "AddEffect" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "MOPixel" }, { "type" : "NEWLINES" , "content" : "\n " },
51
+ { "type" : "TABS" , "content" : "\t " }, { "type" : "WORD" , "content" : "PresetName" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "Foo" }, { "type" : "NEWLINES" , "content" : "\n " },
52
+
53
+ { "type" : "TABS" , "content" : "\t \t " }, { "type" : "WORD" , "content" : "A1" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "X" }, { "type" : "NEWLINES" , "content" : "\n \n " },
54
+ { "type" : "TABS" , "content" : "\t \t " }, { "type" : "WORD" , "content" : "A2" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "X" }, { "type" : "NEWLINES" , "content" : "\n " },
55
+
56
+ { "type" : "TABS" , "content" : "\t \t " }, { "type" : "WORD" , "content" : "B1" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "X" }, { "type" : "NEWLINES" , "content" : "\n " },
57
+ { "type" : "EXTRA" , "content" : " " }, { "type" : "NEWLINES" , "content" : "\n " },
58
+ { "type" : "TABS" , "content" : "\t \t " }, { "type" : "WORD" , "content" : "B2" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "X" }, { "type" : "NEWLINES" , "content" : "\n " },
59
+
60
+ { "type" : "TABS" , "content" : "\t \t " }, { "type" : "WORD" , "content" : "C1" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "X" }, { "type" : "NEWLINES" , "content" : "\n " },
61
+ { "type" : "EXTRA" , "content" : "//foo" }, { "type" : "NEWLINES" , "content" : "\n " },
62
+ { "type" : "TABS" , "content" : "\t \t " }, { "type" : "WORD" , "content" : "C2" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "X" },
63
+ ])
64
+ test ("deindentation_3" , [
65
+ { "type" : "WORD" , "content" : "AddEffect" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "MOPixel" }, { "type" : "NEWLINES" , "content" : "\n " },
66
+ { "type" : "TABS" , "content" : "\t " }, { "type" : "WORD" , "content" : "PresetName" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "Foo" }, { "type" : "NEWLINES" , "content" : "\n " },
67
+
68
+ { "type" : "TABS" , "content" : "\t \t " }, { "type" : "WORD" , "content" : "A1" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "X" }, { "type" : "NEWLINES" , "content" : "\n " },
69
+ { "type" : "EXTRA" , "content" : "\t " }, { "type" : "NEWLINES" , "content" : "\n " },
70
+ { "type" : "TABS" , "content" : "\t \t " }, { "type" : "WORD" , "content" : "A2" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "X" }, { "type" : "NEWLINES" , "content" : "\n " },
71
+
72
+ { "type" : "TABS" , "content" : "\t \t " }, { "type" : "WORD" , "content" : "B1" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "X" }, { "type" : "NEWLINES" , "content" : "\n " },
73
+ { "type" : "EXTRA" , "content" : "\t " }, { "type" : "EXTRA" , "content" : " " }, { "type" : "NEWLINES" , "content" : "\n " },
74
+ { "type" : "TABS" , "content" : "\t \t " }, { "type" : "WORD" , "content" : "B2" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "X" }, { "type" : "NEWLINES" , "content" : "\n " },
75
+
76
+ { "type" : "TABS" , "content" : "\t \t " }, { "type" : "WORD" , "content" : "C1" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "X" }, { "type" : "NEWLINES" , "content" : "\n " },
77
+ { "type" : "EXTRA" , "content" : "\t " }, { "type" : "EXTRA" , "content" : "//foo" }, { "type" : "NEWLINES" , "content" : "\n " },
78
+ { "type" : "TABS" , "content" : "\t \t " }, { "type" : "WORD" , "content" : "C2" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "X" },
79
+ ])
35
80
36
81
37
82
def test (filename , expected ):
38
83
filepath = tests .get_test_path_from_filename (filename )
39
- text = tests .read_test (filepath )
40
84
41
85
tokens = ini_tokenizer .get_tokens (str (filepath ))
42
86
tokens_without_metadata = [ { "type" : token ["type" ], "content" : token ["content" ] } for token in tokens ]
43
87
44
- tests .test (text , tokens_without_metadata , expected )
88
+ tests .test (filename , tokens_without_metadata , expected )
0 commit comments