@@ -98,6 +98,15 @@ def tokenizer_tests():
98
98
{ "type" : "WORD" , "content" : "Foo" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "Bar" }, { "type" : "NEWLINES" , "content" : "\n " },
99
99
{ "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "Baz" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "Bee" },
100
100
])
101
+ test ("datamodule" , [
102
+ { "type" : "WORD" , "content" : "DataModule" }, { "type" : "NEWLINES" , "content" : "\n " },
103
+ { "type" : "TABS" , "content" : "\t " }, { "type" : "WORD" , "content" : "IconFile" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "ContentFile" }, { "type" : "NEWLINES" , "content" : "\n " },
104
+ { "type" : "TABS" , "content" : "\t \t " }, { "type" : "WORD" , "content" : "FilePath" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "Foo" }, { "type" : "NEWLINES" , "content" : "\n " },
105
+ { "type" : "TABS" , "content" : "\t " }, { "type" : "WORD" , "content" : "ModuleName" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "WORD" , "content" : "Bar" },
106
+ ])
107
+ test ("value_on_next_line" , [
108
+ { "type" : "WORD" , "content" : "Foo" }, { "type" : "EXTRA" , "content" : " " }, { "type" : "EQUALS" , "content" : "=" }, { "type" : "NEWLINES" , "content" : "\n " }, { "type" : "WORD" , "content" : "Bar" },
109
+ ])
101
110
102
111
103
112
def test (filename , expected ):
0 commit comments