|
42 | 42 | #+end_block |
43 | 43 | """)) == |
44 | 44 | [Token(K"<block[30]", 1, 13), |
45 | | - Token(K">block[30]", 23, 33)] |
| 45 | + Token(K"<paragraph", 15, 15), |
| 46 | + Token(K">paragraph", 21, 21), |
| 47 | + Token(K">block[30]", 23, 33)] |
46 | 48 | @test collect(Lexer(""" |
47 | 49 | #+BEGIN_BLOCK parameters |
48 | 50 | content |
49 | 51 | #+END_BLOCK |
50 | 52 | """)) == |
51 | 53 | [Token(K"<block[30]", 1, 24), |
52 | | - Token(K">block[30]", 34, 44)] |
| 54 | + Token(K"<paragraph", 26, 26), |
| 55 | + Token(K">paragraph", 32, 32), |
| 56 | + Token(K">block[30]", 34, 44)] |
53 | 57 | @test collect(Lexer(""" |
54 | 58 | #+begin_block |
55 | 59 | #+begin_other |
56 | 60 | content |
57 | 61 | #+end_other |
58 | 62 | #+end_block |
59 | 63 | """)) == |
60 | | - [Token(K"<block[30]", 1, 13) |
61 | | - Token(K"<block[41]", 15, 27) |
62 | | - Token(K">block[41]", 37, 47) |
| 64 | + [Token(K"<block[30]", 1, 13), |
| 65 | + Token(K"<block[41]", 15, 27), |
| 66 | + Token(K"<paragraph", 29, 29), |
| 67 | + Token(K">paragraph", 35, 35), |
| 68 | + Token(K">block[41]", 37, 47), |
63 | 69 | Token(K">block[30]", 49, 59)] |
64 | 70 | end |
65 | 71 | @testset "Dynamic blocks" begin |
|
69 | 75 | #+end: |
70 | 76 | """)) == |
71 | 77 | [Token(K"<dynamic_block", 1, 13), |
| 78 | + Token(K"<paragraph", 15, 15), |
| 79 | + Token(K">paragraph", 21, 21), |
72 | 80 | Token(K">dynamic_block", 23, 28)] |
73 | 81 | end |
74 | 82 | @testset "Lesser blocks" begin |
|
87 | 95 | #+end_src extra |
88 | 96 | """)) == |
89 | 97 | [Token(K"<source_block", 1, 17), |
90 | | - Token(K">source_block", 37, 51)] |
| 98 | + Token(K">source_block", 37, 51), |
| 99 | + Token(K"<paragraph", 53, 53)] |
91 | 100 | @test collect(Lexer(""" |
92 | 101 | #+begin_export html |
93 | 102 | <b>content</b> |
|
105 | 114 | :end: |
106 | 115 | """)) == |
107 | 116 | [Token(K"<drawer", 1, 8), |
| 117 | + Token(K"<paragraph", 10, 10), |
| 118 | + Token(K">paragraph", 16, 16), |
108 | 119 | Token(K">drawer", 18, 22)] |
109 | 120 | @test collect(Lexer(""" |
110 | 121 | :drawer: |
|
114 | 125 | :end: |
115 | 126 | """)) == |
116 | 127 | [Token(K"<drawer", 1, 8), |
| 128 | + Token(K"<paragraph", 10, 10), |
| 129 | + Token(K">paragraph", 16, 16), |
| 130 | + Token(K"<paragraph", 18, 18), |
| 131 | + Token(K">paragraph", 34, 34), |
117 | 132 | Token(K">drawer", 36, 40)] |
118 | 133 | end |
119 | 134 | @testset "Property drawers" begin |
@@ -143,51 +158,84 @@ end |
143 | 158 | end |
144 | 159 | @testset "Footnote defs" begin |
145 | 160 | @test collect(Lexer("[fn:1] stuff")) == |
146 | | - [Token(K"<footnote_definition", 1, 6)] |
| 161 | + [Token(K"<footnote_definition", 1, 6), |
| 162 | + Token(K"<paragraph", 8, 8)] |
147 | 163 | @test collect(Lexer("[fn:1] stuff\n[fn:2] more")) == |
148 | 164 | [Token(K"<footnote_definition", 1, 6), |
| 165 | + Token(K"<paragraph", 8, 8), |
| 166 | + Token(K">paragraph", 12, 12), |
149 | 167 | Token(K">footnote_definition", 12, 12), |
150 | | - Token(K"<footnote_definition", 14, 19)] |
| 168 | + Token(K"<footnote_definition", 14, 19), |
| 169 | + Token(K"<paragraph", 21, 21)] |
151 | 170 | @test collect(Lexer("[fn:1] stuff\n\n\nmore")) == |
152 | 171 | [Token(K"<footnote_definition", 1, 6), |
153 | | - Token(K">footnote_definition", 12, 12)] |
| 172 | + Token(K"<paragraph", 8, 8), |
| 173 | + Token(K">paragraph", 12, 12), |
| 174 | + Token(K">footnote_definition", 12, 12), |
| 175 | + Token(K"<paragraph", 16, 16)] |
154 | 176 | end |
155 | 177 | @testset "Items" begin |
156 | 178 | @test collect(Lexer("+ item")) == |
157 | | - [Token(K"<item[1]", 1, 1)] |
| 179 | + [Token(K"<item[1]", 1, 1), |
| 180 | + Token(K"<paragraph", 3, 3)] |
158 | 181 | @test collect(Lexer(" + item")) == |
159 | | - [Token(K"<item[3]", 3, 3)] |
| 182 | + [Token(K"<item[3]", 3, 3), |
| 183 | + Token(K"<paragraph", 5, 5)] |
160 | 184 | @test collect(Lexer("- item")) == |
161 | | - [Token(K"<item[1]", 1, 1)] |
| 185 | + [Token(K"<item[1]", 1, 1), |
| 186 | + Token(K"<paragraph", 3, 3)] |
162 | 187 | @test collect(Lexer(" * item")) == |
163 | | - [Token(K"item[2]", 2, 2)] |
| 188 | + [Token(K"<item[2]", 2, 2), |
| 189 | + Token(K"<paragraph", 4, 4)] |
164 | 190 | @test collect(Lexer("+ item\nmore")) == |
165 | 191 | [Token(K"<item[1]", 1, 1), |
166 | | - Token(K">item[1]", 6, 6)] |
| 192 | + Token(K"<paragraph", 3, 3), |
| 193 | + Token(K">paragraph", 6, 6), |
| 194 | + Token(K">item[1]", 6, 6), |
| 195 | + Token(K"<paragraph", 8, 8)] |
167 | 196 | @test collect(Lexer("+ item\n more")) == |
168 | | - [Token(K"<item[1]", 1, 1)] |
| 197 | + [Token(K"<item[1]", 1, 1), |
| 198 | + Token(K"<paragraph", 3, 3)] |
169 | 199 | @test collect(Lexer("+ item\n more")) == |
170 | | - [Token(K"<item[1]", 1, 1)] |
| 200 | + [Token(K"<item[1]", 1, 1), |
| 201 | + Token(K"<paragraph", 3, 3)] |
171 | 202 | @test collect(Lexer("+ item\n \n more")) == |
172 | | - [Token(K"<item[1]", 1, 1)] |
| 203 | + [Token(K"<item[1]", 1, 1), |
| 204 | + Token(K"<paragraph", 3, 3), |
| 205 | + Token(K">paragraph", 6, 6), |
| 206 | + Token(K"<paragraph", 11, 11)] |
173 | 207 | @test collect(Lexer("+ item\n\n more")) == |
174 | | - [Token(K"<item[1]", 1, 1)] |
| 208 | + [Token(K"<item[1]", 1, 1), |
| 209 | + Token(K"<paragraph", 3, 3), |
| 210 | + Token(K">paragraph", 6, 6), |
| 211 | + Token(K"<paragraph", 9, 9)] |
175 | 212 | @test collect(Lexer("+ item\n\n\n more")) == |
176 | 213 | [Token(K"<item[1]", 1, 1), |
177 | | - Token(K">item[1]", 6, 6)] |
| 214 | + Token(K"<paragraph", 3, 3), |
| 215 | + Token(K">paragraph", 6, 6), |
| 216 | + Token(K">item[1]", 6, 6), |
| 217 | + Token(K"<paragraph", 10, 10)] |
178 | 218 | @test collect(Lexer(" + item\n more")) == |
179 | 219 | [Token(K"<item[2]", 2, 2), |
180 | | - Token(K">item[2]", 7, 7)] |
| 220 | + Token(K"<paragraph", 4, 4), |
| 221 | + Token(K">paragraph", 7, 7), |
| 222 | + Token(K">item[2]", 7, 7), |
| 223 | + Token(K"<paragraph", 9, 9)] |
181 | 224 | @test collect(Lexer(" + item\n more")) == |
182 | | - [Token(K"<item[2]", 2, 2)] |
| 225 | + [Token(K"<item[2]", 2, 2), |
| 226 | + Token(K"<paragraph", 4, 4)] |
183 | 227 | @test collect(Lexer("1. item")) == |
184 | | - [Token(K"<item[1]", 1, 2)] |
| 228 | + [Token(K"<item[1]", 1, 2), |
| 229 | + Token(K"<paragraph", 4, 4)] |
185 | 230 | @test collect(Lexer("12) item")) == |
186 | | - [Token(K"<item[1]", 1, 3)] |
| 231 | + [Token(K"<item[1]", 1, 3), |
| 232 | + Token(K"<paragraph", 5, 5)] |
187 | 233 | @test collect(Lexer("a. item")) == |
188 | | - [Token(K"<item[1]", 1, 2)] |
| 234 | + [Token(K"<item[1]", 1, 2), |
| 235 | + Token(K"<paragraph", 4, 4)] |
189 | 236 | @test collect(Lexer("ab) item")) == |
190 | | - [Token(K"<item[1]", 1, 3)] |
| 237 | + [Token(K"<item[1]", 1, 3) |
| 238 | + Token(K"<paragraph", 5, 5)] |
191 | 239 | end |
192 | 240 | @testset "Tables" begin |
193 | 241 | @test collect(Lexer("|")) == |
|
205 | 253 | Token(K"<table_cell", 3, 3), |
206 | 254 | Token(K">table_cell", 6, 6), |
207 | 255 | Token(K">table_row", 6, 6), |
208 | | - Token(K">table", 6, 6)] |
| 256 | + Token(K">table", 6, 6), |
| 257 | + Token(K"<paragraph", 8, 8)] |
209 | 258 | @test collect(Lexer("| cell | two | three")) == |
210 | 259 | [Token(K"<table", 1, 1), |
211 | 260 | Token(K"<table_row", 1, 1), |
@@ -246,10 +295,10 @@ end |
246 | 295 | [Token(K"<clock", 1, 1)] |
247 | 296 | @test collect(Lexer("clock: [2019-03-25 Mon 10:49]--[2019-03-25 Mon 11:31] => 0:42")) == |
248 | 297 | [Token(K"<clock", 1, 1)] |
249 | | - @test collect(Lexer("clock: 12:30")) != |
250 | | - [Token(K"<clock", 1, 1)] |
251 | | - @test collect(Lexer("clock: [2024-10-12]--")) != |
252 | | - [Token(K"<clock", 1, 1)] |
| 298 | + @test collect(Lexer("clock: 12:30")) == |
| 299 | + [Token(K"<paragraph", 1, 1)] |
| 300 | + @test collect(Lexer("clock: [2024-10-12]--")) == |
| 301 | + [Token(K"<paragraph", 1, 1)] |
253 | 302 | end |
254 | 303 | @testset "Diary sexp" begin |
255 | 304 | @test collect(Lexer("%%(org-calendar-holiday)")) == |
|
298 | 347 | [Token(K"fixedwidth", 1, 18)] |
299 | 348 | end |
300 | 349 | @testset "Horizontal rule" begin |
301 | | - @test collect(Lexer("----")) == Token[] |
302 | | - @test collect(Lexer("-- ---")) == Token[] |
| 350 | + @test collect(Lexer("----")) == |
| 351 | + [Token(K"<paragraph", 1, 1)] |
| 352 | + @test collect(Lexer("-- ---")) == |
| 353 | + [Token(K"<paragraph", 1, 1)] |
| 354 | + @test collect(Lexer("----- -----")) == |
| 355 | + [Token(K"<paragraph", 1, 1)] |
303 | 356 | @test collect(Lexer("-----")) == |
304 | 357 | [Token(K"hrule", 1, 5)] |
305 | 358 | @test collect(Lexer("------")) == |
|
319 | 372 | stuff |
320 | 373 | \\end{env}fluff |
321 | 374 | """)) == |
322 | | - Token[] |
| 375 | + [Token(K"<paragraph", 1, 1)] |
323 | 376 | @test collect(Lexer(""" |
324 | 377 | \\begin{equation*} |
325 | 378 | \\begin{align} |
|
334 | 387 | x^2 + y^2 = z^2 |
335 | 388 | \\end{equation*} |
336 | 389 | """)) == |
337 | | - [Token(K"latex_environment", 1, 76) |
| 390 | + [Token(K"latex_environment", 1, 76), |
| 391 | + Token(K"<paragraph", 79, 79), |
| 392 | + Token(K">paragraph", 85, 85), |
338 | 393 | Token(K"latex_environment", 88, 136)] |
339 | 394 | end |
340 | 395 | @testset "Type inference" begin |
|
0 commit comments