|
275 | 275 | " and domain_oid in inflight_configuration_data[config_oid][\"domains_temp\"]\n", |
276 | 276 | " ):\n", |
277 | 277 | "\n", |
| 278 | + " # initialize extraction rulesets an empty array if it doesn't exist yet\n", |
| 279 | + " if (\n", |
| 280 | + " not \"extraction_rulesets\"\n", |
| 281 | + " in inflight_configuration_data[config_oid][\"domains_temp\"][domain_oid]\n", |
| 282 | + " ):\n", |
| 283 | + " inflight_configuration_data[config_oid][\"domains_temp\"][domain_oid][\n", |
| 284 | + " \"extraction_rulesets\"\n", |
| 285 | + " ] = []\n", |
| 286 | + "\n", |
278 | 287 | " all_rules = source[\"rules\"]\n", |
279 | 288 | " all_url_filters = source[\"url_filters\"]\n", |
280 | 289 | "\n", |
|
294 | 303 | " \"extracted\": \"extract\",\n", |
295 | 304 | " }\n", |
296 | 305 | "\n", |
297 | | - " ruleset = {}\n", |
| 306 | + " ruleset = []\n", |
298 | 307 | " if all_rules:\n", |
299 | 308 | " ruleset = [\n", |
300 | 309 | " {\n", |
|
309 | 318 | " }\n", |
310 | 319 | " ]\n", |
311 | 320 | "\n", |
312 | | - " # populate the in-memory data structure\n", |
313 | | - " temp_extraction_rulesets = [\n", |
314 | | - " {\n", |
315 | | - " \"url_filters\": url_filters,\n", |
316 | | - " \"rules\": ruleset,\n", |
317 | | - " }\n", |
318 | | - " ]\n", |
| 321 | + " temp_extraction_rulesets = {\n", |
| 322 | + " \"url_filters\": url_filters,\n", |
| 323 | + " \"rules\": ruleset,\n", |
| 324 | + " }\n", |
319 | 325 | "\n", |
320 | 326 | " print(\n", |
321 | 327 | " f\"{extr_count}.) Crawler {config_oid} has extraction rules {temp_extraction_rulesets}\\n\"\n", |
|
324 | 330 | "\n", |
325 | 331 | " inflight_configuration_data[config_oid][\"domains_temp\"][domain_oid][\n", |
326 | 332 | " \"extraction_rulesets\"\n", |
327 | | - " ] = temp_extraction_rulesets" |
| 333 | + " ].append(temp_extraction_rulesets)" |
328 | 334 | ] |
329 | 335 | }, |
330 | 336 | { |
|
0 commit comments