@@ -285,7 +285,12 @@ def scan_vcf(path, target_num_partitions):
285
285
return metadata , vcf .raw_header
286
286
287
287
288
- def scan_vcfs (paths , show_progress , target_num_partitions , worker_processes = 1 ):
288
+ def scan_vcfs (
289
+ paths ,
290
+ show_progress ,
291
+ target_num_partitions ,
292
+ worker_processes = core .DEFAULT_WORKER_PROCESSES ,
293
+ ):
289
294
logger .info (
290
295
f"Scanning { len (paths )} VCFs attempting to split into { target_num_partitions } "
291
296
f" partitions."
@@ -1298,7 +1303,7 @@ def init(
1298
1303
vcfs ,
1299
1304
* ,
1300
1305
column_chunk_size = 16 ,
1301
- worker_processes = 1 ,
1306
+ worker_processes = core . DEFAULT_WORKER_PROCESSES ,
1302
1307
target_num_partitions = None ,
1303
1308
show_progress = False ,
1304
1309
compressor = None ,
@@ -1450,7 +1455,9 @@ def process_partition(self, partition_index):
1450
1455
f"{ num_records } records last_pos={ last_position } "
1451
1456
)
1452
1457
1453
- def explode (self , * , worker_processes = 1 , show_progress = False ):
1458
+ def explode (
1459
+ self , * , worker_processes = core .DEFAULT_WORKER_PROCESSES , show_progress = False
1460
+ ):
1454
1461
self .load_metadata ()
1455
1462
num_records = self .metadata .num_records
1456
1463
if np .isinf (num_records ):
@@ -1518,7 +1525,7 @@ def explode(
1518
1525
vcfs ,
1519
1526
* ,
1520
1527
column_chunk_size = 16 ,
1521
- worker_processes = 1 ,
1528
+ worker_processes = core . DEFAULT_WORKER_PROCESSES ,
1522
1529
show_progress = False ,
1523
1530
compressor = None ,
1524
1531
):
@@ -1543,7 +1550,7 @@ def explode_init(
1543
1550
* ,
1544
1551
column_chunk_size = 16 ,
1545
1552
target_num_partitions = 1 ,
1546
- worker_processes = 1 ,
1553
+ worker_processes = core . DEFAULT_WORKER_PROCESSES ,
1547
1554
show_progress = False ,
1548
1555
compressor = None ,
1549
1556
):
@@ -1605,7 +1612,7 @@ def convert(
1605
1612
* ,
1606
1613
variants_chunk_size = None ,
1607
1614
samples_chunk_size = None ,
1608
- worker_processes = 1 ,
1615
+ worker_processes = core . DEFAULT_WORKER_PROCESSES ,
1609
1616
local_alleles = None ,
1610
1617
show_progress = False ,
1611
1618
icf_path = None ,
@@ -1649,7 +1656,7 @@ def encode(
1649
1656
dimension_separator = None ,
1650
1657
max_memory = None ,
1651
1658
local_alleles = None ,
1652
- worker_processes = 1 ,
1659
+ worker_processes = core . DEFAULT_WORKER_PROCESSES ,
1653
1660
show_progress = False ,
1654
1661
):
1655
1662
# Rough heuristic to split work up enough to keep utilisation high
@@ -1687,7 +1694,7 @@ def encode_init(
1687
1694
max_variant_chunks = None ,
1688
1695
dimension_separator = None ,
1689
1696
max_memory = None ,
1690
- worker_processes = 1 ,
1697
+ worker_processes = core . DEFAULT_WORKER_PROCESSES ,
1691
1698
show_progress = False ,
1692
1699
):
1693
1700
icf_store = IntermediateColumnarFormat (icf_path )
0 commit comments