@@ -422,7 +422,9 @@ def read_input_vectors(
422422 trace_id : Optional [str ] = None ,
423423 ) -> np .array :
424424 logger = setup (config , verbose )
425- logger .debug ("Reading input vectors start_pos: %i, end_pos: %i" , start_pos , end_pos )
425+ logger .debug (
426+ "Reading input vectors start_pos: %i, end_pos: %i" , start_pos , end_pos
427+ )
426428 if source_type == "TILEDB_ARRAY" :
427429 with tiledb .open (source_uri , mode = "r" ) as src_array :
428430 return np .transpose (
@@ -511,7 +513,9 @@ def copy_centroids(
511513 logger = setup (config , verbose )
512514 group = tiledb .Group (array_uri )
513515 centroids_uri = group [CENTROIDS_ARRAY_NAME ].uri
514- logger .debug ("Copying centroids from: %s, to: %s" , copy_centroids_uri , centroids_uri )
516+ logger .debug (
517+ "Copying centroids from: %s, to: %s" , copy_centroids_uri , centroids_uri
518+ )
515519 src = tiledb .open (copy_centroids_uri , mode = "r" )
516520 dest = tiledb .open (centroids_uri , mode = "w" )
517521 src_centroids = src [:, :]
@@ -586,7 +590,9 @@ def init_centroids(
586590 trace_id : Optional [str ] = None ,
587591 ) -> np .array :
588592 logger = setup (config , verbose )
589- logger .debug ("Initialising centroids by reading the first vectors in the source data." )
593+ logger .debug (
594+ "Initialising centroids by reading the first vectors in the source data."
595+ )
590596 with tiledb .scope_ctx (ctx_or_config = config ):
591597 return read_input_vectors (
592598 source_uri = source_uri ,
@@ -922,7 +928,9 @@ def consolidate_partition_udf(
922928 ):
923929 logger = setup (config , verbose )
924930 with tiledb .scope_ctx (ctx_or_config = config ):
925- logger .debug ("Consolidating partitions %d-%d" , partition_id_start , partition_id_end )
931+ logger .debug (
932+ "Consolidating partitions %d-%d" , partition_id_start , partition_id_end
933+ )
926934 group = tiledb .Group (array_uri )
927935 partial_write_array_dir_uri = array_uri + "/" + PARTIAL_WRITE_ARRAY_DIR
928936 partial_write_array_ids_uri = (
@@ -962,12 +970,16 @@ def consolidate_partition_udf(
962970 index_array = tiledb .open (index_array_uri , mode = "r" )
963971 ids_array = tiledb .open (ids_array_uri , mode = "w" )
964972 parts_array = tiledb .open (parts_array_uri , mode = "w" )
965- logger .debug ("Partitions start: %d end: %d" , partition_id_start , partition_id_end )
973+ logger .debug (
974+ "Partitions start: %d end: %d" , partition_id_start , partition_id_end
975+ )
966976 for part in range (partition_id_start , partition_id_end , batch ):
967977 part_end = part + batch
968978 if part_end > partition_id_end :
969979 part_end = partition_id_end
970- logger .debug ("Consolidating partitions start: %d end: %d" , part , part_end )
980+ logger .debug (
981+ "Consolidating partitions start: %d end: %d" , part , part_end
982+ )
971983 read_slices = []
972984 for p in range (part , part_end ):
973985 for partition_slice in partition_slices [p ]:
@@ -985,8 +997,13 @@ def consolidate_partition_udf(
985997 "values"
986998 ]
987999
988- logger .debug ("Ids shape %s, expected size: %d expected range:(%d,%d)" , ids .shape , end_pos - start_pos ,
989- start_pos , end_pos )
1000+ logger .debug (
1001+ "Ids shape %s, expected size: %d expected range:(%d,%d)" ,
1002+ ids .shape ,
1003+ end_pos - start_pos ,
1004+ start_pos ,
1005+ end_pos ,
1006+ )
9901007 if ids .shape [0 ] != end_pos - start_pos :
9911008 raise ValueError ("Incorrect partition size." )
9921009
@@ -1338,7 +1355,10 @@ def consolidate_and_vacuum(
13381355 logger .debug ("input_vectors_per_work_item %d" , input_vectors_per_work_item )
13391356 logger .debug ("input_vectors_work_items %d" , input_vectors_work_items )
13401357 logger .debug ("input_vectors_work_tasks %d" , input_vectors_work_tasks )
1341- logger .debug ("input_vectors_work_items_per_worker %d" , input_vectors_work_items_per_worker )
1358+ logger .debug (
1359+ "input_vectors_work_items_per_worker %d" ,
1360+ input_vectors_work_items_per_worker ,
1361+ )
13421362
13431363 vectors_per_table_partitions = size / partitions
13441364 table_partitions_per_work_item = int (
@@ -1354,10 +1374,15 @@ def consolidate_and_vacuum(
13541374 math .ceil (table_partitions_work_items / MAX_TASKS_PER_STAGE )
13551375 )
13561376 table_partitions_work_tasks = MAX_TASKS_PER_STAGE
1357- logger .debug ("table_partitions_per_work_item %d" , table_partitions_per_work_item )
1377+ logger .debug (
1378+ "table_partitions_per_work_item %d" , table_partitions_per_work_item
1379+ )
13581380 logger .debug ("table_partitions_work_items %d" , table_partitions_work_items )
13591381 logger .debug ("table_partitions_work_tasks %d" , table_partitions_work_tasks )
1360- logger .debug ("table_partitions_work_items_per_worker %d" , table_partitions_work_items_per_worker )
1382+ logger .debug (
1383+ "table_partitions_work_items_per_worker %d" ,
1384+ table_partitions_work_items_per_worker ,
1385+ )
13611386
13621387 logger .debug ("Creating arrays" )
13631388 create_arrays (
0 commit comments