Skip to content

Commit d76e3d5

Browse files
committed
Update run_scratch_add_calibration_logic.py
1 parent ca785e5 commit d76e3d5

File tree

1 file changed

+7
-2
lines changed

1 file changed

+7
-2
lines changed

examples/tmp/run_scratch_add_calibration_logic.py

Lines changed: 7 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -133,11 +133,16 @@ def simulate_calibration(method_metadata, run_toy: bool = False):
133133
if __name__ == "__main__":
134134
metadata_lst = tabarena_method_metadata_collection.method_metadata_lst
135135
metadata_lst = [m for m in metadata_lst if m.method_type == "config"]
136-
run_toy = True
136+
run_toy = True # If True, only calculates using up to 5 configs per method and runs sequentially (debugger friendly).
137+
run_only_small_methods = True # If True, avoids downloading large method results (only runs models that have a single config). If False, will end up downloading 300+ GB of model predictions if not already present.
138+
if run_only_small_methods:
139+
metadata_lst = [m for m in metadata_lst if not m.can_hpo]
137140
if run_toy:
138141
out_dir = "calibration_results_toy"
139142
else:
140143
out_dir = "calibration_results"
144+
if run_only_small_methods:
145+
out_dir += "_only_small"
141146

142147
num_methods = len(metadata_lst)
143148
new_results_lst = []
@@ -160,7 +165,7 @@ def simulate_calibration(method_metadata, run_toy: bool = False):
160165
new_results=all_new_results,
161166
only_valid_tasks=True,
162167
average_seeds=False,
163-
score_on_val=True,
168+
# score_on_val=True, # Uncomment to look at validation scores instead of test scores
164169
)
165170

166171
all_new_results_methods = list(all_new_results["method"].unique())

0 commit comments

Comments
 (0)