Website: Implement a check for differing benchmark results and metadata#144
Conversation
|
The latest updates on your projects. Learn more about Vercel for Git ↗︎
|
|
@siddharth-krishna CI is failing now because there are some benchmarks that appear in benchmark-result.csv but not in result/meta-data.yaml. I think meta-data.yaml should be updated, or we should remove these missing benchmarks from benchmark-result.csv. |
|
Ah yes, nice catch. These are old results (which we are using just for development), I will do a new benchmark run this week that will only run things in the |
1f5440a to
6328c57
Compare
4d3ac70 to
19277ab
Compare
siddharth-krishna
left a comment
There was a problem hiding this comment.
Sorry Jacek, it looks like there are still some mistakes in which benchmarks were removed. If you don't mind, let me do this quickly and push to your branch.
results/benchmark_results.csv
Outdated
| Sienna_modified_RTS_GMLC_DA_sys_NetPTDF_Horizon12_Day29,1-1h,glpk,5.0,2020,TO,Timeout,600,132.948,,, | ||
| Sienna_modified_RTS_GMLC_DA_sys_NetPTDF_Horizon12_Day314,1-1h,glpk,5.0,2020,TO,Timeout,600,132.74,,, | ||
| Sienna_modified_RTS_GMLC_DA_sys_NetPTDF_Horizon12_Day332,1-1h,glpk,5.0,2020,TO,Timeout,600,135.172,,, | ||
| Sienna_modified_RTS_GMLC_DA_sys_NetPTDF_Horizon12_Day332,1-1h,glpk,5.0,2020,unknown,unknown,1.9189174175262451,150.836,0.0,, |
There was a problem hiding this comment.
Hmm I didn't expect any results to be modified -- was this intentional? If not, could we revert the modifications and ensure that the diff only deletes rows corresponding to sizes that were removed from results/metadata.yaml? Thanks
| Sienna_modified_RTS_GMLC_DA_sys_NetTransport_Horizon48_Day332,1-1h,highs,1.8.1,2024,TO,Timeout,600,625.368,,, | ||
| Sienna_modified_RTS_GMLC_DA_sys_NetTransport_Horizon48_Day332,1-1h,scip,9.1.1,2024,TO,Timeout,600,961.792,,, | ||
| genx-1_three_zones,3-1h,highs,1.8.1,2024,TO,Timeout,600,1784.696,,, | ||
| genx-1_three_zones,3-1h,scip,9.1.1,2024,TO,Timeout,600,1320.512,,, |
There was a problem hiding this comment.
It feels like too many rows have been deleted...
|
For the record, I used this script to figure out which benchmark results to keep: import yaml
meta = yaml.safe_load(open("results/metadata.yaml"))
benchs = set()
for b, d in meta['benchmarks'].items():
for s in d['Sizes']:
benchs.add((b, s['Name']))
results = open('results/benchmark_results.csv').readlines()
results1 = []
for line in results:
parts = line.split(',')
if parts[0] == 'Benchmark' or (parts[0], parts[1]) in benchs:
results1.append(line)
with open('results/benchmark_results.csv', "w") as f:
for line in results1:
f.write(line)
results = open('results/benchmark_results_mean_stddev.csv').readlines()
results1 = []
for line in results:
parts = line.split(',')
if parts[0] == 'Benchmark' or (parts[0], parts[1]) in benchs:
results1.append(line)
len(results), len(results1)
with open('results/benchmark_results_mean_stddev.csv', "w") as f:
for line in results1:
f.write(line) |
No description provided.