-
Notifications
You must be signed in to change notification settings - Fork 26
Description
Hello NeuroBench team,
I was reviewing the lstm_benchmark.py example for the Mackey-Glass dataset and noticed a potential issue regarding the evaluation setup. In the benchmark function, specifically at line 168, it appears the train_set is being used for evaluation:
test_set_loader = DataLoader(train_set,
batch_size=mg.testtime_pts,
shuffle=False)In addition, I cannot make sense out of why batch size is set to testtime_pts.
The actual test_set defined above in line 117 is not used anywhere else in the benchmark script.
I have also changed it to use the actual test_set but then performance gets much worse with sMAPEs of around 30.
Is it just me not understanding the code or is there a potential issue? I'd appreciate if you could clarify the reason behind using train_set for this particular evaluation step in the benchmark, or if this might be an oversight?