Skip to content

Using train_set for evaluation in MackeyGlass LSTM benchmark? #262

@paulhaider

Description

@paulhaider

Hello NeuroBench team,

I was reviewing the lstm_benchmark.py example for the Mackey-Glass dataset and noticed a potential issue regarding the evaluation setup. In the benchmark function, specifically at line 168, it appears the train_set is being used for evaluation:

    test_set_loader = DataLoader(train_set,
                                 batch_size=mg.testtime_pts,
                                 shuffle=False)

In addition, I cannot make sense out of why batch size is set to testtime_pts.
The actual test_set defined above in line 117 is not used anywhere else in the benchmark script.

I have also changed it to use the actual test_set but then performance gets much worse with sMAPEs of around 30.

Is it just me not understanding the code or is there a potential issue? I'd appreciate if you could clarify the reason behind using train_set for this particular evaluation step in the benchmark, or if this might be an oversight?

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions