The performance model for most attention implementations is currently broken.
Most of them do not set "dtype_a_b" as needed in SPDA::get_simulation_time here
Even if we were to set dtype_a_b, most of the subclasses also do not set d_h as required in the function's call to
SDPA.get_simulation_time_func here