You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Validate exhaustive autotuning for FP8 Inductor templates (#355)
Summary:
Pull Request resolved: #355
X-link: pytorch/pytorch#161442
Validate exhaustive autotuning for FP8 Inductor templates: scaled MM templates require `block_k >= 32`. Before, exhaustive autotuning defaulted to a limited set of autotuning configs, as limitations for exhaustively autotuning on FP8 shapes had not been tested.
Differential Revision: D80958642
0 commit comments