Skip to content

Commit 83d742c

Browse files
jananisriramfacebook-github-bot
authored andcommitted
Validate exhaustive autotuning for FP8 Inductor templates (#355)
Summary: X-link: pytorch/pytorch#161442 Validate exhaustive autotuning for FP8 Inductor templates: scaled MM templates require `block_k >= 32`. Before, exhaustive autotuning defaulted to a limited set of autotuning configs, as limitations for exhaustively autotuning on FP8 shapes had not been tested. Reviewed By: coconutruben Differential Revision: D80958642
1 parent 07088a7 commit 83d742c

File tree

1 file changed

+4
-0
lines changed

1 file changed

+4
-0
lines changed

tritonbench/operators/fp8_gemm/fp8_gemm.py

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -17,6 +17,10 @@
1717

1818
from .tutorial import matmul as tutorial_matmul
1919

20+
torch._dynamo.config.recompile_limit = (
21+
10000 # Set high recompile limit to allow for exhausting autotuning
22+
)
23+
2024
logger = logging.getLogger(__name__)
2125
try:
2226
from .persistent import (

0 commit comments

Comments
 (0)