Skip to content

Commit fb8f415

Browse files
chunyuan-wsvekars
andauthored
Update prototype_source/max_autotune_on_CPU_tutorial.rst
Co-authored-by: Svetlana Karslioglu <[email protected]>
1 parent 7540b9a commit fb8f415

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

prototype_source/max_autotune_on_CPU_tutorial.rst

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -24,8 +24,8 @@ We have covered most popular data types, including FP32, BF16, FP16, and INT8, w
2424

2525
While the development is still in progress, we have already seen promising speedups over pure ATen-based GEMMs as measured by the three benchmark suites and the inference of LLMs.
2626

27-
How to activate ``max-autotune`` mode
28-
------------
27+
Activating the ``max-autotune`` mode
28+
-------------------------------------
2929
To activate the ``max-autotune`` mode in PyTorch, set the ``mode`` argument to ``max-autotune`` when compiling your model using ``torch.compile``.
3030
If you prefer to bypass the tuning process and always use the CPP template implementations, you can configure this via an environment variable:
3131
``export TORCHINDUCTOR_MAX_AUTOTUNE_GEMM_BACKENDS=CPP``.

0 commit comments

Comments
 (0)