You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Now the ``linear2`` layer runs in FP16 as shown in the above logs.
78
+
Now the ``linear2`` layer runs in FP16 as shown in the above logs.
79
+
80
+
81
+
82
+
FP32 Accumulation
83
+
-----------------
84
+
85
+
When ``use_fp32_acc=True`` is set, Torch-TensorRT will attempt to use FP32 accumulation for matmul layers, even if the input and output tensors are in FP16. This is particularly useful for models that are sensitive to numerical errors introduced by lower-precision accumulation.
86
+
87
+
.. important::
88
+
89
+
When enabling ``use_fp32_acc=True``, **explicit typing must be enabled** by setting ``use_explicit_typing=True``. Without ``use_explicit_typing=True``, the accumulation type may not be properly respected, and you may not see the intended numerical benefits.
0 commit comments