We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 56f5a08 commit 2def489Copy full SHA for 2def489
modelopt/torch/quantization/tensor_quant.py
@@ -53,6 +53,7 @@ def _fp8_eager(x, amax=None):
53
x = x.to(torch.float32) * scale_inv
54
return x.to(dtype)
55
56
+
57
def fp8_eager(x, amax):
58
"""Eager mode implementation of FP8 quantization."""
59
return _fp8_eager(x, amax)
0 commit comments