Skip to content

Commit c80416a

Browse files
committed
minor
Signed-off-by: realAsma <[email protected]>
1 parent 34bcee8 commit c80416a

File tree

1 file changed

+1
-0
lines changed

1 file changed

+1
-0
lines changed

modelopt/torch/quantization/tensor_quant.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -61,6 +61,7 @@ def scaled_e4m3_impl(
6161
None if inputs.device.index == torch.cuda.current_device() else inputs.device.index
6262
):
6363
if amax is None:
64+
# This adds overhead; however this is not a common use case.
6465
amax = torch.tensor(448.0, device=inputs.device, dtype=inputs.dtype)
6566
if amax.numel() == 1:
6667
outputs = cuda_ext_fp8.fake_e4m3fy(inputs, amax)

0 commit comments

Comments
 (0)