Skip to content

Commit c391942

Browse files
Move triton import inside function to work on Windows
Signed-off-by: Keval Morabia <[email protected]>
1 parent 4d1eb0c commit c391942

File tree

1 file changed

+2
-1
lines changed

1 file changed

+2
-1
lines changed

modelopt/torch/quantization/qtensor/nvfp4_tensor.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,6 @@
1919

2020
from ..backends.utils import fp4_compatible
2121
from ..qtensor.base_qtensor import BaseQuantizedTensor
22-
from ..triton.fp4_kernel import fp4_dequantize
2322
from ..utils import reduce_amax, reduce_block_amax, reduce_block_padding
2423

2524
# Define conversion tables
@@ -266,6 +265,8 @@ def _unpack_tensor(input: torch.Tensor):
266265
) from e
267266

268267
if fast:
268+
from ..triton.fp4_kernel import fp4_dequantize
269+
269270
return fp4_dequantize(
270271
self._quantized_data,
271272
kwarg["scale"],

0 commit comments

Comments
 (0)