We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 94ec97d commit 35f3da2Copy full SHA for 35f3da2
modelopt/torch/quantization/nn/modules/tensor_quantizer.py
@@ -960,6 +960,7 @@ def forward(self, inputs):
960
and self._fake_quant
961
):
962
# Reshape is required if the logic is not handled in the simulation kernel
963
+ # Only MX format and NVFP4 reshape are currently supported by the kernel.
964
self._setup_for_blockquant(inputs)
965
setattr(self, "_original_input_shape", inputs.shape)
966
inputs = self._process_for_blockquant(inputs)
0 commit comments