We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 9dfb31c commit 7a2d7c8Copy full SHA for 7a2d7c8
src/compressed_tensors/compressors/quantized_compressors/fp4_quantized.py
@@ -149,6 +149,13 @@ def compress_scale(
149
scale_exp = 127 + torch.floor(torch.log2(scale)).to(torch.int32) - 2
150
return scale_exp.to(quantization_args.scale_dtype)
151
152
+ def decompress_weight(
153
+ self,
154
+ compressed_data: Dict[str, Tensor],
155
+ quantization_args: Optional[QuantizationArgs] = None,
156
+ ) -> torch.Tensor:
157
+ raise NotImplementedError("MXFP4 Decompression is currently not supported")
158
+
159
160
@torch.compile(fullgraph=True, dynamic=True)
161
def pack_fp4_to_uint8(x: torch.Tensor) -> torch.Tensor:
0 commit comments