We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 87c4dc4 commit 83163bcCopy full SHA for 83163bc
src/diffusers/quantizers/finegrained_fp8/finegrained_fp8_quantizer.py
@@ -131,9 +131,8 @@ def create_quantized_param(
131
scale = scale.reshape(scale_orig_shape).squeeze().reciprocal()
132
133
# Load into the model
134
- module._buffers[tensor_name] = quantized_param.to(target_device)
135
- module._buffers["weight_scale_inv"] = scale.to(target_device)
136
- # print("_buffers[0]", module._buffers["weight_scale_inv"])
+ module._parameters[tensor_name] = quantized_param.to(target_device)
+ module._parameters["weight_scale_inv"] = scale.to(target_device)
137
138
def check_if_quantized_param(
139
self,
0 commit comments