Skip to content

Commit ed7b7c7

Browse files
committed
fix lazy loading
1 parent ab0b550 commit ed7b7c7

File tree

1 file changed

+2
-0
lines changed

1 file changed

+2
-0
lines changed

convert_hf_to_gguf.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -352,6 +352,8 @@ def dequant_compressed_tensor(weight: Tensor, scale: Tensor) -> Tensor:
352352
device=weight.device,
353353
dtype=torch.int32,
354354
)
355+
if self.lazy:
356+
unpacked = LazyTorchTensor.from_eager(unpacked)
355357
for i in range(pack_factor):
356358
unpacked[:, i::pack_factor] = (weight >> (num_bits * i)) & mask
357359
# TODO: may need to unpad

0 commit comments

Comments
 (0)