Skip to content
Merged
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
6 changes: 4 additions & 2 deletions src/llmcompressor/modifiers/quantization/gptq/base.py
Original file line number Diff line number Diff line change
Expand Up @@ -2,7 +2,7 @@
from typing import Dict, List, Optional, Tuple, Union

import torch
from compressed_tensors.offload.dist_utils import is_distributed
from compressed_tensors.offload.dist_utils import as_broadcastable, is_distributed
from compressed_tensors.quantization import (
QuantizationConfig,
QuantizationScheme,
Expand Down Expand Up @@ -358,7 +358,9 @@ def _broadcast_quantized_params(self, module_list, module_to_rank):
if getattr(module, attr, None) is not None:
pending_comms.append(
dist.broadcast(
getattr(module, attr), src=src_rank, async_op=True
as_broadcastable(getattr(module, attr)),
src=src_rank,
async_op=True,
)
)
wait_for_comms(pending_comms)
Expand Down
Loading