diff --git a/src/llmcompressor/entrypoints/model_free/helpers.py b/src/llmcompressor/entrypoints/model_free/helpers.py index ef45f09346..9b1b130880 100644 --- a/src/llmcompressor/entrypoints/model_free/helpers.py +++ b/src/llmcompressor/entrypoints/model_free/helpers.py @@ -4,7 +4,7 @@ from typing import Mapping, TypeVar import torch -from compressed_tensors.utils.match import _match_name +from compressed_tensors.utils.match import match_name from loguru import logger from transformers.file_utils import CONFIG_NAME @@ -84,7 +84,7 @@ def natural_key(s: str) -> list[str | int]: for name in names: # match until we get a full set for target in targets: - if _match_name(name, target): + if match_name(name, target): if matches[target] is None: matches[target] = name else: diff --git a/src/llmcompressor/entrypoints/model_free/process.py b/src/llmcompressor/entrypoints/model_free/process.py index 0a3d86efec..9b4412ee1d 100644 --- a/src/llmcompressor/entrypoints/model_free/process.py +++ b/src/llmcompressor/entrypoints/model_free/process.py @@ -5,7 +5,7 @@ import torch from compressed_tensors.quantization import QuantizationScheme -from compressed_tensors.utils.match import _match_name +from compressed_tensors.utils.match import match_name from safetensors.torch import load_file, save_file from torch.nn import Module @@ -31,7 +31,7 @@ def iter_quantizable_tensors( for name in list(tensors.keys()): module_name, param_name = name.rsplit(".", 1) is_linear_weight = param_name == "weight" and not module_name.endswith("norm") - is_ignored = any(_match_name(module_name, ign) for ign in ignore) + is_ignored = any(match_name(module_name, ign) for ign in ignore) if not is_linear_weight or is_ignored: continue