Skip to content

Commit bbb634a

Browse files
dik654brian-dellabetta
authored andcommitted
[GPTQ] Move modifier to top-level for consistent folder structure (vllm-project#2368)
## SUMMARY: Move GPTQModifier from `modifiers/quantization/gptq/` to `modifiers/gptq/` for consistent folder structure with AWQ and AutoRound (related: vllm-project#2306). - Add deprecation wrapper at old import path for backward compatibility - Exclude old GPTQ paths from ModifierFactory to prevent duplicate registration - Update test and example imports to new canonical path ## TEST PLAN: Import verification (all passed): - from llmcompressor.modifiers.gptq import GPTQModifier (new path, no warning) - from llmcompressor.modifiers.quantization import GPTQModifier (BC, no warning) - from llmcompressor.modifiers.quantization.gptq import GPTQModifier (BC, DeprecationWarning) - ModifierFactory.refresh() registers GPTQModifier from new location pytest (11 passed, 3 skipped for GPU): - tests/llmcompressor/transformers/gptq/test_gptq_oneshot.py - tests/llmcompressor/pytorch/modifiers/pruning/sparsegpt/test_pytorch.py - tests/llmcompressor/transformers/compression/test_recipe_parsing.py (requires GPU) ruff check + ruff format passed --------- Signed-off-by: 김대익 <33992354+dik654@users.noreply.github.com> Co-authored-by: Brian Dellabetta <brian-dellabetta@users.noreply.github.com> Signed-off-by: yiliu30 <yi4.liu@intel.com>
1 parent 87252fd commit bbb634a

File tree

41 files changed

+55
-39
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

41 files changed

+55
-39
lines changed

docs/guides/saving_a_model.md

Lines changed: 1 addition & 1 deletion

examples/big_models_with_sequential_onloading/llama3.3_70b.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
from transformers import AutoModelForCausalLM, AutoTokenizer
33

44
from llmcompressor import oneshot
5-
from llmcompressor.modifiers.quantization import GPTQModifier
5+
from llmcompressor.modifiers.gptq import GPTQModifier
66
from llmcompressor.modifiers.transform.smoothquant import SmoothQuantModifier
77

88
# Select model and load it.

examples/multimodal_audio/whisper_example.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@
88
)
99

1010
from llmcompressor import oneshot
11-
from llmcompressor.modifiers.quantization import GPTQModifier
11+
from llmcompressor.modifiers.gptq import GPTQModifier
1212

1313
# Select model and load it.
1414
MODEL_ID = "openai/whisper-large-v3"

examples/multimodal_vision/gemma3_example.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@
44
from transformers import AutoProcessor, Gemma3ForConditionalGeneration
55

66
from llmcompressor import oneshot
7-
from llmcompressor.modifiers.quantization import GPTQModifier
7+
from llmcompressor.modifiers.gptq import GPTQModifier
88

99
# Load model.
1010
model_id = "google/gemma-3-4b-it"

examples/multimodal_vision/idefics3_example.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@
66
from transformers import AutoProcessor, Idefics3ForConditionalGeneration
77

88
from llmcompressor import oneshot
9-
from llmcompressor.modifiers.quantization import GPTQModifier
9+
from llmcompressor.modifiers.gptq import GPTQModifier
1010

1111
# Load model.
1212
model_id = "HuggingFaceM4/Idefics3-8B-Llama3" # or "HuggingFaceTB/SmolVLM-Instruct"

examples/multimodal_vision/internvl3_example.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33
from transformers import AutoModelForImageTextToText, AutoProcessor
44

55
from llmcompressor import oneshot
6-
from llmcompressor.modifiers.quantization import GPTQModifier
6+
from llmcompressor.modifiers.gptq import GPTQModifier
77

88
# Load model.
99
model_id = "OpenGVLab/InternVL3-8B-hf"

examples/multimodal_vision/llama4_example.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33
from transformers import Llama4ForConditionalGeneration, Llama4Processor
44

55
from llmcompressor import oneshot
6-
from llmcompressor.modifiers.quantization import GPTQModifier
6+
from llmcompressor.modifiers.gptq import GPTQModifier
77

88
# Select model and load it.
99
model_id = "meta-llama/Llama-4-Scout-17B-16E-Instruct"

examples/multimodal_vision/llava_example.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@
44
from transformers import AutoProcessor, LlavaForConditionalGeneration
55

66
from llmcompressor import oneshot
7-
from llmcompressor.modifiers.quantization import GPTQModifier
7+
from llmcompressor.modifiers.gptq import GPTQModifier
88

99
# Load model.
1010
model_id = "llava-hf/llava-1.5-7b-hf"

examples/multimodal_vision/medgemma_example.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@
44
from transformers import AutoProcessor, Gemma3ForConditionalGeneration
55

66
from llmcompressor import oneshot
7-
from llmcompressor.modifiers.quantization import GPTQModifier
7+
from llmcompressor.modifiers.gptq import GPTQModifier
88

99
# Load model.
1010
model_id = "google/medgemma-27b-it"

examples/multimodal_vision/mistral3_example.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@
1212
)
1313

1414
from llmcompressor import oneshot
15-
from llmcompressor.modifiers.quantization import GPTQModifier
15+
from llmcompressor.modifiers.gptq import GPTQModifier
1616

1717
# Load model.
1818
model_id = "mistralai/Mistral-Small-3.1-24B-Instruct-2503"

0 commit comments

Comments
 (0)