Skip to content

Commit 6c0c2c9

Browse files
committed
[GPTQ] Move GPTQModifier to modifiers/gptq for consistent folder structure
Signed-off-by: 김대익 <33992354+dik654@users.noreply.github.com>
1 parent 732316c commit 6c0c2c9

File tree

39 files changed

+53
-37
lines changed

39 files changed

+53
-37
lines changed

docs/guides/saving_a_model.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -41,7 +41,7 @@ The simplest approach is to use `oneshot`, which handles both compression and wr
4141
```python
4242
from transformers import AutoModelForCausalLM, AutoTokenizer
4343
from llmcompressor import oneshot
44-
from llmcompressor.modifiers.quantization import GPTQModifier
44+
from llmcompressor.modifiers.gptq import GPTQModifier
4545

4646
# Load model
4747
model = AutoModelForCausalLM.from_pretrained("your-model")

examples/big_models_with_sequential_onloading/llama3.3_70b.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
from transformers import AutoModelForCausalLM, AutoTokenizer
33

44
from llmcompressor import oneshot
5-
from llmcompressor.modifiers.quantization import GPTQModifier
5+
from llmcompressor.modifiers.gptq import GPTQModifier
66
from llmcompressor.modifiers.transform.smoothquant import SmoothQuantModifier
77

88
# Select model and load it.

examples/multimodal_audio/whisper_example.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@
88
)
99

1010
from llmcompressor import oneshot
11-
from llmcompressor.modifiers.quantization import GPTQModifier
11+
from llmcompressor.modifiers.gptq import GPTQModifier
1212

1313
# Select model and load it.
1414
MODEL_ID = "openai/whisper-large-v3"

examples/multimodal_vision/gemma3_example.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@
44
from transformers import AutoProcessor, Gemma3ForConditionalGeneration
55

66
from llmcompressor import oneshot
7-
from llmcompressor.modifiers.quantization import GPTQModifier
7+
from llmcompressor.modifiers.gptq import GPTQModifier
88

99
# Load model.
1010
model_id = "google/gemma-3-4b-it"

examples/multimodal_vision/idefics3_example.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@
66
from transformers import AutoProcessor, Idefics3ForConditionalGeneration
77

88
from llmcompressor import oneshot
9-
from llmcompressor.modifiers.quantization import GPTQModifier
9+
from llmcompressor.modifiers.gptq import GPTQModifier
1010

1111
# Load model.
1212
model_id = "HuggingFaceM4/Idefics3-8B-Llama3" # or "HuggingFaceTB/SmolVLM-Instruct"

examples/multimodal_vision/internvl3_example.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33
from transformers import AutoModelForImageTextToText, AutoProcessor
44

55
from llmcompressor import oneshot
6-
from llmcompressor.modifiers.quantization import GPTQModifier
6+
from llmcompressor.modifiers.gptq import GPTQModifier
77

88
# Load model.
99
model_id = "OpenGVLab/InternVL3-8B-hf"

examples/multimodal_vision/llama4_example.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33
from transformers import Llama4ForConditionalGeneration, Llama4Processor
44

55
from llmcompressor import oneshot
6-
from llmcompressor.modifiers.quantization import GPTQModifier
6+
from llmcompressor.modifiers.gptq import GPTQModifier
77

88
# Select model and load it.
99
model_id = "meta-llama/Llama-4-Scout-17B-16E-Instruct"

examples/multimodal_vision/llava_example.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@
44
from transformers import AutoProcessor, LlavaForConditionalGeneration
55

66
from llmcompressor import oneshot
7-
from llmcompressor.modifiers.quantization import GPTQModifier
7+
from llmcompressor.modifiers.gptq import GPTQModifier
88

99
# Load model.
1010
model_id = "llava-hf/llava-1.5-7b-hf"

examples/multimodal_vision/medgemma_example.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@
44
from transformers import AutoProcessor, Gemma3ForConditionalGeneration
55

66
from llmcompressor import oneshot
7-
from llmcompressor.modifiers.quantization import GPTQModifier
7+
from llmcompressor.modifiers.gptq import GPTQModifier
88

99
# Load model.
1010
model_id = "google/medgemma-27b-it"

examples/multimodal_vision/mistral3_example.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@
1212
)
1313

1414
from llmcompressor import oneshot
15-
from llmcompressor.modifiers.quantization import GPTQModifier
15+
from llmcompressor.modifiers.gptq import GPTQModifier
1616

1717
# Load model.
1818
model_id = "mistralai/Mistral-Small-3.1-24B-Instruct-2503"

0 commit comments

Comments
 (0)