File tree Expand file tree Collapse file tree 2 files changed +1
-3
lines changed
examples/quantization_2of4_sparse_w4a16 Expand file tree Collapse file tree 2 files changed +1
-3
lines changed Original file line number Diff line number Diff line change @@ -58,7 +58,7 @@ Quantization is applied by selecting an algorithm and calling the `oneshot` API.
5858``` python
5959from llmcompressor.modifiers.smoothquant import SmoothQuantModifier
6060from llmcompressor.modifiers.quantization import GPTQModifier
61- from llmcompressor.transformers import oneshot
61+ from llmcompressor import oneshot
6262
6363# Select quantization algorithm. In this case, we:
6464# * apply SmoothQuant to make the activations easier to quantize
Original file line number Diff line number Diff line change 3333bf16 = False # using full precision for training
3434lr_scheduler_type = "cosine"
3535warmup_ratio = 0.1
36- preprocessing_num_workers = 8
3736
3837# this will run the recipe stage by stage:
3938# oneshot sparsification -> finetuning -> oneshot quantization
5352 learning_rate = learning_rate ,
5453 lr_scheduler_type = lr_scheduler_type ,
5554 warmup_ratio = warmup_ratio ,
56- preprocessing_num_workers = preprocessing_num_workers ,
5755)
5856logger .info (
5957 "Note: llcompressor does not currently support running " ,
You can’t perform that action at this time.
0 commit comments