Skip to content

Commit c3a883c

Browse files
committed
minor update
Signed-off-by: Suguna Velury <[email protected]>
1 parent 4057a23 commit c3a883c

File tree

2 files changed

+1
-5
lines changed

2 files changed

+1
-5
lines changed

examples/llm_qat/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -354,7 +354,7 @@ To perform QLoRA training, run:
354354
--lora True
355355
```
356356

357-
After performing QLoRA training the final checkpoint is exported to be ready for deployment. For more details about QLoRA deployment using vLLM dere to the documentation [here](https://docs.vllm.ai/en/latest/features/lora.html). To deploy with vLLM, run:
357+
After performing QLoRA training the final checkpoint exported is ready for deployment using vLLM. For more details about QLoRA deployment using vLLM refer to the documentation [here](https://docs.vllm.ai/en/latest/features/lora.html). To deploy with vLLM, run:
358358

359359
```sh
360360
vllm serve llama3-fp4-qlora/base_model --enable-lora --lora-modules adapter=llama3-fp4-qlora --port 8000 --tokenizer llama3-fp4-qlora

modelopt/torch/quantization/plugins/transformers_trainer.py

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,6 @@
1616
"""ModelOpt plugin for transformers Trainer."""
1717

1818
import gc
19-
import json
2019
import os
2120
import types
2221
from dataclasses import dataclass, field
@@ -219,7 +218,6 @@ def forward_loop(model):
219218
gc.collect()
220219

221220
self._save_modelopt_state_with_weights()
222-
223221
torch.cuda.empty_cache()
224222

225223
if self.accelerator.is_main_process:
@@ -294,8 +292,6 @@ def export_base_model(self):
294292
"""Export the basemodel to HF checkpoint for deployment."""
295293
# Save config.json
296294
if self.accelerator.is_main_process:
297-
with open(f"{self.args.output_dir}/config.json", "w") as f:
298-
json.dump(self.model.config.to_dict(), f, indent=2)
299295
export_hf_checkpoint(self.model, export_dir=f"{self.args.output_dir}/base_model")
300296

301297
def _patch_accelerate_for_fsdp2_fix(self):

0 commit comments

Comments
 (0)