Skip to content

Commit bf0614d

Browse files
authored
Add git-lfs instruction
1 parent e76b480 commit bf0614d

File tree

1 file changed

+4
-0
lines changed

1 file changed

+4
-0
lines changed

README.md

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -111,6 +111,10 @@ We take a public version of the ShareGPT dataset, which is a subset of the Vicun
111111
```bash
112112
git clone https://huggingface.co/datasets/Aeala/ShareGPT_Vicuna_unfiltered
113113
```
114+
Remark: If you haven't installed `git-lfs`, please install it before clone:
115+
```bash
116+
git lfs install
117+
```
114118
#### Train the model
115119
We follow the training setup from [FastChat](https://github.com/lm-sys/FastChat#fine-tuning), but with a much larger learning rate because we freeze the original model and only train the new heads. Here is the training command for the Vicuna-7b model on 4 GPUs. Since we are only training the new heads, the training does not require a lot of memory, and only data parallelism is needed. You can modify the script to fit your own setup. For larger models, we use the same setup. You can also use `--load_in_8bit` or `--load_in_4bit` to load the base model in quantized format.
116120
```bash

0 commit comments

Comments
 (0)