You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+5-2Lines changed: 5 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -39,6 +39,8 @@ To facilitate use by users unfamiliar with deep learning, we provide a Gradio we
39
39
Additionally, we are expanding capabilities for other modalities. Currently, we support full-parameter training and LoRA training for AnimateDiff.
40
40
41
41
## 🎉 News
42
+
- 2024.04.22: Support for inference, fine-tuning, and deployment of **chinese-llama-alpaca-2** series models. This includes:chinese-llama-2-1.3b, chinese-llama-2-7b, chinese-llama-2-13b, chinese-alpaca-2-1.3b, chinese-alpaca-2-7b and chinese-alpaca-2-13b along with their corresponding 16k and 64k long text versions.
43
+
- 2024.04.22: Support for inference and fine-tuning of Llama3 GPTQ-Int4, GPTQ-Int8, and AWQ series models. Support for inference and fine-tuning of chatglm3-6b-128k, Openbuddy-Llama3.
42
44
- 2024.04.20: Support for inference, fine-tuning, and deployment of **Atom** series models. This includes: Atom-7B and Atom-7B-Chat. use [this script](https://github.com/modelscope/swift/blob/main/examples/pytorch/llm/scripts/atom_7b_chat/lora/sft.sh) to train.
43
45
- 2024.04.19: Support for single-card, DDP, ZeRO2, and ZeRO3 training and inference with NPU, please refer to [NPU Inference and Fine-tuning Best Practices](docs/source_en/LLM/NPU-best-practice.md).
44
46
- 2024.04.19: Support for inference, fine-tuning, and deployment of **Llama3** series models. This includes: Llama-3-8B, Llama-3-8B-Instruct, Llama-3-70B, and Llama-3-70B-Instruct. use [this script](https://github.com/modelscope/swift/blob/main/examples/pytorch/llm/scripts/llama3_8b_instruct/lora/sft.sh) to train.
@@ -441,7 +443,7 @@ CUDA_VISIBLE_DEVICES=0 swift deploy \
441
443
| Yuan2 |[Langchao Yuan series models](https://github.com/IEIT-Yuan)| Chinese<br>English | 2B-102B | instruct model |
442
444
| XVerse |[XVerse series models](https://github.com/xverse-ai)| Chinese<br>English | 7B-65B | base model<br>chat model<br>long text model<br>MoE model |
443
445
| LLaMA2 |[LLaMA2 series models](https://github.com/facebookresearch/llama)| English | 7B-70B<br>including quantized versions | base model<br>chat model |
444
-
| LLaMA3 |[LLaMA3 series models](https://github.com/meta-llama/llama3)| English | 8B-70B | base model<br>chat model |
446
+
| LLaMA3 |[LLaMA3 series models](https://github.com/meta-llama/llama3)| English | 8B-70B<br>including quantized versions| base model<br>chat model |
445
447
| Mistral<br>Mixtral |[Mistral series models](https://github.com/mistralai/mistral-src)| English | 7B-22B | base model<br>instruct model<br>MoE model |
446
448
| YI |[01AI's YI series models](https://github.com/01-ai)| Chinese<br>English | 6B-34B<br>including quantized | base model<br>chat model<br>long text model |
447
449
| InternLM<br>InternLM2<br>InternLM2-Math |[Pujiang AI Lab InternLM series models](https://github.com/InternLM/InternLM)| Chinese<br>English | 1.8B-20B | base model<br>chat model<br>math model |
@@ -468,6 +470,7 @@ CUDA_VISIBLE_DEVICES=0 swift deploy \
468
470
| c4ai-command-r |[c4ai](https://cohere.com/command)| Multilingual | 35B-104B | chat model |
469
471
| WizardLM2 |[WizardLM2 series models](https://github.com/nlpxucan/WizardLM)| English | 7B-8x22B<br>including quantized versions | chat model<br>MoE model |
470
472
| Atom |[Atom](https://github.com/LlamaFamily/Llama-Chinese)| Chinese | 7B| base model<br>chat model|
473
+
| Chinese-LLaMA-Alpaca-2 |[Chinese-LLaMA-Alpaca-2](https://github.com/ymcui/Chinese-LLaMA-Alpaca-2)| Chinese | 1.3B-13B| base model<br>chat model<br>long text model|
0 commit comments