1515</p >
1616
1717## Features
18- 1 . supported SFT methods: [ lora ] ( https://arxiv.org/abs/2106.09685 ) , [ qlora ] ( https://arxiv.org/abs/2305.14314 ) , full(full parameter fine-tuning)
18+ 1 . supported SFT methods: [ LoRA ] ( https://arxiv.org/abs/2106.09685 ) , [ QLoRA ] ( https://arxiv.org/abs/2305.14314 ) , full(full parameter fine-tuning)
19192 . supported models:
2020 1 . qwen series: qwen-7b, [ qwen-7b-chat] ( https://github.com/QwenLM/Qwen-7B )
2121 2 . qwen-vl series: qwen-vl, [ qwen-vl-chat] ( https://github.com/QwenLM/Qwen-VL )
22- 3 . baichuan series: baichuan-7b, baichuan-13b, baichuan-13b-chat, baichuan2-7b, baichuan2-7b-chat, baichuan2-13b, baichuan2-13b-chat
23- 4 . chatglm2 series: chatglm2-6b, chatglm2-6b-32k
24- 5 . llama series: llama2-7b, llama2-7b-chat, llama2-13b, llama2-13b-chat, llama2-70b, llama2-70b-chat
25- 6 . openbuddy-llama series: openbuddy-llama2-13b, openbuddy-llama-65b, openbuddy-llama2-70b
26- 7 . internlm series: internlm-7b, internlm-7b-chat, internlm-7b-chat-8k, internlm-20b, internlm-20 -chat
27- 8 . other: polylm-13b, seqgpt-560m
22+ 3 . baichuan series: baichuan-7b, baichuan-13b, baichuan-13b-chat, baichuan2-7b, [ baichuan2-7b-chat] ( https://modelscope.cn/models/baichuan-inc/Baichuan2-7B-Chat/summary ) , baichuan2-13b, baichuan2-13b-chat
23+ 4 . chatglm2 series: [ chatglm2-6b] ( https://modelscope.cn/models/ZhipuAI/chatglm2-6b/summary ) , chatglm2-6b-32k
24+ 5 . llama series: llama2-7b, llama2-7b-chat, llama2-13b, llama2-13b-chat, llama2-70b, [ llama2-70b-chat] ( https://modelscope.cn/models/modelscope/Llama-2-70b-chat-ms/summary )
25+ 6 . openbuddy-llama series: openbuddy-llama2-13b, openbuddy-llama-65b, [ openbuddy-llama2-70b] ( https://modelscope.cn/models/OpenBuddy/openbuddy-llama2-70b-v10.1-bf16/summary )
26+ 7 . internlm series: internlm-7b, internlm-7b-chat, internlm-7b-chat-8k, [ internlm-20b] ( https://modelscope.cn/models/Shanghai_AI_Laboratory/internlm-20b/summary ) , [ internlm-20b -chat] ( https://modelscope.cn/models/Shanghai_AI_Laboratory/internlm-20b-chat/summary )
27+ 8 . other: [ polylm-13b] ( https://modelscope.cn/models/damo/nlp_polylm_13b_text_generation/summary ) , [ seqgpt-560m] ( https://modelscope.cn/models/damo/nlp_seqgpt-560m/summary )
28283 . supported features: quantization, DDP, model parallelism(device map), gradient checkpointing, gradient accumulation, pushing to modelscope hub, custom datasets, multimodal and agent SFT, mutli-round chat, ...
29294 . supported datasets:
30- 1 . NLP: alpaca-en( gpt4), alpaca-zh( gpt4) , finance-en, multi-alpaca-all, code-en, instinwild-en, instinwild-zh, cot-en, cot-zh, firefly-all-zh, poetry-zh, instruct-en, gpt4all-en, cmnli-zh, jd-zh, dureader-robust-zh, medical-en, medical-zh, medical-mini-zh, sharegpt-en, sharegpt-zh, code-python-zh, advertise-gen
30+ 1 . NLP: [ alpaca-en] ( https://modelscope.cn/datasets/AI-ModelScope/alpaca- gpt4-data-en/summary ) (gpt4), [ alpaca-zh] ( https://modelscope.cn/datasets/AI-ModelScope/alpaca- gpt4-data-zh/summary ) (gpt4) , finance-en, multi-alpaca-all, code-en, instinwild-en, instinwild-zh, cot-en, cot-zh, firefly-all-zh, poetry-zh, instruct-en, gpt4all-en, cmnli-zh, [ jd-zh] ( https://modelscope.cn/datasets/DAMO_NLP/jd/summary ) , [ dureader-robust-zh] ( https://modelscope.cn/datasets/modelscope/DuReader_robust-QG/summary ) , medical-en, medical-zh, medical-mini-zh, sharegpt-en, sharegpt-zh, [ code-python-zh] ( https://modelscope.cn/datasets/codefuse-ai/CodeExercise-Python-27k/summary ) , [ advertise-gen] ( https://modelscope.cn/datasets/lvjianjin/AdvertiseGen/summary )
3131 2 . agent: [ damo-agent-zh] ( https://modelscope.cn/datasets/damo/MSAgent-Bench/summary ) , damo-agent-mini-zh
32- 3 . multi-modal: coco-en
33- 4 . other: cls-fudan-news-zh, ner-jave-zh
32+ 3 . multi-modal: [ coco-en] ( https://modelscope.cn/datasets/modelscope/coco_2014_caption/summary )
33+ 4 . other: [ cls-fudan-news-zh] ( https://modelscope.cn/datasets/damo/zh_cls_fudan-news/files ) , [ ner-jave-zh] ( https://modelscope.cn/datasets/damo/zh_ner-JAVE/summary )
34345 . supported templates: chatml(qwen), baichuan, chatglm2, llama, openbuddy-llama, default, default-generation
3535
3636## Prepare the Environment
@@ -53,13 +53,11 @@ pip install matplotlib scikit-learn tqdm tensorboard -U
5353pip install transformers datasets -U
5454pip install accelerate transformers_stream_generator -U
5555
56- pip install ms-swift modelscope -U
56+ pip install modelscope -U
5757# Recommended installation from source code for faster bug fixes
5858git clone https://github.com/modelscope/swift.git
5959cd swift
60- pip install -r requirements.txt
6160pip install .
62- # same as modelscope...(git clone ...)
6361```
6462
6563## Run SFT and Inference
@@ -108,6 +106,7 @@ bash scripts/qwen_7b_chat/full_mp/infer.sh
108106# Recommended experimental environment: A100
109107bash scripts/qwen_7b_chat/full_mp_ddp/sft.sh
110108bash scripts/qwen_7b_chat/full_mp_ddp/infer.sh
109+
111110# For more scripts, please see `scripts/` folder.
112111```
113112
0 commit comments