@@ -62,7 +62,8 @@ Users can check the [documentation of SWIFT](docs/source/GetStarted/快速使用
6262
6363
6464## 🎉 News
65- - 2024.1.26: Support [ yi-vl-6b-chat] ( https://github.com/modelscope/swift/tree/main/examples/pytorch/llm/scripts/yi_vl_6b_chat ) , yi-vl-34b-chat.
65+ - 2024.1.29: Support internlm2-math series: internlm2-math-7b, internlm2-math-7b-chat, internlm2-math-20b, internlm2-math-20b-chat.
66+ - 🔥2024.1.26: Support [ yi-vl-6b-chat] ( https://github.com/modelscope/swift/tree/main/examples/pytorch/llm/scripts/yi_vl_6b_chat ) , yi-vl-34b-chat.
6667- 2024.1.24: Support codefuse-codegeex2-6b-chat, codefuse-qwen-14b-chat.
6768- 2024.1.23: Support orion series: orion-14b, [ orion-14b-chat] ( https://github.com/modelscope/swift/tree/main/examples/pytorch/llm/scripts/orion_14b_chat ) .
6869- 2024.1.20: Support [ xverse-13b-256k] ( https://github.com/modelscope/swift/tree/main/examples/pytorch/llm/scripts/xverse_13b_256k ) , xverse-65b-v2, xverse-65b-chat.
@@ -164,7 +165,7 @@ from swift.llm import (
164165 infer_main, sft_main, app_ui_main, merge_lora_main
165166)
166167
167- model_type = ModelType.qwen_1_8b_chat
168+ model_type = ModelType.qwen_1_8b
168169sft_args = SftArguments(
169170 model_type = model_type,
170171 train_dataset_sample = 2000 ,
@@ -178,7 +179,7 @@ torch.cuda.empty_cache()
178179infer_args = InferArguments(
179180 ckpt_dir = best_model_checkpoint,
180181 load_dataset_config = True ,
181- show_dataset_sample = 10 )
182+ val_dataset_sample = 10 )
182183# merge_lora_main(infer_args)
183184result = infer_main(infer_args)
184185torch.cuda.empty_cache()
@@ -222,6 +223,8 @@ app_ui_main(infer_args)
222223 - [ deepseek-coder] ( https://github.com/deepseek-ai/DeepSeek-Coder ) series: deepseek-coder-1_3b, deepseek-coder-1_3b-instruct, deepseek-coder-6_7b, deepseek-coder-6_7b-instruct, deepseek-coder-33b, deepseek-coder-33b-instruct.
223224 - [ codegeex2] ( https://github.com/THUDM/CodeGeeX2 ) series: codegeex2-6b.
224225 - [ phi] ( https://www.microsoft.com/en-us/research/blog/phi-2-the-surprising-power-of-small-language-models/ ) series: phi2-3b.
226+ - Math:
227+ - [ internlm2-math] ( https://github.com/InternLM/InternLM-Math ) series: internlm2-math-7b, internlm2-math-7b-chat, internlm2-math-20b, internlm2-math-20b-chat.
225228- Supported Datasets: [[ Detailed Info]] ( https://github.com/modelscope/swift/blob/main/docs/source/LLM/%E6%94%AF%E6%8C%81%E7%9A%84%E6%A8%A1%E5%9E%8B%E5%92%8C%E6%95%B0%E6%8D%AE%E9%9B%86.md#%E6%95%B0%E6%8D%AE%E9%9B%86 )
226229 - NLP:
227230 - General: 🔥alpaca-en(gpt4), 🔥alpaca-zh(gpt4), multi-alpaca-all, instinwild-en, instinwild-zh, cot-en, cot-zh, firefly-all-zh, instruct-en, gpt4all-en, sharegpt-en, sharegpt-zh, tutu-v2-sft-mixture, wikipedia-zh, open-orca, open-orca-gpt4, sharegpt-gpt4.
0 commit comments