You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
- 2024.04.18: Supported compatibility with HuggingFace ecosystem using the environment variable `USE_HF`, switching to use models and datasets from HF. Please refer to the [HuggingFace ecosystem compatibility documentation](https://github.com/modelscope/swift/tree/main/docs/source_en/LLM/Compat-HF.md).
44
+
- 2024.04.17: Support the evaluation for OpenAI standard interfaces. Check the [parameter documentation](docs/source_en/LLM/Command-line-parameters.md#eval-parameters) for details.
42
45
- 🔥2024.04.17: Support **CodeQwen1.5-7B** series: CodeQwen1.5-7B, CodeQwen1.5-7B-Chat,CodeQwen1.5-7B-Chat-AWQ, use [this script](https://github.com/modelscope/swift/blob/main/examples/pytorch/llm/scripts/codeqwen1half_7b_chat/lora/sft.sh) to train.
43
46
- 2024.04.16: Supports inference and fine-tuning of llava-v1.6-34b model. For best practice, you can refer to [here](https://github.com/modelscope/swift/tree/main/docs/source_en/Multi-Modal/llava-best-practice.md).
44
47
- 2024.04.13: Support the fine-tuning and inference of Mixtral-8x22B-v0.1 model, use [this script](https://github.com/modelscope/swift/blob/main/examples/pytorch/llm/scripts/mixtral_moe_8x22b_v1/lora_ddp_ds/sft.sh) to start training!
@@ -137,11 +140,11 @@ SWIFT runs in the Python environment. Please ensure your Python version is highe
137
140
138
141
```shell
139
142
# Full capabilities
140
-
pip install ms-swift[all] -U
143
+
pip install 'ms-swift[all]' -U
141
144
# LLM only
142
-
pip install ms-swift[llm] -U
145
+
pip install 'ms-swift[llm]' -U
143
146
# AIGC only
144
-
pip install ms-swift[aigc] -U
147
+
pip install 'ms-swift[aigc]' -U
145
148
# Adapters only
146
149
pip install ms-swift -U
147
150
```
@@ -151,7 +154,7 @@ pip install ms-swift -U
151
154
```shell
152
155
git clone https://github.com/modelscope/swift.git
153
156
cd swift
154
-
pip install -e .[llm]
157
+
pip install -e '.[llm]'
155
158
```
156
159
157
160
SWIFT depends on torch>=1.13, recommend torch>=2.0.0.
@@ -389,7 +409,7 @@ CUDA_VISIBLE_DEVICES=0 swift deploy \
389
409
| XVerse |[XVerse series models](https://github.com/xverse-ai)| Chinese<br>English | 7B-65B | base model<br>chat model<br>long text model<br>MoE model |
390
410
| LLaMA2 |[LLaMA2 series models](https://github.com/facebookresearch/llama)| English | 7B-70B<br>including quantized versions | base model<br>chat model |
391
411
| Mistral<br>Mixtral |[Mistral series models](https://github.com/mistralai/mistral-src)| English | 7B-22B | base model<br>instruct model<br>MoE model |
392
-
| YI |[01AI's YI series models](https://github.com/01-ai)| Chinese<br>English | 6B-34B| base model<br>chat model<br>long text model |
412
+
| YI |[01AI's YI series models](https://github.com/01-ai)| Chinese<br>English | 6B-34B<br>including quantized| base model<br>chat model<br>long text model |
393
413
| InternLM<br>InternLM2<br>InternLM2-Math |[Pujiang AI Lab InternLM series models](https://github.com/InternLM/InternLM)| Chinese<br>English | 1.8B-20B | base model<br>chat model<br>math model |
394
414
| DeepSeek<br>DeepSeek-MoE<br>DeepSeek-Coder<br>DeepSeek-Math |[DeepSeek series models](https://github.com/deepseek-ai)| Chinese<br>English | 1.3B-67B | base model<br>chat model<br>MoE model<br>code model<br>math model |
395
415
| MAMBA |[MAMBA temporal convolution model](https://github.com/state-spaces/mamba)| English | 130M-2.8B | base model |
@@ -412,7 +432,7 @@ CUDA_VISIBLE_DEVICES=0 swift deploy \
412
432
| dbrx |[databricks](https://github.com/databricks/dbrx)| English | 132B | base model<br>chat model |
413
433
| mengzi3 |[Langboat](https://github.com/Langboat/Mengzi3)| Chinese<br>English | 13B | base model |
414
434
| c4ai-command-r |[c4ai](https://cohere.com/command)| Multilingual | 35B-104B | chat model |
415
-
435
+
| WizardLM2 |[WizardLM2 series models](https://github.com/nlpxucan/WizardLM)| English | 7B-8x22B<br>including quantized versions | chat model<br>MoE model |
0 commit comments