Skip to content

Commit 042605f

Browse files
authored
[Doc] Add stable modelslim branch (#2545)
### What this PR does / why we need it? The branch `br_release_MindStudio_8.1.RC2_TR5_20260624` is commercial delivery version of modelslim in Q3, and has been verified available ### Does this PR introduce _any_ user-facing change? ### How was this patch tested? - vLLM version: v0.10.1.1 - vLLM main: vllm-project/vllm@7d67a9d Signed-off-by: wangli <[email protected]>
1 parent 8151a9d commit 042605f

File tree

2 files changed

+8
-10
lines changed

2 files changed

+8
-10
lines changed

docs/source/tutorials/single_npu_qwen3_quantization.md

Lines changed: 3 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -32,11 +32,10 @@ see https://www.modelscope.cn/models/vllm-ascend/Qwen3-8B-W4A8
3232
:::
3333

3434
```bash
35-
git clone https://gitee.com/ascend/msit
36-
cd msit/msmodelslim
35+
# The branch(br_release_MindStudio_8.1.RC2_TR5_20260624) has been verified
36+
git clone -b br_release_MindStudio_8.1.RC2_TR5_20260624 https://gitee.com/ascend/msit
3737

38-
# Optional, this commit has been verified
39-
git checkout f8ab35a772a6c1ee7675368a2aa4bafba3bedd1a
38+
cd msit/msmodelslim
4039

4140
# Install by run this script
4241
bash install.sh

docs/source/user_guide/feature_guide/quantization.md

Lines changed: 5 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -11,11 +11,10 @@ To quantize a model, users should install [ModelSlim](https://gitee.com/ascend/m
1111
Install modelslim:
1212

1313
```bash
14-
git clone https://gitee.com/ascend/msit
15-
cd msit/msmodelslim
14+
# The branch(br_release_MindStudio_8.1.RC2_TR5_20260624) has been verified
15+
git clone -b br_release_MindStudio_8.1.RC2_TR5_20260624 https://gitee.com/ascend/msit
1616

17-
# Optional, this commit has been verified
18-
git checkout f8ab35a772a6c1ee7675368a2aa4bafba3bedd1a
17+
cd msit/msmodelslim
1918

2019
bash install.sh
2120
pip install accelerate
@@ -100,12 +99,12 @@ Enable quantization by specifying `--quantization ascend`, for more details, see
10099
101100
### 1. How to solve the KeyError: 'xxx.layers.0.self_attn.q_proj.weight' problem?
102101
103-
First, make sure you specify `ascend` quantization method. Second, check if your model is converted by this `modelslim-VLLM-8.1.RC1.b020_001` modelslim version. Finally, if it still doesn't work, please
102+
First, make sure you specify `ascend` quantization method. Second, check if your model is converted by this `br_release_MindStudio_8.1.RC2_TR5_20260624` modelslim version. Finally, if it still doesn't work, please
104103
submit a issue, maybe some new models need to be adapted.
105104
106105
### 2. How to solve the error "Could not locate the configuration_deepseek.py"?
107106
108-
Please convert DeepSeek series models using `modelslim-VLLM-8.1.RC1.b020_001` modelslim, this version has fixed the missing configuration_deepseek.py error.
107+
Please convert DeepSeek series models using `br_release_MindStudio_8.1.RC2_TR5_20260624` modelslim, this version has fixed the missing configuration_deepseek.py error.
109108
110109
### 3. When converting deepseek series models with modelslim, what should you pay attention?
111110

0 commit comments

Comments
 (0)