You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
[main][Bugfix] Fix unable to load qwen3_moe quantized weights (#2219)
### What this PR does / why we need it?
Fixes unable to load `qwen3_moe` quantized weights issue due to #1994
### Does this PR introduce _any_ user-facing change?
None
### How was this patch tested?
Add a `qwen3_moe` W8A8 quantized model in
`tests/e2e/multicard/test_qwen3_moe.py`
- vLLM version: v0.10.0
- vLLM main:
vllm-project/vllm@c494f96
---------
Signed-off-by: zhoux77899 <[email protected]>
Copy file name to clipboardExpand all lines: vllm_ascend/models/qwen3_moe.py
+33-4Lines changed: 33 additions & 4 deletions
Original file line number
Diff line number
Diff line change
@@ -1,6 +1,7 @@
1
1
# Copyright (c) 2025 Huawei Technologies Co., Ltd. All Rights Reserved.
2
+
# Copyright 2024 The Qwen team.
2
3
# Copyright 2023 The vLLM team.
3
-
#
4
+
# Copyright 2022 EleutherAI and the HuggingFace Inc. team. All rights reserved. Copyright 2022 EleutherAI and the HuggingFace Inc. team. All rights reserved.
4
5
#
5
6
# Licensed under the Apache License, Version 2.0 (the "License");
6
7
# you may not use this file except in compliance with the License.
0 commit comments