-
Notifications
You must be signed in to change notification settings - Fork 542
[main][Bugfix] Fix unable to load qwen3_moe quantized weights #2219
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Signed-off-by: zhoux77899 <zhouxiang100@huawei.com>
Signed-off-by: zhoux77899 <zhouxiang100@huawei.com>
Signed-off-by: zhoux77899 <zhouxiang100@huawei.com>
Signed-off-by: zhoux77899 <zhouxiang100@huawei.com>
Signed-off-by: zhoux77899 <zhouxiang100@huawei.com>
|
👋 Hi! Thank you for contributing to the vLLM Ascend project. The following points will speed up your PR merge:
If CI fails, you can run linting and testing checks locally according Contributing and Testing. |
| max_model_len=8192, | ||
| dtype=dtype, | ||
| tensor_parallel_size=2, | ||
| quantization="ascend", |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
see: #2223 let's enable aclgraph for test
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Done
Signed-off-by: zhoux77899 <zhouxiang100@huawei.com>
Codecov Report❌ Patch coverage is
Additional details and impacted files@@ Coverage Diff @@
## main #2219 +/- ##
==========================================
- Coverage 76.75% 76.65% -0.11%
==========================================
Files 113 113
Lines 12743 12763 +20
==========================================
+ Hits 9781 9783 +2
- Misses 2962 2980 +18
Flags with carried forward coverage won't be shown. Click here to find out more. ☔ View full report in Codecov by Sentry. 🚀 New features to boost your workflow:
|
Signed-off-by: zhoux77899 <zhouxiang100@huawei.com>
…roject#2219) ### What this PR does / why we need it? Fixes unable to load `qwen3_moe` quantized weights issue due to vllm-project#1994 ### Does this PR introduce _any_ user-facing change? None ### How was this patch tested? Add a `qwen3_moe` W8A8 quantized model in `tests/e2e/multicard/test_qwen3_moe.py` - vLLM version: v0.10.0 - vLLM main: vllm-project/vllm@c494f96 --------- Signed-off-by: zhoux77899 <zhouxiang100@huawei.com>
…roject#2219) ### What this PR does / why we need it? Fixes unable to load `qwen3_moe` quantized weights issue due to vllm-project#1994 ### Does this PR introduce _any_ user-facing change? None ### How was this patch tested? Add a `qwen3_moe` W8A8 quantized model in `tests/e2e/multicard/test_qwen3_moe.py` - vLLM version: v0.10.0 - vLLM main: vllm-project/vllm@c494f96 --------- Signed-off-by: zhoux77899 <zhouxiang100@huawei.com>
…roject#2219) ### What this PR does / why we need it? Fixes unable to load `qwen3_moe` quantized weights issue due to vllm-project#1994 ### Does this PR introduce _any_ user-facing change? None ### How was this patch tested? Add a `qwen3_moe` W8A8 quantized model in `tests/e2e/multicard/test_qwen3_moe.py` - vLLM version: v0.10.0 - vLLM main: vllm-project/vllm@c494f96 --------- Signed-off-by: zhoux77899 <zhouxiang100@huawei.com>
…roject#2219) ### What this PR does / why we need it? Fixes unable to load `qwen3_moe` quantized weights issue due to vllm-project#1994 ### Does this PR introduce _any_ user-facing change? None ### How was this patch tested? Add a `qwen3_moe` W8A8 quantized model in `tests/e2e/multicard/test_qwen3_moe.py` - vLLM version: v0.10.0 - vLLM main: vllm-project/vllm@c494f96 --------- Signed-off-by: zhoux77899 <zhouxiang100@huawei.com>
What this PR does / why we need it?
Fixes unable to load
qwen3_moequantized weights issue due to #1994Does this PR introduce any user-facing change?
None
How was this patch tested?
Add a
qwen3_moeW8A8 quantized model intests/e2e/multicard/test_qwen3_moe.py