Skip to content

Conversation

@rjg-lyh
Copy link
Collaborator

@rjg-lyh rjg-lyh commented May 14, 2025

What this PR does / why we need it?

Fix the bugs when run deepseek model in engine v1.

Does this PR introduce any user-facing change?

No.

How was this patch tested?

CI passed with new added/existing test.

@rjg-lyh rjg-lyh changed the title [BugFix] Fix chunked prefill bugs [BugFix] Fix chunked prefill bugs in engine v1 May 14, 2025
@rjg-lyh rjg-lyh force-pushed the pr-bugfix-dsv branch 2 times, most recently from 1e2aeea to 6725d90 Compare May 14, 2025 08:02
@rjg-lyh rjg-lyh force-pushed the pr-bugfix-dsv branch 10 times, most recently from cb1df18 to 70dc428 Compare May 14, 2025 11:49
@Yikun
Copy link
Collaborator

Yikun commented May 14, 2025

diff --git a/vllm_ascend/patch/platform/patch_common/patch_vllm_config.py b/vllm_ascend/patch/platform/patch_common/patch_vllm_config.py
index 6d606d0..947ec7d 100644
--- a/vllm_ascend/patch/platform/patch_common/patch_vllm_config.py
+++ b/vllm_ascend/patch/platform/patch_common/patch_vllm_config.py
@@ -18,11 +18,10 @@
 # This file is a part of the vllm-ascend project.

 import torch
-
+import vllm.envs as envs
+from vllm.config import CompilationConfig, CompilationLevel, VllmConfig
 from vllm.logger import init_logger
-from vllm.config import (VllmConfig, CompilationConfig, CompilationLevel)
 from vllm.utils import random_uuid
-import vllm.envs as envs

 logger = init_logger(__name__)

Signed-off-by: rjg-lyh <1318825571@qq.com>
self.instance_id = random_uuid()[:5]


VllmConfig.__post_init__ = __post_init__
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

patch vllm config is dengerous. it's always changed by vllm. How we make sure the capability?

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

related PR to fix the issue: vllm-project/vllm#18470

Signed-off-by: rjg-lyh <1318825571@qq.com>
@wangxiyuan wangxiyuan added the ready read for review label May 22, 2025
@wangxiyuan wangxiyuan merged commit b4d6672 into vllm-project:main May 22, 2025
16 checks passed
wangxiyuan pushed a commit that referenced this pull request May 30, 2025
### What this PR does / why we need it?
add basic v1 mtp features
please merge it after
#874 and
#844.

### Does this PR introduce _any_ user-facing change?
now, we supported basic v1 mtp, only supported tp only、eager mode and
k=1
we will continue to expand more scenarios.

### How was this patch tested?
local tested

Signed-off-by: XWFAlone <xuewenfei2@huawei.com>
Co-authored-by: mengwei805 <mengwei25@huawei.com>
Co-authored-by: JC-ut0 <xuyexiong@huawei.com>
momo609 pushed a commit to momo609/vllm-ascend that referenced this pull request Jun 3, 2025
### What this PR does / why we need it?
add basic v1 mtp features
please merge it after
vllm-project#874 and
vllm-project#844.

### Does this PR introduce _any_ user-facing change?
now, we supported basic v1 mtp, only supported tp only、eager mode and
k=1
we will continue to expand more scenarios.

### How was this patch tested?
local tested

Signed-off-by: XWFAlone <xuewenfei2@huawei.com>
Co-authored-by: mengwei805 <mengwei25@huawei.com>
Co-authored-by: JC-ut0 <xuyexiong@huawei.com>
Signed-off-by: wangxiaoxin (A) <w00664509@china.huawei.com>
David9857 pushed a commit to David9857/vllm-ascend that referenced this pull request Jun 3, 2025
### What this PR does / why we need it?
add basic v1 mtp features
please merge it after
vllm-project#874 and
vllm-project#844.

### Does this PR introduce _any_ user-facing change?
now, we supported basic v1 mtp, only supported tp only、eager mode and
k=1
we will continue to expand more scenarios.

### How was this patch tested?
local tested

Signed-off-by: XWFAlone <xuewenfei2@huawei.com>
Co-authored-by: mengwei805 <mengwei25@huawei.com>
Co-authored-by: JC-ut0 <xuyexiong@huawei.com>
@rjg-lyh rjg-lyh deleted the pr-bugfix-dsv branch July 22, 2025 12:13
chopper0126 pushed a commit to chopper0126/vllm-ascend that referenced this pull request Oct 16, 2025
### What this PR does / why we need it?
Fix the bugs when run deepseek model in engine v1.

### Does this PR introduce _any_ user-facing change?
No.

### How was this patch tested?
CI passed with new added/existing test.

---------

Signed-off-by: rjg-lyh <1318825571@qq.com>
chopper0126 pushed a commit to chopper0126/vllm-ascend that referenced this pull request Oct 16, 2025
### What this PR does / why we need it?
add basic v1 mtp features
please merge it after
vllm-project#874 and
vllm-project#844.

### Does this PR introduce _any_ user-facing change?
now, we supported basic v1 mtp, only supported tp only、eager mode and
k=1
we will continue to expand more scenarios.

### How was this patch tested?
local tested

Signed-off-by: XWFAlone <xuewenfei2@huawei.com>
Co-authored-by: mengwei805 <mengwei25@huawei.com>
Co-authored-by: JC-ut0 <xuyexiong@huawei.com>
Angazenn pushed a commit to Angazenn/vllm-ascend that referenced this pull request Oct 21, 2025
### What this PR does / why we need it?
Fix the bugs when run deepseek model in engine v1.

### Does this PR introduce _any_ user-facing change?
No.

### How was this patch tested?
CI passed with new added/existing test.

---------

Signed-off-by: rjg-lyh <1318825571@qq.com>
Angazenn pushed a commit to Angazenn/vllm-ascend that referenced this pull request Oct 21, 2025
### What this PR does / why we need it?
add basic v1 mtp features
please merge it after
vllm-project#874 and
vllm-project#844.

### Does this PR introduce _any_ user-facing change?
now, we supported basic v1 mtp, only supported tp only、eager mode and
k=1
we will continue to expand more scenarios.

### How was this patch tested?
local tested

Signed-off-by: XWFAlone <xuewenfei2@huawei.com>
Co-authored-by: mengwei805 <mengwei25@huawei.com>
Co-authored-by: JC-ut0 <xuyexiong@huawei.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

module:core ready read for review

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants