Skip to content

Conversation

@MengqingCao
Copy link
Collaborator

@MengqingCao MengqingCao commented Jun 9, 2025

What this PR does / why we need it?

  1. run deepseek acc ut per pr --- multicard CI time increased by 9 min
  2. run spec decode e2e test on v1 per pr --- singlecard CI time increased by 3 min (partly is disabled due to not work now)
    3. align the output of whether dbo is enabled or not
    The generated results with and without dbo cannot be aligned.
    https://github.com/vllm-project/vllm-ascend/actions/runs/15822900528/job/44600029405?pr=1136
  3. skip V0 mtp test due to failure in https://github.com/vllm-project/vllm-ascend/actions/runs/16012172833/job/45171988816
  4. fix some version conflicts

How was this patch tested?

CI passed with new added test.

vllm_model.generate_greedy(example_prompts, max_tokens)


def test_models_distributed_DeepSeek():
Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think there is no need to run e2e functional ut as we already have acc ut

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

no, acc test belongs to long-term-test, it is not ran in every commit.

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This won't take a long time as dataset gsmk is small. wydt? cc @Yikun

@github-actions
Copy link

This pull request has conflicts, please resolve those before we can evaluate the pull request.

1 similar comment
@github-actions
Copy link

This pull request has conflicts, please resolve those before we can evaluate the pull request.

@MengqingCao MengqingCao changed the title [CI] Run deepseek acc ut per pr [CI/UT][Refactor] move e2e spec decode and deepseek acc test to per pr Jun 19, 2025
--ignore=tests/e2e/singlecard/long_term/spec_decode/e2e/test_v1_mtp_correctness.py
# ------------ spec decode e2e test on v1 ------------ #
VLLM_USE_MODELSCOPE=True pytest -sv tests/e2e/singlecard/long_term/spec_decode/e2e/test_v1_mtp_correctness.py
# TODO: revert me when test_v1_spec_decode.py::test_ngram_correctness is fixed
Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

test_v1_spec_decode.py::test_ngram_correctness is fixed in #1189. Will revert this when #1189 merged

@github-actions
Copy link

This pull request has conflicts, please resolve those before we can evaluate the pull request.

@codecov
Copy link

codecov bot commented Jun 23, 2025

Codecov Report

❌ Patch coverage is 66.66667% with 1 line in your changes missing coverage. Please review.
✅ Project coverage is 52.36%. Comparing base (c30ddb8) to head (211eabd).
⚠️ Report is 613 commits behind head on main.

Files with missing lines Patch % Lines
vllm_ascend/ops/fused_moe.py 50.00% 1 Missing ⚠️
Additional details and impacted files
@@             Coverage Diff             @@
##             main    #1136       +/-   ##
===========================================
+ Coverage   27.39%   52.36%   +24.96%     
===========================================
  Files          56       78       +22     
  Lines        6191     9631     +3440     
===========================================
+ Hits         1696     5043     +3347     
- Misses       4495     4588       +93     
Flag Coverage Δ
unittests 52.36% <66.66%> (+24.96%) ⬆️

Flags with carried forward coverage won't be shown. Click here to find out more.

☔ View full report in Codecov by Sentry.
📢 Have feedback on the report? Share it here.

🚀 New features to boost your workflow:
  • ❄️ Test Analytics: Detect flaky tests, report on failures, and find test suite problems.

@MengqingCao
Copy link
Collaborator Author

MengqingCao commented Jun 23, 2025

The generated results with and without dbo cannot be aligned.
https://github.com/vllm-project/vllm-ascend/actions/runs/15822900528/job/44600029405?pr=1136

@MengqingCao
Copy link
Collaborator Author

This pr is ready for review cc @Yikun @wangxiyuan

@MengqingCao MengqingCao force-pushed the dsci branch 2 times, most recently from eb46671 to 20bb139 Compare June 28, 2025 10:36
@github-actions
Copy link

This pull request has conflicts, please resolve those before we can evaluate the pull request.

@MengqingCao MengqingCao force-pushed the dsci branch 2 times, most recently from d659d4c to 410650c Compare July 2, 2025 04:56
wangxiyuan pushed a commit that referenced this pull request Jul 2, 2025
### What this PR does / why we need it?
mla attention still using the gpu_input_batch's attr:`swap_states`, which will lead to
an error `AttributeError: 'InputBatch' object has no attribute 'swap_states'`

This PR fixed the mla input patch error
### How was this patch tested?
will be tested by #1136

---------

Signed-off-by: wangli <wangli858794774@gmail.com>
@github-actions
Copy link

github-actions bot commented Jul 2, 2025

This pull request has conflicts, please resolve those before we can evaluate the pull request.

ZhengWG pushed a commit to ZhengWG/vllm-ascend that referenced this pull request Jul 3, 2025
### What this PR does / why we need it?
mla attention still using the gpu_input_batch's attr:`swap_states`, which will lead to
an error `AttributeError: 'InputBatch' object has no attribute 'swap_states'`

This PR fixed the mla input patch error
### How was this patch tested?
will be tested by vllm-project#1136

---------

Signed-off-by: wangli <wangli858794774@gmail.com>
Signed-off-by: ZhengWG <zwg0606@gmail.com>
ZhengWG pushed a commit to ZhengWG/vllm-ascend that referenced this pull request Jul 3, 2025
### What this PR does / why we need it?
mla attention still using the gpu_input_batch's attr:`swap_states`, which will lead to
an error `AttributeError: 'InputBatch' object has no attribute 'swap_states'`

This PR fixed the mla input patch error
### How was this patch tested?
will be tested by vllm-project#1136

---------

Signed-off-by: wangli <wangli858794774@gmail.com>
Signed-off-by: ZhengWG <zwg0606@gmail.com>
  * move e2e spec decode and deepseek acc test to per pr
  * move test_fused_moe_allgather_ep.py to e2e/multicard
  * remove e2e test on deepseek-v2-lite due to already test acc

Signed-off-by: MengqingCao <cmq0113@163.com>
Signed-off-by: MengqingCao <cmq0113@163.com>
Signed-off-by: MengqingCao <cmq0113@163.com>
Signed-off-by: MengqingCao <cmq0113@163.com>
@wangxiyuan wangxiyuan merged commit dd22ac3 into vllm-project:main Jul 4, 2025
20 checks passed
@MengqingCao MengqingCao deleted the dsci branch July 8, 2025 02:14
chopper0126 pushed a commit to chopper0126/vllm-ascend that referenced this pull request Oct 16, 2025
### What this PR does / why we need it?
mla attention still using the gpu_input_batch's attr:`swap_states`, which will lead to
an error `AttributeError: 'InputBatch' object has no attribute 'swap_states'`

This PR fixed the mla input patch error
### How was this patch tested?
will be tested by vllm-project#1136

---------

Signed-off-by: wangli <wangli858794774@gmail.com>
chopper0126 pushed a commit to chopper0126/vllm-ascend that referenced this pull request Oct 16, 2025
vllm-project#1136)

### What this PR does / why we need it?
1. run deepseek acc ut per pr --- multicard CI time increased by 9 min
2. run spec decode e2e test on v1 per pr --- singlecard CI time
increased by 3 min (partly is disabled due to not work now)
~~3. align the output of whether dbo is enabled or not~~
    The generated results with and without dbo cannot be aligned.

https://github.com/vllm-project/vllm-ascend/actions/runs/15822900528/job/44600029405?pr=1136
4. skip V0 mtp test due to failure in
https://github.com/vllm-project/vllm-ascend/actions/runs/16012172833/job/45171988816
5. fix some version conflicts
### How was this patch tested?
CI passed with new added test.

---------

Signed-off-by: MengqingCao <cmq0113@163.com>
Angazenn pushed a commit to Angazenn/vllm-ascend that referenced this pull request Oct 21, 2025
### What this PR does / why we need it?
mla attention still using the gpu_input_batch's attr:`swap_states`, which will lead to
an error `AttributeError: 'InputBatch' object has no attribute 'swap_states'`

This PR fixed the mla input patch error
### How was this patch tested?
will be tested by vllm-project#1136

---------

Signed-off-by: wangli <wangli858794774@gmail.com>
Angazenn pushed a commit to Angazenn/vllm-ascend that referenced this pull request Oct 21, 2025
vllm-project#1136)

### What this PR does / why we need it?
1. run deepseek acc ut per pr --- multicard CI time increased by 9 min
2. run spec decode e2e test on v1 per pr --- singlecard CI time
increased by 3 min (partly is disabled due to not work now)
~~3. align the output of whether dbo is enabled or not~~
    The generated results with and without dbo cannot be aligned.

https://github.com/vllm-project/vllm-ascend/actions/runs/15822900528/job/44600029405?pr=1136
4. skip V0 mtp test due to failure in
https://github.com/vllm-project/vllm-ascend/actions/runs/16012172833/job/45171988816
5. fix some version conflicts
### How was this patch tested?
CI passed with new added test.

---------

Signed-off-by: MengqingCao <cmq0113@163.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants