Skip to content

Conversation

@yiz-liu
Copy link
Collaborator

@yiz-liu yiz-liu commented Apr 23, 2025

What this PR does / why we need it?

Enforce eager mode in the V1 engine ahead of the upcoming CANN and torch_npu releases.

Does this PR introduce any user-facing change?

After this change, users will no longer need to manually set enforce_eager=True.

How was this patch tested?

Test it with regular offline inference examples.

Signed-off-by: Yizhou Liu <liu_yizhou@outlook.com>
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Looks good to me, but the changes in this file seems not related to the topic of this PR

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actually, I spoke with Xiyuan and confirmed that this test case does not require setting the vllm config. As a result, I reverted it to its original state. The reason for this change is that the CI triggered a vllm config-related error while I was addressing the enforce_eager issue, which led me to this test case.

@ganyi1996ppo
Copy link
Collaborator

@wangxiyuan can you please review this PR, we should merge this PR asap.

@MengqingCao
Copy link
Collaborator

lgtm

@wangxiyuan wangxiyuan merged commit d785e78 into vllm-project:main Apr 24, 2025
15 checks passed
@wangxiyuan wangxiyuan changed the title Quick Fix: Enforce eager mode in the V1 engine ahead of the upcoming CANN and torch_npu releases [V1] Make V1 engine backward compatible Apr 24, 2025
ttanzhiqiang pushed a commit to ttanzhiqiang/vllm-ascend that referenced this pull request Apr 27, 2025
### What this PR does / why we need it?
Enforce eager mode in the V1 engine ahead of the upcoming CANN and
torch_npu releases.

### Does this PR introduce _any_ user-facing change?
After this change, users will no longer need to manually set
enforce_eager=True.

### How was this patch tested?
Test it with regular offline inference examples.

Signed-off-by: Yizhou Liu <liu_yizhou@outlook.com>
Angazenn pushed a commit to Angazenn/vllm-ascend that referenced this pull request Oct 21, 2025
### What this PR does / why we need it?
Enforce eager mode in the V1 engine ahead of the upcoming CANN and
torch_npu releases.

### Does this PR introduce _any_ user-facing change?
After this change, users will no longer need to manually set
enforce_eager=True.

### How was this patch tested?
Test it with regular offline inference examples.

Signed-off-by: Yizhou Liu <liu_yizhou@outlook.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants