Skip to content

Conversation

@ZJY0516
Copy link
Contributor

@ZJY0516 ZJY0516 commented Oct 9, 2025

FIX #26454

Reverts #26113.
Fix compile tests in distributed tests CI

@mergify mergify bot added the llama Related to Llama models label Oct 9, 2025
@ZJY0516 ZJY0516 changed the title Revert "[Frontend] CompilationConfig overhaul (#20283): deprecate use_inductor in favor of backend, simplify custom_ops" Revert #26113 "[Frontend] CompilationConfig overhaul (#20283): deprecate use_inductor in favor of backend, simplify custom_ops" Oct 9, 2025
Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request reverts a previous change that overhauled CompilationConfig, re-instating use_inductor as the primary flag over backend. The changes across test files and configuration files are consistent with this revert. However, I found a critical issue in vllm/platforms/cpu.py where a new logic block incorrectly disables custom ops for the 'eager' backend on CPU. My review includes a fix for this issue.

Comment on lines +277 to 279
if compilation_config.use_inductor:
compilation_config.custom_ops = ["none"]

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

critical

The logic to disable custom ops is based on compilation_config.use_inductor, which defaults to True and is not updated based on the backend variable set earlier. This causes custom ops to be incorrectly disabled even when the backend is 'eager', which should support them.

To fix this, use_inductor should be explicitly set based on the chosen backend. This ensures that custom ops are only disabled when inductor is the backend, and use_inductor is consistent for any other logic that might depend on it.

Suggested change
if compilation_config.use_inductor:
compilation_config.custom_ops = ["none"]
compilation_config.use_inductor = (backend == "inductor")
if compilation_config.use_inductor:
compilation_config.custom_ops = ["none"]

…eprecate use_inducto…"

This reverts commit 0c824fc.

Signed-off-by: zjy0516 <riverclouds.zhu@qq.com>
@ZJY0516 ZJY0516 force-pushed the revert-26113-issue-20283-compilation-config branch from 0566bc3 to 8f72d11 Compare October 9, 2025 07:49
@DarkLight1337 DarkLight1337 added the ready ONLY add when PR is ready to merge/full CI is needed label Oct 9, 2025
@ZJY0516
Copy link
Contributor Author

ZJY0516 commented Oct 9, 2025

@DarkLight1337 Only Entrypoints Integration Test (API Server) failed. I think it's not related to this pr. Coudl we merge it?

@vllm-bot vllm-bot merged commit 5728da1 into vllm-project:main Oct 9, 2025
46 of 48 checks passed
@ProExpertProg
Copy link
Collaborator

@DarkLight1337 in the future can we please go through review before reverting a large PR like this? We should forward fix instead of reverting

morrison-turnansky added a commit to morrison-turnansky/vllm that referenced this pull request Oct 9, 2025
…aul (vllm-project#20283): deprecate use_inductor in favor of backend, simplify custom_ops" (vllm-project#26472)"

This reverts commit 5728da1.
morrison-turnansky added a commit to morrison-turnansky/vllm that referenced this pull request Oct 9, 2025
…aul (vllm-project#20283): deprecate use_inductor in favor of backend, simplify custom_ops" (vllm-project#26472)"

This reverts commit 5728da1.
@DarkLight1337
Copy link
Member

DarkLight1337 commented Oct 9, 2025

The usual practice for CI failure is to revert the PR (given that this test has been failing for a while, I thought a fix couldn't be done quickly). I wasn't aware that this PR is so important.

yang926 pushed a commit to yang926/vllm_1008 that referenced this pull request Oct 9, 2025
…m-project#20283): deprecate use_inductor in favor of backend, simplify custom_ops" (vllm-project#26472)

Signed-off-by: zjy0516 <riverclouds.zhu@qq.com>
Signed-off-by: yang926 <yang926@naver.com>
xuebwang-amd pushed a commit to xuebwang-amd/vllm that referenced this pull request Oct 10, 2025
…m-project#20283): deprecate use_inductor in favor of backend, simplify custom_ops" (vllm-project#26472)

Signed-off-by: zjy0516 <riverclouds.zhu@qq.com>
Signed-off-by: xuebwang-amd <xuebwang@amd.com>
Dhruvilbhatt pushed a commit to Dhruvilbhatt/vllm that referenced this pull request Oct 14, 2025
…m-project#20283): deprecate use_inductor in favor of backend, simplify custom_ops" (vllm-project#26472)

Signed-off-by: zjy0516 <riverclouds.zhu@qq.com>
Signed-off-by: Dhruvil Bhatt <bhattdbh@amazon.com>
lywa1998 pushed a commit to lywa1998/vllm that referenced this pull request Oct 20, 2025
…m-project#20283): deprecate use_inductor in favor of backend, simplify custom_ops" (vllm-project#26472)

Signed-off-by: zjy0516 <riverclouds.zhu@qq.com>
alhridoy pushed a commit to alhridoy/vllm that referenced this pull request Oct 24, 2025
…m-project#20283): deprecate use_inductor in favor of backend, simplify custom_ops" (vllm-project#26472)

Signed-off-by: zjy0516 <riverclouds.zhu@qq.com>
xuebwang-amd pushed a commit to xuebwang-amd/vllm that referenced this pull request Oct 24, 2025
…m-project#20283): deprecate use_inductor in favor of backend, simplify custom_ops" (vllm-project#26472)

Signed-off-by: zjy0516 <riverclouds.zhu@qq.com>
Signed-off-by: xuebwang-amd <xuebwang@amd.com>
0xrushi pushed a commit to 0xrushi/vllm that referenced this pull request Oct 26, 2025
…m-project#20283): deprecate use_inductor in favor of backend, simplify custom_ops" (vllm-project#26472)

Signed-off-by: zjy0516 <riverclouds.zhu@qq.com>
Signed-off-by: 0xrushi <6279035+0xrushi@users.noreply.github.com>
0xrushi pushed a commit to 0xrushi/vllm that referenced this pull request Oct 26, 2025
…m-project#20283): deprecate use_inductor in favor of backend, simplify custom_ops" (vllm-project#26472)

Signed-off-by: zjy0516 <riverclouds.zhu@qq.com>
Signed-off-by: 0xrushi <6279035+0xrushi@users.noreply.github.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

llama Related to Llama models ready ONLY add when PR is ready to merge/full CI is needed

Projects

None yet

4 participants