-
-
Notifications
You must be signed in to change notification settings - Fork 11k
[Bug] Dynamo Unsupported due to BasevLLMParameter.torch_function calling disabled super()
#25613
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Signed-off-by: yewentao256 <zhyanwentao@126.com>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Code Review
This pull request introduces a workaround for an issue with TorchDynamo in PyTorch versions older than 2.8.0. The change modifies BasevLLMParameter.__torch_function__ to return NotImplemented for these older versions, preventing a crash. The implementation is correct, but I have one suggestion to improve the logging behavior to avoid potential performance issues and log spam.
Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com> Signed-off-by: Wentao Ye <44945378+yewentao256@users.noreply.github.com>
|
@mgoin @yewentao256 This pr cause the deepseek model crash while loading: The output of tracebackCould it be further checked or tested? (vllm-metax plugin with torch 2.6 got crashed on e2e test, without the modification everything worked well) |
|
+1, hit the similar issue. |
|
@ILikeIneine Thanks for the feedback! Taking a look now |
|
Hey @ILikeIneine @wangxiyuan, is there a reason for your hardware backends to still be using torch 2.6 and 2.7? Generally we expect vllm to be compatible with only the latest version of torch and we don't have a need to support older torch versions before. If we must consider this, then we need to change our policy and CI to help capture issues in upstream. If I understand your comments correctly, you only have an issue with this "bugfix" PR right? We can revert it as it only tries to give a helpful log |
vllm-project/vllm#23991 vllm-project/vllm#25613 --------- Signed-off-by: Chendi Xue <Chendi.Xue@intel.com>
@mgoin Sorry for the delayed response. This is because our hardware backend sometimes can't catch up the latest release of torch. It takes some time to adapt. So the plugin torch version may not always the same as vllm's.(sometimes it's one or two version behind) |
|
yes, we use torch 2.7 now. |
vllm-project/vllm#23991 vllm-project/vllm#25613 --------- Signed-off-by: Chendi Xue <Chendi.Xue@intel.com> Signed-off-by: Iryna Boiko <iboiko@habana.ai>
…lling disabled super() (#25613) Signed-off-by: yewentao256 <zhyanwentao@126.com> Signed-off-by: Wentao Ye <44945378+yewentao256@users.noreply.github.com> Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com> Signed-off-by: yewentao256 <zhyanwentao@126.com>
…lling disabled super() (vllm-project#25613) Signed-off-by: yewentao256 <zhyanwentao@126.com> Signed-off-by: Wentao Ye <44945378+yewentao256@users.noreply.github.com> Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com> Signed-off-by: xuebwang-amd <xuebwang@amd.com>
…lling disabled super() (vllm-project#25613) Signed-off-by: yewentao256 <zhyanwentao@126.com> Signed-off-by: Wentao Ye <44945378+yewentao256@users.noreply.github.com> Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com>
…lling disabled super() (vllm-project#25613) Signed-off-by: yewentao256 <zhyanwentao@126.com> Signed-off-by: Wentao Ye <44945378+yewentao256@users.noreply.github.com> Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com>
…lling disabled super() (vllm-project#25613) Signed-off-by: yewentao256 <zhyanwentao@126.com> Signed-off-by: Wentao Ye <44945378+yewentao256@users.noreply.github.com> Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com> Signed-off-by: xuebwang-amd <xuebwang@amd.com>
Purpose
Added clear instructions for #25604 (comment)
Test
After upgrading to torch 2.8.0, the issue disappear
vllm bench throughput --model Qwen/Qwen3-30B-A3B-FP8 --load-format dummy --input-len 1000 --output-len 100 --trust_remote_code --enable-expert-parallel