-
-
Notifications
You must be signed in to change notification settings - Fork 11k
[Bugfix] Support triton==3.3.0+git95326d9f for RTX 5090 (Unsloth + vLLM compatibility) #15471
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bugfix] Support triton==3.3.0+git95326d9f for RTX 5090 (Unsloth + vLLM compatibility) #15471
Conversation
|
👋 Hi! Thank you for contributing to the vLLM project. 💬 Join our developer Slack at https://slack.vllm.ai to discuss your PR in #pr-reviews, coordinate on features in #feat- channels, or join special interest groups in #sig- channels. Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging. To run CI, PR reviewers can either: Add 🚀 |
|
I have unblocked all the LoRA-related tests to verify if all LoRA-CI can pass correctly. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Seems reasonable to me if existing tests are green since b_ptr is without the comma
Yeah, I think this change doesn't effect on current lora ops. |
…LM compatibility) (vllm-project#15471) Co-authored-by: ServerAI <ai@exc-mad-ai.com> Signed-off-by: Wes Medford <wryanmedford@gmail.com>
|
I am still unable to install unsloth and vllm because using pip to install vllm is untrustworthy; it uninstalls
minimal code to create error: |
I have the same issue |
…LM compatibility) (vllm-project#15471) Co-authored-by: ServerAI <ai@exc-mad-ai.com> Signed-off-by: Louis Ulmer <ulmerlouis@gmail.com>
…LM compatibility) (vllm-project#15471) Co-authored-by: ServerAI <ai@exc-mad-ai.com>
…LM compatibility) (vllm-project#15471) Co-authored-by: ServerAI <ai@exc-mad-ai.com>
…LM compatibility) (vllm-project#15471) Co-authored-by: ServerAI <ai@exc-mad-ai.com> Signed-off-by: Mu Huai <tianbowen.tbw@antgroup.com>
🔧 Fix: Support
triton==3.3.0+git95326d9ffor RTX 5090 (Unsloth + vLLM compatibility)📝 Description
This PR fixes an incompatibility in the Triton kernel logic used in
vllm/lora/ops/triton_ops/kernel_utils.pythat causes vLLM to break when running with:2025.3.182025.3.163.3.0+git95326d9f2.8.0.dev20250324+cu1280.22.0.dev20250324+cu1282.6.0.dev20250324+cu1280.0.30+4fa0149.d20250325All tested using Unsloth and Qwen2.5-3B-Instruct, with an NVIDIA RTX 5090.
🐛 Problem
When running vLLM with Unsloth, the following error occurred:
This was traced to incorrect pointer construction using:
Triton now treats this as a
tuple_type, which breaks downstream logic that expects a pointer object.✅ Fix
We removed unnecessary trailing commas in pointer assignments so that actual
pointer_typeobjects are passed instead of tuples. For example:🚀 Result
This fix makes vLLM fully compatible with Triton 3.3.0+, allowing Unsloth-based LoRA fine-tuning and inference to work out-of-the-box with newer GPUs like the RTX 5090.
🔒 Notes
VLLM_INSTALL_PUNICA_KERNELS=1is set during installation.pip install -e . --no-build-isolation.