-
-
Notifications
You must be signed in to change notification settings - Fork 11.1k
[Misc] Reduce LoRA-related static variable #13166
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Misc] Reduce LoRA-related static variable #13166
Conversation
|
👋 Hi! Thank you for contributing to the vLLM project. 💬 Join our developer Slack at https://slack.vllm.ai to discuss your PR in #pr-reviews, coordinate on features in #feat- channels, or join special interest groups in #sig- channels. Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging. To run CI, PR reviewers can either: Add 🚀 |
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
7e5ad01 to
e12c3db
Compare
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
44665c6 to
fbb7558
Compare
|
This pull request has merge conflicts that must be resolved before it can be |
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
5190aee to
3ce2624
Compare
| """ | ||
| return lora_linear_cls[vllm_linear_cls][fully_sharded] | ||
|
|
||
| return HFCompatibleLinear( |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
After completing this PR, transformers-model will directly support LoRA, but there are some issues with inference that are expected to be resolved in subsequent PRs.
Isotr0py
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Tests passed on my side locally, LGTM!
|
@DarkLight1337 Do you know what's causing the current CI failures? |
|
There were some issues in HF's opt repo yesterday, which should have been fixed. I think re-run these CIs should be just fine? |
Signed-off-by: Louis Ulmer <ulmerlouis@gmail.com>
### What this PR does / why we need it? Because the EOL vLLM-v0.7.3 lacks this PR(vllm-project/vllm#13166), while launching Qwen3+LoRA on vllm-ascend0.7.3, the error **"Qwen3ForCausalLM" object has no attribute "embedding modules**" will be raised. We modify qwen3.py to support Qwen3+LoRA on vllm-ascend v0.7.3 instead. ### Does this PR introduce _any_ user-facing change? No. ### How was this patch tested? --------- Signed-off-by: paulyu <paulyu0307@gmail.com> Co-authored-by: paulyu <paulyu0307@gmail.com>
Motivation
Remove LoRA-related static variable
supported_lora_modules, which not only makes our model implementation cleaner but also enables smoother LoRA supportWork
supported_lora_modules