-
-
Notifications
You must be signed in to change notification settings - Fork 5.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Misc]: LoRA request with Multi GPU does not provide correct responses with num_scheduler_steps config #12487
Closed
1 task done
Labels
Comments
See: #11161 |
jeejeelee
added a commit
that referenced
this issue
Feb 1, 2025
Isotr0py
pushed a commit
to Isotr0py/vllm
that referenced
this issue
Feb 2, 2025
…llm-project#11161) FIX issue vllm-project#9688 vllm-project#11086 vllm-project#12487 --------- Signed-off-by: Jee Jee Li <pandaleefree@gmail.com> Co-authored-by: weilong.yu <weilong.yu@shopee.com> Co-authored-by: Jee Jee Li <pandaleefree@gmail.com> Signed-off-by: Isotr0py <2037008807@qq.com>
youngkent
pushed a commit
to youngkent/vllm
that referenced
this issue
Feb 3, 2025
…llm-project#11161) FIX issue vllm-project#9688 vllm-project#11086 vllm-project#12487 --------- Signed-off-by: Jee Jee Li <pandaleefree@gmail.com> Co-authored-by: weilong.yu <weilong.yu@shopee.com> Co-authored-by: Jee Jee Li <pandaleefree@gmail.com>
srikanthsrnvs
pushed a commit
to srikanthsrnvs/vllm
that referenced
this issue
Feb 3, 2025
…llm-project#11161) FIX issue vllm-project#9688 vllm-project#11086 vllm-project#12487 --------- Signed-off-by: Jee Jee Li <pandaleefree@gmail.com> Co-authored-by: weilong.yu <weilong.yu@shopee.com> Co-authored-by: Jee Jee Li <pandaleefree@gmail.com> Signed-off-by: Srikanth Srinivas <srikanth@astrum.ai>
sahelib25
pushed a commit
to krai/vllm
that referenced
this issue
Feb 3, 2025
…llm-project#11161) FIX issue vllm-project#9688 vllm-project#11086 vllm-project#12487 --------- Signed-off-by: Jee Jee Li <pandaleefree@gmail.com> Co-authored-by: weilong.yu <weilong.yu@shopee.com> Co-authored-by: Jee Jee Li <pandaleefree@gmail.com>
fxmarty-amd
pushed a commit
to fxmarty-amd/vllm
that referenced
this issue
Feb 7, 2025
…llm-project#11161) FIX issue vllm-project#9688 vllm-project#11086 vllm-project#12487 --------- Signed-off-by: Jee Jee Li <pandaleefree@gmail.com> Co-authored-by: weilong.yu <weilong.yu@shopee.com> Co-authored-by: Jee Jee Li <pandaleefree@gmail.com> Signed-off-by: Felix Marty <felmarty@amd.com>
NickLucche
pushed a commit
to NickLucche/vllm
that referenced
this issue
Feb 7, 2025
…llm-project#11161) FIX issue vllm-project#9688 vllm-project#11086 vllm-project#12487 --------- Signed-off-by: Jee Jee Li <pandaleefree@gmail.com> Co-authored-by: weilong.yu <weilong.yu@shopee.com> Co-authored-by: Jee Jee Li <pandaleefree@gmail.com>
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Anything you want to discuss about vllm.
Hello All,
We are encountering a strange issue with our LoRA adapter, when running in multi-GPU setup.
Context:
Base model: Mistral Nemo 12B (https://huggingface.co/nvidia/Mistral-NeMo-12B-Instruct)
Adapter Rank: 8
Vllm Model.json
Multi-lora.json
Now, when we add the num_scheduler_steps configuration to the model.json,
Now the adapter responds with correct response when we don't have 'num_scheduler_steps' in the multi-GPU setup, but when we add this configuration, we don't get the correct response from the adapter any longer, even though everything else remains same.
We are looking at the response from the LoRA targeted request here not the response from Base model.
Has anyone faced similar issue, are there any setting, configurations that need to be made to enable multi-gpu LoRA requests?
Thanks,
Rohit
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: