-
-
Notifications
You must be signed in to change notification settings - Fork 5.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bugfix][Neuron] Fix soft prompt method error in NeuronExecutor #6313
Conversation
@WoosukKwon Thanks for the prompt response to my issue #6269! When the PR is approved, can you please follow up with a procedure to update my install to run the patched vLLM on Neuron systems? I urgently need this for a project I am working on. |
Thanks for the fix. The current neuron CI only tests online inference, the offline inference capability is currently not tested for neuron backend. |
Well, the online inference also appears to be broken: python -m vllm.entrypoints.openai.api_server |
@liangfu As @areanddee pointed out, the error happens when |
…-project#6313) (cherry picked from commit 997df46)
…-project#6313) Signed-off-by: Alvant <alvasian@yandex.ru>
Fixes #6269
However, I'm still not sure how #4645 passed the Neuron CI test.