-
Couldn't load subscription status.
- Fork 26
Fix local development for vllm==0.8.5 #140
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Signed-off-by: Wallas Santos <wallashss@ibm.com>
|
👋 Hi! Thank you for contributing to vLLM support on Spyre. Or this can be done with Now you are good to go 🚀 |
Signed-off-by: Wallas Santos <wallashss@ibm.com>
…to wallas-fix-local Signed-off-by: Wallas Santos <wallashss@ibm.com>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
yep I can confirm this works
…triton Signed-off-by: Wallas Santos <wallashss@ibm.com>
…to wallas-fix-local Signed-off-by: Wallas Santos <wallashss@ibm.com>
Signed-off-by: Wallas Santos <wallashss@ibm.com>
|
I'm also curious to know what changed in 0.8.5 such that this Triton placeholder started behaving incorrectly? I'm assuming this placeholder existed in earlier versions too right? Edit - I think I found the answer - https://github.com/vllm-project/vllm/pull/15099/files So the |
Yeah, this is the issue. I think a back contribution to the upstream would be a full fledge placeholder that every imports work, for instance to be able to import a dummy config; or a environment variable that prevents that, due the side effect it can cause on pytorch. Therefore, for cases like ours, and others it would be better to raise an exception than get an weird behavior in other dependencies like it does to the pytorch itself. |
The problem
The problem started happening when I pulled new code for vllm-spyre and it has updates for vllm. I got this stacktrace:
For newer vllm version (>=0.8.5), vLLM pokes the loaded module and it creates a placeholder for triton. This workaround can make pytorch mess up some code related to torch.compile.
My solution is just delete the triton module, and everything's can work ok again using Dyname Backend as
eagerin case we are developing on mac os.Prashant:
PR for triton placeholder for reference: vllm-project/vllm#15099