-
Notifications
You must be signed in to change notification settings - Fork 6.9k
Fix pickle error with remote code models in vLLM Ray worker process #53815
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fix pickle error with remote code models in vLLM Ray worker process #53815
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Copilot encountered an error and was unable to review this pull request. You can try again by re-requesting a review.
kouroshHakha
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
just one nit, let's also rerun the release tests on this PR before merging.
Signed-off-by: Seiji Eicher <seiji@anyscale.com>
Signed-off-by: Seiji Eicher <seiji@anyscale.com>
Signed-off-by: Seiji Eicher <seiji@anyscale.com>
Signed-off-by: Seiji Eicher <seiji@anyscale.com>
Signed-off-by: Seiji Eicher <seiji@anyscale.com>
fbc1dc5 to
b0c508e
Compare
Signed-off-by: Seiji Eicher <seiji@anyscale.com>
b0c508e to
c7126cf
Compare
…er process (#53815) Signed-off-by: Seiji Eicher <seiji@anyscale.com> Signed-off-by: elliot-barn <elliot.barnwell@anyscale.com>
…er process (ray-project#53815) Signed-off-by: Seiji Eicher <seiji@anyscale.com>
…er process (#53815) Signed-off-by: Seiji Eicher <seiji@anyscale.com> Signed-off-by: elliot-barn <elliot.barnwell@anyscale.com>
Why are these changes needed?
Since #53621, vLLM engines running DeepSeek-V2-Lite and related models that fetch remote code fail with pickle errors, since the engine only registers custom configs to serialize by value (and avoid the pickle error) if
transformers_modulescan be imported.We relied on a call to
AutoProcessor.from_pretrainedto initializetransformers_modules, so thatmaybe_register_config_serialize_by_valuewould execute correctly when AsyncLLM starts. This was removed in #53621. Now we can useinit_hf_modulesso to accomplish the same result more directly.We could also fix this upstream with vllm-project/vllm#19510.
Traceback:
Related issue number
Checks
git commit -s) in this PR.scripts/format.shto lint the changes in this PR.method in Tune, I've added it in
doc/source/tune/api/under thecorresponding
.rstfile.