-
-
Notifications
You must be signed in to change notification settings - Fork 11.1k
[CI/Build] Update the Dockerfile to include Blackwell archs #18092
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
👋 Hi! Thank you for contributing to the vLLM project. 💬 Join our developer Slack at https://slack.vllm.ai to discuss your PR in #pr-reviews, coordinate on features in #feat- channels, or join special interest groups in #sig- channels. Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging. To run CI, PR reviewers can either: Add 🚀 |
simon-mo
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM! Let's see if it builds, and i'll poll contributor on the ubuntu 24 change
1bec9dc to
ecbded0
Compare
Seems like Ubuntu24 upgrade might break people's workflow
|
Is it possible to do this without updating base image? |
tlrmchlsmth
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We need to stick to building with older OSes, unfortunately. The reason for this is that the glibc version is forwards-compatible but not backwards-compatible. If we upgrade to Ubuntu 24.04, then vLLM won't work on 22.04, for instance.
Can the CUDA arch list be extended without upgrading Ubuntu?
|
I think so. Cuda 12.8 still supports 20.04 fortunately. There are two versions there, the base was on 20.04, and vllm-base was on 22.04. Should I revert both? |
|
FYI ubuntu 20.04 will reach its EOL on May 31, 2025, which is only two weeks away. |
I think we should consider the OS upgrade separately. Upgrading to Ubuntu 22.04 will put us on glibc 2.35. This will break vLLM on, for instance:
|
075ba43 to
812fda5
Compare
Signed-off-by: Junhao Li <junhao@ubicloud.com>
|
@0xjunhao I tried running the model OpenGVLab/InternVL3-1B-Instruct on an RTX 5090 with the dockerfile ubicloud/vllm-openai:latest and got this error CUDA error (/__w/xformers/xformers/third_party/flash-attention/hopper/flash_fwd_launch_template.h:175): no kernel image is available for execution on the device |
I believe I could work around this by building the latest xformers from source. |
|
Discussed with Simon, closing this PR. Please refer to PR #18095. |

Update the Dockerfile to include Blackwell archs and use ubuntu 24.04 as base image
Have tested that this works for RTX5090.
Test build FYI: ubicloud/vllm-openai:latest