Skip to content

Conversation

@gshtras
Copy link
Collaborator

@gshtras gshtras commented Sep 22, 2025

Make all whls install at the same command to fix the missing dependency issue when triton_kernels would pull CUDA dependencies from pypi
Split FA into its own build step to run in parallel

Signed-off-by: Gregory Shtrasberg <Gregory.Shtrasberg@amd.com>
@mergify mergify bot added ci/build rocm Related to AMD ROCm labels Sep 22, 2025
Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request refactors the ROCm base Dockerfile to improve the build process. Key changes include installing all wheel files in a single command to prevent dependency conflicts, and separating the FlashAttention build into its own stage to enable parallel builds. These are positive changes that enhance the robustness and efficiency of the Docker image creation. I have identified one high-severity issue related to the filtering of GPU architectures for the FlashAttention build, which could result in an incorrect build configuration. A suggestion to fix this is provided in the detailed comments.

@gshtras gshtras added the ready ONLY add when PR is ready to merge/full CI is needed label Sep 23, 2025
@ProExpertProg ProExpertProg merged commit 302eb94 into vllm-project:main Sep 24, 2025
29 checks passed
yewentao256 pushed a commit that referenced this pull request Oct 3, 2025
…25415)

Signed-off-by: Gregory Shtrasberg <Gregory.Shtrasberg@amd.com>
Signed-off-by: yewentao256 <zhyanwentao@126.com>
xuebwang-amd pushed a commit to xuebwang-amd/vllm that referenced this pull request Oct 10, 2025
…llm-project#25415)

Signed-off-by: Gregory Shtrasberg <Gregory.Shtrasberg@amd.com>
Signed-off-by: xuebwang-amd <xuebwang@amd.com>
choprahetarth pushed a commit to Tandemn-Labs/vllm that referenced this pull request Oct 11, 2025
…llm-project#25415)

Signed-off-by: Gregory Shtrasberg <Gregory.Shtrasberg@amd.com>
lywa1998 pushed a commit to lywa1998/vllm that referenced this pull request Oct 20, 2025
…llm-project#25415)

Signed-off-by: Gregory Shtrasberg <Gregory.Shtrasberg@amd.com>
xuebwang-amd pushed a commit to xuebwang-amd/vllm that referenced this pull request Oct 24, 2025
…llm-project#25415)

Signed-off-by: Gregory Shtrasberg <Gregory.Shtrasberg@amd.com>
Signed-off-by: xuebwang-amd <xuebwang@amd.com>
rtourgeman pushed a commit to rtourgeman/vllm that referenced this pull request Nov 10, 2025
…llm-project#25415)

Signed-off-by: Gregory Shtrasberg <Gregory.Shtrasberg@amd.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ci/build ready ONLY add when PR is ready to merge/full CI is needed rocm Related to AMD ROCm

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants