-
Notifications
You must be signed in to change notification settings - Fork 558
[Doc] Add --shm-size option to Docker command for qwen3 vl 235B #3519
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
👋 Hi! Thank you for contributing to the vLLM Ascend project. The following points will speed up your PR merge:
If CI fails, you can run linting and testing checks locally according Contributing and Testing. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Code Review
This pull request adds the --shm-size option to the Docker run command in the multi_node_qwen3vl.md documentation file. This change aims to prevent potential core dumps of the vllm:EngineCore process under high workloads by increasing the shared memory size. I have added a review comment to address a potential issue.
Added shared memory size option to Docker run command.If shm-size is not specified, docker will use 64MB by default. In this case, vllm:EngineCore process may coredump if workload is high. Signed-off-by: likeful <irayki@gmail.com>
Signed-off-by: likeful <irayki@gmail.com> Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com>
| --device /dev/davinci_manager \ | ||
| --device /dev/devmm_svm \ | ||
| --device /dev/hisi_hdc \ | ||
| --shm-size=1000g # Adjust based on workload. Consider using a smaller value if memory is constrained. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
- This cmd can't be excute after this fix.
| --shm-size=1000g # Adjust based on workload. Consider using a smaller value if memory is constrained. | |
| --shm-size=4g \ |
- What's the recommanded value of
shm-sizeand setting rule? just saw [1] has many different examples.
[1] https://docs.vllm.ai/en/latest/cli/run-batch.html?q=shm-size
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We assume this configuration depends on workloads, in `Qwen3-VL-235B-A22B-Instruct and multiple nodes scenario, suggest set shm-size start from 256g, with 8g or 16g incremental step
Signed-off-by: leijie2015 <irayki@gmail.com>
…-project#3519) ### What this PR does / why we need it? Added shared memory size option to Docker run command.If shm-size is not specified, docker will use 64MB by default. In this case, vllm:EngineCore process may coredump if workload is high. ### Does this PR introduce _any_ user-facing change? No ### How was this patch tested? Done Closes: vllm-project#3513 - vLLM version: v0.11.0rc3 - vLLM main: https://github.com/vllm-project/vllm/commit/v0.11.0 --------- Signed-off-by: likeful <irayki@gmail.com> Signed-off-by: leijie2015 <irayki@gmail.com> Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com>
What this PR does / why we need it?
Added shared memory size option to Docker run command.If shm-size is not specified, docker will use 64MB by default. In this case, vllm:EngineCore process may coredump if workload is high.
Does this PR introduce any user-facing change?
No
How was this patch tested?
Done
Closes: #3513