Skip to content

Conversation

@leijie-cn
Copy link
Contributor

@leijie-cn leijie-cn commented Oct 17, 2025

What this PR does / why we need it?

Added shared memory size option to Docker run command.If shm-size is not specified, docker will use 64MB by default. In this case, vllm:EngineCore process may coredump if workload is high.

Does this PR introduce any user-facing change?

No

How was this patch tested?

Done

Closes: #3513

@github-actions github-actions bot added the documentation Improvements or additions to documentation label Oct 17, 2025
@github-actions
Copy link

👋 Hi! Thank you for contributing to the vLLM Ascend project. The following points will speed up your PR merge:‌‌

  • A PR should do only one thing, smaller PRs enable faster reviews.
  • Every PR should include unit tests and end-to-end tests ‌to ensure it works and is not broken by other future PRs.
  • Write the commit message by fulfilling the PR description to help reviewer and future developers understand.

If CI fails, you can run linting and testing checks locally according Contributing and Testing.

Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request adds the --shm-size option to the Docker run command in the multi_node_qwen3vl.md documentation file. This change aims to prevent potential core dumps of the vllm:EngineCore process under high workloads by increasing the shared memory size. I have added a review comment to address a potential issue.

leijie-cn and others added 2 commits October 17, 2025 20:50
Added shared memory size option to Docker run command.If shm-size is not specified, docker will use 64MB by default.
In this case, vllm:EngineCore process may coredump if workload is high.

Signed-off-by: likeful <irayki@gmail.com>
Signed-off-by: likeful <irayki@gmail.com>
Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com>
--device /dev/davinci_manager \
--device /dev/devmm_svm \
--device /dev/hisi_hdc \
--shm-size=1000g # Adjust based on workload. Consider using a smaller value if memory is constrained.
Copy link
Collaborator

@Yikun Yikun Oct 18, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

  1. This cmd can't be excute after this fix.
Suggested change
--shm-size=1000g # Adjust based on workload. Consider using a smaller value if memory is constrained.
--shm-size=4g \
  1. What's the recommanded value of shm-size and setting rule? just saw [1] has many different examples.

[1] https://docs.vllm.ai/en/latest/cli/run-batch.html?q=shm-size

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

We assume this configuration depends on workloads, in `Qwen3-VL-235B-A22B-Instruct and multiple nodes scenario, suggest set shm-size start from 256g, with 8g or 16g incremental step

Signed-off-by: leijie2015 <irayki@gmail.com>
@leijie-cn leijie-cn changed the title Add --shm-size option to Docker command Fixes #3513 Add --shm-size option to Docker command Oct 20, 2025
@leijie-cn leijie-cn changed the title Fixes #3513 Add --shm-size option to Docker command [Bugfix] Add --shm-size option to Docker command Oct 20, 2025
@Yikun Yikun changed the title [Bugfix] Add --shm-size option to Docker command [Doc] Add --shm-size option to Docker command Oct 20, 2025
@Yikun Yikun changed the title [Doc] Add --shm-size option to Docker command [Bugfix] Add --shm-size option to Docker command for qwen3 vl 235B Oct 20, 2025
@Yikun Yikun changed the title [Bugfix] Add --shm-size option to Docker command for qwen3 vl 235B [Doc] Add --shm-size option to Docker command for qwen3 vl 235B Oct 20, 2025
@Yikun Yikun merged commit 6b68579 into vllm-project:main Oct 20, 2025
13 checks passed
ZYang6263 pushed a commit to rjg-lyh/vllm-ascend that referenced this pull request Oct 23, 2025
…-project#3519)

### What this PR does / why we need it?
Added shared memory size option to Docker run command.If shm-size is not
specified, docker will use 64MB by default. In this case,
vllm:EngineCore process may coredump if workload is high.
### Does this PR introduce _any_ user-facing change?
No
### How was this patch tested?
Done

Closes: vllm-project#3513

- vLLM version: v0.11.0rc3
- vLLM main: https://github.com/vllm-project/vllm/commit/v0.11.0

---------

Signed-off-by: likeful <irayki@gmail.com>
Signed-off-by: leijie2015 <irayki@gmail.com>
Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

documentation Improvements or additions to documentation

Projects

None yet

Development

Successfully merging this pull request may close these issues.

[Bug]: vllm:EngineCore process coredump while testing TextVQA dataset for both Qwen3-VL-30B-A3B-Instruct and Qwen2.5-VL-7B-Instruct

2 participants