Skip to content

Commit 3a4ce2a

Browse files
authored
[Docs] Fix vllm and vllm-ascend version (#107)
### What this PR does / why we need it? Fix vllm and vllm-ascend version | branch/tag | vllm_version | vllm_ascend_version|pip_vllm_ascend_version|pip_vllm_version| |----|----|----|----|----| | main | main | main | v0.7.1rc1 | v0.7.1 | | v0.7.1-dev | v0.7.1 | v0.7.1rc1 | v0.7.1rc1 | v0.7.1 | | v0.7.1rc1 | v0.7.1 | v0.7.1rc1 | v0.7.1rc1 | v0.7.1 | ### Does this PR introduce _any_ user-facing change? No ### How was this patch tested? CI passed Signed-off-by: Yikun Jiang <yikunkero@gmail.com>
1 parent cff03a4 commit 3a4ce2a

File tree

4 files changed

+26
-15
lines changed

4 files changed

+26
-15
lines changed

docs/source/conf.py

Lines changed: 7 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -61,14 +61,17 @@
6161
# Change this when cut down release
6262
myst_substitutions = {
6363
# the branch of vllm, used in vllm clone
64-
# such as 'main', 'v0.7.1'
64+
# - main branch: 'main'
65+
# - vX.Y.Z branch: 'vX.Y.Z'
6566
'vllm_version': 'main',
6667
# the branch of vllm-ascend, used in vllm-ascend clone and image tag
67-
# such as 'main', 'v0.7.1-dev', 'v0.7.1rc1'
68+
# - main branch: 'main'
69+
# - vX.Y.Z branch: latest vllm-ascend release tag
6870
'vllm_ascend_version': 'main',
69-
# the newest release version of vllm, used in quick start or container image tag.
71+
# the newest release version of vllm-ascend and matched vLLM, used in pip install.
7072
# This value should be updated when cut down release.
71-
'vllm_newest_release_version': "v0.7.1rc1",
73+
'pip_vllm_ascend_version': "v0.7.1rc1",
74+
'pip_vllm_version': "v0.7.1",
7275
}
7376

7477
# Add any paths that contain templates here, relative to this directory.

docs/source/installation.md

Lines changed: 7 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -114,15 +114,18 @@ Once it's done, you can start to set up `vllm` and `vllm-ascend`.
114114

115115
You can install `vllm` and `vllm-ascend` from **pre-built wheel**:
116116

117-
```bash
117+
```{code-block} bash
118+
:substitutions:
119+
118120
# Install vllm from source, since `pip install vllm` doesn't work on CPU currently.
119121
# It'll be fixed in the next vllm release, e.g. v0.7.3.
120-
git clone --branch v0.7.1 https://github.com/vllm-project/vllm
122+
git clone --branch |pip_vllm_version| https://github.com/vllm-project/vllm
123+
121124
cd vllm
122125
VLLM_TARGET_DEVICE=empty pip install . --extra-index https://download.pytorch.org/whl/cpu/
123126
124127
# Install vllm-ascend from pypi.
125-
pip install vllm-ascend --extra-index https://download.pytorch.org/whl/cpu/
128+
pip install vllm-ascend==|pip_vllm_ascend_version| --extra-index https://download.pytorch.org/whl/cpu/
126129
127130
# Once the packages are installed, you need to install `torch-npu` manually,
128131
# because that vllm-ascend relies on an unreleased version of torch-npu.
@@ -146,7 +149,7 @@ or build from **source code**:
146149
```{code-block} bash
147150
:substitutions:
148151
149-
git clone --depth 1 --branch |vllm_version| https://github.com/vllm-project/vllm
152+
git clone --depth 1 --branch |vllm_version| https://github.com/vllm-project/vllm
150153
cd vllm
151154
VLLM_TARGET_DEVICE=empty pip install . --extra-index https://download.pytorch.org/whl/cpu/
152155

docs/source/quick_start.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@
1212
:substitutions:
1313
1414
# You can change version a suitable one base on your requirement, e.g. main
15-
export IMAGE=quay.io/ascend:|vllm_newest_release_version|
15+
export IMAGE=quay.io/ascend/vllm-ascend:|vllm_ascend_version|
1616
1717
docker run \
1818
--name vllm-ascend \

docs/source/tutorials.md

Lines changed: 11 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,8 @@
66

77
Run docker container:
88

9-
```bash
9+
```{code-block} bash
10+
:substitutions:
1011
docker run \
1112
--name vllm-ascend \
1213
--device /dev/davinci0 \
@@ -20,7 +21,7 @@ docker run \
2021
-v /etc/ascend_install.info:/etc/ascend_install.info \
2122
-v /root/.cache:/root/.cache \
2223
-p 8000:8000 \
23-
-it quay.io/ascend/vllm-ascend:v0.7.1rc1 bash
24+
-it quay.io/ascend/vllm-ascend:|vllm_ascend_version| bash
2425
```
2526

2627
Setup environment variables:
@@ -67,7 +68,9 @@ Prompt: 'The future of AI is', Generated text: ' following you. As the technolog
6768

6869
Run docker container to start the vLLM server on a single NPU:
6970

70-
```bash
71+
```{code-block} bash
72+
:substitutions:
73+
7174
docker run \
7275
--name vllm-ascend \
7376
--device /dev/davinci0 \
@@ -83,7 +86,7 @@ docker run \
8386
-p 8000:8000 \
8487
-e VLLM_USE_MODELSCOPE=True \
8588
-e PYTORCH_NPU_ALLOC_CONF=max_split_size_mb:256 \
86-
-it quay.io/ascend/vllm-ascend:v0.7.1rc1 \
89+
-it quay.io/ascend/vllm-ascend:|vllm_ascend_version| \
8790
vllm serve Qwen/Qwen2.5-7B-Instruct --max_model_len 26240
8891
```
8992

@@ -131,7 +134,9 @@ INFO 02-13 08:34:35 logger.py:39] Received request cmpl-574f00e342904692a73fb6c1
131134

132135
Run docker container:
133136

134-
```bash
137+
```{code-block} bash
138+
:substitutions:
139+
135140
docker run \
136141
--name vllm-ascend \
137142
--device /dev/davinci0 \
@@ -146,7 +151,7 @@ docker run \
146151
-v /etc/ascend_install.info:/etc/ascend_install.info \
147152
-v /root/.cache:/root/.cache \
148153
-p 8000:8000 \
149-
-it quay.io/ascend/vllm-ascend:v0.7.1rc1 bash
154+
-it quay.io/ascend/vllm-ascend:|vllm_ascend_version| bash
150155
```
151156

152157
Setup environment variables:

0 commit comments

Comments
 (0)