@@ -144,7 +144,7 @@ jobs:
144144 VLLM_USE_MODELSCOPE : True
145145 strategy :
146146 matrix :
147- vllm_version : [main, v0.9.1 ]
147+ vllm_version : [main, ]
148148 steps :
149149 - name : Install packages
150150 run : |
@@ -193,111 +193,111 @@ jobs:
193193 name : vllm-ascend
194194 verbose : true
195195
196- e2e :
197- needs : [lint]
198- # only trigger e2e test on pull request after lint passed
199- if : ${{ needs.lint.result == 'success' && github.event_name == 'pull_request' }}
200- strategy :
201- max-parallel : 2
202- matrix :
203- os : [linux-arm64-npu-1]
204- vllm_version : [main, v0.9.1 ]
205- name : singlecard e2e test
206- runs-on : ${{ matrix.os }}
207- container :
208- # TODO(yikun): Remove m.daocloud.io prefix when infra proxy ready
209- image : m.daocloud.io/quay.io/ascend/cann:8.1.rc1-910b-ubuntu22.04-py3.10
210- env :
211- VLLM_LOGGING_LEVEL : ERROR
212- steps :
213- - name : Check npu and CANN info
214- run : |
215- npu-smi info
216- cat /usr/local/Ascend/ascend-toolkit/latest/"$(uname -i)"-linux/ascend_toolkit_install.info
217-
218- - name : Config mirrors
219- run : |
220- sed -i 's|ports.ubuntu.com|mirrors.tuna.tsinghua.edu.cn|g' /etc/apt/sources.list
221- pip config set global.index-url https://mirrors.tuna.tsinghua.edu.cn/pypi/web/simple
222- apt-get update -y
223- apt install git -y
224- git config --global url."https://gh-proxy.test.osinfra.cn/https://github.com/".insteadOf https://github.com/
225-
226- - name : Checkout vllm-project/vllm-ascend repo
227- uses : actions/checkout@v4
228-
229- - name : Install system dependencies
230- run : |
231- apt-get -y install `cat packages.txt`
232- apt-get -y install gcc g++ cmake libnuma-dev
233-
234- - name : Checkout vllm-project/vllm repo
235- uses : actions/checkout@v4
236- with :
237- repository : vllm-project/vllm
238- ref : ${{ matrix.vllm_version }}
239- path : ./vllm-empty
240-
241- - name : Install vllm-project/vllm from source
242- working-directory : ./vllm-empty
243- run : |
244- VLLM_TARGET_DEVICE=empty pip install -e .
245-
246- - name : Install vllm-project/vllm-ascend
247- env :
248- PIP_EXTRA_INDEX_URL : https://mirrors.huaweicloud.com/ascend/repos/pypi
249- run : |
250- pip install -r requirements-dev.txt
251- pip install -v -e .
252-
253- - name : Run e2e test for V1 Engine
254- env :
255- VLLM_USE_V1 : 1
256- VLLM_WORKER_MULTIPROC_METHOD : spawn
257- VLLM_USE_MODELSCOPE : True
258- run : |
259- pytest -sv tests/e2e/singlecard/test_offline_inference.py
260- # TODO: switch hf to modelscope
261- VLLM_USE_MODELSCOPE=False HF_ENDPOINT=https://hf-mirror.com \
262- pytest -sv tests/e2e/singlecard/test_ilama_lora.py
263- pytest -sv tests/e2e/singlecard/test_guided_decoding.py
264- pytest -sv tests/e2e/singlecard/test_camem.py
265- pytest -sv tests/e2e/singlecard/ \
266- --ignore=tests/e2e/singlecard/test_offline_inference.py \
267- --ignore=tests/e2e/singlecard/test_ilama_lora.py \
268- --ignore=tests/e2e/singlecard/test_guided_decoding.py \
269- --ignore=tests/e2e/singlecard/test_camem.py
270-
271- - name : Run e2e test on V0 engine
272- if : ${{ github.event_name == 'schedule' }}
273- env :
274- VLLM_USE_V1 : 0
275- VLLM_USE_MODELSCOPE : True
276- run : |
277- pytest -sv tests/e2e/singlecard/test_offline_inference.py
278- # TODO: switch hf to modelscope
279- VLLM_USE_MODELSCOPE=False HF_ENDPOINT=https://hf-mirror.com \
280- pytest -sv tests/e2e/singlecard/test_ilama_lora.py
281- pytest -sv tests/e2e/singlecard/test_guided_decoding.py
282- pytest -sv tests/e2e/singlecard/test_camem.py
283- pytest -sv tests/e2e/singlecard/test_prompt_embedding.py
284- pytest -sv tests/e2e/singlecard/ \
285- --ignore=tests/e2e/singlecard/test_offline_inference.py \
286- --ignore=tests/e2e/singlecard/test_ilama_lora.py \
287- --ignore=tests/e2e/singlecard/test_guided_decoding.py \
288- --ignore=tests/e2e/singlecard/test_camem.py \
289- --ignore=tests/e2e/singlecard/test_prompt_embedding.py \
290- --ignore=tests/e2e/singlecard/core/test_ascend_scheduler.py \
291- --ignore=tests/e2e/singlecard/core/test_ascend_scheduler_e2e.py
196+ # e2e:
197+ # needs: [lint]
198+ # # only trigger e2e test on pull request after lint passed
199+ # if: ${{ needs.lint.result == 'success' && github.event_name == 'pull_request' }}
200+ # strategy:
201+ # max-parallel: 2
202+ # matrix:
203+ # os: [linux-arm64-npu-1]
204+ # vllm_version: [main, ]
205+ # name: singlecard e2e test
206+ # runs-on: ${{ matrix.os }}
207+ # container:
208+ # # TODO(yikun): Remove m.daocloud.io prefix when infra proxy ready
209+ # image: m.daocloud.io/quay.io/ascend/cann:8.1.rc1-910b-ubuntu22.04-py3.10
210+ # env:
211+ # VLLM_LOGGING_LEVEL: ERROR
212+ # steps:
213+ # - name: Check npu and CANN info
214+ # run: |
215+ # npu-smi info
216+ # cat /usr/local/Ascend/ascend-toolkit/latest/"$(uname -i)"-linux/ascend_toolkit_install.info
217+
218+ # - name: Config mirrors
219+ # run: |
220+ # sed -i 's|ports.ubuntu.com|mirrors.tuna.tsinghua.edu.cn|g' /etc/apt/sources.list
221+ # pip config set global.index-url https://mirrors.tuna.tsinghua.edu.cn/pypi/web/simple
222+ # apt-get update -y
223+ # apt install git -y
224+ # git config --global url."https://gh-proxy.test.osinfra.cn/https://github.com/".insteadOf https://github.com/
225+
226+ # - name: Checkout vllm-project/vllm-ascend repo
227+ # uses: actions/checkout@v4
228+
229+ # - name: Install system dependencies
230+ # run: |
231+ # apt-get -y install `cat packages.txt`
232+ # apt-get -y install gcc g++ cmake libnuma-dev
233+
234+ # - name: Checkout vllm-project/vllm repo
235+ # uses: actions/checkout@v4
236+ # with:
237+ # repository: vllm-project/vllm
238+ # ref: ${{ matrix.vllm_version }}
239+ # path: ./vllm-empty
240+
241+ # - name: Install vllm-project/vllm from source
242+ # working-directory: ./vllm-empty
243+ # run: |
244+ # VLLM_TARGET_DEVICE=empty pip install -e .
245+
246+ # - name: Install vllm-project/vllm-ascend
247+ # env:
248+ # PIP_EXTRA_INDEX_URL: https://mirrors.huaweicloud.com/ascend/repos/pypi
249+ # run: |
250+ # pip install -r requirements-dev.txt
251+ # pip install -v -e .
252+
253+ # - name: Run e2e test for V1 Engine
254+ # env:
255+ # VLLM_USE_V1: 1
256+ # VLLM_WORKER_MULTIPROC_METHOD: spawn
257+ # VLLM_USE_MODELSCOPE: True
258+ # run: |
259+ # pytest -sv tests/e2e/singlecard/test_offline_inference.py
260+ # # TODO: switch hf to modelscope
261+ # VLLM_USE_MODELSCOPE=False HF_ENDPOINT=https://hf-mirror.com \
262+ # pytest -sv tests/e2e/singlecard/test_ilama_lora.py
263+ # pytest -sv tests/e2e/singlecard/test_guided_decoding.py
264+ # pytest -sv tests/e2e/singlecard/test_camem.py
265+ # pytest -sv tests/e2e/singlecard/ \
266+ # --ignore=tests/e2e/singlecard/test_offline_inference.py \
267+ # --ignore=tests/e2e/singlecard/test_ilama_lora.py \
268+ # --ignore=tests/e2e/singlecard/test_guided_decoding.py \
269+ # --ignore=tests/e2e/singlecard/test_camem.py
270+
271+ # - name: Run e2e test on V0 engine
272+ # if: ${{ github.event_name == 'schedule' }}
273+ # env:
274+ # VLLM_USE_V1: 0
275+ # VLLM_USE_MODELSCOPE: True
276+ # run: |
277+ # pytest -sv tests/e2e/singlecard/test_offline_inference.py
278+ # # TODO: switch hf to modelscope
279+ # VLLM_USE_MODELSCOPE=False HF_ENDPOINT=https://hf-mirror.com \
280+ # pytest -sv tests/e2e/singlecard/test_ilama_lora.py
281+ # pytest -sv tests/e2e/singlecard/test_guided_decoding.py
282+ # pytest -sv tests/e2e/singlecard/test_camem.py
283+ # pytest -sv tests/e2e/singlecard/test_prompt_embedding.py
284+ # pytest -sv tests/e2e/singlecard/ \
285+ # --ignore=tests/e2e/singlecard/test_offline_inference.py \
286+ # --ignore=tests/e2e/singlecard/test_ilama_lora.py \
287+ # --ignore=tests/e2e/singlecard/test_guided_decoding.py \
288+ # --ignore=tests/e2e/singlecard/test_camem.py \
289+ # --ignore=tests/e2e/singlecard/test_prompt_embedding.py \
290+ # --ignore=tests/e2e/singlecard/core/test_ascend_scheduler.py \
291+ # --ignore=tests/e2e/singlecard/core/test_ascend_scheduler_e2e.py
292292
293293 e2e-4-cards :
294- needs : [e2e]
295- if : ${{ needs.e2e.result == 'success' }}
294+ # needs: [e2e]
295+ # if: ${{ needs.e2e.result == 'success' }}
296296 strategy :
297297 max-parallel : 1
298298 matrix :
299299 os : [linux-arm64-npu-4]
300- vllm_version : [main, v0.9.1 ]
300+ vllm_version : [main, ]
301301 name : multicard e2e test
302302 runs-on : ${{ matrix.os }}
303303 container :
0 commit comments