Skip to content

Commit d9cdc65

Browse files
wxsIceyMengqingCao
andauthored
Upgrade to new vllm commit (#3719)
### What this PR does / why we need it? Upgrade to new vllm commit: vllm-project/vllm@c9461e0 - Fix many imports, caused by vllm-project/vllm#26908 - Fix import ```sha256```, caused by vllm-project/vllm#27169 - Remove ```SchedulerConfig.send_delta_data```, caused by vllm-project/vllm#27142 - Fix ```FusedMoE``` because of dual stream execution, caused by vllm-project/vllm#26440 ### Does this PR introduce _any_ user-facing change? N/A ### How was this patch tested? CI passed with new added/existing test. - vLLM version: v0.11.0rc3 - vLLM main: vllm-project/vllm@17c540a --------- Signed-off-by: MengqingCao <cmq0113@163.com> Signed-off-by: Icey <1790571317@qq.com> Co-authored-by: MengqingCao <cmq0113@163.com>
1 parent 226f832 commit d9cdc65

37 files changed

+229
-71
lines changed

.github/workflows/format_pr_body.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,7 @@ jobs:
3636

3737
- name: Get vLLM version
3838
run: |
39-
VLLM_COMMIT=17c540a993af88204ad1b78345c8a865cf58ce44
39+
VLLM_COMMIT=c9461e05a4ed3557cfbf4b15ded1e26761cc39ca
4040
echo "VLLM_COMMIT=https://github.com/vllm-project/vllm/commit/$VLLM_COMMIT" >> $GITHUB_ENV
4141
4242
- name: Checkout repository

.github/workflows/vllm_ascend_test.yaml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -42,7 +42,7 @@ jobs:
4242
lint:
4343
uses: ./.github/workflows/pre-commit.yml
4444
with:
45-
vllm: 17c540a993af88204ad1b78345c8a865cf58ce44
45+
vllm: c9461e05a4ed3557cfbf4b15ded1e26761cc39ca
4646

4747
changes:
4848
runs-on: ubuntu-latest
@@ -83,7 +83,7 @@ jobs:
8383
VLLM_USE_MODELSCOPE: True
8484
strategy:
8585
matrix:
86-
vllm_version: [17c540a993af88204ad1b78345c8a865cf58ce44, v0.11.0]
86+
vllm_version: [c9461e05a4ed3557cfbf4b15ded1e26761cc39ca, v0.11.0]
8787
steps:
8888
- name: Install packages
8989
run: |
@@ -140,7 +140,7 @@ jobs:
140140
name: e2e-light
141141
strategy:
142142
matrix:
143-
vllm_version: [17c540a993af88204ad1b78345c8a865cf58ce44, v0.11.0]
143+
vllm_version: [c9461e05a4ed3557cfbf4b15ded1e26761cc39ca, v0.11.0]
144144
# Note (yikun): If CI resource are limited we can split job into two chain jobs
145145
needs: [lint, changes]
146146
# only trigger e2e test after lint passed and the change is e2e related with pull request.

.github/workflows/vllm_ascend_test_full.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -69,7 +69,7 @@ jobs:
6969
name: e2e-full
7070
strategy:
7171
matrix:
72-
vllm_version: [17c540a993af88204ad1b78345c8a865cf58ce44, v0.11.0]
72+
vllm_version: [c9461e05a4ed3557cfbf4b15ded1e26761cc39ca, v0.11.0]
7373
needs: [changes]
7474
if: ${{ needs.changes.outputs.e2e_tracker == 'true' }}
7575
uses: ./.github/workflows/_e2e_test.yaml

examples/offline_data_parallel.py

Lines changed: 5 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -63,7 +63,11 @@
6363
from vllm import LLM, SamplingParams
6464
from vllm.distributed.parallel_state import ( # noqa E402
6565
destroy_distributed_environment, destroy_model_parallel)
66-
from vllm.utils import get_open_port
66+
from vllm_ascend.utils import vllm_version_is
67+
if vllm_version_is("0.11.0"):
68+
from vllm.utils import get_open_port
69+
else:
70+
from vllm.utils.network_utils import get_open_port
6771

6872
os.environ["VLLM_USE_MODELSCOPE"] = "True"
6973
os.environ["VLLM_WORKER_MULTIPROC_METHOD"] = "spawn"

examples/offline_external_launcher.py

Lines changed: 8 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -65,9 +65,15 @@
6565
import torch
6666
from vllm import LLM, SamplingParams
6767
from vllm.distributed.parallel_state import ( # noqa E402
68-
destroy_distributed_environment, destroy_model_parallel, get_tp_group)
69-
from vllm.utils import get_open_port, GiB_bytes
68+
destroy_distributed_environment, destroy_model_parallel, get_tp_group)
7069
from safetensors.torch import load_file
70+
from vllm_ascend.utils import vllm_version_is
71+
if vllm_version_is("0.11.0"):
72+
from vllm.utils import GiB_bytes, get_open_port
73+
74+
else:
75+
from vllm.utils.mem_constants import GiB_bytes
76+
from vllm.utils.network_utils import get_open_port
7177

7278
os.environ["VLLM_USE_MODELSCOPE"] = "True"
7379
os.environ["VLLM_WORKER_MULTIPROC_METHOD"] = "spawn"

examples/offline_inference_sleep_mode_npu.py

Lines changed: 5 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,11 @@
2020

2121
import torch
2222
from vllm import LLM, SamplingParams
23-
from vllm.utils import GiB_bytes
23+
from vllm_ascend.utils import vllm_version_is
24+
if vllm_version_is("0.11.0"):
25+
from vllm.utils import GiB_bytes
26+
else:
27+
from vllm.utils.mem_constants import GiB_bytes
2428

2529
os.environ["VLLM_USE_MODELSCOPE"] = "True"
2630
os.environ["VLLM_WORKER_MULTIPROC_METHOD"] = "spawn"

examples/offline_weight_load.py

Lines changed: 7 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -66,8 +66,14 @@
6666
from vllm import LLM, SamplingParams
6767
from vllm.distributed.parallel_state import ( # noqa E402
6868
destroy_distributed_environment, destroy_model_parallel, get_tp_group)
69-
from vllm.utils import get_open_port, GiB_bytes
7069
from safetensors.torch import load_file
70+
from vllm_ascend.utils import vllm_version_is
71+
if vllm_version_is("0.11.0"):
72+
from vllm.utils import GiB_bytes, get_open_port
73+
74+
else:
75+
from vllm.utils.mem_constants import GiB_bytes
76+
from vllm.utils.network_utils import get_open_port
7177

7278
os.environ["VLLM_USE_MODELSCOPE"] = "True"
7379
os.environ["VLLM_WORKER_MULTIPROC_METHOD"] = "spawn"

tests/e2e/conftest.py

Lines changed: 6 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -45,7 +45,6 @@
4545
from vllm.outputs import RequestOutput
4646
from vllm.platforms import current_platform
4747
from vllm.transformers_utils.utils import maybe_model_redirect
48-
from vllm.utils import get_open_port
4948

5049
from tests.e2e.model_utils import (TokensTextLogprobs,
5150
TokensTextLogprobsPromptLogprobs)
@@ -55,6 +54,12 @@
5554
# we not explicitly patch here, some of them might be effectiveless
5655
# in pytest scenario
5756
from vllm_ascend.utils import adapt_patch # noqa E402
57+
from vllm_ascend.utils import vllm_version_is
58+
59+
if vllm_version_is("0.11.0"):
60+
from vllm.utils import get_open_port
61+
else:
62+
from vllm.utils.network_utils import get_open_port
5863

5964
adapt_patch(True)
6065
adapt_patch(False)

tests/e2e/multicard/test_single_request_aclgraph.py

Lines changed: 6 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -19,9 +19,14 @@
1919

2020
import openai
2121
import pytest
22-
from vllm.utils import get_open_port
2322

2423
from tests.e2e.conftest import RemoteOpenAIServer
24+
from vllm_ascend.utils import vllm_version_is
25+
26+
if vllm_version_is("0.11.0"):
27+
from vllm.utils import get_open_port
28+
else:
29+
from vllm.utils.network_utils import get_open_port
2530

2631
MODELS = [
2732
"Qwen/Qwen3-30B-A3B",

tests/e2e/nightly/models/test_qwen3_32b.py

Lines changed: 6 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18,10 +18,15 @@
1818

1919
import openai
2020
import pytest
21-
from vllm.utils import get_open_port
2221

2322
from tests.e2e.conftest import RemoteOpenAIServer
2423
from tools.aisbench import run_aisbench_cases
24+
from vllm_ascend.utils import vllm_version_is
25+
26+
if vllm_version_is("0.11.0"):
27+
from vllm.utils import get_open_port
28+
else:
29+
from vllm.utils.network_utils import get_open_port
2530

2631
MODELS = [
2732
"Qwen/Qwen3-32B",

0 commit comments

Comments
 (0)