Skip to content

Commit 5cb546a

Browse files
committed
Bump main
Signed-off-by: Yikun Jiang <yikunkero@gmail.com>
1 parent 12bcbd0 commit 5cb546a

File tree

5 files changed

+17
-11
lines changed

5 files changed

+17
-11
lines changed

.github/workflows/format_pr_body.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,7 @@ jobs:
3636

3737
- name: Get vLLM version
3838
run: |
39-
VLLM_COMMIT=6d8246aaffff3ebec84767e373212a7b8da328e2
39+
VLLM_COMMIT=c60e6137f0bf2034853919b3a9d705d7e06b93cf
4040
echo "VLLM_COMMIT=https://github.com/vllm-project/vllm/commit/$VLLM_COMMIT" >> $GITHUB_ENV
4141
4242
- name: Checkout repository

.github/workflows/vllm_ascend_test.yaml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -82,7 +82,7 @@ jobs:
8282
VLLM_USE_MODELSCOPE: True
8383
strategy:
8484
matrix:
85-
vllm_version: [6d8246aaffff3ebec84767e373212a7b8da328e2, v0.10.2]
85+
vllm_version: [c60e6137f0bf2034853919b3a9d705d7e06b93cf, v0.10.2]
8686
steps:
8787
- name: Install packages
8888
run: |
@@ -140,7 +140,7 @@ jobs:
140140
max-parallel: 2
141141
matrix:
142142
os: [linux-aarch64-a2-1]
143-
vllm_version: [6d8246aaffff3ebec84767e373212a7b8da328e2, v0.10.2]
143+
vllm_version: [c60e6137f0bf2034853919b3a9d705d7e06b93cf, v0.10.2]
144144
name: singlecard e2e test - light
145145
runs-on: ${{ matrix.os }}
146146
container:
@@ -206,7 +206,7 @@ jobs:
206206
max-parallel: 2
207207
matrix:
208208
os: [linux-aarch64-a2-2]
209-
vllm_version: [6d8246aaffff3ebec84767e373212a7b8da328e2, v0.10.2]
209+
vllm_version: [c60e6137f0bf2034853919b3a9d705d7e06b93cf, v0.10.2]
210210
name: multicard e2e test - light
211211
runs-on: ${{ matrix.os }}
212212
container:

.github/workflows/vllm_ascend_test_full.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -72,7 +72,7 @@ jobs:
7272
max-parallel: 2
7373
matrix:
7474
os: [linux-aarch64-a2-1]
75-
vllm_version: [6d8246aaffff3ebec84767e373212a7b8da328e2, v0.10.2]
75+
vllm_version: [c60e6137f0bf2034853919b3a9d705d7e06b93cf, v0.10.2]
7676
name: singlecard e2e test - full
7777
runs-on: ${{ matrix.os }}
7878
container:
@@ -156,7 +156,7 @@ jobs:
156156
max-parallel: 2
157157
matrix:
158158
os: [linux-aarch64-a2-2]
159-
vllm_version: [6d8246aaffff3ebec84767e373212a7b8da328e2, v0.10.2]
159+
vllm_version: [c60e6137f0bf2034853919b3a9d705d7e06b93cf, v0.10.2]
160160
name: multicard e2e test - full
161161
runs-on: ${{ matrix.os }}
162162
container:

tests/e2e/conftest.py

Lines changed: 8 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -32,7 +32,14 @@
3232
BatchEncoding, BatchFeature)
3333
from transformers.models.auto.auto_factory import _BaseAutoModelClass
3434
from vllm import LLM, SamplingParams
35-
from vllm.config import TaskOption, _get_and_verify_dtype
35+
36+
from vllm_ascend.utils import vllm_version_is
37+
38+
if vllm_version_is("0.10.2"):
39+
from vllm.config import TaskOption, _get_and_verify_dtype
40+
else:
41+
from vllm.config.model import TaskOption, _get_and_verify_dtype
42+
3643
from vllm.inputs import TextPrompt
3744
from vllm.outputs import RequestOutput
3845
from vllm.transformers_utils.utils import maybe_model_redirect

vllm_ascend/sample/sampler.py

Lines changed: 3 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,12 +1,11 @@
11
import torch
22
import torch_npu
3-
from vllm.config import LogprobsMode
43
from vllm.v1.sample.ops.topk_topp_sampler import TopKTopPSampler, random_sample
54
from vllm.v1.sample.sampler import Sampler
65

76
from vllm_ascend.utils import is_310p
87

9-
DEFAULT_LOGPROBS_MODE = LogprobsMode.RAW_LOGPROBS
8+
DEFAULT_LOGPROBS_MODE = "raw_logprobs"
109

1110

1211
class AscendSampler(Sampler):
@@ -65,9 +64,9 @@ def forward_native(self, logits, generators, k, p):
6564
"""Override pytorch native implementation to torch_npu"""
6665
logits = self._apply_top_k_top_p(logits, k, p)
6766
logits_to_return = None
68-
if self.logprobs_mode == LogprobsMode.PROCESSED_LOGITS:
67+
if self.logprobs_mode == "processed_logits":
6968
logits_to_return = logits
70-
elif self.logprobs_mode == LogprobsMode.PROCESSED_LOGPROBS:
69+
elif self.logprobs_mode == "processed_logprobs":
7170
logits_to_return = logits.log_softmax(dim=-1, dtype=torch.float32)
7271

7372
probs = logits.softmax(dim=-1, dtype=torch.float32)

0 commit comments

Comments
 (0)