Skip to content

Commit f0f6b9d

Browse files
committed
fix import vllm_version_is
Signed-off-by: Icey <1790571317@qq.com>
1 parent b6a9207 commit f0f6b9d

File tree

14 files changed

+64
-22
lines changed

14 files changed

+64
-22
lines changed

.github/workflows/format_pr_body.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,7 @@ jobs:
3636

3737
- name: Get vLLM version
3838
run: |
39-
VLLM_COMMIT=17c540a993af88204ad1b78345c8a865cf58ce44
39+
VLLM_COMMIT=9fce7bee745230d61c60ad467966790553b0ba48
4040
echo "VLLM_COMMIT=https://github.com/vllm-project/vllm/commit/$VLLM_COMMIT" >> $GITHUB_ENV
4141
4242
- name: Checkout repository

.github/workflows/vllm_ascend_test.yaml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -42,7 +42,7 @@ jobs:
4242
lint:
4343
uses: ./.github/workflows/pre-commit.yml
4444
with:
45-
vllm: 17c540a993af88204ad1b78345c8a865cf58ce44
45+
vllm: 9fce7bee745230d61c60ad467966790553b0ba48
4646

4747
changes:
4848
runs-on: ubuntu-latest
@@ -83,7 +83,7 @@ jobs:
8383
VLLM_USE_MODELSCOPE: True
8484
strategy:
8585
matrix:
86-
vllm_version: [17c540a993af88204ad1b78345c8a865cf58ce44, v0.11.0]
86+
vllm_version: [9fce7bee745230d61c60ad467966790553b0ba48, v0.11.0]
8787
steps:
8888
- name: Install packages
8989
run: |
@@ -146,7 +146,7 @@ jobs:
146146
name: e2e-light
147147
strategy:
148148
matrix:
149-
vllm_version: [17c540a993af88204ad1b78345c8a865cf58ce44, v0.11.0]
149+
vllm_version: [9fce7bee745230d61c60ad467966790553b0ba48, v0.11.0]
150150
# Note (yikun): If CI resource are limited we can split job into two chain jobs
151151
needs: [lint, changes]
152152
# only trigger e2e test after lint passed and the change is e2e related with pull request.

.github/workflows/vllm_ascend_test_full.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -69,7 +69,7 @@ jobs:
6969
name: e2e-full
7070
strategy:
7171
matrix:
72-
vllm_version: [17c540a993af88204ad1b78345c8a865cf58ce44, v0.11.0]
72+
vllm_version: [9fce7bee745230d61c60ad467966790553b0ba48, v0.11.0]
7373
needs: [changes]
7474
if: ${{ needs.changes.outputs.e2e_tracker == 'true' }}
7575
uses: ./.github/workflows/_e2e_test.yaml

vllm_ascend/attention/attention_v1.py

Lines changed: 8 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@
2626
AttentionLayer, AttentionType)
2727
from vllm.config import VllmConfig
2828
from vllm.forward_context import ForwardContext, get_forward_context
29-
from vllm.utils import cdiv, direct_register_custom_op
29+
from vllm.utils import cdiv
3030
from vllm.v1.attention.backends.utils import AttentionCGSupport
3131
from vllm.v1.core.sched.output import SchedulerOutput
3232
from vllm.v1.kv_cache_interface import AttentionSpec
@@ -38,10 +38,16 @@
3838
update_graph_params_workspaces)
3939
from vllm_ascend.ops.attention import vanilla_chunked_prefill
4040
from vllm_ascend.utils import (ACL_FORMAT_FRACTAL_NZ, aligned_16, is_310p,
41-
nd_to_nz_2d, nd_to_nz_spec, version_check)
41+
nd_to_nz_2d, nd_to_nz_spec, version_check,
42+
vllm_version_is)
4243

4344
from ..utils import weak_ref_tensors
4445

46+
if vllm_version_is("0.11.0"):
47+
from vllm.utils import direct_register_custom_op
48+
else:
49+
from vllm.utils.torch_utils import direct_register_custom_op
50+
4551

4652
class AscendAttentionBackend(AttentionBackend):
4753
accept_output_buffer: bool = True

vllm_ascend/models/deepseek_v3_2.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -67,7 +67,6 @@
6767
from vllm_ascend.models.layers.sfa import AscendSFAModules, Indexer
6868
from vllm_ascend.ops.common_fused_moe import AscendFusedMoE
6969
from vllm_ascend.ops.linear import AscendLinearBase
70-
7170
from vllm_ascend.utils import vllm_version_is
7271

7372
if vllm_version_is("0.11.0"):

vllm_ascend/models/layers/mla.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -28,17 +28,18 @@
2828
from vllm.config import CacheConfig, get_current_vllm_config
2929
from vllm.forward_context import ForwardContext, get_forward_context
3030
from vllm.model_executor.layers.quantization import QuantizationConfig
31-
from vllm.utils import direct_register_custom_op
3231

3332
from vllm_ascend.utils import vllm_version_is
3433

3534
if vllm_version_is("0.11.0"):
3635
from vllm.attention import Attention
3736
from vllm.model_executor.layers.mla import \
3837
MultiHeadLatentAttention as MultiHeadLatentAttentionWrapper
38+
from vllm.utils import direct_register_custom_op
3939
else:
4040
from vllm.attention.layer import MLAAttention
4141
from vllm.model_executor.layers.mla import MultiHeadLatentAttentionWrapper
42+
from vllm.utils.torch_utils import direct_register_custom_op
4243

4344

4445
@dataclass

vllm_ascend/models/layers/sfa.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -29,17 +29,18 @@
2929
from vllm.forward_context import ForwardContext, get_forward_context
3030
from vllm.model_executor.layers.linear import ReplicatedLinear
3131
from vllm.model_executor.layers.quantization import QuantizationConfig
32-
from vllm.utils import direct_register_custom_op
3332

3433
from vllm_ascend.utils import vllm_version_is
3534

3635
if vllm_version_is("0.11.0"):
3736
from vllm.attention import Attention
3837
from vllm.model_executor.layers.mla import \
3938
MultiHeadLatentAttention as MultiHeadLatentAttentionWrapper
39+
from vllm.utils import direct_register_custom_op
4040
else:
4141
from vllm.attention.layer import MLAAttention
4242
from vllm.model_executor.layers.mla import MultiHeadLatentAttentionWrapper
43+
from vllm.utils.torch_utils import direct_register_custom_op
4344

4445

4546
@dataclass

vllm_ascend/ops/register_custom_ops.py

Lines changed: 7 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -8,12 +8,17 @@
88
tensor_model_parallel_all_reduce,
99
tensor_model_parallel_reduce_scatter)
1010
from vllm.forward_context import get_forward_context
11-
from vllm.utils import direct_register_custom_op
1211

1312
import vllm_ascend.envs as envs_ascend
1413
from vllm_ascend.ascend_forward_context import MoECommType
1514
from vllm_ascend.ops.weight_prefetch import maybe_npu_prefetch
16-
from vllm_ascend.utils import npu_stream_switch, prefetch_stream
15+
from vllm_ascend.utils import (npu_stream_switch, prefetch_stream,
16+
vllm_version_is)
17+
18+
if vllm_version_is("0.11.0"):
19+
from vllm.utils import direct_register_custom_op
20+
else:
21+
from vllm.utils.torch_utils import direct_register_custom_op
1722

1823

1924
def _maybe_chunk_residual_impl(x: torch.Tensor,

vllm_ascend/patch/platform/patch_common/patch_mamba_config.py

Lines changed: 8 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,9 +3,16 @@
33
from vllm.logger import init_logger
44
from vllm.model_executor.models import ModelRegistry
55
from vllm.model_executor.models.config import MambaModelConfig
6-
from vllm.utils import STR_DTYPE_TO_TORCH_DTYPE, cdiv
6+
from vllm.utils import cdiv
77
from vllm.v1.kv_cache_interface import FullAttentionSpec, MambaSpec
88

9+
from vllm_ascend.utils import vllm_version_is
10+
11+
if vllm_version_is("0.11.0"):
12+
from vllm.utils import STR_DTYPE_TO_TORCH_DTYPE
13+
else:
14+
from vllm.utils.torch_utils import STR_DTYPE_TO_TORCH_DTYPE
15+
916

1017
@classmethod
1118
def verify_and_update_config(cls, vllm_config) -> None:

vllm_ascend/patch/worker/patch_common/patch_weight_loader.py

Lines changed: 7 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,13 @@
33
from vllm.logger import init_logger
44
from vllm.model_executor.layers.linear import UnquantizedLinearMethod
55
from vllm.model_executor.utils import set_weight_attrs
6-
from vllm.utils import GiB_bytes
6+
7+
from vllm_ascend.utils import vllm_version_is
8+
9+
if vllm_version_is("0.11.0"):
10+
from vllm.utils import GiB_bytes
11+
else:
12+
from vllm.utils.mem_constants import GiB_bytes
713

814
logger = init_logger(__name__)
915

0 commit comments

Comments
 (0)