Skip to content

Commit 1c6c6aa

Browse files
try remove version check
Signed-off-by: cty <ctynb@qq.com>
1 parent 4e2e4d6 commit 1c6c6aa

File tree

1 file changed

+10
-10
lines changed

1 file changed

+10
-10
lines changed

vllm_ascend/worker/model_runner_v1.py

Lines changed: 10 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -1214,16 +1214,16 @@ def initialize_kv_cache(self, kv_cache_config: KVCacheConfig) -> None:
12141214
"""
12151215
import torch_npu
12161216
kv_caches: Dict[str, torch.Tensor] = {}
1217-
if not (vllm_version_is("0.8.5") or vllm_version_is("0.8.5.post1")):
1218-
self.input_batch = InputBatch(
1219-
max_num_reqs=self.max_num_reqs,
1220-
max_model_len=self.model_config.max_model_len,
1221-
max_num_batched_tokens=self.max_num_tokens,
1222-
device=self.device,
1223-
pin_memory=True,
1224-
vocab_size=self.model_config.get_vocab_size(),
1225-
kv_cache_config=kv_cache_config,
1226-
)
1217+
# if not (vllm_version_is("0.8.5") or vllm_version_is("0.8.5.post1")):
1218+
# self.input_batch = InputBatch(
1219+
# max_num_reqs=self.max_num_reqs,
1220+
# max_model_len=self.model_config.max_model_len,
1221+
# max_num_batched_tokens=self.max_num_tokens,
1222+
# device=self.device,
1223+
# pin_memory=True,
1224+
# vocab_size=self.model_config.get_vocab_size(),
1225+
# kv_cache_config=kv_cache_config,
1226+
# )
12271227

12281228
for kv_cache_group in kv_cache_config.kv_cache_groups:
12291229
kv_cache_spec = kv_cache_group.kv_cache_spec

0 commit comments

Comments
 (0)