Skip to content

Commit 7d23397

Browse files
fxmarty-amdamitm02
authored andcommitted
[rocm] Fix wrong attention log (vllm-project#18764)
Signed-off-by: Felix Marty <felmarty@amd.com> Signed-off-by: amit <amit.man@gmail.com>
1 parent 5397cda commit 7d23397

File tree

1 file changed

+3
-2
lines changed

1 file changed

+3
-2
lines changed

vllm/platforms/rocm.py

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -194,8 +194,9 @@ def get_attn_backend_cls(cls, selected_backend, head_size, dtype,
194194
f" The selected backend, {selected_backend.name},"
195195
f"is not MLA type while requested for MLA backend.")
196196

197-
selected_backend = (_Backend.ROCM_FLASH if selected_backend
198-
== _Backend.FLASH_ATTN else selected_backend)
197+
if selected_backend is None or selected_backend == _Backend.FLASH_ATTN:
198+
selected_backend = _Backend.ROCM_FLASH
199+
199200
if envs.VLLM_USE_V1:
200201
logger.info("Using Triton Attention backend on V1 engine.")
201202
return ("vllm.v1.attention.backends."

0 commit comments

Comments
 (0)