Skip to content

Commit f4de631

Browse files
committed
fix MHA backend selection on rocm platform
Signed-off-by: tjtanaa <tunjian.tan@embeddedllm.com>
1 parent 4f19e55 commit f4de631

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

vllm/attention/layer.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -308,7 +308,7 @@ def __init__(
308308
backend = backend_name_to_enum(attn_backend.get_name())
309309
if current_platform.is_rocm():
310310
# currently, only torch_sdpa is supported on rocm
311-
backend = _Backend.TORCH_SDPA
311+
self.attn_backend = _Backend.TORCH_SDPA
312312
else:
313313
if backend in {_Backend.FLASH_ATTN, _Backend.FLASH_ATTN_VLLM_V1}:
314314
backend = _Backend.XFORMERS

0 commit comments

Comments
 (0)