We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 4f19e55 commit f4de631Copy full SHA for f4de631
vllm/attention/layer.py
@@ -308,7 +308,7 @@ def __init__(
308
backend = backend_name_to_enum(attn_backend.get_name())
309
if current_platform.is_rocm():
310
# currently, only torch_sdpa is supported on rocm
311
- backend = _Backend.TORCH_SDPA
+ self.attn_backend = _Backend.TORCH_SDPA
312
else:
313
if backend in {_Backend.FLASH_ATTN, _Backend.FLASH_ATTN_VLLM_V1}:
314
backend = _Backend.XFORMERS
0 commit comments