Skip to content

Commit 8108ac8

Browse files
authored
[Bugfix] Fix unsupported FA version check for Turing GPU (#12828)
1 parent afe74f7 commit 8108ac8

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

vllm/attention/backends/utils.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -612,5 +612,5 @@ def flash_attn_version():
612612
return fa_version
613613

614614
VLLM_FLASH_ATTN_VERSION = flash_attn_version()
615-
except ImportError:
615+
except (ImportError, AssertionError):
616616
VLLM_FLASH_ATTN_VERSION = None

0 commit comments

Comments
 (0)