Skip to content

Commit 982937a

Browse files
author
qqma
committed
resolve pre-commit test failure due to E501: line too long
Signed-off-by: qqma <qqma@amazon.com>
1 parent 6e64b12 commit 982937a

File tree

1 file changed

+2
-1
lines changed

1 file changed

+2
-1
lines changed

vllm/v1/attention/backends/mla/flashattn_mla.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -103,7 +103,8 @@ def __init__(self, kv_cache_spec: AttentionSpec, layer_names: list[str],
103103
"Getting flash attention max num splits for "
104104
"cuda graph from environment variable, value=%s",
105105
envs.VLLM_FLASH_ATTN_MAX_NUM_SPLITS_FOR_CUDA_GRAPH)
106-
self.max_num_splits = envs.VLLM_FLASH_ATTN_MAX_NUM_SPLITS_FOR_CUDA_GRAPH
106+
self.max_num_splits = (
107+
envs.VLLM_FLASH_ATTN_MAX_NUM_SPLITS_FOR_CUDA_GRAPH)
107108
else:
108109
self.max_num_splits = _DEFAULT_MAX_NUM_SPLITS_FOR_CUDA_GRAPH
109110

0 commit comments

Comments
 (0)