We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 9c6c81d commit 92709f7Copy full SHA for 92709f7
vllm/v1/attention/backends/flash_attn.py
@@ -7,8 +7,8 @@
7
import numpy as np
8
import torch
9
10
-from vllm import envs
11
from vllm import _custom_ops as ops
+from vllm import envs
12
from vllm.attention.backends.abstract import (AttentionBackend, AttentionImpl,
13
AttentionMetadata, AttentionType,
14
is_quantized_kv_cache)
0 commit comments