We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent b8f17f5 commit 65197a5Copy full SHA for 65197a5
vllm/attention/layers/encoder_only_attention.py
@@ -5,13 +5,13 @@
5
from typing import Optional
6
7
import torch
8
-from transformers import CacheConfig
9
10
from vllm import envs
11
from vllm.attention.backends.abstract import (AttentionBackend,
12
AttentionMetadata, AttentionType)
13
from vllm.attention.layer import Attention
14
from vllm.attention.selector import get_attn_backend
+from vllm.config import CacheConfig
15
from vllm.v1.attention.backends.utils import (CommonAttentionMetadata,
16
subclass_attention_backend)
17
0 commit comments