We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 3483a86 commit eabca2eCopy full SHA for eabca2e
vllm/attention/layers/cross_attention.py
@@ -6,14 +6,13 @@
6
7
import numpy as np
8
import torch
9
-from transformers import CacheConfig
10
11
from vllm import envs
12
from vllm.attention.backends.abstract import (AttentionBackend,
13
AttentionMetadata, AttentionType)
14
from vllm.attention.layer import Attention
15
from vllm.attention.selector import get_attn_backend
16
-from vllm.config import VllmConfig
+from vllm.config import CacheConfig, VllmConfig
17
from vllm.logger import init_logger
18
from vllm.multimodal import MULTIMODAL_REGISTRY
19
from vllm.utils import cdiv
0 commit comments