Skip to content

Commit c4afdb6

Browse files
Move MultiModalConfig from config/__init__.py to config/multimodal.py (#24659)
Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com> Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com>
1 parent b834b4c commit c4afdb6

File tree

11 files changed

+229
-247
lines changed

11 files changed

+229
-247
lines changed

tests/entrypoints/openai/test_lora_resolvers.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@
99

1010
import pytest
1111

12-
from vllm.config import MultiModalConfig
12+
from vllm.config.multimodal import MultiModalConfig
1313
from vllm.engine.multiprocessing.client import MQLLMEngineClient
1414
from vllm.entrypoints.openai.protocol import CompletionRequest, ErrorResponse
1515
from vllm.entrypoints.openai.serving_completion import OpenAIServingCompletion

tests/entrypoints/openai/test_serving_chat.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@
1212
import pytest
1313
import pytest_asyncio
1414

15-
from vllm.config import MultiModalConfig
15+
from vllm.config.multimodal import MultiModalConfig
1616
from vllm.engine.multiprocessing.client import MQLLMEngineClient
1717
from vllm.entrypoints.openai.protocol import ChatCompletionRequest
1818
from vllm.entrypoints.openai.serving_chat import OpenAIServingChat

tests/multimodal/test_cache.py

Lines changed: 12 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -7,6 +7,7 @@
77
import torch
88

99
from vllm.config import ModelConfig, ParallelConfig, VllmConfig
10+
from vllm.multimodal import MULTIMODAL_REGISTRY
1011
from vllm.multimodal.cache import (MultiModalCache,
1112
MultiModalProcessorCacheItem,
1213
MultiModalProcessorCacheItemMetadata,
@@ -17,7 +18,6 @@
1718
MultiModalKwargsItems,
1819
MultiModalSharedField)
1920
from vllm.multimodal.processing import PromptInsertion
20-
from vllm.multimodal.registry import MultiModalRegistry
2121

2222

2323
def _dummy_elem(
@@ -96,7 +96,9 @@ def _create_vllm_config(
9696
enable_ipc: bool,
9797
):
9898
return VllmConfig(
99-
model_config=ModelConfig(mm_processor_cache_gb=mm_processor_cache_gb),
99+
model_config=ModelConfig(
100+
model="llava-hf/llava-onevision-qwen2-0.5b-ov-hf",
101+
mm_processor_cache_gb=mm_processor_cache_gb),
100102
parallel_config=ParallelConfig(
101103
data_parallel_size=1 if enable_ipc else 2),
102104
)
@@ -113,15 +115,16 @@ def _compare_caches(
113115
n_iter: int = 100,
114116
seed: int = 0,
115117
):
116-
mm_registry = MultiModalRegistry()
117-
cache_0_p0 = processor_cache_from_config(config_0, mm_registry)
118-
cache_0_p1 = engine_receiver_cache_from_config(config_0, mm_registry)
119-
cache_1_p0 = processor_cache_from_config(config_1, mm_registry)
120-
cache_1_p1 = engine_receiver_cache_from_config(config_1, mm_registry)
118+
cache_0_p0 = processor_cache_from_config(config_0, MULTIMODAL_REGISTRY)
119+
cache_0_p1 = engine_receiver_cache_from_config(config_0,
120+
MULTIMODAL_REGISTRY)
121+
cache_1_p0 = processor_cache_from_config(config_1, MULTIMODAL_REGISTRY)
122+
cache_1_p1 = engine_receiver_cache_from_config(config_1,
123+
MULTIMODAL_REGISTRY)
121124

122125
cache_size_gb = max(
123-
config_0.model_config.mm_processor_cache_gb,
124-
config_1.model_config.mm_processor_cache_gb,
126+
config_0.model_config.multimodal_config.mm_processor_cache_gb,
127+
config_1.model_config.multimodal_config.mm_processor_cache_gb,
125128
)
126129
item_size_gb = int(cache_size_gb / item_capacity)
127130

tests/test_config.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -6,9 +6,9 @@
66
import pytest
77

88
from vllm.compilation.backends import VllmBackend
9-
from vllm.config import (ModelConfig, PoolerConfig, VllmConfig, get_field,
10-
update_config)
9+
from vllm.config import ModelConfig, PoolerConfig, VllmConfig, update_config
1110
from vllm.config.load import LoadConfig
11+
from vllm.config.utils import get_field
1212
from vllm.model_executor.layers.pooler import PoolingType
1313
from vllm.platforms import current_platform
1414

tests/v1/engine/test_processor_multi_modal_uuids.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,7 @@ def _mk_processor(monkeypatch,
3131
raising=True)
3232
monkeypatch.setattr(ModelConfig,
3333
"__post_init__",
34-
lambda self: None,
34+
lambda self, *args: None,
3535
raising=True)
3636
monkeypatch.setattr(UnspecifiedPlatform,
3737
"is_async_output_supported",

0 commit comments

Comments
 (0)