Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion tests/lora/test_add_lora.py
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,7 @@
from vllm.inputs import TextPrompt
from vllm.lora.request import LoRARequest
from vllm.sampling_params import SamplingParams
from vllm.utils.async_utils import merge_async_iterators
from vllm.utils.asyncio import merge_async_iterators

MODEL_PATH = "zai-org/chatglm3-6b"
LORA_RANK = 64
Expand Down
2 changes: 1 addition & 1 deletion tests/models/multimodal/generation/test_common.py
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@
)

from vllm.platforms import current_platform
from vllm.utils.func import identity
from vllm.utils.functools import identity

from ....conftest import (
IMAGE_ASSETS,
Expand Down
2 changes: 1 addition & 1 deletion tests/utils_/test_async_utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@

import pytest

from vllm.utils.async_utils import merge_async_iterators
from vllm.utils.asyncio import merge_async_iterators


async def _mock_async_iterator(idx: int):
Expand Down
2 changes: 1 addition & 1 deletion tests/utils_/test_func_utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@

import pytest

from vllm.utils.func import deprecate_kwargs, supports_kw
from vllm.utils.functools import deprecate_kwargs, supports_kw

from ..utils import error_on_warning

Expand Down
2 changes: 1 addition & 1 deletion vllm/benchmarks/throughput.py
Original file line number Diff line number Diff line change
Expand Up @@ -34,7 +34,7 @@
from vllm.lora.request import LoRARequest
from vllm.outputs import RequestOutput
from vllm.sampling_params import BeamSearchParams
from vllm.utils.async_utils import merge_async_iterators
from vllm.utils.asyncio import merge_async_iterators


def run_vllm(
Expand Down
2 changes: 1 addition & 1 deletion vllm/entrypoints/chat_utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -51,7 +51,7 @@
from vllm.transformers_utils.processor import cached_get_processor
from vllm.transformers_utils.tokenizer import AnyTokenizer, MistralTokenizer
from vllm.utils import random_uuid
from vllm.utils.func import supports_kw
from vllm.utils.functools import supports_kw

logger = init_logger(__name__)

Expand Down
2 changes: 1 addition & 1 deletion vllm/entrypoints/openai/serving_completion.py
Original file line number Diff line number Diff line change
Expand Up @@ -35,7 +35,7 @@
from vllm.sampling_params import BeamSearchParams, SamplingParams
from vllm.transformers_utils.tokenizer import AnyTokenizer
from vllm.utils import as_list
from vllm.utils.async_utils import merge_async_iterators
from vllm.utils.asyncio import merge_async_iterators

logger = init_logger(__name__)

Expand Down
2 changes: 1 addition & 1 deletion vllm/entrypoints/openai/serving_embedding.py
Original file line number Diff line number Diff line change
Expand Up @@ -40,7 +40,7 @@
)
from vllm.pooling_params import PoolingParams
from vllm.utils import chunk_list
from vllm.utils.async_utils import merge_async_iterators
from vllm.utils.asyncio import merge_async_iterators

logger = init_logger(__name__)

Expand Down
2 changes: 1 addition & 1 deletion vllm/entrypoints/openai/serving_engine.py
Original file line number Diff line number Diff line change
Expand Up @@ -91,7 +91,7 @@
)
from vllm.transformers_utils.tokenizer import AnyTokenizer, MistralTokenizer
from vllm.utils import is_list_of, random_uuid
from vllm.utils.async_utils import (
from vllm.utils.asyncio import (
AsyncMicrobatchTokenizer,
collect_from_async_generator,
make_async,
Expand Down
2 changes: 1 addition & 1 deletion vllm/entrypoints/openai/serving_pooling.py
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,7 @@
from vllm.logger import init_logger
from vllm.outputs import PoolingOutput, PoolingRequestOutput
from vllm.tasks import SupportedTask
from vllm.utils.async_utils import merge_async_iterators
from vllm.utils.asyncio import merge_async_iterators

logger = init_logger(__name__)

Expand Down
2 changes: 1 addition & 1 deletion vllm/entrypoints/openai/serving_score.py
Original file line number Diff line number Diff line change
Expand Up @@ -37,7 +37,7 @@
from vllm.lora.request import LoRARequest
from vllm.outputs import PoolingRequestOutput, ScoringRequestOutput
from vllm.transformers_utils.tokenizer import AnyTokenizer, MistralTokenizer
from vllm.utils.async_utils import make_async, merge_async_iterators
from vllm.utils.asyncio import make_async, merge_async_iterators

logger = init_logger(__name__)

Expand Down
2 changes: 1 addition & 1 deletion vllm/entrypoints/renderer.py
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@
from vllm.inputs.data import TokensPrompt as EngineTokensPrompt
from vllm.inputs.parse import get_prompt_components, parse_raw_prompts
from vllm.transformers_utils.tokenizer import AnyTokenizer
from vllm.utils.async_utils import AsyncMicrobatchTokenizer
from vllm.utils.asyncio import AsyncMicrobatchTokenizer


@dataclass(frozen=True)
Expand Down
2 changes: 1 addition & 1 deletion vllm/executor/executor_base.py
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@
from vllm.lora.request import LoRARequest
from vllm.sequence import ExecuteModelRequest
from vllm.tasks import SupportedTask
from vllm.utils.async_utils import make_async
from vllm.utils.asyncio import make_async
from vllm.v1.outputs import SamplerOutput
from vllm.v1.worker.worker_base import WorkerBase

Expand Down
2 changes: 1 addition & 1 deletion vllm/executor/ray_distributed_executor.py
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,7 @@
get_ip,
get_open_port,
)
from vllm.utils.async_utils import make_async
from vllm.utils.asyncio import make_async
from vllm.v1.outputs import SamplerOutput

if ray is not None:
Expand Down
2 changes: 1 addition & 1 deletion vllm/model_executor/layers/fused_moe/deep_gemm_moe.py
Original file line number Diff line number Diff line change
Expand Up @@ -29,7 +29,7 @@
)
from vllm.utils import has_deep_gemm
from vllm.utils.deep_gemm import m_grouped_fp8_gemm_nt_contiguous
from vllm.utils.func import run_once
from vllm.utils.functools import run_once

logger = init_logger(__name__)

Expand Down
2 changes: 1 addition & 1 deletion vllm/model_executor/models/interfaces.py
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,7 @@
from vllm.inputs.data import PromptType
from vllm.logger import init_logger
from vllm.model_executor.layers.quantization import QuantizationConfig
from vllm.utils.func import supports_kw
from vllm.utils.functools import supports_kw

from .interfaces_base import VllmModel, is_pooling_model

Expand Down
2 changes: 1 addition & 1 deletion vllm/model_executor/models/interfaces_base.py
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@
from typing_extensions import TypeIs, TypeVar

from vllm.logger import init_logger
from vllm.utils.func import supports_kw
from vllm.utils.functools import supports_kw

if TYPE_CHECKING:
from vllm.config import VllmConfig
Expand Down
2 changes: 1 addition & 1 deletion vllm/multimodal/processing.py
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,7 @@
from vllm.transformers_utils.processor import cached_processor_from_config
from vllm.transformers_utils.tokenizer import AnyTokenizer, decode_tokens, encode_tokens
from vllm.utils import flatten_2d_lists, full_groupby
from vllm.utils.func import get_allowed_kwarg_only_overrides
from vllm.utils.functools import get_allowed_kwarg_only_overrides
from vllm.utils.jsontree import JSONTree, json_map_leaves

from .hasher import MultiModalHasher
Expand Down
2 changes: 1 addition & 1 deletion vllm/tracing.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
from collections.abc import Mapping

from vllm.logger import init_logger
from vllm.utils.func import run_once
from vllm.utils.functools import run_once

TRACE_HEADERS = ["traceparent", "tracestate"]

Expand Down
2 changes: 1 addition & 1 deletion vllm/transformers_utils/processor.py
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,7 @@
from transformers.video_processing_utils import BaseVideoProcessor
from typing_extensions import TypeVar

from vllm.utils.func import get_allowed_kwarg_only_overrides
from vllm.utils.functools import get_allowed_kwarg_only_overrides

if TYPE_CHECKING:
from vllm.config import ModelConfig
Expand Down
6 changes: 5 additions & 1 deletion vllm/utils/async_utils.py → vllm/utils/asyncio.py
Original file line number Diff line number Diff line change
@@ -1,6 +1,10 @@
# SPDX-License-Identifier: Apache-2.0
# SPDX-FileCopyrightText: Copyright contributors to the vLLM project
"""Contains helpers related to asynchronous code."""
"""
Contains helpers related to asynchronous code.

This is similar in concept to the `asyncio` module.
"""

import asyncio
import contextlib
Expand Down
File renamed without changes.
4 changes: 2 additions & 2 deletions vllm/v1/engine/async_llm.py
Original file line number Diff line number Diff line change
Expand Up @@ -30,8 +30,8 @@
from vllm.transformers_utils.tokenizer import AnyTokenizer, init_tokenizer_from_configs
from vllm.usage.usage_lib import UsageContext
from vllm.utils import Device, as_list, cdiv
from vllm.utils.async_utils import cancel_task_threadsafe
from vllm.utils.func import deprecate_kwargs
from vllm.utils.asyncio import cancel_task_threadsafe
from vllm.utils.functools import deprecate_kwargs
from vllm.v1.engine import EngineCoreRequest
from vllm.v1.engine.core_client import EngineCoreClient
from vllm.v1.engine.exceptions import EngineDeadError, EngineGenerateError
Expand Down
2 changes: 1 addition & 1 deletion vllm/v1/engine/core_client.py
Original file line number Diff line number Diff line change
Expand Up @@ -29,7 +29,7 @@
get_open_zmq_inproc_path,
make_zmq_socket,
)
from vllm.utils.async_utils import in_loop
from vllm.utils.asyncio import in_loop
from vllm.v1.engine import (
EngineCoreOutputs,
EngineCoreRequest,
Expand Down