Skip to content

Commit 3733339

Browse files
committed
move seed check to platform specific validate
Signed-off-by: NickLucche <nlucches@redhat.com>
1 parent 42aaca7 commit 3733339

File tree

2 files changed

+4
-6
lines changed

2 files changed

+4
-6
lines changed

vllm/platforms/tpu.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@
77
import vllm.envs as envs
88
from vllm.inputs import PromptType
99
from vllm.logger import init_logger
10-
from vllm.sampling_params import SamplingParams
10+
from vllm.sampling_params import SamplingParams, SamplingType
1111

1212
from .interface import Platform, PlatformEnum, _Backend
1313

@@ -149,3 +149,5 @@ def validate_request(
149149
SamplingParams) and params.guided_decoding is not None:
150150
raise ValueError("Structured output is not supported on "
151151
f"{cls.device_name}.")
152+
if params.sampling_type == SamplingType.RANDOM_SEED:
153+
raise ValueError("Torch XLA does not support per-request seed.")

vllm/v1/engine/processor.py

Lines changed: 1 addition & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -14,10 +14,9 @@
1414
from vllm.multimodal.inputs import PlaceholderRange
1515
from vllm.multimodal.processing import EncDecMultiModalProcessor
1616
from vllm.multimodal.utils import merge_and_sort_multimodal_metadata
17-
from vllm.platforms import current_platform
1817
from vllm.pooling_params import PoolingParams
1918
from vllm.prompt_adapter.request import PromptAdapterRequest
20-
from vllm.sampling_params import SamplingParams, SamplingType
19+
from vllm.sampling_params import SamplingParams
2120
from vllm.transformers_utils.tokenizer_group import BaseTokenizerGroup
2221
from vllm.v1.engine import EngineCoreRequest
2322
from vllm.v1.engine.mm_input_cache import MirroredProcessingCache
@@ -78,9 +77,6 @@ def _validate_sampling_params(
7877
params: SamplingParams,
7978
) -> None:
8079
self._validate_structured_output(params)
81-
if (current_platform.is_tpu()
82-
and params.sampling_type == SamplingType.RANDOM_SEED):
83-
raise ValueError("Torch XLA does not support per-request seed.")
8480

8581
if params.allowed_token_ids is None:
8682
return

0 commit comments

Comments
 (0)