Skip to content

Commit e70cc7b

Browse files
committed
fix full graph
Signed-off-by: ganyi <ygan@amd.com>
1 parent 40f896e commit e70cc7b

File tree

7 files changed

+307
-440
lines changed

7 files changed

+307
-440
lines changed

tests/v1/attention/test_batch_reordering.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -43,7 +43,7 @@ class ReorderTestCase:
4343
expected_modified=False,
4444
),
4545
"all_prefills": ReorderTestCase(
46-
requests=[(100, 100), (200, 200), (300, 300)],
46+
requests=[(100, 0), (200, 0), (300, 0)],
4747
expected_order=[0, 1, 2],
4848
expected_modified=False,
4949
),

vllm/config/scheduler.py

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -137,10 +137,6 @@ class SchedulerConfig:
137137
structured outputs, speculative decoding, and pipeline parallelism.
138138
"""
139139

140-
split_prefill_from_chunk: bool = False
141-
"""Whether to split the prefill request into pure prefill and chunked
142-
prefill in a single batch."""
143-
144140
def compute_hash(self) -> str:
145141
"""
146142
WARNING: Whenever a new field is added to this config,

vllm/platforms/rocm.py

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -391,11 +391,6 @@ def check_and_update_config(cls, vllm_config: "VllmConfig") -> None:
391391
):
392392
compilation_config.custom_ops.append("+rms_norm")
393393

394-
if envs.VLLM_ROCM_USE_AITER and envs.VLLM_ROCM_USE_AITER_MHA:
395-
# enable the request reorder if we are using AITER MHA
396-
# for calculation
397-
vllm_config.scheduler_config.split_prefill_from_chunk = True
398-
399394
@classmethod
400395
def verify_model_arch(cls, model_arch: str) -> None:
401396
if model_arch in _ROCM_UNSUPPORTED_MODELS:

0 commit comments

Comments
 (0)