From d90916f4893427fd4cad58fee1788d547cdcf3ca Mon Sep 17 00:00:00 2001 From: zhewenli Date: Tue, 4 Nov 2025 15:28:05 -0800 Subject: [PATCH 1/2] update Signed-off-by: zhewenli --- .buildkite/test-amd.yaml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/.buildkite/test-amd.yaml b/.buildkite/test-amd.yaml index 5abf6122a5c3..6f361f3c93d8 100644 --- a/.buildkite/test-amd.yaml +++ b/.buildkite/test-amd.yaml @@ -618,7 +618,7 @@ steps: - label: LM Eval Small Models # 53min timeout_in_minutes: 75 - mirror_hardwares: [amdexperimental] + mirror_hardwares: [amdexperimental, amdproduction] agent_pool: mi325_1 # grade: Blocking source_file_dependencies: @@ -860,7 +860,7 @@ steps: - cd .. && VLLM_WORKER_MULTIPROC_METHOD=spawn pytest -v -s tests/models/multimodal/generation/test_whisper.py -m core_model # Otherwise, mp_method="spawn" doesn't work - label: Multi-Modal Accuracy Eval (Small Models) # 50min - mirror_hardwares: [amdexperimental] + mirror_hardwares: [amdexperimental, amdproduction] agent_pool: mi325_1 timeout_in_minutes: 70 working_dir: "/vllm-workspace/.buildkite/lm-eval-harness" From fcfae2786ccae431eab29e02f05828b4e1278705 Mon Sep 17 00:00:00 2001 From: zhewenli Date: Tue, 4 Nov 2025 17:35:39 -0800 Subject: [PATCH 2/2] update time Signed-off-by: zhewenli --- .buildkite/test-amd.yaml | 16 ++++++++-------- 1 file changed, 8 insertions(+), 8 deletions(-) diff --git a/.buildkite/test-amd.yaml b/.buildkite/test-amd.yaml index 6f361f3c93d8..d61c3136f3d7 100644 --- a/.buildkite/test-amd.yaml +++ b/.buildkite/test-amd.yaml @@ -48,8 +48,8 @@ steps: commands: - bash standalone_tests/pytorch_nightly_dependency.sh -- label: Async Engine, Inputs, Utils, Worker Test # 36min - timeout_in_minutes: 50 +- label: Async Engine, Inputs, Utils, Worker Test # 10min + timeout_in_minutes: 15 mirror_hardwares: [amdexperimental, amdproduction] agent_pool: mi325_1 # grade: Blocking @@ -616,8 +616,8 @@ steps: - uv pip install --system torchao==0.13.0 - VLLM_TEST_FORCE_LOAD_FORMAT=auto pytest -v -s quantization/ --ignore quantization/test_blackwell_moe.py -- label: LM Eval Small Models # 53min - timeout_in_minutes: 75 +- label: LM Eval Small Models # 15min + timeout_in_minutes: 20 mirror_hardwares: [amdexperimental, amdproduction] agent_pool: mi325_1 # grade: Blocking @@ -627,8 +627,8 @@ steps: commands: - pytest -s -v evals/gsm8k/test_gsm8k_correctness.py --config-list-file=configs/models-small.txt --tp-size=1 -- label: OpenAI API correctness # 22min - timeout_in_minutes: 30 +- label: OpenAI API correctness # 10min + timeout_in_minutes: 15 mirror_hardwares: [amdexperimental, amdproduction] agent_pool: mi325_1 # grade: Blocking @@ -859,10 +859,10 @@ steps: - pytest -v -s models/multimodal -m core_model --ignore models/multimodal/generation/test_whisper.py --ignore models/multimodal/processing - cd .. && VLLM_WORKER_MULTIPROC_METHOD=spawn pytest -v -s tests/models/multimodal/generation/test_whisper.py -m core_model # Otherwise, mp_method="spawn" doesn't work -- label: Multi-Modal Accuracy Eval (Small Models) # 50min +- label: Multi-Modal Accuracy Eval (Small Models) # 10min mirror_hardwares: [amdexperimental, amdproduction] agent_pool: mi325_1 - timeout_in_minutes: 70 + timeout_in_minutes: 15 working_dir: "/vllm-workspace/.buildkite/lm-eval-harness" source_file_dependencies: - vllm/multimodal/