-
-
Notifications
You must be signed in to change notification settings - Fork 11.2k
[CI/Build] Enable some fixed tests in AMD CI #28078
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changes from all commits
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -48,8 +48,8 @@ steps: | |
| commands: | ||
| - bash standalone_tests/pytorch_nightly_dependency.sh | ||
|
|
||
| - label: Async Engine, Inputs, Utils, Worker Test # 36min | ||
| timeout_in_minutes: 50 | ||
| - label: Async Engine, Inputs, Utils, Worker Test # 10min | ||
| timeout_in_minutes: 15 | ||
| mirror_hardwares: [amdexperimental, amdproduction] | ||
| agent_pool: mi325_1 | ||
| # grade: Blocking | ||
|
|
@@ -616,9 +616,9 @@ steps: | |
| - uv pip install --system torchao==0.13.0 | ||
| - VLLM_TEST_FORCE_LOAD_FORMAT=auto pytest -v -s quantization/ --ignore quantization/test_blackwell_moe.py | ||
|
|
||
| - label: LM Eval Small Models # 53min | ||
| timeout_in_minutes: 75 | ||
| mirror_hardwares: [amdexperimental] | ||
| - label: LM Eval Small Models # 15min | ||
| timeout_in_minutes: 20 | ||
| mirror_hardwares: [amdexperimental, amdproduction] | ||
|
Comment on lines
+619
to
+621
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more.
LM Eval Small Models was previously documented to run for ~53 minutes with a 75‑minute timeout. The new configuration reduces the timeout to 20 minutes without altering the invoked pytest command. Unless the test workload was dramatically reduced elsewhere, this will cause deterministic timeouts when enabling the job on AMD production hardware. Useful? React with 👍 / 👎. |
||
| agent_pool: mi325_1 | ||
| # grade: Blocking | ||
| source_file_dependencies: | ||
|
|
@@ -627,8 +627,8 @@ steps: | |
| commands: | ||
| - pytest -s -v evals/gsm8k/test_gsm8k_correctness.py --config-list-file=configs/models-small.txt --tp-size=1 | ||
|
|
||
| - label: OpenAI API correctness # 22min | ||
| timeout_in_minutes: 30 | ||
| - label: OpenAI API correctness # 10min | ||
| timeout_in_minutes: 15 | ||
| mirror_hardwares: [amdexperimental, amdproduction] | ||
|
Comment on lines
+630
to
632
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more.
The OpenAI API correctness step ran in ~22 minutes and previously had a 30‑minute timeout. The change lowers the timeout to 15 minutes while the step continues to run the same Useful? React with 👍 / 👎. |
||
| agent_pool: mi325_1 | ||
| # grade: Blocking | ||
|
|
@@ -859,10 +859,10 @@ steps: | |
| - pytest -v -s models/multimodal -m core_model --ignore models/multimodal/generation/test_whisper.py --ignore models/multimodal/processing | ||
| - cd .. && VLLM_WORKER_MULTIPROC_METHOD=spawn pytest -v -s tests/models/multimodal/generation/test_whisper.py -m core_model # Otherwise, mp_method="spawn" doesn't work | ||
|
|
||
| - label: Multi-Modal Accuracy Eval (Small Models) # 50min | ||
| mirror_hardwares: [amdexperimental] | ||
| - label: Multi-Modal Accuracy Eval (Small Models) # 10min | ||
| mirror_hardwares: [amdexperimental, amdproduction] | ||
| agent_pool: mi325_1 | ||
| timeout_in_minutes: 70 | ||
| timeout_in_minutes: 15 | ||
|
Comment on lines
+862
to
+865
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more.
This step previously recorded a ~50 minute runtime with a 70‑minute timeout. It now expects 10 minutes and times out after 15 minutes while executing the same Useful? React with 👍 / 👎. |
||
| working_dir: "/vllm-workspace/.buildkite/lm-eval-harness" | ||
| source_file_dependencies: | ||
| - vllm/multimodal/ | ||
|
|
||
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This step previously carried a 36‑minute runtime annotation and a 50‑minute timeout. The change drops both to 10 and 15 minutes while the commands still execute the full
multimodalandutils_pytest suites. Nothing in this commit reduces the workload, so on current hardware the job will time out before completion and consistently fail AMD CI.Useful? React with 👍 / 👎.