Skip to content

Conversation

@aarnphm
Copy link
Collaborator

@aarnphm aarnphm commented Sep 6, 2025

This PR re-enables ministral tests for structured outputs with fixed seed in the hope to avoid flakiness in CI runs.

Signed-off-by: Aaron Pham contact@aarnphm.xyz

Signed-off-by: Aaron Pham <contact@aarnphm.xyz>
@aarnphm aarnphm added the ready ONLY add when PR is ready to merge/full CI is needed label Sep 6, 2025
@mergify mergify bot added the v1 label Sep 6, 2025
Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request addresses flaky CI runs by introducing a deterministic seed for tests that use structured outputs. The change re-enables a previously flaky test and adds a seed to the LLM constructor within test_structured_output. This is a good improvement for test stability. My review suggests extending this fix to other non-deterministic tests in the same file to prevent future flakiness and ensure comprehensive CI reliability.

guided_decoding_backend=guided_decoding_backend,
guided_decoding_disable_any_whitespace=(guided_decoding_backend
in {"xgrammar", "guidance"}),
seed=120,
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

high

While adding a seed here helps make this test deterministic, the fix for CI flakiness seems incomplete. Other tests in this file also use non-deterministic sampling parameters but are not seeded, which could lead to continued flakiness.

Specifically, the LLM initializations in the following tests should also have a deterministic seed (e.g., seed=120):

  • test_structured_output_with_reasoning_matrices
  • test_structured_output_auto_mode
  • test_structured_output_batched_with_non_guided_requests

Applying this change consistently will help ensure CI stability.

@DarkLight1337
Copy link
Member

Thanks for fixing!

@DarkLight1337 DarkLight1337 merged commit e675975 into vllm-project:main Sep 7, 2025
26 of 28 checks passed
@aarnphm aarnphm deleted the fix/flaky-structured-outputs-tessts branch September 7, 2025 14:37
eicherseiji pushed a commit to eicherseiji/vllm that referenced this pull request Sep 9, 2025
ChrisYangAI pushed a commit to RichardoMrMu/vllm that referenced this pull request Sep 10, 2025
skyloevil pushed a commit to skyloevil/vllm that referenced this pull request Sep 13, 2025
FeiDaLI pushed a commit to FeiDaLI/vllm that referenced this pull request Sep 25, 2025
xuebwang-amd pushed a commit to xuebwang-amd/vllm that referenced this pull request Oct 10, 2025
…llm-project#24380)

Signed-off-by: Aaron Pham <contact@aarnphm.xyz>
Signed-off-by: xuebwang-amd <xuebwang@amd.com>
xuebwang-amd pushed a commit to xuebwang-amd/vllm that referenced this pull request Oct 24, 2025
…llm-project#24380)

Signed-off-by: Aaron Pham <contact@aarnphm.xyz>
Signed-off-by: xuebwang-amd <xuebwang@amd.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ready ONLY add when PR is ready to merge/full CI is needed v1

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants