Skip to content

[ROCm][CI] Disable async scheduling on ROCm for test_structured_output[meta-llama/Meta-Llama-3.1-8B-Instruct-xgrammar-auto-speculative_config9]#32355

Merged
tjtanaa merged 3 commits intovllm-project:mainfrom
ROCm:micah/disable-async-20260114
Jan 15, 2026

Conversation

@micah-wil
Copy link
Contributor

@micah-wil micah-wil commented Jan 14, 2026

Problem

The issue was exposed in the V1 Test entrypoints test group in AMD CI after #31998 enabled async scheduling by default with spec decoding. The test group has been failing ever since that PR was merged(e.g. in build#2803). It passes if you set async_scheduling=False. The failure can be reproduced with this command:
pytest -v -s tests/v1/entrypoints/llm/test_struct_output_generate.py::test_structured_output[meta-llama/Meta-Llama-3.1-8B-Instruct-xgrammar-auto-speculative_config9].

To unblock AMD CI, I am proposing that we disable async scheduling for this test case until we are able to track down and fix the underlying issue.

Signed-off-by: Micah Williamson <micah.williamson@amd.com>
Signed-off-by: Micah Williamson <micah.williamson@amd.com>
@mergify mergify bot added llama Related to Llama models rocm Related to AMD ROCm structured-output labels Jan 14, 2026
@mergify mergify bot added the v1 label Jan 14, 2026
Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request introduces a temporary fix to disable asynchronous scheduling for a failing test on ROCm. The implementation disables it for all parameterizations of test_structured_output on ROCm. My review suggests making this fix more targeted to only the specific failing test case. This will help maintain test coverage for asynchronous scheduling on ROCm for other, potentially passing, test cases.

@micah-wil
Copy link
Contributor Author

Actively working on a fix in #32303, but adding this workaround in the meantime

@tjtanaa tjtanaa added the ready ONLY add when PR is ready to merge/full CI is needed label Jan 15, 2026
@tjtanaa tjtanaa enabled auto-merge (squash) January 15, 2026 03:21
@tjtanaa tjtanaa merged commit 773d707 into vllm-project:main Jan 15, 2026
23 checks passed
@micah-wil micah-wil deleted the micah/disable-async-20260114 branch January 15, 2026 05:01
sammysun0711 pushed a commit to sammysun0711/vllm that referenced this pull request Jan 16, 2026
…t[meta-llama/Meta-Llama-3.1-8B-Instruct-xgrammar-auto-speculative_config9] (vllm-project#32355)

Signed-off-by: Micah Williamson <micah.williamson@amd.com>
micah-wil added a commit to ROCm/vllm that referenced this pull request Jan 16, 2026
…ed_output[meta-llama/Meta-Llama-3.1-8B-Instruct-xgrammar-auto-speculative_config9] (vllm-project#32355)"

This reverts commit 773d707.

Signed-off-by: Micah Williamson <micah.williamson@amd.com>
akh64bit pushed a commit to akh64bit/vllm that referenced this pull request Jan 16, 2026
…t[meta-llama/Meta-Llama-3.1-8B-Instruct-xgrammar-auto-speculative_config9] (vllm-project#32355)

Signed-off-by: Micah Williamson <micah.williamson@amd.com>
dsuhinin pushed a commit to dsuhinin/vllm that referenced this pull request Jan 21, 2026
…t[meta-llama/Meta-Llama-3.1-8B-Instruct-xgrammar-auto-speculative_config9] (vllm-project#32355)

Signed-off-by: Micah Williamson <micah.williamson@amd.com>
Signed-off-by: dsuhinin <suhinin.dmitriy@gmail.com>
ItzDEXX pushed a commit to ItzDEXX/vllm that referenced this pull request Feb 19, 2026
…t[meta-llama/Meta-Llama-3.1-8B-Instruct-xgrammar-auto-speculative_config9] (vllm-project#32355)

Signed-off-by: Micah Williamson <micah.williamson@amd.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

llama Related to Llama models ready ONLY add when PR is ready to merge/full CI is needed rocm Related to AMD ROCm structured-output v1

Projects

Status: Done

Development

Successfully merging this pull request may close these issues.

3 participants