[Bugfix] Fix test_whisper distributed test stability: torch.compile flakiness and memory utilization#42092
Closed
dzhengAP wants to merge 2 commits intovllm-project:mainfrom
Closed
Conversation
…y_utilization and enforce_eager Signed-off-by: dqzhengAP <dqzheng1996@gmail.com>
Contributor
There was a problem hiding this comment.
Code Review
This pull request modifies the Whisper generation tests by removing the process spawning decorator for distributed tests and adjusting GPU memory utilization settings. Feedback focuses on restoring the spawn decorator and its import to maintain CUDA stability, as well as further reducing the gpu_memory_utilization to 0.65 to ensure reliable execution within CI memory constraints.
… test Signed-off-by: dqzhengAP <dqzheng1996@gmail.com>
b82cab3 to
38d129e
Compare
Contributor
Contributor
Author
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Follow-up to #41423 and #42038.
test_models_distributed in test_whisper.py was failing in CI build #65117 due to two issues:
torch.compile flakiness — with enforce_eager=False, the test triggers torch.compile/AOT cache setup which can fail non-deterministically.
Fix: use enforce_eager=True for the distributed correctness test.
Leftover GPU memory — the Whisper test runs last (command 7/7) in the CI job. Earlier tests leave ~6.6 GiB of GPU memory occupied, causing vLLM's startup memory check to fail: Free memory on device cuda:0 (15.41/22.05 GiB) is less than desired GPU memory utilization (0.92, 20.28 GiB).
Fix: lower gpu_memory_utilization to 0.7 — sufficient for max_model_len=448. Same issue also catched by @SoluMilken
@ProExpertProg @DarkLight1337