Skip to content

Limit gpu utils and lower max BS on test_transcription_api_correctness.py#41649

Merged
DarkLight1337 merged 2 commits into
vllm-project:mainfrom
ekagra-ranjan:er-cohere-asr-ci-oom-2
May 4, 2026
Merged

Limit gpu utils and lower max BS on test_transcription_api_correctness.py#41649
DarkLight1337 merged 2 commits into
vllm-project:mainfrom
ekagra-ranjan:er-cohere-asr-ci-oom-2

Conversation

@ekagra-ranjan
Copy link
Copy Markdown
Contributor

#41478 added a fix to lower BS in the test. The CI passed in that PR but failed later on in https://buildkite.com/vllm/ci/builds/64258/canvas?sid=019df193-a942-4d4b-aeb0-15f160336dfa&tab=output. This PR adds more aggressive limit on max memory size since the CI machine is a 18GB MIG H200.

Signed-off-by: Ekagra Ranjan <3116519+ekagra-ranjan@users.noreply.github.com>
Signed-off-by: Ekagra Ranjan <3116519+ekagra-ranjan@users.noreply.github.com>
Copy link
Copy Markdown

@claude claude Bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Claude Code Review

This pull request is from a fork — automated review is disabled. A repository maintainer can comment @claude review to run a one-time review.

Copy link
Copy Markdown
Contributor

@gemini-code-assist gemini-code-assist Bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request modifies the transcription correctness tests to prevent OOM errors on 18GB GPUs. Specifically, it reduces the MAX_SEQS_FOR_TRANSCRIPTION_TEST from 32 to 8 and introduces a GPU_UTIL_FOR_TRANSCRIPTION_TEST constant set to 0.5, which is now passed as a --gpu_memory_utilization argument to the test server. I have no feedback to provide.

@DarkLight1337 DarkLight1337 enabled auto-merge (squash) May 4, 2026 16:20
@github-actions github-actions Bot added the ready ONLY add when PR is ready to merge/full CI is needed label May 4, 2026
@DarkLight1337 DarkLight1337 merged commit 321fa2d into vllm-project:main May 4, 2026
17 checks passed
chaojun-zhang pushed a commit to chaojun-zhang/vllm that referenced this pull request May 6, 2026
…s.py (vllm-project#41649)

Signed-off-by: Ekagra Ranjan <3116519+ekagra-ranjan@users.noreply.github.com>
Copilot AI pushed a commit to hongbolv/vllm that referenced this pull request May 7, 2026
…s.py (vllm-project#41649)

Signed-off-by: Ekagra Ranjan <3116519+ekagra-ranjan@users.noreply.github.com>
Co-authored-by: hongbolv <33214277+hongbolv@users.noreply.github.com>
ikaadil pushed a commit to ikaadil/vllm that referenced this pull request May 7, 2026
…s.py (vllm-project#41649)

Signed-off-by: Ekagra Ranjan <3116519+ekagra-ranjan@users.noreply.github.com>
Signed-off-by: Ifta Khairul Alam Adil <ikaadil007@gmail.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ready ONLY add when PR is ready to merge/full CI is needed

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants