Qwen3-Omni][Bugfix] Replace vLLM fused layers with HF-compatible numerics in code predictor#2291
Merged
hsliuustc0106 merged 3 commits intoMar 28, 2026
Conversation
Signed-off-by: Junhong Liu <98734602+LJH-LBJ@users.noreply.github.com>
Contributor
|
Please add a nightly-test label. @Gaohan123 @david6666666 @gcanlin @hsliuustc0106 |
Contributor
|
We are working on an accuracy benchmark. Could you add a long-output use case to monitor this scenario? |
Contributor
|
@Sy0307 PTAL |
Contributor
Author
Sure, I can do it in next pr. |
Contributor
|
LGTM. |
hsliuustc0106
approved these changes
Mar 28, 2026
vraiti
pushed a commit
to vraiti/vllm-omni
that referenced
this pull request
Apr 9, 2026
…rics in code predictor (vllm-project#2291) Signed-off-by: Junhong Liu <98734602+LJH-LBJ@users.noreply.github.com>
lengrongfu
pushed a commit
to lengrongfu/vllm-omni
that referenced
this pull request
May 1, 2026
…rics in code predictor (vllm-project#2291) Signed-off-by: Junhong Liu <98734602+LJH-LBJ@users.noreply.github.com>
clodaghwalsh17
pushed a commit
to clodaghwalsh17/nm-vllm-omni-ent
that referenced
this pull request
May 12, 2026
…rics in code predictor (vllm-project#2291) Signed-off-by: Junhong Liu <98734602+LJH-LBJ@users.noreply.github.com>
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
PLEASE FILL IN THE PR DESCRIPTION HERE ENSURING ALL CHECKLIST ITEMS (AT THE BOTTOM) HAVE BEEN CONSIDERED.
Purpose
Resolve: #2286
Fixes audio quality degradation (noise after ~3000 tokens) in Qwen3-Omni's code predictor by replacing vLLM's fused kernels with plain PyTorch equivalents that match HuggingFace reference numerics.
This is the Qwen3-Omni counterpart of PR #2277 (which fixed the same issue for Qwen3-TTS). The Qwen3-Omni code predictor was written following PR #1617's approach and inherits the same precision bug.
RMSNorm(bf16 variance)_RMSNorm(float32 variance)get_rope(bf16 cos/sin)_RotaryEmbedding(float32 cos/sin,torch.autocast(enabled=False))QKVParallelLinear(fused)nn.Linearfor q/k/vMergedColumnParallelLinear+RowParallelLinearnn.Linearfor gate/up/downmode="default"options={"epilogue_fusion": False}[bsz * seq_len]1D flat[bsz, seq_len]2D (HF format)Test Plan
curl -s http://localhost:46354/v1/chat/completions -H "Content-Type: application/json" -d '{
"model": "/workspace/models/Qwen3-Omni-30B-A3B-Instruct",
"messages": [{"role": "user", "content": "Please write a 5000-word novel."}],
"modalities": ["audio"]
}' | jq -r '.choices[0].message.audio.data' | base64 -d > output2.wav
Test Result
There is no noise in the audio. It cannot upload because its size is too big
Essential Elements of an Effective PR Description Checklist
supported_models.mdandexamplesfor a new model. Please runmkdocs serveto sync the documentation editions to./docs.BEFORE SUBMITTING, PLEASE READ https://github.com/vllm-project/vllm-omni/blob/main/CONTRIBUTING.md (anything written below this line will be removed by GitHub Actions)