Skip to content

[diffusion][hot fix] fix accuracy bug caused by PR 14717#18296

Merged
mickqian merged 3 commits intomainfrom
kernel-fix
Feb 5, 2026
Merged

[diffusion][hot fix] fix accuracy bug caused by PR 14717#18296
mickqian merged 3 commits intomainfrom
kernel-fix

Conversation

@yingluosanqian
Copy link
Copy Markdown
Collaborator

@yingluosanqian yingluosanqian commented Feb 5, 2026

Motivation

PR14717 introduces an accuracy issue that causes some models to generate incorrect images, as shown below.

This PR fix it. cc @mickqian @BBuf

flux_image_t2i

image

flux_2_klein_image_t2

image

This issue was caused by a change to the default dtype of the RMSNorm weight. After fixing it, the image quality returns to normal.

image

flux_image_t2i

image

flux_2_klein_image_t2

f8c961f3-1a5c-4fbb-9d39-1b0cc9202beb

Also fix a cutedsl sync issue. Though the model accuracy it's not caused by this issue.

before fix:

E       AssertionError: Tensor-likes are not close!
E       
E       Mismatched elements: 75 / 353894400 (0.0%)
E       Greatest absolute difference: 0.59765625 at index (0, 36219, 342) (up to 0.05 allowed)
E       Greatest relative difference: 13.1875 at index (0, 36219, 480) (up to 0.05 allowed)

python/sglang/jit_kernel/tests/test_fused_norm_scale_shift.py:175: AssertionError
======================================== short test summary info =========================================
FAILED python/sglang/jit_kernel/tests/test_fused_norm_scale_shift.py::TestFusedScaleResidualNormScaleShift::test_dtype_1[dtype1-layer] - AssertionError: Tensor-likes are not close!
===================================== 1 failed, 147 passed in 24.70s =

after fix:

python/sglang/jit_kernel/tests/test_fused_norm_scale_shift.py .................................... [ 24%]
.................................................................................................. [ 90%]
..............                                                                                     [100%]

========================================== 148 passed in 25.08s ==========================================

Modifications

Accuracy Tests

Benchmarking and Profiling

Checklist

Review Process

  1. Ping Merge Oncalls to start the PR flow. See the PR Merge Process.
  2. Get approvals from CODEOWNERS and other reviewers.
  3. Trigger CI tests with comments or contact authorized users to do so.
    • /tag-run-ci-label, /rerun-failed-ci, /tag-and-rerun-ci
  4. After green CI and required approvals, ask Merge Oncalls to merge.

@gemini-code-assist
Copy link
Copy Markdown
Contributor

Warning

You have reached your daily quota limit. Please wait up to 24 hours and I will start processing your requests again!

@github-actions github-actions bot added the diffusion SGLang Diffusion label Feb 5, 2026
@yingluosanqian
Copy link
Copy Markdown
Collaborator Author

fast_hunyuan_video

before fix:

image

after fix:

image

@BBuf
Copy link
Copy Markdown
Collaborator

BBuf commented Feb 5, 2026

/tag-and-rerun-ci

@BBuf BBuf changed the title [diffusion][fix] fix accuracy bug caused by 14717 [diffusion][hot fix] fix accuracy bug caused by 14717 Feb 5, 2026
@BBuf BBuf changed the title [diffusion][hot fix] fix accuracy bug caused by 14717 [diffusion][hot fix] fix accuracy bug caused by PR 14717 Feb 5, 2026
@mickqian mickqian merged commit 4aa03d9 into main Feb 5, 2026
235 of 253 checks passed
@mickqian mickqian deleted the kernel-fix branch February 5, 2026 12:36
charlesHsuGG pushed a commit to charlesHsuGG/sglang that referenced this pull request Feb 9, 2026
Johnsonms pushed a commit to Johnsonms/sglang that referenced this pull request Feb 14, 2026
magicYang1573 pushed a commit to magicYang1573/sglang that referenced this pull request Mar 9, 2026
Wangzheee pushed a commit to Wangzheee/sglang that referenced this pull request Mar 21, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

diffusion SGLang Diffusion run-ci

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants