Skip to content

Update Qwen3 30B H100 Base Configs with HybridEP#2477

Merged
ko3n1g merged 3 commits intomainfrom
rmukundan/qwen3_30b_h100_config_update
Feb 23, 2026
Merged

Update Qwen3 30B H100 Base Configs with HybridEP#2477
ko3n1g merged 3 commits intomainfrom
rmukundan/qwen3_30b_h100_config_update

Conversation

@rhmukundan
Copy link
Copy Markdown
Contributor

@rhmukundan rhmukundan commented Feb 22, 2026

Summary by CodeRabbit

  • Chores
    • Optimized Qwen3 model training configurations to enhance performance efficiency across different parallelism strategies.

Signed-off-by: Raghav Hrishikeshan Mukundan <rmukundan@nvidia.com>
@rhmukundan rhmukundan self-assigned this Feb 22, 2026
@copy-pr-bot
Copy link
Copy Markdown

copy-pr-bot bot commented Feb 22, 2026

This pull request requires additional validation before any workflows can run on NVIDIA's runners.

Pull request vetters can view their responsibilities here.

Contributors can view more details about this message here.

@rhmukundan rhmukundan marked this pull request as ready for review February 22, 2026 16:20
@coderabbitai
Copy link
Copy Markdown
Contributor

coderabbitai bot commented Feb 22, 2026

No actionable comments were generated in the recent review. 🎉


📝 Walkthrough

Walkthrough

Configuration updates for Qwen3 pretraining on H100/GB300: switching MoE token dispatcher type to "flex" in qwen3_llm_pretrain.py, and adjusting parallelism strategy from pipeline-virtual to expert-model-parallel in qwen3_workload_base_configs.py, with corresponding dispatcher backend and overlap settings changes.

Changes

Cohort / File(s) Summary
Qwen3 MoE Token Dispatcher Type
scripts/performance/configs/qwen/qwen3_llm_pretrain.py
Changed moe_token_dispatcher_type from "alltoall" to "flex" for qwen3_30b_a3b_pretrain_config_h100.
Qwen3 Workload Parallelism and MoE Backend
scripts/performance/configs/qwen/qwen3_workload_base_configs.py
Replaced pipeline/virtual parallelism settings with expert_model_parallel_size=16, disabled moe_a2a_overlap, and switched moe_flex_dispatcher_backend from "deepep" to "hybridep" across Qwen3 30B A3B config blocks.

Estimated code review effort

🎯 2 (Simple) | ⏱️ ~12 minutes

Possibly related PRs

Suggested reviewers

  • ko3n1g
  • yaoyu-33
🚥 Pre-merge checks | ✅ 3 | ❌ 1

❌ Failed checks (1 inconclusive)

Check name Status Explanation Resolution
Test Results For Major Changes ❓ Inconclusive PR makes MoE dispatcher configuration changes but test results/performance validation documentation is not accessible in repository context. Provide PR description from #2477 including any test results, performance benchmarks, or convergence validation documentation for dispatcher backend and parallelism strategy changes.
✅ Passed checks (3 passed)
Check name Status Explanation
Description Check ✅ Passed Check skipped - CodeRabbit’s high-level summary is enabled.
Title check ✅ Passed The title clearly and specifically summarizes the main change: updating Qwen3 30B H100 configuration with HybridEP dispatcher backend, which is the primary modification across both configuration files.
Docstring Coverage ✅ Passed Docstring coverage is 100.00% which is sufficient. The required threshold is 80.00%.

✏️ Tip: You can configure your own custom pre-merge checks in the settings.

✨ Finishing Touches
  • 📝 Generate docstrings (stacked PR)
  • 📝 Generate docstrings (commit on current branch)
🧪 Generate unit tests (beta)
  • Create PR with unit tests
  • Post copyable unit tests in a comment
  • Commit unit tests in branch rmukundan/qwen3_30b_h100_config_update

Tip

Issue Planner is now in beta. Read the docs and try it out! Share your feedback on Discord.


Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out.

❤️ Share

Comment @coderabbitai help to get the list of available commands and usage tips.

Copy link
Copy Markdown
Contributor

@ko3n1g ko3n1g left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Can you update golden values of the internal CI?

@rhmukundan rhmukundan added the r0.3.0 Cherry-pick label for r0.3.0 release branch label Feb 23, 2026
@ko3n1g ko3n1g merged commit 4a64507 into main Feb 23, 2026
2 checks passed
@ko3n1g ko3n1g deleted the rmukundan/qwen3_30b_h100_config_update branch February 23, 2026 20:59
ko3n1g pushed a commit that referenced this pull request Feb 24, 2026
Signed-off-by: Raghav Hrishikeshan Mukundan <rmukundan@nvidia.com>
Signed-off-by: oliver könig <okoenig@nvidia.com>
@ko3n1g ko3n1g mentioned this pull request Feb 24, 2026
5 tasks
pengdurice pushed a commit to pengdurice/Megatron-Bridge that referenced this pull request Feb 24, 2026
Signed-off-by: Raghav Hrishikeshan Mukundan <rmukundan@nvidia.com>
Signed-off-by: pengdurice <pengduhit@gmail.com>
copy-pr-bot bot pushed a commit that referenced this pull request Mar 19, 2026
Signed-off-by: Raghav Hrishikeshan Mukundan <rmukundan@nvidia.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

r0.3.0 Cherry-pick label for r0.3.0 release branch

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants