Update Qwen3 30B H100 Base Configs with HybridEP#2477
Conversation
Signed-off-by: Raghav Hrishikeshan Mukundan <rmukundan@nvidia.com>
|
No actionable comments were generated in the recent review. 🎉 📝 WalkthroughWalkthroughConfiguration updates for Qwen3 pretraining on H100/GB300: switching MoE token dispatcher type to "flex" in qwen3_llm_pretrain.py, and adjusting parallelism strategy from pipeline-virtual to expert-model-parallel in qwen3_workload_base_configs.py, with corresponding dispatcher backend and overlap settings changes. Changes
Estimated code review effort🎯 2 (Simple) | ⏱️ ~12 minutes Possibly related PRs
Suggested reviewers
🚥 Pre-merge checks | ✅ 3 | ❌ 1❌ Failed checks (1 inconclusive)
✅ Passed checks (3 passed)
✏️ Tip: You can configure your own custom pre-merge checks in the settings. ✨ Finishing Touches
🧪 Generate unit tests (beta)
Tip Issue Planner is now in beta. Read the docs and try it out! Share your feedback on Discord. Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out. Comment |
ko3n1g
left a comment
There was a problem hiding this comment.
Can you update golden values of the internal CI?
Signed-off-by: Raghav Hrishikeshan Mukundan <rmukundan@nvidia.com> Signed-off-by: oliver könig <okoenig@nvidia.com>
Signed-off-by: Raghav Hrishikeshan Mukundan <rmukundan@nvidia.com> Signed-off-by: pengdurice <pengduhit@gmail.com>
Signed-off-by: Raghav Hrishikeshan Mukundan <rmukundan@nvidia.com>
Summary by CodeRabbit