Support Energon dataset format for updated Qwen3-VL finetuning pipeline#2679
Support Energon dataset format for updated Qwen3-VL finetuning pipeline#2679aub123 wants to merge 1 commit intoNVIDIA-NeMo:mainfrom
Conversation
|
Caution Review failedThe pull request is closed. ℹ️ Recent review info⚙️ Run configurationConfiguration used: Path: .coderabbit.yaml Review profile: CHILL Plan: Pro Run ID: 📒 Files selected for processing (1)
📝 WalkthroughWalkthroughThe Qwen3VL bridge initialization was refactored to shift from generic provider_kwargs-based construction to explicit parameter instantiation with integrated dtype handling and RoPE configuration extraction. New mapping classes were introduced for expert MLPs, and legacy weight alignment utilities were removed. Changes
Estimated code review effort🎯 3 (Moderate) | ⏱️ ~20 minutes Possibly related PRs
Suggested reviewers
✨ Finishing Touches🧪 Generate unit tests (beta)
Tip Try Coding Plans. Let us write the prompt for your AI agent so you can ship faster (with fewer bugs). Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out. Comment |
Motivation
Recent updates to the Qwen3-VL finetuning pipeline introduce changes in the expected multimodal input format. The current implementation does not directly support Energon-format datasets used in our training pipeline.
Changes
qwen3_vl_bridge.pyNotes
This change ensures Energon-format datasets can be directly used in the Qwen3-VL finetuning pipeline without additional preprocessing.
Tested on Qwen3-VL-8B model.
Summary by CodeRabbit
Release Notes
Refactor
Breaking Changes