Conversation
Signed-off-by: Sanyam Kapoor <sanyamk@nvidia.com>
WalkthroughAdds num_workers=10 to DataLoader calls in sft.py and toggles one DataLoader’s drop_last to True. In megatron_policy_worker.py, conditionally applies layernorm_epsilon from config to model configuration if provided. Changes
Sequence Diagram(s)sequenceDiagram
autonumber
participant Trainer as Trainer (sft.py)
participant Data as Dataset
participant DL as DataLoader
Trainer->>Data: Prepare dataset
Trainer->>DL: Create DataLoader(dataset, num_workers=10, drop_last=[false|true])
note over DL: drop_last toggled to True in one loader
DL-->>Trainer: Batches
sequenceDiagram
autonumber
participant Worker as PolicyWorker
participant Cfg as megatron_cfg
participant Model as model_cfg
Worker->>Cfg: Load config
Worker->>Model: Build base model_cfg
alt layernorm_epsilon present
Worker->>Model: Set layernorm_epsilon
note right of Model: Conditional application
else absent
Note right of Worker: Use default epsilon
end
Estimated code review effort🎯 2 (Simple) | ⏱️ ~10 minutes Poem
Pre-merge checks and finishing touches❌ Failed checks (1 warning)
✅ Passed checks (2 passed)
✨ Finishing touches🧪 Generate unit tests
📜 Recent review detailsConfiguration used: CodeRabbit UI Review profile: CHILL Plan: Pro 📒 Files selected for processing (1)
⏰ Context from checks skipped due to timeout of 90000ms. You can increase the timeout in your CodeRabbit configuration to a maximum of 15 minutes (900000ms). (1)
🔇 Additional comments (3)
Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out. Comment |
Signed-off-by: Sanyam Kapoor <sanyamk@nvidia.com>
Signed-off-by: Sanyam Kapoor <sanyamk@nvidia.com> Signed-off-by: SeanNaren <snarenthiran@nvidia.com>
Signed-off-by: Sanyam Kapoor <sanyamk@nvidia.com> Signed-off-by: SeanNaren <snarenthiran@nvidia.com>
Signed-off-by: Sanyam Kapoor <sanyamk@nvidia.com> Signed-off-by: dgitman <dgitman@nvidia.com>
Summary by CodeRabbit
New Features
Chores