Skip to content
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 0 additions & 2 deletions examples/configs/dpo.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -55,8 +55,6 @@ policy:
# makes the training sequence length divisible by the tensor parallel size
# this is useful for sequence parallel training
make_sequence_length_divisible_by: ${policy.dtensor_cfg.tensor_parallel_size}
## NOTE: there is a known issue with gradient clipping when using Dtensor
## if using dtensor, set max_grad_norm to NULL
max_grad_norm: 1.0

optimizer:
Expand Down