Skip to content
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 2 additions & 2 deletions open_instruct/finetune.py
Original file line number Diff line number Diff line change
Expand Up @@ -905,14 +905,14 @@ def main(args: FlatArguments, tc: TokenizerConfig):
avg_loss = sum_loss / total_fwd_passes
metrics_to_log["train_loss"] = avg_loss
else:
avg_loss_per_total_tok = sum_loss / total_tokens_this_log_period
avg_loss = sum_loss / total_tokens_this_log_period
# The loss per pred tok is the closest analogue to what we report as the
# avg_loss in the "mean" case
avg_loss_per_pred_tok = sum_loss / pred_tokens_this_log_period
total_optim_steps = args.logging_steps * accelerator.num_processes
avg_sum_loss = sum_loss / total_optim_steps
metrics_to_log["train_sum_loss"] = avg_sum_loss
metrics_to_log["train_loss_per_total_tok"] = avg_loss_per_total_tok
metrics_to_log["train_loss_per_total_tok"] = avg_loss
metrics_to_log["train_loss_per_pred_tok"] = avg_loss_per_pred_tok
if args.verbose:
sec_per_step = (time.time() - start_time) / (completed_steps - resume_step)
Expand Down