From a5af947ed7d2fecd818a8eb7ae928a1a15b7de4c Mon Sep 17 00:00:00 2001 From: DesmonDay <908660116@qq.com> Date: Thu, 17 Oct 2024 11:23:28 +0800 Subject: [PATCH] bug fix --- paddlenlp/trainer/trainer.py | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/paddlenlp/trainer/trainer.py b/paddlenlp/trainer/trainer.py index 5e7a276b10bb..64d7ef88341d 100644 --- a/paddlenlp/trainer/trainer.py +++ b/paddlenlp/trainer/trainer.py @@ -539,7 +539,9 @@ def _load_from_checkpoint(self, resume_from_checkpoint=None): # Load potential model checkpoint if isinstance(resume_from_checkpoint, bool) and resume_from_checkpoint: uc_async_save = self.args.unified_checkpoint and "async_save" in self.args.unified_checkpoint_config - resume_from_checkpoint = get_last_checkpoint(self.args.output_dir, uc_async_save) + resume_from_checkpoint = get_last_checkpoint( + self.args.output_dir, signal_folder=self.args.output_signal_dir, uc_async_save=uc_async_save + ) if resume_from_checkpoint is None: raise ValueError(f"No valid checkpoint found in output directory ({self.args.output_dir})")