diff --git a/examples/RLHF/ppo_trainer.py b/examples/RLHF/ppo_trainer.py index 32a9c148f193..84577d6d2c01 100644 --- a/examples/RLHF/ppo_trainer.py +++ b/examples/RLHF/ppo_trainer.py @@ -415,7 +415,11 @@ def full_training_step(self: Trainer, inputs: Dict[str, paddle.Tensor], **kwargs self.scaler.step(self.optimizer) self.scaler.update() scale_after = self.scaler._scale - optimizer_was_run = not self.scaler._cache_founf_inf + # Compatible with paddlepaddle 2.6.0 using typo word. + if hasattr(self.scaler, "_cache_founf_inf"): + optimizer_was_run = not self.scaler._cache_founf_inf + else: + optimizer_was_run = not self.scaler._cache_found_inf if not optimizer_was_run: scale_before_value = scale_before.cpu().numpy() scale_after_value = scale_after.cpu().numpy() diff --git a/paddlenlp/trainer/auto_trainer.py b/paddlenlp/trainer/auto_trainer.py index c6d6f2962daf..c9f836f37971 100644 --- a/paddlenlp/trainer/auto_trainer.py +++ b/paddlenlp/trainer/auto_trainer.py @@ -438,7 +438,11 @@ def optimizer_step(self): self.scaler.step(self.optimizer) self.scaler.update() scale_after = self.scaler._scale - optimizer_was_run = not self.scaler._cache_founf_inf + # Compatible with paddlepaddle 2.6.0 using typo word. + if hasattr(self.scaler, "_cache_founf_inf"): + optimizer_was_run = not self.scaler._cache_founf_inf + else: + optimizer_was_run = not self.scaler._cache_found_inf if not optimizer_was_run: scale_before_value = scale_before.cpu().numpy() scale_after_value = scale_after.cpu().numpy() diff --git a/paddlenlp/trainer/trainer.py b/paddlenlp/trainer/trainer.py index 3cc108fb4c01..7fa1cb171e44 100644 --- a/paddlenlp/trainer/trainer.py +++ b/paddlenlp/trainer/trainer.py @@ -1031,7 +1031,11 @@ def _inner_training_loop( self.scaler.step(self.optimizer) self.scaler.update() scale_after = self.scaler._scale - optimizer_was_run = not self.scaler._cache_founf_inf + # Compatible with paddlepaddle 2.6.0 using typo word. + if hasattr(self.scaler, "_cache_founf_inf"): + optimizer_was_run = not self.scaler._cache_founf_inf + else: + optimizer_was_run = not self.scaler._cache_found_inf if not optimizer_was_run: scale_before_value = scale_before.cpu().numpy() scale_after_value = scale_after.cpu().numpy()