From 69f521eaf0a1d99628a100b193e1552c0a8aab8e Mon Sep 17 00:00:00 2001 From: Jirka Borovec Date: Thu, 1 Oct 2020 13:55:40 +0200 Subject: [PATCH] names --- pytorch_lightning/accelerators/ddp2_backend.py | 4 ++-- pytorch_lightning/trainer/logging.py | 4 ++-- pytorch_lightning/utilities/parsing.py | 4 ++-- tests/base/models.py | 4 ++-- 4 files changed, 8 insertions(+), 8 deletions(-) diff --git a/pytorch_lightning/accelerators/ddp2_backend.py b/pytorch_lightning/accelerators/ddp2_backend.py index 4d7b340ee380c..28af6558a0fef 100644 --- a/pytorch_lightning/accelerators/ddp2_backend.py +++ b/pytorch_lightning/accelerators/ddp2_backend.py @@ -47,9 +47,9 @@ def _resolve_task_idx(self): # torchelastic or general non_slurm ddp2 try: self.task_idx = int(os.environ['LOCAL_RANK']) - except Exception as e: + except Exception as exp: m = 'ddp2 only works in SLURM or via torchelastic with the WORLD_SIZE, LOCAL_RANK, GROUP_RANK flags' - raise MisconfigurationException(m) from e + raise MisconfigurationException(m) from exp def train(self): model = self.trainer.model diff --git a/pytorch_lightning/trainer/logging.py b/pytorch_lightning/trainer/logging.py index 1ac4e7e1c865c..b585647fb5a0e 100644 --- a/pytorch_lightning/trainer/logging.py +++ b/pytorch_lightning/trainer/logging.py @@ -141,13 +141,13 @@ def process_dict_result(self, output, train=False): if train: try: loss = output['loss'] - except Exception as e: + except Exception as exp: if isinstance(output, torch.Tensor): loss = output else: raise RuntimeError( 'No `loss` value in the dictionary returned from `model.training_step()`.' - ) from e + ) from exp # when using dp need to reduce the loss if self.use_dp or self.use_ddp2: diff --git a/pytorch_lightning/utilities/parsing.py b/pytorch_lightning/utilities/parsing.py index 82fa6e865f242..c562f780e88ee 100644 --- a/pytorch_lightning/utilities/parsing.py +++ b/pytorch_lightning/utilities/parsing.py @@ -158,8 +158,8 @@ class AttributeDict(Dict): def __getattr__(self, key): try: return self[key] - except KeyError as e: - raise AttributeError(f'Missing attribute "{key}"') from e + except KeyError as exp: + raise AttributeError(f'Missing attribute "{key}"') from exp def __setattr__(self, key, val): self[key] = val diff --git a/tests/base/models.py b/tests/base/models.py index 60dcf1777f1b2..79f407d62aadf 100644 --- a/tests/base/models.py +++ b/tests/base/models.py @@ -10,9 +10,9 @@ try: from test_tube import HyperOptArgumentParser -except ImportError as e: +except ImportError as exp: # TODO: this should be discussed and moved out of this package - raise ImportError('Missing test-tube package.') from e + raise ImportError('Missing test-tube package.') from exp from pytorch_lightning.core.lightning import LightningModule