From b5b77e44b15d1c3f3178a63428967a10d17c6cb0 Mon Sep 17 00:00:00 2001 From: Dang Nguyen Anh Khoa Date: Wed, 4 Dec 2019 19:04:14 +0700 Subject: [PATCH] fix logging error (#575) * fix logging error * no need for the '+' sign * move space to beginning of next line --- pytorch_lightning/callbacks/pt_callbacks.py | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) diff --git a/pytorch_lightning/callbacks/pt_callbacks.py b/pytorch_lightning/callbacks/pt_callbacks.py index c55a488496..1951719877 100644 --- a/pytorch_lightning/callbacks/pt_callbacks.py +++ b/pytorch_lightning/callbacks/pt_callbacks.py @@ -312,16 +312,16 @@ class ModelCheckpoint(Callback): self.best = max(self.best_k_models.values()) if self.verbose > 0: logging.info( - f'\nEpoch {epoch:05d}: {self.monitor} reached', - f'{current:0.5f} (best {self.best:0.5f}), saving model to', - f'{filepath} as top {self.save_top_k}') + f'\nEpoch {epoch:05d}: {self.monitor} reached' + f' {current:0.5f} (best {self.best:0.5f}), saving model to' + f' {filepath} as top {self.save_top_k}') self._save_model(filepath) else: if self.verbose > 0: logging.info( - f'\nEpoch {epoch:05d}: {self.monitor}', - f'was not in top {self.save_top_k}') + f'\nEpoch {epoch:05d}: {self.monitor}' + f' was not in top {self.save_top_k}') else: if self.verbose > 0: