From fd124441e778cba5a12e228961a3010e7999bca2 Mon Sep 17 00:00:00 2001 From: rohitgr7 Date: Sat, 25 Jul 2020 03:27:25 +0530 Subject: [PATCH] Fix logging interval --- pytorch_lightning/trainer/training_loop.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/pytorch_lightning/trainer/training_loop.py b/pytorch_lightning/trainer/training_loop.py index c84769108a24f..df282fcbca08f 100644 --- a/pytorch_lightning/trainer/training_loop.py +++ b/pytorch_lightning/trainer/training_loop.py @@ -632,7 +632,7 @@ def increment_accumulated_grad_global_step(self): def save_train_loop_metrics_to_loggers(self, batch_idx, batch_output): # when metrics should be logged - should_log_metrics = batch_idx % self.row_log_interval == 0 or self.should_stop + should_log_metrics = (batch_idx + 1) % self.row_log_interval == 0 or self.should_stop if should_log_metrics or self.fast_dev_run: # logs user requested information to logger metrics = batch_output.batch_log_metrics