From eb5ea1c6324a3c8039650458a64e4b75285f3e5f Mon Sep 17 00:00:00 2001 From: Jirka Borovec Date: Sat, 30 Jan 2021 01:17:32 +0100 Subject: [PATCH] ro2 --- pytorch_lightning/trainer/training_loop.py | 23 ---------------------- 1 file changed, 23 deletions(-) diff --git a/pytorch_lightning/trainer/training_loop.py b/pytorch_lightning/trainer/training_loop.py index fcdeca9e31448..1f8722dfb9e6e 100644 --- a/pytorch_lightning/trainer/training_loop.py +++ b/pytorch_lightning/trainer/training_loop.py @@ -429,29 +429,6 @@ def _process_training_step_output_1_0(self, training_step_output, split_batch): return training_step_output_for_epoch_end, training_step_output - def _process_result(self, training_step_output, split_batch): - training_step_output.track_batch_size(len(split_batch)) - m = """ - TrainResult and EvalResult were deprecated in 0.9.1 and support will drop in 1.0.0. - Use self.log and .write from the LightningModule to log metrics and write predictions. - training_step can now only return a scalar (for the loss) or a dictionary with anything you want. - - Option 1: - return loss - - Option 2: - return {'loss': loss, 'anything_else': ...} - - Option 3: - return {'loss': loss, 'hiddens': hiddens, 'anything_else': ...} - """ - rank_zero_warn(m) - - training_step_output_for_epoch_end = copy(training_step_output) - training_step_output_for_epoch_end.detach() - - return training_step_output_for_epoch_end - def optimizer_step(self, optimizer, opt_idx, batch_idx, train_step_and_backward_closure): model_ref = self.trainer.get_model()