diff --git a/pytorch_lightning/trainer/training_loop.py b/pytorch_lightning/trainer/training_loop.py index c4e48596e6e70..11b71797ddab4 100644 --- a/pytorch_lightning/trainer/training_loop.py +++ b/pytorch_lightning/trainer/training_loop.py @@ -406,29 +406,6 @@ def _process_training_step_output_1_0(self, training_step_output, split_batch): return training_step_output_for_epoch_end, training_step_output - def _process_result(self, training_step_output, split_batch): - training_step_output.track_batch_size(len(split_batch)) - m = """ - TrainResult and EvalResult were deprecated in 0.9.1 and support will drop in 1.0.0. - Use self.log and .write from the LightningModule to log metrics and write predictions. - training_step can now only return a scalar (for the loss) or a dictionary with anything you want. - - Option 1: - return loss - - Option 2: - return {'loss': loss, 'anything_else': ...} - - Option 3: - return {'loss': loss, 'hiddens': hiddens, 'anything_else': ...} - """ - rank_zero_warn(m) - - training_step_output_for_epoch_end = copy(training_step_output) - training_step_output_for_epoch_end.detach() - - return training_step_output_for_epoch_end - def optimizer_step(self, optimizer, opt_idx, batch_idx, train_step_and_backward_closure): model_ref = self.trainer.get_model()