From 8193bae6bde378e137ab07eb916a4cb515ecbace Mon Sep 17 00:00:00 2001 From: Yuta Hayashibe Date: Fri, 2 Jul 2021 23:48:55 +0900 Subject: [PATCH] Add periods to the documentation (#8252) --- pytorch_lightning/callbacks/model_checkpoint.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/pytorch_lightning/callbacks/model_checkpoint.py b/pytorch_lightning/callbacks/model_checkpoint.py index ad10480987fcd..eae401067b25f 100644 --- a/pytorch_lightning/callbacks/model_checkpoint.py +++ b/pytorch_lightning/callbacks/model_checkpoint.py @@ -102,7 +102,7 @@ class ModelCheckpoint(Callback): saved (``model.save_weights(filepath)``), else the full model is saved (``model.save(filepath)``). every_n_train_steps: Number of training steps between checkpoints. - If ``every_n_train_steps == None or every_n_train_steps == 0``, we skip saving during training + If ``every_n_train_steps == None or every_n_train_steps == 0``, we skip saving during training. To disable, set ``every_n_train_steps = 0``. This value must be ``None`` or non-negative. This must be mutually exclusive with ``train_time_interval`` and ``every_n_val_epochs``. train_time_interval: Checkpoints are monitored at the specified time interval. @@ -111,7 +111,7 @@ class ModelCheckpoint(Callback): guaranteed to execute at the exact time specified, but should be close. This must be mutually exclusive with ``every_n_train_steps`` and ``every_n_val_epochs``. every_n_val_epochs: Number of validation epochs between checkpoints. - If ``every_n_val_epochs == None or every_n_val_epochs == 0``, we skip saving on validation end + If ``every_n_val_epochs == None or every_n_val_epochs == 0``, we skip saving on validation end. To disable, set ``every_n_val_epochs = 0``. This value must be ``None`` or non-negative. This must be mutually exclusive with ``every_n_train_steps`` and ``train_time_interval``. Setting both ``ModelCheckpoint(..., every_n_val_epochs=V)`` and