early_stopping¶
Classes
Monitor a metric and stop training when it stops improving. |
Early Stopping¶
Monitor a metric and stop training when it stops improving.
- class pytorch_lightning.callbacks.early_stopping.EarlyStopping(monitor=None, min_delta=0.0, patience=3, verbose=False, mode='min', strict=True, check_finite=True, stopping_threshold=None, divergence_threshold=None, check_on_train_epoch_end=None)[source]¶
Bases:
pytorch_lightning.callbacks.base.Callback
Monitor a metric and stop training when it stops improving.
- Parameters
min_delta¶ (
float
) – minimum change in the monitored quantity to qualify as an improvement, i.e. an absolute change of less than or equal to min_delta, will count as no improvement.number of checks with no improvement after which training will be stopped. Under the default configuration, one check happens after every training epoch. However, the frequency of validation can be modified by setting various parameters on the
Trainer
, for examplecheck_val_every_n_epoch
andval_check_interval
.Note
It must be noted that the patience parameter counts the number of validation checks with no improvement, and not the number of training epochs. Therefore, with parameters
check_val_every_n_epoch=10
andpatience=3
, the trainer will perform at least 40 training epochs before being stopped.mode¶ (
str
) – one of'min'
,'max'
. In'min'
mode, training will stop when the quantity monitored has stopped decreasing and in'max'
mode it will stop when the quantity monitored has stopped increasing.strict¶ (
bool
) – whether to crash the training if monitor is not found in the validation metrics.check_finite¶ (
bool
) – When setTrue
, stops training when the monitor becomes NaN or infinite.stopping_threshold¶ (
Optional
[float
]) – Stop training immediately once the monitored quantity reaches this threshold.divergence_threshold¶ (
Optional
[float
]) – Stop training as soon as the monitored quantity becomes worse than this threshold.check_on_train_epoch_end¶ (
Optional
[bool
]) – whether to run early stopping at the end of the training epoch. If this isFalse
, then the check runs at the end of the validation.
- Raises
MisconfigurationException – If
mode
is none of"min"
or"max"
.RuntimeError – If the metric
monitor
is not available.
Example:
>>> from pytorch_lightning import Trainer >>> from pytorch_lightning.callbacks import EarlyStopping >>> early_stopping = EarlyStopping('val_loss') >>> trainer = Trainer(callbacks=[early_stopping])
Tip
Saving and restoring multiple early stopping callbacks at the same time is supported under variation in the following arguments:
monitor, mode
Read more: Persisting State
- on_init_end(trainer)[source]¶
Called when the trainer initialization ends, model has not yet been set.
- Return type
- on_load_checkpoint(trainer, pl_module, callback_state)[source]¶
Called when loading a model checkpoint, use to reload state.
- Parameters
pl_module¶ (
LightningModule
) – the currentLightningModule
instance.callback_state¶ (
Dict
[str
,Any
]) – the callback state returned byon_save_checkpoint
.
- Return type
Note
The
on_load_checkpoint
won’t be called with an undefined state. If youron_load_checkpoint
hook behavior doesn’t rely on a state, you will still need to overrideon_save_checkpoint
to return adummy state
.
- on_save_checkpoint(trainer, pl_module, checkpoint)[source]¶
Called when saving a model checkpoint, use to persist state.
- on_train_epoch_end(trainer, pl_module)[source]¶
Called when the train epoch ends.
To access all batch outputs at the end of the epoch, either: :rtype:
None
Implement training_epoch_end in the LightningModule and access outputs via the module OR
Cache data across train batch hooks inside the callback implementation to post-process in this hook.
- property state_key: str¶
Identifier for the state of the callback.
Used to store and retrieve a callback’s state from the checkpoint dictionary by
checkpoint["callbacks"][state_key]
. Implementations of a callback need to provide a unique state key if 1) the callback has state and 2) it is desired to maintain the state of multiple instances of that callback.