Upgrading checkpoints (intermediate)¶
Audience: Users who are upgrading Lightning and their code and want to reuse their old checkpoints.
Resume training from an old checkpoint¶
Next to the model weights and trainer state, a Lightning checkpoint contains the version number of Lightning with which the checkpoint was saved. When you load a checkpoint file, either by resuming training
trainer = Trainer(...)
trainer.fit(model, ckpt_path="path/to/checkpoint.ckpt")
or by loading the state directly into your model,
model = LitModel.load_from_checkpoint("path/to/checkpoint.ckpt")
Lightning will automatically recognize that it is from an older version and migrates the internal structure so it can be loaded properly. This is done without any action required by the user.
Upgrade checkpoint files permanently¶
When Lightning loads a checkpoint, it applies the version migration on-the-fly as explained above, but it does not modify your checkpoint files. You can upgrade checkpoint files permanently with the following command
python -m pytorch_lightning.utilities.upgrade_checkpoint path/to/model.ckpt
or a folder with multiple files:
python -m pytorch_lightning.utilities.upgrade_checkpoint /path/to/checkpoints/folder