deepspeed¶
Functions
Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated |
|
|
|
Utilities that can be used with Deepspeed.
- pytorch_lightning.utilities.deepspeed.convert_zero_checkpoint_to_fp32_state_dict(checkpoint_dir, output_file, tag=None)[source]¶
Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated
state_dict
file that can be loaded withtorch.load(file)
+load_state_dict()
and used for training without DeepSpeed. It gets copied into the top level checkpoint dir, so the user can easily do the conversion at any point in the future. Once extracted, the weights don’t require DeepSpeed and can be used in any application. Additionally the script has been modified to ensure we keep the lightning state inside the state dict for being able to runLightningModule.load_from_checkpoint('...')`
.- Parameters:
checkpoint_dir¶ (_PATH) – path to the desired checkpoint folder. (one that contains the tag-folder, like
global_step14
)output_file¶ (_PATH) – path to the pytorch fp32 state_dict output file (e.g. path/pytorch_model.bin)
tag¶ (str | None) – checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named
latest
in the checkpoint folder, e.g.,global_step14
Examples
>>> from pytorch_lightning.utilities.deepspeed import ( ... convert_zero_checkpoint_to_fp32_state_dict ... ) >>> # Lightning deepspeed has saved a directory instead of a file >>> save_path = "lightning_logs/version_0/checkpoints/epoch=0-step=0.ckpt/" >>> output_path = "lightning_model.pt" >>> convert_zero_checkpoint_to_fp32_state_dict(save_path, output_path) Saving fp32 state dict to lightning_model.pt
- Return type:
None