lightning.fabric.strategies¶
Strategies¶
Base class for all strategies that change the behaviour of the training, validation and test- loop. |
|
Strategy for multi-process single-device training on one or multiple nodes. |
|
Implements data-parallel training in a single process, i.e., the model gets replicated to each device and each gets a split of the data. |
|
Strategy for Fully Sharded Data Parallel provided by torch.distributed. |
|
Provides capabilities to run training using the DeepSpeed library, with training optimizations for large billion parameter models. |
|
Strategy for training multiple TPU devices using the |
|
Strategy for training multiple XLA devices using the |
|
Strategy for training with multiple processes in parallel. |
|
Strategy that handles communication on a single device. |
|
Strategy for training on a single XLA device. |
|
Enables user-defined parallelism applied to a model. |