lightning.fabric.strategies¶
Strategies¶
| Base class for all strategies that change the behaviour of the training, validation and test- loop. | |
| Strategy for multi-process single-device training on one or multiple nodes. | |
| Implements data-parallel training in a single process, i.e., the model gets replicated to each device and each gets a split of the data. | |
| Strategy for Fully Sharded Data Parallel provided by torch.distributed. | |
| Provides capabilities to run training using the DeepSpeed library, with training optimizations for large billion parameter models. | |
| Strategy for training multiple TPU devices using the  | |
| Strategy for training multiple XLA devices using the  | |
| Strategy for training with multiple processes in parallel. | |
| Strategy that handles communication on a single device. | |
| Strategy for training on a single XLA device. | |
| Enables user-defined parallelism applied to a model. |