DeepSpeedPrecisionPlugin
- class pytorch_lightning.plugins.precision.DeepSpeedPrecisionPlugin(precision, amp_type=None, amp_level=None)[source]
Bases:
pytorch_lightning.plugins.precision.precision_plugin.PrecisionPlugin
Precision plugin for DeepSpeed integration.
- Parameters
precision (
Literal
[‘32’, 32, ‘16’, 16, ‘bf16’]) – Full precision (32), half precision (16) or bfloat16 precision (bf16).- Raises
ValueError – If unsupported
precision
is provided.
- backward(tensor, model, optimizer, optimizer_idx, *args, **kwargs)[source]
Performs back-propagation using DeepSpeed’s engine.
- Parameters
tensor (
Tensor
) – the loss tensormodel (
LightningModule
) – the model to be optimizedoptimizer (
Optional
[Steppable
]) – ignored for DeepSpeed*args – additional positional arguments for the
deepspeed.DeepSpeedEngine.backward()
call**kwargs – additional keyword arguments for the
deepspeed.DeepSpeedEngine.backward()
call
- Return type
- clip_gradients(optimizer, clip_val=0.0, gradient_clip_algorithm=GradClipAlgorithmType.NORM)[source]
DeepSpeed handles gradient clipping internally.
- Return type