Training on unreliable mixed GPUs across the internet (Expert)¶
Using Compression to Optimize Communications¶
Below are some ways to reduce communication when training collaboratively. As the size of your model increase, bottlenecks in communication become more apparent.
Compress Gradients & State¶
Hivemind allows you to compress gradients and states before sending them to other machines. This helps reduce the communication overhead substantially when training across the internet.
Below, we enable Float16 compression, which compresses gradients and states to Float16 before sending it to other machines.
Note
Compressing gradients can affect convergence if you’re lowering the precision (i.e training in Float32, but compressing gradients to FP16).
from hivemind import Float16Compression
import pytorch_lightning as pl
from pytorch_lightning.strategies import HivemindStrategy
trainer = pl.Trainer(
strategy=HivemindStrategy(
target_batch_size=target_batch_size,
grad_compression=Float16Compression(),
state_averaging_compression=Float16Compression(),
),
accelerator="gpu",
devices=1,
)
A slightly more advanced scheme is dynamic compression based on value size. Below, we enable 8-bit quantization for large numbers, and Float16 compression for small values, reducing communication bottlenecks even further.
Size Adaptive Compression has been used in a variety of Hivemind applications and has shown success, but does quantize gradients further, meaning we lose precision when compressing.
from hivemind import Float16Compression, Uniform8BitQuantization
import pytorch_lightning as pl
from pytorch_lightning.strategies import HivemindStrategy
# compresses values above threshold with 8bit Quantization, lower with Float16
compression = SizeAdaptiveCompression(
threshold=2**16 + 1, less=Float16Compression(), greater_equal=Uniform8BitQuantization()
)
trainer = pl.Trainer(
strategy=HivemindStrategy(
target_batch_size=target_batch_size,
grad_compression=compression,
state_averaging_compression=compression,
),
accelerator="gpu",
devices=1,
)
PowerSGD¶
PowerSGD is a technique to reduce distributed communication of gradients across processes. In short, PowerSGD uses a low-rank approximation to compress gradients before running an all-reduce step to sync gradients across all processes.
Note
Though PowerSGD can impact convergence, it can also substantially reduce communication between processes.
import pytorch_lightning as pl
from pytorch_lightning.strategies import HivemindStrategy
from functools import partial
from hivemind.optim.power_sgd_averager import PowerSGDGradientAverager
trainer = pl.Trainer(
strategy=HivemindStrategy(
target_batch_size=8192,
grad_averager_factory=partial(PowerSGDGradientAverager, averager_rank=32, min_compression_ratio=0.5),
),
accelerator="gpu",
devices=1,
)