RUS  ENG
Full version
JOURNALS // Doklady Rossijskoj Akademii Nauk. Mathematika, Informatika, Processy Upravlenia // Archive

Dokl. RAN. Math. Inf. Proc. Upr., 2023 Volume 514, Number 2, Pages 126–137 (Mi danma458)

SPECIAL ISSUE: ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING TECHNOLOGIES

Activations and gradients compression for model-parallel training

M. I. Rudakovab, A. N. Beznosikovab, Y. A. Kholodovab, A. V. Gasnikovab

a Innopolis University, Innopolis, Republic of Tatarstan
b Moscow Institute of Physics and Technology, Moscow, Russia

Abstract: Large neural networks require enormous computational clusters of machines. Model-parallel training, when the model architecture is partitioned sequentially between workers, is a popular approach for training modern models. Information compression can be applied to decrease workers' communication time, as it is often a bottleneck in such systems. This work explores how simultaneous compression of activations and gradients in model-parallel distributed training setup affects convergence. We analyze compression methods such as quantization and TopK compression, and also experiment with error compensation techniques. Moreover, we employ TopK with AQ-SGD per-batch error feedback approach. We conduct experiments on image classification and language model fine-tuning tasks. Our findings demonstrate that gradients require milder compression rates than activations. We observe that $K=10\%$ is the highest TopK compression level, which does not harm model convergence severely. Experiments also show that models trained with TopK perform well only when compression is also applied during inference. We find that error feedback techniques do not improve model-parallel training compared to plain compression, but allow model inference without compression with almost no quality drop. Finally, when applied with the AQ-SGD approach, TopK stronger than with $K=30\%$ worsens model performance significantly.

Keywords: distributed learning, model parallelism, activation compression, gradient compression.

UDC: 517.54

Presented: A. L. Semenov
Received: 01.09.2023
Revised: 15.09.2023
Accepted: 18.10.2023

DOI: 10.31857/S2686954323601562


 English version:
Doklady Mathematics, 2023, 108:suppl. 2, S272–S281

Bibliographic databases:


© Steklov Math. Inst. of RAS, 2024