M22: Rate-Distortion Inspired Gradient Compression

Yangyi Liu, Sadaf Salehkalaibar, Stefano Rini, Jun Chen

研究成果: Conference article同行評審

摘要

In federated learning (FL), the communication constraint between the remote users and the Parameter Server (PS) is a crucial bottleneck. This paper proposes M22, a rate-distortion inspired approach to model update compression for distributed training of deep neural networks (DNNs). In particular, (i) we propose a family of distortion measures referred to as "M-magnitude weighted L2"norm, and (ii) we assume that gradient updates follow an i.i.d. distribution with two degrees of freedom - generalized normal and Weibull distributions. To measure the gradient compression performance under a communication constraint, we define the per-bit accuracy as the optimal improvement in accuracy that a bit of communication brings to the centralized model over the training period. Using this performance measure, we systematically benchmark the choice of gradient distributions and the distortion measure. We provide substantial insights on the role of these choices and argue that significant performance improvements can be attained using such a rate-distortion inspired compressor.

指紋

深入研究「M22: Rate-Distortion Inspired Gradient Compression」主題。共同形成了獨特的指紋。

引用此