TY - JOUR
T1 - Lazily Aggregated Quantized Gradient Innovation for Communication-Efficient Federated Learning
AU - Sun, Jun
AU - Chen, Tianyi
AU - Giannakis, Georgios B.
AU - Yang, Qinmin
AU - Yang, Zaiyue
N1 - Publisher Copyright:
© 1979-2012 IEEE.
PY - 2022/4/1
Y1 - 2022/4/1
N2 - This paper focuses on communication-efficient federated learning problem, and develops a novel distributed quantized gradient approach, which is characterized by adaptive communications of the quantized gradients. Specifically, the federated learning builds upon the server-worker infrastructure, where the workers calculate local gradients and upload them to the server; then the server obtain the global gradient by aggregating all the local gradients and utilizes it to update the model parameter. The key idea to save communications from the worker to the server is to quantize gradients as well as skip less informative quantized gradient communications by reusing previous gradients. Quantizing and skipping result in 'lazy' worker-server communications, which justifies the term Lazily Aggregated Quantized (LAQ) gradient. Theoretically, the LAQ algorithm achieves the same linear convergence as the gradient descent in the strongly convex case, while effecting major savings in the communication in terms of transmitted bits and communication rounds. Empirically, extensive experiments using realistic data corroborate a significant communication reduction compared with state-of-the-art gradient- and stochastic gradient-based algorithms.
AB - This paper focuses on communication-efficient federated learning problem, and develops a novel distributed quantized gradient approach, which is characterized by adaptive communications of the quantized gradients. Specifically, the federated learning builds upon the server-worker infrastructure, where the workers calculate local gradients and upload them to the server; then the server obtain the global gradient by aggregating all the local gradients and utilizes it to update the model parameter. The key idea to save communications from the worker to the server is to quantize gradients as well as skip less informative quantized gradient communications by reusing previous gradients. Quantizing and skipping result in 'lazy' worker-server communications, which justifies the term Lazily Aggregated Quantized (LAQ) gradient. Theoretically, the LAQ algorithm achieves the same linear convergence as the gradient descent in the strongly convex case, while effecting major savings in the communication in terms of transmitted bits and communication rounds. Empirically, extensive experiments using realistic data corroborate a significant communication reduction compared with state-of-the-art gradient- and stochastic gradient-based algorithms.
KW - Federated learning
KW - communication-efficient
KW - gradient innovation
KW - quantization
UR - http://www.scopus.com/inward/record.url?scp=85125882680&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85125882680&partnerID=8YFLogxK
U2 - 10.1109/TPAMI.2020.3033286
DO - 10.1109/TPAMI.2020.3033286
M3 - Article
C2 - 33095709
AN - SCOPUS:85125882680
SN - 0162-8828
VL - 44
SP - 2031
EP - 2044
JO - IEEE Transactions on Pattern Analysis and Machine Intelligence
JF - IEEE Transactions on Pattern Analysis and Machine Intelligence
IS - 4
ER -