TY - GEN
T1 - Pruning Analog Over-the-Air Distributed Learning Models with Accuracy Loss Guarantee
AU - Xu, Kailei
AU - Yang, Howard H.
AU - Zhao, Zhongyuan
AU - Hong, Wei
AU - Quek, Tony Q.S.
AU - Peng, Mugen
N1 - Funding Information:
VII. ACKNOWLEDGMENT This work was supported in part by National Natural Science Foundation of China (Grant 61971061), Beijing Natural Science Foundation (Grant L182039), and Zhejiang Provincial Natural Science Foundation of China (Grant LGJ22F010001).
Publisher Copyright:
© 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - Analog over-the-air computing enables a swarm of end-user devices to efficiently conduct distributed learning, where the intermediate parameters of users, such as gradients, are modulated and transmitted via a group of orthogonal waveforms, and can be mixed directly at a server without individually detecting the feedback parameters of each user. Nonetheless, the scarcity of orthogonal waveforms, as well as communication resources of the end-user devices, are throttling this paradigm in adopting complex deep learning models. To balance the tradeoff between communication efficiency and accuracy performance, we study model pruning for analog over-the-air distributed learning in this paper. First, a model pruning scheme is proposed to improve the communication efficiency of analog over-the-air training. An importance measure for model parameter pruning is also designed based on the analog over-the-air aggregated gradient, which can characterize the contribution of each parameter without removing channel fading and electromagnetic interference. Second, an analytical expression of the training error upper bound is derived, which shows the proposed scheme is able to converge even when the aggregated gradient is corrupted by heavy-tailed electromagnetic interference with an infinite variance. Finally, several experimental results are provided to show the performance gains achieved by our proposed scheme, and also verify the correctness of analytical results.
AB - Analog over-the-air computing enables a swarm of end-user devices to efficiently conduct distributed learning, where the intermediate parameters of users, such as gradients, are modulated and transmitted via a group of orthogonal waveforms, and can be mixed directly at a server without individually detecting the feedback parameters of each user. Nonetheless, the scarcity of orthogonal waveforms, as well as communication resources of the end-user devices, are throttling this paradigm in adopting complex deep learning models. To balance the tradeoff between communication efficiency and accuracy performance, we study model pruning for analog over-the-air distributed learning in this paper. First, a model pruning scheme is proposed to improve the communication efficiency of analog over-the-air training. An importance measure for model parameter pruning is also designed based on the analog over-the-air aggregated gradient, which can characterize the contribution of each parameter without removing channel fading and electromagnetic interference. Second, an analytical expression of the training error upper bound is derived, which shows the proposed scheme is able to converge even when the aggregated gradient is corrupted by heavy-tailed electromagnetic interference with an infinite variance. Finally, several experimental results are provided to show the performance gains achieved by our proposed scheme, and also verify the correctness of analytical results.
UR - http://www.scopus.com/inward/record.url?scp=85137262201&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85137262201&partnerID=8YFLogxK
U2 - 10.1109/ICC45855.2022.9839266
DO - 10.1109/ICC45855.2022.9839266
M3 - Conference contribution
AN - SCOPUS:85137262201
T3 - IEEE International Conference on Communications
SP - 5202
EP - 5207
BT - ICC 2022 - IEEE International Conference on Communications
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2022 IEEE International Conference on Communications, ICC 2022
Y2 - 16 May 2022 through 20 May 2022
ER -