TY - GEN
T1 - On the Convergence Rate of Federated Learning over Unreliable Networks
AU - Feng, Chenyuan
AU - Yang, Howard H.
AU - Chen, Zihan
AU - Feng, Daquan
AU - Wang, Zhenzhong
AU - Quek, Tony Q.S.
N1 - Funding Information:
This work was supported in part by the National Key R&D Program of China under Grant 2020YFB1807601, and the Innovation Project of Guangdong Educational Department under Grant 2019KTSCX147. The corresponding author is D. Feng (fdquan@szu.edu.cn).
Publisher Copyright:
© 2021 IEEE.
PY - 2021
Y1 - 2021
N2 - Considering unreliable wireless networks, the efficiency of training a statistical model via Federated Learning (FL) over wireless networks is studied in this work. The system consists of a server and a swarm of end-user devices that are termed as clients. The server allots the statistical model directly onto the clients for local training, based on the dataset that resides on their devices, periodically collects the trained parameters to produce an improved model, and sends it back to the clients for a new round of local computing. This process repeats for multiple rounds until the global model converges. Due to unreliable wireless channels and limited communication resources, the server can only select a handful of clients for parameter updates at each iteration. Analytical expressions are derived to characterize the FL convergence rate by accounting for the transmission reliability and momentum method. The analyses unveil that in networks with reliable connections, model training can be accelerated by expanding the bandwidth to include more clients in each communication round, but it becomes ineffectual when the connections are unreliable. Moreover, it is also confirmed that adopting momentum in the global aggregation stage of FL improves the convergence rate, and such an approach is more insulated from the effects of communication failures. These theoretical findings are validated by extensive empirical simulations.
AB - Considering unreliable wireless networks, the efficiency of training a statistical model via Federated Learning (FL) over wireless networks is studied in this work. The system consists of a server and a swarm of end-user devices that are termed as clients. The server allots the statistical model directly onto the clients for local training, based on the dataset that resides on their devices, periodically collects the trained parameters to produce an improved model, and sends it back to the clients for a new round of local computing. This process repeats for multiple rounds until the global model converges. Due to unreliable wireless channels and limited communication resources, the server can only select a handful of clients for parameter updates at each iteration. Analytical expressions are derived to characterize the FL convergence rate by accounting for the transmission reliability and momentum method. The analyses unveil that in networks with reliable connections, model training can be accelerated by expanding the bandwidth to include more clients in each communication round, but it becomes ineffectual when the connections are unreliable. Moreover, it is also confirmed that adopting momentum in the global aggregation stage of FL improves the convergence rate, and such an approach is more insulated from the effects of communication failures. These theoretical findings are validated by extensive empirical simulations.
UR - http://www.scopus.com/inward/record.url?scp=85124589400&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85124589400&partnerID=8YFLogxK
U2 - 10.1109/ComComAp53641.2021.9652996
DO - 10.1109/ComComAp53641.2021.9652996
M3 - Conference contribution
AN - SCOPUS:85124589400
T3 - 2021 Computing, Communications and IoT Applications, ComComAp 2021
SP - 59
EP - 64
BT - 2021 Computing, Communications and IoT Applications, ComComAp 2021
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2021 Computing, Communications and IoT Applications, ComComAp 2021
Y2 - 26 November 2021 through 28 November 2021
ER -