TY - GEN
T1 - On the Necessity of Aligning Gradients for Wireless Federated Learning
AU - Chang, Wei Ting
AU - Eldin Mohamed, Mohamed Seif
AU - Tandon, Ravi
N1 - Funding Information:
This work has been supported in part by NSF Grants CAREER 1651492, CNS 1715947, CCF 2100013 and the 2018 Keysight Early Career Professor Award.
Publisher Copyright:
© 2021 IEEE.
PY - 2021
Y1 - 2021
N2 - In this paper, we consider the problem of wireless federated learning, where the users wish to jointly train a machine learning model with the help of a parameter server. During the training, the local gradients from the users are aggregated over a wireless channel. Typically, coefficients of the local gradients are aligned by power control techniques to ensure that the estimated aggregated gradient is an unbiased estimator of the true gradient. However, schemes that align gradients require coordination, can be challenging to implement in practice, and often lead to degraded performance due to heterogeneity of users' channels. In this paper, we show that alignment of gradients for wireless FL is not always necessary for convergence. Specifically, we consider non-convex loss functions, and derive conditions under which misaligned wireless gradient aggregation still converges to a stationary point. We also present experimental results to show that transmitting at full power can outperform aligned gradient aggregation depending on the heterogeneity of users' channels.
AB - In this paper, we consider the problem of wireless federated learning, where the users wish to jointly train a machine learning model with the help of a parameter server. During the training, the local gradients from the users are aggregated over a wireless channel. Typically, coefficients of the local gradients are aligned by power control techniques to ensure that the estimated aggregated gradient is an unbiased estimator of the true gradient. However, schemes that align gradients require coordination, can be challenging to implement in practice, and often lead to degraded performance due to heterogeneity of users' channels. In this paper, we show that alignment of gradients for wireless FL is not always necessary for convergence. Specifically, we consider non-convex loss functions, and derive conditions under which misaligned wireless gradient aggregation still converges to a stationary point. We also present experimental results to show that transmitting at full power can outperform aligned gradient aggregation depending on the heterogeneity of users' channels.
KW - Federated learning
KW - Stochastic Gradient Descent
KW - Wireless Aggregation
UR - http://www.scopus.com/inward/record.url?scp=85122797784&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85122797784&partnerID=8YFLogxK
U2 - 10.1109/SPAWC51858.2021.9593177
DO - 10.1109/SPAWC51858.2021.9593177
M3 - Conference contribution
AN - SCOPUS:85122797784
T3 - IEEE Workshop on Signal Processing Advances in Wireless Communications, SPAWC
SP - 226
EP - 230
BT - 2021 IEEE 22nd International Workshop on Signal Processing Advances in Wireless Communications, SPAWC 2021
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 22nd IEEE International Workshop on Signal Processing Advances in Wireless Communications, SPAWC 2021
Y2 - 27 September 2021 through 30 September 2021
ER -