Timezone: »
Adaptive Federated Learning with Auto-Tuned Clients
J. Lyle Kim · Mohammad Taha Toghani · Cesar Uribe · Anastasios Kyrillidis
Event URL: https://openreview.net/forum?id=Y3VS4LrvNc »
Federated learning (FL) is a distributed machine learning framework where the global model of a central server is trained via multiple collaborative steps by participating clients without sharing their data. While being a flexible framework, where the distribution of local data, participation rate, and computing power of each client can greatly vary, such flexibility gives rise to many new challenges, especially in the hyperparameter tuning on both the server and the client side. We propose $\Delta$-SGD, a simple step size rule for SGD that enables each client to use its own step size by adapting to the local smoothness of the function each client is optimizing. We provide empirical results where the benefit of the client adaptivity is shown in various FL scenarios. In particular, our proposed method achieves TOP-1 accuracy in 73\% and TOP-2 accuracy in 100\% of the experiments considered without additional tuning.
Federated learning (FL) is a distributed machine learning framework where the global model of a central server is trained via multiple collaborative steps by participating clients without sharing their data. While being a flexible framework, where the distribution of local data, participation rate, and computing power of each client can greatly vary, such flexibility gives rise to many new challenges, especially in the hyperparameter tuning on both the server and the client side. We propose $\Delta$-SGD, a simple step size rule for SGD that enables each client to use its own step size by adapting to the local smoothness of the function each client is optimizing. We provide empirical results where the benefit of the client adaptivity is shown in various FL scenarios. In particular, our proposed method achieves TOP-1 accuracy in 73\% and TOP-2 accuracy in 100\% of the experiments considered without additional tuning.
Author Information
J. Lyle Kim (Rice University)
Mohammad Taha Toghani (Rice University)
Cesar Uribe (Rice University)
Anastasios Kyrillidis (Rice University)
More from the Same Authors
-
2021 : Mitigating deep double descent by concatenating inputs »
John Chen · Qihan Wang · Anastasios Kyrillidis -
2022 : PARS-Push: Personalized, Asynchronous and Robust Decentralized Optimization »
Mohammad Taha Toghani · Mohammad Taha Toghani · Soomin Lee · Soomin Lee · Cesar Uribe -
2023 : On First-Order Meta-Reinforcement Learning with Moreau Envelopes »
Mohammad Taha Toghani · Sebastian Perez-Salazar · Cesar Uribe -
2023 : On the Performance of Gradient Tracking with Local Updates »
Edward Duc Hien Nguyen · Sulaiman Alghunaim · Kun Yuan · Cesar Uribe -
2021 Workshop: Beyond first-order methods in machine learning systems »
Albert S Berahas · Anastasios Kyrillidis · Fred Roosta · Amir Gholaminejad · Michael Mahoney · Rachael Tappenden · Raghu Bollapragada · Rixon Crane · J. Lyle Kim -
2021 : Computation-Aware Distributed Optimization over Networks: A Hybrid Dynamical Systems Approach »
Daniel Ochoa · Jorge Poveda · Cesar Uribe -
2021 : Model Reference Adaptive Control for Online Policy Adaptation and Network Synchronization »
Miguel F. Arevalo-Castiblanco · Cesar Uribe · Eduardo Mojica-Nava -
2021 : Population Dynamics for Discrete Wasserstein Gradient Flows over Networks »
Gilberto Díaz-García · Cesar Uribe · Nicanor Quijano -
2021 : Spotlight 2 Q&A »
Cesar Uribe -
2021 : Computation of Discrete Flows Over Networks via Constrained Wasserstein Barycenters »
Ferran Arque · Cesar Uribe -
2020 Workshop: Beyond first order methods in machine learning systems »
Albert S Berahas · Amir Gholaminejad · Anastasios Kyrillidis · Michael Mahoney · Fred Roosta -
2020 Poster: Negative Sampling in Semi-Supervised learning »
John Chen · Vatsal Shah · Anastasios Kyrillidis -
2019 Poster: Compressing Gradient Optimizers via Count-Sketches »
Ryan Spring · Anastasios Kyrillidis · Vijai Mohan · Anshumali Shrivastava -
2019 Oral: Compressing Gradient Optimizers via Count-Sketches »
Ryan Spring · Anastasios Kyrillidis · Vijai Mohan · Anshumali Shrivastava