Timezone: »
In this paper, we consider nonconvex minimax optimization, which is gaining prominence in many modern machine learning applications, such as GANs. Large-scale edge-based collection of training data in these applications calls for communication-efficient distributed optimization algorithms, such as those used in federated learning, to process the data. In this paper, we analyze local stochastic gradient descent ascent (SGDA), the local-update version of the SGDA algorithm. SGDA is the core algorithm used in minimax optimization, but it is not well-understood in a distributed setting. We prove that Local SGDA has \textit{order-optimal} sample complexity for several classes of nonconvex-concave and nonconvex-nonconcave minimax problems, and also enjoys \textit{linear speedup} with respect to the number of clients. We provide a novel and tighter analysis, which improves the convergence and communication guarantees in the existing literature. For nonconvex-PL and nonconvex-one-point-concave functions, we improve the existing complexity results for centralized minimax problems. Furthermore, we propose a momentum-based local-update algorithm, which has the same convergence guarantees, but outperforms Local SGDA as demonstrated in our experiments.
Author Information
PRANAY SHARMA (CARNEGIE MELLON UNIVERSITY)
I am a postdoctoral researcher in the Dept. of Electrical and Computer Engineering, at Carnegie Mellon University. I'm working with Prof. Gauri Joshi. In August 2021, I finished my Ph.D. in Electrical Engineering and Computer Science at Syracuse University. My advisor was Prof. Pramod K. Varshney. I finished my B.Tech-M.Tech dual-degree in Electrical Engineering from IIT Kanpur.
Rohan Panda (Carnegie Mellon University)
Gauri Joshi (Carnegie Mellon University)
Pramod K Varshney (Syracuse University)
Related Events (a corresponding poster, oral, or spotlight)
-
2022 Poster: Federated Minimax Optimization: Improved Convergence Analyses and Algorithms »
Tue. Jul 19th through Wed the 20th Room Hall E #605
More from the Same Authors
-
2023 : Towards a Theoretical and Practical Understanding of One-Shot Federated Learning with Fisher Information »
Divyansh Jhunjhunwala · Shiqiang Wang · Gauri Joshi -
2023 Poster: The Blessing of Heterogeneity in Federated Q-Learning: Linear Speedup and Beyond »
Jiin Woo · Gauri Joshi · Yuejie Chi -
2023 Poster: On the Convergence of Federated Averaging with Cyclic Client Participation »
Yae Jee Cho · PRANAY SHARMA · Gauri Joshi · Zheng Xu · Satyen Kale · Tong Zhang -
2022 Poster: Federated Reinforcement Learning: Linear Speedup Under Markovian Sampling »
sajad khodadadian · PRANAY SHARMA · Gauri Joshi · Siva Maguluri -
2022 Oral: Federated Reinforcement Learning: Linear Speedup Under Markovian Sampling »
sajad khodadadian · PRANAY SHARMA · Gauri Joshi · Siva Maguluri -
2021 : Closing Remarks »
Shiqiang Wang · Nathalie Baracaldo · Olivia Choudhury · Gauri Joshi · Peter Richtarik · Praneeth Vepakomma · Han Yu -
2021 Workshop: International Workshop on Federated Learning for User Privacy and Data Confidentiality in Conjunction with ICML 2021 (FL-ICML'21) »
Nathalie Baracaldo · Olivia Choudhury · Gauri Joshi · Peter Richtarik · Praneeth Vepakomma · Shiqiang Wang · Han Yu -
2021 : Opening Remarks »
Shiqiang Wang · Nathalie Baracaldo · Olivia Choudhury · Gauri Joshi · Peter Richtarik · Praneeth Vepakomma · Han Yu -
2020 : Closing remarks »
Nathalie Baracaldo · Olivia Choudhury · Gauri Joshi · Ramesh Raskar · Shiqiang Wang · Han Yu -
2020 : Opening remarks »
Nathalie Baracaldo · Olivia Choudhury · Gauri Joshi · Ramesh Raskar · Shiqiang Wang · Han Yu -
2020 Workshop: Federated Learning for User Privacy and Data Confidentiality »
Nathalie Baracaldo · Olivia Choudhury · Olivia Choudhury · Gauri Joshi · Ramesh Raskar · Gauri Joshi · Shiqiang Wang · Han Yu -
2019 Workshop: Coding Theory For Large-scale Machine Learning »
Viveck Cadambe · Pulkit Grover · Dimitris Papailiopoulos · Gauri Joshi -
2017 Poster: Convergence Analysis of Proximal Gradient with Momentum for Nonconvex Optimization »
Qunwei Li · Yi Zhou · Yingbin Liang · Pramod K Varshney -
2017 Talk: Convergence Analysis of Proximal Gradient with Momentum for Nonconvex Optimization »
Qunwei Li · Yi Zhou · Yingbin Liang · Pramod K Varshney