Timezone: »
Techniques for reducing the variance of gradient estimates used in stochastic programming algorithms for convex finite-sum problems have received a great deal of attention in recent years. By leveraging dissipativity theory from control, we provide a new perspective on two important variance-reduction algorithms: SVRG and its direct accelerated variant Katyusha. Our perspective provides a physically intuitive understanding of the behavior of SVRG-like methods via a principle of energy conservation. The tools discussed here allow us to automate the convergence analysis of SVRG-like methods by capturing their essential properties in small semidefinite programs amenable to standard analysis and computational techniques. Our approach recovers existing convergence results for SVRG and Katyusha and generalizes the theory to alternative parameter choices. We also discuss how our approach complements the linear coupling technique. Our combination of perspectives leads to a better understanding of accelerated variance-reduced stochastic methods for finite-sum problems.
Author Information
Bin Hu (University of Wisconsin-Madison)
Stephen Wright (University of Wisconsin-Madison)
Laurent Lessard (University of Wisconsin-Madison)
Related Events (a corresponding poster, oral, or spotlight)
-
2018 Poster: Dissipativity Theory for Accelerating Stochastic Variance Reduction: A Unified Analysis of SVRG and Katyusha Using Semidefinite Programs »
Wed. Jul 11th 04:15 -- 07:00 PM Room Hall B #137
More from the Same Authors
-
2023 Oral: A Fully First-Order Method for Stochastic Bilevel Optimization »
Jeongyeol Kwon · Dohyun Kwon · Stephen Wright · Robert Nowak -
2023 Poster: Cyclic Block Coordinate Descent With Variance Reduction for Composite Nonconvex Optimization »
Xufeng Cai · Chaobing Song · Stephen Wright · Jelena Diakonikolas -
2023 Poster: Cut your Losses with Squentropy »
Like Hui · Misha Belkin · Stephen Wright -
2023 Poster: A Fully First-Order Method for Stochastic Bilevel Optimization »
Jeongyeol Kwon · Dohyun Kwon · Stephen Wright · Robert Nowak -
2021 Poster: Variance Reduction via Primal-Dual Accelerated Dual Averaging for Nonsmooth Convex Finite-Sums »
Chaobing Song · Stephen Wright · Jelena Diakonikolas -
2021 Oral: Variance Reduction via Primal-Dual Accelerated Dual Averaging for Nonsmooth Convex Finite-Sums »
Chaobing Song · Stephen Wright · Jelena Diakonikolas -
2019 Poster: First-Order Algorithms Converge Faster than $O(1/k)$ on Convex Problems »
Ching-pei Lee · Stephen Wright -
2019 Poster: Bilinear Bandits with Low-rank Structure »
Kwang-Sung Jun · Rebecca Willett · Stephen Wright · Robert Nowak -
2019 Oral: First-Order Algorithms Converge Faster than $O(1/k)$ on Convex Problems »
Ching-pei Lee · Stephen Wright -
2019 Oral: Bilinear Bandits with Low-rank Structure »
Kwang-Sung Jun · Rebecca Willett · Stephen Wright · Robert Nowak -
2019 Poster: Blended Conditonal Gradients »
Gábor Braun · Sebastian Pokutta · Dan Tu · Stephen Wright -
2019 Oral: Blended Conditonal Gradients »
Gábor Braun · Sebastian Pokutta · Dan Tu · Stephen Wright -
2018 Poster: Lyapunov Functions for First-Order Methods: Tight Automated Convergence Guarantees »
Adrien Taylor · Bryan Van Scoy · Laurent Lessard -
2018 Oral: Lyapunov Functions for First-Order Methods: Tight Automated Convergence Guarantees »
Adrien Taylor · Bryan Van Scoy · Laurent Lessard -
2017 Poster: Dissipativity Theory for Nesterov's Accelerated Method »
Bin Hu · Laurent Lessard -
2017 Talk: Dissipativity Theory for Nesterov's Accelerated Method »
Bin Hu · Laurent Lessard