Timezone: »
We present a blended conditional gradient approach for minimizing a smooth convex function over a polytope P, combining the Frank–Wolfe algorithm (also called conditional gradient) with gradient-based steps, different from away steps and pairwise steps, but still achieving linear convergence for strongly convex functions, along with good practical performance. Our approach retains all favorable properties of conditional gradient algorithms, notably avoidance of projections onto P and maintenance of iterates as sparse convex combinations of a limited number of extreme points of P. The algorithm is lazy, making use of inexpensive inexact solutions of the linear programming subproblem that characterizes the conditional gradient approach. It decreases measures of optimality (primal and dual gaps) rapidly, both in the number of iterations and in wall-clock time, outperforming even the lazy conditional gradient algorithms of Braun et al. 2017. We also present a streamlined version of the algorithm that applies when P is the probability simplex.
Author Information
Gábor Braun (Georgia Institute of Technology)
Sebastian Pokutta (Georgia Tech)
Dan Tu (GEORGIA TECH)
Stephen Wright (University of Wisconsin-Madison)
Related Events (a corresponding poster, oral, or spotlight)
-
2019 Poster: Blended Conditonal Gradients »
Wed. Jun 12th 01:30 -- 04:00 AM Room Pacific Ballroom #191
More from the Same Authors
-
2023 Oral: A Fully First-Order Method for Stochastic Bilevel Optimization »
Jeongyeol Kwon · Dohyun Kwon · Stephen Wright · Robert Nowak -
2023 Poster: Cyclic Block Coordinate Descent With Variance Reduction for Composite Nonconvex Optimization »
Xufeng Cai · Chaobing Song · Stephen Wright · Jelena Diakonikolas -
2023 Poster: Cut your Losses with Squentropy »
Like Hui · Misha Belkin · Stephen Wright -
2023 Poster: A Fully First-Order Method for Stochastic Bilevel Optimization »
Jeongyeol Kwon · Dohyun Kwon · Stephen Wright · Robert Nowak -
2021 Poster: Variance Reduction via Primal-Dual Accelerated Dual Averaging for Nonsmooth Convex Finite-Sums »
Chaobing Song · Stephen Wright · Jelena Diakonikolas -
2021 Oral: Variance Reduction via Primal-Dual Accelerated Dual Averaging for Nonsmooth Convex Finite-Sums »
Chaobing Song · Stephen Wright · Jelena Diakonikolas -
2019 Poster: First-Order Algorithms Converge Faster than $O(1/k)$ on Convex Problems »
Ching-pei Lee · Stephen Wright -
2019 Poster: Bilinear Bandits with Low-rank Structure »
Kwang-Sung Jun · Rebecca Willett · Stephen Wright · Robert Nowak -
2019 Oral: First-Order Algorithms Converge Faster than $O(1/k)$ on Convex Problems »
Ching-pei Lee · Stephen Wright -
2019 Oral: Bilinear Bandits with Low-rank Structure »
Kwang-Sung Jun · Rebecca Willett · Stephen Wright · Robert Nowak -
2018 Poster: Dissipativity Theory for Accelerating Stochastic Variance Reduction: A Unified Analysis of SVRG and Katyusha Using Semidefinite Programs »
Bin Hu · Stephen Wright · Laurent Lessard -
2018 Oral: Dissipativity Theory for Accelerating Stochastic Variance Reduction: A Unified Analysis of SVRG and Katyusha Using Semidefinite Programs »
Bin Hu · Stephen Wright · Laurent Lessard -
2017 Poster: Lazifying Conditional Gradient Algorithms »
Gábor Braun · Sebastian Pokutta · Daniel Zink -
2017 Poster: Conditional Accelerated Lazy Stochastic Gradient Descent »
Guanghui · Sebastian Pokutta · Yi Zhou · Daniel Zink -
2017 Poster: Emulating the Expert: Inverse Optimization through Online Learning »
Sebastian Pokutta · Andreas Bärmann · Oskar Schneider -
2017 Talk: Conditional Accelerated Lazy Stochastic Gradient Descent »
Guanghui · Sebastian Pokutta · Yi Zhou · Daniel Zink -
2017 Talk: Lazifying Conditional Gradient Algorithms »
Gábor Braun · Sebastian Pokutta · Daniel Zink -
2017 Talk: Emulating the Expert: Inverse Optimization through Online Learning »
Sebastian Pokutta · Andreas Bärmann · Oskar Schneider