Efficient Dictionary Learning with Gradient Descent
Dar Gilboa · Sam Buchanan · John Wright

Wed Jun 12th 02:40 -- 03:00 PM @ Room 104

Randomly initialized first-order optimization algorithms are the method of choice for solving many high-dimensional nonconvex problems in machine learning, yet general theoretical guarantees cannot rule out convergence to critical points of poor objective value. For some highly structured nonconvex problems however, the success of gradient descent can be understood by studying the geometry of the objective. We study one such problem -- complete orthogonal dictionary learning, and provide converge guarantees for randomly initialized gradient descent to the neighborhood of a global optimum. The resulting rates scale as low order polynomials in the dimension even though the objective possesses an exponential number of saddle points. This efficient convergence can be viewed as a consequence of negative curvature normal to the stable manifolds associated with saddle points, and we provide evidence that this feature is shared by other nonconvex problems of importance as well.

Author Information

Dar Gilboa (Columbia University)
Sam Buchanan (Columbia University)
John Wright (Columbia University, USA)

Related Events (a corresponding poster, oral, or spotlight)

More from the Same Authors