Skip to yearly menu bar Skip to main content


Convergence and Recovery Guarantees of the K-Subspaces Method for Subspace Clustering

Peng Wang · Huikang Liu · Anthony Man-Cho So · Laura Balzano

Hall E #611

Keywords: [ T: Optimization ] [ PM: Spectral Methods ] [ MISC: Unsupervised and Semi-supervised Learning ] [ OPT: Non-Convex ]

Abstract: The K-subspaces (KSS) method is a generalization of the K-means method for subspace clustering. In this work, we present local convergence analysis and a recovery guarantee for KSS, assuming data are generated by the semi-random union of subspaces model, where $N$ points are randomly sampled from $K \ge 2$ overlapping subspaces. We show that if the initial assignment of the KSS method lies within a neighborhood of a true clustering, it converges at a superlinear rate and finds the correct clustering within $\Theta(\log\log N)$ iterations with high probability. Moreover, we propose a thresholding inner-product based spectral method for initialization and prove that it produces a point in this neighborhood. We also present numerical results of the studied method to support our theoretical developments.

Chat is not available.