Abstract:
The Johnson-Lindenstaruss lemma (Johnson & Lindenstrauss, 1984) is a cornerstone result in dimensionality reduction, stating it is possible to embed a set of nn points in dd-dimensional Euclidean space into optimal k=O(ε−2lnn)k=O(ε−2lnn) dimensions, while preserving all pairwise distances to within a factor (1±ε)(1±ε). The seminal Fast Johnson-Lindenstrauss (Fast JL) transform by Ailon and Chazelle (SICOMP'09) supports computing the embedding of a data point in O(dlnd+kln2n)O(dlnd+kln2n) time, where the dlnddlnd term comes from multiplication with a d×dd×d Hadamard matrix and the kln2nkln2n term comes from multiplication with a sparse k×dk×d matrix. Despite the Fast JL transform being more than a decade old, it is one of the fastest dimensionality reduction techniques for many tradeoffs between ε,dε,d and nn. In this work, we give a surprising new analysis of the Fast JL transform, showing that the kln2nkln2n term in the embedding time can be improved to (kln2n)/α(kln2n)/α for an α=Ω(min{ε−1ln(1/ε),lnn})α=Ω(min{ε−1ln(1/ε),lnn}). The improvement follows by using an even sparser matrix. We complement our improved analysis with a lower bound showing that our new analysis is in fact tight.
Chat is not available.