Timezone: »
What learning algorithms can be run directly on compressively-sensed data? In this work, we consider the question of accurately and efficiently computing low-rank matrix or tensor factorizations given data compressed via random projections. We examine the approach of first performing factorization in the compressed domain, and then reconstructing the original high-dimensional factors from the recovered (compressed) factors. In both the matrix and tensor settings, we establish conditions under which this natural approach will provably recover the original factors. While it is well-known that random projections preserve a number of geometric properties of a dataset, our work can be viewed as showing that they can also preserve certain solutions of non-convex, NP-Hard problems like non-negative matrix factorization. We support these theoretical results with experiments on synthetic data and demonstrate the practical applicability of compressed factorization on real-world gene expression and EEG time series datasets.
Author Information
Vatsal Sharan (Stanford University)
Kai Sheng Tai (Stanford University)
Peter Bailis (Stanford University)
Gregory Valiant (Stanford University)
Related Events (a corresponding poster, oral, or spotlight)
-
2019 Oral: Compressed Factorization: Fast and Accurate Low-Rank Factorization of Compressively-Sensed Data »
Wed. Jun 12th 07:00 -- 07:05 PM Room Room 103
More from the Same Authors
-
2021 : Sub-population Guarantees for Importance Weights and KL-Divergence Estimation »
Parikshit Gopalan · Nina Narodytska · Omer Reingold · Vatsal Sharan · Udi Wieder -
2023 Poster: One-sided Matrix Completion from Two Observations Per Row »
Steven Cao · Percy Liang · Greg Valiant -
2021 Poster: Sinkhorn Label Allocation: Semi-Supervised Classification via Annealed Self-Training »
Kai Sheng Tai · Peter Bailis · Gregory Valiant -
2021 Spotlight: Sinkhorn Label Allocation: Semi-Supervised Classification via Annealed Self-Training »
Kai Sheng Tai · Peter Bailis · Gregory Valiant -
2020 Poster: Sample Amplification: Increasing Dataset Size even when Learning is Impossible »
Brian Axelrod · Shivam Garg · Vatsal Sharan · Gregory Valiant -
2019 Poster: LIT: Learned Intermediate Representation Training for Model Compression »
Animesh Koratana · Daniel Kang · Peter Bailis · Matei Zaharia -
2019 Oral: LIT: Learned Intermediate Representation Training for Model Compression »
Animesh Koratana · Daniel Kang · Peter Bailis · Matei Zaharia -
2019 Poster: Equivariant Transformer Networks »
Kai Sheng Tai · Peter Bailis · Gregory Valiant -
2019 Oral: Equivariant Transformer Networks »
Kai Sheng Tai · Peter Bailis · Gregory Valiant -
2019 Poster: Rehashing Kernel Evaluation in High Dimensions »
Paris Siminelakis · Kexin Rong · Peter Bailis · Moses Charikar · Philip Levis -
2019 Poster: Maximum Likelihood Estimation for Learning Populations of Parameters »
Ramya Korlakai Vinayak · Weihao Kong · Gregory Valiant · Sham Kakade -
2019 Oral: Rehashing Kernel Evaluation in High Dimensions »
Paris Siminelakis · Kexin Rong · Peter Bailis · Moses Charikar · Philip Levis -
2019 Oral: Maximum Likelihood Estimation for Learning Populations of Parameters »
Ramya Korlakai Vinayak · Weihao Kong · Gregory Valiant · Sham Kakade -
2017 Poster: Estimating the unseen from multiple populations »
Aditi Raghunathan · Greg Valiant · James Zou -
2017 Poster: Orthogonalized ALS: A Theoretically Principled Tensor Decomposition Algorithm for Practical Use »
Vatsal Sharan · Gregory Valiant -
2017 Talk: Estimating the unseen from multiple populations »
Aditi Raghunathan · Greg Valiant · James Zou -
2017 Talk: Orthogonalized ALS: A Theoretically Principled Tensor Decomposition Algorithm for Practical Use »
Vatsal Sharan · Gregory Valiant