Timezone: »
The large memory requirements of deep neural networks limit their deployment and adoption on many devices.Model compression methods effectively reduce the memory requirements of these models,usually through applying transformations such as weight pruning or quantization.In this paper, we present a novel scheme for lossy weight encoding co-designed with weight simplification techniques.The encoding is based on the Bloomier filter,a probabilistic data structure that can save space at the cost of introducing random errors.Leveraging the ability of neural networks to tolerate these imperfections and by re-training around the errors,the proposed technique, named Weightless, can compress weights by up to 496xwithout loss of model accuracy.This results in up to a 1.51x improvement over the state-of-the-art.
Author Information
Brandon Reagen (Harvard University)
Udit Gupta (Harvard University)
Bob Adolf (Harvard University)
Michael Mitzenmacher (Harvard University)
Alexander Rush (Harvard University)
Gu-Yeon Wei
David Brooks (Harvard University)
Related Events (a corresponding poster, oral, or spotlight)
-
2018 Poster: Weightless: Lossy weight encoding for deep neural network compression »
Fri. Jul 13th 04:15 -- 07:00 PM Room Hall B #138
More from the Same Authors
-
2021 Poster: Gradient Disaggregation: Breaking Privacy in Federated Learning by Reconstructing the User Participant Matrix »
Maximilian Lam · Gu-Yeon Wei · David Brooks · Vijay Janapa Reddi · Michael Mitzenmacher -
2021 Oral: Gradient Disaggregation: Breaking Privacy in Federated Learning by Reconstructing the User Participant Matrix »
Maximilian Lam · Gu-Yeon Wei · David Brooks · Vijay Janapa Reddi · Michael Mitzenmacher -
2019 Poster: Latent Normalizing Flows for Discrete Sequences »
Zachary Ziegler · Alexander Rush -
2019 Oral: Latent Normalizing Flows for Discrete Sequences »
Zachary Ziegler · Alexander Rush -
2019 Poster: Tensor Variable Elimination for Plated Factor Graphs »
Fritz Obermeyer · Elias Bingham · Martin Jankowiak · Neeraj Pradhan · Justin Chiu · Alexander Rush · Noah Goodman -
2019 Oral: Tensor Variable Elimination for Plated Factor Graphs »
Fritz Obermeyer · Elias Bingham · Martin Jankowiak · Neeraj Pradhan · Justin Chiu · Alexander Rush · Noah Goodman -
2018 Poster: Semi-Amortized Variational Autoencoders »
Yoon Kim · Sam Wiseman · Andrew Miller · David Sontag · Alexander Rush -
2018 Poster: Adversarially Regularized Autoencoders »
Jake Zhao · Yoon Kim · Kelly Zhang · Alexander Rush · Yann LeCun -
2018 Oral: Semi-Amortized Variational Autoencoders »
Yoon Kim · Sam Wiseman · Andrew Miller · David Sontag · Alexander Rush -
2018 Oral: Adversarially Regularized Autoencoders »
Jake Zhao · Yoon Kim · Kelly Zhang · Alexander Rush · Yann LeCun -
2017 Poster: Image-to-Markup Generation with Coarse-to-Fine Attention »
Yuntian Deng · Anssi Kanervisto · Jeffrey Ling · Alexander Rush -
2017 Talk: Image-to-Markup Generation with Coarse-to-Fine Attention »
Yuntian Deng · Anssi Kanervisto · Jeffrey Ling · Alexander Rush