Timezone: »
We introduce HyperGAN, a generative model that learns to generate all the parameters of a deep neural network. HyperGAN first transforms low dimensional noise into a latent space, which can be sampled from to obtain diverse, performant sets of parameters for a target architecture. We utilize an architecture that bears resemblance to generative adversarial networks, but we evaluate the likelihood of generated samples with a classification loss. This is equivalent to minimizing the KL-divergence between the distribution of generated parameters, and the unknown true parameter distribution. We apply HyperGAN to classification, showing that HyperGAN can learn to generate parameters which solve the MNIST and CIFAR-10 datasets with competitive performance to fully supervised learning, while also generating a rich distribution of effective parameters. We also show that HyperGAN can also provide better uncertainty estimates than standard ensembles. This is evidenced by the ability of HyperGAN-generated ensembles to detect out of distribution data as well as adversarial examples.
Author Information
Neale Ratzlaff (Oregon State University)
Want to do representation learning for general AI systems.
Fuxin Li (Oregon State University)
Related Events (a corresponding poster, oral, or spotlight)
-
2019 Oral: HyperGAN: A Generative Model for Diverse, Performant Neural Networks »
Wed. Jun 12th 12:15 -- 12:20 AM Room Room 201
More from the Same Authors
-
2021 Poster: Generative Particle Variational Inference via Estimation of Functional Gradients »
Neale Ratzlaff · Jerry Bai · Fuxin Li · Wei Xu -
2021 Spotlight: Generative Particle Variational Inference via Estimation of Functional Gradients »
Neale Ratzlaff · Jerry Bai · Fuxin Li · Wei Xu -
2020 Poster: Implicit Generative Modeling for Efficient Exploration »
Neale Ratzlaff · Qinxun Bai · Fuxin Li · Wei Xu