Timezone: »
Efficient low-variance gradient estimation enabled by the reparameterization trick (RT) has been essential to the success of variational autoencoders. Doubly-reparameterized gradients (DReGs) improve on the RT for multi-sample variational bounds by applying reparameterization a second time for an additional reduction in variance. Here, we develop two generalizations of the DReGs estimator and show that they can be used to train conditional and hierarchical VAEs on image modelling tasks more effectively. We first extend the estimator to hierarchical models with several stochastic layers by showing how to treat additional score function terms due to the hierarchical variational posterior. We then generalize DReGs to score functions of arbitrary distributions instead of just those of the sampling distribution, which makes the estimator applicable to the parameters of the prior in addition to those of the posterior.
Author Information
Matthias Bauer (DeepMind)
Andriy Mnih (DeepMind)
Related Events (a corresponding poster, oral, or spotlight)
-
2021 Spotlight: Generalized Doubly Reparameterized Gradient Estimators »
Thu. Jul 22nd 12:25 -- 12:30 PM Room
More from the Same Authors
-
2023 Poster: Compositional Score Modeling for Simulation-Based Inference »
Tomas Geffner · George Papamakarios · Andriy Mnih -
2021 Poster: The Lipschitz Constant of Self-Attention »
Hyunjik Kim · George Papamakarios · Andriy Mnih -
2021 Spotlight: The Lipschitz Constant of Self-Attention »
Hyunjik Kim · George Papamakarios · Andriy Mnih -
2021 Poster: Scalable Marginal Likelihood Estimation for Model Selection in Deep Learning »
Alexander Immer · Matthias Bauer · Vincent Fortuin · Gunnar Ratsch · Khan Emtiyaz -
2021 Spotlight: Scalable Marginal Likelihood Estimation for Model Selection in Deep Learning »
Alexander Immer · Matthias Bauer · Vincent Fortuin · Gunnar Ratsch · Khan Emtiyaz -
2018 Poster: Disentangling by Factorising »
Hyunjik Kim · Andriy Mnih -
2018 Oral: Disentangling by Factorising »
Hyunjik Kim · Andriy Mnih