Poster
Variational Boosting: Iteratively Refining Posterior Approximations
Andrew Miller · Nicholas J Foti · Ryan P. Adams
Gallery #109
We propose a black-box variational inference method to approximate intractable distributions with an increasingly rich approximating class. Our method, variational boosting, iteratively refines an existing variational approximation by solving a sequence of optimization problems, allowing a trade-off between computation time and accuracy. We expand the variational approximating class by incorporating additional covariance structure and by introducing new components to form a mixture. We apply variational boosting to synthetic and real statistical models, and show that the resulting posterior inferences compare favorably to existing variational algorithms.
Live content is unavailable. Log in and register to view live content