Scaling Graphically Structured Diffusion Models
Christian Weilbach · William Harvey · Hamed Shirzad · Frank Wood
Keywords:
Graphical Models
Diffusion Models
Graph Neural Networks
Bayesian Deep Learning
scaling
sparse attention
Abstract
Applications of the recently introduced graphically structured diffusion model (GSDM) family show that sparsifying the transformer attention mechanism within a diffusion model and meta-training on a variety of conditioning tasks can yield an efficiently learnable diffusion model artifact that is capable of flexible, in the sense of observing different subsets of variables at test-time, amortized conditioning in probabilistic graphical models. While extremely promising in terms of applicability and utility, implementations of GSDMs prior to this work were not scalable beyond toy graphical model sizes. We overcome this limitation by describing and and solving two scaling issues related to GSDMs; one engineering and one methodological. We additionally propose a new benchmark problem of weight inference for a convolutional neural network applied to $14\times14$ MNIST.
Video
Chat is not available.
Successful Page Load