Poster
in
Workshop: Structured Probabilistic Inference and Generative Modeling
Conditional Generative Models are Sufficient to Sample from Any Causal Effect Estimand
Md Musfiqur Rahman · Matt Jordan · Murat Kocaoglu
Keywords: [ causal graphs ] [ Deep Generative Models ] [ Causal Inference ]
Abstract:
Causal inference from observational data has recently found many applications in machine learning. While sound and complete algorithms exist to compute causal effects, many of them assume access to conditional likelihoods, which is difficult to estimate for high-dimensional (particularly image) data. Researchers have alleviated this issue by simulating causal relations with neural models. However, no existing works can effectively deal with causal graphs on image data containing latent confounders, or obtain conditional interventional samples. In this work, we show how to sample from any identifiable interventional distribution given an arbitrary causal graph through a sequence of push-forward computations of conditional generative models, such as diffusion models. Our proposed algorithm follows the recursive steps of the existing likelihood-based identification algorithms to train a set of feed-forward models, and connect them in a specific way to sample from the desired distribution. We conduct experiments on a Colored MNIST dataset having both the treatment ($X$) and the target variables ($Y$) as images and sample from $P(y|do(x))$. Our algorithm also enables us to conduct a causal analysis to evaluate spurious correlations among input features of generative models pre-trained on the CelebA dataset. Finally, we generate high-dimensional interventional samples from the MIMIC-CXR dataset involving text and image variables.
Chat is not available.