Poster
in
Workshop: ICML 2024 Workshop on Foundation Models in the Wild
Not Just Pretty Pictures: Toward Interventional Data Augmentation Using Text-to-Image Generators
Jianhao Yuan · Francesco Pinto · Adam Davies · Phil Torr
Keywords: [ Text-to-Image Generator ] [ synthetic data ] [ robustness ] [ bias ] [ domain generalization ] [ generative model ]
Neural image classifiers are known to undergo severe performance degradation when exposed to inputs that are sampled from environmental conditions that differ from their training data. Given the recent progress in Text-to-Image (T2I) generation, a natural question is how modern T2I generators like Stable Diffusion can be used to simulate arbitrary interventions over such environmental factors in order to augment training data and improve the robustness of downstream classifiers. We experiment across a diverse collection of benchmarks in Single Domain Generalization (SDG), finding that current T2I generators can indeed be used as a powerful interventional data augmentation mechanism, outperforming previously state-of-the-art data augmentation techniques across all datasets. More broadly, our work demonstrates the utility of generative foundation models in synthesizing interventional data that can be used to train more robust machine learning systems, facilitating the application of such technologies in new domains.