Poster
in
Workshop: Challenges in Deployable Generative AI
One-Step Diffusion Distillation via Deep Equilibrium Models
Zhengyang Geng · Ashwini Pokle · Zico Kolter
Keywords: [ Generative Models ] [ Diffusion Models ] [ Deep Equilibrium Models ] [ Distillation ]
Abstract:
Diffusion models excel at producing high-quality samples but naively require hundreds of iterations, prompting multiple attempts to distill this process into a faster network. Existing approaches, however, often require complex multi-stage distillation and perform sub-optimally in single-step image generation. In response, we introduce a simple yet effective means of diffusion distillation---*directly* mapping initial noise to the resulting image. Of particular importance to our approach is to leverage a new Deep Equilibrium (DEQ) model for distillation: the Generative Equilibrium Transformer (GET). Our method enables fully offline training with just noise/image pairs from the diffusion model while achieving superior performance compared to existing one-step methods on comparable training budgets. The DEQ architecture proves crucial, as GET matches a $5\times$ larger ViT in terms of FID scores while striking a critical balance of computational cost and image quality. Code, checkpoints, and datasets will be released.
Chat is not available.