Diffusion Probabilistic Models Generalize when They Fail to Memorize
TaeHo Yoon · Joo Young Choi · Sehyun Kwon · Ernest Ryu
Keywords:
Diffusion models; Diffusion probabilistic models; Data replication; Memorization; Generalization
Abstract
In this work, we study the training of diffusion probabilistic models through a series of hypotheses and carefully designed experiments. We call our key finding the memorization-generalization dichotomy, and it asserts that generalization and memorization are mutually exclusive phenomena. This contrasts with the modern wisdom of supervised learning that deep neural networks exhibit "benign" overfitting and generalize well despite overfitting the data.
Video
Chat is not available.
Successful Page Load