Skip to yearly menu bar Skip to main content


Poster
in
Workshop: ICML 2024 Workshop on Foundation Models in the Wild

On the Discrepancy and Connection between Memorization and Generation in Diffusion Models

Hanyu Wang · Yujin Han · Difan Zou

Keywords: [ Diffusion Models ] [ memorization ] [ trained score ] [ oracle score ] [ generalizability ]


Abstract:

Diffusion models (DMs), as a state-of-the-art generative modeling method, have enjoyed tremendous success in multiple generating tasks. However, the memorization behavior of DMs, that the generation replicates the training data, raises serious privacy concerns and contradicts the actual generalizability of DMs. These prompt us to delve deeper into the generalizability and memorization of DMs, particularly in cases where the closed-form solution of DMs' score function can be explicitly solved. Through a series of comprehensive experiments, we demonstrate the discrepancies and connections between the optimal score and the trained score, noting that the trained one is smoother, which benefits the generalizability of DMs. We also further explore how mixing the optimal score with the trained score during the sampling phase affects generation. Our experimental findings provide novel insights into the understanding of DMs' generalizability.

Chat is not available.