An Expectation-Maximization Algorithm for Training Clean Diffusion Models from Corrupted Observations Weimin Bai Yifei Wang 4 Wenzheng Chen 5,6 He Sun

Neural Information Processing Systems 

Diffusion models excel in solving imaging inverse problems due to their ability to model complex image priors. However, their reliance on large, clean datasets for training limits their practical use where clean data is scarce. In this paper, we propose EMDiffusion, an expectation-maximization (EM) approach to train diffusion models from corrupted observations. Our method alternates between reconstructing clean images from corrupted data using a known diffusion model (Estep) and refining diffusion model weights based on these reconstructions (M-step). This iterative process leads the learned diffusion model to gradually converge to a local optimum, that is, to approximate the true clean data distribution.