Consistent Diffusion Models: Mitigating Sampling Drift by Learning to be Consistent
–Neural Information Processing Systems
Imperfect score-matching leads to a shift between the training and the sampling distribution of diffusion models. Due to the recursive nature of the generation process, errors in previous steps yield sampling iterates that drift away from the training distribution. However, the standard training objective via Denoising Score Matching (DSM) is only designed to optimize over non-drifted data. To train on drifted data, we propose to enforce a Consistency property (CP) which states that predictions of the model on its own generated data are consistent across time. Theoretically, we show that the differential equation that describes CP together with the one that describes a conservative vector field, have a unique solution given some initial condition. Consequently, if the score is learned well on nondrifted points via DSM (enforcing the true initial condition) then enforcing CP on drifted points propagates true score values. Empirically, we show that enforcing CP improves the generation quality for conditional and unconditional generation on CIFAR-10, and in AFHQ and FFHQ.
Neural Information Processing Systems
May-25-2025, 03:22:21 GMT
- Country:
- North America > United States
- California (0.14)
- Texas (0.14)
- North America > United States
- Genre:
- Research Report (0.46)
- Technology: