On Forgetting and Stability of Score-based Generative models
Strasman, Stanislas, Cardoso, Gabriel, Corff, Sylvain Le, Lemaire, Vincent, Ocello, Antonio
Understanding the stability and long-time behavior of generative models is a fundamental problem in modern machine learning. This paper provides quantitative bounds on the sampling error of score-based generative models by leveraging stability and forgetting properties of the Markov chain associated with the reverse-time dynamics. Under weak assumptions, we provide the two structural properties to ensure the propagation of initialization and discretization errors of the backward process: a Lyapunov drift condition and a Doeblin-type minorization condition. A practical consequence is quantitative stability of the sampling procedure, as the reverse diffusion dynamics induces a contraction mechanism along the sampling trajectory. Our results clarify the role of stochastic dynamics in score-based models and provide a principled framework for analyzing propagation of errors in such approaches.
Jan-30-2026
- Country:
- Asia > Afghanistan
- Parwan Province > Charikar (0.04)
- Europe
- France
- Hauts-de-France > Nord
- Lille (0.04)
- Île-de-France > Paris
- Paris (0.04)
- Hauts-de-France > Nord
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- France
- Asia > Afghanistan
- Genre:
- Research Report (0.70)
- Technology: