Goto

Collaborating Authors

 reverse sde


Optimal Transport-Guided Conditional Score-Based Diffusion Model (Appendix) Xiang Gu1, Liwei Y ang

Neural Information Processing Systems

We next explain the rationality of the resampling-by-compatibility presented in Sect. We first i) prove Eq. For Assumption (9), L( π, u, v) is strongly convex as proved in [5]. Eq. (A-19), we have E The codes are in pytorch [7]. The learning rate is 1e-5.




Leveraging Drift to Improve Sample Complexity of Variance Exploding Diffusion Models

Neural Information Processing Systems

Variance exploding (VE) based diffusion models, an important class of diffusion models, have shown state-of-the-art (SOTA) performance. However, only a few theoretical works analyze VE-based models, and those works suffer from a worse forward convergence rate $1/\text{poly}(T)$ than the $\exp{(-T)}$ of variance preserving (VP) based models, where $T$ is the forward diffusion time and the rate measures the distance between forward marginal distribution $q_T$ and pure Gaussian noise. The slow rate is due to the Brownian Motion without a drift term. In this work, we design a new drifted VESDE forward process, which allows a faster $\exp{(-T)}$ forward convergence rate. With this process, we achieve the first efficient polynomial sample complexity for a series of VE-based models with reverse SDE under the manifold hypothesis. Furthermore, unlike previous works, we allow the diffusion coefficient to be unbounded instead of a constant, which is closer to the SOTA models. Besides the reverse SDE, the other common reverse process is the probability flow ODE (PFODE) process, which is deterministic and enjoys faster sample speed. To deepen the understanding of VE-based models, we consider a more general setting considering reverse SDE and PFODE simultaneously, propose a unified tangent-based analysis framework, and prove the first quantitative convergence guarantee for SOTA VE-based models with reverse PFODE.We also show that the drifted VESDE can balance different error terms and improve generated samples without training through synthetic and real-world experiments.



Optimal Transport-Guided Conditional Score-Based Diffusion Model (Appendix) Xiang Gu1, Liwei Y ang

Neural Information Processing Systems

We next explain the rationality of the resampling-by-compatibility presented in Sect. We first i) prove Eq. For Assumption (9), L( π, u, v) is strongly convex as proved in [5]. Eq. (A-19), we have E The codes are in pytorch [7]. The learning rate is 1e-5.





Leveraging Drift to Improve Sample Complexity of Variance Exploding Diffusion Models

Neural Information Processing Systems

Variance exploding (VE) based diffusion models, an important class of diffusion models, have shown state-of-the-art (SOTA) performance. However, only a few theoretical works analyze VE-based models, and those works suffer from a worse forward convergence rate 1/\text{poly}(T) than the \exp{(-T)} of variance preserving (VP) based models, where T is the forward diffusion time and the rate measures the distance between forward marginal distribution q_T and pure Gaussian noise. The slow rate is due to the Brownian Motion without a drift term. In this work, we design a new drifted VESDE forward process, which allows a faster \exp{(-T)} forward convergence rate. With this process, we achieve the first efficient polynomial sample complexity for a series of VE-based models with reverse SDE under the manifold hypothesis.