Accelerating Diffusion Models with Parallel Sampling: Inference at Sub-Linear Time Complexity
Chen, Haoxuan, Ren, Yinuo, Ying, Lexing, Rotskoff, Grant M.
Diffusion models have become a leading method for generativ e modeling of both image and scientific data. As these models are costly to train and evaluate, reducing the inference cost for diffusion models remains a maj or goal. Inspired by the recent empirical success in accelerating diffusion mod els via the parallel sampling technique [1], we propose to divide the sampling proce ss into O (1) blocks with parallelizable Picard iterations within each block. R igorous theoretical analysis reveals that our algorithm achieves null O (poly log d) overall time complexity, marking the first implementation with provable sub-linear complexi ty w.r .t. the data dimension d. Our analysis is based on a generalized version of Girsanov' s theorem and is compatible with both the SDE and probability fl ow ODE implementations. Our results shed light on the potential of fast a nd efficient sampling of high-dimensional data on fast-evolving modern large-me mory GPU clusters.
May-24-2024
- Country:
- Asia > Japan (0.04)
- North America > United States
- California > Santa Clara County > Palo Alto (0.04)
- Genre:
- Research Report > New Finding (0.34)
- Technology: