Coupled Diffusion Sampling for Training-Free Multi-View Image Editing
Alzayer, Hadi, Zhang, Yunzhi, Geng, Chen, Huang, Jia-Bin, Wu, Jiajun
–arXiv.org Artificial Intelligence
We present an inference-time diffusion sampling method to perform multi-view consistent image editing using pre-trained 2D image editing models. These models can independently produce high-quality edits for each image in a set of multi-view images of a 3D scene or object, but they do not maintain consistency across views. Existing approaches typically address this by optimizing over explicit 3D representations, but they suffer from a lengthy optimization process and instability under sparse view settings. We propose an implicit 3D regularization approach by constraining the generated 2D image sequences to adhere to a pre-trained multi-view image distribution. This is achieved through coupled diffusion sampling, a simple diffusion sampling technique that concurrently samples two trajectories from both a multi-view image distribution and a 2D edited image distribution, using a coupling term to enforce the multi-view consistency among the generated images. Diffusion-based image editing models have demonstrated unprecedented realism across diverse tasks via end-to-end training. However, collecting and curating 3D data is significantly more costly than working with 2D data. As a result, recent research has explored test-time optimization methods for multi-view editing that leverage pre-trained 2D image diffusion models (Poole et al., 2023; Haque et al., 2023). Figure 1: Applications of coupled diffusion sampling. Our approach enables lifting off-the-shelf 2D editing models into multi-view by combining the sampling process of 2D diffusion models with multi-view diffusion models to produce view-consistent edits.
arXiv.org Artificial Intelligence
Oct-17-2025
- Country:
- Asia > Japan
- Honshū > Chūbu > Ishikawa Prefecture > Kanazawa (0.04)
- North America > United States
- Maryland > Prince George's County > College Park (0.04)
- Asia > Japan
- Genre:
- Research Report > New Finding (0.46)
- Technology: