TF-ICON: Diffusion-Based Training-Free Cross-Domain Image Composition
Lu, Shilin, Liu, Yanzhu, Kong, Adams Wai-Kin
–arXiv.org Artificial Intelligence
Text-driven diffusion models have exhibited impressive generative capabilities, enabling various image editing tasks. In this paper, we propose TF-ICON, a novel Training-Free Image COmpositioN framework that harnesses the power of text-driven diffusion models for cross-domain image-guided composition. This task aims to seamlessly integrate user-provided objects into a specific visual context. Current diffusion-based methods often involve costly instance-based optimization or finetuning of pretrained models on customized datasets, which can potentially undermine their rich prior. In contrast, TF-ICON can leverage off-the-shelf diffusion models to perform cross-domain image-guided composition without requiring additional training, finetuning, or optimization. Moreover, we introduce the exceptional prompt, which contains no information, to facilitate text-driven diffusion models in accurately inverting real images into latent representations, forming the basis for compositing. Our experiments show that equipping Stable Diffusion with the exceptional prompt outperforms state-of-the-art inversion methods on various datasets (CelebA-HQ, COCO, and ImageNet), and that TF-ICON surpasses prior baselines in versatile visual domains. Code is available at https://github.com/Shilin-LU/TF-ICON
arXiv.org Artificial Intelligence
Oct-10-2023
- Country:
- Asia
- Middle East
- Republic of Türkiye (0.04)
- Saudi Arabia > Northern Borders Province
- Arar (0.04)
- Singapore (0.04)
- Middle East
- Europe > Switzerland
- Asia
- Genre:
- Research Report > New Finding (0.67)
- Industry:
- Media (0.66)
- Technology: