CoBIT: A Contrastive Bi-directional Image-Text Generation Model
You, Haoxuan, Guo, Mandy, Wang, Zhecan, Chang, Kai-Wei, Baldridge, Jason, Yu, Jiahui
–arXiv.org Artificial Intelligence
The field of vision and language has witnessed a proliferation of pre-trained foundation models. Most existing methods are independently pre-trained with contrastive objective like CLIP, image-to-text generative objective like PaLI, or text-to-image generative objective like Parti. However, the three objectives can be pre-trained on the same data, image-text pairs, and intuitively they complement each other as contrasting provides global alignment capacity and generation grants fine-grained understanding. In this work, we present a Contrastive Bi-directional Image-Text generation model (CoBIT), which attempts to unify the three pre-training objectives in one framework. Specifically, CoBIT employs a novel unicoder-decoder structure, consisting of an image unicoder, a text unicoder and a cross-modal decoder. The image/text unicoders can switch between encoding and decoding in different tasks, enabling flexibility and shared knowledge that benefits both image-to-text and text-to-image generations. CoBIT achieves superior performance in image understanding, image-text understanding (Retrieval, Captioning, VQA, SNLI-VE) and text-based content creation, particularly in zero-shot scenarios. For instance, 82.7% in zero-shot ImageNet classification, 9.37 FID score in zero-shot text-to-image generation and 44.8 CIDEr in zero-shot captioning.
arXiv.org Artificial Intelligence
Mar-23-2023
- Country:
- Asia (0.28)
- Europe > Switzerland (0.28)
- Genre:
- Research Report (0.50)
- Industry:
- Information Technology (0.46)
- Technology: