MultiInstruct: Improving Multi-Modal Zero-Shot Learning via Instruction Tuning
Xu, Zhiyang, Shen, Ying, Huang, Lifu
–arXiv.org Artificial Intelligence
Instruction tuning, a new learning paradigm that fine-tunes pre-trained language models on tasks specified through instructions, has shown promising zero-shot performance on various natural language processing tasks. However, it has yet to be explored for vision and multimodal tasks. In this work, we introduce MUL-TIINSTRUCT, the first multimodal instruction tuning benchmark dataset that consists of 62 diverse multimodal tasks in a unified seq-to-seq format covering 10 broad categories. The tasks are derived from 21 existing open-source datasets and each task is equipped with 5 expert-written instructions. We take OFA as the base pre-trained model for multimodal instruction tuning, and to further improve its zero-shot performance, we explore multiple transfer learning strategies to leverage the large-scale NATURAL INSTRUCTIONS dataset. Experimental results demonstrate strong zero-shot performance on various unseen multimodal tasks and the benefit of transfer learning from a text-only instruction dataset. We also design a new evaluation metric - Sensitivity, to evaluate how sensitive the model is to the variety of instructions. Our results indicate that fine-tuning the model on a diverse set of tasks and instructions leads to a reduced sensitivity to variations in instructions for each task.
arXiv.org Artificial Intelligence
Jun-10-2023
- Country:
- Asia > China
- Europe > Ireland
- Leinster > County Dublin > Dublin (0.04)
- North America
- Dominican Republic (0.04)
- United States
- Louisiana > Orleans Parish
- New Orleans (0.04)
- Virginia (0.04)
- Washington > King County
- Seattle (0.14)
- Louisiana > Orleans Parish
- Oceania > Australia
- Queensland > Brisbane (0.04)
- Genre:
- Research Report > New Finding (0.86)
- Technology: