OPT-IML: Scaling Language Model Instruction Meta Learning through the Lens of Generalization
Iyer, Srinivasan, Lin, Xi Victoria, Pasunuru, Ramakanth, Mihaylov, Todor, Simig, Daniel, Yu, Ping, Shuster, Kurt, Wang, Tianlu, Liu, Qing, Koura, Punit Singh, Li, Xian, O'Horo, Brian, Pereyra, Gabriel, Wang, Jeff, Dewan, Christopher, Celikyilmaz, Asli, Zettlemoyer, Luke, Stoyanov, Ves
–arXiv.org Artificial Intelligence
Recent work has shown that fine-tuning large pre-trained language models on a collection of tasks described via instructions, a.k.a. instruction-tuning, improves their zero and few-shot generalization to unseen tasks. However, there is a limited understanding of the performance trade-offs of different decisions made during the instruction-tuning process. These decisions include the scale and diversity of the instruction-tuning benchmark, different task sampling strategies, fine-tuning with and without demonstrations, training using specialized datasets for reasoning and dialogue, and finally, the fine-tuning objectives themselves. In this paper, we characterize the effect of instruction-tuning decisions on downstream task performance when scaling both model and benchmark sizes. To this end, we create OPT-IML Bench: a large benchmark for Instruction Meta-Learning (IML) of 2000 NLP tasks consolidated into task categories from 8 existing benchmarks, and prepare an evaluation framework to measure three types of model generalizations: to tasks from fully held-out categories, to held-out tasks from seen categories, and to held-out instances from seen tasks. Through the lens of this framework, we first present insights about instruction-tuning decisions as applied to OPT-30B and further exploit these insights to train OPT-IML 30B and 175B, which are instruction-tuned versions of OPT. OPT-IML demonstrates all three generalization abilities at both scales on four different evaluation benchmarks with diverse tasks and input formats -- PromptSource, FLAN, Super-NaturalInstructions, and UnifiedSKG. Not only does it significantly outperform OPT on all benchmarks but is also highly competitive with existing models fine-tuned on each specific benchmark. We release OPT-IML at both scales, together with the OPT-IML Bench evaluation framework.
arXiv.org Artificial Intelligence
Jan-30-2023
- Country:
- Africa
- Ethiopia (0.04)
- Middle East > Morocco (0.04)
- Nigeria (0.04)
- Asia
- Afghanistan
- Kabul Province > Kabul (0.04)
- Kandahar Province > Kandahar (0.04)
- China > Hong Kong (0.04)
- Middle East
- Iran > Tehran Province
- Tehran (0.04)
- Saudi Arabia (0.04)
- Iran > Tehran Province
- Pakistan (0.04)
- Russia (0.04)
- Afghanistan
- Europe
- Portugal (0.04)
- Central Europe (0.04)
- Czechia > Prague (0.04)
- Ireland > Leinster
- County Dublin > Dublin (0.04)
- Italy > Calabria
- Catanzaro Province > Catanzaro (0.04)
- Germany > Hesse
- Darmstadt Region > Darmstadt (0.04)
- France (0.04)
- Serbia (0.04)
- Slovakia (0.14)
- United Kingdom > England (0.04)
- Poland
- Greater Poland Province > Poznań (0.04)
- Masovia Province > Warsaw (0.04)
- Russia > Central Federal District
- Moscow Oblast > Moscow (0.04)
- North America
- Canada > Ontario
- Toronto (0.04)
- Costa Rica (0.04)
- United States
- California
- San Diego County > San Diego (0.04)
- San Francisco County > San Francisco (0.04)
- Illinois > Cook County
- Chicago (0.04)
- Iowa (0.04)
- Louisiana > Orleans Parish
- New Orleans (0.04)
- New York > New York County
- New York City (0.04)
- Pennsylvania (0.04)
- Washington > King County
- Seattle (0.04)
- California
- Canada > Ontario
- Oceania > Australia
- South America
- Argentina (0.04)
- Colombia > Meta Department
- Villavicencio (0.04)
- Ecuador (0.04)
- Africa
- Genre:
- Research Report > New Finding (0.67)
- Industry:
- Education (1.00)
- Energy (0.67)
- Government > Regional Government
- Health & Medicine > Therapeutic Area (1.00)
- Leisure & Entertainment
- Media
- Technology: