mmlm
VisLingInstruct: Elevating Zero-Shot Learning in Multi-Modal Language Models with Autonomous Instruction Optimization
Zhu, Dongsheng, Tang, Xunzhu, Han, Weidong, Lu, Jinghui, Zhao, Yukun, Xing, Guoliang, Wang, Junfeng, Yin, Dawei
This paper presents VisLingInstruct, a novel approach to advancing Multi-Modal Language Models (MMLMs) in zero-shot learning. Current MMLMs show impressive zero-shot abilities in multi-modal tasks, but their performance depends heavily on the quality of instructions. VisLingInstruct tackles this by autonomously evaluating and optimizing instructional texts through In-Context Learning, improving the synergy between visual perception and linguistic expression in MMLMs. Alongside this instructional advancement, we have also optimized the visual feature extraction modules in MMLMs, further augmenting their responsiveness to textual cues. Our comprehensive experiments on MMLMs, based on FlanT5 and Vicuna, show that VisLingInstruct significantly improves zero-shot performance in visual multi-modal tasks. Notably, it achieves a 13.1% and 9% increase in accuracy over the prior state-of-the-art on the TextVQA and HatefulMemes datasets.
- South America (0.04)
- North America (0.04)
- Africa (0.04)
- Asia > Myanmar > Tanintharyi Region > Dawei (0.04)
- Media > Film (1.00)
- Leisure & Entertainment (1.00)
- Consumer Products & Services > Food, Beverage, Tobacco & Cannabis > Beverages (0.68)
Beyond Static Models and Test Sets: Benchmarking the Potential of Pre-trained Models Across Tasks and Languages
Ahuja, Kabir, Dandapat, Sandipan, Sitaram, Sunayana, Choudhury, Monojit
Although recent Massively Multilingual Language Models (MMLMs) like mBERT and XLMR support around 100 languages, most existing multilingual NLP benchmarks provide evaluation data in only a handful of these languages with little linguistic diversity. We argue that this makes the existing practices in multilingual evaluation unreliable and does not provide a full picture of the performance of MMLMs across the linguistic landscape. We propose that the recent work done in Performance Prediction for NLP tasks can serve as a potential solution in fixing benchmarking in Multilingual NLP by utilizing features related to data and language typology to estimate the performance of an MMLM on different languages. We compare performance prediction with translating test data with a case study on four different multilingual datasets, and observe that these methods can provide reliable estimates of the performance that are often on-par with the translation based approaches, without the need for any additional translation as well as evaluation costs.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- Asia > Indonesia > Bali (0.04)
- North America > Dominican Republic (0.04)
- (8 more...)
On the Calibration of Massively Multilingual Language Models
Ahuja, Kabir, Sitaram, Sunayana, Dandapat, Sandipan, Choudhury, Monojit
Massively Multilingual Language Models (MMLMs) have recently gained popularity due to their surprising effectiveness in cross-lingual transfer. While there has been much work in evaluating these models for their performance on a variety of tasks and languages, little attention has been paid on how well calibrated these models are with respect to the confidence in their predictions. We first investigate the calibration of MMLMs in the zero-shot setting and observe a clear case of miscalibration in low-resource languages or those which are typologically diverse from English. Next, we empirically show that calibration methods like temperature scaling and label smoothing do reasonably well towards improving calibration in the zero-shot scenario. We also find that few-shot examples in the language can further help reduce the calibration errors, often substantially. Overall, our work contributes towards building more reliable multilingual models by highlighting the issue of their miscalibration, understanding what language and model specific factors influence it, and pointing out the strategies to improve the same.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- Asia > Indonesia > Bali (0.04)
- Asia > India (0.04)
- (7 more...)