Teaching Large Language Models an Unseen Language on the Fly
Zhang, Chen, Liu, Xiao, Lin, Jiuheng, Feng, Yansong
–arXiv.org Artificial Intelligence
Existing large language models struggle to support numerous low-resource languages, particularly the extremely low-resource ones, for which there is minimal training data available for effective parameter updating. We thus investigate whether LLMs can learn a new language on the fly solely through prompting. To study this question, we collect a research suite for Zhuang, a language supported by no LLMs currently. We introduce DiPMT++, a framework for adapting LLMs to unseen languages by in-context learning. Using a dictionary and 5K parallel sentences only, DiPMT++ significantly enhances the performance of GPT-4 from 0 to 16 BLEU for Chinese-to-Zhuang translation and achieves 32 BLEU for Zhuang-to-Chinese translation. We also validate the effectiveness of our framework on Kalamang, another unseen language. Furthermore, we demonstrate the practical utility of DiPMT++ in aiding humans in translating completely unseen languages, which could contribute to the preservation of linguistic diversity.
arXiv.org Artificial Intelligence
Jun-13-2024
- Country:
- Africa > Middle East
- Egypt > Giza Governorate > Giza (0.05)
- Asia
- China
- Beijing > Beijing (0.04)
- Guangxi Province (0.04)
- Indonesia > Bali (0.04)
- Middle East > Jordan (0.04)
- Singapore (0.04)
- China
- Europe
- North America
- Canada > Ontario
- Toronto (0.04)
- Dominican Republic (0.04)
- United States
- Hawaii > Honolulu County
- Honolulu (0.04)
- Minnesota > Hennepin County
- Minneapolis (0.14)
- Pennsylvania (0.04)
- Washington > King County
- Seattle (0.04)
- Hawaii > Honolulu County
- Canada > Ontario
- Africa > Middle East
- Genre:
- Research Report > New Finding (0.46)
- Industry:
- Education (0.66)
- Technology: