Self-Translate-Train: A Simple but Strong Baseline for Cross-lingual Transfer of Large Language Models
Ri, Ryokan, Kiyono, Shun, Takase, Sho
–arXiv.org Artificial Intelligence
Cross-lingual transfer is a promising technique for utilizing data in a source language to improve performance in a target language. However, current techniques often require an external translation system or suffer from suboptimal performance due to over-reliance on cross-lingual generalization of multi-lingual pretrained language models. In this study, we propose a simple yet effective method called Self-Translate-Train. It leverages the translation capability of a large language model to generate synthetic training data in the target language and fine-tunes the model with its own generated data. We evaluate the proposed method on a wide range of tasks and show substantial performance gains across several non-English languages.
arXiv.org Artificial Intelligence
Jun-29-2024
- Country:
- Africa
- Ethiopia > Addis Ababa
- Addis Ababa (0.04)
- Rwanda > Kigali
- Kigali (0.04)
- Ethiopia > Addis Ababa
- Asia
- Europe
- Belgium > Brussels-Capital Region
- Brussels (0.04)
- France (0.04)
- Germany > Berlin (0.04)
- Italy (0.04)
- Middle East > Malta
- Eastern Region > Northern Harbour District > St. Julian's (0.04)
- Belgium > Brussels-Capital Region
- North America
- Canada > Ontario
- Toronto (0.04)
- Dominican Republic (0.04)
- United States
- Louisiana > Orleans Parish
- New Orleans (0.04)
- Minnesota > Hennepin County
- Minneapolis (0.14)
- Pennsylvania (0.04)
- Texas (0.04)
- Louisiana > Orleans Parish
- Canada > Ontario
- Africa
- Genre:
- Research Report > New Finding (0.88)
- Technology: