Optimizing example selection for retrieval-augmented machine translation with translation memories
Bouthors, Maxime, Crego, Josep, Yvon, François
–arXiv.org Artificial Intelligence
Retrieval-augmented machine translation leverages examples from a translation memory by retrieving similar instances. These examples are used to condition the predictions of a neural decoder. We aim to improve the upstream retrieval step and consider a fixed downstream edit-based model: the multi-Levenshtein Transformer. The task consists of finding a set of examples that maximizes the overall coverage of the source sentence. To this end, we rely on the theory of submodular functions and explore new algorithms to optimize this coverage. We evaluate the resulting performance gains for the machine translation task.
arXiv.org Artificial Intelligence
May-23-2024
- Country:
- Asia
- Middle East > UAE
- Abu Dhabi Emirate > Abu Dhabi (0.04)
- Singapore (0.04)
- Middle East > UAE
- Europe
- Belgium > Brussels-Capital Region
- Brussels (0.04)
- Croatia > Dubrovnik-Neretva County
- Dubrovnik (0.04)
- Denmark > Capital Region
- Copenhagen (0.04)
- Finland > Pirkanmaa
- Tampere (0.04)
- France > Île-de-France
- Ireland > Leinster
- County Dublin > Dublin (0.04)
- Italy > Tuscany
- Florence (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- Belgium > Brussels-Capital Region
- North America
- Canada > Ontario
- Toronto (0.04)
- Dominican Republic (0.04)
- United States
- Colorado > Denver County
- Denver (0.04)
- Florida > Broward County
- Fort Lauderdale (0.04)
- Louisiana > Orleans Parish
- New Orleans (0.04)
- Maryland > Baltimore (0.04)
- Oregon > Multnomah County
- Portland (0.04)
- Pennsylvania > Philadelphia County
- Philadelphia (0.04)
- Colorado > Denver County
- Canada > Ontario
- Asia
- Genre:
- Research Report (0.40)
- Technology: