Exploring the traditional NMT model and Large Language Model for chat translation
Yang, Jinlong, Shang, Hengchao, Wei, Daimeng, Guo, Jiaxin, Li, Zongyao, Wu, Zhanglin, Rao, Zhiqiang, Li, Shaojun, Xie, Yuhao, Luo, Yuanchang, Zheng, Jiawei, Wei, Bin, Yang, Hao
–arXiv.org Artificial Intelligence
This paper describes the submissions of Huawei Translation Services Center(HW-TSC) to WMT24 chat translation shared task on English$\leftrightarrow$Germany (en-de) bidirection. The experiments involved fine-tuning models using chat data and exploring various strategies, including Minimum Bayesian Risk (MBR) decoding and self-training. The results show significant performance improvements in certain directions, with the MBR self-training method achieving the best results. The Large Language Model also discusses the challenges and potential avenues for further research in the field of chat translation.
arXiv.org Artificial Intelligence
Sep-24-2024
- Country:
- Asia
- China > Beijing
- Beijing (0.04)
- Middle East > UAE
- Abu Dhabi Emirate > Abu Dhabi (0.04)
- China > Beijing
- Europe > Germany (0.04)
- North America
- Canada > Ontario
- Toronto (0.04)
- United States > Massachusetts
- Suffolk County > Boston (0.04)
- Canada > Ontario
- Asia
- Genre:
- Research Report > New Finding (0.48)
- Technology: