HW-TSC's Submission to the CCMT 2024 Machine Translation Tasks
Wu, Zhanglin, Luo, Yuanchang, Wei, Daimeng, Zheng, Jiawei, Wei, Bin, Li, Zongyao, Shang, Hengchao, Guo, Jiaxin, Li, Shaojun, Zhang, Weidong, Xie, Ning, Yang, Hao
–arXiv.org Artificial Intelligence
This paper presents the submission of Huawei Translation Services Center (HW-TSC) to machine translation tasks of the 20th China Conference on Machine Translation (CCMT 2024). We participate in the bilingual machine translation task and multi-domain machine translation task. For these two translation tasks, we use training strategies such as regularized dropout, bidirectional training, data diversification, forward translation, back translation, alternated training, curriculum learning, and transductive ensemble learning to train neural machine translation (NMT) models based on the deep Transformerbig architecture. Furthermore, to explore whether large language model (LLM) can effectively improve the translation quality of NMT models, we use supervised fine-tuning (SFT) to train llama2-13b as an Automatic post-editing (APE) model to improve the translation results of the NMT model on the multi-domain machine translation task. By using these plyometric strategies, our submission achieves a competitive result in the final evaluation.
arXiv.org Artificial Intelligence
Oct-8-2024