• 中国计算机学会会刊
  • 中国科技核心期刊
  • 中文核心期刊

Computer Engineering & Science ›› 2024, Vol. 46 ›› Issue (04): 743-751.

• Artificial Intelligence and Data Mining • Previous Articles     Next Articles

A neural machine translation method based on language model distillation

SHEN Ying-li1,2,ZHAO Xiao-bing2,3   

  1.  (1.School of Chinese Ethnic Minority Languages and Literatures,Minzu University of China,Beijing 100081;
    2.National Language Resource Monitoring & Research Center of Minority Languages,Beijing 100081;
    3.School of Information Engineering,Minzu University of China,Beijing 100081,China)
  • Received:2023-02-12 Revised:2023-04-20 Accepted:2024-04-25 Online:2024-04-25 Published:2024-04-18

Abstract: The lack of large parallel corpora is one of the key issues in low-resource neural machine translation. This paper proposes a neural machine translation method based on language model distillation, which regularizes neural machine translation training by using a monolingual language model. This method introduces prior knowledge contained in the language model to improve translation results. Specifically, we draw on the idea of knowledge distillation, and use the target-side language model (teacher model) trained on rich monolingual data to construct the regularization factor of the low- resource neural machine translation model (student model), allowing the translation model to learn highly generalized prior knowledge from the language model. Unlike traditional monolingual language models that participate in the decoding process, the language model in this method is only used during training and does not participate in the inference stage, so it can effectively improve decoding speed. Experimental results on two low-resource translation datasets of Uyghur-Chinese and Tibetan-Chinese from the 17th national machine translation conference (CCMT2021) show that compared with the current state-of-the-art language model fusion baseline system, BLEU can be improved by 1.42 points  (Tibetan-Chinese) to 2.11 points (Chinese-Uyghu).

Key words: language model, knowledge distillation, regularization, low-resource neural machine translation