• 中国计算机学会会刊
  • 中国科技核心期刊
  • 中文核心期刊

Computer Engineering & Science ›› 2021, Vol. 43 ›› Issue (01): 134-141.

Previous Articles     Next Articles

Enhancing information transfer in neural machine translation

SHI Xiao-jing,NING Qiu-yi,JI Bai-jun,DUAN Xiang-yu   

  1. (Natural Language Processing Laboratory,Soochow University,Suzhou 215006,China)
  • Received:2020-03-17 Revised:2020-05-08 Accepted:2021-01-25 Online:2021-01-25 Published:2021-01-22

Abstract: In the field of Neural Machine Translation (NMT), the multi-layer neural network model structure can significantly improve the translation performance. However, the structure of multi-layer neural network has an inherent problem with information transfer degeneracy. To alleviate this problem, this paper proposes an information transfer enhancement method by fusing layers information and sublayers information. By introducing a "retention gate" mechanism to control the fused information transfer weight, which is aggregated with the output of the current layer and then serves as the input of the next layer, thus making fuller information transfer between layers. Experiments were carried out on the most advanced NMT model Transformer. Experimental results on the Chinese-English and German-English tasks show that our method improves BLEU score by 0.66, and 0.42 in comparison to the baseline system. 




Key words: neural network, neural machine translation, information transfer, information degene- , racy, residual network, gate mechanism