Computer Engineering & Science
Previous Articles Next Articles
XIAO Xinfeng1,2,LI Shijun2,YU Wei2,LIU Jie2,LIU Beixiong1
Received:
Revised:
Online:
Published:
Abstract:
Current machine translation systems optimize and evaluate the translation process in Indo-European languages to enhance translation accuracy. But researches about Chinese language are few. At present the seq2seq model is the best method in the field of machine translation, which is a neural machine translation model based on the attention mechanism. However, it does not take into account the grammar transformation between different languages. We propose a new optimized English-Chinese translation model. It uses different methods to preprocess texts and initialize embedding layer parameters. Additionally, to improve the seq2seq model structure, a transform layer between the encoder and the decoder is added to deal with grammar transformation problems. Preprocessing can reduce the parameter size and training time of the translation model by 20%, and the translation performance is increased by 0.4 BLEU. The translation performance of the seq2seq model with a transform layer is improved by 0.7 to 1.0 BLEU. Experiments show that compared to the existing seq2seq mainstream model based on the attention mechanism, the training time for English-Chinese translation tasks is the same for corpus of different sizes, but the translation performance of the proposal is improved by 1 to 2 BLEU.
Key words: deep learning, neural machine translation, seq2seq model, attention mechanism, named entity recognition
XIAO Xinfeng1,2,LI Shijun2,YU Wei2,LIU Jie2,LIU Beixiong1.
0 / / Recommend
Add to citation manager EndNote|Ris|BibTeX
URL: http://joces.nudt.edu.cn/EN/
http://joces.nudt.edu.cn/EN/Y2019/V41/I07/1257