[1] |
Bahdanau D,Cho K,Bengio Y.Neural machine translation by jointly learning to align and translate[C]∥Proc of the 3rd International Conference on Learning Representations,2015:1-15.
|
[2] |
Sennrich R,Haddow B,Birch A.Improving neural machine translation models with monolingual data[C]∥Proc of the 54th Annual Meeting of the Association for Computational Linguistics,2016:86-96.
|
[3] |
Brown P F,Della Pietra S A,Della Pietra V J,et al.The mathematics of statistical machine translation:Parameter estimation[J].Computational Linguistics,1993,19(2):263-311.
|
[4] |
Yee K,Ng N,Dauphin Y,et al.Simple and effective noisy channel modeling for neural machine translation[C]∥Proc of the Conference on Empirical Methods in Natural Language Processing and the International Joint Conference on Natural Language Processing,2019:5700-5705.
|
[5] |
Shannon C E.Communication theory of secrecy systems[J].The Bell System Technical Journal,1949,28(4):656-715.
|
[6] |
Gulcehre C,Firat O,Xu K,et al.On using monolingual corpora in neural machine translation[J].arXiv:1503.03535,2015.
|
[7] |
Stahlberg F,Cross J,Stoyanov V.Simple fusion:Return of the language model[C]∥Proc of the 3rd Conference on Machine Translation:Research Papers,2018:204-211.
|
[8] |
Miao M,Meng F,Liu Y,et al.Prevent the language model from being overconfident in neural machine translation[C]∥Proc of the 59th Annual Meeting of the Association for Computational Linguistics,2021:3456-3468.
|
[9] |
Baziotis C, Haddow B,Birch A.Language model prior for low-resource neural machine translation[C]∥Proc of the Conference on Empirical Methods in Natural Language Processing,2020:7622-7634.
|
[10] |
Hinton G,Vinyals O,Dean J.Distilling the knowledge in a neural network[J].arXiv:1503.02531,2015.
|
[11] |
Brown P F,Cocke J,Della Pietra S A,et al.A statistical approach to machine translation[J].Computational Linguistics,1990,16(2):79-85.
|
[12] |
Bengio Y,Ducharme Y,Vincent P,et al.A neural probabilistic language model[J].Journal of Machine Learning Research,2003,3(6):932-938.
|
[13] |
宗成庆.统计自然语言处理[M].第二版.北京:清华大学出版社,2013.
|
|
Zong Cheng-qing.Statistical natural language processing[M].2nd ed. Beijing:Tsinghua University Press,2013.
|
[14] |
Devlin J,Chang M W,Lee K,et al.BERT:Pretraining of deep bidirectional transformers for language understanding[C]∥Proc of the 57th Annual Meeting of the Association for Computational Linguistics,2019:4171-4186.
|
[15] |
Yang Z,Xu Z,Cui Y,et al.CINO:A Chinese minority pre-trained language model[C]∥Proc of the 29th International Conference on Computational Linguistics,2022:3937-3949.
|
[16] |
侯宏旭,孙硕,乌尼尔.蒙汉神经机器翻译研究综述[J].计算机科学,2022,49(1):31-40.
|
|
Hou Hong-xu,Sun Shuo,Wu Ni-er.Survey of Mongolian-Chinese neural machine translation[J].Computer Science,2022,49(1):31-40.
|
[17] |
慈祯嘉措,桑杰端珠,孙茂松,等.融合单语语言模型的藏汉机器翻译方法研究[J].中文信息学报,2019,33(12):61-66.
|
|
Cizhen Jia-cuo,Sangjie Duan-zhu,Sun Mao-song,et al. Tibetan-Chinese machine translation based on Tibetan language model enhanced transformer[J].Journal of Chinese Information Processing,2019,33(12):61-66.
|
[18] |
陈玺,杨雅婷,董瑞.面向汉维机器翻译的BERT嵌入研究[J].计算机工程,2021,47(12):112-117.
|
|
Chen Xi,Yang Ya-ting,Dong Rui.Research on BERT embedding for Chinese-Uyghur machine translation[J].Computer Engineering,2021,47(12):112-117.
|
[19] |
李飞雨,赵亚慧,崔荣一,等.基于强化学习和机器翻译质量评估的中朝机器翻译研究[J].计算机应用研究,2021,38(8):2288-2292.
|
|
Li Fei-yu,Zhao Ya-hui,Cui Rong-yi,et al.Research on Chinese-Korean machine translation based on reinforcement learning and machine translation quality estimation[J].Application Research of Computers,2021,38(8):2288-2292.
|
[20] |
王可超,郭军军,张亚飞,等.基于回译和比例抽取孪生网络筛选的汉越平行语料扩充方法[J].计算机工程与科学,2022,44(10):1861-1868.
|
|
Wang Ke-chao,Guo Jun-jun,Zhang Ya-fei,et al.A Chinese-Vietnamese parallel corpus expansion method based on back translation and proportional extraction siamese network screening[J].Computer Engineering & Science,2022,44(10):1861-1868.
|
[21] |
龙从军,安波.中国少数民族语言文字信息处理的进展[J].暨南学报(哲学社会科学版),2022,44(9):12-23.
|
|
Long Cong-jun,An Bo.Review of information processing of ethnic languages in China[J].Jinan Journal(Philosophy & Social Science Edition),2022,44(9):12-23.
|
[22] |
Kocmi T,Bojar O.Trivial transfer learning for low-resource neural machine translation[C]∥Proc of the 3rd Conference on Machine Translation:Research Papers,2018:244-252.
|
[23] |
Gu J,Wang Y,Chen Y,et al.Meta-learning for low-resource neural machine translation[C]∥Proc of the Conference on Empirical Methods in Natural Language Processing,2018:3622-3631.
|
[24] |
Liu P,Yuan W,Fu J,et al.Pre-train,prompt,and predict:A systematic survey of prompting methods in natural language processing[J].arXiv:2107.13586,2021.
|
[25] |
Szegedy C,Vanhoucke V,Ioffe S,et al.Rethinking the inception architecture for computer vision[C]∥Proc of 2016 IEEE Conference on Computer Vision and Pattern Recognition,2016:2818-2826.
|
[26] |
Muller R,Kornblith S,Hinton G. When does label smoothing help?[C]∥Proc of the 33rd International Conference on Neural Information Processing Systems,2019:4694-4703.
|
[27] |
Sennrich R,Haddow B,Alexandra Birch A. Neural machine translation of rare words with subword units[C]∥Proc of the 54th Annual Meeting of the Association for Computational Linguistics,2016:1715-1725.
|
[28] |
Vaswani A,Shazeer N,Parmar N,et al.Attention is all you need[C]∥Proc of the 31st International Conference on Neural Information Processing Systems,2017:6000-6010.
|
[29] |
Papineni K,Roukos S,Ward T,et al.BLEU:A method for automatic evaluation of machine translation[C]∥Proc of the 40th Annual Meeting of the Association for Computational Linguistics,2002:311-318.
|