[1]Grishman R,Sundheim B.Message understanding conference-6:A brief history[C]∥Proc of the 16th Conference on Computational Linguistics,1996:466-471.
[2]Zong Cheng-qing.Statistical natural language processing[M].2nd Edition.Beijing:Tsinghua University Press,2013.(in Chinese)
[3]McCallum A,Li W.Early results for named entity recognition with conditional random fields,feature induction and web-enhanced lexicons[C]∥Proc of the 7th Conference on Natural Language Learning at HLT-NAACL,2003:188-191.
[4]Craven M,Kumlien J.Constructing biological knowledge bases by extracting information from text sources[C]∥Proc of the Intelligent Systems in Molecular Biology,1999:77-86.
[5]Bunescu R,Mooney R J.Relational Markov networks for collective information extraction[C]∥Proc of the ICML-2004 Workshop on Statistical Relational Learning and Its Connections to Other Fields,2004:1.
[6]Minkov E,Wang R C,Cohen W W.Extracting personal names from email:Applying named entity recognition to informal text[C]∥ Proc of the Conference on Human Language Technology and Empirical Methods in Natural Language Processing,2005:443-450.
[7]Ritter A,Clark S,Etzioni O.Named entity recognition in tweets:An experimental study[C]∥ Proc of the Conference on Empirical Methods in Natural Language Processing,2011:1524-1534.
[8]Tkachenko M,Simanovsky A.Named entity recognition:Exploring features[C]∥Proc of the Konvens,2012:118-127.
[9]Brown P F,Desouza P V,Mercer R L,et al.Class-based n-gram models of natural language[J].Computational Linguistics,1992,18(4):467-479.
[10]Turian J, Ratinov L, Bengio Y.Word representations:A simple and general method for semi-supervised learning[C]∥ Proc of the 48th Annual Meeting of the Association for Computational Linguistics,2010:384-394.
[11]Hinton G E,Salakhutdinov R R.Reducing the dimensionality of data with neural networks[J].Science,2006,313(5786):504-507.
[12]Berger A L,Pietra V J D,Pietra S A D.A maximum entropy approach to natural language processing[J].Computational Linguistics,1996,22(1):39-71.
[13]Melli G,Romming C.An overview of the CPROD1 contest on consumer product recognition within user generated postings and normalization against a large product catalog[C]∥ Proc of the ICDM-2012 Workshop on Consumer Product Contest,2012:861-864.
[14]Hinton G E.Learning distributed representations of concepts[C]∥Proc of the 8th Annual Conference of the Cognitive Science Society,1986:1.
[15]Pennington J,Socher R,Manning C D.Glove:Global vectors for word representation[C]∥Proc of the Empirical Methods in Natural Language Processing,2014:1532-1543.
[16]Soricut R,Och F.Unsupervised morphology induction using word embeddings[C]∥ Proc of the 2015 Annual Conference of the North American Chapter of the ACL,2015:1627-1637.
[17]Collobert R,Weston J,Bottou L,et al.Natural language processing (almost) from scratch[J].The Journal of Machine Learning Research,2011,12(4):2493-2537.
[18]Mnih A,Hinton G.Three new graphical models for statistical language modelling[C]∥ Proc of the 24th International Conference on Machine Learning,2007:641-648.
[19]Mnih A,Hinton G E.A scalable hierarchical distributed language model[C]∥ Proc of the Neural Information Processing Systems,2009:1081-1088.
[20]Li Hang.Statistical learning method [M].Beijing:Tsinghua University Press,2012.(in Chinese)
[21]Borthwick A.A maximum entropy approach to named entity recognition[D].New York:New York University,1999.
[22]Lu Ming,Kang Yu-jie,Yu Neng-hai.Basic grammer rule and maximum entropy based hybrid model for name entity recognition [J].Journal of Chinese Computer Systems,2012,33(3):537-541.(in Chinese)
附中文参考文献:
[2]宗成庆.统计自然语言处理[M].第2版.北京:清华大学出版社,2013.
[20]李航.统计学习方法[M].北京:清华大学出版社,2012.
[22]陆铭,康雨洁,俞能海.简约语法规则和最大熵模型相结合的混合实体识别[J].小型微型计算机系统,2012,33(3):537-541.