• 中国计算机学会会刊
  • 中国科技核心期刊
  • 中文核心期刊

计算机工程与科学 ›› 2025, Vol. 47 ›› Issue (12): 2253-2260.

• 人工智能与数据挖掘 • 上一篇    下一篇

基于大语言模型的面向领域的非连续命名实体识别

唐晋韬,张成贤,鲍琛龙,李文静   

  1. (国防科技大学计算机学院,湖南 长沙 410073)

  • 收稿日期:2024-05-11 修回日期:2024-08-18 出版日期:2025-12-25 发布日期:2026-01-06
  • 基金资助:


Domain oriented discontinuous named entity recognition based on large language model

#br#

TANG  Jintao,ZHANG Chengxian,BAO Chenlong,LI Wenjing#br#
  

  1. (College of Computer Science and Technology,National University of Defense Technology,Changsha 410073,China) 
  • Received:2024-05-11 Revised:2024-08-18 Online:2025-12-25 Published:2026-01-06

摘要: 专业领域中术语间的组成逻辑更加复杂,出现了以非连续命名实体为代表的复杂实体等现象。针对非连续命名实体识别任务,提出一种借助大语言模型的理解与生成能力进行识别的方法。该方法将非连续实体识别建模为句子改写任务,设计规则将非连续命名实体识别数据集转换为句子改写数据集,对大语言模型进行输出微调。在命名实体识别阶段,基于改写后的句子,借助提示学习设计规则指令,通过人物角色对话隐式提示大语言模型数据领域等信息,进一步提升了实体识别的效果。实验表明,在3个数据集上,该方法比基于小模型的现有最好方法在药物不良事件语料库CADEC、共享医疗标注2013版 ShARe13和共享医疗标注2014版ShARe14上,F1值分别提升了3.23%,0.28%和1.04%,验证了大语言模型生成能力有助于专业领域命名实体识别的复杂任务。

关键词: 命名实体识别, 大语言模型, 非连续命名实体

Abstract: In professional fields, the compositional logic between terms is more complex, leading to issues such as complex entities represented by discontinuous named entities. To address the task of discontinuous named entity recognition (DNER), this paper proposes a recognition method that leverages the understanding and generation capabilities of large language models (LLMs). This method  discontinuous entity recognition as a sentence rewriting task: It designs rules to convert discontinuous named entity recognition datasets into sentence rewriting datasets, and performs output fine-tuning on the large language model. In the named entity recognition phase, based on the rewritten sentences, it designs rule-based instructions using prompt learning, and implicitly prompts the large language model with domain-specific information (e.g., the field of the data) through character role dialogue, which further improves the entity recognition performance. Experimental results show that on three datasets, this method improved F1 scores by 3.23%, 0.28%, and 1.04% respectively compared to the state-of-the-art (SOTA) methods based on small models on CSIRO adverse drug event corpus(CADEC), shared annotated resources 2013(ShARe13) and shared annotated resources 2014(ShARe14). These results verify that the generation capability of large models contributes to the complex task of named entity recognition in professional fields.


Key words: named entity recognition, large language model (LLM), discontinuous named entity