• 中国计算机学会会刊
  • 中国科技核心期刊
  • 中文核心期刊

Computer Engineering & Science ›› 2025, Vol. 47 ›› Issue (06): 1133-1140.

• Artificial Intelligence and Data Mining • Previous Articles    

Tibetan long text classification by fusing denoising fine-tuning and graph attention mechanism

JING Rong1,WAN Fucheng1,2,HUANG Rui1,YU Hongzhi1,2,MA Ning1,2    

  1.  (1.Key Laboratory of Linguistic and Cultural Computing Ministry of Education,
    Northwest Minzu University,Lanzhou 730030;
    2.Key Laboratory of Minzu Languages and Cultures Intelligent Information Processing,
    Gansu Province(Northwest Minzu University),Lanzhou 730030,China) 
  • Received:2024-08-27 Revised:2024-09-06 Online:2025-06-25 Published:2025-06-26

Abstract: In Tibetan long text classification tasks, the issue of long-distance dependencies is particularly prominent. Meanwhile, multilingual pre-trained models exhibit certain biases when handling Tibetan text classification tasks. To address these issues, this paper proposes a Tibetan long text classification method based on the pre-trained language model CINO-Large, which integrates denoising fine-tuning and a graph attention network. Firstly, the In-trust loss function is introduced into CINO-Large to enhance the model’s generalization ability in downstream tasks through task-adaptive loss. Secondly, sliding windows and linear classification are introduced into graph structure modeling to selectively increase document-document edges, thereby improving the feature distinguishability among nodes. Finally, the graph attention mechanism is utilized to capture the importance of different nodes in the graph, completing the Tibetan long text classification task. On the TNCC news long text dataset, the classification accuracy of the proposed method reaches 71.66%. Compared to the pre-trained language model CINO-Large, the accuracy, precision, and F1 score of the proposed model are improved by 1.77%, 2.67% and 2.03%, respectively. For some subclasses that are difficult to classify, the F1 score of the proposed method can be significantly improved by approximately 20%.

Key words: pre-training model, denoising fine-tuning, graph attention mechanism, Tibetan long text classification