Text classification model based on GNN and attention mechanism

Addressing the issue of low classification accuracy raised by the poor performance of the model, which is caused by the difficulty in learning from dynamic aggregation unknown neighboring nodes of graph data and insufficient fusion of semantic features, a model named graph attention text classificat...

Full description

Saved in:
Bibliographic Details
Main Authors: ZENG Shuifei, MENG Yao, LIU Jing
Format: Article
Language:zho
Published: Beijing Xintong Media Co., Ltd 2025-05-01
Series:Dianxin kexue
Subjects:
Online Access:http://www.telecomsci.com/thesisDetails#10.11959/j.issn.1000-0801.2025136
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Addressing the issue of low classification accuracy raised by the poor performance of the model, which is caused by the difficulty in learning from dynamic aggregation unknown neighboring nodes of graph data and insufficient fusion of semantic features, a model named graph attention text classification(GATC) based on graph neural network (GNN) and attention mechanism was proposed. Firstly, an inductive learning of graph neural network model was constructed, and dynamic embedding the unknown neighboring node was implemented by using an aggregation function to enhance the model’s generalization ability. Secondly, the reasoning cache size of key-value was reduced by the introduction of multi-head latent attention mechanism that utilized the low-rank key-value joint compression technology, which significantly diminished memory usage and improved the performance of the model. Finally, the integration of GNN and gated recurrent unit (GRU) network models further captured the semantic feature information of structural and temporal attributes for graph data, resulting in achieving efficient feature fusion and improving the classification accuracy of the model. The experimental results show that the proposed method not only is effective, but also improves the accuracy of classification that is increased at least 4.0%, 2.4% and 3.1% on the CSI 100,CSI 300 and Rus 1K datasets, respectively, compared with the algorithm ADGL+MLA (adaptive dynamic graph learning+multi-head latent attention).
ISSN:1000-0801