TY - JOUR
T1 - Integrating information by Kullback–Leibler constraint for text classification
AU - Yin, Shu
AU - Zhu, Peican
AU - Wu, Xinyu
AU - Huang, Jiajin
AU - Li, Xianghua
AU - Wang, Zhen
AU - Gao, Chao
N1 - Publisher Copyright:
© 2023, The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature.
PY - 2023/8
Y1 - 2023/8
N2 - Text classification is an important assignment for various text-related downstream assignments, such as fake news detection, sentiment analysis, and question answering. In recent years, the graph-based method achieves excellent results in text classification tasks. Instead of regarding a text as a sequence structure, this method regards it as a co-occurrence set of words. The task of text classification is then accomplished by aggregating the data from nearby nodes using the graph neural network. However, existing corpus-level graph models are difficult to incorporate the local semantic information and classify new coming texts. To address these issues, we propose a Global–Local Text Classification (GLTC) model, based on the KL constraints to realize inductive learning for text classification. Firstly, a global structural feature extractor and a local semantic feature extractor are designed to capture the structural and semantic information of text comprehensively. Then, the KL divergence is introduced as a regularization term in the loss calculation process, which ensures that the global structural feature extractor can constrain the learning of the local semantic feature extractor to achieve inductive learning. The comprehensive experiments on benchmark datasets present that GLTC outperforms baseline methods in terms of accuracy.
AB - Text classification is an important assignment for various text-related downstream assignments, such as fake news detection, sentiment analysis, and question answering. In recent years, the graph-based method achieves excellent results in text classification tasks. Instead of regarding a text as a sequence structure, this method regards it as a co-occurrence set of words. The task of text classification is then accomplished by aggregating the data from nearby nodes using the graph neural network. However, existing corpus-level graph models are difficult to incorporate the local semantic information and classify new coming texts. To address these issues, we propose a Global–Local Text Classification (GLTC) model, based on the KL constraints to realize inductive learning for text classification. Firstly, a global structural feature extractor and a local semantic feature extractor are designed to capture the structural and semantic information of text comprehensively. Then, the KL divergence is introduced as a regularization term in the loss calculation process, which ensures that the global structural feature extractor can constrain the learning of the local semantic feature extractor to achieve inductive learning. The comprehensive experiments on benchmark datasets present that GLTC outperforms baseline methods in terms of accuracy.
KW - Constraint
KW - Graph neural network
KW - Kullback–Leibler divergence
KW - Text classification
UR - http://www.scopus.com/inward/record.url?scp=85158086909&partnerID=8YFLogxK
U2 - 10.1007/s00521-023-08602-0
DO - 10.1007/s00521-023-08602-0
M3 - 文章
AN - SCOPUS:85158086909
SN - 0941-0643
VL - 35
SP - 17521
EP - 17535
JO - Neural Computing and Applications
JF - Neural Computing and Applications
IS - 24
ER -