计算机科学 ›› 2022, Vol. 49 ›› Issue (8): 230-236.doi: 10.11896/jsjkx.210600042

• 人工智能 • 上一篇    下一篇

基于双图神经网络信息融合的文本分类方法

闫佳丹, 贾彩燕   

  1. 北京交通大学计算机与信息技术学院 北京 100044
    交通数据分析与数据挖掘北京市重点实验室(北京交通大学) 北京 100044
  • 收稿日期:2021-06-04 修回日期:2021-10-23 发布日期:2022-08-02
  • 通讯作者: 贾彩燕(cyjia@bjtu.edu.cn)
  • 作者简介:(18120436@bjtu.edu.cn)
  • 基金资助:
    中央高校基本科研业务费专项资金(2019JBZ110)

Text Classification Method Based on Information Fusion of Dual-graph Neural Network

YAN Jia-dan, JIA Cai-yan   

  1. School of Computer and Information Technology,Beijing Jiaotong University,Beijing 100044,China
    Beijing Key Lab of Traffic Data Analysis and Mining,Beijing Jiaotong University,Beijing 100044,China
  • Received:2021-06-04 Revised:2021-10-23 Published:2022-08-02
  • About author:YAN Jia-dan,born in 1995,postgra-duate.Her main research interests include natural language processing and text classification.
    JIA Cai-yan,born in 1976,Ph.D,professor,is a member of China Computer Federation.Her main research interests include machine learning,complex network analysis and social computing.
  • Supported by:
    Fundamental Research Funds for the Central Universities of Ministry of Education of China(2019JBZ110).

摘要: 近年来,图神经网络在文本分类任务中得到了广泛应用。与图卷积网络相比,基于消息传递的文本级的图神经网络模型具有内存占用少和支持在线检测等优点。然而此类模型通常仅使用词共现信息为语料中的各个文本构建词汇图,导致获取到的信息缺少多样性。文中提出了一种基于双图神经网络信息融合的文本分类方法。该方法在保留原有词共现图的基础上,根据单词间的余弦相似度构建语义图,并通过阈值控制语义图的稀疏程度,更有效地利用了文本的多方位语义信息。此外,测试了直接融合和注意力机制融合两种方式对词汇图和语义图上学习到的文本表示融合的能力。实验使用R8和R52等12个文本分类领域常用的数据集来测试算法的精度,结果表明,与最新的TextLevelGNN,TextING和MPAD这3个文本级的图神经网络模型相比,双图模型能够有效提高文本分类的性能。

关键词: 图神经网络, 文本分类, 信息融合, 语义信息, 注意力机制, 自然语言处理

Abstract: Graph neural networks are recently applied in text classification tasks.Compared with graph convolution network,the text level graph neural network model based on message passing(MP-GNN) has the advantages of low memory usage and supporting online testing.However,MP-GNN model only builds a lexical graph using the word co-occurrence information, and the obtained information lacks diversity.To address this problem,a text classification method based on information fusion of dual-graph neural network is proposed.Besides preserving the original lexical graph built in MP-GNN,this method constructes the semantic graph based on the cosine similarity between pairs of words,and controls the sparsity of the graph through a threshold,which makes more effective use of the multi-directional semantic information of the text.In addition,the ability of direct fusion and attention mechanism fusion are tested to fuse the text representation learned on lexical graph and semantic graph.Experimental results on 12 datasets(R8,R52 and other datasets commonly used for text classification) show that the proposed model demonstrates an obvious improvement on performance of text classification compared with the SOTA(state-of-the-art) methods TextLevelGNN,TextING and MPAD.

Key words: Attention mechanism, Graph neural network, Information fusion, Natural language processing, Semantic information, Text classification

中图分类号: 

  • TP391
[1]SHERVIN M,NAL K,ERIK C,et al.Deep Learning BasedText Classification:A Comprehensive Review[EB/OL].https://arxiv.org/abs/2004.03705v1.
[2]KIM Y.Convolutional Neural Networks for Sentence Classification[C]//Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing(EMNLP).2014:1746-1751.
[3]LIU P,QIU X,HUANG X.Recurrent neural network for text classification with multi-task learning[C]//Proceedings of the 25th International Joint Conference on Artificial Intelligence.2016:2873-2879.
[4]YAO L,MAO C,LUO Y.Graph convolutional networks fortext classification[C]//Proceedings of the AAAI Conference on Artificial Intelligence.2019:7370-7377.
[5]NIKOLENTZOS G,TIXIER A J,VAZIRGIANNIS M.Message Passing Attention Networks for Document Understanding[C]//Proceedings of the 34th AAAI Conference on Artificial Intelligence.2020:8544-8551.
[6]LIU X,YOU X,ZHANG X,et al.Tensor Graph Convolutional Networks for Text Classification[C]//Proceedings of the AAAI Conference on Artificial Intelligence.2020:8409-8416.
[7]WU F,ZHANG T,YU T,et al.Simplifying Graph Convolu-tional Networks[C]//Proceedings of the 36th International Conference on Machine Learning.2019:6861-6871.
[8]HU L,YANG T,SHI C,et al.Heterogeneous graph attention networks for semi-supervised short text classification [C]//Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Confe-rence on Natural Language Processing(EMNLP-IJCNLP).2019:4823-4832.
[9]HUANG L,MA D,LI S,et al.Text level graph neural network for text classification[C]//Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing.2019:3435-3441.
[10]ZHANG Y,YU X,CUI Z,et al.Every Document Owns ItsStructure:Inductive Text Classification via Graph Neural Networks[C]//Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics.2020:334-339.
[11]GREENE D,CUNNINGHAM P.Practical Solutions to theProblem of Diagonal Dominance in Kernel Document Clustering[C]//Proceedings of the 23rd International Conference on Machine Learning.2006:377-384.
[12]LI X,ROTH D.Learning question classifiers:the role of semantic information[C]//Proceedings of the 19th International Conference on Computational Linguistics.2002:1-7.
[13]PANG B,LEE L.Seeing Stars:Exploiting Class Relationships for Sentiment Categorization with Respect to Rating Scales[C]//Proceedings of the 43rd Annual Meeting on Association for Computational Linguistics.2005:115-124.
[14]MAAS A L,DALY R E,PHAM P T,et al.Learning Word Vectors for Sentiment Analysis[C]//Proceedings of the 49th An-nual Meeting of the Association for Computational Linguistics:Human Language Technologies.2011:142-150.
[15]SOCHER R,PERELYGIN A,POTTS C,et al.Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank[C]//Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing.2013:1631-1642.
[16]WIEBE J,WILSON T,CARDIE C.Annotating Expressions of Opinions and Emotions in Language[J].Language Resources and Evaluation,2005,39(2/3):165-210.
[17]PANG B,LEE L.A Sentimental Education:Sentiment Analysis Using Subjectivity Summarization Based on Minimum Cuts[C]//Proceedings of the 42nd Annual Meeting on Association for Computational Linguistics.2004:271-278.
[18]IYYER M,MANJUNATHA V,BOYD-GRABER J,et al.Deep Unordered Composition Rivals Syntactic Methods for Text Classification[C]//Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing.2015:1681-1691.
[19]YANG Z,YANG D,DYER C,et al.Hierarchical Attention Networks for Document Classification[C]//Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies.2016:1480-1489.
[20]PENNINGTON J,SOCHER R,MANNING C.Glove:Globalvectors for word representation[C]//Proceedings of the 2014 Conference on Empirical Methods in Natural Language Proces-sing(EMNLP).2014:1532-1543.
[1] 周芳泉, 成卫青.
基于全局增强图神经网络的序列推荐
Sequence Recommendation Based on Global Enhanced Graph Neural Network
计算机科学, 2022, 49(9): 55-63. https://doi.org/10.11896/jsjkx.210700085
[2] 戴禹, 许林峰.
基于文本行匹配的跨图文本阅读方法
Cross-image Text Reading Method Based on Text Line Matching
计算机科学, 2022, 49(9): 139-145. https://doi.org/10.11896/jsjkx.220600032
[3] 周乐员, 张剑华, 袁甜甜, 陈胜勇.
多层注意力机制融合的序列到序列中国连续手语识别和翻译
Sequence-to-Sequence Chinese Continuous Sign Language Recognition and Translation with Multi- layer Attention Mechanism Fusion
计算机科学, 2022, 49(9): 155-161. https://doi.org/10.11896/jsjkx.210800026
[4] 熊丽琴, 曹雷, 赖俊, 陈希亮.
基于值分解的多智能体深度强化学习综述
Overview of Multi-agent Deep Reinforcement Learning Based on Value Factorization
计算机科学, 2022, 49(9): 172-182. https://doi.org/10.11896/jsjkx.210800112
[5] 饶志双, 贾真, 张凡, 李天瑞.
基于Key-Value关联记忆网络的知识图谱问答方法
Key-Value Relational Memory Networks for Question Answering over Knowledge Graph
计算机科学, 2022, 49(9): 202-207. https://doi.org/10.11896/jsjkx.220300277
[6] 武红鑫, 韩萌, 陈志强, 张喜龙, 李慕航.
监督和半监督学习下的多标签分类综述
Survey of Multi-label Classification Based on Supervised and Semi-supervised Learning
计算机科学, 2022, 49(8): 12-25. https://doi.org/10.11896/jsjkx.210700111
[7] 汪鸣, 彭舰, 黄飞虎.
基于多时间尺度时空图网络的交通流量预测模型
Multi-time Scale Spatial-Temporal Graph Neural Network for Traffic Flow Prediction
计算机科学, 2022, 49(8): 40-48. https://doi.org/10.11896/jsjkx.220100188
[8] 郝志荣, 陈龙, 黄嘉成.
面向文本分类的类别区分式通用对抗攻击方法
Class Discriminative Universal Adversarial Attack for Text Classification
计算机科学, 2022, 49(8): 323-329. https://doi.org/10.11896/jsjkx.220200077
[9] 姜梦函, 李邵梅, 郑洪浩, 张建朋.
基于改进位置编码的谣言检测模型
Rumor Detection Model Based on Improved Position Embedding
计算机科学, 2022, 49(8): 330-335. https://doi.org/10.11896/jsjkx.210600046
[10] 朱承璋, 黄嘉儿, 肖亚龙, 王晗, 邹北骥.
基于注意力机制的医学影像深度哈希检索算法
Deep Hash Retrieval Algorithm for Medical Images Based on Attention Mechanism
计算机科学, 2022, 49(8): 113-119. https://doi.org/10.11896/jsjkx.210700153
[11] 孙奇, 吉根林, 张杰.
基于非局部注意力生成对抗网络的视频异常事件检测方法
Non-local Attention Based Generative Adversarial Network for Video Abnormal Event Detection
计算机科学, 2022, 49(8): 172-177. https://doi.org/10.11896/jsjkx.210600061
[12] 檀莹莹, 王俊丽, 张超波.
基于图卷积神经网络的文本分类方法研究综述
Review of Text Classification Methods Based on Graph Convolutional Network
计算机科学, 2022, 49(8): 205-216. https://doi.org/10.11896/jsjkx.210800064
[13] 侯钰涛, 阿布都克力木·阿布力孜, 哈里旦木·阿布都克里木.
中文预训练模型研究进展
Advances in Chinese Pre-training Models
计算机科学, 2022, 49(7): 148-163. https://doi.org/10.11896/jsjkx.211200018
[14] 金方焱, 王秀利.
融合RACNN和BiLSTM的金融领域事件隐式因果关系抽取
Implicit Causality Extraction of Financial Events Integrating RACNN and BiLSTM
计算机科学, 2022, 49(7): 179-186. https://doi.org/10.11896/jsjkx.210500190
[15] 熊罗庚, 郑尚, 邹海涛, 于化龙, 高尚.
融合双向门控循环单元和注意力机制的软件自承认技术债识别方法
Software Self-admitted Technical Debt Identification with Bidirectional Gate Recurrent Unit and Attention Mechanism
计算机科学, 2022, 49(7): 212-219. https://doi.org/10.11896/jsjkx.210500075
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!