计算机科学 ›› 2020, Vol. 47 ›› Issue (10): 222-227.doi: 10.11896/jsjkx.190900173

• 人工智能 • 上一篇    下一篇

基于交叉注意力机制和新闻正文的评论情感分类

王启发, 王中卿, 李寿山, 周国栋   

  1. 苏州大学计算机科学与技术学院 江苏 苏州215006
  • 收稿日期:2019-09-25 修回日期:2019-12-06 出版日期:2020-10-15 发布日期:2020-10-16
  • 通讯作者: 王中卿(wangzq@suda.edu.cn)
  • 作者简介:20185227011@stu.suda.edu.cn
  • 基金资助:
    国家自然科学基金青年科学基金项目(61806137,61702518);江苏省高等学校自然科学研究面上项目(18KJB520043)

Comment Sentiment Classification Using Cross-attention Mechanism and News Content

WANG Qi-fa, WANG Zhong-qing, LI Shou-shan, ZHOU Guo-dong   

  1. School of Computer Science and Technology,Soochow University,Suzhou,Jiangsu 215006,China
  • Received:2019-09-25 Revised:2019-12-06 Online:2020-10-15 Published:2020-10-16
  • About author:WANG Qi-fa,born in 1994,postgra-duate,is a member of China Computer Federation.His main research interests include natural language processing and emotion analysis.
    WANG Zhong-qing,born in 1987,Ph.D,postgraduate supervisor,is a member of China Computer Federation.His main research interests include natural language understanding and information extraction.
  • Supported by:
    Young Scientists Fund of the National Natural Science Foundation of China (61806137,61702518) and Natural Science Foundation of the Jiangsu Higher Education Institutions of China (18KJB520043)

摘要: 目前,新闻评论已成为新闻的重要衍生数据,新闻评论表达了评论者对新闻事件的观点、立场及个人情感,通过对新闻评论的情感倾向性分析有助于了解社会舆情及动向,因此,新闻评论的情感研究备受广大学者的青睐。通常的新闻评论情感分析只考虑评论文本自身的信息,而新闻评论文本信息和新闻正文信息往往是紧密相关的,基于此,文中提出一种基于交叉注意力机制并结合正文的新闻评论情感分类方法。首先通过双向长短时记忆网络模型分别对新闻正文与评论文本进行特征表示;然后基于交叉注意力机制进一步捕获评论与正文的关系,得到两个更新后的新闻正文与评论文本的向量表示;再将两者拼接得到的语义表示输入全连接层,使用sigmoid激活函数进行分类预测,从而实现新闻评论的情感分类。结果表明,提出的基于交叉注意力机制和新闻正文的评论情感分类模型可以有效地提升新闻评论情感分类的准确率,该模型的F1值较3个基准模型分别提高了1.72%,3.24%和6.21%。

关键词: 情感分类, 双向长短时记忆网络, 注意力机制

Abstract: At present,news comment has become important news derived data.News comment expresses commentators’ views,positions and personal feelings on news events.Through the analysis of sentiment orientation of news comment,it is helpful to understand the social public opinion and trend.Therefore,the sentiment research of news comment is favored by many scholars.The usual news comment sentiment analysis only considers the information of the comment text itself.However,news comment text information is often closely related to news content information.Based on this,this paper proposed a comment sentiment classification method using cross-attention mechanism and combined with news content.Firstly,the bi-directional long short-term memory network model is used to characterize the news content and the comment text respectively.Then,the cross-attention mechanism is used to further capture important information,and obtain the vector representation of the two updated news content texts and comment texts.And then the semantic representation obtained by splicing them together is input into the full connection layer,and sigmoid activation function is used for classification prediction,so as to realize the sentiment classification of news comments.The results show that the model of comment sentiment classification using cross-attention mechanism and news content can effectively improve the accuracy of sentiment classification of news comment,and this model improves by 1.72%,3.24% and 6.21% on F1 respectively compare with the three benchmark models.

Key words: Attention mechanism, Bi-directional Long Short-Term Memory Network, Sentiment Classification

中图分类号: 

  • TP391.1
[1]FAN W,SUN S.Sentiment classification for online comments on Chinese news[C]//International Conference on Computer Appli-cation & System Modeling.IEEE,2010.
[2]NING Y,ZHU T,WANG Y.Affective-word based Chinese text sentiment classification[C]//International Conference on Pervasive Computing & Applications.IEEE,2011.
[3]SU Z,HUA X,ZHANG D,et al.Chinese sentiment classification using a neural network tool-Word2vec[C]//International Conference on Multisensor Fusion & Information Integration for Intelligent Systems.2014.
[4]ZHANG D,XU H,SU Z,et al.Chinese comments sentimentclassification based on word2vec and SVMperf[J].Expert Systems with Applications,2015,42(4):1857-1863.
[5]ABDULLIN Y B,IVANOV V V.Deep learning model for bilingual sentiment classification of short texts[J].Naucˇno-tehnicˇeskijVestnik Informacionnyh Tehnologij,2017,17(1):129-136.
[6]ZHOU S,CHEN Q,WANG X.Active deep networks for semi-supervised sentiment classification[C]//Proceedings of the 23rd International Conference on Computational Linguistics:Posters.Association for Computational Linguistics,2010:1515-1523.
[7]LECUN Y,BOTTOU L,BENGIO Y,et al.Gradient-basedlearning applied to document recognition[J].Proceedings of the IEEE,1998,86(11):2278-2324.
[8]KIM Y.Convolutional Neural Networks for Sentence Classification[C]//Proceedings of the 2014 conference on empirical methods in natural language processing.2014:1746-1751.
[9]CAO Y,XU R,CHEN T.Combining Convolutional Neural Network and Support Vector Machine for Sentiment Classification[C]//Proceedings of the Meeting of the Chinese National Conference on Social Media Processing.2015:144-155.
[10]TANG D,QIN B,LIU T.Document modeling with gated recurrent neural network for sentiment classification[C]//Proceedings of the 2015 conference on empirical methods in natural language processing.2015:1422-1432.
[11]HOCHREITER S,SCHMIDHUBER,JÜRGEN.Long Short-Term Memory[J].Neural Computation,1997,9(8):1735-1780.
[12]WANG X,LIU Y,CHENGJIE S U N,et al.Predicting polarities of tweets by composing word embeddings with long short-term memory[C]//Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1:Long Papers).2015,1:1343-1353.
[13]TAI K S,SOCHER R,MANNING C D.Improved SemanticRepresentations From Tree-Structured Long Short-Term Memory Networks [J].Computer Science,2015,5(1):36.
[14]MNIH V,HEESS N,GRAVES A.Recurrent models of visual attention [J].Advances in Neural Information Processing Systems,2014(3):2204-2212.
[15]BAHDANAU D,CHO K,BENGIO Y.Neural Machine Translation by Jointly Learning to Align and Translate [C]//Proceedings of the ICLR.2015:1-15.
[16]LUONG M,MANNING C.D.Effective Approaches to Attention-based Neural Machine Translation [C]//Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing.2015:1412-1421.
[17]YANG Z,YANG D,DYER C,et al.Hierarchical Attention Networks for Document Classification[C]//Proceedings of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies.2017:1480-1489.
[18]LI Z,WEI Y,ZHANG Y,et al.Hierarchical attention transfer network for cross-domain sentiment classification[C]//Thirty-Second AAAI Conference on Artificial Intelligence.2018.
[19]ZHOU X,WAN X,XIAO J.Attention-based LSTM network for cross-lingual sentiment classification[C]//Proceedings of the 2016 conference on empirical methods in natural language processing.2016:247-256
[1] 饶志双, 贾真, 张凡, 李天瑞.
基于Key-Value关联记忆网络的知识图谱问答方法
Key-Value Relational Memory Networks for Question Answering over Knowledge Graph
计算机科学, 2022, 49(9): 202-207. https://doi.org/10.11896/jsjkx.220300277
[2] 周芳泉, 成卫青.
基于全局增强图神经网络的序列推荐
Sequence Recommendation Based on Global Enhanced Graph Neural Network
计算机科学, 2022, 49(9): 55-63. https://doi.org/10.11896/jsjkx.210700085
[3] 戴禹, 许林峰.
基于文本行匹配的跨图文本阅读方法
Cross-image Text Reading Method Based on Text Line Matching
计算机科学, 2022, 49(9): 139-145. https://doi.org/10.11896/jsjkx.220600032
[4] 周乐员, 张剑华, 袁甜甜, 陈胜勇.
多层注意力机制融合的序列到序列中国连续手语识别和翻译
Sequence-to-Sequence Chinese Continuous Sign Language Recognition and Translation with Multi- layer Attention Mechanism Fusion
计算机科学, 2022, 49(9): 155-161. https://doi.org/10.11896/jsjkx.210800026
[5] 熊丽琴, 曹雷, 赖俊, 陈希亮.
基于值分解的多智能体深度强化学习综述
Overview of Multi-agent Deep Reinforcement Learning Based on Value Factorization
计算机科学, 2022, 49(9): 172-182. https://doi.org/10.11896/jsjkx.210800112
[6] 姜梦函, 李邵梅, 郑洪浩, 张建朋.
基于改进位置编码的谣言检测模型
Rumor Detection Model Based on Improved Position Embedding
计算机科学, 2022, 49(8): 330-335. https://doi.org/10.11896/jsjkx.210600046
[7] 朱承璋, 黄嘉儿, 肖亚龙, 王晗, 邹北骥.
基于注意力机制的医学影像深度哈希检索算法
Deep Hash Retrieval Algorithm for Medical Images Based on Attention Mechanism
计算机科学, 2022, 49(8): 113-119. https://doi.org/10.11896/jsjkx.210700153
[8] 孙奇, 吉根林, 张杰.
基于非局部注意力生成对抗网络的视频异常事件检测方法
Non-local Attention Based Generative Adversarial Network for Video Abnormal Event Detection
计算机科学, 2022, 49(8): 172-177. https://doi.org/10.11896/jsjkx.210600061
[9] 闫佳丹, 贾彩燕.
基于双图神经网络信息融合的文本分类方法
Text Classification Method Based on Information Fusion of Dual-graph Neural Network
计算机科学, 2022, 49(8): 230-236. https://doi.org/10.11896/jsjkx.210600042
[10] 汪鸣, 彭舰, 黄飞虎.
基于多时间尺度时空图网络的交通流量预测模型
Multi-time Scale Spatial-Temporal Graph Neural Network for Traffic Flow Prediction
计算机科学, 2022, 49(8): 40-48. https://doi.org/10.11896/jsjkx.220100188
[11] 金方焱, 王秀利.
融合RACNN和BiLSTM的金融领域事件隐式因果关系抽取
Implicit Causality Extraction of Financial Events Integrating RACNN and BiLSTM
计算机科学, 2022, 49(7): 179-186. https://doi.org/10.11896/jsjkx.210500190
[12] 熊罗庚, 郑尚, 邹海涛, 于化龙, 高尚.
融合双向门控循环单元和注意力机制的软件自承认技术债识别方法
Software Self-admitted Technical Debt Identification with Bidirectional Gate Recurrent Unit and Attention Mechanism
计算机科学, 2022, 49(7): 212-219. https://doi.org/10.11896/jsjkx.210500075
[13] 彭双, 伍江江, 陈浩, 杜春, 李军.
基于注意力神经网络的对地观测卫星星上自主任务规划方法
Satellite Onboard Observation Task Planning Based on Attention Neural Network
计算机科学, 2022, 49(7): 242-247. https://doi.org/10.11896/jsjkx.210500093
[14] 张颖涛, 张杰, 张睿, 张文强.
全局信息引导的真实图像风格迁移
Photorealistic Style Transfer Guided by Global Information
计算机科学, 2022, 49(7): 100-105. https://doi.org/10.11896/jsjkx.210600036
[15] 曾志贤, 曹建军, 翁年凤, 蒋国权, 徐滨.
基于注意力机制的细粒度语义关联视频-文本跨模态实体分辨
Fine-grained Semantic Association Video-Text Cross-modal Entity Resolution Based on Attention Mechanism
计算机科学, 2022, 49(7): 106-112. https://doi.org/10.11896/jsjkx.210500224
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!