计算机科学 ›› 2016, Vol. 43 ›› Issue (12): 277-280.doi: 10.11896/j.issn.1002-137X.2016.12.051

• 智能应用 • 上一篇    下一篇

一种基于联合深度神经网络的食品安全信息情感分类模型

刘金硕,张智   

  1. 武汉大学计算机学院 武汉430072,武汉大学计算机学院 武汉430072
  • 出版日期:2018-12-01 发布日期:2018-12-01
  • 基金资助:
    本文受国家自然科学基金(61303214)资助

Sentiment Analysis on Food Safety News Using Joint Deep Neural Network Model

LIU Jin-shuo and ZHANG Zhi   

  • Online:2018-12-01 Published:2018-12-01

摘要: 针对因中文食品安全文本特征表达困难,而造成语义信息缺失进而导致分类器准确率低下的问题,提出一种基于深度神经网络的跨文本粒度情感分类模型。以食品安全新闻报道为目标语料,采用无监督的浅层神经网络初始化文本的词语级词向量。引入递归神经网络,将预训练好的词向量作为下层递归神经网络(Recursive Neural Network)的输入层,计算得到具备词语间语义关联性的句子特征向量及句子级的情感倾向输出,同时动态反馈调节词向量特征,使其更加接近食品安全特定领域内真实的语义表达。然后,将递归神经网络输出的句子向量以时序逻辑作为上层循环神经网络(Recurrent Neural Network)的输入,进一步捕获句子结构的上下文语义关联信息,实现篇章级的情感倾向性分析任务。实验结果表明,联合深度模型在食品安全新闻报道的情感分类任务中具有良好的效果,其分类准确率和F1值分别达到了86.7%和85.9%,较基于词袋思想的SVM模型有显著的提升。

关键词: 联合神经网络模型,多粒度文本特征,词向量,食品安全,情感倾向性分析

Abstract: Facing the difficulties in feature expression of Chinese food safety text information,and the loss of semantic information with low classification accuracy,a sentimental text classification model based on joint deep neural network was presented.The proposed model utilizes the corpora of food safety document captured from the internet,and word vector from word embedding method as the input for the neural network to get the pre-trained word vector.The pre-trained word vector is further trained dynamically to get the word features and the sentimental classification of the sentence result,which better express the phrase-level sentimental relations for each sentence and the real semantic meaning in the food safety domain.Then the word feature of the sentence is inputted to the recurrent neural network (RNN) to catch the semantic information of the sentence structure further,realizing the sentimental classification of the text.The experiments show that our joint deep neural network model achieves better results in sentiment analysis on food safety information,compared with the bag-of-words based SVM model.The classification accuracy and F1 value reach 86.7% and 85.9% respectively.

Key words: Joint deep neural networks model,Multi-dimensional textual features,Word-embedding,Food safety,Sentiment analysis

[1] Zhao Yan-yan,Qin Bing,Liu Ting.Sentiment Analysis[J].Jour-nal of Software,2010,1(8):1834-1848(in Chinese) 赵妍妍,秦兵,刘挺.文本情感分析[J].软件学报,2010,21(8):1834-1848
[2] Pang B,Lee L.Opinion mining and sentiment analysis[J].Foundations and Trends in Information Retrieval,2008,2(1/2):1-135
[3] Pang B,Lee L.Seeing stars:Exploiting class relationships forsentiment categorization with respect to rating scales[C]∥Proceedings of the 43rd Annual Meeting on Association for Computational Linguistics.Association for Computational Linguistics,2005:115-124
[4] Pang B,et al.A sentimental education:Sentiment analysis using subjectivity summarization based on minimum cuts[C]∥Proceeding of ACL.2004:271-278
[5] Yang B S,Cardie C.Joint Modeling of Opinion Expression Extraction and Attribute Classification[J].Transactions of the Association for Computational Linguistics,2014,2(1):505-516
[6] Zhang Zhi-lin,Zong Cheng-qing.Sentiment Analysis of Chinese Micro Blog Based on Rich-features[J].Journal of Chinese Information Processing,2015,29(4):134-142(in Chinese) 张志琳,宗成庆.基于多样化特征的中文微博情感分类方法研究[J].中文信息学报,2015,29(4):134-142
[7] Liang jun,Chai Yu-mei,Yuan Hui-bin,et al.Deep Learning for Chinese Micro-blog Sentiment Analysis[J].Journal of Chinese Information Processing,2014,28(5):155-160(in Chinese) 梁军,柴玉梅,原慧斌,等.基于深度学习的微博情感分析[J].中文信息学报,2014,28(5):155-160
[8] Maas A L,Daly R E,Pham P T,et al.Learning word vectors for sentiment analysis[C]∥Proceedings of the 49th Annual Mee-ting of the Association for Computational Linguistics:Human Language Technologies.2011,1:142-150
[9] Socher R,Perelygin A,Wu J Y,et al.Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank[C]∥Proceedings of the Empirical Methods in Natural Language Processing.2013:1631-1642
[10] Mikolov T,Zweig G.Context Dependent Recurrent Neural Network Language Model[J].IEEE Spoken Language Technology Workshop,2012,8537(11):234-239
[11] Hinton G.E,et al.Learning Distributed Representations of Concepts[C]∥Proceedings of the 8th Annual Conference of the Cognitive Science Society.1986
[12] Milkolov T,Kombrink S,et al.Extensions of Recurrent Neural Network Language Model [J].IEEE International Conference on Acoustics,Speech & Signal Processing,2011,125(3):5528-5531
[13] Milkolov T,Chen K,Corrando G,et al.Efficient estimation of word representations in vector space[J].arXiv:1301.3781
[14] Boden M.A Guide to Recurrent Neural Networks and Backpropagation [J].Dallas Project Sics Technical Report T Sics,2001,36(6):112-119

No related articles found!
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!