Computer Science ›› 2020, Vol. 47 ›› Issue (11A): 24-27.doi: 10.11896/jsjkx.200400116

• Artificial Intelligence • Previous Articles     Next Articles

CNN_BiLSTM_Attention Hybrid Model for Text Classification

WU Han-yu1,2, YAN Jiang2, HUANG Shao-bin1, LI Rong-sheng1, JIANG Meng-qi1   

  1. 1 College of Computer Science and Technology,Harbin Engineering University,Harbin 150001,China
    2 Big Data Application on Improving Government Governance Capabilities National Engineering Laboratory,CETC Big Data Research Institute Co.,Ltd.,Guiyang 550000,China
  • Online:2020-11-15 Published:2020-11-17
  • About author:WU Han-yu,born in 1996,M.S..His main research interests include natural language processing,deep learning,etc.
  • Supported by:
    This work was supported by the Big Data Application on Improving Government Governance Capabilities National Engineering Laboratory Open Fund Project.

Abstract: Text classification is the basis of many natural language processing tasks.Convolutional neural network (CNN) can be used to extract the phrase level features of text,but it can't capture the structure information of text well;Recurrent neural network (RNN) can extract the global structure information of text,but its ability to capture the key pattern information is insufficient.Attention mechanism can learn the distribution of different words or phrases to the overall semantics of text,key words or phrases will be assigned higher weights,but it is not sensitive to global structure information.In addition,most of the existing models only consider word level information,but ignore phrase level information.In view of the problems in the above models,this paper proposes a hybrid model which integrates CNN,RNN and attention.The model considers the key pattern information and global structure information of different levels at the same time,and fuses them to get the final text representation.Finally,the text representation is input to the softmax layer for classification.Experiments on multiple text classification datasets show that the model can achieve higher accuracy than the existing models.In addition,the effects of different components on the performance of the model are analyzed through experiments.

Key words: Global structure information, Hybrid model, Key pattern information, Text classification, Text representation

CLC Number: 

  • TP391.1
[1] JOACHIMS T.Text categorization with support vector ma-chines:Learning with many relevant features[C]//European Conference on Machine Learning.Berlin:Springer,1998:137-142.
[2] CHEN Z,SHI G,WANG X.Text classification based on Naive Bayes algorithm with feature selection[J].International Information Institute (Tokyo).Information,2012,15(10):4255.
[3] KIM Y.Convolutional neural networks for sentence classification[J].arXiv:1408.5882,2014.
[4] KALCHBRENNER N,GREFENSTETTE E,BLUNSOM P.A convolutional neural network for modelling sentences[J].arXiv:1404.2188,2014.
[5] ZHANG X,ZHAO J,LECUN Y.Character-level convolutional networks for text classification[C]//Advances in Neural Information Processing Systems.2015:649-657.
[6] CONNEAU A,SCHWENK H,BARRAULT L,et al.Very deep convolutional networks for text classification[J].arXiv:1606.01781,2016.
[7] TANG D,QIN B,LIU T.Document modeling with gated recurrent neural network for sentiment classification[C]//Procee-dings of the 2015 Conference on Empirical Methods in Natural Language Processing.2015:1422-1432.
[8] WANG B.Disconnected recurrent neural networks for text categorization[C]//Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1:Long Papers).2018:2311-2320.
[9] YANG Z,YANG D,DYER C,et al.Hierarchical attention networks for document classification[C]//Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies.2016:1480-1489.
[10] LIN Z,FENG M,SANTOS C N,et al.A structured self-attentive sentence embedding[J].arXiv:1703.03130,2017.
[11] LIU Y,ZHAI D H,REN Q N.News Text Classification Based on CNLSTM Model with Attention Mechanism[J].Computer Engineering,2019,45(7):303-308,314.
[12] GU J H,PENG W T,LI N N,et al.Sentiment classificationmethod based on convolution attention mechanism[J].ComputerEmgineering and Design,2020,41(1):95-99.
[13] PENNINGTON J,SOCHER R,MANNING C.Glove:Globalvectors for word representation[C]//Proceedings of the 2014 Conference on Empirical Methods in Natural Language Proces-sing (EMNLP).2014:1532-1543.
[14] PANG B,LEE L.Seeing stars:Exploiting class relationships for sentiment categorization with respect to rating scales[C]//Proceedings of the 43rd Annual Meeting on Association for Computational Linguistics.Association for Computational Linguistics,2005:115-124.
[15] PANG B,LEE L.A sentimental education:Sentiment analysisusing subjectivity summarization based on minimum cuts[C]//Proceedings of the 42nd Annual Meeting on Association for Computational Linguistics.Association for Computational Linguistics,2004:271.
[16] LI X,ROTH D.Learning question classifiers[C]//Proceedings of the 19th international conference on Computational linguistics-Volume 1.Association for Computational Linguistics,2002:1-7.
[17] HU M,LIU B.Mining and summarizing customer reviews[C]//Proceedings of the Tenth ACM SIGKDD International Confe-rence on Knowledge Discovery and Data Mining.ACM,2004:168-177.
[18] SOCHER R,PERELYGIN A,WU J,et al.Recursive deep mo-dels for semantic compositionality over a sentiment treebank[C]//Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing.2013:1631-1642.
[19] KINGMA D P,BA J.Adam:A method for stochastic optimization[J].arXiv:1412.6980,2014.
[20] YIN W,SCHÜTZE H.Multichannel variable-size convolutionfor sentence classification[J].arXiv:1603.04513,2016.
[21] CHO K,VAN MERRIËNBOER B,GULCEHRE C,et al.Learning phrase representations using RNN encoder-decoder for statistical machine translation[J].arXiv:1406.1078,2014.
[22] ZHOU C,SUN C,LIU Z,et al.A C-LSTM Neural Network for Text Classification [J].Computer Science,2015,1(4):39-44.
[23] WANG C,JIANG F,YANG H.A hybrid framework for text modeling with convolutional RNN[C]//Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Disco-very and Data Mining.2017:2061-2069.
[24] ZHOU P,QI Z,ZHENG S,et al.Text classification improved by integrating bidirectional LSTM with two-dimensional max pooling[J].arXiv:1611.06639,2016.
[25] ZHOU F,LI R Y.Convolutional Neural Network Model forText Classification Based on BGRU Pooling[J].Computer Science,2018,45(6):235-240.
[26] ZHENG C,XUE M Y,HONG T T,et al.DC-BiGRU_CNN Model for Short text Classification[J].Computer Science,2019,46(11):186-192.
[1] HAO Zhi-rong, CHEN Long, HUANG Jia-cheng. Class Discriminative Universal Adversarial Attack for Text Classification [J]. Computer Science, 2022, 49(8): 323-329.
[2] TAN Ying-ying, WANG Jun-li, ZHANG Chao-bo. Review of Text Classification Methods Based on Graph Convolutional Network [J]. Computer Science, 2022, 49(8): 205-216.
[3] YAN Jia-dan, JIA Cai-yan. Text Classification Method Based on Information Fusion of Dual-graph Neural Network [J]. Computer Science, 2022, 49(8): 230-236.
[4] WU Hong-xin, HAN Meng, CHEN Zhi-qiang, ZHANG Xi-long, LI Mu-hang. Survey of Multi-label Classification Based on Supervised and Semi-supervised Learning [J]. Computer Science, 2022, 49(8): 12-25.
[5] SHAO Xin-xin. TI-FastText Automatic Goods Classification Algorithm [J]. Computer Science, 2022, 49(6A): 206-210.
[6] DENG Kai, YANG Pin, LI Yi-zhou, YANG Xing, ZENG Fan-rui, ZHANG Zhen-yu. Fast and Transmissible Domain Knowledge Graph Construction Method [J]. Computer Science, 2022, 49(6A): 100-108.
[7] KANG Yan, WU Zhi-wei, KOU Yong-qi, ZHANG Lan, XIE Si-yu, LI Hao. Deep Integrated Learning Software Requirement Classification Fusing Bert and Graph Convolution [J]. Computer Science, 2022, 49(6A): 150-158.
[8] DENG Zhao-yang, ZHONG Guo-qiang, WANG Dong. Text Classification Based on Attention Gated Graph Neural Network [J]. Computer Science, 2022, 49(6): 326-334.
[9] LIU Shuo, WANG Geng-run, PENG Jian-hua, LI Ke. Chinese Short Text Classification Algorithm Based on Hybrid Features of Characters and Words [J]. Computer Science, 2022, 49(4): 282-287.
[10] ZHONG Gui-feng, PANG Xiong-wen, SUI Dong. Text Classification Method Based on Word2Vec and AlexNet-2 with Improved AttentionMechanism [J]. Computer Science, 2022, 49(4): 288-293.
[11] WU Yu-kun, LI Wei, NI Min-ya, XU Zhi-cheng. Anomaly Detection Model Based on One-class Support Vector Machine Fused Deep Auto-encoder [J]. Computer Science, 2022, 49(3): 144-151.
[12] DENG Wei-bin, ZHU Kun, LI Yun-bo, HU Feng. FMNN:Text Classification Model Fused with Multiple Neural Networks [J]. Computer Science, 2022, 49(3): 281-287.
[13] ZHANG Hu, BAI Ping. Graph Convolutional Networks with Long-distance Words Dependency in Sentences for Short Text Classification [J]. Computer Science, 2022, 49(2): 279-284.
[14] DU Shao-hua, WAN Huai-yu, WU Zhi-hao, LIN You-fang. Customs Commodity HS Code Classification Integrating Text Sequence and Graph Information [J]. Computer Science, 2021, 48(4): 97-103.
[15] LU Bo-ren, HU Shi-zhe, LOU Zheng-zheng, YE Yang-dong. Character-level Feature Extraction Method for Railway Text Classification [J]. Computer Science, 2021, 48(3): 220-226.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!