Computer Science ›› 2021, Vol. 48 ›› Issue (6A): 295-298.doi: 10.11896/jsjkx.200500019

• Intelligent Computing • Previous Articles     Next Articles

Relation Classification of Chinese Causal Compound Sentences Based on Transformer Model and Relational Word Feature

YANG Jin-cai, CAO Yuan, HU Quan, SHEN Xian-jun   

  1. School of Computer Science,Central China Normal University,Wuhan 430079,China
  • Online:2021-06-10 Published:2021-06-17
  • About author:YANG Jin-cai,born in 1967,Ph.D,professor,Ph.D supervisor,is a member of China Computer Federation.His main research interests include advanced database and information system,Chinese information processing,artificial intelligence and natural language processing.
  • Supported by:
    National Social Science Fundation of China(19BYY092).

Abstract: Chinese compound sentences have rich and complicated semantic relations.Recognition of relation category of a Chinese compound sentence is judgment of semantic relations of the sentence,and it is very important to analyze the meaning of compound sentences.Causal compound sentences are the most frequently used sentences in Chinese article.In this paper,a corpus of causal compound sentences with two clauses is taken as research object.A deep learning method is used in order to mine the hidden features of compound sentences automatically.At the same time,this paperintegrates significant linguistics knowledge of relational words in the proposed model.Combining word vector of word2vec and relation word feature of one hot coding as input to the model,a transformer model using convolutional neural network as feedforward layer is exploited to identify the relation category of causal compound sentences.Using our model,the F1 value of the experiment reaches 92.13%,which is better than the compa-rative model.

Key words: Causal compound sentences, Deep learning, Relation classification, Transformer model, Word vector

CLC Number: 

  • TP391
[1] JOTY S,GUZMÁN F,MÀRQUEZ L I,et al.Using Discourse Structure for Machine Translation Evaluation[C]//ACL Procee-dings of the 45th Annual Meeting of the Association for Computational Linguistics.2014:687-698.
[2] HUTTUNEN S,VIHAVAINEN A,VON ETTER P,et al.Re-levance Prediction in Information Extraction using Discourse and Lexical Features[C]//Proceedings of the 18th Nordic Confe-rence of Computational Linguistics(NODALIDA 2011).2011:114-121.
[3] VERBERNE S,BOVES L,OOSTDIJK N,et al.Evaluating discourse-based answer extraction for why -question answering[C] //ACM.2007.
[4] RONG L H,ZHOU G D.An Overview of the Study on Chinese Cause-effect Complex Sentence [J].Journal of Changchun Normal University,2011,30(9):47-51.
[5] YANG J C,CHEN Z Z,SHENG X J,et al.Automatic recognition of relation category of non-saturated compound sentences with two clauses [J].Application Research of Computers,2017,34(10):2950-2953.
[6] YANG J C,CHEN Z Z,SHENG X J,et al.Word Semantic Relevancy Computation and Categories Identification[J].Application Research of Computers,2017,44(5):280-284.
[7] HUANG H,CHANG T,CHEN H,et al.Interpretation of Chinese Discourse Connectives for Explicit Discourse Relation Re-cognition[C]//the 25th International Conference on Computational Linguistics.2014,15(1):632-643.
[8] XI X F,ZHOU G D.A Survey on Deep Learning for NaturalLanguage Processing[J].Acta Automatica Sinica,2016,42(10):1445-1465.
[9] RÖNNQVIST S,SCHENK N,CHIARCOS C.A RecurrentNeural Model with Attention for the Recognition of Chinese Implicit Discourse Relations[EB/OL].[2017-04-26].https://arxiv.org/abs/1704.08092.
[10] ZHOU Y,XUE N.The Chinese Discourse TreeBank:a Chinese corpus annotated with discourse relations[J].Language Resources and Evaluation,2015,49(2):397-431.
[11] QIN L,ZHANG Z,ZHAO H.Shallow Discourse Parsing Using Convolutional Neural Network[C]//Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics.2016:70-77.
[12] XING F Y.A Research on Chinese Compound Sentences[M].Beijing:Commercial Press,2003.
[13] VASWANI A,SHAZEER N.Attention Is All You Need[C] //31st Conference on Neural Information Processing Systems.2017.
[14] IOFFE S,SZEGEDY C.Batch Normalization:Accelerating Deep Network Training by Reducing Internal Covariate Shift[EB/OL].[2015-03-02].https://arxiv.org/abs/1502.03167.
[15] MIKOLOV T,SUTSKEVER I,CHEN K,et al.DistributedRepresentations of Words and Phrases and their Compositionality[J].Advances in Neural Information Processing Systems,2013,26:3111-3119.
[16] SRIVASTAVA N,HINTON G,KRIZHEVSKY A,et al.Dropout:A Simple Way to Prevent Neural Networks from Overfitting[J].Journal of Machine Learning Research,2014,15(1):1929-1958.
[17] CARUANA R,LAWRENCE S,GILES L.Overfitting in Neural Nets:Backpropagation,Conjugate Gradient,and Early Stopping[C]//Proceedings of the 13th International Conference on Neural Information Processing Systems(NIPS'00).2001:381-387.
[18] CARO L D,GRELLA M.Sentiment analysis via dependencyparsing[J].Computer Standards & Interfaces,2013,35(5):442-453.
[19] NIEPERT M,AHMED M,KUTZKOV K.Learning Convolu-tional Neural Networks for Graphs [EB/OL].[2016-05-17].https://arxiv.org/abs/1605.05273.
[1] RAO Zhi-shuang, JIA Zhen, ZHANG Fan, LI Tian-rui. Key-Value Relational Memory Networks for Question Answering over Knowledge Graph [J]. Computer Science, 2022, 49(9): 202-207.
[2] TANG Ling-tao, WANG Di, ZHANG Lu-fei, LIU Sheng-yun. Federated Learning Scheme Based on Secure Multi-party Computation and Differential Privacy [J]. Computer Science, 2022, 49(9): 297-305.
[3] XU Yong-xin, ZHAO Jun-feng, WANG Ya-sha, XIE Bing, YANG Kai. Temporal Knowledge Graph Representation Learning [J]. Computer Science, 2022, 49(9): 162-171.
[4] WANG Jian, PENG Yu-qi, ZHAO Yu-fei, YANG Jian. Survey of Social Network Public Opinion Information Extraction Based on Deep Learning [J]. Computer Science, 2022, 49(8): 279-293.
[5] HAO Zhi-rong, CHEN Long, HUANG Jia-cheng. Class Discriminative Universal Adversarial Attack for Text Classification [J]. Computer Science, 2022, 49(8): 323-329.
[6] JIANG Meng-han, LI Shao-mei, ZHENG Hong-hao, ZHANG Jian-peng. Rumor Detection Model Based on Improved Position Embedding [J]. Computer Science, 2022, 49(8): 330-335.
[7] SUN Qi, JI Gen-lin, ZHANG Jie. Non-local Attention Based Generative Adversarial Network for Video Abnormal Event Detection [J]. Computer Science, 2022, 49(8): 172-177.
[8] HU Yan-yu, ZHAO Long, DONG Xiang-jun. Two-stage Deep Feature Selection Extraction Algorithm for Cancer Classification [J]. Computer Science, 2022, 49(7): 73-78.
[9] CHENG Cheng, JIANG Ai-lian. Real-time Semantic Segmentation Method Based on Multi-path Feature Extraction [J]. Computer Science, 2022, 49(7): 120-126.
[10] HOU Yu-tao, ABULIZI Abudukelimu, ABUDUKELIMU Halidanmu. Advances in Chinese Pre-training Models [J]. Computer Science, 2022, 49(7): 148-163.
[11] ZHOU Hui, SHI Hao-chen, TU Yao-feng, HUANG Sheng-jun. Robust Deep Neural Network Learning Based on Active Sampling [J]. Computer Science, 2022, 49(7): 164-169.
[12] SU Dan-ning, CAO Gui-tao, WANG Yan-nan, WANG Hong, REN He. Survey of Deep Learning for Radar Emitter Identification Based on Small Sample [J]. Computer Science, 2022, 49(7): 226-235.
[13] JIANG Sheng-teng, ZHANG Yi-chi, LUO Peng, LIU Yue-ling, CAO Kuo, ZHAO Hai-tao, WEI Ji-bo. Analysis of Performance Metrics of Semantic Communication Systems [J]. Computer Science, 2022, 49(7): 236-241.
[14] ZHU Wen-tao, LAN Xian-chao, LUO Huan-lin, YUE Bing, WANG Yang. Remote Sensing Aircraft Target Detection Based on Improved Faster R-CNN [J]. Computer Science, 2022, 49(6A): 378-383.
[15] WANG Jian-ming, CHEN Xiang-yu, YANG Zi-zhong, SHI Chen-yang, ZHANG Yu-hang, QIAN Zheng-kun. Influence of Different Data Augmentation Methods on Model Recognition Accuracy [J]. Computer Science, 2022, 49(6A): 418-423.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!