Computer Science ›› 2021, Vol. 48 ›› Issue (10): 77-84.doi: 10.11896/jsjkx.210300271

• Artificial Intelligence • Previous Articles     Next Articles

Medical Entity Relation Extraction Based on Deep Neural Network and Self-attention Mechanism

ZHANG Shi-hao, DU Sheng-dong, JIA Zhen, LI Tian-rui   

  1. School of Computing and Artificial Intelligence,Southwest Jiaotong University,Chengdu 611756,China
  • Received:2021-03-29 Revised:2021-05-21 Online:2021-10-15 Published:2021-10-18
  • About author:ZHANG Shi-hao,born in 1996,postgraduate.His main research interests include information extraction and natural language processing.
    LI Tian-rui,born in 1969,Ph.D,professor,Ph.D supervisor,is a distinguished member of China Computer Federation.His main research interests include big data intelligence,rough sets and granular computing.
  • Supported by:
    Sichuan Key R&D Project(2020YFG0035).

Abstract: With the advancement of medical informatization,a large amount of unstructured text data has been accumulated in the medical field.How to mine valuable information from these medical texts is a research hotspot in the field of medical profession and natural language processing.With the development of deep learning,deep neural network is gradually applied to relation extraction task,and “recurrent+CNN” network framework has become the mainstream model in medical entity relation extraction task.However,due to the problems of high entity density and the cross-connection of relationships between entities in medical texts,the “recurrent+CNN” network framework cannot deeply mine the semantic features of medical texts.Based on the “recurrent+CNN” network framework,this paper proposes a Chinese medical entity relation extraction model with multi-channel self-attention mechanism.It includes that BLSTM is used to capture the context information of text sentences,a multi-channel self-attention mechanism is used to mine the global semantic features of sentences,and CNN is used to capture the local phrase features of sentences.The effectiveness of the model is verified by experiments on Chinese medical text dataset.The precision,recall and F1 value of the model are improved compared with the mainstream models.

Key words: Deep learning, Entity relation extraction, Medical text, Multi-channel self-attention

CLC Number: 

  • TP391
[1]GOLSHAN P N,DASHTI H R,AZIZI S,et al.A Study of Recent Contributions on Information Extraction[J].arXiv:1803.05667,2018.
[2]LIU Q,LI Y,DUAN H,et al.Knowledge Graph Construction Techniques[J].Journal of Computer Research and Development,2016,53:582-600.
[3]GRISHMAN R,SUNDHEIM B.Message understanding confe-rence-6:a brief history[C]//Proceedings of the 16th Conference on Computational Linguistics.New York:ACM Press,1996:466-471.
[4]UZUNER O,SOUTH B,SHEN S Y,et al.2010 i2b2/VA challenge on concepts,assertions,and relations in clinical text[J].Journal of the American Medical Informatics Association,2011,18(5):552-556.
[5]NING S M,TENG F,LI T R.Multi-Channel Self-AttentionMechanism for Relation Extraction in Clinical Records[J].Chinese Journal of Computers,2020,43(5):916-929.
[6]HAN X,GAO T Y,LIN Y K,et al.More data,more relations,more context and more openness:a Review and outlook for relation extraction[J].arXiv:2004.03186,2020.
[7]ZHAO S,GRISHMAN R.Extraction relations with integrated information using kernel methods[C]//Proceedings of the 41st Annual Meeting of the Association for Computational Linguistics.Stroudsburg:ACL,2005:419-426.
[8]GUO X Y,HE T T,HU X H,et al.Chinese Named Entity Relation Extraction Based Syntactic and Semantic Features[J].Journal of Chinese Information Processing,2014,28(6):183-189.
[9]KAMBHATLA N.Combining lexical,syntactic,and semanticfeatures with maximum entropy models for extracting relation[C]//Proceedings of ACL on Interactive Poster and Demonstration Sessions.Stroudsburg:ACL,2004:22-26.
[10]ZHOU J.Chinese entity relation extraction based on conditional random fields model[J].Computer Engineering,2010,36(24):192-194.
[11]RINK B,HARABAGIU S,ROBERTS K.Automatic extraction of relations between medical concepts in clinical texts[J].Journal of the American Medical Informatics Association,2011,18(5):594-600.
[12]D'SOUZA J,NG V.Ensemble-Based Medical Relation Classification[C]//25th International Conference on Computational Linguistics.Dublin:COLING,2014:1682-1693.
[13]KIM S,LIU H,YEGANOVA L,et al.Extracting drug-drug interactions from literature using a rich feature-based linear kernel approach[J].Journal of Biomedical Informatics,2015,55(2):23-30.
[14]ZENG D J,LIU K,LAI S W,et al.Relation classification via convolutional deep neural network[C]//Proceedings of the 25th International Conference on Computational Linguistics.Stroudsburg:ACL,2014:2335-2344.
[15]ZHANG D X,WANG D.Relation classification via recurrent neural network[J].arXiv:1508.01006,2015.
[16]ZHANG S,ZHENG D Q,HU X C,et al.Bidirectional Long short-term memory networks for relation classification[C]//Proceedings of the 29th Pacific Asia Conference on Language,Information and Computation.Stroudsburg:ACL,2015:73-78.
[17]ZHU J Z,QIAO J Z,DAI X X,et al.Relation classification via target-concentrated attention CNNs[C]//International Confe-rence on Neural Information Processing.Berlin:Springer,2017:137-146.
[18]ZHOU P,SHI W,TIAN J,et al.Attention-based bidirectional long short-term memory networks for relation classification[C]//Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics.Stroudsburg:ACL,2016:207-212.
[19]LEE J,SEO S,CHOI Y S.Semantic relation classification via bidirectional LSTM networks with entity-aware attention using latent entity typing[J].arXiv:1901.08163,2019.
[20]CAI R,ZHANG X D,WANG H F.Bidirectional recurrent con-volutional neural network for relation classification[C]//Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics.Stroudsburg:ACL,2016:756-765.
[21]ZHANG X B,CHEN F C,HUANG R Y.A combination of RNN and CNN for attention-based relation classification[J].Procedia Computer Science,2018,131:911-917.
[22]TRAN V H,PHI V T,SHINDO H,et al.Relation Classification Using Segment-Level Attention-based CNN and Dependency-based RNN[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics.Stroudsburg:ACL,2019:2793-2798.
[23]SAHU S,ANAND A,ORUGANTY K,et al.Relation extraction from clinical texts using domain invariant convolutional neural network[C]//Proceedings of the 15th Workshop on Biomedical Natural Language Processing.2016:206-215.
[24]ZHOU H W,LANG C K,LIU Z,et al.Knowledge-guided con-volutional networks for chemical-disease relation extraction[J].BMC Bioinformatics,2019,20(1):260-273.
[25]SAHU S,ANAND A.Drug-Drug Interaction Extraction from Biomedical Texts Using Long Short-Term Memory Network[J].Journal of Biomedical Informatics,2018,86:15-24.
[26]BAI T,WANG C,WANG Y,et al.A novel deep learning me-thod for extracting unspecific biomedical relation[J].Concurrency and Computation:Practice and Experience,2020,32:1-11.
[27]RAJ D,SAHU S,ANAND A.Learning local and global contexts using a convolutional recurrent network model for relation classification in biomedical text[C]//Proceedings of the 21st Conference on Computational Natural Language Learning.Vancouver:CoNLL,2017:311-321.
[28]HE B,GUAN Y,DAI R.Convolutional Gated Recurrent Units for Medical Relation Classification[C]//2018 IEEE InternationalConference on Bioinformatics and Biomedicine.New York:IEEE Press,2019:646-650.
[29]LIN Z,FENG M,SANTOS C N,et al.A structured self-attentive sentence embedding[J].arXiv:1703.03130,2017.
[1] XU Yong-xin, ZHAO Jun-feng, WANG Ya-sha, XIE Bing, YANG Kai. Temporal Knowledge Graph Representation Learning [J]. Computer Science, 2022, 49(9): 162-171.
[2] RAO Zhi-shuang, JIA Zhen, ZHANG Fan, LI Tian-rui. Key-Value Relational Memory Networks for Question Answering over Knowledge Graph [J]. Computer Science, 2022, 49(9): 202-207.
[3] TANG Ling-tao, WANG Di, ZHANG Lu-fei, LIU Sheng-yun. Federated Learning Scheme Based on Secure Multi-party Computation and Differential Privacy [J]. Computer Science, 2022, 49(9): 297-305.
[4] WANG Jian, PENG Yu-qi, ZHAO Yu-fei, YANG Jian. Survey of Social Network Public Opinion Information Extraction Based on Deep Learning [J]. Computer Science, 2022, 49(8): 279-293.
[5] HAO Zhi-rong, CHEN Long, HUANG Jia-cheng. Class Discriminative Universal Adversarial Attack for Text Classification [J]. Computer Science, 2022, 49(8): 323-329.
[6] JIANG Meng-han, LI Shao-mei, ZHENG Hong-hao, ZHANG Jian-peng. Rumor Detection Model Based on Improved Position Embedding [J]. Computer Science, 2022, 49(8): 330-335.
[7] SUN Qi, JI Gen-lin, ZHANG Jie. Non-local Attention Based Generative Adversarial Network for Video Abnormal Event Detection [J]. Computer Science, 2022, 49(8): 172-177.
[8] HU Yan-yu, ZHAO Long, DONG Xiang-jun. Two-stage Deep Feature Selection Extraction Algorithm for Cancer Classification [J]. Computer Science, 2022, 49(7): 73-78.
[9] CHENG Cheng, JIANG Ai-lian. Real-time Semantic Segmentation Method Based on Multi-path Feature Extraction [J]. Computer Science, 2022, 49(7): 120-126.
[10] HOU Yu-tao, ABULIZI Abudukelimu, ABUDUKELIMU Halidanmu. Advances in Chinese Pre-training Models [J]. Computer Science, 2022, 49(7): 148-163.
[11] ZHOU Hui, SHI Hao-chen, TU Yao-feng, HUANG Sheng-jun. Robust Deep Neural Network Learning Based on Active Sampling [J]. Computer Science, 2022, 49(7): 164-169.
[12] SU Dan-ning, CAO Gui-tao, WANG Yan-nan, WANG Hong, REN He. Survey of Deep Learning for Radar Emitter Identification Based on Small Sample [J]. Computer Science, 2022, 49(7): 226-235.
[13] WANG Jun-feng, LIU Fan, YANG Sai, LYU Tan-yue, CHEN Zhi-yu, XU Feng. Dam Crack Detection Based on Multi-source Transfer Learning [J]. Computer Science, 2022, 49(6A): 319-324.
[14] CHU Yu-chun, GONG Hang, Wang Xue-fang, LIU Pei-shun. Study on Knowledge Distillation of Target Detection Algorithm Based on YOLOv4 [J]. Computer Science, 2022, 49(6A): 337-344.
[15] ZHOU Zhi-hao, CHEN Lei, WU Xiang, QIU Dong-liang, LIANG Guang-sheng, ZENG Fan-qiao. SMOTE-SDSAE-SVM Based Vehicle CAN Bus Intrusion Detection Algorithm [J]. Computer Science, 2022, 49(6A): 562-570.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!