Computer Science ›› 2025, Vol. 52 ›› Issue (12): 224-230.doi: 10.11896/jsjkx.250600140

• Artificial Intelligence • Previous Articles     Next Articles

Adverse Drug Reaction Relationship Extraction Based on Chain of Thought Enhancement UnderHigh and Low Resources

LI Hao, YANG Yumeng, ZHAO Boyang, ZHENG Puqi, LIN Hongfei   

  1. School of Computer Science and Technology, Dalian University of Technology, Dalian, Liaoning 116081, China
  • Received:2025-06-20 Revised:2025-09-23 Online:2025-12-15 Published:2025-12-09
  • About author:LI Hao,born in 2001,postgraduate.His main research interests include natural language processing,large language models and information extraction.
    LIN Hongfei,born in 1962,Ph.D,professor,Ph.D supervisor.His main research interests include natural language processing,affective computing and computational humor,information retrieval and recommendation systems.
  • Supported by:
    This work was supported by the National Natural Science Foundation of China(62302076,62276043).

Abstract: ADRs refer to harmful and unintended responses that occur when drugs are administered at normal doses for prevention,diagnosis,treatment of diseases,or regulation of physiological functions.These reactions are unexpected and can significantly impact patients’ health and physical condition.To identify ADR-related relational triples from social media and biomedical literature data—providing early warnings for patients and healthcare systems,as well as references for pharmaceutical research and drug development-this paper proposes two frameworks based on generative large language models for both low-resource and high-resource annotated data scenarios:the ADR-GPT framework,which extracts ADR relations using Chain-of-Thought promp-ting,and the CADRE-LLM framework,which leverages Chain-of-Thought-enhanced fine-tuning.It evaluates the performance of these two frameworks on the Twitter and PubMed datasets.CADRE-LLM achieves F1 score improvements of 1.51 percentage points and 1.74 percentage points over previous state-of-the-art models on the respective datasets.The low-resource ADR-GPT framework also demonstrates strong performance,outperforming fully supervised fine-tuned Qwen2.5 on the PubMed dataset.Ablation studies further validate the effectiveness of each module within the two proposed frameworks.

Key words: Relation extraction, Adverse drug reactions, Large language models, Chain of Thought(COT), Low resource

CLC Number: 

  • TP311.13
[1]NATIONAL CENTER FOR ADR MONITORING.Annual report of national adverse drug reaction monitoring(2023) [J].Chinese Journal of Viral Diseases,2023,13(4):245-251.
[2]LI H,QIU Y Z,LIN H F.Adverse drug reaction detection with multi-feature enhancement for social media [C]//Proceedings of the 23rd Chinese National Conference on Computational Linguistics.2024:515-525.
[3]ZHANG T,LIN H,REN Y,et al.Adversarial transfer network with bilinear attention for the detection of adverse drug reactions from social media [J].Applied Soft Computing,2021,106:107358.
[4]ZHANG T,LIN H,REN Y,et al.Adverse drug reaction detection via a multihop self-attention mechanism [J].BMC Bioinformatics,2019,20(1):1-11.
[5]ZHENG S,WANG F,BAO H,et al.Joint Extraction of Entities and Relations Based on a Novel Tagging Scheme[C]//Procee-dings of the 55th Annual Meeting of the Association for Computational Linguistics.2017:1227-1236.
[6]REN F,ZHANG L,ZHAO X,et al.A simple but effective bidirectional framework for relational triple extraction [C]//Proceedings of the 15th ACM Int Conf on Web Search and Data Mining.New York:ACM,2022:824-832.
[7]WANG Y,YU B,ZHANG Y,et al.TPLinker:Single-stage joint extraction of entities and relations through token pair linking [C]//Proceedings of the 28th Int Conf on Computational Linguistics.Stroudsburg,PA:ACL,2020:1572-1582.
[8]SHANG Y,HUANG H,MAO X.Onerel:Joint entity and relation extraction with one module in one step [C]//Proceedings of the 36th AAAI Conf on Artificial Intelligence.Palo Alto,CA:AAAI,2022:11285-11293.
[9]YE H,ZHANG N,DENG S,et al.Contrastive triple extraction with generative transformer [C]//Proceedings of the 35th AAAI Conf on Artificial Intelligence.Palo Alto,CA:AAAI,2021:14257-14265.
[10]DEVLIN J,CHANG M,LEE K,et al.BERT:Pre-training ofdeep bidirectional transformers for language understanding [C]//Proceedings of the 2019 Conf of the North American Chapter of the ACL:Human Language Technologies.Stroudsburg,PA:ACL,2019:4171-4186.
[11]WANG X,ZHOU W,ZU C,et al.InstructUIE:Multi-task instruction tuning for unified information extraction [J].arXiv:2304.08085,2023.
[12]BAI J,BAI S,CHU Y,et al.Qwen technical report [J].arXiv:2309.16609,2023.
[13]LUO L,NING J,ZHAO Y,et al.Taiyi:A bilingual fine-tuned large language model for diverse biomedical tasks [J].Journal of the American Medical Informatics Association,2024,31(9):1865-1874.
[14]WADHWA S,AMIR S,WALLACE B.Revisiting relation ex-traction in the era of large language models [C]//Proceedings of the 61st Annual Meeting of the ACL.Stroudsburg,PA:ACL,2023:15566-15577.
[15]KOJIMA T,GU S,REID M,et al.Large language models are zero-shot reasoners [J].Advances in Neural Information Processing Systems,2022,35:22199-22213.
[16]SHUM K,DIAO S,ZHANG T.Automatic prompt augmenta-tion and selection with chain-of-thought from labeled data [C]//Findings of EMNLP 2023.Stroudsburg,PA:ACL,2023:12113-12139.
[17]REIMERS N,GUREVYCH I.Sentence-BERT:Sentence em-beddings using Siamese BERT-networks [C]//Proceedings of EMNLP-IJCNLP 2019.Stroudsburg,PA:ACL,2019:3982-3992.
[18]DETTMERS T,PAGNONI A,HOLTZMAN A,et al.QLoRA:Efficient finetuning of quantized LLMs [J].Advances in Neural Information Processing Systems,2023,36:10088-10115.
[19]ALVARO N,MIYAO Y,COLLIER N.TwiMed:Twitter andPubMed comparable corpus of drugs,diseases,symptoms,and their relations [J].JMIR Public Health and Surveillance,2017,3(2):e6396.
[20]WEI Z,SU J,WANG Y,et al.A Novel Cascade Binary Tagging Framework for Relational Triple Extraction[C]//Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics.2020:1476-1488.
[21]ACHIAM J,ADLER S,AGARWAL S,et al.GPT-4 technicalreport [J].arXiv:2303.08774,2023.
[22]GRATTAFIORI A,DUBEY A,JAUHRI A,et al.The LLaMA 3 herd of models [J].arXiv:2407.21783,2024.
[1] LIU Le, XIAO Rong, YANG Xiao. Application of Decoupled Knowledge Distillation Method in Document-level RelationExtraction [J]. Computer Science, 2025, 52(8): 277-287.
[2] LI Maolin, LIN Jiajie, YANG Zhenguo. Confidence-guided Prompt Learning for Multimodal Aspect-level Sentiment Analysis [J]. Computer Science, 2025, 52(7): 241-247.
[3] CHEN Jinyin, XI Changkun, ZHENG Haibin, GAO Ming, ZHANG Tianxin. Survey of Security Research on Multimodal Large Language Models [J]. Computer Science, 2025, 52(7): 315-341.
[4] LI Bo, MO Xian. Application of Large Language Models in Recommendation System [J]. Computer Science, 2025, 52(6A): 240400097-7.
[5] ZOU Rui, YANG Jian, ZHANG Kai. Low-resource Vietnamese Speech Synthesis Based on Phoneme Large Language Model andDiffusion Model [J]. Computer Science, 2025, 52(6A): 240700138-6.
[6] HU Caishun. Study on Named Entity Recognition Algorithms in Audit Domain Based on Large LanguageModels [J]. Computer Science, 2025, 52(6A): 240700190-4.
[7] GAO Hongkui, MA Ruixiang, BAO Qihao, XIA Shaojie, QU Chongxiao. Research on Hybrid Retrieval-augmented Dual-tower Model [J]. Computer Science, 2025, 52(6): 324-329.
[8] LI Xiwang, CAO Peisong, WU Yuying, GUO Shuming, SHE Wei. Study on Security Risk Relation Extraction Based on Multi-view IB [J]. Computer Science, 2025, 52(5): 330-336.
[9] GUO Liwei, WU Yonghao, LIU Yong. Semantic Variations Based Defect Generation and Prediction Model Testing [J]. Computer Science, 2025, 52(11A): 241200059-7.
[10] WEI Hao, ZHANG Zongyu, DIAO Hongyue, DENG Yaochen. Review of Application of Information Extraction Technology in Digital Humanities [J]. Computer Science, 2025, 52(11A): 250600198-10.
[11] HUANG Haixin, XU Chenglong, FU Yao. Research on Structured Pruning Algorithm Based on Information Fusion [J]. Computer Science, 2025, 52(11A): 241000041-6.
[12] PAN Jie, WANG Juan, WANG Nan. Large Language Models and Rumors:A Survey on Generation and Detection [J]. Computer Science, 2025, 52(11): 1-12.
[13] FANG Quan, ZHANG Jinlong, WANG Bingqian, HU Jun. Research on Domain Knowledge Question Answering via Large Language Models withCompositional Context Prompting [J]. Computer Science, 2025, 52(11): 13-21.
[14] ZHANG Haoran, HAO Wenning, JIN Dawei, CHENG Kai, ZHAI Ying. DF-RAG:A Retrieval-augmented Generation Method Based on Query Rewriting and Knowledge Selection [J]. Computer Science, 2025, 52(11): 30-39.
[15] ZHOU Yuchen, LI Peng, HAN Keji. Instruct-Malware:Control Flow Graph Based Large Language Model Analysis of Malware [J]. Computer Science, 2025, 52(11): 40-48.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!