Computer Science ›› 2025, Vol. 52 ›› Issue (6A): 240700190-4.doi: 10.11896/jsjkx.240700190

• Large Language Model Technology and Its Application • Previous Articles     Next Articles

Study on Named Entity Recognition Algorithms in Audit Domain Based on Large LanguageModels

HU Caishun   

  1. Naval University of Engineering,Wuhan 430000,China
  • Online:2025-06-16 Published:2025-06-12
  • About author:HU Caishun,born in 1989,postgra-duate.His main research interests include finance and auditing.

Abstract: With the emergence of ChatGPT,large language models have begun to play a significant role across various industries,from general fields to specialized domains.Although there have been methods combining artificial intelligence with auditing,the application of large language models in auditing still needs further research due to the fact that the accuracy of traditional artificial intelligence methods is much lower than that of existing large language models.The use of AI methods to intelligently identify useful entities within text in auditing can greatly enhance work efficiency and reduce errors.Conventional auditing text entity recog-nition algorithms primarily rely on machine learning combined with feature engineering,which generally results in lower accuracy.In light of this,this study investigates the applications of several common open-source models(such as Llama) and closed-source models(such as ChatGPT) in auditing text entity recognition,while integrating contextual learning techniques to improve model recognition performance.The results demonstrate that by employing a sample organization method based on similarity selection,the accuracy of entity recognition can be improved to 98.3%,achieving notable improvements.

Key words: Audit, Large language models, ChatGPT, Named entity recognition, In-context learning

CLC Number: 

  • TP391
[1]CHEN X,OUYANG C,LIU Y,et al.Improving the named enti-ty recognition of Chinese electronic medical records by combining domain dictionary and rules[J].International Journal of Environmental Research and Public Health,2020,17(8):2687-2703.
[2]PATIL N,PATIL A,PAWAR B V.Named entity recognition using conditional random fields[J].Procedia Computer Science,2020,167:1181-1188.
[3]ZHANG Y J,XU Z T,XUE X Y.Fusion of multiple features forChinese named entity recognition based on maximum entropy model[J].Journal of Computer Research and Development,2008,45(6):1004-1010.
[4]LAMPLE G,BALLESTEROS M,SUBRA-MANIAN,et al.Neural architectures for named entity recognition[J].arXiv:1603.01360,2016.
[5]WU J C,ZHAN R Z,YANG S,et al.A survey on llm-gernera-ted text detection:Necessity,methods,and future directions[J].arXiv:2310.14724,2023.
[6]GRAVESA.Long-ShortTerm Memory[J].Neural Computation,1997,9(8):1735-1780.
[7]VASWANI A,SHAZEER N,PARMAR N, et al.Attention is all you need[C]//Proceedings of the 31st International Conference on Neural Information Processing Systems.2017:6000-6010.
[8]DEVLIN J,CHANG M W,LEE K,et al.Bert:Pre-training ofdeep bidirectional transformers for language understanding[J].arXiv:1810.04805,2018.
[9]ACHIAM J,ADLER S,AGARWAL S,et al.Gpt-4 technical report[J].arXiv:2303.08774 2023.
[10]LONG O Y,WU J,JIANG X,et al.Training language models to follow instructions with human feedback[C]//Advances in Neural Information Processing Systems.Morial:MIT Press,2022:27730-27744.
[11]YUAN Z H,SHANG Y,ZHOU Y,et al.Llm inference un-veiled:Survey and roofline model insights[J]. arXiv:2402.16363,2024.
[1] LI Bo, MO Xian. Application of Large Language Models in Recommendation System [J]. Computer Science, 2025, 52(6A): 240400097-7.
[2] ZHENG Xinxin, CHEN Fan, SUN Baodan, GONG Jianguang, JIANG Junhui. Question Answering System for Soybean Planting Management Based on Knowledge Graph [J]. Computer Science, 2025, 52(6A): 240500025-8.
[3] LIN Nan, LIU Zhihui, YANG Cong. Named Entity Recognition Algorithm Based on Pre-training Model and Bidirectional TwoDimensional Convolution [J]. Computer Science, 2025, 52(6A): 240700143-6.
[4] GAO Hongkui, MA Ruixiang, BAO Qihao, XIA Shaojie, QU Chongxiao. Research on Hybrid Retrieval-augmented Dual-tower Model [J]. Computer Science, 2025, 52(6): 324-329.
[5] DUN Jingbo, LI Zhuo. Survey on Transmission Optimization Technologies for Federated Large Language Model Training [J]. Computer Science, 2025, 52(1): 42-55.
[6] LI Tingting, WANG Qi, WANG Jiakang, XU Yongjun. SWARM-LLM:An Unmanned Swarm Task Planning System Based on Large Language Models [J]. Computer Science, 2025, 52(1): 72-79.
[7] CHENG Zhiyu, CHEN Xinglin, WANG Jing, ZHOU Zhongyuan, ZHANG Zhizheng. Retrieval-augmented Generative Intelligence Question Answering Technology Based on Knowledge Graph [J]. Computer Science, 2025, 52(1): 87-93.
[8] HUANG Wei, SHEN Yaodi, CHEN Songling, FU Xiangling. CFGT:A Lexicon-based Chinese Address Element Parsing Model [J]. Computer Science, 2024, 51(9): 233-241.
[9] GUO Zhiqiang, GUAN Donghai, YUAN Weiwei. Word-Character Model with Low Lexical Information Loss for Chinese NER [J]. Computer Science, 2024, 51(8): 272-280.
[10] LIU Yumeng, ZHAO Yijing, WANG Bicong, WANG Chao, ZHANG Baomin. Advances in SQL Intelligent Synthesis Technology [J]. Computer Science, 2024, 51(7): 40-48.
[11] YIN Baosheng, ZHOU Peng. Chinese Medical Named Entity Recognition with Label Knowledge [J]. Computer Science, 2024, 51(6A): 230500203-7.
[12] YUE Meng, ZHU Shibo, HONG Xueting, DUAN Bingyan. Airborne Software Audit Method Based on Trusted Implicit Third Party [J]. Computer Science, 2024, 51(6A): 230400088-6.
[13] LAI Xin, LI Sining, LIANG Changsheng, ZHANG Hengyan. Ontology-driven Study on Information Structuring of Aeronautical Information Tables [J]. Computer Science, 2024, 51(6A): 230800150-7.
[14] YU Bihui, TAN Shuyue, WEI Jingxuan, SUN Linzhuang, BU Liping, ZHAO Yiman. Vision-enhanced Multimodal Named Entity Recognition Based on Contrastive Learning [J]. Computer Science, 2024, 51(6): 198-205.
[15] LIAO Meng, JIA Zhen, LI Tianrui. Chinese Named Entity Recognition Based on Label Information Fusion and Multi-task Learning [J]. Computer Science, 2024, 51(3): 198-204.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!