Computer Science ›› 2026, Vol. 53 ›› Issue (4): 347-355.doi: 10.11896/jsjkx.250800012

• Artificial Intelligence • Previous Articles     Next Articles

Agent4Stu:Efficient LLM-based Student Answer Behavior Simulation Agent

LIU Suyi, LIU Qi, GAO Weibo   

  1. School of Artificial Intelligence and Data Science, University of Science and Technology of China, Hefei 230027, China
  • Received:2025-08-04 Revised:2025-11-07 Online:2026-04-15 Published:2026-04-08
  • About author:LIU Suyi,born in 2003,postgraduate.His main research interests include intelligent education and generative agents.
    GAO Weibo,born in 1997,Ph.D.His main research interests include data mining and generative agents.
  • Supported by:
    National Key Research and Development Program of China(2024YFC3308200),National Natural Science Foundation of China(62337001) and Fundamental Research Funds for the Central Universities.

Abstract: Personalized learning has become a core focus of the digital transformation in education,with its success hinging on the precise understanding and modeling of student response data.However,real-world educational scenarios often face challenges such as sparse learning behaviors,dynamic evolution of learning states,and privacy compliance constraints.Additionally,discrepancies between offline data and online learning behaviors result in insufficient behavioral data and distribution shifts,significantly limiting the modeling capabilities and generalization performance of intelligent education systems.To alleviate these dilemmas,previous studies have attempted to simulate student response behaviors to expand data scale and improve model performance. However,existing methods struggle to simultaneously balance generation quality,efficiency,and resource costs.To overcome these limitations,this paper proposes Agent4Stu,a student response behavior simulation framework that integrates large language models(LLMs) with retrieval-augmented generation(RAG) techniques,enabling low-cost,efficient,and highly generalizable personalized response generation.The framework comprises a pre-built retrieval database with retrieval strategies and an LLM-based agent.The retrieval database is constructed from student response behaviors,and two retrieval strategies are designed:similar-student collaborative retrieval and relevant-fact retrieval.These are combined with each student’s short-term memory to dynamically generate highly relevant prompts.Internally,the agent integrates three core modules,profile,memory,and action,which are responsible for modeling students’ learning characteristics and cognitive abilities,integrating historical experiences with know-ledge from the retrieval database,and simulating students’ responses on specific items based on their profile,memory,and know-ledge mastery.Compared with existing LLM-based student agents,Agent4Stu has a smaller memory footprint and simplified action reasoning,while leveraging a behavior-oriented structured retrieval database to provide auxiliary information.This design enables low-cost,efficient,and highly generalizable personalized response generation.Quantitative and qualitative experiments on two real-world learning datasets demonstrate the effectiveness and superiority of Agent4Stu in simulating student learning response behaviors.

Key words: Data synthesis, Large language models, Retrieval-augmented generation, Agents, Human behavior simulation, Cold-start, Intelligent education

CLC Number: 

  • TP391.1
[1]HUANG H R,LIU J H,PAN J W,et al.Systematic transformation of education for the intelligent era:comprehensive education reform empowered by digitalization[J].E-education Research,2025,46(4):5-12.
[2]FENG M,HEFFERNAN N,KOEDINGER K.Addressing theassessment challenge with an online system that tutors as it assesses[J].User Modeling and User-Adapted Interaction,2009,19:243-266.
[3]NAKAGAWA H,IWASAWA Y,MATSUO Y.Graph-basedknowledge tracing:modeling student proficiency using graph neural network[C]//IEEE/WIC/ACM International Confe-rence on Web Intelligence.2019:156-163.
[4]GAO W,LIU Q,YUE L,et al.Agent4Edu:Generating Learner Response Data by Generative Agents for Intelligent Education Systems[J].arXiv:2501.10332,2025.
[5]XU S,ZHANG X,QIN L.Eduagent:Generative student agentsin learning[J].arXiv:2404.07963,2024.
[6]REDDY S,LEVINE S,DRAGAN A.Accelerating h-umanlearning with deep reinforcement learning[C]//NIPS Workshop:Teaching Machines,Robots,and Humans.2017:5-9.
[7]SU Y,LIU Q,LIU Q,et al.Exercise-enhanced seque-ntial modeling for student performance prediction[C]//Proceedings of the AAAI Conference on Artificial Intelligence.2018.
[8]LIU Q,TONG S,LIU C,et al.Exploiting cognitivestructure for adaptive learning[C]//Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mi-ning.2019:627-635.
[9]ZHAO G,HUANG Z,ZHUANG Y,et al.Simulating student interactions with two-stage imitation lea-rning for intelligent educational systems[C]//Proceeding of the 32nd ACM International Conference on Information and Knowledge Management.2023:3423-3432.
[10]PARK J S,O’BRIEN J,CAI C J,et al.Generative agents:Interactive simulacra of human behavior[C]//Proceedings of the 36th Annual Acm Symposium on User Interface Software and Technology.2023:1-22.
[11]QIAN C,CONG X,YANG C,et al.Communicativeagents forsoftware development[J].arXiv:2307.07924,2023.
[12]LIU J,HUANG Z,XIAO T,et al.SocraticLM:Exploring socratic personalized teaching with large language models[J].Advances in Neural Information Processing Systems,2024,37:85693-85721.
[13]RAM O,LEVINE Y,DALMEDIGOS I,et al.In-context retrieval-augmented language models[J].Transactions of the Association for Computational Linguistics,2023,11:1316-1331.
[14]FENG Z,FENG X,ZHAO D,et al.Retrieval-generation synergy augmented large language models[C]//ICASSP 2024-2024 IEEE International Conference on Acoustics,Speech and Signal Processing(ICASSP).IEEE,2024:11661-11665.
[15]BAKER F B.The basics of item response theory[M/OL].http://ericae.net/irt/baker.
[16]ABDELRAHMAN G,ABDELFATTAH S,WANG Q,et al.Dbe-kt22:A knowledge tracing dataset based on online student evaluation[J].arXiv:2208.12651,2022.
[17]GAO W,LIU Q,YUE L,et al.Collaborative cognitive diagnosis with disentangled representation learning for learner modeling[J].Advances in Neural Information Processing Systems,2024,37:562-588.
[18]HUANG Z,LIU Q,ZHAI C,et al.Exploring multi-objective exercise recommendations in online education systems[C]//Proceedings of the 28th ACM International Conference on Information and Knowledge Management.2019:1261-1270.
[19]PIECH C,BASSEN J,HUANG J,et al.Deep knowledge tracing[C]BAdvances in Neural Information Processing Systems 28.2015.
[20]ZHANG J,SHI X,KING I,et al.Dynamic key-value memorynetworks for knowledge tracing[C]//Proceedings of the 26th International Conference on World Wide Web.2017:765-774.
[21]PANDEY S,KARYPIS G.A self-attentive model for knowledge tracing[J].arXiv:1907.06837,2019.
[1] HU Junjie, CHEN Yujie, HU Yikun, WEN Cheng, CAO Jialun, MA Zhi, SU Jie, SUN Weidi, TIAN Cong, QIN Shengchao. Formal Theorem Proving Empowered by Large Language Model:Survey and Perspectives [J]. Computer Science, 2026, 53(4): 1-23.
[2] XU Cheng, LIU Yuxuan, WANG Xin, ZHANG Cheng, YAO Dengfeng, YUAN Jiazheng. Review of Speech Disorder Assessment Methods Driven by Large Language Models [J]. Computer Science, 2026, 53(3): 307-320.
[3] LI Wenli, FENG Xiaonian, QIAN Tieyun. Few-shot Continuous Toxicity Detection Based on Large Language Model Augmentation [J]. Computer Science, 2026, 53(3): 321-330.
[4] LI Chunying, TANG Zhikang, ZHUANG Zhiwei, LI Wenbo, GUO Yanxi, ZHANG Xiaowei. DCL-FKT:Personalized Knowledge Tracing via Dual Contrastive Learning and ForgettingMechanism [J]. Computer Science, 2026, 53(2): 99-106.
[5] LIU Lilong, LIU Guoming, QI Baoyuan, DENG Xueshan, XUE Dizhan, QIAN Shengsheng. Efficient Inference Techniques of Large Models in Real-world Applications:A Comprehensive Survey [J]. Computer Science, 2026, 53(1): 12-28.
[6] SHAO Xinyi, ZHU Jingwei, ZHANG Liang. LLM-based Business Process Adaptation Method to Respond Long-tailed Changes [J]. Computer Science, 2026, 53(1): 29-38.
[7] LI Maolin, LIN Jiajie, YANG Zhenguo. Confidence-guided Prompt Learning for Multimodal Aspect-level Sentiment Analysis [J]. Computer Science, 2025, 52(7): 241-247.
[8] CHEN Jinyin, XI Changkun, ZHENG Haibin, GAO Ming, ZHANG Tianxin. Survey of Security Research on Multimodal Large Language Models [J]. Computer Science, 2025, 52(7): 315-341.
[9] LI Bo, MO Xian. Application of Large Language Models in Recommendation System [J]. Computer Science, 2025, 52(6A): 240400097-7.
[10] BAI Yuntian, HAO Wenning, JIN Dawei. Study on Open-domain Question Answering Methods Based on Retrieval-augmented Generation [J]. Computer Science, 2025, 52(6A): 240800141-7.
[11] HU Caishun. Study on Named Entity Recognition Algorithms in Audit Domain Based on Large LanguageModels [J]. Computer Science, 2025, 52(6A): 240700190-4.
[12] GAO Hongkui, MA Ruixiang, BAO Qihao, XIA Shaojie, QU Chongxiao. Research on Hybrid Retrieval-augmented Dual-tower Model [J]. Computer Science, 2025, 52(6): 324-329.
[13] LI Hao, YANG Yumeng, ZHAO Boyang, ZHENG Puqi, LIN Hongfei. Adverse Drug Reaction Relationship Extraction Based on Chain of Thought Enhancement UnderHigh and Low Resources [J]. Computer Science, 2025, 52(12): 224-230.
[14] HUANG Haixin, XU Chenglong, FU Yao. Research on Structured Pruning Algorithm Based on Information Fusion [J]. Computer Science, 2025, 52(11A): 241000041-6.
[15] GUO Liwei, WU Yonghao, LIU Yong. Semantic Variations Based Defect Generation and Prediction Model Testing [J]. Computer Science, 2025, 52(11A): 241200059-7.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!