计算机科学 ›› 2026, Vol. 53 ›› Issue (4): 347-355.doi: 10.11896/jsjkx.250800012

• 人工智能 • 上一篇    下一篇

Agent4Stu:基于大语言模型的学生作答行为高效模拟智能体

刘苏熠, 刘淇, 高维博   

  1. 中国科学技术大学人工智能与数据科学学院 合肥 230027
  • 收稿日期:2025-08-04 修回日期:2025-11-07 出版日期:2026-04-15 发布日期:2026-04-08
  • 通讯作者: 高维博(weibogao@mail.ustc.edu.cn)
  • 作者简介:(syliulsy@163.com)
  • 基金资助:
    国家重点研发计划(2024YFC3308200);国家自然科学基金(62337001);中央高校基本科研业务费专项资金

Agent4Stu:Efficient LLM-based Student Answer Behavior Simulation Agent

LIU Suyi, LIU Qi, GAO Weibo   

  1. School of Artificial Intelligence and Data Science, University of Science and Technology of China, Hefei 230027, China
  • Received:2025-08-04 Revised:2025-11-07 Published:2026-04-15 Online:2026-04-08
  • About author:LIU Suyi,born in 2003,postgraduate.His main research interests include intelligent education and generative agents.
    GAO Weibo,born in 1997,Ph.D.His main research interests include data mining and generative agents.
  • Supported by:
    National Key Research and Development Program of China(2024YFC3308200),National Natural Science Foundation of China(62337001) and Fundamental Research Funds for the Central Universities.

摘要: 个性化学习已成为教育数字化转型的核心方向,其关键在于对学生作答数据进行精准理解与建模。然而,受限于学习行为的稀缺性、学习状态的动态演化以及隐私合规等因素,真实教育场景中普遍存在行为数据不足及离线数据与在线学习行为存在分布偏移的问题,严重制约了智能教育系统的建模能力与泛化水平。为缓解上述困境,已有研究尝试通过模拟学生作答行为来扩展数据规模并提升模型性能,然而现有方法难以同时兼顾数据的生成质量、效率与资源成本。为此,提出一种融合大语言模型与检索增强生成技术的学生作答行为模拟智能体框架Agent4Stu,实现低成本、高效率、强泛化能力的个性化作答行为生成。其核心包括预置检索库及检索策略,以及基于大语言模型的智能体。其中,以学生作答行为构建检索库,并设计了两类检索策略,即相似学生协同检索与相关事实检索,同时结合学生个体短期记忆动态生成高相关性的提示信息。智能体内部融合画像、记忆与动作3个核心模块,分别用于刻画学生的学习特征与认知能力,整合历史经验与检索库知识,并基于画像、记忆和知识掌握程度模拟学生针对具体题目的作答行为。与现有学生智能体相比,Agent4Stu 的记忆容量更小,动作推理更简化,依托面向学习行为的结构化检索库提供辅助信息,从而实现了低成本、高效率、强泛化能力的个性化作答行为生成。基于两个真实作答数据集开展的定量与定性实验,验证了 Agent4Stu 在学生作答行为模拟方面的有效性和优越性。

关键词: 数据合成, 大语言模型, 检索增强生成, 智能体, 人类行为模拟, 冷启动, 智能教育

Abstract: Personalized learning has become a core focus of the digital transformation in education,with its success hinging on the precise understanding and modeling of student response data.However,real-world educational scenarios often face challenges such as sparse learning behaviors,dynamic evolution of learning states,and privacy compliance constraints.Additionally,discrepancies between offline data and online learning behaviors result in insufficient behavioral data and distribution shifts,significantly limiting the modeling capabilities and generalization performance of intelligent education systems.To alleviate these dilemmas,previous studies have attempted to simulate student response behaviors to expand data scale and improve model performance. However,existing methods struggle to simultaneously balance generation quality,efficiency,and resource costs.To overcome these limitations,this paper proposes Agent4Stu,a student response behavior simulation framework that integrates large language models(LLMs) with retrieval-augmented generation(RAG) techniques,enabling low-cost,efficient,and highly generalizable personalized response generation.The framework comprises a pre-built retrieval database with retrieval strategies and an LLM-based agent.The retrieval database is constructed from student response behaviors,and two retrieval strategies are designed:similar-student collaborative retrieval and relevant-fact retrieval.These are combined with each student’s short-term memory to dynamically generate highly relevant prompts.Internally,the agent integrates three core modules,profile,memory,and action,which are responsible for modeling students’ learning characteristics and cognitive abilities,integrating historical experiences with know-ledge from the retrieval database,and simulating students’ responses on specific items based on their profile,memory,and know-ledge mastery.Compared with existing LLM-based student agents,Agent4Stu has a smaller memory footprint and simplified action reasoning,while leveraging a behavior-oriented structured retrieval database to provide auxiliary information.This design enables low-cost,efficient,and highly generalizable personalized response generation.Quantitative and qualitative experiments on two real-world learning datasets demonstrate the effectiveness and superiority of Agent4Stu in simulating student learning response behaviors.

Key words: Data synthesis, Large language models, Retrieval-augmented generation, Agents, Human behavior simulation, Cold-start, Intelligent education

中图分类号: 

  • TP391.1
[1]HUANG H R,LIU J H,PAN J W,et al.Systematic transformation of education for the intelligent era:comprehensive education reform empowered by digitalization[J].E-education Research,2025,46(4):5-12.
[2]FENG M,HEFFERNAN N,KOEDINGER K.Addressing theassessment challenge with an online system that tutors as it assesses[J].User Modeling and User-Adapted Interaction,2009,19:243-266.
[3]NAKAGAWA H,IWASAWA Y,MATSUO Y.Graph-basedknowledge tracing:modeling student proficiency using graph neural network[C]//IEEE/WIC/ACM International Confe-rence on Web Intelligence.2019:156-163.
[4]GAO W,LIU Q,YUE L,et al.Agent4Edu:Generating Learner Response Data by Generative Agents for Intelligent Education Systems[J].arXiv:2501.10332,2025.
[5]XU S,ZHANG X,QIN L.Eduagent:Generative student agentsin learning[J].arXiv:2404.07963,2024.
[6]REDDY S,LEVINE S,DRAGAN A.Accelerating h-umanlearning with deep reinforcement learning[C]//NIPS Workshop:Teaching Machines,Robots,and Humans.2017:5-9.
[7]SU Y,LIU Q,LIU Q,et al.Exercise-enhanced seque-ntial modeling for student performance prediction[C]//Proceedings of the AAAI Conference on Artificial Intelligence.2018.
[8]LIU Q,TONG S,LIU C,et al.Exploiting cognitivestructure for adaptive learning[C]//Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mi-ning.2019:627-635.
[9]ZHAO G,HUANG Z,ZHUANG Y,et al.Simulating student interactions with two-stage imitation lea-rning for intelligent educational systems[C]//Proceeding of the 32nd ACM International Conference on Information and Knowledge Management.2023:3423-3432.
[10]PARK J S,O’BRIEN J,CAI C J,et al.Generative agents:Interactive simulacra of human behavior[C]//Proceedings of the 36th Annual Acm Symposium on User Interface Software and Technology.2023:1-22.
[11]QIAN C,CONG X,YANG C,et al.Communicativeagents forsoftware development[J].arXiv:2307.07924,2023.
[12]LIU J,HUANG Z,XIAO T,et al.SocraticLM:Exploring socratic personalized teaching with large language models[J].Advances in Neural Information Processing Systems,2024,37:85693-85721.
[13]RAM O,LEVINE Y,DALMEDIGOS I,et al.In-context retrieval-augmented language models[J].Transactions of the Association for Computational Linguistics,2023,11:1316-1331.
[14]FENG Z,FENG X,ZHAO D,et al.Retrieval-generation synergy augmented large language models[C]//ICASSP 2024-2024 IEEE International Conference on Acoustics,Speech and Signal Processing(ICASSP).IEEE,2024:11661-11665.
[15]BAKER F B.The basics of item response theory[M/OL].http://ericae.net/irt/baker.
[16]ABDELRAHMAN G,ABDELFATTAH S,WANG Q,et al.Dbe-kt22:A knowledge tracing dataset based on online student evaluation[J].arXiv:2208.12651,2022.
[17]GAO W,LIU Q,YUE L,et al.Collaborative cognitive diagnosis with disentangled representation learning for learner modeling[J].Advances in Neural Information Processing Systems,2024,37:562-588.
[18]HUANG Z,LIU Q,ZHAI C,et al.Exploring multi-objective exercise recommendations in online education systems[C]//Proceedings of the 28th ACM International Conference on Information and Knowledge Management.2019:1261-1270.
[19]PIECH C,BASSEN J,HUANG J,et al.Deep knowledge tracing[C]BAdvances in Neural Information Processing Systems 28.2015.
[20]ZHANG J,SHI X,KING I,et al.Dynamic key-value memorynetworks for knowledge tracing[C]//Proceedings of the 26th International Conference on World Wide Web.2017:765-774.
[21]PANDEY S,KARYPIS G.A self-attentive model for knowledge tracing[J].arXiv:1907.06837,2019.
[1] 胡俊杰, 陈宇杰, 胡义坤, 文成, 曹嘉伦, 马智, 苏杰, 孙纬地, 田聪, 秦胜潮.
大模型驱动的形式化定理证明:综述与展望
Formal Theorem Proving Empowered by Large Language Model:Survey and Perspectives
计算机科学, 2026, 53(4): 1-23. https://doi.org/10.11896/jsjkx.251000067
[2] 吴荞睿, 罗丽, 赵才荣.
基于循环一致性约束的LLM增强型语言模型训练框架
LLM-augmented Training Framework with Cycle-Consistency Constraints
计算机科学, 2026, 53(4): 377-383. https://doi.org/10.11896/jsjkx.250600032
[3] 王智彬, 李世鹏, 周宇航, 李雪, 张中辉, 蒋智威, 顾荣, 田臣, 陈贵海, 仲盛.
大语言模型服务系统服务级目标和系统级指标优化
Optimization of Service Level Objectives and System Level Metrics in Large Language ModelServing System
计算机科学, 2026, 53(3): 23-32. https://doi.org/10.11896/jsjkx.250900173
[4] 周悦媛, 卢冠泽, 向佳为, 章家维, 邵恩, 何鑫.
面向海光DCU基于自适应转置的大语言模型训练系统
Training System for Large Language Models Based on Adaptive Transpose on Hygon DCU
计算机科学, 2026, 53(3): 33-40. https://doi.org/10.11896/jsjkx.250600073
[5] 陈涵, 徐泽锋, 蒋究, 樊凡, 章军建, 何楚, 王文伟.
基于大语言模型和深度网络的认知评估量表自动诊断
Large Language Model and Deep Network Based Cognitive Assessment Automatic Diagnosis
计算机科学, 2026, 53(3): 41-51. https://doi.org/10.11896/jsjkx.250600034
[6] 徐嘉雯, 郑云贵, 周伟, 徐尧强, 胡卉芪, 周烜.
SQL-MARS:面向用户模糊需求的 Text2SQL 结构化数据推荐系统
SQL-MARS:Text-to-SQL Structured Data Recommendation System for Ambiguous UserRequirements
计算机科学, 2026, 53(3): 52-63. https://doi.org/10.11896/jsjkx.250700096
[7] 吴贤杰, 李彤亮, 李舟军.
表格问答研究综述
Survey of Table Question Answering Research
计算机科学, 2026, 53(3): 295-306. https://doi.org/10.11896/jsjkx.250900006
[8] 徐成, 刘宇轩, 王欣, 张铖, 姚登峰, 袁家政.
大语言模型驱动的言语障碍评估方法综述
Review of Speech Disorder Assessment Methods Driven by Large Language Models
计算机科学, 2026, 53(3): 307-320. https://doi.org/10.11896/jsjkx.250300125
[9] 翟洁, 李艳豪, 陈乐旋, 郭卫斌.
基于轻量级教育大模型的个性化实践学习资料动态推荐
Dynamic Recommendation of Personalized Hands-on Learning Materials Based on LightweightEducational LLMs
计算机科学, 2026, 53(2): 48-56. https://doi.org/10.11896/jsjkx.250800002
[10] 谢光强, 邱枫阳, 李杨.
基于拓扑虚拟结构洞节点的分布式多智能体快速共识寻求
Fast Consensus Seeking in Distributed Multi-agent System Using Topology Virtual Structural Hole Node
计算机科学, 2026, 53(2): 358-366. https://doi.org/10.11896/jsjkx.241200109
[11] 陈昱胤, 李贯峰, 秦晶, 肖毓航.
知识图谱的复杂逻辑查询方法研究综述
Survey on Complex Logical Query Methods in Knowledge Graphs
计算机科学, 2026, 53(2): 273-288. https://doi.org/10.11896/jsjkx.250400033
[12] 段鹏婷, 温超, 王保平, 王珍妮.
基于协作语义融合的多智能体行为决策方法
Collaborative Semantics Fusion for Multi-agent Behavior Decision-making
计算机科学, 2026, 53(1): 252-261. https://doi.org/10.11896/jsjkx.250300145
[13] 郭陆祥, 王越余, 李芊玥, 李莎莎, 刘晓东, 纪斌, 余杰.
大语言模型智能体操作系统研究综述
Comprehensive Survey of LLM-based Agent Operating Systems
计算机科学, 2026, 53(1): 1-11. https://doi.org/10.11896/jsjkx.250500002
[14] 刘利龙, 刘国明, 齐保元, 邓雪杉, 薛迪展, 钱胜胜.
实际应用场景中的大模型高效推理技术综述
Efficient Inference Techniques of Large Models in Real-world Applications:A Comprehensive Survey
计算机科学, 2026, 53(1): 12-28. https://doi.org/10.11896/jsjkx.250300030
[15] 邵欣怡, 朱经纬, 张亮.
基于大语言模型的业务流程长尾变化应变方法
LLM-based Business Process Adaptation Method to Respond Long-tailed Changes
计算机科学, 2026, 53(1): 29-38. https://doi.org/10.11896/jsjkx.250100001
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!