计算机科学 ›› 2026, Vol. 53 ›› Issue (2): 48-56.doi: 10.11896/jsjkx.250800002
翟洁, 李艳豪, 陈乐旋, 郭卫斌
ZHAI Jie, LI Yanhao, CHEN Lexuan, GUO Weibin
摘要: 人工智能技术在教育领域的深度应用,已成为国家教育数字化转型的核心战略。在计算机实践教学领域,实践学习资料的精准推荐是提升学生学习效能与质量的重要途径。针对高校教育规模化与学生需求多元化之间的矛盾,提出一种基于轻量级教育大模型的个性化实践学习资料推荐模型LightPLRec(Lightweight Personalized Learning Recommender for Dynamic Practice Materials),旨在依据学生个体特征的动态变化智能推荐个性化的实践学习资料。基于低算力需求的轻量级大模型,通过指令微调和强化学习方法构建了面向个性化实践学习资料推荐的教育大模型SPIR(Student Profile & Interest-based Re-commender)。通过整合多源异构数据,深度融入课程知识体系、学科前沿动态、产业发展趋势、国家战略导向,构建了跨学科、多模态的实践学习资料库,并设计了图转主题文本方法gragh2topic。依托于SPIR大模型的强大赋能和多源资料库的坚实支撑,提出了基于智能工作流的资料推荐方法。设计主题分析方法从学生能力评估结果中提取学生的能力特征,应用图卷积网络算法GCN从学生学习行为数据中挖掘学生的兴趣特征,创建了“能力-推荐智能体”和“兴趣-推荐智能体”,构建了双智能体协同驱动的智能化流程体系,实现了从学生个性化画像智能生成到实践学习资料动态推荐的系列工作流任务;并且构建了个性化资料推荐数据集,在该数据集上验证了所提模型的性能显著优于基线模型。其中,以Qwen2.5-3.0B为基模型训练的LightPLRec模型,在能力推荐与兴趣推荐这两项任务中展现出卓越性能,准确率分别高达0.947和0.939,其表现均优于DeepSeek-V3在同一数据集上的测评结果。该研究为教育大模型的垂直场景应用提供了技术范式,同时通过创建个性化实践学习资料动态推荐模型,为践行“因材施教”理念和培育高素质计算机实践人才提供了创新路径。
中图分类号:
| [1]MURALIDHARAN S,TURUVEKERE SREENIVAS S,JOSHIR,et al.Compact language models via pruning and knowledge distillation[J].Advances in Neural Information Processing Systems,2024,37:41076-41102. [2]GRATTAFIORI A,DUBEY A,JAUHRI A,et al.The llama 3 herd of models[J].arXiv:2407.21783,2024. [3]GUNASEKAR S,ZHANG Y,ANEJA J,et al.Textbooks are all you need[J].arXiv:2306.11644,2023. [4]YANG A,YANG B,ZHANG B,et al.Qwen2.5 technical report[J].arXiv:2412.15115,2024. [5]HU S,TU Y,HAN X,et al.Minicpm:Unveiling the potential ofsmall language models with scalable training strategies[J].ar-Xiv:2404.06395,2024. [6]SHENG Y Q,ZENG W X,TANG J Y,et al.Confusing negative commonsense knowledge generation with hierarchy modeling and LLM-enhanced filtering[J].Information Processing & Ma-nagement,2025,62(3):104060. [7]XIANG J,TAO T,GU Y,et al.Language models meet world models:Embodied experiences enhance language models[J].Advances in Neural Information Processing Systems,2023,36:75392-75412. [8]YAO S,ZHAO J,YU D,et al.React:Synergizing reasoning and acting in language models[C]//International Conference on Learning Representations(ICLR).2023. [9]WANG Z,CAI S,CHEN G,et al.Describe,explain,plan and se-lect:Interactive planning with large language models enables open-world multi-task agents[J].arXiv:2302.01560,2023. [10]XIANG J,LIU G,GU Y,et al.Pandora:Towards general world model with natural language actions and video states[J].arXiv:2406.09455,2024. [11]WU Q,BANSAL G,ZHANG J,et al.Autogen:Enabling next-gen LLM applications via multi-agent conversations[C]//First Conference on Language Modeling.2024. [12]HABLER I,HUANG K,NARAJALA V S,et al.Building a se-cure agentic AI application leveraging A2A protocol[J].arXiv:2504.16902,2025. [13]ZENG G,CHEN X,HU J,et al.Routine:A Structural Planning Framework for LLM Agent System in Enterprise[J].arXiv:2507.14447,2025. [14]CHEN C,WU Y,DAI Q,et al.A survey on graph neural networks and graph transformers in computer vision:A task-oriented perspective[J].IEEE Transactions on Pattern Analysis and Machine Intelligence,2024(12):46. [15]HITZLER P,SARKER M.Neuro-symbolic AI=neural+logical+probabilistic AI[J].Neuro-Symbolic Artificial Intelligence:The State of the Art,2022,342:173. [16]YU X,LIU Z,FANG Y,et al.Learning to count isomorphisms with graph neural networks[C]//Proceedings of the AAAI Conference on Artificial Intelligence.2023:4845-4853. [17]TAN F,ZHANG C,LIU L.DyAtGNN:Dynamic AttentionGraph Neural Networks for dynamic graph[J].Knowledge-Based Systems,2025,325:113935. [18]BUTEREZ D,JANET J P,OGLIC D,et al.An end-to-end attention-based approach for learning on graphs[J].Nature Communications,2025,16(1):5244. [19]WU Y,HUANG H,SONG Y,et al.Soft-GNN:towards robust graph neural networks via self-adaptive data utilization[J].Frontiers of Computer Science,2025,19(4):194311. [20]ZHANG T,LIU Y,SHEN Z,et al.Learning from heterogeneity:A dynamic learning framework for hypergraphs[J].IEEE Transactions on Artificial Intelligence,2025,6(6):1513-1528. [21]XIE P Z,LI G J,LI T.Knowledge Tracing Model Based onExercise-Knowledge Point Heterogeneous Graph and Multi-feature Fusion[J].Computer Science,2025,52(3):197-205. [22]CHENG K,PENG L,WANG P,et al.DyGKT:Dynamic Graph Learning for Knowledge Tracing[C]//KDD’24:Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining.2024. [23]ZHAI J,LI Y H,MENG T X.Exploration and practice of personalized computer laboratory teaching based on decision trees and large models[J].Experimental Technology and Management,2023,40(12):8-15. [24]LI Q Y,XIA W,YIN L A,et al.Privileged Knowledge StateDistillation for Reinforcement Learning-based Educational Path Recommendation[C]//KDD’24:The 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining.2024. [25]Education Dialogue Dataset|教育对话数据集|对话生成数据集[DB/OL](2024-10-29)[2025-08-23].https://www.selectdataset.com/dataset/f436a1b97fdc3c9cd38ed8294694b42d. |
|
||