计算机科学 ›› 2026, Vol. 53 ›› Issue (4): 377-383.doi: 10.11896/jsjkx.250600032

• 人工智能 • 上一篇    下一篇

基于循环一致性约束的LLM增强型语言模型训练框架

吴荞睿1, 罗丽2, 赵才荣1   

  1. 1 同济大学计算机科学与技术学院 上海 201804
    2 国网湖南省电力有限公司长沙供电分公司 长沙 410000
  • 收稿日期:2025-06-06 修回日期:2025-07-19 出版日期:2026-04-15 发布日期:2026-04-08
  • 通讯作者: 赵才荣(zhaocairong@tongji.edu.cn)
  • 作者简介:(2331919@tongji.edu.cn)
  • 基金资助:
    国家自然科学基金(62076184,62473286);上海市自然科学基金面上项目(22ZR1466700)

LLM-augmented Training Framework with Cycle-Consistency Constraints

WU Qiaorui1, LUO Li2, ZHAO Cairong1   

  1. 1 School of Computer Science and Technology, Tongji University, Shanghai 201804, China
    2 State Grid Hunan Electric Power Co., Ltd.Changsha Power Supply Branch, Changsha 410000, China
  • Received:2025-06-06 Revised:2025-07-19 Published:2026-04-15 Online:2026-04-08
  • About author:WU Qiaorui,born in 2001,postgra-duate.His main research interests include large language models and model self-optimization.
    ZHAO Cairong,born in 1981,Ph.D,professor,Ph.D supervisor,is a member of CCF(No.24551D).His main research interests include artificial intelligence and computer vision.
  • Supported by:
    National Natural Science Fundation of China(62076184,62473286) and Shanghai Natural Science Foundation(22ZR1466700).

摘要: 针对专利摘要生成任务中的技术特征完整性缺失、合规性不足以及边缘部署效率受限等问题,提出一种基于循环一致性约束的大语言模型增强型语言模型训练框架(LACC)。LACC框架通过构建摘要生成与权利要求扩写的双向可逆任务结构,引入循环一致性约束机制,实现技术特征与法律表述的联合优化。在此基础上,LACC集成了由大语言模型驱动的可控数据增强策略,自动构建高质量专利样本对,并结合动态验证机制,提升了生成文本的技术准确性与合规稳定性。实验结果表明,LACC在中文专利数据集CPTD上的ROUGE-L得分达到56.74,较基线提升8.99个百分点,在循环一致性评分(RCS)方面亦取得显著优势。此外,该方法在边缘设备上的推理延迟控制在420 ms以内,单卡内存占用不超过4.5 GB,具备良好的工业部署潜力。综上所述,LACC框架为专利撰写辅助、法律文本生成与知识产权智能管理等下游任务提供了具备实用价值的技术路径,展现出推动知识产权全生命周期自动化处理的重要应用前景。

关键词: 自然语言处理, 大语言模型, 循环一致性, 数据增强, 知识蒸馏, 协同训练

Abstract: This paper proposes a training framework termed LACC(Large Language Model-Augmented Consistency-Constrained),designed to address key challenges in patent abstract generation,including incomplete coverage of technical features,insufficient legal compliance,and inefficiency in edge deployment.The LACC framework constructs a bidirectional reversible task structure between abstract generation and claim expansion,incorporating a cycle-consistency constraint to jointly optimize technical expression and legal formulation.On this basis,LACC integrates a controllable data augmentation strategy powered by large language models(LLMs) to automatically generate high-quality patent text pairs.A dynamic verification mechanism is further introduced to enhance the technical accuracy and regulatory reliability of generated content.Experimental results on the Chinese patent dataset CPTD demonstrate that LACC achieves a ROUGE-L score of 56.74,outperforming the baseline by 8.99 percentage points,and shows significant improvements in the recurrence consistency score(RCS).Moreover,the framework supports efficient edge deployment,with inference latency controlled within 420 ms and single-GPU memory usage limited to 4.5 GB.Overall,LACC offers a practical and scalable solution for downstream tasks such as patent drafting assistance,legal text generation,and intelligent intellectual property(IP) management,and shows strong potential in enabling the automation of the full lifecycle of IP processing.

Key words: Natural language processing, Large language model, Cycle consistency, Data augmentation, Knowledge distillation, Collaborative training

中图分类号: 

  • TP391
[1]SEMERIKOV S O,VAKALIUK T A,KANEVSKA O B,et al.LLM on the edge:the new frontier[C]//Proceedings of the 5th Edge Computing Workshop(doors 2025),2025:137-161.
[2]FRIHA O,FERRAG M A,KANTARCI B,et al.LLM-basededge intelligence:A comprehensive survey on architectures,applications,security and trustworthiness[J].IEEE Open Journal of the Communications Society,2024,5:5799-5856.
[3]LASKAR M T R,ALQAHTANI S,BARI M S,et al.A systematic survey and critical review on evaluating large language mo-dels:Challenges,limitations,and recommendations[C]//Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing.2024:13785-13816.
[4]ZHU J Y,PARK T,ISOLA P,et al.Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks[C]//2017 IEEE International Conference on Computer Vision(ICCV).IEEE,2017:2242-2251.
[5]KIM B,KIM J,LEE J G,et al.Unsupervised Deformable Image Registration Using Cycle-Consistent CNN[C]//Medical Image Computing and Computer Assisted Intervention-MICCAI 2019.2019.
[6]SHEN T X,LEI T,BARZILAY R,et al.Style Transfer fromNon-Parallel Text by Cross-Alignment[C]//31st Conference on Neural Information Processing Systems(NIPS 2017).2017.
[7]LIAN C,LI X,KONG L,et al.CoCycleReg:Collaborative Cycle-Consistency Method for Multi-Modal Medical Image Re-gistration[J].Neurocomputing,2022,500:799-808.
[8]LI C Y,LI X M,KONG L K,et al.CoCycleReg:Collaborative Cycle-Consistency Method for Multi-Modal Medical Image Re-gistration[J].Neurocomputing,2022,500:799-808.
[9]MA C,ZHANG W E,GUO M Y,et al.Multi-document Summarization Via Deep Learning Techniques:A Survey[J].ACM Computing Surveys,2023,55(5):1-37.
[10]WANG Z,WANG P,LIU K,et al.A comprehensive survey on data augmentation[J].arXiv:2405.09591,2024.
[11]ZHOU Y,GUO C,WANG X,et al.A survey on data augmentation in large model era[J].arXiv:2401.15422,2024.
[12]SCHICK T,SCHÜTZE H.Few-Shot Text Generation with Pattern-Exploiting Training[J].arXiv:2012.11926,2020.
[13]WEI J,ZOU K.EDA:Easy Data Augmentation Techniques for Boosting Performance on Text Classification Tasks[C]//Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Confe-rence on Natural Language Processing.2019:6382-6388.
[14]DING B,QIN C,ZHAO R,et al.Data augmentation usingLLMs:Data perspectives,learning paradigms and challenges[C]//Findings of the Association for Computational Linguistics ACL 2024.2024:1679-1705.
[15]ZELIKMAN E,WU Y H,MU J,et al.STaR:BootstrappingReasoning with Reasoning[C]//NEURIPS 2022.2022.
[16]LOUIS D,VAN H,JAAP S,et al.Robust Deformable ImageRegistration Using Cycle-Consistent Implicit Representations[J].IEEE Transactions on Medical Imaging,2024,43(2):784-793.
[17]RAMAKANTH P,ASLI C,MICHEL G,et al.Data Augmentation for Abstractive Query-Focused Multi-Document Summarization[J].Computing Research Repository,2021,35(15):13666-13674.
[18]GOU J,YU B,MAYBANK S J,et al.Knowledge distillation:A survey[J].International Journal of Computer Vision,2021,129(6):1789-1819.
[19]GEOFFREY H,ORIOL V,JEFF D.Distilling the Knowledge in a Neural Network[J].arXiv:1503.02531,2015.
[20]KUMAR S,ALESSANDRO S,MRINMAYA S.Distilling Reasoning Capabilities into Smaller Language Models[J].arXiv:2212.00193,2022.
[21]LIU Z Y,SOUMYA S,ISABELLE L,et al.Self-Contradictory Reasoning Evaluation and Detection[J].Association for Computational Linguistics,2024(11):3725-3742.
[22]XU X,LI M,TAO C,et al.A survey on knowledge distillation of large language models[J].arXiv:2402.13116,2024.
[23]ZHANG Y,KANG B,HOOI B,et al.Deep long-tailed learning:A survey[J].IEEE Transactions on Pattern Analysis and Machine Intelligence,2023,45(9):10795-10816.
[1] 宋建华, 刘淳, 张龑.
基于结构化知识蒸馏的轻量级伪装目标检测模型
Lightweight Camouflaged Object Detection Model Based on Structured Knowledge Distillation
计算机科学, 2026, 53(4): 299-307. https://doi.org/10.11896/jsjkx.250100105
[2] 胡俊杰, 陈宇杰, 胡义坤, 文成, 曹嘉伦, 马智, 苏杰, 孙纬地, 田聪, 秦胜潮.
大模型驱动的形式化定理证明:综述与展望
Formal Theorem Proving Empowered by Large Language Model:Survey and Perspectives
计算机科学, 2026, 53(4): 1-23. https://doi.org/10.11896/jsjkx.251000067
[3] 刘苏熠, 刘淇, 高维博.
Agent4Stu:基于大语言模型的学生作答行为高效模拟智能体
Agent4Stu:Efficient LLM-based Student Answer Behavior Simulation Agent
计算机科学, 2026, 53(4): 347-355. https://doi.org/10.11896/jsjkx.250800012
[4] 潘嘉豪, 冯翔, 虞慧群.
基于多任务强化学习的优先级加权软模块化方法:SM-PHT
SM-PHT:Robust,Scalable,and Efficient Method for Multi-task Reinforcement Learning
计算机科学, 2026, 53(4): 366-376. https://doi.org/10.11896/jsjkx.250700198
[5] 王智彬, 李世鹏, 周宇航, 李雪, 张中辉, 蒋智威, 顾荣, 田臣, 陈贵海, 仲盛.
大语言模型服务系统服务级目标和系统级指标优化
Optimization of Service Level Objectives and System Level Metrics in Large Language ModelServing System
计算机科学, 2026, 53(3): 23-32. https://doi.org/10.11896/jsjkx.250900173
[6] 周悦媛, 卢冠泽, 向佳为, 章家维, 邵恩, 何鑫.
面向海光DCU基于自适应转置的大语言模型训练系统
Training System for Large Language Models Based on Adaptive Transpose on Hygon DCU
计算机科学, 2026, 53(3): 33-40. https://doi.org/10.11896/jsjkx.250600073
[7] 陈涵, 徐泽锋, 蒋究, 樊凡, 章军建, 何楚, 王文伟.
基于大语言模型和深度网络的认知评估量表自动诊断
Large Language Model and Deep Network Based Cognitive Assessment Automatic Diagnosis
计算机科学, 2026, 53(3): 41-51. https://doi.org/10.11896/jsjkx.250600034
[8] 钱清, 陈辉程, 崔允贺, 唐瑞雪, 付金玫.
多尺度聚合协同轴向语义引导的实体关系联合抽取方法
Joint Entity and Relation Extraction Method with Multi-scale Collaborative Aggregation and Axial-semantic Guidance
计算机科学, 2026, 53(3): 97-106. https://doi.org/10.11896/jsjkx.250500095
[9] 葛泽庆, 黄圣君.
针对多标记表格数据的半监督学习方法
Semi-supervised Learning Method for Multi-label Tabular Data
计算机科学, 2026, 53(3): 151-157. https://doi.org/10.11896/jsjkx.250600149
[10] 吴贤杰, 李彤亮, 李舟军.
表格问答研究综述
Survey of Table Question Answering Research
计算机科学, 2026, 53(3): 295-306. https://doi.org/10.11896/jsjkx.250900006
[11] 徐成, 刘宇轩, 王欣, 张铖, 姚登峰, 袁家政.
大语言模型驱动的言语障碍评估方法综述
Review of Speech Disorder Assessment Methods Driven by Large Language Models
计算机科学, 2026, 53(3): 307-320. https://doi.org/10.11896/jsjkx.250300125
[12] 陈昱胤, 李贯峰, 秦晶, 肖毓航.
知识图谱的复杂逻辑查询方法研究综述
Survey on Complex Logical Query Methods in Knowledge Graphs
计算机科学, 2026, 53(2): 273-288. https://doi.org/10.11896/jsjkx.250400033
[13] 孙明旭, 梁刚, 吴逸飞, 胡海馨.
融合仇恨对象特征与变体词还原机制的中文仇恨言论检测
Chinese Hate Speech Detection Incorporating Hate Object Features and Variant Word Restoration Mechanism
计算机科学, 2026, 53(2): 289-299. https://doi.org/10.11896/jsjkx.241200004
[14] 郭陆祥, 王越余, 李芊玥, 李莎莎, 刘晓东, 纪斌, 余杰.
大语言模型智能体操作系统研究综述
Comprehensive Survey of LLM-based Agent Operating Systems
计算机科学, 2026, 53(1): 1-11. https://doi.org/10.11896/jsjkx.250500002
[15] 刘利龙, 刘国明, 齐保元, 邓雪杉, 薛迪展, 钱胜胜.
实际应用场景中的大模型高效推理技术综述
Efficient Inference Techniques of Large Models in Real-world Applications:A Comprehensive Survey
计算机科学, 2026, 53(1): 12-28. https://doi.org/10.11896/jsjkx.250300030
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!