Computer Science ›› 2026, Vol. 53 ›› Issue (4): 356-365.doi: 10.11896/jsjkx.251100003

• Artificial Intelligence • Previous Articles     Next Articles

Cross-model Collaborative Unsupervised Representation Method for Legal Texts

XU Shenjian   

  1. Digital Society Governance Institute, China University of Political Science and Law, Beijing 100088, China
  • Received:2025-11-03 Revised:2026-01-16 Online:2026-04-15 Published:2026-04-08
  • About author:XU Shenjian,born in 1966,Ph.D,professor.His main research interests include digital society governance legal ethics,procedural law,experiential legal education,judicial system and legal writing.

Abstract: Legal text representation is a fundamental component of legal artificial intelligence systems,directly affecting the performance of downstream tasks such as legal article prediction and case retrieval.However,the professional terminology,complex structure,and reasoning patterns of legal texts often lead to semantic drift in general pre-trained models.Open-source models lack sufficient legal domain knowledge,while closed-source models,despite their strong semantic understanding capabilities,provide representations that are difficult to directly access and reuse.To address these challenges,this paper proposes a cross-model collaborative legal representation framework(CMCLR),which enables collaborative learning between open-source and closed-source models to enhance legal semantic modeling.Specifically,closed-source models are employed to perform dynamic text segmentation and key paragraph identification,producing structured domain-aware signals that guide the fine-tuning of open-source models under collaborative constraints.In addition,unsupervised clustering is introduced to model structural relationships among paragraph-level embeddings,capturing latent semantic associations between legal texts.Experiments conducted on the CAIL2018 legal article classification task demonstrate that CMCLR achieves an accuracy of 90.3%,outperforming representative baseline methods by 2.4 percentage points,while maintaining robust performance across different dataset scales and settings.These results confirm the effectiveness of cross-model collaborative representation learning for deep semantic modeling of legal texts.

Key words: Legal text, Representation, Textual relevance, Legal artificial intelligence, Pretrained models, Cross-model collaborative legal representation(CMCLR)

CLC Number: 

  • TP391
[1]ZHOU W,WANG Z,WEI B.A Generative Model for Automatic Summarization of Legal Judgment Documents[J].ComputerScience,2021,48(12):331-336.
[2]ZHANG H,WANG X,WANG C,et al.A Method for Legal Statute Recommendation on Judgment Documents[J].Computer Science,2019,46(9).
[3]ACHEAMPONG F A,NUNOO-MENSAH H,CHEN W.Trans-former models for text-based emotion detection:a review of BERT-based approaches[J].Artificial Intelligence Review,2021,54(8):5789-5829.
[4]YENDURI G,RAMALINGAM M,SELVI G C,et al.Gpt(generative pre-trained transformer)-a comprehensive review on enabling technologies,potential applications,emerging challenges,and future directions[J].IEEE Access,2024,12:54608-54649.
[5]WANG Z,DING Y,WU C,et al.Causality-inspired legal provision selection with large language model-based explanation[J/OL].Artificial Intelligence and Law,2024:1-25.https://doi.org/10.1007/s10506-024-09429-3
[6]HUANG T,XIE X,LIU X.Multi-level Correlation Matching for Legal Text Similarity Modeling with Multiple Examples[C]//International Conference on Web Information Systems Engineering.Singapore:Springer,2023:621-632.
[7]CHALKIDIS I,FERGADIOTIS M,MALAKASIOTIS P,et al.LEGAL-BERT:The muppets straight out of law school[J].arXiv:2010.02559,2020.
[8]NAVEED H,KHAN A U,QIU S,et al.A comprehensive overview of large language models[J].arXiv:2307.06435,2023.
[9]ACHIAM J,ADLER S,AGARWAL S,et al.Gpt-4 technical report[J].arXiv:2303.08774,2023.
[10]TOUVRON H,LAVRIL T,IZACARD G,et al.Llama:Openand efficient foundation language models[J].arXiv:2302.13971,2023.
[11]YAN L.A Study on the Correlation of Attributive Position and Length in Legal Texts:Taking the Amendment to Criminal Law(XI) as an Example[J].International Journal of Frontiers in Sociology,2023,5(15):120-128.
[12]NALLAPATI R,MANNING C D.Legal docket classification:where machine learning stumbles[C]//Proceedings of the 2008 Conference on Empirical Methods in Natural Language Proces-sing.2008:438-446.
[13]KAUFMAN A R,KRAFT P,SEN M.Improving supreme court forecasting using boosted decision trees[J].Political Analysis,2019,27(3):381-387.
[14]KIM M Y,XU Y,GOEBEL R.Legal question answering using ranking svm and syntactic/semantic similarity[C]//JSAI International Symposium on Artificial Intelligence.Berlin:Springer,2014:244-258.
[15]KAUFMAN A R,KRAFT P,SEN M.Improving supreme court forecasting using boosted decision trees[J].Political Analysis,2019,27(3):381-387.
[16]DEVLIN J,CHANG M W,LEE K,et al.Bert:Pre-training ofdeep bidirectional transformers for language understanding[C]//Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies.2019:4171-4186.
[17]VASWANI A,SHAZEER N,PARMAR N,et al.Attention isall you need[C]//Advances in Neural Information Processing Systems.2017.
[18]CHALKIDIS I,FERGADIOTIS M,MALAKASIOTIS P,et al.LEGAL-BERT:The Muppets straight out of Law School[C]//Findings of the Association for Computational Linguistics:EMNLP 2020.2020.
[19]PAUL S,MANDAL A,GOYAL P,et al.Pre-training trans-formers on indian legal text[J].arXiv:2209.06049,2022.
[20]CHALKIDIS I,DAI X,FERGADIOTIS M,et al.An exploration of hierarchical attention transformers for efficient long document classification[J].arXiv:2210.05529,2022.
[21]PRASAD N,BOUGHANEM M,DKAKI T.Effect of hierarchical domain-specific language models and attention in the classification of decisions for legal cases[C]//CIRCLE(Joint Confe-rence of the Information Retrieval Communities in Europe).2022.
[22]ZHAO J S,SONG M X,GAO X,et al.Research on text representation in natural language processing[J].Journal of Software,2022,33(1):102-128.
[23]HUANG R,XU J.Text classification based on invariant graph convolutional neural networks[J].Computer Science,2024,51(S1):230900018-5.
[24]WEI R M,CHEN R Y,LI H,et al.Technology trend analysis based on deep learning and textometric methods[J].Computer Science,2022,49(S2):211100119-6.
[25]XU Y M,SHI L Y,CAI L Q.A cross-lingual text sentimentanalysis model based on sentiment feature representation[J].Journal of Chinese Information Processing,2022,36(2):129-141.
[26]WU X,JIANG B,ZHONG Y,et al.Multi-target Markov boun-dary discovery:Theory,algorithm,and application[J].IEEE Transactions on Pattern Analysis and Machine Intelligence,2022,45(4):4964-4980.
[27]MCINNES L,HEALY J,SAUL N,et al.UMAP:Uniform Ma-nifold Approximation and Projection[J].Journal of Open Source Software,2018,3(29):861.
[28]ESTER M,KRIEGEL H P,SANDER J,et al.A Density-Based Algorithm for Discovering Clusters in Large Spatial Databases with Noise[C]//Second International Conference on Knowledge Discovery and Data Mining(KDD’96).1996:226-331.
[29]ŁUKASIK S,KOWALSKI P A,CHARYTANOWICZ M,et al.Clustering using flower pollination algorithm and Calinski-Harabasz index[C]//2016 IEEE Congress on Evolutionary Computation(CEC).IEEE,2016:2724-2728.
[30]XIAO C,ZHONG H,GUO Z,et al.Cail2018:A large-scale legal dataset for judgment prediction[J].arXiv:1807.02478,2018.
[31]HOCHREITER S,SCHMIDHUBER J.Long short-term me-mory[J].Neural Computation,1997,9(8):1735-1780.
[32]JACOVI A,SHALOM O S,GOLDBERG Y.UnderstandingConvolutional Neural Networks for Text Classification[C]//Proceedings of the 2018 EMNLP Workshop BlackboxNLP:Ana-lyzing and Interpreting Neural Networks for NLP.2018:56-65.
[33]YANG W,JIA W,ZHOU X,et al.Legal judgment prediction via multi-perspective bi-feedback network[C]//Proceedings of the 28th International Joint Conference on Artificial Intelligence.2019:4085-4091.
[34]ZHENG M,LIU B,SUN L.LawRec:automatic recommendation of legal provisions based on legal text analysis[J].Computatio-nal Intelligence and Neuroscience,2022,2022(1):6313161.
[35]FENG Y,LI C,NG V.Legal judgment prediction via event ex-traction with constraints[C]//Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics.2022:648-664.
[1] HUANG Beibei, LIU Jinfeng. Causal Disentangled Representation Learning with Integrated Sparse Coding [J]. Computer Science, 2026, 53(4): 66-77.
[2] LIU Yichen, LIN Yan, ZHOU Zeyu, GUO Shengnan, LIN Youfang, WAN Huaiyu. Efficient Semantic-aware Trajectory Representation Learning Method via State Space Model [J]. Computer Science, 2026, 53(4): 134-142.
[3] ZHANG Xueqin, WANG Zhineng, LI Jinsheng, LU Yisong, LUO Fei. Key Node Identification in Temporal Social Networks Based on Deep Learning and Multi-feature Fusion [J]. Computer Science, 2026, 53(4): 143-154.
[4] YU Chengcheng, JIANG Yongfa, CHEN Fangshu, WANG Jiahui, MENG Xiankai. Multi-view Exercise Representation and Forgetting Mechanism for Deep KnowledgeTracing [J]. Computer Science, 2026, 53(3): 107-114.
[5] WANG Yiming, JIAO Min, ZHAO Suyun, CHEN Hong, LI Cuiping. Prompt-conditioned Representation Learning with Diffusion Models for Semi-supervised Clustering [J]. Computer Science, 2026, 53(3): 158-165.
[6] ZHANG Jing, PAN Jinghao, JIANG Wenchao. Background Structure-aware Few-shot Knowledge Graph Completion [J]. Computer Science, 2026, 53(2): 331-341.
[7] HUANG Miaomiao, WANG Huiying, WANG Meixia, WANG Yejiang , ZHAO Yuhai. Review of Graph Embedding Learning Research:From Simple Graph to Complex Graph [J]. Computer Science, 2026, 53(1): 58-76.
[8] LIU Hongle, CHEN Juan, FU Cai, HAN Lansheng, GUO Xiaowei, JIANG shuai. Authorship Gender Recognition of Source Code Based on Multiple Mixed Features [J]. Computer Science, 2025, 52(8): 51-61.
[9] YANG Jian, SUN Liu, ZHANG Lifang. Survey on Data Processing and Data Augmentation in Low-resource Language Automatic Speech Recognition [J]. Computer Science, 2025, 52(8): 86-99.
[10] ZHU Rui, YE Yaqin, LI Shengwen, TANG Zijian, XIAO Yue. Dynamic Community Detection with Hierarchical Modularity Optimization [J]. Computer Science, 2025, 52(8): 127-135.
[11] WAN Zhaolin, MA Guangzhe, MI Le, LI Zhiyang, FAN Xiaopeng. Computing 2D Skeleton Using Novel Potential Model [J]. Computer Science, 2025, 52(7): 135-141.
[12] LI Pengyan, WANG Baohui, YE Zihao. Study on Improvements of RippleNet Model Based on Representation Enhancement [J]. Computer Science, 2025, 52(6A): 240800142-9.
[13] LIAO Sirui, HUANG Feihu, ZHAN Pengxiang, PENG Jian, ZHANG Linghao. DCDAD:Differentiated Context Dependency for Time Series Anomaly Detection Method [J]. Computer Science, 2025, 52(6): 106-117.
[14] GUO Xuan, HOU Jinlin, WANG Wenjun, JIAO Pengfei. Dynamic Link Prediction Method for Adaptively Modeling Network Dynamics [J]. Computer Science, 2025, 52(6): 118-128.
[15] TAN Qiyin, YU Jiong, CHEN Zixin. Outlier Detection Method Based on Adaptive Graph Autoencoder [J]. Computer Science, 2025, 52(6): 129-138.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!