Computer Science ›› 2021, Vol. 48 ›› Issue (12): 331-336.doi: 10.11896/jsjkx.210500028

• Artificial Intelligence • Previous Articles     Next Articles

Abstractive Automatic Summarizing Model for Legal Judgment Documents

ZHOU Wei1, WANG Zhao-yu1, WEI Bin2   

  1. 1 School of Information Management for Law,China University of Political Science and Law,Beijing 102249,China
    2 Institute of Digital Jurisprudence,Zhejiang University,Hangzhou 310008,China
  • Received:2021-05-06 Revised:2021-07-15 Online:2021-12-15 Published:2021-11-26
  • About author:ZHOU Wei,born in 1985,assistant professor,Ph.D.His main research in-terests include legal service and judicial management technology,and legal information management.
    WEI Bin,born in 1986,professor of Hundred Talents Program,Ph.D supervisor,is a member of China Computer Federation.His main research interests include AI & Law,knowledge representation and legal logic.
  • Supported by:
    Research and Innovation Project of CUPL(21ZFQ82005),Key R & D Program of Zhejiang Province (2020C01060),Key R & D Projects of the Ministry of Science and Technology(2018YFC0831800),Key Project of National Social Science Foundation(20&ZD047) and Fundamental Research Funds for the Central Universities.

Abstract: At present,the automatic summarization model for Chinese content applied to legal judgement documents mainly adopts the extraction method.However,due to the lengthiness and low level of structure of legal texts,the accuracy and reliability of extraction method is insufficient for practical application.In order to obtain high quality summaries of legal judgment documents,in this paper,we propose an abstractive automatic summarization model based on multi-model fusion.Based on Seq2Seq model,we apply attention mechanism and selective gates to better process the data input.Specifically,we combine Bert pre-trai-ning and reinforcement learning policy to optimize our model.The corpus we built consists of 50 000 legal judgment documents regarding small claims procedure and summary procedure.Evaluations on the corpus demonstrate that the proposed model outperforms all of the baseline model,and the mean ROUGE score is 5.81% higher than that of conventional Seq2seq+Attentionmodel.

Key words: Attention mechanism, Automatic summarization, Judgement documents, Model fusion, Reinforcement lear-ning, Seq2Seq

CLC Number: 

  • TP18
[1]FU Y L.The Functions and Style of Civil Judicial Decisions [J].Social Sciences in China,2000(4):123-133.
[2]Supreme People's Court.Provisions of the Supreme People's Court on the Issuance of Judgments on the Internet by the People's Courts (2016 Revision)[EB/OL].(2016-10-1)[2021-04-25].https://www.pkulaw.com/en_law/e9ea61f2aaa98dfabdfb.html?flag=chinese/.
[3]Supreme People's Court.China Judgement Online[EB/OL]. (2021-04-25)[2021-04-25].https://wenshu.court.gov.cn/.
[4]HOU S L,ZHANG S H,FEI C Q.A Survey to Text Summarization:Popular Datasets and Methods[J].Journal of Chinese Information Processing,2019,33(5):1-16.
[5]LI Q F.Research on the Method of Multi-document Summarization Based on Topic Model[D].Dalian:Dalian Maritime University,2013.
[6]LI F,HUANG J Z,LI Z J,et al.Automatic Summarization Method of News Texts Using Keywords Expansion[J].Journal of Frontiers of Computer Science and Technology,2016,10(3):372-380.
[7]CHEN Y,BANSAL M.Fast abstractive summarization with reinforce-selected sentence rewriting[J]. arXiv:1805.11080.2018.
[8]LUHN H P.The Automatic Creation of Literature Abstracts [J].IBM Journal of Research and Development,1958,2(2):159-165.
[9]EDMUNDSON H P,WYLLYS R E.Automatic abstracting and indexing - survey and recommendations[J].Communications of the ACM,1961,4(5):226-234.
[10]EDMUNDSON H P.New methods in automatic extracting[J].Journal of the ACM (JACM),1969,16(2):264-85.
[11]WANG Y C,XU H M.OA Automatic Abstracting System on Chinese Documents[J].Journal of the China Society for Scienti-fic and Technical Information,1997(2):49-53.
[12]XU Y D,XU Z M,WANG X L,et al.Multi-Document Automa- tic Summarization Technique Based on Information Fusion[J].Chinese Journal of Computers,2007,30(11):2048-2054.
[13]RUSH A M,CHOPRA S,WESTON J.A Neural Attention Model for Abstractive Sentence Summarization[C]//Procee-dings of the 2015 Conference on Empirical Methods in Natural Language Processing.2015:379-389.
[14]HU B,CHEN Q,ZHU F.LCSTS:A Large Scale Chinese Short Text Summarization Dataset[C]//Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing.2015:1967-1972.
[15]YU L.Automatic Chinese Text Summarization Method Based on Convolutional Neural Network[D].Harbin:Harbin Institute of Technology,2017.
[16]ZHOU C D,ZENG B Q,WANG S Y,et al.Chinese Summarization Research on Combination of Local Attention and Convolutional Neural Network[J].Computer Engineering and Applications,2019,55(8):132-137.
[17]MOENS M F,UYTTENDAELE C.Automatic text structuring and categorization as a first step in summarizing legal cases[J].Information Processing & Management,1997,33(6):727-737.
[18]FARZINDAR A,LAPALME G.LetSum:an automatic Legal Text Summarizing system[C]//Legal Knowledge and Information Systems:Jurix 2004,the Seventeenth Annual Conference.2004:11-18.
[19]ZHONG L,ZHONG Z,ZHAO Z,et al.Automatic Summarization of Legal Decisions using Iterative Masking of Predictive Sentences[C]//Proceedings of the Seventeenth International Conference on Artificial Intelligence and Law.2019:163-172.
[20]DEVLIN J,CHANG M W,LEE K,et al.Bert:Pre-training of deep bidirectional transformers for language understanding [J].arXiv:1810.04805,2018.
[21]VASWANI A,SHAZEER N,PARMAR N,et al.Attention is all you need [C]//Advances in Neural Information Processing Systems.Cambridge,2017:5998-6008.
[22]WISEMAN S,RUSH A M.Sequence-to-sequence learning as beam-search optimization[J].arXiv:1606.02960.2016.
[23]ZHANG S,ZHAO T J,YAO C,et al.Research on Sentence Optimum Selection Algorithm for Multi-Document Summarization[J].Journal of Electronics & Information Technology,2008,30(12):2921-2925.
[24]Supreme People's Court.Notice by the Supreme People's Court of Issuing the Formats of Litigation Documents Related to the Pilot Program of the Reform of Separation between Complicated Cases and Simple Ones under Civil Procedure [EB/OL].(2020-09-30)[2021-04-15].https://www.pkulaw.com/chl/cafe4ca0b1059c4fbdfb.html.
[25]FENG D J,YANG L,YAN J F.Research on Automatic Text Summarization Based on Dual-Encoder Structure[J].Computer Engineering,2020,46(6):60-64.
[26]MA S,SUN X,XU J,et al.Improving Semantic Relevance for Sequence-to-Sequence Learning of Chinese Social Media Text Summarization[C]//Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics(Volume 2:Short Papers).2017:635-640.
[1] RAO Zhi-shuang, JIA Zhen, ZHANG Fan, LI Tian-rui. Key-Value Relational Memory Networks for Question Answering over Knowledge Graph [J]. Computer Science, 2022, 49(9): 202-207.
[2] ZHOU Fang-quan, CHENG Wei-qing. Sequence Recommendation Based on Global Enhanced Graph Neural Network [J]. Computer Science, 2022, 49(9): 55-63.
[3] DAI Yu, XU Lin-feng. Cross-image Text Reading Method Based on Text Line Matching [J]. Computer Science, 2022, 49(9): 139-145.
[4] ZHOU Le-yuan, ZHANG Jian-hua, YUAN Tian-tian, CHEN Sheng-yong. Sequence-to-Sequence Chinese Continuous Sign Language Recognition and Translation with Multi- layer Attention Mechanism Fusion [J]. Computer Science, 2022, 49(9): 155-161.
[5] XIONG Li-qin, CAO Lei, LAI Jun, CHEN Xi-liang. Overview of Multi-agent Deep Reinforcement Learning Based on Value Factorization [J]. Computer Science, 2022, 49(9): 172-182.
[6] JIANG Meng-han, LI Shao-mei, ZHENG Hong-hao, ZHANG Jian-peng. Rumor Detection Model Based on Improved Position Embedding [J]. Computer Science, 2022, 49(8): 330-335.
[7] WANG Ming, PENG Jian, HUANG Fei-hu. Multi-time Scale Spatial-Temporal Graph Neural Network for Traffic Flow Prediction [J]. Computer Science, 2022, 49(8): 40-48.
[8] ZHU Cheng-zhang, HUANG Jia-er, XIAO Ya-long, WANG Han, ZOU Bei-ji. Deep Hash Retrieval Algorithm for Medical Images Based on Attention Mechanism [J]. Computer Science, 2022, 49(8): 113-119.
[9] SUN Qi, JI Gen-lin, ZHANG Jie. Non-local Attention Based Generative Adversarial Network for Video Abnormal Event Detection [J]. Computer Science, 2022, 49(8): 172-177.
[10] YAN Jia-dan, JIA Cai-yan. Text Classification Method Based on Information Fusion of Dual-graph Neural Network [J]. Computer Science, 2022, 49(8): 230-236.
[11] JIN Fang-yan, WANG Xiu-li. Implicit Causality Extraction of Financial Events Integrating RACNN and BiLSTM [J]. Computer Science, 2022, 49(7): 179-186.
[12] XIONG Luo-geng, ZHENG Shang, ZOU Hai-tao, YU Hua-long, GAO Shang. Software Self-admitted Technical Debt Identification with Bidirectional Gate Recurrent Unit and Attention Mechanism [J]. Computer Science, 2022, 49(7): 212-219.
[13] PENG Shuang, WU Jiang-jiang, CHEN Hao, DU Chun, LI Jun. Satellite Onboard Observation Task Planning Based on Attention Neural Network [J]. Computer Science, 2022, 49(7): 242-247.
[14] ZHANG Ying-tao, ZHANG Jie, ZHANG Rui, ZHANG Wen-qiang. Photorealistic Style Transfer Guided by Global Information [J]. Computer Science, 2022, 49(7): 100-105.
[15] ZENG Zhi-xian, CAO Jian-jun, WENG Nian-feng, JIANG Guo-quan, XU Bin. Fine-grained Semantic Association Video-Text Cross-modal Entity Resolution Based on Attention Mechanism [J]. Computer Science, 2022, 49(7): 106-112.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!