|
|
Automatic Abstracting Civil Judgment Documents with Two-Stage Procedure |
Wang Yizhen,Ou Shiyan(),Chen Jinju |
School of Information Management, Nanjing University, Nanjing 210023, China |
|
|
Abstract [Objective] This paper tries to automatically summarize the contents of civil judgment documents in the first-instance, aiming to provide concise, readable, coherent, accurate and efficient knowledge services. [Methods] We proposed an automatic abstracting method for judgment documents, which includes extractive summary stage and abstract summary stage. We first added the expanded residual gate convolution to the pre-training model to extract key sentences from the judgment documents. Then, we input the extractive summary to the sequence to sequence model and generated the final judgment document abstracts. [Results] The ROUGE indicators of the proposed model were 50.31, 36.60, and 48.86 with the experimental data sets of judgment documents, which were 25.00, 23.25, 24.66 higher than the results of the benchmark model (LEAD-3). [Limitations] The extractive summary obtained in the first stage is used as the input of the second stage abstract model, which creates cumulative error issue. The overall performance of the proposed model is decided by the extractive model of the first stage. [Conclusions] The proposed model could summarize judgment texts automatically, which solve the information overload issue and help users quickly read judgment documents.
|
Received: 11 November 2020
Published: 27 May 2021
|
|
Fund:The work is supported by the National Social Science Foundation of China(17ATQ001) |
Corresponding Authors:
Ou Shiyan
E-mail: oushiyan@nju.edu.cn
|
[1] |
Edmundson H P. New Methods in Automatic Extracting[J]. Journal of the ACM, 1969,16(2):264-285.
doi: 10.1145/321510.321519
|
[2] |
Liu M, Yu Y, Qi Q, et al. Extractive Single Document Summarization via Multi-feature Combination and Sentence Compression[C]// Proceedings of the 6th CCF International Conference on Natural Language Processing and Chinese Computing. Springer, Cham, 2017: 807-817.
|
[3] |
Kupiec J, Pedersen J, Chen F. A Trainable Document Summarizer[C]// Proceedings of the 18th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Seattle. ACM, 1995: 68-73.
|
[4] |
Conroy J, O’leary D. Text Summarization via Hidden Markov Models[C]// Proceedings of the 24th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, New Orleans, Louisiana, USA. Association for Computing Machinery, 2001: 406-407.
|
[5] |
Alguliyev R M, Aliguliyev R M, Isazade N R, et al. COSUM: Text Summarization Based on Clustering and Optimization[J]. Expert Systems, 2019,36(1):e12340.
doi: 10.1111/exsy.v36.1
|
[6] |
Osborne T J, Nielsen M A. Entanglement in a Simple Quantum Phase Transition[J]. Physical Review A, 2002,66(3):0321103.
|
[7] |
Svore K, Vanderwende L, Burges C. Enhancing Single-document Summarization by Combining RankNet and Third-party Sources[C]// Proceedings of the 2007 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning (EMNLP-CoNLL), Prague, Czech Republic. Association for Computational Linguistics, 2007: 448-457.
|
[8] |
Liu L, Lu Y, Yang M, et al. Generative Adversarial Network for Abstractive Text Summarization[C]//Proceedings of the 2018 AAAI Conference on Artificial Intelligence, New Orleans, Louisiana, USA. AAAI Press, 2018: 8109-8110.
|
[9] |
Al-Sabahi K, Zhang Z, Nadher M. A Hierarchical Structured Self-Attentive Model for Extractive Document Summarization (HSSAS)[J]. IEEE Access, 2018,6:24205-24212.
doi: 10.1109/ACCESS.2018.2829199
|
[10] |
Cho K, van Merrienboer B, Gulcehre C, et al. Learning Phrase Representations Using RNN Encoder-Decoder for Statistical Machine Translation[OL]. arXiv Preprint, arXiv: 1406. 1078.
|
[11] |
Tan J, Wan X, Xiao J. Abstractive Document Summarization with a Graph-based Attentional Neural Model[C]// Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, Vancouver, Canada. Association for Computational Linguistics, 2017: 1171-1181.
|
[12] |
Siddiqui T, Shamsi J A. Generating Abstractive Summaries Using Sequence to Sequence Attention[C]// Proceedings of the 2018 International Conference on Frontiers of Information Technology (FIT), Islamabad, Pakistan. IEEE Computer Society, 2018: 212-217.
|
[13] |
Celikyilmaz A, Bosselut A, He X D, et al. Deep Communicating Agents for Abstractive Summarization[C]// Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, New Orleans. Association for Computational Linguistics, 2018,1:1662-1675.
|
[14] |
江跃华, 丁磊, 李娇娥, 等. 融合词汇特征的生成式摘要模型[J]. 河北科技大学学报, 2019,40(2):152-158.
|
[14] |
( Jiang Yuehua, Ding Lei, Li Jiao’e, et al. Abstractive Summarization Model Considering Hybrid Lexical Features[J]. Journal of Hebei University of Science and Technology, 2019,40(2):152-158.)
|
[15] |
Hachey B, Grover C. Automatic Legal Text Summarization: Experiments with Summary Structuring[C]// Proceedings of the 10th International Conference on Artificial Intelligence and Law, Bologna, Italy. ACM, 2005: 75-84.
|
[16] |
Anand D, Wagh R. Effective Deep Learning Approaches for Summarization of Legal Texts[J]. Journal of King Saud University-Computer and Information Sciences, 2019.
|
[17] |
Devlin J, Chang M W, Lee K, et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding[C]// Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis. Association for Computational Linguistics, 2019: 4171-4186.
|
[18] |
Li Y, Zhang X, Chen D. CSRNet: Dilated Convolutional Neural Networks for Understanding the Highly Congested Scenes[C]// Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake. IEEE, 2018: 1091-1100.
|
[19] |
Dong L, Yang N, Wang W, et al. Unified Language Model Pre-training for Natural Language Understanding and Generation[J]. Advances in Neural Information Processing Systems, 2019,32:13063-13075.
|
[20] |
Gu J, Lu Z, Li H, et al. Incorporating Copying Mechanism in Sequence-to-Sequence Learning[C]// Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, Berlin. ACL, 2016: 1631-1640.
|
[21] |
Lin C. Rouge: A Package for Automatic Evaluation of Summaries[C]// Proceedings of the 2004 Workshop on Text Summarization Branches Out, Spain. 2004: 74-81.
|
[22] |
Zhou Q, Yang N, Wei F, et al. Neural Document Summarization by Jointly Learning to Score and Select Sentences[C]// Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, Melbourne. ACL, 2018: 654-663.
|
[23] |
See A, Liu P J, Manning C D. Get to the Point: Summarization with Pointer-generator Networks[C]// Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, Vancouver, Canada. Association for Computational Linguistics, 2017: 1073-1083.
|
[24] |
Gehrmann S, Deng Y T, Rush A. Bottom-up Abstractive Summarization[C]// Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. ACL, 2018: 4098-4109.
|
|
Viewed |
|
|
|
Full text
|
|
|
|
|
Abstract
|
|
|
|
|
Cited |
|
|
|
|
|
Shared |
|
|
|
|
|
Discussed |
|
|
|
|