Abstract
AbstractAnalyzing and evaluating legal case reports are labor-intensive tasks for judges and lawyers, who usually base their decisions on report abstracts, legal principles, and commonsense reasoning. Thus, summarizing legal documents is time-consuming and requires excellent human expertise. Moreover, public legal corpora of specific languages are almost unavailable. This paper proposes a transfer learning approach with extractive and abstractive techniques to cope with the lack of labeled legal summarization datasets, namely a low-resource scenario. In particular, we conducted extensive multi- and cross-language experiments. The proposed work outperforms the state-of-the-art results of extractive summarization on the Australian Legal Case Reports dataset and sets a new baseline for abstractive summarization. Finally, syntactic and semantic metrics assessments have been carried out to evaluate the accuracy and the factual consistency of the machine-generated legal summaries.
Funder
Alma Mater Studiorum - Università di Bologna
Publisher
Springer Science and Business Media LLC
Subject
Law,Artificial Intelligence
Reference77 articles.
1. Akbik A, Bergmann T, Blythe D, Rasul K, Schweter S, Vollgraf R (2019) FLAIR: an easy-to-use framework for state-of-the-art NLP. In: Ammar W, Louis A, Mostafazadeh N (eds) Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: human language technologies, NAACL-HLT 2019, Minneapolis, MN, USA, June 2–7, 2019, Demonstrations. Association for Computational Linguistics, pp 54–59. https://doi.org/10.18653/v1/n19-4010
2. Bae S, Kim T, Kim J, Lee S (2019) Summary level training of sentence rewriting for abstractive summarization. arXiv:1909.08752
3. Bahdanau D, Cho K, Bengio Y (2015) Neural machine translation by jointly learning to align and translate. In: Bengio Y, LeCun Y (eds) 3rd International conference on learning representations, ICLR 2015, San Diego, CA, USA, May 7–9, 2015, Conference track proceedings. http://arxiv.org/abs/1409.0473
4. Bajaj A, Dangati P, Krishna K, Kumar PA, Uppaal R, Windsor B, Brenner E, Dotterrer D, Das R, McCallum (2021) A Long document summarization in a low resource setting using pretrained language models. In: Kabbara J, Lin H, Paullada A, Vamvas J (eds) Proceedings of the ACL-IJCNLP 2021 student research workshop. ACL, pp 71–80. https://doi.org/10.18653/v1/2021.acl-srw.7
5. Beltagy I, Peters ME, Cohan A (2020) Longformer: the long-document transformer. arXiv:2004.05150
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献