Skip to main content

2021 | OriginalPaper | Buchkapitel

Extractive Summarization of Chinese Judgment Documents via Sentence Embedding and Memory Network

verfasst von : Yan Gao, Zhengtao Liu, Juan Li, Jin Tang

Erschienen in: Natural Language Processing and Chinese Computing

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

A rapidly rising number of open judgment documents has increased the requirement for automatic summarization. Since Chinese judgment documents are characterized by a lengthy and logical structure, extractive summarization is an effective method for them. However, existing extractive models generally cannot capture information between sentences. In order to enable the model to obtain long-term information in the judgment documents, this paper proposes an extractive model using sentence embeddings and a two-layers memory network. A pre-trained language model is used to encode sentences in judgment documents. Then the whitening operation is applied to get isotropic sentence embeddings, which makes the subsequent classification more accurate. These embeddings are fed into a unidirectional memory network to fuse previous sentence embeddings. A bidirectional memory network is followed to introduce position information of sentences. The experimental results show that our proposed model outperforms the baseline methods on the SFZY dataset from CAIL2020.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Polsley, S., Jhunjhunwala, P., Huang, R.: CaseSummarizer: a system for automated summarization of legal texts. In: Proceedings of COLING 2016, the 26th international conference on Computational Linguistics: System Demonstrations, pp. 258–262 (2016) Polsley, S., Jhunjhunwala, P., Huang, R.: CaseSummarizer: a system for automated summarization of legal texts. In: Proceedings of COLING 2016, the 26th international conference on Computational Linguistics: System Demonstrations, pp. 258–262 (2016)
2.
Zurück zum Zitat Liu, C.L., Chen, K.C.: Extracting the gist of Chinese judgments of the supreme court. In: Proceedings of the Seventeenth International Conference on Artificial Intelligence and Law, pp. 73–82 (2019) Liu, C.L., Chen, K.C.: Extracting the gist of Chinese judgments of the supreme court. In: Proceedings of the Seventeenth International Conference on Artificial Intelligence and Law, pp. 73–82 (2019)
3.
Zurück zum Zitat See, A., Liu, P.J., Manning, C.D.: Get to the point: summarization with pointer-generator networks. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 1073–1083 (2017) See, A., Liu, P.J., Manning, C.D.: Get to the point: summarization with pointer-generator networks. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 1073–1083 (2017)
4.
Zurück zum Zitat Mihalcea, R., Tarau, P.: TextRank: bringing order into text. In: Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing, pp. 404–411 (2004) Mihalcea, R., Tarau, P.: TextRank: bringing order into text. In: Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing, pp. 404–411 (2004)
5.
Zurück zum Zitat Liu, H., Yu, H., Deng, Z.H.: Multi-document summarization based on two-level sparse representation model. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 29 (2015) Liu, H., Yu, H., Deng, Z.H.: Multi-document summarization based on two-level sparse representation model. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 29 (2015)
6.
Zurück zum Zitat Li, P., Bing, L., Lam, W., Li, H., Liao, Y.: Reader-aware multi-document summarization via sparse coding. In: Proceedings of the 24th International Conference on Artificial Intelligence, pp. 1270–1276 (2015) Li, P., Bing, L., Lam, W., Li, H., Liao, Y.: Reader-aware multi-document summarization via sparse coding. In: Proceedings of the 24th International Conference on Artificial Intelligence, pp. 1270–1276 (2015)
7.
Zurück zum Zitat Fevry, T., Phang, J.: Unsupervised sentence compression using denoising auto-encoders. In: Proceedings of the 22nd Conference on Computational Natural Language Learning, pp. 413–422 (2018) Fevry, T., Phang, J.: Unsupervised sentence compression using denoising auto-encoders. In: Proceedings of the 22nd Conference on Computational Natural Language Learning, pp. 413–422 (2018)
8.
Zurück zum Zitat Zheng, H., Lapata, M.: Sentence centrality revisited for unsupervised summarization. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 6236–6247 (2019) Zheng, H., Lapata, M.: Sentence centrality revisited for unsupervised summarization. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 6236–6247 (2019)
9.
Zurück zum Zitat Cao, Z., Wei, F., Dong, L., Li, S., Zhou, M.: Ranking with recursive neural networks and its application to multi-document summarization. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 29 (2015) Cao, Z., Wei, F., Dong, L., Li, S., Zhou, M.: Ranking with recursive neural networks and its application to multi-document summarization. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 29 (2015)
10.
Zurück zum Zitat Cheng, J., Lapata, M.: Neural summarization by extracting sentences and words. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 484–494 (2016) Cheng, J., Lapata, M.: Neural summarization by extracting sentences and words. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 484–494 (2016)
11.
Zurück zum Zitat Nallapati, R., Zhai, F., Zhou, B.: SummaRuNNer: a recurrent neural network based sequence model for extractive summarization of documents. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 31 (2017) Nallapati, R., Zhai, F., Zhou, B.: SummaRuNNer: a recurrent neural network based sequence model for extractive summarization of documents. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 31 (2017)
13.
Zurück zum Zitat Bouscarrat, L., Bonnefoy, A., Peel, T., Pereira, C.: STRASS: a light and effective method for extractive summarization based on sentence embeddings. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics: Student Research Workshop, pp. 243–252 (2019) Bouscarrat, L., Bonnefoy, A., Peel, T., Pereira, C.: STRASS: a light and effective method for extractive summarization based on sentence embeddings. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics: Student Research Workshop, pp. 243–252 (2019)
14.
Zurück zum Zitat Yuan, R., Wang, Z., Li, W.: Fact-level extractive summarization with hierarchical graph mask on BERT. In: Proceedings of the 28th International Conference on Computational Linguistics, pp. 5629–5639 (2020) Yuan, R., Wang, Z., Li, W.: Fact-level extractive summarization with hierarchical graph mask on BERT. In: Proceedings of the 28th International Conference on Computational Linguistics, pp. 5629–5639 (2020)
15.
Zurück zum Zitat Zhou, Q., Wei, F., Zhou, M.: At which level should we extract? An empirical analysis on extractive document summarization. In: Proceedings of the 28th International Conference on Computational Linguistics, pp. 5617–5628 (2020) Zhou, Q., Wei, F., Zhou, M.: At which level should we extract? An empirical analysis on extractive document summarization. In: Proceedings of the 28th International Conference on Computational Linguistics, pp. 5617–5628 (2020)
16.
Zurück zum Zitat Rissland, E.L., Ashley, K.D., Loui, R.P.: AI and law: a fruitful synergy. Artif. Intell. 150(1–2), 1–15 (2003)CrossRef Rissland, E.L., Ashley, K.D., Loui, R.P.: AI and law: a fruitful synergy. Artif. Intell. 150(1–2), 1–15 (2003)CrossRef
17.
Zurück zum Zitat Bench-Capon, T., et al.: A history of AI and law in 50 papers: 25 years of the international conference on AI and law. Artif. Intell. Law 20(3), 215–319 (2012)CrossRef Bench-Capon, T., et al.: A history of AI and law in 50 papers: 25 years of the international conference on AI and law. Artif. Intell. Law 20(3), 215–319 (2012)CrossRef
18.
Zurück zum Zitat Surden, H.: Artificial intelligence and law: an overview. Ga. St. UL Rev. 35, 1305 (2018) Surden, H.: Artificial intelligence and law: an overview. Ga. St. UL Rev. 35, 1305 (2018)
19.
Zurück zum Zitat Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)CrossRef Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)CrossRef
20.
Zurück zum Zitat Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp. 4171–4186 (2019) Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp. 4171–4186 (2019)
21.
Zurück zum Zitat Vaswani, A., et al.: Attention is all you need. In: Proceedings of the 31st International Conference on Neural Information Processing Systems, pp. 6000–6010 (2017) Vaswani, A., et al.: Attention is all you need. In: Proceedings of the 31st International Conference on Neural Information Processing Systems, pp. 6000–6010 (2017)
22.
Zurück zum Zitat Reimers, N., et al.: Sentence-BERT: sentence embeddings using Siamese BERT-networks. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics (2019) Reimers, N., et al.: Sentence-BERT: sentence embeddings using Siamese BERT-networks. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics (2019)
24.
Zurück zum Zitat Su, J., Cao, J., Liu, W., Ou, Y.: Whitening sentence representations for better semantics and faster retrieval. CoRR arXiv:2103.15316 (2021) Su, J., Cao, J., Liu, W., Ou, Y.: Whitening sentence representations for better semantics and faster retrieval. CoRR arXiv:​2103.​15316 (2021)
26.
Zurück zum Zitat Lin, C.Y.: ROUGE: a package for automatic evaluation of summaries. In: Text Summarization Branches Out, pp. 74–81 (2004) Lin, C.Y.: ROUGE: a package for automatic evaluation of summaries. In: Text Summarization Branches Out, pp. 74–81 (2004)
27.
Zurück zum Zitat Joulin, A., Grave, É., Bojanowski, P., Mikolov, T.: Bag of tricks for efficient text classification. In: Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 2, Short Papers, pp. 427–431 (2017) Joulin, A., Grave, É., Bojanowski, P., Mikolov, T.: Bag of tricks for efficient text classification. In: Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 2, Short Papers, pp. 427–431 (2017)
28.
Zurück zum Zitat Kim, Y.: Convolutional neural networks for sentence classification (2014) Kim, Y.: Convolutional neural networks for sentence classification (2014)
29.
Zurück zum Zitat Liu, P., Qiu, X., Huang, X.: Recurrent neural network for text classification with multi-task learning. arXiv preprint arXiv:1605.05101 (2016) Liu, P., Qiu, X., Huang, X.: Recurrent neural network for text classification with multi-task learning. arXiv preprint arXiv:​1605.​05101 (2016)
Metadaten
Titel
Extractive Summarization of Chinese Judgment Documents via Sentence Embedding and Memory Network
verfasst von
Yan Gao
Zhengtao Liu
Juan Li
Jin Tang
Copyright-Jahr
2021
DOI
https://doi.org/10.1007/978-3-030-88480-2_33