Skip to main content
Erschienen in: Neural Computing and Applications 8/2021

23.07.2020 | Original Article

SHEG: summarization and headline generation of news articles using deep learning

verfasst von: Rajeev Kumar Singh, Sonia Khetarpaul, Rohan Gorantla, Sai Giridhar Allada

Erschienen in: Neural Computing and Applications | Ausgabe 8/2021

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

The human attention span is continuously decreasing, and the amount of time a person wants to spend on reading is declining at an alarming rate. Therefore, it is imperative to provide a quick glance of important news by generating a concise summary of the prominent news article, along with the most intuitive headline in line with the summary. When humans produce summaries of documents, they not only extract phrases and concatenate them but also produce new grammatical phrases or sentences that coincide with each other and capture the most significant information of the original article. Humans have an incredible ability to create abstractions; however, automatic summarization is a challenging problem. This paper aims to develop an end-to-end methodology that can generate brief summaries and crisp headlines that can capture the attention of readers and convey a significant amount of relevant information. In this paper, we propose a novel methodology known as SHEG, which is designed as a hybrid model. It works by integrating both extractive and abstractive mechanisms using a pipelined approach to produce a concise summary, which is then used for headline generation. Experiments were performed on publicly available datasets, viz. CNN/Daily Mail, Gigaword, and NEWSROOM. The results obtained validate our approach and demonstrate that the proposed SHEG model is effectively producing a concise summary as well as a captivating and fitting headline.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Ayana SS, Liu Z, Sun M (2016) Neural headline generation with minimum risk training. arXiv preprint arXiv:1604.01904 Ayana SS, Liu Z, Sun M (2016) Neural headline generation with minimum risk training. arXiv preprint arXiv:1604.01904
2.
Zurück zum Zitat Bahdanau D, Brakel P, Xu K, Goyal A, Lowe R, Pineau J, Courville A, Bengio Y (2016) An actor-critic algorithm for sequence prediction. arXiv preprint arXiv:1607.07086 Bahdanau D, Brakel P, Xu K, Goyal A, Lowe R, Pineau J, Courville A, Bengio Y (2016) An actor-critic algorithm for sequence prediction. arXiv preprint arXiv:1607.07086
3.
Zurück zum Zitat Bahdanau D, Cho K, Bengio Y (2014) Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473 Bahdanau D, Cho K, Bengio Y (2014) Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473
4.
Zurück zum Zitat Banko M, Mittal VO, Witbrock MJ (2000) Headline generation based on statistical translation. In: Proceedings of the 38th annual meeting on Association for Computational Linguistics. Association for Computational Linguistics, pp 318–325 Banko M, Mittal VO, Witbrock MJ (2000) Headline generation based on statistical translation. In: Proceedings of the 38th annual meeting on Association for Computational Linguistics. Association for Computational Linguistics, pp 318–325
5.
Zurück zum Zitat Carbonell JG, Goldstein J (1998) The use of MMR, diversity-based reranking for reordering documents and producing summaries. SIGIR 98:335–336 Carbonell JG, Goldstein J (1998) The use of MMR, diversity-based reranking for reordering documents and producing summaries. SIGIR 98:335–336
6.
Zurück zum Zitat Chen YC, Bansal M (2018) Fast abstractive summarization with reinforce-selected sentence rewriting. arXiv preprint arXiv:1805.11080 Chen YC, Bansal M (2018) Fast abstractive summarization with reinforce-selected sentence rewriting. arXiv preprint arXiv:1805.11080
7.
Zurück zum Zitat Cheng J, Lapata M (2016) Neural summarization by extracting sentences and words. arXiv preprint arXiv:1603.07252 Cheng J, Lapata M (2016) Neural summarization by extracting sentences and words. arXiv preprint arXiv:1603.07252
8.
Zurück zum Zitat Cheung JCK, Penn G (2014) Unsupervised sentence enhancement for automatic summarization. In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), pp 775–786 Cheung JCK, Penn G (2014) Unsupervised sentence enhancement for automatic summarization. In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), pp 775–786
9.
Zurück zum Zitat Chopra S, Auli M, Rush AM (2016) Abstractive sentence summarization with attentive recurrent neural networks. In: Proceedings of the 2016 conference of the North American chapter of the Association for Computational Linguistics: human language technologies, pp 93–98 Chopra S, Auli M, Rush AM (2016) Abstractive sentence summarization with attentive recurrent neural networks. In: Proceedings of the 2016 conference of the North American chapter of the Association for Computational Linguistics: human language technologies, pp 93–98
10.
Zurück zum Zitat Chung J, Gulcehre C, Cho K, Bengio Y (2014) Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:1412.3555 Chung J, Gulcehre C, Cho K, Bengio Y (2014) Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:1412.3555
11.
Zurück zum Zitat Clarke J, Lapata M (2010) Discourse constraints for document compression. Comput Linguist 36(3):411–441CrossRef Clarke J, Lapata M (2010) Discourse constraints for document compression. Comput Linguist 36(3):411–441CrossRef
12.
Zurück zum Zitat Cohn T, Lapata M (2008) Sentence compression beyond word deletion. In: Proceedings of the 22nd international conference on computational linguistics, vol 1. Association for Computational Linguistics, pp 137–144 Cohn T, Lapata M (2008) Sentence compression beyond word deletion. In: Proceedings of the 22nd international conference on computational linguistics, vol 1. Association for Computational Linguistics, pp 137–144
13.
Zurück zum Zitat Colmenares CA, Litvak M, Mantrach A, Silvestri F (2015) Heads: headline generation as sequence prediction using an abstract feature-rich space. In: Proceedings of the 2015 conference of the North American chapter of the Association for Computational Linguistics: human language technologies, pp 133–142 Colmenares CA, Litvak M, Mantrach A, Silvestri F (2015) Heads: headline generation as sequence prediction using an abstract feature-rich space. In: Proceedings of the 2015 conference of the North American chapter of the Association for Computational Linguistics: human language technologies, pp 133–142
14.
Zurück zum Zitat Devlin J, Chang MW, Lee K, Toutanova K (2018) Bert: pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 Devlin J, Chang MW, Lee K, Toutanova K (2018) Bert: pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805
15.
Zurück zum Zitat Dorr B, Zajic D, Schwartz R (2003) Hedge trimmer: a parse-and-trim approach to headline generation. In: Proceedings of the HLT-NAACL 03 on text summarization workshop, vol 5. Association for Computational Linguistics, pp 1–8 Dorr B, Zajic D, Schwartz R (2003) Hedge trimmer: a parse-and-trim approach to headline generation. In: Proceedings of the HLT-NAACL 03 on text summarization workshop, vol 5. Association for Computational Linguistics, pp 1–8
16.
Zurück zum Zitat Erkan G, Radev DR (2004) Lexrank: graph-based lexical centrality as salience in text summarization. J Artif Intell Res 22:457–479CrossRef Erkan G, Radev DR (2004) Lexrank: graph-based lexical centrality as salience in text summarization. J Artif Intell Res 22:457–479CrossRef
17.
Zurück zum Zitat Fan A, Grangier D, Auli M (2017) Controllable abstractive summarization. arXiv preprint arXiv:1711.05217 Fan A, Grangier D, Auli M (2017) Controllable abstractive summarization. arXiv preprint arXiv:1711.05217
18.
Zurück zum Zitat Filippova K, Alfonseca E, Colmenares CA, Kaiser L, Vinyals O (2015) Sentence compression by deletion with LSTMS. In: Proceedings of the 2015 conference on empirical methods in natural language processing, pp 360–368 Filippova K, Alfonseca E, Colmenares CA, Kaiser L, Vinyals O (2015) Sentence compression by deletion with LSTMS. In: Proceedings of the 2015 conference on empirical methods in natural language processing, pp 360–368
19.
Zurück zum Zitat Ganesan K, Zhai C, Han J (2010) Opinosis: a graph based approach to abstractive summarization of highly redundant opinions. In: Proceedings of the 23rd international conference on computational linguistics (Coling 2010), pp 340–348 Ganesan K, Zhai C, Han J (2010) Opinosis: a graph based approach to abstractive summarization of highly redundant opinions. In: Proceedings of the 23rd international conference on computational linguistics (Coling 2010), pp 340–348
20.
Zurück zum Zitat Gehrmann S, Deng Y, Rush AM (2018) Bottom-up abstractive summarization. arXiv preprint arXiv:1808.10792 Gehrmann S, Deng Y, Rush AM (2018) Bottom-up abstractive summarization. arXiv preprint arXiv:1808.10792
21.
Zurück zum Zitat Grusky M, Naaman M, Artzi Y (2018) Newsroom: a dataset of 1.3 million summaries with diverse extractive strategies. arXiv preprint arXiv:1804.11283 Grusky M, Naaman M, Artzi Y (2018) Newsroom: a dataset of 1.3 million summaries with diverse extractive strategies. arXiv preprint arXiv:1804.11283
22.
Zurück zum Zitat Henß S, Mieskes M, Gurevych I (2015) A reinforcement learning approach for adaptive single-and multi-document summarization. In: GSCL, pp 3–12 Henß S, Mieskes M, Gurevych I (2015) A reinforcement learning approach for adaptive single-and multi-document summarization. In: GSCL, pp 3–12
23.
Zurück zum Zitat Hsu WT, Lin CK, Lee MY, Min K, Tang J, Sun M (2018) A unified model for extractive and abstractive summarization using inconsistency loss. arXiv preprint arXiv:1805.06266 Hsu WT, Lin CK, Lee MY, Min K, Tang J, Sun M (2018) A unified model for extractive and abstractive summarization using inconsistency loss. arXiv preprint arXiv:1805.06266
24.
Zurück zum Zitat Jing H, McKeown KR (2000) Cut and paste based text summarization. In: Proceedings of the 1st North American chapter of the Association for Computational Linguistics conference. Association for Computational Linguistics, pp 178–185 Jing H, McKeown KR (2000) Cut and paste based text summarization. In: Proceedings of the 1st North American chapter of the Association for Computational Linguistics conference. Association for Computational Linguistics, pp 178–185
25.
Zurück zum Zitat Keneshloo Y, Shi T, Ramakrishnan N, Reddy CK (2018) Deep reinforcement learning for sequence to sequence models. arXiv preprint arXiv:1805.09461 Keneshloo Y, Shi T, Ramakrishnan N, Reddy CK (2018) Deep reinforcement learning for sequence to sequence models. arXiv preprint arXiv:1805.09461
26.
Zurück zum Zitat Kim Y (2014) Convolutional neural networks for sentence classification. arXiv preprint arXiv:1408.5882 Kim Y (2014) Convolutional neural networks for sentence classification. arXiv preprint arXiv:1408.5882
27.
Zurück zum Zitat Kingma DP, Ba J (2014) Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 Kingma DP, Ba J (2014) Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980
28.
Zurück zum Zitat Knight K, Marcu D (2000) Statistics-based summarization-step one: sentence compression. AAAI/IAAI 2000:703–710 Knight K, Marcu D (2000) Statistics-based summarization-step one: sentence compression. AAAI/IAAI 2000:703–710
29.
Zurück zum Zitat Lafferty J, McCallum A, Pereira FC (2001) Conditional random fields: probabilistic models for segmenting and labeling sequence data Lafferty J, McCallum A, Pereira FC (2001) Conditional random fields: probabilistic models for segmenting and labeling sequence data
30.
Zurück zum Zitat Li P, Bing L, Lam W (2018) Actor-critic based training framework for abstractive summarization. arXiv preprint arXiv:1803.11070 Li P, Bing L, Lam W (2018) Actor-critic based training framework for abstractive summarization. arXiv preprint arXiv:1803.11070
31.
Zurück zum Zitat Lin CY (2004) Rouge: a package for automatic evaluation of summaries. In: Text summarization branches out, pp 74–81 Lin CY (2004) Rouge: a package for automatic evaluation of summaries. In: Text summarization branches out, pp 74–81
32.
Zurück zum Zitat Lin CY, Hovy E (1997) Identifying topics by position. In: Fifth conference on applied natural language processing, pp 283–290 Lin CY, Hovy E (1997) Identifying topics by position. In: Fifth conference on applied natural language processing, pp 283–290
33.
Zurück zum Zitat Liu Y (2019) Fine-tune bert for extractive summarization. arXiv preprint arXiv:1903.10318 Liu Y (2019) Fine-tune bert for extractive summarization. arXiv preprint arXiv:1903.10318
34.
Zurück zum Zitat Manly B, McDonald L, Thomas DL, McDonald TL, Erickson WP (2007) Resource selection by animals: statistical design and analysis for field studies. Springer, Berlin Manly B, McDonald L, Thomas DL, McDonald TL, Erickson WP (2007) Resource selection by animals: statistical design and analysis for field studies. Springer, Berlin
35.
Zurück zum Zitat Miao Y, Yu L, Blunsom P (2016) Neural variational inference for text processing. In: International conference on machine learning, pp 1727–1736 Miao Y, Yu L, Blunsom P (2016) Neural variational inference for text processing. In: International conference on machine learning, pp 1727–1736
37.
Zurück zum Zitat Mohri M, Pereira F, Riley M (2002) Weighted finite-state transducers in speech recognition. Comput Speech Lang 16(1):69–88CrossRef Mohri M, Pereira F, Riley M (2002) Weighted finite-state transducers in speech recognition. Comput Speech Lang 16(1):69–88CrossRef
38.
Zurück zum Zitat Nallapati R, Zhai F, Zhou B (2017) Summarunner: a recurrent neural network based sequence model for extractive summarization of documents. In: Thirty-first AAAI conference on artificial intelligence Nallapati R, Zhai F, Zhou B (2017) Summarunner: a recurrent neural network based sequence model for extractive summarization of documents. In: Thirty-first AAAI conference on artificial intelligence
39.
Zurück zum Zitat Nallapati R, Zhou B, Gulcehre C, Xiang B et al (2016) Abstractive text summarization using sequence-to-sequence RNNS and beyond. arXiv preprint arXiv:1602.06023 Nallapati R, Zhou B, Gulcehre C, Xiang B et al (2016) Abstractive text summarization using sequence-to-sequence RNNS and beyond. arXiv preprint arXiv:1602.06023
40.
Zurück zum Zitat Mikolov T, Sutskever I, Chen K, Corrado GS, Dean J (2013) Distributed representations of words and phrases and their compositionality. In: Advances in neural information processing systems, pp 3111–3119 Mikolov T, Sutskever I, Chen K, Corrado GS, Dean J (2013) Distributed representations of words and phrases and their compositionality. In: Advances in neural information processing systems, pp 3111–3119
41.
Zurück zum Zitat Paulus R, Xiong C, Socher R (2017) A deep reinforced model for abstractive summarization. arXiv preprint arXiv:1705.04304 Paulus R, Xiong C, Socher R (2017) A deep reinforced model for abstractive summarization. arXiv preprint arXiv:1705.04304
42.
Zurück zum Zitat Radev DR, Hovy E, McKeown K (2002) Introduction to the special issue on summarization. Comput Linguist 28(4):399–408CrossRef Radev DR, Hovy E, McKeown K (2002) Introduction to the special issue on summarization. Comput Linguist 28(4):399–408CrossRef
43.
Zurück zum Zitat Ranzato M, Chopra S, Auli M, Zaremba W (2016) Sequence level training with recurrent neural networks. In: ICLR Ranzato M, Chopra S, Auli M, Zaremba W (2016) Sequence level training with recurrent neural networks. In: ICLR
45.
Zurück zum Zitat Rush AM, Chopra S, Weston J (2015) A neural attention model for abstractive sentence summarization. arXiv preprint arXiv:1509.00685 Rush AM, Chopra S, Weston J (2015) A neural attention model for abstractive sentence summarization. arXiv preprint arXiv:1509.00685
46.
Zurück zum Zitat Ryang S, Abekawa T (2012) Framework of automatic text summarization using reinforcement learning. In: Proceedings of the 2012 joint conference on empirical methods in natural language processing and computational natural language learning. Association for Computational Linguistics, pp 256–265 Ryang S, Abekawa T (2012) Framework of automatic text summarization using reinforcement learning. In: Proceedings of the 2012 joint conference on empirical methods in natural language processing and computational natural language learning. Association for Computational Linguistics, pp 256–265
48.
Zurück zum Zitat Schulman J, Moritz P, Levine S, Jordan M, Abbeel P (2015) High-dimensional continuous control using generalized advantage estimation. arXiv preprint arXiv:1506.02438 Schulman J, Moritz P, Levine S, Jordan M, Abbeel P (2015) High-dimensional continuous control using generalized advantage estimation. arXiv preprint arXiv:1506.02438
49.
Zurück zum Zitat See A, Liu PJ, Manning CD (2017) Get to the point: summarization with pointer-generator networks. arXiv preprint arXiv:1704.04368 See A, Liu PJ, Manning CD (2017) Get to the point: summarization with pointer-generator networks. arXiv preprint arXiv:1704.04368
50.
Zurück zum Zitat Socher R, Pennington J, Huang EH, Ng AY, Manning CD (2011) Semi-supervised recursive autoencoders for predicting sentiment distributions. In: Proceedings of the conference on empirical methods in natural language processing. Association for Computational Linguistics, pp 151–161 Socher R, Pennington J, Huang EH, Ng AY, Manning CD (2011) Semi-supervised recursive autoencoders for predicting sentiment distributions. In: Proceedings of the conference on empirical methods in natural language processing. Association for Computational Linguistics, pp 151–161
51.
Zurück zum Zitat Sutton RS, Barto AG (2011) Reinforcement learning: an introduction Sutton RS, Barto AG (2011) Reinforcement learning: an introduction
52.
Zurück zum Zitat Takase S, Suzuki J, Okazaki N, Hirao T, Nagata M (2016) Neural headline generation on abstract meaning representation. In: Proceedings of the 2016 conference on empirical methods in natural language processing, pp 1054–1059 Takase S, Suzuki J, Okazaki N, Hirao T, Nagata M (2016) Neural headline generation on abstract meaning representation. In: Proceedings of the 2016 conference on empirical methods in natural language processing, pp 1054–1059
53.
Zurück zum Zitat Tan J, Wan X, Xiao J (2017) From neural sentence summarization to headline generation: a coarse-to-fine approach. In: IJCAI, pp 4109–4115 Tan J, Wan X, Xiao J (2017) From neural sentence summarization to headline generation: a coarse-to-fine approach. In: IJCAI, pp 4109–4115
54.
Zurück zum Zitat Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser L, Polosukhin I (2017) Attention is all you need. In: Advances in neural information processing systems, pp 5998–6008 Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser L, Polosukhin I (2017) Attention is all you need. In: Advances in neural information processing systems, pp 5998–6008
55.
Zurück zum Zitat Vinyals O, Fortunato M, Jaitly N (2015) Pointer networks. In: Advances in neural information processing systems, pp 2692–2700 Vinyals O, Fortunato M, Jaitly N (2015) Pointer networks. In: Advances in neural information processing systems, pp 2692–2700
56.
Zurück zum Zitat Williams RJ (1992) Simple statistical gradient-following algorithms for connectionist reinforcement learning. Mach Learn 8(3–4):229–256MATH Williams RJ (1992) Simple statistical gradient-following algorithms for connectionist reinforcement learning. Mach Learn 8(3–4):229–256MATH
57.
Zurück zum Zitat Woodsend K, Feng Y, Lapata M (2010) Generation with quasi-synchronous grammar. In: Proceedings of the 2010 conference on empirical methods in natural language processing. Association for Computational Linguistics, pp 513–523 Woodsend K, Feng Y, Lapata M (2010) Generation with quasi-synchronous grammar. In: Proceedings of the 2010 conference on empirical methods in natural language processing. Association for Computational Linguistics, pp 513–523
58.
Zurück zum Zitat Yin XC, Pei WY, Zhang J, Hao HW (2015) Multi-orientation scene text detection with adaptive clustering. IEEE Trans Pattern Anal Mach Intell 37(9):1930–1937CrossRef Yin XC, Pei WY, Zhang J, Hao HW (2015) Multi-orientation scene text detection with adaptive clustering. IEEE Trans Pattern Anal Mach Intell 37(9):1930–1937CrossRef
59.
Zurück zum Zitat Zajic D, Dorr B, Schwartz R (2002) Automatic headline generation for newspaper stories. In: Workshop on automatic summarization, pp 78–85 Zajic D, Dorr B, Schwartz R (2002) Automatic headline generation for newspaper stories. In: Workshop on automatic summarization, pp 78–85
60.
Zurück zum Zitat Zhou L, Hovy E (2003) Headline summarization at ISI. In: Proceedings of the HLT-NAACL 2003 text summarization workshop and document understanding conference (DUC 2003), pp 174–178 Zhou L, Hovy E (2003) Headline summarization at ISI. In: Proceedings of the HLT-NAACL 2003 text summarization workshop and document understanding conference (DUC 2003), pp 174–178
Metadaten
Titel
SHEG: summarization and headline generation of news articles using deep learning
verfasst von
Rajeev Kumar Singh
Sonia Khetarpaul
Rohan Gorantla
Sai Giridhar Allada
Publikationsdatum
23.07.2020
Verlag
Springer London
Erschienen in
Neural Computing and Applications / Ausgabe 8/2021
Print ISSN: 0941-0643
Elektronische ISSN: 1433-3058
DOI
https://doi.org/10.1007/s00521-020-05188-9

Weitere Artikel der Ausgabe 8/2021

Neural Computing and Applications 8/2021 Zur Ausgabe