Skip to main content
Top

2021 | OriginalPaper | Chapter

NEWS Article Summarization with Pretrained Transformer

Authors : Apar Garg, Saiteja Adusumilli, Shanmukha Yenneti, Tapas Badal, Deepak Garg, Vivek Pandey, Abhishek Nigam, Yashu Kant Gupta, Gyan Mittal, Rahul Agarwal

Published in: Advanced Computing

Publisher: Springer Singapore

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

Pretrained language models have shown tremendous improvement in many NLP applications including text summarization. Text-to-Text transfer transformer (T5) and Bidirectional Encoder Representations from Transformers (BERT) are most recent pretrained language models applied most widely in NLP research domain. In this paper we have shown how T5 and BERT can be applied for text summarization task and can be use for both abstractive and extractive summary generation tool. Our hypothesis is that T5 performance outperform over BART and transformer developed from the scratch. To test our hypothesis used our dataset containing more than 80K news articles and their summaries. This dataset has been tested using BART, Text-to-Text transformer (T5), model generated using transfer learning over T5, and an encoder-decoder based model developed from scratch. The results show that T5 gives better result than other three models used for testing.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Li, H., Zhu, J., Ma, C., Zhang, J., Zong, C.: Read, watch, listen and summarize: multi-modal summarization for asynchronous text, image, audio and video. IEEE Trans. Knowl. Data Eng. 31, 996–1009 (2018)CrossRef Li, H., Zhu, J., Ma, C., Zhang, J., Zong, C.: Read, watch, listen and summarize: multi-modal summarization for asynchronous text, image, audio and video. IEEE Trans. Knowl. Data Eng. 31, 996–1009 (2018)CrossRef
2.
go back to reference Menéndez, H.D., Plaza, L., Camacho, D.: Combining graph connectivity and genetic clustering to improve biomedical summarization. Paper Presented at the 2014 IEEE Congress on Evolutionary Computation (CEC) (2014) Menéndez, H.D., Plaza, L., Camacho, D.: Combining graph connectivity and genetic clustering to improve biomedical summarization. Paper Presented at the 2014 IEEE Congress on Evolutionary Computation (CEC) (2014)
3.
go back to reference Alampalli Ramu, N.A., Bandarupalli, M.S., Nekkanti, M.S.S., Ramesh, G.: Summarization of research publications using automatic extraction. In: Hemanth, D., Shakya, S., Baig, Z. (eds.) Intelligent Data Communication Technologies and Internet of Things, ICICI 2019 (2020) Alampalli Ramu, N.A., Bandarupalli, M.S., Nekkanti, M.S.S., Ramesh, G.: Summarization of research publications using automatic extraction. In: Hemanth, D., Shakya, S., Baig, Z. (eds.) Intelligent Data Communication Technologies and Internet of Things, ICICI 2019 (2020)
4.
go back to reference Lewis, M., et al.: BART Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension. Facebook AI Lewis, M., et al.: BART Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension. Facebook AI
5.
go back to reference Raffel, C., Shazeer, N., Roberts, A., Lee, K., Narang, S.: Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer Raffel, C., Shazeer, N., Roberts, A., Lee, K., Narang, S.: Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
6.
go back to reference Li, P., Lam, W., Bing, L., Wang, Z.: Deep recurrent generative decoder for abstractive text summarization. Key Laboratory on High Confidence Software Technologies (Sub-Lab, CUHK), Ministry of Education, China, Department of Systems Engineering and Engineering Management, The Chinese University of Hong Kong, AI Lab, Tencent Inc., Shenzhen, China Li, P., Lam, W., Bing, L., Wang, Z.: Deep recurrent generative decoder for abstractive text summarization. Key Laboratory on High Confidence Software Technologies (Sub-Lab, CUHK), Ministry of Education, China, Department of Systems Engineering and Engineering Management, The Chinese University of Hong Kong, AI Lab, Tencent Inc., Shenzhen, China
7.
go back to reference Ba-Alwi, F., Gaphari, G.H., Al-Duqaimi, F.: Arabic text summarization using latent semantic analysis. CJAST 10(2), 1–14 (2015) Ba-Alwi, F., Gaphari, G.H., Al-Duqaimi, F.: Arabic text summarization using latent semantic analysis. CJAST 10(2), 1–14 (2015)
10.
go back to reference Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, pp. 5998–6008 (2017) Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, pp. 5998–6008 (2017)
12.
go back to reference Barzilay, R., McKeown, K.R.: Sentence fusion for multidocument news summarization. Comput. Linguist. 31(3), 297–328 (2005)CrossRef Barzilay, R., McKeown, K.R.: Sentence fusion for multidocument news summarization. Comput. Linguist. 31(3), 297–328 (2005)CrossRef
13.
go back to reference Bing, L., Li, P., Liao, Y., Lam, W., Guo, W., Passonneau, R.: Abstractive multidocument summarization via phrase selection and merging. In: ACL, pp. 1587–1597 (2015) Bing, L., Li, P., Liao, Y., Lam, W., Guo, W., Passonneau, R.: Abstractive multidocument summarization via phrase selection and merging. In: ACL, pp. 1587–1597 (2015)
14.
go back to reference Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. arXiv preprint arXiv:1910.10683 (2019) Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. arXiv preprint arXiv:​1910.​10683 (2019)
15.
go back to reference Miao, Y., Blunsom, P.: Language as a latent variable: discrete generative models for sentence compression. In: EMNLP, pp. 319–328 (2016) Miao, Y., Blunsom, P.: Language as a latent variable: discrete generative models for sentence compression. In: EMNLP, pp. 319–328 (2016)
Metadata
Title
NEWS Article Summarization with Pretrained Transformer
Authors
Apar Garg
Saiteja Adusumilli
Shanmukha Yenneti
Tapas Badal
Deepak Garg
Vivek Pandey
Abhishek Nigam
Yashu Kant Gupta
Gyan Mittal
Rahul Agarwal
Copyright Year
2021
Publisher
Springer Singapore
DOI
https://doi.org/10.1007/978-981-16-0401-0_15

Premium Partner