Skip to main content
Top

2024 | OriginalPaper | Chapter

A Second Look on BASS – Boosting Abstractive Summarization with Unified Semantic Graphs

A Replication Study

Authors : Osman Alperen Koraş, Jörg Schlötterer, Christin Seifert

Published in: Advances in Information Retrieval

Publisher: Springer Nature Switzerland

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

We present a detailed replication study of the BASS framework, an abstractive summarization system based on the notion of Unified Semantic Graphs. Our investigation includes challenges in replicating key components and an ablation study to systematically isolate error sources rooted in replicating novel components. Our findings reveal discrepancies in performance compared to the original work. We highlight the significance of paying careful attention even to reasonably omitted details for replicating advanced frameworks like BASS, and emphasize key practices for writing replicable papers.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Appendix
Available only for authorised users
Footnotes
2
  https://static-content.springer.com/image/chp%3A10.1007%2F978-3-031-56066-8_11/MediaObjects/562556_1_En_11_Figbh_HTML.gif https://static-content.springer.com/image/chp%3A10.1007%2F978-3-031-56066-8_11/MediaObjects/562556_1_En_11_Figbi_HTML.gif https://static-content.springer.com/image/chp%3A10.1007%2F978-3-031-56066-8_11/MediaObjects/562556_1_En_11_Figbj_HTML.gif .
 
3
We excluded less than 0.015% of the documents, having no effect on the final score.
 
4
While the BASS paper reports sentence-level R-L scores, they systematically match better with our summary-level R-L\(_{sum}\) scores, which may indicate that prior results are actually R-L\(_{sum}\) scores. Hence we place the scores reported by BASS between columns and compare them with our R-L\(_{sum}\) scores.
 
5
We did not train further, because we already doubled the computational budget used for the original paper.
 
6
The loss curve is decreasing by 0.003 points every 10,000 steps in the range of 300,000 to 450,000 steps, and by 0.002 points every 10,000 steps in the range of 450,000 to 600,000 steps at average.
 
Literature
1.
go back to reference Belz, A., Agarwal, S., Shimorina, A., Reiter, E.: A systematic review of reproducibility research in natural language processing. In: Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, pp. 381–393. Association for Computational Linguistics, Online (2021). https://doi.org/10.18653/v1/2021.eacl-main.29 Belz, A., Agarwal, S., Shimorina, A., Reiter, E.: A systematic review of reproducibility research in natural language processing. In: Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, pp. 381–393. Association for Computational Linguistics, Online (2021). https://​doi.​org/​10.​18653/​v1/​2021.​eacl-main.​29
4.
go back to reference Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, Minnesota (Volume 1: Long and Short Papers), pp. 4171–4186. Association for Computational Linguistics (2019). https://doi.org/10.18653/v1/N19-1423 Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, Minnesota (Volume 1: Long and Short Papers), pp. 4171–4186. Association for Computational Linguistics (2019). https://​doi.​org/​10.​18653/​v1/​N19-1423
6.
go back to reference Dosovitskiy, A., et al.: An image is worth \(16 \times 16\) words: transformers for image recognition at scale. arXiv abs/2010.11929 (2020) Dosovitskiy, A., et al.: An image is worth \(16 \times 16\) words: transformers for image recognition at scale. arXiv abs/2010.11929 (2020)
7.
go back to reference Dou, Z.Y., Liu, P., Hayashi, H., Jiang, Z., Neubig, G.: GSum: a general framework for guided neural abstractive summarization. In: Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 4830–4842. Association for Computational Linguistics, Online (2021). https://doi.org/10.18653/v1/2021.naacl-main.384 Dou, Z.Y., Liu, P., Hayashi, H., Jiang, Z., Neubig, G.: GSum: a general framework for guided neural abstractive summarization. In: Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 4830–4842. Association for Computational Linguistics, Online (2021). https://​doi.​org/​10.​18653/​v1/​2021.​naacl-main.​384
8.
go back to reference Dozat, T., Manning, C.D.: Simpler but more accurate semantic dependency parsing. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, Melbourne, Australia (Volume 2: Short Papers), pp. 484–490. Association for Computational Linguistics (2018). https://doi.org/10.18653/v1/P18-2077 Dozat, T., Manning, C.D.: Simpler but more accurate semantic dependency parsing. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, Melbourne, Australia (Volume 2: Short Papers), pp. 484–490. Association for Computational Linguistics (2018). https://​doi.​org/​10.​18653/​v1/​P18-2077
9.
go back to reference Dror, R., Baumer, G., Shlomov, S., Reichart, R.: The hitchhiker’s guide to testing statistical significance in natural language processing. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, Melbourne, Australia (Volume 1: Long Papers), pp. 1383–1392. Association for Computational Linguistics (2018). https://doi.org/10.18653/v1/P18-1128 Dror, R., Baumer, G., Shlomov, S., Reichart, R.: The hitchhiker’s guide to testing statistical significance in natural language processing. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, Melbourne, Australia (Volume 1: Long Papers), pp. 1383–1392. Association for Computational Linguistics (2018). https://​doi.​org/​10.​18653/​v1/​P18-1128
11.
go back to reference Fan, A., Gardent, C., Braud, C., Bordes, A.: Using local knowledge graph construction to scale Seq2Seq models to multi-document inputs. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, pp. 4186–4196. Association for Computational Linguistics (2019). https://doi.org/10.18653/v1/D19-1428 Fan, A., Gardent, C., Braud, C., Bordes, A.: Using local knowledge graph construction to scale Seq2Seq models to multi-document inputs. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, pp. 4186–4196. Association for Computational Linguistics (2019). https://​doi.​org/​10.​18653/​v1/​D19-1428
16.
go back to reference Huang, L., Wu, L., Wang, L.: Knowledge graph-augmented abstractive summarization with semantic-driven cloze reward. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 5094–5107. Association for Computational Linguistics, Online (2020). https://doi.org/10.18653/v1/2020.acl-main.457 Huang, L., Wu, L., Wang, L.: Knowledge graph-augmented abstractive summarization with semantic-driven cloze reward. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 5094–5107. Association for Computational Linguistics, Online (2020). https://​doi.​org/​10.​18653/​v1/​2020.​acl-main.​457
18.
go back to reference Klicpera, J., Bojchevski, A., Günnemann, S.: Predict then propagate: graph neural networks meet personalized pagerank. In: International Conference on Learning Representations (2018) Klicpera, J., Bojchevski, A., Günnemann, S.: Predict then propagate: graph neural networks meet personalized pagerank. In: International Conference on Learning Representations (2018)
19.
go back to reference Lewis, M., et al.: BART: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 7871–7880. Association for Computational Linguistics, Online (2020). https://doi.org/10.18653/v1/2020.acl-main.703 Lewis, M., et al.: BART: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 7871–7880. Association for Computational Linguistics, Online (2020). https://​doi.​org/​10.​18653/​v1/​2020.​acl-main.​703
21.
go back to reference Liu, Y., Lapata, M.: Text summarization with pretrained encoders. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, Hong Kong, China (EMNLP-IJCNLP), pp. 3730–3740. Association for Computational Linguistics (2019). https://doi.org/10.18653/v1/D19-1387 Liu, Y., Lapata, M.: Text summarization with pretrained encoders. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, Hong Kong, China (EMNLP-IJCNLP), pp. 3730–3740. Association for Computational Linguistics (2019). https://​doi.​org/​10.​18653/​v1/​D19-1387
23.
go back to reference Paulus, R., Xiong, C., Socher, R.: A deep reinforced model for abstractive summarization (2017) Paulus, R., Xiong, C., Socher, R.: A deep reinforced model for abstractive summarization (2017)
24.
go back to reference Qi, P., Huang, Z., Sun, Y., Luo, H.: A knowledge graph-based abstractive model integrating semantic and structural information for summarizing Chinese meetings. In: 2022 IEEE 25th International Conference on Computer Supported Cooperative Work in Design (CSCWD), pp. 746–751 (2022). https://doi.org/10.1109/CSCWD54268.2022.9776298 Qi, P., Huang, Z., Sun, Y., Luo, H.: A knowledge graph-based abstractive model integrating semantic and structural information for summarizing Chinese meetings. In: 2022 IEEE 25th International Conference on Computer Supported Cooperative Work in Design (CSCWD), pp. 746–751 (2022). https://​doi.​org/​10.​1109/​CSCWD54268.​2022.​9776298
25.
go back to reference Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21(1) (2020) Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21(1) (2020)
26.
go back to reference Sharma, E., Li, C., Wang, L.: BIGPATENT: a large-scale dataset for abstractive and coherent summarization. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, pp. 2204–2213. Association for Computational Linguistics (2019). https://doi.org/10.18653/v1/P19-1212 Sharma, E., Li, C., Wang, L.: BIGPATENT: a large-scale dataset for abstractive and coherent summarization. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, pp. 2204–2213. Association for Computational Linguistics (2019). https://​doi.​org/​10.​18653/​v1/​P19-1212
27.
go back to reference Vaswani, A., et al.: Attention is all you need. In: Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS 2017, pp. 6000–6010. Curran Associates Inc., Red Hook (2017) Vaswani, A., et al.: Attention is all you need. In: Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS 2017, pp. 6000–6010. Curran Associates Inc., Red Hook (2017)
28.
go back to reference Velickovic, P., Cucurull, G., Casanova, A., Romero, A., Lio’, P., Bengio, Y.: Graph attention networks. arXiv abs/1710.10903 (2017) Velickovic, P., Cucurull, G., Casanova, A., Romero, A., Lio’, P., Bengio, Y.: Graph attention networks. arXiv abs/1710.10903 (2017)
29.
30.
go back to reference Wu, W., et al.: BASS: boosting abstractive summarization with unified semantic graph. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 6052–6067. Association for Computational Linguistics, Online (2021). https://doi.org/10.18653/v1/2021.acl-long.472 Wu, W., et al.: BASS: boosting abstractive summarization with unified semantic graph. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 6052–6067. Association for Computational Linguistics, Online (2021). https://​doi.​org/​10.​18653/​v1/​2021.​acl-long.​472
32.
33.
go back to reference Ying, C., et al.: Do transformers really perform bad for graph representation? In: Neural Information Processing Systems (2021) Ying, C., et al.: Do transformers really perform bad for graph representation? In: Neural Information Processing Systems (2021)
34.
go back to reference Zhang, J., Zhao, Y., Saleh, M., Liu, P.J.: Pegasus: pre-training with extracted gap-sentences for abstractive summarization. In: Proceedings of the 37th International Conference on Machine Learning, ICML 2020. JMLR.org (2020) Zhang, J., Zhao, Y., Saleh, M., Liu, P.J.: Pegasus: pre-training with extracted gap-sentences for abstractive summarization. In: Proceedings of the 37th International Conference on Machine Learning, ICML 2020. JMLR.org (2020)
36.
go back to reference Zhu, C., et al.: Enhancing factual consistency of abstractive summarization. In: Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 718–733. Association for Computational Linguistics, Online (2021). https://doi.org/10.18653/v1/2021.naacl-main.58 Zhu, C., et al.: Enhancing factual consistency of abstractive summarization. In: Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 718–733. Association for Computational Linguistics, Online (2021). https://​doi.​org/​10.​18653/​v1/​2021.​naacl-main.​58
37.
go back to reference Zhuang, L., Wayne, L., Ya, S., Jun, Z.: A robustly optimized BERT pre-training approach with post-training. In: Proceedings of the 20th Chinese National Conference on Computational Linguistics, Huhhot, China, pp. 1218–1227. Chinese Information Processing Society of China (2021) Zhuang, L., Wayne, L., Ya, S., Jun, Z.: A robustly optimized BERT pre-training approach with post-training. In: Proceedings of the 20th Chinese National Conference on Computational Linguistics, Huhhot, China, pp. 1218–1227. Chinese Information Processing Society of China (2021)
Metadata
Title
A Second Look on BASS – Boosting Abstractive Summarization with Unified Semantic Graphs
Authors
Osman Alperen Koraş
Jörg Schlötterer
Christin Seifert
Copyright Year
2024
DOI
https://doi.org/10.1007/978-3-031-56066-8_11

Premium Partner