Skip to main content

2024 | OriginalPaper | Buchkapitel

A Second Look on BASS – Boosting Abstractive Summarization with Unified Semantic Graphs

A Replication Study

verfasst von : Osman Alperen Koraş, Jörg Schlötterer, Christin Seifert

Erschienen in: Advances in Information Retrieval

Verlag: Springer Nature Switzerland

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

We present a detailed replication study of the BASS framework, an abstractive summarization system based on the notion of Unified Semantic Graphs. Our investigation includes challenges in replicating key components and an ablation study to systematically isolate error sources rooted in replicating novel components. Our findings reveal discrepancies in performance compared to the original work. We highlight the significance of paying careful attention even to reasonably omitted details for replicating advanced frameworks like BASS, and emphasize key practices for writing replicable papers.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Anhänge
Nur mit Berechtigung zugänglich
Fußnoten
2
  https://static-content.springer.com/image/chp%3A10.1007%2F978-3-031-56066-8_11/MediaObjects/562556_1_En_11_Figbh_HTML.gif https://static-content.springer.com/image/chp%3A10.1007%2F978-3-031-56066-8_11/MediaObjects/562556_1_En_11_Figbi_HTML.gif https://static-content.springer.com/image/chp%3A10.1007%2F978-3-031-56066-8_11/MediaObjects/562556_1_En_11_Figbj_HTML.gif .
 
3
We excluded less than 0.015% of the documents, having no effect on the final score.
 
4
While the BASS paper reports sentence-level R-L scores, they systematically match better with our summary-level R-L\(_{sum}\) scores, which may indicate that prior results are actually R-L\(_{sum}\) scores. Hence we place the scores reported by BASS between columns and compare them with our R-L\(_{sum}\) scores.
 
5
We did not train further, because we already doubled the computational budget used for the original paper.
 
6
The loss curve is decreasing by 0.003 points every 10,000 steps in the range of 300,000 to 450,000 steps, and by 0.002 points every 10,000 steps in the range of 450,000 to 600,000 steps at average.
 
Literatur
1.
Zurück zum Zitat Belz, A., Agarwal, S., Shimorina, A., Reiter, E.: A systematic review of reproducibility research in natural language processing. In: Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, pp. 381–393. Association for Computational Linguistics, Online (2021). https://doi.org/10.18653/v1/2021.eacl-main.29 Belz, A., Agarwal, S., Shimorina, A., Reiter, E.: A systematic review of reproducibility research in natural language processing. In: Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, pp. 381–393. Association for Computational Linguistics, Online (2021). https://​doi.​org/​10.​18653/​v1/​2021.​eacl-main.​29
4.
Zurück zum Zitat Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, Minnesota (Volume 1: Long and Short Papers), pp. 4171–4186. Association for Computational Linguistics (2019). https://doi.org/10.18653/v1/N19-1423 Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, Minnesota (Volume 1: Long and Short Papers), pp. 4171–4186. Association for Computational Linguistics (2019). https://​doi.​org/​10.​18653/​v1/​N19-1423
6.
Zurück zum Zitat Dosovitskiy, A., et al.: An image is worth \(16 \times 16\) words: transformers for image recognition at scale. arXiv abs/2010.11929 (2020) Dosovitskiy, A., et al.: An image is worth \(16 \times 16\) words: transformers for image recognition at scale. arXiv abs/2010.11929 (2020)
7.
Zurück zum Zitat Dou, Z.Y., Liu, P., Hayashi, H., Jiang, Z., Neubig, G.: GSum: a general framework for guided neural abstractive summarization. In: Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 4830–4842. Association for Computational Linguistics, Online (2021). https://doi.org/10.18653/v1/2021.naacl-main.384 Dou, Z.Y., Liu, P., Hayashi, H., Jiang, Z., Neubig, G.: GSum: a general framework for guided neural abstractive summarization. In: Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 4830–4842. Association for Computational Linguistics, Online (2021). https://​doi.​org/​10.​18653/​v1/​2021.​naacl-main.​384
8.
Zurück zum Zitat Dozat, T., Manning, C.D.: Simpler but more accurate semantic dependency parsing. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, Melbourne, Australia (Volume 2: Short Papers), pp. 484–490. Association for Computational Linguistics (2018). https://doi.org/10.18653/v1/P18-2077 Dozat, T., Manning, C.D.: Simpler but more accurate semantic dependency parsing. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, Melbourne, Australia (Volume 2: Short Papers), pp. 484–490. Association for Computational Linguistics (2018). https://​doi.​org/​10.​18653/​v1/​P18-2077
9.
Zurück zum Zitat Dror, R., Baumer, G., Shlomov, S., Reichart, R.: The hitchhiker’s guide to testing statistical significance in natural language processing. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, Melbourne, Australia (Volume 1: Long Papers), pp. 1383–1392. Association for Computational Linguistics (2018). https://doi.org/10.18653/v1/P18-1128 Dror, R., Baumer, G., Shlomov, S., Reichart, R.: The hitchhiker’s guide to testing statistical significance in natural language processing. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, Melbourne, Australia (Volume 1: Long Papers), pp. 1383–1392. Association for Computational Linguistics (2018). https://​doi.​org/​10.​18653/​v1/​P18-1128
11.
Zurück zum Zitat Fan, A., Gardent, C., Braud, C., Bordes, A.: Using local knowledge graph construction to scale Seq2Seq models to multi-document inputs. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, pp. 4186–4196. Association for Computational Linguistics (2019). https://doi.org/10.18653/v1/D19-1428 Fan, A., Gardent, C., Braud, C., Bordes, A.: Using local knowledge graph construction to scale Seq2Seq models to multi-document inputs. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, pp. 4186–4196. Association for Computational Linguistics (2019). https://​doi.​org/​10.​18653/​v1/​D19-1428
16.
Zurück zum Zitat Huang, L., Wu, L., Wang, L.: Knowledge graph-augmented abstractive summarization with semantic-driven cloze reward. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 5094–5107. Association for Computational Linguistics, Online (2020). https://doi.org/10.18653/v1/2020.acl-main.457 Huang, L., Wu, L., Wang, L.: Knowledge graph-augmented abstractive summarization with semantic-driven cloze reward. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 5094–5107. Association for Computational Linguistics, Online (2020). https://​doi.​org/​10.​18653/​v1/​2020.​acl-main.​457
18.
Zurück zum Zitat Klicpera, J., Bojchevski, A., Günnemann, S.: Predict then propagate: graph neural networks meet personalized pagerank. In: International Conference on Learning Representations (2018) Klicpera, J., Bojchevski, A., Günnemann, S.: Predict then propagate: graph neural networks meet personalized pagerank. In: International Conference on Learning Representations (2018)
19.
Zurück zum Zitat Lewis, M., et al.: BART: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 7871–7880. Association for Computational Linguistics, Online (2020). https://doi.org/10.18653/v1/2020.acl-main.703 Lewis, M., et al.: BART: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 7871–7880. Association for Computational Linguistics, Online (2020). https://​doi.​org/​10.​18653/​v1/​2020.​acl-main.​703
21.
Zurück zum Zitat Liu, Y., Lapata, M.: Text summarization with pretrained encoders. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, Hong Kong, China (EMNLP-IJCNLP), pp. 3730–3740. Association for Computational Linguistics (2019). https://doi.org/10.18653/v1/D19-1387 Liu, Y., Lapata, M.: Text summarization with pretrained encoders. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, Hong Kong, China (EMNLP-IJCNLP), pp. 3730–3740. Association for Computational Linguistics (2019). https://​doi.​org/​10.​18653/​v1/​D19-1387
23.
Zurück zum Zitat Paulus, R., Xiong, C., Socher, R.: A deep reinforced model for abstractive summarization (2017) Paulus, R., Xiong, C., Socher, R.: A deep reinforced model for abstractive summarization (2017)
24.
Zurück zum Zitat Qi, P., Huang, Z., Sun, Y., Luo, H.: A knowledge graph-based abstractive model integrating semantic and structural information for summarizing Chinese meetings. In: 2022 IEEE 25th International Conference on Computer Supported Cooperative Work in Design (CSCWD), pp. 746–751 (2022). https://doi.org/10.1109/CSCWD54268.2022.9776298 Qi, P., Huang, Z., Sun, Y., Luo, H.: A knowledge graph-based abstractive model integrating semantic and structural information for summarizing Chinese meetings. In: 2022 IEEE 25th International Conference on Computer Supported Cooperative Work in Design (CSCWD), pp. 746–751 (2022). https://​doi.​org/​10.​1109/​CSCWD54268.​2022.​9776298
25.
Zurück zum Zitat Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21(1) (2020) Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21(1) (2020)
26.
Zurück zum Zitat Sharma, E., Li, C., Wang, L.: BIGPATENT: a large-scale dataset for abstractive and coherent summarization. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, pp. 2204–2213. Association for Computational Linguistics (2019). https://doi.org/10.18653/v1/P19-1212 Sharma, E., Li, C., Wang, L.: BIGPATENT: a large-scale dataset for abstractive and coherent summarization. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, pp. 2204–2213. Association for Computational Linguistics (2019). https://​doi.​org/​10.​18653/​v1/​P19-1212
27.
Zurück zum Zitat Vaswani, A., et al.: Attention is all you need. In: Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS 2017, pp. 6000–6010. Curran Associates Inc., Red Hook (2017) Vaswani, A., et al.: Attention is all you need. In: Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS 2017, pp. 6000–6010. Curran Associates Inc., Red Hook (2017)
28.
Zurück zum Zitat Velickovic, P., Cucurull, G., Casanova, A., Romero, A., Lio’, P., Bengio, Y.: Graph attention networks. arXiv abs/1710.10903 (2017) Velickovic, P., Cucurull, G., Casanova, A., Romero, A., Lio’, P., Bengio, Y.: Graph attention networks. arXiv abs/1710.10903 (2017)
29.
30.
Zurück zum Zitat Wu, W., et al.: BASS: boosting abstractive summarization with unified semantic graph. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 6052–6067. Association for Computational Linguistics, Online (2021). https://doi.org/10.18653/v1/2021.acl-long.472 Wu, W., et al.: BASS: boosting abstractive summarization with unified semantic graph. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 6052–6067. Association for Computational Linguistics, Online (2021). https://​doi.​org/​10.​18653/​v1/​2021.​acl-long.​472
32.
33.
Zurück zum Zitat Ying, C., et al.: Do transformers really perform bad for graph representation? In: Neural Information Processing Systems (2021) Ying, C., et al.: Do transformers really perform bad for graph representation? In: Neural Information Processing Systems (2021)
34.
Zurück zum Zitat Zhang, J., Zhao, Y., Saleh, M., Liu, P.J.: Pegasus: pre-training with extracted gap-sentences for abstractive summarization. In: Proceedings of the 37th International Conference on Machine Learning, ICML 2020. JMLR.org (2020) Zhang, J., Zhao, Y., Saleh, M., Liu, P.J.: Pegasus: pre-training with extracted gap-sentences for abstractive summarization. In: Proceedings of the 37th International Conference on Machine Learning, ICML 2020. JMLR.org (2020)
36.
Zurück zum Zitat Zhu, C., et al.: Enhancing factual consistency of abstractive summarization. In: Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 718–733. Association for Computational Linguistics, Online (2021). https://doi.org/10.18653/v1/2021.naacl-main.58 Zhu, C., et al.: Enhancing factual consistency of abstractive summarization. In: Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 718–733. Association for Computational Linguistics, Online (2021). https://​doi.​org/​10.​18653/​v1/​2021.​naacl-main.​58
37.
Zurück zum Zitat Zhuang, L., Wayne, L., Ya, S., Jun, Z.: A robustly optimized BERT pre-training approach with post-training. In: Proceedings of the 20th Chinese National Conference on Computational Linguistics, Huhhot, China, pp. 1218–1227. Chinese Information Processing Society of China (2021) Zhuang, L., Wayne, L., Ya, S., Jun, Z.: A robustly optimized BERT pre-training approach with post-training. In: Proceedings of the 20th Chinese National Conference on Computational Linguistics, Huhhot, China, pp. 1218–1227. Chinese Information Processing Society of China (2021)
Metadaten
Titel
A Second Look on BASS – Boosting Abstractive Summarization with Unified Semantic Graphs
verfasst von
Osman Alperen Koraş
Jörg Schlötterer
Christin Seifert
Copyright-Jahr
2024
DOI
https://doi.org/10.1007/978-3-031-56066-8_11

Premium Partner