Skip to main content
Erschienen in: Discover Computing 2/2022

14.03.2022 | Special Issue on ECIR 2021

Open-domain conversational search assistants: the Transformer is all you need

verfasst von: Rafael Ferreira, Mariana Leite, David Semedo, Joao Magalhaes

Erschienen in: Discover Computing | Ausgabe 2/2022

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

On the quest of providing a more natural interaction between users and search systems, open-domain conversational search assistants have emerged, by assisting users in answering questions about open topics in a conversational manner. In this work, we show how the Transformer architecture achieves state-of-the-art results in key IR tasks, leveraging the creation of conversational assistants that engage in open-domain conversational search with single, yet informative, answers. In particular, we propose a complete open-domain abstractive conversational search agent pipeline to address two major challenges: first, conversation context-aware search and second, abstractive search-answers generation. To address the first challenge, the conversation context is modeled using a query rewriting method that unfolds the context of the conversation up to a specific moment to search for the correct answers. These answers are then passed to a Transformer-based re-ranker to further improve retrieval performance. The second challenge, is tackled with recent Abstractive Transformer architectures to generate a digest of the top most relevant passages. Experiments show that Transformers deliver a solid performance across all tasks in conversational search, outperforming several baselines. This work is an expanded version of Ferreira et al. (Open-domain conversational search assistant with transformers. In: Advances in information retrieval—43rd European conference on IR research, ECIR 2021, virtual event, 28 March–1 April 2021, proceedings, Part I. Springer) which provides more details about the various components of the of the system, and extends the automatic evaluation with a novel user-study, which confirmed the need for the conversational search paradigm, and assessed the performance of our answer generation approach.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
Zurück zum Zitat Belkin, N. J. (1980). Anomalous states of knowledge as a basis for information retrieval. Canadian Journal of Information Science, 5(1), 133–143. Belkin, N. J. (1980). Anomalous states of knowledge as a basis for information retrieval. Canadian Journal of Information Science, 5(1), 133–143.
Zurück zum Zitat Choi, E., He, H., Iyyer, M., Yatskar, M., Yih, W., Choi, Y., Liang, P., & Zettlemoyer, L. (2018). QuAC: Question answering in context. In Proceedings of the 2018 conference on empirical methods in natural language processing, Brussels, Belgium, October 31–November 4, 2018 (pp. 2174–2184). Association for Computational Linguistics. Choi, E., He, H., Iyyer, M., Yatskar, M., Yih, W., Choi, Y., Liang, P., & Zettlemoyer, L. (2018). QuAC: Question answering in context. In Proceedings of the 2018 conference on empirical methods in natural language processing, Brussels, Belgium, October 31–November 4, 2018 (pp. 2174–2184). Association for Computational Linguistics.
Zurück zum Zitat Clarke, C. L. A. (2019). WaterlooClarke at the TREC 2019 conversational assistant track. In E. M. Voorhees & A. Ellis (Eds.), Proceedings of the twenty-eighth Text REtrieval Conference, TREC 2019, Gaithersburg, Maryland, USA, November 13–15, 2019. NIST Special Publication (Vol. 1250). National Institute of Standards and Technology (NIST). Clarke, C. L. A. (2019). WaterlooClarke at the TREC 2019 conversational assistant track. In E. M. Voorhees & A. Ellis (Eds.), Proceedings of the twenty-eighth Text REtrieval Conference, TREC 2019, Gaithersburg, Maryland, USA, November 13–15, 2019. NIST Special Publication (Vol. 1250). National Institute of Standards and Technology (NIST).
Zurück zum Zitat Croft, W. B., & Thompson, R. H. (1987). I3R: A new approach to the design of document retrieval systems. JASIST, 38(6), 389–404.CrossRef Croft, W. B., & Thompson, R. H. (1987). I3R: A new approach to the design of document retrieval systems. JASIST, 38(6), 389–404.CrossRef
Zurück zum Zitat Culpepper, J. S., Diaz, F., & Smucker, M. D. (2018). Research frontiers in information retrieval: Report from the third strategic workshop on information retrieval in Lorne (SWIRL 2018). SIGIR Forum, 52(1), 34–90.CrossRef Culpepper, J. S., Diaz, F., & Smucker, M. D. (2018). Research frontiers in information retrieval: Report from the third strategic workshop on information retrieval in Lorne (SWIRL 2018). SIGIR Forum, 52(1), 34–90.CrossRef
Zurück zum Zitat Dai, Z., Xiong, C., Callan, J., & Liu, Z. (2018). Convolutional neural networks for soft-matching n-grams in ad hoc search. In Proceedings of the eleventh ACM international conference on web search and data mining, WSDM 2018, Marina Del Rey, CA, USA, February 5-9, 2018 (pp. 126–134). ACM. Dai, Z., Xiong, C., Callan, J., & Liu, Z. (2018). Convolutional neural networks for soft-matching n-grams in ad hoc search. In Proceedings of the eleventh ACM international conference on web search and data mining, WSDM 2018, Marina Del Rey, CA, USA, February 5-9, 2018 (pp. 126–134). ACM.
Zurück zum Zitat Dalton, J., Xiong, C., & Callan, J. (2020a). TREC CAsT 2019: The conversational assistance track overview. CoRR, abs/2003.13624. Dalton, J., Xiong, C., & Callan, J. (2020a). TREC CAsT 2019: The conversational assistance track overview. CoRR, abs/2003.13624.
Zurück zum Zitat Dalton, J., Xiong, C., & Callan, J. (2020b). The TREC conversational assistance track (CAsT). Dalton, J., Xiong, C., & Callan, J. (2020b). The TREC conversational assistance track (CAsT).
Zurück zum Zitat Devlin, J., Chang, M., Lee, K., & Toutanova, K. (2019). BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 conference of the North American Chapter of the Association for Computational Linguistics: Human language technologies, NAACL-HLT 2019, Minneapolis, MN, USA, June 2–7, 2019. Long and Short Papers (Vol. 1, pp. 4171–4186). Association for Computational Linguistics. Devlin, J., Chang, M., Lee, K., & Toutanova, K. (2019). BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 conference of the North American Chapter of the Association for Computational Linguistics: Human language technologies, NAACL-HLT 2019, Minneapolis, MN, USA, June 2–7, 2019. Long and Short Papers (Vol. 1, pp. 4171–4186). Association for Computational Linguistics.
Zurück zum Zitat Dietz, L., Gamari, B., & Dalton, J. (2018). TREC CAR 2.1: A data set for complex answer retrieval. Dietz, L., Gamari, B., & Dalton, J. (2018). TREC CAR 2.1: A data set for complex answer retrieval.
Zurück zum Zitat Dinan, E., Roller, S., Shuster, K., Fan, A., Auli, M., & Weston, J. (2019). Wizard of Wikipedia: Knowledge-powered conversational agents. In 7th International conference on learning representations, ICLR 2019, New Orleans, LA, USA, May 6–9, 2019. OpenReview.net. Dinan, E., Roller, S., Shuster, K., Fan, A., Auli, M., & Weston, J. (2019). Wizard of Wikipedia: Knowledge-powered conversational agents. In 7th International conference on learning representations, ICLR 2019, New Orleans, LA, USA, May 6–9, 2019. OpenReview.net.
Zurück zum Zitat Dou, Z., Liu, P., Hayashi, H., Jiang, Z., & Neubig, G. (2021). GSum: A general framework for guided neural abstractive summarization. In Proceedings of the 2021 conference of the North American Chapter of the Association for Computational Linguistics: Human language technologies, NAACL-HLT 2021, Online, June 6–11, 2021 (pp. 4830–4842). Association for Computational Linguistics. Dou, Z., Liu, P., Hayashi, H., Jiang, Z., & Neubig, G. (2021). GSum: A general framework for guided neural abstractive summarization. In Proceedings of the 2021 conference of the North American Chapter of the Association for Computational Linguistics: Human language technologies, NAACL-HLT 2021, Online, June 6–11, 2021 (pp. 4830–4842). Association for Computational Linguistics.
Zurück zum Zitat Elgohary, A., Peskov, D., & Boyd-Graber, J. L. (2019). Can you unpack that? Learning to rewrite questions-in-context. In K. Inui, J. Jiang, V. Ng & X. Wan (Eds.), Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing, EMNLP-IJCNLP 2019, Hong Kong, China, November 3–7, 2019 (pp. 5917–5923). Association for Computational Linguistics. Elgohary, A., Peskov, D., & Boyd-Graber, J. L. (2019). Can you unpack that? Learning to rewrite questions-in-context. In K. Inui, J. Jiang, V. Ng & X. Wan (Eds.), Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing, EMNLP-IJCNLP 2019, Hong Kong, China, November 3–7, 2019 (pp. 5917–5923). Association for Computational Linguistics.
Zurück zum Zitat Ferreira, R., Leite, M., Semedo, D., & Magalhães, J. (2021). Open-domain conversational search assistant with transformers. In Advances in information Retrieval—43rd European conference on IR research, ECIR 2021, virtual event, Proceedings, Part I, March 28–April 1, 2021. Lecture notes in computer science (Vol. 12656, pp. 130–145). Springer. Ferreira, R., Leite, M., Semedo, D., & Magalhães, J. (2021). Open-domain conversational search assistant with transformers. In Advances in information Retrieval—43rd European conference on IR research, ECIR 2021, virtual event, Proceedings, Part I, March 28–April 1, 2021. Lecture notes in computer science (Vol. 12656, pp. 130–145). Springer.
Zurück zum Zitat Gao, J., Galley, M., & Li, L. (2019). Neural approaches to conversational AI. Foundations and Trends in Information Retrieval, 13(2–3), 127–298.CrossRef Gao, J., Galley, M., & Li, L. (2019). Neural approaches to conversational AI. Foundations and Trends in Information Retrieval, 13(2–3), 127–298.CrossRef
Zurück zum Zitat Gardner, M., Grus, J., Neumann, M., Tafjord, O., Dasigi, P., Liu, N. F., Peters, M., Schmitz, M., & Zettlemoyer, L. (2018). AllenNLP: A deep semantic natural language processing platform. In Proceedings of workshop for NLP open source software (NLP-OSS), Melbourne, Australia (pp. 1–6). Association for Computational Linguistics. Gardner, M., Grus, J., Neumann, M., Tafjord, O., Dasigi, P., Liu, N. F., Peters, M., Schmitz, M., & Zettlemoyer, L. (2018). AllenNLP: A deep semantic natural language processing platform. In Proceedings of workshop for NLP open source software (NLP-OSS), Melbourne, Australia (pp. 1–6). Association for Computational Linguistics.
Zurück zum Zitat Guu, K., Lee, K., Tung, Z., Pasupat, P., & Chang, M. (2020). REALM: retrieval-augmented language model pre-training. CoRR, abs/2002.08909. Guu, K., Lee, K., Tung, Z., Pasupat, P., & Chang, M. (2020). REALM: retrieval-augmented language model pre-training. CoRR, abs/2002.08909.
Zurück zum Zitat Han, S., Wang, X., Bendersky, M., & Najork, M. (2020). Learning-to-rank with BERT in TF-ranking. CoRR, abs/2004.08476. Han, S., Wang, X., Bendersky, M., & Najork, M. (2020). Learning-to-rank with BERT in TF-ranking. CoRR, abs/2004.08476.
Zurück zum Zitat Hermann, K. M., Kocisky, T., Grefenstette, E., Espeholt, L., Kay, W., Suleyman, M., & Blunsom, P. (2015). Teaching machines to read and comprehend. In Advances in neural information processing systems (pp. 1693–1701). Hermann, K. M., Kocisky, T., Grefenstette, E., Espeholt, L., Kay, W., Suleyman, M., & Blunsom, P. (2015). Teaching machines to read and comprehend. In Advances in neural information processing systems (pp. 1693–1701).
Zurück zum Zitat Humeau, S., Shuster, K., Lachaux, M., & Weston, J. (2020). Poly-encoders: Architectures and pre-training strategies for fast and accurate multi-sentence scoring. In 8th International conference on learning representations, ICLR 2020, Addis Ababa, Ethiopia, April 26–30, 2020. OpenReview.net. Humeau, S., Shuster, K., Lachaux, M., & Weston, J. (2020). Poly-encoders: Architectures and pre-training strategies for fast and accurate multi-sentence scoring. In 8th International conference on learning representations, ICLR 2020, Addis Ababa, Ethiopia, April 26–30, 2020. OpenReview.net.
Zurück zum Zitat Joshi, M., Chen, D., Liu, Y., Weld, D. S., Zettlemoyer, L., & Levy, O. (2020). SpanBERT: Improving pre-training by representing and predicting spans. Transactions of the Association for Computational Linguistics, 8, 64–77.CrossRef Joshi, M., Chen, D., Liu, Y., Weld, D. S., Zettlemoyer, L., & Levy, O. (2020). SpanBERT: Improving pre-training by representing and predicting spans. Transactions of the Association for Computational Linguistics, 8, 64–77.CrossRef
Zurück zum Zitat Khattab, O., & Zaharia, M. (2020). Colbert: Efficient and effective passage search via contextualized late interaction over BERT. In J. Huang, Y. Chang, X. Cheng, J. Kamps, V. Murdock, J. Wen & Y. Liu (Eds.), Proceedings of the 43rd international ACM SIGIR conference on research and development in Information Retrieval, SIGIR 2020, virtual event, China, July 25–30, 2020 (pp. 39–48). ACM. Khattab, O., & Zaharia, M. (2020). Colbert: Efficient and effective passage search via contextualized late interaction over BERT. In J. Huang, Y. Chang, X. Cheng, J. Kamps, V. Murdock, J. Wen & Y. Liu (Eds.), Proceedings of the 43rd international ACM SIGIR conference on research and development in Information Retrieval, SIGIR 2020, virtual event, China, July 25–30, 2020 (pp. 39–48). ACM.
Zurück zum Zitat Lee, K., Chang, M., & Toutanova, K. (2019). Latent retrieval for weakly supervised open domain question answering. In A. Korhonen, D. R. Traum & L. Màrquez (Eds.), Proceedings of the 57th conference of the Association for Computational Linguistics, ACL 2019, Florence, Italy, July 28–August 2, 2019. Long papers (Vol. 1, pp. 6086–6096). Association for Computational Linguistics. Lee, K., Chang, M., & Toutanova, K. (2019). Latent retrieval for weakly supervised open domain question answering. In A. Korhonen, D. R. Traum & L. Màrquez (Eds.), Proceedings of the 57th conference of the Association for Computational Linguistics, ACL 2019, Florence, Italy, July 28–August 2, 2019. Long papers (Vol. 1, pp. 6086–6096). Association for Computational Linguistics.
Zurück zum Zitat Lee, K., He, L., Lewis, M., & Zettlemoyer, L. (2017). End-to-end neural coreference resolution. In Proceedings of the 2017 conference on empirical methods in natural language processing, EMNLP 2017, Copenhagen, Denmark, September 9–11, 2017 (pp. 188–197). Association for Computational Linguistics. Lee, K., He, L., Lewis, M., & Zettlemoyer, L. (2017). End-to-end neural coreference resolution. In Proceedings of the 2017 conference on empirical methods in natural language processing, EMNLP 2017, Copenhagen, Denmark, September 9–11, 2017 (pp. 188–197). Association for Computational Linguistics.
Zurück zum Zitat Lewis, M., Liu, Y., Goyal, N., Ghazvininejad, M., Mohamed, A., Levy, O., Stoyanov, V., & Zettlemoyer, L. (2020). BART: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In D. Jurafsky, J. Chai, N. Schluter & J. R. Tetreault (Eds.), Proceedings of the 58th annual meeting of the Association for Computational Linguistics, ACL 2020, online, July 5–10, 2020 (pp. 7871–7880). Association for Computational Linguistics. Lewis, M., Liu, Y., Goyal, N., Ghazvininejad, M., Mohamed, A., Levy, O., Stoyanov, V., & Zettlemoyer, L. (2020). BART: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In D. Jurafsky, J. Chai, N. Schluter & J. R. Tetreault (Eds.), Proceedings of the 58th annual meeting of the Association for Computational Linguistics, ACL 2020, online, July 5–10, 2020 (pp. 7871–7880). Association for Computational Linguistics.
Zurück zum Zitat Li, J., Monroe, W., Ritter, A., Jurafsky, D., Galley, M., & Gao, J. (2016). Deep reinforcement learning for dialogue generation. In Proceedings of the 2016 conference on empirical methods in natural language processing, Austin, Texas (pp. 1192–1202). Association for Computational Linguistics. Li, J., Monroe, W., Ritter, A., Jurafsky, D., Galley, M., & Gao, J. (2016). Deep reinforcement learning for dialogue generation. In Proceedings of the 2016 conference on empirical methods in natural language processing, Austin, Texas (pp. 1192–1202). Association for Computational Linguistics.
Zurück zum Zitat Li, J., Monroe, W., Shi, T., Jean, S., Ritter, A., & Jurafsky, D. (2017). Adversarial learning for neural dialogue generation. In Proceedings of the 2017 conference on empirical methods in natural language processing, Copenhagen, Denmark (pp. 2157–2169). Association for Computational Linguistics. Li, J., Monroe, W., Shi, T., Jean, S., Ritter, A., & Jurafsky, D. (2017). Adversarial learning for neural dialogue generation. In Proceedings of the 2017 conference on empirical methods in natural language processing, Copenhagen, Denmark (pp. 2157–2169). Association for Computational Linguistics.
Zurück zum Zitat Lin, S., Yang, J., Nogueira, R., Tsai, M., Wang, C., & Lin, J. (2020). Conversational question reformulation via sequence-to-sequence architectures and pretrained language models. CoRR, abs/2004.01909. Lin, S., Yang, J., Nogueira, R., Tsai, M., Wang, C., & Lin, J. (2020). Conversational question reformulation via sequence-to-sequence architectures and pretrained language models. CoRR, abs/2004.01909.
Zurück zum Zitat Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., Levy, O., Lewis, M., Zettlemoyer, L., & Stoyanov, V. (2019). RoBERTa: A robustly optimized BERT pretraining approach. CoRR, abs/1907.11692. Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., Levy, O., Lewis, M., Zettlemoyer, L., & Stoyanov, V. (2019). RoBERTa: A robustly optimized BERT pretraining approach. CoRR, abs/1907.11692.
Zurück zum Zitat Mazaré, P., Humeau, S., Raison, M., & Bordes, A. (2018). Training millions of personalized dialogue agents. In Proceedings of the 2018 conference on empirical methods in natural language processing, Brussels, Belgium, October 31–November 4, 2018 (pp. 2775–2779). Association for Computational Linguistics. Mazaré, P., Humeau, S., Raison, M., & Bordes, A. (2018). Training millions of personalized dialogue agents. In Proceedings of the 2018 conference on empirical methods in natural language processing, Brussels, Belgium, October 31–November 4, 2018 (pp. 2775–2779). Association for Computational Linguistics.
Zurück zum Zitat Nguyen, T., Rosenberg, M., Song, X., Gao, J., Tiwary, S., Majumder, R., & Deng, L. (2016). MS MARCO: A human generated machine reading comprehension dataset. In Proceedings of the Workshop on Cognitive Computation: Integrating neural and symbolic approaches 2016 co-located with the 30th Annual Conference on Neural Information Processing Systems (NIPS 2016), Barcelona, Spain, December 9, 2016. CEUR workshop proceedings (Vol. 1773). CEUR-WS.org. Nguyen, T., Rosenberg, M., Song, X., Gao, J., Tiwary, S., Majumder, R., & Deng, L. (2016). MS MARCO: A human generated machine reading comprehension dataset. In Proceedings of the Workshop on Cognitive Computation: Integrating neural and symbolic approaches 2016 co-located with the 30th Annual Conference on Neural Information Processing Systems (NIPS 2016), Barcelona, Spain, December 9, 2016. CEUR workshop proceedings (Vol. 1773). CEUR-WS.org.
Zurück zum Zitat NIST. (2019). TREC Washington Post corpus. NIST. (2019). TREC Washington Post corpus.
Zurück zum Zitat Nogueira, R., & Cho, K. (2019). Passage re-ranking with BERT. CoRR, abs/1901.04085. Nogueira, R., & Cho, K. (2019). Passage re-ranking with BERT. CoRR, abs/1901.04085.
Zurück zum Zitat Nogueira, R., Jiang, Z., Pradeep, R., & Lin, J. (2020). Document ranking with a pretrained sequence-to-sequence model. In Proceedings of the 2020 conference on empirical methods in natural language processing: Findings, EMNLP 2020, online event, November 16–20, 2020 (pp. 708–718). Association for Computational Linguistics. Nogueira, R., Jiang, Z., Pradeep, R., & Lin, J. (2020). Document ranking with a pretrained sequence-to-sequence model. In Proceedings of the 2020 conference on empirical methods in natural language processing: Findings, EMNLP 2020, online event, November 16–20, 2020 (pp. 708–718). Association for Computational Linguistics.
Zurück zum Zitat Nogueira, R., Yang, W., Cho, K., & Lin, J. (2019). Multi-stage document ranking with BERT. CoRR, abs/1910.14424. Nogueira, R., Yang, W., Cho, K., & Lin, J. (2019). Multi-stage document ranking with BERT. CoRR, abs/1910.14424.
Zurück zum Zitat Oddy, R. N. (1977). Information retrieval through man–machine dialogue. Journal of Documentation, 33(1), 1–14.CrossRef Oddy, R. N. (1977). Information retrieval through man–machine dialogue. Journal of Documentation, 33(1), 1–14.CrossRef
Zurück zum Zitat Papineni, K., Roukos, S., Ward, T., & Zhu, W.-J. (2002). BLEU: A method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting of the Association for Computational Linguistics, Philadelphia, Pennsylvania, USA (pp. 311–318). Association for Computational Linguistics. Papineni, K., Roukos, S., Ward, T., & Zhu, W.-J. (2002). BLEU: A method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting of the Association for Computational Linguistics, Philadelphia, Pennsylvania, USA (pp. 311–318). Association for Computational Linguistics.
Zurück zum Zitat Qi, W., Yan, Y., Gong, Y., Liu, D., Duan, N., Chen, J., Zhang, R., & Zhou, M. (2020). ProphetNet: Predicting future n-gram for sequence-to-sequence pre-training. In Proceedings of the 2020 conference on empirical methods in natural language processing: Findings, EMNLP 2020, online event, November 16–20, 2020 (pp. 2401–2410). Association for Computational Linguistics. Qi, W., Yan, Y., Gong, Y., Liu, D., Duan, N., Chen, J., Zhang, R., & Zhou, M. (2020). ProphetNet: Predicting future n-gram for sequence-to-sequence pre-training. In Proceedings of the 2020 conference on empirical methods in natural language processing: Findings, EMNLP 2020, online event, November 16–20, 2020 (pp. 2401–2410). Association for Computational Linguistics.
Zurück zum Zitat Qu, C., Yang, L., Chen, C., Qiu, M., Croft, W. B., & Iyyer, M. (2020). Open-retrieval conversational question answering. In Proceedings of the 43rd international ACM SIGIR conference on research and development in information retrieval, SIGIR ’20, New York, NY, USA (pp. 539–548). Association for Computing Machinery. Qu, C., Yang, L., Chen, C., Qiu, M., Croft, W. B., & Iyyer, M. (2020). Open-retrieval conversational question answering. In Proceedings of the 43rd international ACM SIGIR conference on research and development in information retrieval, SIGIR ’20, New York, NY, USA (pp. 539–548). Association for Computing Machinery.
Zurück zum Zitat Qu, C., Yang, L., Qiu, M., Croft, W. B., Zhang, Y., & Iyyer, M. (2019a). BERT with history answer embedding for conversational question answering. In B. Piwowarski, M. Chevalier, É. Gaussier, Y. Maarek, J. Nie & F. Scholer (Eds.), Proceedings of the 42nd international ACM SIGIR conference on research and development in information retrieval, SIGIR 2019, Paris, France, July 21–25, 2019 (pp. 1133–1136). ACM. Qu, C., Yang, L., Qiu, M., Croft, W. B., Zhang, Y., & Iyyer, M. (2019a). BERT with history answer embedding for conversational question answering. In B. Piwowarski, M. Chevalier, É. Gaussier, Y. Maarek, J. Nie & F. Scholer (Eds.), Proceedings of the 42nd international ACM SIGIR conference on research and development in information retrieval, SIGIR 2019, Paris, France, July 21–25, 2019 (pp. 1133–1136). ACM.
Zurück zum Zitat Qu, C., Yang, L., Qiu, M., Zhang, Y., Chen, C., Croft, W. B., & Iyyer, M. (2019b). Attentive history selection for conversational question answering. In W. Zhu, D. Tao, X. Cheng, P. Cui, E. A. Rundensteiner, D. Carmel, Q. He & J. X. Yu (Eds.), Proceedings of the 28th ACM international conference on information and knowledge management, CIKM 2019, Beijing, China, November 3–7, 2019 (pp. 1391–1400). ACM. Qu, C., Yang, L., Qiu, M., Zhang, Y., Chen, C., Croft, W. B., & Iyyer, M. (2019b). Attentive history selection for conversational question answering. In W. Zhu, D. Tao, X. Cheng, P. Cui, E. A. Rundensteiner, D. Carmel, Q. He & J. X. Yu (Eds.), Proceedings of the 28th ACM international conference on information and knowledge management, CIKM 2019, Beijing, China, November 3–7, 2019 (pp. 1391–1400). ACM.
Zurück zum Zitat Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., & Sutskever, I. (2019). Language models are unsupervised multitask learners. OpenAI Blog, 1(8), 9. Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., & Sutskever, I. (2019). Language models are unsupervised multitask learners. OpenAI Blog, 1(8), 9.
Zurück zum Zitat Raffel, C., Shazeer, N., Roberts, A., Lee, K., Narang, S., Matena, M., et al. (2020). Exploring the limits of transfer learning with a unified text-to-text transformer. Journal of Machine Learning Research, 21, 140.1–140.67.MathSciNetMATH Raffel, C., Shazeer, N., Roberts, A., Lee, K., Narang, S., Matena, M., et al. (2020). Exploring the limits of transfer learning with a unified text-to-text transformer. Journal of Machine Learning Research, 21, 140.1–140.67.MathSciNetMATH
Zurück zum Zitat Robertson, S., & Zaragoza, H. (2009). The probabilistic relevance framework: BM25 and beyond. Foundations and Trends in Information Retrieval, 3(4), 333–389.CrossRef Robertson, S., & Zaragoza, H. (2009). The probabilistic relevance framework: BM25 and beyond. Foundations and Trends in Information Retrieval, 3(4), 333–389.CrossRef
Zurück zum Zitat Song, Y., Li, C.-T., Nie, J.-Y., Zhang, M., Zhao, D., & Yan, R. (2018). An ensemble of retrieval-based and generation-based human–computer conversation systems. In Proceedings of the twenty-seventh international joint conference on artificial intelligence, international joint conferences on artificial intelligence organization, IJCAI-18 (pp. 4382–4388). Song, Y., Li, C.-T., Nie, J.-Y., Zhang, M., Zhao, D., & Yan, R. (2018). An ensemble of retrieval-based and generation-based human–computer conversation systems. In Proceedings of the twenty-seventh international joint conference on artificial intelligence, international joint conferences on artificial intelligence organization, IJCAI-18 (pp. 4382–4388).
Zurück zum Zitat Tian, Z., Bi, W., Li, X., & Zhang, N. L. (2019). Learning to abstract for memory-augmented conversational response generation. In Proceedings of the 57th annual meeting of the Association for Computational Linguistics, Florence, Italy (pp. 3816–3825). Association for Computational Linguistics. Tian, Z., Bi, W., Li, X., & Zhang, N. L. (2019). Learning to abstract for memory-augmented conversational response generation. In Proceedings of the 57th annual meeting of the Association for Computational Linguistics, Florence, Italy (pp. 3816–3825). Association for Computational Linguistics.
Zurück zum Zitat Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, L., & Polosukhin, I. (2017). Attention is all you need. In I. Guyon, U. von Luxburg, S. Bengio, H. M. Wallach, R. Fergus, S. V. N. Vishwanathan & R. Garnett (Eds.), Advances in neural information processing systems 30: Annual conference on neural information processing systems 2017, December 4–9, 2017, Long Beach, CA, USA (pp. 5998–6008). Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, L., & Polosukhin, I. (2017). Attention is all you need. In I. Guyon, U. von Luxburg, S. Bengio, H. M. Wallach, R. Fergus, S. V. N. Vishwanathan & R. Garnett (Eds.), Advances in neural information processing systems 30: Annual conference on neural information processing systems 2017, December 4–9, 2017, Long Beach, CA, USA (pp. 5998–6008).
Zurück zum Zitat Voskarides, N., Li, D., Ren, P., Kanoulas, E., & de Rijke, M. (2020). Query resolution for conversational search with limited supervision. Proceedings of the 43rd international ACM SIGIR conference on research and development in information retrieval. Voskarides, N., Li, D., Ren, P., Kanoulas, E., & de Rijke, M. (2020). Query resolution for conversational search with limited supervision. Proceedings of the 43rd international ACM SIGIR conference on research and development in information retrieval.
Zurück zum Zitat Vtyurina, A., Savenkov, D., Agichtein, E., & Clarke, C. L. A. (2017). Exploring conversational search with humans, assistants, and wizards. In Proceedings of the 2017 CHI conference extended abstracts on human factors in computing systems, CHI EA ’17, New York, NY, USA (pp. 2187–2193). Association for Computing Machinery. Vtyurina, A., Savenkov, D., Agichtein, E., & Clarke, C. L. A. (2017). Exploring conversational search with humans, assistants, and wizards. In Proceedings of the 2017 CHI conference extended abstracts on human factors in computing systems, CHI EA ’17, New York, NY, USA (pp. 2187–2193). Association for Computing Machinery.
Zurück zum Zitat Wang, A., Pruksachatkun, Y., Nangia, N., Singh, A., Michael, J., Hill, F., Levy, O., & Bowman, S. R. (2019). SuperGLUE: A stickier benchmark for general-purpose language understanding systems. In Advances in neural information processing systems 32: Annual conference on neural information processing systems 2019, NeurIPS 2019, December 8–14, 2019, Vancouver, BC, Canada (pp. 3261–3275). Wang, A., Pruksachatkun, Y., Nangia, N., Singh, A., Michael, J., Hill, F., Levy, O., & Bowman, S. R. (2019). SuperGLUE: A stickier benchmark for general-purpose language understanding systems. In Advances in neural information processing systems 32: Annual conference on neural information processing systems 2019, NeurIPS 2019, December 8–14, 2019, Vancouver, BC, Canada (pp. 3261–3275).
Zurück zum Zitat Wang, A., Singh, A., Michael, J., Hill, F., Levy, O., & Bowman, S. R. (2018). GLUE: A multi-task benchmark and analysis platform for natural language understanding. In Proceedings of the workshop: Analyzing and interpreting neural networks for NLP, BlackboxNLP@EMNLP 2018, Brussels, Belgium, November 1, 2018 (pp. 353–355). Association for Computational Linguistics. Wang, A., Singh, A., Michael, J., Hill, F., Levy, O., & Bowman, S. R. (2018). GLUE: A multi-task benchmark and analysis platform for natural language understanding. In Proceedings of the workshop: Analyzing and interpreting neural networks for NLP, BlackboxNLP@EMNLP 2018, Brussels, Belgium, November 1, 2018 (pp. 353–355). Association for Computational Linguistics.
Zurück zum Zitat Wolf, T., Debut, L., Sanh, V., Chaumond, J., Delangue, C., Moi, A., Cistac, P., Rault, T., Louf, R., Funtowicz, M., Davison, J., Shleifer, S., von Platen, P., Ma, C., Jernite, Y., Plu, J., Xu, C., Scao, T. L., Gugger, S., Drame, M., Lhoest, Q., & Rush, A. M. (2020). Transformers: State-of-the-art natural language processing. In Proceedings of the 2020 conference on empirical methods in natural language processing: system demonstrations, EMNLP 2020, demos, online November 16–20, 2020 (pp. 38–45). Association for Computational Linguistics. Wolf, T., Debut, L., Sanh, V., Chaumond, J., Delangue, C., Moi, A., Cistac, P., Rault, T., Louf, R., Funtowicz, M., Davison, J., Shleifer, S., von Platen, P., Ma, C., Jernite, Y., Plu, J., Xu, C., Scao, T. L., Gugger, S., Drame, M., Lhoest, Q., & Rush, A. M. (2020). Transformers: State-of-the-art natural language processing. In Proceedings of the 2020 conference on empirical methods in natural language processing: system demonstrations, EMNLP 2020, demos, online November 16–20, 2020 (pp. 38–45). Association for Computational Linguistics.
Zurück zum Zitat Xiong, C., Dai, Z., Callan, J., Liu, Z., & Power, R. (2017). End-to-end neural ad-hoc ranking with kernel pooling. In Proceedings of the 40th international ACM SIGIR conference on research and development in information retrieval, Shinjuku, Tokyo, Japan, August 7–11, 2017 (pp. 55–64). ACM. Xiong, C., Dai, Z., Callan, J., Liu, Z., & Power, R. (2017). End-to-end neural ad-hoc ranking with kernel pooling. In Proceedings of the 40th international ACM SIGIR conference on research and development in information retrieval, Shinjuku, Tokyo, Japan, August 7–11, 2017 (pp. 55–64). ACM.
Zurück zum Zitat Xiong, L., Xiong, C., Li, Y., Tang, K., Liu, J., Bennett, P. N., Ahmed, J., & Overwijk, A. (2021). Approximate nearest neighbor negative contrastive learning for dense text retrieval. In 9th International conference on learning representations, ICLR 2021, virtual event, Austria, May 3–7, 2021. OpenReview.net. Xiong, L., Xiong, C., Li, Y., Tang, K., Liu, J., Bennett, P. N., Ahmed, J., & Overwijk, A. (2021). Approximate nearest neighbor negative contrastive learning for dense text retrieval. In 9th International conference on learning representations, ICLR 2021, virtual event, Austria, May 3–7, 2021. OpenReview.net.
Zurück zum Zitat Yang, Z., Dai, Z., Yang, Y., Carbonell, J. G., Salakhutdinov, R., & Le, Q. V. (2019b). XLNet: Generalized autoregressive pretraining for language understanding. In Advances in neural information processing systems 32: Annual conference on neural information processing systems 2019, NeurIPS 2019, December 8–14, 2019, Vancouver, BC, Canada (pp. 5754–5764). Yang, Z., Dai, Z., Yang, Y., Carbonell, J. G., Salakhutdinov, R., & Le, Q. V. (2019b). XLNet: Generalized autoregressive pretraining for language understanding. In Advances in neural information processing systems 32: Annual conference on neural information processing systems 2019, NeurIPS 2019, December 8–14, 2019, Vancouver, BC, Canada (pp. 5754–5764).
Zurück zum Zitat Yang, P., Fang, H., & Lin, J. (2017). Anserini: Enabling the use of Lucene for information retrieval research. In N. Kando, T. Sakai, H. Joho, H. Li, A. P. de Vries & R. W. White (Eds.), Proceedings of the 40th international ACM SIGIR conference on research and development in information retrieval, Shinjuku, Tokyo, Japan, August 7–11, 2017 (pp. 1253–1256). ACM. Yang, P., Fang, H., & Lin, J. (2017). Anserini: Enabling the use of Lucene for information retrieval research. In N. Kando, T. Sakai, H. Joho, H. Li, A. P. de Vries & R. W. White (Eds.), Proceedings of the 40th international ACM SIGIR conference on research and development in information retrieval, Shinjuku, Tokyo, Japan, August 7–11, 2017 (pp. 1253–1256). ACM.
Zurück zum Zitat Yang, Z., Lan, Q., Guo, J., Fan, Y., Zhu, X., Lan, Y., Wang, Y., & Cheng, X. (2018). A deep top-k relevance matching model for ad hoc retrieval. In Information retrieval—24th China conference, CCIR 2018, proceedings, Guilin, China, September 27–29, 2018. Lecture notes in computer science (Vol. 11168, pp. 16–27). Springer. Yang, Z., Lan, Q., Guo, J., Fan, Y., Zhu, X., Lan, Y., Wang, Y., & Cheng, X. (2018). A deep top-k relevance matching model for ad hoc retrieval. In Information retrieval—24th China conference, CCIR 2018, proceedings, Guilin, China, September 27–29, 2018. Lecture notes in computer science (Vol. 11168, pp. 16–27). Springer.
Zurück zum Zitat Yang, J., Lin, S., Wang, C., Lin, J., & Tsai, M. (2019a). Query and answer expansion from conversation history. In E. M. Voorhees & A. Ellis (Eds.), Proceedings of the twenty-eighth Text REtrieval Conference, TREC 2019, Gaithersburg, Maryland, USA, November 13–15, 2019. NIST special publication (Vol. 1250). National Institute of Standards and Technology (NIST). Yang, J., Lin, S., Wang, C., Lin, J., & Tsai, M. (2019a). Query and answer expansion from conversation history. In E. M. Voorhees & A. Ellis (Eds.), Proceedings of the twenty-eighth Text REtrieval Conference, TREC 2019, Gaithersburg, Maryland, USA, November 13–15, 2019. NIST special publication (Vol. 1250). National Institute of Standards and Technology (NIST).
Zurück zum Zitat Yu, S., Liu, J., Yang, J., Xiong, C., Bennett, P. N., Gao, J., & Liu, Z. (2020). Few-shot generative conversational query rewriting. In Proceedings of the 43rd international ACM SIGIR conference on research and development in Information Retrieval, SIGIR 2020, virtual event, China, July 25–30, 2020 (pp. 1933–1936). ACM. Yu, S., Liu, J., Yang, J., Xiong, C., Bennett, P. N., Gao, J., & Liu, Z. (2020). Few-shot generative conversational query rewriting. In Proceedings of the 43rd international ACM SIGIR conference on research and development in Information Retrieval, SIGIR 2020, virtual event, China, July 25–30, 2020 (pp. 1933–1936). ACM.
Zurück zum Zitat Zhai, C., & Lafferty, J. (2001). A study of smoothing methods for language models applied to ad hoc information retrieval. In Proceedings of the 24th annual international ACM SIGIR conference on research and development in information retrieval, SIGIR’01, New York, NY, USA (pp. 334–342). Association for Computing Machinery. Zhai, C., & Lafferty, J. (2001). A study of smoothing methods for language models applied to ad hoc information retrieval. In Proceedings of the 24th annual international ACM SIGIR conference on research and development in information retrieval, SIGIR’01, New York, NY, USA (pp. 334–342). Association for Computing Machinery.
Zurück zum Zitat Zhang, J., Zhao, Y., Saleh, M., & Liu, P. J. (2020). PEGASUS: Pre-training with extracted gap-sentences for abstractive summarization. In Proceedings of the 37th international conference on machine learning, ICML 2020, virtual event, 13–18 July 2020. Proceedings of machine learning research (Vol. 119, pp. 11328–11339). PMLR. Zhang, J., Zhao, Y., Saleh, M., & Liu, P. J. (2020). PEGASUS: Pre-training with extracted gap-sentences for abstractive summarization. In Proceedings of the 37th international conference on machine learning, ICML 2020, virtual event, 13–18 July 2020. Proceedings of machine learning research (Vol. 119, pp. 11328–11339). PMLR.
Zurück zum Zitat Zhuang, Y., Wang, X., Zhang, H., Xie, J., & Zhu, X. (2017). An ensemble approach to conversation generation. In X. Huang, J. Jiang, D. Zhao, Y. Feng & Y. Hong (Eds.), Natural language processing and Chinese computing—6th CCF international conference, NLPCC 2017, proceedings, Dalian, China, November 8–12, 2017. Lecture notes in computer science (Vol. 10619, pp. 51–62). Springer. Zhuang, Y., Wang, X., Zhang, H., Xie, J., & Zhu, X. (2017). An ensemble approach to conversation generation. In X. Huang, J. Jiang, D. Zhao, Y. Feng & Y. Hong (Eds.), Natural language processing and Chinese computing—6th CCF international conference, NLPCC 2017, proceedings, Dalian, China, November 8–12, 2017. Lecture notes in computer science (Vol. 10619, pp. 51–62). Springer.
Metadaten
Titel
Open-domain conversational search assistants: the Transformer is all you need
verfasst von
Rafael Ferreira
Mariana Leite
David Semedo
Joao Magalhaes
Publikationsdatum
14.03.2022
Verlag
Springer Netherlands
Erschienen in
Discover Computing / Ausgabe 2/2022
Print ISSN: 2948-2984
Elektronische ISSN: 2948-2992
DOI
https://doi.org/10.1007/s10791-022-09403-0

Weitere Artikel der Ausgabe 2/2022

Discover Computing 2/2022 Zur Ausgabe

Guest Editorial: Special Issue on ECIR 2021

Guest editorial: special issue on ECIR 2021

Premium Partner