Skip to main content
Erschienen in: World Wide Web 5/2023

27.07.2023

Knowledge-aware response selection with semantics underlying multi-turn open-domain conversations

verfasst von: Makoto Nakatsuji, Yuka Ozeki, Shuhei Tateishi, Yoshihisa Kano, QingPeng Zhang

Erschienen in: World Wide Web | Ausgabe 5/2023

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Response selection is a critical issue in the AI community, with important applications on the Web. The accuracy of the selected responses, however, tends to be insufficient due to the lack of contextual awareness, especially in open-domain conversations where words tend to have several meanings in different contexts. Our solution, SemSol, is a knowledge-aware response selection model that tackles this problem by utilizing the context-specific semantics behind words that are implicitly shared among users throughout the dialogue. SemSol simultaneously learns word sense disambiguations (WSD) for the words in the dialogue on the basis of an open-domain knowledge graph, i.e. WordNet, while learning the match between the context and the response candidates. Then, SemSol improves the accuracy of the response by exploiting the semantic information in a knowledge graph in accordance with the dialogue context. Our model learns the topics of utterances in the context of the whole training dataset. This topic-level knowledge can provide topic-specific information in the dialogue context. This improves the WSDs and the response selection accuracy. Experiments with two open-domain conversational datasets, Douban (Chinese) and Reddit (English), demonstrated that the SemSol model outperformed state-of-the-art baselines. SemSol is ranked #1 on the Douban leaderboard.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Hardalov, M., Koychev, I., Nakov, P.: Enriched pre-trained transformers for joint slot filling and intent detection. CoRRarXiv:2004.14848 (2020) Hardalov, M., Koychev, I., Nakov, P.: Enriched pre-trained transformers for joint slot filling and intent detection. CoRRarXiv:​2004.​14848 (2020)
2.
Zurück zum Zitat Rastogi, A., Zang, X., Sunkara, S., Gupta, R., Khaitan, P.: Towards scalable multi-domain conversational agents: The schema-guided dialogue dataset. In: Proc. AAAI’20, pp. 8689–8696 (2020) Rastogi, A., Zang, X., Sunkara, S., Gupta, R., Khaitan, P.: Towards scalable multi-domain conversational agents: The schema-guided dialogue dataset. In: Proc. AAAI’20, pp. 8689–8696 (2020)
3.
Zurück zum Zitat Wang, J., Liu, J., Bi, W., Liu, X., He, K., Xu, R., Yang, M.: Improving knowledge-aware dialogue generation via knowledge base question answering. In: Proc. AAAI’20, pp. 9169–9176 (2020) Wang, J., Liu, J., Bi, W., Liu, X., He, K., Xu, R., Yang, M.: Improving knowledge-aware dialogue generation via knowledge base question answering. In: Proc. AAAI’20, pp. 9169–9176 (2020)
4.
Zurück zum Zitat Henderson, M., Casanueva, I., Mrkšić, N., Su, P.-H., Wen, T.-H., Vulić, I.: ConveRT: Efficient and accurate conversational representations from transformers. In: Proc. EMNLP’20, pp. 2161–2174 (2020) Henderson, M., Casanueva, I., Mrkšić, N., Su, P.-H., Wen, T.-H., Vulić, I.: ConveRT: Efficient and accurate conversational representations from transformers. In: Proc. EMNLP’20, pp. 2161–2174 (2020)
5.
Zurück zum Zitat Henderson, M., Vulić, I., Gerz, D., Casanueva, I., Budzianowski, P., Coope, S., Spithourakis, G., Wen, T.-H., Mrkšić, N., Su, P.-H.: Training neural response selection for task-oriented dialogue systems. In: Proc. ACL’19, pp. 5392–5404 (2019) Henderson, M., Vulić, I., Gerz, D., Casanueva, I., Budzianowski, P., Coope, S., Spithourakis, G., Wen, T.-H., Mrkšić, N., Su, P.-H.: Training neural response selection for task-oriented dialogue systems. In: Proc. ACL’19, pp. 5392–5404 (2019)
6.
Zurück zum Zitat Whang, T., Lee, D., Lee, C. Yang, K., Oh, D., Lim, H.: An effective domain adaptive post-training method for BERT in response selection. In: Proc. Interspeech’20, pp. 1585–1589 (2020) Whang, T., Lee, D., Lee, C. Yang, K., Oh, D., Lim, H.: An effective domain adaptive post-training method for BERT in response selection. In: Proc. Interspeech’20, pp. 1585–1589 (2020)
7.
Zurück zum Zitat Abd-alrazaq, A.A., Alajlani, M., Alalwan, A.A., Bewick, B.M., Gardner, P., Househ, M.: An overview of the features of chatbots in mental health: A scoping review. Int. J. Med. Inform. 132, 103978 (2019)CrossRef Abd-alrazaq, A.A., Alajlani, M., Alalwan, A.A., Bewick, B.M., Gardner, P., Househ, M.: An overview of the features of chatbots in mental health: A scoping review. Int. J. Med. Inform. 132, 103978 (2019)CrossRef
8.
Zurück zum Zitat Zhang, Z., Li, J., Zhu, P, Zhao, H., Liu, G.: Modeling multi-turn conversation with deep utterance aggregation. CoRRarXiv:1806.09102 (2018) Zhang, Z., Li, J., Zhu, P, Zhao, H., Liu, G.: Modeling multi-turn conversation with deep utterance aggregation. CoRRarXiv:​1806.​09102 (2018)
9.
Zurück zum Zitat Lowe, R., Pow, N., Serban, I., Pineau, J.: The ubuntu dialogue corpus: A large dataset for research in unstructured multi-turn dialogue systems. CoRRarXiv:1506.08909 (2015) Lowe, R., Pow, N., Serban, I., Pineau, J.: The ubuntu dialogue corpus: A large dataset for research in unstructured multi-turn dialogue systems. CoRRarXiv:​1506.​08909 (2015)
10.
Zurück zum Zitat Chen, L., Zhao, Y., Lyu, B., Jin, L., Chen, Z., Zhu, S., Yu, K.: Neural graph matching networks for chinese short text matching. In: Proc. ACL’20, pp. 6152–6158 (2020) Chen, L., Zhao, Y., Lyu, B., Jin, L., Chen, Z., Zhu, S., Yu, K.: Neural graph matching networks for chinese short text matching. In: Proc. ACL’20, pp. 6152–6158 (2020)
11.
Zurück zum Zitat Tao, C., Wu, W., Xu, C., Hu, W., Zhao, D., Yan, R.: One time of interaction may not be enough: Go deep with an interaction-over-interaction network for response selection in dialogues. In: Proc. ACL’19, pp. 1–11. Association for Computational Linguistics (2019) Tao, C., Wu, W., Xu, C., Hu, W., Zhao, D., Yan, R.: One time of interaction may not be enough: Go deep with an interaction-over-interaction network for response selection in dialogues. In: Proc. ACL’19, pp. 1–11. Association for Computational Linguistics (2019)
12.
Zurück zum Zitat Wu, Y., Wu, W., Xing, C., Xu, C., Li, Z., Zhou, M.: A sequential matching framework for multi-turn response selection in retrieval-based chatbots. Comput. Linguistics 45(1), 163–197 (2019)MathSciNetCrossRef Wu, Y., Wu, W., Xing, C., Xu, C., Li, Z., Zhou, M.: A sequential matching framework for multi-turn response selection in retrieval-based chatbots. Comput. Linguistics 45(1), 163–197 (2019)MathSciNetCrossRef
13.
Zurück zum Zitat Xu, Y., Zhao, H., Zhang, Z.: Topic-aware multi-turn dialogue modeling. In: Proc. AAAI’21, pp 14176–14184 (2021) Xu, Y., Zhao, H., Zhang, Z.: Topic-aware multi-turn dialogue modeling. In: Proc. AAAI’21, pp 14176–14184 (2021)
14.
Zurück zum Zitat Han, J., Hong, T., Kim, B., Ko, Y., Seo, J.: Fine-grained post-training for improving retrieval-based dialogue systems. In: Proc. NAACL-HLT’21, pp. 1549–1558 (2021) Han, J., Hong, T., Kim, B., Ko, Y., Seo, J.: Fine-grained post-training for improving retrieval-based dialogue systems. In: Proc. NAACL-HLT’21, pp. 1549–1558 (2021)
15.
Zurück zum Zitat Zhang, Z., Zhao, H.: Structural pre-training for dialogue comprehension. In: Proc. ACL’21, pp. 5134–5145 (2021) Zhang, Z., Zhao, H.: Structural pre-training for dialogue comprehension. In: Proc. ACL’21, pp. 5134–5145 (2021)
16.
Zurück zum Zitat Whang, T., Lee, D., Oh, D., Lee, C., Han, K., Lee, D.-h., Lee, S.: Do response selection models really know what’s next? utterance manipulation strategies for multi-turn response selection. In: Proc. AAAI’21, pp. 14041–14049 (2021) Whang, T., Lee, D., Oh, D., Lee, C., Han, K., Lee, D.-h., Lee, S.: Do response selection models really know what’s next? utterance manipulation strategies for multi-turn response selection. In: Proc. AAAI’21, pp. 14041–14049 (2021)
17.
Zurück zum Zitat Su, Y., Cai, D., Zhou, Q., Lin, Z., Baker, S., Cao, Y., Shi, S., Collier, N., Wang, Y.: Dialogue response selection with hierarchical curriculum learning. In: Proc. ACL’21, pp. 1740–1751 (2021) Su, Y., Cai, D., Zhou, Q., Lin, Z., Baker, S., Cao, Y., Shi, S., Collier, N., Wang, Y.: Dialogue response selection with hierarchical curriculum learning. In: Proc. ACL’21, pp. 1740–1751 (2021)
18.
Zurück zum Zitat Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, Ł., Polosukhin, I.: Attention is all you need. In: Proc. NIPS’17, vol. 30 (2017) Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, Ł., Polosukhin, I.: Attention is all you need. In: Proc. NIPS’17, vol. 30 (2017)
19.
Zurück zum Zitat Devlin, J., Chang, M.-W., Lee, K., Toutanova, K.: BERT: Pre-training of deep bidirectional transformers for language understanding. In: Proc. NAACL-HLT’19, pp. 4171–4186 (2019) Devlin, J., Chang, M.-W., Lee, K., Toutanova, K.: BERT: Pre-training of deep bidirectional transformers for language understanding. In: Proc. NAACL-HLT’19, pp. 4171–4186 (2019)
20.
Zurück zum Zitat Miller, G.A.: Wordnet: A lexical database for english. Commun. ACM 38(11), 39–41 (1995)CrossRef Miller, G.A.: Wordnet: A lexical database for english. Commun. ACM 38(11), 39–41 (1995)CrossRef
21.
Zurück zum Zitat Wang, S., Bond, F.: Building the Chinese open Wordnet (COW): Starting from core synsets. In: Proceedings of the 11th Workshop on Asian Language Resources, pp. 10–18 (2013) Wang, S., Bond, F.: Building the Chinese open Wordnet (COW): Starting from core synsets. In: Proceedings of the 11th Workshop on Asian Language Resources, pp. 10–18 (2013)
22.
Zurück zum Zitat Wu, Y., Wu, W., Xing, C., Zhou, M., Li, Z.: Sequential matching network: A new architecture for multi-turn response selection in retrieval-based chatbots. In: Proc. ACL’17, pp. 496–505 (2017) Wu, Y., Wu, W., Xing, C., Zhou, M., Li, Z.: Sequential matching network: A new architecture for multi-turn response selection in retrieval-based chatbots. In: Proc. ACL’17, pp. 496–505 (2017)
23.
Zurück zum Zitat Zhang, H., Lan, Y., Pang, L., Chen, H., Ding, Z., Yin, D.: Modeling topical relevance for multi-turn dialogue generation. In: Proc. IJCAI’20, pp. 3737–3743 (2020) Zhang, H., Lan, Y., Pang, L., Chen, H., Ding, Z., Yin, D.: Modeling topical relevance for multi-turn dialogue generation. In: Proc. IJCAI’20, pp. 3737–3743 (2020)
24.
Zurück zum Zitat Nakatsuji, M., Okui, S.: Conclusion-supplement answer generation for non-factoid questions. In: Proc. AAAI’20, pp. 8520–8527 (2020) Nakatsuji, M., Okui, S.: Conclusion-supplement answer generation for non-factoid questions. In: Proc. AAAI’20, pp. 8520–8527 (2020)
25.
Zurück zum Zitat Zhang, Z., Han, X., Liu, Z., Jiang, X., Sun, M., Liu, Q.: ERNIE: Enhanced language representation with informative entities. In: Proc. ACL’19, pp. 1441–1451 (2019) Zhang, Z., Han, X., Liu, Z., Jiang, X., Sun, M., Liu, Q.: ERNIE: Enhanced language representation with informative entities. In: Proc. ACL’19, pp. 1441–1451 (2019)
26.
Zurück zum Zitat Bordes, A., Usunier, N., García-Durán, A., Weston, J., Yakhnenko, O.: Translating embeddings for modeling multi-relational data. In: Proc. NIPS’13, pp. 2787–2795 (2013) Bordes, A., Usunier, N., García-Durán, A., Weston, J., Yakhnenko, O.: Translating embeddings for modeling multi-relational data. In: Proc. NIPS’13, pp. 2787–2795 (2013)
27.
Zurück zum Zitat Liu, Z., Patwary, M., Prenger, R., Prabhumoye, S., Ping, W., Shoeybi, M., Catanzaro, B.: Multi-stage prompting for knowledgeable dialogue generation. In: Proc. ACL’22, pp. 1317–1337 (2022) Liu, Z., Patwary, M., Prenger, R., Prabhumoye, S., Ping, W., Shoeybi, M., Catanzaro, B.: Multi-stage prompting for knowledgeable dialogue generation. In: Proc. ACL’22, pp. 1317–1337 (2022)
28.
Zurück zum Zitat Liu, J., Liu, A., Lu, X., Welleck, S., West, P., Le Bras, R., Choi, Y., Hajishirzi, H.: Generated knowledge prompting for commonsense reasoning. In: Proc. ACL’22, pp. 3154–3169 (2022) Liu, J., Liu, A., Lu, X., Welleck, S., West, P., Le Bras, R., Choi, Y., Hajishirzi, H.: Generated knowledge prompting for commonsense reasoning. In: Proc. ACL’22, pp. 3154–3169 (2022)
29.
Zurück zum Zitat Ouyang, L., Wu, J., Jiang, X., Almeida, D., Wainwright, C.L., Mishkin, P., Zhang, C., Agarwal, S, Slama, K., Ray, A., Schulman, J., Hilton, J., Kelton, F., Miller, L., Simens, M., Askell, A., Welinder, P., Christiano, P., Leike, J., Lowe, R.: Training language models to follow instructions with human feedback (2022) Ouyang, L., Wu, J., Jiang, X., Almeida, D., Wainwright, C.L., Mishkin, P., Zhang, C., Agarwal, S, Slama, K., Ray, A., Schulman, J., Hilton, J., Kelton, F., Miller, L., Simens, M., Askell, A., Welinder, P., Christiano, P., Leike, J., Lowe, R.: Training language models to follow instructions with human feedback (2022)
30.
Zurück zum Zitat Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J.D., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., Agarwal, S., Herbert-Voss, A., Krueger, G., Henighan, T., Child, R., Ramesh, A., Ziegler, D., Wu, J., Winter, C., Hesse, C., Chen, M., Sigler, E., Litwin, M., Gray, S., Chess, B., Clark, J., Berner, C., McCandlish, S., Radford, A., Sutskever, I., Amodei, D.: Language models are few-shot learners. In: Proc. NIPS’20, vol. 33, pp. 1877–1901 (2020) Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J.D., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., Agarwal, S., Herbert-Voss, A., Krueger, G., Henighan, T., Child, R., Ramesh, A., Ziegler, D., Wu, J., Winter, C., Hesse, C., Chen, M., Sigler, E., Litwin, M., Gray, S., Chess, B., Clark, J., Berner, C., McCandlish, S., Radford, A., Sutskever, I., Amodei, D.: Language models are few-shot learners. In: Proc. NIPS’20, vol. 33, pp. 1877–1901 (2020)
31.
Zurück zum Zitat Nakatsuji, M., Fujiwara, Y.: Linked taxonomies to capture users’ subjective assessments of items to facilitate accurate collaborative filtering. Artif. Intell. 207, 52–68 (2014)MathSciNetCrossRef Nakatsuji, M., Fujiwara, Y.: Linked taxonomies to capture users’ subjective assessments of items to facilitate accurate collaborative filtering. Artif. Intell. 207, 52–68 (2014)MathSciNetCrossRef
32.
Zurück zum Zitat Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., Levy, O., Lewis, M., Zettlemoyer, L., Stoyanov, V.: Roberta: A robustly optimized BERT pretraining approach. CoRR, arXiv:1907.11692 (2019) Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., Levy, O., Lewis, M., Zettlemoyer, L., Stoyanov, V.: Roberta: A robustly optimized BERT pretraining approach. CoRR, arXiv:​1907.​11692 (2019)
33.
Zurück zum Zitat Nakatsuji, M., Toda, H., Sawada, H., Zheng, J., Hendler, J.A.: Semantic sensitive tensor factorization. Artif. Intell. 230, 224–245 (2016)MathSciNetCrossRefMATH Nakatsuji, M., Toda, H., Sawada, H., Zheng, J., Hendler, J.A.: Semantic sensitive tensor factorization. Artif. Intell. 230, 224–245 (2016)MathSciNetCrossRefMATH
34.
Zurück zum Zitat Ethayarajh, K.: How contextual are contextualized word representations? Comparing the geometry of BERT, ELMo, and GPT-2 embeddings. In: EMNLP-IJCNLP’19, pp. 55–65 (2019) Ethayarajh, K.: How contextual are contextualized word representations? Comparing the geometry of BERT, ELMo, and GPT-2 embeddings. In: EMNLP-IJCNLP’19, pp. 55–65 (2019)
Metadaten
Titel
Knowledge-aware response selection with semantics underlying multi-turn open-domain conversations
verfasst von
Makoto Nakatsuji
Yuka Ozeki
Shuhei Tateishi
Yoshihisa Kano
QingPeng Zhang
Publikationsdatum
27.07.2023
Verlag
Springer US
Erschienen in
World Wide Web / Ausgabe 5/2023
Print ISSN: 1386-145X
Elektronische ISSN: 1573-1413
DOI
https://doi.org/10.1007/s11280-023-01164-0

Weitere Artikel der Ausgabe 5/2023

World Wide Web 5/2023 Zur Ausgabe

Premium Partner