Skip to main content

2018 | OriginalPaper | Buchkapitel

Multi-turn Inference Matching Network for Natural Language Inference

verfasst von : Chunhua Liu, Shan Jiang, Hainan Yu, Dong Yu

Erschienen in: Natural Language Processing and Chinese Computing

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Natural Language Inference (NLI) is a fundamental and challenging task in Natural Language Processing (NLP). Most existing methods only apply one-pass inference process on a mixed matching feature, which is a concatenation of different matching features between a premise and a hypothesis. In this paper, we propose a new model called Multi-turn Inference Matching Network (MIMN) to perform multi-turn inference on different matching features. In each turn, the model focuses on one particular matching feature instead of the mixed matching feature. To enhance the interaction between different matching features, a memory component is employed to store the history inference information. The inference of each turn is performed on the current matching feature and the memory. We conduct experiments on three different NLI datasets. The experimental results show that our model outperforms or achieves the state-of-the-art performance on all the three datasets.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Abadi, M., et al.: Tensorflow: a system for large-scale machine learning (2016) Abadi, M., et al.: Tensorflow: a system for large-scale machine learning (2016)
2.
Zurück zum Zitat Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning to align and translate. CoRR,abs/1409.0473, September 2014 Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning to align and translate. CoRR,abs/1409.0473, September 2014
3.
Zurück zum Zitat Bowman, S.R., Angeli, G., Potts, C., Manning, C.D.: A large annotated corpus for learning natural language inference. In: EMNLP (2015) Bowman, S.R., Angeli, G., Potts, C., Manning, C.D.: A large annotated corpus for learning natural language inference. In: EMNLP (2015)
4.
Zurück zum Zitat Chen, Q., Zhu, X.D., Ling, Z.H., Wei, S., Jiang, H., Inkpen, D.: Recurrent neural network-based sentence encoder with gated attention for natural language inference. In: EMNLP (2017) Chen, Q., Zhu, X.D., Ling, Z.H., Wei, S., Jiang, H., Inkpen, D.: Recurrent neural network-based sentence encoder with gated attention for natural language inference. In: EMNLP (2017)
5.
Zurück zum Zitat Chen, Q., Zhu, X., Ling, Z.H., Wei, S., Jiang, H., Inkpen, D.: Enhanced LSTM for natural language inference. In: ACL, Vancouver, pp. 1657–1668, July 2017 Chen, Q., Zhu, X., Ling, Z.H., Wei, S., Jiang, H., Inkpen, D.: Enhanced LSTM for natural language inference. In: ACL, Vancouver, pp. 1657–1668, July 2017
6.
Zurück zum Zitat Cheng, J., Dong, L., Lapata, M.: Long short-term memory-networks for machine reading. In: EMNLP, Austin, pp. 551–561, November 2016 Cheng, J., Dong, L., Lapata, M.: Long short-term memory-networks for machine reading. In: EMNLP, Austin, pp. 551–561, November 2016
7.
Zurück zum Zitat Dagan, I., Glickman, O., Magnini, B.: The PASCAL recognising textual entailment challenge. In: Quiñonero-Candela, J., Dagan, I., Magnini, B., d’Alché-Buc, F. (eds.) MLCW 2005. LNCS (LNAI), vol. 3944, pp. 177–190. Springer, Heidelberg (2006). https://doi.org/10.1007/11736790_9CrossRef Dagan, I., Glickman, O., Magnini, B.: The PASCAL recognising textual entailment challenge. In: Quiñonero-Candela, J., Dagan, I., Magnini, B., d’Alché-Buc, F. (eds.) MLCW 2005. LNCS (LNAI), vol. 3944, pp. 177–190. Springer, Heidelberg (2006). https://​doi.​org/​10.​1007/​11736790_​9CrossRef
8.
Zurück zum Zitat Ghaeini, R., et al.: DR-BiLSTM: dependent reading bidirectional LSTM for natural language inference. arXiv preprint arXiv:1802.05577 (2018) Ghaeini, R., et al.: DR-BiLSTM: dependent reading bidirectional LSTM for natural language inference. arXiv preprint arXiv:​1802.​05577 (2018)
9.
Zurück zum Zitat Glickman, O., Dagan, I.: A probabilistic setting and lexical cooccurrence model for textual entailment. In: Proceedings of the ACL Workshop on Empirical Modeling of Semantic Equivalence and Entailment, EMSEE 2005 (2005) Glickman, O., Dagan, I.: A probabilistic setting and lexical cooccurrence model for textual entailment. In: Proceedings of the ACL Workshop on Empirical Modeling of Semantic Equivalence and Entailment, EMSEE 2005 (2005)
10.
Zurück zum Zitat Gong, Y., Luo, H., Zhang, J.: Natural language inference over interaction space. In: ICLR (2018) Gong, Y., Luo, H., Zhang, J.: Natural language inference over interaction space. In: ICLR (2018)
11.
Zurück zum Zitat Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9, 1735–1780 (1997)CrossRef Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9, 1735–1780 (1997)CrossRef
12.
Zurück zum Zitat Khot, T., Sabharwal, A., Clark, P.: SciTail: a textual entailment dataset from science question answering. In: AAAI (2018) Khot, T., Sabharwal, A., Clark, P.: SciTail: a textual entailment dataset from science question answering. In: AAAI (2018)
13.
Zurück zum Zitat Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. CoRR,abs/1412.6980 (2014) Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. CoRR,abs/1412.6980 (2014)
14.
Zurück zum Zitat Lai, A., Bisk, Y., Hockenmaier, J.: Natural language inference from multiple premises. In: IJCNLP, pp. 100–109 (2017) Lai, A., Bisk, Y., Hockenmaier, J.: Natural language inference from multiple premises. In: IJCNLP, pp. 100–109 (2017)
15.
Zurück zum Zitat Liu, P., Qiu, X., Chen, J., Huang, X.: Deep fusion LSTMs for text semantic matching. In: ACL, pp. 1034–1043 (2016) Liu, P., Qiu, X., Chen, J., Huang, X.: Deep fusion LSTMs for text semantic matching. In: ACL, pp. 1034–1043 (2016)
16.
Zurück zum Zitat Liu, P., Qiu, X., Zhou, Y., Chen, J., Huang, X.: Modelling interaction of sentence pair with coupled-LSTMs. In: EMNLP (2016) Liu, P., Qiu, X., Zhou, Y., Chen, J., Huang, X.: Modelling interaction of sentence pair with coupled-LSTMs. In: EMNLP (2016)
17.
Zurück zum Zitat Liu, Y., Sun, C., Lin, L., Wang, X.: Learning natural language inference using bidirectional LSTM model and inner-attention. CoRR,abs/1605.09090 (2016) Liu, Y., Sun, C., Lin, L., Wang, X.: Learning natural language inference using bidirectional LSTM model and inner-attention. CoRR,abs/1605.09090 (2016)
18.
Zurück zum Zitat Luong, T., Pham, H., Manning, C.D.: Effective approaches to attention-based neural machine translation. In: EMNLP, pp. 1412–1421 (2015) Luong, T., Pham, H., Manning, C.D.: Effective approaches to attention-based neural machine translation. In: EMNLP, pp. 1412–1421 (2015)
19.
Zurück zum Zitat MacCartney, B., Galley, M., Manning, C.D.: A phrase-based alignment model for natural language inference. In: EMNLP 2008 (2008) MacCartney, B., Galley, M., Manning, C.D.: A phrase-based alignment model for natural language inference. In: EMNLP 2008 (2008)
20.
Zurück zum Zitat Marelli, M., Menini, S., Baroni, M., Bentivogli, L., bernardi, R., Zamparelli, R.: A sick cure for the evaluation of compositional distributional semantic models. In: LREC (2014) Marelli, M., Menini, S., Baroni, M., Bentivogli, L., bernardi, R., Zamparelli, R.: A sick cure for the evaluation of compositional distributional semantic models. In: LREC (2014)
21.
Zurück zum Zitat McCann, B., Bradbury, J., Xiong, C., Socher, R.: Learned in translation: contextualized word vectors. arXiv preprint arXiv:1708.00107 (2017) McCann, B., Bradbury, J., Xiong, C., Socher, R.: Learned in translation: contextualized word vectors. arXiv preprint arXiv:​1708.​00107 (2017)
22.
Zurück zum Zitat Mou, L., et al.: Natural language inference by tree-based convolution and heuristic matching. In: ACL (2016) Mou, L., et al.: Natural language inference by tree-based convolution and heuristic matching. In: ACL (2016)
23.
Zurück zum Zitat Munkhdalai, T., Yu, H.: Neural semantic encoders. In: EACL, pp. 397–407 (2016) Munkhdalai, T., Yu, H.: Neural semantic encoders. In: EACL, pp. 397–407 (2016)
24.
Zurück zum Zitat Nie, Y., Bansal, M.: Shortcut-stacked sentence encoders for multi-domain inference. In: RepEval@EMNLP (2017) Nie, Y., Bansal, M.: Shortcut-stacked sentence encoders for multi-domain inference. In: RepEval@EMNLP (2017)
25.
Zurück zum Zitat Parikh, A., Täckström, O., Das, D., Uszkoreit, J.: A decomposable attention model for natural language inference. In: EMNLP, Austin, pp. 2249–2255, November 2016 Parikh, A., Täckström, O., Das, D., Uszkoreit, J.: A decomposable attention model for natural language inference. In: EMNLP, Austin, pp. 2249–2255, November 2016
26.
Zurück zum Zitat Pennington, J., Socher, R., Manning, C.: Glove: global vectors for word representation. In: EMNLP (2014) Pennington, J., Socher, R., Manning, C.: Glove: global vectors for word representation. In: EMNLP (2014)
27.
Zurück zum Zitat Rocktäschel, T., Grefenstette, E., Hermann, K.M., Kocisky, T., Blunsom, P.: Reasoning about entailment with neural attention. In: ICLR (2016) Rocktäschel, T., Grefenstette, E., Hermann, K.M., Kocisky, T., Blunsom, P.: Reasoning about entailment with neural attention. In: ICLR (2016)
28.
Zurück zum Zitat Sha, L., Chang, B., Sui, Z., Li, S.: Reading and thinking: re-read LSTM unit for textual entailment recognition. In: COLING, pp. 2870–2879, December 2016 Sha, L., Chang, B., Sui, Z., Li, S.: Reading and thinking: re-read LSTM unit for textual entailment recognition. In: COLING, pp. 2870–2879, December 2016
29.
Zurück zum Zitat Shen, T., Zhou, T., Long, G., Jiang, J., Pan, S., Zhang, C.: DiSAN: directional self-attention network for RNN/CNN-free language understanding. CoRR,abs/1709.04696 (2017) Shen, T., Zhou, T., Long, G., Jiang, J., Pan, S., Zhang, C.: DiSAN: directional self-attention network for RNN/CNN-free language understanding. CoRR,abs/1709.04696 (2017)
30.
Zurück zum Zitat Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: a simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15, 1929–1958 (2014)MathSciNetMATH Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: a simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15, 1929–1958 (2014)MathSciNetMATH
31.
Zurück zum Zitat Tai, K.S., Socher, R., Manning, C.D.: Improved semantic representations from tree-structured long short-term memory networks. In: ACL, Beijing, pp. 1556–1566, July 2015 Tai, K.S., Socher, R., Manning, C.D.: Improved semantic representations from tree-structured long short-term memory networks. In: ACL, Beijing, pp. 1556–1566, July 2015
32.
Zurück zum Zitat Tay, Y., Tuan, L.A., Hui, S.C.: A compare-propagate architecture with alignment factorization for natural language inference. arXiv preprint arXiv:1801.00102 (2017) Tay, Y., Tuan, L.A., Hui, S.C.: A compare-propagate architecture with alignment factorization for natural language inference. arXiv preprint arXiv:​1801.​00102 (2017)
33.
Zurück zum Zitat Wang, S., Jiang, J.: A compare-aggregate model for matching text sequences. CoRR,abs/1611.01747 (2016) Wang, S., Jiang, J.: A compare-aggregate model for matching text sequences. CoRR,abs/1611.01747 (2016)
34.
Zurück zum Zitat Wang, S., Jiang, J.: Learning natural language inference with LSTM. In: NAACL, San Diego, pp. 1442–1451, June 2016 Wang, S., Jiang, J.: Learning natural language inference with LSTM. In: NAACL, San Diego, pp. 1442–1451, June 2016
35.
Zurück zum Zitat Wang, W., Yang, N., Wei, F., Chang, B., Zhou, M.: Gated self-matching networks for reading comprehension and question answering. In: ACL, pp. 189–198 (2017) Wang, W., Yang, N., Wei, F., Chang, B., Zhou, M.: Gated self-matching networks for reading comprehension and question answering. In: ACL, pp. 189–198 (2017)
36.
Zurück zum Zitat Wang, Z., Wael, H., Radu, F.: Bilateral multi-perspective matching for natural language sentences. In: IJCAI, pp. 4144–4150 (2017) Wang, Z., Wael, H., Radu, F.: Bilateral multi-perspective matching for natural language sentences. In: IJCAI, pp. 4144–4150 (2017)
37.
Zurück zum Zitat Weston, J., Chopra, S., Bordes, A.: Memory networks. In: ICLR (2015) Weston, J., Chopra, S., Bordes, A.: Memory networks. In: ICLR (2015)
38.
Zurück zum Zitat Yu, H., Munkhdalai, T.: Neural tree indexers for text understanding. In: ACL, vol. 1, pp. 11–21 (2017) Yu, H., Munkhdalai, T.: Neural tree indexers for text understanding. In: ACL, vol. 1, pp. 11–21 (2017)
Metadaten
Titel
Multi-turn Inference Matching Network for Natural Language Inference
verfasst von
Chunhua Liu
Shan Jiang
Hainan Yu
Dong Yu
Copyright-Jahr
2018
DOI
https://doi.org/10.1007/978-3-319-99501-4_11

Premium Partner