Skip to main content
Top

2018 | OriginalPaper | Chapter

Attention-Based CNN-BLSTM Networks for Joint Intent Detection and Slot Filling

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

Dialogue intent detection and semantic slot filling are two critical tasks in nature language understanding (NLU) for task-oriented dialog systems. In this paper, we present an attention-based encoder-decoder neural network model for joint intent detection and slot filling, which encodes sentence representation with a hybrid Convolutional Neural Networks and Bidirectional Long Short-Term Memory Networks (CNN-BLSTM), and decodes it with an attention-based recurrent neural network with aligned inputs. In the encoding process, our model firstly extracts higher-level phrase representations and local features from each utterance using convolutional neural network, and then propagates historical contextual semantic information with a bidirectional long short-term memory network layer architecture. Accordingly, we could obtain sentence representation by merging the two architectures mentioned above. In the decoding process, we introduce attention mechanism in long short-term memory networks that can provide additional sematic information. We conduct experiment on dialogue intent detection and slot filling tasks with standard data set Airline Travel Information System (ATIS). Experimental results manifest that our proposed model can achieve better overall performance.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Shen, B., Inkpen, D.: Speech intent recognition for robots (2017) Shen, B., Inkpen, D.: Speech intent recognition for robots (2017)
2.
go back to reference Liu, B., Lane, I.: Attention-based recurrent neural network models for joint intent detection and slot filling (2016) Liu, B., Lane, I.: Attention-based recurrent neural network models for joint intent detection and slot filling (2016)
3.
go back to reference Liu, B., Line, I.: Recurrent neural network structured output prediction for spoken language understanding (2015) Liu, B., Line, I.: Recurrent neural network structured output prediction for spoken language understanding (2015)
4.
go back to reference Kim, Y.: Convolutional neural networks for sentence classification. Eprint Arxiv (2014) Kim, Y.: Convolutional neural networks for sentence classification. Eprint Arxiv (2014)
5.
go back to reference Zhang, X., Wang, H.: A joint model of intent determination and slot filling for spoken language understanding, pp. 5690–5694 (2016) Zhang, X., Wang, H.: A joint model of intent determination and slot filling for spoken language understanding, pp. 5690–5694 (2016)
6.
go back to reference Haffner, P., Tur, G., Wright, J.H.: Optimizing SVMs for complex call classification (2003) Haffner, P., Tur, G., Wright, J.H.: Optimizing SVMs for complex call classification (2003)
7.
go back to reference Chen, J., Huang, H., Tian, S., et al.: Feature selection for text classification with Naive Bayes. Expert Syst. Appl. Int. J. 36(3), 5432–5435 (2009)CrossRef Chen, J., Huang, H., Tian, S., et al.: Feature selection for text classification with Naive Bayes. Expert Syst. Appl. Int. J. 36(3), 5432–5435 (2009)CrossRef
8.
go back to reference Graves, A., Jaitly, N., Mohamed, A.R.: Hybrid speech recognition with deep bidirectional LSTM (2014) Graves, A., Jaitly, N., Mohamed, A.R.: Hybrid speech recognition with deep bidirectional LSTM (2014)
9.
go back to reference Xiao, Y., Cho, K.: Efficient character-level document classification by combining convolution and recurrent layers (2016) Xiao, Y., Cho, K.: Efficient character-level document classification by combining convolution and recurrent layers (2016)
10.
go back to reference Xiao, J., Wang, X., Liu, B.: The study of a nonstationary maximum entropy Markov model and its application on the pos-tagging task. ACM Trans. Asian Lang. Inf. Process. 6(2), 7 (2007)CrossRef Xiao, J., Wang, X., Liu, B.: The study of a nonstationary maximum entropy Markov model and its application on the pos-tagging task. ACM Trans. Asian Lang. Inf. Process. 6(2), 7 (2007)CrossRef
11.
go back to reference Raymond, C., Riccardi, G.: Generative and discriminative algorithms for spoken language understanding (2007) Raymond, C., Riccardi, G.: Generative and discriminative algorithms for spoken language understanding (2007)
12.
go back to reference Aliannejadi, M., Kiaeeha, M., Khadivi, S., et al.: Graph-based semi-supervised conditional random fields for spoken language understanding using unaligned data (2017) Aliannejadi, M., Kiaeeha, M., Khadivi, S., et al.: Graph-based semi-supervised conditional random fields for spoken language understanding using unaligned data (2017)
13.
go back to reference Xu, P., Sarikaya, R.: Convolutional neural network based triangular CRF for joint intent detection and slot filling (2014) Xu, P., Sarikaya, R.: Convolutional neural network based triangular CRF for joint intent detection and slot filling (2014)
14.
go back to reference Yao, K., Peng, B., Zhang, Y., et al.: Spoken language understanding using long short-term memory neural networks (2015) Yao, K., Peng, B., Zhang, Y., et al.: Spoken language understanding using long short-term memory neural networks (2015)
15.
go back to reference Vu, N.T., Gupta, P., Adel, H., et al.: Bi-directional recurrent neural network with ranking loss for spoken language understanding (2016) Vu, N.T., Gupta, P., Adel, H., et al.: Bi-directional recurrent neural network with ranking loss for spoken language understanding (2016)
16.
go back to reference Kurata, G., Xiang, B., Zhou, B., et al.: Leveraging sentence-level information with encoder LSTM for natural language understanding (2016) Kurata, G., Xiang, B., Zhou, B., et al.: Leveraging sentence-level information with encoder LSTM for natural language understanding (2016)
17.
go back to reference Zhu, S., Yu, K.: Encoder-decoder with focus-mechanism for sequence labelling based spoken language understanding (2017) Zhu, S., Yu, K.: Encoder-decoder with focus-mechanism for sequence labelling based spoken language understanding (2017)
18.
go back to reference Guo, D., Tur, G., Yih, W.T., et al.: Joint semantic utterance classification and slot filling with recursive neural networks (2015) Guo, D., Tur, G., Yih, W.T., et al.: Joint semantic utterance classification and slot filling with recursive neural networks (2015)
19.
go back to reference Liu, B., Lane, I.: Joint online spoken language understanding and language modeling with recurrent neural networks (2016) Liu, B., Lane, I.: Joint online spoken language understanding and language modeling with recurrent neural networks (2016)
20.
go back to reference Weigelt, S., Hey, T., Landhäußer, M.: Integrating a dialog component into a framework for spoken language understanding (2018) Weigelt, S., Hey, T., Landhäußer, M.: Integrating a dialog component into a framework for spoken language understanding (2018)
21.
go back to reference Zhou, C., Sun, C., Liu, Z., et al.: A C-LSTM neural network for text classification. Comput. Sci. 1(4), 39–44 (2015) Zhou, C., Sun, C., Liu, Z., et al.: A C-LSTM neural network for text classification. Comput. Sci. 1(4), 39–44 (2015)
22.
go back to reference Yao, K., Peng, B., Zhang, Y., et al.: Spoken language understanding using long short-term memory neural networks. In: IEEE – Institute of Electrical & Electronics Engineers, pp. 189–194 (2014) Yao, K., Peng, B., Zhang, Y., et al.: Spoken language understanding using long short-term memory neural networks. In: IEEE – Institute of Electrical & Electronics Engineers, pp. 189–194 (2014)
24.
go back to reference Yin, W., Schütze, H., Xiang, B., et al.: ABCNN: attention-based convolutional neural network for modeling sentence pairs (2015) Yin, W., Schütze, H., Xiang, B., et al.: ABCNN: attention-based convolutional neural network for modeling sentence pairs (2015)
25.
go back to reference Morin, F., Bengio, Y.: Hierarchical probabilistic neural network language model. Aistats (2005) Morin, F., Bengio, Y.: Hierarchical probabilistic neural network language model. Aistats (2005)
26.
go back to reference Graves, A., Mohamed, A.R., Hinton, G.: Speech recognition with deep recurrent neural networks (2013) Graves, A., Mohamed, A.R., Hinton, G.: Speech recognition with deep recurrent neural networks (2013)
27.
go back to reference Hemphill, C.T., Godfrey, J.J., Doddington, G.R.: The ATIS spoken language systems pilot corpus. In: Proceedings of the Darpa Speech & Natural Language Workshop, pp. 96–101 (1990) Hemphill, C.T., Godfrey, J.J., Doddington, G.R.: The ATIS spoken language systems pilot corpus. In: Proceedings of the Darpa Speech & Natural Language Workshop, pp. 96–101 (1990)
28.
go back to reference Jozefowicz, R., Zaremba, W., Sutskever, I.: An empirical exploration of recurrent network architectures. In: International Conference on Machine Learning, pp. 2342–2350. JMLR.org (2015) Jozefowicz, R., Zaremba, W., Sutskever, I.: An empirical exploration of recurrent network architectures. In: International Conference on Machine Learning, pp. 2342–2350. JMLR.org (2015)
Metadata
Title
Attention-Based CNN-BLSTM Networks for Joint Intent Detection and Slot Filling
Authors
Yufan Wang
Li Tang
Tingting He
Copyright Year
2018
DOI
https://doi.org/10.1007/978-3-030-01716-3_21

Premium Partner