Skip to main content
Top

2019 | OriginalPaper | Chapter

Subject Recognition in Chinese Sentences for Chatbots

Authors : Fangyuan Li, Huanhuan Wei, Qiangda Hao, Ruihong Zeng, Hao Shao, Wenliang Chen

Published in: Natural Language Processing and Chinese Computing

Publisher: Springer International Publishing

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

Subject (In this paper, subject means “ https://static-content.springer.com/image/chp%3A10.1007%2F978-3-030-32236-6_46/MediaObjects/490322_1_En_46_Figa_HTML.gif /zhu ti” in Chinese, while we use “grammatical subject” to denote traditional “ https://static-content.springer.com/image/chp%3A10.1007%2F978-3-030-32236-6_46/MediaObjects/490322_1_En_46_Figb_HTML.gif /zhu yu” in Chinese.) recognition plays a significant role in the conversation with a Chatbot. The misclassification of the subject of a sentence leads to the misjudgment of the intention recognition. In this paper, we build a new dataset for subject recognition and propose several systems based on pre-trained language models. We first design annotation guidelines for human-chatbot conversational data, and hire annotators to build a new dataset according to the guidelines. Then, classification methods based on deep neural network are proposed. Finally, extensive experiments are conducted to testify the performance of different algorithms. The results show that our method achieves 88.5% \(F_1\) in the task of subject recognition. We also compare our systems with three other Chatbot systems and find ours perform the best.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Footnotes
1
The term limited refers to specific reference and definite quantity, while non-limited means those generic reference and non-definite quantity. The subject discussed in this paper belongs to the limited, so the non-limited people are usually not regarded as subjects, such as the “men” in “All men must die”.
 
3
A subject-verb relation in LTP. https://​github.​com/​HIT-SCIR/​ltp.
 
Literature
1.
go back to reference Christensen, J., Soderland, S., Etzioni, O., et al.: An analysis of open information extraction based on semantic role labeling. In: K-CAP, pp. 113–120. ACM (2011) Christensen, J., Soderland, S., Etzioni, O., et al.: An analysis of open information extraction based on semantic role labeling. In: K-CAP, pp. 113–120. ACM (2011)
2.
go back to reference Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: Bert: pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018) Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: Bert: pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:​1810.​04805 (2018)
3.
go back to reference Joulin, A., Grave, E., Bojanowski, P., Mikolov, T.: Bag of tricks for efficient text classification. arXiv preprint arXiv:1607.01759 (2016) Joulin, A., Grave, E., Bojanowski, P., Mikolov, T.: Bag of tricks for efficient text classification. arXiv preprint arXiv:​1607.​01759 (2016)
4.
go back to reference Kim, S.M., Hovy, E.: Identifying opinion holders for question answering in opinion texts. In: AAAI 2005 Workshop, pp. 1367–1373 (2005) Kim, S.M., Hovy, E.: Identifying opinion holders for question answering in opinion texts. In: AAAI 2005 Workshop, pp. 1367–1373 (2005)
6.
go back to reference Lai, S., Xu, L., Liu, K., Zhao, J.: Recurrent convolutional neural networks for text classification. In: AAAI (2015) Lai, S., Xu, L., Liu, K., Zhao, J.: Recurrent convolutional neural networks for text classification. In: AAAI (2015)
7.
go back to reference Li, F., et al.: Structure-aware review mining and summarization. In: COLING, pp. 653–661 (2010) Li, F., et al.: Structure-aware review mining and summarization. In: COLING, pp. 653–661 (2010)
8.
go back to reference Miyato, T., Dai, A.M., Goodfellow, I.: Adversarial training methods for semi-supervised text classification. arXiv preprint arXiv:1605.07725 (2016) Miyato, T., Dai, A.M., Goodfellow, I.: Adversarial training methods for semi-supervised text classification. arXiv preprint arXiv:​1605.​07725 (2016)
10.
go back to reference Punyakanok, V., Roth, D., Yih, W.t.: The importance of syntactic parsing and inference in semantic role labeling. Comput. Linguist. 34(2), 257–287 (2008)CrossRef Punyakanok, V., Roth, D., Yih, W.t.: The importance of syntactic parsing and inference in semantic role labeling. Comput. Linguist. 34(2), 257–287 (2008)CrossRef
11.
go back to reference Qi, H., Yang, M., Meng, Y., Han, X., Zhao, T.: Skeleton parsing for specific domain Chinese text. J. Chin. Inf. Process. 18(1), 1–5 (2004). (in Chinese) Qi, H., Yang, M., Meng, Y., Han, X., Zhao, T.: Skeleton parsing for specific domain Chinese text. J. Chin. Inf. Process. 18(1), 1–5 (2004). (in Chinese)
12.
go back to reference Qiu, G., Liu, B., Bu, J., Chen, C.: Opinion word expansion and target extraction through double propagation. Comput. Linguist. 37(1), 9–27 (2011)CrossRef Qiu, G., Liu, B., Bu, J., Chen, C.: Opinion word expansion and target extraction through double propagation. Comput. Linguist. 37(1), 9–27 (2011)CrossRef
13.
go back to reference Rusu, D., Dali, L., Fortuna, B., Grobelnik, M., Mladenic, D.: Triplet extraction from sentences. In: IMSCI, pp. 8–12 (2007) Rusu, D., Dali, L., Fortuna, B., Grobelnik, M., Mladenic, D.: Triplet extraction from sentences. In: IMSCI, pp. 8–12 (2007)
14.
go back to reference Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, pp. 5998–6008 (2017) Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, pp. 5998–6008 (2017)
15.
go back to reference Wang, R., Ju, J., Li, S., Zhou, G.: Feature engineering for CRFs based opinion target extraction. J. Chin. Inf. Process. 26(2), 56–61 (2012). (in Chinese) Wang, R., Ju, J., Li, S., Zhou, G.: Feature engineering for CRFs based opinion target extraction. J. Chin. Inf. Process. 26(2), 56–61 (2012). (in Chinese)
16.
go back to reference Wiegand, M., Klakow, D.: Convolution kernels for opinion holder extraction. In: NAACL-HLT, pp. 795–803 (2010) Wiegand, M., Klakow, D.: Convolution kernels for opinion holder extraction. In: NAACL-HLT, pp. 795–803 (2010)
17.
go back to reference Zhou, H., Huang, M., Zhang, T., Zhu, X., Liu, B.: Emotional chatting machine: emotional conversation generation with internal and external memory. In: AAAI (2018) Zhou, H., Huang, M., Zhang, T., Zhu, X., Liu, B.: Emotional chatting machine: emotional conversation generation with internal and external memory. In: AAAI (2018)
18.
go back to reference Zhou, P., et al.: Attention-based bidirectional long short-term memory networks for relation classification. In: ACL (2016) Zhou, P., et al.: Attention-based bidirectional long short-term memory networks for relation classification. In: ACL (2016)
Metadata
Title
Subject Recognition in Chinese Sentences for Chatbots
Authors
Fangyuan Li
Huanhuan Wei
Qiangda Hao
Ruihong Zeng
Hao Shao
Wenliang Chen
Copyright Year
2019
DOI
https://doi.org/10.1007/978-3-030-32236-6_46

Premium Partner