Skip to main content
Top

2019 | OriginalPaper | Chapter

Syntactic Analysis of Power Grid Emergency Pre-plans Based on Transfer Learning

Authors : He Shi, Qun Yang, Bo Wang, Shaohan Liu, Kai Zhou

Published in: Intelligence Science and Big Data Engineering. Big Data and Machine Learning

Publisher: Springer International Publishing

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

To deal with the emergency pre-plans saved by the power grid dispatch department, so that the dispatcher can quickly retrieve and match similar accidents in the pre-plans, then they can learn from the experience of previous relevant situations, it is necessary to extract the information of the pre-plans and extract its key information. Therefore, deep learning method with strong generalization ability and learning ability and continuous improvement of model can be adopted. However, this method usually requires a large amount of data, but the existing labeling data in the power grid field is limited and the manual method for data labeling is a huge workload. Therefore, in the case of insufficient data, this paper aims to solve how to use deep learning method for effective information extraction? This paper modifies the ULMFiT model and uses it to carry out word vector training, adopting transfer learning method to introduce annotating datasets in the open field and combining with the data in the field of power grid to training model. In this way, the semantic relation of power grid domain is introduced into the syntactic analysis of the pre-plans, and we can further complete the information extraction. Experimental verification is carried out in this paper, the results show that, in the case of insufficient corpus or small amount of annotated data, this method can solve the problem of part of speech analysis errors, it can also improve the accuracy of syntactic analysis, and the experimental verifies the effectiveness of this method.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Lu, Q., Liu, X.Y.: Knowledge map q&a semantic matching model based on transfer learning. Comput. Appl. 335(7), 22–28 (2018)MathSciNet Lu, Q., Liu, X.Y.: Knowledge map q&a semantic matching model based on transfer learning. Comput. Appl. 335(7), 22–28 (2018)MathSciNet
3.
go back to reference Liu, D., Chen, Y., Shen, C.: Study on digital method of power emergency plan. Autom. Power Syst. 33(21) (2009) Liu, D., Chen, Y., Shen, C.: Study on digital method of power emergency plan. Autom. Power Syst. 33(21) (2009)
4.
go back to reference Wu, W.C., Zhang, B.M., Cao, F.C.: Design and key technology of grid emergency command technical support system. Power Syst. Autom. 15, 1–6 (2008) Wu, W.C., Zhang, B.M., Cao, F.C.: Design and key technology of grid emergency command technical support system. Power Syst. Autom. 15, 1–6 (2008)
5.
go back to reference Zhao, X.S., Xie, B.M., Zhang, H.W.: A power grid fault diagnosis method based on deep learning algorithm. Henan Sci. Technol. 23, 53–54 (2016) Zhao, X.S., Xie, B.M., Zhang, H.W.: A power grid fault diagnosis method based on deep learning algorithm. Henan Sci. Technol. 23, 53–54 (2016)
6.
go back to reference Jiang, Q., Shen, L., Zhang, W., He, X.: Research on fault diagnosis method based on deep learning. Comput. Simul. 35(7), 409–413 (2018) Jiang, Q., Shen, L., Zhang, W., He, X.: Research on fault diagnosis method based on deep learning. Comput. Simul. 35(7), 409–413 (2018)
7.
go back to reference Qiu, J., Wang, H.F., Ying, J.L.: Text information mining technology and its application in the life cycle state evaluation of circuit breakers. Power Syst. Autom. 40(6), 107–112 (2016) Qiu, J., Wang, H.F., Ying, J.L.: Text information mining technology and its application in the life cycle state evaluation of circuit breakers. Power Syst. Autom. 40(6), 107–112 (2016)
8.
go back to reference Zhang, Y.K., Zhang, P.Y., Yan, Y.H.: Language model data enhancement technology based on antagonistic training strategy. Acta Automatica Sinica 44(05), 126–135 (2018) Zhang, Y.K., Zhang, P.Y., Yan, Y.H.: Language model data enhancement technology based on antagonistic training strategy. Acta Automatica Sinica 44(05), 126–135 (2018)
9.
go back to reference Zhang, Q.L., Du, J.C., Xu, R.F.: Research on irony recognition based on antagonistic learning. J. Peking Univ. (nat. Sci. Edn.) 55(01), 32–39 (2019) Zhang, Q.L., Du, J.C., Xu, R.F.: Research on irony recognition based on antagonistic learning. J. Peking Univ. (nat. Sci. Edn.) 55(01), 32–39 (2019)
10.
go back to reference Chen, C., Shen, F., Yan, R.Q.: Bearing fault diagnosis based on improved LSSVM migration learning method. J. Instrum. (2017) Chen, C., Shen, F., Yan, R.Q.: Bearing fault diagnosis based on improved LSSVM migration learning method. J. Instrum. (2017)
11.
go back to reference Gu, T.Y., Guo, J.S., Li, Z.X.: Fault probability prediction of airborne equipment based on interpolation-fit-transfer learning algorithm. Syst. Eng. Electron. Technol. 40(1), 114–118 (2018) Gu, T.Y., Guo, J.S., Li, Z.X.: Fault probability prediction of airborne equipment based on interpolation-fit-transfer learning algorithm. Syst. Eng. Electron. Technol. 40(1), 114–118 (2018)
12.
go back to reference Ren, J., Hu, X.F., Li, N.: Transfer learning prediction algorithm based on SDA and SVR hybrid model. Comput. Sci. 45(1) (2018) Ren, J., Hu, X.F., Li, N.: Transfer learning prediction algorithm based on SDA and SVR hybrid model. Comput. Sci. 45(1) (2018)
13.
go back to reference Miwa, M., Sætre, R., Miyao, Y.: A rich feature vector for protein-protein interaction extraction from multiple corpora. In: Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics (2009) Miwa, M., Sætre, R., Miyao, Y.: A rich feature vector for protein-protein interaction extraction from multiple corpora. In: Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics (2009)
14.
go back to reference Li, L.S., Guo, R., Huang, D.G.: Extraction of protein interaction based on transfer learning. J. Chin. Inf. Process. 30(2), 160–167 (2016) Li, L.S., Guo, R., Huang, D.G.: Extraction of protein interaction based on transfer learning. J. Chin. Inf. Process. 30(2), 160–167 (2016)
15.
go back to reference Oquab, M., Bottou, L., Laptev, I.: Learning and transferring mid-level image representations using convolutional neural networks. In: Computer Vision and Pattern Recognition. IEEE (2014) Oquab, M., Bottou, L., Laptev, I.: Learning and transferring mid-level image representations using convolutional neural networks. In: Computer Vision and Pattern Recognition. IEEE (2014)
16.
go back to reference Razavian, A.S., Azizpour, H., Sullivan, J.: CNN features off-the-shelf: an astounding baseline for recognition. In: Computer Vision & Pattern Recognition Workshops (2014) Razavian, A.S., Azizpour, H., Sullivan, J.: CNN features off-the-shelf: an astounding baseline for recognition. In: Computer Vision & Pattern Recognition Workshops (2014)
17.
go back to reference Yosinski, J., Clune, J., Bengio, Y.: How transferable are features in deep neural networks? Eprint Arxiv 27, 3320–3328 (2014) Yosinski, J., Clune, J., Bengio, Y.: How transferable are features in deep neural networks? Eprint Arxiv 27, 3320–3328 (2014)
18.
go back to reference Wu, Y., Ji, Q.: Constrained deep transfer feature learning and its applications. In: Computer Vision and Pattern Recognition, pp. 5101–5109. IEEE (2016) Wu, Y., Ji, Q.: Constrained deep transfer feature learning and its applications. In: Computer Vision and Pattern Recognition, pp. 5101–5109. IEEE (2016)
19.
go back to reference Tzeng, E., Hoffman, J., Zhang, N.: Deep domain confusion: maximizing for domain invariance. Comput. Sci. (2014) Tzeng, E., Hoffman, J., Zhang, N.: Deep domain confusion: maximizing for domain invariance. Comput. Sci. (2014)
20.
go back to reference Wang, N.: Conceptual relationship extraction of basic education geography based on transfer learning. Wuhan University of Technology (2017) Wang, N.: Conceptual relationship extraction of basic education geography based on transfer learning. Wuhan University of Technology (2017)
21.
go back to reference Pan, C.W.: Research on vector optimization method for pre-training Chinese words in transfer learning. Beijing Jiaotong University (2018) Pan, C.W.: Research on vector optimization method for pre-training Chinese words in transfer learning. Beijing Jiaotong University (2018)
22.
go back to reference Peters, M.E., et al.: Deep contextualized word representations. NAACL (2018) Peters, M.E., et al.: Deep contextualized word representations. NAACL (2018)
23.
go back to reference Howard, J., Ruder, S.: Universal Language Model Fine-tuning for Text Classification (2018) Howard, J., Ruder, S.: Universal Language Model Fine-tuning for Text Classification (2018)
24.
go back to reference Zhuang, F.Z.: Research progress in transfer learning. J. Softw. 26(1), 26–39 (2015)MathSciNet Zhuang, F.Z.: Research progress in transfer learning. J. Softw. 26(1), 26–39 (2015)MathSciNet
25.
go back to reference Pan, S.J., Yang, Q.: A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 22(10), 1345–1359 (2010)CrossRef Pan, S.J., Yang, Q.: A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 22(10), 1345–1359 (2010)CrossRef
26.
go back to reference Tan, C., Sun, F., Kong, T.: A Survey on Deep Transfer Learning (2018) Tan, C., Sun, F., Kong, T.: A Survey on Deep Transfer Learning (2018)
27.
go back to reference Zhang, J., Qu, D., Li, Z.: Language model of cyclic neural network based on word vector features. Pattern Recogn. Artif. Intell. 28(4), 000299–305 (2015) Zhang, J., Qu, D., Li, Z.: Language model of cyclic neural network based on word vector features. Pattern Recogn. Artif. Intell. 28(4), 000299–305 (2015)
28.
go back to reference Hinton, G.E.: Learning distributed representations of concepts. In: Proceedings of the Eighth Annual Conference of the Cognitive Science Society, pp. 1–12. Erlbaum, Hillsdale (1986) Hinton, G.E.: Learning distributed representations of concepts. In: Proceedings of the Eighth Annual Conference of the Cognitive Science Society, pp. 1–12. Erlbaum, Hillsdale (1986)
29.
go back to reference Mikolov, T., Chen, K., Corrado, G.: Efficient estimation of word representations in vector space. Comput. Sci. (2013) Mikolov, T., Chen, K., Corrado, G.: Efficient estimation of word representations in vector space. Comput. Sci. (2013)
30.
go back to reference Mikolov, T., Sutskever, I., Kai, C.: Distributed representations of words and phrases and their compositionality. Adv. Neural. Inf. Process. Syst. 26, 3111–3119 (2013) Mikolov, T., Sutskever, I., Kai, C.: Distributed representations of words and phrases and their compositionality. Adv. Neural. Inf. Process. Syst. 26, 3111–3119 (2013)
31.
go back to reference Rong, X.: word2vec Parameter Learning Explained. Comput. Sci. (2014) Rong, X.: word2vec Parameter Learning Explained. Comput. Sci. (2014)
32.
go back to reference Devlin, J., Chang, M.W., Lee, K.: BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding (2018) Devlin, J., Chang, M.W., Lee, K.: BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding (2018)
33.
go back to reference Socher, R., Bauer, J., Manning, C.D., Ng, A.Y.: Parsing with compositional vector grammars. In: Proceedings of 51st Annual Meeting of the Association for Conputational Linguistics (Volume 1: Long papers), pp. 455–465 (2013) Socher, R., Bauer, J., Manning, C.D., Ng, A.Y.: Parsing with compositional vector grammars. In: Proceedings of 51st Annual Meeting of the Association for Conputational Linguistics (Volume 1: Long papers), pp. 455–465 (2013)
34.
go back to reference Klein, D.: Accurate unlexicalized parsing. In: Proceedings of the 41st Meeting of the Association for Computational Linguistics, Sapporo, Japan (2003) Klein, D.: Accurate unlexicalized parsing. In: Proceedings of the 41st Meeting of the Association for Computational Linguistics, Sapporo, Japan (2003)
35.
go back to reference Henderson, J.: Discriminative training of a neural network statistical parser. In: Proceedings of the 42nd Annual Meeting of the Association for Computational Linguistics, Barcelona, Spain. DBLP (2004) Henderson, J.: Discriminative training of a neural network statistical parser. In: Proceedings of the 42nd Annual Meeting of the Association for Computational Linguistics, Barcelona, Spain. DBLP (2004)
37.
go back to reference Yao, Y., Shen, F., Zhang, J., Liu, L., Tang, Z., Shao, L.: Extracting privileged information for enhancing classifier learning. IEEE Trans. Image Process. 28(1), 436–450 (2019)MathSciNetCrossRef Yao, Y., Shen, F., Zhang, J., Liu, L., Tang, Z., Shao, L.: Extracting privileged information for enhancing classifier learning. IEEE Trans. Image Process. 28(1), 436–450 (2019)MathSciNetCrossRef
38.
go back to reference Yao, Y., Zhang, J., Shen, F., Hua, X., Xu, J., Tang, Z.: Exploiting web images for dataset construction: a domain robust approach. IEEE Trans. Multimedia 19(8), 1771–1784 (2017)CrossRef Yao, Y., Zhang, J., Shen, F., Hua, X., Xu, J., Tang, Z.: Exploiting web images for dataset construction: a domain robust approach. IEEE Trans. Multimedia 19(8), 1771–1784 (2017)CrossRef
Metadata
Title
Syntactic Analysis of Power Grid Emergency Pre-plans Based on Transfer Learning
Authors
He Shi
Qun Yang
Bo Wang
Shaohan Liu
Kai Zhou
Copyright Year
2019
DOI
https://doi.org/10.1007/978-3-030-36204-1_14

Premium Partner