Skip to main content
Erschienen in: Neural Processing Letters 2/2018

05.12.2017

First-Order Sensitivity Analysis for Hidden Neuron Selection in Layer-Wise Training of Networks

verfasst von: Bo Li, Cheng Chen

Erschienen in: Neural Processing Letters | Ausgabe 2/2018

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Multilayer neural networks are current trends in machine learning. Although complex architectures bring high performance, having sparse neurons in each layer can save memory, energy, and computational resources. In this paper, we aim to balance benefits between the complexity of architectures and the sparsity of neurons. An algorithm is proposed to prune neurons in multilayer neural networks through the global sensitivity analysis. Motivated by layer-wise training, we construct autoencoders with linear decoders, so mathematical models of multilayer neural networks can be considered as additive models. Hence, a first-order sensitivity analysis method, called random balance designs (RBD), is employed to select redundant neurons in hidden layers. This paper provides a novel framework to apply RBD in multilayer neural networks. Multiple experimental results demonstrate the generality and effectiveness of the proposed approach on structural learning of neural networks. After removing superfluous hidden neurons, higher accuracy can be obtained in most cases with less computation.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Hinton EG, Osindero S, Teh YW (2006) A fast learning algorithm for deep belief nets. Neural Comput 18(7):1527–1554MathSciNetCrossRef Hinton EG, Osindero S, Teh YW (2006) A fast learning algorithm for deep belief nets. Neural Comput 18(7):1527–1554MathSciNetCrossRef
2.
Zurück zum Zitat Schölkopf B, Platt J, Hofmann T (2007) Greedy layer-wise training of deep networks. In: Proceedings of international conference and workshop on neural information processing systems, pp 153–160 Schölkopf B, Platt J, Hofmann T (2007) Greedy layer-wise training of deep networks. In: Proceedings of international conference and workshop on neural information processing systems, pp 153–160
3.
Zurück zum Zitat Lecun Y, Bottou L, Bengio Y, Haffner P (1998) Gradient-based learning applied to document recognition. Proc IEEE 11:2278–2324CrossRef Lecun Y, Bottou L, Bengio Y, Haffner P (1998) Gradient-based learning applied to document recognition. Proc IEEE 11:2278–2324CrossRef
4.
Zurück zum Zitat Sutskever I (2013) Training recurrent neural networks. Ph.D. Dissertation, University of Toronto, Canada Sutskever I (2013) Training recurrent neural networks. Ph.D. Dissertation, University of Toronto, Canada
5.
Zurück zum Zitat Coates A, Huval B, Wang T, Wu D, Catanzaro B, Andrew N (2013) Deep learning with COTS HPC systems. In: Proceedings of the 30th international conference on machine learning, pp 1337–1345 Coates A, Huval B, Wang T, Wu D, Catanzaro B, Andrew N (2013) Deep learning with COTS HPC systems. In: Proceedings of the 30th international conference on machine learning, pp 1337–1345
6.
Zurück zum Zitat Nitish S, Geoffrey H, Alex K, Ilya S, Ruslan S (2014) Dropout: a simple way to prevent neural networks from overfitting. J Mach Learn Res 15(1):1929–1958MathSciNetMATH Nitish S, Geoffrey H, Alex K, Ilya S, Ruslan S (2014) Dropout: a simple way to prevent neural networks from overfitting. J Mach Learn Res 15(1):1929–1958MathSciNetMATH
7.
Zurück zum Zitat Wan L, Zeiler M, Zhang S, Cun YL, Fergus R (2013) Regularization of neural networks using dropconnect.’ In: Proceedings of the 30th international conference on machine learning, pp 1058–1066 Wan L, Zeiler M, Zhang S, Cun YL, Fergus R (2013) Regularization of neural networks using dropconnect.’ In: Proceedings of the 30th international conference on machine learning, pp 1058–1066
8.
9.
Zurück zum Zitat Engelbrecht A (2001) A new pruning heuristic based on variance analysis of sensitivity information. IEEE Trans Neural Netw 12(6):1386–1399CrossRef Engelbrecht A (2001) A new pruning heuristic based on variance analysis of sensitivity information. IEEE Trans Neural Netw 12(6):1386–1399CrossRef
10.
Zurück zum Zitat Mozer MC, Smolensky P (1989) Skeletonization: a technique for trimming the fat from a network via relevance assessment. Adv Neural Inf Process Syst 1:107–115 Mozer MC, Smolensky P (1989) Skeletonization: a technique for trimming the fat from a network via relevance assessment. Adv Neural Inf Process Syst 1:107–115
11.
Zurück zum Zitat Karnin ED (1990) A simple procedure for pruning backpropagation trained neural networks. IEEE Trans Neural Netw 1:239–242CrossRef Karnin ED (1990) A simple procedure for pruning backpropagation trained neural networks. IEEE Trans Neural Netw 1:239–242CrossRef
12.
Zurück zum Zitat Ruck DW, Rogers SK, Kabrisky M (1990) Feature selection using a multilayer perceptron. Neural Netw Comput 2(2):40–48 Ruck DW, Rogers SK, Kabrisky M (1990) Feature selection using a multilayer perceptron. Neural Netw Comput 2(2):40–48
13.
Zurück zum Zitat Tarr G (1991) Multilayered feedforward networks for image segementation. Ph.D. Dissertation, Air Force Institute of Technology Tarr G (1991) Multilayered feedforward networks for image segementation. Ph.D. Dissertation, Air Force Institute of Technology
14.
Zurück zum Zitat Lauret P, Fock E, Mara T (2006) A node pruning algorithm based on a Fourier amplitude sensitivity test method. IEEE Trans Neural Netw 17(2):273–293CrossRef Lauret P, Fock E, Mara T (2006) A node pruning algorithm based on a Fourier amplitude sensitivity test method. IEEE Trans Neural Netw 17(2):273–293CrossRef
15.
Zurück zum Zitat Fock E (2014) Global sensitivity analysis approach for input selection and system identification purposes: a new framework for feedforward neural networks. IEEE Trans Neural Netw Learn Syst 25(8):1484–1495CrossRef Fock E (2014) Global sensitivity analysis approach for input selection and system identification purposes: a new framework for feedforward neural networks. IEEE Trans Neural Netw Learn Syst 25(8):1484–1495CrossRef
16.
Zurück zum Zitat Han H, Qiao J (2010) A self-organizing fuzzy neural network based on a growing-and-pruning algorithm. IEEE Trans Fuzzy Syst 18(6):1129–1143CrossRef Han H, Qiao J (2010) A self-organizing fuzzy neural network based on a growing-and-pruning algorithm. IEEE Trans Fuzzy Syst 18(6):1129–1143CrossRef
17.
Zurück zum Zitat Chen C, Wang F-Y (2013) A self-organizing neuro-fuzzy network based on first order effect sensitivity analysis. Neurocomputing 118:21–32CrossRef Chen C, Wang F-Y (2013) A self-organizing neuro-fuzzy network based on first order effect sensitivity analysis. Neurocomputing 118:21–32CrossRef
18.
Zurück zum Zitat Hastie T, Tibshirani R (1990) Generalized additive models. Taylor and Francis, AbingdonMATH Hastie T, Tibshirani R (1990) Generalized additive models. Taylor and Francis, AbingdonMATH
20.
Zurück zum Zitat Saltelli A, Chan K, Scott E (2000) Sensitivity analysis. Wiley, New YorkMATH Saltelli A, Chan K, Scott E (2000) Sensitivity analysis. Wiley, New YorkMATH
21.
Zurück zum Zitat Saltelli A (2008) Global sensitivity analysis: the primer. Wiley, New YorkMATH Saltelli A (2008) Global sensitivity analysis: the primer. Wiley, New YorkMATH
22.
Zurück zum Zitat Coates A, Lee H, Ng AY (2011) An analysis of single layer networks in unsupervised feature learning. Jf Mach Learn Res 15:215–223 Coates A, Lee H, Ng AY (2011) An analysis of single layer networks in unsupervised feature learning. Jf Mach Learn Res 15:215–223
23.
Zurück zum Zitat Wang F-Y (2010) Parallel control and management for intelligent transportation systems: concepts, architectures, and applications. IEEE Trans Intell Transp Syst 11(3):630–638CrossRef Wang F-Y (2010) Parallel control and management for intelligent transportation systems: concepts, architectures, and applications. IEEE Trans Intell Transp Syst 11(3):630–638CrossRef
24.
Zurück zum Zitat Sun S, Zhang C, Yu G (2006) A bayesian network approach to traffic flow forecasting. IEEE Trans Intell Transp Syst 7(1):124–132CrossRef Sun S, Zhang C, Yu G (2006) A bayesian network approach to traffic flow forecasting. IEEE Trans Intell Transp Syst 7(1):124–132CrossRef
25.
Zurück zum Zitat Pelleg D, Moore A (2000) X-means: extending k-means with efficient estimation of the number of clusters. In: Proceedings of the seventeenth international conference on machine learning. San Francisco: Morgan Kaufmann, pp 727–734 Pelleg D, Moore A (2000) X-means: extending k-means with efficient estimation of the number of clusters. In: Proceedings of the seventeenth international conference on machine learning. San Francisco: Morgan Kaufmann, pp 727–734
Metadaten
Titel
First-Order Sensitivity Analysis for Hidden Neuron Selection in Layer-Wise Training of Networks
verfasst von
Bo Li
Cheng Chen
Publikationsdatum
05.12.2017
Verlag
Springer US
Erschienen in
Neural Processing Letters / Ausgabe 2/2018
Print ISSN: 1370-4621
Elektronische ISSN: 1573-773X
DOI
https://doi.org/10.1007/s11063-017-9764-6

Weitere Artikel der Ausgabe 2/2018

Neural Processing Letters 2/2018 Zur Ausgabe

Neuer Inhalt