Skip to main content

2016 | OriginalPaper | Buchkapitel

Order in the Black Box: Consistency and Robustness of Hidden Neuron Activation of Feed Forward Neural Networks and Its Use in Efficient Optimization of Network Structure

verfasst von : Sandhya Samarasinghe

Erschienen in: Artificial Neural Network Modelling

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Neural networks are widely used for nonlinear pattern recognition and regression. However, they are considered as black boxes due to lack of transparency of internal workings and lack of direct relevance of its structure to the problem being addressed making it difficult to gain insights. Furthermore, structure of a neural network requires optimization which is still a challenge. Many existing structure optimization approaches require either extensive multi-stage pruning or setting subjective thresholds for pruning parameters. The knowledge of any internal consistency in the behavior of neurons could help develop simpler, systematic and more efficient approaches to optimise network structure. This chapter addresses in detail the issue of internal consistency in relation to redundancy and robustness of network structure of feed forward networks (3-layer) that are widely used for nonlinear regression. It first investigates if there is a recognizable consistency in neuron activation patterns under all conditions of network operation such as noise and initial weights. If such consistency exists, it points to a recognizable optimum network structure for given data. The results show that such pattern does exist and it is most clearly evident not at the level of hidden neuron activation but hidden neuron input to the output neuron (i.e., weighted hidden neuron activation). It is shown that when a network has more than the optimum number of hidden neurons, the redundant neurons form clearly distinguishable correlated patterns of their weighted outputs. This correlation structure is exploited to extract the required number of neurons using correlation distance based self organising maps that are clustered using Ward clustering that optimally cluster correlated weighted hidden neuron activity patterns without any user defined criteria or thresholds, thus automatically optimizing network structure in one step. The number of Ward clusters on the SOM is the required optimum number of neurons. The SOM/Ward based optimum network is compared with that obtained using two documented pruning methods: optimal brain damage and variance nullity measure to show the efficacy of the correlation approach in providing equivalent results. Also, the robustness of the network with optimum structure is tested against perturbation of weights and confidence intervals for weights are illustrated. Finally, the approach is tested on two practical problems involving a breast cancer diagnostic system and river flow forecasting.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Anhänge
Nur mit Berechtigung zugänglich
Literatur
1.
Zurück zum Zitat S. Samarasinghe, Neural Networks for Applied Sciences and Engineering-From Fundamentals to Complex Pattern Recognition (CRC Press, 2006) S. Samarasinghe, Neural Networks for Applied Sciences and Engineering-From Fundamentals to Complex Pattern Recognition (CRC Press, 2006)
2.
Zurück zum Zitat C. Bishop, Neural Networks for Pattern Recognition (Clarendon Press, Oxford, UK, 1996)MATH C. Bishop, Neural Networks for Pattern Recognition (Clarendon Press, Oxford, UK, 1996)MATH
3.
Zurück zum Zitat S. Haykin, Neural Networks: A comprehensive Foundation, 2nd edn. (Prentice Hall Inc, New Jersey, USA, 1999)MATH S. Haykin, Neural Networks: A comprehensive Foundation, 2nd edn. (Prentice Hall Inc, New Jersey, USA, 1999)MATH
4.
Zurück zum Zitat R. Reed, Pruning algorithms-A survey. IEEE Trans. Neural Networks 4, 740–747 (1993)CrossRef R. Reed, Pruning algorithms-A survey. IEEE Trans. Neural Networks 4, 740–747 (1993)CrossRef
5.
Zurück zum Zitat Y. Le Cun, J.S. Denker, S.A. Solla, Optimal brain damage, in Advances in Neural Information Processing (2), ed. by D.S. Touretzky (1990), pp. 598–605 Y. Le Cun, J.S. Denker, S.A. Solla, Optimal brain damage, in Advances in Neural Information Processing (2), ed. by D.S. Touretzky (1990), pp. 598–605
6.
Zurück zum Zitat B. Hassibi, D.G. Stork, G.J. Wolff, Optimal brain surgeon and general network pruning. IEEE International Conference on Neural Networks, vol. 1, (San Francisco, 1992), pp. 293–298 B. Hassibi, D.G. Stork, G.J. Wolff, Optimal brain surgeon and general network pruning. IEEE International Conference on Neural Networks, vol. 1, (San Francisco, 1992), pp. 293–298
7.
Zurück zum Zitat B. Hassibi, D.G. Stork, Second-order derivatives for network pruning: Optimal brain surgeon, in Advances in Neural Information Processing Systems, vol. 5, ed. by C. Lee Giles, S.J. Hanson, J.D. Cowan, (1993), pp. 164–171 B. Hassibi, D.G. Stork, Second-order derivatives for network pruning: Optimal brain surgeon, in Advances in Neural Information Processing Systems, vol. 5, ed. by C. Lee Giles, S.J. Hanson, J.D. Cowan, (1993), pp. 164–171
8.
Zurück zum Zitat A.P. Engelbrecht, A new pruning heuristic based on variance analysis of sensitivity information. IEEE Trans. Neural Networks 12(6), 1386–1399 (2001)CrossRef A.P. Engelbrecht, A new pruning heuristic based on variance analysis of sensitivity information. IEEE Trans. Neural Networks 12(6), 1386–1399 (2001)CrossRef
9.
Zurück zum Zitat K. Hagiwara, Regularization learning, early stopping and biased estimator. Neurocomputing 48, 937–955 (2002)CrossRefMATH K. Hagiwara, Regularization learning, early stopping and biased estimator. Neurocomputing 48, 937–955 (2002)CrossRefMATH
10.
Zurück zum Zitat M. Hagiwara, Removal of hidden units and weights for backpropagation networks. Proc. Int. Joint Conf. Neural Networks 1, 351–354 (1993) M. Hagiwara, Removal of hidden units and weights for backpropagation networks. Proc. Int. Joint Conf. Neural Networks 1, 351–354 (1993)
11.
Zurück zum Zitat F. Aires, Neural network uncertainty assessment using Bayesian statistics with application to remote sensing: 1. Network weights. J. Geophys. Res. 109, D10303 (2004). doi:10.1029/2003JD004173 CrossRef F. Aires, Neural network uncertainty assessment using Bayesian statistics with application to remote sensing: 1. Network weights. J. Geophys. Res. 109, D10303 (2004). doi:10.​1029/​2003JD004173 CrossRef
12.
Zurück zum Zitat F. Aires, Neural network uncertainty assessment using Bayesian statistics with application to remote sensing: 2. Output Error. J. Geophys. Res. 109, D10304 (2004). doi:10.1029/2003JD004174 CrossRef F. Aires, Neural network uncertainty assessment using Bayesian statistics with application to remote sensing: 2. Output Error. J. Geophys. Res. 109, D10304 (2004). doi:10.​1029/​2003JD004174 CrossRef
13.
Zurück zum Zitat F. Aires, Neural network uncertainty assessment using Bayesian statistics with application to remote sensing: 3. Network Jacobians. J. Geophys. Res. 109, D10305 (2004). doi:10.1029/2003JD004175 CrossRef F. Aires, Neural network uncertainty assessment using Bayesian statistics with application to remote sensing: 3. Network Jacobians. J. Geophys. Res. 109, D10305 (2004). doi:10.​1029/​2003JD004175 CrossRef
14.
Zurück zum Zitat K. Warne, G. Prasad, S. Rezvani, L. Maguire, Statistical computational intelligence techniques for inferential model development: A comparative evaluation and novel proposition for fusion. Eng. Appl. Artif. Intell. 17, 871–885 (2004)CrossRef K. Warne, G. Prasad, S. Rezvani, L. Maguire, Statistical computational intelligence techniques for inferential model development: A comparative evaluation and novel proposition for fusion. Eng. Appl. Artif. Intell. 17, 871–885 (2004)CrossRef
15.
Zurück zum Zitat I. Rivals, L. Personnaz, Construction of Confidence Intervals for neural networks based on least squares estimation. Neural Networks 13, 463–484 (2000)CrossRef I. Rivals, L. Personnaz, Construction of Confidence Intervals for neural networks based on least squares estimation. Neural Networks 13, 463–484 (2000)CrossRef
16.
Zurück zum Zitat E.J. Teoh, K.C. Tan, C. Xiang, Estimating the number of hidden neurons in a feed forward network using the singular value decomposition IEEE Trans. Neural Networks 17(6), (2006) E.J. Teoh, K.C. Tan, C. Xiang, Estimating the number of hidden neurons in a feed forward network using the singular value decomposition IEEE Trans. Neural Networks 17(6), (2006)
17.
Zurück zum Zitat C. Xian, S.Q. Ding, T.H. Lee, Geometrical interpretation and architecture selection of MLP, IEEE Trans. Neural Networks 16(1), (2005) C. Xian, S.Q. Ding, T.H. Lee, Geometrical interpretation and architecture selection of MLP, IEEE Trans. Neural Networks 16(1), (2005)
18.
Zurück zum Zitat P.A. Castillo, J. Carpio, J.J. Merelo, V. Rivas, G. Romero, A. Prieto, Evolving multilayer perceptrons. Neural Process. Lett. 12(2), 115–127 (2000)CrossRefMATH P.A. Castillo, J. Carpio, J.J. Merelo, V. Rivas, G. Romero, A. Prieto, Evolving multilayer perceptrons. Neural Process. Lett. 12(2), 115–127 (2000)CrossRefMATH
19.
Zurück zum Zitat X. Yao, Evolutionary artificial neural networks. Proc. IEEE 87(9), 1423–1447 (1999)CrossRef X. Yao, Evolutionary artificial neural networks. Proc. IEEE 87(9), 1423–1447 (1999)CrossRef
20.
Zurück zum Zitat S. Samarasinghe, Optimum Structure of Feed Forward Neural Networks by SOM Clustering of Neuron Activations. Proceedings of the International Modelling and Simulation Congress (MODSM) (2007) S. Samarasinghe, Optimum Structure of Feed Forward Neural Networks by SOM Clustering of Neuron Activations. Proceedings of the International Modelling and Simulation Congress (MODSM) (2007)
21.
Zurück zum Zitat Neural Networks for Mathematica, (Wolfram Research, Inc. USA, 2002) Neural Networks for Mathematica, (Wolfram Research, Inc. USA, 2002)
22.
Zurück zum Zitat J. Sietsma, R.J.F. Dow, Creating artificial neural networks that generalize. Neural Networks 4(1), 67–77 (1991)CrossRef J. Sietsma, R.J.F. Dow, Creating artificial neural networks that generalize. Neural Networks 4(1), 67–77 (1991)CrossRef
24.
Zurück zum Zitat J.H. Ward Jr, Hierarchical grouping to optimize an objective function. J. Am Stat. Assoc. 58, 236–244 (1963)MathSciNetCrossRef J.H. Ward Jr, Hierarchical grouping to optimize an objective function. J. Am Stat. Assoc. 58, 236–244 (1963)MathSciNetCrossRef
25.
Zurück zum Zitat K. Hornik, M. Stinchcombe, H. White, Universal approximation of an unknown mapping and its derivatives using multi-layer feedforard networks. Neural Networks 3, 551–560 (1990)CrossRef K. Hornik, M. Stinchcombe, H. White, Universal approximation of an unknown mapping and its derivatives using multi-layer feedforard networks. Neural Networks 3, 551–560 (1990)CrossRef
26.
Zurück zum Zitat A.R. Gallant, H. White, On learning the derivative of an unknown mapping with multilayer feedforward networks. Neural Networks 5, 129–138 (1992)CrossRef A.R. Gallant, H. White, On learning the derivative of an unknown mapping with multilayer feedforward networks. Neural Networks 5, 129–138 (1992)CrossRef
27.
Zurück zum Zitat A. Al-yousef, S. Samarasinghe, Ultrasound based computer aided diagnosis of breast cancer: Evaluation of a new feature of mass central regularity degree. Proceedings of the International Modelling and Simulation Congress (MODSM) (2011) A. Al-yousef, S. Samarasinghe, Ultrasound based computer aided diagnosis of breast cancer: Evaluation of a new feature of mass central regularity degree. Proceedings of the International Modelling and Simulation Congress (MODSM) (2011)
28.
Zurück zum Zitat S. Samarasinghe, Hydrocomplexity: New Tools for Solving Wicked Water Problems Hydrocomplexité: Nouveaux outils pour solutionner des problèmes de l’eau complexes (IAHS Publ. 338) (2010) S. Samarasinghe, Hydrocomplexity: New Tools for Solving Wicked Water Problems Hydrocomplexité: Nouveaux outils pour solutionner des problèmes de l’eau complexes (IAHS Publ. 338) (2010)
Metadaten
Titel
Order in the Black Box: Consistency and Robustness of Hidden Neuron Activation of Feed Forward Neural Networks and Its Use in Efficient Optimization of Network Structure
verfasst von
Sandhya Samarasinghe
Copyright-Jahr
2016
DOI
https://doi.org/10.1007/978-3-319-28495-8_2