Skip to main content
Log in

An analysis of numerical issues in neural training by pseudoinversion

  • Published:
Computational and Applied Mathematics Aims and scope Submit manuscript

Abstract

Some novel strategies have recently been proposed for single hidden layer neural network training that set randomly the weights from input to hidden layer, while weights from hidden to output layer are analytically determined by pseudoinversion. These techniques are gaining popularity in spite of their known numerical issues when singular and/or almost singular matrices are involved. In this paper, we discuss a critical use of Singular Value Analysis for identification of these drawbacks and we propose an original use of regularisation to determine the output weights, based on the concept of critical hidden layer size. This approach also allows to limit the training computational effort. Besides, we introduce a novel technique which relies an effective determination of input weights to the hidden layer dimension. This approach is tested for both regression and classification tasks, resulting in a significant performance improvement with respect to alternative methods.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3

Similar content being viewed by others

References

  • Ajorloo H, Manzuri-Shalmani MT, Lakdashti A (2007) Restoration of damaged slices in images using matrix pseudo inversion. In: Proceedings of the 22nd international symposium on computer and information sciences

  • Bache K, Lichman M (2013) UCI machine learning repository, University of California, School of Information and Computer Sciences, Irvine. http://archive.ics.uci.edu/ml

  • Badeva V, Morozov V (1991) Problèmes incorrectement posés: théorie et applications en identification, filtrage optimal, contrôle optimal, analyse et synthèse de systèmes, reconnaissance d’images. Série Automatique, Masson

    Google Scholar 

  • Bengio Y, Glorot X (2010) Understanding the difficulty of training deep feedforward neural networks. In: Proceedings of AISTATS 2010, vol 9, pp 249–256

  • Bishop CM (2006) Pattern recognition and machine learning (information science and statistics). Springer, Secaucus

    MATH  Google Scholar 

  • Cancelliere R (2001) A high parallel procedure to initialize the output weights of a radial basis function or bp neural network. In: Proceedings of the 5th international workshop on applied parallel computing, New Paradigms for HPC in industry and academia, PARA ’00. Springer, London, pp 384–390

  • Cancelliere R, Gai M, Artières T, Gallinari P (2012) Matrix pseudoinversion for image neural processing. In: Huang T, Zeng Z, Li C, Leung C (eds) Neural information processing, Lecture notes in computer science, vol 7667. Springer, Berlin, pp 116–125

  • Fuhry M, Reichel L (2012) A new tikhonov regularization method. Numer Algorithms 59(3):433–445

    Article  MathSciNet  MATH  Google Scholar 

  • Gallinari P, Cibas T (1999) Practical complexity control in multilayer perceptrons. Sig Process 74:29–46

    Article  MATH  Google Scholar 

  • Golub GH, Van Loan CF (1996) Matrix computations, 3rd edn. Johns Hopkins University Press, Baltimore

    MATH  Google Scholar 

  • Helmy T, Rasheed Z (2009) Multi-category bioinformatics dataset classification using extreme learning machine. In: Proceedings of the eleventh conference on congress on evolutionary computation, CEC’09, Piscataway, NJ, USA, IEEE Press, pp 3234–3240

  • Huang G-B, Zhu Q-Y, Siew C-K (2006) Extreme learning machine: theory and applications. Neurocomputing 70(1):489–501

    Article  Google Scholar 

  • Kohno K, Kawamoto M, Inouye Y (2010) A matrix pseudoinversion lemma and its application to block-based adaptive blind deconvolution for mimo systems. Trans Cir Sys Part I 57(7):1449–1462

    MathSciNet  Google Scholar 

  • LeCun Y, Bottou L, Orr G, Müller K-R (1998) Efficient backprop. Springer, Berlin, pp 9–50

  • Malathi V, Marimuthu N, Baskar S (2010) Intelligent approaches using support vector machine and extreme learning machine for transmission line protection. Neurocomputing 73(1012):2160–2167 Subspace Learning/Selected papers from the European Symposium on Time Series Prediction

    Article  Google Scholar 

  • Miche Y, van Heeswijk M, Bas P, Simula O, Lendasse A (2011) Trop-elm: a double-regularized elm using lars and tikhonov regularization. Neurocomputing 74(16):2413–2421

    Article  Google Scholar 

  • Minhas R, Baradarani A, Seifzadeh S, Wu QJ (2010) Human action recognition using extreme learning machine based on visual vocabularies. Neurocomputing 73(1012):1906–1917 Subspace Learning/Selected papers from the European Symposium on Time Series Prediction

    Article  Google Scholar 

  • Nguyen TD, Pham HTB, Dang VH (2010) An efficient pseudo inverse matrix-based solution for secure auditing. In: Proceedings of the IEEE international conference on computing and communication technologies, research, innovation, and vision for the future, IEEE International Conference

  • Penrose R, Todd JA (1956) On best approximate solutions of linear matrix equations. In: Mathematical Proceedings of the Cambridge Philosophical Society, null, pp 17–19

  • Poggio T, Girosi F (1990) Networks for approximation and learning. Proc IEEE 78(9):1481–1497

    Article  MATH  Google Scholar 

  • Rao C, Mitra S (1971) Generalized inverse of matrices and its applications. In: Wiley Series in Probability and Mathematical Statistics: Applied Probability and Statistics. Wiley, New York

  • Rumelhart DE, Hinton GE, Williams RJ (1986) Learning internal representations by error propagation. In: Parallel Distributed Processing: Explorations in the Microstructure of Cognition, vol 1. MIT Press, Cambridge, pp 318–362

  • Sun Z-L, Choi T-M, Au K-F, Yu Y (2008) Sales forecasting using extreme learning machine with applications in fashion retailing. Dec Support Syst 46(1):411–419

    Article  Google Scholar 

  • Tikhonov AN (1963) Solution of incorrectly formulated problems and the regularization method. Sov Math Dokl 4:1035–1038

    MATH  Google Scholar 

  • Tikhonov AN, Arsenin VY (1977) Solution of Ill-posed Problems. In: Scripta series in mathematics. Winston & Sons, Washington

  • Wang G, Zhao Y, Wang D (2008) A protein secondary structure prediction framework based on the extreme learning machine. Neurocomputing 72(13):262–268 Machine Learning for Signal Processing (MLSP 2006)/Life System Modelling, Simulation, and Bio-inspired Computing (LSMS 2007)

    Article  Google Scholar 

  • Yu D, Deng L (2012) Efficient and effective algorithms for training single-hidden-layer neural networks. Pattern Recogn Lett 33(5):554–558

    Article  MathSciNet  Google Scholar 

Download references

Acknowledgments

The activity has been partially carried on in the context of the Visiting Professor Program of the Gruppo Nazionale per il Calcolo Scientifico (GNCS) of the Italian Istituto Nazionale di Alta Matematica (INdAM).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to R. Cancelliere.

Additional information

Communicated by José Mario Martínez.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Cancelliere, R., Deluca, R., Gai, M. et al. An analysis of numerical issues in neural training by pseudoinversion. Comp. Appl. Math. 36, 599–609 (2017). https://doi.org/10.1007/s40314-015-0246-z

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s40314-015-0246-z

Keywords

Mathematics Subject Classification

Navigation