Skip to main content
Erschienen in: Neural Computing and Applications 3-4/2014

01.09.2014 | Original Article

Advanced learning methods and exponent regularization applied to a high order neural network

verfasst von: Islam El-Nabarawy, Ashraf M. Abdelbar

Erschienen in: Neural Computing and Applications | Ausgabe 3-4/2014

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

High order neural networks (HONNs) are neural networks which employ neurons that combine their inputs non-linearly. The high order network with exponential synaptic links (HONEST) network is a HONN that uses neurons with product units and adaptable exponents. This study examines the use of several advanced learning methods to train the HONEST network: resilient propagation, conjugate gradient, scaled conjugate gradient (SCG), and the Levenberg–Marquardt method. Using a collection of 32 widely-used benchmark datasets, we compare the mean squared error (MSE) performance of the HONEST network across the four algorithms, in addition to backpropagation, and find the SCG method to produce the best performance to a statistically significant extent. Additionally, we investigate the use of a regularization term in the error function, to smooth the magnitudes of the network exponents and nudge the network towards smaller exponents. We find that the use of regularization reduces exponent magnitudes without compromising test set MSE performance.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Rumelhart DE, Hinton GE, McClelland JL (1986) A general framework for parallel distributed processing. Parallel Distrib Process 1(2):45–76 Rumelhart DE, Hinton GE, McClelland JL (1986) A general framework for parallel distributed processing. Parallel Distrib Process 1(2):45–76
2.
Zurück zum Zitat Shin Y, Ghosh J (1991) The pi–sigma network: an efficient higher-order neural network for pattern classification and function approximation. In: Proceedings of the international joint conference on neural networks, IJCNN-91, vol 1, pp 13–18 Shin Y, Ghosh J (1991) The pi–sigma network: an efficient higher-order neural network for pattern classification and function approximation. In: Proceedings of the international joint conference on neural networks, IJCNN-91, vol 1, pp 13–18
3.
Zurück zum Zitat Durbin R, Rumelhart DE (1989) Product units: a computationally powerful and biologically plausible extension to backpropagation networks. Neural Comput 1(1):133–142CrossRef Durbin R, Rumelhart DE (1989) Product units: a computationally powerful and biologically plausible extension to backpropagation networks. Neural Comput 1(1):133–142CrossRef
4.
Zurück zum Zitat Fallahnezhad M, Moradi MH, Zaferanlouei S (2011) A hybrid higher order neural classifier for handling classification problems. Expert Syst Appl 38(1):386–393CrossRef Fallahnezhad M, Moradi MH, Zaferanlouei S (2011) A hybrid higher order neural classifier for handling classification problems. Expert Syst Appl 38(1):386–393CrossRef
5.
Zurück zum Zitat Abdelbar AM, Tagliarini GA (1996) HONEST: a new high order feedforward neural network. In: IEEE international conference on neural networks, vol 2, pp 1257–1262 Abdelbar AM, Tagliarini GA (1996) HONEST: a new high order feedforward neural network. In: IEEE international conference on neural networks, vol 2, pp 1257–1262
6.
Zurück zum Zitat Narayan S (1993) ExpoNet: A generalization of the multi-layer perceptron model. In: Proceedings of the World Congress on Neural Networks, vol 3, pp 494–497 Narayan S (1993) ExpoNet: A generalization of the multi-layer perceptron model. In: Proceedings of the World Congress on Neural Networks, vol 3, pp 494–497
7.
Zurück zum Zitat Rovithakis G, Chalkiadakis I, Zervakis M (2004) High-order neural network structure selection for function approximation applications using genetic algorithms. IEEE Trans Syst Man Cybern B Cybern 34(1):150–158CrossRef Rovithakis G, Chalkiadakis I, Zervakis M (2004) High-order neural network structure selection for function approximation applications using genetic algorithms. IEEE Trans Syst Man Cybern B Cybern 34(1):150–158CrossRef
8.
Zurück zum Zitat Martínez-Estudillo AC, Martínez-Estudillo FJ, Herváz-Martínez C, García-Pedrajas N (2006) Evolutionary product unit based neural networks for regression. Neural Netw 19(4):477–486MATHCrossRef Martínez-Estudillo AC, Martínez-Estudillo FJ, Herváz-Martínez C, García-Pedrajas N (2006) Evolutionary product unit based neural networks for regression. Neural Netw 19(4):477–486MATHCrossRef
9.
Zurück zum Zitat Giles CL, Maxwell T (1987) Learning, invariance, and generalization in high-order neural networks. Appl Opt 26(23):4972–4978CrossRef Giles CL, Maxwell T (1987) Learning, invariance, and generalization in high-order neural networks. Appl Opt 26(23):4972–4978CrossRef
10.
Zurück zum Zitat Abdelbar AM (1998) Achieving superior generalisation with a high order neural network. Neural Comput Appl 7(2):141–146MATHCrossRef Abdelbar AM (1998) Achieving superior generalisation with a high order neural network. Neural Comput Appl 7(2):141–146MATHCrossRef
11.
Zurück zum Zitat Abdelbar AM, Attia S, Tagliarini GA (2002) A hybridization of Bayesian and neural learning. Neurocomputing 48(1):443–453MATHCrossRef Abdelbar AM, Attia S, Tagliarini GA (2002) A hybridization of Bayesian and neural learning. Neurocomputing 48(1):443–453MATHCrossRef
12.
Zurück zum Zitat Tsai HC (2009) Hybrid high order neural networks. Appl Soft Comput 9(3):874–881CrossRef Tsai HC (2009) Hybrid high order neural networks. Appl Soft Comput 9(3):874–881CrossRef
13.
Zurück zum Zitat Tsai HC (2010) Predicting strengths of concrete-type specimens using hybrid multilayer perceptrons with center-unified particle swarm optimization. Expert Syst Appl 37(2):1104–1112CrossRef Tsai HC (2010) Predicting strengths of concrete-type specimens using hybrid multilayer perceptrons with center-unified particle swarm optimization. Expert Syst Appl 37(2):1104–1112CrossRef
14.
Zurück zum Zitat Van Den Bergh F, Engelbrecht AP (2001) Training product unit networks using cooperative particle swarm optimisers. In: Proceedings of the international joint conference on neural networks, IJCNN-01, vol 1, pp 126–131 Van Den Bergh F, Engelbrecht AP (2001) Training product unit networks using cooperative particle swarm optimisers. In: Proceedings of the international joint conference on neural networks, IJCNN-01, vol 1, pp 126–131
15.
Zurück zum Zitat Riedmiller M, Braun H (1993) A direct adaptive method for faster backpropagation learning: the RPROP algorithm. In: IEEE international conference on neural networks, pp 586–591 Riedmiller M, Braun H (1993) A direct adaptive method for faster backpropagation learning: the RPROP algorithm. In: IEEE international conference on neural networks, pp 586–591
16.
Zurück zum Zitat Hestenes MR, Stiefel E (1952) Methods of conjugate gradients for solving linear systems. J Res Natl Bur Stand 49(6):409–436MATHMathSciNetCrossRef Hestenes MR, Stiefel E (1952) Methods of conjugate gradients for solving linear systems. J Res Natl Bur Stand 49(6):409–436MATHMathSciNetCrossRef
17.
Zurück zum Zitat Møller MF (1993) A scaled conjugate gradient algorithm for fast supervised learning. Neural Netw 6(4):525–533CrossRef Møller MF (1993) A scaled conjugate gradient algorithm for fast supervised learning. Neural Netw 6(4):525–533CrossRef
18.
Zurück zum Zitat Levenberg K (1944) A method for the solution of certain problems in least squares. Q Appl Math 2:164–168MATHMathSciNet Levenberg K (1944) A method for the solution of certain problems in least squares. Q Appl Math 2:164–168MATHMathSciNet
19.
20.
Zurück zum Zitat Hagan MT, Menhaj MB (1994) Training feedforward networks with the Marquardt algorithm. IEEE Trans Neural Netw 5(6):989–993CrossRef Hagan MT, Menhaj MB (1994) Training feedforward networks with the Marquardt algorithm. IEEE Trans Neural Netw 5(6):989–993CrossRef
21.
Zurück zum Zitat Suratgar AA, Tavakoli MB, Hoseinabadi A (2005) Modified Levenberg–Marquardt method for neural networks training. World Acad Sci Eng Technol 6:46–48 Suratgar AA, Tavakoli MB, Hoseinabadi A (2005) Modified Levenberg–Marquardt method for neural networks training. World Acad Sci Eng Technol 6:46–48
22.
Zurück zum Zitat Tikhonov A (1963) Solution of incorrectly formulated problems and the regularization method. Sov Math Dokl 5:1035MATH Tikhonov A (1963) Solution of incorrectly formulated problems and the regularization method. Sov Math Dokl 5:1035MATH
23.
Zurück zum Zitat Tsai HC, Wu YW, Tyan YY (2013) Programming squat wall strengths and tuning associated codes with pruned modular neural network. Neural Comput Appl 23(3–4):741–749CrossRef Tsai HC, Wu YW, Tyan YY (2013) Programming squat wall strengths and tuning associated codes with pruned modular neural network. Neural Comput Appl 23(3–4):741–749CrossRef
24.
25.
Zurück zum Zitat Puig V, Witczak M, Nejjari F, Quevedo J, Korbicz J (2007) A GMDH neural network-based approach to passive robust fault detection using a constraint satisfaction backward test. Eng Appl Artif Intell 20(7):886–897CrossRef Puig V, Witczak M, Nejjari F, Quevedo J, Korbicz J (2007) A GMDH neural network-based approach to passive robust fault detection using a constraint satisfaction backward test. Eng Appl Artif Intell 20(7):886–897CrossRef
26.
Zurück zum Zitat Werbos P (1974) Beyond regression: new tools for prediction and analysis in the behavioral sciences. PhD thesis, Committee on Applied Math, Harvard University, Cambridge, MA Werbos P (1974) Beyond regression: new tools for prediction and analysis in the behavioral sciences. PhD thesis, Committee on Applied Math, Harvard University, Cambridge, MA
27.
Zurück zum Zitat Werbos PJ (1994) The roots of backpropagation: from ordered derivatives to neural networks and political forecasting. Wiley-Interscience, New York Werbos PJ (1994) The roots of backpropagation: from ordered derivatives to neural networks and political forecasting. Wiley-Interscience, New York
28.
Zurück zum Zitat Chauvin Y, Rumelhart DE (1995) Backpropagation: theory, architectures, and applications. Lawrence Erlbaum, Hillsdale Chauvin Y, Rumelhart DE (1995) Backpropagation: theory, architectures, and applications. Lawrence Erlbaum, Hillsdale
29.
Zurück zum Zitat Werbos P (1968) The elements of intelligence. Cybernetica (Namur) 3:131–178 Werbos P (1968) The elements of intelligence. Cybernetica (Namur) 3:131–178
30.
Zurück zum Zitat Werbos PJ (1990) Backpropagation through time: what it does and how to do it. Proc IEEE 78(10):1550–1560CrossRef Werbos PJ (1990) Backpropagation through time: what it does and how to do it. Proc IEEE 78(10):1550–1560CrossRef
31.
Zurück zum Zitat Fletcher R (1987) Practical methods of optimization, 2nd edn. Wiley, New YorkMATH Fletcher R (1987) Practical methods of optimization, 2nd edn. Wiley, New YorkMATH
32.
33.
Zurück zum Zitat Falas T, Stafylopatis A (2005) Implementing temporal-difference learning with the scaled conjugate gradient algorithm. Neural Process Lett 22:361–375CrossRef Falas T, Stafylopatis A (2005) Implementing temporal-difference learning with the scaled conjugate gradient algorithm. Neural Process Lett 22:361–375CrossRef
34.
Zurück zum Zitat Lundén J, Koivunen V (2007) Scaled conjugate gradient method for radar pulse modulation estimation. In: IEEE international conference on acoustics, speech, and signal processing, vol 2, pp 297–300 Lundén J, Koivunen V (2007) Scaled conjugate gradient method for radar pulse modulation estimation. In: IEEE international conference on acoustics, speech, and signal processing, vol 2, pp 297–300
35.
Zurück zum Zitat Mehrotra K, Mohan CK, Ranka S (1996) Elements of artificial neural networks. MIT press, Cambridge Mehrotra K, Mohan CK, Ranka S (1996) Elements of artificial neural networks. MIT press, Cambridge
36.
Zurück zum Zitat Haykin SS (2009) Neural networks and learning machines. Prentice Hall, Upper Saddle River Haykin SS (2009) Neural networks and learning machines. Prentice Hall, Upper Saddle River
38.
Zurück zum Zitat Pearson K (1901) Principal components analysis. Lond Edinb Dublin Philos Mag J 6(2):566 Pearson K (1901) Principal components analysis. Lond Edinb Dublin Philos Mag J 6(2):566
39.
Zurück zum Zitat Hall M, Frank E, Holmes G, Pfahringer B, Reutemann P, Witten IH (2009) The WEKA data mining software: an update. ACM SIGKDD Explor Newslett 11(1):10–18CrossRef Hall M, Frank E, Holmes G, Pfahringer B, Reutemann P, Witten IH (2009) The WEKA data mining software: an update. ACM SIGKDD Explor Newslett 11(1):10–18CrossRef
40.
Zurück zum Zitat Wilcoxon F (1945) Individual comparisons by ranking methods. Biom Bull 1(6):80–83CrossRef Wilcoxon F (1945) Individual comparisons by ranking methods. Biom Bull 1(6):80–83CrossRef
41.
Zurück zum Zitat Demšar J (2006) Statistical comparisons of classifiers over multiple data sets. J Mac Learn Res 7:1–30MATH Demšar J (2006) Statistical comparisons of classifiers over multiple data sets. J Mac Learn Res 7:1–30MATH
42.
Zurück zum Zitat García S, Herrera F (2008) An extension on statistical comparisons of classifiers over multiple data sets for all pairwise comparisons. J Mach Learn Res 9(66):2677–2694MATH García S, Herrera F (2008) An extension on statistical comparisons of classifiers over multiple data sets for all pairwise comparisons. J Mach Learn Res 9(66):2677–2694MATH
43.
Zurück zum Zitat Wolberg WH, Mangasarian OL (1990) Multisurface method of pattern separation for medical diagnosis applied to breast cytology. Proc Natl Acad Sci 87(23):9193–9196MATHCrossRef Wolberg WH, Mangasarian OL (1990) Multisurface method of pattern separation for medical diagnosis applied to breast cytology. Proc Natl Acad Sci 87(23):9193–9196MATHCrossRef
44.
Zurück zum Zitat Yeh IC (1998) Modeling of strength of high-performance concrete using artificial neural networks. Cem Concr Res 28(12):1797–1808CrossRef Yeh IC (1998) Modeling of strength of high-performance concrete using artificial neural networks. Cem Concr Res 28(12):1797–1808CrossRef
45.
Zurück zum Zitat Tsanas A, Xifara A (2012) Accurate quantitative estimation of energy performance of residential buildings using statistical machine learning tools. Energy Build 49:560–567CrossRef Tsanas A, Xifara A (2012) Accurate quantitative estimation of energy performance of residential buildings using statistical machine learning tools. Energy Build 49:560–567CrossRef
46.
Zurück zum Zitat Gil D, Girela JL, De Juan J, Jose Gomez-Torres M, Johnsson M (2012) Predicting seminal quality with artificial intelligence methods. Expert Syst Appl 39(16):12,564–12,573CrossRef Gil D, Girela JL, De Juan J, Jose Gomez-Torres M, Johnsson M (2012) Predicting seminal quality with artificial intelligence methods. Expert Syst Appl 39(16):12,564–12,573CrossRef
47.
Zurück zum Zitat Elter M, Schulz-Wendtland R, Wittenberg T (2007) The prediction of breast cancer biopsy outcomes using two CAD approaches that both emphasize an intelligible decision process. Med Phys 34(11):4164–4172CrossRef Elter M, Schulz-Wendtland R, Wittenberg T (2007) The prediction of breast cancer biopsy outcomes using two CAD approaches that both emphasize an intelligible decision process. Med Phys 34(11):4164–4172CrossRef
48.
Zurück zum Zitat Hwang SJ, Fang WH, Lee HJ, Yu HW (2001) Analytical model for predicting shear strength of squat walls. J Struct Eng 127(1):43–50CrossRef Hwang SJ, Fang WH, Lee HJ, Yu HW (2001) Analytical model for predicting shear strength of squat walls. J Struct Eng 127(1):43–50CrossRef
49.
Zurück zum Zitat Yeh IC, Yang KJ, Ting TM (2009) Knowledge discovery on RFM model using Bernoulli sequence. Expert Syst Appl 36(3):5866–5871CrossRef Yeh IC, Yang KJ, Ting TM (2009) Knowledge discovery on RFM model using Bernoulli sequence. Expert Syst Appl 36(3):5866–5871CrossRef
50.
Zurück zum Zitat Cortez P, Cerdeira A, Almeida F, Matos T, Reis J (2009) Modeling wine preferences by data mining from physicochemical properties. Decis Support Syst 47(4):547–553CrossRef Cortez P, Cerdeira A, Almeida F, Matos T, Reis J (2009) Modeling wine preferences by data mining from physicochemical properties. Decis Support Syst 47(4):547–553CrossRef
Metadaten
Titel
Advanced learning methods and exponent regularization applied to a high order neural network
verfasst von
Islam El-Nabarawy
Ashraf M. Abdelbar
Publikationsdatum
01.09.2014
Verlag
Springer London
Erschienen in
Neural Computing and Applications / Ausgabe 3-4/2014
Print ISSN: 0941-0643
Elektronische ISSN: 1433-3058
DOI
https://doi.org/10.1007/s00521-014-1563-7

Weitere Artikel der Ausgabe 3-4/2014

Neural Computing and Applications 3-4/2014 Zur Ausgabe