Skip to main content
Erschienen in: Neural Computing and Applications 11/2020

13.02.2020 | S.I. : Brain inspired Computing&Machine Learning Applied Research-BISMLARE

An advanced active set L-BFGS algorithm for training weight-constrained neural networks

verfasst von: Ioannis E. Livieris

Erschienen in: Neural Computing and Applications | Ausgabe 11/2020

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

In this work, a new advanced active set limited memory BFGS (Broyden–Fletcher–Goldfarb–Shanno) algorithm is proposed for efficiently training weight-constrained neural networks, called AA-L-BFGS. The proposed algorithm possesses the significant property of approximating the curvature of the error function with high-order accuracy by utilizing the theoretically advanced secant condition proposed by Livieris and Pintelas (Appl Math Comput 221:491–502, 2013). Moreover, the global convergence of the proposed algorithm is established provided that the line search satisfies the modified Armijo condition. The presented numerical experiments illustrate the efficiency of the proposed AA-L-BFGS, providing empirical evidence that it significantly accelerates the convergence of the training process.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Anastasiadis AD, Magoulas GD, Vrahatis MN (2005) New globally convergent training scheme based on the resilient propagation algorithm. Neurocomputing 64:253–270CrossRef Anastasiadis AD, Magoulas GD, Vrahatis MN (2005) New globally convergent training scheme based on the resilient propagation algorithm. Neurocomputing 64:253–270CrossRef
2.
Zurück zum Zitat Awan SM, Aslam M, Khan ZA, Saeed H (2014) An efficient model based on artificial bee colony optimization algorithm with neural networks for electric load forecasting. Neural Comput Appl 25(7–8):1967–1978CrossRef Awan SM, Aslam M, Khan ZA, Saeed H (2014) An efficient model based on artificial bee colony optimization algorithm with neural networks for electric load forecasting. Neural Comput Appl 25(7–8):1967–1978CrossRef
3.
Zurück zum Zitat Azar AT (2013) Fast neural network learning algorithms for medical applications. Neural Comput Appl 23(3–4):1019–1034CrossRef Azar AT (2013) Fast neural network learning algorithms for medical applications. Neural Comput Appl 23(3–4):1019–1034CrossRef
4.
Zurück zum Zitat Badem H, Basturk A, Caliskan A, Yuksel ME (2017) A new efficient training strategy for deep neural networks by hybridization of artificial bee colony and limited-memory BFGS optimization algorithms. Neurocomputing 266:506–526CrossRef Badem H, Basturk A, Caliskan A, Yuksel ME (2017) A new efficient training strategy for deep neural networks by hybridization of artificial bee colony and limited-memory BFGS optimization algorithms. Neurocomputing 266:506–526CrossRef
5.
Zurück zum Zitat Bilski K, Smolag J, Galushkin AI (2014) The parallel approach to the conjugate gradient learning algorithm for the feedforward neural networks. In: International conference on artificial intelligence and soft computing. Springer, pp 12–21 Bilski K, Smolag J, Galushkin AI (2014) The parallel approach to the conjugate gradient learning algorithm for the feedforward neural networks. In: International conference on artificial intelligence and soft computing. Springer, pp 12–21
6.
Zurück zum Zitat Demertzis K, Iliadis L (2015) Intelligent bio-inspired detection of food borne pathogen by DNA barcodes: the case of invasive fish species Lagocephalus Sceleratus. In: International conference on engineering applications of neural networks. Springer, pp 89–99 Demertzis K, Iliadis L (2015) Intelligent bio-inspired detection of food borne pathogen by DNA barcodes: the case of invasive fish species Lagocephalus Sceleratus. In: International conference on engineering applications of neural networks. Springer, pp 89–99
7.
Zurück zum Zitat Dolan E, Moré JJ (2002) Benchmarking optimization software with performance profiles. Math Program 91:201–213MathSciNetCrossRef Dolan E, Moré JJ (2002) Benchmarking optimization software with performance profiles. Math Program 91:201–213MathSciNetCrossRef
8.
Zurück zum Zitat Dua D, Taniskidou EK (2017) UCI machine learning repository Dua D, Taniskidou EK (2017) UCI machine learning repository
9.
Zurück zum Zitat Erzin Y, Gul TO (2014) The use of neural networks for the prediction of the settlement of one-way footings on cohesionless soils based on standard penetration test. Neural Comput Appl 24(3–4):891–900CrossRef Erzin Y, Gul TO (2014) The use of neural networks for the prediction of the settlement of one-way footings on cohesionless soils based on standard penetration test. Neural Comput Appl 24(3–4):891–900CrossRef
10.
Zurück zum Zitat Facchinei F, Júdice J, Soares J (1998) An active set Newton algorithm for large-scale nonlinear programs with box constraints. SIAM J Optim 8(1):158–186MathSciNetCrossRef Facchinei F, Júdice J, Soares J (1998) An active set Newton algorithm for large-scale nonlinear programs with box constraints. SIAM J Optim 8(1):158–186MathSciNetCrossRef
11.
Zurück zum Zitat Gatys LA, Ecker AS, Bethge M (2015) Texture synthesis using convolutional neural networks. In: Proceedings of the 28th international conference on neural information processing systems, vol 1, pp 262–270 Gatys LA, Ecker AS, Bethge M (2015) Texture synthesis using convolutional neural networks. In: Proceedings of the 28th international conference on neural information processing systems, vol 1, pp 262–270
12.
Zurück zum Zitat Hertz J, Krogh A, Palmer R (1991) Introduction to the theory of neural computation. Addison-Wesley, Reading Hertz J, Krogh A, Palmer R (1991) Introduction to the theory of neural computation. Addison-Wesley, Reading
13.
Zurück zum Zitat Horton P, Nakai K (1996) A probabilistic classification system for predicting the cellular localization sites of proteins. Ismb 4:109–115 Horton P, Nakai K (1996) A probabilistic classification system for predicting the cellular localization sites of proteins. Ismb 4:109–115
14.
Zurück zum Zitat Iliadis L, Mansfield SD, Avramidis S, El-Kassaby YA (2013) Predicting Douglas-fir wood density by artificial neural networks (ANN) based on progeny testing information. Holzforschung 67(7):771–777CrossRef Iliadis L, Mansfield SD, Avramidis S, El-Kassaby YA (2013) Predicting Douglas-fir wood density by artificial neural networks (ANN) based on progeny testing information. Holzforschung 67(7):771–777CrossRef
15.
Zurück zum Zitat Karras DA, Perantonis SJ (1995) An efficient constrained training algorithm for feedforward networks. IEEE Trans Neural Netw 6(6):1420–1434CrossRef Karras DA, Perantonis SJ (1995) An efficient constrained training algorithm for feedforward networks. IEEE Trans Neural Netw 6(6):1420–1434CrossRef
16.
Zurück zum Zitat Kayaer K, Yıldırım T (2003) Medical diagnosis on Pima Indian diabetes using general regression neural networks. In: Proceedings of the international conference on artificial neural networks and neural information processing (ICANN/ICONIP), pp 181–184 Kayaer K, Yıldırım T (2003) Medical diagnosis on Pima Indian diabetes using general regression neural networks. In: Proceedings of the international conference on artificial neural networks and neural information processing (ICANN/ICONIP), pp 181–184
17.
Zurück zum Zitat Khadse CB, Chaudhari MA, Borghate VB (2016) Conjugate gradient back-propagation based artificial neural network for real time power quality assessment. Int J Electr Power Energy Syst 82:197–206CrossRef Khadse CB, Chaudhari MA, Borghate VB (2016) Conjugate gradient back-propagation based artificial neural network for real time power quality assessment. Int J Electr Power Energy Syst 82:197–206CrossRef
18.
Zurück zum Zitat Kostić S, Vasović D (2015) Prediction model for compressive strength of basic concrete mixture using artificial neural networks. Neural Comput Appl 26(5):1005–1024CrossRef Kostić S, Vasović D (2015) Prediction model for compressive strength of basic concrete mixture using artificial neural networks. Neural Comput Appl 26(5):1005–1024CrossRef
19.
Zurück zum Zitat Li F, Zhang X, Zhang X, Du C, Xu Y, Tian YC (2018) Cost-sensitive and hybrid-attribute measure multi-decision tree over imbalanced data sets. Inf Sci 422:242–256CrossRef Li F, Zhang X, Zhang X, Du C, Xu Y, Tian YC (2018) Cost-sensitive and hybrid-attribute measure multi-decision tree over imbalanced data sets. Inf Sci 422:242–256CrossRef
20.
Zurück zum Zitat Liu DC, Nocedal J (1989) On the limited memory BFGS method for large scale optimization. Math Program 45(1–3):503–528MathSciNetCrossRef Liu DC, Nocedal J (1989) On the limited memory BFGS method for large scale optimization. Math Program 45(1–3):503–528MathSciNetCrossRef
21.
Zurück zum Zitat Livieris IE (2018) Improving the classification efficiency of an ANN utilizing a new training methodology. Informatics 6(1):1CrossRef Livieris IE (2018) Improving the classification efficiency of an ANN utilizing a new training methodology. Informatics 6(1):1CrossRef
22.
Zurück zum Zitat Livieris IE (2019) Forecasting economy-related data utilizing constrained recurrent neural networks. Algorithms 12(85):1MathSciNetMATH Livieris IE (2019) Forecasting economy-related data utilizing constrained recurrent neural networks. Algorithms 12(85):1MathSciNetMATH
23.
Zurück zum Zitat Livieris IE, Pintelas P (2013) A new conjugate gradient algorithm for training neural networks based on a modified secant equation. Appl Math Comput 221:491–502MathSciNetMATH Livieris IE, Pintelas P (2013) A new conjugate gradient algorithm for training neural networks based on a modified secant equation. Appl Math Comput 221:491–502MathSciNetMATH
24.
Zurück zum Zitat Livieris IE, Pintelas P (2015) A new class of nonmonotone conjugate gradient training algorithms. Appl Math Comput 266:404–413MathSciNetMATH Livieris IE, Pintelas P (2015) A new class of nonmonotone conjugate gradient training algorithms. Appl Math Comput 266:404–413MathSciNetMATH
27.
Zurück zum Zitat Maren AJ, Harston CT, Pap RM (2014) Handbook of neural computing applications. Academic Press, CambridgeMATH Maren AJ, Harston CT, Pap RM (2014) Handbook of neural computing applications. Academic Press, CambridgeMATH
28.
Zurück zum Zitat Nguyen D, Widrow B (1990) Improving the learning speed of 2-layer neural network by choosing initial values of adaptive weights. Biol Cybern 59:71–113 Nguyen D, Widrow B (1990) Improving the learning speed of 2-layer neural network by choosing initial values of adaptive weights. Biol Cybern 59:71–113
29.
Zurück zum Zitat Noordewier MO, Towell GG, Shavlik JW (1990) Training knowledge-based neural networks to recognize genes in DNA sequences. In: Proceedings of the 3rd international conference on neural information processing systems, pp 530–536 Noordewier MO, Towell GG, Shavlik JW (1990) Training knowledge-based neural networks to recognize genes in DNA sequences. In: Proceedings of the 3rd international conference on neural information processing systems, pp 530–536
30.
Zurück zum Zitat Perantonis SJ, Karras DA (1995) An efficient constrained learning algorithm with momentum acceleration. Neural Netw 8(2):237–249CrossRef Perantonis SJ, Karras DA (1995) An efficient constrained learning algorithm with momentum acceleration. Neural Netw 8(2):237–249CrossRef
31.
Zurück zum Zitat Prechelt L (1994) PROBEN1-A set of benchmarks and benchmarking rules for neural network training algorithms. Technical report 21/94, Fakultt fr Informatik, University of Karlsruhe Prechelt L (1994) PROBEN1-A set of benchmarks and benchmarking rules for neural network training algorithms. Technical report 21/94, Fakultt fr Informatik, University of Karlsruhe
32.
Zurück zum Zitat Shi ZJ, Wang S (2011) Modified nonmonotone Armijo line search for descent method. Numer Algorithms 57(1):1–25MathSciNetCrossRef Shi ZJ, Wang S (2011) Modified nonmonotone Armijo line search for descent method. Numer Algorithms 57(1):1–25MathSciNetCrossRef
33.
Zurück zum Zitat Wan Zhong, Huang Shuai, Zheng Xiao Dong (2012) New cautious BFGS algorithm based on modified Armijo-type line search. J Inequal Appl 2012(1):241MathSciNetCrossRef Wan Zhong, Huang Shuai, Zheng Xiao Dong (2012) New cautious BFGS algorithm based on modified Armijo-type line search. J Inequal Appl 2012(1):241MathSciNetCrossRef
34.
Zurück zum Zitat Yuan G, Lu X (2011) An active set limited memory BFGS algorithm for bound constrained optimization. Appl Math Model 35:3561–3573MathSciNetCrossRef Yuan G, Lu X (2011) An active set limited memory BFGS algorithm for bound constrained optimization. Appl Math Model 35:3561–3573MathSciNetCrossRef
Metadaten
Titel
An advanced active set L-BFGS algorithm for training weight-constrained neural networks
verfasst von
Ioannis E. Livieris
Publikationsdatum
13.02.2020
Verlag
Springer London
Erschienen in
Neural Computing and Applications / Ausgabe 11/2020
Print ISSN: 0941-0643
Elektronische ISSN: 1433-3058
DOI
https://doi.org/10.1007/s00521-019-04689-6

Weitere Artikel der Ausgabe 11/2020

Neural Computing and Applications 11/2020 Zur Ausgabe

Multi-Source Data Understanding (MSDU)

Mutual kNN based spectral clustering