Skip to main content
Erschienen in: Natural Computing 2/2009

01.06.2009

Negative correlation in incremental learning

verfasst von: Fernanda Li Minku, Hirotaka Inoue, Xin Yao

Erschienen in: Natural Computing | Ausgabe 2/2009

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Negative Correlation Learning (NCL) has been successfully applied to construct neural network ensembles. It encourages the neural networks that compose the ensemble to be different from each other and, at the same time, accurate. The difference among the neural networks that compose an ensemble is a desirable feature to perform incremental learning, for some of the neural networks can be able to adapt faster and better to new data than the others. So, NCL is a potentially powerful approach to incremental learning. With this in mind, this paper presents an analysis of NCL, aiming at determining its weak and strong points to incremental learning. The analysis shows that it is possible to use NCL to overcome catastrophic forgetting, an important problem related to incremental learning. However, when catastrophic forgetting is very low, no advantage of using more than one neural network of the ensemble to learn new data is taken and the test error is high. When all the neural networks are used to learn new data, some of them can indeed adapt better than the others, but a higher catastrophic forgetting is obtained. In this way, it is important to find a trade-off between overcoming catastrophic forgetting and using an entire ensemble to learn new data. The NCL results are comparable with other approaches which were specifically designed to incremental learning. Thus, the study presented in this work reveals encouraging results with negative correlation in incremental learning, showing that NCL is a promising approach to incremental learning.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Fußnoten
1
During negative correlation learning, simple average is used to combine the neural network outputs. However, the combination method used by the ensemble during the test phase can be another one, e.g., majority vote.
 
2
It is possible that a lower number of nodes either benefit or prejudice the result of the learning, depending on the database.
 
Literatur
Zurück zum Zitat Adamczak R, Duch W, Jankowski N (1997) New developments in the feature space mapping model. In: Proceedings of the third conference on neural networks and their applications, Kule, Poland, pp 65–70 Adamczak R, Duch W, Jankowski N (1997) New developments in the feature space mapping model. In: Proceedings of the third conference on neural networks and their applications, Kule, Poland, pp 65–70
Zurück zum Zitat Brown G, Wyatt JL, Tiño P (2005) Managing diversity in regression ensembles. J Mach Learn Res 6:1621–1650MathSciNet Brown G, Wyatt JL, Tiño P (2005) Managing diversity in regression ensembles. J Mach Learn Res 6:1621–1650MathSciNet
Zurück zum Zitat Carpenter GA, Grossberg S, Reynolds JH (1991) ARTMAP: supervied real-time learning and classification of nonstationary data by a self organizing neural network. Neural Networks 4(5):565–588CrossRef Carpenter GA, Grossberg S, Reynolds JH (1991) ARTMAP: supervied real-time learning and classification of nonstationary data by a self organizing neural network. Neural Networks 4(5):565–588CrossRef
Zurück zum Zitat Carpenter GA, Grossberg S, Markuzon N, Reynolds JH (1992) Fuzzy ARTMAP: a neural network architecture for incremental supervised learning of analog multidimensional maps. IEEE Trans Neural Networks 3:698–713CrossRef Carpenter GA, Grossberg S, Markuzon N, Reynolds JH (1992) Fuzzy ARTMAP: a neural network architecture for incremental supervised learning of analog multidimensional maps. IEEE Trans Neural Networks 3:698–713CrossRef
Zurück zum Zitat Chandra A, Yao X (2006) Evolving hybrid ensembles of learning machines for better generalisation. Neurocomputing 69:686–700CrossRef Chandra A, Yao X (2006) Evolving hybrid ensembles of learning machines for better generalisation. Neurocomputing 69:686–700CrossRef
Zurück zum Zitat Chandra A, Chen H, Yao X (2006) Trade-off between diversity and accuracy in ensemble generation. In: Jin Y (ed) Multi-objective machine learning. Springer-Verlag, pp 429–464 Chandra A, Chen H, Yao X (2006) Trade-off between diversity and accuracy in ensemble generation. In: Jin Y (ed) Multi-objective machine learning. Springer-Verlag, pp 429–464
Zurück zum Zitat Dietterich TG (1997) Machine learning research: four current directions. AI Mag 18:97–136 Dietterich TG (1997) Machine learning research: four current directions. AI Mag 18:97–136
Zurück zum Zitat Dietterich TG (1998) Approximate statistical tests for comparing supervised classification learning algorithms. Neural Comput 10:1895–1923CrossRef Dietterich TG (1998) Approximate statistical tests for comparing supervised classification learning algorithms. Neural Comput 10:1895–1923CrossRef
Zurück zum Zitat Dietterich T (2000) An experimental comparison of three methods for constructing ensembles of decision trees: bagging, boosting and randomization. Mach Learn 40(2):1–22CrossRef Dietterich T (2000) An experimental comparison of three methods for constructing ensembles of decision trees: bagging, boosting and randomization. Mach Learn 40(2):1–22CrossRef
Zurück zum Zitat Eiben AE, Smith JE (2003) Introduction to evolutionary computing. Springer-Verlag Berlin Heidelberg, New YorkMATH Eiben AE, Smith JE (2003) Introduction to evolutionary computing. Springer-Verlag Berlin Heidelberg, New YorkMATH
Zurück zum Zitat Freund Y, Schapire R (1997) A decision-theoretic generalization of on-line learning and an application to boosting. J Comp Syst Sci 55(1):119–139MATHCrossRefMathSciNet Freund Y, Schapire R (1997) A decision-theoretic generalization of on-line learning and an application to boosting. J Comp Syst Sci 55(1):119–139MATHCrossRefMathSciNet
Zurück zum Zitat Inoue H, Narihisa H (2000) Improving generalization ability of self-generating neural networks through ensemble averaging. In: Proceedings of the fourth Pacific-Asia conference on knowledge discovery and data mining (LNAI 1805), Kyoto, Japan, pp 177–180 Inoue H, Narihisa H (2000) Improving generalization ability of self-generating neural networks through ensemble averaging. In: Proceedings of the fourth Pacific-Asia conference on knowledge discovery and data mining (LNAI 1805), Kyoto, Japan, pp 177–180
Zurück zum Zitat Inoue H, Narihisa H (2003) Effective pruning method for a multiple classifier system based on self-generating neural networks. In: Proceedings of the 2003 joint international conference (ICANN/ICONIP’03-LNCS 2714), Istanbul, Turkey, pp 11–18 Inoue H, Narihisa H (2003) Effective pruning method for a multiple classifier system based on self-generating neural networks. In: Proceedings of the 2003 joint international conference (ICANN/ICONIP’03-LNCS 2714), Istanbul, Turkey, pp 11–18
Zurück zum Zitat Inoue H, Narihisa H (2005) Self-organizing neural grove and its applications. In: Proceedings of the 2005 international joint conference on neural networks (IJCNN’05), Montreal, Canada, pp 1205–1210 Inoue H, Narihisa H (2005) Self-organizing neural grove and its applications. In: Proceedings of the 2005 international joint conference on neural networks (IJCNN’05), Montreal, Canada, pp 1205–1210
Zurück zum Zitat Islam MM, Yao X, Murase K (2003) A constructive algorithm for training cooperative neural network ensembles. IEEE Trans Neural Networks 14(4):820–834CrossRef Islam MM, Yao X, Murase K (2003) A constructive algorithm for training cooperative neural network ensembles. IEEE Trans Neural Networks 14(4):820–834CrossRef
Zurück zum Zitat Kasabov N (2001) Evolving fuzzy neural networks for supervised/unsupervised online knowledge-based learning. IEEE Trans Syst Man Cybernet – Part B: Cybernet 31(6):902–918CrossRef Kasabov N (2001) Evolving fuzzy neural networks for supervised/unsupervised online knowledge-based learning. IEEE Trans Syst Man Cybernet – Part B: Cybernet 31(6):902–918CrossRef
Zurück zum Zitat Kohonen T (1995) Self-organizing maps. Springer-Verlag, Berlin Kohonen T (1995) Self-organizing maps. Springer-Verlag, Berlin
Zurück zum Zitat Kuncheva L, Whitaker C (2003) Measures of diversity in classifier ensembles and their relationship with the ensemble accuracy. Mach Learn 51(2):181–207MATHCrossRef Kuncheva L, Whitaker C (2003) Measures of diversity in classifier ensembles and their relationship with the ensemble accuracy. Mach Learn 51(2):181–207MATHCrossRef
Zurück zum Zitat Larose DT (2004) Discovering knowledge in data: an introduction to data mining. Wiley-Interscience Larose DT (2004) Discovering knowledge in data: an introduction to data mining. Wiley-Interscience
Zurück zum Zitat Liu Y, Yao X (1999a) Ensemble learning via negative correlation. Neural Networks 12:1399–1404CrossRef Liu Y, Yao X (1999a) Ensemble learning via negative correlation. Neural Networks 12:1399–1404CrossRef
Zurück zum Zitat Liu Y, Yao X (1999b) Simultaneous training of negatively correlated neural networks in an ensemble. IEEE Trans Syst Man Cybernet Part B – Cybernet 29(6):716–725CrossRef Liu Y, Yao X (1999b) Simultaneous training of negatively correlated neural networks in an ensemble. IEEE Trans Syst Man Cybernet Part B – Cybernet 29(6):716–725CrossRef
Zurück zum Zitat Polikar R, Udpa L, Udpa SS, Honavar V (2001) Learn++: an incremental learning algorithm for supervised neural networks. IEEE Trans Syst Man Cybernet – Part C: Appl Rev 31(4):497–508CrossRef Polikar R, Udpa L, Udpa SS, Honavar V (2001) Learn++: an incremental learning algorithm for supervised neural networks. IEEE Trans Syst Man Cybernet – Part C: Appl Rev 31(4):497–508CrossRef
Zurück zum Zitat Prechelt L (1994) PROBEN1 – a set of neural network benchmark problems and benchmarking rules. Technical Report 21/94, FakultSt fnr Informatik, UniversitSt Karlsruhe, Karlsruhe, Germany Prechelt L (1994) PROBEN1 – a set of neural network benchmark problems and benchmarking rules. Technical Report 21/94, FakultSt fnr Informatik, UniversitSt Karlsruhe, Karlsruhe, Germany
Zurück zum Zitat Rätsch G, Onoda T, Müller K-R (2001) Soft margins for AdaBoost. Mach Learn 42(3):287–320MATHCrossRef Rätsch G, Onoda T, Müller K-R (2001) Soft margins for AdaBoost. Mach Learn 42(3):287–320MATHCrossRef
Zurück zum Zitat Rumelhart DE, Hinton GE, Williams RJ (1986) Learning internal representations by error propagation. Parallel Distrib Process: Explor Microstruct Cogn I:318–362 Rumelhart DE, Hinton GE, Williams RJ (1986) Learning internal representations by error propagation. Parallel Distrib Process: Explor Microstruct Cogn I:318–362
Zurück zum Zitat Schapire R (1990) Strength of weak learning. Mach Learn 5:197–227 Schapire R (1990) Strength of weak learning. Mach Learn 5:197–227
Zurück zum Zitat Schapire RE, Freund Y, Bartlett PL, Lee WS (1998) Boosting the margin: a new explanation for the effectiveness of voting methods. Ann Stat 26(5):1651–1686MATHCrossRefMathSciNet Schapire RE, Freund Y, Bartlett PL, Lee WS (1998) Boosting the margin: a new explanation for the effectiveness of voting methods. Ann Stat 26(5):1651–1686MATHCrossRefMathSciNet
Zurück zum Zitat Seipone T, Bullinaria J (2005) Evolving improved incremental learning schemes for neural network systems. In: Proceedings of the 2005 IEEE congress on evolutionary computing (CEC’2005), Piscataway, NJ, pp 273–280 Seipone T, Bullinaria J (2005) Evolving improved incremental learning schemes for neural network systems. In: Proceedings of the 2005 IEEE congress on evolutionary computing (CEC’2005), Piscataway, NJ, pp 273–280
Zurück zum Zitat Tang EK, Suganthan PN, Yao X (2006) An analysis of diversity measures. Mach Learn 62(1):247–271CrossRef Tang EK, Suganthan PN, Yao X (2006) An analysis of diversity measures. Mach Learn 62(1):247–271CrossRef
Zurück zum Zitat Wang Z, Yao X, Xu Y (2004) An improved constructive neural network ensemble approach to medical diagnoses. In: Proceedings of the fifth international conference on intelligent data engineering and automated learning (IDEAL’04), Lecture Notes in Computer Science, vol 3177, Springer, Exeter, UK, pp 572–577 Wang Z, Yao X, Xu Y (2004) An improved constructive neural network ensemble approach to medical diagnoses. In: Proceedings of the fifth international conference on intelligent data engineering and automated learning (IDEAL’04), Lecture Notes in Computer Science, vol 3177, Springer, Exeter, UK, pp 572–577
Zurück zum Zitat Wen WX, Jennings A, Liu H (1992) Learning a neural tree. In: Proceedings of the 1992 international joint conference on neural networks (IJCNN’92), vol 2, Beijing, China, pp 751–756 Wen WX, Jennings A, Liu H (1992) Learning a neural tree. In: Proceedings of the 1992 international joint conference on neural networks (IJCNN’92), vol 2, Beijing, China, pp 751–756
Zurück zum Zitat Witten IH, Frank E (2000) Data mining – pratical machine learning tools and techniques with Java implementations. Morgan Kaufmann Publishers, San Francisco Witten IH, Frank E (2000) Data mining – pratical machine learning tools and techniques with Java implementations. Morgan Kaufmann Publishers, San Francisco
Zurück zum Zitat Zanchettin C, Minku FL, Ludermir TB (2005) Design of experiments in neuro-fuzzy systems. In: Proceedings of the 5th international conference on hybrid intelligent systems, HIS’2005, Rio de Janeiro, Brasil, pp 218–223 Zanchettin C, Minku FL, Ludermir TB (2005) Design of experiments in neuro-fuzzy systems. In: Proceedings of the 5th international conference on hybrid intelligent systems, HIS’2005, Rio de Janeiro, Brasil, pp 218–223
Metadaten
Titel
Negative correlation in incremental learning
verfasst von
Fernanda Li Minku
Hirotaka Inoue
Xin Yao
Publikationsdatum
01.06.2009
Verlag
Springer Netherlands
Erschienen in
Natural Computing / Ausgabe 2/2009
Print ISSN: 1567-7818
Elektronische ISSN: 1572-9796
DOI
https://doi.org/10.1007/s11047-007-9063-7

Weitere Artikel der Ausgabe 2/2009

Natural Computing 2/2009 Zur Ausgabe

Premium Partner