Skip to main content
Top

2016 | OriginalPaper | Chapter

Single Classifier Selection for Ensemble Learning

Authors : Guangtao Wang, Xiaomei Yang, Xiaoyan Zhu

Published in: Advanced Data Mining and Applications

Publisher: Springer International Publishing

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

Ensemble classification is one of representative learning techniques in the field of machine learning, which combines a set of single classifiers together aiming at achieving better classification performance. Not every arbitrary set of single classifiers can obtain a good ensemble classifier. The efficient and necessary condition to construct an accurate ensemble classifier is that the single classifiers should be accurate and diverse. In this paper, we first formally give the definitions of accurate and diverse classifiers and put forward metrics to quantify the accuracy and diversity of the single classifiers; afterwards, we propose a novel parameter-free method to pick up a set of accurate and diverse single classifiers for ensemble. The experimental results on real world data sets show the effectiveness of the proposed method which could improve the performance of the representative ensemble classifier Bagging.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Footnotes
1
In the field of ensemble learning, the independent single classifier is generally called diverse one, and the independence is also named diversity.
 
2
In the table, “F”, “I” and “T” denote the numbers of features, instances and target concept values, respectively.
 
3
In the table, “\( acc_{RF}\)” “\(acc_{BD}(acc_{BN})\)” and “\( acc_{after} \)” denote the classification accuracy of RandomForest, Boosting+DT(Boosting+NB) and Bagging filtered by Algorithm 1.
 
Literature
1.
2.
go back to reference Breiman, L.: Bagging predictors. Mach. Learn. 24(2), 123–140 (1996)MATH Breiman, L.: Bagging predictors. Mach. Learn. 24(2), 123–140 (1996)MATH
3.
go back to reference Brown, G., Wyatt, J., Harris, R., Yao, X.: Diversity creation methods: a survey and categorisation. Inf. Fusion 6(1), 5–20 (2005)CrossRef Brown, G., Wyatt, J., Harris, R., Yao, X.: Diversity creation methods: a survey and categorisation. Inf. Fusion 6(1), 5–20 (2005)CrossRef
4.
go back to reference Cunningham, P., Carney, J.: Diversity versus quality in classification ensembles based on feature selection. In: López de Mántaras, R., Plaza, E. (eds.) ECML 2000. LNCS (LNAI), vol. 1810, pp. 109–116. Springer, Heidelberg (2000). doi:10.1007/3-540-45164-1_12 CrossRef Cunningham, P., Carney, J.: Diversity versus quality in classification ensembles based on feature selection. In: López de Mántaras, R., Plaza, E. (eds.) ECML 2000. LNCS (LNAI), vol. 1810, pp. 109–116. Springer, Heidelberg (2000). doi:10.​1007/​3-540-45164-1_​12 CrossRef
6.
go back to reference Dietterich, T.G.: An experimental comparison of three methods for constructing ensembles of decision trees: bagging, boosting, and randomization. Mach. Learn. 40(2), 139–157 (2000)CrossRef Dietterich, T.G.: An experimental comparison of three methods for constructing ensembles of decision trees: bagging, boosting, and randomization. Mach. Learn. 40(2), 139–157 (2000)CrossRef
7.
go back to reference Dietterich, T.G.: Ensemble learning. In: The Handbook of Brain Theory and Neural Networks, vol. 2, pp. 110–125 (2002) Dietterich, T.G.: Ensemble learning. In: The Handbook of Brain Theory and Neural Networks, vol. 2, pp. 110–125 (2002)
8.
go back to reference Džeroski, S., Ženko, B.: Is combining classifiers with stacking better than selecting the best one? Mach. Learn. 54(3), 255–273 (2004)CrossRefMATH Džeroski, S., Ženko, B.: Is combining classifiers with stacking better than selecting the best one? Mach. Learn. 54(3), 255–273 (2004)CrossRefMATH
9.
go back to reference Fleiss, J.L., Cohen, J., Everitt, B.: Large sample standard errors of kappa and weighted kappa. Psychol. Bull. 72(5), 323 (1969)CrossRef Fleiss, J.L., Cohen, J., Everitt, B.: Large sample standard errors of kappa and weighted kappa. Psychol. Bull. 72(5), 323 (1969)CrossRef
10.
go back to reference Freund, Y., Schapire, R.E., et al.: Experiments with a new boosting algorithm. In: ICML, vol. 96, pp. 148–156 (1996) Freund, Y., Schapire, R.E., et al.: Experiments with a new boosting algorithm. In: ICML, vol. 96, pp. 148–156 (1996)
11.
go back to reference Giacinto, G., Roli, F.: Design of effective neural network ensembles for image classification purposes. Image Vis. Comput. 19(9), 699–707 (2001)CrossRef Giacinto, G., Roli, F.: Design of effective neural network ensembles for image classification purposes. Image Vis. Comput. 19(9), 699–707 (2001)CrossRef
12.
go back to reference Hansen, L.K., Salamon, P.: Neural network ensembles. IEEE Trans. Pattern Anal. Mach. Intell. 10, 993–1001 (1990)CrossRef Hansen, L.K., Salamon, P.: Neural network ensembles. IEEE Trans. Pattern Anal. Mach. Intell. 10, 993–1001 (1990)CrossRef
13.
go back to reference Ho, T.K.: The random subspace method for constructing decision forests. IEEE Trans. Pattern Anal. Mach. Intell. 20(8), 832–844 (1998)CrossRef Ho, T.K.: The random subspace method for constructing decision forests. IEEE Trans. Pattern Anal. Mach. Intell. 20(8), 832–844 (1998)CrossRef
14.
go back to reference Kohavi, R., Wolpert, D.H., et al.: Bias plus variance decomposition for zero-one loss functions. In: ICML, vol. 96, pp. 275–283 (1996) Kohavi, R., Wolpert, D.H., et al.: Bias plus variance decomposition for zero-one loss functions. In: ICML, vol. 96, pp. 275–283 (1996)
15.
go back to reference Kuncheva, L.I., Whitaker, C.J.: Measures of diversity in classifier ensembles and their relationship with the ensemble accuracy. Mach. Learn. 51(2), 181–207 (2003)CrossRefMATH Kuncheva, L.I., Whitaker, C.J.: Measures of diversity in classifier ensembles and their relationship with the ensemble accuracy. Mach. Learn. 51(2), 181–207 (2003)CrossRefMATH
16.
go back to reference Kuncheva, L.I., Whitaker, C.J., Shipp, C.A., Duin, R.P.: Limits on the majority vote accuracy in classifier fusion. Pattern Anal. Appl. 6(1), 22–31 (2003)MathSciNetCrossRefMATH Kuncheva, L.I., Whitaker, C.J., Shipp, C.A., Duin, R.P.: Limits on the majority vote accuracy in classifier fusion. Pattern Anal. Appl. 6(1), 22–31 (2003)MathSciNetCrossRefMATH
17.
18.
go back to reference Lee, J.W., Giraud-Carrier, C.: Automatic selection of classification learning algorithms for data mining practitioners. Intell. Data Anal. 17(4), 665–678 (2013) Lee, J.W., Giraud-Carrier, C.: Automatic selection of classification learning algorithms for data mining practitioners. Intell. Data Anal. 17(4), 665–678 (2013)
19.
go back to reference Peterson, A.H., Martinez, T.R.: Estimating the potential for combining learning models. In: Proceedings of the ICML Workshop on Meta-learning, pp. 68–75 (2005) Peterson, A.H., Martinez, T.R.: Estimating the potential for combining learning models. In: Proceedings of the ICML Workshop on Meta-learning, pp. 68–75 (2005)
20.
go back to reference Sharkey, A.J.: Linear and order statistics combiners for pattern classification. In: Sharkey, A.J. (ed.) Combining Artificial Neural Nets, pp. 127–161. Springer, London (1999)CrossRef Sharkey, A.J.: Linear and order statistics combiners for pattern classification. In: Sharkey, A.J. (ed.) Combining Artificial Neural Nets, pp. 127–161. Springer, London (1999)CrossRef
21.
go back to reference Sim, J., Wright, C.C.: The kappa statistic in reliability studies: use, interpretation, and sample size requirements. Phys. Ther. 85(3), 257–268 (2005) Sim, J., Wright, C.C.: The kappa statistic in reliability studies: use, interpretation, and sample size requirements. Phys. Ther. 85(3), 257–268 (2005)
22.
go back to reference Tumer, K., Ghosh, J.: Error correlation and error reduction in ensemble classifiers. Connection Sci. 8(3–4), 385–404 (1996)CrossRef Tumer, K., Ghosh, J.: Error correlation and error reduction in ensemble classifiers. Connection Sci. 8(3–4), 385–404 (1996)CrossRef
Metadata
Title
Single Classifier Selection for Ensemble Learning
Authors
Guangtao Wang
Xiaomei Yang
Xiaoyan Zhu
Copyright Year
2016
DOI
https://doi.org/10.1007/978-3-319-49586-6_21

Premium Partner