Skip to main content
Top
Published in: Neural Processing Letters 1/2018

08-09-2017

Pruning the Ensemble of ANN Based on Decision Tree Induction

Authors: Sha Ding, Zhi Chen, Shi-yuan Zhao, Tao Lin

Published in: Neural Processing Letters | Issue 1/2018

Log in

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

Ensemble learning is a powerful approach for achieving more accurate predictions compared with single classifier. However, this powerful classification ability is achieved at the expense of heavy storage requirements and computational burdens on the ensemble. Ensemble pruning is a crucial step for the reduction of the predictive overhead without worsening the performance of original ensemble. This paper suggests an efficient and effective ordering-based ensemble pruning based on the induction of decision tree. The suggested method maps the dataset and base classifiers to a new dataset where the ensemble pruning can be transformed to a feature selection problem. Furthermore, a set of accurate, diverse and complementary base classifiers can be selected by the induction of decision tree. Moreover, an evaluation function that deliberately favors the candidate sub-ensembles with an improved performance in classifying low margin instances has also been designed. The comparative experiments on 24 benchmark datasets demonstrate the effectiveness of our proposed method.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Soltanmohammadi E, Naraghi-Pour M, van der Schaar M (2016) Context-based unsupervised ensemble learning and feature ranking. Mach Learn 105(3):459–485MathSciNetMATHCrossRef Soltanmohammadi E, Naraghi-Pour M, van der Schaar M (2016) Context-based unsupervised ensemble learning and feature ranking. Mach Learn 105(3):459–485MathSciNetMATHCrossRef
2.
go back to reference Termenon M, Grana M (2012) A two stage sequential ensemble applied to the classification of Alzheimer’s disease based on MRI features. Neural Process Lett 35(1):1–12CrossRef Termenon M, Grana M (2012) A two stage sequential ensemble applied to the classification of Alzheimer’s disease based on MRI features. Neural Process Lett 35(1):1–12CrossRef
3.
go back to reference Lin SJ, Chen TF (2016) Multi-agent architecture for corporate operating performance assessment. Neural Process Lett 43(1):115–132CrossRef Lin SJ, Chen TF (2016) Multi-agent architecture for corporate operating performance assessment. Neural Process Lett 43(1):115–132CrossRef
4.
go back to reference Laradji IH, Alshayeb M, Ghouti L (2015) Software defect prediction using ensemble learning on selected features. Inf SoftwTechnol 58:388–402 Laradji IH, Alshayeb M, Ghouti L (2015) Software defect prediction using ensemble learning on selected features. Inf SoftwTechnol 58:388–402
5.
go back to reference Kuncheva LI, Whitaker CJ (2003) Measures of diversity in classifier ensembles and their relationship with the ensemble accuracy. Mach Learn 51(2):181–207MATHCrossRef Kuncheva LI, Whitaker CJ (2003) Measures of diversity in classifier ensembles and their relationship with the ensemble accuracy. Mach Learn 51(2):181–207MATHCrossRef
6.
go back to reference Kittler J, Hatef M, Duin RPW, Matas J (1998) On combining classifiers. IEEE Trans Pattern Anal Mach Intell 20(3):226–239CrossRef Kittler J, Hatef M, Duin RPW, Matas J (1998) On combining classifiers. IEEE Trans Pattern Anal Mach Intell 20(3):226–239CrossRef
7.
go back to reference Britto AS, Sabourin R, Oliveira LES (2014) Dynamic selection of classifiers—a comprehensive review. Pattern Recognit 47(11):3665–3680CrossRef Britto AS, Sabourin R, Oliveira LES (2014) Dynamic selection of classifiers—a comprehensive review. Pattern Recognit 47(11):3665–3680CrossRef
8.
go back to reference Haghighi MS, Vahedian A, Yazdi HS (2012) Making diversity enhancement based on multiple classifier system by weight tuning. Neural Process Lett 35(1):61–80CrossRef Haghighi MS, Vahedian A, Yazdi HS (2012) Making diversity enhancement based on multiple classifier system by weight tuning. Neural Process Lett 35(1):61–80CrossRef
9.
go back to reference Freund Y, Schapire RE (1997) A decision-theoretic generalization of on-line learning and an application to boosting. J Comput Syst Sci 55(1):119–139MathSciNetMATHCrossRef Freund Y, Schapire RE (1997) A decision-theoretic generalization of on-line learning and an application to boosting. J Comput Syst Sci 55(1):119–139MathSciNetMATHCrossRef
10.
go back to reference Melville P, Mooney RJ (2005) Creating diversity in ensembles using artificial data. Inf Fusion 6(1):99–111CrossRef Melville P, Mooney RJ (2005) Creating diversity in ensembles using artificial data. Inf Fusion 6(1):99–111CrossRef
11.
go back to reference Wang L, Sugiyama M, Jing Z, Yang C, Zhou ZH, Feng J (2011) A refined margin analysis for boosting algorithms via equilibrium margin. J Mach Learn Res 12(2):1835–1863MathSciNetMATH Wang L, Sugiyama M, Jing Z, Yang C, Zhou ZH, Feng J (2011) A refined margin analysis for boosting algorithms via equilibrium margin. J Mach Learn Res 12(2):1835–1863MathSciNetMATH
12.
go back to reference Sun B, Chen H, Wang J (2015) An empirical margin explanation for the effectiveness of DECORATE ensemble learning algorithm. Knowl Based Syst 78:1–12CrossRef Sun B, Chen H, Wang J (2015) An empirical margin explanation for the effectiveness of DECORATE ensemble learning algorithm. Knowl Based Syst 78:1–12CrossRef
13.
go back to reference Freund Y, Schapire RE (1999) Large margin classification using the perceptron algorithm. Mach Learn 37(3):277–296MATHCrossRef Freund Y, Schapire RE (1999) Large margin classification using the perceptron algorithm. Mach Learn 37(3):277–296MATHCrossRef
14.
go back to reference Hu Q, Zhu P, Yang Y, Yu D (2011) Large-margin nearest neighbor classifiers via sample weight learning. Neurocomputing 74(4):656–660CrossRef Hu Q, Zhu P, Yang Y, Yu D (2011) Large-margin nearest neighbor classifiers via sample weight learning. Neurocomputing 74(4):656–660CrossRef
15.
go back to reference Hu Q, Li L, Wu X, Schaefer G, Yu D (2014) Exploiting diversity for optimizing margin distribution in ensemble learning. Knowl Based Syst 67:90–104CrossRef Hu Q, Li L, Wu X, Schaefer G, Yu D (2014) Exploiting diversity for optimizing margin distribution in ensemble learning. Knowl Based Syst 67:90–104CrossRef
16.
go back to reference Zhou H, Zhao X, Wang X (2014) An effective ensemble pruning algorithm based on frequent patterns. Knowl Based Syst 56(3):79–85CrossRef Zhou H, Zhao X, Wang X (2014) An effective ensemble pruning algorithm based on frequent patterns. Knowl Based Syst 56(3):79–85CrossRef
17.
go back to reference Margineantu DD, Dietterich, TG (1997) Pruning Adaptive Boosting. In: Proceedings of the fourteenth international conference on machine learning. Morgan Kaufmann Publishers Inc, pp 211–218 Margineantu DD, Dietterich, TG (1997) Pruning Adaptive Boosting. In: Proceedings of the fourteenth international conference on machine learning. Morgan Kaufmann Publishers Inc, pp 211–218
18.
go back to reference Martinez-Muoz G, Hernandez-Lobato D, Suarez A (2009) An analysis of ensemble pruning techniques based on ordered aggregation. IEEE Trans Pattern Anal Mach Intell 31(2):245–259CrossRef Martinez-Muoz G, Hernandez-Lobato D, Suarez A (2009) An analysis of ensemble pruning techniques based on ordered aggregation. IEEE Trans Pattern Anal Mach Intell 31(2):245–259CrossRef
19.
go back to reference Martínez-Muñoz G, Suárez A (2006) Pruning in ordered bagging ensembles. In: Proceedings of the 23rd international conference on machine learning. ACM, pp 609–616 Martínez-Muñoz G, Suárez A (2006) Pruning in ordered bagging ensembles. In: Proceedings of the 23rd international conference on machine learning. ACM, pp 609–616
20.
go back to reference Guo L, Boukir S (2013) Margin-based ordered aggregation for ensemble pruning. Pattern Recognit Lett 34(6):603–609CrossRef Guo L, Boukir S (2013) Margin-based ordered aggregation for ensemble pruning. Pattern Recognit Lett 34(6):603–609CrossRef
21.
go back to reference Dai Q, Han XM (2016) An efficient ordering-based ensemble pruning algorithm via dynamic programming. Appl Intell 44(4):816–830MathSciNetCrossRef Dai Q, Han XM (2016) An efficient ordering-based ensemble pruning algorithm via dynamic programming. Appl Intell 44(4):816–830MathSciNetCrossRef
22.
go back to reference Bhardwaj M, Bhatnagar V (2015) Towards an optimally pruned classifier ensemble. Int J Mach Learn Cybern 6(5):1–20CrossRef Bhardwaj M, Bhatnagar V (2015) Towards an optimally pruned classifier ensemble. Int J Mach Learn Cybern 6(5):1–20CrossRef
23.
24.
go back to reference Yin XC, Huang K, Hao HW, Iqbal K, Wang ZB (2014) A novel classifier ensemble method with sparsity and diversity. Neurocomputing 134(134):214–221CrossRef Yin XC, Huang K, Hao HW, Iqbal K, Wang ZB (2014) A novel classifier ensemble method with sparsity and diversity. Neurocomputing 134(134):214–221CrossRef
25.
go back to reference Zhang Y, Burer S, Street WN (2006) Ensemble pruning via semi-definite programming. J Mach Learn Res 7(3):1315–1338MathSciNetMATH Zhang Y, Burer S, Street WN (2006) Ensemble pruning via semi-definite programming. J Mach Learn Res 7(3):1315–1338MathSciNetMATH
26.
go back to reference Dai Q (2013) A novel ensemble pruning algorithm based on randomized greedy selective strategy and ballot. Neurocomputing 122(122):258–265CrossRef Dai Q (2013) A novel ensemble pruning algorithm based on randomized greedy selective strategy and ballot. Neurocomputing 122(122):258–265CrossRef
27.
go back to reference Bakker B, Heskes T (2003) Clustering ensembles of neural network models. Neural Netw 16(2):261–269CrossRef Bakker B, Heskes T (2003) Clustering ensembles of neural network models. Neural Netw 16(2):261–269CrossRef
28.
go back to reference Giacinto G, Roli F, Fumera G (2000) Design of effective multiple classifier systems by clustering of classifiers. In: Proceedings of 15th international conference on pattern recognition, vol 2, pp 160–163 Giacinto G, Roli F, Fumera G (2000) Design of effective multiple classifier systems by clustering of classifiers. In: Proceedings of 15th international conference on pattern recognition, vol 2, pp 160–163
29.
go back to reference Zhang H, Cao L (2014) A spectral clustering based ensemble pruning approach. Neurocomputing 139(139):289–297CrossRef Zhang H, Cao L (2014) A spectral clustering based ensemble pruning approach. Neurocomputing 139(139):289–297CrossRef
30.
go back to reference Partalas I, Tsoumakas G, Vlahavas I (2009) Pruning an ensemble of classifiers via reinforcement learning. Neurocomputing 72(7–9):1900–1909CrossRef Partalas I, Tsoumakas G, Vlahavas I (2009) Pruning an ensemble of classifiers via reinforcement learning. Neurocomputing 72(7–9):1900–1909CrossRef
31.
go back to reference Ykhlef H, Bouchaffra D (2017) An efficient ensemble pruning approach based on simple coalitional games. Inf Fusion 34:28–42CrossRef Ykhlef H, Bouchaffra D (2017) An efficient ensemble pruning approach based on simple coalitional games. Inf Fusion 34:28–42CrossRef
32.
go back to reference Zhao QL, Jiang YH, Xu M (2009) A fast ensemble pruning algorithm based on pattern mining process. Data Min Knowl Discov 19(2):277–292MathSciNetCrossRef Zhao QL, Jiang YH, Xu M (2009) A fast ensemble pruning algorithm based on pattern mining process. Data Min Knowl Discov 19(2):277–292MathSciNetCrossRef
33.
go back to reference Krawczyk B, Woźniak M (2016) Untrained weighted classifier combination with embedded ensemble pruning. Neurocomputing 196:14–22CrossRef Krawczyk B, Woźniak M (2016) Untrained weighted classifier combination with embedded ensemble pruning. Neurocomputing 196:14–22CrossRef
34.
go back to reference r-Aky Z, Reyya S, Windeatt T, Smith R (2015) Pruning of error correcting output codes by optimization of accuracy—diversity trade off. Mach Learn 101(1):1–17MathSciNet r-Aky Z, Reyya S, Windeatt T, Smith R (2015) Pruning of error correcting output codes by optimization of accuracy—diversity trade off. Mach Learn 101(1):1–17MathSciNet
35.
go back to reference Partridge D, Yates WB (1996) Engineering multiversion neural-net systems. Neural Comput 8(4):869–893CrossRef Partridge D, Yates WB (1996) Engineering multiversion neural-net systems. Neural Comput 8(4):869–893CrossRef
36.
go back to reference Breiman L (1996) Bagging predictors. Mach Learn 24(2):123–140MATH Breiman L (1996) Bagging predictors. Mach Learn 24(2):123–140MATH
37.
go back to reference Tsch G, Warmuth MK (2005) Efficient margin maximizing with boosting. J Mach Learn Res 6:2131–2152MathSciNetMATH Tsch G, Warmuth MK (2005) Efficient margin maximizing with boosting. J Mach Learn Res 6:2131–2152MathSciNetMATH
38.
go back to reference Shen C, Li H (2010) Boosting through optimization of margin distributions. IEEE Trans Neural Netw 21(4):659–666CrossRef Shen C, Li H (2010) Boosting through optimization of margin distributions. IEEE Trans Neural Netw 21(4):659–666CrossRef
39.
go back to reference Vapnik V, Chapelle O (2000) Bounds on error expectation for support vector machines. Neural Comput 12(9):2013–2036CrossRef Vapnik V, Chapelle O (2000) Bounds on error expectation for support vector machines. Neural Comput 12(9):2013–2036CrossRef
40.
go back to reference Bache K, Lichman M (2013) UCI machine learning repository Bache K, Lichman M (2013) UCI machine learning repository
41.
go back to reference Hall M, Frank E, Holmes G, Pfahringer B, Reutemann P, Witten IH (2009) The WEKA data mining software: an update. SIGKDD Explor Newsl 11(1):10–18CrossRef Hall M, Frank E, Holmes G, Pfahringer B, Reutemann P, Witten IH (2009) The WEKA data mining software: an update. SIGKDD Explor Newsl 11(1):10–18CrossRef
42.
go back to reference Hodges JL, Lehmann EL (1962) Rank Methods for combination of independent experiments in analysis of variance. Ann Math Stat 33(2):482–497MathSciNetMATHCrossRef Hodges JL, Lehmann EL (1962) Rank Methods for combination of independent experiments in analysis of variance. Ann Math Stat 33(2):482–497MathSciNetMATHCrossRef
43.
44.
go back to reference Whitaker CJ, Kuncheva LI (2002) Examining the relationship between majority vote accuracy and diversity in bagging and boosting. Inform Softw Technol 1(1):1–19 Whitaker CJ, Kuncheva LI (2002) Examining the relationship between majority vote accuracy and diversity in bagging and boosting. Inform Softw Technol 1(1):1–19
45.
go back to reference Tang EK, Suganthan PN, Yao X (2006) An analysis of diversity measures. Mach Learn 65(1):247–271CrossRef Tang EK, Suganthan PN, Yao X (2006) An analysis of diversity measures. Mach Learn 65(1):247–271CrossRef
46.
go back to reference Tsymbal A, Pechenizkiy M, Cunningham P (2005) Diversity in search strategies for ensemble feature selection. Inf Fusion 6(1):83–98CrossRef Tsymbal A, Pechenizkiy M, Cunningham P (2005) Diversity in search strategies for ensemble feature selection. Inf Fusion 6(1):83–98CrossRef
Metadata
Title
Pruning the Ensemble of ANN Based on Decision Tree Induction
Authors
Sha Ding
Zhi Chen
Shi-yuan Zhao
Tao Lin
Publication date
08-09-2017
Publisher
Springer US
Published in
Neural Processing Letters / Issue 1/2018
Print ISSN: 1370-4621
Electronic ISSN: 1573-773X
DOI
https://doi.org/10.1007/s11063-017-9703-6

Other articles of this Issue 1/2018

Neural Processing Letters 1/2018 Go to the issue