Skip to main content
Top

2020 | OriginalPaper | Chapter

An Evolutionary Algorithm for Learning Interpretable Ensembles of Classifiers

Authors : Henry E. L. Cagnini, Alex A. Freitas, Rodrigo C. Barros

Published in: Intelligent Systems

Publisher: Springer International Publishing

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

Ensembles of classifiers are a very popular type of method for performing classification, due to their usually high predictive accuracy. However, ensembles have two drawbacks. First, ensembles are usually considered a ‘black box’, non-interpretable type of classification model, mainly because typically there are a very large number of classifiers in the ensemble (and often each classifier in the ensemble is a black-box classifier by itself). This lack of interpretability is an important limitation in application domains where a model’s predictions should be carefully interpreted by users, like medicine, law, etc. Second, ensemble methods typically involve many hyper-parameters, and it is difficult for users to select the best settings for those hyper-parameters. In this work we propose an Evolutionary Algorithm (an Estimation of Distribution Algorithm) that addresses both these drawbacks. This algorithm optimizes the hyper-parameter settings of a small ensemble of 5 interpretable classifiers, which allows users to interpret each classifier. In our experiments, the ensembles learned by the proposed Evolutionary Algorithm achieved the same level of predictive accuracy as a well-known Random Forest ensemble, but with the benefit of learning interpretable models (unlike Random Forests).

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Alcalá-Fdez, J., et al.: Keel data-mining software tool: data set repository, integration of algorithms and experimental analysis framework. J. Multiple-Valued Logic Soft Comput. 17, 255–287 (2011) Alcalá-Fdez, J., et al.: Keel data-mining software tool: data set repository, integration of algorithms and experimental analysis framework. J. Multiple-Valued Logic Soft Comput. 17, 255–287 (2011)
2.
go back to reference Ali, J., et al.: Random forests and decision trees. Int. J. Comput. Sci. Issues (IJCSI) 9(5), 272 (2012) Ali, J., et al.: Random forests and decision trees. Int. J. Comput. Sci. Issues (IJCSI) 9(5), 272 (2012)
3.
go back to reference Baluja, S., Caruana, R.: Removing the genetics from the standard genetic algorithm. In: Proceedings of the Twelfth International Conference on Machine Learning, pp. 38–46. Elsevier, Tahoe City (1995) Baluja, S., Caruana, R.: Removing the genetics from the standard genetic algorithm. In: Proceedings of the Twelfth International Conference on Machine Learning, pp. 38–46. Elsevier, Tahoe City (1995)
5.
go back to reference Breiman, L., et al.: Classification and Regression Trees. Wadsworth International Group, Belmont (1984)MATH Breiman, L., et al.: Classification and Regression Trees. Wadsworth International Group, Belmont (1984)MATH
6.
go back to reference Carvalho, D.V., Pereira, E.M., Cardoso, J.S.: Machine learning interpretability: a survey on methods and metrics. Electronics 8(8), 832 (2019)CrossRef Carvalho, D.V., Pereira, E.M., Cardoso, J.S.: Machine learning interpretability: a survey on methods and metrics. Electronics 8(8), 832 (2019)CrossRef
7.
go back to reference Cohen, W.W.: Fast effective rule induction. In: Twelfth International Conference on Machine Learning, pp. 115–123 (1995) Cohen, W.W.: Fast effective rule induction. In: Twelfth International Conference on Machine Learning, pp. 115–123 (1995)
9.
go back to reference Fernández-Delgado, M., et al.: Do we need hundreds of classifiers to solve real world classification problems? J. Mach. Learn. Res. 15(1), 3133–3181 (2014)MathSciNetMATH Fernández-Delgado, M., et al.: Do we need hundreds of classifiers to solve real world classification problems? J. Mach. Learn. Res. 15(1), 3133–3181 (2014)MathSciNetMATH
10.
go back to reference Feurer, M., et al.: Efficient and robust automated machine learning. In: Advances in Neural Information Processing Systems, pp. 2962–2970 (2015) Feurer, M., et al.: Efficient and robust automated machine learning. In: Advances in Neural Information Processing Systems, pp. 2962–2970 (2015)
11.
go back to reference Frank, E., Witten, I.H.: Generating accurate rule sets without global optimization. In: Fifteenth International Conference on Machine Learning, pp. 144–151 (1998) Frank, E., Witten, I.H.: Generating accurate rule sets without global optimization. In: Fifteenth International Conference on Machine Learning, pp. 144–151 (1998)
12.
go back to reference Freitas, A.A.: Comprehensible classification models: a position paper. ACM SIGKDD Explor. Newsl. 15(1), 1–10 (2014)CrossRef Freitas, A.A.: Comprehensible classification models: a position paper. ACM SIGKDD Explor. Newsl. 15(1), 1–10 (2014)CrossRef
13.
go back to reference Fürnkranz, J., Kliegr, T., Paulheim, H.: On cognitive preferences and the plausibility of rule-based models (2018). arXiv preprint arXiv:1803.01316 Fürnkranz, J., Kliegr, T., Paulheim, H.: On cognitive preferences and the plausibility of rule-based models (2018). arXiv preprint arXiv:​1803.​01316
14.
go back to reference Galea, M., Shen, Q., Levine, J.: Evolutionary approaches to fuzzy modelling for classification. Knowl. Eng. Rev. 19(1), 27–59 (2004)CrossRef Galea, M., Shen, Q., Levine, J.: Evolutionary approaches to fuzzy modelling for classification. Knowl. Eng. Rev. 19(1), 27–59 (2004)CrossRef
15.
go back to reference Hall, M., et al.: The WEKA data mining software: an update. ACM SIGKDD Explor. Newsl. 11(1), 10–18 (2009)CrossRef Hall, M., et al.: The WEKA data mining software: an update. ACM SIGKDD Explor. Newsl. 11(1), 10–18 (2009)CrossRef
16.
go back to reference Hauschild, M., Pelikan, M.: An introduction and survey of estimation of distribution algorithms. Swarm Evol. Comput. 1(3), 111–128 (2011)CrossRef Hauschild, M., Pelikan, M.: An introduction and survey of estimation of distribution algorithms. Swarm Evol. Comput. 1(3), 111–128 (2011)CrossRef
17.
go back to reference Holland, J.H.: Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence. MIT press, Cambridge (1992)CrossRef Holland, J.H.: Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence. MIT press, Cambridge (1992)CrossRef
18.
go back to reference Huysmans, J., et al.: An empirical evaluation of the comprehensibility of decision table, tree and rule based predictive models. Decis. Supp. Syst. 51(1), 141–154 (2011)CrossRef Huysmans, J., et al.: An empirical evaluation of the comprehensibility of decision table, tree and rule based predictive models. Decis. Supp. Syst. 51(1), 141–154 (2011)CrossRef
20.
go back to reference Kordik, P., Cerny, J., Fryda, T.: Discovering predictive ensembles for transfer learning and meta-learning. Mach. Learn. 107, 177–207 (2018)MathSciNetMATHCrossRef Kordik, P., Cerny, J., Fryda, T.: Discovering predictive ensembles for transfer learning and meta-learning. Mach. Learn. 107, 177–207 (2018)MathSciNetMATHCrossRef
21.
go back to reference Kuncheva, L.I.: Combining Pattern Classifiers: Methods and Algorithms. Wiley, Hoboken (2004)MATHCrossRef Kuncheva, L.I.: Combining Pattern Classifiers: Methods and Algorithms. Wiley, Hoboken (2004)MATHCrossRef
22.
go back to reference Lapuschkin, S., et al.: Unmasking Clever Hans predictors and assessing what machines really learn. Nat. Commun. 10(1), 1096 (2019)CrossRef Lapuschkin, S., et al.: Unmasking Clever Hans predictors and assessing what machines really learn. Nat. Commun. 10(1), 1096 (2019)CrossRef
26.
go back to reference Luštrek, M., et al.: What makes classification trees comprehensible? Exp. Syst. Appl. 62, 333–346 (2016)CrossRef Luštrek, M., et al.: What makes classification trees comprehensible? Exp. Syst. Appl. 62, 333–346 (2016)CrossRef
27.
go back to reference Murphy, K.P.: Machine Learning: A Probabilistic Perspective. MIT Press, Cambridge (2012)MATH Murphy, K.P.: Machine Learning: A Probabilistic Perspective. MIT Press, Cambridge (2012)MATH
28.
29.
go back to reference Quinlan, J.R.: C4.5: Programs for Machine Learning. Morgan Kaufmann Publishers, San Mateo (1993) Quinlan, J.R.: C4.5: Programs for Machine Learning. Morgan Kaufmann Publishers, San Mateo (1993)
30.
go back to reference Rish, I., et al.: An empirical study of the naive Bayes classifier. In: Proceedings of theWorkshop on empirical methods in artificial intelligence, IJCAI 2001, Seattle, USA, vol. 3, pp. 41–46 (2001) Rish, I., et al.: An empirical study of the naive Bayes classifier. In: Proceedings of theWorkshop on empirical methods in artificial intelligence, IJCAI 2001, Seattle, USA, vol. 3, pp. 41–46 (2001)
31.
go back to reference Tan, P.N., Steinbach, M., Kumar, V.: Introduction to Data Mining. Pearson, London (2006) Tan, P.N., Steinbach, M., Kumar, V.: Introduction to Data Mining. Pearson, London (2006)
32.
go back to reference Thornton, C., et al.: Auto-WEKA: combined selection and hyperparameter optimization of classification algorithms. In: International Conference on Knowledge Discovery and Data Mining, pp. 847–855. ACM (2013) Thornton, C., et al.: Auto-WEKA: combined selection and hyperparameter optimization of classification algorithms. In: International Conference on Knowledge Discovery and Data Mining, pp. 847–855. ACM (2013)
33.
go back to reference Xavier-Júnior, J.A.C., et al.: A novel evolutionary algorithm for automated machine learning focusing on classifier ensembles. In: Brazilian Conference on Intelligent Systems. pp. 1–6. IEEE, São Paulo (2018) Xavier-Júnior, J.A.C., et al.: A novel evolutionary algorithm for automated machine learning focusing on classifier ensembles. In: Brazilian Conference on Intelligent Systems. pp. 1–6. IEEE, São Paulo (2018)
34.
go back to reference Zangari, M., et al.: Not all PBILs are the same: unveiling the different learning mechanisms of PBIL variants. Appl. Soft Comput. 53, 88–96 (2017)CrossRef Zangari, M., et al.: Not all PBILs are the same: unveiling the different learning mechanisms of PBIL variants. Appl. Soft Comput. 53, 88–96 (2017)CrossRef
Metadata
Title
An Evolutionary Algorithm for Learning Interpretable Ensembles of Classifiers
Authors
Henry E. L. Cagnini
Alex A. Freitas
Rodrigo C. Barros
Copyright Year
2020
DOI
https://doi.org/10.1007/978-3-030-61377-8_2

Premium Partner