Skip to main content
Erschienen in: Knowledge and Information Systems 9/2020

30.03.2020 | Regular Paper

Bayesian network classifiers using ensembles and smoothing

verfasst von: He Zhang, François Petitjean, Wray Buntine

Erschienen in: Knowledge and Information Systems | Ausgabe 9/2020

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Bayesian network classifiers are, functionally, an interesting class of models, because they can be learnt out-of-core, i.e. without needing to hold the whole training data in main memory. The selective K-dependence Bayesian network classifier (SKDB) is state of the art in this class of models and has shown to rival random forest (RF) on problems with categorical data. In this paper, we introduce an ensembling technique for SKDB, called ensemble of SKDB (ESKDB). We show that ESKDB significantly outperforms RF on categorical and numerical data, as well as rivalling XGBoost. ESKDB combines three main components: (1) an effective strategy to vary the networks that is built by single classifiers (to make it an ensemble), (2) a stochastic discretization method which allows to both tackle numerical data as well as further increases the variance between different components of our ensemble and (3) a superior smoothing technique to ensure proper calibration of ESKDB’s probabilities. We conduct a large set of experiments with 72 datasets to study the properties of ESKDB (through a sensitivity analysis) and show its competitiveness with the state of the art.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Fußnoten
1
The more common representation \(\mathrm{Dir}(\alpha _1,\ldots , \alpha _C)\) is not used here.
 
Literatur
1.
Zurück zum Zitat Bostrom H (2007) Estimating class probabilities in random forests. In: Machine learning and applications, 2007. ICMLA 2007. 6th international conference on, IEEE, pp 211–216 Bostrom H (2007) Estimating class probabilities in random forests. In: Machine learning and applications, 2007. ICMLA 2007. 6th international conference on, IEEE, pp 211–216
2.
Zurück zum Zitat Breiman L (1996) Bagging predictors. Mach Learn 24(2):123–140MATH Breiman L (1996) Bagging predictors. Mach Learn 24(2):123–140MATH
4.
Zurück zum Zitat Buntine W (1991) Theory refinement of Bayesian networks. In: 7th conference on uncertainty in artificial intelligence, Anaheim, CA Buntine W (1991) Theory refinement of Bayesian networks. In: 7th conference on uncertainty in artificial intelligence, Anaheim, CA
5.
Zurück zum Zitat Buntine W (1993) Learning classification trees. Artificial intelligence frontiers in statistics. Springer, Berlin, pp 182–201 Buntine W (1993) Learning classification trees. Artificial intelligence frontiers in statistics. Springer, Berlin, pp 182–201
6.
Zurück zum Zitat Buntine W, Mishra S (2014) Experiments with non-parametric topic models. In: Proceedings of the 20th ACM SIGKDD international conference on knowledge discovery and data mining, pp 881–890 Buntine W, Mishra S (2014) Experiments with non-parametric topic models. In: Proceedings of the 20th ACM SIGKDD international conference on knowledge discovery and data mining, pp 881–890
7.
Zurück zum Zitat Chen T, Guestrin C (2016) XGBoost: a scalable tree boosting system. In: Proceedings of the 22nd ACM SigKDD international conference on knowledge discovery and data mining, ACM, pp 785–794 Chen T, Guestrin C (2016) XGBoost: a scalable tree boosting system. In: Proceedings of the 22nd ACM SigKDD international conference on knowledge discovery and data mining, ACM, pp 785–794
8.
Zurück zum Zitat Chipman HA, George EI, McCulloch RE (1998) Bayesian CART model search. J Am Stat Assoc 93(443):935–948CrossRef Chipman HA, George EI, McCulloch RE (1998) Bayesian CART model search. J Am Stat Assoc 93(443):935–948CrossRef
9.
Zurück zum Zitat Chow C, Liu C (1968) Approximating discrete probability distributions with dependence trees. IEEE Trans Inf Theory 14(3):462–467MathSciNetMATHCrossRef Chow C, Liu C (1968) Approximating discrete probability distributions with dependence trees. IEEE Trans Inf Theory 14(3):462–467MathSciNetMATHCrossRef
10.
Zurück zum Zitat Dash D, Cooper GF (2004) Model averaging for prediction with discrete Bayesian networks. J Mach Learn Res 5:1177–1203MathSciNetMATH Dash D, Cooper GF (2004) Model averaging for prediction with discrete Bayesian networks. J Mach Learn Res 5:1177–1203MathSciNetMATH
11.
Zurück zum Zitat Du L (2011) Non-parametric Bayesian methods for structured topic models. Ph.D. thesis, Australian National University Du L (2011) Non-parametric Bayesian methods for structured topic models. Ph.D. thesis, Australian National University
12.
Zurück zum Zitat Duan Z, Wang L (2017) \(K\)-dependence Bayesian classifier ensemble. Entropy 19(12):651MathSciNet Duan Z, Wang L (2017) \(K\)-dependence Bayesian classifier ensemble. Entropy 19(12):651MathSciNet
13.
Zurück zum Zitat Fayyad U, Irani K (1993) Multi-interval discretization of continuous-valued attributes for classification learning. In: Proceedings of the 13th international joint conference on artificial intelligence, pp 1022–1027 Fayyad U, Irani K (1993) Multi-interval discretization of continuous-valued attributes for classification learning. In: Proceedings of the 13th international joint conference on artificial intelligence, pp 1022–1027
14.
Zurück zum Zitat Freund Y, Schapire RE (1995) A decision-theoretic generalization of on-line learning and an application to boosting. In: European conference on computational learning theory. Springer, pp 23–37 Freund Y, Schapire RE (1995) A decision-theoretic generalization of on-line learning and an application to boosting. In: European conference on computational learning theory. Springer, pp 23–37
15.
Zurück zum Zitat Friedman J, Hastie T, Tibshirani R et al (2000) Additive logistic regression: a statistical view of boosting (with discussion and a rejoinder by the authors). Ann Stat 28(2):337–407MATHCrossRef Friedman J, Hastie T, Tibshirani R et al (2000) Additive logistic regression: a statistical view of boosting (with discussion and a rejoinder by the authors). Ann Stat 28(2):337–407MATHCrossRef
16.
Zurück zum Zitat Friedman N, Geiger D, Goldszmidt M (1997) Bayesian network classifiers. Mach Learn 29(2–3):131–163MATHCrossRef Friedman N, Geiger D, Goldszmidt M (1997) Bayesian network classifiers. Mach Learn 29(2–3):131–163MATHCrossRef
17.
Zurück zum Zitat Hearst MA (1998) Support vector machines. IEEE Intell Syst 13(4):18–28CrossRef Hearst MA (1998) Support vector machines. IEEE Intell Syst 13(4):18–28CrossRef
18.
Zurück zum Zitat Hoeting JA, Madigan D, Raftery AE, Volinsky CT (1999) Bayesian model averaging: a tutorial (with comments by M. Clyde, David Draper and E. I. George, and a rejoinder by the authors). Stat Sci 14(4):382–417MATHCrossRef Hoeting JA, Madigan D, Raftery AE, Volinsky CT (1999) Bayesian model averaging: a tutorial (with comments by M. Clyde, David Draper and E. I. George, and a rejoinder by the authors). Stat Sci 14(4):382–417MATHCrossRef
19.
Zurück zum Zitat Koivisto M, Sood K (2004) Exact Bayesian structure discovery in Bayesian networks. J Mach Learn Res 5:549–573MathSciNetMATH Koivisto M, Sood K (2004) Exact Bayesian structure discovery in Bayesian networks. J Mach Learn Res 5:549–573MathSciNetMATH
20.
Zurück zum Zitat Lewis DD (1998) Naive Bayes at forty: the independence assumption in information retrieval. Springer, Berlin, pp 4–15 Lewis DD (1998) Naive Bayes at forty: the independence assumption in information retrieval. Springer, Berlin, pp 4–15
22.
Zurück zum Zitat Madigan D, York J, Allard D (1995) Bayesian graphical models for discrete data. Int Stat Rev 63(2):215–232MATHCrossRef Madigan D, York J, Allard D (1995) Bayesian graphical models for discrete data. Int Stat Rev 63(2):215–232MATHCrossRef
23.
Zurück zum Zitat Martínez AM, Webb GI, Chen S, Zaidi NA (2016) Scalable learning of Bayesian network classifiers. J Mach Learn Res 17(1):1515–1549 Martínez AM, Webb GI, Chen S, Zaidi NA (2016) Scalable learning of Bayesian network classifiers. J Mach Learn Res 17(1):1515–1549
24.
Zurück zum Zitat Petitjean F, Buntine W, Webb GI, Zaidi N (2018) Accurate parameter estimation for Bayesian network classifiers using hierarchical Dirichlet processes. Mach Learn 107(8):1303–1331MathSciNetMATHCrossRef Petitjean F, Buntine W, Webb GI, Zaidi N (2018) Accurate parameter estimation for Bayesian network classifiers using hierarchical Dirichlet processes. Mach Learn 107(8):1303–1331MathSciNetMATHCrossRef
25.
Zurück zum Zitat Provost F, Domingos P (2003) Tree induction for probability-based ranking. Mach Learn 52(3):199–215MATHCrossRef Provost F, Domingos P (2003) Tree induction for probability-based ranking. Mach Learn 52(3):199–215MATHCrossRef
26.
Zurück zum Zitat Sahami M (1996) Learning limited dependence Bayesian classifiers. KDD 96:335–338 Sahami M (1996) Learning limited dependence Bayesian classifiers. KDD 96:335–338
27.
Zurück zum Zitat Shareghi E, Haffari G, Cohn T (2017) Compressed nonparametric language modelling. In: Proceedings of the 26th international joint conference on artificial intelligence, pp 2701–2707 Shareghi E, Haffari G, Cohn T (2017) Compressed nonparametric language modelling. In: Proceedings of the 26th international joint conference on artificial intelligence, pp 2701–2707
28.
Zurück zum Zitat Teh YW, Jordan MI (2010) Hierarchical Bayesian nonparametric models with applications. Bayesian Nonparametr 1:158–207MathSciNetCrossRef Teh YW, Jordan MI (2010) Hierarchical Bayesian nonparametric models with applications. Bayesian Nonparametr 1:158–207MathSciNetCrossRef
29.
Zurück zum Zitat Tian J, He R, Ram L (2010) Bayesian model averaging using the \(k\)-best Bayesian network structures. In: Proceedings of the 26th conference on uncertainty in artificial intelligence, AUAI Press, UAI’10, pp 589–597 Tian J, He R, Ram L (2010) Bayesian model averaging using the \(k\)-best Bayesian network structures. In: Proceedings of the 26th conference on uncertainty in artificial intelligence, AUAI Press, UAI’10, pp 589–597
30.
Zurück zum Zitat Webb GI, Boughton JR, Wang Z (2005) Not so naive Bayes: aggregating one-dependence estimators. Mach Learn 58(1):5–24MATHCrossRef Webb GI, Boughton JR, Wang Z (2005) Not so naive Bayes: aggregating one-dependence estimators. Mach Learn 58(1):5–24MATHCrossRef
31.
Zurück zum Zitat Zadrozny B, Elkan C (2001) Obtaining calibrated probability estimates from decision trees and naive Bayesian classifiers. ICML Citeseer 1:609–616 Zadrozny B, Elkan C (2001) Obtaining calibrated probability estimates from decision trees and naive Bayesian classifiers. ICML Citeseer 1:609–616
32.
Zurück zum Zitat Zhou ZH (2012) Ensemble methods: foundations and algorithms, 1st edn. CRC Press Zhou ZH (2012) Ensemble methods: foundations and algorithms, 1st edn. CRC Press
Metadaten
Titel
Bayesian network classifiers using ensembles and smoothing
verfasst von
He Zhang
François Petitjean
Wray Buntine
Publikationsdatum
30.03.2020
Verlag
Springer London
Erschienen in
Knowledge and Information Systems / Ausgabe 9/2020
Print ISSN: 0219-1377
Elektronische ISSN: 0219-3116
DOI
https://doi.org/10.1007/s10115-020-01458-z

Weitere Artikel der Ausgabe 9/2020

Knowledge and Information Systems 9/2020 Zur Ausgabe

Premium Partner