Skip to main content
Erschienen in: Artificial Intelligence Review 2/2020

01.03.2019

Online AdaBoost-based methods for multiclass problems

verfasst von: Silas Garrido Teixeira de Carvalho Santos, Roberto Souto Maior de Barros

Erschienen in: Artificial Intelligence Review | Ausgabe 2/2020

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Boosting is a technique forged to transform a set of weak classifiers into a strong ensemble. To achieve this, the components are trained with different data samples and the hypotheses are aggregated in order to perform a better prediction. The use of boosting in online environments is a comparatively new activity, inspired by its success in offline environments, which is emerging to meet new demands. One of the challenges is to make the methods handle significant amounts of information taking into account computational constraints. This paper proposes two new online boosting methods: the first aims to perform a better weight distribution of the instances to closely match the behavior of AdaBoost.M1 whereas the second focuses on multiclass problems and is based on AdaBoost.M2. Theoretical arguments were used to demonstrate their convergence and also that both methods retain the main features of their traditional counterparts. In addition, we performed experiments to compare the accuracy as well as the memory usage of the proposed methods against other approaches using 20 well-known datasets. Results suggest that, in many different situations, the proposed algorithms maintain high accuracies, outperforming the other tested methods.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Anhänge
Nur mit Berechtigung zugänglich
Fußnoten
1
Because it refers to the exact calculation, the symbol \(\approx \) is not used in this equation.
 
2
Since OABM2 has two parameters (w and L) and three values were selected for each one, a total of 10 tests per dataset were performed, including the default values of the method.
 
3
The complexities were defined taking into account the implementations made available by the authors or available in the MOA framework.
 
4
This optimization technique is known as memoization.
 
Literatur
Zurück zum Zitat Barros RSM, Cabral DRL, Gonçalves PM Jr, Santos SGTC (2017) RDDM: reactive drift detection method. Expert Syst Appl 90(C):344–355CrossRef Barros RSM, Cabral DRL, Gonçalves PM Jr, Santos SGTC (2017) RDDM: reactive drift detection method. Expert Syst Appl 90(C):344–355CrossRef
Zurück zum Zitat Beygelzimer A, Hazan E, Kale S, Luo H (2015a) Online gradient boosting. In: Proceedings of 28th international conference on neural information processing systems (NIPS’15). MIT Press, pp 2458–2466 Beygelzimer A, Hazan E, Kale S, Luo H (2015a) Online gradient boosting. In: Proceedings of 28th international conference on neural information processing systems (NIPS’15). MIT Press, pp 2458–2466
Zurück zum Zitat Beygelzimer A, Kale S, Luo H (2015b) Optimal and adaptive algorithms for online boosting. In: Bach F, Blei D (eds) ICML, JMLR workshop and conference proceedings, vol 37, pp 2323–2331 Beygelzimer A, Kale S, Luo H (2015b) Optimal and adaptive algorithms for online boosting. In: Bach F, Blei D (eds) ICML, JMLR workshop and conference proceedings, vol 37, pp 2323–2331
Zurück zum Zitat Bifet A, Holmes G, Pfahringer B, Kirkby R, Gavaldà R (2009) New ensemble methods for evolving data streams. In: Proceedings of 15th ACM international conference on knowledge discovery and data mining, New York, pp 139–148 Bifet A, Holmes G, Pfahringer B, Kirkby R, Gavaldà R (2009) New ensemble methods for evolving data streams. In: Proceedings of 15th ACM international conference on knowledge discovery and data mining, New York, pp 139–148
Zurück zum Zitat Bifet A, Holmes G, Kirkby R, Pfahringer B (2010) MOA: massive online analysis. J Mach Learn Res 11:1601–1604 Bifet A, Holmes G, Kirkby R, Pfahringer B (2010) MOA: massive online analysis. J Mach Learn Res 11:1601–1604
Zurück zum Zitat Brzezinski D, Stefanowski J (2013) Reacting to different types of concept drift: the accuracy updated ensemble algorithm. IEEE Trans Neural Netw Learn Syst 25(1):81–94CrossRef Brzezinski D, Stefanowski J (2013) Reacting to different types of concept drift: the accuracy updated ensemble algorithm. IEEE Trans Neural Netw Learn Syst 25(1):81–94CrossRef
Zurück zum Zitat Chen S, Lin H, Lu C (2012) An online boosting algorithm with theoretical justifications. In: Proceedings of the 29th international conference on machine learning, ICML’12, pp 1873–1880 Chen S, Lin H, Lu C (2012) An online boosting algorithm with theoretical justifications. In: Proceedings of the 29th international conference on machine learning, ICML’12, pp 1873–1880
Zurück zum Zitat Cormen T, Leiserson C, Rivest R, Stein C (2009) Growth of functions. In: Introduction to algorithms, 3rd edn, chap 1. The MIT Press, pp 43–64 Cormen T, Leiserson C, Rivest R, Stein C (2009) Growth of functions. In: Introduction to algorithms, 3rd edn, chap 1. The MIT Press, pp 43–64
Zurück zum Zitat Demšar J (2006) Statistical comparisons of classifiers over multiple data sets. J Mach Learn Res 7:1–30MathSciNetMATH Demšar J (2006) Statistical comparisons of classifiers over multiple data sets. J Mach Learn Res 7:1–30MathSciNetMATH
Zurück zum Zitat Freund Y, Schapire R (1996) Experiments with a new boosting algorithm. In: International conference on machine learning, pp 148–156 Freund Y, Schapire R (1996) Experiments with a new boosting algorithm. In: International conference on machine learning, pp 148–156
Zurück zum Zitat Frías-Blanco I, Campo-Ávila J, Ramos-Jiménez G, Morales-Bueno R, Ortiz-Díaz A, Caballero-Mota Y (2015) Online and non-parametric drift detection methods based on hoeffding’s bounds. IEEE Trans Knowl Data Eng 27(3):810–823CrossRef Frías-Blanco I, Campo-Ávila J, Ramos-Jiménez G, Morales-Bueno R, Ortiz-Díaz A, Caballero-Mota Y (2015) Online and non-parametric drift detection methods based on hoeffding’s bounds. IEEE Trans Knowl Data Eng 27(3):810–823CrossRef
Zurück zum Zitat Frías-Blanco I, Verdecia-Cabrera A, Ortiz-Díaz A, Carvalho A (2016) Fast adaptive stacking of ensembles. In: Proceedings of the 31st ACM symposium on applied computing (SAC’16), Pisa, Italy, pp 929–934 Frías-Blanco I, Verdecia-Cabrera A, Ortiz-Díaz A, Carvalho A (2016) Fast adaptive stacking of ensembles. In: Proceedings of the 31st ACM symposium on applied computing (SAC’16), Pisa, Italy, pp 929–934
Zurück zum Zitat Hulten G, Spencer L, Domingos P (2001) Mining time-changing data streams. In: Proceedings of 7th ACM SIGKDD international conference on knowledge discovery and data mining (KDD), New York, pp 97–106 Hulten G, Spencer L, Domingos P (2001) Mining time-changing data streams. In: Proceedings of 7th ACM SIGKDD international conference on knowledge discovery and data mining (KDD), New York, pp 97–106
Zurück zum Zitat Jung YH, Goetz J, Tewari A (2017) Online multiclass boosting. In: Guyon I, Luxburg UV, Bengio S, Wallach H, Fergus R, Vishwanathan S, Garnett R (eds) Advances in neural information processing systems 30. Curran Associates, Inc., Red Hook, pp 920–929 Jung YH, Goetz J, Tewari A (2017) Online multiclass boosting. In: Guyon I, Luxburg UV, Bengio S, Wallach H, Fergus R, Vishwanathan S, Garnett R (eds) Advances in neural information processing systems 30. Curran Associates, Inc., Red Hook, pp 920–929
Zurück zum Zitat Mukherjee I, Schapire RE (2013) A theory of multiclass boosting. J Mach Learn Res 14(1):437–497MathSciNetMATH Mukherjee I, Schapire RE (2013) A theory of multiclass boosting. J Mach Learn Res 14(1):437–497MathSciNetMATH
Zurück zum Zitat Oza NC, Russell S (2001) Online bagging and boosting. In: Artificial intelligence and statistics, Morgan Kaufmann, pp 105–112 Oza NC, Russell S (2001) Online bagging and boosting. In: Artificial intelligence and statistics, Morgan Kaufmann, pp 105–112
Zurück zum Zitat Schapire R, Freund Y (2012a) Attaining the best possible accuracy. In: Boosting: foundations and algorithms, chap 12. MIT Press, pp 377–413 Schapire R, Freund Y (2012a) Attaining the best possible accuracy. In: Boosting: foundations and algorithms, chap 12. MIT Press, pp 377–413
Zurück zum Zitat Schapire R, Freund Y (2012b) Learning to rank. In: Boosting: foundations and algorithms, chap 11. MIT Press, pp 341–374 Schapire R, Freund Y (2012b) Learning to rank. In: Boosting: foundations and algorithms, chap 11. MIT Press, pp 341–374
Zurück zum Zitat Schapire R, Freund Y (2012c) Loss minimization and generalizations of boosting. In: Boosting: foundations and algorithms, chap 7. MIT Press, pp 175–226 Schapire R, Freund Y (2012c) Loss minimization and generalizations of boosting. In: Boosting: foundations and algorithms, chap 7. MIT Press, pp 175–226
Zurück zum Zitat Schapire R, Freund Y (2012d) Multiclass classification problems. In: Boosting: foundations and algorithms, chap 10. MIT Press, pp 303–340 Schapire R, Freund Y (2012d) Multiclass classification problems. In: Boosting: foundations and algorithms, chap 10. MIT Press, pp 303–340
Zurück zum Zitat Schapire R, Freund Y (2012e) Using confidence-rated weak predictions. In: Boosting: foundations and algorithms, chap 9. MIT Press, pp 271–302 Schapire R, Freund Y (2012e) Using confidence-rated weak predictions. In: Boosting: foundations and algorithms, chap 9. MIT Press, pp 271–302
Zurück zum Zitat Servedio RA (2003) Smooth boosting and learning with malicious noise. J Mach Learn Res 4:633–648MathSciNetMATH Servedio RA (2003) Smooth boosting and learning with malicious noise. J Mach Learn Res 4:633–648MathSciNetMATH
Metadaten
Titel
Online AdaBoost-based methods for multiclass problems
verfasst von
Silas Garrido Teixeira de Carvalho Santos
Roberto Souto Maior de Barros
Publikationsdatum
01.03.2019
Verlag
Springer Netherlands
Erschienen in
Artificial Intelligence Review / Ausgabe 2/2020
Print ISSN: 0269-2821
Elektronische ISSN: 1573-7462
DOI
https://doi.org/10.1007/s10462-019-09696-6

Weitere Artikel der Ausgabe 2/2020

Artificial Intelligence Review 2/2020 Zur Ausgabe