Skip to main content
Erschienen in: Neural Processing Letters 1/2019

13.02.2019

On Incremental Learning for Gradient Boosting Decision Trees

verfasst von: Chongsheng Zhang, Yuan Zhang, Xianjin Shi, George Almpanidis, Gaojuan Fan, Xiajiong Shen

Erschienen in: Neural Processing Letters | Ausgabe 1/2019

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Boosting algorithms, as a class of ensemble learning methods, have become very popular in data classification, owing to their strong theoretical guarantees and outstanding prediction performance. However, most of these boosting algorithms were designed for static data, thus they can not be directly applied to on-line learning and incremental learning. In this paper, we propose a novel algorithm that incrementally updates the classification model built upon gradient boosting decision tree (GBDT), namely iGBDT. The main idea of iGBDT is to incrementally learn a new model but without running GBDT from scratch, when new data is dynamically arriving in batch. We conduct large-scale experiments to validate the effectiveness and efficiency of iGBDT. All the experimental results show that, in terms of model building/updating time, iGBDT obtains significantly better performance than the conventional practice that always runs GBDT from scratch when a new batch of data arrives, while still keeping the same classification accuracy. iGBDT can be used in many applications that require in-time analysis of continuously arriving or real-time user-generated data, such as behaviour targeting, Internet advertising, recommender systems, etc.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
2.
Zurück zum Zitat Babenko B, Yang MH, Belongie S (2009) A family of online boosting algorithms. In: IEEE 12th international conference on computer vision workshops (ICCV workshops). IEEE, pp 1346–1353 Babenko B, Yang MH, Belongie S (2009) A family of online boosting algorithms. In: IEEE 12th international conference on computer vision workshops (ICCV workshops). IEEE, pp 1346–1353
3.
Zurück zum Zitat Beygelzimer A, Hazan E, Kale S, Luo H (2015a) Online gradient boosting. In: NIPS, pp 2458–2466 Beygelzimer A, Hazan E, Kale S, Luo H (2015a) Online gradient boosting. In: NIPS, pp 2458–2466
4.
Zurück zum Zitat Beygelzimer A, Kale S, Luo H (2015b) Optimal and adaptive algorithms for online boosting. In: ICML Beygelzimer A, Kale S, Luo H (2015b) Optimal and adaptive algorithms for online boosting. In: ICML
5.
Zurück zum Zitat Chapelle O, Chang Y (2011) Yahoo! learning to rank challenge overview. In: JMLR proceedings, pp 1–24 Chapelle O, Chang Y (2011) Yahoo! learning to rank challenge overview. In: JMLR proceedings, pp 1–24
6.
Zurück zum Zitat Chen ST, Lin HT, Lu CJ (2012) An online boosting algorithm with theoretical justifications. In: ICML Chen ST, Lin HT, Lu CJ (2012) An online boosting algorithm with theoretical justifications. In: ICML
8.
Zurück zum Zitat Domingos P, Hulten G (2000) Mining high-speed data streams. In: ACM SIGKDD. ACM, pp 71–80 Domingos P, Hulten G (2000) Mining high-speed data streams. In: ACM SIGKDD. ACM, pp 71–80
10.
Zurück zum Zitat Freund Y, Schapire RE (1997) A decision-theoretic generalization of on-line learning and an application to boosting. J Comput Syst Sci 55:119–139MathSciNetCrossRefMATH Freund Y, Schapire RE (1997) A decision-theoretic generalization of on-line learning and an application to boosting. J Comput Syst Sci 55:119–139MathSciNetCrossRefMATH
12.
Zurück zum Zitat Gaber MM, Zaslavsky A, Krishnaswamy S (2005) Mining data streams: a review. ACM Sigmod Rec 34(2):18–26CrossRefMATH Gaber MM, Zaslavsky A, Krishnaswamy S (2005) Mining data streams: a review. ACM Sigmod Rec 34(2):18–26CrossRefMATH
13.
Zurück zum Zitat Grbovic M, Vucetic S (2011) Tracking concept change with incremental boosting by minimization of the evolving exponential loss. In: PKDD. Springer, Berlin, pp 516–532 Grbovic M, Vucetic S (2011) Tracking concept change with incremental boosting by minimization of the evolving exponential loss. In: PKDD. Springer, Berlin, pp 516–532
14.
Zurück zum Zitat Hulten G, Spencer L, Domingos PM (2001) Mining time-changing data streams. In: ACM SIGKDD, pp 97–106 Hulten G, Spencer L, Domingos PM (2001) Mining time-changing data streams. In: ACM SIGKDD, pp 97–106
15.
Zurück zum Zitat Leistner C, Saffari A, Roth PM, Bischof H (2009) On robustness of on-line boosting: a competitive study. In: IEEE 12th international conference on computer vision workshops (ICCV workshops), pp 1362–1369 Leistner C, Saffari A, Roth PM, Bischof H (2009) On robustness of on-line boosting: a competitive study. In: IEEE 12th international conference on computer vision workshops (ICCV workshops), pp 1362–1369
16.
Zurück zum Zitat Liu X, Yu T (2007) Gradient feature selection for online boosting. In: 2007 IEEE 11th international conference on computer vision (ICCV). IEEE, pp 1–8 Liu X, Yu T (2007) Gradient feature selection for online boosting. In: 2007 IEEE 11th international conference on computer vision (ICCV). IEEE, pp 1–8
17.
Zurück zum Zitat Oza NC, Russell S (2001) Experimental comparisons of online and batch versions of bagging and boosting. In: ACM SIGKDD. ACM, pp 359–364 Oza NC, Russell S (2001) Experimental comparisons of online and batch versions of bagging and boosting. In: ACM SIGKDD. ACM, pp 359–364
18.
Zurück zum Zitat Oza NC, Russell SJ (2001) Online bagging and boosting. In: Eighth international workshop on artificial intelligence and statistics, pp 105–112 Oza NC, Russell SJ (2001) Online bagging and boosting. In: Eighth international workshop on artificial intelligence and statistics, pp 105–112
19.
Zurück zum Zitat Pavlov DY, Gorodilov A, Brunk CA (2010) Bagboo: a scalable hybrid bagging-the-boosting model. In: Proceedings of the 19th ACM conference on information and knowledge management, CIKM 2010, Toronto, Ontario, Canada, October 26–30, 2010, pp 1897–1900 Pavlov DY, Gorodilov A, Brunk CA (2010) Bagboo: a scalable hybrid bagging-the-boosting model. In: Proceedings of the 19th ACM conference on information and knowledge management, CIKM 2010, Toronto, Ontario, Canada, October 26–30, 2010, pp 1897–1900
20.
Zurück zum Zitat Pelossof R, Jones M, Vovsha I, Rudin C (2009) Online coordinate boosting. In: IEEE 12th international conference on computer vision workshops (ICCV workshops). IEEE, pp 1354–1361 Pelossof R, Jones M, Vovsha I, Rudin C (2009) Online coordinate boosting. In: IEEE 12th international conference on computer vision workshops (ICCV workshops). IEEE, pp 1354–1361
21.
Zurück zum Zitat Perkins S, Lacker K, Theiler J (2003) Grafting: fast, incremental feature selection by gradient descent in function space. J Mach Learn Res 3(Mar):1333–1356MathSciNetMATH Perkins S, Lacker K, Theiler J (2003) Grafting: fast, incremental feature selection by gradient descent in function space. J Mach Learn Res 3(Mar):1333–1356MathSciNetMATH
22.
Zurück zum Zitat Zhang C, Hao Y, Mazuran M, Zaniolo C, Mousavi H, Masseglia F (2013) Mining frequent itemsets over tuple-evolving data streams. In: Proceedings of the 28th annual ACM symposium on applied computing, SAC’13, Coimbra, Portugal, March 18–22, 2013, pp 267–274 Zhang C, Hao Y, Mazuran M, Zaniolo C, Mousavi H, Masseglia F (2013) Mining frequent itemsets over tuple-evolving data streams. In: Proceedings of the 28th annual ACM symposium on applied computing, SAC’13, Coimbra, Portugal, March 18–22, 2013, pp 267–274
Metadaten
Titel
On Incremental Learning for Gradient Boosting Decision Trees
verfasst von
Chongsheng Zhang
Yuan Zhang
Xianjin Shi
George Almpanidis
Gaojuan Fan
Xiajiong Shen
Publikationsdatum
13.02.2019
Verlag
Springer US
Erschienen in
Neural Processing Letters / Ausgabe 1/2019
Print ISSN: 1370-4621
Elektronische ISSN: 1573-773X
DOI
https://doi.org/10.1007/s11063-019-09999-3

Weitere Artikel der Ausgabe 1/2019

Neural Processing Letters 1/2019 Zur Ausgabe