skip to main content
10.1145/2851613.2851655acmconferencesArticle/Chapter ViewAbstractPublication PagessacConference Proceedingsconference-collections
research-article

Fast adaptive stacking of ensembles

Published:04 April 2016Publication History

ABSTRACT

This paper presents a new ensemble method for learning from non-stationary data streams. In these situations, massive data are constantly generated at high speed and their target function can change over time. The proposed method, named Fast Adaptive Stacking of Ensembles (FASE), uses a meta-classifier to combine the predictions from the base classifiers in the ensemble. FASE maintains a set of adaptive learners, in order to deal with concept drifting data. The new algorithm is able to process the input data in constant time and space computational complexity. It only receives as parameters the confidence level for the change detection mechanism and the number of base classifiers. These characteristics make FASE very suitable for learning from non-stationary data streams. We empirically compare the new algorithm with various state-of-the-art ensemble methods for learning in non-stationary data streams. We use a Naïve Bayes classifier and a Perceptron to evaluate the performance of the algorithms over real-world datasets. The experiment results show that FASE presents higher predictive accuracy in the investigated tasks, being also able to bound its computational cost.

References

  1. M. Baena, J. del Campo, R. Fidalgo, A. Bifet, R. Gavaldà, and R. Morales. Early Drift Detection Method. In 4th Int. Workshop on Knowledge Discovery from Data Streams, 2006.Google ScholarGoogle Scholar
  2. M. Basseville and I. Nikiforov. Detection of Abrupt Changes: Theory and Application. Prentice-Hall, Englewood Cliffs, NJ, 1993. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. S. Bian and W. Wang. On diversity and accuracy of homogeneous and heterogeneous ensembles. Int. J. Hybrid Intell. Syst., 4(2):103--128, Apr. 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. A. Bifet. Adaptive Stream Mining: Pattern Learning and Mining from Evolving Data Streams, volume 207. IOS Press, 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. A. Bifet, E. Frank, G. Holmes, and B. Pfahringer. Ensembles of restricted hoeffding trees. ACM Trans. Intell. Syst. Technol., 3(2):30:1--30:20, 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. A. Bifet and R. Gavaldà. Learning from time-changing data with adaptive windowing. In Proc. 7th SIAM Int. Conf. on Data Mining, 2007.Google ScholarGoogle ScholarCross RefCross Ref
  7. A. Bifet, G. Holmes, R. Kirkby, and B. Pfahringer. MOA: Massive Online Analysis. Journal of Machine Learning Research, 11:1601--1604, 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. A. Bifet, G. Holmes, B. Pfahringer, and R. Gavaldà. New Ensemble Methods for Evolving Data Streams. In Proc. 15th ACM SIGKDD Int. Conf. on Knowledge Discovery and Data Mining, pages 139--148, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. N. Cesa-Bianchi and G. Lugosi. Prediction, Learning, and Games. Cambridge University Press, New York, NY, USA, 2006. Google ScholarGoogle ScholarCross RefCross Ref
  10. W. Fan. Systematic data selection to mine concept-drifting data streams. In Proceedings of the Tenth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD '04, pages 128--137, 2004. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. A. Frank and A. Asuncion. UCI Machine Learning Repository, 2010.Google ScholarGoogle Scholar
  12. I. Frías, J. del Campo, G. Ramos, R. Morales, A. Ortiz, and Y. Caballero. Aprendiendo con detección de cambios online. Computación y Sistemas, 18(1):169--183, 2014. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. I. Frías, J. del Campo, G. Ramos, R. Morales, A. Ortiz, and Y. Caballero. Online and non-parametric drift detection methods based on Hoeffding's bounds. IEEE Transactions on Knowledge and Data Engineering, 14(3):810--823, 2015.Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. J. Gama and M. Gaber. Learning from Data Streams: Processing Techniques in Sensor Networks. Springer-Verlag, 1 edition, 2007. Google ScholarGoogle ScholarCross RefCross Ref
  15. J. Gama, P. Medas, G. Castillo, and P. Rodrigues. Learning with Drift Detection. In Proc. 20th Brazilian Symposium on Artificial Intelligence, pages 286--295, 2004.Google ScholarGoogle ScholarCross RefCross Ref
  16. J. Gama, P. Rodrigues, and G. Castillo. Evaluating Algorithms that Learn from Data Streams. In Proc. 2009 ACM Symposium on Applied Computing, pages 1496--1500, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. J. Gama, R. Sebastião, and P. Rodrigues. On evaluating stream learning algorithms. Machine Learning, 90(3):317--346, 2013. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. J. Gama, R. Sebastião, and P. Rodrigues. Issues in Evaluation of Stream Learning Algorithms. In Proc. 15th ACM SIGKDD Int. Conf. on Knowledge Discovery and Data Mining, pages 329--338, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. M. Harries, C. Sammut, and K. Horn. Extracting Hidden Context. Machine Learning, 32(2):101--126, 1998. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. I. Katakis, G. Tsoumakas, and I. Vlahavas. An Ensemble of Classifiers for coping with Recurring Contexts in Data Streams. In Proc. 18th European Conf. on Artificial Intelligence, pages 763--764, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. L. Kuncheva. Combining Pattern Classifiers: Methods and Algorithms. Wiley-Interscience, 2004. Google ScholarGoogle ScholarCross RefCross Ref
  22. H. Nguyen, Y. Woon, W. Ng, and L. Wan. Heterogeneous ensemble for feature drifts in data streams. In Proceedings of the 16th Pacific-Asia Conference on Advances in Knowledge Discovery and Data Mining - Volume Part II, PAKDD'12, pages 1--12, 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. N. Oza. Online bagging and boosting. In 2005 IEEE Int. Conf. on Systems, Man and Cybernetics, volume 3, pages 2340--2345, 2005.Google ScholarGoogle ScholarCross RefCross Ref
  24. N. Oza and S. Russell. Experimental comparisons of online and batch versions of bagging and boosting. In Proceedings of the Seventh ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD '01, pages 359--364, New York, NY, USA, 2001. ACM. Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. N. Oza and S. Russell. Online Bagging and Boosting. In Proc. 8th Int. Workshop on Artificial Intelligence and Statistics, pages 105--112, 2001.Google ScholarGoogle Scholar

Index Terms

  1. Fast adaptive stacking of ensembles

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      SAC '16: Proceedings of the 31st Annual ACM Symposium on Applied Computing
      April 2016
      2360 pages
      ISBN:9781450337397
      DOI:10.1145/2851613

      Copyright © 2016 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 4 April 2016

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article

      Acceptance Rates

      SAC '16 Paper Acceptance Rate252of1,047submissions,24%Overall Acceptance Rate1,650of6,669submissions,25%

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader