Abstract
We briefly describe our approach for the KDD99 Classification Cup. The solution is essentially a mixture of bagging and boosting. Additionally, asymmetric error costs are taken into account by minimizing the so-called conditional risk. Furthermore, the standard sampling with replacement methodology of bagging was modified to put a specific focus on the smaller but expensive-if-predicted-wrongly classes.
- Duda R. O., Hart P. E.: Pattern Classification and Scene Analysis, Wiley, Chichester/London/New York, 1973.Google Scholar
- MetaL 1999: An ESPRIT Long-Term Research Project: A Meta-Learning Assistant for Providing User Support in Data Mining and Machine Learning, http://www.cs.bris.ac.uk/cgc/METAL/Google Scholar
Index Terms
- Winning the KDD99 classification cup: bagged boosting
Recommendations
Switching class labels to generate classification ensembles
Ensembles that combine the decisions of classifiers generated by using perturbed versions of the training set where the classes of the training examples are randomly switched can produce a significant error reduction, provided that large numbers of ...
Using boosting to prune bagging ensembles
Boosting is used to determine the order in which classifiers are aggregated in a bagging ensemble. Early stopping in the aggregation of the classifiers in the ordered bagging ensemble allows the identification of subensembles that require less memory ...
Class-switching neural network ensembles
This article investigates the properties of class-switching ensembles composed of neural networks and compares them to class-switching ensembles of decision trees and to standard ensemble learning methods, such as bagging and boosting. In a class-...
Comments