ABSTRACT
We address classification problems for which the training instances are governed by a distribution that is allowed to differ arbitrarily from the test distribution---problems also referred to as classification under covariate shift. We derive a solution that is purely discriminative: neither training nor test distribution are modeled explicitly. We formulate the general problem of learning under covariate shift as an integrated optimization problem. We derive a kernel logistic regression classifier for differing training and test distributions.
- Bickel, S., & Scheffer, T. (2007). Dirichlet-enhanced spam filtering based on biased samples. Advances in Neural Information Processing Systems.Google Scholar
- Dudik, M., Schapire, R., & Phillips, S. (2005). Correcting sample selection bias in maximum entropy density estimation. Advances in Neural Information Processing Systems.Google Scholar
- Elkan, C. (2001). The foundations of cost-sensitive learning. Proceedings of the International Joint Conference on Artificial Intellligence. Google ScholarDigital Library
- Heckman, J. (1979). Sample selection bias as a specification error. Econometrica, 47, 153--161.Google ScholarCross Ref
- Huang, J., Smola, A., Gretton, A., Borgwardt, K., & Schöölkopf, B. (2007). Correcting sample selection bias by unlabeled data. Advances in Neural Information Processing Systems.Google Scholar
- Japkowicz, N., & Stephen, S. (2002). The class imbalance problem: A systematic study. Intelligent Data Analysis, 6, 429--449. Google ScholarDigital Library
- Manski, C., & Lerman, S. (1977). The estimation of choice probabilities from choice based samples. Econometrica, 45, 1977--1988.Google ScholarCross Ref
- Shimodaira, H. (2000). Improving predictive inference under covariate shift by weighting the log-likelihood function. Journal of Statistical Planning and Inference, 90, 227--244.Google ScholarCross Ref
- Silverman, B. W. (1986). Density estimation for statistics and data analysis. Chapman & Hall, London.Google Scholar
- Sugiyama, M., & Müüller, K.-R. (2005). Model selection under covariate shift. Proceedings of the International Conference on Artificial Neural Networks. Google ScholarDigital Library
- Xue, Y., Liao, X., Carin, L., & Krishnapuram, B. (2007). Multi-task learning for classification with dirichlet process priors. Journal of Machine Learning Research, 8, 35--63. Google ScholarDigital Library
- Zadrozny, B. (2004). Learning and evaluating classifiers under sample selection bias. Proceedings of the International Conference on Machine Learning. Google ScholarDigital Library
- Discriminative learning for differing training and test distributions
Recommendations
Multiple instance learning by discriminative training of Markov networks
UAI'13: Proceedings of the Twenty-Ninth Conference on Uncertainty in Artificial IntelligenceWe introduce a graphical framework for multiple instance learning (MIL) based on Markov networks. This framework can be used to model the traditional MIL definition as well as more general MIL definitions. Different levels of ambiguity - the portion of ...
Confidence-rated discriminative partial label learning
AAAI'17: Proceedings of the Thirty-First AAAI Conference on Artificial IntelligencePartial label learning aims to induce a multi-class classifier from training examples where each of them is associated with a set of candidate labels, among which only one label is valid. The common discriminative solution to learn from partial label ...
Mismatched training and test distributions can outperform matched ones
In learning theory, the training and test sets are assumed to be drawn from the same probability distribution. This assumption is also followed in practical situations, where matching the training and test distributions is considered desirable. Contrary ...
Comments