ABSTRACT
We consider a setting for discriminative semi-supervised learning where unlabeled data are used with a generative model to learn effective feature representations for discriminative training. Within this framework, we revisit the two-view feature generation model of co-training and prove that the optimum predictor can be expressed as a linear combination of a few features constructed from unlabeled data. From this analysis, we derive methods that employ two views but are very different from co-training. Experiments show that our approach is more robust than co-training and EM, under various data generation conditions.
- Ando, R. K., & Zhang, T. (2005). A framework for learning predictive structures from multiple tasks and unlabeled data. Journal of Machine Learning Research, 6, 1817--1853. Google ScholarDigital Library
- Blum, A., & Mitchell, T. (1998). Combining labeled and unlabeled data with co-training. Proceedings of the eleventh annual conference on Computational learning theory (pp. 92--100). Google ScholarDigital Library
- Dasgupta, S., Littman, M., & McAllester, D. (2001). PAC generalization bounds for co-training. NIPS' 01.Google Scholar
- Nigam, K., McCallum, A. K., Thrun, S., & Mitchell, T. (2000). Text classification from labeled and unlabeled documents using EM. Machine Learning, Special issue on information retrieval, 103--134. Google ScholarDigital Library
- Vapnik, V. (1998). Statistical learning theory. New York: John Wiley & Sons.Google Scholar
- Zhang, T., & Oles, F. J. (2000). A probability analysis on the value of unlabeled data for classification problems. ICML 2000 (pp. 1191--1198).Google Scholar
- Zhu, X., Ghahramani, Z., & Lafferty, J. (2003). Semi-supervised learning using gaussian fields and harmonic functions. ICML 2003.Google ScholarDigital Library
- Two-view feature generation model for semi-supervised learning
Recommendations
Cross-Domain Semi-Supervised Learning Using Feature Formulation
Semi-Supervised Learning (SSL) traditionally makes use of unlabeled samples In this paper, sample and instance are interchangeable terms. by including them into the training set through an automated labeling process. Such a primitive Semi-Supervised ...
Inductive Semi-supervised Multi-Label Learning with Co-Training
KDD '17: Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data MiningIn multi-label learning, each training example is associated with multiple class labels and the task is to learn a mapping from the feature space to the power set of label space. It is generally demanding and time-consuming to obtain labels for training ...
Generalized semi-supervised learning via self-supervised feature adaptation
NIPS '23: Proceedings of the 37th International Conference on Neural Information Processing SystemsTraditional semi-supervised learning (SSL) assumes that the feature distributions of labeled and unlabeled data are consistent which rarely holds in realistic scenarios. In this paper, we propose a novel SSL setting, where unlabeled samples are drawn ...
Comments