ABSTRACT
We present a new approach to multiple instance learning (MIL) that is particularly effective when the positive bags are sparse (i.e. contain few positive instances). Unlike other SVM-based MIL methods, our approach more directly enforces the desired constraint that at least one of the instances in a positive bag is positive. Using both artificial and real-world data, we experimentally demonstrate that our approach achieves greater accuracy than state-of-the-art MIL methods when positive bags are sparse, and performs competitively when they are not. In particular, our approach is the best performing method for image region classification.
- Andrews, S., Tsochantaridis, I., & Hofmann, T. (2003). Support vector machines for multiple-instance learning. Advances in Neural Information Processing Systems 15 (pp. 561--568). Vancouver, BC: MIT Press.Google Scholar
- Bunescu, R. C., & Mooney, R. J. (2006). Subsequence kernels for relation extraction. Advances in Neural Information Processing Systems 18. Vancouver, BC.Google Scholar
- Chapelle, O., & Zien, A. (2005). Semi-supervised classification by low density separation. Proceedings of the Tenth International Workshop on Artificial Intelligence and Statistics.Google Scholar
- Collobert, R., Sinz, F., Weston, J., & Bottou, L. (2006). Large scale transductive SVMs. Journal of Machine Learning Research, 7(Aug), 1687--1712. Google ScholarDigital Library
- Dietterich, T. G., Lathrop, R. H., & Lozano-Perez, T. (1997). Solving the multiple instance problem with axis-parallel rectangles. Artificial Intelligence, 89, 31--71. Google ScholarDigital Library
- Gartner, T., Flach, P., Kowalczyk, A., & Smola, A. (2002). Multi-instance kernels. In Proceedings of the 19th International Conference on Machine Learning (pp. 179--186). Sydney, Australia: Morgan Kaufmann. Google ScholarDigital Library
- Joachims, T. (1999). Transductive inference for text classification using support vector machines. Proceedings of the Sixteenth International Conference on Machine Learning (ICML-99) (pp. 200--209). Bled, Slovenia. Google ScholarDigital Library
- Kuck, H., & de Freitas, N. (2005). Learning about individuals from group statistics. Proceedings of the 21th Annual Conference on Uncertainty in Artificial Intelligence (UAI-05) (p. 332). Edinburgh, Scotland.Google Scholar
- Maron, O. (1998). Learning from ambiguity (Technical Report Doctoral dissertation). Dept. of EECS, MIT, Cambridge, MA. Google ScholarDigital Library
- Platt, J. (1999). Fast training of support vector machines using sequential minimal optimization. In B. Scholkopf, C. J. C. Burges and A. J. Smola (Eds.), Advances in kernel methods - support vector learning, 185--208. Cambridge, MA: MIT Press. Google ScholarDigital Library
- Ray, S., & Craven, M. (2005). Supervised versus multiple instance learning: An empirical comparison. Proceedings of 22nd International Conference on Machine Learning (ICML-2005) (pp. 697--704). Bonn, Germany. Google ScholarDigital Library
- Vapnik, V. N. (1995). The nature of statistical learning theory. Berlin: Springer-Verlag. Google ScholarDigital Library
- Yuille, A. L., & Rangarajan, A. (2002). The concave-convex procedure (CCCP). Advances in Neural Information Processing Systems 14. Cambridge, MA: MIT Press. Google ScholarDigital Library
- Zhang, Q., Goldman, S. A., Yu, W., & Fritts, J. (2002). Content-based image retrieval using multiple-instance learning. Proceedings of 19th International Conference on Machine Learning (ICML-2002) (pp. 682--689). Google ScholarDigital Library
- Multiple instance learning for sparse positive bags
Recommendations
Multiple-Instance Learning from Similar and Dissimilar Bags
KDD '21: Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data MiningMultiple-instance learning (MIL) is an important weakly supervised binary classification problem, where training instances are arranged in bags, and each bag is assigned a positive or negative label. Most of the previous studies for MIL assume that ...
Multiple-Instance Learning From Unlabeled Bags With Pairwise Similarity
In <italic>multiple-instance learning</italic> (MIL), each training example is represented by a bag of instances. A training bag is either negative if it contains no positive instances or positive if it has at least one positive instance. Previous MIL ...
Convex formulation of multiple instance learning from positive and unlabeled bags
AbstractMultiple instance learning (MIL) is a variation of traditional supervised learning problems where data (referred to as bags) are composed of sub-elements (referred to as instances) and only bag labels are available. MIL has a variety ...
Comments