Skip to main content

07.03.2024

Does Confusion Really Hurt Novel Class Discovery?

verfasst von: Haoang Chi, Wenjing Yang, Feng Liu, Long Lan, Tao Qin, Bo Han

Erschienen in: International Journal of Computer Vision

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

When sampling data of specific classes (i.e., known classes) for a scientific task, collectors may encounter unknown classes (i.e., novel classes). Since these novel classes might be valuable for future research, collectors will also sample them and assign them to several clusters with the help of known-class data. This assigning process is known as novel class discovery (NCD). However, category confusion is common in the sampling process and may make the NCD unreliable. To tackle this problem, this paper introduces a new and more realistic setting, where collectors may misidentify known classes and even confuse known classes with novel classes—we name it NCD under unreliable sampling (NUSA). We find that NUSA will empirically degrade existing NCD methods if taking no care of sampling errors. To handle NUSA, we propose an effective solution, named hidden-prototype-based discovery network (HPDN): (1) we try to obtain relatively clean data representations even with the confusedly sampled data; (2) we propose a mini-batch K-means variant for robust clustering, alleviating the negative impact of residual errors embedded in the representations by detaching the noisy supervision timely. Experiments demonstrate that, under NUSA, HPDN significantly outperforms competitive baselines (e.g., \(6\%\) more than the best baseline on CIFAR-10) and remains robust when encountering serious sampling errors.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Fußnoten
1
The novel classes are currently considered as one class.
 
Literatur
Zurück zum Zitat Arthur, D., & Vassilvitskii, S. (2006). k-means++: The advantages of careful seeding (Tech. Rep.). 450 Jane Stanford Way Stanford, CA: Stanford. Arthur, D., & Vassilvitskii, S. (2006). k-means++: The advantages of careful seeding (Tech. Rep.). 450 Jane Stanford Way Stanford, CA: Stanford.
Zurück zum Zitat Canas, G., Poggio, T., & Rosasco, L. (2012). Learning manifolds with k-means and k-flats. Neural Information Processing Systems, 25. Canas, G., Poggio, T., & Rosasco, L. (2012). Learning manifolds with k-means and k-flats. Neural Information Processing Systems, 25.
Zurück zum Zitat Cao, K., Brbic, M., & Leskovec, J. (2022). Open-world semi-supervised learning. In International conference on learning representations. Cao, K., Brbic, M., & Leskovec, J. (2022). Open-world semi-supervised learning. In International conference on learning representations.
Zurück zum Zitat Chen, T., Kornblith, S., Norouzi, M., & Hinton, G. (2020). A simple framework for contrastive learning of visual representations. In International conference on machine learning (pp. 1597–1607). PMLR. Chen, T., Kornblith, S., Norouzi, M., & Hinton, G. (2020). A simple framework for contrastive learning of visual representations. In International conference on machine learning (pp. 1597–1607). PMLR.
Zurück zum Zitat Chi, H., Liu, F., Han, B., Yang, W., Lan, L., Liu, T., Niu, G., Zhou, M., & Sugiyama, M. (2022). Meta discovery: Learning to discover novel classes given very limited data. In International conference on learning representations. Chi, H., Liu, F., Han, B., Yang, W., Lan, L., Liu, T., Niu, G., Zhou, M., & Sugiyama, M. (2022). Meta discovery: Learning to discover novel classes given very limited data. In International conference on learning representations.
Zurück zum Zitat Deng, J., Dong, W., Socher, R., Li, L. , Li, K., & Li, F. (2009). Imagenet: A large-scale hierarchical image database. In IEEE conference on computer vision and pattern recognition (pp. 248–255). IEEE. Deng, J., Dong, W., Socher, R., Li, L. , Li, K., & Li, F. (2009). Imagenet: A large-scale hierarchical image database. In IEEE conference on computer vision and pattern recognition (pp. 248–255). IEEE.
Zurück zum Zitat Fini, E., Sangineto, E., Lathuilière, S., Zhong, Z., Nabi, M., & Ricci, E. (2021). A unified objective for novel class discovery. In IEEE international conference on computer vision (pp. 9284–9292). Fini, E., Sangineto, E., Lathuilière, S., Zhong, Z., Nabi, M., & Ricci, E. (2021). A unified objective for novel class discovery. In IEEE international conference on computer vision (pp. 9284–9292).
Zurück zum Zitat Han, K., Rebuffi, S., Ehrhardt, S., Vedaldi, A., & Zisserman, A. (2020). Automatically discovering and learning new visual categories with ranking statistics. In International conference on learning representations. Han, K., Rebuffi, S., Ehrhardt, S., Vedaldi, A., & Zisserman, A. (2020). Automatically discovering and learning new visual categories with ranking statistics. In International conference on learning representations.
Zurück zum Zitat Han, K., Vedaldi, A., & Zisserman, A. (2019). Learning to discover novel visual categories via deep transfer clustering. In IEEE international conference on computer vision (pp. 8401–8409). Han, K., Vedaldi, A., & Zisserman, A. (2019). Learning to discover novel visual categories via deep transfer clustering. In IEEE international conference on computer vision (pp. 8401–8409).
Zurück zum Zitat Han, B., Yao, Q., Yu, X., Niu, G., Xu,M., Hu,W., Tsang, I., & Sugiyama M. (2018). Co-teaching: Robust training of deep neural networks with extremely noisy labels. Neural Information Processing Systems, 31. Han, B., Yao, Q., Yu, X., Niu, G., Xu,M., Hu,W., Tsang, I., & Sugiyama M. (2018). Co-teaching: Robust training of deep neural networks with extremely noisy labels. Neural Information Processing Systems, 31.
Zurück zum Zitat He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In IEEE conference on computer vision and pattern recognition (pp. 770–778). He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In IEEE conference on computer vision and pattern recognition (pp. 770–778).
Zurück zum Zitat Hsu, Y., Lv, Z., & Kira, Z. (2018). Learning to cluster in order to transfer across domains and tasks. In International conference on learning representations. Hsu, Y., Lv, Z., & Kira, Z. (2018). Learning to cluster in order to transfer across domains and tasks. In International conference on learning representations.
Zurück zum Zitat Hsu, Y., Lv, Z., Schlosser, J., Odom, P., & Kira, Z. (2019). Multi-class classification without multi-class labels. In International conference on learning representations. Hsu, Y., Lv, Z., Schlosser, J., Odom, P., & Kira, Z. (2019). Multi-class classification without multi-class labels. In International conference on learning representations.
Zurück zum Zitat Krause, J., Stark, M., Deng, J., & Fei-Fei, L. (2013). 3D object representations for fine-grained categorization. In IEEE international conference on computer vision workshops (pp. 554–561). Krause, J., Stark, M., Deng, J., & Fei-Fei, L. (2013). 3D object representations for fine-grained categorization. In IEEE international conference on computer vision workshops (pp. 554–561).
Zurück zum Zitat Krizhevsky, A., & Hinton, G. (2009). Learning multiple layers of features from tiny images. Technical Report TR-2009, University of Toronto. Krizhevsky, A., & Hinton, G. (2009). Learning multiple layers of features from tiny images. Technical Report TR-2009, University of Toronto.
Zurück zum Zitat Krizhevsky, A., Sutskever, I., & Hinton, G. E. (2012). Imagenet classification with deep convolutional neural networks. Neural Information Processing Systems, 25. Krizhevsky, A., Sutskever, I., & Hinton, G. E. (2012). Imagenet classification with deep convolutional neural networks. Neural Information Processing Systems, 25.
Zurück zum Zitat Li, X., Liu, T., Han, B., Niu, G., & Sugiyama, M. (2021). Provably end-to-end label-noise learning without anchor points. In International conference on machine learning (pp. 6403–6413). PMLR. Li, X., Liu, T., Han, B., Niu, G., & Sugiyama, M. (2021). Provably end-to-end label-noise learning without anchor points. In International conference on machine learning (pp. 6403–6413). PMLR.
Zurück zum Zitat Li, J. , Socher, R., & Hoi, S. C. H. (2020). DivideMix: Learning with noisy labels as semi-supervised learning. In International conference on learning representations. Li, J. , Socher, R., & Hoi, S. C. H. (2020). DivideMix: Learning with noisy labels as semi-supervised learning. In International conference on learning representations.
Zurück zum Zitat Li, J., Zhang, M., Xu, K., Dickerson, J., & Ba, J. (2021). How does a neural network’s architecture impact its robustness to noisy labels? Neural Information Processing Systems, 34, 9788–9803. Li, J., Zhang, M., Xu, K., Dickerson, J., & Ba, J. (2021). How does a neural network’s architecture impact its robustness to noisy labels? Neural Information Processing Systems, 34, 9788–9803.
Zurück zum Zitat Liu, S., Niles-Weed, J., Razavian, N., & Fernandez-Granda, C. (2020). Early-learning regularization prevents memorization of noisy labels. Advances in Neural Information Processing Systems, 33, 20331–20342. Liu, S., Niles-Weed, J., Razavian, N., & Fernandez-Granda, C. (2020). Early-learning regularization prevents memorization of noisy labels. Advances in Neural Information Processing Systems, 33, 20331–20342.
Zurück zum Zitat MacQueen, J. (1967). Some methods for classification and analysis of multivariate observations. In Proceedings of the fifth Berkeley symposium on mathematical statistics and probability (vol. 1, no. 14, pp. 281–297). MacQueen, J. (1967). Some methods for classification and analysis of multivariate observations. In Proceedings of the fifth Berkeley symposium on mathematical statistics and probability (vol. 1, no. 14, pp. 281–297).
Zurück zum Zitat Mahajan, M., Nimbhorkar, P., & Varadarajan, K. R. (2012). The planar k-means problem is NP-hard. Theoretical Computer Science, 442, 13–21.MathSciNetCrossRef Mahajan, M., Nimbhorkar, P., & Varadarajan, K. R. (2012). The planar k-means problem is NP-hard. Theoretical Computer Science, 442, 13–21.MathSciNetCrossRef
Zurück zum Zitat Ren, M., Zeng, W., Yang, B., & Urtasun, R. (2018). Learning to reweight examples for robust deep learning. In International conference on machine learning (pp. 4334–4343). PMLR. Ren, M., Zeng, W., Yang, B., & Urtasun, R. (2018). Learning to reweight examples for robust deep learning. In International conference on machine learning (pp. 4334–4343). PMLR.
Zurück zum Zitat Rosenberg, A., & Hirschberg, J. (2007). V-measure: A conditional entropy-based external cluster evaluation measure. In EMNLP-CoNLL (pp. 410–420). Rosenberg, A., & Hirschberg, J. (2007). V-measure: A conditional entropy-based external cluster evaluation measure. In EMNLP-CoNLL (pp. 410–420).
Zurück zum Zitat Simonyan, K., & Zisserman, A. (2015). Very deep convolutional networks for large-scale image recognition. In International conference on learning representations. Simonyan, K., & Zisserman, A. (2015). Very deep convolutional networks for large-scale image recognition. In International conference on learning representations.
Zurück zum Zitat Thorndike, R. L. (1953). Who belongs in the family. Psychometrika, 18(4), 267–276.CrossRef Thorndike, R. L. (1953). Who belongs in the family. Psychometrika, 18(4), 267–276.CrossRef
Zurück zum Zitat Van Gansbeke, W., Vandenhende, S., Georgoulis, S., Proesmans, M., & Van Gool, L. (2020). Scan: Learning to classify images without labels. In European conference on computer vision (pp. 268–285). Springer International Publishing. Van Gansbeke, W., Vandenhende, S., Georgoulis, S., Proesmans, M., & Van Gool, L. (2020). Scan: Learning to classify images without labels. In European conference on computer vision (pp. 268–285). Springer International Publishing.
Zurück zum Zitat Vaze, S., Han, K., Vedaldi, A. & Zisserman, A. (2022). Generalized category discovery. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp. 7492–7501). Vaze, S., Han, K., Vedaldi, A. & Zisserman, A. (2022). Generalized category discovery. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp. 7492–7501).
Zurück zum Zitat Wah, C., Branson, S., Welinder, P., Perona, P., & Belongie, S. (2011). The Caltech-UCSD Birds-200-2011 dataset. Technical Report CNS-TR-2011-001. Wah, C., Branson, S., Welinder, P., Perona, P., & Belongie, S. (2011). The Caltech-UCSD Birds-200-2011 dataset. Technical Report CNS-TR-2011-001.
Zurück zum Zitat Xia, X., Liu, T., Han, B., Wang, N., Gong, M., Liu, H., & Sugiyama, M. (2020). Part-dependent label noise: Towards instance-dependent label noise. Neural Information Processing Systems, 33, 7597–7610. Xia, X., Liu, T., Han, B., Wang, N., Gong, M., Liu, H., & Sugiyama, M. (2020). Part-dependent label noise: Towards instance-dependent label noise. Neural Information Processing Systems, 33, 7597–7610.
Zurück zum Zitat Xie, J., Girshick, R. B., & Farhadi, A. (2016). Unsupervised deep embedding for clustering analysis. In International conference on machine learning (pp. 478–487). PMLR. Xie, J., Girshick, R. B., & Farhadi, A. (2016). Unsupervised deep embedding for clustering analysis. In International conference on machine learning (pp. 478–487). PMLR.
Zurück zum Zitat Yang, M., Zhu, Y., Yu, J., Wu, A., & Deng, C. (2022). Divide and conquer: Compositional experts for generalized novel class discovery. In IEEE/CVF conference on computer vision and pattern recognition (pp. 14268–14277). Yang, M., Zhu, Y., Yu, J., Wu, A., & Deng, C. (2022). Divide and conquer: Compositional experts for generalized novel class discovery. In IEEE/CVF conference on computer vision and pattern recognition (pp. 14268–14277).
Zurück zum Zitat Yang, X., Deng, C., Wei, K., Yan, J., & Liu, W. (2020). Adversarial learning for robust deep clustering. Neural Information Processing Systems, 33, 9098–9108. Yang, X., Deng, C., Wei, K., Yan, J., & Liu, W. (2020). Adversarial learning for robust deep clustering. Neural Information Processing Systems, 33, 9098–9108.
Zurück zum Zitat Yao, Y., Liu, T., Han, B., Gong, M., Deng, J., Niu, G., & Sugiyama, M. (2020). Dual T: Reducing estimation error for transition matrix in label-noise learning. Neural Information Processing Systems, 33, 7260–7271. Yao, Y., Liu, T., Han, B., Gong, M., Deng, J., Niu, G., & Sugiyama, M. (2020). Dual T: Reducing estimation error for transition matrix in label-noise learning. Neural Information Processing Systems, 33, 7260–7271.
Zurück zum Zitat Zhan, X., Xie, J., Liu, Z., Ong, Y.-S., & Loy, C. C. (2020). Online deep clustering for unsupervised representation learning. In IEEE conference on computer vision and pattern recognition (pp. 6688–6697). Zhan, X., Xie, J., Liu, Z., Ong, Y.-S., & Loy, C. C. (2020). Online deep clustering for unsupervised representation learning. In IEEE conference on computer vision and pattern recognition (pp. 6688–6697).
Zurück zum Zitat Zhang, S., Khan, S., Shen, Z., Naseer, M., Chen, G., & Khan, F. (2023). PromptCAL: Contrastive affinity learning via auxiliary prompts for generalized novel category discovery. In CVPR (pp. 3479–3488). Zhang, S., Khan, S., Shen, Z., Naseer, M., Chen, G., & Khan, F. (2023). PromptCAL: Contrastive affinity learning via auxiliary prompts for generalized novel category discovery. In CVPR (pp. 3479–3488).
Zurück zum Zitat Zhang, C., Bengio, S., Hardt, M., Recht, B., & Vinyals, O. (2021). Understanding deep learning requires rethinking generalization Understanding deep learning requires rethinking generalization. Communications of the ACM, 643, 107–115. https://doi.org/10.1145/3446776CrossRef Zhang, C., Bengio, S., Hardt, M., Recht, B., & Vinyals, O. (2021). Understanding deep learning requires rethinking generalization Understanding deep learning requires rethinking generalization. Communications of the ACM, 643, 107–115. https://​doi.​org/​10.​1145/​3446776CrossRef
Zurück zum Zitat Zhao, B., & Han, K. (2021). Novel visual category discovery with dual ranking statistics and mutual knowledge distillation. Neural Information Processing Systems, 4, 22982–22994. Zhao, B., & Han, K. (2021). Novel visual category discovery with dual ranking statistics and mutual knowledge distillation. Neural Information Processing Systems, 4, 22982–22994.
Zurück zum Zitat Zhong, Z., Fini, E., Roy, S., Luo, Z., Ricci, E., & Sebe, N. (2021a). Neighborhood contrastive learning for novel class discovery. In IEEE conference on computer vision and pattern recognition (pp. 10867–10875). Zhong, Z., Fini, E., Roy, S., Luo, Z., Ricci, E., & Sebe, N. (2021a). Neighborhood contrastive learning for novel class discovery. In IEEE conference on computer vision and pattern recognition (pp. 10867–10875).
Zurück zum Zitat Zhong, Z., Zhu, L., Luo, Z., Li, S., Yang, Y., & Sebe, N. (2021b). Openmix: Reviving known knowledge for discovering novel visual categories in an open world. In IEEE conference on computer vision and pattern recognition (pp. 9462–9470). Zhong, Z., Zhu, L., Luo, Z., Li, S., Yang, Y., & Sebe, N. (2021b). Openmix: Reviving known knowledge for discovering novel visual categories in an open world. In IEEE conference on computer vision and pattern recognition (pp. 9462–9470).
Metadaten
Titel
Does Confusion Really Hurt Novel Class Discovery?
verfasst von
Haoang Chi
Wenjing Yang
Feng Liu
Long Lan
Tao Qin
Bo Han
Publikationsdatum
07.03.2024
Verlag
Springer US
Erschienen in
International Journal of Computer Vision
Print ISSN: 0920-5691
Elektronische ISSN: 1573-1405
DOI
https://doi.org/10.1007/s11263-024-02012-y

Premium Partner