Skip to main content
Erschienen in: Soft Computing 18/2023

09.02.2023 | Focus

Label contrastive learning for image classification

verfasst von: Han Yang, Jun Li

Erschienen in: Soft Computing | Ausgabe 18/2023

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Image classification is one of the most important research tasks in computer vision. Current image classification methods with supervised learning have achieved good classification accuracy. However, supervised image classification methods mainly focus on the semantic differences at the class level, while lacking attention to the instance level. The core idea of contrastive learning is to compare positive and negative samples in the feature space to learn the feature representation, and the focus on instance-level information can make up for the lack of supervised learning. To this end, in this paper, we combine supervised learning and contrastive learning to propose labeled contrastive learning (LCL). Here, the supervised learning component ensures the distinguishability of different classes, the contrastive learning component enhances the compactness within classes and the separability between classes. In the contrastive learning component, instances with the same label are set as positive samples and instances with different labels are set as negative samples, which avoids the problem of false negative samples (positive samples are mislabeled as negative samples). Also, we applied a dynamic label memory bank and a momentum updated encoder. The experimental results show that LCL can further improve the accuracy of image classification compared with some supervised learning method.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Literatur
Zurück zum Zitat An F (2020) Image classification algorithm based on stacked sparse coding deep learning model-optimized kernel function nonnegative sparse representation. Soft Comput 24(22):16967–16981CrossRef An F (2020) Image classification algorithm based on stacked sparse coding deep learning model-optimized kernel function nonnegative sparse representation. Soft Comput 24(22):16967–16981CrossRef
Zurück zum Zitat Caron M, Misra I, Mairal J et al (2020) Unsupervised learning of visual features by contrasting cluster assignments. Adv Neural Inform Proces Syst 33:9912–9924 Caron M, Misra I, Mairal J et al (2020) Unsupervised learning of visual features by contrasting cluster assignments. Adv Neural Inform Proces Syst 33:9912–9924
Zurück zum Zitat Chen T, Kornblith S, Norouzi M, et al (2020a) A simple framework for contrastive learning of visual representations. In: International conference on machine learning, PMLR, pp 1597–1607 Chen T, Kornblith S, Norouzi M, et al (2020a) A simple framework for contrastive learning of visual representations. In: International conference on machine learning, PMLR, pp 1597–1607
Zurück zum Zitat Chen T, Kornblith S, Swersky K et al (2020) Big self-supervised models are strong semi-supervised learners. Adv Neural Inform Proces Syst 33:22243–22255 Chen T, Kornblith S, Swersky K et al (2020) Big self-supervised models are strong semi-supervised learners. Adv Neural Inform Proces Syst 33:22243–22255
Zurück zum Zitat Chen X, He K (2021) Exploring simple siamese representation learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 15,750–15,758 Chen X, He K (2021) Exploring simple siamese representation learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 15,750–15,758
Zurück zum Zitat Chen X, Fan H, Girshick R, et al (2020c) Improved baselines with momentum contrastive learning. arXiv preprint arXiv:2003.04297 Chen X, Fan H, Girshick R, et al (2020c) Improved baselines with momentum contrastive learning. arXiv preprint arXiv:​2003.​04297
Zurück zum Zitat Chen X, Xie S, He K (2021) An empirical study of training self-supervised visual transformers. arXiv e-prints pp arXiv–2104 Chen X, Xie S, He K (2021) An empirical study of training self-supervised visual transformers. arXiv e-prints pp arXiv–2104
Zurück zum Zitat Ding X, Zhang X, Ma N, et al (2021) Repvgg: Making vgg-style convnets great again. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 13,733–13,742 Ding X, Zhang X, Ma N, et al (2021) Repvgg: Making vgg-style convnets great again. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 13,733–13,742
Zurück zum Zitat Elsayed G, Krishnan D, Mobahi H, et al (2018) Large margin deep networks for classification. Advances in neural information processing systems 31 Elsayed G, Krishnan D, Mobahi H, et al (2018) Large margin deep networks for classification. Advances in neural information processing systems 31
Zurück zum Zitat Goodfellow I, PougeAbadie J, Mirza M, et al (2014) Generative adversarial nets. Advances in neural information processing systems 27 Goodfellow I, PougeAbadie J, Mirza M, et al (2014) Generative adversarial nets. Advances in neural information processing systems 27
Zurück zum Zitat Elsayed G, Krishnan D, Mobahi H, et al (2018) Large margin deep networks for classification. Adv Neural Inform Proces Syst 31 Elsayed G, Krishnan D, Mobahi H, et al (2018) Large margin deep networks for classification. Adv Neural Inform Proces Syst 31
Zurück zum Zitat He K, Zhang X, Ren S, et al (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770–778 He K, Zhang X, Ren S, et al (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770–778
Zurück zum Zitat He K, Fan H, Wu Y, et al (2020) Momentum contrast for unsupervised visual representation learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 9729–9738 He K, Fan H, Wu Y, et al (2020) Momentum contrast for unsupervised visual representation learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 9729–9738
Zurück zum Zitat Huang G, Liu Z, Van Der Maaten L, et al (2017) Densely connected convolutional networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4700–4708 Huang G, Liu Z, Van Der Maaten L, et al (2017) Densely connected convolutional networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4700–4708
Zurück zum Zitat Iandola FN, Han S, Moskewicz MW, et al (2016) Squeezenet: alexnet-level accuracy with 50x fewer parameters and 0.5 mb model size. arXiv preprint arXiv:1602.07360 Iandola FN, Han S, Moskewicz MW, et al (2016) Squeezenet: alexnet-level accuracy with 50x fewer parameters and 0.5 mb model size. arXiv preprint arXiv:​1602.​07360
Zurück zum Zitat Krizhevsky A, Sutskever I, Hinton GE (2012) Imagenet classification with deep convolutional neural networks. Adv Neural Inform Proces Syst 25 Krizhevsky A, Sutskever I, Hinton GE (2012) Imagenet classification with deep convolutional neural networks. Adv Neural Inform Proces Syst 25
Zurück zum Zitat LeCun Y, Bottou L, Bengio Y et al (1998) Gradient-based learning applied to document recognition. Proc IEEE 86(11):2278–2324CrossRef LeCun Y, Bottou L, Bengio Y et al (1998) Gradient-based learning applied to document recognition. Proc IEEE 86(11):2278–2324CrossRef
Zurück zum Zitat Van den Oord A, Li Y, Vinyals O (2018) Representation learning with contrastive predictive coding. arXiv e-prints pp arXiv–1807 Van den Oord A, Li Y, Vinyals O (2018) Representation learning with contrastive predictive coding. arXiv e-prints pp arXiv–1807
Zurück zum Zitat Qin Z, Zhang Z, Chen X, et al (2018) Fd-mobilenet: Improved mobilenet with a fast downsampling strategy. In: 2018 25th IEEE International Conference on Image Processing (ICIP), IEEE, pp 1363–1367 Qin Z, Zhang Z, Chen X, et al (2018) Fd-mobilenet: Improved mobilenet with a fast downsampling strategy. In: 2018 25th IEEE International Conference on Image Processing (ICIP), IEEE, pp 1363–1367
Zurück zum Zitat Radosavovic I, Kosaraju RP, Girshick R, et al (2020) Designing network design spaces. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 10428–10436 Radosavovic I, Kosaraju RP, Girshick R, et al (2020) Designing network design spaces. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 10428–10436
Zurück zum Zitat Ren Y, Bai J, Zhang J (2021) Label contrastive coding based graph neural network for graph classification. database systems for advanced applications Ren Y, Bai J, Zhang J (2021) Label contrastive coding based graph neural network for graph classification. database systems for advanced applications
Zurück zum Zitat Rezende DJ, Mohamed S, Wierstra D (2014) Stochastic backpropagation and variational inference in deep latent gaussian models. In: International conference on machine learning, Citeseer, p 2 Rezende DJ, Mohamed S, Wierstra D (2014) Stochastic backpropagation and variational inference in deep latent gaussian models. In: International conference on machine learning, Citeseer, p 2
Zurück zum Zitat Shah A, Sra S, Chellappa R, et al (2022) Max-margin contrastive learning. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp 8220–8230 Shah A, Sra S, Chellappa R, et al (2022) Max-margin contrastive learning. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp 8220–8230
Zurück zum Zitat Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:​1409.​1556
Zurück zum Zitat Szegedy C, Liu W, Jia Y, et al (2015) Going deeper with convolutions. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1–9 Szegedy C, Liu W, Jia Y, et al (2015) Going deeper with convolutions. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1–9
Zurück zum Zitat Tan M, Le Q (2019) Efficientnet: Rethinking model scaling for convolutional neural networks. In: International conference on machine learning, PMLR, pp 6105–6114 Tan M, Le Q (2019) Efficientnet: Rethinking model scaling for convolutional neural networks. In: International conference on machine learning, PMLR, pp 6105–6114
Zurück zum Zitat Thirumaladevi S, Veera Swamy K, Sailaja M (2022) Improved transfer learning of cnn through fine-tuning and classifier ensemble for scene classification. Soft Comput 26(12):5617–5636CrossRef Thirumaladevi S, Veera Swamy K, Sailaja M (2022) Improved transfer learning of cnn through fine-tuning and classifier ensemble for scene classification. Soft Comput 26(12):5617–5636CrossRef
Zurück zum Zitat Touvron H, Vedaldi A, Douze M, et al (2019) Fixing the train-test resolution discrepancy. Adv Neural Inform Proces Syst 32 Touvron H, Vedaldi A, Douze M, et al (2019) Fixing the train-test resolution discrepancy. Adv Neural Inform Proces Syst 32
Zurück zum Zitat Vincent P, Larochelle H, Bengio Y, et al (2008) Extracting and composing robust features with denoising autoencoders. In: Proceedings of the 25th international conference on Machine learning, pp 1096–1103 Vincent P, Larochelle H, Bengio Y, et al (2008) Extracting and composing robust features with denoising autoencoders. In: Proceedings of the 25th international conference on Machine learning, pp 1096–1103
Zurück zum Zitat Wu Z, Xiong Y, Yu SX, et al (2018) Unsupervised feature learning via non-parametric instance discrimination. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3733–3742 Wu Z, Xiong Y, Yu SX, et al (2018) Unsupervised feature learning via non-parametric instance discrimination. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3733–3742
Zurück zum Zitat Yang C, An Z, Cai L, et al (2022) Mutual contrastive learning for visual representation learning. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp 3045–3053 Yang C, An Z, Cai L, et al (2022) Mutual contrastive learning for visual representation learning. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp 3045–3053
Zurück zum Zitat Zbontar J, Jing L, Misra I, et al (2021) Barlow twins: Self-supervised learning via redundancy reduction. In: International Conference on Machine Learning, PMLR, pp 12310–12320 Zbontar J, Jing L, Misra I, et al (2021) Barlow twins: Self-supervised learning via redundancy reduction. In: International Conference on Machine Learning, PMLR, pp 12310–12320
Zurück zum Zitat Zhang X, Zhou X, Lin M, et al (2018) Shufflenet: An extremely efficient convolutional neural network for mobile devices. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 6848–6856 Zhang X, Zhou X, Lin M, et al (2018) Shufflenet: An extremely efficient convolutional neural network for mobile devices. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 6848–6856
Zurück zum Zitat Zheng L, Xiong J, Zhu Y, et al (2022) Contrastive learning with complex heterogeneity. In: Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp 2594–2604 Zheng L, Xiong J, Zhu Y, et al (2022) Contrastive learning with complex heterogeneity. In: Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp 2594–2604
Metadaten
Titel
Label contrastive learning for image classification
verfasst von
Han Yang
Jun Li
Publikationsdatum
09.02.2023
Verlag
Springer Berlin Heidelberg
Erschienen in
Soft Computing / Ausgabe 18/2023
Print ISSN: 1432-7643
Elektronische ISSN: 1433-7479
DOI
https://doi.org/10.1007/s00500-022-07808-z

Weitere Artikel der Ausgabe 18/2023

Soft Computing 18/2023 Zur Ausgabe

Premium Partner