Skip to main content
Erschienen in: International Journal of Computer Vision 4/2024

08.11.2023

Harmonizing Base and Novel Classes: A Class-Contrastive Approach for Generalized Few-Shot Segmentation

verfasst von: Weide Liu, Zhonghua Wu, Yang Zhao, Yuming Fang, Chuan-Sheng Foo, Jun Cheng, Guosheng Lin

Erschienen in: International Journal of Computer Vision | Ausgabe 4/2024

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Current methods for few-shot segmentation (FSSeg) have mainly focused on improving the performance of novel classes while neglecting the performance of base classes. To overcome this limitation, the task of generalized few-shot semantic segmentation (GFSSeg) has been introduced, aiming to predict segmentation masks for both base and novel classes. However, the current prototype-based methods do not explicitly consider the relationship between base and novel classes when updating prototypes, leading to a limited performance in identifying true categories. To address this challenge, we propose a class contrastive loss and a class relationship loss to regulate prototype updates and encourage a large distance between prototypes from different classes, thus distinguishing the classes from each other while maintaining the performance of the base classes. Our proposed approach achieves new state-of-the-art performance for the generalized few-shot segmentation task on PASCAL VOC and MS COCO datasets.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Literatur
Zurück zum Zitat Boudiaf, M., Kervadec, H., Masud, Z. I., et al. (2021). Few-shot segmentation without meta-learning: A good transductive inference is all you need?. In CVPR, pp. 13979–13988. Boudiaf, M., Kervadec, H., Masud, Z. I., et al. (2021). Few-shot segmentation without meta-learning: A good transductive inference is all you need?. In CVPR, pp. 13979–13988.
Zurück zum Zitat Chen, J., Gao, B. B., Lu, Z., et al. (2021). Apanet: Adaptive prototypes alignment network for few-shot semantic segmentation. arXiv preprint arXiv:2111.12263 Chen, J., Gao, B. B., Lu, Z., et al. (2021). Apanet: Adaptive prototypes alignment network for few-shot semantic segmentation. arXiv preprint arXiv:​2111.​12263
Zurück zum Zitat Chen, L. C., Yang, Y., Wang, J., et al. (2016). Attention to scale: Scale-aware semantic image segmentation. In CVPR, pp. 3640–3649. Chen, L. C., Yang, Y., Wang, J., et al. (2016). Attention to scale: Scale-aware semantic image segmentation. In CVPR, pp. 3640–3649.
Zurück zum Zitat Chen, Z., Fu, Y., Chen, K., et al. (2019a). Image block augmentation for one-shot learning. In AAAI, pp. 3379–3386. Chen, Z., Fu, Y., Chen, K., et al. (2019a). Image block augmentation for one-shot learning. In AAAI, pp. 3379–3386.
Zurück zum Zitat Chen, Z., Fu, Y., Wang, Y. X., et al. (2019b). Image deformation meta-networks for one-shot learning. In CVPR, pp. 8680–8689. Chen, Z., Fu, Y., Wang, Y. X., et al. (2019b). Image deformation meta-networks for one-shot learning. In CVPR, pp. 8680–8689.
Zurück zum Zitat Cheng, G., Lang, C., & Han, J. (2022). Holistic prototype activation for few-shot segmentation. TPAMI, 45(4), 4650–4666. Cheng, G., Lang, C., & Han, J. (2022). Holistic prototype activation for few-shot segmentation. TPAMI, 45(4), 4650–4666.
Zurück zum Zitat Dong, K., Yang, W., Xu, Z., et al. (2021). Abpnet: Adaptive background modeling for generalized few shot segmentation. In ACM MM, pp. 2271–2280. Dong, K., Yang, W., Xu, Z., et al. (2021). Abpnet: Adaptive background modeling for generalized few shot segmentation. In ACM MM, pp. 2271–2280.
Zurück zum Zitat Everingham, M., Van Gool, L., Williams, C. K., et al. (2010). The pascal visual object classes (voc) challenge. IJCV, 88(2), 303–338.CrossRef Everingham, M., Van Gool, L., Williams, C. K., et al. (2010). The pascal visual object classes (voc) challenge. IJCV, 88(2), 303–338.CrossRef
Zurück zum Zitat Finn, C., Abbeel, P., Levine, S. (2017). Model-agnostic meta-learning for fast adaptation of deep networks. In ICML, PMLR, pp. 1126–1135. Finn, C., Abbeel, P., Levine, S. (2017). Model-agnostic meta-learning for fast adaptation of deep networks. In ICML, PMLR, pp. 1126–1135.
Zurück zum Zitat Gairola, S., Hemani, M,. Chopra, A., et al. (2020). Simpropnet: Improved similarity propagation for few-shot image segmentation. In IJCAI. Gairola, S., Hemani, M,. Chopra, A., et al. (2020). Simpropnet: Improved similarity propagation for few-shot image segmentation. In IJCAI.
Zurück zum Zitat Hou, R., Chang, H., Ma, B., et al. (2019). Cross attention network for few-shot classification. In NeurIPS. Hou, R., Chang, H., Ma, B., et al. (2019). Cross attention network for few-shot classification. In NeurIPS.
Zurück zum Zitat Iqbal, E., Safarov, S., Bang, S. (2022). Msanet: Multi-similarity and attention guidance for boosting few-shot segmentation. arXiv preprint arXiv:2206.09667 Iqbal, E., Safarov, S., Bang, S. (2022). Msanet: Multi-similarity and attention guidance for boosting few-shot segmentation. arXiv preprint arXiv:​2206.​09667
Zurück zum Zitat Jamal, M. A., Qi, G. J. (2019). Task agnostic meta-learning for few-shot learning. In CVPR, pp. 11719–11727. Jamal, M. A., Qi, G. J. (2019). Task agnostic meta-learning for few-shot learning. In CVPR, pp. 11719–11727.
Zurück zum Zitat Kang, D., Cho, M. (2022). Integrative few-shot learning for classification and segmentation. In CVPR, pp. 9979–9990. Kang, D., Cho, M. (2022). Integrative few-shot learning for classification and segmentation. In CVPR, pp. 9979–9990.
Zurück zum Zitat Kirillov, A., Girshick, R., He, K., et al. (2019). Panoptic feature pyramid networks. In CVPR, pp. 6399–6408. Kirillov, A., Girshick, R., He, K., et al. (2019). Panoptic feature pyramid networks. In CVPR, pp. 6399–6408.
Zurück zum Zitat Koch, G., Zemel, R., Salakhutdinov, R. (2015). Siamese neural networks for one-shot image recognition. In ICMLW. Koch, G., Zemel, R., Salakhutdinov, R. (2015). Siamese neural networks for one-shot image recognition. In ICMLW.
Zurück zum Zitat Lang, C., Cheng, G., Tu, B., et al. (2022a). Learning what not to segment: A new perspective on few-shot segmentation. In CVPR, pp. 8057–8067. Lang, C., Cheng, G., Tu, B., et al. (2022a). Learning what not to segment: A new perspective on few-shot segmentation. In CVPR, pp. 8057–8067.
Zurück zum Zitat Lang, C., Tu, B., Cheng, G., et al. (2022b). Beyond the prototype: Divide-and-conquer proxies for few-shot segmentation. arXiv preprint arXiv:2204.09903 Lang, C., Tu, B., Cheng, G., et al. (2022b). Beyond the prototype: Divide-and-conquer proxies for few-shot segmentation. arXiv preprint arXiv:​2204.​09903
Zurück zum Zitat Lang, C., Cheng, G., Tu, B., et al. (2023). Base and meta: A new perspective on few-shot segmentation. In TPAMI. Lang, C., Cheng, G., Tu, B., et al. (2023). Base and meta: A new perspective on few-shot segmentation. In TPAMI.
Zurück zum Zitat Li, B., Yang, B., Liu, C., et al. (2021a). Beyond max-margin: Class margin equilibrium for few-shot object detection. In CVPR, pp. 7363–7372. Li, B., Yang, B., Liu, C., et al. (2021a). Beyond max-margin: Class margin equilibrium for few-shot object detection. In CVPR, pp. 7363–7372.
Zurück zum Zitat Li, G., Jampani, V., Sevilla-Lara, L., et al. (2021b). Adaptive prototype learning and allocation for few-shot segmentation. In CVPR, pp. 8334–8343. Li, G., Jampani, V., Sevilla-Lara, L., et al. (2021b). Adaptive prototype learning and allocation for few-shot segmentation. In CVPR, pp. 8334–8343.
Zurück zum Zitat Li, H., Eigen, D., Dodge, S., et al. (2019a). Finding task-relevant features for few-shot learning by category traversal. In CVPR, pp. 1–10. Li, H., Eigen, D., Dodge, S., et al. (2019a). Finding task-relevant features for few-shot learning by category traversal. In CVPR, pp. 1–10.
Zurück zum Zitat Li, W., Wang, L., Xu, J., et al. (2019b). Revisiting local descriptor based image-to-class measure for few-shot learning. In CVPR, pp. 7260–7268. Li, W., Wang, L., Xu, J., et al. (2019b). Revisiting local descriptor based image-to-class measure for few-shot learning. In CVPR, pp. 7260–7268.
Zurück zum Zitat Li, X., Wei, T., Chen, Y. P., et al. (2020). Fss-1000: A 1000-class dataset for few-shot segmentation. In CVPR, pp. 2869–2878. Li, X., Wei, T., Chen, Y. P., et al. (2020). Fss-1000: A 1000-class dataset for few-shot segmentation. In CVPR, pp. 2869–2878.
Zurück zum Zitat Lin, T.Y., Maire, M., Belongie, S., et al. (2014). Microsoft coco: Common objects in context. In Proceedings of the European Conference on Computer Vision, pp. 740–755. Lin, T.Y., Maire, M., Belongie, S., et al. (2014). Microsoft coco: Common objects in context. In Proceedings of the European Conference on Computer Vision, pp. 740–755.
Zurück zum Zitat Liu, B., Ding, Y., Jiao, J., et al. (2021a). Anti-aliasing semantic reconstruction for few-shot semantic segmentation. In CVPR, pp. 9747–9756. Liu, B., Ding, Y., Jiao, J., et al. (2021a). Anti-aliasing semantic reconstruction for few-shot semantic segmentation. In CVPR, pp. 9747–9756.
Zurück zum Zitat Liu, C., Fu, Y., Xu, C., et al. (2021b). Learning a few-shot embedding model with contrastive learning. In AAAI, pp. 8635–8643. Liu, C., Fu, Y., Xu, C., et al. (2021b). Learning a few-shot embedding model with contrastive learning. In AAAI, pp. 8635–8643.
Zurück zum Zitat Liu, L., Cao, J., Liu, M., et al. (2020a). Dynamic extension nets for few-shot semantic segmentation. In ACM MM, pp. 1441–1449. Liu, L., Cao, J., Liu, M., et al. (2020a). Dynamic extension nets for few-shot semantic segmentation. In ACM MM, pp. 1441–1449.
Zurück zum Zitat Liu, W., Zhang, C., Lin, G., et al. (2020b). Crnet: Cross-reference networks for few-shot segmentation. In CVPR. Liu, W., Zhang, C., Lin, G., et al. (2020b). Crnet: Cross-reference networks for few-shot segmentation. In CVPR.
Zurück zum Zitat Liu, W., Kong, X., Hung, T. Y., et al. (2021c). Cross-image region mining with region prototypical network for weakly supervised segmentation. IEEE Transactions on Multimedia. Liu, W., Kong, X., Hung, T. Y., et al. (2021c). Cross-image region mining with region prototypical network for weakly supervised segmentation. IEEE Transactions on Multimedia.
Zurück zum Zitat Liu, W., Zhang, C., Ding, H., et al. (2021d). Few-shot segmentation with optimal transport matching and message flow. IEEE Transactions on Multimedia. Liu, W., Zhang, C., Ding, H., et al. (2021d). Few-shot segmentation with optimal transport matching and message flow. IEEE Transactions on Multimedia.
Zurück zum Zitat Liu, W., Zhang, C., Lin, G., et al. (2022). Crcnet: Few-shot segmentation with cross-reference and region-global conditional networks. International Journal of Computer Vision, 130(12), 3140–3157.CrossRef Liu, W., Zhang, C., Lin, G., et al. (2022). Crcnet: Few-shot segmentation with cross-reference and region-global conditional networks. International Journal of Computer Vision, 130(12), 3140–3157.CrossRef
Zurück zum Zitat Liu, Y., Zhang, X., Zhang, S., et al. (2020c). Part-aware prototype network for few-shot semantic segmentation. In ECCV, pp. 142–158. Liu, Y., Zhang, X., Zhang, S., et al. (2020c). Part-aware prototype network for few-shot semantic segmentation. In ECCV, pp. 142–158.
Zurück zum Zitat Long, J., Shelhamer, E., Darrell, T. (2015). Fully convolutional networks for semantic segmentation. In CVPR, pp. 3431–3440. Long, J., Shelhamer, E., Darrell, T. (2015). Fully convolutional networks for semantic segmentation. In CVPR, pp. 3431–3440.
Zurück zum Zitat Lu, Z., He, S., Zhu, X., et al. (2021). Simpler is better: Few-shot semantic segmentation with classifier weight transformer. In ICCV, pp. 8741–8750. Lu, Z., He, S., Zhu, X., et al. (2021). Simpler is better: Few-shot semantic segmentation with classifier weight transformer. In ICCV, pp. 8741–8750.
Zurück zum Zitat Min, J., Kang, D., Cho, M. (2021). Hypercorrelation squeeze for few-shot segmentation. In Proceedings of the IEEE/CVF international conference on computer vision, pp. 6941–6952. Min, J., Kang, D., Cho, M. (2021). Hypercorrelation squeeze for few-shot segmentation. In Proceedings of the IEEE/CVF international conference on computer vision, pp. 6941–6952.
Zurück zum Zitat Ouyang, C., Biffi, C., Chen, C,. et al. (2020). Self-supervision with superpixels: Training few-shot medical image segmentation without annotation. In ECCV, pp. 762–780. Ouyang, C., Biffi, C., Chen, C,. et al. (2020). Self-supervision with superpixels: Training few-shot medical image segmentation without annotation. In ECCV, pp. 762–780.
Zurück zum Zitat Ravi, S., Larochelle, H. (2016). Optimization as a model for few-shot learning. In ICLR. Ravi, S., Larochelle, H. (2016). Optimization as a model for few-shot learning. In ICLR.
Zurück zum Zitat Siam, M., Doraiswamy, N., Oreshkin, B. N., et al. (2020). Weakly supervised few-shot object segmentation using co-attention with visual and semantic embeddings. In IJCAI. Siam, M., Doraiswamy, N., Oreshkin, B. N., et al. (2020). Weakly supervised few-shot object segmentation using co-attention with visual and semantic embeddings. In IJCAI.
Zurück zum Zitat Tian, Z., Zhao, H., Shu, M., et al. (2020). Prior guided feature enrichment network for few-shot segmentation. TPAMI, 44(2), 1050–1065.CrossRef Tian, Z., Zhao, H., Shu, M., et al. (2020). Prior guided feature enrichment network for few-shot segmentation. TPAMI, 44(2), 1050–1065.CrossRef
Zurück zum Zitat Tian, Z., Lai, X., Jiang, L., et al. (2022). Generalized few-shot semantic segmentation. In CVPR, pp. 11563–11572. Tian, Z., Lai, X., Jiang, L., et al. (2022). Generalized few-shot semantic segmentation. In CVPR, pp. 11563–11572.
Zurück zum Zitat Wang, H., Zhang, X., Hu, Y., et al. (2020). Few-shot semantic segmentation with democratic attention networks. In ECCV, pp. 730–746. Wang, H., Zhang, X., Hu, Y., et al. (2020). Few-shot semantic segmentation with democratic attention networks. In ECCV, pp. 730–746.
Zurück zum Zitat Wang, K., Liew, J., Zou, Y., et al. (2019). Panet: Few-shot image semantic segmentation with prototype alignment. In ICCV, pp. 9197–9206. Wang, K., Liew, J., Zou, Y., et al. (2019). Panet: Few-shot image semantic segmentation with prototype alignment. In ICCV, pp. 9197–9206.
Zurück zum Zitat Wang, W., Zhou, T., Yu, F., et al. (2021). Exploring cross-image pixel contrast for semantic segmentation. In CVPR, pp. 7303–7313. Wang, W., Zhou, T., Yu, F., et al. (2021). Exploring cross-image pixel contrast for semantic segmentation. In CVPR, pp. 7303–7313.
Zurück zum Zitat Wu, Z., Shi, X., Lin, G., et al. (2021). Learning meta-class memory for few-shot semantic segmentation. In ICCV, pp. 517–526. Wu, Z., Shi, X., Lin, G., et al. (2021). Learning meta-class memory for few-shot semantic segmentation. In ICCV, pp. 517–526.
Zurück zum Zitat Xie, G. S., Liu, J., Xiong, H., et al. (2021a). Scale-aware graph neural network for few-shot semantic segmentation. In CVPR, pp. 5475–5484. Xie, G. S., Liu, J., Xiong, H., et al. (2021a). Scale-aware graph neural network for few-shot semantic segmentation. In CVPR, pp. 5475–5484.
Zurück zum Zitat Xie, G. S., Xiong, H., Liu, J., et al. (2021b). Few-shot semantic segmentation with cyclic memory network. In ICCV, pp. 7293–7302. Xie, G. S., Xiong, H., Liu, J., et al. (2021b). Few-shot semantic segmentation with cyclic memory network. In ICCV, pp. 7293–7302.
Zurück zum Zitat Yang, L., Zhuo, W., Qi, L., et al. (2021). Mining latent classes for few-shot segmentation. In ICCV, pp. 8721–8730. Yang, L., Zhuo, W., Qi, L., et al. (2021). Mining latent classes for few-shot segmentation. In ICCV, pp. 8721–8730.
Zurück zum Zitat Yang, X., Wang, B., Chen, K., et al. (2020). Brinet: Towards bridging the intra-class and inter-class gaps in one-shot segmentation. arXiv preprint arXiv:2008.06226 Yang, X., Wang, B., Chen, K., et al. (2020). Brinet: Towards bridging the intra-class and inter-class gaps in one-shot segmentation. arXiv preprint arXiv:​2008.​06226
Zurück zum Zitat Zhang, B., Xiao, J., Qin, T. (2021). Self-guided and cross-guided learning for few-shot segmentation. In CVPR, pp. 8312–8321. Zhang, B., Xiao, J., Qin, T. (2021). Self-guided and cross-guided learning for few-shot segmentation. In CVPR, pp. 8312–8321.
Zurück zum Zitat Zhao, H., Shi, J., Qi, X., et al. (2017). Pyramid scene parsing network. In CVPR, pp. 2881–2890. Zhao, H., Shi, J., Qi, X., et al. (2017). Pyramid scene parsing network. In CVPR, pp. 2881–2890.
Zurück zum Zitat Zhao, X., Vemulapalli, R., Mansfield, P. A., et al. (2021). Contrastive learning for label efficient semantic segmentation. In CVPR, pp. 10623–10633. Zhao, X., Vemulapalli, R., Mansfield, P. A., et al. (2021). Contrastive learning for label efficient semantic segmentation. In CVPR, pp. 10623–10633.
Zurück zum Zitat Zhong, Y., Yuan, B., Wu, H., et al. (2021). Pixel contrastive-consistent semi-supervised semantic segmentation. In CVPR, pp. 7273–7282. Zhong, Y., Yuan, B., Wu, H., et al. (2021). Pixel contrastive-consistent semi-supervised semantic segmentation. In CVPR, pp. 7273–7282.
Zurück zum Zitat Zhu, K., Zhai, W., Zha, Z. J., et al. (2020). Self-supervised tuning for few-shot segmentation. arXiv preprint arXiv:2004.05538 Zhu, K., Zhai, W., Zha, Z. J., et al. (2020). Self-supervised tuning for few-shot segmentation. arXiv preprint arXiv:​2004.​05538
Zurück zum Zitat Zhuge, Y., Shen, C. (2021). Deep reasoning network for few-shot semantic segmentation. In ACM MM, pp. 5344–5352. Zhuge, Y., Shen, C. (2021). Deep reasoning network for few-shot semantic segmentation. In ACM MM, pp. 5344–5352.
Metadaten
Titel
Harmonizing Base and Novel Classes: A Class-Contrastive Approach for Generalized Few-Shot Segmentation
verfasst von
Weide Liu
Zhonghua Wu
Yang Zhao
Yuming Fang
Chuan-Sheng Foo
Jun Cheng
Guosheng Lin
Publikationsdatum
08.11.2023
Verlag
Springer US
Erschienen in
International Journal of Computer Vision / Ausgabe 4/2024
Print ISSN: 0920-5691
Elektronische ISSN: 1573-1405
DOI
https://doi.org/10.1007/s11263-023-01939-y

Weitere Artikel der Ausgabe 4/2024

International Journal of Computer Vision 4/2024 Zur Ausgabe

Premium Partner