Skip to main content
Erschienen in: International Journal of Computer Vision 11/2023

09.07.2023

Attribute-Image Person Re-identification via Modal-Consistent Metric Learning

verfasst von: Jianqing Zhu, Liu Liu, Yibing Zhan, Xiaobin Zhu, Huanqiang Zeng, Dacheng Tao

Erschienen in: International Journal of Computer Vision | Ausgabe 11/2023

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Attribute-image person re-identification (AIPR) is a cross-modal retrieval task that searches person images who meet a list of attributes. Due to large modal gaps between attributes and images, current AIPR methods generally depend on cross-modal feature alignment, but they do not pay enough attention to similarity metric jitters among varying modal configurations (i.e., attribute probe vs. image gallery, image probe vs. attribute gallery, image probe vs. image gallery, and attribute probe vs. attribute gallery). In this paper, we propose a modal-consistent metric learning (MCML) method that stably measures comprehensive similarities between attributes and images. Our MCML is with favorable properties that differ in two significant ways from previous methods. First, MCML provides a complete multi-modal triplet (CMMT) loss function that pulls the distance between the farthest positive pair as close as possible while pushing the distance between the nearest negative pair as far as possible, independent of their modalities. Second, MCML develops a modal-consistent matching regularization (MCMR) to reduce the diversity of matching matrices and guide consistent matching behaviors on varying modal configurations. Therefore, our MCML integrates the CMMT loss function and MCMR, requiring no complex cross-modal feature alignments. Theoretically, we offer the generalization bound to establish the stability of our MCML model by applying on-average stability. Experimentally, extensive results on PETA and Market-1501 datasets show that the proposed MCML is superior to the state-of-the-art approaches.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Fußnoten
1
We use \(i \in [ m ]\) to denote that i is generated from \([ m ] = \{ {1,2,...,m} \}\). The same definition is also applied to \(l_i\!\in \! [c].\)
 
2
The single-modal HMT loss function means only images are applied to the HMT loss function, while the cross-modal HMT loss function means both images and attributes are applied to the HMT loss function.
 
Literatur
Zurück zum Zitat Andrew, G., Arora, R., Bilmes, J., & Livescu, K. (2013). Deep canonical correlation analysis. In ICML (pp. 1247–1255). Andrew, G., Arora, R., Bilmes, J., & Livescu, K. (2013). Deep canonical correlation analysis. In ICML (pp. 1247–1255).
Zurück zum Zitat Bousquet, O., Klochkov, Y., & Zhivotovskiy, N. (2020). Sharper bounds for uniformly stable algorithms. In PMLR conference on learning theory (pp. 610–626). Bousquet, O., Klochkov, Y., & Zhivotovskiy, N. (2020). Sharper bounds for uniformly stable algorithms. In PMLR conference on learning theory (pp. 610–626).
Zurück zum Zitat Cao, Y. T., Wang, J., & Tao, D. (2020). Symbiotic adversarial learning for attribute-based person search. In ECCV. Cao, Y. T., Wang, J., & Tao, D. (2020). Symbiotic adversarial learning for attribute-based person search. In ECCV.
Zurück zum Zitat Deng, Y., Luo, P., Loy, C. C., & Tang, X. (2014). Pedestrian attribute recognition at far distance. In ACMMM (pp. 789–792). Deng, Y., Luo, P., Loy, C. C., & Tang, X. (2014). Pedestrian attribute recognition at far distance. In ACMMM (pp. 789–792).
Zurück zum Zitat Dong, Q., Gong, S., & Zhu, X. (2019). Person search by text attribute query as zero-shot learning. In CVPR (pp. 3652–3661). Dong, Q., Gong, S., & Zhu, X. (2019). Person search by text attribute query as zero-shot learning. In CVPR (pp. 3652–3661).
Zurück zum Zitat Eisenschtat, A., & Wolf, L. (2017). Linking image and text with 2-way nets. In CVPR (pp. 4601–4611). Eisenschtat, A., & Wolf, L. (2017). Linking image and text with 2-way nets. In CVPR (pp. 4601–4611).
Zurück zum Zitat Feldman, V., & Vondrak, J. (2018). Generalization bounds for uniformly stable algorithms. In NeurIPS (pp. 9770–9780). Feldman, V., & Vondrak, J. (2018). Generalization bounds for uniformly stable algorithms. In NeurIPS (pp. 9770–9780).
Zurück zum Zitat Feldman, V., & Vondrak, J. (2019). High probability generalization bounds for uniformly stable algorithms with nearly optimal rate. In PMLR conference on learning theory (pp. 1270–1279). Feldman, V., & Vondrak, J. (2019). High probability generalization bounds for uniformly stable algorithms with nearly optimal rate. In PMLR conference on learning theory (pp. 1270–1279).
Zurück zum Zitat Felix, R., Kumar, V. B., Reid, I., & Carneiro, G. (2018). Multi-modal cycle-consistent generalized zero-shot learning. In ECCV (pp. 21–37). Felix, R., Kumar, V. B., Reid, I., & Carneiro, G. (2018). Multi-modal cycle-consistent generalized zero-shot learning. In ECCV (pp. 21–37).
Zurück zum Zitat Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., & Bengio, Y. (2014). Generative adversarial nets. In NIPS (pp. 2672–2680). Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., & Bengio, Y. (2014). Generative adversarial nets. In NIPS (pp. 2672–2680).
Zurück zum Zitat He, K., Fan, H., Wu, Y., Xie, S., & Girshick, R. (2020). Momentum contrast for unsupervised visual representation learning. In CVPR (pp. 9729–9738). He, K., Fan, H., Wu, Y., Xie, S., & Girshick, R. (2020). Momentum contrast for unsupervised visual representation learning. In CVPR (pp. 9729–9738).
Zurück zum Zitat He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In CVPR (pp. 770–778). He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In CVPR (pp. 770–778).
Zurück zum Zitat Hubert Tsai, Y. H., Huang, L. K., & Salakhutdinov, R. (2017). Learning robust visual-semantic embeddings. In ICCV (pp. 3571–3580). Hubert Tsai, Y. H., Huang, L. K., & Salakhutdinov, R. (2017). Learning robust visual-semantic embeddings. In ICCV (pp. 3571–3580).
Zurück zum Zitat Iodice, S., & Mikolajczyk, K. (2020). Text attribute aggregation and visual feature decomposition for person search. In BMVC (2020). Iodice, S., & Mikolajczyk, K. (2020). Text attribute aggregation and visual feature decomposition for person search. In BMVC (2020).
Zurück zum Zitat Ioffe, S., & Szegedy, C. (2015). Batch normalization: Accelerating deep network training by reducing internal covariate shift. In ICML (pp. 448–456). Ioffe, S., & Szegedy, C. (2015). Batch normalization: Accelerating deep network training by reducing internal covariate shift. In ICML (pp. 448–456).
Zurück zum Zitat Jeong, B., Park, J., & Kwak, S. (2021). Asmr: Learning attribute-based person search with adaptive semantic margin regularizer. In ICCV (pp. 12016–12025). Jeong, B., Park, J., & Kwak, S. (2021). Asmr: Learning attribute-based person search with adaptive semantic margin regularizer. In ICCV (pp. 12016–12025).
Zurück zum Zitat Ji, Z., He, E., Wang, H., & Yang, A. (2019). Image-attribute reciprocally guided attention network for pedestrian attribute recognition. Pattern Recognition Letters, 120, 89–95.CrossRef Ji, Z., He, E., Wang, H., & Yang, A. (2019). Image-attribute reciprocally guided attention network for pedestrian attribute recognition. Pattern Recognition Letters, 120, 89–95.CrossRef
Zurück zum Zitat Ji, Z., Hu, Z., He, E., Han, J., & Pang, Y. (2020). Pedestrian attribute recognition based on multiple time steps attention. Pattern Recognition Letters, 138, 170–176.CrossRef Ji, Z., Hu, Z., He, E., Han, J., & Pang, Y. (2020). Pedestrian attribute recognition based on multiple time steps attention. Pattern Recognition Letters, 138, 170–176.CrossRef
Zurück zum Zitat Ji, Z., Sun, Y., Yu, Y., Pang, Y., & Han, J. (2019). Attribute-guided network for cross-modal zero-shot hashing. IEEE Transactions on Neural Networks and Learning Systems, 31(1), 321–330.CrossRef Ji, Z., Sun, Y., Yu, Y., Pang, Y., & Han, J. (2019). Attribute-guided network for cross-modal zero-shot hashing. IEEE Transactions on Neural Networks and Learning Systems, 31(1), 321–330.CrossRef
Zurück zum Zitat Layne, R., Hospedales, T.M., & Gong, S. (2012a). Towards person identification and re-identification with attributes. In ECCV (pp. 402–412). Layne, R., Hospedales, T.M., & Gong, S. (2012a). Towards person identification and re-identification with attributes. In ECCV (pp. 402–412).
Zurück zum Zitat Layne, R., Hospedales, T. M., Gong, S., & Mary, Q. (2012b). Person re-identification by attributes. In BMVC (p. 8). Layne, R., Hospedales, T. M., Gong, S., & Mary, Q. (2012b). Person re-identification by attributes. In BMVC (p. 8).
Zurück zum Zitat Lei, Y., Ledent, A., & Kloft, M. (2020). Sharper generalization bounds for pairwise learning. NeurIPS 33. Lei, Y., Ledent, A., & Kloft, M. (2020). Sharper generalization bounds for pairwise learning. NeurIPS 33.
Zurück zum Zitat Li, D., Chen, X., & Huang, K. (2015a). Multi-attribute learning for pedestrian attribute recognition in surveillance scenarios. In ACPR (pp. 111–115). Li, D., Chen, X., & Huang, K. (2015a). Multi-attribute learning for pedestrian attribute recognition in surveillance scenarios. In ACPR (pp. 111–115).
Zurück zum Zitat Li, D., Chen, X., & Huang, K. (2015b). Multi-attribute learning for pedestrian attribute recognition in surveillance scenarios. In ACPR (pp. 111–115). IEEE. Li, D., Chen, X., & Huang, K. (2015b). Multi-attribute learning for pedestrian attribute recognition in surveillance scenarios. In ACPR (pp. 111–115). IEEE.
Zurück zum Zitat Li, S., Xiao, T., Li, H., Yang, W., & Wang, X. (2017). Identity-aware textual-visual matching with latent co-attention. In ICCV (pp. 1890–1899). Li, S., Xiao, T., Li, H., Yang, W., & Wang, X. (2017). Identity-aware textual-visual matching with latent co-attention. In ICCV (pp. 1890–1899).
Zurück zum Zitat Li, W., Zhu, X., & Gong, S. (2020). Scalable person re-identification by harmonious attention. International Journal of Computer Vision, 128(6), 1635–1653.CrossRef Li, W., Zhu, X., & Gong, S. (2020). Scalable person re-identification by harmonious attention. International Journal of Computer Vision, 128(6), 1635–1653.CrossRef
Zurück zum Zitat Li, Z., Min, W., Song, J., Zhu, Y., Kang, L., Wei, X., Wei, X., & Jiang, S. (2022). Rethinking the optimization of average precision: Only penalizing negative instances before positive ones is enough. In AAAI (Vol. 36, pp. 1518–1526). Li, Z., Min, W., Song, J., Zhu, Y., Kang, L., Wei, X., Wei, X., & Jiang, S. (2022). Rethinking the optimization of average precision: Only penalizing negative instances before positive ones is enough. In AAAI (Vol. 36, pp. 1518–1526).
Zurück zum Zitat Lin, X., Ren, P., Xiao, Y., Chang, X., & Hauptmann, A. (2021). Person search challenges and solutions: A survey. Lin, X., Ren, P., Xiao, Y., Chang, X., & Hauptmann, A. (2021). Person search challenges and solutions: A survey.
Zurück zum Zitat Lin, Y., Zheng, L., Zheng, Z., Wu, Y., Hu, Z., Yan, C., & Yang, Y. (2019). Improving person re-identification by attribute and identity learning. Pattern Recognition, 95, 151–161.CrossRef Lin, Y., Zheng, L., Zheng, Z., Wu, Y., Hu, Z., Yan, C., & Yang, Y. (2019). Improving person re-identification by attribute and identity learning. Pattern Recognition, 95, 151–161.CrossRef
Zurück zum Zitat Liu, L., Zhang, H., Xu, X., Zhang, Z., & Yan, S. (2019). Collocating clothes with generative adversarial networks cosupervised by categories and attributes: A multidiscriminator framework. IEEE Transactions on Neural Networks and Learning Systems, 31(9), 3540–3554.MathSciNetCrossRef Liu, L., Zhang, H., Xu, X., Zhang, Z., & Yan, S. (2019). Collocating clothes with generative adversarial networks cosupervised by categories and attributes: A multidiscriminator framework. IEEE Transactions on Neural Networks and Learning Systems, 31(9), 3540–3554.MathSciNetCrossRef
Zurück zum Zitat Liu, P., Liu, X., Yan, J., & Shao, J. (2018). Localization guided learning for pedestrian attribute recognition. In BMVC. Liu, P., Liu, X., Yan, J., & Shao, J. (2018). Localization guided learning for pedestrian attribute recognition. In BMVC.
Zurück zum Zitat Liu, X., Zhao, H., Tian, M., Sheng, L., Shao, J., Yi, S., Yan, J., & Wang, X. (2017). Hydraplus-net: Attentive deep features for pedestrian analysis. In ICCV (pp. 350–359). Liu, X., Zhao, H., Tian, M., Sheng, L., Shao, J., Yi, S., Yan, J., & Wang, X. (2017). Hydraplus-net: Attentive deep features for pedestrian analysis. In ICCV (pp. 350–359).
Zurück zum Zitat Luo, H., Jiang, W., Gu, Y., Liu, F., Liao, X., Lai, S., & Gu, J. (2019). A strong baseline and batch normalization neck for deep person re-identification. IEEE Transactions on Multimedia. Luo, H., Jiang, W., Gu, Y., Liu, F., Liao, X., Lai, S., & Gu, J. (2019). A strong baseline and batch normalization neck for deep person re-identification. IEEE Transactions on Multimedia.
Zurück zum Zitat Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., & Antiga, L., et al. (2019). Pytorch: An imperative style, high-performance deep learning library. In NeurIPS (pp. 8026–8037). Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., & Antiga, L., et al. (2019). Pytorch: An imperative style, high-performance deep learning library. In NeurIPS (pp. 8026–8037).
Zurück zum Zitat Schroff, F., Kalenichenko, D., & Philbin, J. (2015). Facenet: A unified embedding for face recognition and clustering. In CVPR (pp. 815–823). Schroff, F., Kalenichenko, D., & Philbin, J. (2015). Facenet: A unified embedding for face recognition and clustering. In CVPR (pp. 815–823).
Zurück zum Zitat Schumann, A., & Stiefelhagen, R. (2017). Person re-identification by deep learning attribute-complementary information. In CVPR Workshop (pp. 20–28). Schumann, A., & Stiefelhagen, R. (2017). Person re-identification by deep learning attribute-complementary information. In CVPR Workshop (pp. 20–28).
Zurück zum Zitat Su, C., Zhang, S., Xing, J., Gao, W., & Tian, Q. (2016). Deep attributes driven multi-camera person re-identification. In ECCV (pp. 475–491). Su, C., Zhang, S., Xing, J., Gao, W., & Tian, Q. (2016). Deep attributes driven multi-camera person re-identification. In ECCV (pp. 475–491).
Zurück zum Zitat Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V., & Rabinovich, A. (2015). Going deeper with convolutions. In CVPR (pp. 1–9). Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V., & Rabinovich, A. (2015). Going deeper with convolutions. In CVPR (pp. 1–9).
Zurück zum Zitat Tan, Z., Yang, Y., Wan, J., Guo, G., & Li, S. Z. (2020). Relation-aware pedestrian attribute recognition with graph convolutional networks. In AAAI (pp. 12055–12062). Tan, Z., Yang, Y., Wan, J., Guo, G., & Li, S. Z. (2020). Relation-aware pedestrian attribute recognition with graph convolutional networks. In AAAI (pp. 12055–12062).
Zurück zum Zitat Tan, Z., Yang, Y., Wan, J., Hang, H., Guo, G., & Li, S. Z. (2019). Attention-based pedestrian attribute analysis. Transactions on Image Processing, 28(12), 6126–6140.MathSciNetCrossRefMATH Tan, Z., Yang, Y., Wan, J., Hang, H., Guo, G., & Li, S. Z. (2019). Attention-based pedestrian attribute analysis. Transactions on Image Processing, 28(12), 6126–6140.MathSciNetCrossRefMATH
Zurück zum Zitat Vaquero, D. A., Feris, R. S., Tran, D., Brown, L., Hampapur, A., & Turk, M. (2009). Attribute-based people search in surveillance environments. In Workshop on applications of computer vision (pp. 1–8). Vaquero, D. A., Feris, R. S., Tran, D., Brown, L., Hampapur, A., & Turk, M. (2009). Attribute-based people search in surveillance environments. In Workshop on applications of computer vision (pp. 1–8).
Zurück zum Zitat Wang, B., Yang, Y., Xu, X., Hanjalic, A., & Shen, H. (2017). Adversarial cross-modal retrieval. In ACM MM (pp. 154–162). Wang, B., Yang, Y., Xu, X., Hanjalic, A., & Shen, H. (2017). Adversarial cross-modal retrieval. In ACM MM (pp. 154–162).
Zurück zum Zitat Wang, J., Zhu, X., Gong, S., & Li, W. (2018). Transferable joint attribute-identity deep learning for unsupervised person re-identification. In CVPR (pp. 2275–2284). Wang, J., Zhu, X., Gong, S., & Li, W. (2018). Transferable joint attribute-identity deep learning for unsupervised person re-identification. In CVPR (pp. 2275–2284).
Zurück zum Zitat Wang, W., Arora, R., Livescu, K., & Bilmes, J. (2015). On deep multi-view representation learning. In ICML (pp. 1083–1092). Wang, W., Arora, R., Livescu, K., & Bilmes, J. (2015). On deep multi-view representation learning. In ICML (pp. 1083–1092).
Zurück zum Zitat Wang, X., Han, X., Huang, W., Dong, D., & Scott, M. R. (2019). Multi-similarity loss with general pair weighting for deep metric learning. In CVPR (pp. 5022–5030). Wang, X., Han, X., Huang, W., Dong, D., & Scott, M. R. (2019). Multi-similarity loss with general pair weighting for deep metric learning. In CVPR (pp. 5022–5030).
Zurück zum Zitat Wu, M., Huang, D., Guo, Y., & Wang, Y. (2019). Distraction-aware feature learning for human attribute recognition via coarse-to-fine attention mechanism. In AAAI. Wu, M., Huang, D., Guo, Y., & Wang, Y. (2019). Distraction-aware feature learning for human attribute recognition via coarse-to-fine attention mechanism. In AAAI.
Zurück zum Zitat Xu, B., Wang, N., Chen, T., & Li, M. (2015). Empirical evaluation of rectified activations in convolutional network. arXiv preprint arXiv:1505.00853. Xu, B., Wang, N., Chen, T., & Li, M. (2015). Empirical evaluation of rectified activations in convolutional network. arXiv preprint arXiv:​1505.​00853.
Zurück zum Zitat Yang, Y., Tan, Z., Tiwari, P., Pandey, H. M., Wan, J., Lei, Z., Guo, G., & Li, S. Z. (2021). Cascaded split-and-aggregate learning with feature recombination for pedestrian attribute recognition. International Journal of Computer Vision (pp. 1–14). Yang, Y., Tan, Z., Tiwari, P., Pandey, H. M., Wan, J., Lei, Z., Guo, G., & Li, S. Z. (2021). Cascaded split-and-aggregate learning with feature recombination for pedestrian attribute recognition. International Journal of Computer Vision (pp. 1–14).
Zurück zum Zitat Ye, M., Shen, J., Lin, G., Xiang, T., Shao, L., & Hoi, S.C. (2021). Deep learning for person re-identification: A survey and outlook. IEEE Transactions on Pattern Analysis and Machine Intelligence (pp. 1–1). Ye, M., Shen, J., Lin, G., Xiang, T., Shao, L., & Hoi, S.C. (2021). Deep learning for person re-identification: A survey and outlook. IEEE Transactions on Pattern Analysis and Machine Intelligence (pp. 1–1).
Zurück zum Zitat Yin, J., Wu, A., & Zheng, W. S. (2020). Fine-grained person re-identification. International Journal of Computer Vision, 128(6), 1654–1672.CrossRef Yin, J., Wu, A., & Zheng, W. S. (2020). Fine-grained person re-identification. International Journal of Computer Vision, 128(6), 1654–1672.CrossRef
Zurück zum Zitat Yin, Z., Zheng, W. S., Wu, A., Yu, H. X., Wan, H., Guo, X., Huang, F., & Lai, J. (2018). Adversarial attribute-image person re-identification. In IJCAI (pp. 1100–1106). Yin, Z., Zheng, W. S., Wu, A., Yu, H. X., Wan, H., Guo, X., Huang, F., & Lai, J. (2018). Adversarial attribute-image person re-identification. In IJCAI (pp. 1100–1106).
Zurück zum Zitat Yu, K., Leng, B., Zhang, Z., Li, D., & Huang, K. (2017). Weakly-supervised learning of mid-level features for pedestrian attribute recognition and localization. In ECCV. Yu, K., Leng, B., Zhang, Z., Li, D., & Huang, K. (2017). Weakly-supervised learning of mid-level features for pedestrian attribute recognition and localization. In ECCV.
Zurück zum Zitat Zeng, H., Ai, H., Zhuang, Z., & Chen, L. (2020). Multi-task learning via co-attentive sharing for pedestrian attribute recognition. In ICME (pp. 1–6). Zeng, H., Ai, H., Zhuang, Z., & Chen, L. (2020). Multi-task learning via co-attentive sharing for pedestrian attribute recognition. In ICME (pp. 1–6).
Zurück zum Zitat Zhan, Y., Yu, J., Yu, T., & Tao, D. (2019). On exploring undetermined relationships for visual relationship detection. In CVPR (pp. 5128–5137). Zhan, Y., Yu, J., Yu, T., & Tao, D. (2019). On exploring undetermined relationships for visual relationship detection. In CVPR (pp. 5128–5137).
Zurück zum Zitat Zhan, Y., Yu, J., Yu, T., & Tao, D. (2020). Multi-task compositional network for visual relationship detection. International Journal of Computer Vision, 128(8), 2146–2165.CrossRef Zhan, Y., Yu, J., Yu, T., & Tao, D. (2020). Multi-task compositional network for visual relationship detection. International Journal of Computer Vision, 128(8), 2146–2165.CrossRef
Zurück zum Zitat Zhan, Y., Yu, J., Yu, Z., Zhang, R., Tao, D., & Tian, Q. (2018). Comprehensive distance-preserving autoencoders for cross-modal retrieval. In ACM international conference on multimedia (pp. 1137–1145). Zhan, Y., Yu, J., Yu, Z., Zhang, R., Tao, D., & Tian, Q. (2018). Comprehensive distance-preserving autoencoders for cross-modal retrieval. In ACM international conference on multimedia (pp. 1137–1145).
Zurück zum Zitat Zhang, J., Chen, Z., & Tao, D. (2021). Towards high performance human keypoint detection. International Journal of Computer Vision, 129(9), 2639–2662.CrossRef Zhang, J., Chen, Z., & Tao, D. (2021). Towards high performance human keypoint detection. International Journal of Computer Vision, 129(9), 2639–2662.CrossRef
Zurück zum Zitat Zhang, S., Song, Z., Cao, X., Zhang, H., & Zhou, J. (2019). Task-aware attention model for clothing attribute prediction. IEEE Transactions on Circuits and Systems for Video, 30(4), 1051–1064.CrossRef Zhang, S., Song, Z., Cao, X., Zhang, H., & Zhou, J. (2019). Task-aware attention model for clothing attribute prediction. IEEE Transactions on Circuits and Systems for Video, 30(4), 1051–1064.CrossRef
Zurück zum Zitat Zheng, L., Shen, L., Tian, L., Wang, S., Wang, J., & Tian, Q. (2015). Scalable person re-identification: A benchmark. In ICCV (pp. 1116–1124). Zheng, L., Shen, L., Tian, L., Wang, S., Wang, J., & Tian, Q. (2015). Scalable person re-identification: A benchmark. In ICCV (pp. 1116–1124).
Zurück zum Zitat Zhu, J., Liao, S., Lei, Z., & Li, S. Z. (2017). Multi-label convolutional neural network based pedestrian attribute classification. Image and Vision Computing, 58, 224–229.CrossRef Zhu, J., Liao, S., Lei, Z., & Li, S. Z. (2017). Multi-label convolutional neural network based pedestrian attribute classification. Image and Vision Computing, 58, 224–229.CrossRef
Zurück zum Zitat Zhu, J., Liao, S., Yi, D., Lei, Z., & Li, S.Z. (2015). Multi-label cnn based pedestrian attribute learning for soft biometrics. In ICB (pp. 535–540). Zhu, J., Liao, S., Yi, D., Lei, Z., & Li, S.Z. (2015). Multi-label cnn based pedestrian attribute learning for soft biometrics. In ICB (pp. 535–540).
Zurück zum Zitat Zhu, J., Zeng, H., Huang, J., Zhu, X., Lei, Z., Cai, C., & Zheng, L. (2019). Body symmetry and part-locality-guided direct nonparametric deep feature enhancement for person reidentification. IEEE Internet of Things Journal, 7(3), 2053–2065.CrossRef Zhu, J., Zeng, H., Huang, J., Zhu, X., Lei, Z., Cai, C., & Zheng, L. (2019). Body symmetry and part-locality-guided direct nonparametric deep feature enhancement for person reidentification. IEEE Internet of Things Journal, 7(3), 2053–2065.CrossRef
Zurück zum Zitat Zhu, J., Zeng, H., Liao, S., Lei, Z., Cai, C., & Zheng, L. (2017). Deep hybrid similarity learning for person re-identification. IEEE Transactions on Circuits and Systems for Video Technology, 28(11), 3183–3193.CrossRef Zhu, J., Zeng, H., Liao, S., Lei, Z., Cai, C., & Zheng, L. (2017). Deep hybrid similarity learning for person re-identification. IEEE Transactions on Circuits and Systems for Video Technology, 28(11), 3183–3193.CrossRef
Metadaten
Titel
Attribute-Image Person Re-identification via Modal-Consistent Metric Learning
verfasst von
Jianqing Zhu
Liu Liu
Yibing Zhan
Xiaobin Zhu
Huanqiang Zeng
Dacheng Tao
Publikationsdatum
09.07.2023
Verlag
Springer US
Erschienen in
International Journal of Computer Vision / Ausgabe 11/2023
Print ISSN: 0920-5691
Elektronische ISSN: 1573-1405
DOI
https://doi.org/10.1007/s11263-023-01841-7

Weitere Artikel der Ausgabe 11/2023

International Journal of Computer Vision 11/2023 Zur Ausgabe

S.I. : Traditional Computer Vision in the Age of Deep Learning

Improving Domain Adaptation Through Class Aware Frequency Transformation

Premium Partner