Skip to main content
Erschienen in: International Journal of Machine Learning and Cybernetics 2/2024

05.08.2023 | Original Article

MSIF: multi-spectrum image fusion method for cross-modality person re-identification

verfasst von: Qingshan Chen, Zhenzhen Quan, Yifan Zheng, Yujun Li, Zhi Liu, Mikhail G. Mozerov

Erschienen in: International Journal of Machine Learning and Cybernetics | Ausgabe 2/2024

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Sketch-RGB cross-modality person re-identification (ReID) is a challenging task that aims to match a sketch portrait drawn by a professional artist with a full-body photo taken by surveillance equipment to deal with situations where the monitoring equipment is damaged at the accident scene. However, sketch portraits only provide highly abstract frontal body contour information and lack other important features such as color, pose, behavior, etc. The difference in saliency between the two modalities brings new challenges to cross-modality person ReID. To overcome this problem, this paper proposes a novel dual-stream model for cross-modality person ReID, which is able to mine modality-invariant features to reduce the discrepancy between sketch and camera images end-to-end. More specifically, we propose a multi-spectrum image fusion (MSIF) method, which aims to exploit the image appearance changes brought by multiple spectrums and guide the network to mine modality-invariant commonalities during training. It only processes the spectrum of the input images without adding additional calculations and model complexity, which can be easily integrated into other models. Moreover, we introduce a joint structure via a generalized mean pooling (GMP) layer and a self-attention (SA) mechanism to balance background and texture information and obtain the regional features with a large amount of information in the image. To further shrink the intra-class distance, a weighted regularized triplet (WRT) loss is developed without introducing additional hyperparameters. The model was first evaluated on the PKU Sketch ReID dataset, and extensive experimental results show that the Rank-1/mAP accuracy of our method is 87.00%/91.12%, reaching the current state-of-the-art performance. To further validate the effectiveness of our approach in handling cross-modality person ReID, we conducted experiments on two commonly used IR-RGB datasets (SYSU-MM01 and RegDB). The obtained results show that our method achieves competitive performance. These results confirm the ability of our method to effectively process images from different modalities.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Weitere Produktempfehlungen anzeigen
Literatur
2.
Zurück zum Zitat Bryan B, Gong Y, Zhang Y, Poellabauer C (2019) Second-order nonlocal attention networks for person re-identification. In: Proc. ICCV’19, Seoul, Korea, Oct. 2019, pp 3759–3768 Bryan B, Gong Y, Zhang Y, Poellabauer C (2019) Second-order nonlocal attention networks for person re-identification. In: Proc. ICCV’19, Seoul, Korea, Oct. 2019, pp 3759–3768
3.
Zurück zum Zitat Chen T, Ding S, Xie J, Yuan Y, Chen W, Yang Y, Ren Z, Wang Z (2019) Abd-net: Attentive but diverse person re-identification. In: Proc. ICCV’19, Seoul, Korea, Oct. 2019, pp 8350–8360 Chen T, Ding S, Xie J, Yuan Y, Chen W, Yang Y, Ren Z, Wang Z (2019) Abd-net: Attentive but diverse person re-identification. In: Proc. ICCV’19, Seoul, Korea, Oct. 2019, pp 8350–8360
4.
Zurück zum Zitat Luo X, Jiang M, Kong J (2022) Selective relation-aware representations for person re-identification. Int J Mach Learn Cybern 13(11):3523–3541CrossRef Luo X, Jiang M, Kong J (2022) Selective relation-aware representations for person re-identification. Int J Mach Learn Cybern 13(11):3523–3541CrossRef
6.
Zurück zum Zitat Chen Y, Xia R, Zou K, Yang K (2023) FFTI: Image inpainting algorithm via features fusion and two-steps inpainting. J Vis Commun Image Represent 91:103776CrossRef Chen Y, Xia R, Zou K, Yang K (2023) FFTI: Image inpainting algorithm via features fusion and two-steps inpainting. J Vis Commun Image Represent 91:103776CrossRef
8.
Zurück zum Zitat Xia R, Chen Y, Ren B (2022) Improved anti-occlusion object tracking algorithm using unscented Rauch-Tung-Striebel smoother and kernel correlation filter. J King Saud Univ-Comput Inform Sci 34(8):6008–6018 Xia R, Chen Y, Ren B (2022) Improved anti-occlusion object tracking algorithm using unscented Rauch-Tung-Striebel smoother and kernel correlation filter. J King Saud Univ-Comput Inform Sci 34(8):6008–6018
9.
Zurück zum Zitat Wang Z, Ye M, Yang F, Bai X, Satoh S (2018) Cascaded sr-gan for scale-adaptive low resolution person re-identification. In: Proc. IJCAI’18, Stockholm, Sweden, Jul. 2018, pp 3891–3897 Wang Z, Ye M, Yang F, Bai X, Satoh S (2018) Cascaded sr-gan for scale-adaptive low resolution person re-identification. In: Proc. IJCAI’18, Stockholm, Sweden, Jul. 2018, pp 3891–3897
11.
Zurück zum Zitat Chen Y-C, Li Y-J, X. Du, and Y.-C. F. Wang, “Learning resolution invariant deep representations for person re-identification. In: Proc. AAAI’19, Hawat, USA, Jau. 2019, pp. 8215–8222. Chen Y-C, Li Y-J, X. Du, and Y.-C. F. Wang, “Learning resolution invariant deep representations for person re-identification. In: Proc. AAAI’19, Hawat, USA, Jau. 2019, pp. 8215–8222.
12.
Zurück zum Zitat Cheng Z, Dong Q, Gong S, Zhu X (2020) Inter-task association critic for cross-resolution person re-identification. In: Proc. CVPR’20, Seattle, WA, USA, Jun. 2020, pp 2602–2612 Cheng Z, Dong Q, Gong S, Zhu X (2020) Inter-task association critic for cross-resolution person re-identification. In: Proc. CVPR’20, Seattle, WA, USA, Jun. 2020, pp 2602–2612
13.
Zurück zum Zitat Hou R, Chang H, Ma B, Huang R, Shan S (2021) Bicnet-tks: learning efficient spatial-temporal representation for video person reidentification. In: Proc. CVPR’21, Jun. 2021, pp 2014–2023 Hou R, Chang H, Ma B, Huang R, Shan S (2021) Bicnet-tks: learning efficient spatial-temporal representation for video person reidentification. In: Proc. CVPR’21, Jun. 2021, pp 2014–2023
14.
Zurück zum Zitat Chen Y, Liu L, Phonevilay V, Gu K, Xia R, Xie J, Zhang Q, Yang K (2021) Image super-resolution reconstruction based on feature map attention mechanism. Appl Intell 51:4367–4380CrossRef Chen Y, Liu L, Phonevilay V, Gu K, Xia R, Xie J, Zhang Q, Yang K (2021) Image super-resolution reconstruction based on feature map attention mechanism. Appl Intell 51:4367–4380CrossRef
15.
Zurück zum Zitat Chen Y, Phonevilay V, Tao J, Chen X, Xia R, Zhang Q, Yang K, Xiong J, Xie J (2021) The face image super-resolution algorithm based on combined representation learning. Multimed Tools Appl 80:30839–30861CrossRef Chen Y, Phonevilay V, Tao J, Chen X, Xia R, Zhang Q, Yang K, Xiong J, Xie J (2021) The face image super-resolution algorithm based on combined representation learning. Multimed Tools Appl 80:30839–30861CrossRef
17.
Zurück zum Zitat Wu A, Zheng W-S, Yu H-X, Gong S, Lai J (2017) Rgb-infrared crossmodality person re-identification. In: Proc. ICCV’17, Venice, Italy, Oct. 2017, pp 5390–5399 Wu A, Zheng W-S, Yu H-X, Gong S, Lai J (2017) Rgb-infrared crossmodality person re-identification. In: Proc. ICCV’17, Venice, Italy, Oct. 2017, pp 5390–5399
18.
Zurück zum Zitat Ye M, Lan X, Li J, Yuen P (2018) Hierarchical discriminative learning for visible thermal person re-identification. In: Proc. AAAI’18, New Orleans, Louisiana, USA, Feb. 2018, pp 7501–7508 Ye M, Lan X, Li J, Yuen P (2018) Hierarchical discriminative learning for visible thermal person re-identification. In: Proc. AAAI’18, New Orleans, Louisiana, USA, Feb. 2018, pp 7501–7508
19.
Zurück zum Zitat Hao Y, Wang N, Li J, Gao X (2019) Hsme: hypersphere manifold embedding for visible thermal person re-identification. In: Proc. AAAI’19, Hawat, USA, Jau. 2019, pp 8385–8392 Hao Y, Wang N, Li J, Gao X (2019) Hsme: hypersphere manifold embedding for visible thermal person re-identification. In: Proc. AAAI’19, Hawat, USA, Jau. 2019, pp 8385–8392
20.
Zurück zum Zitat Zhang P, Xu J, Wu Q, Huang Y, Zhang J (2019) Top-push constrained modality-adaptive dictionary learning for cross-modality person reidentification. IEEE Trans Circuits Syst Video Technol 30(12):4554–4566CrossRef Zhang P, Xu J, Wu Q, Huang Y, Zhang J (2019) Top-push constrained modality-adaptive dictionary learning for cross-modality person reidentification. IEEE Trans Circuits Syst Video Technol 30(12):4554–4566CrossRef
21.
Zurück zum Zitat Ye M, Shen J, Shao L (2020) Visible-infrared person re-identification via homogeneous augmented tri-modal learning. IEEE Trans Inf Forensics Secur 16:728–739CrossRef Ye M, Shen J, Shao L (2020) Visible-infrared person re-identification via homogeneous augmented tri-modal learning. IEEE Trans Inf Forensics Secur 16:728–739CrossRef
22.
Zurück zum Zitat Feng Z, Lai J, Xie X (2020) Learning modality-specific representations for visible-infrared person re-identification. IEEE Trans Image Process 29:579–590MathSciNetCrossRef Feng Z, Lai J, Xie X (2020) Learning modality-specific representations for visible-infrared person re-identification. IEEE Trans Image Process 29:579–590MathSciNetCrossRef
23.
Zurück zum Zitat Choi S, Lee S, Kim Y, Kim T, Kim C (2020) Hi-cmd: Hierarchical cross-modality disentanglement for visible-infrared person reidentification. In: Proc. CVPR’20, Seattle, WA, USA, Jun. 2020, pp 10257–10266 Choi S, Lee S, Kim Y, Kim T, Kim C (2020) Hi-cmd: Hierarchical cross-modality disentanglement for visible-infrared person reidentification. In: Proc. CVPR’20, Seattle, WA, USA, Jun. 2020, pp 10257–10266
24.
Zurück zum Zitat Chen Y, Wan L, Li Z, Jing Q, Sun Z (2021) Neural feature search for rgb-infrared person re-identification. In: Proc. CVPR’21, Jun. 2021, pp 587–597 Chen Y, Wan L, Li Z, Jing Q, Sun Z (2021) Neural feature search for rgb-infrared person re-identification. In: Proc. CVPR’21, Jun. 2021, pp 587–597
25.
Zurück zum Zitat Niu K, Huang Y, Ouyang W, Wang L (2020) Improving description-based person re-identification by multi-granularity image-text alignments. IEEE Trans Image Process 29:5542–5556CrossRef Niu K, Huang Y, Ouyang W, Wang L (2020) Improving description-based person re-identification by multi-granularity image-text alignments. IEEE Trans Image Process 29:5542–5556CrossRef
26.
Zurück zum Zitat Wu J, Liu H, Jiang J, Qi M, Ren B, Li X, Wang Y (2020) Person attribute recognition by sequence contextual relation learning. IEEE Trans Circuits Syst Video Technol 30(10):3398–3412CrossRef Wu J, Liu H, Jiang J, Qi M, Ren B, Li X, Wang Y (2020) Person attribute recognition by sequence contextual relation learning. IEEE Trans Circuits Syst Video Technol 30(10):3398–3412CrossRef
27.
Zurück zum Zitat Zheng K, Liu W, Liu J, Zha Z-J, Mei T (2020) Hierarchical gumbel attention network for text-based person search. In: Proc. ACM’20, Seattle, USA, Oct. 2020, pp 3441–3449 Zheng K, Liu W, Liu J, Zha Z-J, Mei T (2020) Hierarchical gumbel attention network for text-based person search. In: Proc. ACM’20, Seattle, USA, Oct. 2020, pp 3441–3449
28.
Zurück zum Zitat Zhang P, Ouyang D, Chen F, Shao J (2021) Multiplicative angular margin loss for text-based person search. In: Proc. ACM Asia’21, Singapore, Mar. 2021, pp 1–7 Zhang P, Ouyang D, Chen F, Shao J (2021) Multiplicative angular margin loss for text-based person search. In: Proc. ACM Asia’21, Singapore, Mar. 2021, pp 1–7
29.
Zurück zum Zitat Pang L, Wang Y, Song Y-Z, Huang T, Tian Y (2018) Cross-domain adversarial feature learning for sketch re-identification. In: Proc. ACM’18, Seoul, Korea, Oct. 2018, pp 609–617 Pang L, Wang Y, Song Y-Z, Huang T, Tian Y (2018) Cross-domain adversarial feature learning for sketch re-identification. In: Proc. ACM’18, Seoul, Korea, Oct. 2018, pp 609–617
30.
Zurück zum Zitat Gui S, Zhu Y, Qin X, Ling X (2020) Learning multi-level domain invariant features for sketch re-identification. Neurocomputing 403:294–303CrossRef Gui S, Zhu Y, Qin X, Ling X (2020) Learning multi-level domain invariant features for sketch re-identification. Neurocomputing 403:294–303CrossRef
31.
Zurück zum Zitat Yang F, Wu Y, Wang Z, Li X, Sakti S, Nakamura S (2021) Instance level heterogeneous domain adaptation for limited-labeled sketch-to-photo retrieval. IEEE Trans Multimed 23:2347–2360CrossRef Yang F, Wu Y, Wang Z, Li X, Sakti S, Nakamura S (2021) Instance level heterogeneous domain adaptation for limited-labeled sketch-to-photo retrieval. IEEE Trans Multimed 23:2347–2360CrossRef
32.
Zurück zum Zitat Yuan B, Chen B, Tan Z, Shao X, Bao B-K (2022) Unbiased feature enhancement framework for cross-modality person re-identification. Multimed Syst 28(3):749–759CrossRef Yuan B, Chen B, Tan Z, Shao X, Bao B-K (2022) Unbiased feature enhancement framework for cross-modality person re-identification. Multimed Syst 28(3):749–759CrossRef
33.
Zurück zum Zitat Zhu F, Zhu Y, Jiang X, Ye J (2022) Cross-domain attention and center loss for sketch re-identification. IEEE Trans Inf Forensics Secur 17:3421–3432CrossRef Zhu F, Zhu Y, Jiang X, Ye J (2022) Cross-domain attention and center loss for sketch re-identification. IEEE Trans Inf Forensics Secur 17:3421–3432CrossRef
34.
Zurück zum Zitat Rachmadi RF, Nugroho SMS, Purnama IKE (2022) Revisiting dropout regularization for cross-modality person re-identification. IEEE Access 10:102195–102209CrossRef Rachmadi RF, Nugroho SMS, Purnama IKE (2022) Revisiting dropout regularization for cross-modality person re-identification. IEEE Access 10:102195–102209CrossRef
35.
Zurück zum Zitat Chen Q, Quan Z, Zhao K, Zheng Y, Liu Z, Li Y (2022) A cross modality sketch person re-identification model based on cross-spectrum image generation. In: Proc. IFTC’21. Springer, Shanghai, China, Dec. 2022, pp 312–324 Chen Q, Quan Z, Zhao K, Zheng Y, Liu Z, Li Y (2022) A cross modality sketch person re-identification model based on cross-spectrum image generation. In: Proc. IFTC’21. Springer, Shanghai, China, Dec. 2022, pp 312–324
36.
Zurück zum Zitat Chen C, Ye M, Qi M, Du B (2022) Sketch transformer: Asymmetrical disentanglement learning from dynamic synthesis. In: Proc. ACM’22, Lisboa Portugal, Oct. 2022, pp 4012–4020 Chen C, Ye M, Qi M, Du B (2022) Sketch transformer: Asymmetrical disentanglement learning from dynamic synthesis. In: Proc. ACM’22, Lisboa Portugal, Oct. 2022, pp 4012–4020
37.
Zurück zum Zitat Liu J, Ni B, Yan Y, Zhou P, Cheng S, Hu J (2018) Pose transferrable person re-identification. In: Proc. CVPR’18, Salt Lake City, UT, USA, Jun. 2018, pp 4099–4108 Liu J, Ni B, Yan Y, Zhou P, Cheng S, Hu J (2018) Pose transferrable person re-identification. In: Proc. CVPR’18, Salt Lake City, UT, USA, Jun. 2018, pp 4099–4108
38.
Zurück zum Zitat Hou R, Ma B, Chang H, Gu X, Shan S, Chen X (2019) Interaction-and-aggregation network for person re-identification. In: Proc. CVPR’19, Long Beach, CA, USA, Jun. 2019, pp 9309–9318 Hou R, Ma B, Chang H, Gu X, Shan S, Chen X (2019) Interaction-and-aggregation network for person re-identification. In: Proc. CVPR’19, Long Beach, CA, USA, Jun. 2019, pp 9309–9318
39.
Zurück zum Zitat Luo H, Jiang W, Zhang X, Fan X, Qian J, Zhang C (2019) Alignedreid++: dynamically matching local information for person reidentification. Pattern Recogn 94:53–61CrossRef Luo H, Jiang W, Zhang X, Fan X, Qian J, Zhang C (2019) Alignedreid++: dynamically matching local information for person reidentification. Pattern Recogn 94:53–61CrossRef
40.
Zurück zum Zitat Tian M, Yi S, Li H, Li S, Zhang X, Shi J, Yan J, Wang X (2018) Eliminating background-bias for robust person re-identification. In: Proc. CVPR’18, Salt Lake City, UT, USA, Jun. 2018, pp 5794–5803 Tian M, Yi S, Li H, Li S, Zhang X, Shi J, Yan J, Wang X (2018) Eliminating background-bias for robust person re-identification. In: Proc. CVPR’18, Salt Lake City, UT, USA, Jun. 2018, pp 5794–5803
42.
Zurück zum Zitat Ding C, Wang K, Wang P, Tao D (2020) Multi-task learning with coarse priors for robust part-aware person re-identification. IEEE Trans Pattern Anal Mach Intell 44:1474–1488CrossRef Ding C, Wang K, Wang P, Tao D (2020) Multi-task learning with coarse priors for robust part-aware person re-identification. IEEE Trans Pattern Anal Mach Intell 44:1474–1488CrossRef
43.
Zurück zum Zitat Niu J, Huang C, Li J, Fan M (2018) Parallel computing techniques for concept-cognitive learning based on granular computing. Int J Mach Learn Cybern 9(11):1785–1805CrossRef Niu J, Huang C, Li J, Fan M (2018) Parallel computing techniques for concept-cognitive learning based on granular computing. Int J Mach Learn Cybern 9(11):1785–1805CrossRef
44.
Zurück zum Zitat Shen Y, Li H, Yi S, Chen D, Wang X (2018) Person re-identification with deep similarity-guided graph neural network. In: Proceedings of the European conference on computer vision (ECCV), pp 486–504 Shen Y, Li H, Yi S, Chen D, Wang X (2018) Person re-identification with deep similarity-guided graph neural network. In: Proceedings of the European conference on computer vision (ECCV), pp 486–504
45.
Zurück zum Zitat Miao J, Wu Y, Liu P, Ding Y, Yang Y (2019), Pose-guided feature alignment for occluded person re-identification. In: Proc. ICCV’19, Seoul, Korea, Oct. 2019, pp 542–551 Miao J, Wu Y, Liu P, Ding Y, Yang Y (2019), Pose-guided feature alignment for occluded person re-identification. In: Proc. ICCV’19, Seoul, Korea, Oct. 2019, pp 542–551
46.
Zurück zum Zitat Luo C, Chen Y, Wang N, Zhang Z (2019) Spectral feature transformation for person re-identification. In: Proc. ICCV’19, Seoul, Korea, Oct. 2019, pp 4975–4984 Luo C, Chen Y, Wang N, Zhang Z (2019) Spectral feature transformation for person re-identification. In: Proc. ICCV’19, Seoul, Korea, Oct. 2019, pp 4975–4984
47.
Zurück zum Zitat Xu J, Zhao R, Zhu F, Wang H, Ouyang W (2018) Attention-aware compositional network for person re-identification. In: Proc. CVPR’18, Salt Lake City, UT, USA, Jun. 2018, pp 2119–2128 Xu J, Zhao R, Zhu F, Wang H, Ouyang W (2018) Attention-aware compositional network for person re-identification. In: Proc. CVPR’18, Salt Lake City, UT, USA, Jun. 2018, pp 2119–2128
48.
Zurück zum Zitat Zhong Z, Zheng L, Zheng Z, Li S, Yang Y (2019) Camstyle: A novel data augmentation method for person re-identification. IEEE Trans Image Process 28(3):1176–1190MathSciNetCrossRef Zhong Z, Zheng L, Zheng Z, Li S, Yang Y (2019) Camstyle: A novel data augmentation method for person re-identification. IEEE Trans Image Process 28(3):1176–1190MathSciNetCrossRef
49.
Zurück zum Zitat Xie K, Wu Y, Xiao J, Li J, Xiao G, Cao Y (2021) Unsupervised person re-identification via k-reciprocal encoding and style transfer. Int J Mach Learn Cybern 12(10):2899–2916CrossRef Xie K, Wu Y, Xiao J, Li J, Xiao G, Cao Y (2021) Unsupervised person re-identification via k-reciprocal encoding and style transfer. Int J Mach Learn Cybern 12(10):2899–2916CrossRef
50.
Zurück zum Zitat Wei L, Zhang S, Gao W, Tian Q (2018) Person transfer gan to bridge domain gap for person re-identification. In: Proc. CVPR’18, Salt Lake City, UT, USA, Jun. 2018, pp 79–88 Wei L, Zhang S, Gao W, Tian Q (2018) Person transfer gan to bridge domain gap for person re-identification. In: Proc. CVPR’18, Salt Lake City, UT, USA, Jun. 2018, pp 79–88
51.
Zurück zum Zitat Wang G, Yang S, Liu H, Wang Z, Yang Y, Wang S, Yu G, Zhou E, Sun J (2020) High-order information matters: Learning relation and topology for occluded person re-identification. In: Proc. CVPR’20, Seattle, WA, USA, Jun. 2020, pp 6449–6458 Wang G, Yang S, Liu H, Wang Z, Yang Y, Wang S, Yu G, Zhou E, Sun J (2020) High-order information matters: Learning relation and topology for occluded person re-identification. In: Proc. CVPR’20, Seattle, WA, USA, Jun. 2020, pp 6449–6458
52.
Zurück zum Zitat Kiani Galoogahi H, Sim T (2012) Face photo retrieval by sketch example. In: Proc. ACM’12, Nara, Japan, Oct. 2012, pp 949–952 Kiani Galoogahi H, Sim T (2012) Face photo retrieval by sketch example. In: Proc. ACM’12, Nara, Japan, Oct. 2012, pp 949–952
53.
Zurück zum Zitat Zhang W, Wang X, Tang X (2011) Coupled information-theoretic encoding for face photo-sketch recognition. In: Proc. CVPR’11, Colorado Springs, CO, USA, Jun. 2011, pp 513–520 Zhang W, Wang X, Tang X (2011) Coupled information-theoretic encoding for face photo-sketch recognition. In: Proc. CVPR’11, Colorado Springs, CO, USA, Jun. 2011, pp 513–520
54.
Zurück zum Zitat Rahman A, Beg M (2015) Face sketch recognition using sketching with words. Int J Mach Learn Cybern 6(4):597–605CrossRef Rahman A, Beg M (2015) Face sketch recognition using sketching with words. Int J Mach Learn Cybern 6(4):597–605CrossRef
55.
Zurück zum Zitat Bhunia AK, Yang Y, Hospedales TM, Xiang T, Song Y-Z (2020) Sketch less for more: On-the-fly fine-grained sketch-based image retrieval. In: Proc. CVPR’20, Seattle, WA, USA, Jun. 2020, pp 9779–9788 Bhunia AK, Yang Y, Hospedales TM, Xiang T, Song Y-Z (2020) Sketch less for more: On-the-fly fine-grained sketch-based image retrieval. In: Proc. CVPR’20, Seattle, WA, USA, Jun. 2020, pp 9779–9788
56.
Zurück zum Zitat Pang K, Li K, Yang Y, Zhang H, Hospedales TM, Xiang T, Song Y-Z (2019) Generalising fine-grained sketch-based image retrieval. In: Proc. CVPR’19, Long Beach, CA, USA, Jun. 2019, pp 677–686 Pang K, Li K, Yang Y, Zhang H, Hospedales TM, Xiang T, Song Y-Z (2019) Generalising fine-grained sketch-based image retrieval. In: Proc. CVPR’19, Long Beach, CA, USA, Jun. 2019, pp 677–686
57.
Zurück zum Zitat Ye M, Wang Z, Lan X, Yuen PC (2018) Visible thermal person re-identification via dual-constrained top-ranking. In: Proc. IJCAI’18, Stockholm, Jul. 2018, pp 1092–1099 Ye M, Wang Z, Lan X, Yuen PC (2018) Visible thermal person re-identification via dual-constrained top-ranking. In: Proc. IJCAI’18, Stockholm, Jul. 2018, pp 1092–1099
58.
Zurück zum Zitat Ye M, Lan X, Wang Z, Yuen PC (2019) Bi-directional center-constrained top-ranking for visible thermal person re-identification. IEEE Trans Inf Forensics Secur 15:407–419CrossRef Ye M, Lan X, Wang Z, Yuen PC (2019) Bi-directional center-constrained top-ranking for visible thermal person re-identification. IEEE Trans Inf Forensics Secur 15:407–419CrossRef
60.
Zurück zum Zitat Zhang D, Zhang Z, Ju Y, Wang C, Xie Y, Qu Y (2022) Dual mutual learning for cross-modality person re-identification. IEEE Trans Circuits Syst Video Technol 32(8):5361–5373CrossRef Zhang D, Zhang Z, Ju Y, Wang C, Xie Y, Qu Y (2022) Dual mutual learning for cross-modality person re-identification. IEEE Trans Circuits Syst Video Technol 32(8):5361–5373CrossRef
61.
Zurück zum Zitat Zheng X, Chen X, Lu X (2022) Visible-infrared person re-identification via partially interactive collaboration. IEEE Trans Image Process 31:6951–6963CrossRef Zheng X, Chen X, Lu X (2022) Visible-infrared person re-identification via partially interactive collaboration. IEEE Trans Image Process 31:6951–6963CrossRef
62.
Zurück zum Zitat Hao X, Zhao S, Ye M, Shen J (2021) Cross-modality person reidentification via modality confusion and center aggregation. In: Proc. ICCV’21, Mar. 2021, pp. 16403–16412 Hao X, Zhao S, Ye M, Shen J (2021) Cross-modality person reidentification via modality confusion and center aggregation. In: Proc. ICCV’21, Mar. 2021, pp. 16403–16412
63.
Zurück zum Zitat Sun J, Li Y, Chen H, Peng Y, Zhu J (2023) Visible-infrared person re-identification model based on feature consistency and modal indistinguishability. Mach Vis Appl 34(1):14CrossRef Sun J, Li Y, Chen H, Peng Y, Zhu J (2023) Visible-infrared person re-identification model based on feature consistency and modal indistinguishability. Mach Vis Appl 34(1):14CrossRef
64.
Zurück zum Zitat Wang Z, Wang Z, Zheng Y, Chuang Y-Y, Satoh S (2019) Learning to reduce dual-level discrepancy for infrared-visible person reidentification. In: Proc. CVPR’19, Long Beach, CA, USA, Jun. 2019, pp 618–626 Wang Z, Wang Z, Zheng Y, Chuang Y-Y, Satoh S (2019) Learning to reduce dual-level discrepancy for infrared-visible person reidentification. In: Proc. CVPR’19, Long Beach, CA, USA, Jun. 2019, pp 618–626
65.
Zurück zum Zitat Wang G, Zhang T, Cheng J, Liu S, Yang Y, Hou Z (2019) Rgb-infrared cross-modality person re-identification via joint pixel and feature alignment. In: Proc. ICCV’19, Seoul, Korea, Oct. 2019, pp 3622–3631 Wang G, Zhang T, Cheng J, Liu S, Yang Y, Hou Z (2019) Rgb-infrared cross-modality person re-identification via joint pixel and feature alignment. In: Proc. ICCV’19, Seoul, Korea, Oct. 2019, pp 3622–3631
66.
Zurück zum Zitat Wang G-A, Zhang T, Yang Y, Cheng J, Chang J, Liang X, Hou Z-G (2020) Cross-modality paired-images generation for rgb-infrared person re-identification. In: Proc. AAAI’20, New York, USA, Feb. 2020, pp 12144–12151 Wang G-A, Zhang T, Yang Y, Cheng J, Chang J, Liang X, Hou Z-G (2020) Cross-modality paired-images generation for rgb-infrared person re-identification. In: Proc. AAAI’20, New York, USA, Feb. 2020, pp 12144–12151
67.
Zurück zum Zitat Zeng X, Long J, Tian S, Xiao G (2023) Random area pixel variation and random area transform for visible-infrared cross-modal pedestrian re-identification. Expert Syst Appl 215:119307CrossRef Zeng X, Long J, Tian S, Xiao G (2023) Random area pixel variation and random area transform for visible-infrared cross-modal pedestrian re-identification. Expert Syst Appl 215:119307CrossRef
68.
Zurück zum Zitat Li D, Wei X, Hong X, Gong Y (2020) Infrared-visible cross-modal person re-identification with an x modality. In: Proc. AAAI’20, New York, USA, Feb. 2020, pp. 4610–4617 Li D, Wei X, Hong X, Gong Y (2020) Infrared-visible cross-modal person re-identification with an x modality. In: Proc. AAAI’20, New York, USA, Feb. 2020, pp. 4610–4617
69.
Zurück zum Zitat Ye M, Shen J, Crandall DJ, Shao L, Luo J (2020) Dynamic dual-attentive aggregation learning for visible-infrared person reidentification. In: Proc. ECCV’20, Aug. 2020, pp 229–247 Ye M, Shen J, Crandall DJ, Shao L, Luo J (2020) Dynamic dual-attentive aggregation learning for visible-infrared person reidentification. In: Proc. ECCV’20, Aug. 2020, pp 229–247
70.
Zurück zum Zitat Wei Z, Yang X, Wang N, Gao X (2021) Syncretic modality collaborative learning for visible infrared person re-identification. In: Proc. ICCV’21, Mar. 2021, pp 225–234 Wei Z, Yang X, Wang N, Gao X (2021) Syncretic modality collaborative learning for visible infrared person re-identification. In: Proc. ICCV’21, Mar. 2021, pp 225–234
71.
Zurück zum Zitat Alehdaghi M, Josi A, Cruz RM, Granger E (2022) Visible-infrared person re-identification using privileged intermediate information. In: Proc. ECCV’22 Workshops: Tel Aviv, Israel, Oct., 2022, pp 720–737 Alehdaghi M, Josi A, Cruz RM, Granger E (2022) Visible-infrared person re-identification using privileged intermediate information. In: Proc. ECCV’22 Workshops: Tel Aviv, Israel, Oct., 2022, pp 720–737
72.
Zurück zum Zitat He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proc. CVPR’16, Las Vegas, USA, Jun. 2016, pp 770–778 He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proc. CVPR’16, Las Vegas, USA, Jun. 2016, pp 770–778
74.
Zurück zum Zitat Zhang H, Goodfellow I, Metaxas D, Odena A (2019) Self-attention generative adversarial networks. In: Proc. PMLR’19, 2019, pp 7354–7363 Zhang H, Goodfellow I, Metaxas D, Odena A (2019) Self-attention generative adversarial networks. In: Proc. PMLR’19, 2019, pp 7354–7363
75.
Zurück zum Zitat Radenović F, Tolias G, Chum O (2019) Fine-tuning cnn image retrieval with no human annotation. IEEE Trans Pattern Anal Mach Intell 41(7):1655–1668CrossRef Radenović F, Tolias G, Chum O (2019) Fine-tuning cnn image retrieval with no human annotation. IEEE Trans Pattern Anal Mach Intell 41(7):1655–1668CrossRef
77.
Zurück zum Zitat Nguyen DT, Hong HG, Kim KW, Park KR (2017) Person recognition system based on a combination of body images from visible light and thermal cameras. Sensors 17(3):605CrossRef Nguyen DT, Hong HG, Kim KW, Park KR (2017) Person recognition system based on a combination of body images from visible light and thermal cameras. Sensors 17(3):605CrossRef
78.
Zurück zum Zitat Paszke A, Gross S, Massa F, Lerer A, Bradbury J, Chanan G, Killeen T, Lin Z, Gimelshein N, Antiga L et al (2019) Pytorch: an imperative style, high-performance deep learning library. Adv Neural Inf Process Syst 32:8026–8037 Paszke A, Gross S, Massa F, Lerer A, Bradbury J, Chanan G, Killeen T, Lin Z, Gimelshein N, Antiga L et al (2019) Pytorch: an imperative style, high-performance deep learning library. Adv Neural Inf Process Syst 32:8026–8037
79.
Zurück zum Zitat Yu Q, Liu F, Song Y-Z, Xiang T, Hospedales TM, Loy C-C (2016) Sketch me that shoe. In: Proc. CVPR’16, Las Vegas, NV, USA, Jun. 2016, pp 799–807 Yu Q, Liu F, Song Y-Z, Xiang T, Hospedales TM, Loy C-C (2016) Sketch me that shoe. In: Proc. CVPR’16, Las Vegas, NV, USA, Jun. 2016, pp 799–807
80.
Zurück zum Zitat Sangkloy P, Burnell N, Ham C, Hays J (2016) The sketchy database: learning to retrieve badly drawn bunnies. ACM Trans Graph 35(4):1–12CrossRef Sangkloy P, Burnell N, Ham C, Hays J (2016) The sketchy database: learning to retrieve badly drawn bunnies. ACM Trans Graph 35(4):1–12CrossRef
82.
Zurück zum Zitat Lin H, Fu Y, Lu P, Gong S, Xue X, Jiang Y-G (2019) Tc-net for isbir: Triplet classification network for instance-level sketch-based image retrieval. In: Proc. ACM’19, Nice, France, Oct. 2019, pp 1676–1684 Lin H, Fu Y, Lu P, Gong S, Xue X, Jiang Y-G (2019) Tc-net for isbir: Triplet classification network for instance-level sketch-based image retrieval. In: Proc. ACM’19, Nice, France, Oct. 2019, pp 1676–1684
83.
Zurück zum Zitat Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, Erhan D, Vanhoucke V, Rabinovich A (2015) Going deeper with convolutions. In: Proc. CVPR’15, Boston, Massachusetts, USA, Jun. 2015, pp 1–9 Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, Erhan D, Vanhoucke V, Rabinovich A (2015) Going deeper with convolutions. In: Proc. CVPR’15, Boston, Massachusetts, USA, Jun. 2015, pp 1–9
84.
Zurück zum Zitat Kang JK, Hoang TM, Park KR (2019) Person re-identification between visible and thermal camera images based on deep residual cnn using single input. IEEE Access 7:57972–57984CrossRef Kang JK, Hoang TM, Park KR (2019) Person re-identification between visible and thermal camera images based on deep residual cnn using single input. IEEE Access 7:57972–57984CrossRef
85.
Zurück zum Zitat Lu Y, Wu Y, Liu B, Zhang T, Li B, Chu Q, Yu N (2020) Crossmodality person re-identification with shared-specific feature transfer. In: Proc. CVPR’20, Seattle, WA, USA, Jun. 2020, pp 13379–13389 Lu Y, Wu Y, Liu B, Zhang T, Li B, Chu Q, Yu N (2020) Crossmodality person re-identification with shared-specific feature transfer. In: Proc. CVPR’20, Seattle, WA, USA, Jun. 2020, pp 13379–13389
86.
Zurück zum Zitat Dai P, Ji R, Wang H, Wu Q, Huang Y (2018) Cross-modality person re-identification with generative adversarial training. In: Proc. IJCAI’18, Stockholm, Jul. 2018, pp 677–683 Dai P, Ji R, Wang H, Wu Q, Huang Y (2018) Cross-modality person re-identification with generative adversarial training. In: Proc. IJCAI’18, Stockholm, Jul. 2018, pp 677–683
87.
Zurück zum Zitat Zhong X, Lu T, Huang W, Ye M, Jia X, Lin C-W (2022) Grayscale enhancement colorization network for visible-infrared person re-identification. IEEE Trans Circuits Syst Video Technol 32(3):1418–1430CrossRef Zhong X, Lu T, Huang W, Ye M, Jia X, Lin C-W (2022) Grayscale enhancement colorization network for visible-infrared person re-identification. IEEE Trans Circuits Syst Video Technol 32(3):1418–1430CrossRef
88.
Zurück zum Zitat Xu X, Liu S, Zhang N, Xiao G, Wu S (2022) Channel exchange and adversarial learning guided cross-modal person re-identification. Knowl-Based Syst 257:109883CrossRef Xu X, Liu S, Zhang N, Xiao G, Wu S (2022) Channel exchange and adversarial learning guided cross-modal person re-identification. Knowl-Based Syst 257:109883CrossRef
90.
Zurück zum Zitat Zheng L, Shen L, Tian L, Wang S, Wang J, Tian Q (2015) Scalable person re-identification: a benchmark. In: Proc. ICCV’15, Santiago, Chile, Dec. 2015, pp 1116–1124 Zheng L, Shen L, Tian L, Wang S, Wang J, Tian Q (2015) Scalable person re-identification: a benchmark. In: Proc. ICCV’15, Santiago, Chile, Dec. 2015, pp 1116–1124
91.
Zurück zum Zitat Ghosh A, Zhang R, Dokania PK, Wang O, Efros AA, Torr PH, Shechtman E (2019) Interactive sketch & fill: Multiclass sketch-to-image translation. In: Proc. ICCV’19, Seoul, Korea, Oct. 2019, pp 1171–1180 Ghosh A, Zhang R, Dokania PK, Wang O, Efros AA, Torr PH, Shechtman E (2019) Interactive sketch & fill: Multiclass sketch-to-image translation. In: Proc. ICCV’19, Seoul, Korea, Oct. 2019, pp 1171–1180
Metadaten
Titel
MSIF: multi-spectrum image fusion method for cross-modality person re-identification
verfasst von
Qingshan Chen
Zhenzhen Quan
Yifan Zheng
Yujun Li
Zhi Liu
Mikhail G. Mozerov
Publikationsdatum
05.08.2023
Verlag
Springer Berlin Heidelberg
Erschienen in
International Journal of Machine Learning and Cybernetics / Ausgabe 2/2024
Print ISSN: 1868-8071
Elektronische ISSN: 1868-808X
DOI
https://doi.org/10.1007/s13042-023-01932-4

Weitere Artikel der Ausgabe 2/2024

International Journal of Machine Learning and Cybernetics 2/2024 Zur Ausgabe

Neuer Inhalt