MFCNet: Mining Features Context Network for RGB–IR Person Re-Identification
Abstract
:1. Introduction
- A new end-to-end deep learning network structure is proposed to enhance the discriminant ability of feature representations for RGB–IR Re-ID problems. This approach provides a strong baseline for future research in this field.
- This paper introduces a new attention module combining spatial attention and global attention into the field of RGB–IR Re-ID. Based on this design, the SGA module combined with the RestNet50 backbone network can extract discriminative features by mining the context of features and can effectively avoid the effect of background clutter.
- Our proposed SGA module, in an end-to-end manner with the two-stream CNN structure, outperforms the state-of-the art approaches by large margins on two public RGB–IR Re-ID datasets. We also experimentally verified the improvement of the network by embedding different positions and numbers of SGA modules.
2. Related Works
2.1. RGB–RGB Re-ID
2.2. RGB–IR Re-ID Based on CNN Networks
2.3. RGB–IR Re-ID Based on GAN Networks
2.4. Attention Mechanisms
3. Our Method
3.1. Problem Description
3.2. The Proposed MFCNet
4. Results
4.1. Experimental Settings
4.2. Experimental Parameters
4.3. Comparison with State-of-the-Art Methods
5. Discussion
5.1. Ablation Experiment
- In the two modes, compared to the baseline method, MFCNet achieves an improvement in the rank-1 accuracy and mAP. Notably, the rank-1 accuracy increased by 5.76% and 7.56% between the proposed method and other methods. The results show that the SGA module designed in this paper can identify more discriminative features than can other methods to improve the hit rate.
- The spatial attention branch in SGA module helps improve network performance. In all-search and indoor-search modes, the spatial attention mechanism increases the hit rate for r = 1 by 2–5%. Specifically, the spatial attention mechanism can retain the key information form an image while ignoring noncritical information and interference;
- The global attention branch in SGA module yields a 3–5% improvement in the hit rate for r = 1 in the two modes. This result verifies that the global attention mechanism can mine feature context, which helps the network extract robust features and improve network performance.
- As shown in Table 1, Table 2 and Table 3, for the same method, indoor-search performs better than all-search. This is because images of indoor have less background variation and a more stable light condition and person pose, which makes matching easier. In the first row of the Figure 1, the first two images were taken indoors, and the second two were taken outdoors. We can see that the photos taken indoors are purer than those taken outdoors.
5.2. Supplementary Experiment
- Compared with the baseline (M1), the SGA module could improve the performance of the network. At r = 1, the results of M2-M10 are better than the baseline.
- The network of two SGA modules gives better results. For long-range dependency, two SGA modules can better preserve this relationship, which could mine the discriminative features.
- Experiments prove that the extracted features of the network proposed in this paper are more discriminative. The rank 1 accuracy reaches 51.64%. The framework in this paper is optimal.
5.3. Visualization Experiment
5.4. Comparison of Different Attention Mechanisms
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Zheng, L.; Yang, Y.; Tian, Q. SIFT meets CNN: A decade survey of instance retrieval. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 1224–1244. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Ye, M.; Liang, C.; Yu, Y.; Wang, Z.; Leng, Q.; Xiao, C.; Chen, J.; Hu, R. Person reidentification via ranking aggregation of similarity pulling and dissimilarity pushing. IEEE Trans. Multimed. 2016, 18, 2553–2566. [Google Scholar] [CrossRef]
- Liao, S.; Hu, Y.; Zhu, X.; Li, S.Z. Person re-identification by local maximal occurrence representation and metric learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
- Wei, L.; Zhang, S.; Gao, W.; Tian, Q. Person transfer gan to bridge domain gap for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
- Zheng, Z.; Yang, X.; Yu, Z.; Zheng, L.; Yang, Y.; Kautz, J. Joint discriminative and generative learning for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
- Hou, R.; Ma, B.; Chang, H.; Gu, X.; Shan, S.; Chen, X. Vrstc: Occlusion-free video person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
- Yu, H.; Zheng, W.; Wu, A.; Guo, X.; Gong, S.; Lai, J. Unsupervised person re-identification by soft multilabel learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
- Zhao, Y.; Shen, X.; Jin, Z.; Lu, H.; Hua, X.-S. Attribute-driven feature disentangling and temporal aggregation for video person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
- Wu, A.; Zheng, W.; Yu, H.; Gong, S.; Lai, J. Rgb-infrared cross-modality person re-identification. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
- Xiang, X.; Lv, N.; Yu, Z.; Zhai, M.; Saddik, A. Cross-Modality Person Re-Identification Based on Dual-Path Multi-Branch Network. IEEE Sens. J. 2019, 19, 11706–11713. [Google Scholar] [CrossRef]
- Chen, D.; Li, H.; Liu, X.; Shen, Y.; Shao, J.; Yuan, Z.; Wang, X. Improving deep visual representation for person re-identification by global and local image-language association. In Proceedings of the 15th European Conference on Computer Vision, ECCV 2018, Munich, Germany, 8–14 September 2018. [Google Scholar]
- Gong, Y.; Zhang, Y.; Poellabauer, C. Second-order non-local attention networks for person re-identification. In Proceedings of the International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019. [Google Scholar]
- Hou, R.; Ma, B.; Chang, H.; Gu, X.; Shan, S.; Chen, X. Interaction-and-aggregation network for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
- Wang, J.; Zhu, X.; Gong, S.; Li, W. Transferable joint attribute-identity deep learning for unsupervised person re-identification. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
- Fang, P.; Zhou, J.; Roy, S.; Petersson, L.; Harandi, M. Bilinear attention networks for person retrieval. In Proceedings of the International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019. [Google Scholar]
- Li, S.; Bak, S.; Carr, P.; Wang, X. Diversity regularized spatiotemporal attention for video-based person re-identification. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
- Liu, C.-T.; Wu, C.-W.; Wang, Y.-C.F.; Chien, S.-Y. Spatially and temporally efficient non-local attention network for video-based person re-identification. In Proceedings of the 30th British Machine Vision Conference 2019, BMVC 2019, Cardiff, UK, 9–12 September 2019. [Google Scholar]
- Ye, M.; Shen, J.; Lin, G.; Xiang, T.; Shao, L.; Hoi, S.C.H. Deep Learning for Person Re-Identification: A Survey and Outlook. Available online: https://arxiv.org/pdf/2001.04193 (accessed on 14 November 2021).
- Ye, M.; Shen, J.; Crandall, D.J.; Shao, L.; Luo, J. Dynamic Dual-Attentive Aggregation Learning for Visible-Infrared Person Re-identification. In Proceedings of the 2020 European Conference on Computer Vision, ECCV 2020, Glasgow, UK, 23–28 August 2020. [Google Scholar]
- Chen, L.; Zhang, H.; Xiao, J.; Nie, L.; Shao, J.; Chua, T. Sca-cnn: Spatial and channel-wise attention in convolutional networks for image captioning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
- Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-Excitation Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 2011–2023. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Jaderberg, M.; Simonyan, K.; Andrew Zisserman, A.; Kavukcuoglu, K. Spatial transformer networks. In Proceedings of the Neural Information Processing Systems, Montreal, QB, Canada, 7–12 December 2015. [Google Scholar]
- Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-local neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
- Ye, M.; Lan, X.; Wang, Z.; Yuen, P.C. Bi-Directional Center-Constrained Top-Ranking for Visible Thermal Person Re-Identification. IEEE Trans. Inf. Forensics Secur. 2020, 15, 407–419. [Google Scholar] [CrossRef]
- Hao, Y.; Wang, N.; Li, J.; Gao, X. HSME: Hypersphere Manifold Embedding for Visible Thermal Person Re-Identification. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019. [Google Scholar]
- Liu, H.; Cheng, J.; Wen, W.; Su, Y.; Bai, H. Enhancing the discriminative feature learning for visible-thermal cross-modality person re-identification. Neurocomputing 2020, 398, 11–19. [Google Scholar] [CrossRef] [Green Version]
- Zhao, Y.; Lin, J.; Xuan, Q.; Xi, X. HPLIN: A feature learning framework for cross-modality person re-identification. IET Image Process. 2020, 13, 2897–2904. [Google Scholar] [CrossRef]
- Zhu, Y.; Yang, Z.; Wang, L.; Zhao, S.; Hu, X.; Tao, D. Hetero-Center Loss for Cross-Modality Person Re-Identification. Neurocomputing 2020, 386, 97–109. [Google Scholar] [CrossRef] [Green Version]
- Feng, Z.; Lai, J.; Xie, X. Learning Modality-Specific Representations for Visible-Infrared Person Re-Identification. IEEE Trans. Image Process. 2020, 29, 579–590. [Google Scholar] [CrossRef] [PubMed]
- Lu, Y.; Wu, Y.; Liu, B.; Zhang, T.; Li, B.; Chu, Q.; Yu, N. Cross-Modality Person Re-Identification With Shared-Specific Feature Transfer. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
- Li, D.; Wei, X.; Hong, X.; Gong, Y. Infrared-Visible Cross-Modal Person Re-Identification with an X Modality. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020. [Google Scholar]
- Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the Advances in Neural Information Processing Systems 27 (NIPS 2014), Montreal, QB, Canada, 8–13 December 2014. [Google Scholar]
- Dai, P.; Ji, R.; Wang, H.; Wu, Q.; Huang, Y. Cross-Modality Person Re-Identification with Generative Adversarial Training. In Proceedings of the 2018 International Joint Conference on Artificial Intelligence, IJCAI 2018, Stockholm, Sweden, 13–19 July 2018. [Google Scholar]
- Wang, G.; Zhang, T.; Cheng, J.; Liu, S.; Yang, Y.; Hou, Z. Rgb-infrared cross-modality person re-identification via joint pixel and feature alignment. In Proceedings of the International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019. [Google Scholar]
- Wang, Z.; Wang, Z.; Zheng, Y.; Chuang, Y.; Satoh, S. Learning to Reduce Dual-Level Discrepancy for Infrared-Visible Person Re-Identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
- Choi, S.; Lee, S.; Kim, Y.; Kim, T.; Kim, C. Hi-CMD: Hierarchical Cross-Modality Disentanglement for Visible-Infrared Person Re-Identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
- Wang, G.; Yang, T.; Yang, Y.; Yang, T.; Cheng, J.; Chang, J.; Liang, X.; Hou, Z. Cross-Modality Paired-Images Generation for RGB-Infrared Person Re-Identification. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
- Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual Attention Network for Scene Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
- Park, J.; Woo, S.; Lee, J.; Kweon, I. BAM: Bottleneck attention module. arXiv 2018, arXiv:1807.06514. [Google Scholar]
- Park, J.; Woo, S.; Lee, J.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the 15th European Conference on Computer Vision, ECCV 2018, Munich, Germany, 8–14 September 2018. [Google Scholar]
- Cao, Y.; Xu, J.; Lin, S.; Wei, F.; Hu, H. GCNet: Non-local Networks Meet Squeeze-Excitation Networks and Beyond. In Proceedings of the International Conference on Computer Vision Workshop, Seoul, Korea, 27 October–2 November 2019. [Google Scholar]
- Filip, R.; Giorgos, T.; Ondrej, C. Fine-tuning CNN Image Retrieval with No Human Annotation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 41, 1655–1668. [Google Scholar]
- Zhang, S.; Yang, Y.; Wang, P.; Zhang, X.; Zhang, Y. Attend to the Difference: Cross-Modality Person Re-identification via Contrastive Correlation. arXiv 2019, arXiv:1910.11656. [Google Scholar] [CrossRef] [PubMed]
- Dalal, N.; Triggs, B. Histograms of Oriented Gradients for Human Detection. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–26 June 2005. [Google Scholar]
- Ye, M.; Lan, X.; Li, J.; Yuen, P.C. Hierarchical discriminative learning for visible thermal person re-identification. In Proceedings of the The Thirty-Second AAAI Conference on Artificial Intelligence (AAAI-18), New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
- Ye, M.; Lan, X.; Leng, Q.; Shen, J. Cross-modality person re-identification via modality-aware collaborative ensemble learning. IEEE Trans. Image Process. 2020, 29, 9387–9399. [Google Scholar] [CrossRef] [PubMed]
- Gao, G.; Shao, H.; Yu, Y.; Wu, F.; Yang, M. Leaning Compact and Representative Features for Cross-Modality Person Re-Identification. arXiv 2021, arXiv:2103.14210. [Google Scholar]
- Liang, W.; Wang, G.; Lai, J.; Xie, X. Homogeneous-to-Heterogeneous: Unsupervised Learning for RGB-Infrared Person Re-Identification. IEEE Trans. Image Process. 2021, 30, 6392–6407. [Google Scholar] [CrossRef] [PubMed]
- Wei, Z.; Yang, X.; Wang, N.; Gao, X. Flexible Body Partition-Based Adversarial Learning for Visible Infrared Person Re-Identification. IEEE Trans. Neural Netw. Learn. Syst. 2021, 1–12. [Google Scholar] [CrossRef] [PubMed]
- Chen, Y.; Wan, L.; Li, Z.; Jing, Q.; Sun, Z. Neural Feature Search for RGB-Infrared Person Re-Identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021. [Google Scholar]
- Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
- Ma, A.J.; Yuen, P.C.; Li, J. Domain transfer support vector ranking for person re-identification without target camera label information. In Proceedings of the 2013 IEEE International Conference on Computer Vision (ICCV), Sydney, Australia, 1–8 December 2013. [Google Scholar]
- Li, H.; Chen, Y.; Tao, D.; Yu, Z.; Qi, G. Attribute-Aligned Domain-Invariant Feature Learning for Unsupervised Domain Adaptation Person Re-Identification. IEEE Trans. Inf. Forensics Secur. 2021, 16, 1480–1494. [Google Scholar] [CrossRef]
- Li, H.; Dong, N.; Yu, Z.; Tao, D.; Qi, G. Triple Adversarial Learning and Multi-view Imaginative Reasoning for Unsupervised Domain Adaptation Person Re-identification. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, C2. [Google Scholar] [CrossRef]
- Zhu, Z.; Luo, Y.; Chen, S.; Qi, G.; Mazur, N.; Zhong, C.; Li, Q. Camera style transformation with preserved self-similarity and domain-dissimilarity in unsupervised person re-identification. J. Vis. Commun. Image Represent. 2021, 80, 103303. [Google Scholar] [CrossRef]
Settings | All-Search | Indoor Search | ||||||
---|---|---|---|---|---|---|---|---|
Methods | r = 1 | r = 10 | r = 20 | mAP | r = 1 | r = 10 | r = 20 | mAP |
HOG [45] | 2.76 | 18.3 | 32 | 4.24 | 3.22 | 24.7 | 44.6 | 7.25 |
LOMO [3] | 3.64 | 23.2 | 37.3 | 4.53 | 5.75 | 34.4 | 54.9 | 10.2 |
One-stream [9] | 12.04 | 49.68 | 66.74 | 13.67 | 16.94 | 63.55 | 82.1 | 22.95 |
Two-stream [9] | 11.65 | 47.99 | 65.5 | 12.85 | 15.6 | 61.18 | 81.02 | 21.49 |
Zero-padding [9] | 14.8 | 54.12 | 71.33 | 15.95 | 20.58 | 68.38 | 85.79 | 26.92 |
Tone [46] | 12.52 | 50.72 | 68.6 | 14.42 | 20.82 | 68.86 | 84.46 | 26.38 |
HCML [46] | 14.32 | 53.16 | 69.17 | 16.16 | 24.52 | 73.25 | 86.73 | 30.08 |
cmGAN [33] | 26.97 | 67.51 | 80.56 | 31.49 | 31.63 | 77.23 | 89.18 | 42.19 |
BDTR [24] | 27.32 | 66.96 | 81.07 | 27.32 | 31.92 | 77.18 | 89.28 | 41.86 |
HSME [25] | 20.68 | 32.74 | 77.95 | 23.12 | - | - | - | - |
D2RL [35] | 28.9 | 70.6 | 82.4 | 29.2 | - | - | - | - |
MAC [47] | 33.26 | 79.04 | 90.09 | 36.22 | 36.43 | 62.36 | 71.63 | 37.03 |
MSR [29] | 37.35 | 83.4 | 93.34 | 38.11 | 39.64 | 89.29 | 97.66 | 50.88 |
AlignGAN [34] | 42.4 | 85 | 93.7 | 40.7 | 45.9 | 87.6 | 94.4 | 54.3 |
HPLIN [27] | 41.36 | 84.78 | 94.31 | 42.95 | 45.77 | 91.82 | 98.46 | 56.52 |
AGW [15] | 47.5 | 84.39 | 92.14 | 47.65 | 54.17 | 91.14 | 95.98 | 62.97 |
LCRF [48] | 43.23 | 82.78 | 90.91 | 43.09 | 50.07 | 90.63 | 96.99 | 58.88 |
H2H [49] | 45.47 | 72.78 | 82.28 | 47.99 | - | - | - | - |
FBP-AL [50] | 54.14 | 86.04 | 93.03 | 50.20 | - | - | - | - |
Ours | 51.64 | 86.8 | 94.66 | 49.99 | 57.29 | 92.89 | 96.38 | 64.91 |
Settings | Visible to Infrared | |||
---|---|---|---|---|
Methods | r = 1 | r = 10 | r = 20 | mAP |
HOG [45] | 13.49 | 33.22 | 43.66 | 10.31 |
LOMO [3] | 0.85 | 2.47 | 4.10 | 2.28 |
Zero padding [9] | 17.75 | 34.21 | 44.35 | 18.90 |
HCML [46] | 24.44 | 47.53 | 56.78 | 20.08 |
BDTR [24] | 33.56 | 58.61 | 67.43 | 32.76 |
HSME [25] | 50.85 | 73.36 | 81.66 | 47.00 |
D2RL [35] | 43.4 | 66.1 | 76.3 | 44.1 |
MAC [47] | 36.43 | 62.36 | 71.63 | 37.03 |
MSR [29] | 48.43 | 70.32 | 79.95 | 48.67 |
EDFL [26] | 52.58 | 72.10 | 81.47 | 52.98 |
AlignGAN [34] | 57.9 | - | - | 53.6 |
Xmodal [31] | 62.21 | 83.13 | 91.72 | 60.18 |
H2H [49] | 62.27 | 77.56 | 83.72 | 61.90 |
Ours | 69.76 | 86.07 | 91.73 | 52.52 |
Settings | All-Search | Indoor Search | ||||||
---|---|---|---|---|---|---|---|---|
Methods | r = 1 | r = 10 | r = 20 | mAP | r = 1 | r = 10 | r = 20 | mAP |
B | 45.88 | 82.96 | 90.38 | 46.36 | 49.73 | 90.90 | 96.97 | 59.21 |
B+S | 48.51 | 84.41 | 91.69 | 46.12 | 54.30 | 90.62 | 97.10 | 61.10 |
B+G | 49.65 | 84.12 | 92.72 | 49.01 | 54.57 | 91.44 | 96.24 | 62.21 |
B+S+G | 51.64 | 86.80 | 94.66 | 49.99 | 57.29 | 92.89 | 96.38 | 64.91 |
Setting | All-Search | |||||||
---|---|---|---|---|---|---|---|---|
Design | stage1 | stage2 | stage3 | stage4 | r = 1 | r = 10 | r = 20 | mAP |
M1 | × | × | × | × | 45.88 | 82.96 | 90.38 | 46.36 |
M2 | √ | × | × | × | 46.23 | 82.16 | 90.14 | 45.39 |
M3 | × | √ | × | × | 46.70 | 83.64 | 90.66 | 46.47 |
M4 | × | × | √ | × | 47.15 | 84.72 | 92.53 | 46.75 |
M5 | × | × | × | √ | 47.75 | 85.46 | 93.29 | 44.73 |
M6 | √ | √ | × | × | 48.28 | 83.91 | 91.87 | 47.12 |
M7 | √ | × | √ | × | 48.46 | 84.30 | 92.06 | 44.95 |
M8 | √ | × | × | √ | 48.09 | 85.35 | 92.35 | 47.50 |
M9 | × | √ | √ | × | 51.64 | 86.80 | 94.66 | 49.99 |
M10 | × | √ | × | √ | 48.83 | 85.43 | 92.01 | 48.13 |
M11 | √ | √ | √ | × | 50.17 | 85.46 | 92.30 | 49.08 |
M12 | √ | × | √ | √ | 49.65 | 86.41 | 93.48 | 49.16 |
M13 | × | √ | √ | √ | 49.33 | 85.25 | 92.82 | 48.55 |
M14 | √ | √ | √ | √ | 49.22 | 84.30 | 91.11 | 48.66 |
Settings | All-Search | |||
---|---|---|---|---|
Methods | r = 1 | r = 10 | r = 20 | mAP |
SE | 46.98 | 82.68 | 90.82 | 46.76 |
CBAM | 47.35 | 84.54 | 91.87 | 46.97 |
Non-local | 47.50 | 84.39 | 92.14 | 47.65 |
SGA | 51.64 | 86.80 | 94.66 | 49.99 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Mei, J.; Xu, H.; Li, Y.; Bian, M.; Huang, Y. MFCNet: Mining Features Context Network for RGB–IR Person Re-Identification. Future Internet 2021, 13, 290. https://doi.org/10.3390/fi13110290
Mei J, Xu H, Li Y, Bian M, Huang Y. MFCNet: Mining Features Context Network for RGB–IR Person Re-Identification. Future Internet. 2021; 13(11):290. https://doi.org/10.3390/fi13110290
Chicago/Turabian StyleMei, Jing, Huahu Xu, Yang Li, Minjie Bian, and Yuzhe Huang. 2021. "MFCNet: Mining Features Context Network for RGB–IR Person Re-Identification" Future Internet 13, no. 11: 290. https://doi.org/10.3390/fi13110290
APA StyleMei, J., Xu, H., Li, Y., Bian, M., & Huang, Y. (2021). MFCNet: Mining Features Context Network for RGB–IR Person Re-Identification. Future Internet, 13(11), 290. https://doi.org/10.3390/fi13110290