Next Article in Journal
DEM Simulation and Experimental Validation of the Performance of an Orchard Deep Applicator for Manure
Next Article in Special Issue
Convolved Feature Vector Based Adaptive Fuzzy Filter for Image De-Noising
Previous Article in Journal
Displacement Analyses of Main Structure of Parallel Pit Excavation and Analysis of Countermeasures
Previous Article in Special Issue
A Real-Time Remote Respiration Measurement Method with Improved Robustness Based on a CNN Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CNN Attention Enhanced ViT Network for Occluded Person Re-Identification

Shanghai Marine Intelligent Information and Navigation Remote Sensing Engineering Technology Research Center, Shanghai Ocean University, Shanghai 201306, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(6), 3707; https://doi.org/10.3390/app13063707
Submission received: 5 February 2023 / Revised: 13 March 2023 / Accepted: 13 March 2023 / Published: 14 March 2023
(This article belongs to the Special Issue New Trends in Image Processing III)

Abstract

:
Person re-identification (ReID) is often affected by occlusion, which makes most of the features extracted by ReID models contain a lot of identity-independent noise. Recently, the use of Vision Transformer (ViT) has enabled significant progress in various visual artificial intelligence tasks. However, ViT suffers from insufficient local information extraction capability, which should be of concern to researchers in the field of occluded ReID. This paper conducts a study to exploit the potential of attention mechanisms to enhance ViT in ReID tasks. In this study, an Attention Enhanced ViT Network (AET-Net) is proposed for occluded ReID. We use ViT as the backbone network to extract image features. Even so, occlusion and outlier problems still exist in ReID. Then, we combine the spatial attention mechanism into the ViT architecture, by which we enhance the attention of ViT patch embedding vectors to important regions. In addition, we design a MultiFeature Training Module to optimize the network by the construction of multiple classification features and calculation of the multi-feature loss to enhance the performance of the model. Finally, the effectiveness and superiority of the proposed method are demonstrated by broad experiments on both occluded and non-occluded datasets.

1. Introduction

With the rise of digital cities, intelligent security projects cover numerous fields, and intelligent monitoring systems play a huge role in maintaining the convenience and security stability of society. Person re-identification (ReID) aims to correlate identities captured by different image sensor devices (like CCD or CMOS) using computer vision-related techniques. However, ReID also faces many problems with uncontrollable factors in complex environments, such as low-resolution of images, light variations, pose changes, camera viewpoint changes, occlusion and inaccurate detection making Re-ID still difficult to solve [1,2,3]. The various challenges are shown in Figure 1. Among them, local occlusions often occur in our real life, which significantly influences the accuracy of models for identity recognition [4,5,6]. Therefore, the construction of a ReID model, which can solve the occlusion problem, is crucial [7,8,9].
Based on a detailed observation of the occlusion situation, the types of occlusions can be classified into two types. The first type is external occlusion which does not belong to the human himself (e.g., a car or a railing). These can be easily separated. The second type is additional occlusion. These occlusions belong to the human (e.g., a backpack or umbrella) and change with the posture and viewing angle of the human. As a result, additional occlusions are difficult to separate from the humans themselves.
The current research on deep learning for ReID is mainly divided into two types: supervised learning and unsupervised learning [1], but unsupervised learning is not as effective as supervised learning in terms of recognition accuracy due to the limitation of model training [10,11]. However, on the other hand, for the scarcity and hard-to-label ReID datasets, some researchers have done in-depth research on data sampling, weakly supervised learning and unsupervised learning [12,13,14]. In recent years, with the help of convolutional neural networks, supervised learning models have been broken through in solving the occlusion ReID problem [2,3,4,5,6,7,8,9]. Some researchers extracted local regions of humans by combining with target segmentation [15,16,17] and semantic analysis [2,7,18,19,20,21] models to accomplish identity matching, but the introduction of additional models made the cost of occluded ReID much higher. Additionally, other researchers also match detailed information by region segmentation, but such an approach does not work well in the case of occlusion [15,18,22,23]. Because of the randomness, occlusions make these works often fail to match during model identification. Furthermore, lots of works have introduced attention mechanisms into the ReID model in order to filter more significant parts from a huge amount of information, aiming to focus the attention of the model to identify information and ignore secondary information [5,15,16,24,25,26].
It has been studied that CNNs use convolution to limit the extraction of image features by the model to small regions, and the use of down-sampling modules would tend to lose the local details of humans [24]. In recent years, the Transformer, popular in Natural Language Processing (NLP) [27,28], was introduced to the ReID task, achieving comparable performance to CNN-based methods with the help of a self-attention mechanism [5,24,29]. However, ViT suffers from insufficient local information extraction capability [30,31], which should be concerning for researchers in the field of occluded ReID.
We hope to improve ViT’s ability to focus on important areas by incorporating CNN’s spatial attention mechanism. This method uses the convolution process to extract local importance distribution, similar to how the CNN attention mechanism enhances attention to important information. In addition, the channel attention of CNN can extract the channel distribution state of the current feature vector, and it can enhance the attention of the model to the important channel features according to the backpropagation algorithm. Based on the above, this work proposed an Attention Enhanced ViT Network (AET-Net) to address the challenges of ReID occlusion, the general procedure is shown in Figure 2. Additionally, the specific methodology is presented in Section 3, following the related works of Section 2. Thereafter the experimental settings and results are shown in Section 4. The paper’s contributions are summarised as follows:
  • We proposed a ViT-based ReID framework, named AET-Net, to address the occlusion problem.
  • Spatial Attention Enhanced Module (SAEM) is designed to strengthen the ViT patch embedding structure to enhance its attention to important local features of the image.
  • MultiFeature Training Module (MFTM) exploits multiple losses to optimize the model and also to keep the model from over-biasing the attention feature.
  • AET-Net achieves superior performance on the occluded dataset and competitive performance on the non-occluded datasets.

2. Related Work

2.1. Occluded Person ReID Based on Deep Learning

Occluded ReID aims to find the unoccluded region in the image for identity matching, and the incomplete information makes this task more of a challenge. However, most of the ReID models are based on the existing complete human information in the image and are highly sensitive to the occlusion situation [16,17,20,23,32,33]. We divide the occluded ReID models into two types: partial models and attention models. The process of these two types of models is shown in Figure 3.

2.1.1. Partial Models

Robust local features have been demonstrated to help occluded ReID [2,16,18,20,22]. Previous methods mainly construct local regions to extract local features by human segmentation, human semantic analysis, and horizontal stripe segmentation. Song et al. [16] designed the MGCAM, which uses a binary segmentation mask to segment human regions. Models for solving occlusion like PDC [20], HOReID [2], PGFA [7], and Spindle Net [18] are used to construct local regions by generating human local key points through semantic analysis models. Sun et al. [22] proposed PCB that divides human images into multiple horizontal local regions from top to bottom. In addition, the alignment of the human body has also been shown to improve the accuracy of the model [6,7,9,34,35,36]. However, the accuracy of these strategies is affected by the division method and pre-trained models.

2.1.2. Attention Models

Attention mechanisms have been studied in progressive deep learning research that attempts to focus the model’s attention on locally important information, which plays the role of enhancing recognition features and suppressing irrelevant features [37]. A large number of studies have introduced it into ReID tasks and achieved excellent performance. Li et al. [38] proposed a lightweight attention network, HA-CNN, which learns both hard regional attention and soft pixel attention in human images to obtain invariant feature representations. Cai et al. [15] proposed MMGA in order to extract discriminable features in human images by learning global and local attention jointly. Zhang et al. [25] proposed a powerful module for Relation-aware Global Attention (RGA) that strengthens the discriminative ability of features themselves by extracting feature-to-feature relationships. Affected by Transformer in the NLP field, He et al. [24] proposed TransReID, which is the first Transformer-based framework for ReID. High performance was obtained with the latter on several public datasets, including occluded datasets. Jia et al. [5] proposed a DRL-Net, which also uses the Transformer architecture to achieve advanced performance on occluded ReID datasets.

2.2. Transformer in Computer Vision

The performance of the Transformer in NLP has brought a huge revolution to computer vision [39,40,41,42,43,44,45]. Dosovitskiy et al. [39] propose a pure Transformer model referred to as a Vision Transformer (ViT), which solves the input problem of image classification and it can extract the hidden relationships between sequence vectors through self-attention achieve. It has advanced performance on several image classification benchmarks. However, in the process of patch embedding, local feature extraction in ViT is to use convolution, which does not have strong robustness [30,31]. To solve this problem, researchers have made attempts to integrate the advantages of Transformer architecture with CNN structure and proposed various methods to enhance the local feature extraction capability of ViT. It has been shown that Transformer can strengthen the CNN convolution performance (VT [40], BoT-Net [41]). In addition, DEIT [43], ConViT [44], and ResT [45] also demonstrate that Transformer architectures can be enhanced by CNN structures to achieve more advanced performance.
Different from the existing ReID model, our proposed AET-Net attempts to use the Transformer for feature learning of ReID and constructs CNN attention modules applicable to ViT for improving the robustness and classification ability of ViT global features by spatial enhancement and multi-loss optimization.

3. Method

With this section, the proposed framework for AET-Net is introduced, including the ViT [39] architecture as a backbone network and the SAEM to enhance the ViT concentrating on more relevant regions for the ReID task. As well, MFTM is devised by utilizing channel attention for the construction of multi-loss, thus further optimizing the network. In addition, MFTM allows the model not to be overly biased towards attentional features. These two modules are to be trained jointly through an end-to-end method. The overall architecture of AET-Net is shown in Figure 4.

3.1. Feature Extraction Backbone

This section introduces the backbone network of AET-Net based on ViT, an architecture that is the basis for extracting reliable global features of humans. The utilization of ViT is inspired by the nature of strong global correlation in Transformer, its success in NLP and applicable computer vision. Based on the above ideas and inspired by the work of He [24], we use ViT as the backbone network of AET-Net, which consists of four main components.
Patch Embedding: Given a human image vector x i n R H × W × C where H, W, and C denote the height, width, and number of image channels, respectively. To solve the input problem of the Transformet in computer vision, ViT proposes patch embedding, which uses a fixed-size K × K convolution kernel to partition the image into N = HW / K 2 non-overlapping image blocks and embed them as feature sequence vectors f = { f 1 , f 2 , , f N } into the Transformer Encoder. The process of patch embedding is shown in the green area of Figure 4.
Class Token and Positional Embedding: ViT adds an additional learnable embedding class token noted as classification representation before the feature sequence vector, denoted by f c l s , which will be used to represent the state of the transformer encoder output. The class token is shown as [cls] in the blue part of Figure 4. Furthermore, to keep the image location information, a learnable bilinear 2D interpolation position embedding v p o s is applied to ViT. The positional embedding is shown in the lower right of the blue area in Figure 4. The process of obtaining the standard input vectors of the Transformer Encoder by positional embedding can be expressed as:
f i n = c a t ( f c l s ; f ) + v p o s ,
where f i n R ( N + 1 ) × D denotes the standard input vectors of the Transformer Encoder, D denotes the convolutional kernel dimension of patch embedding, and c a t ( ) denotes the concatenating operation.
Transformer Encoder: A transformer encoder is composed of L transformer layers joined in a series; each transformer layer comprises a multi-headed self-attention layer (MSA) and an MLP layer.
Supervision Learning: Similar to the general ReID framework, Cross-Entropy loss and Triplet loss are used by AET-Net to guide the learning of the entire network [5,10,22,35,36], but the difference is that AET-Net constructs attentional features from channel attention, with multi-feature corresponding to losses, and the details of how the different losses are calculated are in Section 3.4.

3.2. Spatial Attention Enhancement Module

Although ViT can obtain more robust global features, its approach of obtaining sequence vectors through patch embedding would have an insufficient focus on local regions and remain challenging in the case of occluded ReID. The SAEM is inspired by the CNN attention mechanism, which aims to enhance ViT’s attention to important local information by strengthening the regional information of image feature vectors after patch embedding [46]. Specifically, we generate a 2D spatial attention map produced by exploiting the spatial relationship between the sequence vectors of the patch embedding, and this 2D spatial attention map emphasizes “where” the given image is important. The detailed process of SAEM is shown in Figure 5a.
To calculate the spatial attention map, in the first place, we reshape the sequence vector map f R B × N × D after patch embedding in N dimensions as f R B × H K × W K × D , where B is the number of images inferred from one batch, K is the convolutional size of patch embedding, N = H W / K 2 is the number of sequence vector, and f is denoted as a generalization of the information of N × K × K regions. Second, average pooling and maximum pooling are performed separately along the channel dimension, and the results of both are concatenated to produce a feature descriptor. Along the channel dimension, pooling has been shown to be effective in highlighting local information [22]. A convolution is used on the feature descriptors to generate spatial features corresponding to the f channel dimension. Finally, the spatial feature is put into a sigmoid activation function to map into a probability distribution for obtaining a spatial attention map. The process of applying the spatial attention map to the sequence vector can be simply defined as:
f s a = s i g m o i d ( S A ( f ) ) × f ,
where f s a denotes the spatial attention-enhanced feature, s i g m o i d ( ) denotes the probability mapping function, and S A ( ) denotes the generation process of spatial attention features, the specific generation process of which can be defined as:
S A ( x ) = C o v 7 × 7 ( c a t ( A v g P o o l ( x ) ; M a x P o o l ( x ) ) ) ,
where A v g P o o l ( ) and M a x P o o l ( ) denote the average pooling and maximum pooling, respectively, c a t ( ) denotes the splicing operation, and C o v 7 × 7 denotes the 7 × 7 convolution process.
The feature vector f s a generated by SAEM is a weighted sum with the original feature vector f by the weighted hyperparameter λ ; this process can be defined as:
f s a = λ × f s a + ( 1 λ ) × f ,
where f s a denotes the weighted vector. It’s going to be reshaped as B × N × D and spliced with the classification representation f c l s . The splicing result and the position embedding are added and input to the Transformer encoder. In the ablation study, we analyzed the weighted hyperparameters λ . The overall process of SAEM is shown in the blue part of Figure 4.

3.3. MultiFeature Training Module

The ViT has gained popularity for its unique global feature extraction, but its fixed training mode of using individual global features for classification brings challenges compared with the flexible and versatile characteristics of CNN. It has been studied that multi-feature of deep learning can correspond to the same label, and combining multiple features for the model training could help to improve its effectiveness [1,22,47].
We construct a MultiFeature Training Module (MFTM) for ViT to improve the effectiveness of AET-Net. As shown in the upper half of the blue, orange, and gray areas in Figure 4. The original ViT directly uses the classification representation g c l s R 1 × C after Transformer Encoder as the global features, but directly using them for classification will inevitably lose some information of the remaining N encoding vectors. In detail, MFTM takes into consideration two characteristics of the global feature. First, the dimension of the global feature from spatial is 1. At the same time, it is strongly representative of the global feature as it integrates information from all channels in the entire image. However, when the occlusion exists, a large amount of occlusion and complex background is blended in this feature, which is unfavorable for the identity recognition of the occlusion condition. In other words, although the global features of the Transformer Encoder are highly representative, too much noise is involved in its channel dimension. In contrast to spatial attention, channel attention aims to emphasize the “what” is important for a given image. The proposed MFTM conducts its training process by constructing attention global features jointly with the original ViT global features, and the process of attention global features construction is shown in Figure 5b. It is divided into two stages: computation of channel attention map and generation of global attention features.
For the calculation of the channel attention map, we first perform spatial compression on the reshaped vector f after patch embedding; two feature descriptors are generated by aggregating spatial information using mean and maximum pooling. Next, two feature descriptors are fed into a shared multilayer perceptron to generate two channel features, and then the two channel features are combined into one whole channel feature. Finally, the synthesized channel features are fed into a sigmoid function which yields a channel attention map. The calculation process of a channel attention map is represented as follows:
w c a = s i g m o i d ( C A ( f ) ) ,
where w c a denotes the generated channel attention map, f denotes the reshaped vector f after patch embedding, and C A ( ) denotes the extraction process of channel attention. The detailed process of C A ( ) can be expressed as follows:
C A ( x ) = M L P ( A v g P o o l ( x ) ) + M L P ( M a x P o o l ( x ) ) ,
where M L P ( ) denotes a multilayer perceptron shared by two feature descriptors. In this way, the global features of the Transformer encoder output will be applied to the channel attention map generating the attention global feature. This process can be simply defined as:
g a t t n = w c a × g ,
where g a t t n denotes the attention global feature and g denotes the global feature output by Transformer Encoder. The generated g a t t n and g further calculate Cross-Entropy loss and Triplet loss, respectively. Some studies have shown that the use of the attention mechanism may cause the model to ignore the attention to some important secondary information [25,28,37]. To make the model not overly biased to the attention global feature, we introduce a weight hyperparameter δ to suppress the attention global feature, including its losses. Specifically, during training, δ is used to balance the losses generated by the attention global feature g a t t n and the global feature g . During testing, δ is used to balance g a t t n and g , which can be defined as follows:
z = δ × z a t t n + ( 1 δ ) × z ,
where z denotes the balanced parameter term, δ denotes the weight hyperparameter, which has been extensively investigated in our ablation studies, and z a t t n denotes the attention parameter term, which is generally the attention global feature or its output losses. z denotes the non-attention parameter term, which is generally the global feature or its losses.

3.4. Loss Functions

Similar to the regular ReID task, Cross-Entropy loss and Triplet loss were constructed to optimize the network. Unlike that, we construct two kinds of losses for both the attention global feature and the global feature (as shown in the gray part of Figure 4).

3.4.1. SoftMax Cross-Entropy Loss

For the global feature g and the attention global feature g a t t n , we calculate the Cross-Entropy Loss L I D and L I D _ a t t n for the two features, respectively. The process of Cross-Entropy loss calculation is defined as:
L I D = i y ( i ) log p ( i ) ,
where y ( i ) denotes the real label of the i-th image; p ( i ) denotes the identity prediction probability distribution of the i-th image generated by two global features through their respective fully connected layers. The weights of L I D and L I D _ a t t n are controlled by the weight hyperparameter δ , as shown in (8).

3.4.2. Triplet Loss

The triplet loss was constructed so that on hard samples, the model performed better, and we used the general supervised learning process to construct triplet losses L t r i and L t r i _ a t t n , by using the global feature and attention global feature, respectively. Triplet loss is generally able to close the in-class spacing and widen the out-of-class spacing. More precisely, the process of building mixed feature Triplet loss with soft margin is given by:
L t r i = L n [ 1 + exp ( f A f P 2 2 f A f N 2 2 ) ] ,
where L t r i denotes L t r i or L t r i _ a t t n . f A , f P , and f N denote the anchor sample features, the positive sample features, and the negative sample features, respectively. These feature vectors are derived from the global feature g and the attention global feature g a t t n .

3.4.3. Total Loss

For recognizing identities, the help of the important secondary regions is sometimes necessary, and the introduction of attention may make the model overly dependent on the most important local region. Therefore, we introduce the weight hyperparameter φ to improve the constraints on the attention global feature. AET-Net’s total loss is defined as:
L t o t a l = φ × L I D _ t o t a l + ( 1 φ ) × L t r i _ t o t a l ,
wherein φ is the weighting parameter used to balance the two types of losses, which is set as 0.5 in the experiment. L I D _ t o t a l and L t r i _ t o t a l denote the balanced total Cross-Entropy loss and total Triplet loss, respectively. AET-Net trains the network end-to-end by minimizing L t o t a l .

4. Results and Discussion

4.1. Datasets and Evaluation Protocols

4.1.1. Datasets

Our experiments were performed on the three public ReID datasets Market-1501 [48], DukeMTMC-ReID [49], and Occluded-DukeMTMC [7]; details of these datasets are given in Table 1.
The Market-1501 dataset consists of 32,668 images from 6 cameras of 1501 people, with 12,936 images from 751 people serving as the training set and 19,732 images from another 750 people serving as the test set.
The DukeMTMC-ReID dataset contains 36,411 images from 1812 humans captured by eight cameras. A total of 16,522 images from 702 humans were randomly selected as the training set, while another 19,889 Images from 702 humans were used as the test set.
The Occluded-Duke dataset was used to study occluded ReID, and it is a subset of the DukeMTMC-ReID. In contrast to other datasets, Occluded-Duke is selected from DukeMTMC with 9%/100%/10% occluded images, according to train/query/gallery, respectively.

4.1.2. Evaluation Protocols

The model’s performance is assessed using standard evaluation metrics, such as mean average precision (mAP) [48] and cumulative matching characteristics (CMC) [50] at Rank-1, which are commonly used in general ReID methods.
The performance of the ReID model can be comprehensively evaluated by the mean average precision (mAP), which measures how well the model places the correct predicted results of a query image at the top of the retrieved list. To calculate mAP, the average precision (AP) of the model for each identity prediction result must first be computed. The AP calculation process can be described as follows:
A P i = i = 1 N p i I i ,
where P i represents the count of images that contain the i-th identity target, I i denotes the number of images containing the i-th identity target, and N represents the total number of identities.
The mAP is the average of AP scores weighted by the number of images containing each identity, and this process can be mathematically represented as:
m A P = i = 1 N A P i N ,
where i = 1 N A P i denotes the average precision of each identity.
In general, the AP metric evaluates the performance of a learned model on a single category, while the mAP metric evaluates the model’s overall performance across all identity targets.
Rank-n is a frequently used evaluation metric in ReID, which is derived from the cumulative distribution curve (CMC). The value of n denotes a particular situation, which involves ranking the metric distance between the query set features and the gallery set features and measuring the accuracy of the query’s correct result ranked before the n-th position. This means that the identity of the query set is found in the gallery set with the most similar image of the human person. In a ReID task, the value of n is usually set to 1, 5, 10, or 20. In this particular paper, the value of n is set to 1.

4.2. Implementation Details

Data Augmentation: During the training and inference phases, all images are resized to a fixed size. In addition, training images are augmented with random horizontal flips, random cropping, and random erasures, with a probability of 0.5 for each.
Backbone: AET-Net is done under the Pytorch version 1.8.0 deep learning framework, the patch size is set to 16 × 16 , and the convolutional kernel depth D of patch embedding is set to 768. The ViT initial weights are obtained after being first pre-trained on ImageNet21K and then finetuned on Image-1K. Note that label smoothing is not used.
Training: The experiments were conducted on an Nvidia GeForce 3090 GPU, with a batch size of 64 and 4 instances, and a total training time was 69 min and 58 s. In order to compare with the ViT-based baseline model TransReID, the training was carried out for 120 epochs, using SGD as the optimizer, the initial learning rate was 0.008, and the cosine learning rate schedule was used for decay.

4.3. Comparison with State-of-the-Art Methods

In order to showcase the effectiveness of our supervised learning ReID model, AET-Net, we performed comparisons with current dominant models on three public datasets, which are divided into partial models and attention models. Although several state-of-the-art models utilize the Attention mechanism or Transformer architecture, there is no integration of both for ReID models, which are still more sensitive to the occlusion problem.

4.3.1. Results for the Non-Occluded Dataset

In Table 2, we show the results of AET-Net and other models on two non-occluded datasets, Market-1501 and DukeMTMC. AET-Net achieved an 87.5% mAP and a 94.8% Rank-1 accuracy on the Market-1501 dataset. It surpassed the mAP of partial models in the table, which are considered state-of-the-art, and also demonstrated superior performance in Rank-1 compared to most models. On DukeMTMC, AET-Net also performed fairly well, reporting 80.1% mAP and 89.5% Rank-1 accuracy. Exceeding all state-of-the-art models in the table, as well, there is a +0.8%/+3.5% mAP and +0.7%/+1.4% Rank-1 improvement compared to TransReID and DRL-Net. This indicates the effectiveness of AET-Net in solving the ReID task. This slight improvement is due to the attention mechanism that enhances the ViT’s ability to extract local details.

4.3.2. Results for the Occluded Dataset

The results of AET-Net on the Occluded-Duke dataset are presented in Table 3, along with those of other classic models for comparison. The Occluded-Duke images have more background clutter, occlusions, and other confusing features, which lead to lower accuracy for all models. However, the proposed AET-Net baseline model reported 54.5% mAP and 64.5% Rank-1 accuracy, outperforming the best partial model HOReID by 10.7% mAP and 9.4% Rank-1 and exceeding the best attention model PAT by 0.9% in mAP. In particular, AET-Net shows certain improvements compared to TransReID, DRL-Net, and PAT based on Transformer architecture. These outcomes suggest that our proposed AET-Net has improved the recognition performance and classification ability of the model.

4.4. Ablation Study and Visualization

We performed a comprehensive ablation study on Market-1501 and Occluded-Duke to verify the efficacy of each component in AET-Net and to design the optimal architecture. In order to ensure impartial comparison, none of the models used rearrange. The results of the ablation study are presented in Table 4.

4.4.1. Transformer Architecture

Table 4 presents a comparison of the results between the ViT-based ReID baseline model and the CNN-based model on the Market-1501 and Occluded-Duke. The results show that the ViT-based model achieves comparable performance to the CNN-based model on the non-occluded datasets. However, on the occluded dataset, the ViT-based model far outperforms the CNN-based model. From these results, it can be concluded that using the Transformer architecture to extract potential relationships among the local region of human targets would help to address the occlusion challenge in the ReID task.

4.4.2. Spatial Attention Enhancement Module

We conducted several ablation studies on the SAEM of AET-Net to verify its efficacy. To explore the effect of the weighted hyperparameter λ on SAEM, we gradually increased the parameter λ on the Occluded-Duke dataset to enhance the effect of the spatial attention feature, and the results are shown in Figure 6a. SAEM performance is best on the Occluded-Duke dataset when λ takes the value of 0.9. The results in SAEM ablation studies on the Market-1501 and Occluded-Duke are illustrated in Table 5, corresponding to index-1 to index-3, and it can be seen from the results that under both the occluded and non-occluded datasets, SAEM shows a slight performance improvement. The results testify that the global feature classification ability extracted by the model is improved under the effect of spatial attention.

4.4.3. Multi-Feature Training Module

In the same way as that of SAEM, we conduct the ablation study on MFTM for AET-Net to verify its validity. The analysis results of the balance parameters δ on Occluded-Duke are shown in Figure 6b. MFTM performs best on the Occluded-Duke dataset when δ takes the value of 0.6. Table 5 shows the results of the ablation study on MFTM, corresponding to index-4 and index-5. This demonstrates that MFTM provides better performance compared to SAEM, leading to a more significant improvement in the model’s performance.

4.4.4. Integration of the Two Modules

To verify that the two attention modules together can better improve the performance of the ViT architecture, we conduct an experiment by uniting the two attention modules, and the performance outcomes are presented in Table 5, corresponding to index-6 and index-7. It is shown that the performance under the integrated action of the two attention modules exceeds that of the individual module. In particular, the enhancement effect is higher on the occluded dataset compared with the non-occluded dataset.

4.4.5. Inferential Costs

We compute the inference time, computation and number of parameters of AET-Net by the following experiments. Furthermore, to decrease the inference time error resulting from the difference in memory IO of the device, the input for the model tests were all set to 64 × 256 × 128 . The results of this analysis are shown in Table 6. To make the comparison more explicit, the inference cost of ResNet50 is expressed as 1x.
The results show that the addition of SAEM accelerates the inference speed of the model, although the effect was not significant, which provides a feasible way of Transformer based models’ acceleration. The attention model has a slightly higher cost in terms of flops and parameters compared to the baseline model, owing to the inclusion of additional operations. Compared with SAEM, MFTM requires more complex computations and network nodes for refinement operations in the channel dimension, which is more costly. Finally, the attention enhancement models have the advantage of being lightweight compared to the Jigsaw Patch Module in TransReID.

4.4.6. Visualization

We conducted feature map visualization to further demonstrate the recognition performance of AET-Net. As shown in Figure 7, four humans’ images from the Occluded-Duke dataset, which are no occlusion, external occlusion, additional occlusion, and multiple occlusions, illustrate the attention of the model to the images in different scenes.
The visualization results demonstrate that the CNN-based model (ResNet50) distributes the perceptual fields mostly in small regions using convolution. In contrast, the self-attention mechanism in ViT allows it to capture inter-region dependencies, enabling the model to focus on multiple regions of the image. However, ViT is limited by patch embedding, and the model has a lot of attention scattered in the occluded regions or background regions, which is not adequate for a ReID task. The added SAEM changes the patch embedding state of ViT, which gives the model more attention in the regions of identity discrimination. In addition, MFTM by fusing different features, so the model’s attention will be distributed in a great number of discriminable regions of the image. Moreover, the joint application of SAEM and MFTM will integrate the two module features so that the model attention is heavily concentrated on the human regions. The results show that the AET-Net indirectly improves the extraction of local features by enhancing the model’s attention to the identification regions.

5. Conclusions

In this work, we propose a CNN attention-enhanced ViT network for occluded person re-identification. Firstly, the AET-Net uses the Vision Transformer as the backbone network for feature extraction, which can extract the implied relationships between regions, and this improves mAP accuracy by 9.3% and Rank-1 accuracy by 5.4% on the Occluded-Duke dataset compared to the CNN-based model HoReID. Secondly, we designed the spatial attention enhancement module to make the ViT model more focused on locally important regions. Additionally, we design the multi-feature training module to guide channel attention by computing multiple losses, which can reduce the model’s over-bias on attention features. We have conducted extensive studies on three benchmark datasets, and the proposed model outperforms the baseline model TransReID by 0.7% of mAP on the non-occluded dataset Market-1501. The experimental results show that AET-Net has good performance on occluded ReID datasets, exceeding the baseline model by 1.4% for mAP and 4% for Rank-1. This work provides a new idea for transformer-based ReID tasks.

Author Contributions

J.W. and P.L. conceived and designed the framework of the study. R.Z. (Rongfeng Zhao) completed the data collection and processing. R.Z. (Ruyan Zhou) and Y.H. completed the data analysis. J.W. and P.L. completed the algorithm design and model construction and were the lead authors of the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

The work was funded by the National Natural Science Foundation of China, grant/award number: 61806123 and the National Key R&D Program of China, grant/award number: 2019YFD0900805.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All data in this paper are public datasets that can be downloaded by all researchers through search engines.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ye, M.; Shen, J.; Lin, G. Deep learning for person re-identification: A survey and outlook. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 2872–2893. [Google Scholar] [CrossRef] [PubMed]
  2. Wang, G.A.; Yang, S.; Liu, H. High-order information matters: Learning relation and topology for occluded person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 6449–6458. [Google Scholar]
  3. Wang, P.; Ding, C.; Shao, Z.; Hong, Z.; Zhang, S.; Tao, D. Quality-aware part models for occluded person re-identification. IEEE Trans. Multimed. 2022. [Google Scholar] [CrossRef]
  4. Huang, H.; Li, D.; Zhang, Z. Adversarial occluded samples for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 5098–5107. [Google Scholar]
  5. Jia, M.; Cheng, X.; Lu, S. Learning disentangled representation implicitly via transformer for occluded person re-identification. IEEE Trans. Multimed. 2022. [Google Scholar] [CrossRef]
  6. Jia, M.; Cheng, X.; Zhai, Y. Matching on sets: Conquer occluded person re-identification without alignment. Proc. AAAI Conf. Artif. Intell. 2021, 35, 1673–1681. [Google Scholar] [CrossRef]
  7. Miao, J.; Wu, Y.; Liu, P. Pose-guided feature alignment for occluded person re-identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 542–551. [Google Scholar]
  8. Li, Y.; He, J.; Zhang, T. Diverse part discovery: Occluded person re-identification with part-aware transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 2898–2907. [Google Scholar]
  9. He, L.; Wang, Y.; Liu, W. Foreground-aware pyramid reconstruction for alignment-free occluded person re-identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8450–8459. [Google Scholar]
  10. Wang, H.; Lu, J.; Pang, F. Bi-directional Style Adaptation Network for Person Re-Identification. IEEE Sens. J. 2022, 22, 12339–12347. [Google Scholar] [CrossRef]
  11. Hu, Z.; Hou, W.; Liu, X. Deep Batch Active Learning and Knowledge Distillation for Person Re-Identification. IEEE Sens. J. 2022, 22, 14347–14355. [Google Scholar] [CrossRef]
  12. Wu, Y.; Lin, Y.; Dong, X. Exploit the unknown gradually: One-shot video-based person re-identification by stepwise learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 5177–5186. [Google Scholar]
  13. Wu, Y.; Lin, Y.; Dong, X. Progressive learning for person re-identification with one example. IEEE Trans. Image Process. 2019, 28, 2872–2881. [Google Scholar] [CrossRef]
  14. Lin, Y.; Wu, Y.; Yan, C. Unsupervised person re-identification via cross-camera similarity exploration. IEEE Trans. Image Process. 2020, 29, 5481–5490. [Google Scholar] [CrossRef]
  15. Cai, H.; Wang, Z.; Cheng, J. Multi-scale body-part mask guided attention for person re-identification. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Long Beach, CA, USA, 16–17 June 2019; pp. 1555–1564. [Google Scholar] [CrossRef] [Green Version]
  16. Song, C.; Huang, Y.; Ouyang, W. Mask-guided contrastive attention model for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1179–1188. [Google Scholar]
  17. Liu, M.; Yan, X.; Wang, C. Segmentation mask-guided person image generation. Appl. Intell. 2020, 51, 1161–1176. [Google Scholar] [CrossRef]
  18. Zhao, H.; Tian, M.; Sun, S. Spindle net: Person re-identification with human body region guided feature decomposition and fusion. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1077–1085. [Google Scholar]
  19. Su, C.; Li, J.; Zhang, S. Pose-driven deep convolutional model for person re-identification. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 3960–3969. [Google Scholar]
  20. Kalayeh, M.M.; Basaran, E.; Gökmen, M. Human semantic parsing for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1062–1071. [Google Scholar]
  21. Zhu, K.; Guo, H.; Liu, Z. Identity-guided human semantic parsing for person re-identification. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 346–363. [Google Scholar]
  22. Sun, Y.; Zheng, L.; Yang, Y. Beyond part models: Person retrieval with refined part pooling (and a strong convolutional baseline). In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 480–496. [Google Scholar]
  23. Wang, G.; Yuan, Y.; Chen, X. Learning discriminative features with multiple granularities for person re-identification. In Proceedings of the 26th ACM International Conference on Multimedia, Seoul, Republic of Korea, 22–26 October 2018; pp. 274–282. [Google Scholar]
  24. He, S.; Luo, H.; Wang, P. Transreid: Transformer-based object re-identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 15013–15022. [Google Scholar]
  25. Zhang, Z.; Lan, C.; Zeng, W. Relation-aware global attention for person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3186–3195. [Google Scholar]
  26. Chen, B.; Deng, W.; Hu, J. Mixed high-order attention network for person re-identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 371–381. [Google Scholar]
  27. Vaswani, A.; Shazeer, N.; Parmar, N. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  28. Galassi, A.; Lippi, M.; Torroni, P. Attention in natural language processing. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4291–4308. [Google Scholar] [CrossRef]
  29. Zhao, J.; Wang, H.; Zhou, Y. Spatial-Channel Enhanced Transformer for Visible-Infrared Person Re-Identification. IEEE Trans. Multimed. 2022. [Google Scholar] [CrossRef]
  30. Khan, S.; Naseer, M.; Hayat, M. Transformers in vision: A survey. ACM Comput. Surv. 2022, 54, 1–41. [Google Scholar] [CrossRef]
  31. Liu, Y.; Zhang, Y.; Wang, Y. A survey of visual transformers. arXiv 2021, arXiv:2111.06091. [Google Scholar]
  32. Zhang, S.; Zhang, Q.; Yang, Y. Person re-identification in aerial imagery. IEEE Trans. Multimed. 2020, 23, 281–291. [Google Scholar] [CrossRef] [Green Version]
  33. Yang, W.; Huang, H.; Zhang, Z. Towards rich feature discovery with class activation maps augmentation for person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1389–1398. [Google Scholar]
  34. Zhao, L.; Li, X.; Zhuang, Y. Deeply-learned part-aligned representations for person re-identification. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 3219–3228. [Google Scholar]
  35. He, L.; Liang, J.; Li, H. Deep spatial feature reconstruction for partial person re-identification: Alignment-free approach. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7073–7082. [Google Scholar]
  36. Zhang, X.; Luo, H.; Fan, X. Alignedreid: Surpassing human-level performance in person re-identification. arXiv 2017, arXiv:1711.08184. [Google Scholar]
  37. Guo, M.H.; Xu, T.X.; Liu, J.J. Attention mechanisms in computer vision: A survey. Comput. Vis. Media 2022, 8, 331–368. [Google Scholar] [CrossRef]
  38. Li, W.; Zhu, X.; Gong, S. Harmonious attention network for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 2285–2294. [Google Scholar]
  39. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  40. Wu, B.; Xu, C.; Dai, X. Visual transformers: Token-based image representation and processing for computer vision. arXiv 2020, arXiv:2006.03677. [Google Scholar]
  41. Srinivas, A.; Lin, T.-Y.; Parmar, N. Bottleneck transformers for visual recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 16519–16529. [Google Scholar]
  42. Liu, Z.; Lin, Y.; Cao, Y. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  43. Touvron, H.; Cord, M.; Douze, M. Training data-efficient image transformers & distillation through attention. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 10347–10357. [Google Scholar]
  44. D’Ascoli, S.; Touvron, H.; Leavitt, M.L. Convit: Improving vision transformers with soft convolutional inductive biases. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 2286–2296. [Google Scholar]
  45. Zhang, Q.; Yang, Y.B. ResT: An efficient transformer for visual recognition. Adv. Neural Inf. Process. Syst. 2021, 34, 15475–15485. [Google Scholar]
  46. Woo, S.; Park, J.; Lee, J.Y. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  47. Xu, Y.; Jiang, Z.; Men, A. Multi-view feature fusion for person re-identification. Knowl. Based Syst. 2021, 229, 107344. [Google Scholar] [CrossRef]
  48. Zheng, L.; Shen, L.; Tian, L. Scalable person re-identification: A benchmark. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1116–1124. [Google Scholar]
  49. Zheng, Z.; Zheng, L.; Yang, Y. Unlabeled samples generated by gan improve the person re-identification baseline in vitro. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 3754–3762. [Google Scholar]
  50. Wang, X.; Doretto, G.; Sebastian, T. Shape and appearance context modeling. In Proceedings of the 2007 IEEE 11th International conference on Computer Vision, Rio De Janeiro, Brazil, 14–21 October 2007; pp. 1–8. [Google Scholar]
  51. Miao, J.; Wu, Y.; Yang, Y. Identifying visible parts via pose estimation for occluded person re-identification. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 4624–4634. [Google Scholar] [CrossRef] [PubMed]
  52. Luo, H.; Gu, Y.; Liao, X. Bag of tricks and a strong baseline for deep person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–17 June 2019. [Google Scholar]
Figure 1. Challenges in ReID tasks. (a) low resolution; (b) light variations; (c) change of posture; (d) change of camera view; (e) occlusion; (f) inaccurate detection.
Figure 1. Challenges in ReID tasks. (a) low resolution; (b) light variations; (c) change of posture; (d) change of camera view; (e) occlusion; (f) inaccurate detection.
Applsci 13 03707 g001
Figure 2. The process diagram of the attention mechanism to enhance Transformer architecture.
Figure 2. The process diagram of the attention mechanism to enhance Transformer architecture.
Applsci 13 03707 g002
Figure 3. The process of the two types of ReID models. (a) partial models’ process; (b) attention models’ process.
Figure 3. The process of the two types of ReID models. (a) partial models’ process; (b) attention models’ process.
Applsci 13 03707 g003
Figure 4. Framework of proposed AET-Net. SAEM changes the state of the spatial distribution of embedding vectors by extracting spatial attention maps (blue). MFTM optimizes the network by constructing multiple features and calculating multiple losses (yellow and gray).
Figure 4. Framework of proposed AET-Net. SAEM changes the state of the spatial distribution of embedding vectors by extracting spatial attention maps (blue). MFTM optimizes the network by constructing multiple features and calculating multiple losses (yellow and gray).
Applsci 13 03707 g004
Figure 5. The extraction framework diagram of spatial attention map and channel attention map. (a) The extraction framework of spatial attention map. (b) The extraction framework of channel attention map.
Figure 5. The extraction framework diagram of spatial attention map and channel attention map. (a) The extraction framework of spatial attention map. (b) The extraction framework of channel attention map.
Applsci 13 03707 g005
Figure 6. Ablation study results of weight hyperparameter λ and δ . (a) Ablation study of λ on Occluded-Duke. (b) Ablation study of δ on Occluded-Duke.
Figure 6. Ablation study results of weight hyperparameter λ and δ . (a) Ablation study of λ on Occluded-Duke. (b) Ablation study of δ on Occluded-Duke.
Applsci 13 03707 g006
Figure 7. Visualization of models’ attention heat maps. (a) Input image, (b) CNN-based model, (c) ViT-based model, (d) SAEM, (e) MFTM, (f) SAEM + MFTM. The red regions represent a high level of attention by the model, and the blue regions indicate a low level of attention.
Figure 7. Visualization of models’ attention heat maps. (a) Input image, (b) CNN-based model, (c) ViT-based model, (d) SAEM, (e) MFTM, (f) SAEM + MFTM. The red regions represent a high level of attention by the model, and the blue regions indicate a low level of attention.
Applsci 13 03707 g007
Table 1. Dataset information. ID denotes the total number of identities in the dataset. The images denote the total number of images in the dataset.
Table 1. Dataset information. ID denotes the total number of identities in the dataset. The images denote the total number of images in the dataset.
DatasetsID/ImageTrain (ID/Image)QueryGallery
Market-15011501/32,668751/12,936750/3368750/19,732
DukeMTMC-ReID1404/36,441702/16,522702/22281110/17,661
Occluded-Duke1404/35,489702/15,618519/22101110/17,661
Table 2. The performance comparison between our proposed model and state-of-the-art models on DukeMTMC and Market-1501.
Table 2. The performance comparison between our proposed model and state-of-the-art models on DukeMTMC and Market-1501.
TypesModelsDukeMTMCMarket-1501
mAP/%Rank-1/%mAP/%Rank-1/%
Partial ModelsSpindle [18]---76.9
Part Aligned [34]--63.481.0
PDC [19]--63.484.1
DSR [35]--64.283.6
MGCAM [16]--74.383.8
Aligned Reid [36]--79.391.8
PGFA [7]65.582.676.891.2
PGFA-PE [51]72.686.281.392.7
PCB [22]66.181.877.492.3
PCB+RPP [22]69.283.381.693.8
HOReID [2]75.686.984.994.2
FPR [9]78.488.686.695.4
MGN [23]78.488.786.995.7
Attention ModelsHACNN [38]63.880.582.893.8
MHN-6 [26]77.289.185.095.1
CAM [33]72.985.884.594.7
PAT [8]78.288.888.095.4
DRL-Net [5]76.688.186.994.7
TransReID [24]79.388.886.894.7
OursAET-Net80.189.587.594.8
Table 3. The performance comparison between our proposed model and state-of-the-art models on the Occluded-Duke.
Table 3. The performance comparison between our proposed model and state-of-the-art models on the Occluded-Duke.
TypesModelsOccluded-Duke
mAP/%Rank-1/%
Partial ModelsPart Aligned [34]20.228.8
DSR [35]30.440.8
Ad-Occluded [4]32.244.5
PGFA [7]37.351.4
PCB [22]42.633.7
HOReID [2]43.855.1
PGFA-PE [51]43.556.6
Attention ModelsHACNN [38]26.034.4
DRL-Net [5]50.865.0
ISP [21]52.362.8
PAT [8]53.664.5
TransReID [24]53.160.5
OursAET-Net54.564.5
Table 4. Comparison results of Transformer architecture and CNN-based architecture.
Table 4. Comparison results of Transformer architecture and CNN-based architecture.
TypeModelsMarket-1501Occluded-Duke
mAP/%Rank-1/%mAP/%Rank-1/%
CNN-BasedHOReID [2]84.994.243.855.1
HACNN [38]82.893.82634.4
TransformerViT-based (baseline)86.894.753.160.5
Table 5. Results of ablation study of aet-net on Market-1501 and Occluded-duke datasets. λ | δ denotes the use of the attention-weight hyperparameter.
Table 5. Results of ablation study of aet-net on Market-1501 and Occluded-duke datasets. λ | δ denotes the use of the attention-weight hyperparameter.
IndexSAEMMFTM λ | δ Market-1501Occluded-Duke
mAP/%Rank-1/%mAP/%Rank-1/%
1 86.894.753.160.5
2 86.494.551.558.1
3 87.294.753.961.6
4 87.094.152.359.9
5 87.594.854.462.4
6 87.494.753.561.2
787.594.854.564.5
Table 6. Comparison of inferential costs on a different model. Baseline + S + M denotes baseline, and both modules are used together.
Table 6. Comparison of inferential costs on a different model. Baseline + S + M denotes baseline, and both modules are used together.
ModelsInference TimeFlopsParameters
ResNet50 [52]1x1x1x
Baseline0.4338x2.7105x3.6411x
TransReID + JPM0.4595x2.9401x3.9427x
Baseline + SAEM0.4212x2.7105x3.6411x
Baseline + MFTM0.3955x2.7106x3.6442x
Baseline + S + M0.4159x2.7106x3.6442x
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, J.; Li, P.; Zhao, R.; Zhou, R.; Han, Y. CNN Attention Enhanced ViT Network for Occluded Person Re-Identification. Appl. Sci. 2023, 13, 3707. https://doi.org/10.3390/app13063707

AMA Style

Wang J, Li P, Zhao R, Zhou R, Han Y. CNN Attention Enhanced ViT Network for Occluded Person Re-Identification. Applied Sciences. 2023; 13(6):3707. https://doi.org/10.3390/app13063707

Chicago/Turabian Style

Wang, Jing, Peitong Li, Rongfeng Zhao, Ruyan Zhou, and Yanling Han. 2023. "CNN Attention Enhanced ViT Network for Occluded Person Re-Identification" Applied Sciences 13, no. 6: 3707. https://doi.org/10.3390/app13063707

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop