Next Article in Journal
IBD: An Interpretable Backdoor-Detection Method via Multivariate Interactions
Next Article in Special Issue
Emotion Detection Using Deep Normalized Attention-Based Neural Network and Modified-Random Forest
Previous Article in Journal
Eddy Current Measurement for Planar Structures
Previous Article in Special Issue
Real-Time and Efficient Multi-Scale Traffic Sign Detection Method for Driverless Cars
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

End-to-End Network for Pedestrian Detection, Tracking and Re-Identification in Real-Time Surveillance System

1
School of Computer and Control Engineering, Yantai University, Yantai 264005, China
2
Computer School, Beijing Information Science and Technology University, Beijing 100101, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2022, 22(22), 8693; https://doi.org/10.3390/s22228693
Submission received: 16 October 2022 / Revised: 7 November 2022 / Accepted: 7 November 2022 / Published: 10 November 2022
(This article belongs to the Special Issue Sensing Technologies for Image/Video Analysis)

Abstract

:
Surveillance video has been widely used in business, security, search, and other fields. Identifying and locating specific pedestrians in surveillance video has an important application value in criminal investigation, search and rescue, etc. However, the requirements for real-time capturing and accuracy are high for these applications. It is essential to build a complete and smooth system to combine pedestrian detection, tracking and re-identification to achieve the goal of maximizing efficiency by balancing real-time capture and accuracy. This paper combined the detector and Re-ID models into a single end-to-end network by introducing a new track branch to YOLOv5 architecture for tracking. For pedestrian detection, we employed the weighted bi-directional feature pyramid network (BiFPN) to enhance the network structure based on the YOLOv5-Lite, which is able to further improve the ability of feature extraction. For tracking, based on Deepsort, this paper enhanced the tracker, which uses the Noise Scale Adaptive (NSA) Kalman filter to track, and adds adaptive noise to strengthen the anti-interference of the tracking model. In addition, the matching strategy is further updated. For pedestrian re-identification, the network structure of Fastreid was modified, which can increase the feature extraction speed of the improved algorithm by leaps and bounds. Using the proposed unified network, the parameters of the entire model can be trained in an end-to-end method with the multi-loss function, which has been demonstrated to be quite valuable in some other recent works. Experimental results demonstrate that pedestrians detection can obtain a 97% mean Average Precision (mAP) and that it can track the pedestrians well with a 98.3% MOTA and a 99.8% MOTP on the MOT16 dataset; furthermore, high pedestrian re-identification performance can be achieved on the VERI-Wild dataset with a 77.3% mAP. The overall framework proposed in this paper has remarkable performance in terms of the precise localization and real-time detection of specific pedestrians across time, regions, and cameras.

1. Introduction

Video surveillance systems play a pivotal role in traffic, security, commercial, criminal investigation, and other fields [1,2,3]. With the rapid growth of the video data volume, how to quickly locate relevant data in unrelated mass data is a crucial issue. Manually searching or browsing for specific pedestrians in such a large amount of data is usually not feasible, as it is very time consuming, laborious, and most often unsuccessful. This makes the video surveillance system much less effective. Although the development of face recognition technology is more mature, in the case of the high-density crowds, low-resolution images, or lower camera angles and other circumstances, the use of face recognition technology often does not have the ideal effect. However, pedestrian re-identification technology can effectively locate and identify specific pedestrians in real-time surveillance videos, which is of great importance in the investigation of criminal cases and the search for missing persons. The core algorithms include object detection, multi-object tracking, and pedestrian re-identification.
In recent years, within the domain of visual analytics, there has been a surge in interest in visual surveillance searches, also known as forensic visual searches. However, there are few systems that perform search tasks in the context of surveillance. A related case is the IBM Smart Surveillance System [4], which is capable of indexing videos based on multiple search criteria, thus enabling various query types such as the primary object color, object size and type, and visual features of faces.
The visual search task is accomplished by detecting pedestrians in the video, generating a discriminative representation of specific pedestrians, and tracking them across time and regions. These representations are utilized in a query-by-example manner to compare the search images with the pool of detected pedestrians, so that a list of potential matches, sorted by similarity, is generated. In this paper, we implement an end-to-end pedestrian tracking and re-identification algorithm based on deep learning networks. It combines the YOLOv5-Lite detection algorithm and Deepsort tracking algorithm to detect specific pedestrians across time, regions, and cameras. It can achieve better results through a series of improvements, which can be commonly applied to intelligent monitoring and security systems.
The contributions of this paper are as follows:
  • The network structure of the YOLOv5-Lite model is improved, and BiFPN is used for cross-scale feature fusion, which significantly improves the performance of pedestrian detection.
  • The network structure of the Fastreid algorithm is improved to increase the speed of the extraction of pedestrian features; thus, the overall pedestrian re-identification efficiency is improved.
  • The tracking strategy is optimized by using an improved Kalman filter algorithm and adding linear compensation. The improved Deepsort algorithm is used to track the re-identified pedestrian, and the tracking performance is significantly improved in all metrics.
Following is a brief summary of the rest of the paper. Section 2 introduces related works about object detection, pedestrian tracking and pedestrian re-identification. Section 3 describes the main framework of this paper. Section 4 details the improvements of the methods in detail. Section 5 presents the experimental results. Finally, Section 6 presents the conclusion.

2. Related Work

2.1. Object Detection

The development history of target detection can be categorised in AlexNet [5], with traditional feature descriptor algorithms dominating before 2012 and deep learning algorithms dominating after 2013. Traditional object detection algorithms search the entire image for regions with a class-specific maximum response using image descriptors such as HOG [6] and SIFT [7]. Numerous works based on CNN have been proposed as a result of the success of deep learning on object detection [8,9,10,11,12,13,14,15,16]. At present, CNN-based object detection can be classified into anchor-based and anchor-free methods.

2.1.1. Anchor-Based Methods

Anchor-based methods employ the concepts of traditional sliding windows and proposal-based detectors, such as Fast R-CNN [17]. Their main approach is to a priori introduce anchor boxes, which are essentially pre-defined ideas, for bounding box regression. Anchor-based object detection can be broadly classified into two-stage and one-stage approaches. For two-stage detectors, anchors serve as regression references and classification candidates. Two-stage detectors have taken the lead thanks to the development of Faster R-CNN. It uses a region proposal network (RPN) and a region-wise prediction network (R-CNN) [18] to detect objects. Several algorithms are then put forward to improve its performance, including network structure redesign and improvement [19,20,21,22], the context and attention mechanism [23,24,25], pre-training strategies [26], feature fusion and data augmentation [27,28], loss function [29,30], and an improved proposal [31,32]. At present, the two-stage methods based on standard detection benchmarks still hold the most advanced results. For one-stage detectors, anchors act as the reference boxes for final selections. One-stage detectors have drawn significant interest due to their great computational efficiency since the emergence of SSD [33]. SSD distributes anchor boxes across multi-scale layers within the convolutional neural network to predict object categories and anchor box offsets directly. Since then, a large amount of work has been proposed to improve its performance in various areas, including merging context data from various layers [34], network architecture redesign [13,14], designing new loss functions [35], reintegration and the matching of anchor boxes [36], and feature enhancement and alignment [37,38]. Currently, one-stage methods can perform almost as well as two-stage methods while executing at a faster inference speed.

2.1.2. Anchor-Free Methods

Unlike anchor-based detectors, anchor-free detectors do not require predetermined anchor boxes. In fact, there is a lengthy history of the anchor-free methods. There are primarily two types of anchor-free detectors suggested: center-based methods and keypoint-based methods. The former determines positives by using the object’s center points instead of anchor boxes, and then predicts the object bounding box for detection. The latter first detects a few pre-defined or self-taught keypoints and then generates bounding boxes to detect the objects. Early center-based anchor-free methods include YOLOv1 [39], DenseBox [40], and UnitBox [41]. Due to the relatively small number of positive samples, the recall rate for these detectors is low. In FCOS [9], all points within the object’s bounding box are considered positive samples. Besides detecting positive points, it also measures the distance between the points and the boundary box. For the purpose of detecting pedestrians with a fixed aspect ratio, CSP [42] only defines the object’s center points as positives. FoveaBox [43] treats the position in the center of the object as a positive with four distances to detect the object. It is simpler to predict the final class probability directly without the need for a centered voting. Keypoint-based anchor-free methods regard object detection as a keypoints localization problem, such as CornerNet [44], CenterNet [45], ExtremeNet [46] and RepPoint [47]. CornerNet recognizes the top-left and bottom-right corners of an object as a pair of critical points that define its bounding box. As opposed to a pair of keypoints, CenterNet expands CornetNet to a triplet to increase precision and recall. ExtremeNet generates the object bounding box by detecting four extreme points and one center point. RepPoints generates sets of points to represent objects using deformable convolutional networks (DCN) [48]. Additionally, some publications [49,50] attempt to extend the anchor-free idea to instance segmentation.

2.2. Multi-Object Tracking

Multi-Object Tracking (MOT) methods can be divided into tracking-by-detection methods and joint-detection-tracking methods. Tracking-by-detection methods [51,52,53,54,55] obtain the localization of objects before associating them with details about their appearance, motion, etc. With the rapid development of object detection techniques [10,14,44,45], tracking-by-detection methods have been a major task for MOT for many years. Simple Online and Real-time Tracking (SORT) [56] is a simple and effective method for tracking multiple objects using the Kalman filter and Hungarian algorithms. However, with more occlusion and various camera angles, the SORT algorithm’s effectiveness declines. To enhance the SORT algorithm, the developers of the Deepsort algorithm [57] added a new distance metric based on the “appearance” of the item. In recent years, a number methods of joint-detection-tracking [58,59,60,61,62] have been proposed to enhance detection and several other components, such as motion, embedding, and association models, jointly. The joint trackers provide equivalent performance at a minimal cost of computation. However, competition between various components and a lack of training data are the two main issues joint trackers are confronted with. The two issues place a cap on the maximum tracking precision. Therefore, the tracking-by-detection paradigm is still the best option in terms of tracking accuracy, as evidenced by the success of the most recent SORT-like frameworks [51,52,57,63].

2.3. Pedestrian Re-Identification

Pedestrian re-identification (Re-ID) is a technology used in video surveillance for public security and safety that seeks to recognize the same person in a collection of videos from non-overlapping camera perspectives. The process can be broken down into the following two steps: determining the embedding of features and matching them using a distance metric. It was discovered that merging image-level data with information on human parts could strengthen the robustness of Re-ID models [64,65,66]. Furthermore, numerous part-based strategies have shown significant advancements [3,67,68,69]. In order to handle partial occlusion, attention-based methods are becoming increasingly popular [70,71]. They construct useful video representations by choosing discriminative frames from video sequences. Several closely related works are as follows: Zhang et al. [70] proposed a deep graph model called Heterogeneous Local Graph Attention Networks (HLGAT) to model the inter-local and intra-local relationships with the attention mechanism in the completed local graph at the same time. Chen et al. [67] proposed a bidirectional interaction network for pedestrian Re-ID that considers multiple convolutional features as responses to various body part properties and exploits discriminative representations for person identities using inter-layer interaction. In [71], Second-Order Non-Local Attention Networks (SONA-Net) are proposed to directly model long-range relationships via second-order feature statistics for pedestrian re-identification. To address occlusion issues, Wu et al. [68] proposed a multi-level Context-aware Part Attention (CPA) model to learn discriminative and robust local part features for pedestrian Re-ID.

3. Overview of the Framework

The overall process of the system consisting of YOLOv5-Lite, Deepsort and Fastreid algorithms is shown in Figure 1. Firstly, the target pedestrian image is intercepted; then, after the Fastreid feature extraction model, feature extraction is performed on the intercepted pedestrian base library to generate the corresponding .npy file, which is read into the video to be detected. YOLOv5-Lite object detection is used if the similarity is greater than the threshold γ , the target pedestrian is determined to be a target pedestrian, and then the target pedestrian is continuously tracked by the Deepsort algorithm, while the similarity is less than the threshold if the ID-Switch phenomenon (i.e., tracking failure) occurs while the Deepsort algorithm is tracking the target pedestrian; then, the YOLOv5-Lite algorithm is used again to detect all the pedestrians in the video and continue the next process. After this, the whole process is displayed through simple visualization.

4. Methods

4.1. Pedestrian Detection

Although YOLOv5-Lite is an improvement on YOLOv5, and the detection speed is significantly improved, the detection speed is not satisfactory in real-time video inference. The accuracy of the model trained on the training set still needs to be improved, and there may be problems of inconsistent feature information at each scale. Therefore, the model network structure still has room for improvement.
In order to accelerate the speed of real-time video inference, solve the possible inconsistency of multi-scale feature information in the model, and improve the accuracy of the model, BiFPN [10] (a weighted bi-directional feature pyramid network) is introduced, for with the structure is shown in Figure 2, which can make multi-scale feature fusion faster and more convenient. The main implementations are as follows.
  • Adding residual connections: The intention is to enhance the representation of features by implementing simple residual operations, adding a jump connection between an input node and an output node at the same level, and fusing more features without considerably increasing the computational cost.
  • Removing the nodes of single input edges: Since the nodes of single input edges are not fused, they have less information and do not contribute to a considerable extent to the final fusion, and removing them also reduces the computation.
  • Weight fusion: In simple terms, this is the addition of a weight for each scale feature of the fusion, by adjusting the contribution of each scale to improve the detection speed by Fast-softmax [72]. The actual process is the fusion of the attention mechanism with FPN [73].
Overall, BiFPN achieves a combination of cross-scale bidirectional connectivity and fast normalization. For example, the two fusion features in Figure 3 at layer 6 are Equations (1) and (2).
P 6 t d = Conv w 1 · P 6 i n + w 2 · Resize P 7 i n w 1 + w 2 + ϵ ,
P 6 out = Conv w 1 · P 6 in + w 2 · P 6 t d + w 3 · Resize P 5 out w 1 + w 2 + w 3 + ϵ ,
where P 6 td is the intermediate characteristic of layer 6 in the top-down path, P 6 out is the output characteristic of layer 6 in the bottom-up path, and all other nodes are constructed in a similar manner. The R e s i z e is usually a downsampling or upsampling operation; w is the parameter we learn to distinguish the importance of different features in the feature fusion process; and C o n v is a convolutional op for feature processing.
One of the major features of BiFPN is to add weights to features at different scales. Compared with the traditional practice of directly stacking features at different scales, BiFPN can input different feature weights and let the network learn by itself, using Softmax-based fusion, as shown in Equation (3), w i and w j are learnable weights, and the softmax function is applied to normalize the weights to between 0 and 1.
O = i e w i j e w j .
For the branch model v5Lite-g of YOLOv5-Lite, the network header is modified by replacing all its Concat with BiFPN_Concat, and the model is changed, as shown in Figure 3.

4.2. Pedestrian Tracking

The Deepsort algorithm can be summarized in a two-branch structural framework, where the appearance features are extracted in the appearance branch part by detecting each frame, followed by a feature extraction model, which utilizes a feature set to store the features of the last 100 frames of each trajectory set.
With the new detection, the minimum cosine distance between the feature set R i of the ith trajectory and the feature f i of the jth detection is calculated as shown in Equation (4), with f j denoting the feature vector of the jth detection and f k ( i ) denoting the feature vector of the corresponding trace, as the feature vector that retains the past k successful traces.
d ( i , j ) = min 1 f j T f k ( i ) f k ( i ) R i ,
This distance is used as the matching cost in the association process. In the motion branch, the Kalman filtering algorithm is used to predict the position of the trajectory in the current frame. The Marxian distance is then used to measure the spatio-temporal difference between the trajectory and the detection. Deepsort uses this motion distance to filter out unlikely associations. However, the Kalman filter used by Deepsort is not robust and ignores information on the detection noise scale.
It then undergoes a matching cascade algorithm that treats the association task as a series of subproblems, rather than a global assignment problem. The core idea is to provide greater matching priority to the more familiar objects. Each association subproblem is solved using the Hungarian algorithm. However, as the tracker becomes more powerful, the matching cascade algorithm becomes more robust to confusable associations; thus, the additional prior constraints will limit the accuracy of matching.
In order to solve the problems of the ordinary Kalman filter becoming easily disturbed by the quality detection and that it ignores the information on the detection noise scale, this paper replaces the ordinary Kalman filter with the NSA Kalman filter [74] algorithm and introduces an adaptive calculation of the noise covariance R ˜ k , as shown in Equation (5).
R ˜ k = 1 c k R k ,
where R k is a predetermined constant measurement noise covariance, c k is the detection confidence score under state k, and instead of using only the appearance feature distance, both appearance and motion information are considered in the matching process.
The cost matrix C is the weighted sum of the appearance cost A a and the action cost A m , as in Equation (6).
C = λ A a + ( 1 λ ) A m ,
where the weight factor λ is set to 0.98. In addition, to solve the problem that additional prior constraints will limit the matching accuracy, the ordinary global linear assignment is used instead of the matching cascade.
The algorithm improvement also includes updating the appearance state e i t of the ith trajectory at frame t in an exponential moving average (EMA) fashion, as shown in Equation (7).
e i t = α e i t 1 + ( 1 α ) f i t ,
where f i t is the appearance embedding of the current matching detection and α = 0.9 is the momentum term. The EMA update strategy not only improves the matching quality, but also reduces the time loss.
In summary, all the improved strategies are shown in Figure 4 compared with the pre-improvement situation.

4.3. Pedestrian Re-Identification

The Fastreid algorithm uses a feature extraction model with a .pth suffix, which is the model file of Pytorch. This introduces the ONNX (Open Neural Network Exchange) deep learning model format, which optimizes the network inference speed and makes the model network structure more fine-grained.
The model files are converted from those with a .pth suffix to model files with an .onnx suffix according to the model conversion program.
The YOLOv5-Lite, Deepsort, and Fastreid algorithms are integrated into the pedestrian detection and re-identification system. Through the visual analysis and reasoning of the real-time video, it is found that the detection and reasoning speed of the experimental video is very slow, and the frame jamming phenomenon is more serious when each frame of the video is detected. Considering that the system may be bloated, some codes need to be optimized to improve the detection frame rate.
The pedestrian detection is changed to inter-frame detection, i.e., all pedestrians in the video are detected by the YOLOv5-Lite detection model in every other frame, and the frame rate module is added to the real-time video visualization interface to display the frame rate, which provides convenience for the model to make comparison tests based on the frame rate in the process of continuous iterative improvement.

5. Experiment Results

All comparison tests were conducted in the following hardware and software environments.
  • Operating system: Linux
  • Python version: 3.8.10
  • Number of CPU: 24
  • Number of GPU: 1
  • GPU type: NVIDIA GeForce RTX 3060
The pedestrian detection dataset used for YOLOv5-Lite model training consists of a total of 3900 images containing only labeled pedestrians (19,000 pedestrians in total), of which 550 are used as the test set and 550 as the validation set; all data are stored in the YOLO format before training, with a training batch_size of 16, with each being trained for 150 sessions. In addition to this, to eliminate the bias in the dataset, we also conducted a comparative experiment on the performance of our model on an additional dataset, which we will describe in detail in the next experiments.
Experiments on multi-target tracking were conducted on datasets MOT15, MOT16, and MOT17Det, which are a series of datasets from the Multi-Target Tracking MOT Challenge series that provides a measure of multi-target detection tracking methods.

5.1. YOLOv5-Lite Experimental Analysis

The evaluation indicators of the comparison test for target detection are detailed below.
For recall, which assesses the ability of the model to find all positive samples, the higher the indicator, the better the model, and the formula is (8). TP means the true positive rate, which is the number of instances which are relevant and which the model correctly identified as relevant and FN means the false negative rate, that is the number of instances which are relevant and which the model incorrectly identified as not relevant.
R e c a l l = T P T P + F N
Precision, which is the proportion of correctly predicted samples out of all predicted samples in the model, the higher the indicator, the better the model, as shown in Equation (9). FP means the false positive rate, which is is the number of instances which are not relevant but which the model incorrectly identified as relevant.
P r e c i s i o n = T P T P + F P
The mean average precision (mAP), first calculated for each category and then averaged over them, is commonly used in target detection as mAP_0.5, the average mAP over thresholds greater than 0.5, and mAP_0.5:0.95, which denotes the average mAP over different IoU thresholds, starting from 0.5 to 0.95 at 0.05 intervals (0.5, 0.55, 0.6, 0.65, 0.7, 0.75, 0.8, 0.85, 0.9, 0.95) on the mean mAP, as shown in Equation (10).
m A P = 1 N i N A P i
The improved network models were trained on the above hardware and software platforms using the pedestrian detection dataset (a total of 3900 images containing only labeled pedestrians, including a total of 19,000 pedestrians); each was trained for 150 rounds, using mAP_0.5:0.95, precision, mAP_0.5, and recall as the model’s superiority and inferiority. The optimal model before and after the improvement is shown in Table 1, and the overall size of the improved model increased slightly with the input size of 640 × 640. The accuracy of the model increased by 3% from 0.89 to 0.92, indicating that the accuracy of the improved model has increased to a certain extent, and compared with other algorithms, there is a high level of improvement in the model size and accuracy, and the performance is better in the test set.
Subsequently, in order to verify the generalization ability and practical effect of the model, we performed comparison experiments on different pedestrian datasets for the improved YOLOv5-lite model before and after.
One of the datasets used in the comparison experiments is the WiderPerson dataset, which is a benchmark dataset for pedestrian detection in more crowded scenes; its images are selected from a variety of scenes and are no longer limited to traffic scenes. The other dataset is the CUHK Occlusion dataset, which is from the Chinese University of Hong Kong and contains pedestrian images from various scenes. The division of the two datasets is shown in the following Table 2.
Under the above dataset division, we trained YOLOv4, YOLOv3, the original YOLOv5-Lite model and the improved YOLOv5-Lite model with the same training parameters: batch_size = 16, epochs = 150. The trained precision, recall, and mAP_0.5, and mAP_0.5:0.95 results are shown in Table 3 and Table 4.
The experimental results show that the improved YOLOv5-Lite model is the best in terms of the precision metric on the dataset WiderPerson, reaching 77%. Additionally, on the dataset CUHK Occlusion, our model has the highest precision, reaching 93%, as well as the first in Recall value, reaching an excellent 80%, which is sufficient to prove that our model has a better detection capability and excellent performance for different datasets.

5.2. Deepsort Experimental Analysis

The evaluation metrics for the multi-objective tracking comparison test are as follows.
FP: The total number of false positives, the number of positive samples predicted by the model but false in reality. The number of FP in the whole video is the sum of the number of FP in each frame. The lower the indicator, the better the performance.
FN: The total number of false negatives, the number of negative samples predicted by the model. The number of FN in the whole video is the sum of the number of FN in each frame. The lower the index, the better the performance.
FM (Fragm): The total number of fragmentations; each tracking object is considered a fragmentation after the tracking is interrupted and resumed.
IDs: The total number of ID changes during tracking; the lower the indicator the better the performance.
GT (Ground Truth) refers to the real tag or real object.
MOTA: MOTA is used to evaluate the accuracy of tracking, negatively correlated with the number of occurrences of FNs, FPs, IDs, and Identity Switch (IDSW). Ground-truth tracks(gtDet) are represented as the sum of the gt of each frame in a video, using Equation (11), where the higher the performance of the indicator the better.
M O T A = 1 | F N | + | F P | + | I D S W | | g t D e t |
MOTP: This is used to measure the position error in the tracking process, where C t represents how many of the predicted tracks match the GT track successfully in frame t. Where d t , i stand for the distance measured by IOU or the Euclidean distance between the trajectory i and GT. If the IOU is greater than a certain threshold or the Euclidean distance is less than a certain threshold, it is regarded as matching. MOTP pays more attention to the detection quality than evaluating the tracking effect. The higher the index, the better the performance. The formula is as follows:
M O T P = t , i d t , i t c t
MT (Mostly Tracked trajectories): The number of GT trajectories with more than 80% of the total frames successfully tracked, where the higher the indicator, the better the performance.
ML (Mostly Lost trajectories): The number of GT trajectories where the number of successfully tracked frames is less than 20% of the total number of frames, and the lower the indicator, the better the performance.
IDP: The fraction of correctly identified detections, where the higher the indicator is the better, as shown in Equation (13). Identity true positives (IDTPs) correspond to the overlapping parts of trajectories. Identity false negatives (IDFNs) represent the area that does not intersect with the matching trajectories and remaining unmatched trajectories. Unmatched trajectories are referred to as Identity False Positives (IDFPs).
I D P = I D T P I D T P + I D F P
IDR: Recall value for correct identification, where the higher the performance of the indicator the better, as shown in Equation (14).
I D R = I D T P I D T P + I D F N
IDF1: The reconciled mean value of IDP and IDR, where the higher the performance of the indicator the better, as in Equation (15).
I D F 1 = 2 1 I D P + 1 I D R = 2 I D T P 2 I D T P + I D F P + I D F N
MODA (Multiple Object Detection Accuracy) and MODP (Multiple Object Detection Precision) are also applied in target detection to measure the performance of the model, where t denotes the frame, m t denotes the number of missed detects in frame t (missed detects), and f p t denotes the number of false positives (false alarms). c ( m ) represents the custom loss function for missed detections, and c f represents the custom loss function for false alarms. N G t represents the number of ground truth objects in frame t, as shown in Equation (16).
M O D A ( t ) = 1 c m m t + c f f p t N G t
O v e r l a p p i n g R a t i o = i = 1 N m a p p e d t G i ( t ) D i ( t ) G i ( t ) D i ( t )
In O v e r l a p p i n g R a t i o , G i ( t ) represents the ground truth label of the i-th object in frame t. D i ( t ) represents the result of the detection. N m a p p e d t represents the total number of objects detected in this frame, as shown in Equation (17). Dividing the overlapping ratio by the quantity N m a p p e d t provides the indicator MODP, as shown in Equation (18).
M O D P ( t ) = O v e r l a p R a t i o N m a p p e d t
FAR (False Acceptance Rate): The number of false recognitions per frame, where the lower the index, the better the performance.
RCll (Recall): Recall rate, as shown in Equation (8).
PRCn (Precision): Accuracy rate, as shown in Equation (9).
We conducted comparative experiments on the MOT15, MOT16 and MOT17 datasets, respectively, and the results are shown in Table 5, Table 6, Table 7, Table 8 and Table 9. Regardless of the dataset, our model achieves the highest performance in the metrics of IDF1, IDP, and IDR, indicating the best detection score for correct identification, in addition to our model’s MOTA and MOTP, which all show the best results in the test. On dataset MOT16, our improved model achieves an impressive score of 99.8 in the MOTP metric, indicating that the precision of our multi-objective tracking is the highest in the current dataset compared to that of other models of the experiment. In addition, on the MOT15 dataset, our IDs metric was reduced to 134, indicating that our improved model greatly reduces the occurrence of the ID-switch phenomenon; at the same time, our improved model also has a substantial reduction in its FP and FN metrics, the false alarm rate of the model has been improved, and the performance and accuracy reached the best results at this stage.
The analysis shows that after the algorithm improvement, the tracking performance and accuracy of Deepsort improved substantially, and the performance is better on the same data set than after the improvement.

5.3. Fastreid Experimental Analysis

The model with the .pth suffix before the conversion and the model with the .onnx suffix after the conversion were used to extract features from 12 pedestrian photos, and the time used for calculation was 2.76 s before the improvement, while the feature extraction time of the improved model with the onnx format was only 0.67 s, which represents a significant reduction in feature extraction time.
In the dataset VERI-Wild compared the Fastreid model with other popular pedestrian re-identification models. As can be seen in Table 10, the mAP metric of the Fastreid model is consistently higher than that of other models, and its application to our system can achieve the desired accuracy results.

5.4. System Experimental Analysis

After the improvement of the system, pedestrian detection is changed to inter-frame detection. Under the experimental hardware and software platform of 4.3.1, the detection frame rate of the experimental video is increased from 7∼10 frames to 15∼25 frames, and the detection speed is significantly improved, and the real-time effect can be achieved.

5.5. System Showcase

Figure 5 shows two images of pedestrians to be searched for. Under the experimental hardware and software platform of 4.3.1, the effects of real-time video detection, tracking and re-identification are shown in Figure 6, and the average frame rate is 20∼25 FPS. The whole process from the appearance to the disappearance of the target is correctly tracked, as seen in Figure 6a. Moreover, it is still accurately re-identified and tracked in the new scene in Figure 6c. In Figure 6b, the person wearing red clothes can be tracked perfectly even if the level of confusion increases. When this person appears in Figure 6d, he is first accurately re-identified, and then the whole walk is tracked. Therefore, the proposed system can achieve a real-time accurate search and localization of specific pedestrians across time, regions and cameras.

6. Conclusions

In this paper, we introduce an efficient and powerful framework for real-time video pedestrian detection, re-identification and tracking. Using the improved YOLOv5-Lite algorithm with higher accuracy, the improvement as well as mainstream recognition algorithms, and the BiFPN module to solve the possible inconsistency of multi-scale feature information of the model, the detection performance of the improved YOLOv5-Lite algorithm is validated against current academic benchmarks on our adopted dataset, showing that the improved algorithm is even better. Using the improved Deepsort tracking algorithm, we introduce the NSA Kalman filtering algorithm to eliminate the interference of quality detection, the EMA update strategy to improve the matching quality to reduce the time loss, and use the ordinary global linear assignment instead of the matching cascade to solve the problem that additional prior constraints can limit the matching accuracy, and validate our improved effect using the MOT16 dataset. The results show that our method has a considerable performance improvement in each tracking metric compared to that before the improvement, and our method is more advanced and robust. Finally, we also made corresponding improvements to the Fastreid algorithm and the overall system, which were well tested and showed that our improvements were quite effective.

Author Contributions

Conceptualization, X.W., Y.S. and W.Y.; software, M.L.; investigation, J.X. and J.L.; formal analysis, Y.S. and W.Y.; writing—original draft preparation, M.L. and X.W.; writing—review and editing, J.Z., W.Y. and Y.S.; supervision, Jindong Zhao, J.X. and J.L.; funding acquisition, X.W., J.X. and J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of Shandong Province (ZR2020QF108, ZR2022QF037, ZR2020MF148, ZR2020QF046, ZR2022MF238), and the National Natural Science Foundation of China (62272405, 62103350, 62072391, 62066013, 62273290), and the Youth Innovation Science and Technology Support Program of Shandong Province under Grant 2021KJ080, Yantai Science and Technology Innovation Development Plan Project under Grant 2021YT06000645, the Open Foundation of State key Laboratory of Networking and Switching Technology (Beijing University of Posts and Telecommunications) under Grant SKLNST-2022-1-12.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets are available on Github at https://github.com/xuanwang-91/Framework-for-Pedestrian-Detection-Tracking-and-Re-identifcation.git accessed on 10 October 2022.

Acknowledgments

We would like to thank anonymous reviewers for their supportive comments to improve our manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, X.; Song, H.; Cui, H. Pedestrian abnormal event detection based on multi-feature fusion in traffic video. Optik 2018, 154, 22–32. [Google Scholar] [CrossRef]
  2. Shu, X.; Yuan, D.; Liu, Q.; Liu, J. Adaptive weight part-based convolutional network for person re-identification. Multimed. Tools Appl. 2020, 79, 23617–23632. [Google Scholar] [CrossRef]
  3. Zhang, Z.; Zhang, H.; Liu, S.; Xie, Y.; Durrani, T.S. Part-guided graph convolution networks for person re-identification. Pattern Recognit. 2021, 120, 108155. [Google Scholar] [CrossRef]
  4. Hampapur, A.; Brown, L.; Feris, R.; Senior, A.; Shu, C.F.; Tian, Y.; Zhai, Y.; Lu, M. Searching surveillance video. In Proceedings of the 2007 IEEE Conference on Advanced Video and Signal Based Surveillance, London, UK, 5–7 September 2007; pp. 75–80. [Google Scholar]
  5. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 5, 1106–1114. [Google Scholar] [CrossRef] [Green Version]
  6. Prasanna, D.; Prabhakar, M. An effiecient human tracking system using Haar-like and hog feature extraction. Clust. Comput. 2019, 22, 2993–3000. [Google Scholar] [CrossRef]
  7. Ng, P.C.; Henikoff, S. SIFT: Predicting amino acid changes that affect protein function. Nucleic Acids Res. 2003, 31, 3812–3814. [Google Scholar] [CrossRef] [Green Version]
  8. Tokmakov, P.; Li, J.; Burgard, W.; Gaidon, A. Learning to track with object permanence. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 10860–10869. [Google Scholar]
  9. Tian, Z.; Shen, C.; Chen, H.; He, T. Fcos: A simple and strong anchor-free object detector. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 1922–1933. [Google Scholar] [CrossRef]
  10. Tan, M.; Pang, R.; Le, Q.V. Efficientdet: Scalable and efficient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 10781–10790. [Google Scholar]
  11. Duan, K.; Xie, L.; Qi, H.; Bai, S.; Huang, Q.; Tian, Q. Corner proposal network for anchor-free, two-stage object detection. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 399–416. [Google Scholar]
  12. Long, X.; Deng, K.; Wang, G.; Zhang, Y.; Dang, Q.; Gao, Y.; Shen, H.; Ren, J.; Han, S.; Ding, E.; et al. PP-YOLO: An effective and efficient implementation of object detector. arXiv 2020, arXiv:2007.12099. [Google Scholar]
  13. Zhu, X.; Lyu, S.; Wang, X.; Zhao, Q. TPH-YOLOv5: Improved YOLOv5 based on transformer prediction head for object detection on drone-captured scenarios. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 2778–2788. [Google Scholar]
  14. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  15. Chen, Q.; Wang, Y.; Yang, T.; Zhang, X.; Cheng, J.; Sun, J. You only look one-level feature. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13039–13048. [Google Scholar]
  16. Panigrahi, S.; Raju, U. MS-ML-SNYOLOv3: A robust lightweight modification of SqueezeNet based YOLOv3 for pedestrian detection. Optik 2022, 260, 169061. [Google Scholar] [CrossRef]
  17. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  18. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  19. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Cai, Z.; Vasconcelos, N. Cascade r-cnn: Delving into high quality object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6154–6162. [Google Scholar]
  21. Li, Y.; Chen, Y.; Wang, N.; Zhang, Z. Scale-aware trident networks for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 6054–6063. [Google Scholar]
  22. Sun, P.; Zhang, R.; Jiang, Y.; Kong, T.; Xu, C.; Zhan, W.; Tomizuka, M.; Li, L.; Yuan, Z.; Wang, C.; et al. Sparse r-cnn: End-to-end object detection with learnable proposals. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 14454–14463. [Google Scholar]
  23. Dong, H.; Song, K.; He, Y.; Xu, J.; Yan, Y.; Meng, Q. PGA-Net: Pyramid feature fusion and global context attention network for automated surface defect detection. IEEE Trans. Ind. Inform. 2019, 16, 7448–7458. [Google Scholar] [CrossRef]
  24. Leng, J.; Liu, Y. Context augmentation for object detection. Appl. Intell. 2022, 52, 2621–2633. [Google Scholar] [CrossRef]
  25. Xiong, J.; Zhu, L.; Ye, L.; Li, J. Attention aware cross faster RCNN model and simulation. Wirel. Netw. 2021, 1–13. [Google Scholar] [CrossRef]
  26. Luo, J.Q.; Fang, H.S.; Shao, F.M.; Zhong, Y.; Hua, X. Multi-scale traffic vehicle detection based on faster R–CNN with NAS optimization and feature enrichment. Def. Technol. 2021, 17, 1542–1554. [Google Scholar] [CrossRef]
  27. Zhao, G.; Ge, W.; Yu, Y. GraphFPN: Graph feature pyramid network for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 2763–2772. [Google Scholar]
  28. Xie, J.; Pang, Y.; Nie, J.; Cao, J.; Han, J. Latent Feature Pyramid Network for Object Detection. IEEE Trans. Multimed. 2022. [Google Scholar] [CrossRef]
  29. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollar, P. Focal Loss for Dense Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 318–327. [Google Scholar] [CrossRef] [Green Version]
  30. Wan, J.; Liu, Z.; Chan, A.B. A generalized loss function for crowd counting and localization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Montreal, QC, Canada, 10–17 October 2021; pp. 1974–1983. [Google Scholar]
  31. Ge, Z.; Jie, Z.; Huang, X.; Li, C.; Yoshie, O. Delving deep into the imbalance of positive proposals in two-stage object detection. Neurocomputing 2021, 425, 107–116. [Google Scholar] [CrossRef]
  32. Xia, R.; Li, G.; Huang, Z.; Meng, H.; Pang, Y. CBASH: Combined Backbone and Advanced Selection Heads with Object Semantic Proposals for Weakly Supervised Object Detection. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 6502–6514. [Google Scholar] [CrossRef]
  33. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 21–37. [Google Scholar]
  34. Pang, Y.; Wang, T.; Anwer, R.M.; Khan, F.S.; Shao, L. Efficient featurized image pyramid network for single shot detector. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 7336–7344. [Google Scholar]
  35. Wu, S.; Yang, J.; Wang, X.; Li, X. Iou-balanced loss functions for single-stage object detection. Pattern Recognit. Lett. 2022, 156, 96–103. [Google Scholar] [CrossRef]
  36. Zhang, X.; Wan, F.; Liu, C.; Ji, X.; Ye, Q. Learning to match anchors for visual object detection. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3096–3109. [Google Scholar] [CrossRef] [PubMed]
  37. Chen, S.B.; Dai, B.M.; Tang, J.; Luo, B.; Wang, W.Q.; Lv, K. A refined single-stage detector with feature enhancement and alignment for oriented objects. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 8898–8908. [Google Scholar] [CrossRef]
  38. Luo, Z.; Yu, L.; Mo, X.; Li, Y.; Jia, L.; Fan, H.; Sun, J.; Liu, S. EBSR: Feature enhanced burst super-resolution with deformable alignment. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Montreal, QC, Canada, 10–17 October 2021; pp. 471–478. [Google Scholar]
  39. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  40. Huang, L.; Yang, Y.; Deng, Y.; Yu, Y. Densebox: Unifying landmark localization with end to end object detection. arXiv 2015, arXiv:1509.04874. [Google Scholar]
  41. Yu, J.; Jiang, Y.; Wang, Z.; Cao, Z.; Huang, T. Unitbox: An advanced object detection network. In Proceedings of the 24th ACM international conference on Multimedia, Amsterdam, The Netherlands, 15–19 October 2016; pp. 516–520. [Google Scholar]
  42. Liu, W.; Liao, S.; Ren, W.; Hu, W.; Yu, Y. High-level semantic feature detection: A new perspective for pedestrian detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 5187–5196. [Google Scholar]
  43. Kong, T.; Sun, F.; Liu, H.; Jiang, Y.; Li, L.; Shi, J. Foveabox: Beyound anchor-based object detection. IEEE Trans. Image Process. 2020, 29, 7389–7398. [Google Scholar] [CrossRef]
  44. Law, H.; Deng, J. Cornernet: Detecting objects as paired keypoints. In Proceedings of the European conference on computer vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 734–750. [Google Scholar]
  45. Duan, K.; Bai, S.; Xie, L.; Qi, H.; Huang, Q.; Tian, Q. Centernet: Keypoint triplets for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Long Beach, CA, USA, 16–20 June 2019; pp. 6569–6578. [Google Scholar]
  46. Zhou, X.; Zhuo, J.; Krahenbuhl, P. Bottom-up object detection by grouping extreme and center points. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 850–859. [Google Scholar]
  47. Yang, Z.; Liu, S.; Hu, H.; Wang, L.; Lin, S. Reppoints: Point set representation for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Long Beach, CA, USA, 16–20 June 2019; pp. 9657–9666. [Google Scholar]
  48. Dai, J.; Qi, H.; Xiong, Y.; Li, Y.; Zhang, G.; Hu, H.; Wei, Y. Deformable convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 764–773. [Google Scholar]
  49. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. Scaled-yolov4: Scaling cross stage partial network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Montreal, QC, Canada, 10–17 October 2021; pp. 13029–13038. [Google Scholar]
  50. Li, W.; Huang, L. YOLOSA: Object detection based on 2D local feature superimposed self-attention. arXiv 2022, arXiv:2206.11825. [Google Scholar] [CrossRef]
  51. Cao, J.; Weng, X.; Khirodkar, R.; Pang, J.; Kitani, K. Observation-Centric SORT: Rethinking SORT for Robust Multi-Object Tracking. arXiv 2022, arXiv:2203.14360. [Google Scholar]
  52. Du, Y.; Song, Y.; Yang, B.; Zhao, Y. Strongsort: Make deepsort great again. arXiv 2022, arXiv:2202.13514. [Google Scholar]
  53. Chen, M.; Liao, Y.; Liu, S.; Wang, F.; Hwang, J.N. TR-MOT: Multi-Object Tracking by Reference. arXiv 2022, arXiv:2203.16621. [Google Scholar]
  54. Pang, J.; Qiu, L.; Li, X.; Chen, H.; Li, Q.; Darrell, T.; Yu, F. Quasi-dense similarity learning for multiple object tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Montreal, QC, Canada, 10–17 October 2021; pp. 164–173. [Google Scholar]
  55. Dadgar, A.; Baleghi, Y.; Ezoji, M. Multi-view data fusion in multi-object tracking with probability density-based ordered weighted aggregation. Optik 2022, 262, 169279. [Google Scholar] [CrossRef]
  56. Bewley, A.; Ge, Z.; Ott, L.; Ramos, F.; Upcroft, B. Simple online and realtime tracking. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 3464–3468. [Google Scholar]
  57. Wojke, N.; Bewley, A.; Paulus, D. Simple online and realtime tracking with a deep association metric. In Proceedings of the 2017 IEEE international conference on image processing (ICIP), Beijing, China, 17–20 September 2017; pp. 3645–3649. [Google Scholar]
  58. Li, J.; Ding, Y.; Wei, H. SimpleTrack: Rethinking and Improving the JDE Approach for Multi-Object Tracking. Sensors 2022, 22, 5863. [Google Scholar] [CrossRef]
  59. Liang, C.; Zhang, Z.; Zhou, X.; Li, B.; Zhu, S.; Hu, W. Rethinking the competition between detection and ReID in multiobject tracking. IEEE Trans. Image Process. 2022, 31, 3182–3196. [Google Scholar] [CrossRef] [PubMed]
  60. Zhang, Y.; Wang, C.; Wang, X.; Zeng, W.; Liu, W. Fairmot: On the fairness of detection and re-identification in multiple object tracking. Int. J. Comput. Vis. 2021, 129, 3069–3087. [Google Scholar] [CrossRef]
  61. Hyun, J.; Kang, M.; Wee, D.; Yeung, D.Y. Detection Recovery in Online Multi-Object Tracking with Sparse Graph Tracker. arXiv 2022, arXiv:2205.00968. [Google Scholar]
  62. Luo, R.; Wei, J.; Lin, Q. VariabilityTrack: Multi-Object Tracking with Variable Speed Object Movement. arXiv 2022, arXiv:2203.06424. [Google Scholar]
  63. Zhang, Y.; Sun, P.; Jiang, Y.; Yu, D.; Yuan, Z.; Luo, P.; Liu, W.; Wang, X. Bytetrack: Multi-object tracking by associating every detection box. arXiv 2021, arXiv:2110.06864. [Google Scholar]
  64. Guo, J.; Yuan, Y.; Huang, L.; Zhang, C.; Yao, J.G.; Han, K. Beyond human parts: Dual part-aligned representations for person re-identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Long Beach, CA, USA, 16–20 June 2019; pp. 3642–3651. [Google Scholar]
  65. Sun, Y.; Xu, Q.; Li, Y.; Zhang, C.; Li, Y.; Wang, S.; Sun, J. Perceive where to focus: Learning visibility-aware part-level features for partial person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 393–402. [Google Scholar]
  66. He, T.; Shen, X.; Huang, J.; Chen, Z.; Hua, X.S. Partial person re-identification with part-part correspondence learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Montreal, QC, Canada, 10–17 October 2021; pp. 9105–9115. [Google Scholar]
  67. Chen, X.; Zheng, X.; Lu, X. Bidirectional interaction network for person re-identification. IEEE Trans. Image Process. 2021, 30, 1935–1948. [Google Scholar] [CrossRef]
  68. Wu, D.; Ye, M.; Lin, G.; Gao, X.; Shen, J. Person re-identification by context-aware part attention and multi-head collaborative learning. IEEE Trans. Inf. Forensics Secur. 2021, 17, 115–126. [Google Scholar] [CrossRef]
  69. Jin, H.; Lai, S.; Qian, X. Occlusion-sensitive person re-identification via attribute-based shift attention. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 2170–2185. [Google Scholar] [CrossRef]
  70. Zhang, Z.; Zhang, H.; Liu, S. Person re-identification using heterogeneous local graph attention networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Montreal, QC, Canada, 10–17 October 2021; pp. 12136–12145. [Google Scholar]
  71. Xia, B.N.; Gong, Y.; Zhang, Y.; Poellabauer, C. Second-order non-local attention networks for person re-identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Long Beach, CA, USA, 16–20 June 2019; pp. 3760–3769. [Google Scholar]
  72. Hussain, M.A.; Tsai, T.H. An efficient and fast softmax hardware architecture (EFSHA) for deep neural networks. In Proceedings of the 2021 IEEE 3rd International Conference on Artificial Intelligence Circuits and Systems (AICAS), Washington, DC, USA, 6–9 June 2021; pp. 1–4. [Google Scholar]
  73. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  74. Du, Y.; Wan, J.; Zhao, Y.; Zhang, B.; Tong, Z.; Dong, J. GIAOTracker: A comprehensive framework for MCMOT with global information and optimizing strategies in VisDrone 2021. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 2809–2819. [Google Scholar]
  75. Yang, L.; Luo, P.; Loy, C.C.; Tang, X. A large-scale car dataset for fine-grained categorization and verification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  76. Hongye, L.; Tian, Y.; Wang, Y.; Pang, L.; Huang, T. Deep Relative Distance Learning: Tell the Difference between Similar Vehicles. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  77. Lou, Y.; Bai, Y.; Liu, J.; Wang, S.; Duan, L.Y. VERI-Wild: A Large Dataset and a New Method for Vehicle Re-Identification in the Wild. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  78. Alfasly, S.A.S.; Hu, Y.; Li, H.; Tiancai, L.; Jin, X.; Beibei, L.; Qingli, Z. Multi-Label-Based Similarity Learning for Vehicle Re-Identification. IEEE Access 2019, 7, 162605–162616. [Google Scholar] [CrossRef]
Figure 1. Overall system flow chart.
Figure 1. Overall system flow chart.
Sensors 22 08693 g001
Figure 2. BiFPN [10] Network Architecture.
Figure 2. BiFPN [10] Network Architecture.
Sensors 22 08693 g002
Figure 3. YOLOv5-Lite modified network structure diagram.
Figure 3. YOLOv5-Lite modified network structure diagram.
Sensors 22 08693 g003
Figure 4. Deepsort Improvement Strategy Map.
Figure 4. Deepsort Improvement Strategy Map.
Sensors 22 08693 g004
Figure 5. The images of pedestrians to be detected, from left to right, are named bag, red.
Figure 5. The images of pedestrians to be detected, from left to right, are named bag, red.
Sensors 22 08693 g005
Figure 6. Real-time pedestrian re-identification in different area cross camera. (a,b): Display of pedestrian re-identification in area 1. (c,d): Display of pedestrian re-identification in area 2.
Figure 6. Real-time pedestrian re-identification in different area cross camera. (a,b): Display of pedestrian re-identification in area 1. (c,d): Display of pedestrian re-identification in area 2.
Sensors 22 08693 g006
Table 1. Comparison of detection results with other methods.
Table 1. Comparison of detection results with other methods.
ModelInput SizeFlopsSizePrecisionRecallmAP_0.5mAP_0.5:0.95FPS
YOLOv5 640 × 640 16.5 G 96.1 0.89 0.91 0.95 0.83 28
tph-YOLOv5 640 × 640 16.2 G 92.7 0.87 0.90 0.96 0.84 25
YOLOv5-Lite 640 × 640 15.6 G 1.10 0.89 0.93 0.97 0.87 42
Ours 640 × 640 15.7 G 1.11 0.92 0.93 0.97 0.87 41
Table 2. Experimental data division for different data sets.
Table 2. Experimental data division for different data sets.
DatasetTotal NumberTrain SetTest Set
WiderPerson13,38290004382
CUHK Occlusion1063850213
Table 3. Comparison of the detection results with other methods on the dataset WiderPerson.
Table 3. Comparison of the detection results with other methods on the dataset WiderPerson.
ModelInput SizeFlopsPrecisionRecallmAP_0.5mAP_0.5:0.95
YOLOv4-tiny 640 × 640 6.48 G 0.27 0.61 0.52 0.24
YOLOv3-tiny 640 × 640 13.0 G 0.75 0.54 0.62 0.32
YOLOv5-Lite 640 × 640 15.6 G 0.76 0.64 0.70 0.40
Ours 640 × 640 15.7 G 0.77 0.62 0.69 0.40
Table 4. Comparison of the detection results with other methods on the CUHK Occlusion dataset.
Table 4. Comparison of the detection results with other methods on the CUHK Occlusion dataset.
ModelInput SizeFlopsPrecisionRecallmAP_0.5mAP_0.5:0.95
YOLOv4-tiny 640 × 640 6.48 G 0.28 0.52 0.42 0.20
YOLOv3-tiny 640 × 640 13.0 G 0.88 0.77 0.84 0.47
YOLOv5-Lite 640 × 640 15.6 G 0.92 0.76 0.88 0.52
Ours 640 × 640 15.7 G 0.93 0.80 0.88 0.52
Table 5. Comparison of tracking metrics on MOT15 dataset and deepsort algorithm.
Table 5. Comparison of tracking metrics on MOT15 dataset and deepsort algorithm.
MethodIDF1IDPIDRRCllPRCnFARGT
Sort 45.9 55.4 39.2 56.0 79.2 1.07 500
Deepsort 41.1 43.5 39.0 54.5 60.8 2.55 500
Ours 86.6 87.1 86.1 97.5 98.7 0.09 500
Table 6. Comparison of tracking metrics on MOT15 dataset and deepsort algorithm (continued).
Table 6. Comparison of tracking metrics on MOT15 dataset and deepsort algorithm (continued).
MethodMTMLFPFNIDsFMMOTAMOTP
Sort120145587117,573544984 39.9 72.8
Deepsort11419514,02318,1505001117 18.1 72.2
Ours42027521100713482 95.8 99.8
Table 7. Comparison of tracking metrics on MOT16 dataset and deepsort algorithm.
Table 7. Comparison of tracking metrics on MOT16 dataset and deepsort algorithm.
MethodIDF1IDPIDRRCllPRCnFARGT
Sort 29.7 45.0 22.2 37.6 76.3 2.43 517
Deepsort 37.4 74.7 24.9 31.3 94.0 0.42 517
Ours 93.6 93.9 93.3 98.9 99.6 0.09 517
Table 8. Comparison of tracking metrics on MOT16 dataset and deepsort algorithm (continued).
Table 8. Comparison of tracking metrics on MOT16 dataset and deepsort algorithm (continued).
MethodMTMLFPFNIDsFMMOTAMOTP
Sort4127112,91668,90410901493 24.9 77.9
Deepsort30307221475,8172391190 29.1 78.5
Ours4901489117025845 98.3 99.8
Table 9. Comparison of tracking metrics on MOT17Det dataset and deepsort algorithm.
Table 9. Comparison of tracking metrics on MOT17Det dataset and deepsort algorithm.
MethodRCllPRCnFARFPFNMODAMODP
Sort99.155.010.1453,91762617.8 96.1
Deepsort99.399.30.0843949298.6 95.0
Ours99.299.30.0948950898.5 99.6
Table 10. Comparison of the state-of-the-art vehicle Re-Id methods on the VERI-Wild dataset.
Table 10. Comparison of the state-of-the-art vehicle Re-Id methods on the VERI-Wild dataset.
Methods Small Medium Large
mAPR-1mAPR-1mAPR-1
GoogLeNet [75]24.357.324.253.221.544.6
DRDL [76]22.557.019.351.914.844.6
FDA-Net [77]35.164.029.857.822.849.4
MLSL [78]46.386.042.483.036.677.5
Fastreid87.796.483.595.177.392.5
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lei, M.; Song, Y.; Zhao, J.; Wang, X.; Lyu, J.; Xu, J.; Yan, W. End-to-End Network for Pedestrian Detection, Tracking and Re-Identification in Real-Time Surveillance System. Sensors 2022, 22, 8693. https://doi.org/10.3390/s22228693

AMA Style

Lei M, Song Y, Zhao J, Wang X, Lyu J, Xu J, Yan W. End-to-End Network for Pedestrian Detection, Tracking and Re-Identification in Real-Time Surveillance System. Sensors. 2022; 22(22):8693. https://doi.org/10.3390/s22228693

Chicago/Turabian Style

Lei, Mingwei, Yongchao Song, Jindong Zhao, Xuan Wang, Jun Lyu, Jindong Xu, and Weiqing Yan. 2022. "End-to-End Network for Pedestrian Detection, Tracking and Re-Identification in Real-Time Surveillance System" Sensors 22, no. 22: 8693. https://doi.org/10.3390/s22228693

APA Style

Lei, M., Song, Y., Zhao, J., Wang, X., Lyu, J., Xu, J., & Yan, W. (2022). End-to-End Network for Pedestrian Detection, Tracking and Re-Identification in Real-Time Surveillance System. Sensors, 22(22), 8693. https://doi.org/10.3390/s22228693

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop