Next Article in Journal
Reducing Manual Annotation Costs for Cell Segmentation by Upgrading Low-Quality Annotations
Previous Article in Journal
Noninvasive Quantification of Glucose Metabolism in Mice Myocardium Using the Spline Reconstruction Technique
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Efficient Data Association for Multi-Object Tracking: Augmented with SSIM-Based Ambiguity Elimination

by
Aswathy Prasannakumar
* and
Deepak Mishra
Department of Avionics, Indian Institute of Space Science and Technology, Trivandrum 695547, Kerala, India
*
Author to whom correspondence should be addressed.
J. Imaging 2024, 10(7), 171; https://doi.org/10.3390/jimaging10070171
Submission received: 24 May 2024 / Revised: 6 July 2024 / Accepted: 10 July 2024 / Published: 16 July 2024
(This article belongs to the Section Computer Vision and Pattern Recognition)

Abstract

:
Recently, to address the multiple object tracking (MOT) problem, we harnessed the power of deep learning-based methods. The tracking-by-detection approach to multiple object tracking (MOT) involves two primary steps: object detection and data association. In the first step, objects of interest are detected in each frame of a video. The second step establishes the correspondence between these detected objects across different frames to track their trajectories. This paper proposes an efficient and unified data association method that utilizes a deep feature association network (deepFAN) to learn the associations. Additionally, the Structural Similarity Index Metric (SSIM) is employed to address uncertainties in the data association, complementing the deep feature association network. These combined association computations effectively link the current detections with the previous tracks, enhancing the overall tracking performance. To evaluate the efficiency of the proposed MOT framework, we conducted a comprehensive analysis of the popular MOT datasets, such as the MOT challenge and UA-DETRAC. The results showed that our technique performed substantially better than the current state-of-the-art methods in terms of standard MOT metrics.

1. Introduction

Vision-based multiple object tracking (MOT) is a longstanding research problem with broad applications in computer vision such as intelligent surveillance systems, robotics, human–computer interaction, medical image processing, and autonomous driving. The MOT algorithm provides a robust framework for real-time monitoring and analysis of multiple moving objects, enabling accurate tracking and prediction of their movements in various dynamic scenarios. The tracking-by-detection paradigm is widely recognized as the most effective approach to multiple object tracking (MOT). It involves utilizing an efficient object-detection algorithm to identify objects within each frame of a video sequence. Subsequently, a data association algorithm is employed to establish associations between detections across frames, thereby creating object trajectories [1,2,3]. Although various approaches have been presented to handle the problem, MOT is still a challenging research area due to factors like object occlusions, the varying number of objects per frame, abrupt appearance changes, etc.
The state-of-the-art MOT concepts have become more potent with recent advances in deep learning. MOT treats object detection and data association as two independent tasks. Modern advancements in deep learning have led to the development of highly effective off-the-shelf object detectors capable of accurately detecting various objects in complex scenes [4,5,6,7,8,9,10]. Once object detections are obtained in each frame, the subsequent task of data association focuses on linking these detections across consecutive frames to establish object trajectories over time. Data association remains a challenging task in its own right and has not fully leveraged the advancements in deep learning. The standard process of data association typically involves extracting representative features from individual detections and then matching them with existing object trajectories using specified similarity metrics. Deep learning networks offer enhanced capability for learning robust feature representations of objects. In our study, we employed a modified VGGNet deep network for hierarchical feature learning and extraction from all detected objects. This deep feature extractor allowed us to capture distinctions and variations in object appearances, improving the accuracy and reliability of object associations over time.
After extracting features for detections in each frame, the next task is to associate these detections with previously tracked objects. This association involves comparing the extracted features of detections with those of existing object trajectories to find the most suitable matches. Detections with the highest similarity scores are linked to their corresponding trajectories. Our research mainly focuses on enhancing the data association task within multiple object tracking (MOT) frameworks. We propose an efficient association framework that integrates a deep feature association network (deepFAN) and the Structural Similarity Index Metric (SSIM) [11,12,13]. This framework jointly calculates association scores for object detection–target pairs. The deepFAN learns the complex feature association function to encode the association score between detections and tracked targets, while the SSIM handles uncertainties in association by comparing its feature similarities. By combining the deep learning capabilities of the deep feature association network with SSIM, our approach aims to improve the accuracy and reliability of object associations across frames.
Traditionally, affinities in multiple object tracking (MOT) are calculated by exhaustively evaluating all possible pairs of detection and target features. In contrast, our proposed MOT framework integrates a neighborhood detection estimation (NDE) module to refine this process, selecting a more reliable subset of detection–target pairs. The NDE module enhances efficiency by focusing on nearby or contextually relevant detections rather than evaluating every possible permutation. This filtering step improves the quality of associations by prioritizing those with a higher likelihood of accuracy. In our framework, the deep feature association network (deepFAN) and the Structural Similarity Index Metric (SSIM) jointly determine the association score for these refined pairs. Furthermore, the training method we employed for the deep feature association network (deepFAN) enables efficient object association across multiple frames in a video sequence, ensuring reliable trajectory tracking. During training, the network is exposed to input frame pairs that are not necessarily consecutive. This strategy proves beneficial by allowing the framework to link objects across non-adjacent frames. This capability reduces instances of identity switches and fragmented object trajectories. By integrating the NDE module and optimizing deepFAN training, our MOT framework enhances tracking accuracy while maintaining computational efficiency.
This study presents a systematic approach to estimating an efficient association matrix for multiple object tracking (MOT), which effectively summarizes the correspondence between current frame detections and previously estimated target trajectories. Leveraging the capabilities of deep learning architectures, our proposed framework integrates innovative components aimed at enhancing MOT performance. Key components of our approach include the following:
  • The proposed data association framework employs the deep feature association network (deepFAN) along with the Structural Similarity Index Metric (SSIM) to estimate an efficient association matrix. This combination improves the robustness of object associations by leveraging deep learning for feature extraction and similarity evaluation.
  • In the proposed data association framework, a neighborhood-detection-estimation (NDE) scheme is introduced to select a reliable subset of detection–target pairs. This neighborhood detection estimation, along with post-processing steps within the deep feature association network, contributes to enhancing the computational efficiency. Experimental evaluations highlight that the proposed approach minimizes incorrect associations, thereby improving overall tracking performance.
  • A specialized training strategy is developed for the deep feature association network (deepFAN), allowing the network to utilize non-consecutive frame pairs for the effective learning of the data association function. This method improves the overall ability of the network to link objects across frames, thereby reducing identity switches and fragmented trajectories.
We validated the effectiveness of each component through ablative experiments on the MOT validation dataset. Additionally, comprehensive analyses on the benchmark datasets, including MOT15, MOT17, MOT20, and UA-DETRAC, demonstrated that our method achieved competitive and state-of-the-art results across various MOT evaluation metrics. The MOT metric scores for identity switches, fragmentation, and false negatives were reduced, indicating the reduction in the wrong association among detection target pairs.
The rest of the article is organized as follows: Section 2 reviews the existing literature on multiple object tracking (MOT). Section 3 details the methodology employed in the online multiple object tracking framework. In Section 4, we present the experimental findings and comparative results and discuss them in depth. Finally, Section 5 concludes the study with a summary of the findings and suggestions for future research directions.

2. Related Works

To obtain a comprehensive overview of the multiple object tracking (MOT) problem, we refer to foundational studies [14,15,16]. Within MOT frameworks, the tracking-by-detection approach is the most commonly utilized method [1,2,3]. The effectiveness of this approach relies heavily on the quality of object detections and the accuracy of trajectory estimation. The recent advancements in deep learning have significantly improved object detectors [4,5,6,7,8,9,10], leading to better object detection performance and, consequently, enhancing the overall efficiency of the MOT framework.
This discussion will focus specifically on data association approaches used for trajectory estimation in MOT. An essential step for any data association method is computing representative features of the detections in each frame. Several approaches exist for determining representation models, including appearance-based [17,18,19], motion-based [20,21,22], and composite models [23,24]. For MOT frameworks, deep learning-based feature extraction methods provide robust and discriminative representation models for object detections, which significantly boost tracking performance. Typically, pre-trained classification or object detection models are employed for feature extraction in tracking tasks [25,26,27,28,29]. In particular, ShiJie et al. [30] proposed a deep affinity network that jointly learns representational features and their affinities with targets. Our proposed MOT framework adopts the feature extraction model utilized in [30].
The study by Emami et al. [31] views data association as a multidimensional assignment problem and consolidates many popular learning algorithms employed for MOT data association. Researchers have explored various methodologies, including non-probabilistic algorithms, probabilistic graphical models, Markov Chain Monte Carlo (MCMC), and deep learning techniques to solve the data association problem. Among non-probabilistic approaches, the Greedy Randomized Adaptive Search Procedure (GRASP) is frequently used for multi-sensor multi-object tracking [32]. In probabilistic graphical models, common techniques include network optimization [33,34], conditional random fields [35,36,37], and belief propagation [38,39]. Additionally, MCMC has been a valuable tool for data association in multiple object tracking [40,41].
In recent years, there have been numerous successful attempts to formulate data association in MOT using deep learning methods. The Deep Affinity Network (DAN) proposed by ShiJie et al. [30] represents an end-to-end trainable deep network that jointly learns feature modeling and association estimation. Similarly, FAMNet [42] leverages deep networks for both feature extraction and association estimation. Yihong et al. introduced the Deep Hungarian Network (DHN) [43], which predicts associations from a cost matrix derived from detections and targets. The Dual-Matching Attention Network (DMAN) [44] employs spatial and temporal attention mechanisms to predict and refine association assignments. The integration of deep models such as Recurrent Neural Networks (RNNs) [45], autoencoders [46], and Generative Adversarial Networks (GANs) [47] into the data association problem has led to significant improvements in MOT performance.
This work presents a systematic approach to data association within the MOT framework that harnesses the power of deep learning models. The proposed MOT algorithm for track association enhances both computational efficiency and tracking accuracy. By leveraging the potential of deep learning models, our method aims to address the complexities and challenges associated with data association in MOT, ensuring more reliable and effective tracking outcomes.

3. Methodology: Online Multiple Object Tracking Framework

In the tracking-by-detection paradigm of MOT, the process involves two distinct modules: the object detector and the object tracker. The object detector initially identifies target objects by generating bounding boxes in each video frame. From these bounding boxes, we compute the center locations of objects, C D f for frame I f . Our proposed MOT framework is designed to seamlessly integrate with existing multi-object detection methods. We evaluated our approach across various online challenges in multiple object tracking, where state-of-the-art object detectors provide the initial object detections. Specifically, we utilized detections from prominent MOT challenges such as MOT15, MOT17, MOT20 [48,49], and UA-DETRAC [50,51]. Each challenge provides video sequences annotated with detections generated by specific detectors designated for the challenge.
The block diagram representation of the proposed MOT framework is shown in Figure 1. One of the significant components of the proposed MOT framework is a deep feature extractor using a modified VGGNet architecture. The architecture of the feature extractor employed in the proposed framework is based on the state-of-the-art MOT framework described in Reference [30]. The system is expertly developed to efficiently extract comprehensive and compact features from the input object detections. The pretrained VGGNet architecture is fine-tuned within the context of multiple object tracking using training sequences of the MOT benchmark. As shown in Figure 1, the representative feature of each object is obtained by passing the current video frame I f and object centers C D f through the deep feature extractor. For each object detection, a 520-dimensional feature vector is obtained. We refer to [30] for the architectural details of the modified VGGNet feature extractor.
Let D f = { d i f } i = 1 N d represent the set of detections given in frame I f , where N d is the number of available detections. We acquire a detection feature matrix, F D f R 520 × N d , by accumulating the 520-dimensional feature vector for N d detections for each input frame I f . This detection feature matrix F D f is then made available for the data association task.

3.1. Data Association Methodology

In this section, we extend our discussion on the proposed data association framework that incorporates a deep feature association network and the Structural Similarity Index Metric along with neighborhood detection estimation to tackle the problem. The association algorithm identifies a correspondence between the object detections in the current frame and existing trajectories from the previous frames. This involves comparing the extracted features of detections with those of existing object trajectories to find the most suitable matches. Detections with the highest similarity scores are linked to their corresponding trajectories. Here, we employed a deep feature association network (deepFAN) that consists of a pre-trained CNN-based compression network and an image similarity metric, the Structural Similarity Index Metric (SSIM), to estimate the data association efficiently. This part of the MOT framework computes a feature association matrix A , which encodes the pairwise similarities of the detections and the pre-existing targets.

3.1.1. Neighborhood Detection Estimation

Generally, the data association matrix models a global relationship between all the detections in the current frame and the tracked targets from the previous frames. In the proposed method, instead of considering all the combinations, only the reliable detection–target pairs are chosen for the association task. The neighborhood detection estimation methods are employed to identify those detection–target pairs. This method is based on the assumption that the objects are in a smooth motion, i.e., the location of the objects did not drift drastically in subsequent video frames. Therefore, we have to consider only the detections in the neighborhood area of the targets for the data association.
Let F T L f 1 R 520 × N T L represent the set of target feature vectors in the previous frame I f 1 , including the feature vectors of tracked and lost targets.
F T L f 1 = { F T f 1 , F L f 1 } , where F T f 1 R 520 × N T , F L f 1 R 520 × N L , N T L = N T + N L .
In Equation (1), F T f 1 and F L f 1 are the feature matrices that consist of the feature vectors of the tracked and lost targets from the previous frame I f 1 and N T and N L are the number of active tracked and lost targets.
The neighborhood-detection-estimation algorithm simply relies on the distance between the centers of the detection and target feature vectors. In order to find the distance, we need to define a distance metric. Here, we are adopting the Euclidean distance with an additional scaling factor. Let C D = { C D x , C D y } and C T L = { C T L x , C T L y } be the centers of the detections and targets. The scaled Euclidean distance E s between a detection and a target with centers ( c D x , c D y ) and ( c t x , c t y ) is defined as
E s = ( c D x c t x ) 2 + ( c D y c t y ) 2 I x 2 + I y 2 ,
where ( I x , I y ) represents the size of the video frame.
Optical flow-based motion prediction: From the object detection bounding boxes, we can determine the center locations of all the detections in the image frame, C D f . Further, we have the locations of the tracked and lost targets C T L f 1 in the frame I f 1 as feedback information from the previous target trajectories. The possible locations of these targets in the present frame I f , C ^ T L f , are estimated using the optical flow motion model. Specifically, knowing the target center in I f 1 , c t f 1 = { c t x f 1 , c t y f 1 } , we compute its corresponding location c ^ t f in the following frame ( I f ) using the Lucas–Kanade optical flow method with pyramids [52].
c ^ t f = c t f 1 + v = ( c t x f 1 + v x , c t y f 1 + v y ) ,
where v = ( v x , v y ) is the optical flow at c t f 1 . Using optical-flow-based motion prediction, the location of a lost target is continuously updated. Consequently, if the target is occluded in one frame and reappears at a different location in subsequent frames, this motion prediction aids in estimating the likely location of the lost target. This approach improves the efficiency of reidentifying the lost target, leading to more accurate and reliable tracking performance.
Using Equation (2), we calculate the distance between each existing target and all detections D f and select only those targets within the distance threshold, T e , to prioritize nearby detections. The network then encodes the feature vectors of all the possible pairings between the targets and the respective neighboring detections into a tensor, termed the feature permutation matrix Φ N × N × ( 520 × 2 ) . For clarity, the dimension of the tensor Φ is described as W i d t h × H e i g h t × D e p t h , where the width represents the targets and the height represents the detections. The feature vector of each target is concatenated with the feature vector of each one of its neighboring detections and arranged in Φ along its depth dimension. For each image frame in the video sequence, the number of targets and detections will vary. To maintain consistency in the tensor dimension, we introduce additional zero vectors into the matrix, ensuring that the size consistently remains at N × N × 1040 . The value chosen for N limits the maximum number of object detections in each frame, and through our analysis, N = 80 was found to be a generous bound for the MOT benchmark datasets.

3.1.2. Deep Feature Association Network

The objective of this component in the proposed MOT framework is to estimate the affinities between the selected detection–target pairs using the extracted feature vectors. This sub-network maps the tensor Φ R N × N × 1040 into a feature association matrix A F R N × N . In the association matrix A F , the columns account for the detections in the current frame and rows account for the active targets, both tracked and lost, from the previous trajectory. Besides, the scalar score in the matrix A F ( i , j ) indicates the confidence of the j t h detection and i t h target ( d j f and T L i f 1 ) associated with the same identity.
We refer to the major component of this module as the deep compression network due to its functionality. The architecture of the deep compression network is inspired by the work presented in [30]. The input to this network is the tensor Φ R N × N × 1040 , which accumulates the feature vectors of target–detection pairs. The output is an association matrix A F R N × N that encodes the similarity scores of these pairs. The specifications of the deep compression network architecture are detailed in Table 1. This network employs a five-layer convolutional neural network with 1 × 1 kernels for the task. As the tensor Φ passes through the network, it undergoes gradual dimension reduction along the depth dimension via the 1 × 1 kernels. These convolutional kernels enable the computation of similarity scores for each object pair without interference from neighboring objects.
Training deep compression network: During the training process, the deep compression network learns the association function, which estimates the feature association matrix A F R N × N from the tensor Φ R N × N × 1040 for reliable online multiple object tracking. The approach used to train the compression network is illustrated in Figure 2. When we employ the proposed MOT framework (Figure 1) for online tracking, the feature extractor functions as a single-stream model. Additionally, during the tracking process, the input frames are presented in the order of the original video. We develop a specialized training strategy for the deep compression network, which enables the network to effectively learn the data association function by utilizing non-consecutive frame pairs from the video sequence. As a result, the network learns to reliably associate objects in a given frame with those in multiple previous frames, benefiting the framework by reducing identity switches and fragmented target trajectories. As shown in Figure 2, during training, we configured the network as a two-stream network of modified VGGNet with shared parameters. The feature extractor receives two frames, I f and I f p , separated by p frames (i.e., not adjacent frames), as well as the centers of object detection, C f and C f p , of pre-detected objects within those frames. These frame pairs are processed by modified VGGNets, which extract a 520-dimensional feature vector for each object detection in the input frames. We obtain feature matrices, F D f and F D f p , which accumulate the feature vectors for detections in each input frame I f and I f p . Since the input frames are non-adjacent, neighborhood detection estimation (NDE) is not applicable and is excluded from the training pipeline. The network arranges the columns of F f and F f p to concatenate the columns of the two feature matrices along the depth dimension of the tensor Φ R N × N × 1040 in all possible permutations. To maintain consistency in the tensor dimensions, additional zero vectors are introduced, ensuring that the size remains N × N × 1040 . This tensor is then forward-passed through the compression network, which utilizes five convolutional layers with 1 × 1 kernels to map and estimate the feature association matrix A F R N × N . For computing the error of the network during the learning process, we define a loss function J with the help of ground truth trajectories. A ground truth target association matrix G R N × N is constructed as a binary matrix encoding the correspondence between the objects detected in frames I f p and I f . If the i t h target in I f p corresponds to the j t h target in I f , then the entry to the matrix G ( i , j ) f p , f is non-zero; otherwise, it is zero. The ground truth target association matrix G is subsequently compared with the network-predicted feature association matrix A F , for the loss computation. The loss function of our training network is defined as
J ( G , A F ) = i , j = 1 : N G ( log A F ) i , j = 1 : N G ,
where the symbol ⊙ represents the Hadamard product. The log operation on A F is performed elementwise, and i , j = 1 : N finds the sum of all elements in the Hadamard product matrix. In the loss function, instead of computing the distance metric between the predicted association matrix A F and ground truth association matrix G , the probabilities encoded by the relevant coefficients of A F are maximized. During learning, the parameters of the compression network are updated by minimizing the loss over the training samples. The trained compression network is employed in online multiple object tracking.
Referring to Section 3.1.1, for consistency, additional zero vectors are introduced in the tensor Φ , so that the size will always be N × N × ( 520 × 2 ) . Therefore, in the association matrix A F R N × N , there are irrelevant values corresponding to the appended zero vectors. To reduce the irrelevant information and to normalize the matrix, we performed the following three post-processing steps over the feature association matrix A F :
(i) 
Truncation : Since we have only N d detections and N T L active targets, the  matrix A F R N × N is resized by truncating the matrix to N T L × N d .
(ii) 
Rowwise Softmax : This operation normalizes the rows of the association matrix by fitting a separate probability distribution. The output row values are between the range [0, 1], and the total sums up to 1. Thus, each row of the resulting association matrix encodes the association probability between each active target in I f 1 and all the detections in I f .
(iii) 
Thresholding : The association matrix values indicate the similarity between the detection and target objects. For a reliable data association, the values above the threshold T a are retained, and all other values below the threshold are set to zero.
These post-processing steps obtained for us an updated feature association matrix A F N T L × N d , which was further passed to the SSIM for the association update.

3.1.3. Structural Similarity Index Metric for Association Update

The ultimate aim of the data association module is to develop a robust association model that delivers the most relevant information for achieving accurate multiple object tracking (MOT) performance. In the association matrix, a non-zero association value indicates a potential match between the corresponding target–detection pair. Traditionally, the detection with the highest association score is linked to the target trajectory. However, when multiple detections have similar or nearly equivalent association scores, uncertainties arise, leading to unreliable associations between detections and targets.
To address this issue, our proposed method incorporates the Structural Similarity Index Metric (SSIM) [11,12,13]. The SSIM is a widely recognized perceptual metric that measures the similarity between two images by leveraging their structural characteristics. By integrating the SSIM, we enhance the decision-making process for target associations. The proposed MOT framework considers the association results derived from the SSIM to make the final decision on the target association. This metric evaluates the effective similarity between the target and detection pairs, thereby improving the accuracy and reliability of the associations. We reduced the chance of wrong associations, which can happen when multiple detections have association scores that are very close to each other by using the SSIM. This makes sure that the detections are more accurately aligned with their targets.
Let ( T L ) i be the i t h active target and { d k } k = 1 K be the detections corresponding to the non-zero association scores with the i t h target. Also, let d m a x represent the detection with the highest score and A F ( i , m a x ) be the highest score. As stated before, if there are other detections with similar or closer scores to this highest association score A F ( i , m a x ) , uncertainties occur in the target association. For the target ( T L ) i , first, the set of detections with uncertainty D s is estimated as follows.
D s i = { d k with A F ( i , k ) ( A F ( i , m a x ) 0.1 ) } , D s = { D s i } i = 1 N T L .
If the association matrix A F contains any zero rows, then the corresponding detection set in D s becomes an empty set. The SSIM module calculates the similarity score between the target and each detection in D s . The output of the SSIM module is another SSIM association matrix A S R N T L × N d in which rows and columns represent the same active targets and detections as in A F , but the entries replace the SSIM score of each valid pair, i.e.,
A S ( i , j ) = S S I M ( T L i , d j ) , if d j D s i 0 , otherwise
The SSIM-based association matrix, A s , is utilized alongside A F to establish the final track association, A . The track association matrix is the result of adding both matrices A F and A s  together.

3.2. Track Association

In a multiple object tracking scenario, an object detected in a video sequence has to undergo different state transitions. When the object detector detects the object for the first time, a new track is initialized in the trajectory list. Now, the object is in the tracked state and remains in the same state until re-detected in the subsequent frames. When the object gets occluded or goes out of the camera’s field of view, the object is transferred to the lost state. If the lost target re-appears, then the state is updated as tracked, and the tracking process resumes. The trajectory of the lost target is terminated if it stays long in the lost state. The data association algorithm in MOT helps to find the state of each detection in the video sequence. It estimates the correspondence between the object detections in the current frame and existing targets.
After accomplishing the training of the deep compression network with MOT datasets, we employed the trained network in the proposed MOT framework. Algorithm 1 summarizes the online tracking process in the proposed method. The objective of the MOT problem is to find the trajectory of all the possible targets present in the given input image sequence. Here, the MOT framework expects the present image frame I f and the object detection centers C D f as its inputs. The detection feature matrix F D f computed by the VGGNet feature extractor along with the target feature vector matrix F T L f 1 are utilized to create the feature permutation tensor Φ by a concatenation operation. We stored the feature vectors of the active targets, both tracked and lost targets, from the previous frame to find the association in the current frames. The tensor Φ forward-passed through the compression network is mapped to the association matrix A F as described in Section 3.1.2. Along with A F , the SSIM-based association matrix A s is also utilized for finding the final track association, A . The track association method adapted in our framework is performed as follows.
Algorithm 1: Online multiple object tracking.
Input: Video sequence, V = { I f | f = 1 , 2 , , F } and object detections D f
Output: Set of object trajectories, T = { τ i } i = 1 N ,
1:
Initialization: T
2:
for Video frame I f in V do
3:
   Feature extraction
4:
    Input : I f and C D f
5:
    Output : F D f R 520 × N d
6:
   if ( f = = 1 ) then
7:
     Initialize trajectory τ i 1 for each detection,
8:
     state==tracked;
9:
   else
10:
     Neighborhood estimation detection
11:
      Input : F D f and F T L f 1
12:
      Output : Tensor, Φ R N × N × 1040
13:
     Feature association network
14:
      Input : Φ
15:
      Output : A F
16:
     Structural Similarity Index Metric
17:
     for each active target, ( T L ) i , do
18:
        find D s i detections with uncertainty.
19:
     end for
20:
      Input : ( T L ) i and D s i , i = 1 : N T L
21:
      Output : A s
22:
     Final track association matrix
23:
      Input : A F and A s
24:
      Output : Final track association matrix, A = A F + A s
25:
     Target association
26:
     Hungarian algorithm assigns detection to active targets.
27:
      Input : A
28:
      Output : trajectory, τ f
29:
     if tracked track τ j ( f 1 ) not assigned to detection then
30:
        state == lost;
31:
     end if
32:
     if lost track τ j ( f 1 ) assigned to detection d m  then
33:
        state ==tracked;
34:
     else
35:
        state==inactive (if length of lost frames > N i n a c t , terminate track);
36:
     end if
37:
     for detections not covered by tracked and lost targets do
38:
        Initialize trajectory τ i f .
39:
        state==tracked.
40:
     end for
41:
   end if
42:
end for.
43:
return trajectories of the objects, T .
In the first frame I 1 , we initialize the trajectory list T with tracks { τ i } i = 1 N d by considering all the detections present in it as new tracked targets. Here, a track τ i is an ordered set of the states of the i t h target in the video sequence.
τ i = { s i f e , , s i f t } , s i = ( c x , c y , w , h )
In Equation (7), f e and f t are the entry and terminate frame for the i t h target, ( c x , c y ) is the center of the target, and ( w , h ) are the width and height of the target. For each new target entry, the track is initialized with τ i = s i f e . The trajectory list is updated after each input frame by employing the Hungarian algorithm [53] on the final association matrix A . In the track association part, the targets under the tracked state get higher priority. In this process, the targets that are associated with the detections are labeled as tracked, and the targets without association are labeled as lost. If the target stays in the lost state for a long time (say N i n a c t as the length of frames; here, we chose the value as 20 frames), it is considered that the object has entered an inactive state, and we terminate the trajectory corresponding to that object. Finally, we initialize new trajectories for the detections that are not associated with the tracked targets.

4. Experiment Results and Discussion

In this section, we experimentally demonstrate the performance of the proposed deep MOT framework on the popular MOT benchmark datasets using the standard metrics. Here, we present the implementation details of our MOT framework, followed by the benchmark datasets and metrics used for performance analysis. We first conducted an ablation study on the validation dataset to understand the behavior of our approach better. Further, to obtain an authoritative reference when addressing MOT problems, the proposed framework was evaluated on the test datasets and the results compared with the state-of-the-art methods.
MOT benchmark datasets: The three popular MOT datasets, namely MOT15, MOT17, and MOT20 from the MOT Challenge [48,49], and UA-DETRAC [50,51] were employed here to test the performance of the proposed approach. These are the centralized benchmark datasets used to evaluate the tracking techniques in online multiple-object-tracking challenges. The annotated training video sequence, which includes the object detections and the ground truth labels in each frame, is used to train the models. The video sequences in the test datasets provided only object detections, whereas the ground truth labels remained unrevealed. Once the new MOT tracker has been submitted for performance analysis, the online MOT challenge hosting server evaluates the tracking results based on the standard MOT metrics [54].

4.1. Implementation Details

The proposed MOT framework was implemented in a Python framework, and the training was conducted on an NVIDIA Geforce Titan Xp 12GB GPU. We performed the training of the deep compression network on the MOT17 training dataset using the SGD optimizer. The hyperparameter values finally used in the training process were as follows: a batch size of 8, momentum of 0.9, an initial learning rate of 0.01, a weight decay of 0.0001, and the number of epochs per model of 120.
In the proposed MOT framework, the decision for the state transition of a target from lost to inactive is based on the hyperparameter N i n a c t , which is the maximum number of frames the target stays in the lost state before being transferred into an inactive state. In our analysis, we kept the value for N i n a c t at 20. We chose N = 80 as a generous bound for the MOT benchmark datasets, because it limits the maximum number of object detections in each frame. The feature extractor network has an input frame size of 900 × 900 . Therefore, the network first resizes all the training and testing data to these dimensions before passing them through. The two threshold parameters used in this proposed framework are distance threshold, T e and association threshold T a . The optimum value for the evaluation metrics obtained with the value of T e is equal to 0.35. In the thresholding step implemented as a post-processing part of the feature association matrix A F , we used a association threshold T a . For T a equal to 0.4, the proposed MOT framework obtained the optimum performance. The selection of T e and T a is explained in Section 4.2.

4.2. Ablation Study

In this section, to gain a deeper insight into the proposed MOT framework, we experimentally evaluated the contribution of different tracking components. Since the ground truth annotations are not provided for the MOT test datasets, the ablation study was conducted on the MOT15, MOT17, and MOT20 training datasets. We split the MOT training dataset into training and validation datasets. The splitting of the dataset is presented in Table 2. The proposed framework was trained on the training sequences, and the performance was evaluated on the validation sequences, as provided in Table 2.

Significance of the Proposed Tracking Components

This section follows the detailed analysis and discussion on the results obtained for the analyses of the three main components, (i) neighborhood estimation detection, (ii) feature association network, and (iii) SSIM association update. To investigate the significance of each component, we conducted several experiments by disabling one element at a time and studying the performance for the MOT metrics. Table 3 consolidates the evaluation results of the variants of the proposed method on all MOT evaluation metrics that demonstrate the significance of each module in the framework:
(i) 
Neighborhood detection estimation:
As we discussed earlier, using NDE, we limited the search space for the association of the particular target into its neighborhood, assuming that the target will not move drastically from its position in a single frame change. The neighborhood of the target object was set to a limit using a distance threshold T e . Figure 3 shows the MOTA and IDF1 with different values for distance threshold T e . The optimum value for the evaluation metrics obtained with the value of T e is equal to 0.35, and we used this value of T e for the further experiments.
To demonstrate the significance of the proposed NDE in the MOT framework, we compared the performance of the trackers with and without NDE. Figure 4 shows three essential MOT metrics, MOTA, MOTP, and IDF1, evaluated on both the MOT17 and MOT20 validation datasets. Also, Table 3 tabulates the experimental results on all MOT metrics evaluated on the MOT17 and MOT20 validation datasets. The MOTA metric measures the overall accuracy of the detection and tracking, whereas the IDF1 scores highly depend on the association accuracy. The MOTP deals with the detection output. It is evident from the MOT scores that the scores improved with NDE. The MOTA is a metric derived from three types of detection–association errors: false positives, false negatives (missed targets), and identity switches. Since the NDE employed in the proposed method helps to choose only the reliable pairs for the association, it reduces the chance of wrong associations during the track estimation. It is clear from the results that NDE helps to reduce the wrong association, thereby reducing the identity switches, fragmentation, false negatives, and false positives, which in effect improves the MOTA. Also, the improvement in the IDF1 score also justifies that, with NDE, the association accuracy is improved.
(ii) 
Deep feature association network:
The deep feature association network (deepFAN) estimates the association matrix that encodes the association scores of each detection–target pair. The module includes three post-processing steps that remove the irrelevant information from the association matrix, improving the trajectory estimation. In the thresholding step, we used a hyperparameter, threshold T a . Figure 5 plots the MOTA and IDF1 scores of the proposed MOT framework with different values of T a , and an optimum result was obtained for T a equal to 0.4.
Figure 6 and Table 3 show the performance of the proposed training strategy on the MOT17 and MOT20 validation sequences in terms of the MOT metrics. The deep network was trained on the MOT dataset with the strategy that the input frames need not be sequential, i.e, non-consecutive input frames. Therefore, the data association model becomes robust to the tracking challenges such as appearance variation, illumination changes, scale changes, etc. It also helps in the re-identification of the lost targets and handles object occlusions, thereby reducing the identity switches and fragmentation issues in MOT. The experimental results showed that it improves the overall MOT performance.
(iii) 
SSIM association update:
The SSIM introduced in the proposed model can be considered as a second opinion when an ambiguity in association occurs. Figure 7 and Table 3 show the importance of SSIM association by evaluating the model on MOT metrics with the MOT17 and MOT20 validation data sequences. As the performance of the data association algorithm improves, we will obtain better association estimation, which will enhance the tracker’s tracking performance. It is observed from the results that the SSIM enhances the performance of the data association algorithm. It reduces the false negatives and identity switches and, hence, the MOTA. Also, the high IDF1 score validates the significance of SSIM association in the refinement of the association matrix.

4.3. MOT Benchmark Evaluation

This section shows the experimental evaluation of the proposed method on the benchmark datasets. Table 4 summarizes and compares our results with state-of-the-art algorithms on MOT benchmark datasets and Table 5 on UA-DETRAC. Here, we also show the effects of systematically adding neighborhood detection estimation, non-sequential training, and SSIM association update in the proposed tracker.
The benchmark evaluation results show that the proposed MOT framework performs very well in terms of the MOT evaluation metrics. We would like to emphasize that the metric scores for identity switches, fragmentation, and false negatives are reduced, indicating the reduction in the wrong association among detection target pairs. This results in better accuracy (MOTA). Also, the IDF1 score is improved, which is a clear indication of the association accuracy. This shows the robustness and efficiency of the proposed data association method.
We compared our results with recent state-of-the-art methods. The benchmark evaluation result depicts that the proposed data association method outperforms the state-of-the-art DAN model [30]. In particular, the nearest neighborhood estimation employed for detection–target feature pair selection reduces the association mismatch and improves the computational efficiency. The post-processing steps after deepFAN also help enhance the association accuracy and reduce the computational complexity. Here, the employment of the SSIM reduces the ambiguity in the data association.
The tracking results of the proposed tracker with the UA-DETRAC dataset are summarized in Table 5. Here, we opted for the EB detector [63] for a fair comparison. Since the trackers in Table 5 used different detectors, the name of the tracker is given along with the detector used. The proposed method gives better results on the UA-DETRAC evaluation compared with other approaches and can also be effectively used for vehicle tracking.

5. Conclusions

Developing a better data association framework is very crucial for robust multiple object tracking. This research work proposes two important contributions to enhance the data association. The first one is by introducing neighborhood detection estimation (NDE) only to retain reliable detection–target pairs. Secondly, the SSIM association component is proposed to reject ambiguous associations with high or near high association scores. A comprehensive evaluation strategy was adopted to understand and study the impacts of our technical contributions on popular multiple object tracking benchmarks. Further, we carried out a systematic ablation study to pinpoint the benefits of each proposal. We compared our proposals with recent multiple object tracking frameworks. Our studies found that the proposed tracker gave very low identity switches, which is one of the crucial factors in ranking various trackers. Further, the proposed tracker also achieved very high overall MOTA and IDF1 scores. Another factor that we wish to highlight here is that the proposed framework rejects ambiguous associations and employs only the neighboring detections for data associations. Ultimately, this leads to achieving higher tracking speed, which is another important factor in multiple object tracking. In the future, we would like to deploy this tracker in real-time tracking scenarios by augmenting a dedicated object-detection module along with the proposed tracker for real-world applications.

Author Contributions

A.P. developed the approach for this research and drafted this manuscript. D.M. helped revise the drafted manuscript and provided valuable advice for this research. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The source code has been uploaded to GitHub, which can be accessed at the following link: “https://github.com/aswathyiist123/MOT_deepFAN_SSIM (accessed on 1 July 2024)”.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Tian, Y.; Dehghan, A.; Shah, M. On detection, data association and segmentation for multi-target tracking. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 2146–2160. [Google Scholar] [CrossRef]
  2. Wen, L.; Du, D.; Li, S.; Bian, X.; Lyu, S. Learning nonuniform hypergraph for multi-object racking. arXiv 2018, arXiv:1812.03621. [Google Scholar]
  3. Sheng, H.; Zhang, Y.; Chen, J.; Xiong, Z.; Zhang, J. Heterogeneous association graph fusion for target association in multiple object tracking. IEEE Trans. Circuits Syst. Video Technol. 2018, 29, 3269–3280. [Google Scholar] [CrossRef]
  4. Felzenszwalb, P.F.; Girshick, R.B.; McAllester, D.; Ramanan, D. Object Detection with Discriminatively Trained Part-Based Models. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 1627–1645. [Google Scholar] [CrossRef]
  5. He, K.; Zhang, X.; Ren, S. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef]
  6. Yang, F.; Choi, W.; Lin, Y. Exploit all the layers: Fast and accurate CNN object detector with scale dependent pooling and cascaded rejection classifiers. In Proceedings of the IEEE Conference Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2129–2137. [Google Scholar]
  7. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  8. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015. [Google Scholar]
  9. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards realtime object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28, 91–99. [Google Scholar] [CrossRef]
  10. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  11. Kobayashi, T. Structured Feature Similarity with Explicit Feature Map; National Institute of Advanced Industrial Science and Technology Umezono: Tsukuba, Japan, 2016. [Google Scholar]
  12. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image Quality Assessment: From Error Visibility to Structural Similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
  13. Premaratne, P.; Premaratne, M. New Structural Similarity Measure for Image Comparison. Commun. Comput. Inf. Sci. 2012, 304, 292–297. [Google Scholar]
  14. Luo, W.; Xing, J.; Milan, A.; Zhang, X.; Liu, W.; Zhao, X.; Kim, T.K. Multiple Object Tracking: A Literature Review. arXiv 2017, arXiv:1409.7618v4. pp. 1–18. [Google Scholar] [CrossRef]
  15. Ciaparrone, G.; Sánchez, F.L.; Tabik, F.L.S.; Troiano, L.; Tagliaferri, R.; Herrera, F. Deep learning in video multi-object tracking: A survey. Neurocomputing 2020, 381, 61–88. [Google Scholar] [CrossRef]
  16. Park, Y.; Dang, L.M.; Lee, S.; Han, D.; Moon, H. Multiple Object Tracking in Deep Learning Approaches: A Survey. Electronics 2021, 10, 2406. [Google Scholar] [CrossRef]
  17. Kim, C.; Li, F.; Ciptadi, A.; Rehg, J.M. Multiple hypothesis tracking revisited. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 4696–4704. [Google Scholar]
  18. Zhang, S.; Lan, X.; Yao, H.; Zhou, H.; Tao, D.; Li, X. A Biologically Inspired Appearance Model for Robust Visual Tracking. IEEE Trans. Neural Netw. Learn. Syst. 2016, 28, 2357–2370. [Google Scholar] [CrossRef]
  19. Nam, H.; Han, B. Learning Multi-Domain Convolutional Neural Networks for Visual Tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 4293–4302. [Google Scholar]
  20. Breitenstein, M.D.; Reichlin, F.; Leibe, B.; Koller-Meier, E.; VanGool, L. Robust tracking-by-detection using a detector confidence particle filter. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision (ICCV), Kyoto, Japan, 27 September–4 October 2009; pp. 1515–1522. [Google Scholar]
  21. Kutschbach, T.; Bochinski, E.; Eiselein, V.; Sikora, T. Sequential sensor fusion combining probability hypothesis density and kernelized correlation filters for multi-object tracking in video data. In Proceedings of the 2017 14th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Lecce, Italy, 29 August–1 September 2017. [Google Scholar]
  22. Fu, Z.; Feng, P.; Naqvi, S.M.; Chambers, J.A. Particle PHD filter based multi-target tracking using discriminative groupstructured dictionary learning. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 4376–4380. [Google Scholar]
  23. Bae, S.H.; Yoon, K.J. Confidence-Based Data Association and Discriminative Deep Appearance Learning for Robust Online Multi-Object Tracking. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 595–610. [Google Scholar] [CrossRef] [PubMed]
  24. Wen, L.; Li, W.; Yan, J.; Lei, Z.; Yi, D.; Li, S.Z. Multiple target tracking based on undirected hierarchical relation hypergraph. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014; Volume 1, pp. 1282–1289. [Google Scholar]
  25. Schulter, S.; Vernaza, P.; Choi, W.; Chandraker, M. Deep network flow for multi-object tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6951–6960. [Google Scholar]
  26. Feichtenhofer, C.; Pinz, A.; Zisserman, A. Detect to Track and Track to Detect. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; Volume 2017, pp. 3057–3065. [Google Scholar]
  27. Insafutdinov, E.; Andriluka, M.; Pishchulin, L.; Tang, S.; Levinkov, E.; Andres, B.; Schiele, B. rtTrack: Articulated Multi-person Tracking in the Wild. arXiv 2017, arXiv:1612.01465. [Google Scholar]
  28. Kim, C.; Li, F.; Rehg, J.M. Multi-object tracking with neural gating using bilinear lstm. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 200–215. [Google Scholar]
  29. Zhang, D.; Zheng, Z.; Wang, T.; He, Y. HROM: Learning High-Resolution Representation and Object-Aware Masks for Visual Object Tracking. Sensors 2020, 20, 4807. [Google Scholar] [CrossRef] [PubMed]
  30. Sun, S.; Akhtar, N.; Song, H.; Mian, A.; Shah, M. Deep affinity network for multiple object tracking. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 104–119. [Google Scholar] [CrossRef] [PubMed]
  31. Emami, P.; Pardalos, P.M.; Elefteriadou, L.; Ranka, S. Machine Learning Methods for Solving Assignment Problems in Multi Target Tracking. arXiv 2018, 1–35. [Google Scholar]
  32. Murphey, R.A.; Pardalos, P.M.; Pitsoulis, L.S. A greedy randomized adaptive search procedure for the multitarget multisensor tracking problem. Netw. Des. Connect. Facil. Locat. 1997, 40, 277–302. [Google Scholar]
  33. Berclaz, J.; Fleuret, F.; Turetken, E.; Fua, P. Multiple object tracking using k-shortest paths optimization. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 1806–1819. [Google Scholar] [CrossRef]
  34. Yang, M.; Wu, Y.; Jia, Y. A hybrid data association framework for robust online multi-object tracking. IEEE Trans. Image Process. 2017, 26, 5667–5679. [Google Scholar] [CrossRef]
  35. Choi, W. Near-online multi-target tracking with aggregated local flow descriptor. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR’15), Boston, MA, USA, 7–12 June 2015; pp. 3029–3037. [Google Scholar]
  36. Milan, A.; Schindler, K.; Roth, S. Multi-target tracking by discrete-continuous energy minimization. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 2054–2068. [Google Scholar] [CrossRef] [PubMed]
  37. Yang, B.; Huang, C.; Nevatia, R. Learning affinities and dependencies for multi-target tracking using a CRF model. In Proceedings of the 2011 IEEE Conference on Computer Vision and Pattern Recognition (CVPR’11), Colorado Springs, CO, USA, 20–25 June 2011; pp. 1233–1240. [Google Scholar]
  38. Meyer, F.; Braca, P.; Willett, P.; Hlawatsch, F. Tracking an unknown number of targets using multiple sensors: A belief propagation method. In Proceedings of the 19th International Conference on Information Fusion (FUSION’16), Heidelberg, Germany, 5–8 July 2016; pp. 719–726. [Google Scholar]
  39. Chen, L.; Wainwright, M.J.; Cetin, M.; Willsky, A.S. Data association based on optimization in graphical models with application to sensor networks. Math. Comput. Model. 2006, 43, 1114–1135. [Google Scholar] [CrossRef]
  40. Oh, S.; Russell, S.; Sastry, S. Markov chain Monte Carlo data association for general multiple-target tracking problems. In Proceedings of the 43rd IEEE Conference on Decision and Control (CDC’04), Bahamas, Nassau, 14–17 December 2004; Volume 1, pp. 735–742. [Google Scholar]
  41. Pasula, H.; Russell, S.; Ostland, M.; Ritov, Y. Tracking many objects with many sensors. In Proceedings of the 1999 International Joint Conference on Artificial Intelligence (IJCAI’99), Stockholm, Sweden, 31 July–6 August 1999; Volume 99, pp. 1160–1171. [Google Scholar]
  42. Chu, P.; Ling, H. FAMNet: Joint Learning of Feature, Affinity and Multi-Dimensional Assignment for Online Multiple Object Tracking. In Proceedings of the 2019 IEEE International Conference on Computer Vision (ICCV’19), Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  43. Xu, Y.; Ban, Y.; Alameda-Pineda, X.; Horaud, R. DeepMOT: A Differentiable Framework for Training Multiple Object Trackers. arXiv 2019, arXiv:1906.06618. [Google Scholar]
  44. Zhu, J.; Yang, H.; Liu, N.; Kim, M.; Zhang, W.; Yang, M.H. Online multi-object tracking with dual matching attention networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 366–382. [Google Scholar]
  45. Ondruska, P.; Posner, I. Deep Tracking: Seeing Beyond Seeing Using Recurrent Neural Networks. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence (AAAI’16), Phoenix, AZ, USA, 12–17 February 2016; AAAI Press: Washington, DC, USA, 2016; pp. 3361–3367. [Google Scholar]
  46. Steenkiste, S.V.; Chang, M.; Greff, K.; Schmidhuber, J. Relational neural expectation maximization: Unsupervised discovery of objects and their interactions. arXiv 2018, arXiv:1802.10353. [Google Scholar]
  47. Fernando, T.; Denman, S.; Sridharan, S.; Fookes, C. Tracking by prediction: A deep generative model for mutli-person localisation and tracking. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; pp. 1122–1132. [Google Scholar]
  48. Leal-Taixe, L.; Milan, A.; Reid, I.; Roth, S.; Schindler, K. MOTChallenge 2015: Towards a Benchmark for Multi-Target Tracking. arXiv 2015, arXiv:1504.01942. [Google Scholar]
  49. Milan, A.; Leal-Taixe, L.; Reid, I.; Roth, S.; Schindler, K. MOT16: A Benchmark for Multi-Object Tracking. arXiv 2016, arXiv:1603.00831. [Google Scholar]
  50. Wen, L.; Du, D.; Cai, Z.; Lei, Z.; Chang, M.-C.; Qi, H.; Lim, J.; Yang, M.-H.; Lyu, S. UA-DETRAC: A New Benchmark and Protocol for Multi-Object Detection and Tracking. Comput. Vis. Image Underst. 2015, 193, 102907. [Google Scholar] [CrossRef]
  51. Lyu, S.; Chang, M.C.; Du, D.; Wen, L.; Qi, H.; Li, Y.; Wei, Y.; Ke, L.; Hu, T.; Del Coco, M.; et al. UA-DETRAC 2018: Report of AVSS2018 & IWT4S challenge on advanced traffic monitoring. In Proceedings of the 15th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Auckland, New Zealand, 27–30 November 2018. [Google Scholar]
  52. Bouguet, J.Y. Pyramidal implementation of the affine Lucas Kanade feature tracker description of the algorithm. Intel Corp. 2001, 5, 1–10. [Google Scholar]
  53. Munkres, J. Algorithms for the Assignment and Transportation Problems. J. Soc. Ind. Appl. Math. 1957, 5, 32–38. [Google Scholar] [CrossRef]
  54. Bernardin, K.; Stiefelhagen, R. Evaluating Multiple Object Tracking Performance: The CLEAR MOT Metrics. EURASIP J. Image Video Process. 2008, 2008, 246309. [Google Scholar] [CrossRef]
  55. Brasó, G.; Leal-Taixé, L. Learning a Neural Solver for Multiple Object Tracking. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 6246–6256. [Google Scholar]
  56. Stadler, D.; Beyerer, J. Improving Multiple Pedestrian Tracking by Track Management and Occlusion Handling. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 10953–10962. [Google Scholar] [CrossRef]
  57. Stadler, D.; Beyerer, J. Multi-Pedestrian Tracking with Clusters. In Proceedings of the 2021 17th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Virtual, 16–19 November 2021. [Google Scholar]
  58. Ren, W.; Wang, X.; Tian, J.; Tang, Y.; Chan, A.B. Tracking-by-Counting: Using Network Flows on Crowd Density Maps for Tracking Multiple Targets. IEEE Trans. Image Process. 2020, 30, 1439–1452. [Google Scholar] [CrossRef] [PubMed]
  59. Bergmann, P.; Meinhardt, T.; Leal-Taixe, L. Tracking without bells and whistles. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  60. Yoon, Y.C.; Kim, D.Y.; Song, Y.M.; Yoon, K.; Jeon, M. Online Multiple Pedestrian Tracking Using Deep Temporal Appearance Matching Association. Inf. Sci. 2021, 561, 326–351. [Google Scholar] [CrossRef]
  61. Zhang, Y.; Sheng, H.; Wu, Y.; Wang, S.; Lyu, W.; Ke, W.; Xiong, Z. Long-Term Tracking with Deep Tracklet Association. IEEE Trans. Image Process. 2020, 29, 6694–6706. [Google Scholar] [CrossRef]
  62. Papakis, I.; Sarkar, A.; Karpatne, A. GCNNMatch: Graph Convolutional Neural Networks for Multi-Object Tracking via Sinkhorn Normalization. arXiv 2020, arXiv:abs/2010.00067. [Google Scholar]
  63. Wang, L.; Lu, Y.; Wang, H.; Zheng, Y.; Ye, H.; Xue, X. Evolving boxes for fast vehicle detection. In Proceedings of the IEEE International Conference on Multimedia and Expo, Hong Kong, China, 10–14 July 2017; pp. 1135–1140. [Google Scholar]
  64. Cai, Z.; Saberian, M.; Vasconcelos, N. Learning complexity aware cascades for deep pedestrian detection. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 3361–3369. [Google Scholar]
  65. Bochinski, E.; Eiselein, V.; Sikora, T. High-Speed tracking-by-detection without using image information. In Proceedings of the IEEE 2017 14th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Lecce, Italy, 29 August–1 September 2017. [Google Scholar]
  66. Bae, S.H.; Yoon, K.J. Robust online multi-object tracking based on tracklet confidence and online discriminative appearance learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2014 (CVPR), Columbus, OH, USA, 23–28 June 2014; pp. 1218–1225. [Google Scholar]
Figure 1. Representation of the proposed MOT framework: Inputs to the framework are the current frame I f and the centers of the object detections C D f ; the output is the estimated trajectory of the all the targets for frames till I f . The two main proposals in the framework are neighborhood detection estimation (NDE) and data association framework with deepFAN and SSIM. The detection feature matrix F D f obtained from the deep feature extractor and matrix with existing targets feature vector F T L f 1 are given to NDE to find reliable detection-target pairs and encode them on 3D tensor Φ . The input to the data association network is Φ , and the output is the association matrix, A R N T L × N d ; the scalar scores, A ( i , j ) represents the association score between the j t h detection and i t h target. Trajectory list T updated using association matrix A for I f .
Figure 1. Representation of the proposed MOT framework: Inputs to the framework are the current frame I f and the centers of the object detections C D f ; the output is the estimated trajectory of the all the targets for frames till I f . The two main proposals in the framework are neighborhood detection estimation (NDE) and data association framework with deepFAN and SSIM. The detection feature matrix F D f obtained from the deep feature extractor and matrix with existing targets feature vector F T L f 1 are given to NDE to find reliable detection-target pairs and encode them on 3D tensor Φ . The input to the data association network is Φ , and the output is the association matrix, A R N T L × N d ; the scalar scores, A ( i , j ) represents the association score between the j t h detection and i t h target. Trajectory list T updated using association matrix A for I f .
Jimaging 10 00171 g001
Figure 2. Approach to train the deep feature association network (deepFAN): Though an online MOT framework, the feature extractor is deployed as a single-stream model, and during training, it is considered as a two-stream network with shared parameters. The inputs to the feature extractor are the those two frames ( I f and I f p , which are p frames apart, meaning they need not be adjacent frames), and we need to find the association between the detections and the centers of the object detections ( C f and C f p for I f and I f p ). Since the input frames are non-adjacent, neighborhood detection estimation (NDE) is not valid and is not applied in the training pipeline. With the supervision of ground truth G f p , f , the cost function J is computed, and the weights of the deep compression network in deepFAN are updated.
Figure 2. Approach to train the deep feature association network (deepFAN): Though an online MOT framework, the feature extractor is deployed as a single-stream model, and during training, it is considered as a two-stream network with shared parameters. The inputs to the feature extractor are the those two frames ( I f and I f p , which are p frames apart, meaning they need not be adjacent frames), and we need to find the association between the detections and the centers of the object detections ( C f and C f p for I f and I f p ). Since the input frames are non-adjacent, neighborhood detection estimation (NDE) is not valid and is not applied in the training pipeline. With the supervision of ground truth G f p , f , the cost function J is computed, and the weights of the deep compression network in deepFAN are updated.
Jimaging 10 00171 g002
Figure 3. Performance analysis of proposed MOT framework with different values for NDE distance threshold, T e . MOTA and IDF1 with different values for distance threshold T e on the MOT17 validation dataset are evaluated to find the optimum value of T e .
Figure 3. Performance analysis of proposed MOT framework with different values for NDE distance threshold, T e . MOTA and IDF1 with different values for distance threshold T e on the MOT17 validation dataset are evaluated to find the optimum value of T e .
Jimaging 10 00171 g003
Figure 4. Analysis of neighborhood detection estimation on MOT17 and MOT20 validation sequences.
Figure 4. Analysis of neighborhood detection estimation on MOT17 and MOT20 validation sequences.
Jimaging 10 00171 g004
Figure 5. Performance analysis of proposed MOT framework with different values for threshold T a . The MOTA and IDF1 with different values for threshold T a on the MOT17 validation dataset are evaluated to find the optimum value of T a .
Figure 5. Performance analysis of proposed MOT framework with different values for threshold T a . The MOTA and IDF1 with different values for threshold T a on the MOT17 validation dataset are evaluated to find the optimum value of T a .
Jimaging 10 00171 g005
Figure 6. Analysis of deep feature association network on MOT17 and MOT20 validation sequences.
Figure 6. Analysis of deep feature association network on MOT17 and MOT20 validation sequences.
Jimaging 10 00171 g006
Figure 7. Analysis of Structural Similarity Index Metric on MOT17 and MOT20 validation sequences.
Figure 7. Analysis of Structural Similarity Index Metric on MOT17 and MOT20 validation sequences.
Jimaging 10 00171 g007
Table 1. Architectural details of the deep compression network. Here, we use stride = 1 and ReLU activation in each layer. BN indicates batch normalization, and Y/N denotes whether B is applied or not.
Table 1. Architectural details of the deep compression network. Here, we use stride = 1 and ReLU activation in each layer. BN indicates batch normalization, and Y/N denotes whether B is applied or not.
IndexInput ChannelsOutput ChannelsBN
01040512Y
3512256Y
6256128Y
912864N
11641N
Table 2. Training and validation data sequences for ablation study of proposed MOT framework on the MOT17 and MOT20 benchmark.
Table 2. Training and validation data sequences for ablation study of proposed MOT framework on the MOT17 and MOT20 benchmark.
TrainingValidation
MOT15
TUD-StadtmitteTUD-Campus
ETH-BahnhofETH-Sunnyday, ETH-Pedcross2
ADL-Rundle-6ADL-Rundle-8, Venice-2
KITTI-13KITTI-17
MOT 17
MOT17-02MOT17-04
MOT17-05MOT17-09
MOT17-10MOT17-11, MOT17-13
MOT 20
MOT20-01MOT20-02
MOT20-03MOT20-05
Table 3. Analysis of the proposed framework on the MOT validation datasets and comparison with different proposed tracker variants by disabling different components. (The best values are in boldface. ↑ indicates that a higher value is better, and ↓ indicates a lower value is better.)
Table 3. Analysis of the proposed framework on the MOT validation datasets and comparison with different proposed tracker variants by disabling different components. (The best values are in boldface. ↑ indicates that a higher value is better, and ↓ indicates a lower value is better.)
TrackerMOTA↑IDF1↑MT↑ML↓FP↓FN↓Recall↑Precision↑FAF↓IDSw↓Frag↓
MOT20
Proposed MOT53.651.031.628.112,094135,54855.485.51.212641853
Proposed MOT without NDE51.249.329.630.714,296154,78052.079.31.215732098
Proposed MOT trained with sequential frames52.650.930.929.412,944139,54755.395.61.514631921
Proposed MOT without SSIM51.748.927.331.214,991141,70353.197.42.122343084
MOT17
Proposed MOT61.665.634.424.9836176,12362.188.81.1695969
Proposed MOT without NDE56.363.633.225.6966286,42558.785.71.28341167
Proposed MOT trained with sequential frames59.264.233.425.2869177,45359.387.41.17261098
Proposed MOT without SSIM54.762.332.426.510,23189,65361.588.41.58321217
MOT15
Proposed MOT46.262.322.215.4213410,16255.485.50.8121286
Proposed MOT without NDE38.271.918.020.0309410,94352.079.31.0236341
Proposed MOT trained with sequential frames43.274.221.116.4265210,29554.882.41.0134291
Proposed MOT without SSIM38.173.119.221.8285111,34353.682.11.1216312
Table 4. Comparison of the proposed MOT framework on the MOT test dataset with state-of-the-art trackers. (Red for the best values and blue for second place. NA represents the values that are not available in the publications. ↑ indicates that a higher value is better, and ↓ indicates a lower value is better.)
Table 4. Comparison of the proposed MOT framework on the MOT test dataset with state-of-the-art trackers. (Red for the best values and blue for second place. NA represents the values that are not available in the publications. ↑ indicates that a higher value is better, and ↓ indicates a lower value is better.)
TrackerMOTA↑IDF1↑MT↑ML↓FP↓FN↓Recall↑Precision↑FAF↓IDSw↓Frag↓
MOT20
Proposed MOT58.959.730.719.231,063158,87661.789.24.518423126
Proposed MOT without NDE55.457.929.220.733,473161,87560.588.15.621874125
Proposed MOT trained with sequential frames57.558.630.919.831,974159,17458.688.54.919253215
Proposed MOT without SSIM56.157.628.720.133,542168,65467.987.75.322133982
MPN Track [55]57.659.138.222.516,953201,38461.194.93.812101420
TMOH [56]60.161.246.717.838,043165,89967.890.28.523424326
MPTC [57]60.659.751.116.745,318153,97870.288.910.145335163
TBC [58]54.550.133.419.737,937195,24262.389.58.524492580
MOT17
Proposed MOT58.660.824.129.120,230212,34559.593.90.811221943
Proposed MOT without NDE56.458.322.631.921,237222,74654.092.50.917822153
Proposed MOT trained with sequential frames57.860.123.529.920,934215,14557.593.30.813521986
Proposed MOT without SSIM54.952.820.931.122,237229,44755.992.91.126683469
DAN [30]52.449.521.430.725,423234,59258.476.9NA843114,797
Tractor++ [59]53.552.349.536.612,201248,04756.096.30.720724611
DMAN [44]48.255.719.338.326,218263,60853.392.01.521945378
DEEP TAMA [60]50.353.519.237.525,479252,99655.292.41.421923978
FAMNet [42]52.048.718.133.414,138253,61655.195.60.830725318
TT17 [61]54.963.124.438.120,236233,29558.794.21.110882392
MOT15
Proposed MOT52.560.033.825.8683721,21864.885.31.2370784
Proposed MOT without NDE49.458.829.428.210,77423,20464.184.61.26281090
Proposed MOT trained with sequential frames51.159.031.127.6807021,29264.684.71.7677922
Proposed MOT without SSIM47.547.823.426.4953125,50258.583.71.310401350
MPN Track [55]51.558.631.225.9762021,78064.683.91.3375872
Tracker++ [59]46.647.618.227.9462426,89656.288.20.812901702
GNN Match [62]46.743.221.828.2664325,31158.884.51.18201371
Table 5. Comparison of the proposed MOT framework on the test dataset, UA-DETRAC, with state-of-the-art trackers. (Red for the best values and blue for second place. NA represents the values that are not available in the publications.)
Table 5. Comparison of the proposed MOT framework on the test dataset, UA-DETRAC, with state-of-the-art trackers. (Red for the best values and blue for second place. NA represents the values that are not available in the publications.)
UA-DETRAC
TrackerPR-MOTA↑PR-MOTP↑PR-MT↑BPR-FPPR-FN↓PR-IDSw↓BHz
EB [63] + Proposed MOT23.430.917.516.88253.617,532.6462.2721.112.7
EB + Proposed MOT without NDE21.128.516.517.99757.519,572.9537.9794.510.9
EB + Proposed MOT trained with sequential frames21.729.116.918.19034.218,854.5489.1774.112.1
EB + Proposed MOT without SSIM20.828.116.018.710,054.320,834.6549.7875.414.5
EB + DAN [30]20.226.314.518.19747.8135,978.1518.2NA6.3
compACT [64] + FAMNet [42]19.836.717.118.214,988.6164,432.6617.4970.2NA
EB + IOUT [65]19.428.917.718.414,796.5171,806.82311.32445.9NA
R-CNN [7] + IOUT16.038.313.820.722,535.1193,041.95029.45795.7NA
compACT + CMOT [66]12.636.116.118.657,885.9167,110.8285.31516.83.8
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Prasannakumar, A.; Mishra, D. Deep Efficient Data Association for Multi-Object Tracking: Augmented with SSIM-Based Ambiguity Elimination. J. Imaging 2024, 10, 171. https://doi.org/10.3390/jimaging10070171

AMA Style

Prasannakumar A, Mishra D. Deep Efficient Data Association for Multi-Object Tracking: Augmented with SSIM-Based Ambiguity Elimination. Journal of Imaging. 2024; 10(7):171. https://doi.org/10.3390/jimaging10070171

Chicago/Turabian Style

Prasannakumar, Aswathy, and Deepak Mishra. 2024. "Deep Efficient Data Association for Multi-Object Tracking: Augmented with SSIM-Based Ambiguity Elimination" Journal of Imaging 10, no. 7: 171. https://doi.org/10.3390/jimaging10070171

APA Style

Prasannakumar, A., & Mishra, D. (2024). Deep Efficient Data Association for Multi-Object Tracking: Augmented with SSIM-Based Ambiguity Elimination. Journal of Imaging, 10(7), 171. https://doi.org/10.3390/jimaging10070171

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop