Next Article in Journal
A Bio-Inspired Vibration Energy Harvesting System with Internal Resonance and Slapping Mechanism for Enhanced Low-Frequency Power Generation
Previous Article in Journal
Differentiating Human Falls from Daily Activities Using Machine Learning Methods Based on Accelerometer and Altimeter Sensor Fusion Feature Engineering
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Object Tracking with Confidence-Based Trajectory Prediction Scheme

1
Intelligent Policing Key Laboratory of Sichuan Province, Luzhou 646000, China
2
College of Computer Science, Sichuan University, Chengdu 610065, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(23), 7221; https://doi.org/10.3390/s25237221
Submission received: 24 October 2025 / Revised: 19 November 2025 / Accepted: 21 November 2025 / Published: 26 November 2025
(This article belongs to the Section Sensing and Imaging)

Abstract

Multi-Object Tracking (MOT) aims to associate multiple objects across consecutive video sequences and maintain continuous and stable trajectories. Currently, much attention has been paid to data association problems, where many methods filter detection boxes for object matching based on the confidence scores (CS) of the detectors without fully utilizing the detection results. Kalman filter (KF) is a traditional means for sequential frame processing, which has been widely adopted in MOT. It matches and updates a predicted trajectory with a detection box in video. However, under crowded scenes, the noise will create low-confidence detection boxes, causing identity switch (IDS) and tracking failure. In this paper, we thoroughly investigate the limitations of existing trajectory prediction schemes in MOT and prove that KF can still achieve competitive results in video sequence processing if proper care is taken to handle the noise. We propose a confidence-based trajectory prediction scheme (dubbed ConfMOT) based on KF. The CS of the detection results is used to adjust the noise during updating KF and to predict the trajectories of the tracked objects in videos. While a cost matrix (CM) is constructed to measure the cost of successful matching of unreliable objects. Meanwhile, each trajectory is labeled with a unique CS, while the lost trajectories that have not been updated for a long time will be removed. Our tracker is simple yet efficient. Extensive experiments have been conducted on mainstream datasets, where our tracker has exhibited superior performance to other advanced competitors.

1. Introduction

MOT has been highly spotlighted recently; it deals with the detection of multiple objects across video frames and identifier assignment for each trajectory. It plays a vital role in intelligent video surveillance, military surveillance, autonomous driving, etc. Over the past few years, MOT has benefited greatly from the rapid development of advanced object detection schemes. Primarily, the tracking-by-detection (TBD) paradigm has achieved tremendous success in MOT. Most existing works divide MOT into detection and association tasks and solve them independently under the TBD framework [1,2]. Typically, they first apply object detectors to detect known targets and infer the trajectories via different matching strategies for cross-frame data association. For instance, MOTRv2 [3] integrates YOLOX with the MOTR tracker, where YOLOX creates high-quality object proposals to facilitate the association process by MOTR. However, TBD has a notable problem: the tracking performance is highly reliant on the detection results (i.e., the core impact on the tracking performance is the threshold of the detector and subsequent matching mechanisms). Many existing works only retain detection results above the threshold for association and tracking and discard detection results below this threshold directly. But a fixed threshold of the detector has difficulty coping with complex situations where multiple objects interact. Simply discarding those objects will cause a large number of missed detections, which are irreversible, reducing overall tracking performance. Classical TBD-based methods include SORT [4], DeepSORT [1], and ByteTrack [2]. Among them, SORT [4] simply uses intersection over union (IoU) and motion information to measure matching similarity, and it creates new trajectories and rejects lost ones. DeepSORT [1] employs a matching cascade and calculates Mahalanobis distance between predicted track boxes and detection boxes (instead of IoU), since the accuracy of the inactive tracks decreases over time. However, the Mahalanobis distance is only used to avoid irrational assignments, which is not suitable for matching. ByteTrack [2] associates the low-confidence detections that are unmatched in the first stage, which indeed pushes the record of MOT to a new level and therefore has been adopted by many works. However, high-confidence detections are usually preferred over low-confidence ones, while low-confidence detections will no longer be assigned to inactive trajectories. Fortunately, with the continuous development of object detection, the detection results are becoming increasingly reliable. Under such cases, we deal with the detections with different CS in a separate way. We match the high-confidence trajectories in the first stage and remedy the potential trajectories with low CS in the second stage.
Apart from detection, frequent object occlusions under crowded scenes are still the major challenge for MOT [5], causing IDS. Various appearance models with effective feature learning methods are employed to calculate feature similarity so as to correct object ID when occlusions occur [6]. TransCenter [7] advocates the use of image-related dense detection queries and efficient sparse tracking queries under their query learning networks (QLNs). AMtrack [8] is built upon an encoder–decoder Transformer architecture, which realizes data association across frames via evolving a set of track predictions. However, it has a slower inference speed, and its predictive ability on linear motion is not as good as that of KF. DeNoising-MOT [9] learns the denoising process in an encoder–decoder structure and develops a cascaded mask strategy to prevent mutual suppression between neighboring trajectories. However, as the scenes become more complex, the tracking performances of the above methods will still be severely affected by occlusion. Although recent improvements in detector performance are conducive to correct object associations, even the best detectors make false predictions. There remains room for improvement in existing methods using detections.
The linear KF algorithm has long been widely used in visual object tracking [2,10]. During the update process, the measurement noise is set to a constant value. For a reliable detection result, the measurement noise should be low accordingly. To obtain more accurate update results, we adjust the measurement noise based on the CS (output by the detector) so as to ensure the quality of the tracking results. During the matching process between the prediction and detection boxes, IoU and feature similarities are usually used to form CM (to measure the discrepancy between the predicted target position and the detection box) of the Hungarian algorithm. When computing CM, the reliability of the current bounding box is often ignored. We therefore fuse the CS into CM to improve its reliability. In some cases, some trajectories exist for only a period of time but discontinue in the following frames, which disturbs the subsequent trajectory matching and causes incorrect association. We regard them as lost and hence delete them in due course. Through analyzing the connection between the detection and associations, we believe the detection results have not been fully utilized. Our results show that the performance of the current TBD-based methods can be further improved by combining the CS of the detection results. Therefore, an appropriate detection processing scheme from a tracking perspective is required to bridge the gap between detection and tracking applications.
In a nutshell, we have made contributions in three key stages of MOT, including data association, trajectory prediction, and trajectory management:
1.
In the data association stage, confidence scores are integrated into the cost matrix to improve matching accuracy and achieve more reliable assignment results.
2.
In the trajectory prediction stage, a confidence-based trajectory prediction scheme has been proposed based on KF, which achieves more accurate prediction performance by controlling the measurement noise in KF.
3.
In the trajectory management stage, a trajectory deletion scheme has been proposed to determine the duration of trajectories and delete less reliable trajectories to avoid possible incorrect matches.
The experimental result shows that our tracking scheme further improves the performance of the current TBD-based methods. On MOT Challenge benchmarks, ConfMOT ranks among the top on MOT17 [11] and MOT20 [12]. ConfMOT achieves 64.5% HOTA and 80.5% MOTA on the MOT17 dataset, and 62.9% HOTA and 78.3% MOTA on the MOT20.
The rest of the paper is organized as follows: Section 2 briefly reviews related work, Section 3 describes our proposed method in detail. Experimental results are provided in Section 4 with ablation studies. Finally, a conclusion is drawn in Section 5.

2. Related Works

With the rapid development of object detection [13], the current research hotspots of visual tracking [2,14] have been shifted from designing complex and individual trackers to subsequent data association schemes that are built upon advanced detectors. The main goal is to maintain the correct trajectories of the objects.

2.1. Motion Models

Most of the recent TBD-based methods are based on motion models (under the assumption of constant velocity). As one of the classical models, KF is a Bayes filter that predicts and updates in a recursive way. The ground-truth state is an unobserved Markov process, while the measurement is observed from a hidden Markov model. The measurement noise is the noise that exists in the signal observation process. OC-SORT [14] uses virtual trajectories to smooth parameters for solving the cumulative error of KF. ByteTrack [2], on the other hand, uses a normal linear KF to associate every detection box to recover true objects and filter out the background elements. However, the above two methods apply the same KF to all targets, achieving trajectory prediction quickly and efficiently, but ignoring the differences between various detections. You et al. [15] propose a novel Spatial–Temporal Topology-based Detector (STTD) algorithm and introduce a topology structure for the dynamics of moving targets. It indeed reduces false positives, but since it only considers group motion between targets to build topology, it hence overlooks camera motion. Li et al. [6] present a TBD-based framework for MOT, where a deep association network is developed (followed by detection) to infer the association degree of objects. Despite some advantages in correlation operation, the object detection and tracking tasks are somewhat independent of each other, where two deep feature extractions are needed in each stage. GIAOTracker [16] proposes a modified KF with confidence scores. Khurana et al. [17] adopt depth estimates from a monocular depth estimator to forecast the trajectories of occluded people. However, they rely heavily on appearance features, which introduces a high computational cost. Wang et al. [18] published a novel approach to tackle long-term vehicle tracking without appearance information. But it cannot be extended to other types of objects. Tracktor++ [19] employs camera motion compensation for frame alignment. The former method performs global association by developing an appearance-free link model to address missing detection and missing association problems. While the latter employs global information and optimization strategies to associate object trajectories. But it is not applicable to online real-time tracking.

2.2. Cost Matrix

In order to update the objects’ trajectories, the prediction boxes obtained from KF prediction phase need to be matched with the detection boxes obtained from the detector. OC-SORT [14] claims that using IoU alone is sufficient to generate good results if the detector is accurate enough. But it ignores the impact of detection differences on the cost matrix. MOTDT [20] constructs the cost matrix of the tracking processes with appearance features and IoU respectively. FairMOT [21], on the other hand, fuses IoU and appearance feature, but it does not consider the negative impact of an unreliable cost matrix. The above 2 methods combine multiple indicators to form a cost matrix, which can solve the confused assignment problem to some extent. However, the extraction of appearance information also slows down the tracking speed in return.

2.3. Trajectories Management

Most trackers set different durations for lost trajectories for different datasets. For example, JDE [10] and FairMOT [21] save the unmatched trajectories for 30 frames, while ByteTrack [2] keeps lost trajectories for 14 frames for MOT17-06 and 25 frames for MOT17-14 according to the length of the videos. However, these parameter settings require pre-processing of the video, since the time of disappearing trajectory varies due to different occlusion situations. It would be impractical to make different settings manually for different datasets. From a practical perspective, a unified trajectory processing scheme should be developed that is applicable to different datasets, while different time intervals should be assigned to different kinds of lost trajectories.

3. Method

The block diagram of our tracker is drawn in Figure 1. Basically, the CS is incorporated into the tracking process, where the detection boxes with high and low scores are taken care of by the first and the second association processes, respectively, to ensure correct associations of trajectories.

3.1. Confidence-Based Adaptive KF

KF (with a constant velocity) is commonly adopted in object tracking tasks in building a motion model. It follows the Markov assumption that the current system state is only related to the state of its previous moment, which is irrelevant to other earlier states. Then the previous state vector and the state observed at the current frame are needed to estimate the current state. Specifically, at time t, the hidden state of an object X t is written as
x t = [ u t , v t , a t , h t , u ˙ t , v ˙ t , a ˙ t , h ˙ t ] T ,
where ( u t , v t ) denotes the center position of the bounding box, h t denotes the height of the detection box, a t is the aspect ratio of the width and the height. u ˙ t , v ˙ t , a ˙ t , h ˙ t represent their corresponding derivatives.
KF consists of a prediction step and an update step. In the prediction step, KF generates the estimated state variables (along with their uncertainties), which will then be updated with a weighted average of the estimated state and measurement. Since there is no active control in multi-object tracking, the state vector at frame t is represented by
x ^ t | t 1 = F t x t 1 | t 1 , P t | t 1 = F t P t 1 | t 1 F t T + Q t ,
P t | t 1 is the covariance matrix of the posterior state. F t is the state transition matrix and Q t is the covariance of noise, which are calculated as
F t = I 4 × 4 I 4 × 4 O 4 × 4 I 4 × 4 , Q t = diag ( ( δ p w t 1 | t 1 ) 2 , ( δ p h t 1 | t 1 ) 2 , ( δ p w t 1 | t 1 ) 2 , ( δ p h t 1 | t 1 ) 2 , ( δ v w t 1 | t 1 ) 2 , ( δ v h t 1 | t 1 ) 2 , ( δ v w t 1 | t 1 ) 2 , ( δ v h t 1 | t 1 ) 2 )
δ v and δ p are the parameters of the noise. Meanwhile, the observation state vector (i.e., measurement by the detector) is expressed by
z t = [ z x , z y , z w , z h ] T = H x t + r t , H = [ I 4 × 4 O 4 × 4 ] ,
Here r t is the measurement noise. H is the observation matrix. For classical KF, the update process is implemented by computing the Kalman gain as
K t = P t | t 1 H T ( H P t | t 1 H T + R t ) 1 ,
Then the current object state and covariance matrix of the posterior state are updated as
x t | t = x ^ t | t 1 + K t ( z t H x ^ t | t 1 ) , P t | t = ( I K t H ) P t | t 1 ,
For existing methods (e.g., SORT [4], ByteTrack [2] and OC-SORT [14]), the covariance of the measurement noise R t is a fixed value for all detections. However, detection reliability varies significantly: low-confidence detections frequently contain larger localization errors. Assigning the same R t to all detections forces KF to over-trust unreliable observations, resulting in trajectory drift and ID switches. To mitigate this issue, we introduce an adaptive scheme based on the confidence score. The update process of KF is expressed as
R t = ( 1 c t ) R ,
where c t ∈ [0, 1] is the confidence score of the matched detection and R is the base covariance of the measurement noise. Therefore, Equation (5) will be modified into
K t = P t | t 1 H T ( H P t | t 1 H T + ( 1 c t ) R ) 1 ,
As can be seen, a high-confidence detection ( c t → 1) makes a lower R t and higher Kalman gain, while the update process relies heavily on the measurement. Conversely, a low-confidence detection ( c t → 0) makes a higher R t but a lower Kalman gain, while the update process will be affected more by the predicted state, preventing unreliable detections from destroying the trajectories.
Our proposed adaptive mechanism enables KF to dynamically balance prediction and measurement based on the confidence scores of detection, thereby improving robustness against noisy detections and reducing ID switches in crowded or occluded environments.

3.2. Data Association

The posterior state of frame t  ( x ^ t | t 1 ) is propagated from frame t     1 through KF. The predicted bounding boxes are then used to compute IoU with all detections at frame t to form CM. Then the trajectories will be associated with the detection boxes via the Hungarian algorithm based on the CM and will be updated via KF.
The detection boxes with low confidence scores are regarded as unreliable, which may contain large localization errors and have a higher cost for successful matching. If they contribute equally to CM, they may incorrectly alter trajectory estimates. To alleviate this issue, ConfMOT rectifies IoU-based costs by using the CS of each detection as follows:
C i , j = 1 1 + S j 2 × ( 1 c i , j i o u ) ,
where S j is the CS of the jth detection and c i , j i o u is the IoU between the ith predicted box and the jth detection. The above equation assigns a higher cost to unreliable detections (low S j ) and therefore lowers their chance of being selected during the matching process and encourages high-confidence detections to dominate the association process.

3.3. Deletion Strategy

In previous methods, the lost trajectories were usually kept for a certain number of frames, which may be recovered if matches were found or otherwise deleted. However, the trajectories with lower CSs are essentially less reliable. We therefore create track score for each trajectory by summing up the CsS of an object on its last tracklet. For a tracklet that lasts for n frames, its track score is expressed as
S t r a c k = i = 1 n S i ,
where S i represents the CS of an object in the ith frame of the tracklet. The lower the track score, the less reliable the trajectory will be. Such a trajectory will be assigned a shorter time of duration before it is regarded as lost. For the lost trajectories, we use the following equation to determine the time to delete them:
S del = min α S track + S last , 1 log 1 + β T umatch ,
where S l a s t is the CS of the last frame of the tracklet and T u m a t c h is the number of frames when the trajectory does not find a match in the following frames. α and β are two parameters.
The design of Equation (11) is quite intuitive: the first term min α S track + S last ,   1 depicts the reliability of a trajectory. α S track + S last aggregates both the accumulated CS along the tracklet and the CS of the last observation. It may become a large number due to a possibly large S track ; in that situation, the trajectory will not be able to be deleted. We therefore apply both α and min( , 1) to restrict it to [0, 1]. The addition of S last prevents newly generated but highly reliable trajectories from being penalized only because their durations are short. The second term log 1 + β T umatch imposes a time-dependent penalty on lost trajectories. It is monotonically increasing with T umatch . Since log operation increases slowly at the beginning and becomes more aggressively when T umatch becomes large. It allows the short-term occluded objects but deletes long unmatched trajectories. The parameters α and β realize the trade-off between the contribution of historical confidence and the speed of temporal decay.
Overall, S del is an interpretable confidence-like score. It assigns longer survival time to reliable trajectories or those that are only shortly unmatched, which shortens the unreliable trajectories or those that remain unmatched for a long time.

3.4. Tracking Procedure

The pseudo-code of tracking is shown in Algorithm 1, and the corresponding diagram has been drawn in Figure 2 to better understand the process. Firstly, we adopt YOLOX as our detector to obtain the bounding boxes and the corresponding CS, where the detection boxes are divided into two parts: D h i g h (yellow in Figure 2) and D l o w (green in Figure 2) based on confidence thresholds τ h i g h and τ l o w (which are set to 0.6 and 0.1, respectively, according to OC-SORT [14]). The bad detection boxes with CS lower than τ l o w are removed. Then, we use KF to predict the new locations of each trajectory in T a l l (labeled 1, 2, and 3 in Figure 2).
The first association is performed between the high score boxes D h i g h and all trajectories T a l l . We only use the IoU CM via Equation (9) and utilize the Hungarian algorithm to assign the detection boxes to corresponding trajectories. In particular, if the IoU cost between the detection box and the prediction box is greater than a threshold, then the matching will be rejected. The unmatched detections in the first association are kept in D r e m a i n , and the unmatched trajectories (e.g., No. 3 in Figure 2) are kept in T r e m a i n (lines 14 to 18 in Algorithm 1), which will be matched with the low score boxes in the second stage association.
In the second association, T r e m a i n will be associated with the low score boxes D l o w . The unmatched detection boxes in D l o w are regarded as the background, and unmatched trajectories are marked as lost and kept in T l o s t (lines 19 to 21 in Algorithm 1), which will be deleted after a certain period of time.
The successfully matched trajectories are updated by KF through Equation (8). For trajectories in T l o s t , their CS ( S d e l ) will be calculated via Equation (11). If they are lower than τ d e l , they will be moved from T l o s t to T d e l . If a lost trajectory is recovered, it will be moved from T l o s t to T , and the track score will be recalculated.
Algorithm 1: Pseudo-code of tracking
Sensors 25 07221 i001
Finally, for detection boxes in D r e m a i n , we initialize new trajectories if their CSs are higher than τ n e w and exist for two consecutive frames (lines 29 to 31 in Algorithm 1).

4. Experiment

4.1. Datasets

The training process of MOT17 [11] is conducted jointly on CrowHuman [22], CityPersons [23], ETHZ [24], and the training set of MOT17. And we use CrowdHuman and the training set of MOT20 [12] to train MOT20, while testing is carried out on MOT17 and MOT20 under the “private detection” protocol. For the ablation study, we combine CrowdHuman and half of the training set of MOT17 as the training set and test our tracker on the second half of MOT17. The initial parameters of our model are pre-trained on COCO [25].

4.2. Metrics

We adopt Higher-Order Tracking Accuracy (HOTA), Multi-Object Tracking Accuracy (MOTA), ID F1 Score (IDF1) (the ratio of correctly identified detections over the average number of ground-truth and computed detections), and Number of Identity Switches (IDS) as the main metrics to evaluate the performance of our model. Considering MOTA is mainly used to assess detection performance while IDF1 characterizes the ability to keep constant identities, we use HOTA as the final evaluation metric for ranking, which is a more comprehensive indicator to reflect the general performance of Detection Accuracy, association, and localization. In addition, we adopt Association Accuracy (AssA), Association Precision (AssPr), Association Recall (AssRe), Detection Accuracy (DetA), and Localization Accuracy (LocA) to further compare the trackers with similar performances. AssA, AssPr, and AssRe are used to measure association performance, while DetA is used for detection quality.

4.3. Implementation Details

All the experiments are implemented in PyTorch 1.13.1 and under NVIDIA Tesla V100 GPU. We use YOLOX as the detector with YOLOX as the backbone. The datasets and training strategy are described in Section 4.1. The optimizer is SGD with a weight decay of 5 × 10 4 and a momentum of 0.9, and the initial learning rate is set to 10 3 . The parameters of the tracker are the same as the baseline. The high- and low-confidence thresholds τ h i g h and τ l o w are empirically set to 0.6 and 0.1, respectively. If the CSs of unmatched detections are higher than 0.7 [2], we initialize new trajectories starting from them. The lost trajectories will be deleted if their CSs are lower than 0.1. GSI is adopted as a post-processing method.

4.4. Experimental Results

Extensive experiments were conducted on MOT17 [11] and MOT20 [12] to testify to the effectiveness of our tracker.

4.4.1. Results on MOT17

The video sequences in MOT17 are filmed by both static and moving cameras. The comparative results on MOT17 are listed in Table 1. Generally, ByteTrack [2], OC-SORT [14], SCTrack [26], etc., and ConfMOT outperform other trackers with some margins. Among them, ConfMOT surpasses ByteTrack by 1.4% and 2.0% in HOTA and MOTA, respectively, with lower IDS as well. Meanwhile, we lead OC-SORT in HOTA, MOTA, and IDF1. Our obvious advantage in HOTA is attributed to the proposed early trajectory deletion strategy, which ensures the tracking accuracy.
To further investigate the subtle differences among ByteTrack [2], OC-SORT [14], and our tracker, we compare another five metrics: AssA, AssPr, AssRe, DetA and LocA. As mentioned in Section 4.2, the first three metrics reflect the data association ability of a tracker, while DetA and LocA demonstrate the accuracies of object detection and localization, respectively. As shown in Table 2, except that we lag behind OC-SORT in AssPr, ConfMOT outperforms the other two trackers in all other metrics. The reason is that OC-SORT is stronger than our trackers in association ability, but weaker in detection. It is worth noting that ConfMOT ranks first in AssRe, which reflects the accurate prediction of the target trajectories.
A group of visualization results of ConfMOT on the test set of MOT17 is shown in Figure 3. The MOT17-01 scene is a normal outdoor scenario with pedestrians walking around. Apparently, the color of the bounding boxes does not change or swap, indicating constant identities during the tracking process. MOT17-03 is a crowded outdoor scene, in which our tracker still completes the tracking task without obvious miss detections or ID switches. MOT17-06 is a normal street scene, and it is worth noting that the old man with dark blue box number 27 (in the middle of the scene) always remains unchanged after several overlaps by other pedestrians, which indicates the ability of our tracker to deal with occlusions. In MOT17-07, the video is taken by moving cameras, and there is a man sitting in the left corner in the first frame, who is moving closer in the following frames with a constant ID. MOT17-8 shows a tracking scenario with varying lighting conditions. Still, our tracker demonstrates robustness in tracking multiple targets with fixed IDS. Lastly, MOT17-14 is a video sequence containing small targets that is captured on a moving box. Clearly, we can correctly detect and locate the small targets over a moving viewpoint.

4.4.2. Results on MOT20

MOT20 is a more challenging benchmark with crowded scenes. As shown in Table 1, ConfMOT ranks first in HOTA, MOTA and IDF1 due to our proposed deletion strategies, but it also causes slightly higher IDS, since the lost tracks are difficult to match. Like MOT17, we also compare AssA, AssPr, AssRe, DetA, and LocA among OC-SORT, ByteTrack, and ConfMOT. As shown in Table 3, ConfMOT surpasses both OC-SORT and ByteTrack in all five metrics, which reflects stronger association accuracy and more precise localization. The deletion of unreliable lost trajectories in the early stage ensures the overall tracking performances, while the integration of the confidence-based cost matrix guarantees the robustness of the tracker during the long period of the tracking process.
A group of visualization results of ConfMOT on the test set of MOT20 is shown in Figure 4, including indoor and outdoor scenarios under different lighting conditions. As mentioned earlier, MOT20 includes crowded scenes. MOT20-04 is an outdoor scene at night, while MOT20-06 and MOT20-08 are daytime scenarios. No matter under the scene, our tracker maintains excellent tracking performances, which is less sensitive to lighting conditions, and the ID of each target remains largely unchanged.

4.4.3. Comprehensive Analysis

As mentioned earlier, ByteTrack [2], OC-SORT [14] and our tracker achieve better results than the other competitors. Here, we further analyze the advantages/disadvantages of the three methods. ByteTrack [2] utilizes pure IoU to process all the detections. In contrast, we process the detection boxes using the cost matrix, which is adjusted by the confidence score. Our scheme matches the high-quality detection boxes in a more effective way, resulting in more stable assignment results. As a result, we achieve the best overall secondary indexes (including AssA, AssPr, AssRe, DetA and LocA). Besides, we obtain a more precise object location due to the proposed modified KF, which has been verified by the highest DetA (Detection Accuracy) we have reached. In summary, we have achieved the best overall results on MOT-17 and on MOT-20.

4.4.4. Visualization of Trajectories

To further demonstrate the superiority of our tracker, a comparison among ByteTrack [2], OC-SORT [14] and ConfMOT is made in Figure 5, in which the trajectory of each object of the last 50 frames is drawn with different colors. Obviously, OC-SORT (shown in Figure 5a) has many sporadic segments of trajectories, which reflect target loss under the crowded scene, while it has almost no trajectories on the left-hand side of the scene. ByteTrack (shown in Figure 5b) has more trajectories, but there are some zigzags and stagnation, indicating incorrect matching and ID switches. In comparison, as shown in Figure 5c, the trajectories of ConfMOT appear more stable, with almost identical trajectory lengths and no broken segments. In addition, the colors of the trajectories do not change within the dense areas, indicating the robustness of ConfMOT under crowded scenes.

4.5. Ablation Studies

In the ablation experiment, we use YOLOX as the backbone and train our model jointly on CrowdHuman and the first half of the training set of MOT17, while the second half of the training set of MOT17 is used for validation. In this section, we will firstly verify the effectiveness of the core modules of our tracker, including KF, CM+CS, and the deletion strategy (DS). The comparative results on the MOT17 validation set is shown in Table 4.

4.5.1. KF Based on the Confidence Score

As shown in Table 4, when KF (with the CS) is integrated into the tracking network, we enjoy a slight increase in both HOTA and IDF1. By incorporating confidence, KF assigns weights to different detections based on their CS, making it more robust to noise and uncertain detections. During the update of KF, the high-quality detection boxes will be assigned higher weights than the prediction boxes, making the detection boxes more dependent on the detection results. Conversely, for low-quality detection results, the detection boxes will depend more on the prediction boxes, which will be assigned higher weights.

4.5.2. Deletion Strategy Based on the Confidence Score

As shown in Table 4, when the proposed lost trajectory deletion strategy has been added to the tracking architecture, the HOTA and IDF1 values increase by 0.2% and 0.4%, respectively. These results illustrate the fact that when a lost trajectory has been deleted and the subsequent detected target cannot find a match, and therefore, a new trajectory will be created. Moreover, the increased IDF1 suggests that the deleted trajectories are indeed unreliable, and our deletion strategy reduces the likelihood of incorrect matches between a detected target and the wrong trajectory. Intuitively, if a trajectory remains unmatched for a certain period of time and is re-matched again, it is quite possible that it matches with a different target (instead of the previous one). Therefore, our deletion strategy becomes very necessary to avoid target drifting caused by re-matching.

4.5.3. Cost Matrix Based on the Confidence Score

When calculating the matching cost between the prediction and detection boxes, such a cost will be higher for low-quality detection boxes. As shown in Table 4, after the CS is incorporated into CM, both HOTA and MOTA increase by 0.4%. This result indicates that the integration of the confidence score indeed reduces the probability of matching low-quality detection boxes and continuation of incorrect trajectories.

5. Conclusions

In this paper, we thoroughly analyze the limitations of current TBD-based trackers and present ConfMOT to ensure the stable continuation of trajectories among different objects. To realize the goal, we adopt YOLOX as our detector to obtain the detection results. We adopt a two-stage association strategy and predict the trajectories based on the confidence score and Kalman filtering technique. Then, a cost matrix is constructed to measure the cost of unreliable matching. Finally, a deletion strategy is proposed to determine the duration of a trajectory and delete the less reliable trajectories. Extensive experiments have been conducted on the MOT17 and MOT20 datasets, along with ablation studies to testify to the efficacy of our tracker. The results prove that our trackers have obvious advantages in the HOTA metric compared to other methods, due to the cost matrix and trajectory deletion strategy we designed. However, the performance of our tracker in IDS is mediocre (ID switch caused by occlusion), since we delete unreliable trajectories at an early stage.
In future work, we will be engaged in developing a light-weight appearance feature extraction module to improve the robustness of the model in dealing with fast appearance changes. We will also employ specific motion modeling schemes to ensure the stability and continuation of the target trajectory.

Author Contributions

K.Y.: Funding, Resources, Conceptualization; J.L.: Methodology, Validation, Visualization; Y.Z.: Administration, Writing, Supervision. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Intelligent Policing Key Laboratory of Sichuan Province, No. ZNJW2024KFMS004.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Wojke, N.; Bewley, A.; Paulus, D. Simple online and realtime tracking with a deep association metric. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 3645–3649. [Google Scholar]
  2. Zhang, Y.; Sun, P.; Jiang, Y.; Yu, D.; Weng, F.; Yuan, Z.; Luo, P.; Liu, W.; Wang, X. Bytetrack: Multi-object tracking by associating every detection box. In Proceedings of the Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, 23–27 October 2022; Proceedings, Part XXII. Springer: Berlin/Heidelberg, Germany, 2022; pp. 1–21. [Google Scholar]
  3. Zhang, Y.; Wang, T.; Zhang, X. Motrv2: Bootstrapping end-to-end multi-object tracking by pretrained object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 22056–22065. [Google Scholar]
  4. Bewley, A.; Ge, Z.; Ott, L.; Ramos, F.; Upcroft, B. Simple online and realtime tracking. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 3464–3468. [Google Scholar]
  5. Sun, L.; Zhang, J.; Gao, D.; Fan, B.; Fu, Z. Occlusion-aware visual object tracking based on multi-template updating Siamese network. Digit. Signal Process. 2024, 148, 104440. [Google Scholar] [CrossRef]
  6. Li, H.; Liu, Y.; Liang, X.; Yuan, Y.; Cheng, Y.; Zhang, G.; Tamura, S. Multi-object tracking via deep feature fusion and association analysis. Eng. Artif. Intell. 2023, 124, 106527. [Google Scholar] [CrossRef]
  7. Xu, Y.; Ban, Y.; Delorme, G.; Gan, C.; Rus, D.; Alameda-Pineda, X. Transcenter: Transformers with dense representations for multiple-object tracking. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 7820–7835. [Google Scholar] [CrossRef] [PubMed]
  8. Liu, Z.; Huang, X.; Sun, J.; Zhang, X. AMtrack: Anti-occlusion multi-object tracking algorithm. Signal Image Video Process. 2024, 18, 9305–9318. [Google Scholar] [CrossRef]
  9. Fu, T.; Wang, X.; Yu, H.; Niu, K.; Li, B.; Xue, X. Denoising-mot: Towards multiple object tracking with severe occlusions. In Proceedings of the 31st ACM International Conference on Multimedia, Ottawa, ON, Canada, 29 October–3 November 2023; pp. 2734–2743. [Google Scholar]
  10. Wang, Z.; Zheng, L.; Liu, Y.; Li, Y.; Wang, S. Towards real-time multi-object tracking. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XI 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 107–122. [Google Scholar]
  11. Milan, A.; Leal-Taixé, L.; Reid, I.; Roth, S.; Schindler, K. Mot16: A benchmark for multi-object tracking. arXiv 2016, arXiv:1603.00831. [Google Scholar] [CrossRef]
  12. Dendorfer, P.; Rezatofighi, H.; Milan, A.; Shi, J.; Cremers, D.; Reid, I.; Roth, S.; Schindler, K.; Leal-Taixé, L. Mot20: A benchmark for multi object tracking in crowded scenes. arXiv 2020, arXiv:2003.09003. [Google Scholar] [CrossRef]
  13. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. Yolox: Exceeding yolo series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar] [CrossRef]
  14. Cao, J.; Pang, J.; Weng, X.; Khirodkar, R.; Kitani, K. Observation-centric sort: Rethinking sort for robust multi-object tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 9686–9696. [Google Scholar]
  15. You, S.; Yao, H.; Xu, C. Multi-object tracking with spatial-temporal topology-based detector. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 3023–3035. [Google Scholar] [CrossRef]
  16. Du, Y.; Wan, J.; Zhao, Y.; Zhang, B.; Tong, Z.; Dong, J. Giaotracker: A comprehensive framework for mcmot with global information and optimizing strategies in visdrone 2021. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 2809–2819. [Google Scholar]
  17. Khurana, T.; Dave, A.; Ramanan, D. Detecting invisible people. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 3174–3184. [Google Scholar]
  18. Wang, G.; Gu, R.; Liu, Z.; Hu, W.; Song, M.; Hwang, J.-N. Track without appearance: Learn box and tracklet embedding with local and global motion patterns for vehicle tracking. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 9876–9886. [Google Scholar]
  19. Bergmann, P.; Meinhardt, T.; Leal-Taixe, L. Tracking without bells and whistles. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 941–951. [Google Scholar]
  20. Chen, L.; Ai, H.; Zhuang, Z.; Shang, C. Real-time multiple people tracking with deeply learned candidate selection and person re-identification. In Proceedings of the 2018 IEEE International Conference on Multimedia and Expo (ICME), San Diego, CA, USA, 23-27 July 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–6. [Google Scholar]
  21. Zhang, Y.; Wang, C.; Wang, X.; Zeng, W.; Liu, W. Fairmot: On the fairness of detection and re-identification in multiple object tracking. Int. Comput. Vis. 2021, 129, 3069–3087. [Google Scholar] [CrossRef]
  22. Shao, S.; Zhao, Z.; Li, B.; Xiao, T.; Yu, G.; Zhang, X.; Sun, J. Crowdhuman: A benchmark for detecting human in a crowd. arXiv 2018, arXiv:1805.00123. [Google Scholar] [CrossRef]
  23. Zhang, S.; Benenson, R.; Schiele, B. Citypersons: A diverse dataset for pedestrian detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3213–3221. [Google Scholar]
  24. Ess, A.; Leibe, B.; Schindler, K.; Gool, L.V. A mobile vision system for robust multi-person tracking. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; IEEE: Piscataway, NJ, USA, 2008; pp. 1–8. [Google Scholar]
  25. Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part V 13. Springer: Berlin/Heidelberg, Germany, 2014; pp. 740–755. [Google Scholar]
  26. Li, H.; Qin, S.; Li, S.; Gao, Y.; Wu, Y. Synergistic-aware cascaded association and trajectory refinement for multi-object tracking. Image Vis. Comput. 2025, 162, 105695. [Google Scholar] [CrossRef]
  27. Psalta, A.; Tsironis, V.; Karantzalos, K. Transformer-based assignment decision network for multiple object tracking. Comput. Vis. Image Underst. 2024, 241, 103957. [Google Scholar] [CrossRef]
  28. Li, Y.-F.; Ji, H.-B.; Zhang, W.-B.; Lai, Y.-K. Learning discriminative motion models for multiple object tracking. IEEE Trans. Multimed. 2024, 26, 11372–11385. [Google Scholar] [CrossRef]
  29. Sun, P.; Cao, J.; Jiang, Y.; Zhang, R.; Xie, E.; Yuan, Z.; Wang, C.; Luo, P. Transtrack: Multiple object tracking with transformer. arXiv 2020, arXiv:2012.15460. [Google Scholar]
  30. Cai, J.; Xu, M.; Li, W.; Xiong, Y.; Xia, W.; Tu, Z.; Soatto, S. Memot: Multi-object tracking with memory. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 8090–8100. [Google Scholar]
  31. Gao, R.; Wang, L. Memotr: Long-term memory-augmented transformer for multi-object tracking. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 4–6 October 2023; pp. 9901–9910. [Google Scholar]
  32. Luo, R.; Song, Z.; Ma, L.; Wei, J.; Yang, W.; Yang, M. Diffusiontrack: Diffusion model for multi-object tracking. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 26–27 February 2024; Volume 38, pp. 3991–3999. [Google Scholar]
  33. Hu, M.; Zhu, X.; Wang, H.; Cao, S.; Liu, C.; Song, Q. Stdformer: Spatial-temporal motion transformer for multiple object tracking. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 6571–6594. [Google Scholar] [CrossRef]
  34. Yu, E.; Li, Z.; Han, S.; Wang, H. Relationtrack: Relation-aware multiple object tracking with decoupled representation. IEEE Trans. Multimed. 2022, 25, 2686–2697. [Google Scholar] [CrossRef]
  35. Zhou, C.; Jiang, M.; Kong, J. Bgtracker: Cross-task bidirectional guidance strategy for multiple object tracking. IEEE Trans. Multimed. 2023, 25, 8132–8144. [Google Scholar] [CrossRef]
  36. Liu, Y.; Wu, J.; Fu, Y. Collaborative tracking learning for frame-rate-insensitive multi-object tracking. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 4–6 October 2023; pp. 9964–9973. [Google Scholar]
  37. Chen, D.; Shen, H.; Shen, Y. Jdt-nas: Designing efficient multi-object tracking architectures for non-gpu computers. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 7541–7553. [Google Scholar] [CrossRef]
  38. Deng, K.; Zhang, C.; Chen, Z.; Hu, W.; Li, B.; Lu, F. Jointing recurrent across-channel and spatial attention for multi-object tracking with block-erasing data augmentation. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 4054–4069. [Google Scholar] [CrossRef]
  39. Kong, J.; Mo, E.; Jiang, M.; Liu, T. Motfr: Multiple object tracking based on feature recoding. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 7746–7757. [Google Scholar] [CrossRef]
  40. Wang, Q.; Zheng, Y.; Pan, P.; Xu, Y. Multiple object tracking with correlation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 3876–3886. [Google Scholar]
  41. Chu, P.; Wang, J.; You, Q.; Ling, H.; Liu, Z. Transmot: Spatial-temporal graph transformer for multiple object tracking. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2023; pp. 4870–4880. [Google Scholar]
  42. Stadler, D.; Beyerer, J. Modelling ambiguous assignments for multi-person tracking in crowds. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 133–142. [Google Scholar]
  43. Lv, W.; Zhang, N.; Zhang, J.; Zeng, D. One-shot multiple object tracking with robust id preservation. IEEE Trans. Circuits Syst. Video Technol. 2023, 34, 4473–4488. [Google Scholar] [CrossRef]
  44. Seidenschwarz, J.; Brasó, G.; Serrano, V.C.; Elezi, I.; Leal-Taixé, L. Simple cues lead to a strong multi-object tracker. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 13813–13823. [Google Scholar]
  45. Yan, S.; Wang, Z.; Huang, Y.; Liu, Y.; Liu, Z.; Yang, F.; Lu, W.; Li, D. GGSTrack: Geometric graph with spatio-temporal convolution for multi-object tracking. Neurocomputing 2025, 653, 131234. [Google Scholar] [CrossRef]
  46. Zhao, T.; Yang, G.; Li, Y.; Lu, M.; Sun, H. Multi-object tracking using score-driven hierarchical association strategy between predicted tracklets and objects. Image Vis. Comput. 2024, 152, 105303. [Google Scholar] [CrossRef]
  47. Kim, V.; Jung, G.; Lee, S.-W. AM-SORT: Adaptable motion predictor with historical trajectory embedding for multi-object tracking. In Proceedings of the International Conference on Pattern Recognition and Artificial Intelligence, Jeju Island, Republic of Korea, 18–21 June 2024; pp. 92–107. [Google Scholar]
Figure 1. Block diagram of the proposed ConfMOT. We adopt a two-stage association scheme, where we predict and update object trajectories based on the confidence score of the detections.
Figure 1. Block diagram of the proposed ConfMOT. We adopt a two-stage association scheme, where we predict and update object trajectories based on the confidence score of the detections.
Sensors 25 07221 g001
Figure 2. Illustration of the processing flow of the proposed method. The high-confidence detections will be processed in the first-stage association. The low-confidence ones will be sent to the second-stage association. The unmatched trajectories will be deleted after a certain period of time.
Figure 2. Illustration of the processing flow of the proposed method. The high-confidence detections will be processed in the first-stage association. The low-confidence ones will be sent to the second-stage association. The unmatched trajectories will be deleted after a certain period of time.
Sensors 25 07221 g002
Figure 3. Demonstration of tracking results on the test set of MOT17; our tracker successively associates multiple objects under different scenes.
Figure 3. Demonstration of tracking results on the test set of MOT17; our tracker successively associates multiple objects under different scenes.
Sensors 25 07221 g003
Figure 4. Demonstration of tracking results on the test set of MOT20; our tracker successively associates multiple objects even in crowded scenes with fewer ID switches.
Figure 4. Demonstration of tracking results on the test set of MOT20; our tracker successively associates multiple objects even in crowded scenes with fewer ID switches.
Sensors 25 07221 g004
Figure 5. Comparison of trajectories generated by three tracking methods. Compared with ByteTrack and OC-SORT, we generate more stable trajectories with no broken segments.
Figure 5. Comparison of trajectories generated by three tracking methods. Compared with ByteTrack and OC-SORT, we generate more stable trajectories with no broken segments.
Sensors 25 07221 g005
Table 1. Evaluation on the test sets of MOT17 and MOT20. We compare our method with recent methods. The best results are shown in red and the second-best results are in blue.
Table 1. Evaluation on the test sets of MOT17 and MOT20. We compare our method with recent methods. The best results are shown in red and the second-best results are in blue.
MethodsMOT17MOT20
HOTAMOTAIDF1IDSFPSHOTAMOTAIDF1IDSFPS
TADN [27]-69.0%60.8%---68.7%61.0%--
DMMTracker [28]52.1%67.1%64.3%313516.148.7%62.5%60.5%20439.7
TransTrack [29]54.1%75.2%63.5%360310.048.5%65.0%59.4%36087.2
TransCenter [7]54.5%73.2%62.2%46141.043.5%61.9%50.4%46531.0
MeMOT [30]56.9%72.5%69.0%2724-54.1%63.7%66.1%1938-
AMtrack [8]58.6%74.4%71.5%4740-56.8%73.2%69.2%1870-
DNMOT [9]58.0%75.6%68.1%2529-58.6%70.5%73.2%987-
MeMOTR [31]58.8%72.8%71.5%-------
FairMOT [21]59.3%73.7%72.3%330325.954.6%61.8%67.3%524313.2
DiffusionTrack [32]60.8%77.9%73.8%3819-55.3%72.8%66.3%4117-
STDFormer-LMPH [33]60.9%78.4%73.1%5091-60.2%76.2%72.1%5245-
RelationTrack [34]61.0%73.8%74.4%13747.456.5%67.2%70.5%42432.7
BGTracker [35]61.0%75.6%73.8%373520.757.5%71.6%71.8%247112.8
ColTrack [36]61.0%78.8%73.9%1881------
JDT-NAS-T1 [37]-74.3%72.0%281813.3-----
DcMOT [38]61.3%74.5%75.2%268220.453.8%59.7%67.4%563610.6
MOTFR [39]61.8%74.4%76.3%265222.257.2%69.0%71.7%364813.3
CorrTracker [40]-76.5%73.6%336914.8-65.2%69.1%51838.5
TransMOT [41]-76.7%75.1%2346--77.5%75.2%1615-
MAA [42]62.0%79.4%75.9%1452189.157.3%73.9%71.2%133114.7
MOTRv2 [3]62.0%78.6%75.0%--61.0%76.2%73.1%--
PID-MOT [43]62.1%74.7%76.3%156319.757.0%67.5%71.3%10158.7
GHOST [44]62.8%78.7%77.1%2325-61.2%73.7%75.2%1264-
GGSTrack [45]62.8%80.2%-168958.061.8%75.1%-149815.3
ScoreMOT [46]63.0%79.8%76.7%400725.662.3%77.7%75.6%144016.2
ByteTrack [2]63.1%80.3%77.3%219629.661.3%77.8%75.2%122317.5
OC-SORT [14]63.2%78.0%77.5%195029.062.1%75.5%75.9%91318.7
AM-SORT [47]63.3%78.0%77.8%--62.0%75.5%76.1%--
SCTrack [26]63.5%79.4%77.7%2022-61.4%75.6%76.1%837-
ConfMOT64.5%80.5%79.3%198026.162.9%78.3%76.1%135915.2
Table 2. Further comparison of different tracking methods on the MOT17 dataset. The best results are highlighted in red and the second-best results are in blue.
Table 2. Further comparison of different tracking methods on the MOT17 dataset. The best results are highlighted in red and the second-best results are in blue.
TrackerAssA (%)↑AssPr (%)↑AssRe (%)↑DetA (%)↑LocA (%)↑
ByteTrack [2]62.076.068.264.583.0
OC-SORT [14]63.480.867.563.283.4
ConfMOT (Ours)63.877.970.064.983.4
Table 3. Further comparison with ByteTrack on the testing set of MOT20. The best results are shown in bold.
Table 3. Further comparison with ByteTrack on the testing set of MOT20. The best results are shown in bold.
TrackerAssPr↑AssRe↑DetA↑LocA↑
ByteTrack [2]59.6%74.6%66.2%63.4%83.6%
OC-SORT [14]60.5%75.1%67.1%64.2%83.9%
ConfMOT61.4%77.1%67.8%64.6%84.7%
Table 4. Ablation study on the MOT17 validation set. The best results are shown in bold.
Table 4. Ablation study on the MOT17 validation set. The best results are shown in bold.
KFDSCM+CSHOTA↑MOTA↑IDF1↑
67.8%77.9%79.6%
68.1%77.9%79.8%
68.3%78.0%80.2%
68.7%78.4%80.5%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yi, K.; Li, J.; Zhang, Y. Multi-Object Tracking with Confidence-Based Trajectory Prediction Scheme. Sensors 2025, 25, 7221. https://doi.org/10.3390/s25237221

AMA Style

Yi K, Li J, Zhang Y. Multi-Object Tracking with Confidence-Based Trajectory Prediction Scheme. Sensors. 2025; 25(23):7221. https://doi.org/10.3390/s25237221

Chicago/Turabian Style

Yi, Kai, Jiarong Li, and Yi Zhang. 2025. "Multi-Object Tracking with Confidence-Based Trajectory Prediction Scheme" Sensors 25, no. 23: 7221. https://doi.org/10.3390/s25237221

APA Style

Yi, K., Li, J., & Zhang, Y. (2025). Multi-Object Tracking with Confidence-Based Trajectory Prediction Scheme. Sensors, 25(23), 7221. https://doi.org/10.3390/s25237221

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop