Next Article in Journal
Comparative Analysis of Beamforming Techniques and Beam Management in 5G Communication Systems
Previous Article in Journal
Advances in Interferometric Synthetic Aperture Radar Technology and Systems and Recent Advances in Chinese SAR Missions
Previous Article in Special Issue
A Dynamic Kalman Filtering Method for Multi-Object Fruit Tracking and Counting in Complex Orchards
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Self-Supervised Visual Tracking via Image Synthesis and Domain Adversarial Learning

1
Guangzhou Institute of Technology, Xidian University, Guangzhou 510555, China
2
School of Electrical and Information Engineering, Yunnan Minzu University, Kunming 650504, China
3
School of Science, Harbin Institute of Technology, Shenzhen 518055, China
4
National Center for Applied Mathematics, Chongqing Normal University, Chongqing 401331, China
*
Authors to whom correspondence should be addressed.
Sensors 2025, 25(15), 4621; https://doi.org/10.3390/s25154621
Submission received: 19 June 2025 / Revised: 17 July 2025 / Accepted: 23 July 2025 / Published: 25 July 2025
(This article belongs to the Special Issue AI-Based Computer Vision Sensors & Systems)

Abstract

With the widespread use of sensors in applications such as autonomous driving and intelligent security, stable and efficient target tracking from diverse sensor data has become increasingly important. Self-supervised visual tracking has attracted increasing attention due to its potential to eliminate reliance on costly manual annotations; however, existing methods often train on incomplete object representations, resulting in inaccurate localization during inference. In addition, current methods typically struggle when applied to deep networks. To address these limitations, we propose a novel self-supervised tracking framework based on image synthesis and domain adversarial learning. We first construct a large-scale database of real-world target objects, then synthesize training video pairs by randomly inserting these targets into background frames while applying geometric and appearance transformations to simulate realistic variations. To reduce domain shift introduced by synthetic content, we incorporate a domain classification branch after feature extraction and adopt domain adversarial training to encourage feature alignment between real and synthetic domains. Experimental results on five standard tracking benchmarks demonstrate that our method significantly enhances tracking accuracy compared to existing self-supervised approaches without introducing any additional labeling cost. The proposed framework not only ensures complete target coverage during training but also shows strong scalability to deeper network architectures, offering a practical and effective solution for real-world tracking applications.

1. Introduction

Visual object tracking is an essential issue in the field of computer vision involving the task of estimating the state of the target in subsequent frames based on a given target object in the initial frame. The issue of achieving stable and efficient target tracking from these diverse sensor data has become a research hotspot in recent years with the wide application of various sensors such as cameras, infrared sensors, etc., in fields such as autonomous driving, intelligent security, and augmented reality. The proven success of transformer architectures [1] across multiple computer vision domains has made them a popular choice for the design of visual tracking algorithms [2,3,4]. Despite their success, transformer-based tracking methods necessitate a substantial volume of video data for effective training. Due to the expense and time-consuming nature of manually annotating data, it is crucial to investigate self-supervised visual target tracking as a viable alternative. Given the current widespread use of deep neural network models, there are two key issues that self-supervised learning needs to address: first, designing a pretext task that enables the model to learn discriminative feature representation of the tracked target; and second, finding more effective ways to train deep neural networks. Researchers have conducted extensive exploration to address both of these issues.
In the design of pretext tasks, existing self-supervised methods such as UDT [5], Self-SDCT [6], and CycleSiam [7] perform bidirectional tracking by randomly cropping the template region in the initial frame, then calculating the consistency loss of the targets in the initial frame before and after tracking as the supervision signal. On the other hand, the S2SiamFC tracker [8] generates training pairs by randomly cropping the template region in the image and pairing it with the original image. This approach utilizes an adversarial masking strategy to blur the appearance of the target, enabling the model to learn contextual information related to the target. Additionally, the USOT tracker [9] employs unsupervised optical flow and dynamic programming methods to crop moving objects instead of random cropping while adding a cycle memory learning scheme to enable the model can be trained on a longer time span. Despite the aforementioned methods using different cropping strategies in the initial frame, they can still fail to completely enclose the entire target and often capture only a part of it. Consequently, the model is unable to learn complete information related to the target, which leads to the complete target not being included during actual testing. Moreover, these models are highly susceptible to interference from similar objects, resulting in poor robustness.
As with other self-supervised methods such as MAST [10], differentiable memory modules have been proposed to enhance self-supervised tracking algorithms. Approaches such as JLSTC [11], ContrastCorr [12], and others [13,14,15,16] explore the leveraging of spatiotemporal contextual information in videos for self-supervised representation learning. In addition, Li et al. [17] synthesized training data by applying transformations on given labeled targets, thereby enhancing the robustness of the tracker. These approaches do not directly conduct end-to-end training on unlabeled data, and achieve commendable performance when applied to trackers; however, they are only suitable for training shallow networks. When applied to deep networks such as DiMP [18] and ToMP [2], their performance often significantly lags behind supervised learning methods.
Existing self-supervised methods often rely on incomplete object representations during training, resulting in inaccurate localization during inference. Moreover, their effectiveness significantly degrades when applied to deep neural networks. In this work, we propose a self-supervised visual tracking method via image synthesis and domain adversarial learning to solve the issues mentioned above. First, to tackle the problem of current methods being unable to crop a complete target in the initial frame, we introduce a real-world target and then synthesize it into both the initial and subsequent frames using image synthesis techniques. In this way, we establish an extensive target object database that can ensure the quality and quantity of synthesized targets. Examples of specific objects in the database are presented in Figure 1. Second, our proposed method is fully capable of training deep neural networks. By synthesizing targets in video frames, we can obtain complete and accurate target bounding boxes. This allow the model to be trained in a supervised manner, which can be applied to the training of deep networks. Third, in addition to directly using synthesized data for training, we employ domain adversarial training to reduce the difference between synthesized and real data. Such differences include obvious edges of synthesized targets and irregular color transitions between synthesized targets and original images. These differences may lead to the model overfitting on the training data, as learning these differences can cause the model to converge quickly without actually learning any target-related features. Through domain adversarial learning, the model can learn truly discriminative features related to the target. Finally, we fine-tune the tracker with a small amount of labeled data to further improve its tracking performance on downstream tasks.
Extensive experiments demonstrate that our self-supervised tracker achieves competitive tracking performance with supervised trackers without increasing the cost of data annotation. The main contributions of this work are summarized as follows:
  • We propose a novel self-supervised tracking framework that leverages image synthesis and domain adversarial learning to enable effective training of deep tracking networks by using complete target instances without manual annotations.
  • We incorporate domain adversarial learning into the training pipeline to reduce the domain gap between synthetic and real data, which mitigates overfitting and facilitates the extraction of robust and discriminative features.
  • Extensive experiments on multiple benchmarks demonstrate that our method achieves comparable performance to supervised methods.

2. Related Work

2.1. Visual Tracking Method

In recent years, single-stage visual trackers have become mainstream in the field of visual object tracking [19,20,21]. These models utilize a unified encoder to simultaneously perform feature extraction and feature interaction, simplifying the overall architecture and improving efficiency. OSTrack [22] was the first to successfully adopt this design, demonstrating significant gains in both tracking accuracy and speed. Building upon this foundation, subsequent works [23,24,25] have proposed various enhancements to the attention mechanism to further refine the interaction between the template and search regions. Despite their strong performance, these one-stage trackers often struggle with long-term tracking scenarios, particularly when the target undergoes significant appearance changes or occlusions. To address this limitation, several recent approaches [26,27,28] have introduced multi-template mechanisms, enabling the model to maintain memory of the target over time. For instance, ODTrack [28] learns the joint dependencies between multiple templates and the current search region within the encoder, while MambaLCT [29] compresses historical target representations into a compact set of feature tokens, which are then updated using the Mamba architecture, improving long-term reasoning capability. Similarly, LMTrack [30] retrieves template patches from previous frames and performs joint modeling with the current search region to enhance robustness in challenging conditions.
While these advanced trackers achieve state-of-the-art performance, their success relies heavily on large-scale annotated datasets. With the increasing complexity and parameter count of modern tracking models, the demand for labeled data has grown significantly, posing a barrier to scalability. In this paper, we propose a self-supervised training framework that leverages synthetic images to replace manually annotated data. This approach greatly reduces the dependency on human-labeled data, providing an efficient and scalable solution for training deep tracking models without compromising performance.

2.2. Self-Supervised Visual Tracking

In recent years, trackers [2,31,32] have become increasingly reliant on large-scale datasets for training (e.g., ODTrack [28], SWinTrack [33]), highlighting the advantages of self-supervised learning that eliminates the need for labeled data. Among existing self-supervised tracking approaches, several different methods [5,7,9,34] have adopted a strategy that randomly crops the template region in the initial frame as the target, then enforces consistency by computing a loss between the target appearance before and after bidirectional tracking. Other methods [11,12,34] exploit the spatiotemporal context across video frames; for instance by colorizing objects in future frames utilizing information extracted from previous frames. Although these methods have achieved some success, they fail to capture the complete target during training; as a result, the tracker often focuses on only a part of the target during inference, leading to degraded tracking performance.
Recently, several trackers such as SimTrack [35], STTrack [36], and MixFormer [37] have pioneered the use of transformer-based architectures and achieved remarkable performance. For this type of tracker, MAT [38] applies random masking to both the template and search regions and enforces mutual reconstruction based on their cross-region correlations. Diff-Tracker [39] leverages the rich knowledge embedded in pretrained diffusion models to perform unsupervised visual tracking, including their understanding of image semantics and structural information. The method proposed in [17] generates labeled data for training trackers by employing a crop–transform–paste strategy. ARPTrack [40] is conditioned on the target’s historical appearance to generate future video frames using a diffusion model, enabling unsupervised pretraining of the tracker on these synthesized sequences. Furthermore, methods have utilized the SAM model [41] to help reduce the use of labeled data thanks to its powerful performance and zero-shot generalization ability. For instance, SAM2MOT [42] directly applies the pretrained SAM to the multi-object tracking task without any fine-tuning while maintaining high performance and significantly reducing the demand for labeled data. The method proposed in [43] uses the SAM model to generate pseudolabels, helping to to lower the cost of labeled data.
Unlike the aforementioned methods, in this study we propose constructing a target object database. We then randomly select target objects from this database and combine them into video frames to obtain accurate target bounding boxes, enabling the use of complete targets to train the model in a supervised manner.

2.3. Image Synthesis Method

With the rapid development of deep learning, the demand for data in model training is increasing. However, acquiring labeled data is expensive and time-consuming. Constrained by the scarcity of training data, many researchers resort to image synthesis methods to augment training data [44,45,46]. Zhang et al. [44] utilized image synthesis to generate thermal infrared images, facilitating end-to-end training of thermal infrared trackers and alleviating the scarcity of thermal infrared image data. Li et al. [17] enhanced and transformed the targets in the first frame of each video sequence, synthesizing them into subsequent frames to enhance the tracker’s robustness against interference from similar objects. Kerim et al. [47] employed image synthesis to generate a person-tracking dataset under various adverse weather conditions, addressing the insufficient pedestrian data in harsh weather conditions. Khoreva et al. [48] synthesized plausible future video frames using annotations provided on the first frame of each video, enabling model training with only a small amount of labeled data. Song et al. [49] proposed a self-supervised object synthesis framework that can comprehensively solve the object synthesis task within a unified model while changing the perspective, geometric shape, color, and shadow of the generated objects without the need for manual annotation; Chen et al. [50] utilized artificial intelligence to synthesize data to address the problem of fault data in the railway system being difficult to access. These methods all directly utilize synthesized data or mix synthesized data with real data to train models, but ignore the differences between synthesized and real data.
In this work, we incorporate domain adversarial training into the regular training process to prevent models from quickly overfitting on synthesized data and failing to learn genuine discriminative features, which can help the model to effectively extract the most discriminative features.

2.4. Domain Adversarial Learning

Domain adversarial learning aims to reduce the distributional discrepancy between the source and target domains by learning domain-invariant representations. A representative approach in this field is the Domain Adversarial Neural Network [51], which introduces a gradient reversal layer to enable joint optimization of the feature extractor and domain classifier in an end-to-end manner. To improve training stability and generalization performance, the method proposed in [52] introduced Smooth Domain Adversarial Training, which regularizes the adversarial process and enhances the effectiveness of domain alignment. Zhou et al. [53] extended adversarial learning to the domain generalization setting by designing a learning objective that encourages the classifier to correctly predict labels while simultaneously deceiving the domain discriminator. In addition to conventional classification tasks, Uplavikar et al. [54] utilized adversarial learning to enhance underwater images by aligning visual characteristics across different aquatic environments, while Chen et al. [55] combined self-training with domain adversarial learning to address the limitation that traditional methods align feature distributions without ensuring their discriminative capacity in the target domain. From a theoretical perspective, recent work by [56] has introduced a novel generalization bound for domain adaptation based on the f-divergence. This development unifies and extends existing theoretical results, including those of DANNs. In the context of reinforcement learning, domain adversarial optimization has been employed to improve the robustness of learned representations against visual domain shifts, enabling zero-shot generalization to unseen environments [57]. In industrial settings, the Domain Adversarial Graph Convolutional Network [58] jointly models class labels, domain labels, and the structural relationships among data samples to tackle mechanical fault diagnosis under varying working conditions.
However, despite such a wide range of applications, domain adversarial learning has rarely been explored in tracking. In this work, we utilize domain adversarial learning to mitigate the differences between synthesized and real data in the process of self-supervised training. We hope that this can provide some inspiration for future exploration.

3. Methods

We propose a self-supervised training method for visual tracking based on image synthesis and domain adversarial training. Figure 2 illustrates the training process. Our method begins by constructing a target object database, which is used to synthesize training data with unlabeled video sequences. The construction process is described in detail in Section 3.1. Next, Section 3.2 elaborates on how we synthesize training pairs by embedding target objects from the database into video frames. These synthetic pairs allow us to perform self-supervised training using complete target objects, addressing the limitation that existing methods typically rely on partial or randomly cropped targets. Finally, Section 3.3 presents our integration of domain adversarial learning into the training pipeline from Section 3.2. Incorporating domain adversarial training effectively mitigates the discrepancy between synthetic and real-world images, enhancing the generalization capability of the tracker.

3.1. Construction of the Target Object Database

The idea of synthesizing a target object into a video frame to train the object tracking model using the complete target object in the unlabeled data is a straightforward one. Building upon this thought and aiming to bring the synthesized data closer to the real data, we propose synthesizing real-world targets into video frames for tracking. To achieve this, we collect target segmentation labels from CoCo [59] and YouTube-VOS [60] datasets and filter out labels that do not meet the requirements, such as targets that are incomplete or too small. Subsequently, we establish a database of about a million different target objects. Figure 1 illustrates some of the target objects in the database.
As can be seen from Figure 1, our target objects are all real objects in natural scenes, meaning that they can be directly used by the tracker as the tracking target. Using this target database, it is possible to train the tracker on any original video sequence. Specifically, we can select targets from the target object database and synthesize them into the video sequence, from which we obtain the bounding boxes of the targets in the video sequence. In this way, we transform the unlabeled video sequence into the labeled video sequence without any extra cost, after which the tracker can be trained according to the normal procedure.

3.2. Image Synthesis Method

The process of our image synthesis method is illustrated in the image synthesis part of Figure 2. The main idea involves randomly selecting a target from the target object database and simultaneously extracting two frames from the video sequence. Then, the target and the transformed target are synthesized into the two frames as template frames and test frames, respectively. Finally, the template frame and the test frame are paired to form a training pair. The detailed generation steps for the template frame and the test frame are described below.
For the generation of the template frame I t e m p , we randomly select a target z from the target object database Z and an original template frame x t from the video sequence X = { x i } . The subscripts i and t denote the respective frame sequence numbers. Next, we synthesize the target z into the video frame x t in a pasted fashion to obtain the final template frame for training. The paste operation is denoted as P ( · ) . In mathematical language, the above process can be described as follows:
z = S ( Z ) ,
x t = S ( X ) ,
I t e m p = P ( x t , z , ( x t , y t ) ) ,
where the function S ( · ) represents the random selection operation. For the image synthesis operation P ( x t , z , ( x t , y t ) ) , we first select a location ( x t , y t ) within the video frame for synthesizing. Then, the mask of the target object m a s k z is padded to match the size of the video frame based on the selected location. Finally, the synthetic image is generated as follows: I t e m p = m a s k z · z + ( 1 m a s k z ) · x t , where the value of mask z is 1 at the target position and 0 elsewhere.
The generation process of the test frames is similar to that of the template frames. First, we randomly extract the original test frame x ( t + n ) from the video sequence. To simulate the change of the target in the real world, we apply random transformation operations to targets before synthesizing them into the test frame. These operations include blurring, scaling, occlusion, deformation, and distractors (similar object interference). The synthesis process for the test frame can be represented as follows:
z = T j ( z ) ,
x t + n = S ( X ) ,
I t e s t = P ( x t + n , z , ( x t + n , y t + n ) ) ,
where the symbol z represents the target after applying the j-th transformation operation T j ( z ) and the letter n is an integer between 1 and 30 which ensures the sampling distance between the original template frame and the original test frame does not exceed 30 frames. An example of the image synthesis effect before and after target transformation is shown in Figure 3; for detailed parameter settings and strategies, please refer to the Experimental Details section.
Ultimately, our approach enables synthesizing target objects at designated locations in unlabeled video frames, thereby generating pseudo-labels and constructing training samples. This allows us to perform self-supervised training of the tracker using unlabeled video data. Unlike existing self-supervised methods that often rely on partial targets during training which leading to the model focusing on only a part region of the target during inference, our method leverages complete target objects for training and effectively avoiding this limitation and improving tracking accuracy.

3.3. Domain Adversarial Training

If the training process is conducted solely on synthesized data, the model will be prone to overfitting due to the tendency to capture differences between the synthesized target and the video frames, especially at the edges of the target. In this case, the model will perform very well during training but poorly during testing, as it cannot extract discriminative features related to the target. Specifically, on the test image, the model cannot capture the edges formed by the synthetic target, meaning that it has difficulty tracking the target; at the same time, because the discriminative features related to the target are not learned, the tracking is extremely susceptible to interference from similar targets. In this work, we address this issue by employing domain adversarial training, which is a technique derived from domain adaptation theory.
Concretely, we add a domain classifier behind the feature extractor (Encoder); the detailed method structure is demonstrated in the forward tracking part of Figure 2. The domain classifier consists of two fully connected layers, and primarily aims to distinguish whether the input images belong to synthesized or real images. In terms of input, the template branch and test branch input both real images and synthetic images; only synthetic images are used when training the tracker, while both synthetic and real images are used when training the domain classifier. The specific training process involves two alternating steps. In the first step, consisting of domain classifier training, the forward process calculates the domain classification loss of all input images using binary cross-entropy loss, and the tracker’s loss is ignored. In the process of updating the parameters, we freeze all parameters of the tracker and update the parameters of the domain classifier in the direction of reducing the domain classification loss. The second step consists of tracker training, in which the forward process simultaneously computes the tracking loss on synthesized images and the domain classification loss on all input images. During parameter updating, we freeze all parameters of the domain classifier and the parameters of the tracker are optimized in the direction of increasing the domain classification loss and decreasing the tracking loss. It is important to note that the calculation of the tracking loss depends on the loss calculation method of the original tracker.
To facilitate understanding, we use mathematical notation to restate the above process. We use the symbol l t ( · ) to represent the total tracking loss of the tracker and the symbol l c ( · ) to represent the classification loss of the domain classifier. Then, the domain classifier training loss can be represented as follows:
u t e m p = M ( x t , l t e m p ) ,
u t e s t = M ( x t + n , l t e s t ) ,
L d o m a i n = l c ( u t e m p , y c ) + l c ( u t e s t , y c ) ,
where M ( · ) represents the image mixing operation and y c is the domain label indicating whether the image belongs to synthesized or real images. Similarly, the tracker training loss can be represented as follows:
L t r a c k e r = α · l t ( I t e s t , y t ) β · L d o m a i n
where y t represents the label of the test frame, while α and β are scalar weightings used to regulate the contribution ratio of each loss.
The main idea of the whole process is to force the tracker to avoid tracking by exploiting the difference between the synthetic target and the real target. When this difference is used for tracking, the domain classifier distinguishes whether the image is real or synthetic, meaning that the classification loss is reduced. However, because we optimize the domain classifier in the direction of increasing the classification loss during training, the tracker must extract discriminative features related to the object in order to track effectively. Ultimately, by integrating domain adversarial learning into the training process of the tracker, it is possible to prevent the tracker from overfitting on synthetic data, ensuring good generalization performance in real-world testing.

4. Experiments

In this section, we employ the ToMP tracker as the baseline and train it using our proposed method. The training details are outlined in Section 4.1. To validate the effectiveness of our proposed method, we conduct comparative experiments between our tracker and the original ToMP tracker in Section 4.2. Subsequently, in Section 4.3.1 we perform ablation experiments on various components of our method to demonstrate their indispensability. In Section 4.3.2, we visualize the features extracted by the model before and after adversarial training to show the effectiveness of the adversarial training process in our tracker. In Section 4.4, we verified the effectiveness and competitiveness of our tracker when trained with limited budget samples by comparing it to state-of-the-art trackers trained with extensive labeled training samples on five standard benchmark datasets (OTB100 [61], UAV123 [62], GOT10k [63], LaSOT [64], and TrackingNet [65]). Finally, in Section 4.5 we visualize the tracking results of our tracker to other representative trackers on the LaSOT dataset to demonstrate its effectiveness in complex tracking scenarios.

4.1. Experimental Details

We trained ToMP [2] on the data collected from five datasets, including LaSOT [64], GOT-10k [63], ImageNet VID [66], YouTube-VOS [60], and TrackingNet [65]. It is worth noting that the ground truth labels of these datasets are not available in our method. During training, we utilized AdamW on an NVIDIA GeForce RTX 3090 GPU. Additionally, to ensure a more natural color transition between the target and background in synthesized images, we preprocessed the target objects using a Gaussian kernel before compositing them, allowing the target objects to appear in the image linearly.
The training process consisted of 340 epochs, with each epoch comprising 30,000 images and a batch size set to 20. Throughout the entire process, the domain classifier and tracker were alternately trained. In each alternating cycle, the domain classifier was trained for two epochs and the tracker was trained for six epochs. During the training of the tracker, the parameters of the domain classifier were frozen and the hyperparameters in the loss function were set to α = 100 and β = 1. Additionally, ToMP has two sets of parameters trained with different learning rates; for the first four epochs, these were set to 2 ×   10 5 and 1 ×   10 5 , respectively, then decreased to 4 ×   10 6 and 2 ×   10 6 after the fourth epoch. During training of the domain classifier, the parameters of the tracker were frozen and the learning rate was set to 1 ×   10 5 for the first epoch and decreased to 2 ×   10 6 for the second epoch. Finally, after the end of adversarial training, we fine-tuned the tracker on the labeled training dataset of the original tracker until convergence.
For the detailed parameter settings of the target transformations in the synthetic image generation process described in Section 3.2, we implemented the occlusion transformation using a Gaussian filter kernel with a size of 9. The scaling factor for scale transformation was randomly sampled from the range [ 0.7 , 1.3 ] . Affine transformations were used to simulate the deformation of the target. To introduce distractor variations, we additionally synthesized and placed two visually similar targets within the same region. Finally, we applied color augmentation to simulate additional blur effects. Specifically, the contrast was randomly adjusted within the range [ 0.9 , 1.1 ] and the brightness was randomly varied within the range [ 10 , 10 ] .

4.2. Contrast Experiment

To verify that our method can improve the tracker’s performance without increasing data collection costs, we compared our tracker with the original ToMP [2]. Specifically, both our tracker and the original tracker were trained to converge on the training data, then evaluated on three benchmarks: OTB100 [61], LaSOT [64], and UAV123 [62]. We conducted two sets of comparative experiments based on the number of labeled datasets used during training: GOT-10k [63] and GOT-10k+LaSOT. The experimental results are shown in Figure 4.
The evaluation results when using only GOT-10k as the training dataset are shown in Figure 4a. It can be observed that the AUC scores of ToMP improved by 1% on OTB100, 4% on LaSOT, and 1% on UAV123 after training with our method. This indicates that our method can indeed improve the model’s performance without increasing data collection costs. To validate the effectiveness of our method with more annotated data, we conducted a second set of comparative experiments with increased training data. The experimental results are shown in Figure 4b. Similar to Figure 4a, our tracker showed improvements on all three benchmarks. This demonstrates that our method can enhance the tracker’s performance regardless of whether a small amount or larger amount of labeled data is used for training.

4.3. Ablation Study

4.3.1. Effectiveness of Each Component

To validate the effectiveness and importance of each component in our method, we conducted ablation experiments on its two parts, namely, image transformation and domain adversarial training. Models were trained using unlabeled data from the GOT-10k, LaSOT, and TrackingNet datasets. After pretraining the models on unlabeled datasets using our method, the models were then trained on the labeled data until convergence. The detailed evaluation results are shown in Table 1.
As can be seen from Table 1, the worst performance is achieved when neither module is used. When using either module individually, the performance improves significantly. Comparing the second and fourth rows, it can be seen that the success rate curve (AUC) is improved by 2.3% and that the position prediction accuracy is improved by 1.9% after using the image transformation operations. This indicates that the image transformation operations help the tracker to learn the discriminative features of the target. Comparing the third row with the fourth row shows that the AUC is improved by 2.4% and that the location prediction accuracy is improved by 1.4% after using the domain adversarial training module. The larger improvement of the tracker before and after using domain adversarial training indicates that domain adversarial training is effective in mitigating the adverse effects of synthetic data. To illustrate the effect of domain adversarial more clearly, we visualize the features extracted by the model before and after domain adversarial training in Section 4.3.2.

4.3.2. Feature Visualization of the Effectiveness of Domain Adversarial Training

To provide a clearer demonstration of the impact of domain adversarial training on the model, we utilize the t-SNE clustering method [67] to visualize the features extracted by the model before and after adversarial training. Specifically, we mixed 1000 real images with 1000 synthetic images and fed them into the model, then extracted the features output by the encoder of the model and performed two-dimensional t-SNE clustering on them to obtain the clustering results. Finally, we normalized these two-dimensional clustering results for visualization. The detailed results are shown in Figure 5.
Comparing the two images in Figure 5 clearly shows that the features of the real images and synthetic images can be distinctly separated prior to adversarial training. This indicates that the model extracts significantly different features from the synthetic and real images. In other words, the model overfits to the synthetic dataset, and its performance on synthetic images does not accurately represent its performance on real images; consequently, a model trained on synthetic images will perform poorly on real images. On the other hand, a significant reduction in the differences between the features extracted from the two types of images can be observed after adversarial training. This suggests that the model extracts features shared by both synthetic and real images for tracking, meaning that its performance on synthetic images is now indicative of its performance on real images. From the feature visualization, it can be concluded that adversarial training effectively mitigates the overfitting phenomenon observed in the model trained on synthetic data.

4.4. State-of-the-Art Comparison

Next, we fine-tuned our tracker on labeled data to convergence. In this section, we present comparisons of our tracker with some state-of-the-art trackers on five challenging benchmark datasets to verify the effectiveness of our proposed approach. The benchmark datasets are GOT-10k [63], TrackingNet [65], OTB100 [61], LaSOT [64], and UAV123 [62].
GOT-10k [63]. To demonstrate the superior performance of our tracker, we evaluated it on the GOT-10k dataset and compared it with other trackers. The GOT-10K test set consists of 180 video sequences, and the test method is similar to the test method on the TrackingNet data. We uploaded tracking results of our tracker on the GOT-10k test set to the official website for evaluation and compared our results with those of several state-of-the-art trackers. The evaluation metrics used in GOT-10K include the Average Overlap (AO), Success Rate at threshold 0.5 (SR0.5), and Success Rate at threshold 0.75 (SR0.75). The compared state-of-the-art trackers include BANDT [68], SiamTPN [69], SiamGAT [70], SiamR-CNN [71], PrDiMP-50 [72], Ocean [73], SiamFC++ [74], DiMP-50 [18], SiamRPN++ [75], and STMTrack [76]. Detailed performance comparison results are provided in Table 2.
From Table 2, it is evident that our tracker achieves the highest scores in AO and SR0.5 among the various trackers, ranking second in SR0.75. Specifically, our tracker outperforms the second-ranked SiamR-CNN by 1.5% in AO and surpasses the third-ranked BANDT by 1.9%. In terms of SR0.5, our tracker exhibits a notable performance advantage of 3.6% over the second-ranked tracker. Lastly, in terms of SR0.75, our tracker trails the top-performing tracker by only 0.5%. This analysis demonstrates that our tracker holds a significant competitive edge among these state-of-the-art trackers, highlighting its superior performance across multiple evaluation metrics.
TrackingNet [65]. We also evaluated our tracker on the large-scale TrackingNet dataset, which consists of 511 video sequences without publicly available ground truth labels. Specifically, we tested our tracker on the test split of TrackingNet, then submitted the tracking results to the official website for evaluation. Subsequently, we compared our tracker with other state-of-the-art trackers, including BANDT [68], SiamTPN [69], TrTr [3], TrDiMP [79], HiFT [80], KeepTrack [81], PDiMP-50 [72], Ocean [73], SiamFC++ [74], DiMP-50 [18], SiamRPN++ [75], and SiamDW [77]. The comparison is based on metrics such as Precision, Normalized Precision, and AUC scores. Detailed results are presented in Table 3.
From Table 3, it can be observed that our tracker achieves the best performance across all three of the Precision, Normalized Precision, and AUC score metrics. Compared to the second-ranked tracker, BANDT, our tracker outperforms it by 0.4%, 0.8%, and 0.2%, respectively, in these three metrics; furthermore, our tracker maintains a significant lead over the third-ranked tracker, demonstrating that the tracker trained using our method can achieve competitive tracking performance compared to state-of-the-art trackers.
To demonstrate the effectiveness of our approach, we compared a version of our tracker without fine-tuning to the other self-supervised methods. The comparison results are shown in the starred results in the table. From these results, it can be seen that our method demonstrates a significantly higher performance advantage compared to existing methods, further proving the superiority of our approach.
OTB100 [61]. We evaluated our tracker on the classic benchmark dataset OTB100, which consists of 100 annotated short video sequences. We compared our tracker with several state-of-the-art trackers: SeqTrack [82], DropTrack [83], OSTrack [22], GRM [23], SwinTrack [33], CSWinTT [84], STARK [85], TransT [4], Ocean [73], SiamFC++ [74], DiMP-50 [18], and SiamDW [77]. Similar to the TrackingNet benchmark, the evaluation metrics include Precision and AUC scores. Detailed comparison results are presented in Table 4.
From the data in Table 4, it can be observed that our tracker achieves the highest score in terms of AUC scores, outperforming DropTrack and TransT by a margin of 0.1%. Although our tracker does not achieve the highest score in Precision, the difference from the top three trackers (Ocean, DropTrack, and SwinTrack) is relatively small. The lower performance in Precision may be attributed to the inability of domain adversarial training to eliminate overfitting of the model on synthetic data. Despite performing well on synthetic images during training, the model still struggles to accurately predict the position of the target in real images during testing. This observation is supported by the feature visualization experiment in Section 4.3.2, where some disparities still exist after domain adversarial training despite the significant reduction in differences between synthetic and real images. In short, the performance comparison data indicate that our tracker achieves competitive results on OTB100 compared to state-of-the-art trackers, demonstrating the effectiveness and feasibility of our approach.
LaSOT [64]. To verify that our tracker can maintain its excellent performance in long video tracking, we evaluated it on the test split of the LaSOT dataset. This split comprises 280 long videos with an average of 2512 frames per video. To compare our tracker with other top-performing trackers, we visualize the Success and Precision plots of each tracker in Figure 6. These trackers include TrSiam [79], PrDiMP-50 [72], DiMP-50 [18], Ocean [73], LTMU [86], and ATOM [78], among others.
From Figure 6, it is evident that our tracker demonstrates a strong competitive edge compared to others. In the Success plots, our tracker and TrSiam lead significantly ahead of other trackers, with our tracker trailing TrSiam by only 0.2%. Moreover, in the Precision plots, our tracker outperforms all other trackers, ranking first. This series of leading results convincingly demonstrates that our tracker maintains a robust competitive edge in long video tracking.
UAV123. The UAV123 dataset primarily comprises 91 drone videos, and is specifically designed to assess the robustness of a tracker against rapid changes in tracking perspectives. We evaluated our tracker on the UAV123 dataset and compared its performance with other top-performing trackers, including SiamTPN [69], TrTr [3], Ocean [73], and SiamBAN [87], among others. Detailed comparison results are presented in Table 5.
From the data in Table 5, it can be observed that while our tracker exhibits some weaknesses compared to the evaluations on the previous datasets, it remains highly competitive. In terms of Precision, our tracker shows relatively lower performance, which can be attributed to incomplete elimination of the impact of synthetic data during the adversarial training process. This may result in the feature extractor failing to effectively capture discriminative features within small targets, adversely affecting the prediction of target positions. However, our tracker still demonstrates excellent performance in terms of AUC score, ranking second among all compared trackers and with only a 0.7% difference from the top performer. These results effectively demonstrate the robustness of our tracker in handling rapid changes in tracking perspectives.
Table 3. Comparison results on the TrackingNet [65] dataset. The top three scores are highlighted in red, blue, and green, respectively, while * indicates that labeled data were not used for training.
Table 3. Comparison results on the TrackingNet [65] dataset. The top three scores are highlighted in red, blue, and green, respectively, while * indicates that labeled data were not used for training.
TypeSupervised MethodSelf-Supervised Method
TrackersOursBANDT
[68]
TrTr
[3]
TrDiMP
[79]
KeepTrack
[81]
PrDiMP-50
[72]
SiamFC++
[74]
DiMP-50
[18]
SiamRPN++
[75]
SiamDW
[77]
OursECO *
[88]
LUDT *
[89]
USOT *
[9]
Precision (%)74.974.567.473.173.870.468.770.569.456.360.448.949.555.1
Norm.prec (%)83.582.779.583.383.581.680.080.180.071.373.262.163.368.2
Success (AUC) (%)78.778.570.778.478.175.875.474.073.361.166.556.156.359.9
Table 4. Comparison results on the OTB100 [61] dataset. The top three scores are highlighted in red, blue, and green, respectively.
Table 4. Comparison results on the OTB100 [61] dataset. The top three scores are highlighted in red, blue, and green, respectively.
TrackersOursSeqTrack
[82]
DropTrack
[83]
OSTrack
[22]
GRM
[23]
SwinTrack
[33]
CSWinTT
[84]
STARK
[85]
TransT
[4]
Ocean
[73]
SiamFC++
[74]
DiMP-50
[18]
SiamDW
[77]
Precision (%)90.089.191.088.790.090.287.288.489.992.089.688.889.2
Success (AUC) (%)69.668.369.568.168.969.167.168.069.568.468.368.067.0
Table 5. Comparison results on the UAV123 [62] dataset. The top three scores are highlighted in red, blue, and green, respectively.
Table 5. Comparison results on the UAV123 [62] dataset. The top three scores are highlighted in red, blue, and green, respectively.
TrackersOursSiamTPN
[69]
TrTr
[3]
Ocean
[73]
SiamFC++
[74]
SiamBAN
[87]
DiMP-50
[18]
SiamDW
[77]
HiFT
[80]
Precision (%)82.782.383.982.381.083.385.877.678.7
Success(AUC) (%)64.163.663.362.162.363.164.853.658.9

4.5. Qualitative Comparison

Finally, to demonstrate the effectiveness of our tracker in complex scenarios, we visualize the tracking results of our tracker and four other representative trackers on the LaSOT dataset. Detailed results are shown in Figure 7. From the visualizations, in the first and fourth rows it can be seen that our tracker exhibits better robustness when facing interference from similar objects. In addition to interference from similar objects, target deformations and occlusions are common occurrences in real-world scenarios. From the results in the second and third rows, it is evident that our tracker continues to accurately track the target, demonstrating strong robustness even when the tracked target undergoes significant deformation or is occluded by other objects. These visualization results indicate that our tracker maintains strong robustness in various complex scenarios.

5. Limitations, Discussion, and Future Work

Although our method successfully eliminates the need for manual annotations and mitigates domain discrepancies between synthetic and real images through domain adversarial learning, a significant performance gap remains compared to fully supervised trackers. This limitation arises from two main factors. First, while synthetic training data can effectively provide complete target coverage, they may not fully capture the complex dynamics, appearance variations, and background clutter present in real-world videos, limiting the model’s generalization ability. Second, domain adversarial training, although beneficial, cannot entirely remove subtle distribution differences, and in some cases may lead to unstable convergence or reduced feature discriminability. As a result, the learned representations may fall short in challenging tracking scenarios involving occlusion, motion blur, or similar distractors.
Essentially, training a visual object tracker requires knowledge of the target’s location in video frames, including both the template and the search frames. This requirement fundamentally conflicts with the definition of self-supervised learning, which assumes no access to labeled data. Therefore, we argue that self-supervised training for visual tracking should be regarded as a form of pretraining rather than as a standalone training paradigm. This perspective aligns with recent self-supervised approaches such as MAE [90], where the goal is to learn transferable representations that can benefit downstream tasks.
In our future work, we plan to introduce more advanced image synthesis techniques such as conditional diffusion models; specifically, these techniques allow the diffusion model to generate more realistic images by conditioning the generation process on the target template and its spatial location, then treating the video frames to be synthesized as noisy inputs. This is expected to further narrow the gap between synthetic and real data, thereby enhancing the generalization ability of self-supervised training. Additionally, we plan to adopt more powerful domain adaptation strategies and explore multimodal self-supervised learning approaches, with the goal of extending the applicability of self-supervised training to a broader range of real-world scenarios.

6. Conclusions

With the continuous growth of applications based on sensors, the demand for visual tracking has become increasingly urgent. This paper explores methods for training an efficient deep visual tracking model in the absence of a large amount of labeled data. To this end, we propose a self-supervised method based on image synthesis and domain adversarial training. The proposed method aims to conduct self-supervised training using complete targets while also achieving effects on deep networks. Our method first builds a diverse database of target objects and synthesizes training pairs by embedding complete targets into unlabeled video frames, thereby ensuring precise bounding box supervision. Then, we incorporate domain adversarial training into the model optimization process to narrow the domain gap between synthetic data and real data. Experimental results show that our method achieves an AUC score of 66.5% on the TrackingNet dataset without using the labeled data, showing competitive performance compared to supervised benchmarks, while significantly reducing the gap between self-supervised tracking and supervised tracking. The proposed method can provide an efficient solution for AI-driven visual tasks where labeled data are scarce and unlabeled data remain underexploited.

Author Contributions

Conceptualization, G.G. and Q.L.; methodology, G.G. and D.Y.; validation, G.G. and Q.L.; formal analysis, G.G. and Q.L.; investigation, S.Z. and J.T.; writing—original draft preparation, G.G., D.Y., X.Z., S.Z. and J.T.; writing—review and editing, Q.L., D.Y., X.Z., S.Z. and J.T.; visualization, G.G. and Q.L.; supervision, D.Y. and X.Z.; project administration, D.Y. and S.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Natural Science Foundation of China under Grant Nos. 62202362 and 62302073, by the China Postdoctoral Science Foundation under Grant Nos. 2022TQ0247 and 2023M742742, by the Guangdong Basic and Applied Basic Research Foundation under Grant Nos. 2024A1515011626 and 2025A1515012949, by the Foundation of Yunnan Key Laboratory of Unmanned Autonomous Systems under Grant No. 202502ZD02, and by the Science and Technology Projects in Guangzhou under Grant No. 2023A04J0397.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The LaSOT dataset is available at http://vision.cs.stonybrook.edu/~lasot/, the TrackingNet dataset is available at https://eval.ai/web/challenges/challenge-page/1805/overview, the OTB100 dataset is available at https://github.com/prosti221/OTB-dataset, and UAV123 dataset is available at https://irip-buaa.github.io/posts/UAV123/.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.u.; Polosukhin, I. Attention is All you Need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30, pp. 1–11. [Google Scholar]
  2. Mayer, C.; Danelljan, M.; Bhat, G.; Paul, M.; Paudel, D.P.; Yu, F.; Van Gool, L. Transforming model prediction for tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 8731–8740. [Google Scholar]
  3. Zhao, M.; Okada, K.; Inaba, M. Trtr: Visual tracking with transformer. arXiv 2021, arXiv:2105.03817. [Google Scholar] [CrossRef]
  4. Chen, X.; Yan, B.; Zhu, J.; Wang, D.; Yang, X.; Lu, H. Transformer tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 8126–8135. [Google Scholar]
  5. Wang, N.; Song, Y.; Ma, C.; Zhou, W.; Liu, W.; Li, H. Unsupervised deep tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 1308–1317. [Google Scholar]
  6. Yuan, D.; Chang, X.; Huang, P.Y.; Liu, Q.; He, Z. Self-supervised deep correlation tracking. IEEE Trans. Image Process. 2021, 30, 976–985. [Google Scholar] [CrossRef]
  7. Yuan, W.; Wang, M.Y.; Chen, Q. Self-supervised object tracking with cycle-consistent siamese networks. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems, Las Vegas, NV, USA, 24 October–24 January 2021; pp. 10351–10358. [Google Scholar]
  8. Sio, C.H.; Ma, Y.J.; Shuai, H.H.; Chen, J.C.; Cheng, W.H. S2siamfc: Self-supervised fully convolutional siamese network for visual tracking. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 1948–1957. [Google Scholar]
  9. Zheng, J.; Ma, C.; Peng, H.; Yang, X. Learning to track objects from unlabeled videos. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 13546–13555. [Google Scholar]
  10. Lai, Z.; Lu, E.; Xie, W. Mast: A memory-augmented self-supervised tracker. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 6479–6488. [Google Scholar]
  11. Li, X.; Liu, S.; De Mello, S.; Wang, X.; Kautz, J.; Yang, M.H. Joint-task self-supervised learning for temporal correspondence. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; Volume 32. [Google Scholar]
  12. Wang, N.; Zhou, W.; Li, H. Contrastive transformation for self-supervised correspondence learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 8 February 2021; Volume 35, pp. 10174–10182. [Google Scholar]
  13. Ndayikengurukiye, D.; Mignotte, M. CoSOV1Net: A Cone- and Spatial-Opponent Primary Visual Cortex-Inspired Neural Network for Lightweight Salient Object Detection. Sensors 2023, 23, 6450. [Google Scholar] [CrossRef]
  14. Emily L., D.; Vighnesh, B. Unsupervised learning of disentangled representations from video. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  15. Zhao, D.; Li, Y.; Li, J.; Duan, X.; Ma, N.; Wang, Y. A Target Tracking Method Based on a Pyramid Channel Attention Mechanism. Sensors 2025, 25, 3214. [Google Scholar] [CrossRef] [PubMed]
  16. Shu, X.; Huang, F.; Qiu, Z.; Zhang, X.; Yuan, D. Learning Unsupervised Cross-Domain Model for TIR Target Tracking. Mathematics 2024, 12, 2882. [Google Scholar] [CrossRef]
  17. Li, X.; Pei, W.; Wang, Y.; He, Z.; Lu, H.; Yang, M.H. Self-supervised tracking via target-aware data synthesis. IEEE Trans. Neural Netw. Learn. Syst. 2023, 35, 9186–9197. [Google Scholar] [CrossRef] [PubMed]
  18. Bhat, G.; Danelljan, M.; Gool, L.V.; Timofte, R. Learning discriminative model prediction for tracking. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6182–6191. [Google Scholar]
  19. Shi, L.; Zhong, B.; Liang, Q.; Li, N.; Zhang, S.; Li, X. Explicit visual prompts for visual object tracking. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 26 February 2024; Volume 38, pp. 4838–4846. [Google Scholar]
  20. Lin, L.; Fan, H.; Zhang, Z.; Wang, Y.; Xu, Y.; Ling, H. Tracking meets lora: Faster training, larger model, stronger performance. In Proceedings of the European Conference on Computer Vision, Milan, Italy, 29 September–4 October 2024; pp. 300–318. [Google Scholar]
  21. Bai, Y.; Zhao, Z.; Gong, Y.; Wei, X. Artrackv2: Prompting autoregressive tracker where to look and how to describe. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 19048–19057. [Google Scholar]
  22. Ye, B.; Chang, H.; Ma, B.; Shan, S.; Chen, X. Joint feature learning and relation modeling for tracking: A one-stream framework. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; pp. 341–357. [Google Scholar]
  23. Gao, S.; Zhou, C.; Zhang, J. Generalized relation modeling for transformer tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 18686–18695. [Google Scholar]
  24. Cai, Y.; Liu, J.; Tang, J.; Wu, G. Robust object modeling for visual tracking. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 9589–9600. [Google Scholar]
  25. Li, Y.; Liu, M.; Wu, Y.; Wang, X.; Yang, X.; Li, S. Learning adaptive and view-invariant vision transformer for real-time UAV tracking. In Proceedings of the Forty-first International Conference on Machine Learning, Vienna, Austria, 21–27 July 2024. [Google Scholar]
  26. Kang, B.; Chen, X.; Lai, S.; Liu, Y.; Liu, Y.; Wang, D. Exploring enhanced contextual information for video-level object tracking. In Proceedings of the AAAI Conference on Artificial Intelligence, Philadelphia, PA, USA, 25 February–4 March 2025; Volume 39, pp. 4194–4202. [Google Scholar]
  27. Xie, J.; Zhong, B.; Liang, Q.; Li, N.; Mo, Z.; Song, S. Robust tracking via mamba-based context-aware token learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Philadelphia, PA, USA, 25 February–4 March 2025; Volume 39, pp. 8727–8735. [Google Scholar]
  28. Zheng, Y.; Zhong, B.; Liang, Q.; Mo, Z.; Zhang, S.; Li, X. Odtrack: Online dense temporal token learning for visual tracking. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 26 February 2024; Volume 38, pp. 7588–7596. [Google Scholar]
  29. Li, X.; Zhong, B.; Liang, Q.; Li, G.; Mo, Z.; Song, S. MambaLCT: Boosting Tracking via Long-term Context State Space Model. In Proceedings of the AAAI Conference on Artificial Intelligence, Philadelphia, PA, USA, 25 February–4 March 2025; Volume 39, pp. 4986–4994. [Google Scholar]
  30. Xu, C.; Zhong, B.; Liang, Q.; Zheng, Y.; Li, G.; Song, S. Less is More: Token Context-aware Learning for Object Tracking. arXiv 2025, arXiv:2501.00758. [Google Scholar] [CrossRef]
  31. Chen, H.; Chen, Z.; Yu, H. Enhanced YOLOv5: An Efficient Road Object Detection Method. Sensors 2023, 23, 8355. [Google Scholar] [CrossRef]
  32. Yuan, D.; Liao, D.; Huang, F.; Qiu, Z.; Shu, X.; Tian, C.; Liu, Q. Hierarchical Attention Siamese Network for Thermal Infrared Target Tracking. IEEE Trans. Instrum. Meas. 2024, 73, 1–11. [Google Scholar] [CrossRef]
  33. Lin, L.; Fan, H.; Zhang, Z.; Xu, Y.; Ling, H. Swintrack: A simple and strong baseline for transformer tracking. In Proceedings of the Advances in Neural Information Processing Systems, New Orleans, LA, USA, 28 November–9 December 2022; Volume 35, pp. 16743–16754. [Google Scholar]
  34. Li, W.; Xie, J.; Loy, C.C. Correlational image modeling for self-supervised visual pre-training. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 15105–15115. [Google Scholar]
  35. Chen, B.; Li, P.; Bai, L.; Qiao, L.; Shen, Q.; Li, B.; Gan, W.; Wu, W.; Ouyang, W. Backbone is all your need: A simplified architecture for visual object tracking. In Proceedings of the European conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; pp. 375–392. [Google Scholar]
  36. Yuan, D.; Zhang, H.; Liu, Q.; Chang, X.; He, Z. Transformer-Based RGBT Tracking with Spatio-Temporal Information Fusion. IEEE Sens. Journal 2025, 25, 25386–25396. [Google Scholar] [CrossRef]
  37. Cui, Y.; Jiang, C.; Wang, L.; Wu, G. Mixformer: End-to-end tracking with iterative mixed attention. In Proceedings of the IEEE/CVF conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 13608–13618. [Google Scholar]
  38. Zhao, H.; Wang, D.; Lu, H. Representation learning for visual object tracking by masked appearance transfer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 18696–18705. [Google Scholar]
  39. Zhang, Z.; Xu, L.; Peng, D.; Rahmani, H.; Liu, J. Diff-tracker: Text-to-image diffusion models are unsupervised trackers. In Proceedings of the European Conference on Computer Vision, Milan, Italy, 29 September–4 October 2024; pp. 319–337. [Google Scholar]
  40. Liang, S.; Bai, Y.; Gong, Y.; Wei, X. Autoregressive Sequential Pretraining for Visual Tracking. In Proceedings of the Computer Vision and Pattern Recognition Conference, Nashville, TN, USA, 11–15 June 2025; pp. 7254–7264. [Google Scholar]
  41. Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.Y.; et al. Segment anything. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 4015–4026. [Google Scholar]
  42. Jiang, J.; Wang, Z.; Zhao, M.; Li, Y.; Jiang, D. SAM2MOT: A Novel Paradigm of Multi-Object Tracking by Segmentation. arXiv 2025, arXiv:2504.04519. [Google Scholar]
  43. Kostiv, M.; Adamovskyi, A.; Cherniavskyi, Y.; Varenyk, M.; Viniavskyi, O.; Krashenyi, I.; Dobosevych, O. Self-Supervised Real-Time Tracking of Military Vehicles in Low-FPS UAV Footage. In Proceedings of the 2025 International Conference on Military Communication and Information Systems (ICMCIS), Oerias, Portugal, 13–14 May 2025; pp. 1–8. [Google Scholar]
  44. Zhang, L.; Gonzalez-Garcia, A.; Van De Weijer, J.; Danelljan, M.; Khan, F.S. Synthetic data generation for end-to-end thermal infrared tracking. IEEE Trans. Image Process. 2018, 28, 1837–1850. [Google Scholar] [CrossRef]
  45. Jiang, H.; Fels, S.; Little, J.J. Optimizing multiple object tracking and best view video synthesis. IEEE Trans. Multimed. 2008, 10, 997–1012. [Google Scholar] [CrossRef]
  46. Jiang, Y.; Han, D.; Cui, M.; Fan, Y.; Zhou, Y. A Video Target Tracking and Correction Model with Blockchain and Robust Feature Location. Sensors 2023, 23, 2408. [Google Scholar] [CrossRef] [PubMed]
  47. Kerim, A.; Celikcan, U.; Erdem, E.; Erdem, A. Using synthetic data for person tracking under adverse weather conditions. Image Vis. Comput. 2021, 111, 104187. [Google Scholar] [CrossRef]
  48. Khoreva, A.; Benenson, R.; Ilg, E.; Brox, T.; Schiele, B. Lucid data dreaming for object tracking. In Proceedings of the The DAVIS Challenge on Video Object Segmentation, Honolulu, HI, USA, 21–26 July 2017; pp. 1–6. [Google Scholar]
  49. Song, Y.; Zhang, Z.; Lin, Z.; Cohen, S.; Price, B.; Zhang, J.; Kim, S.Y.; Aliaga, D. Objectstitch: Object compositing with diffusion model. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 18310–18319. [Google Scholar]
  50. Chen, Z.; Yang, J.; Li, F.; Feng, Z.; Chen, L.; Jia, L.; Li, P. Foreign Object Detection Method for Railway Catenary Based on a Scarce Image Generation Model and Lightweight Perception Architecture. In Proceedings of the 2025 International Conference on Military Communication and Information Systems (ICMCIS), Oerias, Portugal, 13–14 May 2025. [Google Scholar] [CrossRef]
  51. Ganin, Y.; Ustinova, E.; Ajakan, H.; Germain, P.; Larochelle, H.; Laviolette, F.; March, M.; Lempitsky, V. Domain-adversarial training of neural networks. J. Mach. Learn. Res. 2016, 17, 1–35. [Google Scholar]
  52. Rangwani, H.; Aithal, S.K.; Mishra, M.; Jain, A.; Radhakrishnan, V.B. A closer look at smoothness in domain adversarial training. In Proceedings of the International Conference on Machine Learning, Baltimore, MD, USA, 17–23 July 2022; pp. 18378–18399. [Google Scholar]
  53. Zhou, K.; Yang, Y.; Hospedales, T.; Xiang, T. Deep domain-adversarial image generation for domain generalisation. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 13025–13032. [Google Scholar]
  54. Uplavikar, P.M.; Wu, Z.; Wang, Z. All-in-One Underwater Image Enhancement Using Domain-Adversarial Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–20 June 2019; pp. 1–8. [Google Scholar]
  55. Chen, M.; Zhao, S.; Liu, H.; Cai, D. Adversarial-learned loss for domain adaptation. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 3521–3528. [Google Scholar]
  56. Acuna, D.; Zhang, G.; Law, M.T.; Fidler, S. f-Domain Adversarial Learning: Theory and Algorithms. In Proceedings of the 38th International Conference on Machine Learning, Virtual, 18–24 July 2021; Volume 139, pp. 66–75. [Google Scholar]
  57. Li, B.; François-Lavet, V.; Doan, T.; Pineau, J. Domain Adversarial Reinforcement Learning. arXiv 2021, arXiv:2102.07097. [Google Scholar] [CrossRef]
  58. Li, T.; Zhao, Z.; Sun, C.; Yan, R.; Chen, X. Domain Adversarial Graph Convolutional Network for Fault Diagnosis Under Variable Working Conditions. IEEE Trans. Instrum. Meas. 2021, 70, 1–10. [Google Scholar] [CrossRef]
  59. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common objects in context. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 740–755. [Google Scholar]
  60. Xu, N.; Yang, L.; Fan, Y.; Yue, D.; Liang, Y.; Yang, J.; Huang, T. Youtube-vos: A large-scale video object segmentation benchmark. arXiv 2018, arXiv:1809.03327. [Google Scholar]
  61. Wu, Y.; Lim, J.; Yang, M.H. Object tracking benchmark. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1834–1848. [Google Scholar] [CrossRef]
  62. Mueller, M.; Smith, N.; Ghanem, B. A benchmark and simulator for UAV tracking. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 445–461. [Google Scholar]
  63. Huang, L.; Zhao, X.; Huang, K. GOT-10k: A large high-diversity benchmark for generic object tracking in the wild. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 1562–1577. [Google Scholar] [CrossRef] [PubMed]
  64. Fan, H.; Lin, L.; Yang, F.; Chu, P.; Deng, G.; Yu, S.; Bai, H.; Xu, Y.; Liao, C.; Ling, H. LaSOT: A high-quality benchmark for large-scale single object tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 5374–5383. [Google Scholar]
  65. Muller, M.; Bibi, A.; Giancola, S.; Alsubaihi, S.; Ghanem, B. TrackingNet: A large-scale dataset and benchmark for object tracking in the wild. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 300–317. [Google Scholar]
  66. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef]
  67. Cieslak, M.C.; Castelfranco, A.M.; Roncalli, V.; Lenz, P.H.; Hartline, D.K. t-Distributed Stochastic Neighbor Embedding (t-SNE): A tool for eco-physiological transcriptomic analysis. Mar. Genom. 2020, 51, 100723. [Google Scholar] [CrossRef] [PubMed]
  68. Yang, K.; Zhang, H.; Shi, J.; Ma, J. Bandt: A border-aware network with deformable transformers for visual tracking. IEEE Trans. Consum. Electron. 2023, 69, 377–390. [Google Scholar] [CrossRef]
  69. Xing, D.; Evangeliou, N.; Tsoukalas, A.; Tzes, A. Siamese transformer pyramid networks for real-time UAV tracking. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 2139–2148. [Google Scholar]
  70. Guo, D.; Shao, Y.; Cui, Y.; Wang, Z.; Zhang, L.; Shen, C. Graph attention tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 9543–9552. [Google Scholar]
  71. Voigtlaender, P.; Luiten, J.; Torr, P.H.; Leibe, B. Siam R-CNN: Visual tracking by re-detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 6578–6588. [Google Scholar]
  72. Danelljan, M.; Gool, L.V.; Timofte, R. Probabilistic regression for visual tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 7183–7192. [Google Scholar]
  73. Zhang, Z.; Peng, H. Ocean: Object-aware Anchor-free Tracking. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020. [Google Scholar]
  74. Xu, Y.; Wang, Z.; Li, Z.; Yuan, Y.; Yu, G. Siamfc++: Towards robust and accurate visual tracking with target estimation guidelines. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12549–12556. [Google Scholar]
  75. Li, B.; Wu, W.; Wang, Q.; Zhang, F.; Xing, J.; Yan, J. SiamRPN++: Evolution of Siamese visual tracking with very deep networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 4282–4291. [Google Scholar]
  76. Fu, Z.; Liu, Q.; Fu, Z.; Wang, Y. Stmtrack: Template-free visual tracking with space-time memory networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 13774–13783. [Google Scholar]
  77. Zhang, Z.; Peng, H. Deeper and wider Siamese networks for real-time visual tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 4591–4600. [Google Scholar]
  78. Danelljan, M.; Bhat, G.; Khan, F.S.; Felsberg, M. ATOM: Accurate tracking by overlap maximization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 4660–4669. [Google Scholar]
  79. Wang, N.; Zhou, W.; Wang, J.; Li, H. Transformer meets tracker: Exploiting temporal context for robust visual tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 1571–1580. [Google Scholar]
  80. Cao, Z.; Fu, C.; Ye, J.; Li, B.; Li, Y. Hift: Hierarchical feature transformer for aerial tracking. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 15457–15466. [Google Scholar]
  81. Mayer, C.; Danelljan, M.; Paudel, D.P.; Van Gool, L. Learning target candidate association to keep track of what not to track. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 13444–13454. [Google Scholar]
  82. Chen, X.; Peng, H.; Wang, D.; Lu, H.; Hu, H. Seqtrack: Sequence to sequence learning for visual object tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 14572–14581. [Google Scholar]
  83. Wu, Q.; Yang, T.; Liu, Z.; Wu, B.; Shan, Y.; Chan, A.B. Dropmae: Masked autoencoders with spatial-attention dropout for tracking tasks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 14561–14571. [Google Scholar]
  84. Song, Z.; Yu, J.; Chen, Y.P.P.; Yang, W. Transformer tracking with cyclic shifting window attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 8791–8800. [Google Scholar]
  85. Yan, B.; Peng, H.; Fu, J.; Wang, D.; Lu, H. Learning spatio-temporal transformer for visual tracking. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 10448–10457. [Google Scholar]
  86. Dai, K.; Zhang, Y.; Wang, D.; Li, J.; Lu, H.; Yang, X. High-performance long-term tracking with meta-updater. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 6298–6307. [Google Scholar]
  87. Chen, Z.; Zhong, B.; Li, G.; Zhang, S.; Ji, R. Siamese Box Adaptive Network for Visual Tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 6668–6677. [Google Scholar]
  88. Danelljan, M.; Bhat, G.; Shahbaz Khan, F.; Felsberg, M. Eco: Efficient convolution operators for tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6638–6646. [Google Scholar]
  89. Wang, N.; Zhou, W.; Song, Y.; Ma, C.; Liu, W.; Li, H. Unsupervised deep representation learning for real-time tracking. Int. J. Comput. Vis. 2021, 129, 400–418. [Google Scholar] [CrossRef]
  90. He, K.; Chen, X.; Xie, S.; Li, Y.; Dollár, P.; Girshick, R. Masked autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 16000–16009. [Google Scholar]
Figure 1. Overview of the target object database. The figure shows a random selection of target objects from the target object database, with the objects in each column belonging to the same category. This database is used to synthesize realistic training pairs in our self-supervised tracking framework.
Figure 1. Overview of the target object database. The figure shows a random selection of target objects from the target object database, with the objects in each column belonging to the same category. This database is used to synthesize realistic training pairs in our self-supervised tracking framework.
Sensors 25 04621 g001
Figure 2. Overview of the proposed self-supervised framework. In the image synthesis part, our method randomly selects target objects from the target object database and synthesizes them into video frames to form training pairs. During the synthesis of test frames, we apply image transformations to the target objects. During the training process, we add a domain classifier (blue) behind the backbone of the tracker, which is used for domain adversarial training. Its primary function is to perform domain classification on the images inputted to the network. The other part (yellow) represents the tracker used for training.
Figure 2. Overview of the proposed self-supervised framework. In the image synthesis part, our method randomly selects target objects from the target object database and synthesizes them into video frames to form training pairs. During the synthesis of test frames, we apply image transformations to the target objects. During the training process, we add a domain classifier (blue) behind the backbone of the tracker, which is used for domain adversarial training. Its primary function is to perform domain classification on the images inputted to the network. The other part (yellow) represents the tracker used for training.
Sensors 25 04621 g002
Figure 3. Main steps of image transformation operations. The left part is the object extracted from the target object database and the video frame extracted from the video, while the right side lists the five transformation operations and their effects.
Figure 3. Main steps of image transformation operations. The left part is the object extracted from the target object database and the video frame extracted from the video, while the right side lists the five transformation operations and their effects.
Sensors 25 04621 g003
Figure 4. Performance comparison between the original ToMP and our tracker, using AUC scores as the performance metric. The left shows the comparison of tracker performance trained only on GOT-10k, while the right shows the comparison of tracker performance trained on both GOT-10k and LaSOT.
Figure 4. Performance comparison between the original ToMP and our tracker, using AUC scores as the performance metric. The left shows the comparison of tracker performance trained only on GOT-10k, while the right shows the comparison of tracker performance trained on both GOT-10k and LaSOT.
Sensors 25 04621 g004
Figure 5. t-SNE feature clustering visualization graph: (a) the feature visualization graph before adversarial training and (b) the feature visualization graph after adversarial training. In each graph, red points represent features from synthetic images, while green points represent features from real images.
Figure 5. t-SNE feature clustering visualization graph: (a) the feature visualization graph before adversarial training and (b) the feature visualization graph after adversarial training. In each graph, red points represent features from synthetic images, while green points represent features from real images.
Sensors 25 04621 g005
Figure 6. Plot showing the results on the LaSOT [64] testing set in terms of Precision and Success (AUC) scores.
Figure 6. Plot showing the results on the LaSOT [64] testing set in terms of Precision and Success (AUC) scores.
Sensors 25 04621 g006
Figure 7. Qualitative comparison results between our tracker and four other representative trackers on the LaSOT benchmark.
Figure 7. Qualitative comparison results between our tracker and four other representative trackers on the LaSOT benchmark.
Sensors 25 04621 g007
Table 1. Ablation study results on the OTB100 dataset [61]. Checkmarks indicate that the module is used, and the highest score is indicated in bold.
Table 1. Ablation study results on the OTB100 dataset [61]. Checkmarks indicate that the module is used, and the highest score is indicated in bold.
Domain AdversarialImage TransformationPrecision (%)AUC (%)
86.265.1
88.167.3
88.667.2
90.069.6
Table 2. Comparison results on the GOT10k [63] test set. The top three scores are highlighted in red, blue, and green, respectively.
Table 2. Comparison results on the GOT10k [63] test set. The top three scores are highlighted in red, blue, and green, respectively.
TrackersAO (%)SR0.50 (%)SR0.75 (%)
BANDT [68]64.573.854.2
SiamTPN [69]57.668.644.1
SiamGAT [70]62.774.348.8
SiamR-CNN [71]64.972.859.7
PrDiMP-50 [72]63.473.854.3
Ocean [73]61.172.147.3
SiamFC++ [74]59.569.547.3
DiMP-50 [18]61.171.749.2
SiamDW [77]42.948.314.7
ATOM [78]55.663.440.2
SiamRPN++ [75]51.761.632.5
STMTrack [76]64.273.757.5
Ours66.477.959.2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Geng, G.; Zhou, S.; Tang, J.; Zhang, X.; Liu, Q.; Yuan, D. Self-Supervised Visual Tracking via Image Synthesis and Domain Adversarial Learning. Sensors 2025, 25, 4621. https://doi.org/10.3390/s25154621

AMA Style

Geng G, Zhou S, Tang J, Zhang X, Liu Q, Yuan D. Self-Supervised Visual Tracking via Image Synthesis and Domain Adversarial Learning. Sensors. 2025; 25(15):4621. https://doi.org/10.3390/s25154621

Chicago/Turabian Style

Geng, Gu, Sida Zhou, Jianing Tang, Xinming Zhang, Qiao Liu, and Di Yuan. 2025. "Self-Supervised Visual Tracking via Image Synthesis and Domain Adversarial Learning" Sensors 25, no. 15: 4621. https://doi.org/10.3390/s25154621

APA Style

Geng, G., Zhou, S., Tang, J., Zhang, X., Liu, Q., & Yuan, D. (2025). Self-Supervised Visual Tracking via Image Synthesis and Domain Adversarial Learning. Sensors, 25(15), 4621. https://doi.org/10.3390/s25154621

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop