Next Article in Journal
Toward High-Resolution UAV Imagery Open-Vocabulary Semantic Segmentation
Previous Article in Journal
Trajectory Planning of Unmanned Aerial Vehicles in Complex Environments Based on Intelligent Algorithm
Previous Article in Special Issue
Lightweight UAV Detection Method Based on IASL-YOLO
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing the Scale Adaptation of Global Trackers for Infrared UAV Tracking

1
College of Aerospace Science and Engineering, National University of Defense Technology, Changsha 410073, China
2
Hunan Provincial Key Laboratory of Image Measurement and Vision Navigation, Changsha 410073, China
*
Author to whom correspondence should be addressed.
Drones 2025, 9(7), 469; https://doi.org/10.3390/drones9070469
Submission received: 9 June 2025 / Revised: 27 June 2025 / Accepted: 28 June 2025 / Published: 1 July 2025
(This article belongs to the Special Issue UAV Detection, Classification, and Tracking)

Abstract

Tracking unmanned aerial vehicles (UAVs) in infrared video is an essential technology for the anti-UAV task. Given frequent UAV target disappearances caused by occlusion or moving out of view, global trackers, which have the unique ability to recapture targets, are widely used in infrared UAV tracking. However, global trackers perform poorly when dealing with large target scale variation because they cannot maintain approximate consistency between target sizes in the template and the search region. To enhance the scale adaptation of global trackers, we propose a plug-and-play scale adaptation enhancement module (SAEM). This can generate a scale adaptation enhancement kernel according to the target size in the previous frame, and then perform implicit scale adaptation enhancement on the extracted target template features. To optimize training, we introduce an auxiliary branch to supervise the learning of SAEM and add Gaussian noise to the input size to improve its robustness. In addition, we propose a one-stage anchor-free global tracker (OSGT), which has a more concise structure than other global trackers to meet the real-time requirement. Extensive experiments on three Anti-UAV Challenge datasets and the Anti-UAV410 dataset demonstrate the superior performance of our method and verify that our proposed SAEM can effectively enhance the scale adaptation of existing global trackers.

1. Introduction

Infrared UAV tracking is used to continuously locate a specific UAV in an infrared video and estimate its scale [1]. Benefiting from the excellent imaging ability of infrared cameras under all-weather and low-light conditions, this technology is widely applied to the anti-UAV task to protect individual privacy and public safety [2,3]. In essence, infrared UAV tracking belongs to the research field of single object tracking [4]. However, unlike tracking generic objects such as pedestrians and vehicles, tracking high-speed moving UAVs in infrared videos presents more challenges:
(1)
Infrared UAV tracking is susceptible to occlusion, thermal crossover, and interference in complex scenarios such as trees, buildings, heavy clouds, and strong clutter.
(2)
Due to the rapid movement of the UAV target or the instability of the infrared camera platform, the position of the UAV target will change drastically between two adjacent frames or even move out of view.
(3)
The target scale variation is dramatic when the camera adjusts its focal length or the target moves rapidly closer or farther away, especially in the UAV-to-UAV task [5].
According to the size of the search region, single-object trackers can be categorized into local trackers [6,7] and global trackers [8,9,10]. Local trackers crop a small patch from the current frame (Frame T) as the search region according to the target position and size in the previous frame, as shown in Figure 1a. This helps to reduce computation and background clutter. However, local trackers rely on tracking a target stably. When a target disappears, it is liable to select the wrong local search region, which leads to the failure of subsequent tracking. Global trackers use the whole current frame as the search region, as shown in Figure 1b, so they have the advantage of recapturing the disappeared target. For the above challenge (1) and (2), where targets frequently disappear due to occlusion or moving out of view, global trackers demonstrate greater robustness than local trackers, making them the preferred choice for infrared UAV tracking. Table 1 shows a comparison between the two categories of trackers.
For challenge (3), the key to achieving scale adaptation is to maintain approximate consistency between the sizes of the target in the template and the search region, which is beneficial for matching the features of both for robust tracking. Traditional local trackers [11,12] employ pyramid sampling to obtain multi-scale search regions to deal with the target scale variation. Local trackers [13,14] based on deep learning adjust the template and search region to fixed sizes by a resizing operation, as shown in Figure 1a. This operation makes the target size in the search region approximately consistent with the target size in the template, which avoids the undesirable effects of target scale variation on tracking. However, global trackers cannot follow the above methods. On the one hand, if the whole current frame is resized according to the target scale change factor, the size of the image input to the model will increase dramatically, which leads to a sharp increase in the computation. On the other hand, if the target template is directly resized, the features of the resized template must be extracted every frame during online tracking, which also leads to a serious reduction in the running efficiency. Therefore, the existing global trackers have poor scale adaptation and cannot handle the target scale variation well, especially when the scale changes drastically. As shown in Figure 1b, the global tracker incorrectly identifies the tower structure as the target, because the labeled target in the template is more approximate in size to the tower structure rather than the ground truth in the search region.
In the field of arbitrary-scale image super-resolution, some methods [15,16] can adaptively adjust the model parameters based on the required scaling factor. This processing effectively establishes the connection between the scaling factors and the scale information of the image features, enabling direct feature processing for generating high-quality images with arbitrary sizes. Inspired by this mechanism, we propose an implicit scale adaptation enhancement method for the target template using the target size in the previous frame as a guide, as shown in Figure 1b. Specifically, we propose a scale adaptation enhancement module (SAEM), which can generate a scale adaptation enhancement kernel according to the previous target size, and then directly process the scale information of the extracted target template features. This is defined as the implicit method, which requires less computation than the explicit method that needs to first resize the target template and then extract its features in every frame. During training, we set up an auxiliary branch that adopts the above explicit method to supervise the learning of SAEM, and we also add Gaussian noise to the input size to improve the robustness of SAEM to handle inaccurate input sizes in complex scenarios. During the online tracking, an adaptive threshold is proposed to more accurately judge whether the target is disappearing. This is used to prevent SAEM from receiving the incorrect size input caused by target disappearance. Moreover, SAEM is a plug-and-play module that can be embedded into other global trackers to enhance their scale adaptation, especially when the typical scale change is more than 10× in infrared UAV tracking.
In addition, we propose a concise one-stage anchor-free global tracker (OSGT) to meet the real-time requirement of the infrared UAV tracking task. It is superior to the complex and inefficient two-stage anchor-based structure commonly used in existing global trackers [10,17]. In detail, it combines the feature fusion module from GlobalTrack [8], improved by using hierarchical cross-correlation, and the output head of FCOS [18], without the centerness branch. OSGT can run in real time at 30.9 fps in the infrared video with a resolution of 512 × 640. After SAEM is embedded, it can still run at 27.3 fps.
In summary, the main contributions of the work in this paper are as follows:
(1)
We propose a plug-and-play scale adaptation enhancement module, which can implicitly resize the target template to enhance the scale adaptation of existing global trackers for the infrared UAV tracking task.
(2)
During training, we design an auxiliary branch to supervise the learning of SAEM and add Gaussian noise to the input size to enhance its robustness. During online tracking, an adaptive threshold is proposed to accurately judge target disappearance and prevent SAEM from being affected by incorrect input size.
(3)
We propose a one-stage anchor-free global tracker with a simpler structure, which can track UAVs in real time.
The rest of this paper is organized as follows. In Section 2, we briefly review the related works. In Section 3, we describe the proposed method in detail. In Section 4, experimental results are presented and analyzed. Finally, we conclude this paper in Section 5.

2. Related Work

This section focuses on approaches closely related to our work, including single object tracking, global tracker, infrared UAV tracking, and scale-arbitrary image super-resolution.

2.1. Single Object Tracking

Single object tracking requires a target labeled in the first frame to be continuously located and scale-estimated in subsequent video frames. The target types are not limited, and the types of videos are usually RGB and infrared. Traditional target tracking algorithms use particle filtering [19], structured SVM [20], and correlation filtering [21]. Among them, correlation filtering tackles the visual tracking by solving the ridge regression in the Fourier domain, which has been widely developed because of its high efficiency. Benefiting from the development of deep learning, researchers have designed many high-performance deep trackers. These trackers usually use the Siamese network [22,23] to extract the features of the target template and the search region, respectively, and then fuse them to search for the target. According to the feature fusion methods, deep trackers can be roughly categorized into CNN-based and Transformer-based. The former mostly uses cross-correlation [22,24] and discriminative filtering [25,26], which has a simpler network structure and faster running speed. The latter utilizes the Transformer attention mechanism to fuse the features [27,28], which exhibits attractive performance. However, these trackers need to use multiple Transformer modules in series, which leads to more computation and memory occupation. OSTrack [7] and MixFormer [29] use Transformer-based one-stream frameworks, which combine feature extraction and feature fusion into one stage to deeply explore more suitable visual features for tracking. Moreover, ARTrack [30] and AQATrack [31] use the autoregressive method to introduce the temporal motion information into the tracking model and achieve excellent performance.

2.2. Global Tracker

Unlike local trackers, which use a small image patch as the search region, global trackers search for targets over the entire current frame [8,32]. Therefore, global trackers have the ability to recapture the target, which can effectively deal with rapid target movement and temporary target disappearance caused by occlusion and moving out of view. This advantage makes them perform well when tracking UAVs in complex infrared scenarios [9,10,17,33,34]. In contrast, the local tracker needs to attach a global detection module to recapture the reappearing target [35,36], which makes the model structure so complex that it needs to be trained in modules or stages. In fact, the global tracker is also often used as the global detection module of the local tracker; for example, LTMU [36] uses GlobalTrack to recapture the reappearing target. Nevertheless, the existing global trackers have poor scale adaptation because they cannot keep the approximate consistency between the target sizes in the template and the search region, which makes them unable to cope well with the drastic target scale variation.

2.3. Infrared UAV Tracking

Compared with RGB cameras, infrared cameras are more suitable for the anti-UAV task because they can work well under all-weather and low-light conditions. However, tracking UAVs in infrared video is still challenging due to the complexity and diversity of application scenarios. In order to accurately track small UAV targets, TransIST [37] combines multi-scale attention and a side window filter into a Transformer-based tracking model. Considering the camera motion, Zhao et al. [38] extract feature points and calculate the homography matrix to compensate for the motion between frames. For the problem of background interference [39], Fang et al. [9] adopt Kalman filtering to estimate target motion, while SiamDT [34] detects and memorizes interference objects in the scene to exclude incorrect candidate targets. In the common case where the target disappears during UAV tracking, Wu et al. [40] search for the target by expanding the search region, while Zhao et al. [38] and Yu et al. [41] use the global detection module to recapture the target, and SiamSTA [17], CAMTracker [10], and SiamDT adopt the global tracking scheme. When the target scale changes, our work solves the problem of poor scale adaptation of global trackers to track UAVs more robustly.

2.4. Scale-Arbitrary Image Super-Resolution

Our proposed SAEM refers to some image super-resolution research results, which are briefly introduced here. Image super-resolution aims to reconstruct a high-resolution image from a low-resolution one, and is widely used in remote sensing, medical, surveillance, etc. Recently, researchers have proposed some deep super-resolution models such as EDSR [42] and SwinIR [43]. However, these methods usually need to set a fixed scale factor, such as ×2 or ×4, which lacks flexibility. To address this problem, MetaSR [15] generates convolution kernels based on the desired factor and pixel coordinates for achieving image super-resolution with arbitrary magnification. ArbRCAN [16] combines experts learned during training based on the desired factor and then generates a convolution kernel to achieve scale-arbitrary super-resolution. Inspired by these methods, we improve the scale-aware convolutional layer of ArbRCAN to implicitly resize the target template in the global tracker.

3. Methodology

We propose an efficient one-stage anchor-free global tracker (OSGT) and introduce a scale adaptation enhancement module (SAEM) to implicitly resize the target template to solve the problem that global trackers do not cope well with target scale variation. As shown in Figure 2, our proposed OSGT adopts a Siamese network. Firstly, the multi-level features of the first frame (Frame 1, template) and the current frame (Frame T, search region) are extracted by parameter-shared ResNet50 and FPN (feature pyramid network). Secondly, the feature fusion part fuses the features of the target template and the search region. Finally, the output heads generate the multi-level score maps and regression maps, and the bounding box corresponding to the position with the largest score is taken as the tracking result of the current frame. In addition, our proposed SAEM is embedded into the feature fusion part, and directly processes the extracted target template features with the input of the initial target size and the target size in the previous frame (Frame T-1). During the training, an auxiliary branch is designed to supervise the learning of SAEM, and Gaussian noise is added to the input size to enhance its robustness. During online tracking, an adaptive threshold is proposed to more accurately judge target disappearance and effectively avoid SAEM being affected by incorrect input size.

3.1. One-Stage Anchor-Free Global Tracker

GlobalTrack, SiamSTA, and SiamDT, based on RCNN [44], adopt the two-stage anchor-based framework, including the region proposal stage and the target regression stage. These methods have complex structures and require two rounds of feature fusion between the target template and the search region, making them unable to run in real time. In the field of object detection, one-stage anchor-free detection algorithms such as FCOS and CenterNet [45] use simpler structures that can run faster and perform well. Therefore, we combine the feature fusion module of GlobalTrack and the output head of FCOS to propose a concise and efficient one-stage anchor-free global tracker.

3.1.1. Feature Extraction

ResNet50 is used to extract the features of the first and current frame for obtaining four levels of features with different resolutions, and their sizes are [ H / l , W / l ] , where l { 4 , 8 , 16 , 32 } , and then FPN is used for multi-scale feature enhancement. Note that the features of the first frame are extracted only once as the template during the online tracking.
Our OSGT is trained to use feature maps with different resolutions to track UAV targets of corresponding sizes, as shown in Figure 2, where the feature map of level L i ( i = 0 , 1 , 2 , 3 correspond to l ) processes the targets with sizes belonging to [ 0 , 10 ) , [ 10 , 20 ) , [ 20 , 40 ) and [ 40 , + ) , respectively. In this way, small targets with fewer appearance features can be tracked using low-level texture features, while large targets with rich appearance features can be tracked using high-level semantic features.

3.1.2. Feature Fusion

The feature fusion part of GlobalTrack uses RoI Align [46] (region of interest), based on the ground truth, to extract a level of the target features, with the level index determined by the target size. Instead, we extract all four levels of the target features z i 256 × 7 × 7 , and then fuse z i and the features of the current frame x i 256 × H / l × W / l using hierarchical cross-correlation to obtain the fusion feature x ^ i 256 × H / l × W / l . When SAEM is not embedded, the above process is calculated as follows:
x ^ i = C i ( z i , x i ) = F i 1 × 1 F i 7 × 7 ( z i ) x i ,   i = 0 , 1 , 2 , 3
where F i 7 × 7 is a 7 × 7 convolutional layer with padding 0 for converting z i to a 256 × 1 × 1 convolutional kernel, is the convolutional operation and F i 1 × 1 is a 1 × 1 convolutional layer for transforming the number of channels. The hierarchical cross-correlation fuses the features of the target template and the search region separately according to the feature depth, which avoids the semantic confusion caused by correlating features from different levels. In addition, extracting all levels of the target features can provide our proposed SAEM with rich multi-scale information.

3.1.3. Output Head

Our proposed OSGT adopts the head of FCOS, but removes the centerness branch and some convolutional layers. In detail, the fusion feature x ^ i is processed with two CGRs (Conv+GroupNorm+ReLU) instead of the four CGRs used in FCOS, which is sufficient for the infrared UAV tracking task and conducive to improving the running speed. Then, a classification branch (consisting of a 3 × 3 convolutional layer) is set up to obtain the score map s i 1 × H / l × W / l , and a regression branch (consisting of a 3 × 3 convolutional layer) is set up to obtain the bounding box b i 4 × H / l × W / l , where the first dimension denotes the upper left and lower right corner points of the target ( x 0 , y 0 , x 1 , y 1 ) . Finally, we adopt the Top-1 prediction to take the bounding box corresponding to the position with the highest score as the tracking result of the current frame.

3.2. Enhancing the Scale Adaptation of Global Tracker

Figure 3 shows the changes in four levels of the target features z i after 2× upsampling and 2× downsampling. This indicates that the target features not only contain appearance information such as shape, texture, and brightness, but also are closely related to the target size. Therefore, the global tracker tends to search for the object in the current frame that is close in size to the target template. However, when the target scale changes greatly, existing global trackers cannot adjust the target template based on the scale change, resulting in tracking failure in case of being disturbed by other objects of similar size.
In order to enhance the scale adaptation of global trackers, the target templates need to be dynamically adjusted to keep approximately the same size as the target in the search region. Inspired by the scale-arbitrary image super-resolution, we propose a plug-and-play scale adaptation enhancement module that can be embedded in the feature fusion part, as shown in Figure 2. This can generate the scale adaptation enhancement kernel according to the target size of the previous frame, and then directly process the extracted features of the target template to resize it implicitly, as shown in Figure 4.

3.2.1. Scale Adaptation Enhancement Module

SAEM aims to directly adjust the scale information of the target template features based on the target size of the previous frame [ h T - 1 , w T - 1 ] , as shown in the upper part of Figure 4. Firstly, the four levels of the target template features z i are concatenated into Z 1024 × 7 × 7 . Next, Z is convolved with a scale adaptation enhancement kernel (SAEK) to obtain Z ^ . Finally, we divide Z ^ into z ^ i 256 × 7 × 7 by the reshape operation. The concatenation operation can ensure that each z ^ i can obtain information from all four levels of z i . The above process is calculated as follows:
Z ^ = SAEK ( h 1 , w 1 , h T 1 , w T 1 ) Z Z = Concat ( z 0 , z 1 , z 2 , z 3 ) z ^ 0 , z ^ 1 , z ^ 2 , z ^ 3 = Reshape ( Z ^ ) .
After SAEM is embedded, the feature fusion is calculated as follows:
x ^ i = C i ( z ^ i , x i ) = F i 1 × 1 F i 7 × 7 ( z ^ i ) x i ,   i = 0 , 1 , 2 , 3 .
The SAEK is obtained based on the initial size of the target template [ h 1 , w 1 ] and the target size of the previous frame [ h T - 1 , w T - 1 ] . Referring to ArbRCAN for scale-arbitrary super-resolution, the SAEK is designed to contain M (the default value is 12) experts, and each expert is a learnable tensor with dimensions 1024 × 1024 × 3 × 3, as shown in the lower part of Figure 4. During training, experts learn and save knowledge that can be used to adjust the scale information of target template features. In addition, [ h 1 , w 1 ] and [ h T - 1 , w T - 1 ] are concatenated and input to the fully connected network to obtain the expert weight, and then the experts are weighted and summed to obtain SAEK. Therefore, the parameters of SAEK can change dynamically according to the input target size. The above process is calculated as follows:
SAEK ( h 1 , w 1 , h T 1 , w T 1 ) = j = 1 M W e i g h t j × E x p e r t j W e i g h t = FC ( Concat ( h 1 , w 1 , h T 1 , w T 1 ) ) .
As mentioned above, we concatenate [ h 1 , w 1 ] and [ h T - 1 , w T - 1 ] as the input, which is different from the ratio input [ h T - 1 / h 1 , w T - 1 / w 1 ] in ArbRCAN. The advantage of our method is that it can provide a base scale for the model to avoid confusion. For example, a target with a size [ 8 , 10 ] is upsampled ×2 to [ 16 , 20 ] , and the other target with a size [ 32 , 40 ] is also upsampled ×2 to [ 64 , 80 ] . Although these two operations have the same magnification factor, different experts (learned knowledge) should be used, and their SAEKs should also be different.

3.2.2. Supervision and Gaussian Noise

The essence of SAEM is to map the template features with the initial size [ h 1 , w 1 ] to new template features with the size [ h T - 1 , w T - 1 ] . The latter can also be obtained through an explicit method that involves: (1) resizing the template according to the target size in the previous frame, and (2) extracting its target features. These extracted features can serve as the ground truth for training the SAEM. In this way, we set up an auxiliary branch to supervise the learning of SAEM, as shown in the blue dashed box in Figure 5. Firstly, Template is resized by a factor of [ h T - 1 / h 1 , w T - 1 / w 1 ] to obtain Template*; then, the target features in Template* are extracted to obtain z i ; finally, we take z i as the ground truth and use the smooth L1 loss to calculate the loss L sup between z ^ i and z i . For training our model, the loss is calculated as follows:
L total = 1 4 i = 0 3 L cls ( s i , s i ) + L reg ( b i , b i ) + L sup ( z ^ i , z i ) ,
where L cls ( · ) is the focal loss for classification, L reg ( · ) is the giou loss for regression, and s i 1 × H / l × W / l and b i 4 × H / l × W / l denote the ground truth of score and bounding box at the i-th layer, respectively. The total loss L total is the average of four-layer losses.
During online tracking, the previous target size [ h T - 1 , w T - 1 ] input to SAEM may be inaccurate due to occlusion and background clutter. To enhance the robustness of SAEM, [ h T - 1 , w T - 1 ] is multiplied by a random value to simulate the size bias during the training. Figure 6 shows histograms of the change ratios for target width and height between adjacent frames on the Anti-UAV dataset (excluding the sequences belonging to target scale change scenarios). Under normal circumstances, the target size hardly changes between adjacent frames, with a ratio of approximately 1.0. When the target is occluded or interfered by background clutter, only part of the target remains visible, resulting in a ratio of less than 1.0. Conversely, when the target exits the occlusion or clutter region and becomes fully visible again, the ratio exceeds 1.0. Among them, the latter two abnormal states occur less frequently, and their ratios are basically within the range of [0.7, 1.3], approximately conforming to a Gaussian distribution with mean μ = 1.0 and standard deviation σ = 0.1 . Therefore, we use Gaussian noise and perform a clamp operation to limit its range within [ μ 3 σ , μ + 3 σ ] to avoid extreme values with small probability, as shown below:
[ h T - 1 , w T - 1 ] = [ h T - 1 , w T - 1 ] × g noise ,   g noise [ μ 3 σ , μ + 3 σ ] ,
where [ h T - 1 , w T - 1 ] is the previous target size after adding Gaussian noise, as shown in the red dashed box in Figure 5. The above process can be regarded as adding noise to the exact [ h T - 1 , w T - 1 ] to achieve a data augmentation, which can effectively reduce the sensitivity of SAEM to the input size.

3.2.3. An Adaptive Threshold for Judging Target Disappearance

During online tracking, the previous target size [ h T - 1 , w T - 1 ] entered into SAEM is initialized with the target annotation in the first frame, and it is updated frame by frame in subsequent tracking. However, [ h T - 1 , w T - 1 ] should stop updating when the target disappears due to occlusion or moving out of view. Otherwise, if the wrong [ h T - 1 , w T - 1 ] is input to SAEM, it will lead to tracking failure and incorrect recapture. Therefore, it is important to accurately judge whether the target exists for online tracking. A common method is to set a specific threshold for the maximum score, s max ; for example, if s max > 0.5 , the target is present, and vice versa. Nevertheless, as shown in Figure 7, the smaller the target size, the harder it is to be tracked, and the lower the maximum score s max . If 0.5 is used as the threshold, targets with a size smaller than 10 will be misjudged as having disappeared; if a small threshold is used, such as 0.3, there will be more false positive cases when judging larger targets. Therefore, a fixed threshold is not appropriate. To address this problem, we propose an adaptive method to judge whether a target exists or not, as follows:
e = 1 ,   s max > 0.5 s T = 1 max 0 ,   o t h e r w i s e
where ‘1’ means the target is present, ‘0’ means the target is absent, and s T = 1 max represents the target score obtained by inputting the first frame into the tracker and finding the corresponding score based on the position of the real target. s T = 1 max provides a base value for the threshold, which can avoid misjudging the existence of small and dim targets compared with directly using 0.5 as the threshold.

4. Experiment

4.1. Datasets and Evaluation Metrics

We use the public Anti-UAV Challenge and Anti-UAV410 datasets for training and testing. These datasets are closely related to the infrared UAV tracking task.
Our proposed OSGT is trained on the most comprehensive 3rd Anti-UAV training dataset, which contains 150 infrared videos, and tested on the 1st [1], 2nd Anti-UAV test-dev [47], and 3rd Anti-UAV validation datasets. These three test datasets contain 100, 140, and 50 infrared videos, respectively, with increasing scene complexity and tracking difficulty. Each video contains 1000 to 1500 frames, and the resolution of each frame is 512 × 640. These datasets cover UAVs at different scales and in various scenarios such as cloudy skies, cities, and mountains.
The Anti-UAV410 dataset [34] is augmented with a large number of diverse and complex scenarios based on the 1st and 2nd Anti-UAV datasets. These scenarios included a wide range of backgrounds such as buildings, mountains, forests, urban areas, clouds, water surfaces, and others. The Anti-UAV410 dataset is divided into three sets: the training set, which consists of 200 videos; the validation set, which consists of 90 videos; and the test set, which consists of 120 videos. These videos contain 1069 frames on average, and their resolution is 512 × 640. There are some differences in the data distribution between the Anti-UAV410 dataset and the 3rd Anti-UAV dataset. The latter has a higher proportion of difficult scenarios, which increases the difficulty of tracking UAVs.
On the three Anti-UAV Challenge test datasets, two commonly used evaluation metrics, namely precision plot and success plot, are utilized to assess the performance of location and scale estimation of all trackers through One-Pass Evaluation (OPE). Specifically, the precision plot calculates the percentage of frames in which the estimated target location falls within a given distance threshold from the ground truth. As shown in Equation (8), precision (P, with a threshold of 20 pixels) and normal precision (PNorm) can be used to measure the accuracy of locating the target. The success plot measures the fraction of successful frames where the Intersection over Union (IoU) between the predicted bounding box and the ground truth is greater than a threshold ranging from 0 to 1. As shown in Equation (9), the area under curve (AUC) of the success plot and 50% overlap precision (OP50) can be used to evaluate the accuracy of estimating the target scale. The above metrics are widely used in tracking benchmarks.
P = N d 20 N t o t a l × 100 % ,   P N o r m = N d n o r m 0.2 N t o t a l × 100 % ,
A U C = 0 1 N I o U > u N t o t a l   d u × 100 % ,   O P 50 = N I o U > 0.5 N t o t a l × 100 % ,
where N t o t a l is the total number of frames, N d 20 is the number of frames that satisfy the error distance d is not greater than 20 pixels, d n o r m is the normalized error distance, N I o U > u is the number of frames where IoU is greater than u , and N I o U > 0.5 is the number of frames where IoU is greater than 0.5.
According to [34], on the Anti-UAV410 dataset, we use the state accuracy (SA) to comprehensively evaluate the performance of all trackers, which is calculated as follows:
S A = t I o U t × v t + ( 1 e t ) × ( 1 v t ) T × 100 % ,
where v t and e t are the ground truth and predicted value of whether the target exists in the t-th frame, respectively. This metric not only requires the tracker to accurately predict the position and scale of the target, but also to determine the state of the target’s presence in the current frame, and make a judgment when the target disappears from the field of view.

4.2. Implementation Details

We use the SGD optimizer to train our proposed model for 24 epochs, and each epoch contains about 1200 iterations. The initial learning rate is 0.005, and is divided by 10 at the 8th and 16th epochs. The batch size is set to 12. We use an AMD EPYC 7302 CPU and four NVIDIA RTX 2080Ti GPUs for training and use one of them for testing.
During training, we use common data augmentation techniques such as horizontal flipping and photometric distortion, and also use random scaling to increase the proportion of data target scale variation. Our proposed auxiliary branch supervision and Gaussian noise are only used to train SAEM, while they are removed during testing.

4.3. Quantitative Evaluation

Our method is tested against 17 high-performance deep trackers, including 11 local trackers—ATOM [25], DiMP [26], PrDiMP [48], KYS [49], STARK [27], AiATrack [50], TOMP [28], MixFormer [29], OSTrack [7], SeqTrack [51], AQATrack [31]—and 6 trackers with target redetection capability—DaSiamRPN [24], SiamSTA [17], LTMU [36], GlobalTrack [8], CAMTracker [10], and SiamDT [34]. Among them, SiamSTA, GlobalTrack, and SiamDT are global trackers; DaSiamRPN is able to expand its search region when the target disappears; LTMU is attached with a global search module; and SiamSTA, CAMTracker, and SiamDT are specifically designed for the infrared UAV tracking task. For a fair comparison, these trackers are retrained on the 3rd Anti-UAV training dataset and the Anti-UAV410 training dataset, respectively.

4.3.1. Comparison Results on Anti-UAV Challenge Datasets

Table 2 shows the quantitative comparison results between our proposed method and 14 deep trackers. Figure 8 shows the corresponding precision plots (ranked based on P) and success plots (ranked based on AUC). Overall, the trackers with target recapture capability perform better than local trackers, and this proves that global search is essential for handling challenging scenarios in the infrared UAV tracking task. In addition, OSTrack is the best local tracker, while DaSiamRPN expands the search region to recapture the target, and its performance is lower than that of other trackers that directly redetect the target over the whole frame. Our proposed OSGT outperforms the state-of-the-art SiamDT in almost all four evaluation metrics on all three test datasets. Specifically, P is improved by 1.8% on the 2nd Anti-UAV test-dev, while AUC is improved by 1.9% and P is improved by 1.6% on the most difficult 3rd Anti-UAV val. After SAEM is embedded, the performance of OSGT is further improved, and it reaches the best place in most of the metrics on the three test datasets. In particular, AUC, OP50, P, and PNorm are improved by 1.3%, 1.5%, 1.5% and 1.2% on the 3rd Anti-UAV val, respectively. It indicates that the proposed SAEM is beneficial in improving the performance of OSGT.
It is noted that OSGT has a slight performance degradation on the 1st test-dev dataset after embedding SAEM. This is because this dataset consists mostly of simple scenarios without obvious target scale variation, but the template features still change slightly through SAEM, which affects the precision of location and scale estimation to a certain extent.

4.3.2. Comparison Results on Anti-UAV410 Dataset

Table 3 shows the quantitative comparison results between our proposed method and 8 deep trackers. OSGT outperforms the baseline GlobalTrack. After embedding SAEM, OSGT achieves state-of-the-art performance, outperforming SiamDT by improving SA by 0.79%. OSGT and SiamDT, respectively, focus on solving different problems of global trackers in the infrared UAV tracking task. SiamDT attempts to enhance the anti-interference ability of global trackers in complex backgrounds. To be specific, it detects and memorizes interference objects in the scene to exclude wrong candidate targets. In contrast, OSGT attempts to improve the scale adaptation of global trackers by implicitly resizing the template when the target scale changes. These two algorithms effectively improve the performance of global trackers from different aspects.

4.3.3. Inference Performance Comparison

Table 4 shows the inference performance of our proposed OSGT compared to four trackers with target redetection capability on an NVIDIA RTX 2080Ti GPU or an AMD EPYC 7302 CPU. These statistics include the preprocessing time (about 10 ms) for image loading, normalization, etc. On the GPU, the running speed of OSGT reaches 30.9 fps, which meets the real-time requirement and far exceeds SiamDT and LTMU. After SAEM is embedded, OSGT can still run in real time at 27.3 fps. It verifies that OSGT and SAEM are efficient. In comparison, all methods demonstrate relatively slow running speeds on the CPU, which reflects that developing lightweight models is still a challenge for practical applications in the infrared UAV tracking task.

4.4. Qualitative Evaluation

Figure 9 shows the visualization results of our method, OSTrack, GlobalTrack, LTMU, and SiamDT in four different scenarios. The first row shows that OSGT can robustly track targets in complex backgrounds, and particularly, it can estimate the size of the target more accurately. The second row shows that OSGT can track small targets well: in detail, the first two images show the case where the target position changes rapidly caused by camera movement, and the last two images show the case where the target temporarily disappears due to moving out of view, both of which are better handled by the global tracker than the local tracker by utilizing global search mechanism. The target scales in the third and fourth rows change from small to large and from large to small, respectively. GlobalTrack, SiamDT, and OSGT without SAEM track wrong targets such as flying birds and towers that are similar in size and appearance to the target template. OSTrack and LTMU have good scale adaptation by resizing the search region, but they still fail to track targets when the target scale changes drastically, as shown in the fourth row. After SAEM is embedded, OSGT can timely adjust the scale information of the target template features according to the target scale change, which improves the scale adaptation of OSGT. In addition, it can also be regarded as a specific scale constraint, which allows OSGT to effectively avoid being distracted by other objects with undesired scales.

4.5. Model Analysis

4.5.1. Ablation Study

Table 5 shows the effect of SAEM, supervision, and Gaussian noise on OSGT’s performance, and the latter two are used when training SAEM. The role of Gaussian noise is to simulate inaccurate input to SAEM. Obviously, training without Gaussian noise causes AUC and OP50 to severely decrease by 3.2% and 2.2%, respectively. This is because occlusion and background clutter lead to imprecise estimation of the target size in previous frames during online tracking, as shown in Figure 10. If SAEM processes the target template features according to the imprecise size, it will increase the error of the estimated bounding box. More seriously, if the auxiliary branch is used to supervise module training, SAEM will be more dependent on the exact input size, and the reduction in AUC and OP50 will be larger and reach 4.6% and 3.9%, respectively. This proves the importance of Gaussian noise to improve the robustness of estimating bounding boxes.
Under the premise of introducing Gaussian noise, using the auxiliary branch to supervise SAEM training can further improve the performance of OSGT. To be specific, AUC, OP50, P, and PNorm are improved by 0.6%, 0.5%, 0.9% and 0.8%, respectively. Compared with SAEM, which implicitly adjusts the scale information of the extracted target template features, the auxiliary branch adopts a more intuitive approach of first resizing the target template and then extracting its features. It can provide a ground truth for the output of SAEM during training, which significantly reduces the difficulty of module training. In summary, SAEM can better improve the scale adaptation of the global tracker only when both the supervision and Gaussian noise are used during training.

4.5.2. Different Backbone Networks

Table 6 shows the impact of using four backbone networks on the performance of OSGT, including Darknet, SwinTransformer [52], ResNeXt50 [53], and ResNet50. Among these, Darknet is the backbone of the earlier version of YOLO, SwinTransformer adopts a structure based on Transformer, and ResNeXt50 is an improved version of ResNet50 that introduces group convolution. Overall, ResNeXt50 achieves the best performance, while ResNet50 also performs well as the default backbone for OSGT. Darknet and SwinTransformer perform poorly due to their efficiency-oriented design and the unsuitable window mechanism for small targets, respectively.

4.5.3. Different Multi-Level Settings

Figure 11 shows the OSGT’s performance in tracking UAV targets of different sizes when using different multi-level settings. Among them, ‘1, 2, 3’ indicates using the last three levels to track targets with sizes belonging to [ 0 , 20 ) , [ 20 , 40 ) and [ 40 , + ) , respectively; ‘2, 3’ indicates using the last two levels to track targets with sizes belonging to [ 0 , 40 ) and [ 40 , + ) , respectively; and ‘3’ indicates that only the last level is used to track targets with sizes belonging to [ 0 , + ) —all targets. Obviously, using the lower level can effectively improve the OSGT’s performance for tracking smaller targets. This is because the lower level uses the feature map with higher resolution, which contains richer texture information and is more suitable for tracking targets with small size and weak appearance features. The higher level uses the feature map with lower resolution, which contains more semantic information and is suitable for OSGT to track large targets. Note that when only the last level is used, the learning of the model for small targets will degrade its tracking performance for large targets. In summary, using four levels enables OSGT to optimally track infrared UAV targets of different sizes.

4.5.4. Effectiveness of SAEM

The purpose of SAEM is to implicitly adjust the scale information of the extracted template features according to the target scale change. The explicit way to achieve this goal is to first resize the template and then extract its target features, which is also how the auxiliary branch supervision works for training SAEM. Figure 12 visualizes the template features obtained using the two methods. The results of both are approximately the same, which proves that SAEM achieves the expected effect.

4.5.5. Compatibility of SAEM

Our proposed SAEM is a plug-and-play module that can be used with other global trackers to improve their scale adaptation. Table 7 shows the impact of embedding SAEM into accessible global trackers, such as GlobalTrack and SiamDT. After SAEM is embedded, all metrics of GlobalTrack and SiamDT are significantly improved. To show the increment of scale adaptation, we test 3 global trackers using SAEM in the scale variation scenario, as shown in Figure 13. To be specific, P and AUC of GlobalTrack are improved by 2.1% and 4.1%, respectively; P and AUC of SiamDT are improved by 3.7% and 2.7%, respectively; and P and AUC of OSGT are improved by 5.9% and 2.7%, respectively. This not only verifies the effectiveness of SAEM in improving the scale adaptation of OSGT but also proves that SAEM can be embedded into other global trackers to effectively improve their ability to cope with target scale variation.

4.5.6. Number of Experts in SAEM

In Section 3.2.1, the expert in SAEK is the knowledge learned during training to adjust the scale information of the extracted target template features, and its specific form is the learnable model parameters. The impact of the number of experts on the model and its performance is shown in Table 8. When four more experts are added to SAEK, the number of model parameters increases by 0.13 K, the calculation amount increases by 0.28 K, the time consumption increases by about 0.8 ms, and the running speed decreases by about 0.6 fps. The more experts, the higher the tracking performance of the model, but too many experts degrade the performance due to model overfitting and increased training difficulty. The tracker performs best when the number of experts is 12. Overall, SAEM requires very little computation and time consumption.
Furthermore, Table 9 shows the impact of the number of experts on the model performance for different target scale variations (defined as the ratio between maximum and minimum target sizes in a sequence). Note that performance metrics vary significantly among scale variation intervals due to different tracking difficulties, but this does not affect the comparison within each scale variation interval. For small-scale variations within [1, 5), the number of experts has minimal impact on the model performance, with differences in both AUC and P remaining below 1%. For moderate scale variations within [5, 10), models with 12 or 16 experts outperform those with 4 or 8 experts by approximately 2% across nearly all metrics. For large-scale variations within [10,15], models employing 12 or 16 experts demonstrate significantly superior performance, especially the 12-expert configuration, achieving optimal results. In conclusion, larger target scale variations typically require a larger number of experts, and 12 experts can best handle the scale variations within [5,15].

4.5.7. Different Input Forms in SAEM

SAEM requires the change information of the target scale as input. There are usually two input forms: one is the ratio form [ h T - 1 / h 1 , w T - 1 / w 1 ] in ArbRCAN, and the other is our proposed concatenation form [ h 1 , w 1 , h T - 1 , w T - 1 ] . Table 10 shows the influence of these two input forms on the model performance. Compared with the ratio form, the concatenation form can provide the model with a base scale to avoid confusion and get better results.

4.5.8. Using Different Thresholds to Judge the Target Disappearance

In Section 3.2.3, considering the target is occluded or out of view, the algorithm needs to accurately judge the disappearance of the target and stop updating the previous target size input to SAEM. Otherwise, when the target reappears, it cannot be recaptured correctly due to inputting the wrong size to SAEM. Table 11 shows the effect of using different thresholds to judge the target disappearance on the model performance. When the threshold is 0, it means that the algorithm does not judge whether the target exists and updates the input of SAEM every frame. In this case, the performance of the tracker decreases drastically. As shown on the left of Figure 14, the previous target size is inaccurate due to the target’s disappearance, and if the size input of SAEM is updated, the tracker cannot track the target correctly when the target reappears. Our proposed adaptive threshold takes the maximum score of the first frame s T = 1 max as the base value, which can more accurately judge whether the small and dim target exists. Compared with using 0.5 as the threshold, our method improves AUC, OP50, and PNorm by 0.5%, 0.9% and 0.2%, respectively. As shown on the right of Figure 14, the target size is small, and its maximum score is 0.3531. If 0.5 is used as the threshold, the tracker will misjudge that the target has disappeared and stop updating the input of SAEM, which means that the target size cannot be correctly estimated in the subsequent tracking. In contrast, our method can adjust the threshold to 0.1766 and correctly determine the existence of small targets.

4.5.9. Comparison Results of Different Template Update Methods

Template update is a commonly used trick to deal with target appearance variations during online tracking. It is typically a temporal appearance update method, where the tracker replaces the original template with the tracking result every N frames [27]. In contrast, we only update the scale information of the target template feature, including an explicit update method (first resizing the target template and then extracting its features) and an implicit update method (using our proposed SAEM).
Table 12 shows the effect of different template update methods on the OSGT’s tracking results. The temporal appearance update (with N = 100) causes the degradation of the tracking performance. This is because most infrared UAV tracking scenarios are highly complex and prone to disturbing the tracking result, which makes the updated template unreliable, leading to subsequent tracking failure. In contrast, both scale update methods significantly improve the OSGT’s performance by enhancing its scale adaptation. However, compared with our proposed implicit scale update, the explicit scale update exhibits lower efficiency because it needs to extract features from the resized template every frame.

4.5.10. Tracking Failure Cases

The global tracker inherently possesses the ability to recapture targets by taking the whole frame as its search region. This advantage enables it to effectively handle target disappearance caused by occlusion and moving out of view. However, a large search region introduces more background interference, especially in extremely complex scenarios. As shown in Figure 15, GlobalTrack, SiamDT, and our OSGT+SAEM are interfered with by distant similar objects, resulting in tracking failure. In contrast, OStrack and LTMU, which employ local tracking schemes, can avoid this problem. However, if they get close to these similar objects, they will also be affected. Table 13 shows the quantitative comparison results of our method and the above 4 deep trackers in two tracking failure scenarios. All trackers demonstrate low performance metrics, even though SiamDT is specifically designed for anti-interference. Therefore, in addition to the poor scale adaptation, anti-interference is another critical research issue for global trackers, along with local trackers.

5. Conclusions

This paper focuses on solving the problem of poor scale adaptation of global trackers in the infrared UAV tracking task. To address this problem, we propose a plug-and-play scale adaptation enhancement module, which can implicitly resize the target template according to the target size in the previous frame. Moreover, we optimize the learning of SAEM by setting the auxiliary branch supervision and Gaussian noise. During the online tracking, an adaptive threshold is proposed to judge target disappearance and avoid SAEM being affected by the wrong input size. In addition, we propose a one-stage anchor-free global tracker, which has a concise structure to track UAVs in real time in infrared videos. On three Anti-UAV Challenge datasets and the Anti-UAV410 dataset, our proposed tracker is compared with 17 deep trackers and shows excellent performance, especially when dealing with the dramatic target scale variation.
Our future work will explore combining the anti-interference and scale adaptation of the global tracker to deal with more complex tracking scenarios.

Author Contributions

Methodology, Z.F.; software, Z.F. and D.L.; validation, W.Z. and D.L.; investigation, W.Z. and Q.Y.; writing, E.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Jiang, N.; Wang, K.; Peng, X.; Yu, X.; Wang, Q.; Xing, J.; Li, G.; Guo, G.; Ye, Q.; Jiao, J.; et al. Anti-UAV: A Large-scale Benchmark for Vision-based UAV Tracking. IEEE Trans. Multimed. 2023, 25, 486–500. [Google Scholar] [CrossRef]
  2. Yang, H.; Liang, B.; Feng, S.; Jiang, J.; Fang, A.; Li, C. Lightweight UAV Detection Method Based on IASL-YOLO. Drones 2025, 9, 325. [Google Scholar] [CrossRef]
  3. Ye, Z.; You, J.; Gu, J.; Kou, H.; Li, G. Modeling and Simulation of Urban Laser Countermeasures Against Low-Slow-Small UAVs. Drones 2025, 9, 419. [Google Scholar] [CrossRef]
  4. Javed, S.; Danelljan, M.; Khan, F.; Khan, M.; Felsberg, M.; Matas, J. Visual Object Tracking with Discriminative Filters and Siamese Networks: A Survey and Outlook. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 6552–6574. [Google Scholar] [CrossRef] [PubMed]
  5. Li, J.; Ye, D.H.; Kolsch, M.; Wachs, J.P.; Bouman, C.A. Fast and Robust UAV to UAV Detection and Tracking from Video. IEEE Trans. Emerg. Top. Comput. 2021, 10, 1519–1531. [Google Scholar] [CrossRef]
  6. Danelljan, M.; Bhat, G.; Khan, F.S.; Felsberg, M. ECO: Efficient Convolution Operators for Tracking. In Proceedings of the 2017 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6931–6939. [Google Scholar]
  7. Ye, B.; Chang, H.; Ma, B.; Shan, S.; Chen, X. Joint Feature Learning and Relation Modeling for Tracking: A One-Stream Framework. In Proceedings of the European Conference on Computer Vision (ECCV 2022), Tel Aviv, Israel, 23–27 October 2022; pp. 341–357. [Google Scholar]
  8. Huang, L.; Zhao, X.; Huang, K. Globaltrack: A Simple and Strong Baseline for Long-term Tracking. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 11037–11044. [Google Scholar]
  9. Fang, H.; Wang, X.; Liao, Z.; Chang, Y.; Yan, L. A Real-Time Anti-Distractor Infrared UAV Tracker with Channel Feature Refinement Module. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), Montreal, BC, Canada, 11–17 October 2021; pp. 1240–1248. [Google Scholar]
  10. Wang, Z.; Hu, Y.; Yang, J.; Zhou, G.; Liu, F.; Liu, Y. A Contrastive-Augmented Memory Network for Anti-UAV Tracking in TIR Videos. Remote Sens. 2024, 16, 4775. [Google Scholar] [CrossRef]
  11. Li, Y.; Zhu, J. A Scale Adaptive Kernel Correlation Filter Tracker with Feature Integration. In Proceedings of the European Conference on Computer Vision (ECCV 2015), Zurich, Switzerland, 6–12 September 2015; pp. 254–265. [Google Scholar]
  12. Danelljan, M.; Hager, G.; Khan, F.S.; Felsberg, M. Discriminative Scale Space Tracking. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1561–1575. [Google Scholar] [CrossRef]
  13. Zhang, Z.; Peng, H.; Fu, J.; Li, B.; Hu, W. Ocean: Object-Aware Anchor-Free Tracking. In Proceedings of the European Conference on Computer Vision (ECCV 2020), Glasgow, UK, 23–28 August 2020; pp. 771–787. [Google Scholar]
  14. Chen, X.; Yan, B.; Zhu, J.; Wang, D.; Yang, X.; Lu, H. Transformer Tracking. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 19–25 June 2021; pp. 8122–8131. [Google Scholar]
  15. Hu, X.; Mu, H.; Zhang, X.; Wang, Z.; Tan, T.; Sun, J. Meta-SR: A Magnification-Arbitrary Network for Super-Resolution. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 1575–1584. [Google Scholar]
  16. Wang, L.; Wang, Y.; Lin, Z.; Yang, J.; An, W.; Guo, Y. Learning a Single Network for Scale-Arbitrary Super-Resolution. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 4781–4790. [Google Scholar] [CrossRef]
  17. Huang, B.; Chen, J.; Xu, T.; Wang, Y.; Jiang, S.; Wang, Y.; Wang, L.; Li, J. SiamSTA: Spatio-temporal Attention based Siamese Tracker for Tracking UAVs. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), Montreal, BC, Canada, 11–17 October 2021; pp. 1204–1212. [Google Scholar]
  18. Tian, Z.; Shen, C.; Chen, H.; He, T. FCOS: Fully Convolutional One-Stage Object Detection. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9626–9635. [Google Scholar]
  19. Li, S.; Zhao, S.; Cheng, B.; Zhao, E.; Chen, J. Robust Visual Tracking via Hierarchical Particle Filter and Ensemble Deep Features. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 179–191. [Google Scholar] [CrossRef]
  20. Hare, S.; Golodetz, S.; Saffari, A.; Vineet, V.; Cheng, M.-M.; Hicks, S.L.; Torr, P.H.S. Struck: Structured Output Tracking with Kernels. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 2096–2109. [Google Scholar] [CrossRef] [PubMed]
  21. Henriques, J.F.; Caseiro, R.; Martins, P.; Batista, J. Exploiting the Circulant Structure of Tracking-by-Detection with Kernels. In Proceedings of the European Conference on Computer Vision (ECCV 2012), Florence, Italy, 7–13 October 2012; pp. 702–715. [Google Scholar]
  22. Li, B.; Wu, W.; Wang, Q.; Zhang, F.; Xing, J.; Yan, J. SiamRPN++: Evolution of Siamese Visual Tracking with Very Deep Networks. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–21 June 2019; pp. 4277–4286. [Google Scholar]
  23. Xie, X.; Xi, J.; Yang, X.; Lu, R.; Xia, W. STFTrack: Spatio-Temporal-Focused Siamese Network for Infrared UAV Tracking. Drones 2023, 7, 296. [Google Scholar] [CrossRef]
  24. Zhu, Z.; Wang, Q.; Li, B.; Wu, W.; Yan, J.; Hu, W. Distractor-Aware Siamese Networks for Visual Object Tracking. In Proceedings of the European Conference on Computer Vision (ECCV 2018), Munich, Germany, 8–14 September 2018; pp. 103–119. [Google Scholar]
  25. Danelljan, M.; Bhat, G.; Khan, F.S.; Felsberg, M. ATOM: Accurate Tracking by Overlap Maximization. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–21 June 2019; pp. 4655–4664. [Google Scholar]
  26. Bhat, G.; Danelljan, M.; Van Gool, L.; Timofte, R. Learning Discriminative Model Prediction for Tracking. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6181–6190. [Google Scholar] [CrossRef]
  27. Yan, B.; Peng, H.; Fu, J.; Wang, D.; Lu, H. Learning Spatio-Temporal Transformer for Visual Tracking. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Virtual, 11–17 October 2021; pp. 10428–10437. [Google Scholar]
  28. Mayer, C.; Danelljan, M.; Bhat, G.; Paul, M.; Paudel, D.P.; Yu, F.; Van Gool, L. Transforming model prediction for tracking. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 8731–8740. [Google Scholar]
  29. Cui, Y.; Song, T.; Wu, G.; Wang, L. MixFormerV2: Efficient Fully Transformer Tracking. arXiv 2023, arXiv:2305.15896. [Google Scholar]
  30. Wei, X.; Bai, Y.; Zheng, Y.; Shi, D.; Gong, Y. Autoregressive Visual Tracking. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 9697–9706. [Google Scholar]
  31. Xie, J.; Zhong, B.; Mo, Z.; Zhang, S.; Shi, L.; Song, S.; Ji, R. Autoregressive Queries for Adaptive Tracking with Spatio-Temporal Transformers. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024; pp. 19300–19309. [Google Scholar]
  32. Voigtlaender, P.; Luiten, J.; Torr, P.H.S.; Leibe, B. Siam R-CNN: Visual Tracking by Re-Detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 14–19 June 2020; pp. 6577–6587. [Google Scholar] [CrossRef]
  33. Huang, B.; Dou, Z.; Chen, J.; Li, J.; Shen, N.; Wang, Y.; Xu, T. Searching Region-Free and Template-Free Siamese Network for Tracking Drones in TIR Videos. IEEE Trans. Geosci. Remote Sens. 2023, 62, 5000315. [Google Scholar] [CrossRef]
  34. Huang, B.; Li, J.; Chen, J.; Wang, G.; Zhao, J.; Xu, T. Anti-UAV410: A Thermal Infrared Benchmark and Customized Scheme for Tracking Drones in the Wild. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 2852–2865. [Google Scholar] [CrossRef] [PubMed]
  35. Yan, B.; Zhao, H.; Wang, D.; Lu, H.; Yang, X. ‘Skimming-Perusal’ Tracking: A Framework for Real-Time and Robust Long-Term Tracking. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 2385–2393. [Google Scholar]
  36. Dai, K.; Zhang, Y.; Wang, D.; Li, J.; Lu, H.; Yang, X. High-Performance Long-Term Tracking with Meta-Updater. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 14–19 June 2020; pp. 6297–6306. [Google Scholar] [CrossRef]
  37. Qian, K.; Zhu, D.; Wu, Y.; Shen, J.; Zhang, S. TransIST: Transformer Based Infrared Small Target Tracking Using Multi-Scale Feature and Exponential Moving Average Learning. Infrared Phys. Technol. 2025, 145, 105674. [Google Scholar] [CrossRef]
  38. Zhao, J.; Zhang, X.; Zhang, P. A Unified Approach for Tracking UAVs in Infrared. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), Virtual, 11–17 October 2021; pp. 1213–1222. [Google Scholar]
  39. Liu, S.; Huang, W.; Chen, X. Is Brain-inspired Intelligence a New Dawn for Infrared Imaging Missile Anti-interference Strategies? Chin. J. Aeronaut. 2025, 38, 103462. [Google Scholar] [CrossRef]
  40. Wu, H.; Li, W.; Li, W.; Liu, G. A Real-Time Robust Approach for Tracking UAVs in Infrared Videos. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Virtual, 14–19 June 2020; pp. 4448–4455. [Google Scholar] [CrossRef]
  41. Yu, Q.; Ma, Y.; He, J.; Yang, D.; Zhang, T. A Unified Transformer Based Tracker for Anti-UAV Tracking. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Vancouver, BC, Canada, 18–22 June 2023; pp. 3035–3045. [Google Scholar]
  42. Lim, B.; Son, S.; Kim, H.; Nah, S.; Lee, K.M. Enhanced Deep Residual Networks for Single Image Super-Resolution. In Proceedings of the 2017 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1132–1140. [Google Scholar]
  43. Liang, J.; Cao, J.; Sun, G.; Zhang, K.; Van Gool, L.; Timofte, R. SwinIR: Image Restoration Using Swin Transformer. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), Virtual, 11–17 October 2021; pp. 1833–1844. [Google Scholar] [CrossRef]
  44. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar] [CrossRef]
  45. Zhou, X.; Wang, D.; Krähenbühl, P. Objects as Points. arXiv 2019, arXiv:1904.07850. [Google Scholar]
  46. He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef]
  47. Zhao, J.; Wang, G.; Li, J.; Jin, L.; Fan, N.; Wang, M.; Wang, X.; Yong, T.; Deng, Y.; Guo, Y.; et al. The 2nd Anti-UAV Workshop & Challenge: Methods and Results. arXiv 2021, arXiv:2108.09909. [Google Scholar]
  48. Danelljan, M.; Van Gool, L.; Timofte, R. Probabilistic Regression for Visual Tracking. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 14–19 June 2020; pp. 7181–7190. [Google Scholar] [CrossRef]
  49. Bhat, G.; Danelljan, M.; Van Gool, L.; Timofte, R. Know Your Surroundings: Exploiting Scene Information for Object Tracking. In Proceedings of the European Conference on Computer Vision (ECCV 2020), Virtual, 23–28 August 2020; pp. 205–221. [Google Scholar]
  50. Gao, S.; Zhou, C.; Ma, C.; Wang, X.; Yuan, J. AiATrack: Attention in Attention for Transformer Visual Tracking. In Proceedings of the European Conference on Computer Vision (ECCV 2022), Tel Aviv, Israel, 23–27 October 2022; pp. 146–164. [Google Scholar]
  51. Chen, X.; Peng, H.; Wang, D.; Lu, H.; Hu, H. SeqTrack: Sequence to Sequence Learning for Visual Object Tracking. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 14572–14581. [Google Scholar] [CrossRef]
  52. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  53. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. In Proceedings of the 2017 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1492–1500. [Google Scholar]
Figure 1. Comparison of the local tracker and the global tracker in handling target scale variation. (a) Local trackers can use the resizing operation in the preprocessing to cope with target scale variation; (b) global trackers perform poorly when dealing with target scale variation. The red arrow indicates that our proposed SAEM, which is embedded into the global tracker, uses the previous target size to implicitly resize the template to improve the scale adaptation.
Figure 1. Comparison of the local tracker and the global tracker in handling target scale variation. (a) Local trackers can use the resizing operation in the preprocessing to cope with target scale variation; (b) global trackers perform poorly when dealing with target scale variation. The red arrow indicates that our proposed SAEM, which is embedded into the global tracker, uses the previous target size to implicitly resize the template to improve the scale adaptation.
Drones 09 00469 g001
Figure 2. The overall framework of OSGT consists of feature extraction, feature fusion, and output head. In addition, our proposed SAEM is embedded into the feature fusion part, as shown in the red dashed box.
Figure 2. The overall framework of OSGT consists of feature extraction, feature fusion, and output head. In addition, our proposed SAEM is embedded into the feature fusion part, as shown in the red dashed box.
Drones 09 00469 g002
Figure 3. Visualization feature maps of the target template with different scales. For the convenience of display, we arrange the 256 × 7 × 7 tensor into 16 × 16 small squares with the size 7 × 7.
Figure 3. Visualization feature maps of the target template with different scales. For the convenience of display, we arrange the 256 × 7 × 7 tensor into 16 × 16 small squares with the size 7 × 7.
Drones 09 00469 g003
Figure 4. Our proposed scale adaptation enhancement module.
Figure 4. Our proposed scale adaptation enhancement module.
Drones 09 00469 g004
Figure 5. The auxiliary branch for supervising the learning of SAEM and Gaussian noise is added to the previous target size to enhance the robustness of SAEM.
Figure 5. The auxiliary branch for supervising the learning of SAEM and Gaussian noise is added to the previous target size to enhance the robustness of SAEM.
Drones 09 00469 g005
Figure 6. Histograms of the change ratios for target width and height between adjacent frames on the Anti-UAV dataset.
Figure 6. Histograms of the change ratios for target width and height between adjacent frames on the Anti-UAV dataset.
Drones 09 00469 g006
Figure 7. The red line shows the relationship between the target size and the maximum score s max , the green dashed line represents the threshold of 0.5, the blue dashed line represents our proposed adaptive threshold, and the gray shaded area shows the misjudgment when using 0.5 as the threshold.
Figure 7. The red line shows the relationship between the target size and the maximum score s max , the green dashed line represents the threshold of 0.5, the blue dashed line represents our proposed adaptive threshold, and the gray shaded area shows the misjudgment when using 0.5 as the threshold.
Drones 09 00469 g007
Figure 8. Precision plots and success plots of our method and 13 deep trackers on three Anti-UAV Challenge test datasets.
Figure 8. Precision plots and success plots of our method and 13 deep trackers on three Anti-UAV Challenge test datasets.
Drones 09 00469 g008aDrones 09 00469 g008b
Figure 9. Visualization results of our method, OSTrack, GlobalTrack, LTMU, and SiamDT.
Figure 9. Visualization results of our method, OSTrack, GlobalTrack, LTMU, and SiamDT.
Drones 09 00469 g009aDrones 09 00469 g009b
Figure 10. Using Gaussian noise during training enhances the performance of our model to accurately estimate the target size.
Figure 10. Using Gaussian noise during training enhances the performance of our model to accurately estimate the target size.
Drones 09 00469 g010
Figure 11. The impact of different multi-level settings on OSGT’s performance.
Figure 11. The impact of different multi-level settings on OSGT’s performance.
Drones 09 00469 g011
Figure 12. The template feature maps are obtained by using SAEM and the explicit method, respectively. Some feature variations are shown in the black dashed circle.
Figure 12. The template feature maps are obtained by using SAEM and the explicit method, respectively. Some feature variations are shown in the black dashed circle.
Drones 09 00469 g012
Figure 13. The enhanced scale adaptation of OSGT, GlobalTrack, and SiamDT after embedding SAEM.
Figure 13. The enhanced scale adaptation of OSGT, GlobalTrack, and SiamDT after embedding SAEM.
Drones 09 00469 g013
Figure 14. The influence of using different thresholds to judge the target disappearance on tracking results.
Figure 14. The influence of using different thresholds to judge the target disappearance on tracking results.
Drones 09 00469 g014
Figure 15. Tracking failure cases due to background interference.
Figure 15. Tracking failure cases due to background interference.
Drones 09 00469 g015
Table 1. Comparison of local and global trackers in the infrared UAV tracking task.
Table 1. Comparison of local and global trackers in the infrared UAV tracking task.
SchemeSearch RegionEfficiencyScale AdaptationOcclusion or
Moving out of View
Fast Target or
Camera Motion
Local trackerLocal patchHighDrones 09 00469 i001Drones 09 00469 i002Drones 09 00469 i002
Global trackerWhole frameLowDrones 09 00469 i002Drones 09 00469 i001Drones 09 00469 i001
Table 2. Quantitative comparison results of our method and 14 deep trackers on three Anti-UAV Challenge test datasets.
Table 2. Quantitative comparison results of our method and 14 deep trackers on three Anti-UAV Challenge test datasets.
MethodsPublication1st Anti-UAV test-dev2nd Anti-UAV test-dev3rd Anti-UAV val
AUCOP50PPNormAUCOP50PPNormAUCOP50PPNorm
ATOM [25]CVPR 201961.677.979.378.954.168.872.569.543.154.758.557.6
DiMP [26]ICCV 201966.884.085.284.959.174.677.775.347.458.864.462.1
PrDiMP [48]CVPR 202069.287.789.188.761.378.182.279.049.062.166.464.2
KYS [49]ECCV 202067.384.585.885.559.675.378.476.049.060.967.163.5
STARK [27]ICCV 202169.587.489.488.562.078.382.279.148.862.169.064.0
TOMP [28]CVPR 202265.882.083.082.857.872.174.372.943.855.260.857.8
OSTrack [7]ECCV 202272.491.393.692.762.779.583.479.951.964.868.767.2
SeqTrack [51]CVPR 202355.371.473.272.950.163.766.965.243.555.362.057.9
AQATrack [31]CVPR 202470.388.990.989.960.977.080.778.047.559.666.262.3
DaSiamRPN [24]ECCV 201868.788.190.787.957.774.577.274.842.053.059.655.7
GlobalTrack [8]AAAI 202075.695.597.596.465.583.189.385.253.066.374.770.5
LTMU [36]CVPR 202075.895.396.796.268.686.488.388.155.469.273.372.3
SiamSTA # [17]ICCVW 202172.696.965.588.8
SiamDT [34]PAMI 202476.496.297.797.268.587.189.489.153.367.175.070.3
OSGT76.296.698.097.368.688.391.289.855.270.576.675.2
OSGT+SAEM76.496.297.997.369.488.991.790.556.572.078.176.4
The best three results are shown in red, blue, and green fonts. The gray background indicates the results of trackers with target redetection capability. # means the values are taken from their publications.
Table 3. Quantitative comparison results of our method and 8 deep trackers on the Anti-UAV410 test dataset.
Table 3. Quantitative comparison results of our method and 8 deep trackers on the Anti-UAV410 test dataset.
MethodsPrDiMP
[48]
STARK
[27]
AiATrack
[50]
OSTrack
[7]
MixFormer
[29]
GlobalTrack
[8]
CAMTracker #
[10]
SiamDT #
[34]
OSGTOSGT
+SAEM
PublicationCVPR 2020 ICCV 2021ECCV
2022
ECCV 2022CVPR
2023
AAAI
2020
RS
2024
PAMI 2024
SA54.6957.1559.5660.1559.6566.4567.1068.1967.0368.98
The gray background indicates the results of trackers with target redetection capability. # means the values are taken from their publications.
Table 4. Inference performance of our method compared to 4 trackers with target redetection capability on an NVIDIA RTX 2080Ti GPU or an AMD EPYC 7302 CPU.
Table 4. Inference performance of our method compared to 4 trackers with target redetection capability on an NVIDIA RTX 2080Ti GPU or an AMD EPYC 7302 CPU.
MethodsDaSiamRPNGlobalTrackLTMUSiamDTOSGTOSGT+SAEM
Speed (GPU)22.722.31.59.130.927.3
Speed (CPU)2.02.00.53.22.2
Table 5. The influence of SAEM, supervision, and Gaussian noise on model performance. The model is tested on 3rd Anti-UAV val.
Table 5. The influence of SAEM, supervision, and Gaussian noise on model performance. The model is tested on 3rd Anti-UAV val.
OSGTSAEMSupervisionGaussian NoiseAUCOP50PPNorm
Drones 09 00469 i001 55.270.576.675.2
Drones 09 00469 i001Drones 09 00469 i001 52.769.376.875.3
Drones 09 00469 i001Drones 09 00469 i001Drones 09 00469 i001 51.968.176.874.5
Drones 09 00469 i001Drones 09 00469 i001 Drones 09 00469 i00155.971.577.075.6
Drones 09 00469 i001Drones 09 00469 i001Drones 09 00469 i001Drones 09 00469 i00156.572.078.176.4
Table 6. The impact of using different backbone networks on OSGT’s performance.
Table 6. The impact of using different backbone networks on OSGT’s performance.
MetricsDarknetSwinTransformerResNeXt50ResNet50
AUC50.154.455.355.2
OP5062.968.769.870.5
P69.575.877.676.6
PNorm66.972.575.475.2
Table 7. The impact of embedding SAEM into different global trackers.
Table 7. The impact of embedding SAEM into different global trackers.
MetricsGlobalTrackGlobalTrack
+SAEM
SiamDTSiamDT
+SAEM
OSGTOSGT
+SAEM
AUC53.054.4 (1.4 ↑)53.355.3 (2.0 ↑)55.256.5 (1.3 ↑)
OP5066.368.2 (1.9 ↑)67.169.5 (2.4 ↑)70.572.0 (1.5 ↑)
P74.776.1 (1.4 ↑)75.076.1 (1.1 ↑)76.678.1 (1.5 ↑)
PNorm70.573.1 (2.6 ↑)70.372.8 (2.5 ↑)75.276.4 (1.2 ↑)
Table 8. The impact of the number of experts on the model and its performance.
Table 8. The impact of the number of experts on the model and its performance.
ExpertsParams.FLOPsTimeAUCOP50PPNormSpeed
41.35 K2.58 K1.72 ms55.670.676.974.628.5
81.48 K2.86 K2.53 ms55.871.476.976.027.9
121.61 K3.14 K3.31 ms56.572.078.176.427.3
161.74 K3.42 K4.11 ms56.272.077.575.226.6
Table 9. The impact of the number of experts on the model’s AUC/OP50/P/PNorm for different target scale variations.
Table 9. The impact of the number of experts on the model’s AUC/OP50/P/PNorm for different target scale variations.
Experts[1, 5)[5, 10)[10, 15]
446.4/71.0/73.0/55.421.3/35.0/41.8/42.855.4/74.0/79.8/65.5
846.2/70.5/72.0/57.120.9/35.2/40.6/41.754.6/72.5/76.1/63.5
1245.8/70.4/72.2/54.623.3/37.8/43.7/45.160.0/83.1/89.8/75.4
1646.3/71.6/73.0/57.523.0/37.8/43.8/40.158.0/78.8/81.8/67.5
Table 10. The influence of different input forms in SAEM on model performance.
Table 10. The influence of different input forms in SAEM on model performance.
Input FormsAUCOP50PPNorm
Ratio form56.272.077.475.6
Concatenation form56.572.078.176.4
Table 11. The effect of using different thresholds to judge the target disappearance on model performance.
Table 11. The effect of using different thresholds to judge the target disappearance on model performance.
ThresholdAUCOP50PPNorm
0.055.470.577.074.3
0.556.071.178.476.2
0.5 s T = 1 max 56.572.078.176.4
Table 12. Comparison results of different template update methods.
Table 12. Comparison results of different template update methods.
Template Update MethodsAUCOP50PPNormSpeed
None55.270.576.675.230.9
Temporal appearance update54.970.276.474.729.9
Explicit scale update56.271.477.776.023.3
Implicit scale update (SAEM)56.572.078.176.427.3
Table 13. Quantitative comparison results of our method and 4 deep trackers in two tracking failure scenarios.
Table 13. Quantitative comparison results of our method and 4 deep trackers in two tracking failure scenarios.
MethodsAUCOP50PPNorm
OSTrack4.3/23.98.0/37.99.5/39.911.9/23.7
LTMU16.3/29.030.2/41.536.4/40.930.6/30.5
GlobalTrack4.6/36.88.7/61.612.0/62.816.1/42.9
SiamDT6.0/18.610.3/30.116.1/30.515.7/20.4
OSGT+SAEM18.2/41.230.0/62.829.1/63.028.0/45.1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Feng, Z.; Zhang, W.; Pan, E.; Liu, D.; Yu, Q. Enhancing the Scale Adaptation of Global Trackers for Infrared UAV Tracking. Drones 2025, 9, 469. https://doi.org/10.3390/drones9070469

AMA Style

Feng Z, Zhang W, Pan E, Liu D, Yu Q. Enhancing the Scale Adaptation of Global Trackers for Infrared UAV Tracking. Drones. 2025; 9(7):469. https://doi.org/10.3390/drones9070469

Chicago/Turabian Style

Feng, Zicheng, Wenlong Zhang, Erting Pan, Donghui Liu, and Qifeng Yu. 2025. "Enhancing the Scale Adaptation of Global Trackers for Infrared UAV Tracking" Drones 9, no. 7: 469. https://doi.org/10.3390/drones9070469

APA Style

Feng, Z., Zhang, W., Pan, E., Liu, D., & Yu, Q. (2025). Enhancing the Scale Adaptation of Global Trackers for Infrared UAV Tracking. Drones, 9(7), 469. https://doi.org/10.3390/drones9070469

Article Metrics

Back to TopTop