Next Article in Journal
Decrypting Spatiotemporal Dynamics and Optimization Pathway of Ecological Resilience Under a Panarchy-Inspired Framework: Insights from the Wuhan Metropolitan Area
Previous Article in Journal
Machine Learning Approaches for Assessing Avocado Alternate Bearing Using Sentinel-2 and Climate Variables—A Case Study in Limpopo, South Africa
Previous Article in Special Issue
EAS-Det: Edge-Aware Semantic Feature Fusion for Robust 3D Object Detection in LiDAR Point Clouds
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

TAF-YOLO: A Small-Object Detection Network for UAV Aerial Imagery via Visible and Infrared Adaptive Fusion

1
College of Missile Engineering, Rocket Force University of Engineering, Xi’an 710025, China
2
School of Remote Sensing Information Engineering, Wuhan University, Wuhan 430079, China
3
School of Electrical and Control Engineering, Xuzhou University of Technology, Xuzhou 221018, China
4
Department of Geography and Resource Management, The Chinese University of Hong Kong, Hong Kong 999077, China
5
State Key Laboratory of Integrated Services Networks, Xidian University, Xi’an 710071, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(24), 3936; https://doi.org/10.3390/rs17243936
Submission received: 28 October 2025 / Revised: 30 November 2025 / Accepted: 3 December 2025 / Published: 5 December 2025

Highlights

What are the main findings?
  • An early-fusion multimodal small-object detection framework, TAF-YOLO, is proposed for UAV aerial imagery. It effectively integrates complementary visible and infrared information at the pixel level, thereby improving small-object detection accuracy while reducing missed detections and false alarms.
  • TAF-YOLO removes the PANet from the baseline model, reduces redundant information and introduces lightweight DSAB into a multimodal detection framework, achieving efficient and accurate detection of small objects.
What are the implications of the main findings?
  • The proposed framework shows that early fusion of visible and infrared modalities improves robustness and precision in UAV-based small-object detection, suggesting a promising direction for multimodal sensing in complex environments.
  • The integration of lightweight modules and removal of redundant network structures provide evidence that high detection accuracy can be achieved with reduced computational cost, supporting broader deployment in real-time and resource-constrained scenarios.

Abstract

Detecting small objects from UAV-captured aerial imagery is a critical yet challenging task, hindered by factors such as small object size, complex backgrounds, and subtle inter-class differences. Single-modal methods lack the robustness for all-weather operation, while existing multimodal solutions are often too computationally expensive for deployment on resource-constrained UAVs. To this end, we propose TAF-YOLO, a lightweight and efficient multimodal detection framework designed to balance accuracy and efficiency. First, we propose an early fusion module, the Two-branch Adaptive Fusion Network (TAFNet), which adaptively integrates visible and infrared information at both pixel and channel levels before the feature extractor, maximizing complementary data while minimizing redundancy. Second, we propose a Large Adaptive Selective Kernel (LASK) module that dynamically expands the receptive field using multi-scale convolutions and spatial attention, preserving crucial details of small objects during downsampling. Finally, we present an optimized feature neck architecture that replaces PANet’s bidirectional path with a more efficient top-down pathway. This is enhanced by a Dual-Stream Attention Bridge (DSAB) that injects high-level semantics into low-level features, improving localization without significant computational overhead. On the VEDAI benchmark, TAF-YOLO achieves 67.2% mAP50, outperforming the CFT model by 2.7% and demonstrating superior performance against seven other YOLO variants. Our work presents a practical and powerful solution that enables real-time, all-weather object detection on resource-constrained UAVs.

1. Introduction

With the rapid development of unmanned aerial vehicle (UAV) technology [1] and the miniaturization of onboard sensors [2], UAV-based visual perception has become a key tool in applications such as disaster monitoring [3], agricultural management [4], and urban surveillance [5]. Compared with satellite or ground-based observation, UAV imagery offers flexible viewpoints and high temporal resolution but also presents unique challenges for object detection: targets are typically small, densely distributed, and surrounded by complex backgrounds [6]. In this context, we define small objects as targets occupying less than 1% of the image area or fewer than 32 × 32 pixels [7]. Their extremely limited spatial footprint makes them easily confused with background textures, and variations in UAV altitude or environment further change their scale and appearance. As a result, reliably detecting such small objects in UAV imagery remains a highly challenging task [8].
Although recent advances in deep learning, particularly in Convolutional Neural Networks (CNNs) and Transformer-based architectures, have led to remarkable progress in UAV remote sensing detection [9,10,11,12], methods relying solely on visible images are sensitive to challenging conditions such as low light, heavy fog, or rain, where their performance degrades sharply. In contrast, infrared (IR) imaging, which captures thermal radiation from objects, is robust to lighting variations and capable of all-weather operation, effectively highlighting targets obscured by poor illumination or occlusion. However, IR imagery inherently suffers from low spatial resolution and a lack of color and texture details. Therefore, visible and infrared modalities are naturally complementary: the former provides high-frequency texture and color information, while the latter contributes stable, illumination-invariant thermal cues.
Modern UAVs are often equipped with dual-sensor gimbals integrating both visible and infrared cameras, providing an ideal foundation for multimodal fusion (MF). Effectively fusing these two modalities can significantly enhance the robustness and reliability of UAV perception systems, which is crucial for improving the detection of small objects in complex environments. Consequently, building an efficient end-to-end visible-infrared fusion detection network that fully leverages inter-modality complementarity is an inevitable requirement for the advancement of intelligent visual perception technology.
Currently, deep learning-based visible-infrared fusion (VIF) methods [13] are primarily categorized into early (pixel-level), intermediate (feature-level), and late (decision-level) fusion. Early fusion [14], though computationally efficient, often relies on simple operations like addition or concatenation, failing to adequately model the deeper relationships between modalities. Intermediate [15] and late fusion [16] strategies (e.g., those using Transformer-based cross-attention [17,18]) better capture cross-modal interactions but still face challenges such as feature space misalignment, information loss, and insufficient adaptation to UAV-specific scenarios. Existing fusion frameworks generally suffer from three key limitations: overly simplistic early fusion strategies, information loss in later fusion stages, and inadequate adaptation to UAV small-object scenarios, where feature degradation is particularly severe.
To address these issues, this paper proposes TAF-YOLO, a multimodal small-object detection network designed specifically for UAV aerial imagery. Before feature extraction, TAFNet adaptively fuses visible and infrared information at both the pixel and channel levels. This early fusion strategy effectively integrates complementary information while suppressing redundant features, providing a high-quality input for subsequent detection. To mitigate feature degradation during downsampling, the LASK module dynamically adjusts its receptive field using multi-scale depthwise convolutions and spatial attention. Combined with an anti-aliasing filtering strategy, it preserves edge information and enhances the feature representation of small objects. In the network’s neck, we introduce the lightweight DSAB module for the first time. It injects high-level semantics into shallow features via semantic-guided foreground-background routing and multi-scale context aggregation, significantly improving localization precision for small objects while maintaining semantic integrity.
Through these designs, TAF-YOLO achieves an excellent balance between detection accuracy and computational efficiency. The source code will be made available at https://github.com/Yecis/TAF-YOLO (accessed on 30 November 2025). The main contributions of this paper are summarized as follows:
  • We propose TAF-YOLO, a novel end-to-end multimodal fusion framework specifically engineered for robust small-object detection in complex UAV scenarios. By integrating a series of targeted optimizations, the framework achieves state-of-the-art accuracy among YOLO-based visible–infrared models while maintaining a UAV-friendly computational budget.
  • To overcome ineffective early fusion and feature degradation during downsampling in multimodal feature fusion and propagation, we introduce a set of innovative modules. We introduce an efficient early fusion network, TAFNet, to maximize modal complementarity while preserving critical details at the input stage; the LASK module in the backbone to mitigate feature degradation of small objects via adaptive receptive fields and an anti-aliasing strategy; and the DSAB module in the neck to enhance localization precision through targeted cross-layer semantic injection.
  • Extensive experiments on public UAV visible–infrared datasets are conducted to validate the effectiveness of the proposed modules and overall framework. The results demonstrate that, compared to current state-of-the-art methods, TAF-YOLO achieves superior performance in both accuracy and efficiency.
The rest of this article is arranged as follows. The related works are briefly reviewed in Section 2. The details of the proposed method are described in Section 3. Experiments and evaluations are provided in Section 4. The discussion in Section 5. A conclusion about this article is given in Section 6.

2. Related Work

2.1. Deep Learning Based Object Detection

Object detection aims to automatically identify and locate specific objects in spatial images. With the development of deep neural networks, deep learning-based object detection technology has been widely applied on intelligent platforms such as UAVs. Currently, the predominant deep learning detection algorithms can be categorized into the CNN-based two-stage detectors, CNN-based one-stage detectors and the more recent Transformer-based detectors.
Two-stage detectors, represented by R-CNN [19], Fast R-CNN [20], Faster R-CNN [21], and Cascade R-CNN [22], first generate region proposals and then perform classification and regression in separate steps to complete detection. These approaches offer high accuracy but are less suitable for embedded or UAV platforms due to the high computational demands. In contrast, one-stage detectors such as SSD [23] and RetinaNet [24] leverage dense anchor priors or focal loss to achieve a significant speed boost while maintaining accuracy. Meanwhile, Transformer-based detectors, such as the DETR family [25,26], reformulate object detection as a set prediction problem, relying on global self-attention mechanisms and a Hungarian algorithm-based matching strategy to realize an end-to-end framework that obviates the need for non-maximum suppression (NMS). Furthermore, models such as Swin Transformer [27] and ViT [28] have been extensively employed as backbones or detection heads, integrating long-range dependency modeling with hierarchical feature representation to enhance both accuracy and scalability.
In one-stage detection methods, many studies have adopted the YOLO series as the foundational framework for UAV-based small-object detection owing to its efficient detection capability. From the original YOLO [29] to the widely deployed YOLOv5 [30] and YOLOv8 [31] models in both academia and industry, the series has undergone continuous evolution. Recent variants—such as YOLOv10 [32], which emphasizes NMS-free operation and high accuracy, as well as the lightweight YOLOv11 [33] and the attention-based YOLOv12 [34]—further refine the architecture and extend its applicability across diverse real-world scenarios.
Small objects in UAV aerial imagery are densely distributed and complex in background, causing the traditional detection algorithms to be prone to false and missed. Wu et al. [35] proposed MV-YOLO which leverages the Mamba Self-Attention Vision Module (MSAVM) to enhance long-range interactions, combined with the Context Clue Guided Module (CCGM) to refine the representation of small objects, thereby achieving efficient and accurate detection of small objects. Liu et al. [36] proposed MFAE-YOLO, which introduces a Global Feature Fusion Processing (GFFP) module to enhance global feature extraction and an Accurate Intersection over Union (AIoU) loss function to improve the precision of bounding box regression, thereby improving the detection accuracy of small objects in densely packed scenarios. Cao et al. [37] proposed GCL-YOLO, a lightweight YOLO network based on GhostConv, which introduces a new prediction head special for UAV small objects and uses the Focal-Efficient Intersection over Union (Focal-EIOU) loss as the localization loss function, improving detection accuracy while being lightweight. Detecting small objects in infrared images, especially in low-contrast and complex backgrounds, remains challenging. Tang et al. [38] proposed IRSTD-YOLO, an infrared small target detection network, which introduces the infrared small target enhancement (IRSTE) module and the edge and feature extraction (EFE) module, enhancing small target representation and detection robustness. Hao et al. [39] proposed YOLO-SR, which combines image super-resolution technology with YOLO, leveraging the structure-preserving super-resolution (SPSR) to preprocess the input infrared images before they enter the detection network, which enhances the accuracy of the network.
Although the above methods have achieved significant improvements in detection accuracy and speed, single-modal detectors remain vulnerable to external factors such as lighting, weather variations, and occlusion. The limited representational capacity of a single modality makes it difficult to maintain consistent and reliable detection performance under complex environmental conditions. In contrast, multimodal detection techniques provide an effective solution by integrating complementary information from multiple sources.

2.2. Multimodal Fusion Object Detection

With the development of sensor technology, multimodal data has been widely applied in fields such as autonomous driving [40] and remote sensing [41]. Multimodal object detection addresses the detection failure problem caused by light change, occlusion or noise in a single mode by jointly optimizing the data from different physical sensors such as optical and infrared sensors. It fully leverages the complementary characteristics of different modal data to effectively improve the robustness and accuracy of detection. Different sensors in multimodal object detection each have their own advantages but also exhibit inherent limitations. Visible-light and infrared cameras are particularly suitable for UAV platforms due to their compact size, low weight, low power consumption, and real-time imaging capabilities. Moreover, these two modalities are highly complementary in terms of information: visible-light images provide rich texture and structural details, while infrared images offer strong robustness under low-light and complex background conditions. Therefore, the VIF has become the mainstream approach in current research on multimodal small-object detection for UAVs.
In object detection tasks, features at different levels play complementary roles in determining the final detection performance. Low-level features typically contain rich local texture, edge, and spatial location information, while high-level features encode more global contextual and semantic information. The ability to effectively integrate these features at appropriate stages directly impacts the perception and detection capabilities of detector. Currently, there are three fusion paradigms: early fusion, intermediate fusion, and late fusion.
Intermediate fusion generally refers to the process of extracting features separately from the original data of different modalities and integrating them by leveraging the complementary advantages of distinct feature maps. This approach aims to achieve a balance between fine-grained local details and broader semantic context, thereby enhancing the overall representation capability of the model. Wang et al. [42] proposed a cross-scale dynamic convolution-driven YOLO fusion network, which introduces a novel cross-scale dynamic convolution fusion module to adaptively extract and fuse bimodal features concerning on data distribution. Shen et al. [43] proposed a novel feature fusion framework of dual cross-attention transformers, which utilizes an iterative interaction mechanism to share parameters among block-wise multimodal transformers, reducing model complexity and computational cost. Zeng et al. [44] proposed MMI-Det, a multi-modal fusion method for visible and infrared object detection, which introduces the fusion focus module and the information guided module, thereby enhancing the ability to perceive and combine visible-infrared modal information. Intermediate fusion effectively combines complementary features of different modalities during the later stage of feature extraction, effectively utilizing local and global information. Intermediate fusion integrates multimodal information at the feature level, effectively capturing complementary representations and balancing local detail with global semantics. However, since features are extracted independently before integration, this strategy often leads to the loss of modality-specific information and incomplete preservation of intrinsic feature relationships.
Late fusion improves detection accuracy by integrating the predictions of separate modality networks. This preserves the distinctive features of each modality and reduces cross-modal interference. Tang et al. [45] proposed a novel decision fusion module (DFM) for making an intermediate decision based on the features in the current layer and then fuse its results with the original features before passing them to the next layers to avoid information loss during feature abstraction and fully utilize lower layer features.
In contrast, early fusion combines visible and infrared images at the input level, preserving complementary texture, color, and thermal information. This approach enables more complete feature extraction and reduces information loss, while maintaining a simpler and more efficient network structure. Zhang et al. [46] proposed SuperYOLO, which utilizes a symmetric compact MF to extract supplementary information from various data and the assisted super resolution (SR) learning, showing a favorable accuracy–speed tradeoff. Chen et al. [47] proposed an MF framework—Fused Modality Enhancement Network—which consists of two fusion modules: the Common Feature Enhancement Module and the Difference Feature Enhancement Module. This method achieves accurate detection of low-light objects at night while reducing the computational complexity of the model.
Considering the above factors, this study adopts an early fusion strategy. On resource-constrained UAV platforms, early fusion offers a lightweight and efficient solution by requiring only a single feature extraction pathway, thus reducing computational and memory overhead. Meanwhile, by integrating visible and infrared data at the input level, this approach fully exploits their complementary characteristics.

3. Methodology

3.1. Overview of Our Method

The comprehensive framework of our proposed TAF-YOLO network, designed for multimodal small-object detection in UAV aerial imagery, is illustrated in Figure 1.
First, we explore different MF methods and propose TAFNet. By adaptively fusing visible and infrared modality images at both the pixel and channel levels before the backbone network, we achieved full integration of complementary features while preserving fine-grained spatial details that are crucial for small objects. Secondly, the proposed LASK module replaces the standard convolutional layers at the P3 and P4 stages in the YOLOv12 backbone to alleviate feature degradation during downsampling. By employing multi-scale depthwise convolutions and adaptive spatial attention, it selectively emphasizes informative regions and integrates an anti-aliasing filter to preserve edge and texture information, thereby enhancing the representation of small objects. Finally, the feature path simplification strategy retains a pure top-down feature fusion pathway. On this basis, the DSAB introduces a lightweight semantic guidance strategy that injects high-level semantics into low-level features and performs multi-scale context aggregation. The top-down feature fusion pathway combined with the lightweight semantic guidance strategy improves the semantic completeness of small object features, thereby boosting detection accuracy and maintaining computational efficiency.

3.2. Two-Branch Adaptive Fusion Network

The diversity of multimodal data can effectively enhance informational complementarity, thereby significantly improving the accuracy and perceptual capability of object detection models. In UAV object detection tasks, the method of feature fusion among different modalities greatly influences detection performance. Intermediate fusion during the feature extraction stage may cause information loss or distortion. This is especially detrimental for small-object detection in UAV imagery, because even the slightest detail loss can significantly reduce recognition accuracy. Similarly, although late fusion can integrate multimodal information at the output level, it only operates at the final stage, failing to fully exploit the complementary nature of low-level features across modalities, while also increasing computational complexity and processing time.
However, directly concatenating or summing these features often causes redundant information propagation and suboptimal modality weighting, as different regions may rely on different modalities. To address this issue, we propose a Two-branch Adaptive Fusion Network (TAFNet), which dynamically adjusts the contribution of each modality in a pixel-level and channel-level adaptive mechanism. The TAFNet enables the network to emphasize the most informative modality at each spatial location while suppressing redundant features, thereby achieving high-quality fusion with minimal computational cost. Before fusion, we perform camera calibration to remove lens distortion and then use a homography transformation to align the RGB image to the thermal image. Finally, the aligned RGB image is cropped to match the resolution of the infrared image, ensuring that both modalities share a consistent spatial scale. As shown in Figure 2, the input modalities of RGB and infrared are denoted as I R G B C v × H × W , I I R C i × H × W , respectively. To align both modalities into a common feature space, each branch is processed by a 3 × 3 convolution, batch normalization, and SiLU activation:
M R G B = ϕ ( B ( Conv 1 ( I R G B ) ) ) ,   M I R = ϕ ( B ( Conv 2 ( I I R ) ) )
where Conv 1 and Conv 2 represent the 3 × 3 convolution, B represents batch normalization and ϕ ( ) denotes the SiLU function.
After alignment, both features share the same intermediate dimensionality d , ensuring the two features have a consistent expression space before subsequent fusion. Then, the two aligned features are concatenated along the channel dimension and passed through a 1 × 1 convolution to generate gating logits:
A = Conv 3 ( Concat ( M R G B , M I R ) ) 2 × H × W
Here, A = [ A R G B , A I R ] represents the unnormalized modality scores at each spatial location, Conv 3 denotes the 1 × 1 convolution operations, and Concat represents the concatenation operation.
A modality-wise softmax is then applied along the modality dimension (channel axis) to obtain normalized pixel-wise weight maps W R G B , W I R 1 × H × W :
W R G B = e A R G B e A R G B + e A I R ,   W I R = e A I R e A R G B + e A I R
which ensures W R G B + W I R = 1 for every pixel. The softmax operation compares modal scores at each spatial location and converts them into a probability-like allocation of attention between modalities, thus realizing pixel level mode selection. The fused feature is obtained as follows:
F g a t e = W R G B F R G B + W I R F I R
where denotes element-wise multiplication with broadcasting. This soft gating strategy allows the network to adaptively select the most reliable modality based on spatial context, achieving dynamic weighted fusion.
To further enhance inter-channel dependencies and suppress noisy responses, we introduce a lightweight Squeeze-and-Excitation (SE) calibration after fusion. The final fused feature refined as follows:
F = ϕ ( B ( Conv 4 ( F g a t e ) ) )
O f u s i o n = Conv 5 F ( σ ( W 2 ReLU ( W 1 G A P ( F ) ) ) )
where represents the element-wise multiplication between the original feature map and the excitation weights, G A P ( ) represents the global average pooling operation, σ ( ) represents the Sigmoid function, W 1 and W 2 denote two fully connected layers forming a bottleneck structure, Conv 4 and Conv 5 represent the 3 × 3 and 1 × 1 convolution operations, respectively.
The unique advantage of the TAFNet lies in its ability to combine the spatial adaptivity of attention mechanisms with the efficiency of channel recalibration. Traditional fusion approaches often fail to adapt to local modality variations, leading to loss of fine-grained details crucial for small-object detection. In contrast, TAFNet dynamically generates pixel-wise modality weights through spatial Softmax gating, achieving adaptive feature fusion without redundant computation. Furthermore, the embedded SE structure enhances inter-channel dependency modeling, allowing subtle features from different modalities to be selectively amplified.

3.3. Large Adaptive Selective Kernel Module

As shown in Figure 3, the proposed backbone network is designed to extract both fine-grained texture features and high-level semantic information from UAV aerial images fused by TAFNet. It consists of three main types of modules: the C3k2 block, the LASK (Large Adaptive Selective Kernel) module, and the A2C2f (Area-Attention Enhanced Cross-Feature module) structure. The C3k2 blocks in the shallow layers are responsible for capturing detailed spatial textures and low-level features, which are crucial for detecting small objects with weak visual cues.
To address the feature degradation caused by conventional downsampling, we replace the standard convolutional layers at the P3 and P4 stages with the proposed LASK module. This module dynamically adjusts its receptive field through multi-scale depth-wise convolutions and an adaptive spatial attention mechanism, enabling the network to selectively focus on informative regions while suppressing background noise. Moreover, the LASK module integrates an anti-aliasing filter that preserves structural details and reduces information loss during stride = 2 downsampling, which is particularly beneficial for small and densely distributed objects in UAV aerial images. In deeper stages, the A2C2f modules employ area attention and multi-branch feature fusion to enhance global contextual perception and semantic consistency, helping the model distinguish small objects from cluttered backgrounds.
The proposed LASK is illustrated in Figure 4, aiming to introduce a multi-scale selective convolution mechanism coupled with an anti-aliasing downsampling strategy. Unlike standard convolutional blocks, which use a single receptive field and straightforward stride = 2 downsampling, LASK performs multi-scale feature extraction, adaptive spatial weighting, and low-pass filtering before subsampling.
To capture both local and global spatial context, LASK employs multiple depth-wise convolutional branches with different receptive fields. Specifically, three depth-wise convolutions are applied to the same input feature map X C × H × W :
U 1 = F DW 3 × 3 ( X ) ,   U 2 = F DW 5 × 5 ( X ) ,   U 3 = F DW 7 × 7 , d = 3 ( X )
where F DW k × k ( ) represents a depth-wise convolution with kernel size k and dilation rate d .
The three branches, respectively, emphasize local texture ( U 1 ), mid-range context ( U 2 ), and long-range semantic structure ( U 3 ). And U 1 , U 2 , U 3 C × H × W , each maintaining the same number of channels C as the input. Because these branches are depth-wise and channel-aligned, each channel in U i corresponds to the same semantic component across different receptive fields. In this setting, element-wise summation is the most natural and semantically consistent fusion strategy. The multi-scale feature maps are first aggregated through element-wise summation to form U ˜ = U 1 + U 2 + U 3 . Unlike channel concatenation, which would artificially expand the dimensionality and require additional projection to realign the channels, element-wise addition preserves a one-to-one correspondence between feature channels across scales. This design ensures that each channel of U i integrates multi-scale spatial responses that describe the same semantic entity under different receptive contexts. Then, channel-wise average pooling and max pooling are applied to capture spatial descriptors that reflect different types of activation statistics. These two pooled feature maps are concatenated and passed through a pair of large 7 × 7 convolutional layers to form a spatial attention tensor:
S = σ F 7 × 7 ( Concat ( P avg ( U ˜ ) , P max ( U ˜ ) ) )
where σ ( ) represents the Sigmoid function; F 7 × 7 ( ) denotes a two-layer transformation composed of a 7 × 7 convolution, a SiLU nonlinearity, and another 7 × 7 convolution; and P avg ( ) and P max ( ) represent channel-wise average pooling and max pooling operation, respectively.
The resulting attention map S = [ S 1 ,   S 2 ,   S 3 ] 3 × H × W generates three spatial attention coefficients, each corresponding to one of the convolution branches. This mechanism adaptively assigns higher importance to the most informative regions at each receptive field scale. Each feature branch is then modulated by its corresponding attention mask, and all branches are fused to form the final aggregated feature:
F = i = 1 3 ( S i U i )
where denotes element-wise multiplication. This selective fusion process effectively allows the model to emphasize spatially salient features across multiple scales while suppressing redundant or noisy background activations. As a result, the fused feature map F carries richer semantic cues with enhanced discriminative power for object localization and recognition.
In conventional CNNs, stride = 2 convolutions can introduce aliasing artifacts that blur object boundaries and degrade high-frequency textures. To address this issue, LASK introduces an anti-aliasing filter L ( ) before downsampling. This filter is implemented as a depth-wise convolution using a fixed Gaussian-like kernel. The filtered feature F ^ is obtained as F ^ = L ( F ) = F K 5 × 5 . The low-pass filter smooths high-frequency components while retaining important edge information.
K 5 × 5 = 1 256 1 4 6 4 1 4 16 24 16 4 6 24 36 24 6 4 16 24 16 4 1 4 6 4 1
This anti-aliasing operation reduces spatial distortion during sub-sampling, leading to more stable feature representations in deeper layers. After anti-aliasing, a point-wise convolution ( 1 × 1 ) is applied to mix channel information and restore the desired output dimensionality. Subsequently, Batch Normalization and the SiLU activation function are employed to stabilize training and introduce nonlinearity. The entire computation pipeline of LASK can be summarized as
Y = ϕ ( B ( F PW 1 × 1 ( L ( i = 1 3 S i F DW k i ( X ) ) ) ) )
where ϕ ( ) denotes the SiLU function, B represents batch normalization, and F PW 1 × 1 ( ) represents 1 × 1 point-wise convolution.
The output Y maintains the same spatial resolution and channel configuration as the standard convolutional layer, allowing LASK to seamless integration into existing YOLO architectures. In this work, the proposed LASK module is selectively integrated into the P3 and P4 stages of the YOLOv12 backbone. Because P5 primarily captures high-level semantic information with low spatial resolution, it contributes little to fine-grained localization, whereas P1 and P2 mainly contain noisy low-level textures with insufficient representational depth; therefore, applying LASK in these stages yields limited improvement for small-object feature extraction and is computationally inefficient.

3.4. Dual-Stream Attention Bridge

To enhance small-object detection on UAV platforms with limited computational budgets, the neck design in this work follows a progressive strategy combining feature path simplification strategy and the proposed Dual-Stream Attention Bridge (DSAB). First, as shown in Figure 1, the feature path simplification strategy retains a pure top-down feature fusion pathway, replacing the bidirectional aggregation in PANet to prevent the over-propagation and dilution of fine-grained details during feature transmission while significantly reducing computational overhead. Then, the proposed DSAB injects high-level semantic guidance from P5 into the low-level feature map P3. Through semantic-guided foreground and background routing and lightweight multi-scale context aggregation, the DSAB selectively enhances informative regions and suppresses irrelevant background noise, enabling the network to focus on fine structural boundaries and small-object regions. In this way, the feature path simplification strategy guarantees efficient and undistorted information flow, while DSAB complements it by reintroducing rich semantics in a targeted and controllable manner. Together, these two components form a balanced neck architecture that maintains computational efficiency and substantially improves detection accuracy for small objects.
As shown in Figure 5, given the high-level semantic feature map P 5 and the low-level feature map P 3 , our DSAB aims to generate an enhanced representation E 3 . We first align the high-level feature and generate a semantic attention map via sigmoid activation. Then the attention map is then used to derive the foreground stream A F and background stream A B :
A F = σ ( U ( Conv ( P 5 ) ) ) A B = E A F
where Conv ( ) represents the convolution operation, U ( ) denotes the upsample operation, and σ is the sigmoid function producing a spatial attention prior.
Low-level features P 3 are modulated into two semantic streams. Then each stream is processed through a Lite-ASPP module—comprising three depthwise separable convolutions with different dilation rates—to extract multi-scale contextual information, which is crucial for small-object recognition:
E 3 = Φ ( [ i = 1 3 D i ( P 3 A F ) , i = 1 3 D i ( P 3 A B ) ] )
where denotes element-wise multiplication; i = 1 3 D i ( ) represents concatenation the all results of { D i ( ) } i = 1 3 operation. { D i ( ) } i = 1 3 denotes a depthwise separable convolution with dilation rates of 1, 3, and 5. They all have 3 × 3 kernels with 128 channels followed by a SiLU layer. The operator [   ,   ] denotes channel-wise concatenation. Φ ( ) denotes the final fusion operator, where the concatenation features are passed through a 3 × 3 convolution followed by a SiLU activation.

4. Experiment

4.1. Settings

  • Hardware/Software Environment. The following experiments were performed on a computer with NVIDIA RTX 3090 GPU, 24 GB of memory, and Python code running on Ubuntu 24.04. With Visual Studio Code 1.103.2, CUDA 11.6, PyTorch 1.12, Python 3.8 and other commonly used deep learning and image processing libraries.
  • Dataset. The VEDAI [48] dataset contains a total of 1246 pairs of RGB and IR images, covering nine object categories such as cars, trucks, vans, tractors, pick-ups, and other road vehicles. The images were captured by UAV and aircraft platforms over semi-urban and rural areas, featuring complex backgrounds such as roads, buildings, vegetation, and shadowed regions. Each image has a resolution of 1024 × 1024 pixels, and the objects of interest are generally small in scale, often occupying less than 1% of the total image area. These characteristics make the dataset particularly challenging for small-object detection, as the vehicles exhibit significant variations in size, shape, and orientation, and are often affected by background clutter, illumination variations, and low inter-class contrast.
  • Implementation Details. We trained the network using the stochastic gradient descent (SGD) optimizer with the following hyperparameters. The initial learning rate is set to 0.01, the weight decay to 0.0005, and the momentum to 0.937. The model is trained for 300 iterations with a batch size of 8, and the input image size is set to 1024.
  • Evaluation Metrics. In this work, we use Precision, Recall, and mean Average Precision (mAP), GFLOPs and Parameters as evaluation metrics to assess the performance of the detector as defined below:
Precision = T P T P + F P
Recall = T P T P + F N
AP c = 0 1 P ( r ) d r
mAP = 1 N c = 1 N AP c
GFLOPs = l L MACs l 10 9
where T P , F P , and F N denote true positive, false positive, and false negative, respectively. AP c represents average precision for class c . P ( r ) is the precision at the recall r and N is the number of classes. MACs l represents the Multiply–Accumulate Operations for layer l .

4.2. Ablation Study

To assess the performance of the proposed method, we conducted ablation experiments on the VEDAI dataset, comparing multiple models across various metrics and marking the best results in bold.

4.2.1. Comparisons of Different Fusion Stages and Approaches

To evaluate the effectiveness of the proposed fusion method, we compared feature concatenation (Concat) with our TAFNet across six different detectors. Additionally, the various intermediate fusion strategies and the late fusion strategy were compared, with the network architecture illustrated in Figure 6. The experimental results of different fusion methods are presented in Table 1.
Model size and inference performance were assessed using parameter counts (Params) and GFLOPs. Experimental results indicate that, compared with the early fusion method based on simple feature concatenation, our proposed approach achieves a noticeable improvement in detection accuracy with only a marginal increase in the number of parameters and GFLOPs. By evaluating multiple YOLO-based baseline models, we observe that the proposed TAFNet on the vast majority of baseline models consistently outperforms the simple concatenation-based early fusion strategy in both mAP50 and mAP50:95 metrics, achieving the best detection performance. Specifically, Ours + TAFNet improves precision, recall, mAP50, and mAP50:95 by 4.5%, 4.2%, 3.1%, and 2.1%, respectively, compared with Our + Concat, indicating a significant performance improvement.
Moreover, our experimental results reveal a clear performance gap between the two early fusion strategies and the intermediate and late fusion strategies in terms of both detection accuracy and computational efficiency. These findings indicate that early fusion strategies are more effective in preserving low-level features, such as edges and textures, which are crucial for small object detection in UAV aerial imagery. In contrast, intermediate fusion tends to introduce redundant information and feature loss, potentially obscuring the representation of small objects. Late fusion fails to sufficiently exploit the complementary information between modalities while incurring additional computational overhead. Overall, these results demonstrate that the proposed TAFNet early fusion strategy achieves an excellent trade-off between detection accuracy and computational cost, making it particularly suitable for multimodal UAV-based small object detection tasks.

4.2.2. Ablation Study of Each Component

Table 2 presents a comprehensive ablation study evaluating the individual and combined contributions of the proposed LASK, DSAB, and the effect of removing the PANet on the TAF-YOLO model’s performance. The metrics considered include model Params, GFLOPs, Precision (Pre), Recall (Rec), mAP@0.5 (mAP50), and mAP@0.5:0.95 (mAP50:95).
As shown in Table 2, adding LASK to the baseline leads to a notable improvement in Precision (from 57.5% to 61.8%). While mAP50 and mAP50:95 show a slight increase (0.4% and 0.2%, respectively), surprisingly, the GFLOPs slightly reduce from 6.3 to 6.1, despite a minor increase in parameters (from 2.810 M to 3.024 M). This suggests that LASK enhances the model’s ability to dynamically select informative receptive fields while maintaining edge and texture integrity. Removing PANet from the baseline significantly reduces model complexity, decreasing parameters from 2.810 M to 1.831 M (a 34.8% reduction) and GFLOPs from 6.3 to 4.9 (a 22.3% reduction). Crucially, removing PANet also leads to a slight improvement in mAP50 (from 62.4% to 63.3%) and mAP50:95 (from 41.2% to 41.9%). This indicates that the original PANet architecture is inefficient and introduces redundant features that are not suitable for small-object detection tasks. Removing it can improve efficiency and slightly enhance performance. When PANet is removed, adding LASK further improves Precision (from 58.8% to 62.1%) and also yields a marginal increase in mAP50 (0.4%) and mAP50:95 (0.5%). The GFLOPs again show a slight reduction (from 4.9 to 4.7), reinforcing LASK’s efficiency benefits. Adding DSAB while removing PANet significantly boosts mAP50 (from 62.4% to 66.9%) and mAP50:95 (from 41.2% to 43.6%). This indicates that DSAB can effectively utilize high-level semantics to guide low-level features, enabling the network to better perceive fine structural boundaries and small object regions. However, this requires acceptable computational increase: parameters increase to 4.305 M and GFLOPs to 8.6. The best performing configuration, which incorporates both LASK and DSAB while removing PANet, achieves the highest scores across all accuracy metrics: Precision (64.8%), Recall (62.7%), mAP50 (67.2%), and mAP50:95 (44.3%).

4.3. Comparison of Experiment with Previous Methods

We compared the performance of state-of-the-art methods, including YOLOv5 [30], YOLOv6 [49], YOLOv8 [31], YOLOv10 [32], YOLOv11 [33], YOLOv12 [34], CFT [18], DEYOLO [50], and our proposed TAF-YOLO method on the VEDAI dataset across various scenarios. Figure 7 presents some visual comparison, where the red circles indicate missed detection, the yellow circles represent false detection, and the blue circles highlight false alarm. It can be observed that under scenarios with complex backgrounds or occlusions, other methods tend to produce a considerable number of missed and false detections. In UAV aerial imagery, vehicles usually occupy small regions, and their appearances across different categories often exhibit high similarity. Although classes such as car, pickup, camping, and truck are relatively abundant in the dataset and achieve better detection precision overall, large variations in object scale and inter-class resemblance still make accurate classification challenging. Most existing algorithms can accurately locate vehicles, but they have difficulty distinguishing fine-grained differences, which leads to frequent misclassifications, especially between cars and pickups.
The visualization results clearly demonstrate the superior performance of our proposed TAF-YOLO. It effectively addresses the critical challenges of small object detection, significant scale variation, and high inter-class similarity in UAV aerial imagery. Other methods often suffer from missed detections, frequent misclassifications (especially between cars and pickups) and false alarms in cluttered scenes. In contrast, TAF-YOLO provides more accurate and robust detection results. This indicates that our method’s architecture and fusion strategy are particularly well-suited for extracting discriminative features necessary for precise localization and fine-grained classification in these demanding scenarios.
Table 3 summarizes the performance of the above methods. The MF model achieved significantly better AP values than the single modal (RGB or IR) model in most categories, indicating that the proposed TAFNet can effectively utilize early feature fusion strategies. However, it is worth noting that for the “car” category, the detection performance of the fusion model is actually lower than using only RGB modes. This phenomenon may be related to the information differences between different modalities and the performance of specific categories (such as cars) under different lighting conditions. In some cases, infrared information may not provide sufficient details or have significant differences from visible light images, thereby affecting the detection accuracy after fusion. However, overall, the improvement in mAP50 and mAP50:95 still indicates that TAFNet exhibits superior performance in most categories. Meanwhile, the increase in parameter count and computational complexity is very limited, indicating that the method has achieved an ideal balance between accuracy and efficiency.
The visualization results of several failed detections presented in Figure 8 reveal the limitations of TAF-YOLO in highly complex scenarios. Missed detections typically occur when objects are extremely small or blend seamlessly with the background. A representative example is the “tractors” in Figure 8a, located at the top-left corner of the image along a dirt path, which remains undetected due to its minute size and similar texture to the surrounding ground. And as shown in Figure 8d, a “car” is erroneously identified as a “pickup,” clearly demonstrating the model’s difficulty in fine-grained classification, particularly between visually similar vehicles like cars and pickups, where subtle structural differences are flattened and obscured from an overhead view. In summary, the failure cases indicate the extreme complexity and inherent challenges of object detection in UAV aerial imagery, encompassing the diminutive size of objects, the intrinsic high visual similarity among different vehicle classes, and the strong interference from diverse and complex backgrounds (e.g., farmland, urban structures, shadows, road textures).

4.4. Generalization Experiment

To validate the generalization capability of the proposed TAF-YOLO, we conduct a comparative evaluation of TAF-YOLO and other detectors on the public DroneVehicle dataset [51]. DroneVehicle captures complex airborne imaging scenarios, where vehicles exhibit significant variations in size, orientation, and background clutter due to changes in drone altitude and camera tilt. The dataset provides paired RGB and IR images, enabling the exploration of MF detector’s robustness under diverse environmental conditions. The dataset encompasses diverse illumination situations (daytime, dusk, and nighttime) and viewing angles, simulating real-world drone surveillance tasks. Such complementary characteristics make DroneVehicle well-suited for evaluating MF detecor and generalization ability across domains. For our experiments, we follow the common split setting and select 2800 image pairs for training, 800 for validation, and 400 for testing. The annotations cover five vehicle categories—car, truck, bus, van, and freight-car—each appearing at varying scales and densities. The presence of occlusions, dense traffic, and multi-scale objects poses additional challenges for detection algorithms.
Table 4 presents in detail the comparison results of different algorithms on the DroneVehicle dataset, while Figure 9 highlights the visualization experimental results of CFT, DEYOLO, and TAF-YOLO in night scenes, ground texture interference scenes, and low-light scenes. Experimental results on the public dataset DroneVehicle show that our proposed TAF-YOLO achieved mAP50 and mAP50:95 values of 65.8% and 46.7%, respectively. This validates the efficacy and superiority of the proposed approach for detecting small objects in UAV aerial images, further demonstrating its excellent generalization capability. The visualization results demonstrate that the proposed TAF-YOLO excels in detecting objects under challenging conditions, such as night scenes, ground texture interference, and low-light environments. Compared to the state-of-the-art methods, CFT and DEYOLO, TAF-YOLO significantly reduces false detections and false alarms while also improving both positional accuracy and confidence levels. In addition, TAF-YOLO achieves a favorable balance between accuracy and computational efficiency. With only 4.518 M parameters and 8.4 GFLOPs, it delivers performance comparable to larger models such as CFT and DEYOLO, demonstrating the proposed method’s potential for deployment in real-time UAV detection tasks with limited computational resources.
To further verify the robustness and statistical reliability of the proposed TAF-YOLO, we conducted repeated experiments on the VEDAI dataset using five different random seeds, while keeping all hyperparameters and training configurations identical to those described in Section 4.1. Table 5 summarizes the corresponding Precision, Recall, mAP50, and mAP50:95 results.
As shown in Table 5, the performance variations across different seeds are very small, with fluctuations of only ±0.3% in Precision, ±0.7% in Recall, ±0.7% in mAP50, and ±0.5% in mAP50:95. The overall mean values (Pre 68.64%, Rec 64.12%, mAP50 67.34%, mAP50:95 44.32%) remain highly consistent with the single-run performance reported in Table 1. These results demonstrate that the improvements achieved by the proposed TAFNet and the complete TAF-YOLO architecture are stable and reproducible across different random initializations, confirming that the observed accuracy gains are not the result of a favorable random seed but reflect genuine architectural advantages.

5. Discussion

The proposed TAF-YOLO effectively leverages TAFNet for pixel-level and channel-level adaptive fusion of visible and infrared inputs, enabling comprehensive integration of complementary information while minimizing redundancy. It demonstrates strong detection performance in nighttime, background-interference, and low-light environments, reducing both missed detection and false alarm rates. However, its performance in extreme scenarios still requires further improvement. As shown in Figure 10a,b, in extreme modality imbalance scenarios, the visible light modality fails to provide effective information, while the infrared modality exhibits strong background noise, leading to false alarms. In the multimodal information scarcity scenarios shown in Figure 10c,d, both modalities fail to provide useful information, resulting in missed detections.
Although TAF-YOLO achieves strong overall performance, the results also reveal that fine-grained classification remains challenging for categories with highly similar visual appearances, such as car and pickup. This difficulty is inherent to UAV top-down imaging, where objects occupy only a few pixels and distinctive structural cues are often compressed or obscured. In addition, modality characteristics can further exacerbate this issue: visible images may lose texture details under low illumination, while infrared images sometimes lack the fine contours required for subtle inter-class discrimination. These factors jointly limit the separability of visually similar classes.
And the LASK module dynamically adjusts receptive fields through multi-scale depthwise convolutions and spatial attention, preserving fine edge and texture details during downsampling. By replacing the traditional convolution operations in P3 and P4 of the original baseline model with this approach, we achieve promising results. However, we have not yet established that replacing the baseline convolutional blocks with the LASK module will consistently outperform the original convolutions across different datasets and a rchitectures. Additional experiments are needed to verify LASK’s robustness and generalization as a direct replacement for standard convolutional layers.
The Ground Sampling Distance (GSD) is crucial for determining the spatial resolution of an image, particularly in the context of small object detection. In our study, the UAV should fly at an altitude between 60 and 100 m, ensuring an optimal GSD for detecting small objects in UAV aerial imagery. At this altitude, the GSD is approximately 3–5 cm per pixel, which strikes a balance between spatial resolution and the extent of the area captured. At this GSD level, a small target of around 10 cm × 10 cm still spans 2–3 pixels, meeting the minimum requirement for detectors such as YOLO or SSD to extract effective features. Moreover, at 60–100 m, the UAV can fly stably at speeds ≤5 m/s with exposure times ≤1/2000 s, keeping motion blur below 0.5 pixel, thus preserving image sharpness. This altitude minimizes the risk of motion blur and interference from environmental factors such as weather, while maintaining computational efficiency on resource-constrained UAV platforms.
To ensure that the conclusions drawn from the VEDAI dataset are not limited by its sample size, we further performed a dataset size adequacy discussion from multiple perspectives. First, although VEDAI contains 1246 pairs of visible–infrared images, the proposed TAF-YOLO was additionally evaluated on the DroneVehicle dataset, which provides more than twice the number of training samples and features substantially different scene distributions, object scales, and imaging conditions. As shown in Table 1 and Table 4, TAFNet consistently improves detection accuracy across twelve YOLO baselines on VEDAI and maintains the same performance advantages on DroneVehicle. This cross-dataset consistency suggests that the improvements introduced by our fusion strategy are not sensitive to dataset size or distribution and are therefore unlikely to be artifacts of sample insufficiency or dataset-specific bias. Second, we examined the training dynamics of TAF-YOLO. The training loss and validation mAP curves exhibit smooth and stable convergence without late-epoch oscillations, indicating that the dataset provides enough samples to support reliable optimization.
While classical image processing approaches are computationally lightweight, they are inherently limited to simple or binary targets and cannot provide robust multimodal recognition in complex multimodal UAV imagery such as VEDAI. For this reason, and because the objective of this study is to investigate multimodal fusion within the YOLO framework, YOLO-based baselines form the most appropriate and fair point of comparison. At the same time, the computational cost of neural methods is now compatible with modern UAV hardware. The proposed TAF-YOLO requires 8.4 GFLOPs per inference. Assuming a 50-FPS camera, full-frame processing would require approximately 420 GFLOPs/s. Modern lightweight embedded processors, such as the NVIDIA Jetson Nano (472 GFLOPs FP16) and Jetson TX2 (1.3 TFLOPs FP16), can comfortably satisfy these demands. Based on the required 8.4 GFLOPs per inference, TAF-YOLO achieves approximately 56 FPS on the Jetson Nano and approximately 155 FPS on the Jetson TX2. This confirms that the proposed method achieves a strong balance between accuracy and computational efficiency, enabling real-time operation on resource-constrained UAV platforms.

6. Conclusions

In this paper, we presented TAF-YOLO, an efficient multimodal small-object detection framework tailored for UAV aerial imagery. By performing early adaptive fusion of visible and infrared modalities through the proposed TAFNet, the framework effectively integrates complementary texture and thermal information while suppressing redundancy. The LASK module dynamically enhances receptive field adaptability and preserves fine-grained edge and texture features, significantly improving the perception of small objects. Furthermore, the incorporation of DSAB and a simplified feature fusion path strengthens semantic interaction and localization precision while maintaining lightweight computational costs. Extensive experiments on the VEDAI dataset demonstrate that TAF-YOLO achieves superior performance in both detection accuracy and efficiency, confirming its suitability for real-time, all-weather UAV perception. However, in extreme detection scenarios characterized by strong background interference and extremely small inter-class differences, the algorithm still requires enhanced feature extraction and background suppression capabilities. Future research will focus on further enhancing the performance of TAF-YOLO in extreme scenarios with significant modal imbalance to improve generalization across diverse UAV platforms and complex environments, extending the applicability of TAF-YOLO to broader multimodal UAV remote sensing tasks.

Author Contributions

Conceptualization, R.L. and Y.Y.; methodology, Z.Z. (Zhanhong Zhuo); software, S.W.; validation, R.L., Z.Z. (Zhanhong Zhuo) and S.W.; investigation, R.L.; resources, Z.Z. (Zhanhong Zhuo); writing—original draft preparation, Z.Z. (Zhanhong Zhuo); writing—review and editing, Z.Z. (Zhanhong Zhuo) and R.L.; visualization, J.Z.; supervision, R.L.; project administration, Z.Z. (Zhi Zheng); funding acquisition, X.Y. and R.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported in part by the National Natural Science Foundation of China under grant number 62276274 and 62573423, in part by the Key R&D Program of Shaanxi Province 2024CY2-GJHX-42, and in part by the Shannxi Sanqin Elite Special Support Program 2024-SQ-001.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
UAVUnmanned Aerial Vehicle
CNNsConvolutional Neural Networks
IRInfrared
MFMultimodal Fusion
VIFVisible and Infrared Image Fusion
NMSNon-Maximum Suppression
TAFNetTwo-branch Adaptive Fusion Network
LASKLarge Adaptive Selective Kernel
DSABDual-Stream Attention Bridge
GSDGround Sampling Distance
A2C2fArea-Attention Enhanced Cross-Feature module

References

  1. Osco, L.P.; Marcato, J., Jr.; Marques Ramos, A.P.; de Castro Jorge, L.A.; Fatholahi, S.N.; de Andrade Silva, J.; Matsubara, E.T.; Pistori, H.; Gonçalves, W.N.; Li, J. A review on deep learning in UAV remote sensing. Int. J. Appl. Earth Obs. Geoinf. 2021, 102, 102456. [Google Scholar] [CrossRef]
  2. Aasen, H.; Honkavaara, E.; Lucieer, A.; Zarco-Tejada, P.J. Quantitative Remote Sensing at Ultra-High Resolution with UAV Spectroscopy: A Review of Sensor Technology, Measurement Procedures, and Data Correction Workflows. Remote Sens. 2018, 10, 1091. [Google Scholar] [CrossRef]
  3. Bo, W.; Liu, J.; Fan, X.; Tjahjadi, T.; Ye, Q.; Fu, L. BASNet: Burned Area Segmentation Network for Real-Time Detection of Damage Maps in Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5627913. [Google Scholar] [CrossRef]
  4. Yu, X.; Jiang, T.; Zhu, Y.; Li, L.; Fan, F.; Jin, X. FEL-YoloV8: A New Algorithm for Accurate Monitoring Soybean Seedling Emergence Rates and Growth Uniformity. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5634217. [Google Scholar] [CrossRef]
  5. Xiao, Y.; Wang, J.; Zhao, Z.; Jiang, B.; Li, C.; Tang, J. UAV Video Vehicle Detection: Benchmark and Baseline. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5609814. [Google Scholar] [CrossRef]
  6. Luo, M.; Zhao, R.; Zhang, S.; Chen, L.; Shao, F.; Meng, X. IM-CMDet: An Intramodal Enhancement and Cross-Modal Fusion Network for Small Object Detection in UAV Aerial Visible-Infrared Imagery. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5008316. [Google Scholar] [CrossRef]
  7. Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. In Proceedings of the Computer Vision—ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Springer: Cham, Switzerland, 2014; pp. 740–755. [Google Scholar]
  8. Ying, X.; Xiao, C.; An, W.; Li, R.; He, X.; Li, B.; Cao, X.; Li, Z.; Wang, Y.; Hu, M.; et al. Visible-Thermal Tiny Object Detection: A Benchmark Dataset and Baselines. IEEE Trans. Pattern Anal. Mach. Intell. 2025, 47, 6088–6096. [Google Scholar] [CrossRef]
  9. Du, B.; Huang, Y.; Chen, J.; Huang, D. Adaptive Sparse Convolutional Networks with Global Context Enhancement for Faster Object Detection on Drone Images. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 13435–13444. [Google Scholar]
  10. Wang, K.; Fu, X.; Huang, Y.; Cao, C.; Shi, G.; Zha, Z.J. Generalized UAV Object Detection via Frequency Domain Disentanglement. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 1064–1073. [Google Scholar]
  11. Zou, Z.; Hu, X.; Zhong, P. Active Object Detection for UAV Remote Sensing via Behavior Cloning and Enhanced Q-Network with Shallow Features. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5628116. [Google Scholar] [CrossRef]
  12. Jiao, Z.; Wang, M.; Qiao, S.; Zhang, Y.; Huang, Z. Transformer-Based Object Detection in Low-Altitude Maritime UAV Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2025, 63, 4210413. [Google Scholar] [CrossRef]
  13. Zhang, X.; Demiris, Y. Visible and Infrared Image Fusion Using Deep Learning. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 10535–10554. [Google Scholar] [CrossRef] [PubMed]
  14. Xu, S.; Chen, X.; Li, H.; Liu, T.; Chen, Z.; Gao, H.; Zhang, Y. Airborne Small Target Detection Method Based on Multimodal and Adaptive Feature Fusion. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5637215. [Google Scholar] [CrossRef]
  15. Liu, J.; Fan, X.; Jiang, J.; Liu, R.; Luo, Z. Learning a Deep Multi-Scale Feature Ensemble and an Edge-Attention Guidance for Image Fusion. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 105–119. [Google Scholar] [CrossRef]
  16. Li, Q.; Zhang, C.; Hu, Q.; Fu, H.; Zhu, P. Confidence-Aware Fusion Using Dempster-Shafer Theory for Multispectral Pedestrian Detection. IEEE Trans. Multimed. 2023, 25, 3420–3431. [Google Scholar] [CrossRef]
  17. Park, S.; Vien, A.G.; Lee, C. Cross-Modal Transformers for Infrared and Visible Image Fusion. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 770–785. [Google Scholar] [CrossRef]
  18. Fang, Q.; Han, D.; Wang, Z. Cross-modality fusion transformer for multispectral object detection. arXiv 2021, arXiv:2111.00273. [Google Scholar] [CrossRef]
  19. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  20. Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  21. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef]
  22. Cai, Z.; Vasconcelos, N. Cascade R-CNN: Delving into High Quality Object Detection. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6154–6162. [Google Scholar]
  23. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the Computer Vision—ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  24. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2999–3007. [Google Scholar]
  25. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-End Object Detection with Transformers. In Proceedings of the Computer Vision—ECCV 2020, Glasgow, UK, 23–28 August 2020; pp. 213–229. [Google Scholar]
  26. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable DETR: Deformable transformers for end-to-end object detection. arXiv 2020, arXiv:2010.04159. [Google Scholar]
  27. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer using Shifted Windows. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 9992–10002. [Google Scholar]
  28. Dosovitskiy, A. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  29. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  30. Tian, Z.; Shen, C.; Chen, H.; He, T. FCOS: A Simple and Strong Anchor-Free Object Detector. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 1922–1933. [Google Scholar] [CrossRef]
  31. Varghese, R.; S, M. YOLOv8: A Novel Object Detection Algorithm with Enhanced Performance and Robustness. In Proceedings of the 2024 International Conference on Advances in Data Engineering and Intelligent Computing Systems (ADICS), Chennai, India, 18–19 April 2024; pp. 1–6. [Google Scholar]
  32. Chen, H.; Chen, K.; Ding, G.; Han, J.; Lin, Z.; Liu, L.; Wang, A. Yolov10: Real-time end-to-end object detection. In Proceedings of the 2024 Advances in Neural Information Processing Systems (NeurIPS), Vancouver, BC, Canada, 10–15 December 2024; Volume 37, pp. 107984–108011. [Google Scholar]
  33. Khanam, R.; Hussain, M. Yolov11: An overview of the key architectural enhancements. arXiv 2024, arXiv:2410.17725. [Google Scholar] [CrossRef]
  34. Tian, Y.; Ye, Q.; Doermann, D. Yolov12: Attention-centric real-time object detectors. arXiv 2025, arXiv:2502.12524. [Google Scholar]
  35. Wu, S.; Lu, X.; Guo, C.; Guo, H. MV-YOLO: An Efficient Small Object Detection Framework Based on Mamba. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5632814. [Google Scholar] [CrossRef]
  36. Liu, Y.; Cheng, X.; Xu, N.; Wang, L.; Wang, X.; Zhong, X. MFAE-YOLO: Multifeature Attention-Enhanced Network for Remote Sensing Images Object Detection. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5631214. [Google Scholar] [CrossRef]
  37. Cao, J.; Bao, W.; Shang, H.; Yuan, M.; Cheng, Q. GCL-YOLO: A GhostConv-Based Lightweight YOLO Network for UAV Small Object Detection. Remote Sens. 2023, 15, 4932. [Google Scholar] [CrossRef]
  38. Tang, Y.; Xu, T.; Qin, H.; Li, J. IRSTD-YOLO: An Improved YOLO Framework for Infrared Small Target Detection. IEEE Geosci. Remote Sens. Lett. 2025, 22, 7001405. [Google Scholar] [CrossRef]
  39. Hao, X.; Luo, S.; Chen, M.; He, C.; Wang, T.; Wu, H. Infrared small target detection with super-resolution and YOLO. Opt. Laser Technol. 2024, 177, 111221. [Google Scholar] [CrossRef]
  40. Caesar, H.; Bankiti, V.; Lang, A.H.; Vora, S.; Liong, V.E.; Xu, Q.; Krishnan, A.; Pan, Y.; Baldan, G.; Beijbom, O. nuScenes: A Multimodal Dataset for Autonomous Driving. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 11618–11628. [Google Scholar]
  41. Hong, D.; Gao, L.; Yokoya, N.; Yao, J.; Chanussot, J.; Du, Q.; Zhang, B. More Diverse Means Better: Multimodal Deep Learning Meets Remote-Sensing Imagery Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 4340–4354. [Google Scholar] [CrossRef]
  42. Wang, Z.; Liao, X.; Yuan, J.; Yao, Y.; Li, Z. CDC-YOLOFusion: Leveraging Cross-Scale Dynamic Convolution Fusion for Visible-Infrared Object Detection. IEEE Trans. Intell. Veh. 2025, 10, 2080–2093. [Google Scholar] [CrossRef]
  43. Shen, J.; Chen, Y.; Liu, Y.; Zuo, X.; Fan, H.; Yang, W. ICAFusion: Iterative cross-attention guided feature fusion for multispectral object detection. Pattern Recognit. 2024, 145, 109913. [Google Scholar] [CrossRef]
  44. Zeng, Y.; Liang, T.; Jin, Y.; Li, Y. MMI-Det: Exploring Multi-Modal Integration for Visible and Infrared Object Detection. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 11198–11213. [Google Scholar] [CrossRef]
  45. Tang, K.; Ma, Y.; Miao, D.; Song, P.; Gu, Z.; Tian, Z.; Wang, W. Decision Fusion Networks for Image Classification. IEEE Trans. Neural Netw. Learn, Syst. 2025, 36, 3890–3903. [Google Scholar] [CrossRef] [PubMed]
  46. Zhang, J.; Lei, J.; Xie, W.; Fang, Z.; Li, Y.; Du, Q. SuperYOLO: Super Resolution Assisted Object Detection in Multimodal Remote Sensing Imagery. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5605415. [Google Scholar] [CrossRef]
  47. Chen, Y.; Da, F. YOLO-FMEN: Pixel-Level Image Fusion-Based Nighttime Object Detection Network. In Proceedings of the 2025 44th Chinese Control Conference (CCC), Chongqing, China, 28–30 July 2025; pp. 9179–9186. [Google Scholar]
  48. Razakarivony, S.; Jurie, F. Vehicle detection in aerial imagery: A small target detection benchmark. J. Vis. Commun. Image Represent. 2016, 34, 187–203. [Google Scholar] [CrossRef]
  49. Li, C.; Li, L.; Jiang, H.; Weng, K.; Geng, Y.; Li, L.; Ke, Z.; Li, Q.; Cheng, M.; Nie, W. YOLOv6: A single-stage object detection framework for industrial applications. arXiv 2022, arXiv:2209.02976. [Google Scholar] [CrossRef]
  50. Chen, Y.; Wang, B.; Guo, X.; Zhu, W.; He, J.; Liu, X.; Yuan, J. DEYOLO: Dual-Feature-Enhancement YOLO for Cross-Modality Object Detection. In Pattern Recognition; Springer: Cham, Switzerland, 2025; pp. 236–252. [Google Scholar]
  51. Sun, Y.; Cao, B.; Zhu, P.; Hu, Q. Drone-Based RGB-Infrared Cross-Modality Vehicle Detection Via Uncertainty-Aware Learning. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 6700–6713. [Google Scholar] [CrossRef]
Figure 1. Overall framework of the TAF-YOLO. The modules highlighted in red text or enclosed by red boxes denote the innovative components proposed in this work.
Figure 1. Overall framework of the TAF-YOLO. The modules highlighted in red text or enclosed by red boxes denote the innovative components proposed in this work.
Remotesensing 17 03936 g001
Figure 2. The Schematic of TAFNet.
Figure 2. The Schematic of TAFNet.
Remotesensing 17 03936 g002
Figure 3. Backbone structure of TAF-YOLO.
Figure 3. Backbone structure of TAF-YOLO.
Remotesensing 17 03936 g003
Figure 4. The illustration of the LASK module.
Figure 4. The illustration of the LASK module.
Remotesensing 17 03936 g004
Figure 5. The object concealment failure results.
Figure 5. The object concealment failure results.
Remotesensing 17 03936 g005
Figure 6. The network architecture of different fusion strategies. (a) Intermediate fusion strategies. (b) Late fusion strategy.
Figure 6. The network architecture of different fusion strategies. (a) Intermediate fusion strategies. (b) Late fusion strategy.
Remotesensing 17 03936 g006
Figure 7. The visual comparison of different inpainting methods. (af) represent different detection scenarios. For each subfigure, the columns from left to right correspond to RGB Ground Truth, IR Ground Truth, YOLOv5, YOLOv6, YOLOv8, YOLOv10, YOLOv11, YOLOv12, CFT, DEYOLO, and Ours. The red circles indicate missed detection, the yellow circles represent false detection, and the blue circles highlight false alarm.
Figure 7. The visual comparison of different inpainting methods. (af) represent different detection scenarios. For each subfigure, the columns from left to right correspond to RGB Ground Truth, IR Ground Truth, YOLOv5, YOLOv6, YOLOv8, YOLOv10, YOLOv11, YOLOv12, CFT, DEYOLO, and Ours. The red circles indicate missed detection, the yellow circles represent false detection, and the blue circles highlight false alarm.
Remotesensing 17 03936 g007
Figure 8. The failure cases of MF detection methods. (ad) represent different detection scenarios. The red circles indicate missed detection, the yellow circles represent false detection, and the blue circles highlight false alarm.
Figure 8. The failure cases of MF detection methods. (ad) represent different detection scenarios. The red circles indicate missed detection, the yellow circles represent false detection, and the blue circles highlight false alarm.
Remotesensing 17 03936 g008
Figure 9. Generalization experiments in different scenarios on DroneVehicle dataset. (a,b) are night scenes, (c) is a ground texture interference scene, and (d) is a low-light scene. The yellow circles represent false detection, and the blue circles highlight false alarm.
Figure 9. Generalization experiments in different scenarios on DroneVehicle dataset. (a,b) are night scenes, (c) is a ground texture interference scene, and (d) is a low-light scene. The yellow circles represent false detection, and the blue circles highlight false alarm.
Remotesensing 17 03936 g009
Figure 10. Multimodal fusion under extreme scenarios. (a,b) depict a modality imbalance scenario. (c,d) illustrate a multimodal information scarcity scenario.
Figure 10. Multimodal fusion under extreme scenarios. (a,b) depict a modality imbalance scenario. (c,d) illustrate a multimodal information scarcity scenario.
Remotesensing 17 03936 g010
Table 1. Comparison results of different fusion stages and approaches on the VEDIA datasets.
Table 1. Comparison results of different fusion stages and approaches on the VEDIA datasets.
MethodPreRecmAP50mAP50:95ParamsGFLOPs
Early
Fusion
YOLOv5n + Concat60.658.263.637.22.189 M6.0
YOLOv5n + TAFNet62.956.063.439.92.307 M6.4
YOLOv6n + Concat57.056.356.935.94.160 M11.7
YOLOv6n + TAFNet59.161.060.639.14.235 M11.9
YOLOv8n + Concat65.452.360.938.52.691 M7.0
YOLOv8n + TAFNet66.054.861.940.63.007 M7.4
YOLOv10n + Concat68.650.460.238.32.710 M8.3
YOLOv10n + TAFNet71.652.362.039.32.698 M8.7
YOLOv11n + Concat62.059.661.640.62.412 M6.0
YOLOv11n + TAFNet62.660.462.841.22.728 M6.4
YOLOv12n + Concat54.659.360.237.32.509 M5.9
YOLOv12n + TAFNet57.562.662.441.22.810 M6.3
Ours + Concat64.359.564.142.24.231 M8.0
Ours + TAFNet68.863.767.244.34.518 M8.4
Intermediate
Fusion
Fusion160.259.362.139.84.639 M39.0
Fusion262.659.662.740.74.654 M13.0
Fusion361.761.463.141.24.827 M13.8
Fusion467.360.863.641.55.297 M14.6
Late Fusion60.254.359.837.211.709 M16.8
Note: Bold values indicate the best results.
Table 2. The ablation experiment TAF-YOLO.
Table 2. The ablation experiment TAF-YOLO.
LASKDSABRemove PANetParamsGFLOPsPreRecmAP50mAP50:95
1 2.810 M6.357.562.662.441.2
2 3.024 M6.161.862.262.841.4
3 4.305 M8.662.762.566.943.6
4 1.831 M4.958.862.063.341.9
5 2.044 M4.762.160.863.742.4
64.518 M8.464.862.767.244.3
Note: Bold values indicate the best results and “√” indicates that the corresponding module is used.
Table 3. The quantitative comparison of different methods on VEDAI.
Table 3. The quantitative comparison of different methods on VEDAI.
MethodCarPickupCampingTruckOthersTractorBoatVanmAP50mAP50:95ParamsGFLOPs
YOLOv5nRGB88.369.460.659.234.649.765.152.960.037.42.189 M5.9
IR85.764.967.560.322.253.635.565.756.934.52.189 M5.9
Multi88.172.065.265.447.253.863.351.863.439.92.307 M6.4
YOLOv6nRGB87.172.060.159.244.255.235.155.358.536.14.160 M11.6
IR86.765.652.865.622.255.114.552.251.831.74.160 M11.6
Multi84.568.760.959.453.760.027.867.960.639.14.335 M11.9
YOLOv8nRGB87.876.155.054.738.662.151.545.358.935.52.691 M6.9
IR85.266.663.663.330.541.435.752.554.833.42.691 M6.9
Multi83.173.661.864.341.668.144.939.061.940.63.007 M7.4
YOLOv10nRGB84.969.961.359.917.649.448.950.555.333.92.710 M8.2
IR81.568.555.153.823.836.930.951.950.331.22.710 M8.2
Multi81.266.961.663.837.948.557.678.462.039.32.842 M8.7
YOLOv11nRGB88.871.775.957.331.459.440.749.359.335.42.412 M5.9
IR84.769.164.973.630.437.339.847.755.933.92.412 M5.9
Multi84.974.167.372.044.872.162.248.862.841.22.728 M6.6
YOLOv12nRGB87.071.663.561.142.856.946.742.159.035.22.509 M5.8
IR80.864.757.066.022.434.629.137.149.029.92.509 M5.8
Multi83.373.475.758.045.660.454.259.962.441.22.810 M6.3
CFTMulti85.673.265.265.644.364.254.260.864.542.147.06 M117.2
DEYOLOnMulti85.270.570.068.350.367.060.261.565.743.26.09 M-
TAF-YOLO
(Ours)
RGB88.570.463.462.143.257.252.245.662.240.44.231 M7.9
IR84.568.760.959.430.239.627.839.660.338.34.231 M7.9
Multi87.072.375.771.347.665.467.962.967.244.34.518 M8.4
Note: Bold values indicate the best results, and “-“ indicates that the results are not reported.
Table 4. The quantitative comparison of different methods on DroneVehicle.
Table 4. The quantitative comparison of different methods on DroneVehicle.
MethodCarsTrucksBusesVansFreight carsmAP50mAP50:95ParamsGFLOPs
YOLOv5nMulti91.846.384.345.344.662.141.42.307 M6.4
YOLOv8nMulti92.050.484.346.645.262.942.13.007 M7.4
YOLOv10nMulti91.344.883.640.244.460.341.12.842 M8.9
YOLOv11nMulti91.336.480.633.741.857.438.32.728 M6.6
YOLOv12nMulti91.445.184.542.345.061.840.42.810 M6.3
CFTMulti92.455.588.644.646.365.445.747.06 M117.2
DEYOLOnMulti91.051.885.244.146.964.443.66.09 M-
TAF-YOLOMulti91.758.689.246.944.365.846.74.518 M8.4
Note: Bold values indicate the best results, and “-“ indicates that the results are not reported.
Table 5. Performance of TAF-YOLO with different random seeds on the VEDAI dataset.
Table 5. Performance of TAF-YOLO with different random seeds on the VEDAI dataset.
Random SeedPreRecmAP50mAP50:95
168.863.767.244.3
268.063.967.043.9
368.764.167.344.3
468.764.267.344.4
569.064.767.944.7
Mean Value68.6464.1267.3444.32
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhuo, Z.; Lu, R.; Yao, Y.; Wang, S.; Zheng, Z.; Zhang, J.; Yang, X. TAF-YOLO: A Small-Object Detection Network for UAV Aerial Imagery via Visible and Infrared Adaptive Fusion. Remote Sens. 2025, 17, 3936. https://doi.org/10.3390/rs17243936

AMA Style

Zhuo Z, Lu R, Yao Y, Wang S, Zheng Z, Zhang J, Yang X. TAF-YOLO: A Small-Object Detection Network for UAV Aerial Imagery via Visible and Infrared Adaptive Fusion. Remote Sensing. 2025; 17(24):3936. https://doi.org/10.3390/rs17243936

Chicago/Turabian Style

Zhuo, Zhanhong, Ruitao Lu, Yongxiang Yao, Siyu Wang, Zhi Zheng, Jing Zhang, and Xiaogang Yang. 2025. "TAF-YOLO: A Small-Object Detection Network for UAV Aerial Imagery via Visible and Infrared Adaptive Fusion" Remote Sensing 17, no. 24: 3936. https://doi.org/10.3390/rs17243936

APA Style

Zhuo, Z., Lu, R., Yao, Y., Wang, S., Zheng, Z., Zhang, J., & Yang, X. (2025). TAF-YOLO: A Small-Object Detection Network for UAV Aerial Imagery via Visible and Infrared Adaptive Fusion. Remote Sensing, 17(24), 3936. https://doi.org/10.3390/rs17243936

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop