Next Article in Journal
Estimating Rice SPAD Values via Multi-Sensor Data Fusion of Multispectral and RGB Cameras Using Machine Learning with a Phenotyping Robot
Previous Article in Journal
Mitigating Integrity Risk in SBAS Positioning Using Enhanced IGG III Robust Estimation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MFAFNet: A Multi-Feature Attention Fusion Network for Infrared Small Target Detection

1
Xi’an Institute of Optics and Precision Mechanics, Chinese Academy of Sciences, Xi’an 710119, China
2
University of Chinese Academy of Sciences, Beijing 100190, China
3
Xi’An Key Laboratory of Spacecraft Optical Imaging and Measurement Technology, Xi’an 710119, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(17), 3070; https://doi.org/10.3390/rs17173070
Submission received: 24 July 2025 / Revised: 23 August 2025 / Accepted: 25 August 2025 / Published: 3 September 2025

Abstract

Infrared small target detection is a critical task in remote sensing applications, such as aerial reconnaissance, maritime surveillance, and early-warning systems. However, due to the inherent characteristics of remote sensing imagery, such as complex backgrounds, low contrast, and limited spatial resolution-detecting small-scale, dim infrared targets remains highly challenging. To address these issues, we propose MFAFNet, a novel Multi-Feature Attention Fusion Network tailored for infrared remote sensing scenarios. The network comprises three key modules: a Feature Interactive Fusion Module (FIFM), a Patch Attention Block (PAB), and an Asymmetric Contextual Fusion Module (ACFM). FIFM enhances target saliency by integrating the original infrared image with two locally enhanced feature maps capturing different receptive field scales. PAB exploits global contextual relationships by computing inter-pixel correlations across multi-scale patches, thus improving detection robustness in cluttered remote scenes. ACFM further refines feature representation by combining shallow spatial details with deep semantic cues, alleviating semantic gaps across feature hierarchies. Experimental results on two public remote sensing datasets, SIRST-Aug and IRSTD-1k, demonstrate that MFAFNet achieves excellent performance, with mean IoU values of 0.7465 and 0.6701, respectively, confirming its effectiveness and generalizability in infrared remote sensing image analysis.

1. Introduction

Infrared small target detection (IRSTD) is a crucial task in the field of remote sensing, with widespread applications in early warning systems, ground surveillance, maritime monitoring, and forest fire prevention [1]. As an essential component of infrared search and tracking (IRST) systems, this task benefits from the unique advantages of infrared imaging—such as long-range perception, all-weather operability, and passive sensing-making it highly suitable for remote sensing platforms, including unmanned aerial vehicles, spaceborne satellites, and high-altitude balloons.
Despite these advantages, infrared small target detection in remote sensing imagery faces persistent challenges distinct from conventional object detection. Specifically, due to long-distance imaging, large-scale scene variability, and low signal-to-clutter ratio (SCR), the target typically occupies only a few pixels and lacks distinctive shape, color, or texture information. Furthermore, background clutter and sensor-induced noise often overwhelm the weak target signals, making reliable detection particularly difficult in complex or heterogeneous environments [2].
The current infrared small target detection methods can be broadly classified into two categories: single-frame-based (SFB) and multi-frame-based (MFB) methods. For example, Han et al. [3] combine the salience of the time domain and the phase spectrum of the quaternion Fourier transform to achieve high-performance small target detection. Sun et al. [4] combine the pixel detection and trained LightGBM model to detect targets in individual frames, establishing short-strict and long-loose constraints to distinguish the target trajectory. MFB methods make use of spatial and temporal information in order to achieve detection, while SFB methods are capable of detecting small targets in each individual frame with minimal computational cost.
In the past decades, the majority of SFB methods have been model-driven due to the lack of public datasets, including filtering-based methods, local-contrast-based methods, and low-rank-based methods [5,6], local-contrast-based methods [7,8,9], and low-rank-based methods [10,11,12]. However, these model-driven methods heavily rely on prior assumptions. Constrained by the fact that these prior assumptions do not match all the complex scenes, these methods cannot be applied in the actual detection task. Furthermore, model-driven methods are typically susceptible to hyper-parameter sensitivity, which renders the utilization of handcrafted features and fixed hyper-parameters in diverse scenarios challenging.
With the recent emergence of open-source infrared datasets, data-driven approaches based on deep learning have gained momentum. In particular, convolutional neural networks (CNNs) have demonstrated promising capabilities in learning discriminative features directly from infrared imagery [13,14]. These methods offer improved generalization and scene adaptability. However, CNN-based architectures may suppress weak target features in deeper layers due to their tendency to prioritize prominent background patterns. Moreover, the reliance on large, high-quality training datasets poses a major constraint in infrared remote sensing, where annotated data is costly and scarce.
To address the limitations of CNNs, transformer-based models have been introduced into IRSTD, leveraging their global attention mechanisms to enhance long-range dependency modeling [15,16]. However, transformer architectures often overlook fine-grained local details and are computationally intensive—an issue particularly problematic for small targets embedded in high-resolution remote sensing imagery.
To address the challenges of detecting small infrared targets in complex remote sensing scenes, we propose MFAFNet, a Multi-Feature Attention Fusion Network that integrates the local detail modeling capability of CNNs with the global contextual reasoning of transformers. The network introduces a Multi-Feature Interactive Fusion Module to enhance target saliency through progressive fusion of multi-scale features, a Patch Attention Block to capture inter-pixel dependencies via multi-scale attention, and an improved Asymmetric Contextual Fusion Module (ACFM) to align spatial and semantic information across network layers. This unified design enables MFAFNet to effectively suppress clutter and highlight small targets, significantly improving detection performance in infrared remote sensing imagery. The main contributions of this paper are summarized as follows:
(1)
We propose MFAFNet, a novel multi-feature attention fusion network designed specifically for infrared small target detection in remote sensing imagery. By integrating CNN-based local perception and transformer-inspired global reasoning within a unified architecture, MFAFNet effectively addresses the challenges of low signal-to-clutter ratio and limited spatial resolution.
(2)
We design a Multi-Feature Interactive Fusion Module (FIFM) that progressively combines the original input with multi-scale local-enhanced feature maps. This fusion enhances the representation of weak targets by emphasizing their saliency and suppressing background interference.
(3)
We introduce a Patch Attention Block (PAB) with cascaded channel-spatial attention, enabling the model to capture long-range dependencies and contextual relationships among patches at multiple scales. This improves target localization in cluttered and heterogeneous infrared scenes.
(4)
We enhance the Asymmetric Contextual Fusion Module to integrate low-level spatial features with high-level semantic information. By aligning hierarchical feature distributions, this module reduces semantic gaps and supports more accurate target discrimination.
The remainder of this paper is organized as follows. Section 2 reviews related work in infrared small target detection. Section 3 presents the architecture and modules of the proposed MFAFNet in detail. Section 4 reports the experimental setup and comparative evaluation results. Section 5 concludes the paper and outlines future research directions.

2. Related Work

In recent years, infrared small target detection has attracted increasing attention in the remote sensing community due to its critical role in aerial surveillance, early warning, and environmental monitoring. To better understand the limitations of existing methods and motivate our proposed approach, we review representative model-driven algorithms, deep learning-based strategies, and transformer-enhanced techniques.

2.1. Traditional Model-Driven Approaches

In the field of infrared small target detection, conventional model-driven methods have long dominated due to the absence of large-scale, publicly available datasets. These approaches typically rely on handcrafted priors to exploit the local discontinuity between the target and its background. Among them, human visual system (HVS)-inspired methods have received notable attention, aiming to enhance target saliency by modeling local contrast and suppressing background clutter [17].
A representative example is the Local Contrast Measure (LCM) [7], which employs a sliding window to compute pixel-wise contrast and applies adaptive thresholding to locate potential targets. This method has inspired several extensions, such as Improved LCM (ILCM) [18], Relative LCM (RLCM) [19], Weighted Strengthened LCM (WSLCM) [20], and Tri-layer LCM (TLLCM) [21], each aiming to improve robustness across variable scenes. Wei et al. [8] introduced MPCM, a multi-scale patch-based contrast descriptor, which was further refined by incorporating the Local Energy Factor (LEF) [22]. Zhang et al. [9] proposed a combined Local Intensity and Gradient (LIG) framework to improve both target enhancement and background suppression.
While these methods are interpretable and computationally efficient, they rely heavily on strict prior assumptions, such as uniform background, high SCR, or fixed target morphology-which do not generalize well to real-world remote sensing imagery. As a result, their performance drops significantly in complex scenes with dynamic textures, multi-scale clutter, and unknown noise patterns, as shown in Figure 1.
To overcome these limitations, we adopt a data-driven, learning-based framework that abandons fixed priors and instead learns discriminative features directly from annotated samples across diverse backgrounds and target conditions.

2.2. Data-Driven Deep Learning Approaches

With the increasing availability of annotated infrared datasets, deep learning-based methods have rapidly emerged as alternatives to traditional algorithms. These data-driven methods can be broadly categorized into detection-based and segmentation-based paradigms.
The pioneering work of Liu et al. [23] introduced a multilayer perceptual detection network for small targets in infrared imagery, marking the transition to deep learning in SIRST. More recent efforts favor segmentation-based frameworks due to their ability to generate pixel-level target masks, which are particularly suitable for small and ambiguous targets in cluttered remote sensing environments. Dai et al. [14] proposed segmentation networks with asymmetric context modules to fuse shallow and deep features. Wang et al. [13] decomposed the detection task into two subtasks-reducing miss detections and false alarms-using generative adversarial networks (GANs) to optimize the tradeoff.
Other studies have explored hybrid strategies that integrate handcrafted priors with learned features. For instance, Hou et al. [24] constructed a mapping network from handcrafted descriptors to probabilistic maps, while Yang et al. [25] designed a depth-aware fusion module to combine original and smoothed features. Chen et al. [26] employed a supervised attention module trained on a target spread map to suppress irrelevant background noise.
The integration of attention mechanisms and transformers has further advanced IRSTD performance. Liu et al. [27] introduced a transformer with a feature enhancement module to model long-range feature dependencies. Li et al. [28] proposed a cascaded channel-spatial attention module (CSAM) to selectively enhance multi-scale features. IAANet [15] adopts a coarse-to-fine pipeline where CNNs locate rough regions and transformers refine pixel-level details. Similarly, AGPCNet [16] combines convolutional encoding with non-local attention for contextual refinement.
Despite their progress, existing CNN-based methods often suffer from limited receptive fields, which restrict their ability to model long-range dependencies. Moreover, as target features become progressively weaker in deep layers, they are often overwhelmed by dominant background structures. These methods also struggle with balancing generalization and precision, particularly under low SCR and scene heterogeneity.
To mitigate these issues, we design a multi-feature fusion mechanism that explicitly preserves and enhances small target features across network layers. Our architecture leverages both local detail modeling via CNNs and global semantic reasoning through attention mechanisms to maintain target discriminability throughout the network.

2.3. Transformer-Based Methods in Vision

Inspired by the success of self-attention in NLP [29], researchers have extended transformer architectures to computer vision tasks. Vision Transformer (ViT) [30] applies non-overlapping image patches for classification, while non-local blocks [31] generalize long-range interaction modeling in feature space. Although transformers are effective in capturing global context, they typically require large datasets and introduce substantial computational overhead. Moreover, their limited capacity for modeling local structures poses challenges for detecting small, dim targets.
To mitigate these issues, hybrid models combining CNNs with transformer-based attention have been proposed [32,33]. For instance, Liu et al. [34] leverage transformer layers to explore semantic dependencies while enhancing feature discriminability through channel-aware refinement. These hybrid networks aim to balance global perception and local precision, making them increasingly attractive for infrared small target detection in remote sensing scenarios.
Although transformers offer strong global modeling capacity, they typically require large-scale training data and introduce significant computational cost. Their attention mechanisms also tend to underperform in capturing fine-grained local details, which are crucial for detecting dim and tiny targets. In remote sensing, where high-resolution imagery and computational constraints are prevalent, these limitations hinder practical deployment.
We propose a hybrid fusion network, MFAFNet, that integrates CNNs and transformer-inspired attention in a layered and lightweight manner. Instead of relying on pure self-attention, we use multi-scale patch attention and asymmetric fusion modules to selectively aggregate spatial details and semantic contexts, ensuring both efficiency and precision in infrared remote sensing imagery.

3. Methods

In this section, we present the architecture of the proposed Multi-Feature Attention Fusion Network for infrared small target detection. We begin by describing the overall framework, followed by detailed explanations of the Multi-Feature Interactive Fusion Module and the Patch Attention Block. Finally, we introduce the enhanced Asymmetric Contextual Fusion Module designed to refine semantic integration and produce accurate detection results.

3.1. Overall Architecture

The overall architecture of the proposed method is illustrated in Figure 2. It consists of three main modules: a multi-feature fusion module to extract the features, a patch attention block to obtain long-range associations, and an improved asymmetric fusion mapping network to up-sample and generate the detection results.
As shown in Figure 2, MFAFNet takes three inputs: the raw infrared image, a local intensity image, and a local gradient image. The latter two are derived via human visual system-inspired preprocessing, which enhances small target saliency. These inputs are passed through a ResNet-50 backbone, where multi-feature fusion is embedded between residual stages through the FIFM module, allowing progressive enhancement of spatial and morphological features. The resulting feature maps are fed into the PAB module, which models inter-pixel dependencies across spatially distributed patches to capture long-range contextual associations. Finally, a U-Net-style skip connection is employed to merge low-level and high-level features. These are processed by the ACFM, which addresses feature misalignment and generates a refined binary detection map through fully connected layers. The final output highlights the spatial location of infrared small targets.

3.2. Multi-Feature Interactive Fusion Module

In ResNet50, an image is fed into a sequence of convolution layers with a down-sample, which submerges the infrared small target. Inspired by the HVS-based methods, we introduced the local intensity and gradient feature into neural network, which can enhance the salience of the target and improve the detection performance. The intensity and gradient features are fused with the morphological feature in the original image. The fusion is carried out through the redesigned multi-feature interactive fusion module.
In general, the grayscale distribution of the infrared small target exhibits an approximate centrosymmetric shape and radiates, which indicates that the small target can be described by a two-dimensional Gaussian function (2-D Gaussian function). Consequently, the local intensity and gradient operation can effectively distinguish the candidate targets. Figure 3 shows the intensity and gradient features of the infrared small target and some background. It can be observed that the majority of pixels belonging to the target have been enhanced, while those belonging to the clutter have been suppressed. Nevertheless, the grayscale distributions of certain background clutter exhibit similarities with those of the small target. The processed results still contain a substantial amount of background clutter. Conversely, a single local feature can enhance the pixels of the background, which introduces greater intrusion. As a result, we propose the multi-feature fusion module, which comprehensively utilizes multiple features and incorporation of diverse constraints. The intensity and gradient feature calculations are defined as follows:
I = m a x ( 0 , f 0 f ¯ )
G = i = 1 4 G i , i f   G m i n G m a x > k 0 , o t h e r w i s e
Our model employs a convolutional neural network (CNN), adapting ResNet20 as the underlying architecture. The key distinction lies in the multi-feature interactive fusion module, which serves to extract comprehensive features pertaining to the target. The feature-fusion strategy in each layer enables the efficient fusion of the local salient feature and the original morphological feature, thereby enhancing the semantic content of the infrared small target in the feature map. The fusion strategy is implemented in an additive mode in the current study. Consequently, our method entails the fusion of each local feature with the original feature map, followed by the superposition of the results.
The multi-feature interactive fusion module is employed in the feature extraction, as shown in Figure 4. The FIFM fuses the original image and two additional local feature images at varying depths. In our method, the original feature, intensity feature, and gradient feature maps are employed as the input. Considering the identical sampling rate of the original feature map and the two additional local feature maps, the FIFM has three input feature maps O t ,   I t ,   G t R C × W × H with the same size. The FIFM initially performs a fusion on the original feature map and one of the multiple features. Subsequently, the two sets of fusion results shown in Figure 4 are added to the output O t of the FIFM in layer t and the input O t + 1 of the convolutional layer t + 1 . The other two sets of fusion results shown in Figure 4 were employed as the input I t + 1 ,   G t + 1 of the convolutional layer t + 1 . The O t + 1 ,   I t + 1 ,   G t + 1 R C × W × H have the same size with O t ,   I t ,   G t .
O t = O t · I t + O t · G t
I t = O t · I t
G t = O t · G t
where O t ,   I t ,   G t are the original feature map and two other local feature maps in the layer t . O t ,   I t ,   G t are feature maps.
The feature fusion layer injects the rich features of the small target in intensity images and gradient images into the original feature map. Concurrently, the noise and clutter can be mitigated by the disparity between the two local feature images. To further stabilize this process, we introduce a threshold parameter k when computing gradient features, which physically corresponds to the minimum detectable edge contrast in infrared imagery. Gradients below this threshold are typically caused by background noise rather than actual targets. The value of k is determined based on the gradient magnitude histogram, ensuring genuine small target edges (spanning 2 × 2–3 × 3 pixels) are retained while weak clutter is suppressed.

3.3. Patch Attention Block

The patch attention block is the fundamental module of the network, as shown in Figure 5. The PAB incorporates a multi-scale non-local block with a cascaded channel and spatial attention module (CSAM).
The receptive field of the CNNs is the reason why the context module should be introduced to obtain the long-range dependencies between pixels. The non-local network can utilize the global computation in order to capture the correlation between each pixel and other pixels. As shown in Figure 6, the non-local block can be embedded into the neural network and effectively improves the performance as a key component. Although the non-local block is able to aggregate the pixels of the same type, it has a lower calculation efficiency.
The non-local block is applied directly to the original feature X R C × W × H with high space complexity. In the meantime, the smaller targets occupy a few pixels in the images. Consequently, the multi-scale non-local block is employed to reduce the calculation, and the feature map is divided X into s × s patches of size w × h , where w = W s ,   h = H s . In our method, the patch factor s is:
s = [ 10 ,   6 ,   5 ,   4 ]
Subsequently, each patch is fed into a non-local block in order to calculate the dependencies of pixels within a local range. Patches from the same feature map share weights, and these outputs are concentrated to generate a local associated feature map P R C × W × H . The replacement of the full feature map X by the local patch results in a reduction of the receptive field of the network. This reduction is achieved by calculating the dependencies between pixels in the local patch, with the objective of aggregating pixels of the same type. In each patch, the local feature map is capable of distinguishing the type of pixels, reducing the influence of noise and clutter. In this way, the algorithmic complexity and computational time are markedly reduced by computing the local association in each patch. The chosen patch sizes [4,5,6,10] are not arbitrary but are designed to match the scale of infrared small targets. After downsampling, one feature pixel corresponds to ~8 × 8 pixels of the original image, meaning that patches of size 4–6 effectively cover targets of 2 × 2–3 × 3 pixels plus contextual surroundings, while larger patches (e.g., 10) extend the field of view for robustness in cluttered scenes.
In each patch with the same size w × h , the output of the non-local block has been concentrated into a new associated feature map P R C × W × H . Meanwhile, the input feature map X are also fed into multiple PABs of the different patch scales in parallel. For each PAB, the result can be formulated as { P A B s 1 , P A B s 2 , } , where s i is the size of the patch. The local associated feature map and the original feature map X have been combined along the scale axis to generate the new feature map L R ( ( n + 1 ) × C ) × W × H , and the shape of the feature map has been added in dimension.
For the concatenated feature map L R ( ( n + 1 ) × C ) × W × H , a cascaded channel and spatial attention module is employed to achieve adaptive feature enhancement. As illustrated in Figure 7, the feature map has undergone modification. L R C × W × H as the final feature map. The CSAM is comprised of two cascaded attention units: a one-dimensional channel attention and a two-dimensional spatial attention. The channel attention can be obtained as follows:
L = σ M L P P m a x L + M L P P a v g L L
where denotes the element-wise multiplication, σ denotes the sigmoid function. The Ρ m a x and Ρ a v g denote the max pooling and the average pooling, respectively.
In contrast to the channel attention, the spatial attention can be succinctly described as follows:
L = σ f 7 × 7 P m a x L , P a v g L L
where f 7 × 7 represents a convolution with a 7 × 7 kernel. The pooling operation is applied in both the spatial and channel dimensions to generate the weight matrix M C R C × 1 × 1 , M S R 1 × W × H , which is the result of the channel attention and spatial attention.
It is worth noting that, although non-local attention is typically high in complexity, the patch-wise multi-scale design used in PAB restricts computation to local regions, thereby reducing complexity from quadratic to linear with respect to feature map size and ensuring practical efficiency.

3.4. Asymmetric Contextual Fusion Module

To integrate the subtle features and the coarse semantic features from the higher layers, we develop a novel asymmetric contextual fusion module, which is inspired by the ACM. As shown in Figure 8, the low-level semantics X l and the deep-level semantics X d are fed into ACFM as input. Rather than ACM, our approach initially fuses the low-level and deep-level semantics, subsequently generating the pixel and channel attention weight matrices to guide the fusion.
W c a = σ B W 2 δ B W 1 P X d
W p a = σ B W 2 δ B W 1 X l
The point-attention mechanism, as defined in Equation (10), is employed in this study. We use 1 × 1 convolution dimensionality reduction and the channel-attention mechanism in Equation (9). Where p , B , δ , σ denote the global average pooling, batch normalization, the rectified linear unit, sigmoid function, respectively. Based on the weight matrix W c a , W p a , The low-level and deep-level semantics are unified by introducing a constraint that is applied separately to each level. As illustrated in Equation (11), our module is capable of resolving the discrepancy between the low-level semantics. X l and the deep-level semantics X d , and reduce the difference caused by separate constraints.
X A C F M = W c a X l + W p a X d
As shown in Figure 9, the heatmap comparison provides direct qualitative evidence of the effectiveness of the ACFM. Before fusion, the low-level spatial features exhibit strong responses along local textures and background clutter, which leads to noisy activations and potential false alarms. In contrast, the high-level semantic features focus more on the approximate target location but suffer from blurred boundaries and incomplete structural representation, reflecting a semantic gap between the two levels. After applying ACFM, the fused heatmap demonstrates a clear improvement: activations are concentrated within the true target regions, boundaries become sharper, and irrelevant background responses are significantly suppressed. This indicates that ACFM successfully aligns low-level details with high-level semantics, thereby reducing the semantic gap and enhancing the discriminative capability of the network in cluttered infrared environments.

3.5. Loss Function

To optimize the performance of MFAFNet in detecting small infrared targets under complex backgrounds, we construct a hybrid loss function that incorporates both pixel-level accuracy and structure-aware constraints. Given the severe foreground-background imbalance and the sparse distribution of small targets, traditional binary cross-entropy loss is insufficient to guide precise learning. To address this, we integrate Soft-IoU loss and a semantic consistency loss, and introduce a feature supervision loss at intermediate layers to improve training stability and target localization.
(1)
Binary Cross-Entropy (BCE) Loss
Binary Cross-Entropy is used to supervise pixel-level classification. For a predicted probability map Y ^ , and ground truth label map Y, the BCE loss is defined as:
L B C E = 1 N i = 1 N Y i · l o g Y i ^ + 1 Y i · l o g 1 Y i ^
where N is the total number of pixels. This term ensures the pixel-wise correctness of the final binary prediction.
(2)
Soft-IoU Loss
To address the foreground-background imbalance and better model region-level accuracy, we incorporate Soft-IoU loss, which directly optimizes the Intersection-over-Union between the prediction and ground truth:
L I o U = 1 i Y i ^ · Y i i Y i ^ + Y i Y i ^ · Y i + ϵ
where ϵ is a small constant to avoid division by zero. Soft-IoU encourages spatial alignment between predicted blobs and actual target regions, which is particularly critical for sparse and small objects.
(3)
Multi-Scale Feature Supervision Loss
To guide the intermediate feature learning in FIFM and PAB, we introduce a deep supervision strategy. Let Y ^ l be the auxiliary output of the intermediate feature map from layer l, and Y be the ground truth label map resized to the same resolution. The supervision loss is defined as:
L f e a t = l = 1 L α l · L B C E Y ^ l , Y l
where α l is the weighting factor for each layer, and L is the total number of supervised layers. This term enforces target-aware feature enhancement at different depths.
(4)
Semantic Alignment Loss
The Asymmetric Contextual Fusion Module aligns low-level spatial features and high-level semantic information. To reduce semantic inconsistency, we introduce a semantic alignment constraint:
L a l i g n = μ C 1 F l o w μ C 2 F h i g h 2 2
where F l o w and F h i g h are the low-level and high-level features, C 1 and C 2 are projection layers (e.g., 1 × 1 convolutions), and μ (⋅) is a non-linear activation (e.g., ReLU). This loss enforces consistency in semantic representations across scales.
The total loss function used to train MFAFNet is a weighted combination of the above components:
L a l i g n = λ 1 L B C E + λ 2 L I o U + λ 3 L f e a t + λ 4 L a l i g n
where λ 1 , λ 2 ,   λ 3 ,   λ 4 are hyperparameters that control the contribution of each loss term.

4. Experiments

In this section, we employ experimental techniques to assess the efficacy of MFAFNet. We commence by providing an overview of the implementation details. Subsequently, the experimental setup is presented, including the evaluation metrics, public datasets, and state-of-the-art methods used for comparison. In parallel, we conduct ablation studies for each module with the objective of evaluating the practicality of each module. Finally, a comparison is made between the proposed method and state-of-the-art methods in both the visual and numerical fields.

4.1. Implementation Details and Setting

The input images used in our method vary in size; therefore, a preprocessing module is employed to uniformly resize all three input types (raw infrared, intensity, and gradient images) to a resolution of 256 × 256. These preprocessed images are then stacked into batches and fed into the backbone network for feature extraction. The feature extraction backbone is based on ResNet-50, and three parallel extraction streams-corresponding to each image modality-are constructed using identical architectures. Each stream begins with a convolutional layer with a kernel size of 7, followed by three residual blocks. Notably, each residual layer is embedded with a Multi-Feature Interactive Fusion Module to enhance feature integration across modalities. The output feature map has a spatial resolution of 32 × 32. Within the Patch Attention Block, multi-scale patches of varying sizes (e.g., 1 × 1, 3 × 3, and 5 × 5) are utilized to model contextual dependencies at different receptive fields. Given the significant class imbalance between small targets and background regions, we adopt the Soft-IoU loss function to improve training stability and enhance sensitivity to small foreground objects. The entire network is trained for 100 epochs using the AdaGrad optimizer, with an initial learning rate of 0.001 and a batch size of 16, without any data augmentation strategies. All experiments are conducted on a workstation equipped with an NVIDIA RTX 4090 GPU.
In this study, we adopt commonly used pixel-level evaluation metrics, including Precision, Recall, Intersection over Union (IoU), F-measure, and Area Under the Curve (AUC), to comprehensively assess the performance of infrared small target detection methods on benchmark datasets.
The IoU is a pixel-level evaluation metric, which is calculated by the ratio of the intersection and the union areas between the predicted result and the label.
I o U = A r e a   o f   O v e r l a p A r e a   o f   U n i o n
P r e c i s i o n and R e c a l l , respectively, represent the ratio of correctly classified pixels to labelled positive targets and predicted positive targets. The F-measure is employed to assess the relationship between P r e c i s i o n and R e c a l l .
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F m e a s u r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
Furthermore, the area under the curve (AUC) is introduced as a quantitative metric for receiver operating characteristic (ROC) to evaluate the comprehensive potential of a method. The true positive rate (TPR) and the false positive rate (FPR) are defined as follows:
F P R = F P F P + T N
T P R = T P T O + F N
Since there are few public infrared small target detection datasets, part of the ISTD dataset comes from artificial synthesis with target and background. Through the SIRST dataset [14], which is widely used, and a real IRST dataset with high-quality images, it has only 427 images in total, which leads to problems, such as unstable network training and easy over-fitting. Hence, we consider the SIRST-Aug dataset [16], which is based on SIRST and enhanced (clipping, inversion, displacement, etc.). SIRST-Aug includes 8525 images as a training set and 545 images as a test set with the fixed resolution 256 × 256, which is enough to train neural networks.
On the other hand, we also adopt the IRSTD-1k dataset [35], captured by an infrared camera in the real world. The IRSTD-1k comprises 1000 infrared images and labels at the pixel level, all with a resolution of 512 × 512. We divided the IRSTD-1k dataset into a training set, a validation set, and a test set, with a proportion of 50%, 30% and 20%, respectively. The validation set is used exclusively for hyperparameter tuning and model selection, while the test set is kept unseen during training to provide an unbiased performance evaluation. The dataset includes different kinds of targets, such as drones, vessels, and vehicles. Meanwhile, it covers lots of different real scenes and backgrounds, including the sea, river, sky, and city, with random clutter and noise.
Figure 10 shows the representation images of the SIRST-Aug and IRSTD-1k dataset. In addition, the detail of these two datasets is listed in Table 1. Compared in these two datasets, we can evaluate the effectiveness of the proposed method in the real dataset and the generated dataset, respectively.

4.2. Ablation Study

To analyze the rationality of each component of MFAFNet, we run our models with different settings.
Effectiveness of local feature. As shown in Table 2, we verify their effectiveness by introducing two different local features. In MIFM, the intensity feature and gradient feature, respectively, are removed. Meanwhile, we are keeping the same segmentation task, evaluation metrics, and the dataset.
From Table 2, we can see that when a gradient feature and an intensity feature are introduced into the same backbone, the performance of the model can be improved in the same dataset. This improvement in performance can be attributed to two factors. Firstly, the local feature can enhance the infrared small target in the original image, which has greater contrast with the background. Secondly, a single local feature may similarly enhance the noise and clutter, resulting in a slight performance degradation. To address this, we introduce a different local feature to constrain the enhanced clutter with an appropriate complexity.
Effectiveness of local patch size in PAB. Given that the feature map size is 32 × 32 with down-sampling occurring three times, it can be inferred that each pixel in the feature map represents an 8 × 8 region in the original image. To investigate the effect of the patch size, multiple sets of comparison experiments were set up. It can be posited that the appropriate patch size should be capable of covering the entire target area and a certain background. To test this hypothesis, an experiment was designed with different patch sizes. As shown in Table 3, it can be observed that the combination of patch size and PAB exhibits the most optimal performance. This indicates that the utilization of different patch sizes can facilitate the more comprehensive coverage of diverse targets within the dataset.
Effectiveness of ACFM. To explicitly investigate the contribution of the ACFM, which is designed to bridge shallow and deep features via pixel-wise and channel-wise attention, we conduct a dedicated ablation study. While the previous section validates the effectiveness of the FIFM and the PAB, the standalone impact of ACFM was previously embedded within the complete pipeline, making it difficult to discern its individual role. To isolate the effect of the Asymmetric Contextual Fusion Module (ACFM) and examine its interaction with other components, we design four variants of MFAFNet: (1) w/o ACFM, which replaces the module with a direct summation of low-level and deep-level features without attention; (2) w/o ACFM + FIFM, which removes both ACFM and FIFM to assess the synergy between semantic alignment and multi-modal fusion; (3) w/o ACFM + PAB, which eliminates ACFM and PAB to evaluate the role of ACFM in supporting global-local integration; and (4) w/o ACFM + FIFM + PAB, which retains only the backbone and segmentation head, removing all three modules to serve as a minimal reference model. We compare these variants against the full MFAFNet using the SIRST-Aug dataset. The results are shown in Table 4.
From Table 4, removing only the ACFM module results in a performance drop of 4.1% in F-measure and 4.3% in mIoU, demonstrating its significant standalone impact on detection quality. The observed reduction confirms that the attention-based asymmetric fusion of shallow and deep features improves both target localization and boundary delineation, especially in infrared imagery where semantic granularity is unevenly distributed. Furthermore, when ACFM is removed in conjunction with FIFM or PAB, the performance deteriorates even further. This indicates that ACFM is not only individually effective but also works synergistically with other modules. For example, without ACFM and FIFM, the mIoU drops to 0.6802, suggesting that FIFM benefits from the refined semantic alignment provided by ACFM. Likewise, removing ACFM and PAB causes a greater performance decline than removing either one alone, highlighting that ACFM enhances the contextual modeling effect of PAB by providing more semantically aligned feature inputs.
Ablation on Hybrid Loss Weights. To investigate the effect of the weighting factors (λ1–λ4) in Equation (16), we conduct an ablation study on the SIRST-Aug dataset. Specifically, we vary the weights assigned to each component loss while keeping the others fixed. The results are summarized in Table 5.
From Table 5, it is clear that the balanced setting (all weights = 1) achieves the best tradeoff across metrics. Increasing the contribution of a single term often biases the optimization towards pixel accuracy or region alignment but degrades overall generalization. This confirms that the hybrid loss benefits from a synergistic balance, and further hyperparameter tuning yields only marginal improvements.

4.3. Comparison to Excellent Methods

To comprehensively evaluate the effectiveness of the proposed MFAFNet, we compare it with a series of representative infrared small target detection methods specifically designed for remote sensing scenarios. The baseline methods include both model-driven approaches, such as LEF [22], WSLCM [20], PSTNN [12], and MSLSTIPT [36], which rely on filtering, local contrast, or low-rank priors; and data-driven deep learning methods, including AGPCNet [16], DNANet [28], ACM [14], MSHNet [37], RISTDnet [24], Trans-IRSD [27], IRSAM [38], STPSA-Net [39], and MiM-ISTD [40]. All comparative models are implemented using the parameter settings provided in their respective original papers.
Among the classical approaches, LEF introduces a local energy factor to jointly evaluate local dissimilarity and brightness contrast, while WSLCM incorporates a strengthened local contrast measure and a region-aware weighting function to suppress noise in complex backgrounds. RIPT models target-background separation as a robust low-rank tensor recovery problem using patch-based tensor representation, and PSTNN improves upon RIPT by introducing a non-convex low-rank regularizer and a background-aware prior map, enabling a better balance between detection accuracy and computational efficiency. MSLSTIPT further explores spatial-temporal structure by leveraging multisubspace learning in tensor form, effectively handling heterogeneous scenes.
In the realm of learning-based methods, ACM proposes an asymmetric contextual modulation structure that fuses global feedback and channel attention to highlight small targets. AGPCNet constructs an attention-guided pyramid framework, combining patch-level association and global context modeling through a multi-scale structure. DNANet introduces a dense nested attention design to maintain target features across layers, while MSHNet focuses on optimizing the loss function via a scale- and location-sensitive formulation (SLS) to improve detection robustness. RISTDnet integrates handcrafted features with deep CNNs to build a likelihood estimation framework suitable for low-SNR conditions. Trans-IRSD employs a transformer backbone with feature enhancement modules to extract global dependencies and mitigate missed detections in complex infrared backgrounds. IRSAM adapts the Segment Anything Model (SAM) to thermal imaging, combining Perona-Malik diffusion and a granularity-aware decoder to bridge the domain gap. STPSA-Net encodes infrared images into semantic tokens and applies patch-wise spatial attention to enhance representation and localization.
Furthermore, MiM-ISTD adopts a state-space model inspired by Mamba to overcome the quadratic complexity of traditional transformers. It introduces a nested Mamba-in-Mamba structure, where Outer Mamba captures global dependencies across visual “sentences” (patches), and Inner Mamba captures fine-grained local structures within sub-patches. This design ensures efficient inference on high-resolution infrared images, achieving substantial reductions in computational cost and memory usage while maintaining state-of-the-art accuracy.
(1)
Quantitative evaluation: The results of different methods on two public datasets, SIRST-Aug and IRSTD-1k, are presented in Table 5. The maximum value of each column is indicated in bold black. For the model-driven method, the obtained predicts indicate the probability of detecting targets. Hence, we set an adaptive threshold T to remove low-response areas:
T = M a x [ M a x G × 0.7 ,   0.5 × σ G + a v g ( G ) ]
where M a x G represents the large value of the output image, and σ G and a v g ( G ) mean the standard deviation and average value. For data-driven methods, we adopt the fixed value, which is the same as in their original papers. Following common evaluation practice in infrared small target detection, we adopt adaptive thresholding (Equation (23)) for model-driven methods and fixed thresholds from original papers for deep learning-based methods. This choice reflects the methodological nature of each category: model-driven methods typically produce continuous response maps requiring post-hoc threshold tuning, while learning-based methods are trained end-to-end with fixed output heads and thresholds jointly optimized during training. Using fixed thresholds for deep models ensures consistency with their reported performance and published baselines. We acknowledge that the adaptive thresholding might not fully compensate for the sensitivity of handcrafted methods, potentially leading to lower recall. However, our goal here is not to penalize model-driven methods, but to provide a realistic and reproducible comparative benchmark under commonly accepted evaluation conditions.
From the data presented in Table 6, it can be observed that our MFAFNet methods demonstrate a clear superiority over model-driven methods. This is because model-driven methods are typically designed for specific scenarios, and the performance of these methods is highly dependent on the manually selected hyperparameters (e.g., patch size, threshold value), which limit their generalizability. Moreover, we can observe that most model-driven methods exhibit high precision and low recall, indicating that these methods are able to correctly detect a limited number of pixels within the target region, rather than the entire target. As shown in Figure 11, Figure 12, Figure 13 and Figure 14, model-driven methods obtain the correct target region with a few pixels, which leads to lower mIoU than data-driven methods. The reason for this phenomenon is that model-driven methods introduce various constraints to suppress the background while suppressing the target.
As shown in Table 6, the improvement achieved by MFAFNet over other data-driven methods is obvious. On the SIRST-Aug dataset, MFAFNet achieves a precision of 0.8376, a recall of 0.8652, an mIoU of 0.7465, and an F-measure of 0.8512, representing the highest performance among all competing methods. Compared with MiM-ISTD, which achieves an F-measure of 0.8431 and an mIoU of 0.7288, MFAFNet demonstrates improved detection stability and boundary segmentation accuracy. While IRSAM shows strong recall at 0.8493 and an F-measure of 0.8256, its mIoU of 0.7025 is still noticeably lower than that of MFAFNet, indicating inferior localization capability for small targets. STPSA-Net, as a transformer-based method emphasizing semantic token refinement, achieves competitive results, with an F-measure of 0.7954 and an mIoU of 0.6603, yet still falls short of MFAFNet in all metrics.
Classical model-driven methods such as LEF, WSLCM, RIPT, PSTNN, and MSLSTIPT exhibit relatively high precision but suffer from extremely low recall and poor mIoU values. For instance, RIPT reports a precision of 0.9484 but only achieves a recall of 0.0800 and an mIoU of 0.0797, resulting in a low F-measure of 0.1476. Similar trends are observed for PSTNN and WSLCM, where conservative detection strategies result in severe under-detection of small targets. MSLSTIPT, though designed for heterogeneous backgrounds, yields the lowest recall of 0.0400 and the lowest F-measure of 0.0766, indicating poor adaptation to the single-frame detection setting.
Among other deep learning methods, DNANet, AGPC, and ACM present severe imbalances between precision and recall. DNANet reaches a relatively high precision of 0.8420 but achieves a recall of only 0.0304 and an mIoU of 0.0302. AGPC and ACM similarly exhibit limited recall, which compromises their overall detection performance. MSHNet, which focuses on loss function design, shows improved balance between precision and recall compared to other learning-based baselines, but its F-measure of 0.2328 remains significantly lower than that of MFAFNet.
On the IRSTD-1k dataset, MFAFNet continues to demonstrate superior generalization performance. It achieves a precision of 0.7907, a recall of 0.8146, an mIoU of 0.6701, and an F-measure of 0.8025, outperforming all compared methods in terms of both localization accuracy and overall segmentation quality. MiM-ISTD ranks second with an F-measure of 0.7879 and an mIoU of 0.6500, confirming its robustness but still slightly inferior to MFAFNet. IRSAM achieves an F-measure of 0.7526 and an mIoU of 0.6034, showing strong consistency but relatively weaker region precision. STPSA-Net obtains an F-measure of 0.7417 and an mIoU of 0.5895, indicating that its semantic refinement strategy is effective but less capable in handling cluttered backgrounds compared to MFAFNet.
Model-driven methods on IRSTD-1k continue to exhibit significant performance gaps. LEF, PSTNN, and RIPT report mIoU values below 0.09, with F-measure scores ranging from 0.1073 to 0.1495. Although these methods achieve moderate precision, their recall remains too low to support practical applications in real-world remote sensing. RISTDnet performs relatively better among model-driven methods, reaching an F-measure of 0.3024 and an mIoU of 0.1782, but still lags far behind the learning-based models.
Overall, MFAFNet demonstrates state-of-the-art performance on both benchmark datasets. Its consistently high scores across all evaluation metrics validate the effectiveness of its multi-feature fusion, patch-level attention, and semantic contextual integration strategies. The results confirm that MFAFNet provides a robust and accurate solution for infrared small target detection under complex remote sensing conditions.
(2)
Qualitative evaluation: To further illustrate the effectiveness of our method in complex infrared imaging scenarios, qualitative detection results are presented on two benchmark datasets, SIRST-Aug and IRSTD-1k, as shown in Figure 11, Figure 12, Figure 13 and Figure 14. Four representative scenes are selected from each dataset, covering typical challenges such as low contrast, cluttered backgrounds, and small-scale targets. The output results of ten representative methods, including both model-driven and learning-based approaches, are displayed in a unified 3D visualization format for intuitive comparison. Each method’s output is marked in the upper-left corner of the corresponding subfigure to ensure clarity.
The visual results indicate that traditional model-driven methods such as LEF, RIPT, and PSTNN tend to produce overly smooth responses or fail to highlight small targets, especially under low signal-to-clutter conditions. These methods often suppress background noise effectively but sacrifice target recall, resulting in incomplete or missing detections. For example, in cluttered background scenarios, PSTNN exhibits weak responses and underestimates target size due to its strong regularization priors.
Deep learning-based methods such as DNANet, AGPC, and ACM provide more distinct responses in the target regions but often introduce significant false alarms in background areas. This phenomenon is particularly evident in densely textured scenes, where the lack of fine-grained semantic supervision leads to over-activation around noise edges and structural clutter.
In contrast, recent Transformer-based or attention-enhanced methods like MiM-ISTD, IRSAM, and STPSA-Net show noticeable improvements in target saliency and background suppression. MiM-ISTD demonstrates strong target localization with relatively clean backgrounds but occasionally produces blurred target edges. IRSAM and STPSA-Net offer sharper segmentation boundaries; however, in low-contrast regions, they may still fail to differentiate faint targets from thermal clutter.
Our proposed MFAFNet achieves the most accurate and visually consistent results across all scenes. It effectively highlights true target regions while suppressing irrelevant background responses, even in scenes with multiple interfering artifacts or weak target contrast. The predicted masks align closely with the ground truth in both shape and position, demonstrating high confidence and completeness. The integration of multi-feature fusion and hierarchical attention mechanisms enables MFAFNet to maintain semantic consistency while capturing local structural cues critical for small target delineation.
Overall, the qualitative results corroborate the findings from quantitative evaluations. MFAFNet offers superior target saliency, precise localization, and robust background discrimination, validating its practical applicability for infrared small target detection in diverse remote sensing environments.

4.4. Efficiency and Deployment Evaluation

To further verify the practical applicability of MFAFNet under real-world constraints, especially in computationally limited remote sensing platforms, we conducted an in-depth evaluation of its inference efficiency, model complexity, and feasibility of lightweight deployment. This section complements the performance assessment by adding resource-oriented metrics, including FLOPs, model parameters, inference time, and memory footprint. Additionally, we explore the potential of replacing the ResNet-50 backbone with a lightweight alternative (MobileNetV2) to validate real-time suitability. As shown in Table 7, MFAFNet achieves a favorable balance between accuracy and efficiency, significantly reducing computational cost compared with transformer-based methods while maintaining superior detection performance.
Compared to transformer-based MiM-ISTD, MFAFNet reduces both FLOPs and inference time by ~43%, while still achieving higher mIoU and F-measure. The original MFAFNet takes only 24.7 ms per image, corresponding to ~40 FPS, which is acceptable for most real-time monitoring scenarios.
When integrated with MobileNetV2, MFAFNet-Lite achieves a significant reduction in model size and latency with only a slight performance drop (e.g., F-measure decreases by 1.8% on SIRST-Aug). This result suggests that the architecture is amenable to efficient deployment on edge devices such as onboard UAV systems or embedded GPU platforms.

5. Conclusions

In this paper, we propose MFAFNet, a novel infrared small target detection framework designed to enhance detection accuracy in complex remote sensing scenarios. The network introduces a Multi-Feature Interactive Fusion Module, which is embedded beneath the convolutional layers to effectively extract and integrate local saliency features inspired by the human visual system. This fusion strategy enriches the semantic representation of small targets while suppressing background interference. To further improve contextual understanding, we design a Patch Attention Block that captures long-range dependencies through multi-scale patch-wise attention. A Cascaded Channel and Spatial Attention Module is incorporated to adaptively refine the aggregated features across scales. Additionally, an Asymmetric Contextual Fusion Module is employed to integrate low-level spatial details with high-level semantic cues, enabling more precise and robust target localization. Ablation studies confirm the effectiveness of each individual component within the network. Comprehensive experiments on both synthetic and real-world infrared datasets demonstrate that MFAFNet achieves superior performance compared to state-of-the-art methods, with higher detection accuracy and fewer false alarms under challenging conditions.

Author Contributions

Conceptualization, W.C.; Data curation, H.W.; Formal analysis, S.D.; Funding acquisition, H.W.; Investigation, Y.C. and H.W.; Methodology, Z.Z., S.D. and Y.C.; Project administration, W.C.; Resources, Z.Z.; Software, Z.Z. and W.C.; Supervision, Y.C.; Writing—original draft, Z.Z.; Writing—review and editing, S.D. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the West Light Foundation of the Chinese Academy of Sciences under the Grant XAB2022YN06.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ma, J.; Pan, N.; Yin, D.; Wang, D.; Zhou, J. MEFA-Net: Multilevel feature extraction and fusion attention network for infrared small-target detection. Remote Sens. 2025, 17, 2502. [Google Scholar] [CrossRef]
  2. Li, S.; Huang, J.; Duan, Q.; Li, Z. WT-HMFF: Wavelet transform convolution and hierarchical multi-scale feature fusion network for detecting infrared small targets. Remote Sens. 2025, 17, 2268. [Google Scholar] [CrossRef]
  3. Han, Y.; Zhang, P.; Fei, C.; Wang, X. Infrared small target detection based on spatio-temporal saliency in video sequence. In Proceedings of the IEEE International Computer Conference on Wavelet Active Media Technology and Information Processing, Chengdu, China, 18–20 December 2015; pp. 279–282. [Google Scholar]
  4. Sun, X.; Guo, L.; Zhang, W.; Wang, Z.; Yu, Q. Small aerial target detection for airborne infrared detection systems using LightGBM and trajectory constraints. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 9959–9973. [Google Scholar] [CrossRef]
  5. Deshpande, S.D.; Er, M.H.; Venkateswarlu, R.; Chan, P. Max-mean and max-median filters for detection of small targets. In Signal and Data Processing of Small Targets 1993, Proceedings of the Optical Engineering and Photonics in Aerospace Sensing, Orlando, FL, USA, 20–22 July 1999; SPIE: Bellingham, WA, USA, 1993; pp. 74–83. [Google Scholar]
  6. Wang, X.; Lv, G.; Xu, L. Infrared dim target detection based on visual attention. Infrared Phys. Technol. 2012, 55, 513–521. [Google Scholar] [CrossRef]
  7. Chen, C.; Li, H.; Wei, Y.; Xia, T.; Tang, Y. A local contrast method for small infrared target detection. IEEE Trans. Geosci. Remote Sens. 2013, 52, 574–581. [Google Scholar] [CrossRef]
  8. Wei, Y.; You, X.; Li, H. Multiscale patch-based contrast measure for small infrared target detection. Pattern Recognit. 2016, 58, 216–226. [Google Scholar] [CrossRef]
  9. Zhang, H.; Zhang, L.; Yuan, D.; Chen, H. Infrared small target detection based on local intensity and gradient properties. Infrared Phys. Technol. 2018, 89, 88–96. [Google Scholar] [CrossRef]
  10. Zhang, L.; Peng, L.; Zhang, T.; Cao, S.; Peng, Z. Infrared small target detection via non-convex rank approximation minimization joint/2,1 norm. Remote Sens. 2018, 10, 1821. [Google Scholar] [CrossRef]
  11. Qin, Y.; Bruzzone, L.; Gao, C.; Li, B. Infrared small target detection based on facet kernel and random walker. IEEE Trans. Geosci. Remote Sens. 2019, 57, 7104–7118. [Google Scholar] [CrossRef]
  12. Zhang, L.; Peng, Z. Infrared small target detection based on partial sum of the tensor nuclear norm. Remote Sens. 2019, 11, 382. [Google Scholar] [CrossRef]
  13. Guan, T.; Chang, S.; Deng, Y.; Xue, F.; Wang, C.; Jia, X. Oriented SAR ship detection based on edge deformable convolution and point set representation. Remote Sens. 2025, 17, 1612. [Google Scholar] [CrossRef]
  14. Dai, Y.; Wu, Y.; Zhou, F.; Barnard, K. Asymmetric contextual modulation for infrared small target detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Virtual Conference, 5–9 January 2021; pp. 950–959. [Google Scholar]
  15. Wang, K.; Du, S.; Liu, C.; Cao, Z. Interior Attention-Aware Network for Infrared Small Target Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 3163410. [Google Scholar] [CrossRef]
  16. Zhang, T.; Cao, S.; Pu, T.; Peng, Z. AGPCNet: Attention-guided pyramid context networks for infrared small target detection. arXiv 2021, arXiv:2111.03580. [Google Scholar] [CrossRef]
  17. Deng, H.; Sun, X.; Liu, M.; Ye, C.; Zhou, X. Small infrared target detection based on weighted local difference measure. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4204–4214. [Google Scholar] [CrossRef]
  18. Han, J.; Ma, Y.; Zhou, B.; Fan, F.; Liang, K.; Fang, Y. A robust infrared small target detection algorithm based on human visual system. IEEE Geosci. Remote Sens. Lett. 2014, 11, 2168–2172. [Google Scholar]
  19. Han, J.; Liang, K.; Zhou, B.; Zhu, X.; Zhao, J.; Zhao, L. Infrared small target detection utilizing the multiscale relative local contrast measure. IEEE Geosci. Remote Sens. Lett. 2018, 15, 612–616. [Google Scholar] [CrossRef]
  20. Han, J.; Moradi, S.; Faramarzi, I.; Zhang, H.; Zhao, Q.; Zhang, X.; Li, N. Infrared small target detection based on the weighted strengthened local contrast measure. IEEE Geosci. Remote Sens. Lett. 2020, 18, 1670–1674. [Google Scholar] [CrossRef]
  21. Han, J.; Moradi, S.; Faramarzi, I.; Liu, C.; Zhang, H.; Zhao, Q. A local contrast method for infrared small-target detection utilizing a tri-layer window. IEEE Geosci. Remote Sens. Lett. 2019, 17, 1822–1826. [Google Scholar] [CrossRef]
  22. Xia, C.; Li, X.; Zhao, L.; Shu, R. Infrared small target detection based on multiscale local contrast measure using local energy factor. IEEE Geosci. Remote Sens. Lett. 2019, 17, 157–161. [Google Scholar] [CrossRef]
  23. Liu, M.; Du, H.; Zhao, Y.; Dong, L.; Hui, M.; Wang, S.X. Image small target detection based on deep learning with SNR controlled sample generation. Curr. Trends Comput. Sci. Mech. Autom. 2017, 1, 211–220. [Google Scholar]
  24. Hou, Q.; Wang, Z.; Tan, F.; Zhao, Y.; Zheng, H.; Zhang, W. RISTDnet: Robust infrared small target detection network. IEEE Geosci. Remote Sens. Lett. 2022, 19, 7000805. [Google Scholar] [CrossRef]
  25. Yang, Z.; Ma, T.; Ku, Y.; Ma, Q.; Fu, J. DFFIR-net: Infrared dim small object detection network constrained by gray-level distribution model. IEEE Trans. Instrum. Meas. 2022, 71, 5026215. [Google Scholar] [CrossRef]
  26. Chen, F.; Gao, C.; Liu, F.; Zhao, Y.; Zhou, Y.; Meng, D.; Zuo, W. Local patch network with global attention for infrared small target detection. IEEE Trans. Aerosp. Electron. Syst. 2022, 58, 3979–3991. [Google Scholar] [CrossRef]
  27. Liu, F.; Gao, C.; Chen, F.; Meng, D.; Zuo, W.; Gao, X. Infrared small and dim target detection with transformer under complex backgrounds. IEEE Trans. Image Process. 2023, 32, 5921–5932. [Google Scholar] [CrossRef] [PubMed]
  28. Li, B.; Xiao, C.; Wang, L.; Wang, Y.; Lin, Z.; Li, M.; An, W.; Guo, Y. Dense nested attention network for infrared small target detection. IEEE Trans. Image Process. 2022, 32, 1745–1758. [Google Scholar] [CrossRef]
  29. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Aidan, N.; Kziser, L. Attention is all you need. Adv. Neural Inf. Process Syst. 2017, 30, 1–11. [Google Scholar]
  30. Alexey, D.; Lucas, B.; Alexander, K.; Dirk, W.; Zhai, X.; Thomas, U.; Mostafa, D.; Matthias, M.; Georg, H.; Sylvain, G.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  31. Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-local neural networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7794–7803. [Google Scholar]
  32. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable detr: Deformable transformers for end-to-end object detection. arXiv 2020, arXiv:2010.04159. [Google Scholar]
  33. Dai, Z.; Cai, B.; Lin, Y.; Chen, J. Up-detr: Unsupervised pre-training for object detection with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Online, 19–25 June 2021; pp. 1601–1610. [Google Scholar]
  34. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  35. Zhang, M.; Zhang, R.; Yang, Y.; Bai, H.; Zhang, J.; Guo, J. ISNet: Shape matters for infrared small target detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–23 June 2022; pp. 877–886. [Google Scholar]
  36. Sun, Y.; Yang, J.; An, W. Infrared dim and small target detection via multiple subspace learning and spatial-temporal patch-tensor model. IEEE Trans. Geosci. Remote Sens. 2020, 59, 3737–3752. [Google Scholar] [CrossRef]
  37. Liu, Q.; Liu, R.; Zheng, B. Infrared small target detection with scale and location sensitivity. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 17490–17499. [Google Scholar]
  38. Zhang, M.; Wang, Y.; Guo, J. IRSAM: Advancing segment anything model for infrared small target detection. In Proceedings of the European Conference on Computer Vision, Milan, Italy, 29 September–4 October 2024; pp. 233–249. [Google Scholar]
  39. Liu, S.; Qiao, B.J.; Li, S. Patch spatial attention networks for semantic token transformer in infrared small target detection. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5003014. [Google Scholar] [CrossRef]
  40. Chen, T.; Ye, Z.; Tan, Z.; Gong, T.; Wu, Y.; Chu, C.; Liu, B.; Yu, N.; Ye, J. MiM-ISTD: Mamba-in-mamba for efficient infrared small-target detection. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5007613. [Google Scholar] [CrossRef]
Figure 1. The results of HVS methods include MPCM, LIG, and RLCM.
Figure 1. The results of HVS methods include MPCM, LIG, and RLCM.
Remotesensing 17 03070 g001
Figure 2. Overall architecture of proposed MFAFNet. The network is composed of three functional modules: FIFM for feature-level enhancement, PAB for contextual modeling, and ACFM for final semantic fusion and upsampling.
Figure 2. Overall architecture of proposed MFAFNet. The network is composed of three functional modules: FIFM for feature-level enhancement, PAB for contextual modeling, and ACFM for final semantic fusion and upsampling.
Remotesensing 17 03070 g002
Figure 3. Original images, gradient feature, intensity feature, and LIG images.
Figure 3. Original images, gradient feature, intensity feature, and LIG images.
Remotesensing 17 03070 g003
Figure 4. An illustration of the FIFM among each layer. (ac) have the output of FIFM and the input of the next layer with the same size.
Figure 4. An illustration of the FIFM among each layer. (ac) have the output of FIFM and the input of the next layer with the same size.
Remotesensing 17 03070 g004
Figure 5. Illustration of patch attention block.
Figure 5. Illustration of patch attention block.
Remotesensing 17 03070 g005
Figure 6. Illustration of Non-Local network.
Figure 6. Illustration of Non-Local network.
Remotesensing 17 03070 g006
Figure 7. Illustration of cascaded channel and spatial attention module.
Figure 7. Illustration of cascaded channel and spatial attention module.
Remotesensing 17 03070 g007
Figure 8. Illustration of asymmetric contextual fusion module.
Figure 8. Illustration of asymmetric contextual fusion module.
Remotesensing 17 03070 g008
Figure 9. Heatmap comparison of feature activations before and after ACFM. ACFM reduces the semantic gap by aligning low-level and high-level representations, resulting in sharper target boundaries and better clutter suppression.
Figure 9. Heatmap comparison of feature activations before and after ACFM. ACFM reduces the semantic gap by aligning low-level and high-level representations, resulting in sharper target boundaries and better clutter suppression.
Remotesensing 17 03070 g009
Figure 10. Representative sample of datasets.
Figure 10. Representative sample of datasets.
Remotesensing 17 03070 g010
Figure 11. Detection result of each comparison method in SIRST-Aug (Scene 1) and relative 3-D display.
Figure 11. Detection result of each comparison method in SIRST-Aug (Scene 1) and relative 3-D display.
Remotesensing 17 03070 g011
Figure 12. Detection result of each comparison method in SIRST-Aug (Scene 2) and relative 3-D display.
Figure 12. Detection result of each comparison method in SIRST-Aug (Scene 2) and relative 3-D display.
Remotesensing 17 03070 g012
Figure 13. Detection result of each comparison method in IRSTD-1k (Scene 1) and relative 3-D display.
Figure 13. Detection result of each comparison method in IRSTD-1k (Scene 1) and relative 3-D display.
Remotesensing 17 03070 g013
Figure 14. Detection result of each comparison method in IRSTD-1k (Scene 2) and relative 3-D display.
Figure 14. Detection result of each comparison method in IRSTD-1k (Scene 2) and relative 3-D display.
Remotesensing 17 03070 g014
Table 1. Details of Datasets.
Table 1. Details of Datasets.
DatasetsTraining ImagesTesting ImagesImage SizeSmallest Target Size
SIRST-Aug8525545256 × 2562 × 2
IRSTD-1k800201512 × 5123 × 3
Table 2. Ablation Study on the MIFM.
Table 2. Ablation Study on the MIFM.
MIFMSIRST-AugIRSTD-1k
IoUFmeasureIoUFmeasure
Without I and G0.68610.81380.62170.7713
With I0.72520.84400.65140.7885
With G0.71010.83370.64130.7803
With I and G0.74650.85120.67010.8025
Table 3. Ablation Study on the patch size.
Table 3. Ablation Study on the patch size.
Patch SizeSIRST-AugIRSTD-1k
IoUFmeasureIoUFmeasure
[3]0.69860.82890.62520.7754
[3, 5]0.71540.83650.63260.7791
[3, 5, 6]0.72660.84020.65370.7965
[3, 5, 6, 10]0.74650.85120.67010.8025
Table 4. Isolated and Joint Impact of ACFM on Detection Performance (SIRST-Aug).
Table 4. Isolated and Joint Impact of ACFM on Detection Performance (SIRST-Aug).
MethodPrecisionRecallmIoUF-Measure
w/o ACFM0.80570.81500.70360.8102
w/o ACFM + FIFM0.78210.80380.68020.7928
w/o ACFM + PAB0.79460.79910.68550.7968
w/o ACFM + FIFM + PAB0.78120.79250.68240.7868
Full MFAFNet0.83760.86520.74650.8512
Table 5. Ablation on Loss Weights (SIRST-Aug).
Table 5. Ablation on Loss Weights (SIRST-Aug).
Configurationλ1
(BCE)
λ2
(IoU)
λ3
(Feature)
λ4
(Semantic)
F-MeasuremIoU
Equal Weights (default)11110.85120.7465
Emphasis on BCE21110.83650.7281
Emphasis on IoU12110.84270.7346
Emphasis on Feature Supervision11210.84420.7391
Emphasis on Semantic Alignment11120.84750.7410
Table 6. Comparison with State-of-the-art Methods on SIRST-Aug and IRSTD-1k.
Table 6. Comparison with State-of-the-art Methods on SIRST-Aug and IRSTD-1k.
MethodsSIRST-AugIRSTD-1k
PrecisionRecallmIoUFmeasurePrecisionRecallmIoUFmeasure
LEF0.68370.39290.33250.49900.08320.32480.07090.1324
WSLCM0.93340.10380.10300.18680.54430.09520.08820.1621
RIPT0.94840.08000.07970.14760.57650.05910.05670.1073
PSTNN0.94210.10430.10360.18780.48380.08840.08080.1495
MSLSTIPT0.88100.04000.03980.07660.66020.04700.04580.0877
ACM0.61450.10610.09940.18090.81170.11690.11380.2044
DNANet0.84200.03040.03020.05870.83240.05400.05350.1015
AGPC0.95310.06050.06030.11380.62080.04650.04530.0866
MSHNet0.95750.13250.13170.23280.25940.20720.13020.2303
Trans-IRSD0.96700.07400.07380.13740.60200.10660.09960.1812
RISTDnet0.65950.12290.11550.20720.42360.23520.17820.3024
IRSAM0.80310.84930.70250.82560.71070.79980.60340.7526
STPSA-Net0.83840.75660.66030.79540.76530.71950.58950.7417
MiM-ISTD0.83230.85420.72880.84310.78500.79080.65000.7879
MFAFNet0.83760.86520.74650.85120.79070.81460.67010.8025
Table 7. Summarizes the computational and memory consumption metrics of MFAFNet and other representative methods on the SIRST-Aug dataset, measured on an RTX 4090 GPU.
Table 7. Summarizes the computational and memory consumption metrics of MFAFNet and other representative methods on the SIRST-Aug dataset, measured on an RTX 4090 GPU.
MethodParams (M)FLOPs (G)Inference Time
(ms/img)
GPU Memory (MB)
DNANet9.221.813.61790
Trans-IRSD38.7102.348.23110
MiM-ISTD27.361.537.42685
MFAFNet16.834.924.72163
MFAFNet-Lite
(MobileNetV2)
6.113.711.21420
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, Z.; Chen, W.; Dong, S.; Chen, Y.; Wang, H. MFAFNet: A Multi-Feature Attention Fusion Network for Infrared Small Target Detection. Remote Sens. 2025, 17, 3070. https://doi.org/10.3390/rs17173070

AMA Style

Zhao Z, Chen W, Dong S, Chen Y, Wang H. MFAFNet: A Multi-Feature Attention Fusion Network for Infrared Small Target Detection. Remote Sensing. 2025; 17(17):3070. https://doi.org/10.3390/rs17173070

Chicago/Turabian Style

Zhao, Zehao, Weining Chen, Seng Dong, Yaohong Chen, and Hao Wang. 2025. "MFAFNet: A Multi-Feature Attention Fusion Network for Infrared Small Target Detection" Remote Sensing 17, no. 17: 3070. https://doi.org/10.3390/rs17173070

APA Style

Zhao, Z., Chen, W., Dong, S., Chen, Y., & Wang, H. (2025). MFAFNet: A Multi-Feature Attention Fusion Network for Infrared Small Target Detection. Remote Sensing, 17(17), 3070. https://doi.org/10.3390/rs17173070

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop