Next Article in Journal
Autonomous Cyber-Physical Systems Enabling Smart Positive Energy Districts
Previous Article in Journal
Standardization of Safety Separation for Multi-Category Unmanned Aerial Vehicles in Low-Altitude Airspace Operations
Previous Article in Special Issue
Fourier Ptychographic Microscopy with Optical Aberration Correction and Phase Unwrapping Based on Semi-Supervised Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

LightMFF: A Simple and Efficient Ultra-Lightweight Multi-Focus Image Fusion Network

1
State Key Laboratory of Ocean Sensing & Ocean College, Zhejiang University, Zhoushan 316021, China
2
Hainan Institute, Zhejiang University, Sanya 572025, China
3
Donghai Laboratory, Zhoushan 316021, China
4
Hainan Observation and Research Station of Ecological Environment and Fishery Resource in Yazhou Bay, Sanya 572024, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2025, 15(13), 7500; https://doi.org/10.3390/app15137500
Submission received: 11 June 2025 / Revised: 28 June 2025 / Accepted: 2 July 2025 / Published: 3 July 2025
(This article belongs to the Special Issue Advances in Optical Imaging and Deep Learning)

Abstract

In recent years, deep learning-based multi-focus image fusion (MFF) methods have demonstrated remarkable performance. However, their reliance on complex network architectures often demands substantial computational resources, limiting practical applications. To address this, we propose LightMFF, an ultra-lightweight fusion network that achieves superior performance with minimal computational overhead. Our core insight is to reformulate the multi-focus fusion problem from a classification perspective to a refinement perspective, where coarse initial decision maps and explicit edge information are leveraged to guide the final decision map generation. This novel formulation enables a significantly simplified architecture, requiring only 0.02 M parameters while maintaining state-of-the-art fusion quality. Extensive experiments demonstrate that LightMFF achieves real-time performance at 0.02 s per image pair with merely 0.06 G FLOPs, representing a 98.05% reduction in computational cost compared to prior approaches. Crucially, LightMFF consistently surpasses existing methods across standard fusion quality metrics.

1. Introduction

Multi-focus image fusion (MFF) is a crucial technique in optical imaging that addresses the limited depth of field by integrating multiple source images, each focused at a different depth, into a single all-in-focus representation. By extracting and combining the optimally focused regions, MFF algorithms produce a visually coherent and comprehensive fused image. This approach has broad applications, including microscopy [1], textile quality inspection [2], and autonomous driving [3].
Traditional MFF methods are categorized into two main groups: transform domain-based and spatial domain-based approaches. Transform domain methods, such as multi-scale decomposition and sparse representation, project images into alternative domains for processing [4,5,6,7,8,9]. While effective, these methods often suffer from intensity inconsistencies [10] and noise sensitivity [11]. Spatial domain methods directly process pixels using techniques like block-based [12,13], region-based [14,15], and pixel-level strategies [16,17]. However, these methods typically rely on handcrafted features that lack expressiveness, especially in homogeneous regions [18], and can introduce artifacts along object boundaries [19]. More recent advancements, such as MDLSR_RFM [20], have improved traditional methods by using multi-dictionary linear sparse representation for focus detection and edge-weighting matting for boundary refinement. PADCDTNP [21] introduced a self-adjusting, dual-channel neural system for more accurate focused region identification. Despite these improvements, the increasing complexity of traditional methods leads to greater computational demands.
The advent of deep learning has revolutionized MFF. By leveraging the parallel processing capabilities of modern GPUs, deep learning-based methods have surpassed many traditional approaches in both fusion quality and processing efficiency. These methods learn from large datasets, automating feature extraction and fusion and eliminating the need for complex, handcrafted schemes.
Early deep learning methods framed MFF as a classification task. They used convolutional neural networks (CNNs) to generate a decision map for subsequent fusion [10,19,22,23]. By guiding network optimization with annotated decision maps, these approaches significantly reduced task complexity and improved fusion results within the same computational budget. This fusion paradigm remains a popular research topic due to its interpretability, staged optimization capabilities, and synergy with traditional image processing techniques. More recently, a representative and innovative study introduced ZMFF [11], an unsupervised and untrained deep learning model for MFF that uses a deep image prior and a deep mask prior to effectively fuse images without requiring training data. However, this method requires multiple optimization iterations, leading to very low processing efficiency and limited practical utility.
Subsequently, the focus has shifted to developing general end-to-end MFF frameworks [24,25,26,27,28]. These methods aim to perform the entire fusion task—from feature extraction to reconstruction—within a single network, which eliminates the need for decision map post-processing and better handles complex scenes. However, to enhance fusion quality, some of these newer methods have sacrificed computational efficiency. For example, SwinMFF [26] achieves high pixel-value fidelity but has high computational requirements due to its self-attention mechanism. The multi-step sampling of diffusion model-based methods like FusionDiff [29] places significant strain on the hardware. The resulting high computational demands, energy consumption, and hardware costs severely limit the practical deployment of these models in resource-constrained environments, such as autonomous driving systems and edge devices.
To address the high complexity and computational overhead of end-to-end models, which can cause latency in real-time systems, we propose an ultra-lightweight MFF network with minimal computational overhead. We adopt a decision-map-based fusion paradigm to achieve this goal. Our major contributions are summarized as follows:
  • We introduce a novel task transformation paradigm that reframes MFF from direct focus classification to a decision map refinement process. By providing an initial coarse decision map and explicit edge maps as prior guidance and boundary cues, our approach simplifies network optimization and improves fusion quality.
  • We present ULU-Net, a novel U-Net variant designed specifically for MFF. This architecture achieves competitive performance with an ultra-low parameter count, significantly reducing model complexity.
  • We propose LightMFF, an ultra-lightweight MFF framework that combines our novel task transformation paradigm with the ULU-Net architecture. Extensive quantitative and qualitative experiments demonstrate that LightMFF achieves state-of-the-art fusion quality while requiring significantly fewer parameters and computational resources than existing methods.
The remainder of this paper is organized as follows: Section 2 reviews related works, covering traditional, deep learning-based, and lightweight MFF methods. Section 3 details our proposed LightMFF method, including its theoretical foundations, network architecture, and training strategy. Section 4 presents comprehensive experimental results and comparisons with state-of-the-art methods. Section 5 discusses the limitations and future directions of our work. Finally, Section 6 concludes the paper.

2. Related Works

2.1. Decision Map-Based Deep Learning Methods

Decision map-based deep learning methods have significantly advanced MFF by providing interpretable and precise fusion guidance. This approach typically involves a neural network learning to identify focused or defocused regions, which then informs the final fusion. A notable advantage of this approach is that the models are generally smaller than end-to-end fusion networks. This is because they only need to learn to classify or regress a decision map for each pixel or patch, which is a simpler task than learning to directly synthesize a complete fused image. Consequently, these networks are more efficient and easier to train.
A pioneering work by Liu et al. in 2017 [22] introduced a CNN-based method to generate decision maps through binary patch classification, laying the groundwork for deep learning in MFF. Subsequent efforts enhanced this foundation: ECNN [23] utilizes an ensemble of CNNs for more robust focus detection, while DRPL [19] reframes focus detection as a regression task to learn continuous focus measurements, significantly improving accuracy in boundary regions.
Recent years have seen further innovations in this paradigm. Unsupervised learning frameworks like SESF [30] emerged, eliminating the need for labeled data. Generative adversarial networks (GANs) were leveraged in MFIF-GAN [31] to improve decision map quality. MSFIN [10] introduced multi-scale feature interaction for more precise fusion. ZMFF [11] proposed a zero-shot learning framework using a deep image prior, enhancing generalization to unseen scenes.
Most recently, MFIF-STCU-Net [32] proposed synthesizing training datasets using a depth estimation model and achieved state-of-the-art performance with a multi-level feature fusion (MFF) model that uses a hybrid U-Net and Transformer architecture. DMANet [33] introduced a novel framework that integrates explicit defocus blur modeling into the MFF process, improving both interpretability and performance. LSKN-MFIF [34] is a new MFF network that uses a dynamically adjusting “large selective kernel” module to capture both global and local features, leading to more accurate fusion results that outperform existing methods.

2.2. End-to-End Deep Learning Methods

End-to-end deep learning methods for MFF directly learn the mapping from source images to the fused result, bypassing the explicit generation of a decision map. This paradigm gained significant traction around 2020 with the introduction of IFCNN [24], a general image fusion framework, and U2Fusion [25], a unified unsupervised framework that enhanced training flexibility.
Following these foundational works, subsequent advancements have introduced diverse architectures and learning strategies. In 2021, SDNet [35] achieved real-time performance through a squeeze-and-decomposition network, while MFF-GAN [36] leveraged unsupervised generative adversarial networks to improve visual quality. The year 2022 saw the integration of Transformer-based architectures with SwinFusion [27], which effectively captured long-range dependencies and significantly enhanced fusion performance for complex scenes.
More recent developments have pushed the boundaries of this field. MUFusion [37] introduced a self-evolutionary training formula with a novel memory unit, leveraging intermediate fusion results for supervision. Further explorations include SwinMFF [26] for enhanced pixel-level fidelity and FusionDiff [29], which applies denoising diffusion probabilistic models to MFF. The latest works address specific limitations in the field: DDBFusion [28] improves feature extraction with a dual decomposition approach, and StackMFF [38] extends MFF to image stacks using 3D CNNs and synthesized training data. Additionally, MMAE [39] introduced a mask attention mechanism to filter redundant information, and LFDT-Fusion [40] proposed an efficient latent feature-guided diffusion model that operates in a compressed latent space.
However, a core challenge for end-to-end methods is their inherent demand for complex networks and substantial computational resources. Unlike decision map-based approaches that decompose the fusion problem, these models must simultaneously perform both feature extraction and image reconstruction, a significantly more complex task. This makes them less suitable for lightweight applications. For instance, Transformer-based approaches like SwinMFF [26] and SwinFusion [27] often require tens of millions of parameters. Similarly, diffusion-based methods such as FusionDiff [29] typically demand extensive computational power due to their iterative nature. Even relatively compact architectures like IFCNN [24] require a considerable number of parameters to maintain fusion quality. This fundamental limitation stems directly from the end-to-end paradigm, where generating a pixel-perfect fusion result from source images is an inherently complex task, posing significant challenges for ultra-lightweight deployments. For resource-constrained scenarios, decision map-based methods offer a more promising direction by decomposing the fusion problem into more manageable sub-tasks.

2.3. Lightweight Multi-Focus Image Fusion Networks

While deep learning-based MFF approaches have shown great promise, their often complex architectures can be computationally demanding, hindering practical application. This has led to an increasing focus on developing lightweight MFF networks. For example, Jin et al. [41] proposed a lightweight scheme that leverages the Laplacian pyramid transform and an adaptive pulse-coupled neural network with local spatial frequency. This approach aimed to improve fusion efficiency and performance by processing fewer sub-images. Similarly, Zhou et al. [42] introduced LNMF. By classifying in-focus and out-of-focus regions, LNMF significantly reduces computational complexity and memory consumption. More recently, Nie et al. [43] presented MLNet, a multi-domain lightweight network. MLNet employs discrete cosine transform-based and local binary pattern-based convolutions to effectively capture both frequency and spatial domain features with a minimal parameter count.
In the field of multi-modal image fusion, which is closely related to the field of MFF, there are also works focusing on lightweight fusion networks. For example, Wu et al. [44] presented a lightweight image fusion algorithm specifically designed for merging visible light and infrared images by utilizing Depthwise Separable Convolution. In the field of medical image fusion, MACTFusion [45] proposed a novel lightweight cross Transformer based on a cross multi-axis attention mechanism. They mainly perform as lightweight on the network, and the proposed method not only performs as lightweight on the network but also indirectly allows the use of a lighter network through the transformation of the fusion paradigm.
Building upon these efforts, our proposed LightMFF also focuses on an ultra-lightweight architecture. Figure 1 presents a comparison between our proposed method and several prominent learning-based existing MFF methods, evaluating both fusion quality and processing efficiency. To the best of our knowledge, the proposed LightMFF is currently the best method with the lowest computational requirements while maintaining good overall performance.

3. Method

3.1. Motivation

The primary objective of decision map-based MFF is to generate a binary decision map that accurately reflects the focus properties of source images. Conventionally, networks are trained to distinguish between focused and defocused regions through loss optimization. We propose a novel paradigm that transforms this learning objective by introducing an initial coarse decision map as prior guidance. This approach shifts the network’s task from direct focus/defocus classification to decision map refinement, where the network learns to optimize an imperfect decision map using the source images as a reference. This task transformation significantly reduces the learning complexity and, consequently, the required model capacity. Furthermore, by incorporating explicit edge maps as additional input, we provide crucial boundary cues that further simplify the optimization process, enabling an even more compact network architecture. The proposed fusion framework is illustrated in Figure 2.

3.2. Focus Property and Edge Detection

To provide prior guidance through an initial coarse decision map, we adopt the pixel-level spatial frequency gradient, as proposed in SESF-Fuse [30], as our initial focus measure. The computational formulation is as follows:
RF i ( x , y ) = a = r r b = r r I i ( x + a , y + b ) I i ( x + a , y + b 1 ) 2
CF i ( x , y ) = a = r r b = r r I i ( x + a , y + b ) I i ( x + a 1 , y + b ) 2
SF i ( x , y ) = CF i ( x , y ) 2 + RF i ( x , y ) 2 ( 2 r + 1 ) 2
where RF and CF represent the row and column frequency vectors, respectively, and r denotes the kernel radius. By computing the pixel-level spatial frequency for source images I i , where i { A , B } , we obtain their respective spatial frequency maps SF i . The initial decision maps IDM A and IDM B are then generated through pixel-wise comparison of SF A and SF B :
IDM A ( x , y ) = 1 , if   SF A ( x , y ) SF B ( x , y ) 0 , otherwise
IDM B ( x , y ) = 1 IDM A ( x , y )
Additionally, to provide explicit boundary information and facilitate the refinement of decision maps, especially in edge regions, we employ the Sobel operators [47] for edge detection. For source images I A and I B , the edge maps Edge A and Edge B are computed using horizontal and vertical Sobel kernels:
S x = 1 0 1 2 0 2 1 0 1 , S y = 1 2 1 0 0 0 1 2 1
Edge i = ( I i S x ) 2 + ( I i S y ) 2 , i { A , B }
where ∗ denotes the convolution operation. The edge maps highlight structural boundaries in the source images by computing the magnitude of gradients in both horizontal and vertical directions, providing crucial guidance for decision map refinement in regions where focus properties transition.

3.3. Network Architecture

We propose ULU-Net, an ultra-lightweight U-Net variant specifically designed for MFF, as shown in Figure 3. The network employs an encoder-decoder architecture that takes source images, initial decision maps, and edge maps as input to generate refined decision maps. Unlike conventional U-Net architectures that typically use large channel numbers (e.g., 64→128→256→512), we carefully design a minimal channel progression (8→16→24→32→48→64) that significantly reduces parameters while maintaining essential feature extraction capability. The encoder consists of six stages with this progressive channel expansion, while the decoder has five stages for feature reconstruction. To further enhance parameter efficiency, we employ element-wise addition instead of channel concatenation for skip connections, preventing channel number expansion and thus reducing the parameters in subsequent layers.
The input layer concatenates source images ( I A , I B ), initial decision maps ( IDM A , IDM B ), and edge maps ( Edge A , Edge B ) into a multi-channel tensor:
X input = Concat ( I A , I B , IDM A , IDM B , Edge A , Edge B )
To achieve parameter efficiency while maintaining feature extraction capability, we introduce InceptionDWConv2d [48], a lightweight module inspired by the Inception architecture. The module splits input channels with a branching ratio of 0.125 and processes them through parallel pathways:
Y out = Concat ( X identity , D W C o n v 3 × 3 ( X branch1 ) ,             D W C o n v 1 × 11 ( X branch2 ) , D W C o n v 11 × 1 ( X branch3 ) )
where D W C o n v represents depthwise convolution with specified kernel sizes. This design effectively captures multi-scale and directional features while minimizing computational cost.
The entire network contains only 15K parameters, which is merely 1/200 of MobileNetV3’s parameter count [49], making it remarkably lightweight compared to existing methods. This ultra-compact design is achieved without sacrificing performance through our careful architectural choices and the effective utilization of focus property and edge information as prior guidance.

3.4. Loss Function

To effectively train our lightweight network for decision map refinement, we propose a combined loss function that leverages two complementary components: Binary Focal Loss (BFL) [50] and Binary Dice Loss (BDL) [51]. The total loss L total is defined as:
L total = λ 1 L BFL + λ 2 L BDL
where λ 1 = 15 and λ 2 = 1 are the weighting coefficients that balance the contribution of each loss term. In our context, BFL helps the network focus on hard-to-classify regions. Formally, BFL is defined as:
L BFL = α y ( 1 p ) γ log ( p ) ( 1 α ) ( 1 y ) p γ log ( 1 p )
where p represents the predicted probability, y is the ground truth label, α is the balancing factor (set to 0.25), and γ is the focusing parameter (set to 2.0). The focal loss introduces modulating factors ( 1 p ) γ for positive samples and p γ for negative samples. This automatically down-weights the contribution of easy examples and focuses on hard examples during training. This property is particularly beneficial for our task, as it helps the network concentrate on refining the decision boundaries where the focus state is ambiguous.
To complement the focal loss, we incorporate the Binary Dice Loss, which is particularly effective for handling imbalanced segmentation tasks. BDL measures the overlap between predicted and ground truth regions:
L BDL = 1 2 i = 1 N p i y i + ϵ i = 1 N p i 2 + i = 1 N y i 2 + ϵ
where p i and y i are the predicted and ground truth values at position i, N is the total number of pixels, and ϵ is a small constant to ensure numerical stability. The Binary Dice Loss is particularly advantageous for our decision map refinement task, as it naturally optimizes for spatial overlap between predicted and ground truth regions.

3.5. Fusion Scheme

Figure 4 illustrates the complete pipeline from the input image pair to the final fused result. Given a pair of source images I A and I B , along with their coarse initial decision maps and edge maps as inputs, the proposed ULU-Net first generates a continuous-valued decision map, Output. To obtain more robust fusion results, we process this raw output through several stages:
First, we apply a binarization operation to convert the continuous-valued Output to a binary decision map, Output’:
Output ( x , y ) = 1 , if   Output ( x , y ) T 0 , otherwise
where T is the binarization threshold, empirically set to 0.5.
Next, to ensure spatial consistency and eliminate isolated noise, we perform morphological operations on Output’ to obtain the final decision map, DM :
DM = δ ( ϵ ( Output ) )
where ϵ ( · ) and δ ( · ) denote the morphological erosion and dilation operations, respectively, with a 5 × 5 structuring element.
Finally, the fused image I F is generated by combining the source images according to the refined decision map:
I F ( x , y ) = DM ( x , y ) · I A ( x , y ) + ( 1 DM ( x , y ) ) · I B ( x , y )

4. Experiments

4.1. Experimental Setup

Training Strategy. We use the DUTS dataset (https://saliencydetection.net/duts/) accessed on 2 June 2025 [52], which comprises 15,572 carefully curated images. This dataset is divided into 10,553 training samples and 5019 validation samples, each resized to a resolution of 256 × 256. To generate the training pairs, we first convert the ground truth annotations into binary masks. These masks then guide the application of variable Gaussian blur to specific regions within the source images, using kernel sizes that range from 3 to 21. This process yields realistic multi-focus image pairs. The detailed procedure for generating these training data can be found in SwinMFF [26]. For network optimization, we utilized the AdamW optimizer, initialized with a learning rate of 1 × 10−3. A CosineAnnealingLR scheduler was implemented to progressively reduce the learning rate over the course of training. The model was trained for 200 epochs with a batch size of 32. All training was performed on a high-performance computing platform equipped with dual Nvidia A6000 GPUs, supported by an Intel(R) Xeon(R) Platinum 8375C CPU operating at 2.90 GHz. The entire framework was developed in PyTorch (version 2.4.1).
Datasets for Evaluation. To rigorously assess LightMFF’s performance, we conducted extensive experiments using three widely recognized MFF benchmarks: Lytro [46], MFFW [53], and MFI-WHU [36]. The Lytro dataset consists of 20 image pairs captured with a Lytro light-field camera and was utilized for both qualitative and quantitative evaluation. The MFFW dataset includes 13 image pairs notable for their pronounced defocus spread effects, allowing for qualitative evaluation under challenging focus conditions. Extending our evaluation scope, the MFI-WHU dataset provides 120 synthetically generated image pairs, created using Gaussian blur, which encompass diverse scenarios, including high-contrast scenes, and was also used for qualitative assessment.
Methods for Comparison. To thoroughly evaluate the efficacy of the proposed LightMFF method, we performed extensive comparisons against state-of-the-art image fusion approaches across four primary categories. For transform domain methods, our comparisons included DWT [7], DTCWT [54], NSCT [8], CVT [55], GFF [56], SR [57], ASR [58], MWGF [59], ICA [60], and NSCT-SR [61]. In the spatial domain category, we evaluated SSSDI [62], QUADTREE [63], DSIFT [61], SRCF [46], GFDF [64], BRW [65], MISF [66], and MDLSR_RFM [20]. Regarding end-to-end deep learning approaches, we included IFCNN-MAX [24], U2Fusion [25], SDNet [35], MFF-GAN [36], SwinFusion [27], MUFusion [37], FusionDiff [29], SwinMFF [26], and DDBFusion [28]. Additionally, we compared with decision map-based deep learning methods such as CNN [22], ECNN [23], DRPL [19], SESF [30], MFIF-GAN [31], MSFIN [10], GACN [67], and ZMFF [11]. For traditional methods, we either employed the default parameters as specified in their original implementations or directly used the fusion results made publicly available by their respective authors. For deep learning-based methods, we utilized the pre-trained models provided by their authors to ensure a fair comparison. This comprehensive comparison of 37 methods across four distinct categories allows for a thorough assessment of the proposed method’s capabilities. It is important to note that while some related work discusses lightweight networks, we could not include them in our comparisons due to the unavailability of their codebases. Nevertheless, many highly compact networks, such as MFF-GAN [36] and SESF [30], are already featured in our evaluation.
Evaluation metrics. To comprehensively evaluate the performance of our method, we employ six complementary metrics that assess different aspects of the fusion results: Edge-based Similarity Metric Q A B / F  [68], Normalized Mutual Information Metric Q M I  [69], Phase Congruency-based Metric Q P  [70], Structural Similarity Metrics Q W and Q E  [71], and Human Perception-inspired Metric Q C B  [72]. For all metrics, higher values (↑) indicate better fusion performance. These metrics collectively provide a comprehensive evaluation framework, covering edge preservation, information content, structural similarity, and perceptual quality of the fusion results.

4.2. Experimental Results

Qualitative comparison. The decision maps offer more direct insights into a network’s performance than the fused images themselves. For qualitative comparisons of decision maps, we exclusively present the initial decision maps generated by each network, which helps us avoid variations that might come from different post-processing techniques. We have visualized these decision maps, which are responsible for generating the fusion results, in Figure 5. Visual inspection suggests that LightMFF produces the most visually compelling decision maps among all the approaches we compared.
Compared to the Lytro dataset [46], the MFFW dataset [53] offers more challenging scenarios due to its more pronounced defocus dispersion effects. To further validate our approach under these demanding conditions, we conducted additional visual comparisons on this dataset, as shown in Figure 6. The comparison reveals that most methods, with the exceptions of our proposed LightMFF and MFIF-GAN [31], show noticeably degraded performance on this dataset when compared to their results on the Lytro dataset. Several approaches struggle with large-area misclassification, which severely impacts the quality of their fusion results. In contrast, LightMFF maintains exceptional performance, consistently producing the most accurate decision maps among all compared methods.
In Figure 7, we compared the fusion results of different methods using the “coffee cup” example from the MFFW dataset. As the figure illustrates, particularly in the boundary regions of the coffee cup, only ECNN [23] and our proposed LightMFF accurately differentiate between focused and defocused areas along the cup edges. Both ECNN and LightMFF achieve clean fusion results without notable artifacts. In contrast, other approaches exhibit visible fusion artifacts in these challenging regions. This robust performance demonstrates LightMFF’s superior capability in handling scenes with strong defocus effects.
Additionally, we conducted qualitative comparisons on representative image pairs from the MFI-WHU dataset [36]. The corresponding decision maps are shown in Figure 8. The selected cases represent three challenging scenarios: small object detection (first row), high-contrast scenes (second row), and large homogeneous regions (third and fourth rows). The proposed LightMFF consistently outperforms all competing methods across all test cases, demonstrating superior accuracy in every scenario.
Quantitative comparison. We conduct comprehensive quantitative evaluations on the widely-used Lytro dataset [46] to assess our method’s performance. Table 1 thoroughly compares our approach against 36 other methods, highlighting the best-performing methods in bold and the second-best in underline. For each metric, an upward arrow indicates that a higher value is better. Methods highlighted with a background color represent our proposed approach or the complete proposed model.As shown in Table 1, the proposed LightMFF achieves state-of-the-art performance across most metrics. Specifically, LightMFF obtains the highest scores in five out of six metrics: Q A B / F (0.7588), Q M I (1.1462), Q P (0.8450), Q E (0.9061), and Q C B (0.8067). For the remaining metric Q W , our method achieves the second-best score (0.9400), only marginally lower than GACN (0.9405) [67] by 0.0005.
Efficiency Analysis. We conduct a comprehensive efficiency analysis comparing LightMFF against state-of-the-art methods across model size, computational complexity (FLOPs), and inference time, as detailed in Table 2. The inference time is measured on the MFI-WHU dataset [36], representing the average time required to process an image pair.
The proposed LightMFF achieves remarkable efficiency improvements across all metrics. With only 0.02 M parameters, LightMFF is approximately 2.5× smaller than its closest end-to-end competitor (MFF-GAN [36] with 0.05M parameters) and 3.5× smaller than the most compact decision map-based method (SESF [30]/GACN [67] with 0.07 M parameters). In terms of computational complexity, LightMFF requires merely 0.06 G FLOPs, representing a 98.05% reduction in computational cost compared to the most compact prior method (MFF-GAN [36] with 3.08 G FLOPs). The practical benefits of our efficient design are evident in the inference time measurements, where LightMFF executes in 0.02 s on GPU, achieving a 3× speedup over MFF-GAN (0.06 s) [36] and an 8× improvement over GACN (0.16 s) [67].

4.3. Performance Under Extreme Conditions

To validate the robustness of the proposed method under extreme conditions where input image pairs consist of completely sharp and completely blurred images, we conducted a challenging experiment using images from the MFI-WHU dataset [36], as illustrated in Figure 9. For each image, we synthesized a corresponding completely blurred version using Gaussian blur kernels.
The paired images, comprising the original sharp image and its completely blurred counterpart, were then fed into the proposed model to generate the decision maps and fusion results shown in the figure. Remarkably, the proposed model maintains excellent performance even under these extreme conditions, successfully distinguishing between defocused and focused regions while producing accurate decision maps for effective image fusion. This demonstrates the model’s robustness and reliability across diverse imaging conditions.

4.4. Performance in Real-World Scenarios

Motivated by findings in SAMF [3] that MFF can improve object detection accuracy in autonomous driving systems, we evaluated our LightMFF network on the Road-MF dataset [3]. This dataset contains 80 pairs of real-world multi-focus images captured in diverse road scenes, making it an ideal choice for assessing our method’s practical applicability.
As shown in Figure 10, our proposed method performs better than DRPL [19] in real-world road scenes. LightMFF notably shows fewer misclassifications in both sky regions and with objects closer to the ground. Even with the inherent challenges of real-world scenes, LightMFF achieves exceptional performance while remaining computationally efficient. This dual advantage highlights the practical value of our lightweight architecture, especially for real-world applications like autonomous driving systems where both accuracy and real-time processing are crucial.

4.5. Performance Under Image Degradation

To investigate the performance of the proposed model when input images suffer from various degradation conditions, we present a qualitative comparison in Figure 11 using the same input image pair under different scenarios, including variations in contrast and brightness, as well as corruption by Gaussian noise and salt-and-pepper noise.
The experimental results demonstrate that the proposed model exhibits remarkable robustness to brightness and contrast variations, with virtually no impact on fusion performance. However, the proposed method lacks robustness against high-frequency noise. When input images are corrupted by Gaussian noise or salt-and-pepper noise, the high-frequency noise components significantly degrade the quality of the initial decision map, thereby adversely affecting the fusion results. Nevertheless, as illustrated in Figure 11h,i, this issue can be effectively mitigated through simple filtering operations.

4.6. Ablation Study

To validate the effectiveness of our proposed components and design choices, we conduct comprehensive ablation studies on the Lytro dataset [46]. Specifically, we investigate how the initial decision map prior and edge information contribute to the network’s performance. We design four experimental settings:
  • Group 1 (Baseline): Only source images are fed into the network, representing the conventional approach where the network directly learns to distinguish focused from defocused regions.
  • Group 2 (w/Decision Map): Source images and the initial decision map are used as input. This setting evaluates the effectiveness of our proposed paradigm shift from direct classification to decision map refinement.
  • Group 3 (w/Edge Map): Source images are complemented with edge maps. This setting examines the impact of explicit boundary information without decision map guidance.
  • Group 4 (Full Model): Our complete model, incorporating both the initial decision map and edge information, demonstrates the synergistic effect of all components.
Table 3 and Figure 12 showcase the results of our ablation study. The baseline model (Group 1) performed less effectively, highlighting the inherent difficulties of directly classifying focus/defocus regions with an ultra-lightweight architecture. Adding the initial decision map (Group 2) significantly boosted performance. This confirms our hypothesis that reframing the learning objective to decision map refinement genuinely reduces task complexity. The edge information (Group 3) also brought noticeable improvements over the baseline, affirming that explicit boundary cues help the network better identify focus transitions. These ablation studies confirm that our task transformation approach and targeted guidance signals successfully reduce computational complexity while maintaining superior multi-focus fusion (MFF) performance.

5. Discussion

Despite the demonstrated effectiveness of our proposed method in MFF, we have identified three primary limitations:
First, our decision map-based paradigm, while computationally efficient, might not perform as robustly as end-to-end approaches in handling complex scenes [26].
Second, our current post-processing strategy relies on basic morphological operations to refine the decision map. More advanced techniques, such as image matting-based methods [73], can indeed effectively reduce artifacts along fusion boundaries. However, they introduce substantial computational overhead—typically adding 2–3 s per image pair to processing time. This additional cost conflicts with our goal of maintaining computational efficiency. Therefore, we chose lightweight morphological operations as a compromise. While our current post-processing approach generally improves fusion quality, it may occasionally yield suboptimal results in complex local regions with fine details, as illustrated in Figure 13.
Third, our method is susceptible to high-frequency noise. Because the initial decision map is generated based on a focus measure that is sensitive to local variations, the presence of high-frequency noise in the source images can lead to an erroneous initial decision map, thereby further affecting the final decision map. This sensitivity highlights a need for more robust focus measure mechanisms that are less influenced by noise, or for a pre-processing step to denoise the source images.
Our work opens several promising directions for future research. We could investigate more sophisticated yet efficient post-processing techniques to further enhance fusion quality in fine-detail regions. Additionally, developing a more noise-robust focus measure or integrating a denoising module could significantly improve the reliability of the decision map generation. Extending our lightweight architecture design principles to other image fusion tasks (e.g., multi-exposure, infrared-visible) might lead to a more general framework for efficient image fusion. Finally, exploring the integration of our method with specific downstream tasks like object detection could significantly boost its practical utility in real-world applications.

6. Conclusions

In this paper, we have presented LightMFF, an ultra-lightweight network for MFF. The key contributions are as follows: (1) proposing a novel task transformation paradigm that reformulates multi-focus fusion from focus classification to decision map refinement, which significantly reduces model complexity while maintaining performance; (2) introducing an ultra-lightweight architecture that achieves state-of-the-art performance with only 0.02 M parameters and 0.06 G FLOPs, which represents a 98.05% reduction in computational cost compared to prior methods; and (3) demonstrating through extensive experiments that our method achieves competitive or superior performance on standard benchmarks while maintaining real-time inference capability (0.02 s per image pair).
The proposed LightMFF demonstrates that it is possible to achieve state-of-the-art performance in MFF with significantly reduced computational resources, challenging the conventional wisdom that deeper and more complex networks are necessary for superior performance. Our approach’s efficiency makes it particularly suitable for resource-constrained scenarios such as mobile devices, embedded systems, and real-time applications. We believe that this work represents a significant step toward making advanced MFF techniques more accessible and practical for real-world deployment.

Author Contributions

Conceptualization, X.X., Z.L. and B.G.; methodology, X.X., Z.L., B.G., and S.H.; software, X.X. and Z.L.; validation, X.X., B.G., and Y.G.; formal analysis, S.H. and Y.G.; investigation, B.G., S.H., and Y.B.; resources, S.H., Y.G., and P.L.; data curation, X.X.; writing—original draft preparation, X.X. and Z.L.; writing—review and editing, X.X., Z.L., B.G., S.H., and Y.B.; visualization, X.X.; supervision, P.L.; project administration, P.L. and Y.B.; funding acquisition, B.G., S.H., and P.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Key R&D Program of China (No. 2022-34), the Project of Sanya Yazhou Bay Science and Technology City (No. SCKJ-JYRC-2023-59), Innovational Fund for Scientific and Technological Personnel of Hainan Province (NO. KJRC2023D19) and the Research Startup Funding from Hainan Institute of Zhejiang University (No. 0208-6602-A12204). The authors would like to thank Hainan Observation and Research Station of Ecological Environment and Fishery Resource in Yazhou Bay for providing computing resources for this research.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The code of this study will be publicly accessible in a GitHub repository at https://github.com/Xinzhe99/LightMFF (accessed on 2 June 2025).

Acknowledgments

The authors would like to thank Hainan Observation and Research Station of Ecological Environment and Fishery Resource in Yazhou Bay for providing computing resources for this research.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study, in the collection, analyses, or interpretation of data, in the writing of the manuscript, or in the decision to publish the results.

References

  1. Xie, X.; Guo, B.; Li, P.; Jiang, Q. Underwater Three-Dimensional Microscope for Marine Benthic Organism Monitoring. In Proceedings of the OCEANS 2024-Singapore, Singapore, 15–18 April 2024; pp. 1–4. [Google Scholar]
  2. Chen, Y.; Deng, N.; Xin, B.J.; Xing, W.Y.; Zhang, Z.Y. Structural characterization and measurement of nonwoven fabrics based on multi-focus image fusion. Measurement 2019, 141, 356–363. [Google Scholar] [CrossRef]
  3. Li, X.; Li, X.; Tan, H.; Li, J. SAMF: Small-area-aware multi-focus image fusion for object detection. In Proceedings of the ICASSP 2024–2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seoul, Republic of Korea, 14–19 April 2024; pp. 3845–3849. [Google Scholar]
  4. Burt, P.J.; Adelson, E.H. The Laplacian pyramid as a compact image code. In Readings in Computer Vision; Elsevier: Amsterdam, The Netherlands, 1987; pp. 671–679. [Google Scholar]
  5. Burt, P.J.; Kolczynski, R.J. Enhanced image capture through fusion. In Proceedings of the 1993 (4th) International Conference on Computer Vision, Berlin, Germany, 11–14 May 1993; pp. 173–182. [Google Scholar]
  6. Lewis, J.J.; O’Callaghan, R.J.; Nikolov, S.G.; Bull, D.R.; Canagarajah, N. Pixel-and region-based image fusion with complex wavelets. Inf. Fusion 2007, 8, 119–130. [Google Scholar] [CrossRef]
  7. Li, H.; Manjunath, B.; Mitra, S.K. Multisensor image fusion using the wavelet transform. Graph. Model. Image Process. 1995, 57, 235–245. [Google Scholar] [CrossRef]
  8. Yang, B.; Li, S.; Sun, F. Image fusion using nonsubsampled contourlet transform. In Proceedings of the Fourth International Conference on Image and Graphics (ICIG 2007), Chengdu, China, 22–24 August 2007; pp. 719–724. [Google Scholar]
  9. Zhang, Q.; Guo, B.L. Multifocus image fusion using the nonsubsampled contourlet transform. Signal Process. 2009, 89, 1334–1346. [Google Scholar] [CrossRef]
  10. Liu, Y.; Wang, L.; Cheng, J.; Chen, X. Multiscale feature interactive network for multifocus image fusion. IEEE Trans. Instrum. Meas. 2021, 70, 1–16. [Google Scholar] [CrossRef]
  11. Hu, X.; Jiang, J.; Liu, X.; Ma, J. ZMFF: Zero-shot multi-focus image fusion. Inf. Fusion 2023, 92, 127–138. [Google Scholar] [CrossRef]
  12. Sujatha, K.; Shalini Punithavathani, D. Optimized ensemble decision-based multi-focus imagefusion using binary genetic Grey-Wolf optimizer in camera sensor networks. Multimed. Tools Appl. 2018, 77, 1735–1759. [Google Scholar] [CrossRef]
  13. Kausar, N.; Majid, A.; Javed, S.G. A novel ensemble approach using individual features for multi-focus image fusion. Comput. Electr. Eng. 2016, 54, 393–405. [Google Scholar] [CrossRef]
  14. Huang, Y.; Li, W.; Gao, M.; Liu, Z. Algebraic multi-grid based multi-focus image fusion using watershed algorithm. IEEE Access 2018, 6, 47082–47091. [Google Scholar] [CrossRef]
  15. Duan, J.; Chen, L.; Chen, C.P. Multifocus image fusion with enhanced linear spectral clustering and fast depth map estimation. Neurocomputing 2018, 318, 43–54. [Google Scholar] [CrossRef]
  16. Jagtap, N.; Thepade, S. High-quality image multi-focus fusion to address ringing and blurring artifacts without loss of information. Vis. Comput. 2022, 38, 4353–4371. [Google Scholar] [CrossRef]
  17. Kong, W.; Lei, Y. Multi-focus image fusion using biochemical ion exchange model. Appl. Soft Comput. 2017, 51, 314–327. [Google Scholar] [CrossRef]
  18. Duan, Z.; Luo, X.; Zhang, T. Combining transformers with CNN for multi-focus image fusion. Expert Syst. Appl. 2024, 235, 121156. [Google Scholar] [CrossRef]
  19. Li, J.; Guo, X.; Lu, G.; Zhang, B.; Xu, Y.; Wu, F.; Zhang, D. DRPL: Deep regression pair learning for multi-focus image fusion. IEEE Trans. Image Process. 2020, 29, 4816–4831. [Google Scholar] [CrossRef]
  20. Wang, J.; Qu, H.; Zhang, Z.; Xie, M. New insights into multi-focus image fusion: A fusion method based on multi-dictionary linear sparse representation and region fusion model. Inf. Fusion 2024, 105, 102230. [Google Scholar] [CrossRef]
  21. Li, B.; Zhang, L.; Liu, J.; Peng, H.; Wang, Q.; Liu, J. Multi-focus image fusion with parameter adaptive dual channel dynamic threshold neural P systems. Neural Netw. 2024, 179, 106603. [Google Scholar] [CrossRef]
  22. Liu, Y.; Chen, X.; Peng, H.; Wang, Z. Multi-focus image fusion with a deep convolutional neural network. Inf. Fusion 2017, 36, 191–207. [Google Scholar] [CrossRef]
  23. Amin-Naji, M.; Aghagolzadeh, A.; Ezoji, M. Ensemble of CNN for multi-focus image fusion. Inf. Fusion 2019, 51, 201–214. [Google Scholar] [CrossRef]
  24. Zhang, Y.; Liu, Y.; Sun, P.; Yan, H.; Zhao, X.; Zhang, L. IFCNN: A general image fusion framework based on convolutional neural network. Inf. Fusion 2020, 54, 99–118. [Google Scholar] [CrossRef]
  25. Xu, H.; Ma, J.; Jiang, J.; Guo, X.; Ling, H. U2Fusion: A unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 502–518. [Google Scholar] [CrossRef]
  26. Xie, X.; Guo, B.; Li, P.; He, S.; Zhou, S. SwinMFF: Toward high-fidelity end-to-end multi-focus image fusion via swin transformer-based network. Vis. Comput. 2024, 41, 1–24. [Google Scholar] [CrossRef]
  27. Ma, J.; Tang, L.; Fan, F.; Huang, J.; Mei, X.; Ma, Y. SwinFusion: Cross-domain long-range learning for general image fusion via swin transformer. IEEE/CAA J. Autom. Sin. 2022, 9, 1200–1217. [Google Scholar] [CrossRef]
  28. Zhang, Z.; Li, H.; Xu, T.; Wu, X.J.; Kittler, J. DDBFusion: An unified image decomposition and fusion framework based on dual decomposition and Bézier curves. Inf. Fusion 2025, 114, 102655. [Google Scholar] [CrossRef]
  29. Li, M.; Pei, R.; Zheng, T.; Zhang, Y.; Fu, W. FusionDiff: Multi-focus image fusion using denoising diffusion probabilistic models. Expert Syst. Appl. 2024, 238, 121664. [Google Scholar] [CrossRef]
  30. Ma, B.; Zhu, Y.; Yin, X.; Ban, X.; Huang, H.; Mukeshimana, M. Sesf-fuse: An unsupervised deep model for multi-focus image fusion. Neural Comput. Appl. 2021, 33, 5793–5804. [Google Scholar] [CrossRef]
  31. Wang, Y.; Xu, S.; Liu, J.; Zhao, Z.; Zhang, C.; Zhang, J. MFIF-GAN: A new generative adversarial network for multi-focus image fusion. Signal Process. Image Commun. 2021, 96, 116295. [Google Scholar] [CrossRef]
  32. Shao, X.; Jin, X.; Jiang, Q.; Miao, S.; Wang, P.; Chu, X. Multi-focus image fusion based on transformer and depth information learning. Comput. Electr. Eng. 2024, 119, 109629. [Google Scholar] [CrossRef]
  33. Quan, Y.; Wan, X.; Tang, Z.; Liang, J.; Ji, H. Multi-Focus Image Fusion via Explicit Defocus Blur Modelling. Proc. Aaai Conf. Artif. Intell. 2025, 39, 6657–6665. [Google Scholar] [CrossRef]
  34. Zhai, H.; Zhang, G.; Zeng, Z.; Xu, Z.; Fang, A. LSKN-MFIF: Large selective kernel network for multi-focus image fusion. Neurocomputing 2025, 635, 129984. [Google Scholar] [CrossRef]
  35. Zhang, H.; Ma, J. SDNet: A versatile squeeze-and-decomposition network for real-time image fusion. Int. J. Comput. Vis. 2021, 129, 2761–2785. [Google Scholar] [CrossRef]
  36. Zhang, H.; Le, Z.; Shao, Z.; Xu, H.; Ma, J. MFF-GAN: An unsupervised generative adversarial network with adaptive and gradient joint constraints for multi-focus image fusion. Inf. Fusion 2021, 66, 40–53. [Google Scholar] [CrossRef]
  37. Cheng, C.; Xu, T.; Wu, X.J. MUFusion: A general unsupervised image fusion network based on memory unit. Inf. Fusion 2023, 92, 80–92. [Google Scholar] [CrossRef]
  38. Xie, X.; Qingyan, J.; Chen, D.; Guo, B.; Li, P.; Zhou, S. StackMFF: End-to-end multi-focus image stack fusion network. Appl. Intell. 2025, 55, 503. [Google Scholar] [CrossRef]
  39. Wang, X.; Fang, L.; Zhao, J.; Pan, Z.; Li, H.; Li, Y. MMAE: A universal image fusion method via mask attention mechanism. Pattern Recognit. 2025, 158, 111041. [Google Scholar] [CrossRef]
  40. Yang, B.; Jiang, Z.; Pan, D.; Yu, H.; Gui, G.; Gui, W. LFDT-Fusion: A latent feature-guided diffusion Transformer model for general image fusion. Inf. Fusion 2025, 113, 102639. [Google Scholar] [CrossRef]
  41. Jin, X.; Hou, J.; Nie, R.; Yao, S.; Zhou, D.; Jiang, Q.; He, K. A lightweight scheme for multi-focus image fusion. Multimed. Tools Appl. 2018, 77, 23501–23527. [Google Scholar] [CrossRef]
  42. Zhou, Y.; Liu, K.; Dou, Q.; Liu, Z.; Jeon, G.; Yang, X. LNMF: Lightweight network for multi-focus image fusion. Multimed. Tools Appl. 2022, 81, 22335–22353. [Google Scholar] [CrossRef]
  43. Nie, X.; Hu, B.; Gao, X. MLNet: A multi-domain lightweight network for multi-focus image fusion. IEEE Trans. Multimed. 2022, 25, 5565–5579. [Google Scholar] [CrossRef]
  44. Wu, Z.; Chen, J.; Tan, L.; Gong, H.; Zhou, Y.; Shi, G. A lightweight GAN-based image fusion algorithm for visible and infrared images. In Proceedings of the 2024 4th International Conference on Computer Science and Blockchain (CCSB), Shenzhen, China, 6–8 September 2024; pp. 466–470. [Google Scholar]
  45. Xie, X.; Zhang, X.; Tang, X.; Zhao, J.; Xiong, D.; Ouyang, L.; Yang, B.; Zhou, H.; Ling, B.W.K.; Teo, K.L. MACTFusion: Lightweight cross transformer for adaptive multimodal medical image fusion. IEEE J. Biomed. Health Inform. 2024, 29, 3317–3328. [Google Scholar] [CrossRef]
  46. Nejati, M.; Samavi, S.; Shirani, S. Multi-focus image fusion using dictionary-based sparse representation. Inf. Fusion 2015, 25, 72–84. [Google Scholar] [CrossRef]
  47. Sobel, I.; Feldman, G. A 3x3 isotropic gradient operator for image processing. Stanf. Artif. Intell. Proj. (Sail) 1968, 1968, 271–272. [Google Scholar]
  48. Yu, W.; Zhou, P.; Yan, S.; Wang, X. Inceptionnext: When inception meets convnext. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 5672–5683. [Google Scholar]
  49. Howard, A.; Sandler, M.; Chu, G.; Chen, L.C.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudevan, V.; et al. Searching for mobilenetv3. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1314–1324. [Google Scholar]
  50. Ross, T.Y.; Dollár, G. Focal loss for dense object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2980–2988. [Google Scholar]
  51. Sudre, C.H.; Li, W.; Vercauteren, T.; Ourselin, S.; Jorge Cardoso, M. Generalised dice overlap as a deep learning loss function for highly unbalanced segmentations. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support: Third International Workshop, DLMIA 2017, and 7th International Workshop, ML-CDS 2017, Held in Conjunction with MICCAI 2017, Québec City, QC, Canada, 14 September 2017; Proceedings 3; Springer: Berlin/Heidelberg, Germany, 2017; pp. 240–248. [Google Scholar]
  52. Wang, L.; Lu, H.; Wang, Y.; Feng, M.; Wang, D.; Yin, B.; Ruan, X. Learning to detect salient objects with image-level supervision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 136–145. [Google Scholar]
  53. Xu, S.; Wei, X.; Zhang, C.; Liu, J.; Zhang, J. MFFW: A new dataset for multi-focus image fusion. arXiv 2020, arXiv:2002.04780. [Google Scholar]
  54. Rockinger, O. Image sequence fusion using a shift-invariant wavelet transform. In Proceedings of the International Conference on Image Processing, Santa Barbara, CA, USA, 26–29 October 1997; Volume 3, pp. 288–291. [Google Scholar]
  55. Haghighat, M.B.A.; Aghagolzadeh, A.; Seyedarabi, H. Multi-focus image fusion for visual sensor networks in DCT domain. Comput. Electr. Eng. 2011, 37, 789–797. [Google Scholar] [CrossRef]
  56. Li, S.; Kang, X.; Hu, J. Image fusion with guided filtering. IEEE Trans. Image Process. 2013, 22, 2864–2875. [Google Scholar]
  57. Olshausen, B.A.; Field, D.J. Emergence of simple-cell receptive field properties by learning a sparse code for natural images. Nature 1996, 381, 607–609. [Google Scholar] [CrossRef] [PubMed]
  58. Liu, Y.; Wang, Z. Simultaneous image fusion and denoising with adaptive sparse representation. IET Image Process. 2015, 9, 347–357. [Google Scholar] [CrossRef]
  59. Zhou, Z.; Li, S.; Wang, B. Multi-scale weighted gradient-based fusion for multi-focus images. Inf. Fusion 2014, 20, 60–72. [Google Scholar] [CrossRef]
  60. Paul, S.; Sevcenco, I.S.; Agathoklis, P. Multi-exposure and multi-focus image fusion in gradient domain. J. Circuits Syst. Comput. 2016, 25, 1650123. [Google Scholar] [CrossRef]
  61. Liu, Y.; Liu, S.; Wang, Z. A general framework for image fusion based on multi-scale transform and sparse representation. Inf. Fusion 2015, 24, 147–164. [Google Scholar] [CrossRef]
  62. Guo, D.; Yan, J.; Qu, X. High quality multi-focus image fusion using self-similarity and depth information. Opt. Commun. 2015, 338, 138–144. [Google Scholar] [CrossRef]
  63. De, I.; Chanda, B. Multi-focus image fusion using a morphology-based focus measure in a quad-tree structure. Inf. Fusion 2013, 14, 136–146. [Google Scholar] [CrossRef]
  64. Qiu, X.; Li, M.; Zhang, L.; Yuan, X. Guided filter-based multi-focus image fusion through focus region detection. Signal Process. Image Commun. 2019, 72, 35–46. [Google Scholar] [CrossRef]
  65. Ma, J.; Zhou, Z.; Wang, B.; Miao, L.; Zong, H. Multi-focus image fusion using boosted random walks-based algorithm with two-scale focus maps. Neurocomputing 2019, 335, 9–20. [Google Scholar] [CrossRef]
  66. Zhan, K.; Kong, L.; Liu, B.; He, Y. Multimodal image seamless fusion. J. Electron. Imaging 2019, 28, 023027. [Google Scholar] [CrossRef]
  67. Ma, B.; Yin, X.; Wu, D.; Shen, H.; Ban, X.; Wang, Y. End-to-end learning for simultaneously generating decision map and multi-focus image fusion result. Neurocomputing 2022, 470, 204–216. [Google Scholar] [CrossRef]
  68. Xydeas, C.S.; Petrovic, V. Objective image fusion performance measure. Electron. Lett. 2000, 36, 308–309. [Google Scholar] [CrossRef]
  69. Qu, G.; Zhang, D.; Yan, P. Information measure for performance of image fusion. Electron. Lett. 2002, 38, 1. [Google Scholar] [CrossRef]
  70. Zhao, J.; Laganiere, R.; Liu, Z. Performance assessment of combinative pixel-level image fusion based on an absolute feature measurement. Int. J. Innov. Comput. Inf. Control 2007, 3, 1433–1447. [Google Scholar]
  71. Piella, G.; Heijmans, H. A new quality metric for image fusion. In Proceedings of the 2003 International Conference on Image Processing (Cat. No. 03CH37429), Barcelona, Spain, 14–17 September 2003; Volume 3, pp. 3–173. [Google Scholar]
  72. Chen, Y.; Blum, R.S. A new automated quality assessment algorithm for image fusion. Image Vis. Comput. 2009, 27, 1421–1432. [Google Scholar] [CrossRef]
  73. Levin, A.; Lischinski, D.; Weiss, Y. A closed-form solution to natural image matting. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 30, 228–242. [Google Scholar] [CrossRef]
Figure 1. Quantitative comparison of model size (indicated by bubble size), computational complexity (FLOPs), and fusion quality ( Q a b / f ) among different deep learning-based MFF methods on the Lytro dataset [46].
Figure 1. Quantitative comparison of model size (indicated by bubble size), computational complexity (FLOPs), and fusion quality ( Q a b / f ) among different deep learning-based MFF methods on the Lytro dataset [46].
Applsci 15 07500 g001
Figure 2. Architecture of the proposed method.
Figure 2. Architecture of the proposed method.
Applsci 15 07500 g002
Figure 3. The designed U-shaped ultra-light network ULU-Net.
Figure 3. The designed U-shaped ultra-light network ULU-Net.
Applsci 15 07500 g003
Figure 4. The adopted fusion scheme.
Figure 4. The adopted fusion scheme.
Applsci 15 07500 g004
Figure 5. Comparison of decision maps generated by different methods on the Lytro dataset [46].
Figure 5. Comparison of decision maps generated by different methods on the Lytro dataset [46].
Applsci 15 07500 g005
Figure 6. Comparison of decision maps generated by different methods on the MFFW dataset [53].
Figure 6. Comparison of decision maps generated by different methods on the MFFW dataset [53].
Applsci 15 07500 g006
Figure 7. Fusion results of various SOTA methods on the “Coffee cup” example from the MFFW dataset [53].
Figure 7. Fusion results of various SOTA methods on the “Coffee cup” example from the MFFW dataset [53].
Applsci 15 07500 g007
Figure 8. Comparison of decision maps generated by different methods on the MFI-WHU dataset [36].
Figure 8. Comparison of decision maps generated by different methods on the MFI-WHU dataset [36].
Applsci 15 07500 g008
Figure 9. Performance evaluation under extreme conditions. The proposed method demonstrates robust performance when processing image pairs consisting of completely sharp and completely blurred images. From left to right: original sharp image, synthetically blurred counterpart, generated decision map, and final fusion result. The model successfully distinguishes between focused and defocused regions even in these challenging scenarios.
Figure 9. Performance evaluation under extreme conditions. The proposed method demonstrates robust performance when processing image pairs consisting of completely sharp and completely blurred images. From left to right: original sharp image, synthetically blurred counterpart, generated decision map, and final fusion result. The model successfully distinguishes between focused and defocused regions even in these challenging scenarios.
Applsci 15 07500 g009
Figure 10. Qualitative comparison of the proposed LightMFF and DRPL on the Road-MF dataset [3]. FR and DM denote the final fusion results and decision maps (after post-processing), respectively.
Figure 10. Qualitative comparison of the proposed LightMFF and DRPL on the Road-MF dataset [3]. FR and DM denote the final fusion results and decision maps (after post-processing), respectively.
Applsci 15 07500 g010
Figure 11. Performance comparison of the model under different image degradation conditions. (a) Original image; (b) High brightness; (c) Low brightness; (d) High contrast; (e) Low contrast; (f) Gaussian noise; (g) Salt-and-pepper noise; (h) Gaussian filtering for Gaussian noise removal; (i) Median filtering for salt-and-pepper noise removal.
Figure 11. Performance comparison of the model under different image degradation conditions. (a) Original image; (b) High brightness; (c) Low brightness; (d) High contrast; (e) Low contrast; (f) Gaussian noise; (g) Salt-and-pepper noise; (h) Gaussian filtering for Gaussian noise removal; (i) Median filtering for salt-and-pepper noise removal.
Applsci 15 07500 g011
Figure 12. Visualization of ablation study results on the Lytro dataset [46].
Figure 12. Visualization of ablation study results on the Lytro dataset [46].
Applsci 15 07500 g012
Figure 13. Limitations of current decision map post-processing methods.
Figure 13. Limitations of current decision map post-processing methods.
Applsci 15 07500 g013
Table 1. Quantitative comparison of different MFF methods on the Lytro dataset [46].
Table 1. Quantitative comparison of different MFF methods on the Lytro dataset [46].
Method Q AB / F Q MI Q P Q W Q E Q CB
Methods based on image transform domain
DWT [7]0.68500.86770.28780.89770.83560.6117
DTCWT [54]0.69290.89920.29250.89870.84080.6234
NSCT [8]0.69010.90390.29280.90300.84130.6174
CVT [55]0.72430.89680.79660.93880.90230.7277
DCT [7]0.70310.93830.78250.90930.80730.6624
GFF [56]0.69981.00200.29520.89820.83510.6518
SR [57]0.69441.00030.29210.89840.83090.6406
ASR [58]0.69511.00240.29260.89860.83080.6413
MWGF [59]0.70371.05450.31760.89130.81070.6758
ICA [60]0.67660.86870.29640.90840.82190.5956
NSCT-SR [61]0.69951.01890.29490.90000.83850.6501
Methods based on image spatial domain
SSSDI [62]0.69661.03510.29150.89610.82790.6558
QUADTREE [63]0.70271.06300.29400.89620.82650.6681
DSIFT [61]0.70461.06420.29540.89770.83540.6675
SRCF [46]0.70361.05900.29540.89780.83690.6669
GFDF [64]0.70491.05240.29740.89890.83990.6657
BRW [65]0.70401.05160.29640.89840.83710.6650
MISF [66]0.69841.03910.29450.89290.80630.6607
MDLSR_RFM [20]0.75181.12330.82940.93940.90210.8064
End-to-end methods based on deep learning
IFCNN-MAX [24]0.67840.88630.29620.90130.83240.5986
U2Fusion [25]0.61900.78030.29940.89090.71080.5159
SDNet [35]0.64410.84640.30720.89340.74640.5739
MFF-GAN [36]0.62220.79300.28400.88870.76600.5399
SwinFusion [27]0.65970.84040.31170.90110.74600.5745
MUFusion [37]0.66140.80300.71600.90890.80360.6758
FusionDiff [29]0.67440.86920.29000.89800.82610.5747
SwinMFF [26]0.73210.96050.82220.93900.89860.7543
DDBFusion [28]0.50260.81520.56100.83910.49470.6057
Decision map-based methods using deep learning
CNN [22]0.70191.04240.29680.89760.83110.6628
ECNN [23]0.70301.07230.29450.89460.81690.6698
DRPL [19]0.75741.14050.84350.93970.90600.8035
SESF [30]0.70311.05240.29500.89770.83530.6657
MFIF-GAN [31]0.70291.06180.29600.89820.83950.6660
MSFIN [10]0.70451.06010.29730.89900.84360.6664
GACN [67]0.75811.13340.84430.94050.90130.8024
ZMFF [11]0.66350.86940.28900.89510.82530.6136
LightMFF0.75881.14620.84500.94000.90610.8067
Table 2. Comparison of computational efficiency across different learning-based MFF methods.
Table 2. Comparison of computational efficiency across different learning-based MFF methods.
MethodModel Size (M)FLOPs (G)Time (s)
End-to-end methods based on deep learning
IFCNN-MAX [24]0.088.540.09
U2Fusion [25]0.6686.400.16
SDNet [35]0.078.810.10
MFF-GAN [36]0.053.080.06
SwinFusion [27]0.9363.731.79
MUFusion [37]2.1624.070.72
FusionDiff [29]26.9058.1381.47
SwinMFF [26]41.2522.380.46
DDBFusion [28]10.92184.931.69
Decision map-based methods using deep learning
CNN [22]8.76142.230.06
ECNN [23]1.5914.93125.53
DRPL [19]1.07140.490.22
SESF [30]0.074.900.26
MFIF-GAN [31]3.82693.030.32
MSFIN [10]4.5926.761.10
GACN [67]0.0710.890.16
ZMFF [11]6.33464.53165.38
LightMFF0.020.060.02
Reduction (%)60.00%98.05%66.67%
Table 3. Ablation quantitative results on the Lytro dataset [46]. Group 1: baseline with only source images; Group 2: with initial decision map; Group 3: with edge maps; Group 4: full model with both components.
Table 3. Ablation quantitative results on the Lytro dataset [46]. Group 1: baseline with only source images; Group 2: with initial decision map; Group 3: with edge maps; Group 4: full model with both components.
Method Q AB / F Q MI Q P Q W Q E Q CB
Group 10.75781.14480.84370.93630.90210.8036
Group 20.75871.14600.84470.93890.90550.8060
Group 30.75841.14590.84430.93730.90530.8057
Group 40.75881.14620.84500.94000.90610.8067
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xie, X.; Lin, Z.; Guo, B.; He, S.; Gu, Y.; Bai, Y.; Li, P. LightMFF: A Simple and Efficient Ultra-Lightweight Multi-Focus Image Fusion Network. Appl. Sci. 2025, 15, 7500. https://doi.org/10.3390/app15137500

AMA Style

Xie X, Lin Z, Guo B, He S, Gu Y, Bai Y, Li P. LightMFF: A Simple and Efficient Ultra-Lightweight Multi-Focus Image Fusion Network. Applied Sciences. 2025; 15(13):7500. https://doi.org/10.3390/app15137500

Chicago/Turabian Style

Xie, Xinzhe, Zijian Lin, Buyu Guo, Shuangyan He, Yanzhen Gu, Yefei Bai, and Peiliang Li. 2025. "LightMFF: A Simple and Efficient Ultra-Lightweight Multi-Focus Image Fusion Network" Applied Sciences 15, no. 13: 7500. https://doi.org/10.3390/app15137500

APA Style

Xie, X., Lin, Z., Guo, B., He, S., Gu, Y., Bai, Y., & Li, P. (2025). LightMFF: A Simple and Efficient Ultra-Lightweight Multi-Focus Image Fusion Network. Applied Sciences, 15(13), 7500. https://doi.org/10.3390/app15137500

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop