Next Article in Journal
Deep Learning-Based Min-Entropy-Accelerated Evaluation for High-Speed Quantum Random Number Generation
Previous Article in Journal
CurveMark: Detecting AI-Generated Text via Probabilistic Curvature and Dynamic Semantic Watermarking
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

KCUNET: Multi-Focus Image Fusion via the Parallel Integration of KAN and Convolutional Layers

1
School of Physics and Electronics, Shandong Normal University, Jinan 250358, China
2
School of Management Engineering, Shandong Jianzhu University, Jinan 250101, China
3
Institute of Information Science, Beijing Jiaotong University, Beijing 100044, China
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(8), 785; https://doi.org/10.3390/e27080785
Submission received: 20 May 2025 / Revised: 19 July 2025 / Accepted: 22 July 2025 / Published: 24 July 2025
(This article belongs to the Section Signal and Data Analysis)

Abstract

Multi-focus image fusion (MFIF) is an image-processing method that aims to generate fully focused images by integrating source images from different focal planes. However, the defocus spread effect (DSE) often leads to blurred or jagged focus/defocus boundaries in fused images, which affects the quality of the image. To address this issue, this paper proposes a novel model that embeds the Kolmogorov–Arnold network with convolutional layers in parallel within the U-Net architecture (KCUNet). This model keeps the spatial dimensions of the feature map constant to maintain high-resolution details while progressively increasing the number of channels to capture multi-level features at the encoding stage. In addition, KCUNet incorporates a content-guided attention mechanism to enhance edge information processing, which is crucial for DSE reduction and edge preservation. The model’s performance is optimized through a hybrid loss function that evaluates in several aspects, including edge alignment, mask prediction, and image quality. Finally, comparative evaluations against 15 state-of-the-art methods demonstrate KCUNet’s superior performance in both qualitative and quantitative analyses.

1. Introduction

Depth of field (DOF) is defined as the distance between the nearest and furthest points that appear in focus, with objects within this range maintaining sharpness while those outside it gradually blur. In photography, the limitations of DOF prevent optical lenses from capturing images where all elements are simultaneously in focus. However, achieving a fully focused image is often desired in practice. To address this challenge, multi-focus image fusion (MFIF) has been developed as a solution. This technique combines multiple source images, each with different focus points, to produce a single image that is sharp across all focal planes, thereby enriching the scene’s information. By doing so, MFIF effectively extends the DOF of optical lenses, enabling imaging systems to overcome DOF constraints and produce higher-quality images.
MFIF algorithms have been extensively studied for decades and can generally be categorized into traditional approaches and deep learning-based approaches [1]. Traditional approaches are further divided into transform domain-based methods and spatial domain-based methods. Transform domain-based approaches involve converting the image from the spatial domain to a transform domain (e.g., wavelet transform, DCT transform, contourlet transform, etc.), extracting image features, and processing them using the characteristics of the transform domain. The final fused result is obtained by transforming the processed data back to the spatial domain through an inverse transformation. However, this method often yields unrealistic results. On the other hand, spatial domain-based approaches include three main categories: pixel-based [2,3], block-based [4,5], and region-based [6,7] methods. These methods differentiate focusing conditions at the pixel, block, or region level, respectively. Pixel-based methods are prone to mismatch issues, while block-based techniques are often sensitive to block size. Additionally, the region segmentation process in region-based methods can significantly impact their efficiency and performance.
In recent years, deep learning technology has advanced rapidly, and deep learning-based MFIF has attracted widespread attention. Numerous experimental results demonstrate that the quality of fusion outcomes achieved through deep learning significantly surpasses that of traditional techniques. Currently, MFIF methods based on deep learning include decision map-based methods [8] and end-to-end methods [9] according to the model process. The former generates a probability map, known as a decision map, by learning the features of the image, which indicates the level of pixel or region focus activity. This map is then used to determine whether to retain a region and to wait or select the original image to generate the fused image. While this approach offers flexibility in designing specific processing strategies for different focus regions, it requires additional post-processing steps, thereby increasing the computational complexity. The latter directly extracts features from the input image and merges them into an image without generating decision maps. However, the fused result has artifacts that affect fusion quality. Deep learning approaches can be broadly classified into supervised [10] and unsupervised [11] learning paradigms based on the availability of labeled training data. Supervised learning relies on annotated datasets where each input sample is paired with corresponding output labels, typically requiring substantial amounts of labeled data. However, acquiring authentic MFIF datasets for training remains challenging due to practical imaging constraints. While unsupervised learning eliminates the need for manual labeling by extracting features directly from unlabeled data, this approach often suffers from high computational demands and inconsistent fusion performance, potentially leading to detail loss or artifact generation. Given these considerations, we propose a supervised deep learning model capable of generating both decision maps and fusion results for use in MFIF.
In 2015, O. Ronneberger et al. [12] introduced the U-Net model, which has since become widely adopted in image processing and spawned numerous variants. Cicek et al. [13] adapted the U-Net architecture for 3D image segmentation, significantly improving 3D-imaging accuracy. Zhou et al. [14] further advanced the architecture by developing U-Net++, which integrates nested structures and dense feature fusion mechanisms. Recent advances in transformer technology have led to novel hybrid architectures that combine U-Net’s precise localization capabilities with transformers’ ability to capture long-range dependencies. Cao et al. [15] proposed Swin-UNet, a medical image segmentation network that incorporates the Swin transformer architecture. Currently, the Kolmogorov–Arnold network (KAN) model is being used to enhance the ability of models to capture complex features and patterns due to its unique structural design. By leveraging its superior nonlinear modeling capabilities, KAN can effectively capture intricate relationships within images, enabling more efficient global feature extraction.
Building upon these advancements, this paper proposes KCUNet, an innovative U-Net variant that embeds the KAN model in parallel with a convolutional layer into the encoder of the UNet network. The KAN model, which has attracted significant research attention this year, combines an attention mechanism that can better capture key information and a kernel approach that enhances the nonlinear representation of the model. In the KCUNet architecture, while convolutional layers extract fundamental features, parallel KAN layers provide complementary contextual understanding, collectively strengthening the model’s ability to handle complex visual scenes. In addition, the traditional U-Net network extracts features by downsampling and upsampling, but continuous downsampling loses some important spatial information and details, leading to blurring of image boundaries and details. Therefore, the coding part of this paper no longer reduces the feature map size. The principal contributions of this work are as follows.
  • This paper proposes a novel MFIF network based U-Net called KCUNet, which a parallel KAN layer alongside convolutional layers in the encoder, maintaining consistent spatial dimensions of feature maps while progressively increasing channel depth to capture multi-scale features effectively.
  • This paper designs a spatial–channel guided attention network (S–CGANet), which boosts edge preservation by combining Laplace pyramid-based high-frequency feature extraction with channel guided attention (CGA), refining boundary localization and reducing defocus effects in image fusion.
  • This paper designs a hybrid loss function that comprehensively considers the performance of the model in several aspects including edge alignment, mask prediction and fused images.
  • This paper demonstrates the performance of KCUNet through extensive comparative experiments and ablation studies. Compared with 15 state-of-the-art (SOTA) methods, KCUNet shows superior performance in focal region detection, visual perception analysis and quantitative scoring.
The remainder of this paper is structured as follows. A review of some related work is provided in Section 2. The proposed model structure and loss function are described in detail in Section 3. Experimental results are presented and discussed analytically in Section 4. The paper is concluded in Section 5.

2. Related Work

2.1. Traditional MFIF Methods

In the 1980s, MFIF technology began to emerge, with traditional methods broadly categorized into transform domain-based and spatial domain-based approaches. During the early development of transform domain methodology, Burt et al. [16] pioneered the application of the Laplace pyramid approach to MFIF. This method involved recursively applying Gaussian filtering and downsampling to decompose an image into multiple layers of varying resolutions, enabling multi-scale image analysis. Subsequently, Toet et al. [17] introduced the ratio of low-pass pyramid (RP)-based method, which utilized the ratio between low-pass pyramids of two images to determine fusion weights. The advent of the wavelet transform later led to the development of various wavelet-based methods. Lewis et al. [18] proposed a technique based on the dual-tree complex wavelet transform (DTCWT), capable of capturing both magnitude and phase information to better preserve local features and edge details. Zhang et al. [19] introduced the non-subsampled wavelet transform (NSCT) method, which avoided information loss during image decomposition and reconstruction by eliminating subsampling. Yang et al. [20] incorporated sparse representations into image integration for the first time. He et al. [21] later proposed guided image filtering (GFF), a novel filtering technique that leverages a guiding image to preserve edge and texture information while effectively suppressing noise.
To address the limitations of transform-domain methods in image fusion, researchers have developed spatial-domain approaches. Li et al. [6] introduced a spatial frequency (SF)-based method, which evaluates image sharpness by analyzing spatial frequency information, allowing the identification of focus and defocus regions in images. Further advancing the field, Li et al. [4] proposed an MFIF method incorporating region segmentation, which effectively balances local details and global features by partitioning the image into distinct regions. Zhang et al. [22] developed a boundary detection-based MFIF technique, employing a novel multi-scale morphological focus metric to extract boundary information and guide the fusion process. Panigrahy et al. [23] introduced a parameter-adaptive method that utilizes fractal dimension as a feature metric, dynamically adjusting the parameters of a two-channel impulse-coupled neural network to optimize fusion performance. Additionally, Kurban [24] proposed the Gaussian of differences (GD), a computationally efficient pixel-based framework for general image fusion. This approach enhances edge preservation through first-order derivative edge detection while incorporating neighborhood pixel contributions via Gaussian filtering. To ensure adaptability across diverse images, the method employs a pattern search algorithm for automatic parameter optimization.

2.2. Deep Learning-Based Methods

Recent years have witnessed the emergence of numerous deep learning-based algorithms in image fusion. In 2017, Liu et al. [8] pioneered the application of deep learning in MFIF by proposing a CNN-based model that integrates region-blocking techniques to guide activity level measurement and fusion rule generation. Zhang et al. [25] later introduced a universal end-to-end model (IFCNN), capable of handling various fusion tasks beyond MFIF. This approach directly learns fused images from input data without requiring complex post-processing. Ma et al. [11] proposed an unsupervised deep learning model (SESF-Fuse), which extracts deep features directly from the input image by training the encoder-decoder network in an unsupervised manner, which is then used for image fusion. Similarly, Jung et al. [26] proposed a model that handles all fusion stages comprehensively. This method leverages structural tensor representations of multi-channel image contrast as the basis for an unsupervised loss function, eliminating dependence on labeled data while overcoming the computational limitations of traditional approaches. Yan et al. [27] introduced a feed-forward fully convolutional network (MFNet) capable of processing variable-sized images. This model employs structural similarity (SSIM) as its loss function to optimize output quality. Ma et al. [28] investigated an end-to-end (MFIF) approach employing multi-scale generative adversarial networks (MsGAN). In this framework, source images are decomposed into multi-scale sub-images, which undergo processing through depth-variant GANs for feature extraction. The fused image is subsequently reconstructed via inverse wavelet transform, utilizing CNN-based multi-scale decomposition to optimize detail preservation. More recently, Cheng et al. [29] presented a framework (MUFusion) that adaptively refines fusion performance by incorporating intermediate training results as additional guidance, thereby enhancing the final output through multi-stage feature learning.

2.3. Defocus Spread Effect (DSE)

DSE is a prevalent issue that can affect the quality of image fusion. It is a fact that focused and defocused regions are not always separated by a clear boundary. When the image’s foreground is sharply focused and the background is blurred, the focused foreground is not affected by the background out of focus, and then the focus/defocus boundary (FDB) is a sharp outline, as shown in Figure 1a. But if the focus is reversed, the defocused foreground will penetrate into the background, then the FDB is an indeterminate region, as shown in Figure 1b. Therefore, how to avoid DSE is one of the main research directions at present.
In previous studies, researchers have explored various approaches to address the DSE in image fusion. Ma et al. [30] proposed a model to simulate DSE and generated training data to capture the differences in DSE when either the foreground or background is defocused. They also introduced a cascade-based boundary-aware convolutional network (MMF-Net), which refines fusion results along the FDB by generating fusion bootstrap maps through two sub-networks. Xu et al. [31] employed a gradient ascent technique to enhance the structural similarity index of fused images, ensuring that each segment closely matches its corresponding region in the source images, thereby reducing the effect of DSE. Wang et al. [32] developed MFIF-GAN, a network that generates focus maps with slightly expanded foreground regions to counteract the effects of DSE. Unlike the above methods, we use the Laplace pyramid to identify and extract high-frequency features, including edges and contours of the image. Additionally, the CGA module is employed to augment the model’s focus on edge information and retain the edge information at different levels, effectively mitigating DSE. Further details are discussed in Section 3.2.2.

2.4. Kolmogorov–Arnold Network (KAN)

Recent advancements in neural network architectures have led to the development of innovative models that prioritize both interpretability and performance. Among these, KAN has emerged as a promising approach, attracting significant attention due to its enhanced interpretability and efficient data-fitting capabilities with fewer parameters.
Drawing on the KAN representation theorem, MIT researchers developed KAN as an innovative network architecture [33]. The architecture’s fundamental innovation lies in its departure from conventional neural network design through the replacement of traditional weight parameters and the repositioning of activation functions. In contrast to standard multi-layer perceptrons (MLPs), KAN places learnable activation functions on the ‘edges’ of the network instead of using fixed activation functions at the nodes. Specifically, KAN eliminates fixed activation functions entirely, instead implementing them through learnable univariate functions. This architectural choice endows KAN with superior computational efficiency, accuracy, and interpretability compared to MLPs. The training process of KAN involves parameter updates through loss function computation and gradient backpropagation. This training framework enables KAN to autonomously determine optimal univariate functions that most effectively approximate the target function. Furthermore, KAN demonstrates greater parameter efficiency by eliminating the extensive weight parameters characteristic of traditional MLPs, achieving complex nonlinear transformations through fewer spline function parameters instead.
Building upon this foundation, we integrate KAN with widely used CNN and U-Net architectures, seeking to further investigate KAN’s efficacy and potential advantages in MFIF tasks.

3. Proposed Method

This section first provides a comprehensive overview of KCUNet, the proposed framework in this study. Subsequently, it elaborates on the network’s key components and the employed loss function in greater detail.

3.1. Problem Formulation

As mentioned in the previous section, MFIF aims at fusing multiple source images { I 1 , I 2 , , I N } into one fully focused image F. For a typical dual image fusion scene, the process can be mathematically expressed as follows:
F = I A × D M + I B × ( 1 D M ) ,
where I A and I B R C × H × W denote the foreground and background focus images, respectively (C, H, and W are the number of channels, height, and width of the image, respectively). D M [ 0 , 1 ] 1 × H × W is a decision map that identifies the focus region in I A . The ideal D M should satisfy the following:
DM ( x , y ) = 1 , I A ( x , y ) is in focus 0 , I B ( x , y ) is in focus .
Here, the accuracy of D M directly affects the quality of image fusion. Generating an accurate D M requires multi-scale feature analysis, while preserving clear boundaries depends on a precise D M as a foundation. To address these challenges, we propose KCUNet, a parallel KAN-convolutional architecture that synergistically resolves these key issues. By integrating local detail features with global contextual information, the network ensures accurate capture and preservation of critical details across different scales, thereby enhancing both the visual quality and informational integrity of the fused results.

3.2. Overview

As shown in Figure 2, this paper proposes KCUNet, a novel MFIF network that embeds KAN with convolutional layers in parallel within a U-Net framework. The architecture consists of three primary components: an encoding stage, a decoding stage, and an attention module.
The encoding stage comprises four encoders, each containing parallel–connected KAN and convolutional layers. Differing from the traditional U-Net architecture, KCUNet maintains the spatial sizes of feature maps throughout the encoding stage while incrementally increasing the channel count at high resolution. This approach enables each encoding layer to learn richer feature representations at consistent spatial dimensions. With increasing network depth, the expanding channel capacity allows the model to capture increasingly detailed features across different levels, thereby supplying more comprehensive feature information for subsequent fusion processes. In the decoding stage, an inverse strategy is adopted. Specifically, the channel count is progressively reduced while preserving the feature map dimensions. This design enables the model to more attentively process and integrate key features from the encoder while simultaneously reducing computational complexity and parameter volume.
In addition, the decoding part combines the edge information extracted by the attention module for feature enhancement and the Decoder x-1 layer to convert the feature map into a binary image. Specifically, the output feature maps of the encoder are fused with the enhanced feature maps from the attention module and the source image processed by the Laplace operation. That is, the feature map from Encoder 4 undergoes channel reduction via Decoder 4-2, and then is converted into a binary image by Decoder 4-1 and fed into the attention module. The attention module combined this output with the features extracted by Encoder 3 and the source image processed by Laplace operation, and then passed them to Decoder 3-2. The process is repeated three times until Decoder 1-1 produces the preliminary decision map. Finally, a small-area removal strategy is applied to generate the ultimate decision map, which is then used to generate the fusion result according to Equation (1). Table 1 shows the specifications of each module.

3.2.1. Encoding Component

The encoder consists of parallel–connected KAN layers and convolutional layers. While convolutional layers exhibit robustness to variations in object location and scale within the image due to their translation-invariant nature, they struggle with global information processing because of their reliance on local receptive fields when processing the image. In contrast, KAN leverages its nonlinear modeling capability to capture complex nonlinear relationships in images, enabling effective extraction of global features. Specifically, KAN extracts each channel’s information pixel by pixel and converts it to 1D data, which is then processed using the KAN layer and restored to 2D. By combining convolutional layers and KAN layers in a parallel dual-channel structure, the model retains both local feature sensitivity and global information awareness. As illustrated in Figure 3, the convolutional layer consists of two cascaded 3 × 3 convolutions. The features extracted from this layer are concatenated with those from the KAN layer along the channel dimension. The merged features are then further processed using a 3 × 3 convolution followed by a 1 × 1 convolution for refined feature extraction.

3.2.2. Attention Module

This section further elaborates on the principle of S–CGANet. As a core component of the KCUNet architecture, the S–CGANet module exhibits robust edge-guidance capabilities and operates across multiple levels. It extracts features from low to high levels while enhancing the model’s attention to edge-related information. The module primarily employs the Laplace pyramid method to capture high-frequency variations in images and utilizes the CGA module to precisely focus on dispersion boundaries. This approach significantly improves the accuracy and robustness of feature extraction.
The CGA module, introduced in DEA-Net [34], is an innovative attention mechanism that generates a unique spatial importance map (SIM) for each channel in a CNN. Its key strength lies in leveraging input feature information to guide SIM generation, enabling the model to more accurately detect and emphasize feature-relevant regions in every channel. This mechanism first constructs an initial attention map by applying operations such as global maximum pooling and global average pooling across channel and spatial dimensions. The channel and spatial attention components are then combined through addition to produce a preliminary SIM. Finally, the CGA refines this coarse SIM for each channel using channel shuffling and grouped convolution, yielding channel-specific SIMs that enhance the model’s ability to focus on the most discriminative feature information.
The Laplace pyramid method builds an efficient multi-scale representation by progressively reducing image resolution and storing inter-layer differences, thereby improving both efficiency and accuracy in image analysis. When processing source image pairs, the Laplace pyramid preserves critical edge information. Among them, the Laplace operator, as a second-order derivative edge detection tool, captures high-frequency features, such as edges and contours, by analyzing second-order variations in the input image. These extracted high-frequency features enable the identification of FDB and effectively mitigating DSE. It should be noted that while the Laplace operator exhibits sensitivity to high-frequency noise, our design effectively addresses this limitation through two key mechanisms. Firstly, the multi-scale decomposition framework of the Laplace pyramid inherently suppresses high-frequency noise at coarse scales while preserving meaningful edge information. Secondly, the CGA module dynamically attenuates noise-dominated feature channels through its channel-adaptive SIMs.
As illustrated in Figure 4, the S–CGANet module enhances the model’s ability to focus on edge-related information by integrating diverse features and leveraging the robust edge-guidance capability of the CGA. This mechanism preserves critical high-frequency details across different layers, particularly the edge information that tends to diminish as network depth increases. Specifically, the S–CGANet module at each layer processes three inputs: the embedded features from the encoder, the predicted features from the higher layers of the decoder, and the high-frequency features extracted by the Laplace operator. The predicted features at the decoder’s output generate an inverse attention map through a single subtraction operation, while the boundary attention map is produced by applying the Laplace operator. These two attention maps, along with the high-frequency features extracted by the Laplace operator, are fed into the CGA attention module after undergoing element-wise multiplication with the embedded features from the current encoder. This integration occurs through sequential operations of concatenation, convolution, and element-wise multiplication. Consequently, the S–CGANet module achieves more precise localization of defocus boundaries, significantly improving the accuracy and robustness of feature extraction. This approach retains more edge features while mitigating the impact of DSE on the MFIF task.

3.3. Loss Function

In deep learning, the loss function plays a crucial role in determining model performance, as it guides the learning direction of the model and drives the model’s predictions toward ground truth values. To comprehensively consider the effectiveness of KCUNet, we propose a hybrid loss function L t o t a l , which guides the training of the model by combining different loss components, while constraining the output decision map and fusion results. Specifically, it consists of three main components, each of which is optimized for a different aspect of model performance, as defined below:
L t o t a l = λ 1 L E A + λ 2 L M a s k + λ 3 L Q g ,
where λ i is the hyperparameter for adjusting the influence of each loss component on the overall loss. This design allows flexibility in adjusting how much attention the model pays to different performance metrics during training, which will be set to 1 in this paper.

3.3.1. Edge Alignment Loss

L E A = α 1 L B C E + α 2 L D i c e .
This part of the loss function focuses on the performance of KCUNet in terms of edge alignment. In deep networks, some important detailed information may be lost as the depth of the network increases. We apply the loss function on multiple decoding layers and use this loss function to measure the difference between the predicted probability distributions and the true labels at different levels, which makes the model focus on the detailed information at each stage and reduces the information loss. It consists of L B C E and L D i c e , two sub-losses corresponding to the bounded cross-entropy loss and the Dice loss. The weights of α i are controlled and are set to 0.5 in this paper. The combination of these two losses allows the model to be more accurate in predicting edges while maintaining the consistency of the region. The weights of this part of the loss are controlled by λ 1 .
L D i c e is defined in Equation (3) as the loss function in semantic segmentation:
L D i c e = 1 2 i N P p i g i + 1 i N P p i 2 + i N P g i 2 + 1 ,
which sums the predicted binary segmentation map p i and the ground truth map g i over N p pixels. The addition of 1 helps to address the issue of gradient loss. This loss function measures the agreement between the predicted decision map and the true values, i.e., the segmentation accuracy of the target region. By comparing their similarity, it constrains the model parameters and improves the network’s ability to identify different targets.
L B C E is a classical binary loss defined as follows:
L BCE = i , j H , W G ( i , j ) ln DM ( i , j ) + 1 G ( i , j ) ln 1 DM ( i , j ) n ,
where G ( i , j ) represents the ground truth binary mask, and D M ( i , j ) is the predicted decision map. L B C E can access the correct classification of each pixel and improves the network’s ability to distinguish between sharp and blurred pixels.

3.3.2. Mask Loss

L M a s k = L D i c e + L S S I M .
The mask loss evaluates the model’s performance in mask prediction by measuring the discrepancy between the decision map and the ground truth labels. This loss component combines Dice loss and structural similarity index (SSIM) loss to enhance both accuracy and structural preservation in mask segmentation. The parameter λ 2 regulates the weight of the mask loss.
The L S S I M loss function constrains the brightness, contrast, and structural metrics of the image, ensuring that the fusion result maintains greater similarity with the input image. It is computed using the following formula:
L S S I M I A , I B , F = 1 1 N P ω = 1 N P S I A , I B , F ω ,
where I A and I B serve as the input images, with F being the fused result, N P represents the total count of sliding windows, ω denotes a specific local window, and S is defined as follows:
S I A , I B , F ω = S S I M I A , F ω i f s t d I A , ω s t d I B , ω S S I M I B , F ω o t h e r w i s e ,
where S S I M ( I A , F | ω ) refers to the SSIM index calculated between image patch I A and F within a local window ω , ω = 11 × 11 in this paper, and s t d ( I A , | ω ) represents the standard deviation of image patch I A within the local window ω .

3.3.3. Quality Loss

L Q g is derived from the gradient-aware loss function in GACN [9], enhancing gradient preservation in the fusion output by regulating edge information retention at the pixel level. For detailed formulation and parameters, please consult the original GACN reference. The weights are controlled by λ 3 .

4. Experiment

4.1. Execution Details

Dataset selection plays a crucial role in model performance, yet acquiring real MFIF datasets for training remains challenging due to practical shooting constraints. To address this, we employ the VOC 2012 dataset [35], which comprises 2913 images with corresponding segmentation masks. We synthesize MFIF training data by applying Gaussian filtering to create two variants of each image: one with blurred foreground and another with blurred background. The α -matte boundary defocus spread model [30] is then used to simulate the DSE. The synthesized dataset is split in a 7:3 ratio, yielding 2016 image pairs for training and 897 pairs for validation. Figure 5 illustrates samples from our synthetic MFIF dataset.
During the training process, the input image size is first resized to a uniform 256 × 256 pixels to enhance the model’s computational efficiency. Next, the image is randomly cropped to 156 × 156 pixels to simulate diverse scenarios and increase data diversity. Additionally, the input data undergo random horizontal and vertical flipping, as well as random rotation at angles selected from 0, 30, 60, 90, 120, 150, 180, 210, 240, 270, 300, and 330 degrees. This helps the model adapt to varying angles and orientations, thereby improving its generalization capability. To further enhance robustness, data augmentation techniques such as random blurring, random erasure, random offset, and Gaussian noise addition are applied. These operations strengthen the model’s ability to handle various image disturbances, ensuring stable performance even with incomplete information or in complex environments. In contrast, no preprocessing is performed during the validation phase to accurately simulate the model’s performance on real-world datasets.
KCUNet was trained using the PyTorch framework with the following configurations: an initial learning rate of 0.0001, 50 epochs, and a learning rate reduction by a factor of 0.8 every two epochs. The weight decay was set to 0.0001, and the Adam optimizer was employed to adaptively adjust the learning rate for each parameter. A batch size of eight was used during training. The model was trained and tested on an NVIDIA Quadro P5000 GPU with 16GB of memory.

4.2. Test Datasets

For the MFIF task, the test datasets consist of both real-world and synthetic images. This study evaluates model performance using the Lytro dataset and the MFFW dataset for real image fusion, while the MFI-WHU dataset is employed to assess synthetic image fusion performance.
The Lytro camera is a light-field imaging device that records not only light intensity but also the direction of light. This unique capability enables post-capture focus adjustment in the resulting images. The Lytro dataset [36] contains 20 image pairs, each with dimensions of 520 × 520 pixels, all captured by the Lytro light-field camera under identical scene conditions but with varying focal points.
The MFFW dataset, specifically designed for MFIF, contains 12 real multi-focus image pairs exhibiting significant DSE. These images are from real-world scenes with complex backgrounds and variable lighting conditions, presenting a rigorous challenge for fusion algorithms.
For synthetic image evaluation, the MFI-WHU dataset [37] is utilized. Constructed by Zhang et al., this dataset derives from the COCO datasets [38,39], where 120 clear images were selected and processed using Gaussian blur and manual decision maps to generate multi-focus pairs. From these, 18 image pairs are chosen for testing in this study.

4.3. Assessment Methods

Assessing the quality of image fusion in the MFIF can be categorized into two approaches: objective and subjective methods. The objective approach, known as quantitative evaluation, employs evaluation metrics to assess fusion quality. These metrics fall into four main categories: information theory-based, image feature-based, structural similarity-based, and human perception-inspired metrics [40]. Each category provides a distinct perspective for evaluating fusion algorithm performance. The subjective approach, also referred to as qualitative evaluation, relies on human observers to visually inspect fused images and make judgments based on personal perception. While straightforward, this method is inherently subjective and prone to bias. Consequently, subjective evaluation is typically used in conjunction with objective methods.
We select eight fusion metrics from four categories to assess the performance of KCUNet. Theory-based information, including M I [41] and N C I E [42], human perception-inspired fusion metrics, including Q C B [43], image structural similarity-based metrics, including S S I M [44], image feature-based metrics, including Q M [45], Q G [46], G L D [47], and M S D [47]. For the metrics, higher values indicate higher performance. The objective evaluation provides a direct numerical assessment of the model’s fusion efficiency. To ensure fairness, default parameters are applied consistently across all metrics. All calculations are performed using MATLAB R2023a.

4.4. Comparison with SOTA Methods

We conduct a comprehensive comparison between KCUNet and 15 SOTA methods, encompassing both traditional and deep learning-based approaches. Traditional methods include BFMF [22] and DCT-EOL [48], deep learning based methods include CNN [8], ECNN [49], IFCNN [25], DRPL [10], MFF-Net [30], U2Fusion [50], MFF-GAN [37], MFIF-GAN [32], SESF [11], SwinFusion [51], MiT [52], ZMFF [53], and FusionDiff [54].

4.4.1. Subjective Analyses

For visual comparison, representative images from three benchmark datasets were selected: the 3rd and 17th images from the Lytro dataset, the 4th image from the MFFW dataset, and the 8th image from the MFI-WHU dataset. Figure 6 presents the comparative visual results between KCUNet and the 15 state-of-the-art methods, demonstrating the performance differences across different fusion scenarios.
To better visualize the subtle differences between the various methods, we select specific local regions from the images and enlarge them using bilinear interpolation. A smaller red rectangle in each figure marks the selected region for magnification, while the enlarged view is displayed in the upper left corner of the image, corresponding to the larger red rectangle. Source A and Source B represent the near-focus and far-focus images, respectively, while the remaining images display fusion results from different models, with their corresponding method names labeled below. In the third image of the Lytro dataset, the edges in the BFMF result appear blurred, while the hat edges in the DCT-EOL, DRPL, MFIF-GAN, and FusionDiff results exhibit varying degrees of artifacts. In the seventeenth image, the BFMF method causes the doll’s eyes to appear blurred and discolored, whereas the DCT-EOL, FusionDiff, ECNN, and Swin Fusion methods introduce jagged edges, and the DRPL method produces artifacts. For the fourth image in the MFFW dataset, significant diffusion blurring is observed at the old man’s sleeve in the results of BFMF, CBF, DRPL, and NSCT-SR, while IFCNN, MFF-GAN, U2Fusion, and ZMFF exhibit severe diffusion blurring. Additionally, MiT, U2Fusion, and MMFNet display noticeable ghosting artifacts. In the eighth image of the MFI-WHU dataset, BFMF, CNN, ECNN, and SESF show severe blurring around the light. Furthermore, the U2Fusion method produces fusion results with significant color discrepancies.
The above comparison demonstrates that KCUNet exhibits significant advantages in visual performance, particularly in preserving edge details and color accuracy, resulting in more natural and realistic fused images. Consequently, KCUNet excels in addressing DSE and offers an effective solution for MFIF.

4.4.2. Objective Analyses

Objectively, Table 2 provides the average measurement results for all images in the testing dataset. Compared to other SOTA methods, KCUNet obtains the highest scores on seven out of eight evaluation metrics for both the Lytro and MFFW datasets, while closely approaching the highest score on the remaining metric. On the MFI-WHU dataset, KCUNet ranks first in five out of eight metrics. Both Lytro and MFFW are real images captured by a camera, and in particular, there is significant DSF in the MFFW dataset, and KCUNet’s good performance in both datasets suggests the effective mitigation of DSE. Among the specific evaluation metrics, it is noteworthy that KCUNet achieves the highest scores on MI and NCIE, which indicates that the fusion results retain a substantial amount of information from the source image. The highest score for Q G indicates that the fusion results successfully inherit the rich detail and edge information from the source image. The highest scores for Q M and MSD indicate KCUNet effectively maintains multiscale structural information, further validating its ability to retain hierarchical features from source images. Overall, the highest and near-highest values for the other evaluation metrics further emphasize that the method produces fusion results with high similarity to the source image with fewer artifacts. However, there remains room for improvement in Q C B and GLD, likely due to a trade-off where enhancing fine details slightly compromises image smoothness, leading to minor visual distortions. This limitation will be thoroughly investigated in future work to further enhance fusion quality.
In addition, Table 2 presents the average fusion time for each method. While KCUNet demonstrates superior performance in fusion quality, its computational efficiency does not show a significant advantage over most competing methods. This limitation likely stems from the KAN model’s dense connection architecture, which increases computational complexity and reduces processing speed.

4.5. Ablation Experiments

4.5.1. Effect of KAN

To validate the critical role of the KAN layer, we conduct comparative experiments between KCUNet and a purely convolutional layer encoder across three benchmark datasets (Lytro, MFFW, and MFI-WHU). As evidenced by Table 3, KCUNet’s fusion results consistently outperform the KAN-free variant across all eight evaluation metrics. Among them, the differences in Q M , G L D , Q G , and Q C B are more obvious, which indicates that the KAN layer has a better ability to extract edge and detail information and can extract more multi-scale information. Figure 7 demonstrates the differences in the experimental results. It can be seen that the KAN structure can better capture the complex transition features of DSE due to its nonlinear modeling capability. Additionally, by processing channel information at the pixel level, KAN compensates for traditional convolution’s limitations in global feature extraction. The parallel design of KAN and convolutional layers further achieves a balance between local feature sensitivity and global context awareness. While KAN increases computation time by approximately 3–4 times, the performance gains outweigh this cost. Future research could extend the KAN structure to other image processing tasks requiring fine boundary processing, particularly in scenarios where preserving both local details and global structure is crucial.

4.5.2. Impact of Encoder and Attention Modules

KCUNet primarily consists of three components: the encoding part, the decoding part, and the attention part. To investigate the impact of the designed encoder and attention module, we compare image fusion results using KCUNet with the following three approaches:
(i) Image fusion using only the decoder. This evaluates the fused image quality without the assistance of the encoder and attention module.
(ii) Image fusion using both the encoder and decoder. By integrating the encoder with the decoder, the role of the proposed encoder can be assessed.
(iii) Image fusion using both the attention module and decoder. This combines the attention module with the decoder to verify the contribution of the attention module.
Figure 8 shows the values of the eight fusion metrics obtained when fusing the Lytro, MFFW and MFI-WHU datasets using KCUNet and these approaches. Where circles represent the Lytro dataset, triangles represent the MFFW dataset, and rectangles represent the MFI-WHU dataset. As shown in the figure, most metrics exhibit lower values when only the decoder is employed, suggesting that the decoder alone processes limited information in the fusion results. After incorporating the encoder, the metric values increase, demonstrating that the encoder enhances feature extraction and thereby improves the information content of the fused images. However, when the attention module is used alongside the decoder, all metric values reach their minimum, indicating that the attention module fails to correctly learn which features are important. In contrast, KCUNet achieves the highest values for most fusion metrics, confirming that the model’s design is well-justified.

4.5.3. Effects of Different Loss Functions

To investigate the impact of different loss components on decision graph inference, we conduct a comparative analysis using representative samples from the Lytro dataset, as illustrated in Figure 9. The first column displays the foreground-focused source image, followed by four decision graph results. These comparisons demonstrate that KCUNet’s performance, which employs a high-fidelity fusion strategy guided by decision diagrams, is directly influenced by the accuracy of these decision diagrams. The experimental results demonstrate that removing L E A leads to blurred focus boundaries, as this loss function plays a crucial role in capturing subtle edge features through its combination of multiscale structural similarity and regional consistency constraints. Without it, the network struggles to accurately detect fine edge details. Similarly, the absence of L M a s k results in increased misclassified pixels and coarse edges, since this component maintains structural smoothness and regional consistency in the decision graph. Its removal weakens global constraints on focus region delineation, producing irregular boundaries and scattered misclassified points. Eliminating L Q g causes a significant rise in decision graph noise, attributable to the loss of its gradient-aware mechanism that normally preserves high-frequency information. This degradation highlights its importance in processing texture-rich regions effectively. By contrast, when all loss components are incorporated, the decision graph achieves optimal performance, underscoring the complementary and essential nature of each term in ensuring high-quality results.

4.6. Hyperparameter Tuning Experiments

This section examines the influence of channel quantity and feature map size on the model’s fusion performance. To investigate these factors, three experimental configurations were designed:
(i) Progressive Channel Variation with Feature Map Scaling: In the encoder, the channel count progressively increases while the feature map dimensions are halved at each deeper layer. Conversely, in the decoder, the channel count gradually decreases, and the feature map size incrementally expands to restore the original dimensions.
(ii) Fixed Channel and Feature Map Configuration: Both the encoder and decoder maintain a constant channel count and feature map size throughout the network.
(iii) Constant Channel Count with Feature Map Scaling: The number of channels remains unchanged across all layers, while the encoder reduces the feature map size by half at each stage, and the decoder progressively upscales it to recover the original resolution.
Table 4 presents the detailed parameters of input channel counts and feature map sizes for each encoder and decoder in the three experimental configurations. Figure 10 displays the values of eight fusion metrics obtained by applying KCUNet and the three experimental models to the Lytro, MFFW, and MFI-WHU datasets. In the figure, circles, triangles, and squares denote the fusion performance curves for the Lytro, MFFW, and MFI-WHU datasets, respectively. The results demonstrate that maintaining a constant feature map size yields superior performance compared to reducing it, suggesting that downscaling feature maps leads to information degradation. Additionally, dynamically adjusting the channel count while preserving the feature map size further enhances fusion performance. To assess the impact of maintaining a constant feature map size on computational complexity, we conduct a comparative analysis between the traditional downsampling architecture and the proposed method using the Lytro dataset. The experiments show that the average fusion time of the traditional downsampling method is 1.21 s, which is only 15.7% of that of our method (7.69 s), but its fusion quality metrics lag behind across the board: our method achieves a significant improvement in key metrics, such as the increase in MI from 1.170 to 1.177 (+0.6%), the increase in Q G from 0.747 to 0.756 (+1.2%), and the increase in Q M from 2.460 to 2.508 (+2.0%). These findings suggest that the enhanced capabilities in focus boundary detection and detail preservation come at the cost of reduced computational efficiency.

5. Conclusions

This paper introduces KCUNet, a novel deep learning model for MFIF. Built upon the U-Net architecture, KCUNet incorporates parallel connections between the KAN layer and the convolutional layer during the encoding stage to efficiently extract image features while maintaining constant spatial dimensions of feature maps. This design preserves high-resolution details while progressively increasing channel counts to capture multi-level feature representations. During the decoding phase, the model adopts an inverse strategy by progressively decreasing channel counts while maintaining constant feature map sizes. This approach enables focused processing and effective fusion of key features derived from the encoder. The integration of a CGA mechanism in the attention module enhances edge-related feature emphasis, effectively reducing DSE and preserving critical edge information. In addition, the proposed hybrid loss function comprehensively evaluates model performance across multiple aspects, including edge alignment, mask prediction, and fused image quality. Extensive comparative experiments with 15 SOTA methods demonstrate KCUNet’s superior performance in both qualitative and quantitative assessments. While achieving notable results in MFIF tasks, the model shows limitations in certain complex scenarios (e.g., strong defocus background effects and multi-source image fusion) and requires further improvement in computational efficiency. These aspects represent important directions for future research and model optimization.
In conclusion, KCUNet achieves significant progress in the MFIF domain by delivering high-quality, fully focused images. Its success stems from the integration of KAN with convolutional layers, an innovative attention module, and a carefully designed loss function. Future work will focus on enhancing the network’s efficiency, extending its application to other image processing tasks, and exploring advanced attention mechanisms for further performance gains.

Author Contributions

Conceptualization, J.F. and X.N.; Data curation, R.W. (Ruiqing Wang); Formal analysis, X.L. and Z.Z.; Funding acquisition, S.H.; Investigation, S.T.; Methodology, R.W. (Ruxian Wang); Project administration, W.L.; Resources, J.W.; Software, R.W. (Ruxian Wang); Supervision, J.F.; Validation, R.W. (Ruiqing Wang); Visualization, X.N.; Writing—original draft, R.W. (Ruxian Wang); Writing—review and editing, J.F. All authors will be updated at each stage of manuscript processing, including submission, revision, and revision reminder, via emails from our system or the assigned Assistant Editor. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of China under Grant 62002208 and Grant 42271093.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The original data used during this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhang, X. Deep learning-based multi-focus image fusion: A survey and a comparative study. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 4819–4838. [Google Scholar] [CrossRef] [PubMed]
  2. Li, S.; Kang, X.; Hu, J.; Yang, B. Image matting for fusion of multi-focus images in dynamic scenes. Inf. Fusion 2013, 14, 147–162. [Google Scholar] [CrossRef]
  3. Liu, Y.; Liu, S.; Wang, Z. Multi-focus image fusion with dense SIFT. Inf. Fusion 2015, 23, 139–155. [Google Scholar] [CrossRef]
  4. Li, S.; Yang, B. Multifocus image fusion using region segmentation and spatial frequency. Image Vis. Comput. 2008, 26, 971–979. [Google Scholar] [CrossRef]
  5. Duan, J.; Chen, L.; Chen, C.P. Multifocus image fusion with enhanced linear spectral clustering and fast depth map estimation. Neurocomputing 2018, 318, 43–54. [Google Scholar] [CrossRef]
  6. Li, S.; Kwok, J.T.; Wang, Y. Combination of images with diverse focuses using the spatial frequency. Image Vis. Comput. 2001, 2, 169–176. [Google Scholar] [CrossRef]
  7. Bai, X.; Zhang, Y.; Zhou, F.; Xue, B. Quadtree-based multi-focus image fusion using a weighted focus-measure. Image Vis. Comput. 2015, 22, 105–118. [Google Scholar] [CrossRef]
  8. Liu, Y.; Chen, X.; Peng, H.; Wang, Z. Multi-focus image fusion with a deep convolutional neural network. Inf. Fusion 2017, 36, 191–207. [Google Scholar] [CrossRef]
  9. Ma, B.; Yin, X.; Wu, D.; Shen, H.; Ban, X.; Wang, Y. End-to-end learning for simultaneously generating decision map and multi-focus image fusion result. Neurocomputing 2022, 470, 204–216. [Google Scholar] [CrossRef]
  10. Li, J.; Guo, X.; Lu, G.; Zhang, B.; Xu, Y.; Wu, F.; Zhang, D. DRPL: Deep regression pair learning for multi-focus image fusion. IEEE Trans. Image Process. 2020, 29, 4816–4831. [Google Scholar] [CrossRef] [PubMed]
  11. Ma, B.; Zhu, Y.; Yin, X.; Ban, X.; Huang, H.; Mukeshimana, M. Sesf-fuse: An unsupervised deep model for multi-focus image fusion. Neural Comput. Appl. 2021, 33, 5793–5804. [Google Scholar] [CrossRef]
  12. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; proceedings, part III 18. Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar] [CrossRef]
  13. Çiçek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning dense volumetric segmentation from sparse annotation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2016: 19th International Conference, Athens, Greece, 17–21 October 2016; Proceedings, Part II 19. Springer: Berlin/Heidelberg, Germany, 2016; pp. 424–432. [Google Scholar] [CrossRef]
  14. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support: 4th International Workshop, DLMIA 2018, and 8th International Workshop, ML-CDS 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, 20 September 2018; proceedings 4. Springer: Berlin/Heidelberg, Germany, 2018; pp. 3–11. [Google Scholar] [CrossRef]
  15. Cao, H.; Wang, Y.; Chen, J.; Jiang, D.; Zhang, X.; Tian, Q.; Wang, M. Swin-unet: Unet-like pure transformer for medical image segmentation. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 205–218. [Google Scholar] [CrossRef]
  16. Burt, P.J.; Adelson, E.H. Merging images through pattern decomposition. In Proceedings of the Applications of Digital Image Processing VIII, San Diego, CA, USA, 20–22 August 1985; Volume 575, pp. 173–181. [Google Scholar] [CrossRef]
  17. Toet, A. Image fusion by a ratio of low-pass pyramid. Pattern Recognit. Lett. 1989, 9, 245–253. [Google Scholar] [CrossRef]
  18. Lewis, J.J.; O’ Callaghan, R.J.; Nikolov, S.G.; Bull, D.R.; Canagarajah, N. Pixel-and region-based image fusion with complex wavelets. Inf. Fusion 2007, 8, 119–130. [Google Scholar] [CrossRef]
  19. Zhang, Q.; Guo, B.l. Multifocus image fusion using the nonsubsampled contourlet transform. Signal Process. 2009, 89, 1334–1346. [Google Scholar] [CrossRef]
  20. Yang, B.; Li, S. Multifocus image fusion and restoration with sparse representation. IEEE Trans. Instrum. Meas. 2009, 59, 884–892. [Google Scholar] [CrossRef]
  21. He, K.; Sun, J.; Tang, X. Guided image filtering. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 1397–1409. [Google Scholar] [CrossRef] [PubMed]
  22. Zhang, Y.; Bai, X.; Wang, T. Boundary finding based multi-focus image fusion through multi-scale morphological focus-measure. Inf. Fusion 2017, 35, 81–101. [Google Scholar] [CrossRef]
  23. Panigrahy, C.; Seal, A.; Mahato, N.K. Fractal dimension based parameter adaptive dual channel PCNN for multi-focus image fusion. Opt. Lasers Eng. 2020, 133, 106141. [Google Scholar] [CrossRef]
  24. Kurban, R. Gaussian of differences: A simple and efficient general image fusion method. Entropy 2023, 25, 1215. [Google Scholar] [CrossRef] [PubMed]
  25. Zhang, Y.; Liu, Y.; Sun, P.; Yan, H.; Zhao, X.; Zhang, L. IFCNN: A general image fusion framework based on convolutional neural network. Inf. Fusion 2020, 54, 99–118. [Google Scholar] [CrossRef]
  26. Jung, H.; Kim, Y.; Jang, H.; Ha, N.; Sohn, K. Unsupervised deep image fusion with structure tensor representations. IEEE Trans. Image Process. 2020, 29, 3845–3858. [Google Scholar] [CrossRef] [PubMed]
  27. Yan, X.; Gilani, S.Z.; Qin, H.; Mian, A. Structural similarity loss for learning to fuse multi-focus images. Sensors 2020, 20, 6647. [Google Scholar] [CrossRef] [PubMed]
  28. Ma, X.; Wang, Z.; Hu, S.; Kan, S. Multi-focus image fusion based on multi-scale generative adversarial network. Entropy 2022, 24, 582. [Google Scholar] [CrossRef] [PubMed]
  29. Cheng, C.; Xu, T.; Wu, X.J. MUFusion: A general unsupervised image fusion network based on memory unit. Inf. Fusion 2023, 92, 80–92. [Google Scholar] [CrossRef]
  30. Ma, H.; Liao, Q.; Zhang, J.; Liu, S.; Xue, J.H. An α-matte boundary defocus model-based cascaded network for multi-focus image fusion. IEEE Trans. Image Process. 2020, 29, 8668–8679. [Google Scholar] [CrossRef] [PubMed]
  31. Xu, S.; Ji, L.; Wang, Z.; Li, P.; Sun, K.; Zhang, C.; Zhang, J. Towards reducing severe defocus spread effects for multi-focus image fusion via an optimization based strategy. IEEE Trans. Comput. Imaging 2020, 6, 1561–1570. [Google Scholar] [CrossRef]
  32. Wang, Y.; Xu, S.; Liu, J.; Zhao, Z.; Zhang, C.; Zhang, J. MFIF-GAN: A new generative adversarial network for multi-focus image fusion. Signal Process. Image Commun. 2021, 96, 116295. [Google Scholar] [CrossRef]
  33. Liu, Z.; Wang, Y.; Vaidya, S.; Ruehle, F.; Halverson, J.; Soljačić, M.; Hou, T.Y.; Tegmark, M. KAN: Kolmogorov-Arnold Networks. In Proceedings of the Thirteenth International Conference on Learning Representations (ICLR 2025), Singapore, 24 April 2025. [Google Scholar]
  34. Chen, Z.; He, Z.; Lu, Z.M. DEA-Net: Single image dehazing based on detail-enhanced convolution and content-guided attention. IEEE Trans. Image Process. 2024, 33, 1002–1015. [Google Scholar] [CrossRef]
  35. Everingham, M.; Eslami, S.A.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The pascal visual object classes challenge: A retrospective. Int. J. Comput. Vis. 2015, 111, 98–136. [Google Scholar] [CrossRef]
  36. Nejati, M.; Samavi, S.; Shirani, S. Multi-focus image fusion using dictionary-based sparse representation. Inf. Fusion 2015, 25, 72–84. [Google Scholar] [CrossRef]
  37. Zhang, H.; Le, Z.; Shao, Z.; Xu, H.; Ma, J. MFF-GAN: An unsupervised generative adversarial network with adaptive and gradient joint constraints for multi-focus image fusion. Inf. Fusion 2021, 66, 40–53. [Google Scholar] [CrossRef]
  38. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; proceedings, part v 13. Springer: Berlin/Heidelberg, Germany, 2014; pp. 740–755. [Google Scholar] [CrossRef]
  39. Cai, J.; Gu, S.; Zhang, L. Learning a deep single image contrast enhancer from multi-exposure images. IEEE Trans. Image Process. 2018, 27, 2049–2062. [Google Scholar] [CrossRef] [PubMed]
  40. Liu, Z.; Blasch, E.; Xue, Z.; Zhao, J.; Laganiere, R.; Wu, W. Objective assessment of multiresolution image fusion algorithms for context enhancement in night vision: A comparative study. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 34, 94–109. [Google Scholar] [CrossRef] [PubMed]
  41. Qu, G.; Zhang, D.; Yan, P. Information measure for performance of image fusion. Electron. Lett. 2002, 38, 313–315. [Google Scholar] [CrossRef]
  42. Stathaki, T. Image Fusion: Algorithms and Applications; Elsevier: Amsterdam, The Netherlands, 2011. [Google Scholar]
  43. Chen, Y.; Blum, R.S. A new automated quality assessment algorithm for image fusion. Image Vis. Comput. 2009, 27, 1421–1432. [Google Scholar] [CrossRef]
  44. Yang, C.; Zhang, J.Q.; Wang, X.R.; Liu, X. A novel similarity based quality metric for image fusion. Inf. Fusion 2008, 9, 156–160. [Google Scholar] [CrossRef]
  45. Wang, P.w.; Liu, B. A novel image fusion metric based on multi-scale analysis. In Proceedings of the 2008 9th International Conference on Signal Processing, Beijing, China, 26–29 October 2008; pp. 965–968. [Google Scholar] [CrossRef]
  46. Xydeas, C.S.; Petrovic, V. Objective image fusion performance measure. Electron. Lett. 2000, 36, 308–309. [Google Scholar] [CrossRef]
  47. Zhao, W.; Wang, D.; Lu, H. Multi-focus image fusion with a natural enhancement via a joint multi-level deeply supervised convolutional neural network. IEEE Trans. Circuits Syst. Video Technol. 2018, 29, 1102–1115. [Google Scholar] [CrossRef]
  48. Amin-Naji, M.a.; Aghagolzadeh, A. Multi-focus image fusion in DCT domain using variance and energy of Laplacian and correlation coefficient for visual sensor networks. J. AI Data Min. 2018, 6, 233–250. [Google Scholar] [CrossRef]
  49. Amin-Naji, M.; Aghagolzadeh, A.; Ezoji, M. Ensemble of CNN for multi-focus image fusion. Inf. Fusion 2019, 51, 201–214. [Google Scholar] [CrossRef]
  50. Xu, H.; Ma, J.; Jiang, J.; Guo, X.; Ling, H. U2Fusion: A unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 502–518. [Google Scholar] [CrossRef] [PubMed]
  51. Ma, J.; Tang, L.; Fan, F.; Huang, J.; Mei, X.; Ma, Y. SwinFusion: Cross-domain Long-range Learning for General Image Fusion via Swin Transformer. IEEE/CAA J. Autom. Sin. 2022, 9, 1200–1217. [Google Scholar] [CrossRef]
  52. Karacan, L. Multi-image transformer for multi-focus image fusion. Signal Process. Image Commun. 2023, 119, 117058. [Google Scholar] [CrossRef]
  53. Hu, X.; Jiang, J.; Liu, X.; Ma, J. ZMFF: Zero-shot multi-focus image fusion. Inf. Fusion 2023, 92, 127–138. [Google Scholar] [CrossRef]
  54. Li, M.; Pei, R.; Zheng, T.; Zhang, Y.; Fu, W. FusionDiff: Multi-focus image fusion using denoising diffusion probabilistic models. Expert Syst. Appl. 2024, 238, 121664. [Google Scholar] [CrossRef]
Figure 1. (a) No DSE and (b) with DSE. The red box in the upper left corner shows a zoomed-in view of the water droplet.
Figure 1. (a) No DSE and (b) with DSE. The red box in the upper left corner shows a zoomed-in view of the water droplet.
Entropy 27 00785 g001
Figure 2. Overall integration strategy of KCUNet. The network takes as input a pair of source images in the RGB format, where Source A captures the foreground in sharp focus, while Source B highlights the background with clarity. Conv, BN, and ReLU represent the convolutional layer, batchnorm layer, and rectified linear unit, respectively. The number following each layer indicates its output channels.
Figure 2. Overall integration strategy of KCUNet. The network takes as input a pair of source images in the RGB format, where Source A captures the foreground in sharp focus, while Source B highlights the background with clarity. Conv, BN, and ReLU represent the convolutional layer, batchnorm layer, and rectified linear unit, respectively. The number following each layer indicates its output channels.
Entropy 27 00785 g002
Figure 3. Detailed architecture of the encoder block. The kernel, stride, and padding sizes of the convolutional layer are provided by the array in brackets in each block.
Figure 3. Detailed architecture of the encoder block. The kernel, stride, and padding sizes of the convolutional layer are provided by the array in brackets in each block.
Entropy 27 00785 g003
Figure 4. The architecture of the S–CGANet block. Laplace denotes the application of the Laplace operator to generate a bounded attention map, and Reverse denotes the generation of reverse attention by a one-subtraction operation. H, W, and N denote the height, width, and channel counts of the input. Orange arrows denote feature maps from the encoder, black arrows represent high-frequency features extracted by the Laplace pyramid, and blue arrows indicate predicted features from the decoder.
Figure 4. The architecture of the S–CGANet block. Laplace denotes the application of the Laplace operator to generate a bounded attention map, and Reverse denotes the generation of reverse attention by a one-subtraction operation. H, W, and N denote the height, width, and channel counts of the input. Orange arrows denote feature maps from the encoder, black arrows represent high-frequency features extracted by the Laplace pyramid, and blue arrows indicate predicted features from the decoder.
Entropy 27 00785 g004
Figure 5. Training examples derived from the VOC2012 dataset. The top row displays the image with near-focus, while the bottom row shows the image with far-focus.
Figure 5. Training examples derived from the VOC2012 dataset. The top row displays the image with near-focus, while the bottom row shows the image with far-focus.
Entropy 27 00785 g005
Figure 6. The fusion results for all algorithms with detail enlargement on Lytro (3rd), Lytro (17th), MFFW (4th), and MFI-WHU (8th).
Figure 6. The fusion results for all algorithms with detail enlargement on Lytro (3rd), Lytro (17th), MFFW (4th), and MFI-WHU (8th).
Entropy 27 00785 g006aEntropy 27 00785 g006b
Figure 7. The differences in the experimental results. Red represents the Lytro dataset, blue represents the MFFW dataset, and green represents the MFI-WHU dataset.
Figure 7. The differences in the experimental results. Red represents the Lytro dataset, blue represents the MFFW dataset, and green represents the MFI-WHU dataset.
Entropy 27 00785 g007
Figure 8. The values of the eight fusion metrics when fusing the different datasets using KCUNet and each of these three approaches.
Figure 8. The values of the eight fusion metrics when fusing the different datasets using KCUNet and each of these three approaches.
Entropy 27 00785 g008
Figure 9. Comparison of decision maps generated by different loss functions. (a) with the complete loss function, (b) without edge alignment loss, (c) without mask loss, and (d) without quality loss generated using different loss function components.
Figure 9. Comparison of decision maps generated by different loss functions. (a) with the complete loss function, (b) without edge alignment loss, (c) without mask loss, and (d) without quality loss generated using different loss function components.
Entropy 27 00785 g009
Figure 10. The values of the eight fusion metrics when fusing different datasets using KCUNet and the three sets of experimental models, respectively.
Figure 10. The values of the eight fusion metrics when fusing different datasets using KCUNet and the three sets of experimental models, respectively.
Entropy 27 00785 g010
Table 1. Output channel specifications across model modules.
Table 1. Output channel specifications across model modules.
EncoderS–CGANetDecoder
Eecoder 116KAN layer3S–CGANet 116Conv layer 116Decoder 1-11
3 × 3 Conv + 3 × 3 Conv3
Conv layer 21Decoder 1-216
3 × 3 Conv + 1 × 1 Conv16
Eecoder 232KAN Layer16S–CGANet 232Conv layer 132Decoder 2-11
3 × 3 Conv + 3 × 3 Conv16
Conv layer 21Decoder 2-232
3 × 3 Conv + 1 × 1 Conv32
Eecoder 348KAN Layer32S–CGANet 348Conv layer 148Decoder 3-11
3 × 3 Conv + 3 × 3 Conv32
Conv layer 21Decoder 3-248
3 × 3 Conv + 1 × 1 Conv48
Eecoder 464KAN Layer48/Decoder 4-11
3 × 3 Conv + 3 × 3 Conv48
Decoder 4-264
3 × 3 Conv + 1 × 1 Conv64
Table 2. Average scores of the fused images of each algorithm for Lytro, MFFW, and MFI-WHU datasets. The red number is the best, the blue number is the second best, and the green number is the third best.
Table 2. Average scores of the fused images of each algorithm for Lytro, MFFW, and MFI-WHU datasets. The red number is the best, the blue number is the second best, and the green number is the third best.
(a) Lytro
MI NCIE Q G GLD SSIM Q CB MSD Q M Time (s)
BFMF [22]1.1207860.8415650.7486620.8299800.9779150.7978080.3971882.0564380.82
DCT-EOL [48]1.1353650.8419710.7509890.8255810.9756090.7959570.3978042.2086930.34
CNN [8]1.1096870.8401680.7515970.8327010.9780430.7998810.3962862.04122089.34
DRPL [10]1.0915450.8387830.7525930.8398410.9817420.7942020.3966231.7029780.29
ECNN [49]1.1273250.8413590.7477710.8180950.9756740.7958080.3969672.20038372.85
FusionDiff [54]1.1339190.8419580.7510380.8259590.9755110.7950620.3975902.2128815.63
IFCNN [25]0.9268480.8292660.7251880.8026730.9515720.7234200.3824351.1859220.08
MFF-GAN [37]0.8046530.8237090.6592470.7124640.8821230.6456810.3781130.6411941.06
MFIF-GAN [32]1.1313160.8417180.7525630.8306680.9793510.8005590.3981512.1799601.26
MiT [52]1.1496380.8426430.7509320.8317390.9689420.7859130.3981522.368812.89
Swin Fusion [51]1.1535130.8413520.7533910.8401630.9814110.7938560.3990572.175921.62
SESF [11]1.1022580.8397210.7464650.8250420.9757650.7962750.3947752.0947750.31
U2Fusion [50]0.7724680.8220850.6092940.6657090.7912300.5682140.3994140.4984872.57
MMF-Net [30]0.9719460.8321440.7228710.812610.9514210.7508050.3837431.392920.51
ZMFF [50]0.8838460.8270760.7030350.7852890.9313680.7411610.3918490.599906359.26
Proposed1.1769070.8446140.7562600.8329720.9836030.8028010.4007742.5080827.69
(b) MFFW
MINCIE Q G GLDSSIM Q CB MSD Q M Time (s)
BFMF [22]0.7815350.8206000.6189440.5575550.9327780.6803190.3669490.4714191.37
DCT-EOL [48]0.7367330.8188800.6238270.5394490.7635400.6652720.3577930.4964580.56
CNN [8]0.7677080.8199030.6278820.5633300.7826540.6750260.3640130.476062119.71
DRPL [10]0.8264840.8223310.6805200.6462200.8400190.6957050.3740990.7965020.74428
ECNN [49]0.7448420.8191790.6228120.5477080.7598780.6693050.3596070.495281102.47
FusionDiff [54]1.0749700.8192790.7227780.6724280.8722190.7045300.3599231.9441506.87
IFCNN [25]0.7501730.8191720.6062580.5263170.7580450.6261190.3615000.3935330.13
MFF-GAN [37]0.7065320.8178210.5715070.4809300.7128130.5841290.3670980.3417731.71
MFIF-GAN [32]0.7715740.8200450.6305380.5713610.7903280.6881560.3645160.4920211.88
MiT [52]0.9455370.8213660.7181290.7288540.9016840.7023520.3641242.285314.37
SwinFusion [51]1.0163550.8317120.7126340.6953880.9246380.7033840.3653212.153741.83
SESF [11]0.7383240.8188980.6247140.5514590.8272080.6733020.3542210.4745400.74
U2Fusion [50]0.6970700.8170800.5360630.4699570.6671860.5494750.3897980.3113363.94
MMF-Net [30]0.8158160.8219020.6836510.6459860.8380080.6906070.3720840.7636921.15
ZMFF [50]0.7639810.8198970.6540510.6232210.8456440.67512450.3732130.478555373.33
Proposed1.1488290.8399200.7320710.6974610.9367930.7148250.4050852.4981528.87
(c) MFI-WHU
MINCIE Q G GLDSSIM Q CB MSD Q M Time (s)
BFMF [22]1.1628220.8445790.7264020.798410.9868690.8314810.3870282.2672571.03
DCT-EOL [48]1.176270.8449350.7221510.7613140.9798790.8113060.3900872.3842310.39
CNN [8]1.156160.8440160.7293260.7992110.987450.8310870.3864532.25610992.52
DRPL [10]1.0892680.8389040.7254690.7936220.9851090.8231780.3827711.583520.56
ECNN [49]1.1770140.8452990.7293970.7944530.9872180.8302740.3882922.38092683.34
FusionDiff [54]1.1610530.8442750.7292860.7983140.9871220.8318010.3868092.3044517.25
IFCNN [25]0.8977870.82780.6944250.7813410.9577260.7404760.3638741.0023060.22
MFF-GAN [37]0.7573890.821760.6380630.7166020.8853040.6362030.3517680.5262231.41
MFIF-GAN [32]1.1670110.8451520.7248620.7825650.9825380.824580.3881392.3181451.09
MiT [52]1.1622590.8371360.7261490.7962720.9812220.8119810.3887312.381853.53
SwinFusion [51]1.0892680.8389040.7254690.7936220.9851020.8231780.3896681.583521.73
SESF [11]1.149440.8436430.722590.7954140.9850440.8238070.3872532.2693560.67
U2Fusion [50]0.6796150.8186770.5436440.6157930.7585410.5135380.3661020.3401313.17
MMF-Net [30]1.063770.8371740.7232020.7964560.9836240.8186790.380841.3734670.59
ZMFF [50]0.7755620.8223460.6331810.6936810.6778820.7136610.3619730.370791389.21
Proposed1.2053820.8477250.7300770.7945140.9873270.8308580.3923062.4775467.97
Table 3. Average scores of the fused images of each algorithm for Lytro, MFFW, and MFI-WHU datasets. The red number is the best, the yellow number is the second best, and the green number is the third best.
Table 3. Average scores of the fused images of each algorithm for Lytro, MFFW, and MFI-WHU datasets. The red number is the best, the yellow number is the second best, and the green number is the third best.
MINCIE Q G GLDSSIM Q CB MSD Q M
LytroProposed1.1769070.8446140.7562600.8329720.9836030.8028010.4007742.508082
No KAN1.1711630.8441480.7446050.8169120.9786400.7930310.3987912.411444
Difference0.0057440.0004660.0116550.0160600.0049620.0097710.0019830.096637
MFFWProposed1.1488290.8399200.7320710.6974610.9367930.7148250.4050852.498152
No KAN1.1466060.8395390.7255260.6849550.9344550.7054030.4039142.453557
Difference0.0022230.0003810.0065450.0125060.0023390.0094220.0011710.044595
MFI-WHUProposed1.2053820.8477250.7300770.7945140.9873270.8308580.3923062.477546
No KAN1.2002480.8469460.7252820.7899170.9848020.8275450.3896692.437547
Difference0.0051340.0007790.0047950.0045970.0025250.0033130.0026370.040000
Table 4. Parameters specific to each encoder and decoder, including input channel count and size of input feature maps. H and W are the feature map height and width, respectively.
Table 4. Parameters specific to each encoder and decoder, including input channel count and size of input feature maps. H and W are the feature map height and width, respectively.
ExperimentParameterEncoder1Encoder2Encoder3Encoder4Decoder4Decoder3Decoder2Decoder1
ProposedChannels316324864483216
SizeH,WH,WH,WH,WH,WH,WH,WH,W
Experiment1Channels316324864483216
SizeH,WH/2,W/2,H/4,W/4H/8,W/8H/8,W/8H/4,W/4H/2,W/2H,W
Experiment2Channels33333333
SizeH,WH,WH,WH,WH,WH,WH,WH,W
Experiment3Channels33333333
SizeH,WH/2,W/2,H/4,W/4H/8,W/8H/8,W/8H/4,W/4H/2,W/2H,W
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Fang, J.; Wang, R.; Ning, X.; Wang, R.; Teng, S.; Liu, X.; Zhang, Z.; Lu, W.; Hu, S.; Wang, J. KCUNET: Multi-Focus Image Fusion via the Parallel Integration of KAN and Convolutional Layers. Entropy 2025, 27, 785. https://doi.org/10.3390/e27080785

AMA Style

Fang J, Wang R, Ning X, Wang R, Teng S, Liu X, Zhang Z, Lu W, Hu S, Wang J. KCUNET: Multi-Focus Image Fusion via the Parallel Integration of KAN and Convolutional Layers. Entropy. 2025; 27(8):785. https://doi.org/10.3390/e27080785

Chicago/Turabian Style

Fang, Jing, Ruxian Wang, Xinglin Ning, Ruiqing Wang, Shuyun Teng, Xuran Liu, Zhipeng Zhang, Wenfeng Lu, Shaohai Hu, and Jingjing Wang. 2025. "KCUNET: Multi-Focus Image Fusion via the Parallel Integration of KAN and Convolutional Layers" Entropy 27, no. 8: 785. https://doi.org/10.3390/e27080785

APA Style

Fang, J., Wang, R., Ning, X., Wang, R., Teng, S., Liu, X., Zhang, Z., Lu, W., Hu, S., & Wang, J. (2025). KCUNET: Multi-Focus Image Fusion via the Parallel Integration of KAN and Convolutional Layers. Entropy, 27(8), 785. https://doi.org/10.3390/e27080785

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop