Next Article in Journal
Machine Learning-Based Risk Stratification for Sudden Cardiac Death Using Clinical and Device-Derived Data
Previous Article in Journal
Time Series Models of the Human Heart in Patients with Heart Failure: Toward a Digital Twin Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Scale Interactive Network with Color Attention for Low-Light Image Enhancement

1
Guangdong Cardiovascular Institute, Guangdong Provincial People’s Hospital, Guangdong Academy of Sciences, Guangzhou 510080, China
2
Department of Radiology, Guangdong Provincial People’s Hospital (Guangdong Academy of Medical Sciences), Southern Medical University, Guangzhou 510080, China
3
Guangdong Provincial Key Laboratory of Artificial Intelligence in Medical Image Analysis and Application, Guangzhou 510080, China
4
School of Computer and Information Security, Guilin University of Electronic Technology, Guilin 541004, China
*
Author to whom correspondence should be addressed.
Sensors 2026, 26(1), 83; https://doi.org/10.3390/s26010083 (registering DOI)
Submission received: 10 November 2025 / Revised: 12 December 2025 / Accepted: 18 December 2025 / Published: 22 December 2025
(This article belongs to the Section Sensing and Imaging)

Abstract

Enhancing low-light images is crucial in computer vision applications. Most existing learning-based models often struggle to balance light enhancement and color correction, while images typically contain different types of information at different levels. Hence, we proposed a multi-scale interactive network with color attention named MSINet to effectively explore these different types of information for lowlight image enhancement (LLIE) tasks. Specifically, the MSINet first employs the CNN-based branch built upon stacked residual channel attention blocks (RCABs) to fully explore the image local features. Meanwhile, the Transformer-based branch constructed by Transformer blocks contains cross-scale attention (CSA) and multi-head self-attention (MHSA) to mine the global features. Notably, the local and global features extracted by each RCAB and Transformer block are interacted with by the fusion module. Additionally, the color correction branch (CCB) based upon self-attention (SA) can learn the color distribution information from the lowlight input for further guaranteeing the color fidelity of the final output. Extensive experiments have demonstrated that our proposed MSINet outperforms state-of-the-art LLIE methods in light enhancement and color correction.

1. Introduction

Images captured under low-light environments typically suffer from low brightness, low signal-to-noise ratio, and color distortion, which can significantly reduce the performance of high-level vision tasks such as object detection, segmentation, and scene understanding [1,2]. In the early stage, charge-coupled devices (CCD), complementary metal-oxide-semiconductor (CMOS) sensors and other advanced specialised-hardwares are used to obtain high-quality images under suboptimal lighting conditions [3]. But these hardwares are expensive and have operational complexity, limiting their application in the realworld. As a result, researchers gradually focus on designing software-driven low-light image enhancement (LLIE) approaches to restore degraded images with low illumination [4].
Numerous traditional LLIE methods containing histogram equalization (HE) [5], Retinex-based methods [6], and domain transformation-based methods [7] have been proposed for promoting the brightness and contrast of lowlight images. HE-based methods directly adjust the image pixels in a pixel-to-pixel manner to improve the contrast and illumination of lowlight images, but they may lead to over-enhancement and detail loss. Retinex-based methods typically decompose images into illumination and reflectance components to enhance the images’ visibility depending on the specific guidelines. For example, single-scale Retinex (SSR) and multi-scale Retinex with color restoration (MSRCR) [8] try to analyze lowlight images. They tend to introduce noticeable color deviations and blurry detail in enhanced images due to inaccurate estimation of illumination and reflectance components. The domain transformation-based methods transfer the original image into gradient, wavelet domains for detail enhancements, but suffer from a dramatic drop in restoration performance under complexity scenarios.
With the advancement of computing resources, learning-based methods [2,4,9], such as convolutional neural networks (CNNs), generative adversarial networks (GANs), and diffusion model, have proven effective in various computer vision tasks, with the help of their powerful feature extraction and representation capability. LLNet [10], the most groundbreaking learning-based LLIE work, stacked sparse denoising autoencoders for light improvement and denoising simultaneously. In recent years, researchers have proposed many advanced data-driven methods, such as JED [11], SID [12], IPT [13], PairLIE [14], etc., to entirely utilize the inherent relations among low-/normal-illumination images in training datasets, and these LLIE models exhibit outstanding performance in generating visually pleasing results with clearer details. Most above-listed LLIE models gain improvement by injecting the image pyramid [15], Retinex theory [16], meta-learning strategy [9], and other advanced technologies or substantially increasing the depth of the network. However, these LLIE methods seldom utilize the correlation and complementarity of global and local features. In addition, most existing methods only focus on enhancing brightness and contrast, while ignoring the color distribution of the original image.
To address these challenges, this paper proposes a multi-scale interactive network with color attention for lighting up degraded images captured under suboptimal lighting conditions, named MSINet. The proposed MSINet can simultaneously capture local textures and global context to effectively enhance the brightness and details of the inputs as well as remove their color deviation. Our MSINet contains three parallel branches, i.e., CNN-based branch, Transformer-based branch, and Color correction branch. The former can extract local features from the original input by stacking CNN blocks, and the Transformer-based branch based on stacked cross-scale Transformer blocks, including multi-head self-attention (MHSA) and cross-scale attention (CSA), can dig up global features. Meanwhile, the fusion block is introduced to realize the interaction of global and local features for analyzing their correlation and complementarity. The color correction branch built upon the self-attention can fully explore the color distribution in the original images.
We emphasize the primary contributions of this work as follows.
(1)
We propose an efficient and robust LLIE method named MSINet, which integrates CNN and Transformer structures for balancing local detail extraction and global feature encoding. Extensive experiments show our MSINet can generate visually pleasing images.
(2)
We proposed a cross-scale Transformer module combining cross-scale attention (CSA) and multi-head self-attention (MHSA) to enhance the model’s multi-scale feature learning. Meanwhile, the CNN-based branch can fully explore the local feature.
(3)
We proposed a self-attention-based color correction branch to dig up color distribution weighting for color correction in the LLIE tasks. Additionally, we design a fusion block to analyze the correlation and complementarity of global–local features.
We demonstrate the organization of the remainder of this paper as follows. The previous LLIE methods related to our MSINet are reviewed in Section 2. In Section 3, the architecture of our model is demonstrated. In Section 4, we evaluate the proposed method on public benchmarks. Additionally, the analysis of detail enhancement, computational complexity, ablation study, and applications as well as limitations and future work are presented. Finally, the conclusions are given.

2. Related Works

2.1. CNN-Based LLIE Enhancement

CNN-based LLIE approaches [17,18] have demonstrated dramatic improvements due to their powerful nonlinear representation ability, which can learn nonlinear mapping from the lowlight inputs to their corresponding normal-light versions. For example, Liu et al. [7] designed a Guided Filter-inspired Network (GFNet) for lighting up low-light RAW image enhancement in a guided filter (GF)-like manner. Lim et al. [19] applied the Laplacian pyramid in a multi-scale network to adjust global illumination and restore fine details. Zhang et al. [20] proposed a deep color-consistent network to enhance the naturalness of the image by preserving color information. Although the above-listed methods can light up low-illumination images, they suffer from poor interpretability. Hence, the physical model including Retinex, dehazing model, is used to enhance the interpretability of the LLIE models [17,21]. Wei et al. [18] first injected the Retinex theory into the traditional convolutional neural networks (CNNs) to develop an end-to-end architecture called RetinexNet. Subsuquently, URetinex-Net [16], including initialization, unfolding optimization, and illumination adjustment modules as well as CRetinex [22] decomposing an image into reflectance, color shift, and illumination, is proposed for the LLIE tasks. Recently, multi-scale feature fusion [23,24], perceptual fidelity estimation techniques [25], meta-learning [9], collaborative learning [26], and other advanced technologies are employed to design more efficient LLIE methods. But most of these existing LLIE methods heavily rely on the high-quality paired datasets, limiting their applicability in the real world. To address this challenge, Jiang et al. [27] developed an unsupervised method based on GANs, which integrates a global–local discriminator to improve performance. Guo et al. [28] used a lightweight network that learns enhancement curves and iteratively enhances images. Yao et al. [29] presented a gradient-aware and contrastive-adaptive (GACA) learning framework by estimating more accurate gradient information and introducing a regularization constraint. Yan et al. [30] proposed a Horizontal/Vertical-Intensity (HVI) color space based on the polarized HS maps and learnable intensity. These self-/un-supervised learning-based methods inevitably generate color deviation and blurry details in enhanced images.

2.2. Transformer-Based LLIE Methods

The vision transformer (ViT) model based on the self-attention mechanism demonstrates powerful potential in capturing the global dependency of the input feature [31,32]. Cai et al. [33] proposed a one-stage Retinex-based transformer to estimate the illumination information for lighting up the low-light image. Zhang et al. [34] injected the multiple heads into a single network to perform denoising, luminance adjustment, refinement, and detail enhancement for the LLIE. Wu et al. [35] presented two key innovations: an improved Gaussian filtering-based image enhancement module and a hierarchical feature extraction network. Dang et al. [36] employed a lightweight model PPformer to extract both local and non-local information, and further fused them by the dual cross-attention mechanism for the LLIE tasks. Pei et al. [37] proposed a fast Fourier transform embedded noise-aware CNN-Transformer, which removes noise in both the spatial and frequency domains. Brateanu et al. [38] proposed a lightweight transformer-based network containing Channel-Wise Denoiser (CWD), Multi-Stage Squeeze and Excite Fusion (MSEF), and Multi-Headed Self-Attention (MHSA) for promoting the quality of low-light images. Wen et al. [39] performed the pure CNN-based estimator to generate a light-up feature map and a lit-up image, and the restorer based on the U-shaped network equipped with an Illumination-Guided Dual Attention Block (IGDAB) was used to denoise the lit-up image. Dong et al. [40] presented a new multi-scale CNN-Transformer hybrid framework guided by structure priors. Notably, the illumination-invariant edge detectors based on the UNet encoder–decoder architecture with the CNN-Transformer hybrid structure were used to extract robust structure priors. Most ViT-based methods focus on channel modeling to reduce expensive computational costs; however, they introduce spatial illumination inconsistencies, artifacts, and blurry details in restored images.

2.3. Diffusion-Based LLIE Methods

The diffusion model is a generative framework that gradually adds noise to disrupt the data structure and then learns the reverse denoising process, which has been wildly used in image restoration [41,42], medical image processing [43], and so on. Yi et al. [44] formulated the low-light image enhancement problem into Retinex decomposition and conditional image generation. Subsequently, they proposed a LLIE method named Diff-Retinex++ containing the Denoising Diffusion Model (DDM) and the Retinex-Driven Mixture of Experts Model (RMoE) [45]. Lin et al. [46] proposed the Attribute Guidance Diffusion framework (AGLLDiff), a training-free method for effective real-world LLIE. Yang et al. [47] employed the Diffusion-guided Degradation Calibration (DDC) module to narrow the gap between real-world and training low-light degradation, further developing the Fine-grained Target domain Distillation (FTD) module to find a more visual-friendly solution space. Huang et al. [48] integrated a size-agnostic diffusion process with a reverse process reconstruction loss to more accurately recover fine details. Hu et al. [49] employed the conditional correlation module (CCM) to effectively integrate color and illumination priors, and the residual decomposition network (RDN) was introduced to generate the reflectance image representing the color object. Jiang et al. [50] injected the wavelet transformation into the conditional diffusion model to achieve stable denoising and reduce randomness during inference. Jin et al. [51] proposed Dual-Conditional Guidance Sparse Diffusion (DCGSD), a physically explainable and prior guidance model, to light up low-illumination images. Although diffusion-based LLIE approaches can generate visually pleasing images from the lowlight inputs, they suffer from time-consuming, excessive computational resource consumption, and unstable restoration.

3. Methodology

We present the motivation for our proposed method in Figure 1; it can effectively enhance low-light images and preserve color fidelity and fine details. The MSINet is composed of three main components: a CNN-based branch, a Transformer-based branch, and a color correction branch. Specifically, the CNN-based branch first extracts shallow features through a convolutional layer with a size of 1 × 1 . Then, to fully explore the local image features, the shallow features is fed to the successive stacked Residual Channel Attention Blocks (RCABs) (pure CNN blocks). This stage can be formulated as
F L = H CNN I low = H RCAB Con 1 I low ,
where I low is the lowlight images, H CNN · represents the CNN-based branch, Con 1 represents the convolutional layer with a size of 1 × 1 , and H RCAB · represents the pure CNN block.
The Transformer-based branch performs the successive stacked pure Transformer blocks based on cross-scale attention (CSA) and multi-head self-attention (MHSA) on the inputs to their global features. Meanwhile, the global-local features are aggregated by the fusion blocks to explore their complementarity and correlation.
F G = H Trans I low F Fu = H FB F L , F G ,
where F G represents the global features extracted by the Transformer-based branch H Trans · , and F Fu represents the fused features generated by the fusion block H FB · .
Finally, the color correction branch employs two successive Conv layers with kernel sizes of 3 × 3 and 1 × 1 to detect low-level features, and the learnable self-attention mechanism is used, fully exploring color distribution. We further yield visually pleasing images with vivid color by performing color distribution weighting on the fused feature maps.
F out = F Fu H Trans Conv 3 Conv 1 I low .
where F out is the final output, ⊗ represents the multiplication operation, and Conv 3 represents the convolutional layer with a size of 3 × 3 .

3.1. CNN-Based Branch

The CNN shows powerful potential local perception ability for lighting up low-illumination images. In our proposed MSINet, we design a pure CNN-based branch by stacking some Residual Channel Attention Blocks (RCABs). Given a lowlight input I low , the convolutional layer with a size of 1 × 1 is performed to extract the shallow features of local regions. Subsequently, we feed these features into the stacked Residual Channel Attention Blocks (RCABs) to fully dig up the local image features. And the processing procedure can be defined as
F SF = con 1 × 1 I low F L i = F RCAB i F L i 1 , 1 i N .
where F R C A B i denotes the i t h Residual Channel Attention Block (RCAB), and F L i represents the local features generated by the i t h Residual Channel Attention Block (RCAB). Notably, each RCAB only contains a series of Conv layers, ReLU, and global pooling, and the skip connection is also introduced into the RCAB to explore the model’s hierarchical feature extraction ability.

3.2. Transformer-Based Branch

The Transformer-based branch is designed to analyze the long-range dependency of features for perceiving the global image features. This branch is a pure Transformer dependent upon a stacked transformer block (as illustrated in Figure 2a), and each block contains a cross-scale attention (CSA) module and multi-head self-attention (MSHA) to fully explore the self-correlation and scale correlation of the features. Firstly, the input image F e R h × w × c is converted into multiple tokens T 0 R n × d , and each token represents a part of the image with dimension d. Then, the token T 0 R n × d is fed into the transformer block, and the operations within each involve several steps, namely
T i 1 = FFN LN T i 1 + T i 1 , T i 1 = MHSA LN T i 1 + T i 1 , T i = FFN LN T i 1 + T i 1 , T i 1 = CSA LN T i 1 + T i 1 .
where LN · represents the normalization layer, FFN · represents the feedforward neural network, and MHSA · and CSA · represent the MHSA and CSA, respectively.
CSA: The CSA can promote the model’s ability to represent cross-scale features, and its structure is shown in Figure 2b. We detail a description of the procedure of the CSA as follows: the input tokens are embedded into T R m × d , and then split along the axial dimension (i.e., the last dimension) into two parts, T a R m × d / 2 and T b R m × d / 2 . These parts are then used to generate T R m × d , which is subsequently reconstructed to retain the structure of the original token while obtaining larger tokens. In this process, the stride s is closely related to both the number of tokens and their dimensionality.
n = h t s + 1 × w t s + 1 , d = d × t 2 2 × t 2 = c 2 × t 2 .
With the help of the CSA, the MSINet can generate a large number of overlapping tokens at different scales, which helps in discovering cross-scale repetitive structures in the image. To better exploit image blocks at different scales and transfer large-scale features to smaller blocks, the model uses tokens of a larger size ( n > n ) during reconstruction. The network then processes T s and T to calculate the cross-scale attention weights, and further extract richer information from these tokens. Specifically, the first step is to generate query, key, and value from the two token sets T s and T . And T s corresponds to ( q s , k s , v s ) R n × d / 2 , while T corresponds to ( q , k , v ) R n × d / 2 . Finally, by adjusting the dimensionality, the CSA improves the efficient information transfer between different scales, optimizing the image restoration process without additional computational burden.
As we know, the size of the large and small tokens plays a key role. Hence, we perform the CSA module with different combinations of large and small token sizes on the LOL-v1 dataset. The PSNR scores of our MSINet with different token sizes are presented in Table 1. It can be observed that the small and large token sizes are respectively set to 3 and 4 and can generate higher PSNR scores.
Fusion module: For fully capturing feature representation at different branches, we propose a multi-branch feature fusion module, which horizontally connects the intermediate features of each branch to enable more efficient feature interaction and integration. This design allows the model to leverage the unique strengths of each branch, improving the overall performance of the network. Figure 1 illustrates the specific structure and implementation of the fusion module.
Specifically, for the intermediate features T i and F i output from the i t h RCAB and Transformer block, feature fusion is performed through the fusion module H fuse , which combines these features from different branches by capturing cross-branch dependency. The fusion is mathematically expressed as follows:
M i = H fise i rearrange T i F i , 1 i N
where M i R 2 c × h × w represents the fused feature representation, where the ‘rearrange‘ operation indicates a reordering of image features, and the symbol ‖ denotes concatenation along the channel dimension. Additionally, the fusion module H fuse uses a 1 × 1 convolutional layer to improve feature fusion along the channel dimension. Except for the final fusion module (i.e., i = N ), the fused feature M i is evenly split along the channel dimension into two parts, denoted as M T i R c × h × w and M F i R c × h × w .

3.3. Color Correction Branch

This paper introduces a unique color restoration branch designed to enhance the model’s ability to perceive and correct local color information in images. Specifically, the branch first extracts shallow features from the input image by combining deep convolution layers with standard convolution, and then expands the feature channels to enhance the image’s color representation. These enhanced shallow features are then passed through a specially designed Color Attention module, which uses learned query (Q), key (K), and value (V) to efficiently associate features. The Transformer-based self-attention mechanism in this module comprehensively explores and extracts the color features of the image.
This module effectively captures local color features in the image and performs color correction. Compared to traditional image enhancement methods, it more accurately restores the image’s color details. Specifically, by adjusting the convolution kernels and sizes of each convolution layer, the color restoration branch adapts the image’s color balance, reducing color deviations caused by lighting and noise. Finally, through the combination of multiple convolution and Transformer modules, the branch recovers the color details of the image and enhances its visual quality.

3.4. Loss Function

This paper employs L 1 and visual geometry group (VGG) perceptual loss functions to create our proposed method for enhancing lowlight images and detail enhancement. Among them, L 1 loss is robust and less sensitive to outliers, which can promote stable and faster convergence of the model by calculating the pixel-wise differences between the predicted image and its ground truth. And it can be expressed as
L 1 = 1 N i = 1 N y i y ^ i ,
where N is the total number of samples. y and y ^ represent the reference and output images, respectively.
VGG loss can measure the difference between the predicted image and its ground truth in deep feature space to make the former exhibit visually satisfactory perception. The VGG loss function L v g g can be defined as
L v g g = Ψ I Ψ ( I ^ ) ,
where Ψ ( · ) denotes the pre-trained VGG network. I and I ^ denote the feature maps extracted by the pre-trained VGG from the ground truth and enhanced images, respectively.
Finally, the total loss function L T o t a l used in this paper can be defined as
L T o t a l = L 1 + λ L v g g .
where λ is a weighting factor, empirically finding that the total loss function L T o t a l with λ = 1 × 10 5 can guarantee our method generates visually pleasing images.

4. Experimental Results and Analysis

This section first describes the implementation details and experimental settings. Next, we perform comprehensive evaluations on paired and unpaired datasets to verify the effectiveness of our MSINet. Finally, the analysis of detail enhancement, computational complexity, ablation study, and applications as well as limitations and future work is performed.

4.1. Implementation Details

We use the MIT-Adobe 5K [52], LOL-v1 [18], and LOL-v2 [53] datasets for verifying the performance in the LLIE tasks. Among them, the MIT-Adobe 5K [52] contains 4500 pairs of low-/normal-light training images and 500 pairs of low-/normal-light testing image pairs, the LOL-v1 [18] contains 485 pairs of low-/normal-light training image pairs and 15 pairs of low-/normal-light testing image pairs, and the LOL-v2 [53] contains 900 pairs of low-/normal-light training images and 100 pairs of low-/normal-light testing image pairs. Note that the training and testing images are resized to the size of 512 × 512 × 3 . Additionally, we also perform our proposed LLIE model on four unpaired datasets, including DICM, LIME, VV, and MEF, to further test its robustness in real-world applications.
All validation experiments are implemented in the Pytorch framework on an NVIDIA Tesla P100 GPU. We augment the training dataset through rotating by 90 and horizontal flipping. During training, the ADAM optimizer with β 1 = 0.9 , β 1 = 0.9999 , and ε = 10 8 is used to train the parameters of our model. The initial learning rate is set to 1 × 10 3 and fine-tuned to 1 × 10 4 after 50 epochs. A batch size of 8 is applied.

4.2. Experimental Settings

Our method is compared with eighteen LLIE methods, including traditional methods, LIME [54], JED [11]; supervised learning-based methods, SID [12], IPT [13], RetinexNet [18], STANet [55], MIRNet [56], DRBN [57], KinD [58], LPNet [15], UFormer [16], PairLIE [14] and RetFormer [33], LightenDiffusion [41], QuadPrior [59], CIDNet [30], and END [60]; and self/un-supervised learning-based methods, RUAS [61], DSLR [19], Zero-DCE [28], URetinex-Net [16]. Notably, these above-listed comparison methods use publicly available source codes with recommended parameters to reproduce the enhanced results.
Except for comparison of visual perception, we also adopt commonly used image quality evaluation metrics to quantitatively assess their performance. The peak signal-to-noise ratio (PSNR), structural similarity index (SSIM), and perceptual image patch similarity (LPIPS) are used to evaluate the MIT-Adobe 5K [52], LOL-v1 [18], and LOL-v2 [53] datasets. While the former two metrics are full-reference evaluation, the LPIPS is no-reference evaluation. Among them, a higher PSNR or SSIM score indicates more realistic restoration results, while a lower LPIPS score suggests a better visual perception. For the DICM, LIME, VV, and MEF datasets, the natural image quality evaluator (NIQE), perceptual index (PI), and no-reference image quality metric (NIQMC) are used to assess the enhanced images. And these three metrics are no-reference evaluation. Among them, a lower NIQE or PI score suggests a better natural-looking and visual perception. A higher NIQMC score indicates a greater amount of image information.

4.3. Comprehensive Evaluation on Paired Datasets

The visual evaluation of our MSINet is compared with state-of-the-art LLIE approaches on the 5K, LOL-v1, and LOL-v2 datasets to verify their performance in the LLIE tasks.
Qualitative Evaluation. On the MIT-Adobe 5K dataset, as shown in Figure 3, LIME [54] and RetinexNet [18] struggle with overexposure and unnatural color shifts. And the latter yields unnatural-looking visual experience and blurry details. DSLR [19] creates high-contrast images with observable artifact haloes. STANet [55] introduces local overenhancement and unwanted color deviation. Although LPNet [15] effectively lights up the brightness of lowlight images, it is unsatisfactory in removing local underenhancement. Zero-DCE [28] and URetinex-Net [16] yield hazy-like and visually unnatural-looking images from the lowlight inputs, and they also fail in removing color distortion and inherent noise. UFormer [16] shows satisfactory performance in the LLIE task, but some UFormer-enhanced images exhibit hazy-like (e.g., the second row in Figure 3i) and local dark areas (e.g., the third row in Figure 3i). PairLIE [14] struggles to balance the contrast stretch and detail enhancement for the low-illumination images. In contrast, our proposed method achieves a more balanced enhancement performance, characterized by refined texture rendition and minimized perceptual distortion, thereby yielding superior visual fidelity.
Figure 4 and Figure 5 present the enhanced results randomly selected from the LOL-v1 and LOL-v2 datasets. For the LOL-v1 dataset, RetinexNet [18] and Zero-DCE [28] generate a greenish tone and unsatisfactory visual perception. LIME [54] cannot yield high-contrast images or remove local darkness. DSLR [19], URetinex-Net [16], and LPNet [15] fail to achieve color correction, and the former two methods inject blurry details and edges in enhanced images. STANet [55] and PairLIE [14] promote the quality of lowlight images, but successfully remove artifact haloes. For the LOL-v2 dataset, all comparison methods cannot light up the partially dark areas. Among them, the performance of the UFormer [16] is the poorest, followed by that of PairLIE [14]. URetinex-Net [16] generates high-brightness images with unsatisfactory contrast. LightenDiffusion [41] introduces color deviation (e.g., the sky part of the image in Figure 5g), and QuadPrior [59] fails in removing inherent noise and making details clearer. CIDNet [30] injects observable color deviation and partial darkness. In comparison, our method effectively removes color deviations, highlights the structural details, and improves visibility without over-enhancement or oversaturation.
Quantitative Evaluation. From the quantitative evaluation scores on the LOL-v1, LOL-v2, and MIT-Adobe 5K datasets in Table 2, it can be seen that our MSINet has higher scores of the PSNR, SSIM, and LPIPS on the MIT-Adobe 5K dataset than the compared methods. On the LOL-v1 dataset, the proposed MSINet has comparable and higher values of the PSNR and SSIM. Additionally, our MSINet achieves the second-best performance on the LOL-v2 dataset. To clearly demonstrate the PSNR, SSIM, and LPIPS scores on the LOL-v1, LOL-v2, and MIT-Adobe 5K datasets, we further drew their bar charts in Figure 6. The qualitative and quantitative results suggest that our MSINet can effectively produce high visibility and natural color with significant enhancement of the contrast, brightness, and texture details.

4.4. Comprehensive Evaluation on Unpaired Datasets

We perform them on unpaired datasets, including DICM, LIME, VV, and MEF, to further evaluate the robustness of our MSINet and other compared LLIE methods in both qualitative and quantitative assessments.
Qualitative Evaluation. Figure 7 demonstrates enhanced results generated by these state-of-the-art approaches randomly selected from the DICM, LIME, VV, and MEF benchmarks. The following observations from these enhanced images can be easily obtained: The original lowlight images suffer from local low-contrast and unsatisfactory contrast and illumination as well as blurry details. RetinexNet [18] successfully lights up lowlight images, but generates unnatural-looking visual experience and unwanted artefact haloes. LIME [54] consistently yields visually pleasing results without fine-tuning its parameters on these four datasets. But the LIME-enhanced images still confront lower contrast and color deviation. DSLR [19] inevitably introduces blocking effects in enhanced results (e.g., the third/forth rows in Figure 7d) and fails in removing local extremely lowlight areas. URetinex-Net [16] and UFommer [31] show unsatisfactory performance in removing local over-enhancement. In addition, UFommer [31] introduces an observable checkerboard effect in some enhanced images. Zero-DCE [28] can effectively remove unwanted local extremely-low-illumination areas, while it fails in detail enhancement and noise suppression. RetFormer [33], UFormer [16], DSLR [19], PairLIE [14], and STANet [55] cannot tackle local under-exposure and fail in local detail boosting. STANet [55] introduces color distortion (e.g., the second/third rows in Figure 7e) and local over-enhancement (e.g., the third row in Figure 7e) for some low-illumination images. PairLIE [14] can yield observable artefact haloes, as demonstrated in the first row in Figure 7j. RetFormer [33] also fails in removing color distortion. LPNet [15] generates high-contrast images and clearer details, while unsuccessfully removing local over-enhancement (e.g., the third row in Figure 7f) and enhancing the contrast of the dark areas. In comparison, our model produces visually appealing results with better control over exposure, preserving the details in both bright and dark regions. The image sharpness and clarity are enhanced while maintaining a natural appearance. The fine textures, especially in the foreground and background, are well-preserved, and there are no noticeable color shifts or over-enhancement.
Quantitative Evaluation. We further employ the NIQE, PI, and NIQMC to evaluate the performance of these LLIE methods quantitatively. The NIQE, PI, and NIQMC scores of different LLIE methods on the DICM, LIME, VV, and MEF datasets are shown in Table 3. From the quantitative evaluation scores in Table 3, it can be found that our MSINet achieves the best scores in NIQE (2.816), PI (2.553), and NIQMC (5.219) on the DICM dataset. On other datasets, our method also yields lower NIQMC and the second-best PI scores than the compared LLIE methods. For clearly demonstrating the NIQE, PI, and NIQMC scores on the DICM, LIME, VV, and MEF datasets, we further drew their bar charts in Figure 8. In conclusion, the qualitative and quantitative results suggest that our MSINet works better in color correction, detail boosting, and contrast stretch for the LLIE tasks.

4.5. Comprehensive Evaluation of Detail Enhancement

High-quality images with clearer details play an important role in object detection and scene understanding. We compare our proposed MSINet with other compared LLIE in terms of detail boosting. As illustrated in Figure 3, Figure 4, Figure 5, Figure 6 and Figure 7, our proposed MSINet significantly lights up lowlight images and removes color deviation from a global view. From a local view, our method can effectively make the fine structural details clearer than other compared methods. Furthermore, we introduce the average gradient (AG), local variance (LVar) and local standard deviation (LSTD) to verify the detail enhancement capability of the model. Among them, the AG can measure the detail information, while the LVar and LSTD can measure the spatial variations in noise and features. Table 4 presents the AG, LVar, and LSTD scores of different LLIE methods. From Table 4, it can be easily found that our proposed MSINet yields more satisfactory AG, LVar, and LSTD scores than the compared approaches. The quantitative and qualitative results indicate that our MSINet is superior compared to LLIE methods in detail enhancement.

4.6. Comprehensive Evaluation of Color Correction

Most LLIE methods can light up images and enhance details, but inevitably introduce color deviations. In our experiments, we further verify the color correction ability of our MSINet on the LOL-v1 dataset. Additionally, the Δ E [7] and Bhattacharyya Distance (BD) [63] are introduced to quantitatively evaluate the color correction capability of the compared LLIE methods and our MSINet. Among them, the former is a metric for color distortion, and the latter can measure the color difference between the enhanced images and their reference ones. Table 4 presents the average Δ E and Bhattacharyya Distance (BD) scores of different LLIE methods. It can be easily observed that our method yields lower BD and Δ E scores than the compared methods, indicating that the enhanced images generated by the MSINet share similar color distribution with the reference image.
Following references [64,65], we present the visual and corresponding histogram of the original, enhanced, and reference images (as shown in Figure 9) to further demonstrate the physical process of our MSINet for the LLIE tasks. From Figure 9, we can observe that the original image exhibits low illumination and color deviation, and its corresponding histogram is distributed in the area with smaller pixel values. The image processed by our MSINet shows high brightness and vivid color, and its corresponding histogram is distributed more uniformly by pixel interpolation for the lowlight input. The most likely reason is that the Bayer filter sensors make only the smaller value pixels receive the wavelength light under suboptimal lighting conditions. Our proposed MSINet reduces the difference in spectral response among the wavelengths of the R, G, and B channels for removing the color deviation from the enhanced results.

4.7. Comprehensive Evaluation of Computational Complexity

Table 5 presents the Param, Flops, and runtime of our MSINet and all the above-listed comparison LLIE methods on the LOL-v1 dataset to comprehensively evaluate their computational complexity. It can be easily found that Zero-DCE [28] enjoys lower computational complexity and faster inference speed. LightenDiffusion [41] encounters a heavy computational burden, limiting its application in the real-world. In contrast, our proposed MSINet can balance the computational complexity as well as inference speed and enhanced results for the LLIE tasks.

4.8. Ablation Study

We further perform an ablation study on our proposed MSINet to test the effectiveness of each component. The ablation studies include (a) our method without residual channel Attention block (-w/o RCAB), (b) our method without Cross-Scale Attention Module (-w/o CSA), and our method without color correction Branch (-w/o CCB).
Figure 10 shows the visual comparisons on the MIT-Adobe 5K, LOL-v1, and LOL-v2 datasets. The following visual results can be observed: (1) -w/o RCAB fails in detail boosting for the LLIE tasks and introduces unnatural-looking visual experience in some enhanced images; (2) -w/o CSA lights up low-illumination images and promotes contrast, but the enhanced images suffer from blurry boundaries and details; (3) -w/o CCB generates observed color deviation in enhanced images; (4) the full model (MSINet) with all key components can effectively yield visually pleasing images with vivid color and clearer details.
Furthermore, we present the quantitative scores of the ablated models for the MIT-Adobe 5K, LOL-v1, and LOL-v2 datasets in Table 6. It can be found that our MSINet can create more satisfactory quantitative scores than the ablated models across three public benchmarks, benefiting from each key component.

4.9. Generalization of Our Proposed Method

We first perform our method on underwater images to test the generalization of our proposed method. The training/testing datasets are randomly selected from the UCCS benchmark in the ratio of 7:3. Figure 11 demonstrates the enhanced results randomly selected from the UCCS benchmark. Intuitively, the input underwater images exhibit unsatisfactory visual perception, whereas our method can effectively promote the quality of the underwater images. We also present the NIQE score of the input and its corresponding enhanced images, and the latter has a superior NIQE score.
The proposed MSINet also works well on medical images including endoscopic and pathological images with low contrast. Notably, the endoscopic images are randomly selected from the CVC-ClinicDB dataset (https://tianchi.aliyun.com/dataset/93690 (accessed on 20 October 2025)), and the pathological images are randomly selected from the LUAD-HistoSeg dataset (https://drive.google.com/drive/folders/1E3Yei3Or3xJXukHIybZAgochxfn6FJpr (accessed on 20 October 2025)). As illustrated in the first rows of Figure 12, the original endoscopic images and pathological images encounter low contrast and blurry details, which may deliver compromised information for disease diagnosis, prognosis analysis, and therapeutic effect prediction in clinical analysis. On the contrary, our MSINet shows satisfactory performance in contrast stretch and detail boosting for endoscopic images and pathological images. In addition, we also present the NIQE score of the original input and its corresponding enhanced results. It can be easily found that the enhanced results generated by the MSINet can generate more satisfactory NIQE scores. Experiments suggest that our MSINet exhibits solid generalization in restoring the quality of the endoscopic images and pathological images.

4.10. Limitations and Future Work

The LLIE method can enhance lowlight images and further promote the performance of object detection, image classification, and other advanced computer vision tasks. Our proposed MSINet can work well in yielding more satisfactory results with vivid color and clearer details from low-illumination images in most situations, whereas it fails in noise reduction, local exposure control for lowlight images with local over-exposure, boosted noise, and so on. For example, Figure 13 illustrates some failure instances created by our proposed MSINet. Intuitively, the enhanced images exhibit vivid color and high contrast, but they also contain observed noise, artefact halos, and local over-enhancement. The reason may be that our MSINet directly processes high-frequency components with the inherent noise of the lowlight input. Additionally, our method does not take into account the enhancement of images with local overexposure. In the future, we will promote learning and a specialised denoising module for tackling these challenging issues.

5. Conclusions

This paper presents a multi-scale interactive network with color attention named MSINet for light enhancement and color correction. The MSINet is a three-branch CNN-Transformer hybrid structure containing CNN-based branch, Transformer-based branch, and a color correction branch. The CNN-based branch, a pure CNN network, can fully explore the local image features by the stacked residual channel attention blocks (RCABs), and the Transformer-based branch, a pure Transformer network, can mine the global features by the stacked Transformer blocks containing cross-scale attention (CSA) and multi-head self-attention (MHSA). Meanwhile, we design a fusion module to integrate the global and local features extracted by each RCAB and Transformer block. Additionally, we further employ the color correction branch based on self-attention (SA) to learn the color distribution information for removing the color deviation. Experimental results on different datasets have shown that our MSINet can generate visually pleasing images with clearer details and more vivid color compared with LLIE methods.

Author Contributions

Conceptualization, H.L., C.Q. and Z.W.; methodology, H.L.; software, H.L. and C.Q.; validation, H.L.; investigation, H.L. and C.Q.; writing—original draft preparation, H.L.; writing—review and editing, Z.L. and Z.W.; visualization, H.L. and C.Q.; supervision, Z.L.; funding acquisition, Z.L. and H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under grants (82502451, 82272075, 62462017), the Guangxi Science and Technology Project (AB21220037), the Guangxi Natural Science Foundation under grants (2025GXNSFBA069390), and the Innovation Project of Guangxi Graduate Education (2025YCXB008).

Data Availability Statement

Publicly available datasets were analyzed in this study. These data can be found here: LOL-v1 benchmark https://daooshee.github.io/BMVC2018website/ (accessed on 1 October 2025), LOL-v2 benchmark https://www.kaggle.com/datasets/tanhyml/lol-v2-dataset (accessed on 1 October 2025), MIT-Adobe 5K benchmark https://data.csail.mit.edu/graphics/fivek/ (accessed on 1 October 2025), and LIME, MEF, DICM, and VV https://drive.google.com/drive/folders/1lp6m5JE3kf3M66Dicbx5wSnvhxt90V4T (accessed on 1 October 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Nasir, M.F.; Rehman, M.U.; Hussain, I. A self-attention guided approach for advanced underwater image super-resolution with depth awareness. IEEE Open J. Comput. Soc. 2025, 6, 1715–1725. [Google Scholar] [CrossRef]
  2. Li, M.; Jia, T.; Wang, H.; Ma, B.; Lu, H.; Lin, S.; Cai, D.; Chen, D. AO-DETR: Anti-overlapping DETR for X-Ray prohibited items detection. IEEE Trans. Neural Networks Learn. Syst. 2025, 36, 12076–12090. [Google Scholar] [CrossRef] [PubMed]
  3. Li, C.; Guo, C.; Loy, C.C. Learning to enhance low-light image via zero-reference deep curve estimation. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 4225–4238. [Google Scholar] [CrossRef] [PubMed]
  4. Huang, C.; Wang, Y.; Jiang, Y.; Li, M.; Huang, X.; Wang, S.; Pan, S.; Zhou, C. Flow2GNN: Flexible two-way flow message passing for enhancing GNNs beyond homophily. IEEE Trans. Cybern. 2024, 54, 6607–6618. [Google Scholar] [CrossRef]
  5. Li, C.; Liu, J.; Zhu, J.; Zhang, W.; Bi, L. Mine image enhancement using adaptive bilateral gamma adjustment and double plateaus histogram equalization. Multimed. Tools Appl. 2022, 81, 12643–12660. [Google Scholar] [CrossRef]
  6. Yang, W.; Wang, W.; Huang, H.; Wang, S.; Liu, J. Sparse gradient regularized deep retinex network for robust low-light image enhancement. IEEE Trans. Image Process. 2021, 30, 2072–2086. [Google Scholar] [CrossRef]
  7. Liu, X.; Zhao, Q. Guided filter-inspired network for low-light RAW image enhancement. Sensors 2025, 25, 2637. [Google Scholar] [CrossRef]
  8. Wang, F.; Zhang, B.; Zhang, C.; Yan, W.; Zhao, Z.; Wang, M. Low-light image joint enhancement optimization algorithm based on frame accumulation and multi-scale Retinex. Ad Hoc Netw. 2021, 113, 102398. [Google Scholar] [CrossRef]
  9. Liao, H.; Xia, J.; Yang, Z.; Pan, F.; Liu, Z.; Liu, Y. Meta-learning based domain prior with application to optical-ISAR image translation. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 7041–7056. [Google Scholar] [CrossRef]
  10. Lore, K.G.; Akintayo, A.; Sarkar, S. LLNet: A deep autoencoder approach to natural low-light image enhancement. Pattern Recognit. 2017, 61, 650–662. [Google Scholar] [CrossRef]
  11. Ren, X.; Li, M.; Cheng, W.H.; Liu, J. Joint enhancement and denoising method via sequential decomposition. In Proceedings of the 2018 IEEE International Symposium on Circuits and Systems (ISCAS), Florence, Italy, 27–30 May 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–5. [Google Scholar]
  12. Chen, C.; Chen, Q.; Do, M.N.; Koltun, V. Seeing motion in the dark. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3185–3194. [Google Scholar]
  13. Chen, H.; Wang, Y.; Guo, T.; Xu, C.; Deng, Y.; Liu, Z.; Ma, S.; Xu, C.; Xu, C.; Gao, W. Pre-trained image processing transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 12299–12310. [Google Scholar]
  14. Huang, Y.; Liu, Z.; Lu, H.; Wang, W.; Lan, R. Brighten up images via dual-branch structure-texture awareness feature interaction. IEEE Signal Process. Lett. 2023, 31, 46–50. [Google Scholar] [CrossRef]
  15. Li, J.; Li, J.; Fang, F.; Li, F.; Zhang, G. Luminance-aware pyramid network for low-light image enhancement. IEEE Trans. Multimed. 2020, 23, 3153–3165. [Google Scholar] [CrossRef]
  16. Wu, W.; Weng, J.; Zhang, P.; Wang, X.; Yang, W.; Jiang, J. Uretinex-net: Retinex-based deep unfolding network for low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5901–5910. [Google Scholar]
  17. Qian, S.; Chen, Y.; Wang, W.; Zhang, G.; Li, L.; Hao, Z.; Wang, Y. Physics-guided deep neural networks for bathymetric mapping using Sentinel-2 multi-spectral imagery. Front. Mar. Sci. 2025, 12, 1636124. [Google Scholar] [CrossRef]
  18. Wei, C.; Wang, W.; Yang, W.; Liu, J. Deep retinex decomposition for low-light enhancement. arXiv 2018, arXiv:1808.04560. [Google Scholar] [CrossRef]
  19. Lim, S.; Kim, W. DSLR: Deep stacked Laplacian restorer for low-light image enhancement. IEEE Trans. Multimed. 2020, 23, 4272–4284. [Google Scholar] [CrossRef]
  20. Zhang, Z.; Zheng, H.; Hong, R.; Xu, M.; Yan, S.; Wang, M. Deep color consistent network for low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 1899–1908. [Google Scholar]
  21. Wang, W.; Yin, B.; Li, L.; Li, L.; Liu, H. A Low Light Image Enhancement Method Based on Dehazing Physical Model. Comput. Model. Eng. Sci. (CMES) 2025, 143, 1595–1616. [Google Scholar] [CrossRef]
  22. Xu, H.; Zhang, H.; Yi, X.; Ma, J. CRetinex: A progressive color-shift aware retinex model for low-light image enhancement. Int. J. Comput. Vis. 2024, 132, 3610–3632. [Google Scholar] [CrossRef]
  23. Wu, J.; Ai, H.; Zhou, P.; Wang, H.; Zhang, H.; Zhang, G.; Chen, W. Low-light image dehazing and enhancement via multi-feature domain fusion. Remote Sens. 2025, 17, 2944. [Google Scholar] [CrossRef]
  24. Tu, B.; Zhou, T.; Liu, B.; He, Y.; Li, J.; Plaza, A. Multi-scale autoencoder suppression strategy for hyperspectral image anomaly detection. IEEE Trans. Image Process. 2025, 34, 5115–5130. [Google Scholar] [CrossRef] [PubMed]
  25. Gu, K.; Liu, H.; Liu, Y.; Qiao, J.; Zhai, G.; Zhang, W. Perceptual information fidelity for quality estimation of industrial images. IEEE Trans. Circuits Syst. Video Technol. 2025, 35, 477–491. [Google Scholar] [CrossRef]
  26. Huang, C.; Gao, C.; Li, M.; Li, Y.; Wang, X.; Jiang, Y.; Huang, X. Correlation information enhanced graph anomaly detection via hypergraph transformation. IEEE Trans. Cybern. 2025, 55, 2865–2878. [Google Scholar] [CrossRef]
  27. Jiang, Y.; Gong, X.; Liu, D.; Cheng, Y.; Fang, C.; Shen, X.; Yang, J.; Zhou, P.; Wang, Z. EnlightenGAN: Deep light enhancement without paired supervision. IEEE Trans. Image Process. 2021, 30, 2340–2349. [Google Scholar] [CrossRef]
  28. Guo, C.; Li, C.; Guo, J.; Loy, C.C.; Hou, J.; Kwong, S.; Cong, R. Zero-reference deep curve estimation for low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 1780–1789. [Google Scholar]
  29. Yao, Z.; Su, J.N.; Fan, G.; Gan, M.; Chen, C.P. GACA: A gradient-aware and contrastive-adaptive learning framework for low-light image enhancement. IEEE Trans. Instrum. Meas. 2024, 73, 1–14. [Google Scholar] [CrossRef]
  30. Yan, Q.; Feng, Y.; Zhang, C.; Pang, G.; Shi, K.; Wu, P.; Dong, W.; Sun, J.; Zhang, Y. HVI: A new color space for low-light image enhancement. In Proceedings of the Computer Vision and Pattern Recognition Conference, Nashville, TN, USA, 11–15 June 2025; pp. 5678–5687. [Google Scholar]
  31. Wang, Z.; Cun, X.; Bao, J.; Zhou, W.; Liu, J.; Li, H. Uformer: A general u-shaped transformer for image restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 17683–17693. [Google Scholar]
  32. Zhang, Y.; Wang, Z.; Huang, M.; Li, M.; Zhang, J.; Wang, S.; Zhang, J.; Zhang, H. S2DBFT: Spectral–Spatial Dual-Branch Fusion Transformer for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2025, 63, 1–17. [Google Scholar] [CrossRef]
  33. Cai, Y.; Bian, H.; Lin, J.; Wang, H.; Timofte, R.; Zhang, Y. Retinexformer: One-stage retinex-based transformer for low-light image enhancement. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 4–6 October 2023; pp. 12504–12513. [Google Scholar]
  34. Zhang, S.; Meng, N.; Lam, E.Y. LRT: An efficient low-light restoration transformer for dark light field images. IEEE Trans. Image Process. 2023, 32, 4314–4326. [Google Scholar] [CrossRef]
  35. Wu, P.; Zuo, Z.; Su, S.; Zhao, B. Infrared target detection based on image enhancement and an improved feature extraction network. Drones 2025, 9, 695. [Google Scholar] [CrossRef]
  36. Dang, J.; Zhong, Y.; Qin, X. PPformer: Using pixel-wise and patch-wise cross-attention for low-light image enhancement. Comput. Vis. Image Underst. 2024, 241, 103930. [Google Scholar] [CrossRef]
  37. Pei, X.; Huang, Y.; Su, W.; Zhu, F.; Liu, Q. FFTFormer: A spatial-frequency noise aware CNN-Transformer for low light image enhancement. Knowl.-Based Syst. 2025, 314, 113055. [Google Scholar] [CrossRef]
  38. Brateanu, A.; Balmez, R.; Avram, A.; Orhei, C.; Ancuti, C. LYT-NET: Lightweight YUV Transformer-Based Network for Low-Light Image Enhancement. IEEE Signal Process. Lett. 2025, 32, 2065–2069. [Google Scholar] [CrossRef]
  39. Wen, Y.; Xu, P.; Li, Z.; ATO, W.X. An illumination-guided dual attention vision transformer for low-light image enhancement. Pattern Recognit. 2025, 158, 111033. [Google Scholar] [CrossRef]
  40. Dong, W.; Min, Y.; Zhou, H.; Chen, J. Towards Scale-Aware Low-Light Enhancement via Structure-Guided Transformer Design. In Proceedings of the Computer Vision and Pattern Recognition Conference, Nashville, TN, USA, 11–15 June 2025; pp. 1460–1470. [Google Scholar]
  41. Jiang, H.; Luo, A.; Liu, X.; Han, S.; Liu, S. Lightendiffusion: Unsupervised low-light image enhancement with latent-retinex diffusion models. arXiv 2024, arXiv:2407.08939. [Google Scholar]
  42. Li, L.; Wang, X.; Huang, D.; He, Y.; Zhong, Z.; Xia, Q. KEDM: Knowledge-embedded diffusion model for infrared image festriping. IEEE Photonics J. 2025, 17, 1–9. [Google Scholar] [CrossRef]
  43. Zhang, L.; Wu, F.; Bronik, K.; Papiez, B.W. DiffuSeg: Domain-driven diffusion for medical image segmentation. IEEE J. Biomed. Health Inform. 2025, 29, 3619–3631. [Google Scholar] [CrossRef] [PubMed]
  44. Yi, X.; Xu, H.; Zhang, H.; Tang, L.; Ma, J. Diff-retinex: Rethinking low-light image enhancement with a generative diffusion model. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 4–6 October 2023; pp. 12302–12311. [Google Scholar]
  45. Yi, X.; Xu, H.; Zhang, H.; Tang, L.; Ma, J. Diff-Retinex++: Retinex-driven reinforced diffusion model for low-light image enhancement. IEEE Trans. Pattern Anal. Mach. Intell. 2025, 47, 6823–6841. [Google Scholar] [CrossRef] [PubMed]
  46. Lin, Y.; Ye, T.; Chen, S.; Fu, Z.; Wang, Y.; Chai, W.; Xing, Z.; Li, W.; Zhu, L.; Ding, X. Aglldiff: Guiding diffusion models towards unsupervised training-free real-world low-light image enhancement. In Proceedings of the AAAI Conference on Artificial Intelligence, Philadelphia, PA, USA, 25 February–4 March 2025; Volume 39, pp. 5307–5315. [Google Scholar]
  47. Yang, S.; Zhang, X.; Wang, Y.; Yu, J.; Wang, Y.; Zhang, J. Difflle: Diffusion-based domain calibration for weak supervised low-light image enhancement. Int. J. Comput. Vis. 2025, 133, 2527–2546. [Google Scholar] [CrossRef]
  48. Huang, Y.; Liao, X.; Liang, J.; Shi, B.; Xu, Y.; Le Callet, P. Detail-preserving diffusion models for low-light image enhancement. IEEE Trans. Circuits Syst. Video Technol. 2025, 35, 3396–3409. [Google Scholar] [CrossRef]
  49. Hu, R.; Luo, T.; Jiang, G.; Chen, Y.; Xu, H.; Liu, L.; He, Z. DiffDark: Multi-prior integration driven diffusion model for low-light image enhancement. Pattern Recognit. 2025, 168, 111814. [Google Scholar] [CrossRef]
  50. Jiang, H.; Luo, A.; Fan, H.; Han, S.; Liu, S. Low-light image enhancement with wavelet-based diffusion models. ACM Trans. Graph. (TOG) 2023, 42, 1–14. [Google Scholar] [CrossRef]
  51. Jin, H.; Wang, J.; Zuo, F.; Su, H.; Xiao, Z.; Wang, B.; Zhang, Y. DCGSD: Low-light image enhancement with dual-conditional guidance sparse diffusion model. IEEE Trans. Circuits Syst. Video Technol. 2025, 35, 7792–7806. [Google Scholar] [CrossRef]
  52. Wang, R.; Zhang, Q.; Fu, C.W.; Shen, X.; Zheng, W.S.; Jia, J. Underexposed photo enhancement using deep illumination estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6849–6857. [Google Scholar]
  53. Wang, Y.; Wan, R.; Yang, W.; Li, H.; Chau, L.P.; Kot, A. Low-light image enhancement with normalizing flow. In Proceedings of the AAAI Conference on Artificial Intelligence, Pomona, CA, USA, 24–28 October 2022; Volume 36, pp. 2604–2612. [Google Scholar]
  54. Guo, X. LIME: A method for low-light image enhancement. In Proceedings of the 24th ACM International Conference on Multimedia, Amsterdam, The Netherlands, 15–19 October 2016; pp. 87–91. [Google Scholar]
  55. Xu, K.; Chen, H.; Xu, C.; Jin, Y.; Zhu, C. Structure-texture aware network for low-light image enhancement. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 4983–4996. [Google Scholar] [CrossRef]
  56. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.H.; Shao, L. Learning enriched features for real image restoration and enhancement. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XXV 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 492–511. [Google Scholar]
  57. Yang, W.; Wang, S.; Fang, Y.; Wang, Y.; Liu, J. From fidelity to perceptual quality: A semi-supervised approach for low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 3063–3072. [Google Scholar]
  58. Zhang, Y.; Zhang, J.; Guo, X. Kindling the darkness: A practical low-light image enhancer. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 1632–1640. [Google Scholar]
  59. Wang, W.; Yang, H.; Fu, J.; Liu, J. Zero-reference low-light enhancement via physical quadruple priors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–18 June 2024; pp. 26057–26066. [Google Scholar]
  60. Wang, H.; Yan, X.; Hou, X.; Zhang, K.; Dun, Y. Extracting noise and darkness: Low-light image enhancement via dual prior guidance. IEEE Trans. Circuits Syst. Video Technol. 2025, 35, 1700–1714. [Google Scholar] [CrossRef]
  61. Liu, R.; Ma, L.; Zhang, J.; Fan, X.; Luo, Z. Retinex-inspired unrolling with cooperative prior architecture search for low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 10561–10570. [Google Scholar]
  62. Xu, X.; Wang, R.; Fu, C.W.; Jia, J. Snr-aware low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 17714–17724. [Google Scholar]
  63. Choi, E.; Lee, C. Feature extraction based on the Bhattacharyya distance. Pattern Recognit. 2003, 36, 1703–1709. [Google Scholar] [CrossRef]
  64. Cheremkhin, P.; Lesnichii, V.; Petrov, N. Use of spectral characteristics of DSLR cameras with Bayer filter sensors. In Journal of Physics: Conference Series; IOP Publishing: Bristol, UK, 2014; Volume 536, p. 012021. [Google Scholar]
  65. Ren, Y.; Ying, Z.; Li, T.H.; Li, G. LECARM: Low-light image enhancement using the camera response model. IEEE Trans. Circuits Syst. Video Technol. 2019, 29, 968–981. [Google Scholar] [CrossRef]
Figure 1. Overview of the proposed model (MSINet), which includes three parrallel branches: CNN-based branch, Transformer-based branch, and color correction branch.
Figure 1. Overview of the proposed model (MSINet), which includes three parrallel branches: CNN-based branch, Transformer-based branch, and color correction branch.
Sensors 26 00083 g001
Figure 2. Structure of the Transformer block (a) and cross-scale attention (CSA) (b).
Figure 2. Structure of the Transformer block (a) and cross-scale attention (CSA) (b).
Sensors 26 00083 g002
Figure 3. Visual comparisons on the MIT-Adobe 5K dataset. (a) The input samples are randomly selected from the MIT-Adobe 5K dataset. The enhanced results generated by (b) LIME [54], (c) RetinexNet [18], (d) DSLR [19], (e) STANet [55], (f) LPNet [15], (g) Zero-DCE [28], (h) URetinex-Net [16], (i) UFormer [16], (j) PairLIE [14], (k) ours, and (l) Ground truth (GT).
Figure 3. Visual comparisons on the MIT-Adobe 5K dataset. (a) The input samples are randomly selected from the MIT-Adobe 5K dataset. The enhanced results generated by (b) LIME [54], (c) RetinexNet [18], (d) DSLR [19], (e) STANet [55], (f) LPNet [15], (g) Zero-DCE [28], (h) URetinex-Net [16], (i) UFormer [16], (j) PairLIE [14], (k) ours, and (l) Ground truth (GT).
Sensors 26 00083 g003
Figure 4. Visual comparisons on the LOL-v1 dataset. (a) The input samples are randomly selected from the LOL-v1 dataset. The enhanced results generated by (b) LIME [54], (c) RetinexNet [18], (d) DSLR [19], (e) STANet [55], (f) LPNet [15], (g) Zero-DCE [28], (h) URetinex-Net [16], (i) UFormer [16], (j) PairLIE [14], (k) ours, and (l) Ground truth (GT).
Figure 4. Visual comparisons on the LOL-v1 dataset. (a) The input samples are randomly selected from the LOL-v1 dataset. The enhanced results generated by (b) LIME [54], (c) RetinexNet [18], (d) DSLR [19], (e) STANet [55], (f) LPNet [15], (g) Zero-DCE [28], (h) URetinex-Net [16], (i) UFormer [16], (j) PairLIE [14], (k) ours, and (l) Ground truth (GT).
Sensors 26 00083 g004
Figure 5. Visual comparisons on the LOL-v2 dataset. (a) The input samples are randomly selected from the LOL-v2 dataset. The enhanced results generated by (b) LIME [54], (c) RetinexNet [18], (d) DSLR [19], (e) STANet [55], (f) LPNet [15], (g) Zero-DCE [28], (h) URetinex-Net [16], (i) UFormer [16], (j) PairLIE [14], (k) ours, and (l) Ground truth (GT).
Figure 5. Visual comparisons on the LOL-v2 dataset. (a) The input samples are randomly selected from the LOL-v2 dataset. The enhanced results generated by (b) LIME [54], (c) RetinexNet [18], (d) DSLR [19], (e) STANet [55], (f) LPNet [15], (g) Zero-DCE [28], (h) URetinex-Net [16], (i) UFormer [16], (j) PairLIE [14], (k) ours, and (l) Ground truth (GT).
Sensors 26 00083 g005
Figure 6. The bar charts of the PSNR, SSIM, and LPIPS of different LLIE methods on paired datasets.
Figure 6. The bar charts of the PSNR, SSIM, and LPIPS of different LLIE methods on paired datasets.
Sensors 26 00083 g006
Figure 7. Visual comparisons on paired datasets. From top to bottom, (a) the inputs are randomly selected from the DICM, LIME, VV, and MEF, respectively. Their corresponding enhanced results generated by (b) LIME [54], (c) RetinexNet [18], (d) DSLR [19], (e) STANet [55], (f) LPNet [15], (g) Zero-DCE [28], (h) URetinex-Net [16], (i) UFormer [16], (j) PairLIE [14], (k) RetFormer [33], and (l) ours.
Figure 7. Visual comparisons on paired datasets. From top to bottom, (a) the inputs are randomly selected from the DICM, LIME, VV, and MEF, respectively. Their corresponding enhanced results generated by (b) LIME [54], (c) RetinexNet [18], (d) DSLR [19], (e) STANet [55], (f) LPNet [15], (g) Zero-DCE [28], (h) URetinex-Net [16], (i) UFormer [16], (j) PairLIE [14], (k) RetFormer [33], and (l) ours.
Sensors 26 00083 g007
Figure 8. The bar charts of the NIQE, PI, and NIQMC of different LLIE methods on unpaired datasets.
Figure 8. The bar charts of the NIQE, PI, and NIQMC of different LLIE methods on unpaired datasets.
Sensors 26 00083 g008
Figure 9. Visual and corresponding histogram comparisons of the original, enhanced, and reference images.
Figure 9. Visual and corresponding histogram comparisons of the original, enhanced, and reference images.
Sensors 26 00083 g009
Figure 10. Qualitative ablation results for each key component of our method on paired LOL-v2 datasets. (a) Raw images. (b) -w/o RCAB. (c) -w/o CSA. (d) -w/o CCB. (e) MSINet (full model).
Figure 10. Qualitative ablation results for each key component of our method on paired LOL-v2 datasets. (a) Raw images. (b) -w/o RCAB. (c) -w/o CSA. (d) -w/o CCB. (e) MSINet (full model).
Sensors 26 00083 g010
Figure 11. Samples of our enhanced results (bottom) were randomly selected from the UCCS benchmark. -/- suggests the NIQE score of the input and its corresponding enhanced images.
Figure 11. Samples of our enhanced results (bottom) were randomly selected from the UCCS benchmark. -/- suggests the NIQE score of the input and its corresponding enhanced images.
Sensors 26 00083 g011
Figure 12. Visual results of our proposed method for enhancing pathological images (the former two columns) and endoscopic images (the latter two columns) with low contrast. From top to bottom, the original images as well as their corresponding enhanced images yielded by our MSINet. -/- suggests the NIQE score of the original input and its corresponding enhanced results.
Figure 12. Visual results of our proposed method for enhancing pathological images (the former two columns) and endoscopic images (the latter two columns) with low contrast. From top to bottom, the original images as well as their corresponding enhanced images yielded by our MSINet. -/- suggests the NIQE score of the original input and its corresponding enhanced results.
Sensors 26 00083 g012
Figure 13. Visual results of our proposed method for some failure instances. From left to right, the original images as well as their correspongding enhanced image yielded by our method. The circle and arrow are used to indicate the areas where image enhancement fails.
Figure 13. Visual results of our proposed method for some failure instances. From left to right, the original images as well as their correspongding enhanced image yielded by our method. The circle and arrow are used to indicate the areas where image enhancement fails.
Sensors 26 00083 g013
Table 1. The PSNR score of the CSA with different combinations of large and small token sizes on the LOL-v1 dataset. Bold text means the best performance.
Table 1. The PSNR score of the CSA with different combinations of large and small token sizes on the LOL-v1 dataset. Bold text means the best performance.
Token Sizes12345
119.8820.1620.7320.8921.03
220.1622.1222.7522.8322.96
320.7322.7523.0324.7224.62
420.8922.8324.7224.6124.32
521.0322.9624.6224.3224.27
622.2623.4124.1024.2824.19
Table 2. Quantitative analysis of different comparison methods on paired benchmarks. Red/blue text means the best/second-best performance.
Table 2. Quantitative analysis of different comparison methods on paired benchmarks. Red/blue text means the best/second-best performance.
MethodsLOL-v1LOL-v2MIT-Adobe 5K
PSNR SSIM LPIPS PSNR SSIM LPIPS PSNR SSIM LPIPS
LIME [54]14.260.51870.265918.240.62240.298717.310.7567 0.1089 
JED [11]17.150.57730.274119.050.64310.302517.720.74760.1139
RetinexNet [18]16.760.42390.423917.410.78110.207412.300.62980.2861
DSLR [19]16.350.61290.230719.370.72730.314624.010.87340.1164
STANet [55]22.030.82130.107319.360.59110.322715.980.78540.1039
Zero-DCE [28]16.430.54210.303416.430.54740.329116.390.78390.1435
DRBN [57]19.860.80460.154720.380.84910.297719.450.75210.1275
KinD [58]21.870.80770.215719.360.82410.265721.950.77260.0833
SNR [62]21.310.85640.152924.140.90280.065820.440.86410.0942
LPNet [15]21.430.80190.095522.090.90140.163624.190.89160.0793
URetinex-Net [16]21.330.83460.108423.500.82570.228221.330.82960.0861
UFormer [31]19.250.76350.302918.820.77450.313423.640.83660.1563
PairLlE [14]19.680.82330.163721.140.81070.234314.690.63980.1589
Retformer [33]22.430.818322.940.89050.214723.670.8274
LightenDiffusion [41]21.650.80470.194723.480.81550.256123.950.89270.0846
QuadPrior [59]21.590.81420.237516.100.76240.224023.610.82770.1564
CIDNet [30]22.590.86710.176625.700.94240.056224.430.89360.0846
END [60]24.570.82070.203116.170.77140.232325.340.83610.0791
Ours24.720.90780.103924.140.90770.100925.590.93460.0681
Table 3. Quantitative assessments of different LLIE methods on the DICM, LIME, VV, and MEF datasets. Red/blue text means the best/second-best performance.
Table 3. Quantitative assessments of different LLIE methods on the DICM, LIME, VV, and MEF datasets. Red/blue text means the best/second-best performance.
MethodsDICMLIMEVVMEF
NIQE PI NIQMC NIQE PI NIQMC NIQE PI NIQMC NIQE PI NIQMC
LIME [54]3.8363.2644.9524.6372.9944.8815.6723.0235.3263.4992.7784.979
JED [11]4.2873.3194.8935.3043.3844.8954.0142.9825.0354.7413.3094.951
RetinexNet [18]3.8623.1814.6345.5183.4274.5313.2782.2444.9594.3552.9024.572
DSLR [19]3.5133.3265.0974.3723.2625.0153.46262.7535.1384.4923.3425.175
LPNet [15]3.7523.5334.8413.7523.5334.8414.9963.3594.9774.1162.9585.144
Zero-DCE [28]3.1693.0554.8963.1693.0554.8963.2612.5015.3703.3692.4294.943
URetinex-Net [16]3.4253.1465.1133.4253.1465.1133.3232.5735.0943.8412.8585.083
MIRNet [56]3.3843.5285.0614.6253.0284.9373.5132.7745.1064.0453.0975.036
UFommer [31]5.0542.9544.8857.9273.0614.8935.3082.9634.6874.2582.9464.857
PairllE [14]4.0162.8934.7914.5192.8134.7734.2692.8464.2934.3372.8614.839
LightenDiffusion [41]3.7413.1325.1063.9683.0164.7622.9682.5534.8373.8432.9614.927
Ours2.8162.5535.2194.8263.7585.3644.6812.4385.5224.6372.4395.274
Table 4. Quantitative comparison of existing LLIE methods on the LOL-v1 dataset. Bold text means the best performance.
Table 4. Quantitative comparison of existing LLIE methods on the LOL-v1 dataset. Bold text means the best performance.
MethodAGLVarLSTD Δ E BD
RetinexNet [18]5.1493.1011.0627.3420.273
KinD [58]6.1232.3421.0026.2060.231
END [60]6.5542.6572.3745.1100.159
Zero-DCE [28]5.2211.6971.0107.0910.139
SNR [62]6.3352.9971.8922.3100.173
CIDNet [30]6.6873.0280.9943.7400.211
LightenDiffusion [41]5.9085.0171.9305.1070.193
EnlightenGAN [27]5.7121.1012.3714.5290.123
PairLIE [14]5.8262.0171.9094.9370.187
DSLR [19]5.2904.0072.0624.3870.208
UFormer [31]5.2503.3942.1074.3200.118
URetinex-Net [31]5.9371.9974.0047.8630.211
Ours7.3581.0750.9373.8710.109
Table 5. Computational complexity comparison of existing LLIE methods on the LOL-v1 dataset. Bold text means the best performance.
Table 5. Computational complexity comparison of existing LLIE methods on the LOL-v1 dataset. Bold text means the best performance.
MethodParam (M)Flops (G)Time (s)
RetinexNet [18]1.236.790.5217
KinD [58]8.497.440.6445
END [60]8.36270.420.7963
Zero-DCE [28]1.215.210.0079
SNR [62]4.0126.350.5141
CIDNet [30]1.988.030.7869
LightenDiffusion [41]101.71210G1.2001
EnlightenGAN [27]8.647.880.6501
PairLIE [14]2.1619.240.2971
DSLR [19]14.3122.950.9210
UFormer [31]5.2010.680.6298
URetinex-Net [31]26.2790.610.8902
Ours1.998.240.1009
Table 6. Ablation study on the MIT-Adobe 5K, LOL-v1, and LOL-v2 datasets. Red/blue text means the best/second-best performance.
Table 6. Ablation study on the MIT-Adobe 5K, LOL-v1, and LOL-v2 datasets. Red/blue text means the best/second-best performance.
Ablated ModelMIT-Adobe 5KLOL-v1LOL-v2
PSNR SSIM LPIPS PSNR SSIM LPIPS PSNR SSIM LPIPS
-w/o RCAB21.300.7380.07120.810.7930.18723.120. 8120.167
-w/o CSA22.520.8040.08923.190.8060.17423.910.8780.157
-w/o CCB23.090.8310.10622.140.8550.18723.680.8930.169
full model25.590.9340.06724.720.9030.10224.140.9080.101
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lu, H.; Qian, C.; Wang, Z.; Liu, Z. Multi-Scale Interactive Network with Color Attention for Low-Light Image Enhancement. Sensors 2026, 26, 83. https://doi.org/10.3390/s26010083

AMA Style

Lu H, Qian C, Wang Z, Liu Z. Multi-Scale Interactive Network with Color Attention for Low-Light Image Enhancement. Sensors. 2026; 26(1):83. https://doi.org/10.3390/s26010083

Chicago/Turabian Style

Lu, Haoxiang, Changna Qian, Ziming Wang, and Zhenbing Liu. 2026. "Multi-Scale Interactive Network with Color Attention for Low-Light Image Enhancement" Sensors 26, no. 1: 83. https://doi.org/10.3390/s26010083

APA Style

Lu, H., Qian, C., Wang, Z., & Liu, Z. (2026). Multi-Scale Interactive Network with Color Attention for Low-Light Image Enhancement. Sensors, 26(1), 83. https://doi.org/10.3390/s26010083

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop