Next Article in Journal
FloCyT: A Flow-Aware Centroid Tracker for Cell Analysis in High-Speed Capillary-Driven Microfluidic Flow
Previous Article in Journal
LDLK-U-Mamba: An Efficient and Highly Accurate Method for 3D Rock Pore Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DCD-Net: Decoupling-Centric Decomposition Network for Low-Light Image Enhancement

College of Mechanical and Electronic Engineering, Nanjing Forestry University, Nanjing 210037, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(22), 7038; https://doi.org/10.3390/s25227038 (registering DOI)
Submission received: 16 October 2025 / Revised: 13 November 2025 / Accepted: 17 November 2025 / Published: 18 November 2025
(This article belongs to the Section Sensing and Imaging)

Abstract

This paper presents a Decoupling-Centric Decomposition network for Low-Light Image Enhancement (DCD-Net). The DCDNet addresses two key challenges: (1) existing methods center on how to design the enhancement network and ignore the decomposition network’s critical role to decouple reflectance and illuminations as the first step and (2) existing decomposition networks process images directly (or with pre-denoising), ignore their compliance with the Retinex theory. Specifically, by centering on illumination–reflectance decoupling refinement, DCDNet operates without reliance on supplementary enhancement networks. It consists of a preprocessing network and a decomposition network. The preprocessing network adopts a self-supervised learning mechanism to suppress Retinex-incompatible features in the input image, thereby improving the quality of Retinex decomposition. Within the decomposition network, the reflectance net is designed to suppress the contamination of illumination on reflectance restoration by the Dual-Gated Directional Reflectance Module (DGD-RM) and Reflectance-Guided Multi-head Self-Attention (RG-MSA), while the illumination net utilizes DCT to achieve local–global illumination estimates. Comprehensive experiments were conducted on two benchmark datasets (LOL and MIT) and five unpaired datasets. The quantitative results on different datasets are as follows (measured by PSNR and SSIM): LOL v1 (20.87, 0.770) and MIT (21.66, 0.864). The average NIQE across five unpaired datasets is 3.420. Both qualitative and quantitative analyses demonstrated the superiority of DCDNet over state-of-the-art methods. Moreover, ablation studies demonstrated the effectiveness of each module in DCDNet.

1. Introduction

Low-light images are characterized by low brightness, low contrast, and high noise, which fail to satisfy human visual perception requirements. Critically, this degradation can severely degrade the performance of high-level computer vision tasks [1,2,3]. Therefore, low-light image enhancement (LLIE) is an important yet challenging task in computer vision. It aims to enhance visual quality by improving image contrast and restoring detail information, and to provide reliable inputs for downstream vision tasks.
With the advancement of deep learning, convolutional neural networks (CNNs) have been widely applied to LLIE. Existing methods primarily fall into two categories: the first category directly learns end-to-end (Figure 1a) mappings from low-light to normal-light images using CNNs [4,5], whose data-driven nature often leads to color distortion due to the lack of physical constraints. The second category combines Convolutional Neural networks (CNNs) with Retinex theory (Figure 1b). These methods decompose an input image into illumination (L) and reflectance (R) components using dedicated decomposition networks, i.e., I = R L , where ⊙ denotes element-wise multiplication. Wei et al. [6] proposed RetinexNet as a seminal work, completing the decomposition–enhancement pipeline for real-world scenes. Subsequent research has progressively optimized the enhancement network architecture to achieve better performance in LLIE tasks [7,8,9]. Some studies adopt denoising networks as an important step for images [10,11]. Zhu et al. [11] introduced RRDNet, which incorporates a noise component in addition to illumination and reflectance during decomposition. Recently, a third category of methods leveraging CNN-Transformer hybrid architectures has emerged, aiming to harness the local feature extraction prowess of CNNs and the global contextual modeling capabilities of Transformers [12,13]. However, these end-to-end methods often result in highly complex models with limited physical interpretability. Furthermore, as most employ a supervised learning paradigm, their performance is constrained by the availability of paired training data, which consequently limits their generalization capability on real-world unlabeled datasets.
Recently, Fu et al. [14] proposed the PairLIE method, introduces paired low-light images as input, innovatively leveraging the physical prior of reflectance consistency to achieve low-light image enhancement without relying on complex enhancement networks. Furthermore, it replaces the denoising network with a Retinex decomposition-adapted disturbance-removal network, providing important theoretical foundations and architectural references for our research. However, this approach exhibits two notable limitations: firstly, it fails to adequately address the critical role of the illumination component in the enhancement process; secondly, relying solely on physical priors and relatively simple network designs is insufficient for achieving comprehensive feature extraction of the reflectance component, particularly in removing illumination artifacts from it.
Despite architectural differences, these methods share a common limitation: sub-optimal constraints during decomposition lead to incomplete decoupling of illumination and reflectance. This compels heavy reliance on enhancement networks to compensate for decomposition flaws, ultimately dictating final output quality. Additionally, using denoising methods as a preprocessing step tends to cause blurring, while employing it as a postprocessing step can lead to noise amplification—both scenarios fail to adequately adhere to the adaptability required by the Retinex decomposition theory.
To address the aforementioned issues, we propose DCDNet. Our network centers on the decomposition network to optimize illumination–reflectance decoupling without relying on additional enhancement networks. It comprises two core components: (1) the Dense Preprocessing network (DPnet) and (2) a decomposition network containing a reflectance network (Rnet) and an illumination network (Lnet). DPnet leverages dense connectivity for cross-layer feature reuse. By hierarchical feature extraction, it progressively eliminates Retinex-incompatible features, thereby optimizing the subsequent decomposition efficacy. As an unsupervised model, DCDNet operates without normal-light reference images. To address this, it leverages the physical prior of reflectance consistency between paired low-light images and employs a CNN-Transformer architecture to extract reflectance features by integrating both local and global information. Rnet integrates a CNN-Transformer architecture. Its CNN stage employs a Dual-Gated Directional Reflectance Module (DGD-RM), leveraging directional convolution’s high-frequency sensitivity to mitigate the contamination of illumination on reflectance. The transformer stage utilizes Reflectance-Guided Multi-Head Self-Attention (RG-MSA) for adaptive feature restoration. LNet integrates the Discrete Cosine Transform (DCT) to introduce detailed local illumination information, thereby achieving a joint local–global illumination estimation.
In summary, the contributions of this paper are as follows:
  • We propose DCDNet, an LLIE framework that eliminates additional enhancement networks by refining illumination–reflectance decoupling through its dedicated decomposition network.
  • To achieve local and global feature extraction for unsupervised reflectance estimation, our Rnet leverages a CNN-Transformer hybrid network guided by the reflectance consistency prior of image pairs.
  • Dense Preprocessing network (DPnet) uses a densely connected architecture for cross-layer feature reuse, optimizing Retinex decomposition. Within its Rnet component, the Dual-Gated Directional Reflectance Module (DGD-RM) mitigates illumination via directional convolution and gating. The Reflectance-Guided Multi-Head Self-Attention (RG-MSA) mechanism employs reflectance as a key cue to guide long-range feature restoration. Lnet utilizes the Discrete Cosine Transform (DCT) to introduce local illumination information to achieve a local–global illumination estimate.
  • Extensive quantitative and qualitative experiments demonstrate that our final trained DCDNet surpasses existing methods across multiple datasets.

2. Related Works

2.1. Low-Light Image Enhancement

Traditional LLIE methods, including histogram equalization [15] and Retinex methods [16,17]. Histogram equalization methods enhance contrast by redistributing the brightness histogram. However, these methods are prone to causing local overexposure and artifacts. Methods based on the Retinex decompose the input image into illumination and reflectance. This is achieved by introducing various handcrafted prior constraints to optimize the decomposition process. Due to their heavy reliance on handcrafted priors and manual parameter tuning, these models suffer from poor adaptability and are prone to causing color distortion and unnatural textures [18].
In recent years, deep learning techniques have been widely applied in the field of LLIE. Several methods [6,7,8,9,19] have significantly improved enhancement quality based on a decomposition–enhancement network architecture. Wu et al. [9] proposed URetinex-Net, which models the Retinex theory as an iteratively unfoldable network structure, incorporating it as a physical prior in the deep network to guide more accurate image decomposition and enhancement. Other approaches [11,20,21,22] address the LLIE task by designing specialized decomposition networks coupled with carefully constructed loss functions. Zhu et al. [11] introduced RRDNet, which incorporates a noise component in addition to illumination and reflectance during decomposition, effectively suppressing the amplification of noise during image contrast stretching. Recently, transformers have been widely used in LLIE tasks [12,13,23,24,25,26]. Xu et al. [12] proposed an SNR-aware CNN-Transformer hybrid model, SNR-Net, for low-light image enhancement. However, its high computational complexity limits further application. Building upon this, Cai et al. [13] introduced a single-stage Retinex framework and an illumination-guided Transformer architecture, Retinexformer, which successfully applied the Transformer to the LLIE task for the first time and significantly reduced computational overhead. Nevertheless, under complex lighting conditions, existing methods often struggle to achieve a good balance between denoising and enhancement. Li et al. [26] proposed DEnet, an interpretable framework based on Discrete Cosine Transform (DCT) frequency-domain decomposition and self-supervised training with sub-image pairs. It employs a degradation-aware Transformer network to achieve integrated restoration of noise and illumination degradations in real-world low-light images. However, in real-world scenarios with severe noise or extreme illumination unevenness, the enhanced output still exhibits issues such as partial detail loss or residual noise. Unlike most previous methods that use a single low-light image as input, Fu et al. [14] were the first to propose using paired low-light images for training, a method named PairLIE, aiming to provide a general solution for low-light enhancement. However, due to its relatively simple network structure and insufficient consideration of the crucial role of illumination in the enhancement process, the method still falls short in balancing image enhancement and noise suppression, thereby limiting its overall performance.

2.2. Discrete Cosine Transform (DCT)

The Discrete Cosine Transform (DCT), being the core of image compression standards like JPEG, holds a foundational importance in the field of image processing. Its exceptional energy compaction property condenses image information into low-frequency and high-frequency components: low frequencies carry the fundamental structural contours and illumination information, while high frequencies contain details, textures, and noise. In recent years, researchers have deeply explored this frequency-domain prior, integrating it with deep learning to address image restoration problems [26,27,28,29,30]. Galteri et al. [27] achieved face restoration by learning to separate and process DCT components in an end-to-end manner. Li et al. [26], the DCT to separate images into different frequency components through frequency domain decomposition, constructing physical priors that guide the network to more effectively process noise and details during the enhancement process. These methods collectively corroborate that the fine-grained manipulation of high-frequency DCT and low-frequency DCT components is key to guiding models in balancing macro-structure preservation and micro-detail recovery.

3. Methods

In this section, we first introduce the workflow of the proposed DCDNet for LLIE, followed by detailing the loss functions of the DCDNet.

3.1. Network Architecture

As shown in Figure 2, DCDNet consists of a Dense Preprocessing network (DPnet) and a Decomposition network. The Decomposition network consists of the Reflectance network (Rnet) and the illumination network (Lnet). DPnet aims to eliminate features incompatible with Retinex decomposition. Rnet and Lnet serve as decomposition networks to obtain the reflectance and illumination components, respectively. Table 1 illustrates the dimensional changes and primary convolutional choices of each network module. The module details will be elaborated in subsequent sections. We show the pseudocode of DCDNet in Figure 3.
In the training phase, paired low-light images I 1 and I 2 are processed by the DPnet to obtain optimized versions i 1 and i 2 . These are then decomposed into illumination ( L 1 and L 2 ) and reflectance ( R 1 and R 2 ) using Lnet and Rnet, respectively. In the testing phase, an input low-light image I is processed through DCDNet to obtain the final illumination L and reflectance R. The enhanced image I e n is then reconstructed using the following formulation:
I e n = f ( L ) R = L λ R
where λ is the illumination correction factor, I e n denotes the enhanced image.

3.2. DCDNet Algorithm Description

The proposed DCDNet framework consists of three main components: DPnet for Retinex decomposition, Lnet for illumination estimation, and Rnet for reflectance processing. The complete algorithm proceeds as follows:
Network Components:
  • DPnet: Optimizes Retinex decomposition using a dense structure
  • Lnet: Illumination network combining DCT and CNN operations
  • Rnet: Reflectance network incorporating DGD-RM, RG-MSA, and CNN modules
Forward Propagation:
  • Given input image I input , compute feature representation: X = DPnet ( I input )
  • Estimate illumination map: L = Lnet ( X )
  • Extract reflectance component: R = Rnet ( X )
  • Return the triple: ( X , L , R )
DPnet Architecture:
  • Extract initial features: x 1 = ReLU ( Conv 3 × 3 ( I ) )
  • Process with residual connections: x 2 = ReLU ( Conv 3 × 3 ( x 1 ) + x 1 )
  • Continue dense propagation: x 3 = ReLU ( Conv 3 × 3 ( x 2 ) + x 2 + x 1 )
  • Further feature aggregation: x 4 = ReLU ( Conv 3 × 3 ( x 3 ) + x 3 + x 2 + x 1 )
  • Generate final output: x = Sigmoid ( Conv 3 × 3 ( x 4 ) )
Lnet Architecture:
  • Apply DCT transformation: x dct = DCT ( x )
  • Concatenate features: x concat = Concat ( x , x dct )
  • Fuse features: x fused = Conv 1 × 1 ( x concat )
  • Generate illumination map: L = Sigmoid ( CNN 4 - layer ( x fused ) )
Rnet Architecture:
  • Apply DGD-RM module: x 0 = DGD - RM ( x )
  • Process with RG-MSA: x 1 = RG - MSA ( X , x 0 )
  • Generate reflectance map: R = Sigmoid ( CNN 3 - layer ( x relight ) )

3.3. Dense Preprocessing Network

Serving as the preprocessing network, DPnet aims to eliminate features incompatible with Retinex decomposition, thereby optimizing the decomposition process. As shown in Figure 3, the network takes a low-light image I R H × W × 3 as input. The network employs a densely connected architecture comprising five 3 × 3 convolutions for progressive feature extraction. Layers 1–4 implement cross-layer feature reuse through element-wise addition of features from all preceding layers, where each layer output f i is computed as follows:
f i = R e L U ( C o n v i ( f i 1 ) + j = 1 i 1 f j )
where i = 2 , 3 , 4 , f 1 = R e L U ( C o n v 1 ( I ) ) . Finally, a 3 × 3 convolution followed by a Sigmoid function normalizes the output i R H × W × 3 to the range [ 0 , 1 ] .

3.4. Decomposition Network

The decomposition network comprises reflectance network (Rnet) and illumination network (Lnet).

3.4.1. Reflectance Network

As shown in Figure 4, the Rnet (Figure 4a) integrates CNN-Transformer architecture, composed of DGD-RM (Figure 4b) and RG-MSA (Figure 4c).
In DGD-RM (Figure 4b), the input features i 1 R H × W × 3 undergo feature extraction via 3 × 3 convolution, expanding the channel dimension to yield R R H × W × C . Subsequently, a 7 × 7 depth-wise separable convolution extracts long-range contextual features R 1 R H × W × C . Two parallel 1 × 1 convolutions project R 1 into R 2 R H × W × 3 C and R 3 R H × W × 3 C . After applying ReLU6 activation to R 2 , the two features undergo element-wise multiplication. A final 1 × 1 convolution projects the result back to R 4 R H × W × C . This gating mechanism implements channel-adaptive weighting, suppressing low-frequency illumination artifacts and enhancing high-frequency reflectance through the element-wise multiplication. Subsequently, bidirectional feature compression employs dual-branch directional convolutions as follows: a 7 × 1 and 1 × 7 depth-wise separable convolution layer process the input to produce R w R H × W × C 2 and R h R H × W × C 2 . These outputs are concatenated along the channel dimension and fused by a 1 × 1 convolution generating R 5 R H × W × C to effectively suppress low-frequency illumination artifacts. To preserve ungated detail features, a residual connection adds the output features to the original input. Finally, 3 × 3 convolution projects the features to match the RGB channels, outputting the reflectance map R ^ R H × W × 3 .
RG-MSA (Figure 4c) processes spatial features as tokens for self-attention computation. The preprocessed feature map i 1 R H × W × 3 is reshaped into primary tokens X R H W × C . To address spatially varying reflectance distortion caused by divergent illumination, the DGD-RM mitigates illumination interference, generating refined reflectance features R ^ R H × W × 3 . RG-MSA leverages this reflectance information to guide region-adaptive feature restoration by aligning R ^ with X through channel expansion via 1 × 1 convolution and reshaping into reflectance prior tokens Y R H W × C . Then X and Y are split into k heads as follows:
X = [ X 1 , X 2 , , X k ]
Y = [ Y 1 , Y 2 , , Y k ]
where X i , Y i R H W × d k , d k = c k , and i = 1 , 2 , , k . We show the situation with k = 1 and omit some details for simplification. For each h e a d i , three fully connected ( f c ) layers without bias are used to linearly project X i into query elements, Q i R H W × d k , key elements K i R H W × d k , and value elements V i R H W × d k as follows:
Q i = X i W Q i T , K i = X i W K i T , V i = X i W V i T
where W Q i , W K i , and W V i R d k × d k represent the learnable parameters of the f c layers and T denotes the matrix transpose. Then the self-attention for each h e a d i is formulated as follows:
Attention ( Q i , K i , V i , Y i ) = ( Y i V i ) softmax K i T Q i α i
where α i R 1 is a learnable scaling parameter. ⊙ is element-wise multiplication. The k heads are then concatenated, processed through f c layer, augmented with learnable positional encoding P R H W × C , and reshaped to output feature F o u t R H × W × C . Finally, the features undergo refinement through two 3 × 3 convolutions with ReLU activation applied only to the first layer, followed by normalization via Sigmoid function to constrain the output to the range 0 , 1 . This yields the final reflectance features R R H × W × 3 .

3.4.2. Illumination Network

The Lnet (Figure 5) integrates a DCT module with CNNs. Unlike existing networks that typically estimate global illumination but ignore local illumination information, Lnet extracts low-frequency features i l R H × W × 3 from the input i 1 R H × W × 3 using DCT, utilizing them as local illumination information compensation. These features are channel-wise concatenated with i 1 , fused via 1 × 1 convolution, and processed through four consecutive 3 × 3 convolutional layers. The first three layers use ReLU activation, while the final layer employs Sigmoid activation to normalize the output L R H × W × 1 to 0 , 1 .

3.5. Loss Function

In this section, we present the loss functions of DCDNet, including Retinex Loss, Reflectance Loss, and Overall Loss.

3.5.1. Retinex Loss

Following Retinex theory, we decompose a low-light image into illumination and reflectance components. Inspired by PairLIE [14], we introduce two regularization losses: (1) the Preprocessing Loss L P , which first eliminates incompatible features from the input image to optimize decomposition, and (2) the Retinex Decomposition Loss L R to ensure physically plausible decomposition. These losses are formulated as follows:
L P = | | I 1 i 1 | | 2 2
L R = R L i 2 2 + R i / s t o p g r a d ( L ) 2 2 + L L 0 2 2 + L 1
where i denotes the preprocessed image obtained after optimization through the L P loss, ∇ represents the horizontal and vertical gradients. The reconstructed image is computed as | | R L i | | 2 2 . L 0 denotes the initial illumination estimation, computed as the maximum value across RGB channels. After obtaining the illumination, the reflectance component is calculated through pixel-wise division between the preprocessed image and its illumination map, i.e., | | R i / s t o p g r a d ( L ) | | 2 2 . s t o p g r a d ( L ) represents stopping the gradient of the illumination map.

3.5.2. Reflectance Loss

Spatial Consistency Loss L S p a encourages spatial coherence by preserving differences between adjacent regions in the low-light image pairs reflectance features.
L S p a = 1 K i = 1 K j Ω ( i ) R 1 i R 1 j R 2 i R 2 j 2
where k is the number of local regions, and Ω i is the four neighboring regions (top, down, left, right) centered at the region i. We denote R 1 and R 2 as the average intensity values of local regions within the paired reflectance maps, respectively.
Reflectance Consistency Loss L C is calculated based on low-light image pairs and the Retinex theory. Reflectance represents the physical properties of the image. So the low-light image pairs share the same R.
L C = | | R 1 R 2 | | 2 2
where R 1 and R 2 refer to the reflectance components of paired low-light images. L C makes the network predict the same reflectance components.

3.5.3. Overall Loss

The overall loss function for training our DCDNet is a liner combination of each loss as follows:
L A l l = w 1 L P + w 2 L R + w 3 L C + w 4 L S p a
where w 1 , w 2 , w 3 , and w 4 denote the weights.

4. Experiments

In this section, we present the experimental setup and results of DCDNet.

4.1. Datasets and Metrics

We collect low-light image pairs from SICE (part2) [31] and LOL (training set) [6], which contain multi-exposure images. we collect 324 sequences (a total of 1000 low-light images). During the training phase, we randomly select two images from each sequence to form a pair. Then, we benchmark our method on the following datasets: the original LOL [6] (commonly referred to as LOL v1), its extended version LOL v2 [32] (which includes the real-captured LOL v2-real and synthetically generated LOL v2-synthetic subsets) and another 115 randomly sampled image pairs from the MITAdobe FiveK Dataset (MIT for short) as the benchmark evaluation dataset. Additionally, for no-reference evaluation, five widely accepted benchmark datasets are adopted, including MEF [33], DICM [34], LIME [35], Fusion [6], and VV [36]. For evaluation metrics, we adopt four full-reference metrics, namely the Peak Signal-to-Noise Ratio (PSNR), the Structural (SSIM) [37], the Learned Perceptual Image Patch Similarity (LPIPS) [38] and one no-reference metric, namely the Natural Image Quality Evaluator (NIQE) [39].

4.2. Implementation Details

We implement DCDNet with PyTorch 2.0.0 and Python 3.9.19. We use ADAM with the initial learning rate of 1 × 10 4 to optimize the network. The number of training epochs is set to 400. The learning rate is half-decayed per 100 epochs. Different datasets empirically provide different correction factors λ . The correction coefficient λ is set to 0.18 by default. For scenarios under extremely low illumination, such as those in the LOL v1 dataset, λ is set to 0.14. For the MIT dataset, which exhibits relatively adequate illumination, λ is set to 0.35. As for the hyper-parameters w 1 , w 2 , w 3 , and w 4 in Equation (11), we set w 1 = 500 , w 2 = w 3 = w 4 = 1 . Finally, all experiments are conducted on RTX 4060 GPU.

4.3. Comparison Results

We compare the proposed model with state-of-the-art methods, including Retinex-Net [6], MBLLEN [5], KinD [7], URetinex-Net [9], SNR-Aware [12], Retinexformer [13], CIDNet [40], DiffLight [41], CWNet [42], RRDNet [11], EnlightenGAN [20], UEGAN [43], ZeroDCE [22], RUAS [28], SCI [44], PairLIE [14], ResQ-Net [45], and DEnet [26]. The first six methods are supervised, and the remaining methods are unsupervised together with ours.

4.3.1. Qualitative Comparisons

Table 2 presents a quantitative comparison of method performance on reference-based datasets (LOL v1, LOL-v2-real, LOL-v2-synthetic, and MIT) using PSNR, SSIM, and LPIPS metrics. As an unsupervised approach, our method achieves state-of-the-art (SOTA) performance on LOL v1, LOL v2-real, and LOL v2-synthetic among unsupervised methods, and attains the best overall results on the MIT dataset.
Table 3, we evaluate method performance on five no-reference benchmarks using the NIQE metric. Our method delivers optimal or competitive results on the DICM, LIME, and MEF datasets, and achieves the best average performance across all five datasets.
Table 4 compares model complexity to assess computational efficiency. This analysis focuses specifically on deep-learning-based methods, which benefit from GPU acceleration. Our method exhibits the lowest FLOPs among all supervised approaches, while some unsupervised zero-reference methods (e.g., Zero-DCE, SCI, and ResQ-Net) naturally excel in model complexity. Our approach still surpasses several classical unsupervised methods such as EnlightenGAN, RRDNet, and DENet. Notably, compared to PairLIE—an unsupervised method with comparable performance—our method achieves comprehensive performance gains while significantly reducing model complexity, thanks to a more efficient network architecture that substantially cuts down the number of channels. This reflects our deliberate effort to strike an optimal balance between performance and efficiency. Although methods such as Zero-DCE, RUAS, and SCI exhibit lower FLOPs and Params compared to our approach, their reliance on sequential iterative optimization or recursive structures, as well as their foundation on traditional CNN architectures with heavy dependence on convolution, limits GPU parallelization, resulting in higher inference times than our method. Specifically, our DPnet enhances Retinex decomposition through densely connected convolutional modules, enabling parallel execution of Rnet and Lnet during the decomposition phase. Rnet adopts a hybrid CNN-Transformer design, where the CNN component utilizes multi-scale parallel convolutional paths to maximize parallelism. This network architecture improves computational continuity and operator fusion, thereby reducing inference time on hardware.
Additionally, as our method is built on a Transformer-based architecture, we conduct an in-depth qualitative comparison with several existing Transformer-based methods: SNR-Aware, Retinexformer, DENet, CIDNet and ours, and diffusion-based methon: DiffLight.
When compared to supervised Transformer methods such as SNR-Aware, Retinexformer, CIDNet, DENet, DiffLight and CWNet. Our approach—which follows an unsupervised paradigm and does not rely on ground-truth illumination labels during training—exhibits slightly lower performance on reference-based datasets. This outcome is consistent with the typical performance gap between supervised and unsupervised frameworks, which represents an area for future optimization. More notably, in comparison with the unsupervised method DE-net, our approach achieves superior performance across all four reference-based datasets, demonstrating clear advantages among unsupervised peers.
On non-reference datasets, our model exhibits more marked advantages; our unsupervised method achieves superior performance compared to CWNet, which exhibits the best overall performance among the supervised methods. This demonstrates the strong generalization capability of our method.
In terms of model complexity, our approach demonstrates significant efficiency advantages. On key metrics including FLOPs, number of parameters, and inference time, our method notably outperforms advanced supervised Transformer methods (such as Retinexformer, SNR-Aware, and CIDNet) as well as the unsupervised DE-net. It is particularly noteworthy that the diffusion-based DiffLight method exhibits the highest FLOPs. This efficiency advantage demonstrates that our method achieves highly competitive performance with substantially reduced computational overhead.

4.3.2. Visual Comparisons

In addition to the quantitative evaluations conducted earlier, Figure 6, Figure 7, Figure 8, Figure 9, Figure 10 and Figure 11 present visual comparisons to qualitatively compare enhancement results across different methods.
Figure 6 and Figure 7 demonstrate examples of noise suppression and detail restoration on LOL v1 and LOL v2, respectively. Although our network does not incorporate additional denoising or enhancement modules, it achieves more accurate restoration of reflectance and illumination components in low-light images through a more rational decomposition network design. Specifically, by suppressing interference from illumination information in the reflectance decomposition network, we obtain purer image reflectance. All three examples demonstrate that our method can effectively suppress noise in dark regions generated during image enhancement. Compared to other methods, our approach preserves clearer and more complete edge structures. Furthermore, in the complex scenario shown in the third row, our method not only effectively suppresses noise but also better retains critical details in the image.
Figure 8 shows examples in scenarios on unpaired datasets. We selected the top 5 methods based on the results in Table 3 for visual presentation. Most of these scenes exhibit uneven illumination distribution, thus requiring enhancement methods that combine global and local illumination information. Benefiting from our illumination network incorporating low-frequency DCT components as supplementary local illumination information, our method achieves visually superior results compared to other approaches. For example, in the third-row example, our method effectively avoids underexposure issues in the dark area of the lower-left corner. In the fourth-row example, our method successfully prevents overexposure problems in the bright areas of the light source.
Figure 9 shows the decomposed outputs of the method for different correction factors ( λ ). The reflectance component reveals the image’s textures and details, while the illumination component remains piecewise smooth and texture-free, indicating that our method accurately decomposes low-light images. We further present enhanced results under different correction factors. When the correction factor λ is less than 0.2, the image becomes overexposed; when λ exceeds 0.4, the image appears severely underexposed. For each dataset during the testing phase, we assign specific correction factors; for instance, λ = 0.3 for the MIT dataset. The values of λ were selected to optimize enhancement performance for the specific illumination characteristics of each dataset. Users can adjust the correction factor λ during testing according to their personal preferences for image style.
Figure 10 shows the different illumination and reflectance components decomposed by our method and PairLIE. By suppressing interference from illumination information in the reflectance decomposition network, the specific performance of the reflectance maps demonstrates that, compared to PairLIE, our results contain richer textures and details while reducing potential noise-induced artifacts. In addition, our method shows advantages in characterizing both global illumination features and local detail variations. In the first two examples, our method captures local illumination details that PairLIE fails to represent, such as the books on the shelf and the stone piers at the doorway. In the next two examples, our results better align with the global characteristics of illumination, such as the light variations on the wall.
Figure 11 shows face detection in dark environments using challenging DARK FACE samples containing 28 low-illumination faces. Employing the Dual Shot Face Detector (DSFD) [46], the raw detection on this challenging sample yields very poor performance with only 1 face detected and 27 missed. URetinexNet and RRDNet detected 12 and 13 faces, respectively. ResQ, ZeroDCE, RUAS, SCI, and PairLIE detected 22, 20, 18, 18, and 19 faces, respectively. URetinexNet failed to recognize valid faces due to excessive blurriness in the enhanced images. RRDNet yielded unsatisfactory detection results because of the relatively dark enhanced images. Although ZeroDCE, RUAS, SCI, and PairLIE produced visually pleasing images, practical detection results demonstrate their ineffectiveness in detecting darker faces. Our method effectively illuminates faces in extremely dark regions while preserving details in well-lit areas, ultimately detecting 22 out of 28 faces and achieving the best detection performance, which validates the practicality of our approach.

4.4. Ablation Study

We conduct ablation studies to validate module effectiveness. For simplicity, evaluations use the LOL v1 dataset, and the respective performances are quantitatively evaluated using the combination of PSNR, SSIM, and LPIPS. Constrained by decomposition theory, illumination and reflectance networks remain intact. PairLIE’s [14] standard-convolution implementation serves as baseline for module-level substitution. Since RG-MSA takes the output of DGD-RM as its input, we conduct ablation studies on both modules as an integrated whole. (a) without DGD-RM and RG-MSA (in reflectance branch); (b) without DCT (in illumination branch); (c) without L S p a (in reflectance branch); and (d) DPnet compare to PairLIE’s Pnet (in Projection branch).
Quantitative results on the LOL v1, LOL v2-real, LOL v2-syn, and MIT datasets in Table 5 confirm that each component critically enhances DCDNet’s performance. Visual results on the LIME dataset in Figure 12, removal of DGD-RM and RG-MSA induces severe color distortion in outputs due to reflectance contamination from the illumination component. When the DCT module is omitted, color reproduction remains largely intact; however, zoomed-in views reveal edge distortion around structural borders. Without L S p a , images maintain overall naturalness but exhibit reduced contrast, particularly in background elements such as picture frames.
Figure 13 illustrate the reconstruction errors on LOL v1, LOL v2-real, LOL v2-syn, and MIT. The blue line represents decomposition using the original low-light image called baseline, i.e., | L R I | ; both orange and green lines denote decomposition using preprocessed images, i.e., | L R i | . The orange line corresponds to the Pnet proposed in PairLIE [14], while the green line represents our DPnet. As can be seen from the figures, the reconstruction errors for the majority of images across the four datasets are lower than those of the baseline and DPnet. Table 6 quantitatively demonstrates that our average reconstruction error is also lower than that of the baseline and DPnet. Experimental results demonstrate that our DPnet not only exhibits a distinct advantage in optimizing Retinex decomposition but also outperforms the baseline and PairLIE methods in terms of both per-image optimization and the average error.

4.5. High-Frequency DCT and Low-Frequency DCT

There are two types of DCT: high-frequency DCT (HDCT for short) and low-frequency DCT (LDCT for short). We have experimentally validated the selection of low-frequency DCT in our DCDNet. In this part, we keep the rest of the network unchanged and separately introduce HDCT and LDCT to observe their specific contributions to the network performance.
Table 7 quantitatively compares HDCT and LDCT’s performances on reference datasets (LOL v1, LOL v2-real, LOL v2-synthetic, and MIT) using PSNR, SSIM, and LPIPS metrics. LDCT achieves the optimal performance on the LOL v1, LOL v2-real, and LOL v2-synthetic datasets, and also demonstrates superior overall performance compared to HDCT on the MIT dataset.
Figure 14 shows that when HDCT introduces illumination reference, it additionally introduces corresponding noise, leading to illumination artifacts in the illumination component. For example, in the first row, several non-existent noise bands appear at the door, and in the second row, the partitions at the roulette wheel are ambiguous. Figure 15 demonstrates that HDCT, compared to LDCT, has advantages in local detail restoration when incorporating illumination reference. For instance, in the first row, the text on the book is clearer, and in the second row, the houses in the background are better recovered.
Experimental results indicate that both LDCT and HDCT contribute positively to image illumination estimation. LDCT is particularly effective at recovering global illumination structures, while HDCT, by leveraging high-frequency priors, shows advantages in restoring fine details. However, under the same network configuration, our comprehensive evaluation reveals that LDCT achieves overall superior performance. We hypothesize two primary reasons for this: (1) The high-frequency priors in HDCT are prone to amplifying noise, which can lead to visible artifacts in the estimated illumination. (2) The Retinex theory posits that the illumination layer primarily consists of low-frequency components, making LDCT a more physically plausible choice.
Based on this analysis, we adopt LDCT in our illumination estimation network.

4.6. Analysis of Failure Cases

To comprehensively evaluate the performance of DCDNet, we provide some failure cases and conduct an analysis. As shown in Figure 16, when processing extremely low-light areas in images, our method may exhibit localized detail loss and color deviation issues. During the training phase, our network takes paired low-light images as input and leverages the physical prior of reflectance consistency to extract reflectance through the designed Rnet. However, in extremely low-light conditions, severe noise interference leads to significant degradation of image details. Under such circumstances, using paired inputs may introduce additional degradation issues. Moreover, due to the absence of normally lit images in the training data as reference, the model struggles to effectively distinguish between noise and authentic image information. Since our DPnet is primarily designed to optimize Retinex decomposition rather than denoising, this ultimately results in the loss of local details. Regarding the color deviation issue, the Lnet incorporates local illumination information of the image through DCT, achieving local–global illumination extraction. However, in regions degraded by noise, the relative magnitudes of the RGB channel values are altered, leading to a fundamental shift in the dominant color of the image. This causes deviations in our illumination extraction method, which relies on the maximum RGB channel values, ultimately resulting in color distortion in the enhanced image.

5. Conclusions

This paper proposes DCDNet, an unsupervised LLIE method. Departing from the existing CNN-Retinex framework, it eliminates the need for enhancement networks through a decomposition network that refines illumination–reflectance decoupling. The preprocessing network DPNet delivers optimized inputs to decomposition network. DCT incorporates edge-structural priors during illumination estimation, while DGD-RM and RG-MSA jointly suppress illumination contamination on reflectance restoration. Extensive experiments on public benchmarks show that DCDNet outperforms the state-of-the-art methods significantly. In future work, considering that high-frequency DCT is beneficial for LLIE tasks in illumination networks, the illumination information obtained from high-frequency DCT and low-frequency DCT can be properly combined.

Author Contributions

Conceptualization, W.W.; methodology, W.W. and C.X.; software, W.W.; validation, Y.Z.; formal analysis, W.W.; investigation, C.X.; resources, C.X.; data curation, Y.Z.; writing—original draft preparation, W.W.; writing—review and editing, C.X.; visualization, M.Z.; supervision, C.X.; project administration, C.X.; funding acquisition, C.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Postgraduate Research and Practice Innovation Program of Jiangsu Province under Grant SJCX25_0430.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Loh, Y.P.; Chan, C.S. Getting to Know Low-Light Images with the Exclusively Dark Dataset. Comput. Vis. Image Underst. 2019, 178, 30–42. [Google Scholar] [CrossRef]
  2. Lin, F.; Rojas, J.D.; Dayton, P.A. Super Resolution Contrast Ultrasound Imaging: Analysis of Imaging Resolution and Application to Imaging Tumor Angiogenesis. In Proceedings of the 2016 IEEE International Ultrasonics Symposium (IUS), Tours, France, 18–21 September 2016; IEEE: New York, NY, USA, 2016; pp. 1–4. [Google Scholar]
  3. Wang, H.; Chen, Y.; Cai, Y.; Chen, L.; Li, Y.; Sotelo, M.A.; Li, Z. SFNet-N: An Improved SFNet Algorithm for Semantic Segmentation of Low-Light Autonomous Driving Road Scenes. IEEE Trans. Intell. Transp. Syst. 2022, 23, 21405–21417. [Google Scholar] [CrossRef]
  4. Lore, K.G.; Akintayo, A.; Sarkar, S. LLNet: A Deep Autoencoder Approach to Natural Low-Light Image Enhancement. arXiv 2015, arXiv:1511.03995. [Google Scholar] [CrossRef]
  5. Lv, F.; Lu, F.; Wu, J.; Lim, C. MBLLEN: Low-Light Image/Video Enhancement Using CNNs. In Proceedings of the BMVC, Newcastle, UK, 3–6 September 2018; Northumbria University: Tyne, UK, 2018; Volume 220, p. 4. [Google Scholar]
  6. Wei, C.; Wang, W.; Yang, W.; Liu, J. Deep Retinex Decomposition for Low-Light Enhancement. arXiv 2018, arXiv:1808.04560. [Google Scholar] [CrossRef]
  7. Zhang, Y.; Zhang, J.; Guo, X. Kindling the Darkness: A Practical Low-Light Image Enhancer. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 1632–1640. [Google Scholar]
  8. Hai, J.; Xuan, Z.; Yang, R.; Hao, Y.; Zou, F.; Lin, F.; Han, S. R2RNet: Low-Light Image Enhancement via Real-Low to Real-Normal Network. J. Vis. Commun. Image Represent. 2023, 90, 103712. [Google Scholar] [CrossRef]
  9. Wu, W.; Weng, J.; Zhang, P.; Wang, X.; Yang, W.; Jiang, J. URetinex-Net: Retinex-Based Deep Unfolding Network for Low-Light Image Enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5901–5910. [Google Scholar]
  10. Ren, X.; Li, M.; Cheng, W.H.; Liu, J. Joint Enhancement and Denoising Method via Sequential Decomposition. In Proceedings of the 2018 IEEE International Symposium on Circuits and Systems (ISCAS), Florence, Italy, 27–30 May 2018; IEEE: New York, NY, USA, 2018; pp. 1–5. [Google Scholar]
  11. Zhu, A.; Zhang, L.; Shen, Y.; Ma, Y.; Zhao, S.; Zhou, Y. Zero-Shot Restoration of Underexposed Images via Robust Retinex Decomposition. In Proceedings of the 2020 IEEE International Conference on Multimedia and Expo (ICME), London, UK, 6–10 July 2020; IEEE Computer Society: New York, NY, USA, 2020; pp. 1–6. [Google Scholar]
  12. Xu, X.; Wang, R.; Fu, C.W.; Jia, J. SNR-Aware Low-Light Image Enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 17714–17724. [Google Scholar]
  13. Cai, Y.; Bian, H.; Lin, J.; Wang, H.; Timofte, R.; Zhang, Y. Retinexformer: One-Stage Retinex-Based Transformer for Low-Light Image Enhancement. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 12504–12513. [Google Scholar]
  14. Fu, Z.; Yang, Y.; Tu, X.; Huang, Y.; Ding, X.; Ma, K.K. Learning a Simple Low-Light Image Enhancer from Paired Low-Light Instances. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 22252–22261. [Google Scholar]
  15. Pizer, S.M.; Amburn, E.P.; Austin, J.D.; Cromartie, R.; Geselowitz, A.; Greer, T.; ter Haar Romeny, B.; Zimmerman, J.B.; Zuiderveld, K. Adaptive Histogram Equalization and Its Variations. Comput. Vis. Graph. Image Process. 1987, 39, 355–368. [Google Scholar] [CrossRef]
  16. Land, E.H.; McCann, J.J. Lightness and Retinex Theory. J. Opt. Soc. Am. 1971, 61, 1–11. [Google Scholar] [CrossRef]
  17. Land, E.H. The Retinex Theory of Color Vision. Sci. Am. 1977, 237, 108–129. [Google Scholar] [CrossRef]
  18. Zheng, S.; Ma, Y.; Pan, J.; Lu, C.; Gupta, G. Low-Light Image and Video Enhancement: A Comprehensive Survey and Beyond. arXiv 2022, arXiv:2212.10772. [Google Scholar]
  19. Wu, W.; Weng, J.; Zhang, P.; Wang, X.; Yang, W.; Jiang, J. Interpretable Optimization-Inspired Unfolding Network for Low-Light Image Enhancement. IEEE Trans. Pattern Anal. Mach. Intell. 2025, 47, 2545–2562. [Google Scholar] [CrossRef]
  20. Jiang, Y.; Gong, X.; Liu, D.; Cheng, Y.; Fang, C.; Shen, X.; Yang, J.; Zhou, P.; Wang, Z. EnlightenGAN: Deep Light Enhancement without Paired Supervision. IEEE Trans. Image Process. 2021, 30, 2340–2349. [Google Scholar] [CrossRef]
  21. Zhang, L.; Zhang, L.; Liu, X.; Shen, Y.; Zhang, S.; Zhao, S. Zero-Shot Restoration of Back-Lit Images Using Deep Internal Learning. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 1623–1631. [Google Scholar]
  22. Guo, C.; Li, C.; Guo, J.; Loy, C.C.; Hou, J.; Kwong, S.; Cong, R. Zero-Reference Deep Curve Estimation for Low-Light Image Enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1780–1789. [Google Scholar]
  23. Brateanu, A.; Balmez, R.; Orhei, C.; Ancuti, C.; Ancuti, C. Enhancing Low-Light Images with Kolmogorov-Arnold Networks in Transformer Attention. Sensors 2025, 25, 327. [Google Scholar] [CrossRef] [PubMed]
  24. Brateanu, A.; Balmez, R.; Avram, A.; Orhei, C.; Ancuti, C. LYT-Net: Lightweight YUV Transformer-Based Network for Low-Light Image Enhancement. IEEE Signal Process. Lett. 2025, 32, 2065–2069. [Google Scholar] [CrossRef]
  25. In, H.; Kweon, J.; Moon, C. Squeeze-EnGAN: Memory Efficient and Unsupervised Low-Light Image Enhancement for Intelligent Vehicles. Sensors 2025, 25, 1825. [Google Scholar] [CrossRef] [PubMed]
  26. Li, H.; Hu, X.; Wang, H. Interpretable Unsupervised Joint Denoising and Enhancement for Real-World Low-Light Scenarios. In Proceedings of the Thirteenth International Conference on Learning Representations, Singapore, 24–28 April 2025. [Google Scholar]
  27. Zhou, P.; Yu, N.; Wu, Z.; Davis, L.S.; Shrivastava, A.; Lim, S.N. Deep Video Inpainting Detection. arXiv 2021, arXiv:2101.11080. [Google Scholar] [CrossRef]
  28. Liu, R.; Ma, L.; Zhang, J.; Fan, X.; Luo, Z. Retinex-Inspired Unrolling with Cooperative Prior Architecture Search for Low-Light Image Enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 10561–10570. [Google Scholar]
  29. Zeng, Q.; Hui, B.; Liu, Z.; Xu, Z.; He, M. A Method Combining Discrete Cosine Transform with Attention for Multi-Temporal Remote Sensing Image Matching. Sensors 2025, 25, 1345. [Google Scholar] [CrossRef]
  30. Hasan, N.; Islam, M.S.; Chen, W.; Kabir, M.A.; Al-Ahmadi, S. Encryption Based Image Watermarking Algorithm in 2DWT-DCT Domains. Sensors 2021, 21, 5540. [Google Scholar] [CrossRef]
  31. Cai, J.; Gu, S.; Zhang, L. Learning a Deep Single Image Contrast Enhancer from Multi-Exposure Images. IEEE Trans. Image Process. 2018, 27, 2049–2062. [Google Scholar] [CrossRef]
  32. Yang, W.; Wang, W.; Huang, H.; Wang, S.; Liu, J. Sparse Gradient Regularized Deep Retinex Network for Robust Low-Light Image Enhancement. IEEE Trans. Image Process. 2021, 30, 2072–2086. [Google Scholar] [CrossRef]
  33. Ma, K.; Zeng, K.; Wang, Z. Perceptual Quality Assessment for Multi-Exposure Image Fusion. IEEE Trans. Image Process. 2015, 24, 3345–3356. [Google Scholar] [CrossRef]
  34. Lee, C.; Lee, C.; Kim, C.S. Contrast Enhancement Based on Layered Difference Representation of 2D Histograms. IEEE Trans. Image Process. 2013, 22, 5372–5384. [Google Scholar] [CrossRef] [PubMed]
  35. Guo, X.; Li, Y.; Ling, H. LIME: Low-Light Image Enhancement via Illumination Map Estimation. IEEE Trans. Image Process. 2016, 26, 982–993. [Google Scholar] [CrossRef] [PubMed]
  36. Vonikakis, V.; Kouskouridas, R.; Gasteratos, A. On the Evaluation of Illumination Compensation Algorithms. Multimed. Tools Appl. 2018, 77, 9211–9231. [Google Scholar] [CrossRef]
  37. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image Quality Assessment: From Error Visibility to Structural Similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
  38. Zhang, R.; Isola, P.; Efros, A.A.; Shechtman, E.; Wang, O. The Unreasonable Effectiveness of Deep Features as a Perceptual Metric. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 586–595. [Google Scholar]
  39. Mittal, A.; Soundararajan, R.; Bovik, A.C. Making a “Completely Blind” Image Quality Analyzer. IEEE Signal Process. Lett. 2012, 20, 209–212. [Google Scholar] [CrossRef]
  40. Yan, Q.; Feng, Y.; Zhang, C.; Pang, G.; Shi, K.; Wu, P.; Dong, W.; Sun, J.; Zhang, Y. HVI: A New Color Space for Low-Light Image Enhancement. In Proceedings of the Computer Vision and Pattern Recognition Conference, Nashville, TN, USA, 11–15 June 2025; pp. 5678–5687. [Google Scholar]
  41. Feng, Y.; Hou, S.; Lin, H.; Zhu, Y.; Wu, P.; Dong, W.; Sun, J.; Yan, Q.; Zhang, Y. DiffLight: Integrating Content and Detail for Low-Light Image Enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 6143–6152. [Google Scholar]
  42. Zhang, T.; Liu, P.; Lu, Y.; Cai, M.; Zhang, Z.; Zhang, Z.; Zhou, Q. CWNet: Causal Wavelet Network for Low-Light Image Enhancement. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Honolulu, HI, USA, 19–25 October 2025; pp. 8789–8799. [Google Scholar]
  43. Ni, Z.; Yang, W.; Wang, S.; Ma, L.; Kwong, S. Towards Unsupervised Deep Image Enhancement with Generative Adversarial Network. IEEE Trans. Image Process. 2020, 29, 9140–9151. [Google Scholar] [CrossRef]
  44. Ma, L.; Ma, T.; Liu, R.; Fan, X.; Luo, Z. Toward Fast, Flexible, and Robust Low-Light Image Enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 5637–5646. [Google Scholar]
  45. Xie, C.; Fei, L.; Tao, H.; Hu, Y.; Zhou, W.; Hoe, J.T.; Hu, W.; Tan, Y.P. Residual Quotient Learning for Zero-Reference Low-Light Image Enhancement. IEEE Trans. Image Process. 2024, 34, 365–378. [Google Scholar] [CrossRef]
  46. Li, J.; Wang, Y.; Wang, C.; Tai, Y.; Qian, J.; Yang, J.; Wang, C.; Li, J.; Huang, F. DSFD: Dual Shot Face Detector. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5060–5069. [Google Scholar]
Figure 1. Comparison of existing learning frameworks with ours for LLIE. (a) End-to-end learning framework, accepted by LLNet and MBLLEN. (b) CNN-Retinex learning framework, accepted by Retinex-Net, KinD, URetinex-Net, and PairLIE. They require enhancement networks to optimize and refine the decomposed components after the decomposition network. (c) Our approach eliminates the need for additional enhancement networks by optimizing component decoupling within the decomposition network.
Figure 1. Comparison of existing learning frameworks with ours for LLIE. (a) End-to-end learning framework, accepted by LLNet and MBLLEN. (b) CNN-Retinex learning framework, accepted by Retinex-Net, KinD, URetinex-Net, and PairLIE. They require enhancement networks to optimize and refine the decomposed components after the decomposition network. (c) Our approach eliminates the need for additional enhancement networks by optimizing component decoupling within the decomposition network.
Sensors 25 07038 g001
Figure 2. The architecture of DCDNet. During training, DPnet preprocesses low-light inputs to optimize Retinex decomposition, while Rnet and Lnet represent the reflectance and illumination decomposition networks, respectively. Optimization uses four losses: L P (Preprocessing), L R (Retinex Decomposition), L S p a (Spatial Consistency), and L C (Reflectance Consistency). During testing, trained DPnet, Lnet, and Rnet decompose inputs into reflectance and illumination; illumination is adjusted and recombined with reflectance to output enhanced images.
Figure 2. The architecture of DCDNet. During training, DPnet preprocesses low-light inputs to optimize Retinex decomposition, while Rnet and Lnet represent the reflectance and illumination decomposition networks, respectively. Optimization uses four losses: L P (Preprocessing), L R (Retinex Decomposition), L S p a (Spatial Consistency), and L C (Reflectance Consistency). During testing, trained DPnet, Lnet, and Rnet decompose inputs into reflectance and illumination; illumination is adjusted and recombined with reflectance to output enhanced images.
Sensors 25 07038 g002
Figure 3. Overall Architecture of DPnet. DPnet aims to eliminate features incompatible with Retinex decomposition.
Figure 3. Overall Architecture of DPnet. DPnet aims to eliminate features incompatible with Retinex decomposition.
Sensors 25 07038 g003
Figure 4. (a) Shows the overall architecture of Rnet, whose key components are the DGD-RM (b) and the RG-MSA (c). Detailed structures of these two components are illustrated on the right. The RG-MSA utilizes the reflectance estimated by the DGD-RM to guide the self-attention computation. Beginning with 3 × 3 convolution, the structure culminates in the output F o u t .
Figure 4. (a) Shows the overall architecture of Rnet, whose key components are the DGD-RM (b) and the RG-MSA (c). Detailed structures of these two components are illustrated on the right. The RG-MSA utilizes the reflectance estimated by the DGD-RM to guide the self-attention computation. Beginning with 3 × 3 convolution, the structure culminates in the output F o u t .
Sensors 25 07038 g004
Figure 5. Overall architecture of Lnet. Lnet incorporates a global illumination prior through DCT.
Figure 5. Overall architecture of Lnet. Lnet incorporates a global illumination prior through DCT.
Sensors 25 07038 g005
Figure 6. Visual comparison of the compared methods on LLIE, including ZeroDCE, RUAS, SCI, PairLIE, ResQ, DENet, and ours. The final result corresponds to the ground truth image. The dataset comes from LOL v1. The areas within the red boxes are selected for magnification, and the detailed views are presented beneath.
Figure 6. Visual comparison of the compared methods on LLIE, including ZeroDCE, RUAS, SCI, PairLIE, ResQ, DENet, and ours. The final result corresponds to the ground truth image. The dataset comes from LOL v1. The areas within the red boxes are selected for magnification, and the detailed views are presented beneath.
Sensors 25 07038 g006
Figure 7. Visual comparisons of enhancement results by using different methods: ZeroDCE, RUAS, SCI, PairLIE, ResQ, DENet, and ours. The final result corresponds to the ground truth image. The dataset comes from LOL v2. The areas within the red boxes are selected for magnification, and the detailed views are presented beneath.
Figure 7. Visual comparisons of enhancement results by using different methods: ZeroDCE, RUAS, SCI, PairLIE, ResQ, DENet, and ours. The final result corresponds to the ground truth image. The dataset comes from LOL v2. The areas within the red boxes are selected for magnification, and the detailed views are presented beneath.
Sensors 25 07038 g007
Figure 8. Visual comparisons of enhancement results by using different methods: EnlightenGAN, URetinexNet, PairLIE, ResQ, and ours. The datasets from top to bottom are MEF, DICM, Fusion, and LIME, respectively.
Figure 8. Visual comparisons of enhancement results by using different methods: EnlightenGAN, URetinexNet, PairLIE, ResQ, and ours. The datasets from top to bottom are MEF, DICM, Fusion, and LIME, respectively.
Sensors 25 07038 g008
Figure 9. Visualization of the decomposed outputs. Enhanced results under different correction factors λ .
Figure 9. Visualization of the decomposed outputs. Enhanced results under different correction factors λ .
Sensors 25 07038 g009
Figure 10. The decomposition results using our method and PairLIE. From left to right: the enhanced image ( I e n ), illumination (L), and reflectance (R). The dataset comes from LOL v1, MIT and LOL v2 (LOL v2-real and LOL v2-syn).
Figure 10. The decomposition results using our method and PairLIE. From left to right: the enhanced image ( I e n ), illumination (L), and reflectance (R). The dataset comes from LOL v1, MIT and LOL v2 (LOL v2-real and LOL v2-syn).
Sensors 25 07038 g010
Figure 11. Dark face detection on a challenging example from the DARK FACE dataset, showing results from unlabeled, low, ZeroDCE, RUAS, RRDNet, SCI, URetinexNet, PairLIE, ResQ, and ours. The green boxes indicate the regions to be magnified. The enlarged views are presented below, with red boxes highlighting the detected faces.
Figure 11. Dark face detection on a challenging example from the DARK FACE dataset, showing results from unlabeled, low, ZeroDCE, RUAS, RRDNet, SCI, URetinexNet, PairLIE, ResQ, and ours. The green boxes indicate the regions to be magnified. The enlarged views are presented below, with red boxes highlighting the detected faces.
Sensors 25 07038 g011
Figure 12. Ablation studies results on LIME dataset. (a) w/o DGD-RM & RG-MSA; (b) w/o DCT; (c) w/o L S p a ; (d) DCDNet.
Figure 12. Ablation studies results on LIME dataset. (a) w/o DGD-RM & RG-MSA; (b) w/o DCT; (c) w/o L S p a ; (d) DCDNet.
Sensors 25 07038 g012
Figure 13. Visualization of reconstruction errors on LOL v1, LOL v2-real, LOL v2-syn, and MIT. Baseline decomposes original low-light images (blue); PairLIE uses its Pnet on projected images (orange); our method employs DPnet on projected images (green).
Figure 13. Visualization of reconstruction errors on LOL v1, LOL v2-real, LOL v2-syn, and MIT. Baseline decomposes original low-light images (blue); PairLIE uses its Pnet on projected images (orange); our method employs DPnet on projected images (green).
Sensors 25 07038 g013
Figure 14. Illuminationmaps from HDCT vs. LDCT. LDCT is better. From left to right: low, HDCT, and LDCT. The red boxes indicate the regions to be magnified; the zoomed-in views are presented on the right.
Figure 14. Illuminationmaps from HDCT vs. LDCT. LDCT is better. From left to right: low, HDCT, and LDCT. The red boxes indicate the regions to be magnified; the zoomed-in views are presented on the right.
Sensors 25 07038 g014
Figure 15. Illuminationmaps from HDCT vs. LDCT. HDCT is better. From left to right: low, HDCT, and LDCT. The red boxes indicate the regions to be magnified; the zoomed-in views are presented on the right.
Figure 15. Illuminationmaps from HDCT vs. LDCT. HDCT is better. From left to right: low, HDCT, and LDCT. The red boxes indicate the regions to be magnified; the zoomed-in views are presented on the right.
Sensors 25 07038 g015
Figure 16. Comparison chart of failure case results. Among them, from left to right, are the low-light image, the enhanced image, and the label image in that order. The datasets come from LOL v1 and LOL v2.
Figure 16. Comparison chart of failure case results. Among them, from left to right, are the low-light image, the enhanced image, and the label image in that order. The datasets come from LOL v1 and LOL v2.
Sensors 25 07038 g016
Table 1. Network architecture specifications. The 32 indicates the use of the DGD-RM and RG-MSA modules here. The 6 indicates the use of the DCT module here.
Table 1. Network architecture specifications. The 32 indicates the use of the DGD-RM and RG-MSA modules here. The 6 indicates the use of the DCT module here.
NetworkDimensional ChangesLayer DepthConvolution Kernel
DPnet3 (input) — 16 — 16 — 16 — 3 (output)4 3 × 3
Rnet3 (input) — 32 — 3 — 32 — 3 (output)4 3 × 3
Lnet3 (input) — 6 — 32 — 32 — 32 — 1 (output)5 3 × 3 & 1 × 1
Table 2. Conduct a quantitative evaluation of comparative methods on three reference benchmark datasets. The best and second-best results are marked in red and blue, respectively. ↑ (or ↓) means that the larger (or smaller), the better. “Supervised” denotes supervised methods and “Unsupervised” denotes unsupervised methods.
Table 2. Conduct a quantitative evaluation of comparative methods on three reference benchmark datasets. The best and second-best results are marked in red and blue, respectively. ↑ (or ↓) means that the larger (or smaller), the better. “Supervised” denotes supervised methods and “Unsupervised” denotes unsupervised methods.
MethodLOL v1LOL v2-realLOL v2-syntheticMIT
PSNR ↑SSIM ↑LPIPS ↓PSNR ↑SSIM ↑LPIPS ↓PSNR ↑SSIM ↑LPIPS ↓PSNR ↑SSIM ↑LPIPS ↓
Supervised
Retinex-Net [6]14.980.3270.62416.100.4010.54314.040.2730.81814.730.7380.382
MBLLEN [5]17.860.6270.22517.870.6940.27215.780.5010.59918.010.7510.283
KinD [7]16.150.6200.42119.000.7370.27916.310.5680.55217.840.7680.250
URetinex-Net [9]17.280.6970.31121.090.7570.10017.700.6290.44818.560.8220.188
SNR-Aware [12]24.610.8420.23321.480.8490.23724.140.9280.189---
Retinexformer [13]23.930.8310.12522.450.8440.16525.670.9300.059---
CIDNet [40]23.810.8700.08624.110.8670.11625.120.9390.045---
DiffLight [41]25.850.8760.082--------
CWNet [42]23.600.8490.06527.390.9000.03825.500.9360.020---
Unsupervised
EnlightGAN [20]17.560.6650.31618.680.6730.30115.180.4820.62015.560.8000.213
RRDNet [11]10.160.4060.48313.700.5080.29010.500.3100.72919.600.8300.229
UEGAN [43]8.240.2220.68912.810.2680.4108.900.1930.70019.540.8700.199
ZeroDCE [22]15.330.5670.33518.460.5800.30512.990.3650.68417.960.8430.226
RUAS [28]16.150.4620.36814.890.4550.37212.190.2970.7488.460.5400.585
SCI [44]14.780.5220.34017.300.5340.30812.360.3230.70119.520.8850.176
PairLIE [14]19.510.7360.24818.230.7350.26016.110.5660.56014.470.7590.198
ResQ [45]18.280.5960.32319.670.5730.34715.650.3940.68720.870.9150.107
DENet [26]19.800.7500.25320.220.7930.26618.240.7960.26820.300.8760.193
Ours20.870.7700.19720.350.8100.19119.700.8100.20021.660.8640.164
Table 3. Conduct quantitative evaluation of comparative methods using NIQE on five no-reference benchmark datasets. The best and second-best results are marked in red and blue, respectively.
Table 3. Conduct quantitative evaluation of comparative methods using NIQE on five no-reference benchmark datasets. The best and second-best results are marked in red and blue, respectively.
MethodNo-Reference Benchmark DatasetAverage
MEFDICMLIMEFusionVV
EnlightenGAN [20]3.4203.5684.0613.6542.8233.505
RRDNet [11]3.7816.7276.1255.7812.9795.078
UEGAN [43]5.1324.0464.5404.2283.6964.328
RUAS [28]5.1095.7274.6976.0805.3465.392
URetinexNet [9]3.7893.4594.3413.8183.0193.685
SNR-Aware [12]4.1804.7125.7424.3239.8725.765
PairLIE [14]4.1643.5194.5155.0023.6544.171
Retinexformer [13]4.3223.8534.3103.7623.0943.869
ResQ [45]3.4773.3884.0343.6322.7323.453
CWNet [42]3.5683.7954.1384.0253.2143.747
Ours3.4623.0274.0003.6622.9513.420
Table 4. Quantitative evaluation of deep-learning-based methods on model complexity in terms of Flops, Params, and inference time (Gpu seconds). “Supervised” denotes supervised methods and “Unsupervised” denotes unsupervised methods.
Table 4. Quantitative evaluation of deep-learning-based methods on model complexity in terms of Flops, Params, and inference time (Gpu seconds). “Supervised” denotes supervised methods and “Unsupervised” denotes unsupervised methods.
MethodFLOPs (G)Params (M)Time (S)
Supervised
Retinex-Net [6]136.020.83830.0864
MBLLEN [5]60.020.45020.2111
KinD [7]29.138.54020.1529
URetinex-Net [9]58.270.36210.0176
SNR-Aware [12]27.884.01320.0266
Retinexformer [13]17.021.60570.0234
CIDNet [40]7.571.88-
DifLight [41]168.3--
CWNet [42]11.31.23-
Unsupervised
EnlightenGAN [20]61.018.63600.0701
RRDNet [11]30.660.12821.5677
UEGAN [43]32.7216.61490.0435
ZeroDCE [22]5.210.07890.0204
RUAS [28]0.210.00340.0165
SCI [44]0.080.00040.0160
PairLIE [14]22.350.34180.1985
ResQ [45]0.490.01230.0178
DENet [26]20.400.3120.0517
Ours7.000.1070.0085
Table 5. Ablation studies results on LOL v1, LOL v2-real, LOL v2-syn, and MIT datasets. The best results is marked in red, respectively. ↑ (or ↓) means that the larger (or smaller), the better.
Table 5. Ablation studies results on LOL v1, LOL v2-real, LOL v2-syn, and MIT datasets. The best results is marked in red, respectively. ↑ (or ↓) means that the larger (or smaller), the better.
SettingLOL v1LOL v2-realLOL v2-synMIT
PSNR ↑SSIM ↑LPIPS ↓PSNR ↑SSIM ↑LPIPS ↓PSNR ↑SSIM ↑LPIPS ↓PSNR ↑SSIM ↑LPIPS ↓
(a) w/o DGD-RM & RG-MSA19.720.7070.24319.900.7690.23819.030.7320.23020.810.8360.192
(b) w/o DCT20.360.7360.23219.540.7550.22119.300.7520.23521.200.8470.179
(c) w/o L S p a 20.450.7470.22120.130.7980.20119.470.7950.20221.450.8500.178
DCDNet20.870.7700.19720.350.8100.19119.700.8100.20021.660.8640.164
Table 6. Conduct quantitative evaluation of comparative baseline, Pnet and DPnet (ours), on four reference benchmark datasets. The better results are marked in red, respectively. ↓ means that the smaller the better.
Table 6. Conduct quantitative evaluation of comparative baseline, Pnet and DPnet (ours), on four reference benchmark datasets. The better results are marked in red, respectively. ↓ means that the smaller the better.
MethodLOL v1LOL v2-realLOL v2-syntheticMIT
Reconstruction Error ↓Reconstruction Error ↓Reconstruction Error ↓Reconstruction Error ↓
Baseline0.01010.01040.02210.0217
Pnet0.00920.00960.02160.0211
DPnet (Ours)0.00700.00710.01680.0149
Table 7. Conduct quantitative evaluation of comparative HDCT and LDCT on three reference benchmark datasets. The better results are marked in red, respectively. ↑ (or ↓) means that the larger (or smaller), the better.
Table 7. Conduct quantitative evaluation of comparative HDCT and LDCT on three reference benchmark datasets. The better results are marked in red, respectively. ↑ (or ↓) means that the larger (or smaller), the better.
MethodLOL v1LOL v2-RealLOL v2-SyntheticMIT
PSNR ↑SSIM ↑LPIPS ↓PSNR ↑SSIM ↑LPIPS ↓PSNR ↑SSIM ↑LPIPS ↓PSNR ↑SSIM ↑LPIPS ↓
HDCT20.650.7500.21720.160.7850.21019.230.8040.20221.330.8590.160
LDCT (our)20.870.7700.19720.350.8100.19119.700.8100.20021.660.8640.164
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, W.; Zhu, Y.; Zhang, M.; Xie, C. DCD-Net: Decoupling-Centric Decomposition Network for Low-Light Image Enhancement. Sensors 2025, 25, 7038. https://doi.org/10.3390/s25227038

AMA Style

Wang W, Zhu Y, Zhang M, Xie C. DCD-Net: Decoupling-Centric Decomposition Network for Low-Light Image Enhancement. Sensors. 2025; 25(22):7038. https://doi.org/10.3390/s25227038

Chicago/Turabian Style

Wang, Wei, Yi Zhu, Mingming Zhang, and Chao Xie. 2025. "DCD-Net: Decoupling-Centric Decomposition Network for Low-Light Image Enhancement" Sensors 25, no. 22: 7038. https://doi.org/10.3390/s25227038

APA Style

Wang, W., Zhu, Y., Zhang, M., & Xie, C. (2025). DCD-Net: Decoupling-Centric Decomposition Network for Low-Light Image Enhancement. Sensors, 25(22), 7038. https://doi.org/10.3390/s25227038

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop