Next Article in Journal
Deep Learning Techniques for Prostate Cancer Analysis and Detection: Survey of the State of the Art
Previous Article in Journal
Synthetic Scientific Image Generation with VAE, GAN, and Diffusion Model Architectures
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DFCNet: Dual-Stage Frequency-Domain Calibration Network for Low-Light Image Enhancement

School of Automation, Nanjing University of Science and Technology, Nanjing 210094, China
*
Author to whom correspondence should be addressed.
J. Imaging 2025, 11(8), 253; https://doi.org/10.3390/jimaging11080253
Submission received: 3 July 2025 / Revised: 19 July 2025 / Accepted: 25 July 2025 / Published: 28 July 2025
(This article belongs to the Section Image and Video Processing)

Abstract

Imaging technologies are widely used in surveillance, medical diagnostics, and other critical applications. However, under low-light conditions, captured images often suffer from insufficient brightness, blurred details, and excessive noise, degrading quality and hindering downstream tasks. Conventional low-light image enhancement (LLIE) methods not only require annotated data but also often involve heavy models with high computational costs, making them unsuitable for real-time processing. To tackle these challenges, a lightweight and unsupervised LLIE method utilizing a dual-stage frequency-domain calibration network (DFCNet) is proposed. In the first stage, the input image undergoes the preliminary feature modulation (PFM) module to guide the illumination estimation (IE) module in generating a more accurate illumination map. The final enhanced image is obtained by dividing the input by the estimated illumination map. The second stage is used only during training. It applies a frequency-domain residual calibration (FRC) module to the first-stage output, generating a calibration term that is added to the original input to darken dark regions and brighten bright areas. This updated input is then fed back to the PFM and IE modules for parameter optimization. Extensive experiments on benchmark datasets demonstrate that DFCNet achieves superior performance across multiple image quality metrics while delivering visually clearer and more natural results.

1. Introduction

Over the past decade, the widespread adoption of imaging technologies in domains including surveillance systems, medical diagnostics, and autonomous driving has established image quality as a critical determinant of overall system performance [1]. In extremely low-light environments, however, image acquisition often results in insufficient brightness, reduced contrast, blurred textures, and significant noise. These degradations compromise both visual appearance and the reliability of downstream vision tasks. For example, in surveillance systems, blurred low-light images may cause target detection and recognition failures [2]. In medical imaging, poor image quality impedes lesion localization and reduces diagnostic accuracy [3]. In autonomous driving scenarios, failure to perceive clear road conditions directly compromises safety [4].
Conventional low-light image enhancement (LLIE) methods, like histogram equalization (HE) and gamma correction, perform basic brightness adjustments but rely exclusively on global gray-level mapping. Such methods lack adaptability to complex illumination distributions and often require manual parameter tuning, rendering them ineffective in diverse, dynamic real-world environments. Over the past few years, approaches grounded in deep learning have shown remarkable progress within LLIE by jointly modeling local details and global illumination features. However, the majority of current approaches are built upon supervised learning with paired low- and normal-light images [5], which involves substantial manual labor for data collection and annotation. Furthermore, many models are computationally intensive and parameter-heavy [6], posing challenges for deployment on resource-constrained devices and limiting their practicality in real-time systems.
To tackle these limitations, a lightweight and unsupervised LLIE framework named dual-stage frequency-domain calibration network (DFCNet) is proposed in this work. The framework employs a dual-stage enhancement architecture: during training, a collaborative dual-stage optimization enhances performance, while at inference, only a single-stage forward pass is needed, effectively balancing accuracy and efficiency. A preliminary feature modulation (PFM) module and an illumination estimation (IE) module are further designed to jointly generate dynamic guidance maps, enabling end-to-end optimization from feature enhancement to illumination estimation. Finally, the frequency-domain residual calibration (FRC) module is applied to compute a calibration term from the stage I output. This calibration term is then fed back into the PFM and IE modules. By using this calibration term to update their parameters, the network acquires the capability to produce superior enhancements in a single inference pass—achieving more natural results and suppressing artifacts without any additional runtime cost.
Extensive experimental results demonstrate that DFCNet outperforms leading supervised and unsupervised baseline methods in enhancement quality, detail preservation, and computational efficiency, thereby showcasing strong practical value and deployment potential. To summarize, the main contributions of this work are outlined as follows:
  • An unsupervised dual-stage framework is proposed to eliminate reliance on paired training data, thus striking a balance between enhancement quality and real-time efficiency.
  • An FRC module is designed to enhance structural consistency while suppressing noise and artifacts by exploiting residual information in the frequency domain.
  • A lightweight PFM module and an IE module are designed to collaboratively facilitate accurate illumination estimation via dynamic feature guidance mechanisms.
The subsequent parts of this work are structured as follows: Section 2 surveys related research in the LLIE domain. Section 3 elaborates on the proposed DFCNet architecture, its submodules, and the loss functions used. Section 4 discusses experimental findings and ablation analyses. Section 5 draws the conclusion.

2. Related Work

In recent years, LLIE has attracted significant attention due to its critical role in improving the visibility and quality of images captured under challenging illumination conditions. The main objective of LLIE is to restore clear and natural images from degraded low-light inputs, thereby facilitating subsequent vision tasks such as detection and recognition. Existing LLIE methods can generally be divided into three categories: conventional enhancement methods, supervised learning-based methods, and unsupervised learning-based methods.

2.1. Conventional Methods

Conventional LLIE methods typically rely on point-wise gray-level mappings (e.g., HE) and Retinex-based decomposition approaches [7]. Although most gray-level transformations can adjust brightness and contrast, they often need to be combined with other techniques under poor lighting conditions. HE-based methods enhance contrast by dividing the histogram into subbands and equalizing each separately, but they often fail to maintain global illumination consistency. Retinex-based methods improve low-light images by decomposing them into illumination and reflectance components, effectively preserving details [8]. Jobson et al. [9] employed Single-Scale Retinex (SSR) to apply the Retinex theory for intensity consistency and lightness reproduction in gray-level images. A recursive bilateral filter was proposed by Li et al. [10] to significantly expedite the decomposition procedure for enhancing efficiency in LLIE, while Guo et al. [11] proposed the LIME method, which estimates pixel-wise illumination and refines it to reduce computational overhead.

2.2. Supervised Deep Learning-Driven Methods

Supervised methods have recently advanced LLIE. These approaches learn low-light to normal-light mappings using large, paired datasets. Lore et al. [12] pioneered LLNet with autoencoders for joint contrast improvement and noise reduction. Lv et al. [13] proposed MBLLEN, which utilizes multi-branch fusion to eliminate noise. Chen et al. [14] proposed SID, exploiting RAW image linearity for extreme LLIE. Lim et al. [15] adopted Laplacian pyramids in DSLR for multi-scale recovery. Building on Retinex theory, RetinexNet by Wei et al. [5] decomposes images into illumination and reflectance for targeted adjustment. Zhang et al. [16] refined this idea in KinD with dedicated decomposition submodules. URetinex was introduced by Wu et al. [17], which integrates denoising and enhancement through adaptive unfolding. Cheng et al. [18] fused light-guided features in LCUN to improve illumination accuracy, while Wang et al. [19] leveraged color priors in LCDPNet for exposure correction. Cai et al. [20] formulated Retinexformer to capture non-local interactions, and Bai et al. [21] accelerated RetinexMamba inference by using state-space modeling. Yao et al. [22] proposed DFFN, a supervised method that fuses spatial and frequency-domain features through dual-domain blocks, adopts a dual-stage design for amplitude and phase optimization, and uses cross-stage fusion to enhance low-light remote sensing images.
Despite their success, supervised methods face limitations. They rely heavily on paired low-/normal-light data, which are rarely available in real-world scenarios. Moreover, synthetic data often poorly reflect real-world conditions, limiting scalability.

2.3. Unsupervised Deep Learning-Driven Methods

To overcome the limitations of supervised methods, unsupervised LLIE techniques have emerged as promising alternatives. These methods utilize unpaired datasets, eliminating dependency on ground truth references. Wang et al. [23] proposed a framework called GLADNet, which estimates global illumination via an encoder–decoder architecture to recover detail losses from initial rescaling. EnlightenGAN, introduced by Jiang et al. [6], adopts global and local discriminators with self-supervised regularization, content preservation losses, and attention mechanisms. Zero-DCE, proposed by Guo et al. [24], predicts image-adaptive curves using lightweight networks trained with no-reference losses. RRDNet, introduced by Zhu et al. [25], leverages triple-branch CNNs for joint denoising and restoration. Yang et al. [26] designed DRBN, applying perception-driven linear transformations guided by learned quality assessment networks. RetinexDIP, proposed by Zhao et al. [27], recasts Retinex decomposition as a generative modeling task. Wang et al. [28] created MAGAN, incorporating multi-scale attention for simultaneous enhancement/denoising. Developed by Ma et al. [29], SCI implements a weight-shared and self-calibrated module to learn illumination. Kandula et al. [30] introduced region-adaptive single-input multiple-output generation for user-customized enhancement. CLIP-LIT, presented by Liang et al. [31], leverages CLIP’s semantic priors via prompt learning for illumination adjustment. Jiang et al. [32] proposed UDCN, which decomposes images into illumination and reflectance components, corrects illumination via an adaptive pyramid-based strategy, and removes noise through a dedicated noise removal module. Fei et al. [33] proposed GDP, a generative diffusion prior framework that leverages a pre-trained DDPM model for unsupervised posterior modeling. Luo et al. [34] proposed CRF-inspired plug-and-play frameworks with lightweight curves for self-supervised adaptation. Peng et al. [35] developed DeULLE leveraging luminance masking and illumination-reflectance decoupling under bidirectional cycle supervision.
Although unsupervised methods alleviate the need for paired datasets, many existing architectures remain computationally expensive, making them unsuitable for deployment on resource-constrained devices. Consequently, there is a strong practical imperative to develop efficient, lightweight, unsupervised low-light enhancement approaches that can run in real-time with minimal memory and power overhead. Such a solution would enable broader adoption in applications ranging from on-device surveillance and smartphone photography to battery-powered autonomous systems.

3. Proposed Method

3.1. Overview of the DFCNet

The proposed DFCNet in this work employs a dual-stage architecture, depicted in Figure 1. It comprises three core elements: PFM, IE, and FRC, which work together to enhance low-light images efficiently.
During stage I, the input image undergoes PFM processing, where multi-scale convolutional layers extract and fuse features to generate modulated feature maps. These modulated features are then passed to IE, which follows a U-Net-style [36] architecture and incorporates both spatial and channel attention mechanisms to produce an estimated illumination map. The enhanced image is generated via pixel-wise division of the input image by the illumination map, where each pixel in the input is divided by its corresponding pixel in the illumination map.
During stage II, the FRC is designed to further optimize PFM and IE. Specifically, the enhanced image from stage I is passed through FRC to generate a calibration term. This term is added to the original input image, forming the new input for stage II. The pipeline of stage I is then repeated using this updated input. During training, both stages are constrained by separate loss terms to guide the optimization process and improve overall enhancement quality. It is worth noting that during inference, only the first-stage procedure is executed. That is, the input image undergoes processing via PFM and IE to obtain the illumination map, and the final enhanced result is derived via pixel-wise division of the input image by this map. This design ensures high enhancement quality while maintaining inference efficiency.

3.2. Preliminary Feature Modulation

Deep feature extraction and adaptive modulation enhance discriminative representations for subsequent illumination estimation and enhancement [37]. To enhance adaptability to diverse low-light scenarios, we designed the PFM with two parallel branches. The main branch conducts progressive feature extraction through multi-layer convolutional layers with ReLU activations to capture fine-grained details and structural information. The downsampling branch enlarges the receptive field using max-pooling, followed by convolutional operations and bilinear upsampling to restore resolution. Features from both branches are fused in a channel-oriented manner for multi-scale fusion, then processed via convolutional layers with ReLU to generate a guidance map. The input image is adaptively modulated through pixel-wise division by this guidance map, resulting in an optimized feature map. This dynamic modulation mechanism adapts subsequent enhancement processes to input-specific feature distributions, significantly enhancing enhancement effectiveness in low-light conditions.

3.3. Illumination Estimation

Retinex theory provides an interpretable framework for LLIE and has demonstrated strong generalization capabilities in deep learning applications [38]. It models an input image as the element-wise multiplication of reflectance and illumination: S = R L , where R captures the object’s intrinsic properties, and L represents the illumination affected by external lighting conditions.
To accurately estimate illumination, which is critical for effective reflectance recovery and detail enhancement [39], we designed IE based on the U-Net architecture. The encoder acquires hierarchical features via stacked convolution-ReLU blocks, followed by max pooling to downsample the spatial resolution. A bottleneck layer with channel attention [40] adaptively emphasizes informative channels, capturing global illumination context. The decoder restores spatial resolution through upsampling and integrates shallow details using skip connections with spatial attention [41], enhancing structural awareness. Finally, a Sigmoid-activated convolution layer outputs the estimated illumination map I 3 × H × W .
Given the uneven lighting in low-light scenes, a residual enhancement strategy is adopted at the output stage. The input features and the predicted illumination map are fused via weighted addition, correcting estimation errors and reinforcing structural detail for more natural enhancement.

3.4. Frequency-Domain Residual Calibration

Spatial-domain illumination correction methods like the ICN in UDCN [32] have shown effectiveness in adjusting local brightness via adaptive pyramidal processing. Similarly, DFCNet also employs a calibration mechanism, but shifts the focus to the frequency domain. Recent studies have revealed a strong correlation between illumination distributions in low-light conditions and the magnitude characteristics in the frequency domain [42]. As shown in Figure 2, which compares the magnitudes and phases of low-light, normal-light, and enhanced images, explicitly modeling these magnitude features facilitates significant improvements in image brightness and contrast. To improve the modeling of global structures and periodic noise in LLIE networks, FRC is introduced in stage II. In contrast to DFFN [22], which fuses spatial and frequency branches across stages for joint image reconstruction, DFCNet employs a decoupling strategy: its calibration terms yielded from the FRC module serve solely to refine the parameters of the modules in the spatial branch, not to participate directly in image reconstruction. Specifically, FRC mines global magnitude characteristics in the frequency domain to produce a calibration term from the first-stage output. This term is added to the original low-light input, suppressing under-exposed regions and amplifying well-lit areas to yield an updated input. The refined image is then re-fed into the PFM and IE modules, guiding them to learn more robust illumination-aware features while incurring no additional computational overhead at inference. Figure 3 illustrates the magnitude adjustment block (MAB), the central component of the FRC, designed to exploit frequency-domain information for LLIE.
The module takes the reflectance map R 3 × H × W from stage I as input. After shallow encoding, the feature map F i n is transformed into the frequency domain via the fast Fourier transform (FFT):
F f r e q   = F ( F i n   ) = M e j P
where M represents magnitude and P represents phase. The magnitude component M is then calibrated by a nonlinear unit consisting of channel-expansion convolution, LeakyReLU activation, and channel-reduction convolution.
M = W 2   LeakyReLU ( W 1   M )
This enables adaptive modeling and rescaling of frequency component magnitudes to enhance global structure and noise representation. M is then recombined with P to form the calibrated frequency feature.
F f r e q   = M cos ( P ) + j M sin ( P ) = M e j P
The inverse fast Fourier transform (IFFT) maps it back to the spatial domain as F c a l . For stable information flow and efficient training, F c a l is fused with F i n via a gated residual connection:
F o u t = F i n + F c a l F i n
where denotes element-wise multiplication. After iterative MAB processing, final features are processed by a convolutional layer followed by a sigmoid activation function to output the calibration term C 3 × H × W , which is afterwards added to the original input image. This resultant sum acts as the input for stage II, enabling further refinement of the global structure and details of the illumination map.
Operating exclusively in stage II, FRC yields calibration terms to aid PFM and IE in learning refined parameters, enabling higher-quality enhancement. Leveraging the strong correlation between frequency-domain magnitude features and low-light illumination distribution, FRC utilizes MAB for efficient frequency-domain modeling and global calibration. This spatial-frequency collaborative modeling strikes a balance between local details and global structures. For clarity, the pipeline of FRC is detailed in Algorithm 1.
Algorithm 1 Frequency-Domain Residual Calibration (FRC)
InputReflectance map R 3 × H × W
OutputCalibration term C 3 × H × W
1: Spatial Encoding F i n   = ReLU   ( Conv   ( R ) )
2: for  i = 0   to   k  do
3:  FFT F f r e q   = F ( F i n   )
4:  Magnitude & Phase Separation M ,   P = F f r e q   ,   angle   ( F f r e q )
5:  Magnitude Calibration M = Conv   ( LeakyReLU   ( Conv   ( M ) ) )
6:  Frequency Recombination F f r e q   = M cos ( P ) + j M sin ( P )
7:  IFFT F c a l   = F 1 ( F f r e q   )
8:  Residual Fusion with Gating F o u t = F i n + F c a l F i n
9:  Update Input F i n     F o u t
10: end for
11: Output Calibration Term C = Sigmoid   ( Conv   ( F o u t ) )

3.5. Loss Function

To comprehensively regulate the network, we propose a multi-component loss function to address pixel fidelity, structural smoothness, and noise suppression. Training proceeds in two stages, with losses computed independently at each stage and then summed to form the final optimization objective:
L   total   t = t = 1 2 α L   illu   t + β L   smooth   t + γ L   tv   t
where α , β , and γ balance the illumination consistency loss, structural smoothness loss, and total variation loss, respectively, and t indicates the stage. This combined loss supervision ensures that the network not only boosts brightness and contrast but also preserves structural consistency—thereby balancing subjective visual quality with objective evaluation metrics. The total loss is composed of the following three terms.

3.5.1. Illumination Consistency Loss

To ensure that the illumination map at each stage faithfully reflects the input’s overall brightness and structure, we employ a pixel-level loss between the stage’s input and its predicted illumination map:
L   illu   t = i = 1 N X i t I i t 2
where X i t  denotes the value at pixel  i  of input to stage  t I i t is the corresponding predicted illumination map, and N represents the total pixel count. Such a loss effectively averts over-enhancement.

3.5.2. Structural Smoothness Loss

Smooth prediction of illumination is important for maintaining the consistency of the spatial structure of illumination maps [43]. To maintain spatial consistency in the illumination map and suppress noise, we adopt a weighted smoothness loss inspired by SCI [29], which helps suppress local noise while preserving edge structures:
L   smooth   t = i = 1 N j N i w i , j I i t I j t
w i , j = exp Y i t Y j t 2 2 σ 2
where N ( i ) denotes the 5 × 5 neighborhood of pixel i , w i , j represents the weight computed from the Y channel (after converting the input image to the standard YUV color space), and σ represents the Gaussian kernel’s standard deviation.

3.5.3. Total Variation Loss

To boost perceptual quality by reducing artifacts and noise, a total variation loss is introduced:
L   tv   t = i = 1 N c { r , g , b } h I i , c t 2 + v I i , c t 2
where I i , c t represents the value at pixel i and channel c (RGB) of the predicted illumination map; h denotes the horizontal operators; and v denotes vertical gradient operators.

4. Experiments

4.1. Experiment Settings

The proposed DFCNet was trained on the LOL-v2-real dataset [44], an expanded version of the original LOL dataset [5]. The extended dataset includes 689 low-light/normal-light image pairs for training and 100 pairs for testing, up from the initial 500 pairs. As an unsupervised learning approach is adopted, only the low-light images are utilized during training, without leveraging their paired normal-light counterparts. Experiments were executed on a computer outfitted with a NVIDIA GeForce RTX 4060 Ti GPU (16G) using PyTorch 2.1.0. The training configurations are as follows: a batch size of 8, an initial learning rate of 0.0002, the Adam optimizer (with β 1 = 0.9, β 2 = 0.999, ϵ = 1 × 10−8), and a total of 300 training epochs. DFCNet is highly lightweight, with only 0.0529M parameters, even including stage II’s FRC.

4.2. Comparative Methods and Evaluation Metrics

For a thorough evaluation of DFCNet’s performance, we compared it against several mainstream LLIE methods, including RetinexNet [5], DSLR [15], RRDNet [25], Zero-DCE [24], EnlightenGAN [6], RetinexDIP [27], LCDPNet [19], SCI [29], CLIP-LIT [31], and GDP [33]. All these methods provide publicly available code with recommended parameters and their corresponding training datasets. They encompass both supervised and unsupervised algorithms, enabling a multifaceted assessment of the strengths and weaknesses of our method.
Evaluation was conducted using both no-reference and full-reference strategies. For perceptual quality, we adopted the no-reference metrics naturalness image quality evaluator (NIQE) [45] and blind/referenceless image spatial quality evaluator (BRISQUE) [46], applied to all test images. For quantitative analysis on paired datasets, we additionally report the full-reference metrics peak-signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM). Recognizing the incompleteness of these metrics in capturing human visual perception, we also include subjective visual comparisons across various scenarios.
The evaluation datasets include LOL [5], and 50 images are randomly sampled from each of MIT [47], ExDark [48], and DarkFace [49] for evaluation. Among them, LOL and MIT are paired datasets that provide low-light images along with their corresponding normal-light references, making them suitable for quantitative evaluation using full-reference metrics. In contrast, ExDark and DarkFace contain only real-world low-light images without reference counterparts, thus better representing practical deployment scenarios and serving as the basis for no-reference evaluation.

4.3. Results

The quantitative results comparing our DFCNet to ten other methods on the LOL and MIT datasets (both featuring paired low-/normal-light images) are presented in Table 1. Red and blue are used to mark the best and second-best performers, respectively. On the LOL dataset, DFCNet achieved the second-best performance for both PSNR and NIQE. On the MIT dataset, DFCNet attained the best scores in both PSNR (0.32 higher than the second best) and NIQE (0.14 lower than the second best), while also securing the second-best result for BRISQUE. These results demonstrate DFCNet’s consistent effectiveness across datasets in both reference-based and perceptual quality metrics.
As illustrated in Figure 4, both EnlightenGAN and DFCNet effectively enhanced low-light images. However, DFCNet delivered more significant brightness improvement while better preserving fine details. In contrast, RetinexNet’s results appeared blurry, and DSLR not only suffered from insufficient brightness enhancement but also introduced uneven color artifacts. While other methods provided some enhancement, they generally lacked sufficient brightness improvement, resulting in suboptimal visual results. Figure 5 shows that EnlightenGAN, GDP, and DFCNet all preserved texture information in both overall and zoomed-in views. Nevertheless, EnlightenGAN and DSLR exhibited localized color anomalies, whereas GDP suffered from local overexposure. By comparison, RetinexNet and Zero-DCE produced excessively bright images with low contrast, noticeable distortion, and substantial detail loss. The remaining methods not only failed to effectively enhance image brightness but also struggled with edge preservation between distinct regions, further compromising visual clarity.
The quantitative results comparing our DFCNet to ten other methods on the ExDark and DarkFace datasets are presented in Table 2. As shown, DFCNet achieves the best NIQE score on ExDark (0.05 lower than the second-best method) and attains the second-best BRISQUE score, indicating its superior ability to enhance perceptual quality under real-world low-light conditions. On the DarkFace dataset, DFCNet demonstrates an even more pronounced advantage, achieving the best results for both NIQE and BRISQUE metrics, with scores lower than the second-best method by 0.10 and 1.21, respectively. This consistently strong performance across two challenging datasets, both of which contain diverse and complex low-light scenarios, highlights DFCNet’s robustness and generalization ability in objective image quality evaluation.
Figure 6 visually illustrates the effectiveness of DFCNet. It shows that DFCNet not only restores global brightness but also maintains local contrast by preserving edges between adjacent regions—an essential characteristic for achieving visually pleasing and informative enhancement results in low-light scenarios. In contrast, although RetinexNet manages to improve brightness to some extent, it introduces substantial noise, thereby degrading visual quality and potentially interfering with subsequent image analysis tasks. EnlightenGAN, meanwhile, produces results with a yellowish color cast and degraded fine details, compromising both color fidelity and image realism. Other methods continue to struggle with inadequate brightness enhancement, with RRDNet exhibiting the most deficiencies. Figure 7 further confirms the advantages of our method. DFCNet consistently delivers well-illuminated and visually clear results. By comparison, RetinexNet generates outputs with uneven color distribution, leading to unnatural appearance. In the zoomed-in views, although LCDPNet and CLIP-LIT succeed in brightening local dark areas to some extent, their overall brightness remains inferior to that of DFCNet, indicating limited effectiveness in holistic illumination correction. Notably, while Zero-DCE, EnlightenGAN, SCI, RetinexDIP, and GDP suffer primarily from insufficient global brightness, DSLR and RRDNet not only exhibit inadequate enhancement but also introduce noticeable haze, which further compromises image sharpness and clarity.
In summary, DFCNet offers a balanced trade-off between brightness enhancement, detail preservation, and visual naturalness, providing an effective framework for practical LLIE applications.

4.4. Complexity Analysis

In practical applications of LLIE, the model parameter count and inference speed are critical factors determining the ability to meet real-time requirements. Large models significantly increase memory consumption, while slow inference speed directly impacts user experience or the performance of downstream tasks. To comprehensively assess the overall performance of each method under real-time constraints, we fed low-light images of size 600 × 400 pixels (from the LOL dataset) and 1080 × 720 pixels (from the DarkFace dataset) into their respective models with a batch size of 1, simulating the scenario of processing individual frames in a continuous video stream. It should be noted that RRDNet, RetinexDIP, and GDP do not rely on pre-trained models but instead require hundreds or even thousands of iterative steps to generate enhanced outputs. Due to GDP’s extremely high parameter count and the computational cost of its diffusion process, its FLOPs and inference time are not directly comparable with other models. The results are presented in Table 3, while more intuitive computational efficiency comparisons are visualized in Figure 8. Our DFCNet exhibits exceptional competitiveness on both key metrics: it boasts a remarkably low parameter count of merely 0.0428M and requires only 0.0038 s to enhance a single image on the 600 × 400 LOL dataset (with 0.0051 s on the 1080 × 720 DarkFace dataset). This extremely low parameter count makes it the second lightest model in this comparison, significantly lower than EnlightenGAN (8.637M), while achieving comparable enhancement quality. In terms of computational cost (floating point operations, FLOPs), DFCNet shows favorable efficiency across resolutions, with 4.0263G on LOL and 13.0454G on DarkFace, ranking competitively. However, the inference time of the second-ranked method in parameter compactness, RetinexDIP, is exceedingly slow (reaching 16.2311 s per image on LOL and 16.8708 s on DarkFace), severely limiting its practical deployment. Although the parameter count and per-image inference time of DFCNet are higher than those of the top-ranked SCI, DFCNet achieves superior enhancement results. In summary, DFCNet’s performance demonstrates its ability to handle real-time LLIE tasks efficiently, striking an optimal balance between model compactness and computational efficiency across diverse low-light image resolutions.

4.5. Ablation Study

The effectiveness of the proposed PFM, IE, and FRC modules within DFCNet was evaluated through eight ablation experiments conducted on the LOL dataset. Performance was assessed using PSNR, SSIM, NIQE, and BRISQUE metrics. Detailed experimental configurations and corresponding results are presented in Table 4. To further illustrate the impact of each module on visual quality, visual comparisons of enhancement results are provided in Figure 9, where the right panel displays zoomed-in details of key regions. Baseline1 replaces the IE module with standard convolutional layers (employing ReLU activations) of equivalent depth for illumination estimation. Baseline2, in contrast, substitutes the FRC module with a multi-head spatial-frequency convolution block. Visual enhancement results, illustrated in Figure 9, demonstrate DFCNet’s superior performance. Compared to the cascaded convolutions in Baseline1, our IE module achieves more accurate illumination prediction, yielding enhanced images with higher overall brightness. Unlike the multi-head convolution block in Baseline2, whose redundant heads amplify noise within the sparse frequency spectrum of static images (leading to color fringing and blurred edges), our single-branch FRC effectively suppresses these artifacts, preserving clean and color-consistent boundaries. Guided by the PFM module, the IE module focuses on capturing extremely dark regions, enabling better illumination prediction while preserving image details during the enhancement process. The inclusion of the FRC module effectively corrects color casts in the enhanced images, producing more natural and visually pleasing results. The synergistic integration of all three modules achieves optimal performance across all evaluation metrics, highlighting their complementary roles in LLIE.

5. Conclusions

To address the issue of degraded image quality under low-light imaging conditions, an innovative unsupervised LLIE method named DFCNet is proposed. DFCNet effectively overcomes the limitations of traditional supervised approaches, particularly their inability to meet real-time processing requirements. It achieves this through a dual-stage workflow. In the first stage, DFCNet utilizes PFM to modulate low-light images, guiding the network toward more accurate illumination estimation. The reflectance map, representing the enhanced output, is then obtained by dividing each pixel of the input image by the corresponding value in the estimated illumination map. In the second stage, FRC is introduced to further optimize network performance, significantly improving the quality of the enhanced images. DFCNet not only improves the usability of low-light images but also demonstrates strong potential in providing more reliable visual data in practical applications such as security surveillance and autonomous driving. Future work will explore deploying DFCNet in broader real-world scenarios to further extend its applicability and impact.

Author Contributions

Conceptualization, H.Z.; methodology, H.Z.; software, H.Z. and Y.M.; validation, J.L. and L.L.; investigation, Y.M. and Y.L.; resources, J.L.; data curation, H.Z. and L.L.; writing—original draft preparation, H.Z.; writing—review and editing, H.Z., J.L., Y.M. and L.L.; visualization, H.Z. and Y.L.; funding acquisition, J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Cui, Y.; Ren, W.; Cao, X.; Knoll, A. Image Restoration via Frequency Selection. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 46, 1093–1108. [Google Scholar] [CrossRef]
  2. Munsif, M.; Khan, S.U.; Khan, N.; Baik, S.W. Attention-Based Deep Learning Framework for Action Recognition in a Dark Environment. Hum. Centric Comput. Inf. Sci. 2024, 14, 4. [Google Scholar]
  3. Herath, H.; Herath, H.; Madusanka, N.; Lee, B.-I. A Systematic Review of Medical Image Quality Assessment. J. Imaging 2025, 11, 100. [Google Scholar] [CrossRef] [PubMed]
  4. Zhang, J.; Cheng, B.; Zhang, T.; Zhao, Y.; Fu, T.; Wu, Z.; Tao, X. MIMO-Uformer: A Transformer-Based Image Deblurring Network for Vehicle Surveillance Scenarios. J. Imaging 2024, 10, 274. [Google Scholar] [CrossRef]
  5. Wei, C.; Wang, W.; Yang, W.; Liu, J. Deep Retinex Decomposition for Low-Light Enhancement. arXiv 2018, arXiv:1808.04560. [Google Scholar] [CrossRef]
  6. Jiang, Y.; Gong, X.; Liu, D.; Cheng, Y.; Fang, C.; Shen, X.; Yang, J.; Zhou, P.; Wang, Z. Enlightengan: Deep Light Enhancement Without Paired Supervision. IEEE Trans. Image Process. 2021, 30, 2340–2349. [Google Scholar] [CrossRef]
  7. Tatana, M.M.; Tsoeu, M.S.; Maswanganyi, R.C. Low-Light Image and Video Enhancement for More Robust Computer Vision Tasks: A Review. J. Imaging 2025, 11, 125. [Google Scholar] [CrossRef]
  8. Guo, J.; Ma, J.; García-Fernández, Á.F.; Zhang, Y.; Liang, H. A Survey on Image Enhancement for Low-Light Images. Heliyon 2023, 9, e14558. [Google Scholar] [CrossRef] [PubMed]
  9. Jobson, D.J.; Rahman, Z.; Woodell, G.A. Properties and Performance of a Center/Surround Retinex. IEEE Trans. Image Process. 1997, 6, 451–462. [Google Scholar] [CrossRef]
  10. Li, D.; Zhang, Y.; Wen, P.; Bai, L. A Retinex Algorithm for Image Enhancement Based on Recursive Bilateral Filtering. In Proceedings of the 2015 11th International Conference on Computational Intelligence and Security (CIS), Shenzhen, China, 19–20 December 2015; IEEE: New York, NY, USA, 2015; pp. 154–157. [Google Scholar]
  11. Guo, X.; Li, Y.; Ling, H. LIME: Low-Light Image Enhancement via Illumination Map Estimation. IEEE Trans. Image Process. 2016, 26, 982–993. [Google Scholar] [CrossRef] [PubMed]
  12. Lore, K.G.; Akintayo, A.; Sarkar, S. LLNet: A Deep Autoencoder Approach to Natural Low-Light Image Enhancement. Pattern Recognit. 2017, 61, 650–662. [Google Scholar] [CrossRef]
  13. Lv, F.; Lu, F.; Wu, J.; Lim, C. MBLLEN: Low-Light Image/Video Enhancement Using Cnns. In Proceedings of the Bmvc, Newcastle, UK, 3–6 September 2018; Volume 220, p. 4. [Google Scholar]
  14. Chen, C.; Chen, Q.; Xu, J.; Koltun, V. Learning to See in the Dark. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3291–3300. [Google Scholar]
  15. Lim, S.; Kim, W. DSLR: Deep Stacked Laplacian Restorer for Low-Light Image Enhancement. IEEE Trans. Multimed. 2020, 23, 4272–4284. [Google Scholar] [CrossRef]
  16. Zhang, Y.; Zhang, J.; Guo, X. Kindling the Darkness: A Practical Low-Light Image Enhancer. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 1632–1640. [Google Scholar]
  17. Wu, W.; Weng, J.; Zhang, P.; Wang, X.; Yang, W.; Jiang, J. Uretinex-Net: Retinex-Based Deep Unfolding Network for Low-Light Image Enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5901–5910. [Google Scholar]
  18. Cheng, D.; Chen, L.; Lv, C.; Guo, L.; Kou, Q. Light-Guided and Cross-Fusion U-Net for Anti-Illumination Image Super-Resolution. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 8436–8449. [Google Scholar] [CrossRef]
  19. Wang, H.; Xu, K.; Lau, R.W. Local Color Distributions Prior for Image Enhancement. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: New York, NY, USA, 2022; pp. 343–359. [Google Scholar]
  20. Cai, Y.; Bian, H.; Lin, J.; Wang, H.; Timofte, R.; Zhang, Y. Retinexformer: One-Stage Retinex-Based Transformer for Low-Light Image Enhancement. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 12504–12513. [Google Scholar]
  21. Bai, J.; Yin, Y.; He, Q.; Li, Y.; Zhang, X. Retinexmamba: Retinex-Based Mamba for Low-Light Image Enhancement. arXiv 2024, arXiv:2405.03349. [Google Scholar]
  22. Yao, Z.; Fan, G.; Fan, J.; Gan, M.; Chen, C.L.P. Spatial-Frequency Dual-Domain Feature Fusion Network for Low-Light Remote Sensing Image Enhancement. IEEE Trans. Geosci. Remote Sens. 2024, 62, 4706516. [Google Scholar] [CrossRef]
  23. Wang, W.; Wei, C.; Yang, W.; Liu, J. Gladnet: Low-Light Enhancement Network with Global Awareness. In Proceedings of the 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018), Xi’an, China, 15–19 May 2018; IEEE: New York, NY, USA, 2018; pp. 751–755. [Google Scholar]
  24. Guo, C.; Li, C.; Guo, J.; Loy, C.C.; Hou, J.; Kwong, S.; Cong, R. Zero-Reference Deep Curve Estimation for Low-Light Image Enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1780–1789. [Google Scholar]
  25. Zhu, A.; Zhang, L.; Shen, Y.; Ma, Y.; Zhao, S.; Zhou, Y. Zero-Shot Restoration of Underexposed Images via Robust Retinex Decomposition. In Proceedings of the 2020 IEEE International Conference on Multimedia and Expo (ICME), London, UK, 6–10 July 2020; IEEE: New York, NY, USA, 2020; pp. 1–6. [Google Scholar]
  26. Yang, W.; Wang, S.; Fang, Y.; Wang, Y.; Liu, J. Band Representation-Based Semi-Supervised Low-Light Image Enhancement: Bridging the Gap Between Signal Fidelity and Perceptual Quality. IEEE Trans. Image Process. 2021, 30, 3461–3473. [Google Scholar] [CrossRef]
  27. Zhao, Z.; Xiong, B.; Wang, L.; Ou, Q.; Yu, L.; Kuang, F. RetinexDIP: A Unified Deep Framework for Low-Light Image Enhancement. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 1076–1088. [Google Scholar] [CrossRef]
  28. Wang, R.; Jiang, B.; Yang, C.; Li, Q.; Zhang, B. MAGAN: Unsupervised Low-Light Image Enhancement Guided by Mixed-Attention. Big Data Min. Anal. 2022, 5, 110–119. [Google Scholar] [CrossRef]
  29. Ma, L.; Ma, T.; Liu, R.; Fan, X.; Luo, Z. Toward Fast, Flexible, and Robust Low-Light Image Enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5637–5646. [Google Scholar]
  30. Kandula, P.; Suin, M.; Rajagopalan, A.N. Illumination-Adaptive Unpaired Low-Light Enhancement. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 3726–3736. [Google Scholar] [CrossRef]
  31. Liang, Z.; Li, C.; Zhou, S.; Feng, R.; Loy, C.C. Iterative Prompt Learning for Unsupervised Backlit Image Enhancement. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 8094–8103. [Google Scholar]
  32. Jiang, Q.; Mao, Y.; Cong, R.; Ren, W.; Huang, C.; Shao, F. Unsupervised Decomposition and Correction Network for Low-Light Image Enhancement. IEEE Trans. Intell. Transp. Syst. 2022, 23, 19440–19455. [Google Scholar] [CrossRef]
  33. Fei, B.; Lyu, Z.; Pan, L.; Zhang, J.; Yang, W.; Luo, T.; Zhang, B.; Dai, B. Generative Diffusion Prior for Unified Image Restoration and Enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 9935–9946. [Google Scholar]
  34. Wang, W.; Luo, R.; Yang, W.; Liu, J. Unsupervised Illumination Adaptation for Low-Light Vision. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 5951–5966. [Google Scholar] [CrossRef]
  35. Peng, B.; Zhang, J.; Zhang, Z.; Huang, Q.; Chen, L.; Lei, J. Unsupervised Low-Light Image Enhancement via Luminance Mask and Luminance-Independent Representation Decoupling. IEEE Trans. Emerg. Top. Comput. Intell. 2024, 8, 3029–3039. [Google Scholar] [CrossRef]
  36. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–Miccai 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Springer: New York, NY, USA, 2015; pp. 234–241. [Google Scholar]
  37. Sun, L.; Dong, J.; Tang, J.; Pan, J. Spatially-Adaptive Feature Modulation for Efficient Image Super-Resolution. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–3 October 2023; pp. 13190–13199. [Google Scholar]
  38. Rowlands, D.A.; Finlayson, G.D. Optimisation of Convolution-Based Image Lightness Processing. J. Imaging 2024, 10, 204. [Google Scholar] [CrossRef] [PubMed]
  39. Wang, R.; Zhang, Q.; Fu, C.-W.; Shen, X.; Zheng, W.-S.; Jia, J. Underexposed Photo Enhancement Using Deep Illumination Estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 6849–6857. [Google Scholar]
  40. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  41. Zhu, X.; Cheng, D.; Zhang, Z.; Lin, S.; Dai, J. An Empirical Study of Spatial Attention Mechanisms in Deep Networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6688–6697. [Google Scholar]
  42. Li, C.; Guo, C.-L.; Zhou, M.; Liang, Z.; Zhou, S.; Feng, R.; Loy, C.C. Embedding Fourier for Ultra-High-Definition Low-Light Image Enhancement. arXiv 2023, arXiv:2302.11831. [Google Scholar]
  43. Zhang, Y.; Guo, X.; Ma, J.; Liu, W.; Zhang, J. Beyond Brightening Low-Light Images. Int. J. Comput. Vis. 2021, 129, 1013–1037. [Google Scholar] [CrossRef]
  44. Yang, W.; Wang, W.; Huang, H.; Wang, S.; Liu, J. Sparse Gradient Regularized Deep Retinex Network for Robust Low-Light Image Enhancement. IEEE Trans. Image Process. 2021, 30, 2072–2086. [Google Scholar] [CrossRef]
  45. Wang, S.; Zheng, J.; Hu, H.-M.; Li, B. Naturalness Preserved Enhancement Algorithm for Non-Uniform Illumination Images. IEEE Trans. Image Process. 2013, 22, 3538–3548. [Google Scholar] [CrossRef]
  46. Mittal, A.; Moorthy, A.K.; Bovik, A.C. No-Reference Image Quality Assessment in the Spatial Domain. IEEE Trans. Image Process. 2012, 21, 4695–4708. [Google Scholar] [CrossRef]
  47. Bychkovsky, V.; Paris, S.; Chan, E.; Durand, F. Learning Photographic Global Tonal Adjustment with a Database of Input/Output Image Pairs. In Proceedings of the Cvpr 2011, Colorado Springs, CO, USA, 20–25 June 2011; IEEE: New York, NY, USA, 2011; pp. 97–104. [Google Scholar]
  48. Loh, Y.P.; Chan, C.S. Getting to Know Low-Light Images with the Exclusively Dark Dataset. Comput. Vis. Image Underst. 2019, 178, 30–42. [Google Scholar] [CrossRef]
  49. Yang, W.; Yuan, Y.; Ren, W.; Liu, J.; Scheirer, W.J.; Wang, Z.; Zhang, T.; Zhong, Q.; Xie, D.; Pu, S. Advancing Image Understanding in Poor Visibility Environments: A Collective Benchmark Study. IEEE Trans. Image Process. 2020, 29, 5737–5752. [Google Scholar] [CrossRef]
Figure 1. Overview of the proposed DFCNet, including components such as preliminary feature modulation (PFM), illumination estimation (IE), frequency-domain residual calibration (FRC), and the overall workflow. Here, H and W denote the input image’s height and width, respectively. Arrows indicate the flow direction of feature maps between components. The dual-stage process is visually distinguished by colored lines: pink lines indicate the workflow shared by both stages, and green lines mark processes exclusive to the second stage.
Figure 1. Overview of the proposed DFCNet, including components such as preliminary feature modulation (PFM), illumination estimation (IE), frequency-domain residual calibration (FRC), and the overall workflow. Here, H and W denote the input image’s height and width, respectively. Arrows indicate the flow direction of feature maps between components. The dual-stage process is visually distinguished by colored lines: pink lines indicate the workflow shared by both stages, and green lines mark processes exclusive to the second stage.
Jimaging 11 00253 g001
Figure 2. Comparison of magnitudes and phases of low-light, normal-light, and enhanced images.
Figure 2. Comparison of magnitudes and phases of low-light, normal-light, and enhanced images.
Jimaging 11 00253 g002
Figure 3. Schematic of the MAB.
Figure 3. Schematic of the MAB.
Jimaging 11 00253 g003
Figure 4. Visual comparison of mainstream methods and DFCNet on the LOL dataset. Right panels show local enlargements of regions marked by green boxes in left panels.
Figure 4. Visual comparison of mainstream methods and DFCNet on the LOL dataset. Right panels show local enlargements of regions marked by green boxes in left panels.
Jimaging 11 00253 g004
Figure 5. Visual comparison of mainstream methods and DFCNet on the MIT dataset. Right panels show local enlargements of regions marked by green boxes in left panels.
Figure 5. Visual comparison of mainstream methods and DFCNet on the MIT dataset. Right panels show local enlargements of regions marked by green boxes in left panels.
Jimaging 11 00253 g005
Figure 6. Visual comparison of mainstream methods and DFCNet on the ExDark dataset. Right panels show local enlargements of regions marked by green boxes in left panels.
Figure 6. Visual comparison of mainstream methods and DFCNet on the ExDark dataset. Right panels show local enlargements of regions marked by green boxes in left panels.
Jimaging 11 00253 g006
Figure 7. Visual comparison of mainstream methods and DFCNet on the DarkFace dataset. Right panels show local enlargements of regions marked by green boxes in left panels.
Figure 7. Visual comparison of mainstream methods and DFCNet on the DarkFace dataset. Right panels show local enlargements of regions marked by green boxes in left panels.
Jimaging 11 00253 g007
Figure 8. Comparison of computational efficiency across different methods, where the bubble size represents model size (m).
Figure 8. Comparison of computational efficiency across different methods, where the bubble size represents model size (m).
Jimaging 11 00253 g008
Figure 9. Visualization results of the ablation experiments on PFM, IE, and FRC conducted on the LOL dataset. Right panels show local enlargements of regions marked by green boxes in left panels.
Figure 9. Visualization results of the ablation experiments on PFM, IE, and FRC conducted on the LOL dataset. Right panels show local enlargements of regions marked by green boxes in left panels.
Jimaging 11 00253 g009
Table 1. Qualitative results of mainstream methods and DFCNet on LOL and MIT datasets. Red and blue numbers are used to mark the best and second-best performers, respectively. ↑ Represents the higher the value is better, whereas ↓ means the lower the value is better.
Table 1. Qualitative results of mainstream methods and DFCNet on LOL and MIT datasets. Red and blue numbers are used to mark the best and second-best performers, respectively. ↑ Represents the higher the value is better, whereas ↓ means the lower the value is better.
MethodsVenueLOLMIT
PSNR ↑SSIM ↑NIQE ↓BRISQUE ↓PSNR ↑SSIM ↑NIQE ↓BRISQUE ↓
RetinexNetBMVC’1816.77 ± 2.380.564 ± 0.0778.39 ± 0.5335.43 ± 7.6914.29 ± 1.890.698 ± 0.0874.58 ± 0.9229.24 ± 8.59
DSLRTMM’2014.95 ± 4.160.600 ± 0.1318.18 ± 0.4833.78 ± 8.7518.17 ± 3.880.769 ± 0.0473.83 ± 0.7932.66 ± 10.05
RRDNetICME’2010.70 ± 3.650.436 ± 0.1487.82 ± 0.9129.32 ± 8.2716.16 ± 3.690.664 ± 0.0684.08 ± 0.8830.20 ± 10.22
Zero-DCECVPR’2014.86 ± 4.270.562 ± 0.1247.85 ± 1.0324.98 ± 8.1016.07 ± 1.950.738 ± 0.0613.65 ± 0.7624.31 ± 8.52
EnlightenGANTIP’2117.48 ± 3.530.651 ± 0.1105.40 ± 0.6923.77 ± 5.8917.95 ± 2.440.791 ± 0.0603.19 ± 0.7028.14 ± 9.49
RetinexDIPTCSVT’2111.68 ± 3.860.487 ± 0.1367.57 ± 1.0822.55 ± 6.8918.41 ± 3.110.836 ± 0.0493.23 ± 0.6722.86 ± 8.83
LCDPNetECCV’2214.51 ± 4.940.575 ± 0.0987.44 ± 0.7922.63 ± 7.3717.07 ± 2.520.770 ± 0.0653.24 ± 0.7425.75 ± 10.03
SCICVPR’2214.81 ± 4.080.548 ± 0.1377.81 ± 1.0627.32 ± 6.6018.06 ± 3.100.815 ± 0.0633.55 ± 0.7925.55 ± 8.26
CLIP-LITICCV’2312.39 ± 3.620.493 ± 0.1217.90 ± 0.9728.10 ± 8.0817.19 ± 2.940.773 ± 0.0663.38 ± 0.7623.90 ± 9.27
GDPCVPR’2315.90 ± 2.380.542 ± 0.0777.95 ± 0.7330.14 ± 7.6918.26 ± 2.890.799 ± 0.0873.22 ± 0.9224.76 ± 8.59
Ours-17.04 ± 3.320.592 ± 0.1077.12 ± 0.7924.06 ± 7.6818.73 ± 2.260.787 ± 0.0693.05 ± 0.7123.11 ± 8.48
Table 2. Qualitative results of mainstream methods and DFCNet on ExDark and DarkFace datasets. Red and blue numbers are used to mark the best and second-best performers, respectively. ↓ means the lower the value is better.
Table 2. Qualitative results of mainstream methods and DFCNet on ExDark and DarkFace datasets. Red and blue numbers are used to mark the best and second-best performers, respectively. ↓ means the lower the value is better.
MethodsVenueExDarkDarkFace
NIQE ↓BRISQUE ↓NIQE ↓BRISQUE ↓
RetinexNetBMVC’184.44 ± 0.6831.64 ± 9.054.42 ± 0.2234.67 ± 2.83
DSLRTMM’203.90 ± 0.6428.65 ± 8.553.74 ± 0.3834.90 ± 4.17
RRDNetICME’203.65 ± 0.9920.77 ± 7.453.97 ± 0.4430.34 ± 3.59
Zero-DCECVPR’203.03 ± 0.5827.72 ± 7.043.50 ± 0.4528.18 ± 5.16
EnlightenGANTIP’213.12 ± 0.6319.61 ± 6.813.26 ± 0.3326.61 ± 3.22
RetinexDIPTCSVT’213.32 ± 0.6325.76 ± 6.813.58 ± 0.4324.50 ± 4.27
LCDPNetECCV’223.27 ± 0.8120.22 ± 7.543.23 ± 0.2928.62 ± 4.92
SCICVPR’222.97 ± 0.6621.65 ± 7.153.36 ± 0.4424.71 ± 3.26
CLIP-LITICCV’233.16 ± 0.6623.65 ± 8.423.73 ± 0.4832.46 ± 6.35
GDPCVPR’233.21 ± 0.6825.45 ± 7.053.68 ± 0.2229.16 ± 3.83
Ours-2.92 ± 0.5720.03 ± 6.983.13 ± 0.3723.29 ± 2.40
Table 3. Model sizes, FLOPs, and single-image inference times of different models under real-time constraints (GPU-based). Red and blue numbers are used to mark the best and second-best performers, respectively.
Table 3. Model sizes, FLOPs, and single-image inference times of different models under real-time constraints (GPU-based). Red and blue numbers are used to mark the best and second-best performers, respectively.
MethodsVenueSIZE(M)LOL (600 × 400)DarkFace (1080 × 720)
FLOPs (G)TIME (S)FLOPs (G)TIME (S)
RetinexNetBMVC’180.8383136.01510.1131441.45130.5533
DSLRTMM’2014.931343.03190.0729141.00710.0451
RRDNetICME’200.128230.650090.470299.3212272.6918
Zero-DCECVPR’200.079419.00810.006161.58590.0076
EnlightenGANTIP’218.637061.02610.0098196.56700.1028
RetinexDIPTCSVT’210.70723.413416.231111.254916.8708
LCDPNetECCV’220.28185.00680.038215.98250.0286
SCICVPR’220.00030.12820.00250.41520.0033
CLIP-LITICCV’230.278866.67010.0095216.01110.0165
Ours-0.04284.02630.003813.04540.0051
Table 4. Qualitative results of ablation experiments on the LOL dataset. √ Indicates the module is used. Red and blue numbers are used to mark the best and second-best performers, respectively. ↑ Represents the higher the value is better, whereas ↓ means the lower the value is better.
Table 4. Qualitative results of ablation experiments on the LOL dataset. √ Indicates the module is used. Red and blue numbers are used to mark the best and second-best performers, respectively. ↑ Represents the higher the value is better, whereas ↓ means the lower the value is better.
BaselinePFMIEFRCPSNR ↑SSIM ↑NIQE ↓BRISQUE ↓
Baseline1 10.0490.3818.55631.650
Baseline2 12.4140.4607.37426.561
12.3480.4568.13928.632
13.4960.4677.93528.247
12.9190.4757.44226.354
15.9850.5747.64427.468
15.4550.5697.30925.919
17.0390.5927.11524.060
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, H.; Li, J.; Mao, Y.; Liu, L.; Lu, Y. DFCNet: Dual-Stage Frequency-Domain Calibration Network for Low-Light Image Enhancement. J. Imaging 2025, 11, 253. https://doi.org/10.3390/jimaging11080253

AMA Style

Zhou H, Li J, Mao Y, Liu L, Lu Y. DFCNet: Dual-Stage Frequency-Domain Calibration Network for Low-Light Image Enhancement. Journal of Imaging. 2025; 11(8):253. https://doi.org/10.3390/jimaging11080253

Chicago/Turabian Style

Zhou, Hui, Jun Li, Yaming Mao, Lu Liu, and Yiyang Lu. 2025. "DFCNet: Dual-Stage Frequency-Domain Calibration Network for Low-Light Image Enhancement" Journal of Imaging 11, no. 8: 253. https://doi.org/10.3390/jimaging11080253

APA Style

Zhou, H., Li, J., Mao, Y., Liu, L., & Lu, Y. (2025). DFCNet: Dual-Stage Frequency-Domain Calibration Network for Low-Light Image Enhancement. Journal of Imaging, 11(8), 253. https://doi.org/10.3390/jimaging11080253

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop