Next Article in Journal
Research on Time–Frequency Domain Characteristics Analysis of Fault Arc Under Different Connection Methods
Previous Article in Journal
An Zero-Point Drift Suppression Method for eLoran Signal Based on a Segmented Inaction Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Noise-Resilient Low-Light Image Enhancement with CLIP Guidance and Pixel-Reordering Subsampling

School of Integrated Technology, College of Computing, Yonsei University, 402, 12, Baumoe-ro 6-gil, Seocho-gu, Seoul 06763, Republic of Korea
Electronics 2025, 14(24), 4839; https://doi.org/10.3390/electronics14244839 (registering DOI)
Submission received: 16 November 2025 / Revised: 4 December 2025 / Accepted: 4 December 2025 / Published: 8 December 2025

Abstract

Low-light image enhancement (LLIE) is an essential task for improved image quality that ultimately supports crucial downstream tasks such as autonomous driving and mobile photography. Despite notable advances achieved by traditional, Retinex-based methods, existing approaches still struggle to maintain globally consistent illumination and to suppress sensor noise under extremely dark conditions. To overcome these limitations, we propose a noise-resilient LLIE framework that integrates a CLIP-guided loss (CLIP-LLA) and a pixel-reordering subsampling (PRS) scheme into the Retinexformer backbone. The CLIP-LLA loss exploits the semantic prior of a large-scale vision–language model to align enhanced outputs within the manifold of well-illuminated natural images, leading to faithful global tone rendering and perceptual realism. In parallel, the PRS-based multi-scale training strategy effectively regularizes the network by augmenting structural diversity, thereby improving denoising capability without architectural modification or inference cost. Extensive experiments on both sRGB and RAW benchmarks validate the effectiveness of our design. The proposed method achieves consistent improvements over state-of-the-art techniques, including a + 2.73 dB PSNR gain on the SMID dataset and superior perceptual scores, while maintaining computational efficiency. These results demonstrate that fusing foundation-model priors with transformer-based Retinex frameworks offers a practical and scalable pathway toward perceptually faithful low-light image enhancement.

1. Introduction

Low-light image enhancement (LLIE) is one of the most fundamental restoration tasks in computer vision. It directly influences the reliability of downstream applications such as nighttime surveillance [1,2], autonomous driving [3,4], and remote sensing [5,6]. Real-world low-light scenes suffer from severe degradations—including sensor noise, detail loss, and color imbalance—that cannot be remedied by naïve exposure or gamma correction alone. This makes LLIE a core challenge for robust visual perception in both scientific and consumer imaging pipelines.
Early efforts in LLIE explored a wide spectrum of enhancement paradigms, ranging from histogram equalization [7] and image-adaptive lookup-table approaches [8] to learning-based mappings [9,10]. Among them, Retinex-based methods [11,12,13,14,15] hold particular significance because they explicitly model illumination–reflectance interaction—a physical interpretation of image formation that promotes decomposition stability and generalization across domains. This formulation not only bridges traditional physics-based reasoning and modern deep learning, but also provides a principled way to incorporate semantic priors and transformer-based architectures (e.g., RetinexFormer [16]). Consequently, Retinex-guided frameworks serve as a unifying platform capable of accommodating varied enhancement strategies while preserving interpretable illumination estimation.
Despite these advances, current LLIE models still struggle to produce globally consistent illumination and to suppress sensor-induced noise in extreme darkness. Pixel-level optimization schemes tend to overfit to dataset-specific brightness distributions and lack semantic awareness, resulting in color shift and unnatural tone rendering. Furthermore, diffusion-based approaches [17,18,19,20] achieve strong restoration quality but often hallucinate structures and incur excessive computational cost, limiting their practical use in real-time or mobile scenarios. These limitations highlight the need for models that combine high-level semantic priors with efficient transformer-based architectures for robust, perceptually faithful enhancement.
To address these challenges, we extend the RetinexFormer framework with two key components, as shown in Figure 1. First, we introduce a CLIP-guided low-light assessment loss (CLIP-LLA) that leverages the vision–language priors of CLIP, or contrastive language-image pre-training [21], to align enhanced images with the semantic distribution of well-illuminated natural scenes. By optimizing in the CLIP embedding space, the network learns global tone consistency and color semantics that are difficult to capture through pixel-wise objectives alone. Second, we propose a pixel-reordering subsampling (PRS) strategy for multi-scale training. This strategy creates diverse sub-samples from single images without altering the model architecture, serving as an implicit regularizer that reduces noise and enhances local detail preservation. Together, these components address the two dominant limitations in LLIE—global tone ambiguity and noise resilience—through a synergistic combination of semantic guidance and multi-scale representation learning.
The main contributions of this work are summarized as follows:
  • We introduce the CLIP-LLA loss to inject semantic and perceptual priors from a foundation model into LLIE training, achieving faithful illumination and color consistency.
  • We design a pixel-reordering subsampling (PRS) scheme that enhances robustness and noise suppression through multi-scale data regularization.
  • We conduct extensive experiments on representative benchmarks such as LOL variations and SID. The results demonstrate consistent performance boosts, over 2.7 dB PSNR and 0.010 LPIPS improvement on SMID benchmark, validating the effectiveness of our proposed method.

2. Related Works

Traditional algorithmic approaches, such as gamma correction [22] and histogram equalization [7], operate by remapping pixel intensities to adjust the dynamic range. However, these methods are insufficient for aesthetically enhancing real-world RGB images that contain complex color components. To address this limitation, Retinex-based techniques [11,23] have been introduced, which decompose low-light images into reflection and illumination maps. Despite their effectiveness, these approaches face significant challenges: designing explicit priors [12] that can generalize across diverse scenes is inherently difficult, and they remain vulnerable to noise artifacts [13] introduced by low exposure conditions.
The introduction of deep learning to LLIE began with autoencoder-based models trained on synthetically darkened data [9]. Multi-branch CNNs further addressed brightness restoration and noise reduction [10]. To ease paired-data dependency, supervised and adversarial methods exploited expert-enhanced references [24,25], while zero-reference strategies estimated pixel-wise curves without ground truth [26]. Retinex theory was soon incorporated into learning frameworks: Retinex-Net jointly decomposed reflectance and illumination [14], and KinD extended this with practical pipelines for color fidelity and denoising [15]. These works established Retinex-guided enhancement as a strong foundation for LLIE.
Transformers, first introduced for sequence modeling [27], were later adapted to vision tasks in ViT [28] and enhanced by Swin Transformer with shifted-window attention [29]. In image restoration, architectures like Uformer [30] and Restormer [31] enabled efficient local–global modeling. LLIE soon adopted these designs, with LLFormer exploiting axis-based attention for high-resolution enhancement [32]. Retinexformer advanced this trend by combining Retinex decomposition with Transformer attention [16], guiding illumination estimation and achieving strong performance under uneven lighting.
RetinexFormer has become the dominant baseline for LLIE, inspiring a range of studies aimed at improving its training inefficiency. RetinexFormer+ [33] incorporates a dual-channel paradigm and a Nested U-Net [34] to increase efficiency, while RetinexMamba [35] follows the recent shift toward state-space models [36] by substituting the transformer components with Mamba blocks [37]. Spike-RetinexFormer [38] leverages spiking neural networks [39] to reduce energy consumption. These variants typically deliver comparable or slightly improved pixel-wise metrics while lowering training cost. Additionally, several works seek to extend RetinexFormer to real-world deployment, including methods that reduce domain gaps using unpaired real data [40] and unsupervised fine-tuning techniques based solely on extra-low-light images [41]. Through these developments, RetinexFormer continues to serve as a strong and widely adopted golden standard for LLIE.
In the field of generative vision, the success of diffusion models has brought increasing attention to generation-based approaches for low-light image enhancement. Among these, Diff-Retinex [17] employs the diffusion process for the reconstruction of both the illumination map and the reflectance map. Other approaches [18,19,20] have demonstrated performance improvements by directly learning a representation of the degradation itself. Nevertheless, a few of those diffusion-based methods tend to produce hallucinations, non-existent structures in extremely low-light regions. Furthermore, a critical limitation of these methods lies in their excessive computational cost and prolonged inference time, rendering them impractical for real-world applications.
By jointly training LLIE and deblurring, very recent deterministic LLIE models DarkIR [42] and URWKV [43] report more efficient training and considerably better pixel-based evaluation metrics in several benchmark datasets, while the training part of their model is not available publicly yet. Thus, we selected the relatively popular and widely verified benchmark model Retinexformer.

3. Method

3.1. Motivation and Preliminaries

In LLIE tasks, incorporating a prior that enables the separation of illumination and reflectance contributes to improved training stability and better generalization. Consequently, retinex theory aims to decompose a low-light image I R H × W × 3 into a reflectance image R R H × W × 3 and an illumination map L R H × W as
I = R L ,
where ⊙ denotes the element-wise multiplication. Traditional approaches estimate L using handcrafted priors such as the Gray-world or Max-RGB assumptions. In contrast, CNN-based supervised methods adopt a data-driven strategy to predict the normal-light illumination map, which is subsequently multiplied with the reflectance component R to assume the normal-light image I l u .
To account for the inevitable corruption in low-light images, Retinexformer [16] models the reflectance R as the normal-light target image, obtained by removing the overall corruption C from the light-up image I l u R H × W × 3 . The light-up image is formulated as
I l u = I L 0 ,
where the network aims to predict the light-up map L 0 R H × W . The obtained I l u effectively serves as a post-processing method for corruption reduction. To simultaneously estimate the corruption term C , Retinexformer employs a U-net structure with skip connections and transformer blocks, in which the self-attention mechanism is modified to utilize the illumination prior for guidance.
Despite its effectiveness, Retinexformer’s optimization remains limited by pixel-level supervision, which cannot capture the global semantics or perceptual realism of well-lit natural scenes. Moreover, its training instability under low SNR inputs often leads to oscillatory convergence (Figure 2), causing inconsistencies between pixel-level and perceptual metrics.
To overcome these issues, our method introduces two complementary components: (1) CLIP-guided low-light assessment (CLIP-LLA) to align restored images with the semantic manifold of well-lit images via foundation-model priors and (2) pixel-reordering subsampling (PRS) to improve generalization and noise resilience through efficient multi-scale regularization. These two components are designed to be modular and orthogonal, improving Retinexformer’s perceptual alignment and local detail robustness without modifying its inference pipeline.

3.2. CLIP-LLA Loss

CLIP [21] is a large-scale vision–language foundation model designed to align visual and textual embeddings. We leverage this semantic alignment capability to design a CLIP-guided loss (CLIP-LLA) that encourages the enhanced image to resemble a semantically well-lit image both textually and visually.
Textual Guidance. We take inspiration from CLIP-IQA [44], which exploited CLIP’s capability and introduced single-image visual quality assessment. Specifically, the approach employs predefined positive–negative keyword pairs and evaluates image quality by measuring the cosine similarity of the image embedding to the positive keywords and its dissimilarity to the negative keywords, thereby quantifying the overall positivity of the image.
Following CLIP-IQA [44], we compute the cosine similarity between L2-normalized image features I feat R 1 × d and text embeddings T feat R 2 × d of positive–negative prompt pairs as follows:
z = I feat T feat × t ,
where z = [ z good , z bad ] represents the logits corresponding to a positive–negative prompt pair. And t is a temperature scaling factor for CLIP, which is fixed to 100.
The probabilities are obtained through the softmax function, p = [ p good , p bad ] = softmax ( z ) . The CLIP loss is then defined as
L CLIP - IQA = 1 1 N i = 1 N p good ( i ) ,
where N denotes the number of prompt pairs and p good ( i ) is the probability assigned to the positive prompt in the i-th pair.
For designing the Proposed CLIP-Low Light Assessment loss, or CLIP-LLA loss, the five general prompt pair z g e n e r a l for overall quality enhancement are set as following:
  • {“Good image”, “bad image”},
  • {“Sharp image”, “blurry image”},
  • {“sharp edges”, “blurry edges”},
  • {“High resolution image”, “low resolution image”},
  • {“Noise-free image”, “noisy image”}.
Additionally, we utilize a task-specific prompt pair z L L I E , namely
  • {“A well lit image”, “A low light image”}.
Visual–Semantic Guidance. In addition to the aforementioned textual guiding method, we propose a variation which exploits the average feature of the target reference image R feat as an additional positive–negative pair. Specifically, we form the logits
T feat vis = R feat + R feat R 2 × d , z mix = I feat ( T feat vis ) × t ,
which yields p mix = [ p ref , p non - ref ] through the softmax function.
This loss could be used solely as a visual variation of the loss, or can be considered if it was another additional prompt pair for calculating CLIP-LLA, written as
L CLIP - LLA - mix = 1 1 N + 1 i = 1 N p good ( i ) + p ref ,
which formulates the mixed variation.
Then, L CLIP is summed with the original Retinexformer training loss λ L 1 to construct the final training objective as
L F i n a l = L L 1 + λ L CLIP ,
where λ denotes the CLIP-LLA loss weight, which is set to 1 in our experiments. The visual guidance directs the learning process toward the ground truth of the overall dataset within a high-level perceptual embedding space. With textual guidance, the joint objective balances pixel fidelity and perceptual alignment, ensuring that enhanced images not only match the ground truth in the intensity space but also in the semantic embedding space of high-quality images.

3.3. Pixel-Reordering Subsampling (PRS): Multi-Scale Training for Noise Robustness

To enhance the local detail preservation of Retinexformer and guide consistent output, we introduce pixel-reordering subsampling (PRS). This approach is inspired by pretext task design strategies in self-supervised learning, drawing upon the image partitioning and rearrangement strategy of PsyNet [45] as well as the subsampler-based training set generation method utilized in single-image denoising methods [46,47].
PRS is a data augmentation technique that downsamples both the input and target images into four half-resolution images through pixel rearrangement. Specifically, an image I R H × W × 3 is divided into a 2 × 2 grid, and for each cell, pixel values are sampled from a fixed position among the four candidates. This procedure yields four half-resolution images { I k R H 2 × W 2 × 3 } k = 1 4 , which are then concatenated into sub-pixels, as illustrated in Figure 3. PRS is applied simultaneously to both the low-light input I and the ground-truth target R , and the resulting pairs are treated analogously to the original dataset for training.
Unlike PsyNet, where symmetries may allow semantically identical regions to be exploited, PRS arranges sub-images in a parallel manner without symmetry. Although training patches may cross sub-image boundaries, the presence of these seams increases local patch variation, which can positively influence the performance of vision models [48]. Moreover, PRS alleviates the reduced sampling frequency of near-boundary regions that commonly arises in random cropping. It is important to note that PRS is employed solely during training; inference proceeds without modification.
We fix the ordering of 4 sub-images to match the relative position of pixel. Practically, PRS can be performed and saved in advance of the training process, where benchmark datasets of LLIE are relatively small. Thereby, PRS introduces no additional training time or computation power, while effectively doubling the dataset and preventing overfit to a noise characteristic in a certain scale.

3.4. Summary of Proposed Algorithm

The proposed LLIE network is trained from scratch. Before training, PRS can optionally be precomputed and stored as an additional augmentation of the training dataset. The network is then trained in a fully supervised manner using paired low-light and ground-truth well-lit images, enabling the network to predict the latter from the former. If PRS is not precomputed, it is applied on the fly to 50 % of the training pairs so that the model learns both scales consistently.
The training objective consists of the standard L1 loss, which enforces pixel-wise fidelity between the prediction and the ground truth, combined with the proposed CLIP-LLA loss that encourages perceptual alignment. After training, the network infers well-lit images directly from low-light inputs without requiring PRS during inference.

4. Experiment

4.1. Dataset and Implementation Details

4.1.1. sRGB Datasets

We employ the widely used benchmark LOL dataset for both training and evaluation. The dataset is divided into three variants: v1 [14], v2-real, and v2-synthetic [49]. Following the protocol proposed in the original work, we adopt the official training–test splits of 485:15, 689:100, and 900:100, respectively.

4.1.2. RAW Datasets

We utilize four RAW datasets: SID, SMID, SDSD-indoor, and SDSD-outdoor. The SID dataset [50] is designed to address extreme low-light conditions and consists of 2697 paired samples captured with a Sony α 7S II camera. Following the standard protocol, the dataset is split into 2099 training pairs and 598 test pairs. The SMID dataset [51] provides real noisy–clean pairs acquired through dual smartphone captures, and it is widely recognized as a benchmark for image denoising. We use 15,763 pairs for training and 5063 pairs for testing. The SDSD dataset [52] is derived from video sequences recorded with a smartphone camera and decomposed into frame-level pairs. It contains realistic degradations including camera shake, light leakage, and sensor noise. Specifically, SDSD-indoor comprises 7778 training pairs and 360 test pairs, whereas SDSD-outdoor contains 13,690 training pairs and 390 test pairs.

4.1.3. Evaluation Metrics

For the evaluation, four metrics are employed, which can be divided into two groups. PSNR and SSIM [53] are the most commonly adopted indicators for assessing LLIE performance, although they often diverge considerably from human perceptions. To assess reconstruction fidelity, we employ the Peak Signal-to-Noise Ratio (PSNR) and the Structural Similarity Index Measure (SSIM). PSNR quantifies the pixel-wise error between a reference image I and a reconstructed image I ^ as follows:
PSNR = 10 log 10 L 2 MSE , MSE = 1 N i = 1 N ( I i I ^ i ) 2 ,
where L denotes the maximum possible pixel value and N the number of pixels. A higher PSNR indicates a smaller reconstruction error. SSIM, in contrast, measures perceptual similarity considering luminance, contrast, and structural information as follows:
SSIM ( I , I ^ ) = ( 2 μ I μ I ^ + C 1 ) ( 2 σ I I ^ + C 2 ) ( μ I 2 + μ I ^ 2 + C 1 ) ( σ I 2 + σ I ^ 2 + C 2 ) ,
where μ , σ 2 , and σ I I ^ denote the mean, variance, and covariance of the image intensities. Larger SSIM values (closer to 1) represent better perceptual quality.
Although SSIM measures perceptual structures and visual quality more faithfully, it is still lacking in terms of alignment to human perception, as it overlooks uniform regions and distortion near strong edges [54,55]. To complement this limitation, two additional perceptually oriented metrics are introduced. LPIPS [56] quantifies the distance between feature maps extracted from a pre-trained network, thereby yielding results that are more consistent with human visual perception. DISTS [57], on the other hand, explicitly separates the measurement of structural and textural information, leading to a reliable alignment with human preference when evaluating distortions.

4.1.4. Training and Evaluation Protocol

The models were independently trained and evaluated on each dataset. For patch extraction, we set the patch size to 256 × 256 for LOL-v2-real and LOL-v2-synthetic, and to 128 × 128 for the remaining datasets. For each iteration, each of the training datasets is augmented by PRS, with a probability of 0.5 . In experiments based on RAW datasets, the number of training iterations and the phase division of the cosine annealing schedule were kept identical to those used for the real-image datasets. Batch size and learning rate were halved for RAW dataset-based experiments, while all other configurations followed those of RetinexFormer. Notably, during reproduction, we observed that performance deviated substantially from the results reported in the original paper for certain datasets. The varying resolution of the dataset being used across different benchmark reports could cause PSNR differences on the LOLv1 and LOLv2-real datasets. Consequently, we re-trained and evaluated the model and dataset using the official code and configurations released by each of the authors, reporting those results herein.

4.2. Low-Light Image Enhancement

The quantitative comparison of various state-of-the-art LLIE methods and the proposed approach are presented in Table 1. Our method outperforms other approaches, including KinD [15], Restormer [31], and MIRNet [58], on the majority of the benchmark datasets. A key advantage of the proposed method is that it retains the efficiency of the baseline model, Retinexformer, particularly its low parameter count (Param) and FLOPs during inference. Furthermore, all three variations of the CLIP-LLA Loss, as detailed in the relevant section, demonstrated overall performance improvements compared to the baseline Retinexformer. Notably, the mixed variation, which leverages both textual and visual guidance, achieved significant gains. Specifically, it exhibited a PSNR increase of + 0.38 dB on the LOL-v2-synthetic dataset and a more substantial increase of + 2.73 dB on the SMID dataset.
Table 2 summarizes the extent to which the proposed method improves upon the baseline in terms of perceptually accurate evaluation metrics. In most cases, the proposed approach achieves significant improvements, including on the SDSD dataset where the pixel-wise low-level metric did not improve. The greatest effectiveness is observed in the mixed configuration, where both textual and visual guidance are employed simultaneously. Compared to the naive Retinexformer, our method outperforms in 13 out of 14 quantitative metrics across seven datasets.
The quantitative results are presented in Figure 4 and Figure 5. As shown in Figure 4, KinD suffers from reduced local saturation, resulting in tonal inconsistencies with the ground truth. SNR-Net, in contrast, tends to suppress high-frequency details, such as sharp object boundaries. Retinexformer, which serves as the most direct point of comparison with our approach, fails to preserve fine details and produces an overall reddish global tone that deviates from the reference. In comparison, the proposed method yields results that are most consistent with the ground truth, both in terms of tonal fidelity and textural detail.
Figure 5 and Figure 6 illustrate the experimental results on the LOLv2-synthetic and SMID datasets, respectively. The LOLv2-synthetic dataset presents a considerable challenge due to its wide dynamic range. On this dataset, the proposed method produces reconstructions that most closely approximate the ground truth, particularly with respect to sky coloration and image sharpness. The SMID dataset, by contrast, is characterized by severe noise-induced distortions, making it a highly demanding benchmark. Within this setting, the proposed method is uniquely capable of accurately restoring stripe-like structures, even within extremely dark regions, thereby demonstrating its robustness under adverse imaging conditions. Additional qualitative results are displayed in Appendix A.

4.3. Ablation Study

We further include an ablation study focusing on the design of individual modules within our proposed framework. Table 3 presents the results when utilizing only one of the two key components: the proposed CLIP-LLA loss or the pixel reordering subsampling (PRS). The results show that employing a CLIP-based loss alone negatively impacts computational comparison metrics of PSNR and SSIM. Conversely, incorporating multi-scale training plays a crucial role in preventing pixel-wise overfitting to the domain characteristics of the training dataset, leading to a significant performance improvement in perceptually accurate metrics. When both the CLIP-LLA loss and the PRS module are integrated into the final model, we observe a considerable performance gain, in line with their complementary effects.

5. Discussion and Conclusions

In this paper, we propose a methodology that integrates the rich visual–semantic representations of the image–text foundation model CLIP into the training scheme of Retinexformer, while enhancing robustness through a multi-scale dataset augmentation strategy. By introducing CLIP-based perceptual supervision, the model effectively compensates for the information deficiency inherent in pixel-wise losses, which often approximate flat regions and fail to reconstruct fine details. Furthermore, the inherently ill-posed global tone prediction is guided more accurately through CLIP’s high-level semantic alignment between visual and linguistic domains. Complementarily, the proposed multi-scale training scheme prevents overfitting to the fixed-resolution noise characteristics of the dataset. The corresponding results show improved generalization and perceptual coherence. As a result, the proposed approach achieves substantial gains in both global tone alignment and fine-detail restoration, consistently outperforming the baseline across seven benchmark datasets.
Currently, state-of-the-art (SotA) methods in image restoration, particularly in low-light image enhancement (LLIE), are primarily evaluated using pixel-wise comparison metrics such as PSNR and SSIM. When models are trained with an L-distance loss against the ground truth, they are effectively optimized directly for these metrics, which often conflicts with perceptually meaningful fidelity [53]. This trade-off on restoration models has been thoroughly analyzed and empirically demonstrated in prior literature [66,67]. Aside from LLIE models pursuing higher PSNR, the PSNR metric itself is known to be overly sensitive to minor, perceptually negligible details, such as global tone shifts or slight brightness inconsistencies. Future models are therefore expected to avoid overfitting to such metrics and instead aim for a more balanced optimization that aligns with human perception.
It should be noted that the LLIE benchmarks typically span fewer than 500 scenes, which stands in contrast to the scalability that transformer-based architectures are designed to leverage. This limitation motivates the introduction of LLIE networks that achieve competitive performance through more efficient architectures and training schemes. In theory, the proposed CLIP-LLA and PRS modules are not confined to Retinexformer, which opens its applicability to general end-to-end LLIE networks, potentially enhancing both robustness and perceptual quality.

Funding

This research was supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the MSIP (No. RS-2025-00520207); and by an IITP grant funded by the Korea government (MSIT) and a KEIT grant funded by the Korea government (MOTIE) (No. 2022-0-01045).

Data Availability Statement

The datasets used in this study are publicly available. Specifically, the LOL dataset [14], LOL-v2 dataset [49], SID dataset [50], SMID dataset [51], and SDSD dataset [52] were utilized for experimental validation. No new data were created or analyzed in this study. All datasets cited above are publicly accessible and can be obtained from the corresponding references.

Conflicts of Interest

The author declares no conflicts of interest.

Appendix A. Detail on Network Framework and Algorithm

The detailed architecture of the proposed framework is displayed on Figure A1. Based on the baseline RetinexFormer architecture, the network consists of an Illuumination estimator, and multiple stages of an IGA (Illumination-Guided Attention) block to predict the residual of the enhanced image. Multiple IGA blocks form an encoder–bottleneck–decoder structure that downsamples the feature maps to one-quarter resolution before upsampling them again. Downsampling is performed using stride-2 4 × 4 convolution layers, while upsampling employs 2 × 2 stride-2 transposed convolutions followed by a 1 × 1 convolution for concatenating the encoded features.
Figure A1. The detailed architecture the proposed framework utilizes. Please refer to Figure 3 and Section 3.3 for details on pixel-reordering subsampling (PRS).
Figure A1. The detailed architecture the proposed framework utilizes. Please refer to Figure 3 and Section 3.3 for details on pixel-reordering subsampling (PRS).
Electronics 14 04839 g0a1
Compared to a standard transformer block followed by a feed-forward network, IGAB implements following modifications to suit the LLIE task:
  • Utilization of illumination features. IG-MSA multiplies the value vectors by the illumination map from the estimator before attention; standard blocks operate solely on the token features.
  • Positional branch. IGAB adds a depthwise 3 × 3 convolutional positional stream to the attention output; transformers rely on fixed or learned positional embeddings without extra conv branches.
  • Channel-wise FFN. IGAB’s feed-forward path is all convolutional ( 1 × 1 depthwise 3 × 3 1 × 1 with GELU); transformers use linear layers.
  • PreNorm placement. IGAB wraps only the FFN with PreNorm; attention path skips LayerNorm to keep illumination modulation raw, whereas standard blocks apply LayerNorm before both attention and FFN.
  • Residual flow. IGAB applies residual adds after both attention + positional fusion and the conv MLP; transformers use two residuals but without the illumination gating step.

Appendix B. Additional Qualitative Results

We post additional comparisons on different benchmark datasets in Figure A2Figure A4.
Figure A2. Qualitative results on RAW dataset SDSD-outdoor. Proposed method preserves texture and prevents over-flattening, especially on the zoomed-in section. It also offers the closest match to the ground truth in terms of global brightness and tonal appearance.
Figure A2. Qualitative results on RAW dataset SDSD-outdoor. Proposed method preserves texture and prevents over-flattening, especially on the zoomed-in section. It also offers the closest match to the ground truth in terms of global brightness and tonal appearance.
Electronics 14 04839 g0a2
Figure A3. Qualitative results on RAW dataset SDSD-indoor. The baseline RetinexFormer exhibits flickering blotchy artifacts on the white wall, and produces faint white halos around sharp edges in the zoomed-in region. In contrast, our method recovers the scene more faithfully, without such artifacts.
Figure A3. Qualitative results on RAW dataset SDSD-indoor. The baseline RetinexFormer exhibits flickering blotchy artifacts on the white wall, and produces faint white halos around sharp edges in the zoomed-in region. In contrast, our method recovers the scene more faithfully, without such artifacts.
Electronics 14 04839 g0a3aElectronics 14 04839 g0a3b
Figure A4. Qualitative results on real-world indoor dataset LOLv1. Considering the bright yellow tone of the plates, Ours—Mixed recovers the closest brightness and color, while perceptually outperforming other methods in overall noise suppression.
Figure A4. Qualitative results on real-world indoor dataset LOLv1. Considering the bright yellow tone of the plates, Ours—Mixed recovers the closest brightness and color, while perceptually outperforming other methods in overall noise suppression.
Electronics 14 04839 g0a4

Appendix C. Influence of Loss Weight λ

The influence of the CLIP-LLA loss weight λ , used during training, is summarized in Table A1. Overall, larger values of λ tend to improve perceptual metrics, whereas excessively large values (e.g., λ = 4 ) lead to degradation across metrics. We observe that λ = 1 provides a favorable balance, maintaining both pixel-wise accuracy and perceptual similarity. Although the optimal value of λ may vary depending on the dataset characteristics, λ = 1 consistently produced a fine trade-off in our experiments. All qualitative and quantitative results presented in the main paper were obtained using this setting.
Table A1. Effect of CLIP-LLA loss weight λ on each performance metrics, conducted with LOL-v2-synthetic dataset. The best result in each row is bolded.
Table A1. Effect of CLIP-LLA loss weight λ on each performance metrics, conducted with LOL-v2-synthetic dataset. The best result in each row is bolded.
λ 0.250.5124
PSNR↑25.4425.2425.9925.7825.68
SSIM↑0.93180.93320.95560.93350.9324
LPIPS↓0.05670.05600.05460.05320.0544
DISTS↓0.06470.06440.06390.06270.0645

Appendix D. Detailed Computational Efficiency

For comparison, selected representative networks with comparable performances have been displayed in the qualitative comparison. Table A2 presents their computational efficiency measured under the same hardware setup. All experiments were performed on a single NVIDIA A6000 GPU with 600 × 400 RGB inputs, matching the configuration of LOL-v1 and LOL-v2-real. Relative to RetinexFormer, the proposed approach delivers improved performance without adding any extra inference time and computational burden.
Table A2. Comparison of computational efficiency among different LLIE models. The best result in each row is bolded.
Table A2. Comparison of computational efficiency among different LLIE models. The best result in each row is bolded.
ModelInference Time (s)FLOPs (G)Memory (MB)#Params (M)
MIRNet0.5178515431.8
KinD0.5335.01548.02
SNR-Net0.1126.41494.01
RetinexFormer0.3315.65721.61
Ours0.3415.65721.61

References

  1. Loh, Y.P.; Chan, C.S. Getting to know low-light images with the exclusively dark dataset. Comput. Vis. Image Underst. 2019, 178, 30–42. [Google Scholar] [CrossRef]
  2. Jia, X.; Zhu, C.; Li, M.; Tang, W.; Zhou, W. LLVIP: A visible-infrared paired dataset for low-light vision. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 3496–3504. [Google Scholar]
  3. Li, J.; Li, B.; Tu, Z.; Liu, X.; Guo, Q.; Juefei-Xu, F.; Xu, R.; Yu, H. Light the night: A multi-condition diffusion framework for unpaired low-light enhancement in autonomous driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024; pp. 15205–15215. [Google Scholar]
  4. Xian, X.; Zhou, Q.; Qin, J.; Yang, X.; Tian, Y.; Shi, Y.; Tian, D. CROSE: Low-light enhancement by CROss-SEnsor interaction for nighttime driving scenes. Expert Syst. Appl. 2024, 248, 123470. [Google Scholar] [CrossRef]
  5. Zhao, X.; Huang, L.; Li, M.; Han, C.; Nie, T. Atmospheric Scattering Model and Non-Uniform Illumination Compensation for Low-Light Remote Sensing Image Enhancement. Remote Sens. 2025, 17, 2069. [Google Scholar] [CrossRef]
  6. Wu, J.; Ai, H.; Zhou, P.; Wang, H.; Zhang, H.; Zhang, G.; Chen, W. Low-Light Image Dehazing and Enhancement via Multi-Feature Domain Fusion. Remote Sens. 2025, 17, 2944. [Google Scholar] [CrossRef]
  7. Pizer, S.M.; Amburn, E.P.; Austin, J.D.; Cromartie, R.; Geselowitz, A.; Greer, T.; ter Haar Romeny, B.; Zimmerman, J.B.; Zuiderveld, K. Adaptive histogram equalization and its variations. Comput. Vis. Graph. Image Process. 1987, 39, 355–368. [Google Scholar] [CrossRef]
  8. Zeng, H.; Cai, J.; Li, L.; Cao, Z.; Zhang, L. Learning image-adaptive 3d lookup tables for high performance photo enhancement in real-time. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 2058–2073. [Google Scholar] [CrossRef]
  9. Lore, K.G.; Akintayo, A.; Sarkar, S. LLNet: A deep autoencoder approach to natural low-light image enhancement. Pattern Recognit. 2017, 61, 650–662. [Google Scholar] [CrossRef]
  10. Lv, F.; Lu, F.; Wu, J.; Lim, C. MBLLEN: Low-light image/video enhancement using cnns. In Proceedings of the BMVC, Newcastle upon Tyne, UK, 3–6 September 2018; Volume 220, p. 4. [Google Scholar]
  11. Land, E.H.; McCann, J.J. Lightness and retinex theory. J. Opt. Soc. Am. 1971, 61, 1–11. [Google Scholar] [CrossRef]
  12. Guo, X.; Li, Y.; Ling, H. LIME: Low-light image enhancement via illumination map estimation. IEEE Trans. Image Process. 2016, 26, 982–993. [Google Scholar] [CrossRef]
  13. Li, M.; Liu, J.; Yang, W.; Sun, X.; Guo, Z. Structure-Revealing Low-Light Image Enhancement Via Robust Retinex Model. IEEE Trans. Image Process. 2018, 27, 2828–2841. [Google Scholar] [CrossRef]
  14. Wei, C.; Wang, W.; Yang, W.; Liu, J. Deep retinex decomposition for low-light enhancement. arXiv 2018, arXiv:1808.04560. [Google Scholar] [CrossRef]
  15. Zhang, Y.; Zhang, J.; Guo, X. Kindling the darkness: A practical low-light image enhancer. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 1632–1640. [Google Scholar]
  16. Cai, Y.; Bian, H.; Lin, J.; Wang, H.; Timofte, R.; Zhang, Y. Retinexformer: One-stage retinex-based transformer for low-light image enhancement. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 12504–12513. [Google Scholar]
  17. Yi, X.; Xu, H.; Zhang, H.; Tang, L.; Ma, J. Diff-retinex: Rethinking low-light image enhancement with a generative diffusion model. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 12302–12311. [Google Scholar]
  18. Wang, T.; Zhang, K.; Zhang, Y.; Luo, W.; Stenger, B.; Lu, T.; Kim, T.K.; Liu, W. LLDiffusion: Learning degradation representations in diffusion models for low-light image enhancement. Pattern Recognit. 2025, 166, 111628. [Google Scholar] [CrossRef]
  19. Zhou, D.; Yang, Z.; Yang, Y. Pyramid diffusion models for low-light image enhancement. arXiv 2023, arXiv:2305.10028. [Google Scholar] [CrossRef]
  20. Jiang, H.; Luo, A.; Fan, H.; Han, S.; Liu, S. Low-light image enhancement with wavelet-based diffusion models. ACM Trans. Graph. 2023, 42, 1–14. [Google Scholar] [CrossRef]
  21. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning, Virtual Event, 18–24 July 2021; pp. 8748–8763. [Google Scholar]
  22. Poynton, C. Digital Video and HD: Algorithms and Interfaces; Elsevier: Amsterdam, The Netherlands, 2012. [Google Scholar]
  23. Jobson, D.J.; Rahman, Z.u.; Woodell, G.A. A multiscale retinex for bridging the gap between color images and the human observation of scenes. IEEE Trans. Image Process. 1997, 6, 965–976. [Google Scholar] [CrossRef] [PubMed]
  24. Wang, R.; Zhang, Q.; Fu, C.W.; Shen, X.; Zheng, W.S.; Jia, J. Underexposed photo enhancement using deep illumination estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 6849–6857. [Google Scholar]
  25. Jiang, Y.; Gong, X.; Liu, D.; Cheng, Y.; Fang, C.; Shen, X.; Yang, J.; Zhou, P.; Wang, Z. Enlightengan: Deep light enhancement without paired supervision. IEEE Trans. Image Process. 2021, 30, 2340–2349. [Google Scholar] [CrossRef] [PubMed]
  26. Guo, C.; Li, C.; Guo, J.; Loy, C.C.; Hou, J.; Kwong, S.; Cong, R. Zero-reference deep curve estimation for low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual Event, 14–19 June 2020; pp. 1780–1789. [Google Scholar]
  27. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  28. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  29. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual Event, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  30. Wang, Z.; Cun, X.; Bao, J.; Zhou, W.; Liu, J.; Li, H. Uformer: A general u-shaped transformer for image restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 17683–17693. [Google Scholar]
  31. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.H. Restormer: Efficient transformer for high-resolution image restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 5728–5739. [Google Scholar]
  32. Wang, T.; Zhang, K.; Shen, T.; Luo, W.; Stenger, B.; Lu, T. Ultra-high-definition low-light image enhancement: A benchmark and transformer-based method. AAAI Conf. Artif. Intell. 2023, 37, 2654–2662. [Google Scholar] [CrossRef]
  33. Liu, S.; Zhang, H.; Li, X.; Yang, X. Retinexformer+: Retinex-Based Dual-Channel Transformer for Low-Light Image Enhancement. Comput. Mater. Contin. 2025, 82, 1969–1984. [Google Scholar] [CrossRef]
  34. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Proceedings of the International Workshop on Deep Learning in Medical Image Analysis, Granada, Spain, 20 September 2018; Springer: Cham, Switzerland, 2018; pp. 3–11. [Google Scholar]
  35. Bai, J.; Yin, Y.; He, Q.; Li, Y.; Zhang, X. Retinexmamba: Retinex-based mamba for low-light image enhancement. In Proceedings of the International Conference on Neural Information Processing, Auckland, New Zealand, 2–6 December 2024; pp. 427–442. [Google Scholar]
  36. Gu, A.; Goel, K.; Ré, C. Efficiently modeling long sequences with structured state spaces. arXiv 2021, arXiv:2111.00396. [Google Scholar]
  37. Gu, A.; Dao, T. Mamba: Linear-time sequence modeling with selective state spaces. In Proceedings of the First Conference on Language Modeling, Philadelphia, PA, USA, 7–9 October 2024. [Google Scholar]
  38. Wang, H.; Liang, X.; Han, J.; Geng, W. Spike-RetinexFormer: Rethinking Low-light Image Enhancement with Spiking Neural Networks. In Proceedings of the Thirty-Ninth Annual Conference on Neural Information Processing Systems, San Diego, CA, USA, 2–7 December 2025. [Google Scholar]
  39. Tavanaei, A.; Ghodrati, M.; Kheradpisheh, S.R.; Masquelier, T.; Maida, A. Deep learning in spiking neural networks. Neural Netw. 2019, 111, 47–63. [Google Scholar] [CrossRef]
  40. Uddin, S.; Hussain, B.; Fareed, S.; Arif, A.; Ali, B. Real-world adaptation of retinexformer for low-light image enhancement using unpaired data. Int. J. Ethical AI Appl. 2025, 1, 1–6. [Google Scholar] [CrossRef]
  41. Xu, S.; Chen, Q.; Hu, H.; Peng, L.; Tao, W. An unsupervised fine-tuning strategy for low-light image enhancement. J. Vis. Commun. Image Represent. 2025, 110, 104480. [Google Scholar] [CrossRef]
  42. Feijoo, D.; Benito, J.C.; Garcia, A.; Conde, M.V. Darkir: Robust low-light image restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 11–15 June 2025; pp. 10879–10889. [Google Scholar]
  43. Xu, R.; Niu, Y.; Li, Y.; Xu, H.; Liu, W.; Chen, Y. URWKV: Unified RWKV Model with Multi-state Perspective for Low-light Image Restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 11–15 June 2025; pp. 21267–21276. [Google Scholar]
  44. Wang, J.; Chan, K.C.; Loy, C.C. Exploring clip for assessing the look and feel of images. AAAI Conf. Artif. Intell. 2023, 37, 2555–2563. [Google Scholar] [CrossRef]
  45. Baek, K.; Lee, M.; Shim, H. Psynet: Self-supervised approach to object localization using point symmetric transformation. AAAI Conf. Artif. Intell. 2020, 34, 10451–10459. [Google Scholar] [CrossRef]
  46. Huang, T.; Li, S.; Jia, X.; Lu, H.; Liu, J. Neighbor2neighbor: Self-supervised denoising from single noisy images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual Event, 19–25 June 2021; pp. 14781–14790. [Google Scholar]
  47. Mansour, Y.; Heckel, R. Zero-shot noise2noise: Efficient image denoising without any data. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 14018–14027. [Google Scholar]
  48. Yun, S.; Han, D.; Oh, S.J.; Chun, S.; Choe, J.; Yoo, Y. Cutmix: Regularization strategy to train strong classifiers with localizable features. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6023–6032. [Google Scholar]
  49. Yang, W.; Wang, W.; Huang, H.; Wang, S.; Liu, J. Sparse gradient regularized deep retinex network for robust low-light image enhancement. IEEE Trans. Image Process. 2021, 30, 2072–2086. [Google Scholar] [CrossRef]
  50. Chen, C.; Chen, Q.; Do, M.N.; Koltun, V. Seeing motion in the dark. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3185–3194. [Google Scholar]
  51. Chen, C.; Chen, Q.; Xu, J.; Koltun, V. Learning to see in the dark. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 3291–3300. [Google Scholar]
  52. Wang, R.; Xu, X.; Fu, C.W.; Lu, J.; Yu, B.; Jia, J. Seeing dynamic scene in the dark: A high-quality video dataset with mechatronic alignment. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Online, 11–17 October 2021; pp. 9700–9709. [Google Scholar]
  53. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
  54. Pambrun, J.F.; Noumeir, R. Limitations of the SSIM quality metric in the context of diagnostic imaging. In Proceedings of the 2015 IEEE International Conference on Image Processing (ICIP), Quebec City, QC, Canada, 27–30 September 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 2960–2963. [Google Scholar]
  55. Nilsson, J.; Akenine-Möller, T. Understanding ssim. arXiv 2020, arXiv:2006.13846. [Google Scholar] [CrossRef]
  56. Zhang, R.; Isola, P.; Efros, A.A.; Shechtman, E.; Wang, O. The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 586–595. [Google Scholar]
  57. Ding, K.; Ma, K.; Wang, S.; Simoncelli, E.P. Image quality assessment: Unifying structure and texture similarity. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 2567–2581. [Google Scholar] [CrossRef]
  58. Yang, W.; Wang, S.; Fang, Y.; Wang, Y.; Liu, J. Band representation-based semi-supervised low-light image enhancement: Bridging the gap between signal fidelity and perceptual quality. IEEE Trans. Image Process. 2021, 30, 3461–3473. [Google Scholar] [CrossRef] [PubMed]
  59. Kosugi, S.; Yamasaki, T. Unpaired image enhancement featuring reinforcement-learning-controlled image editing software. AAAI Conf. Artif. Intell. 2020, 34, 11296–11303. [Google Scholar] [CrossRef]
  60. Moran, S.; Marza, P.; McDonagh, S.; Parisot, S.; Slabaugh, G. Deeplpf: Deep local parametric filters for image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Online, 14–19 June 2020; pp. 12826–12835. [Google Scholar]
  61. Chen, H.; Wang, Y.; Guo, T.; Xu, C.; Deng, Y.; Liu, Z.; Ma, S.; Xu, C.; Xu, C.; Gao, W. Pre-trained image processing transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Online, 19–25 June 2021; pp. 12299–12310. [Google Scholar]
  62. Liu, R.; Ma, L.; Zhang, J.; Fan, X.; Luo, Z. Retinex-inspired unrolling with cooperative prior architecture search for low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual Event, 19–25 June 2021; pp. 11197–11206. [Google Scholar]
  63. Xu, K.; Yang, X.; Yin, B.; Lau, R.W. Learning to restore low-light images via decomposition-and-enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Online, 14–19 June 2020; pp. 2281–2290. [Google Scholar]
  64. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.H.; Shao, L. Learning enriched features for real image restoration and enhancement. In Proceedings of the European Conference on Computer Vision, Online, 23–28 August 2020; pp. 492–511. [Google Scholar]
  65. Xu, X.; Wang, R.; Fu, C.W.; Jia, J. Snr-aware low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 17714–17724. [Google Scholar]
  66. Blau, Y.; Michaeli, T. The perception-distortion tradeoff. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6228–6237. [Google Scholar]
  67. Cohen, R.; Kligvasser, I.; Rivlin, E.; Freedman, D. Looks too good to be true: An information-theoretic analysis of hallucinations in generative restoration models. Adv. Neural Inf. Process. Syst. 2024, 37, 22596–22623. [Google Scholar]
Figure 1. Based on the end-to-end LLIE framework, we propose (i) multi-scale training based on a pixel-reordering subsampler, and also a guidance of training procedure by exploiting CLIP’s rich features in both (ii) textual and (iii) visual spaces.
Figure 1. Based on the end-to-end LLIE framework, we propose (i) multi-scale training based on a pixel-reordering subsampler, and also a guidance of training procedure by exploiting CLIP’s rich features in both (ii) textual and (iii) visual spaces.
Electronics 14 04839 g001
Figure 2. Validation set performance changes of the original Retinexformer on the LOLv2-Real dataset, in terms of evaluation metrics during training. The upper arrow indicates that a higher value means better performance, and vice versa.
Figure 2. Validation set performance changes of the original Retinexformer on the LOLv2-Real dataset, in terms of evaluation metrics during training. The upper arrow indicates that a higher value means better performance, and vice versa.
Electronics 14 04839 g002
Figure 3. Pixel-reordering subsampling (PRS) for efficient multi-scale training. Pixels of both the input and target images are rearranged in a 4-color-checkerboard-like manner to form 4 half-scale images. The numbers are arbitrarily assigned to denote the pixel at the corresponding location. Best viewed in color.
Figure 3. Pixel-reordering subsampling (PRS) for efficient multi-scale training. Pixels of both the input and target images are rearranged in a 4-color-checkerboard-like manner to form 4 half-scale images. The numbers are arbitrarily assigned to denote the pixel at the corresponding location. Best viewed in color.
Electronics 14 04839 g003
Figure 4. Qualitative results on real-world dataset LOLv2-real. The proposed method most appropriately recovers the global tone, and detailed texture.
Figure 4. Qualitative results on real-world dataset LOLv2-real. The proposed method most appropriately recovers the global tone, and detailed texture.
Electronics 14 04839 g004
Figure 5. Qualitative results on sRGB dataset LOLv2-synthetic. Proposed method accurately matches the color and sharpness of the target.
Figure 5. Qualitative results on sRGB dataset LOLv2-synthetic. Proposed method accurately matches the color and sharpness of the target.
Electronics 14 04839 g005
Figure 6. Qualitative results on RAW dataset SMID. The proposed method most accurately recovers the linear pattern in the dark area.
Figure 6. Qualitative results on RAW dataset SMID. The proposed method most accurately recovers the linear pattern in the dark area.
Electronics 14 04839 g006
Table 1. Quantitative comparison results on the LOL, SID, SMID, and SDSD datasets and their variations. Best results for each column are notated in bold.
Table 1. Quantitative comparison results on the LOL, SID, SMID, and SDSD datasets and their variations. Best results for each column are notated in bold.
MethodComplexityLOL-v1LOL-v2-realLOL-v2-synSIDSMIDSDSD-inSDSD-out
FLOPSParamsPSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIM
SID [50]13.737.7614.350.43613.240.44215.040.61016.970.59124.780.71823.290.70324.900.693
3DLUT [8]0.0750.5914.350.44517.590.72118.040.80020.110.59223.860.67821.660.65521.890.649
DeepUPE [24]21.11.0214.380.44613.270.45215.080.62317.010.60423.910.69021.700.66221.940.698
RF [59]46.2321.5415.230.45214.050.45815.970.63216.440.59623.110.68120.970.65521.210.689
DeepLPF [60]5.861.7715.280.47314.100.48016.020.58718.070.60024.360.68822.210.66422.760.658
IPT [61]6887115.3116.270.50419.800.81318.300.81120.530.56127.030.78326.110.83127.550.850
UFormer [30]125.2916.360.77118.820.77119.660.87118.540.57727.200.79223.170.85923.850.748
RetinexNet [14]587.470.8416.770.56015.470.56717.130.79816.480.57822.830.68420.840.61720.960.629
SparseRN [49]53.262.3317.200.64020.060.81622.050.90518.680.60625.480.76623.250.86325.280.804
EnGAN [25]61.01114.3517.480.65018.230.61716.570.73417.230.54322.620.67420.020.60420.100.616
RUAS [62]0.830.00318.230.72018.370.72316.550.65218.440.58125.880.74423.170.69623.840.743
FIDE [63]28.518.6218.270.66516.850.67815.200.61218.340.57824.420.69222.410.65922.200.629
DRBN [58]48.615.2720.130.83020.290.83123.220.92719.020.57726.600.78124.080.86825.770.841
KinD [15]34.998.0220.860.79014.740.64113.290.57818.020.58322.180.63421.950.67221.970.654
Restormer [31]144.2526.1322.430.82319.940.82721.410.83022.270.64926.970.75825.670.82724.790.802
MIRNet [64]78531.7624.140.83020.020.82021.940.87620.840.60525.660.76224.380.86427.130.837
SNR-Net [65]26.354.0124.610.84221.480.84924.140.92822.870.62528.490.80529.440.89428.660.866
RetinexFormer [16]15.571.6122.780.88320.060.86225.610.95524.380.67729.120.81327.380.88628.540.864
Ours—Textual15.571.6122.710.87620.980.85124.700.91924.120.66828.600.80227.840.87628.450.856
Ours—Visual 23.470.88320.230.86425.640.95424.220.67031.340.83228.080.88628.210.862
Ours—Mixed 22.660.88220.230.84425.990.95624.490.67531.850.83426.650.88128.870.867
Table 2. Perceptual evaluation results (LPIPS [56] and DISTS [57]) comparing our proposed methods against the RetinexFormer baseline. Note that lower values indicate better performance for both LPIPS and DISTS. The best result in each column is bolded.
Table 2. Perceptual evaluation results (LPIPS [56] and DISTS [57]) comparing our proposed methods against the RetinexFormer baseline. Note that lower values indicate better performance for both LPIPS and DISTS. The best result in each column is bolded.
MethodLOL-v1LOL-v2-realLOL-v2-synSIDSMIDSDSD-inSDSD-out
LPIPSDISTSLPIPSDISTSLPIPSDISTSLPIPSDISTSLPIPSDISTSLPIPSDISTSLPIPSDISTS
RetinexFormer0.1450.1410.1680.1480.0590.0670.3570.2090.1640.1330.1370.1230.1670.117
Ours—Textual0.1600.1390.1950.1560.1240.1340.3580.2100.1790.1360.1520.1200.1650.106
Ours—Visual0.1490.1420.1630.1420.0570.0670.3520.2110.1570.1290.1330.1210.1760.120
Ours—Mixed0.1400.1370.1640.1420.0550.0640.3550.2100.1540.1310.1290.1190.1640.118
Table 3. Ablation study of proposed method on LOLv2-syn dataset. Mixed variation is used for the proposed CLIP-LLA loss. The best result in each column is bolded.
Table 3. Ablation study of proposed method on LOLv2-syn dataset. Mixed variation is used for the proposed CLIP-LLA loss. The best result in each column is bolded.
ModelPSNR (dB)↑SSIM↑LPIPS↓DISTS↓
Baseline25.610.95450.05940.0673
w/o PRS25.300.93130.06030.0671
w/o CL25.100.93080.05840.0656
Ours—Mixed25.990.95560.05460.0639
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Song, S. Noise-Resilient Low-Light Image Enhancement with CLIP Guidance and Pixel-Reordering Subsampling. Electronics 2025, 14, 4839. https://doi.org/10.3390/electronics14244839

AMA Style

Song S. Noise-Resilient Low-Light Image Enhancement with CLIP Guidance and Pixel-Reordering Subsampling. Electronics. 2025; 14(24):4839. https://doi.org/10.3390/electronics14244839

Chicago/Turabian Style

Song, Seongjong. 2025. "Noise-Resilient Low-Light Image Enhancement with CLIP Guidance and Pixel-Reordering Subsampling" Electronics 14, no. 24: 4839. https://doi.org/10.3390/electronics14244839

APA Style

Song, S. (2025). Noise-Resilient Low-Light Image Enhancement with CLIP Guidance and Pixel-Reordering Subsampling. Electronics, 14(24), 4839. https://doi.org/10.3390/electronics14244839

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop