You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

27 November 2025

Boosting Diffusion Networks with Deep External Context-Aware Encoders for Low-Light Image Enhancement

,
and
School of Artificial Intelligence, Beijing University of Posts and Telecommunications, Beijing 100876, China
*
Author to whom correspondence should be addressed.
This article belongs to the Section Sensing and Imaging

Abstract

Low-light image enhancement (LLIE) requires modeling spatially extensive and interdependent degradations across large pixel regions, while directly equipping diffusion-based LLIE with heavy global modules inside the iterative denoising backbone leads to prohibitive computational overhead. To enhance long-range context modeling without inflating the per-step cost of diffusion, we propose ECA-Diff, a diffusion framework augmented with a deep External Context-Aware Encoder (ECAE). A latent-space context network built with hybrid Transformer–Convolution blocks extracts holistic cues from the input, generates multi-scale context features once, and injects them into the diffusion backbone as lightweight conditional guidance across all sampling steps. In addition, a CIELAB-space Luminance-Adaptive Chromaticity Loss regularizes conditional diffusion training and mitigates the cool color cast frequently observed in low-luminance regions. Experiments on paired and unpaired benchmarks show that ECA-Diff consistently outperforms recent state-of-the-art LLIE methods in both full-reference (PSNR/SSIM/LPIPS) and no-reference (NIQE/BRISQUE) metrics, with the external context path introducing only modest overhead relative to the baseline diffusion backbone. These results indicate that decoupling global context estimation from the iterative denoising process is an effective way to boost diffusion-based LLIE and provides a general compute-once conditioning paradigm for low-level image restoration.

1. Introduction

Images captured under low-light conditions often suffer from low signal-to-noise ratios (SNRs) and poor visibility because of the limited number of photons reaching the image sensor. These degradations compromise the usability of low-light images in both individual and industrial contexts, and constrain the performance of high-level vision-based downstream tasks. Although physics-based solutions can alleviate these issues to some extent, they suffer from inherent limitations: long exposure requires static scenes, high ISO induces severe noise, and high-end sensors are costly. As a result, algorithmic approaches to addressing the task of low-light image enhancement (LLIE) have become efficient and essential alternatives.
Due to the ill-posed nature of degraded low-light images, accurately restoring color and structural details remains a long-standing challenge. Over the years, a wide range of LLIE methods have been developed. Traditional approaches, such as histogram equalization (HE) [1,2] and Retinex-based methods [3,4], primarily focus on adjusting brightness, tone, or contrast to align with human visual perception while neglecting the presence of noise, which makes them ineffective under extremely low illumination or short exposure conditions. The recent comprehensive development of deep learning-based methods has significantly advanced the field of LLIE. Chen et al. [5] propose reconstructing high-quality images from extreme low-light raw-format sensor data with an end-to-end neural network to deal with the full low-light image processing pipeline. Following this work, a variety of works such as [6,7,8,9,10], apply end-to-end deep neural networks to handle LLIE tasks. At the same time, a series of methods combine the design of deep networks with the Retinex theory, e.g., [11,12,13]. Also, many state-of-the-art deep learning-based technologies have been introduced into this field, such as transformer [14,15,16,17], normalizing flow [18] and diffusion models [19,20,21,22,23]. Among these, diffusion models have recently emerged as a promising direction in LLIE due to their superior generative capabilities. The inherent denoising property of diffusion models makes them particularly effective in handling the noise prevalent in LLIE, yielding perceptually pleasing and smooth results that often outperform other learning-based methods.
However, most diffusion-based LLIE methods retain the DDPM [24] convention of a shallow U-Net [25] as the denoising backbone. Such backbones are efficient at estimating manually added white (i.i.d.) Gaussian noise but, due to their limited receptive field, struggle to capture the non-local and interdependent degradations characteristic of low-light scenes. Consistent with this, CNN-based studies [6,9,10] show that long-range, context-aware representations are essential for globally lifting illumination, suppressing widespread noise, recovering locally missing details, and maintaining color consistency. Recent diffusion variants illuminate the trade-off from two directions. PyDiff [19] demonstrates that conditional diffusion without explicit global correction can exhibit color bias; it mitigates this via an auxiliary refinement network executed alongside the denoiser at each sampling step, an approach that ties global correction to the iterative loop and thus scales inference cost with the number of steps. Conversely, CLE [21] injects a global target, namely image-level mean intensity, as exposure guidance, achieving accurate overall brightness, but the absence of explicit long-range modeling limits large-scale color consistency and fine-detail recovery. Taken together, these observations expose a central gap: diffusion-based LLIE needs global context that is both effective and decoupled from per-step inference, enabling large-scale structural and color correction without prohibitive computational overhead.
To this end, we propose Boosting Diffusion Networks with a Deep External Context-Aware Encoder (ECA-Diff) for LLIE, a framework that augments diffusion backbones with an External Context-Aware Encoder (ECAE) to strengthen global context modeling. During inference, ECAE is executed once per image and its features are reused at every sampling step, supplying stable global guidance with negligible per-step overhead. Inspired by Latent Diffusion Models (LDMs) [26], ECAE models global context in a compact latent space to extract information efficiently, and the resulting context-aware features are injected into the diffusion backbone as conditional guidance. Unlike LDMs, our denoising backbone operates in the full-resolution image space to preserve the fidelity required for LLIE. By leveraging the rich low-dimensional representations from ECAE, ECA-Diff broadens the contextual receptive field and extends the semantic capacity of the denoiser backbone, enabling more accurate modeling of long-range structures and more faithful restoration of global color distributions in LLIE tasks.
Specifically, the ECAE in ECA-Diff comprises a pretrained VAE [26] encoder that maps the input image to a compact latent space, followed by a Latent Context Network (LCN) to extract higher-level global context efficiently. Each LCN stage stacks multiple Hybrid Transformer–Convolution (HTC) blocks, in which a Transformer branch and a large-kernel convolutional branch run in parallel and are subsequently fused via Bidirectional Cross-Attention (Bi-CA). This hybrid design in latent space supplies global context-aware guidance while preserving fine-grained spatial cues, without incurring excessive computational cost. For multi-scale guidance, the LCN features are progressively upsampled, temporally modulated by the current diffusion timestep, and concatenated with the backbone features, thereby seamlessly injecting global context into the denoising trajectory.
In addition, we introduce an auxiliary color regularizer, the Luminance-Adaptive Chromaticity Loss (LACL), to curb the cool, low-saturation cast frequently observed in low-luminance regions when training solely with the conditional noise objective and standard RGB-domain losses. LACL is computed in the CIELAB [27] space, leveraging its near-perceptual uniformity to decouple lightness from color, and imposes two simple constraints with a luminance-adaptive weighting: a hue-angle consistency term and a relative anti-desaturation term that penalizes chroma deficits without over-saturating bright areas. This auxiliary loss complements our main objective with negligible overhead and yields more faithful global color reproduction across scenes.
Extensive experiments on both paired and unpaired LLIE benchmarks show that ECA-Diff delivers consistent gains over state-of-the-art methods in quantitative metrics and perceived visual quality. The main contributions of this work are summarized as follows:
  • We propose ECA-Diff, a diffusion-based LLIE framework that augments the diffusion backbone with an External Context-Aware Encoder to enhance long-range modeling. The encoder is executed once per image and reused across all sampling steps, thereby strengthening global color restoration and structural fidelity with negligible per-step overhead.
  • We design an LCN that operates in a compact latent space to extract context-aware features for the diffusion backbone. The LCN comprises HTC blocks in which Transformer and large-kernel CNN branches run in parallel and are fused via bidirectional cross-attention, enabling strong global representation learning while preserving local detail with low computational overhead.
  • We introduce LACL, an auxiliary loss in the CIELAB space that regularizes diffusion training by enforcing hue-angle consistency and penalizing desaturation via luminance-adaptive weighting. It improves global color fidelity and enhances chroma reconstruction in dark regions for LLIE.

3. Methodology

This section details ECA-Diff, an External Context-Aware Encoder (ECAE) guided diffusion framework for low-light image enhancement (LLIE). We begin by reviewing the diffusion formulation, then outline the overall ECA-Diff pipeline to provide a high-level workflow. Next, we describe each component in turn, including the ECAE, the Latent Context Network (LCN), the Hybrid Transformer–Convolution (HTC) block augmented with Bidirectional Cross-Attention (Bi-CA), and the Time-modulated Context Fusion (TCF) block. Subsequently, we introduce the Luminance-Adaptive Chromaticity Loss (LACL) and specify the overall training objective and inference strategy. Finally, we present the datasets and the evaluation protocol used in our experiments.

3.1. Preliminaries: Diffusion Models

Denoising Diffusion Probabilistic Model (DDPM). Diffusion models are a class of generative models that synthesize data by learning to invert a gradual Gaussian noising process, progressively transforming noise into samples from the target distribution. The classic DDPM [24] instantiates this idea with a fixed, typically increasing variance schedule { β t } t = 1 T ( 0 , 1 ) and a noise-prediction network ϵ θ that parameterizes the reverse process. The forward process gradually adds Gaussian noise to a clean image y 0 :
q ( y t y t 1 ) = N y t ; α t y t 1 , ( 1 α t ) I ,
where α t = 1 β t and I is the identity matrix. Let α ¯ t = i = 1 t α i , with α ¯ 0 = 1 ; then the marginal distribution q ( y t y 0 ) can be expressed as
q ( y t y 0 ) = N y t ; α ¯ t y 0 , ( 1 α ¯ t ) I .
The variance schedule { β t } t = 1 T is carefully designed such that α ¯ t 0 as t T , and hence q ( y T ) N ( 0 , I ) .
In the reverse process, a neural network ϵ θ is trained to predict the noise component to progressively convert the standard Gaussian into the target data distribution:
p θ ( y t 1 y t ) = N y t 1 ; μ θ ( y t , t ) , β ˜ t I ,
where
μ θ ( y t , t ) = 1 α t y t β t 1 α ¯ t ϵ θ ( y t , t ) , β ˜ t = 1 α ¯ t 1 1 α ¯ t β t .
During inference, the reverse diffusion process starts by sampling y T N ( 0 , I ) and then iteratively applies the learned denoising transitions to recover a clean image y 0 .
Training objective. Given standard Gaussian noise ϵ N ( 0 , I ) and the noisy sample at time t generated by the forward diffusion, y t = α ¯ t y 0 + 1 α ¯ t ϵ , a noise-prediction network ϵ θ ( y t , t ) is trained to estimate ϵ . The parameters θ are optimized by minimizing a simplified noise-prediction loss:
L ddpm = E y 0 , t , ϵ ϵ ϵ θ y t , t 2 2 .
Denoising Diffusion Implicit Model (DDIM) Sampling. With the non-stochastic ( η = 0 ) setting, where η [ 0 , 1 ] controls the stochasticity of the generalized reverse update, DDIM [46] renders the reverse diffusion deterministic by conditioning the update on a fixed signal trajectory, thereby removing stochastic noise injection and accelerating sampling. The update rule is given by
y t 1 = α ¯ t 1 y t 1 α ¯ t ϵ θ ( y t , t ) α ¯ t + 1 α ¯ t 1 ϵ θ ( y t , t ) .
In this work, we adopt DDIM for inference, while training strictly follows the standard DDPM strategy with the simplified noise-prediction objective in Equation (5).

3.2. External Context-Aware Encoder Guided Diffusion Model

High-fidelity low-light image reconstruction requires preserving fine details throughout the generative process. While latent diffusion models (LDMs) [26] achieve efficiency by performing denoising in a compressed latent space, their autoencoder compression and latent-space processing can discard fine structural details. Therefore, in this work, we employ a diffusion backbone that operates directly in pixel space at the native resolution, thereby maintaining richer visual information for subsequent reconstruction. In addition to preserving fine details, effective LLIE also demands robust modeling of global context to ensure color consistency, global denoising, and coherent illumination restoration. However, achieving this within diffusion models via internal global-context modules typically requires complex components, which substantially increase the per-step computational cost, which accumulates over sampling steps to raise the total inference time. To address these challenges, we propose ECA-Diff, a standard denoising diffusion backbone enhanced with an External Context-Aware Encoder (ECAE). This encoder is run once per image to compute deep multi-scale latent condition features that are subsequently reused across all sampling steps, thereby providing robust global context guidance with minimal per-step overhead.
As illustrated in Figure 1, our system comprises a denoising diffusion backbone and an external encoder ECAE. The backbone follows the standard DDPM formulation [24]; following prior practice [21], we implement a conditional variant in which the U-Net [25] denoiser receives the low-light input x together with the time embedding t and predicts the noise ϵ θ ( y t , t , x ) . For inference, we adopt DDIM [46] (non-stochastic, η = 0) to deterministically update from y T to y 0 . In addition to conditioning on x , ECAE constructs a time-invariant set of multi-scale deep context features { c s } s = 1 S , where S matches the number of backbone resolution scales. These features are computed once per image and cached before sampling begins, and then injected at matching resolutions via TCF blocks. The TCF block comprises a time-modulation function g s ( c s , t ) that synchronizes the injected features with the reverse diffusion process, yielding the conditional denoiser ϵ θ ( y t , t , x , C ) , where C g s ( c s , t ) s = 1 S .
Figure 1. Architecture of ECA-Diff. LCN denotes the Latent Context Network; TCF denotes time-modulated context fusion block. The timestep t is embedded at each backbone stage and jointly used in the TCF modules to modulate the fused features. (a) Denoising Diffusion Network: starting from y T N ( 0 , I ) , each step takes ( y t , x , t ) and the context features c s from ECAE, predicts noise ϵ θ ( y t , t ) , and updates the state to y t 1 . (b) External Context-aware Encoder (ECAE): the input x is encoded once by a pretrained VAE E to produce a latent z; (c) within (b), the LCN (built with HTC blocks) converts z into multi-scale features { c s } s = 1 S , which are cached and reused at every sampling step by the TCFs.
Given the low-light input x , the target normal-light image y 0 , and the cached time-modulated context C , the training objective for noise prediction is
L ddpm cond = E y 0 , t , ϵ ϵ ϵ θ y t , t , x , C 2 2 , y t = α ¯ t y 0 + 1 α ¯ t ϵ , ϵ N ( 0 , I ) .
The DDIM-style conditional reverse update is
y t 1 = α ¯ t 1 y t 1 α ¯ t ϵ θ ( y t , t , x , C ) α ¯ t + 1 α ¯ t 1 ϵ θ ( y t , t , x , C ) .

3.3. External Context-Aware Encoder

The ECAE is designed to externally enhance the backbone diffusion model by supplying deep context-aware representations, thereby mitigating color shifts and improving overall LLIE performance. As shown in Figure 1b, the ECAE consists of three components: (i) a pretrained latent encoder that maps the input x into a compact latent code z; (ii) a lightweight LCN composed of HTC blocks that extract both long-range and local cues from z and transform them into multi-scale context-aware features { c s } s = 1 S ; and (iii) a time-modulated fusion module, TCF, that injects these features into the backbone denoiser.

3.3.1. Pre-Trained Latent Encoder

Although a latent-space diffusion scheme is not adopted in the ECA-Diff backbone to avoid potential information loss, the ECAE is specifically designed for global context modeling. In this setting, employing the pretrained VAE encoder from LDM [26] provides a lightweight and computationally efficient solution for information integration and dimensionality reduction. Benefiting from large-scale pretraining (e.g., on ImageNet [47]), this encoder produces semantically rich latent features without incurring substantial additional training cost. Architecturally, the VAE encoder in LDM comprises a convolutional downsampling path augmented with residual blocks and a self-attention layer at the bottleneck. This design helps preserve essential local structures while capturing a degree of global dependencies, albeit with some inevitable loss of fine-grained details. Given a low-light image x , the encoder outputs a latent feature map formulated as
z = E ( x ) ,
which matches the spatial resolution of the backbone denoiser’s bottleneck and has four channels. The latent code z thus provides a compact representation that facilitates deeper context modeling in the subsequent LCN module.
Note that the pretrained VAE encoder is frozen during the early stages of ECA-Diff training to fully exploit the priors learned from large-scale pretraining; we subsequently unfreeze it and fine-tune with a smaller learning rate to better adapt to the target dataset and the backbone network.

3.3.2. Latent Context Network

The latent representation z produced by the preceding VAE encoder provides an initial feature map at low spatial resolution (one-eighth of the input size) with four channels, and remains compatible with the ECA-Diff backbone bottleneck. To construct deeper and more comprehensive global context, we introduce the Latent Context Network (LCN), which employs a U-Net [25] with two levels of downsampling and upsampling, augmented with skip connections to preserve spatial details, as illustrated in Figure 1c. Each stage of the LCN is built upon multiple HTC blocks, which serve as the primary computational units for extracting rich global context features (detailed in the next subsection).
To effectively leverage the deep context features extracted by the LCN for guiding the backbone in predicting the noise ϵ θ for a given low-light input x , we generate and inject multi-scale context features into the corresponding stages of the denoiser. Specifically, the LCN first produces a stage-1 context feature map c 1 , formulated as
c 1 = LCN ( z ) .
Subsequently, a series of transposed convolutions with a 2 × upsampling factor are applied to generate multi-scale context features { c s } s = 1 S that are spatially aligned with the backbone stages.

3.3.3. Hybrid Transformer–Convolution Block

The proposed Hybrid Transformer–Convolution (HTC) block serves as the core computational unit of each LCN stage, designed to integrate the complementary strengths of convolutional and transformer-based representations. Operating in the low-resolution, low-channel latent space generated by the VAE encoder, HTC blocks efficiently capture both long-range dependencies and fine-grained local patterns by combining large-kernel convolutions with pixel-level self-attention.
As illustrated in Figure 2a, the HTC adopts a parallel architecture consisting of a convolution branch and a transformer branch. The convolution branch comprises two residual blocks with relatively large kernels, thereby expanding the receptive field while preserving local detail fidelity. The transformer branch includes two self-attention blocks that specialize in modeling global dependencies across spatial positions. To promote mutual feature refinement between the two branches, Bidirectional Cross-Attention (Bi-CA) is applied: convolution-branch queries are refined using transformer-derived keys and values, while transformer-branch queries are refined using CNN-derived keys and values. This reciprocal interaction allows each branch to benefit from the complementary information of the other, leading to richer and more context-aware representations. Compared with naive element-wise addition or channel concatenation, which mix features in a fixed manner, Bi-CA performs content-adaptive, spatially varying fusion so that each location selectively aggregates the most informative responses from the complementary branch. After the cross-attention step, the enhanced convolutional and transformer features are concatenated along the channel dimension, fused by a 1 × 1 convolution, and added to the original input through a residual connection. This fusion strategy consolidates multi-scale contextual cues while stabilizing training by preserving the original signal pathway. In subsequent stages of the LCN, stacked HTC blocks progressively enrich the latent representation, providing robust and semantically meaningful context features to guide the ECA-Diff backbone.
Figure 2. HTC and TCF blocks. (a) HTC: A Hybrid Transformer–Convolution block with bidirectional cross-attention (Bi-CA). Q/KV box colors match their source branch. Queries Q from one branch attend to key–value pairs ( K , V ) from the other branch; the two branches exchange attention, and their outputs are fused via concatenation followed by a 1 × 1 convolution. (b) TCF: time-modulated context fusion. The stage-s context map c s R H × W × 4 is modulated by FiLM-style, time-conditioned scale and shift, then concatenated with the backbone features and compressed to H × W × C . Symbols denote elementwise addition (Add), concatenation (Concat), and multiplication (Mul).

3.3.4. Time-Modulated Context Fusion into the Denoiser

Following a conditioning paradigm similar to LDM [26], we inject multi-scale image features into the backbone denoiser. The proposed Time-modulated Context Fusion (TCF) block is illustrated in Figure 2b. To balance efficiency and accuracy, TCF replaces cross-attention with lightweight, concatenation-based fusion. To synchronize guidance with the reverse diffusion process, we apply feature-wise linear modulation (FiLM) [48]: given the backbone’s time embedding t, small modulators predict per-channel scale and shift, producing step-aware context c ˜ s = g s ( c s , t ) . At each stage s, c ˜ s is concatenated with the backbone activations and compressed with a 1 × 1 projection to match dimensionality before being passed through the denoiser. This design provides global illumination and color guidance while preserving local detail, with only a small additional per-step cost. Denoting the condition set as ( x , C ) , where C = { c ˜ s } s = 1 S , the conditional reverse diffusion follows Equations (3) and (4) with the noise predictor ϵ θ ( y t , t , x , C ) .

3.4. Luminance-Adaptive Chromaticity Loss

Optimizing only the conditional noise-prediction objective L ddpm cond in Equation (7) on paired low-light data is, in principle, sufficient to train ECA-Diff for high-quality LLIE. However, in practice we consistently observe a cool color cast with reduced saturation, most notably in low-luminance regions. This artifact often persists even when auxiliary RGB-domain losses are applied to the denoised estimate y ^ 0 , including pixelwise 1 , structural similarity index measure (SSIM) [49], and a perceptual term [50]. To explicitly regularize color tone, we introduce the Luminance-Adaptive Chromaticity Loss (LACL), computed in the CIELAB (L * a * b * ) space [27]. Owing to CIELAB’s near-perceptual uniformity, lightness is decoupled from chromatic components, enabling precise constraints on hue and chroma. LACL complements RGB-domain losses and, in this work, consists of two terms: (i) a hue-angle consistency term and (ii) a relative anti-desaturation term that penalizes chroma deficits. In addition, we apply a luminance-adaptive weight that increases as the target lightness decreases, thereby reinforcing color reconstruction in darker regions while avoiding unnecessary penalties in bright areas. We thus define LACL as follows.
Let y ^ 0 , y [ 0 , 1 ] 3 × H × W denote the predicted and reference sRGB images. We convert them to CIELAB: ( L p * , a p * , b p * ) = Lab ( y ^ 0 ) ; ( L g * , a g * , b g * ) = Lab ( y ) . Here, L * is perceptual lightness (0 for black, 100 for white), a * encodes the green (−) ↔ red (+) axis, and b * encodes the blue (−)↔ yellow (+) axis; subscripts p and g refer to predicted and ground-truth quantities, respectively. We define chroma C * = a * 2 + b * 2 . Throughout, · denotes the per-image spatial average, i.e., f = 1 H W i = 1 H j = 1 W f i j .
Hue loss via unit-vector cosine. We measure a cosine mismatch between unit-normalized chromaticity vectors to penalize hue discrepancies:
L hue = w hue 1 cos ( a p * , b p * ) C p * + ε , ( a g * , b g * ) C g * + ε .
Here cos ( · , · ) is the cosine similarity between two 2D vectors, and ε > 0 prevents division by zero and stabilizes near-gray regions ( C * 0 ).
Desaturation penalty. To discourage chroma collapse while avoiding over-saturation, we use
L desat = w desat max 0 , C g * C p * C g * + ε .
Luminance- and chroma-aware weights. We emphasize darker and more colorful regions while de-emphasizing near-gray pixels using the scalar gates defined as follows:
w hue : = w L ( 1 s C ) , w desat : = w L m ( C g * ) .
In Equation (13), w L = 0.5 + 0.5 1 L g * 100 α L , increasing the weight as luminance decreases; s C = exp C g * 2 2 σ C 2 , close to 1 for near-gray regions ( C g * 0 ) and decaying with chroma; m ( C g * ) = C g * C g * + τ C γ C , a monotonic saturating function of chroma where τ C sets the knee and γ C the slope. Unless stated otherwise, hyperparameters are fixed across all experiments: α L = 1.5, σ C = 20, τ C = 7.5, γ C = 1.4, and ε = 10 6 .
In summary, w hue = w L ( 1 s C ) focuses hue supervision on dark, chromatic pixels while downweighting near-gray regions, whereas w desat = w L m ( C g * ) strengthens anti-desaturation constraints in dark, highly saturated areas.
Overall objective. The LACL objective is finally defined as
L LACL = λ h L hue + λ d L desat ,
where λ h = 1.0 and λ d = 0.8 by default.

3.5. Overall Optimization Objective

To stabilize training and improve convergence, we optimize the conditional objective L ddpm cond together with LACL and three RGB-domain losses applied to y ^ 0 . The auxiliary RGB losses are summarized below:
(i)
Pixelwise loss: L RGB - 1 = y ^ 0 y 1 ;
(ii)
Structural loss: L SSIM = 1 SSIM ( y ^ 0 , y )  [49];
(iii)
Perceptual loss: L perc = l F ϕ l ( y ^ 0 ) ϕ l ( y ) 2  [50], where ϕ l ( · ) are VGG-16 features [51] pretrained on ImageNet [47], and F indexes a small set of layers.
Therefore, the final training objective is
L train = L ddpm cond + λ RGB - 1 L RGB - 1 + λ SSIM L SSIM + λ perc L perc + λ LACL L LACL .
Here, λ RGB - 1 , λ SSIM , λ perc , and λ LACL are scalar weights tuned on a validation set.

3.6. Datasets and Evaluation Protocol

Datasets. We adopt widely used LLIE benchmarks covering both paired and unpaired settings. The paired benchmarks include LOL [11] and LOL-v2 [52]. The LOL dataset comprises 500 real-captured low/normal-light pairs, of which 485 are used for training and 15 for testing. To assess performance on a larger and more diverse real-captured corpus, we additionally use the Real subset of LOL-v2, which contains 689 pairs for training and 100 pairs for testing. We follow the official splits for both datasets. To evaluate cross-dataset generalization, we also report results on several unpaired collections commonly used in LLIE, including DICM [2], LIME [4], MEF [53], NPE [54], and VV [55].
Preprocessing and Augmentation. For each paired sample, we apply identical augmentations—random 128 × 128 cropping and horizontal/vertical flips (p = 0.5). The low-light input and the reference image are both normalized to [ 0 , 1 ] and linearly remapped to [ 1 , 1 ] to match the diffusion model’s working dynamic range. During evaluation, we disable cropping and other geometric transforms; numerical preprocessing remains identical to the training pipeline.
Evaluation Metrics. We evaluate with full-reference and no-reference image quality assessment (IQA) metrics for paired and unpaired datasets, respectively. Following common LLIE practice, we report Peak Signal-to-Noise Ratio (PSNR) and SSIM [49] on LOL and LOL-v2 Real datasets. PSNR measures fidelity by mapping the mean squared error (MSE) between the enhanced result and the reference onto a logarithmic decibel scale, while SSIM is a perception-oriented metric that jointly considers luminance, contrast, and structural consistency. Unless otherwise noted, all values are computed on 8-bit sRGB images, per image and then averaged over the test set. To better reflect human visual preference, we also report the Learned Perceptual Image Patch Similarity (LPIPS) [50] on paired datasets, using the official implementation with the AlexNet [56] backbone. This complements PSNR/SSIM by comparing deep feature representations across multiple scales. For unpaired datasets (DICM, LIME, MEF, NPE, VV), we evaluate using BRISQUE [57] and NIQE [58], two widely used no-reference IQA measures that correlate with human judgments without requiring ground-truth images. Briefly, BRISQUE (opinion-aware) regresses natural scene statistics features against human opinion scores, making it sensitive to artifacts such as noise, blur, and contrast shifts; NIQE (opinion-unaware) measures the deviation of an image’s statistics from a pristine natural-image model. We use standard implementations and report per-image scores averaged over each dataset.
Implementation Details. We implement all models in PyTorch 1.13.1 and train on a single NVIDIA RTX 4090 GPU. Unless noted, training uses mini-batches of 12 randomly cropped 256 × 256 patches. We adopt AdamW with weight decay 1 × 10 4 and three parameter groups: (i) the diffusion backbone with learning rate l r = 5 × 10 5 , (ii) the LCN head with learning rate 10 × l r , and (iii) the fine-tuned parameters of the pretrained VAE encoder with learning rate 0.5 × l r . We apply cosine annealing over the full schedule with warm-up (multiplier 2.0) for the first 10% of the 8000 training epochs. A dropout rate of 0.15 is used in the network blocks, and the global gradient norm is clipped to 1.0 at each step.
For the diffusion setup, we use a 1000-step diffusion horizon with a linear β schedule from 1 × 10 4 to 2 × 10 2 during training. Unless otherwise stated, inference adopts DDIM sampling with 100 steps.
The denoising backbone takes a 6-channel input formed by concatenating the low-light RGB image and the current noisy latent y t . The backbone is a U-Net [25] with a base channel width of 128 and channel multipliers [ 1 , 2 , 3 , 4 ] . It performs three downsampling and three symmetric upsampling operations, yielding four resolution scales and nine stages in total (encoder: 4; bottleneck: 1; decoder: 4), with two residual blocks per stage. Consequently, in the multi-scale conditioning set { c s } s = 1 S we use S = 4. The LCN within ECAE consumes the 4-channel latent produced by the pretrained VAE encoder E ; it uses a base channel width of 64 with channel multipliers [ 2 , 4 , 4 ] , and two HTC blocks are adopted per stage. During training, E is frozen for the first 2000 epochs and then fine-tuned with half the backbone learning rate. Loss terms and their default hyperparameters follow the corresponding subsection in the method section.

4. Results

Building on the datasets and evaluation protocols detailed in Section 3.6, we present quantitative and qualitative results for ECA-Diff. For the paired benchmarks LOL and LOL-v2 Real, we compare PSNR, SSIM (higher is better), and LPIPS (lower is better), together with visual examples illustrated in Table 1, Figure 3 and Figure 4. For the real-world unpaired sets DICM, LIME, MEF, NPE, and VV, we report no-reference BRISQUE and NIQE (both lower is better) and show qualitative comparisons in Table 2 and Figure 5. For a detailed comparison of model size and inference latency against representative transformer- and diffusion-based LLIE methods, please refer to Section 5.3 and Table 3.
Table 1. Paired quantitative results on LOL and LOL-v2 Real. Metrics are PSNR/SSIM ↑ (higher is better) and LPIPS ↓ (lower is better). Best is red bold; second best is blue underlined. A dash (“–”) denotes unavailable results. Scores are taken from original papers or reproduced using authors’ official pretrained models.
Figure 3. Visual comparison across multiple scenes on LOL dataset. Each group shows Input, Kind++ [13], LLFlow [18], RetinexFormer [16], CLEDiff [21], DiffLL [20], ECA-Diff (Ours), and the Reference. The region highlighted by the red box is enlarged below to show finer details. Please zoom in for best viewing.
Figure 4. Visual comparison across multiple scenes on LOL-v2 Real dataset. Each group shows Input, Kind++ [13], LLFlow [18], RetinexFormer [16], CLEDiff [21], DiffLL [20], ECA-Diff (Ours), and the Reference. The region highlighted by the red and green square boxes is enlarged below to show finer details. Please zoom in for best viewing.
Table 2. No-reference quantitative results on real-world unpaired benchmarks. Metrics are NIQE ↓ and BRIS ↓ (BRIS abbreviates BRISQUE); ↓ indicates that lower is better. AVG. denotes the average across all datasets. Best and second best results are indicated by red bold and blue underlined text, respectively.
Figure 5. Qualitative comparison on the unpaired benchmarks LIME, MEF, and NPE. For each dataset, one representative image is shown and six methods are displayed in a fixed order: Input, KinD++ [13], LLFlow [18], CLEDiff [21], DiffLL [20], and ECA-Diff (Ours). Dataset labels are overlaid at the top-left of the first panel in each group. Zoom in for best viewing.
Table 3. Comparison of LLIE performance (PSNR/SSIM ↑), model complexity, and inference efficiency on LOL. T image denotes the average end-to-end inference time per image. For diffusion-based methods with N steps, we also report the per-step sampling latency T step . Bold values denote the best results.

4.1. Paired Benchmarks (LOL/LOL-v2)

Quantitative Results. We compare ECA-Diff against state-of-the-art LLIE methods across multiple families: CNN-based methods including Zero-DCE [59], RetinexNet [11], and KinD++ [13]; Transformer-based methods including SNR-Aware [14], LLFormer [15], RetinexFormer [16], and KAN-T [17]; Normalizing flow method LLFlow [18]; Diffusion-based methods including DiffLL [20], AnlightenDiff [22], and CLEDiff [21].
Across both paired datasets (LOL and LOL-v2 Real), ECA-Diff achieves state-of-the-art performance on all three metrics (PSNR/SSIM ↑, LPIPS ↓). On LOL, it reaches 27.79 dB PSNR, 0.887 SSIM, and 0.074 LPIPS, exceeding the second-best methods by +0.61 dB PSNR over RetinexFormer (27.18 dB), +0.014 SSIM over LLFlow (0.873), and 0.041 LPIPS over LLFlow (0.115). These gains correspond to higher pixel-level fidelity (PSNR), stronger structural consistency (SSIM), and better perceptual quality (LPIPS). On LOL-v2 real, ECA-Diff attains 28.90 dB PSNR, 0.899 SSIM, and 0.086 LPIPS. The margins are +0.02 dB PSNR over DiffLL (28.88 dB), +0.015 SSIM over KAN-T (0.884), and 0.014 LPIPS over DiffLL (0.100). Although the PSNR margin over DiffLL is small, which suggests comparable fidelity, the consistently higher SSIM and lower LPIPS nevertheless demonstrate stronger detail enhancement and superior perceptual naturalness.
For a fair comparison, we follow the same diffusion framework and training pipeline as CLEDiff when constructing our baseline. Relative to this baseline, ECA-Diff removes the self-attention blocks inside the iterative backbone and instead introduces the external ECAE with TCF-based multi-scale injection. Under these matched settings, ECA-Diff improves CLEDiff by +2.21 dB PSNR/+0.072 SSIM/ 0.077 LPIPS on LOL, and by +2.21 dB/+0.070/ 0.075 on LOL-v2 real, highlighting the effectiveness of the proposed external context-aware guidance.
Qualitative Results. Representative visual comparisons on LOL and LOL-v2 are provided in Figure 3 and Figure 4, respectively. In Figure 3, the first two sets show that ECA-Diff recovers finer details and more faithful colors than competing methods. In the third set (red zoom box), ECA-Diff yields the sharpest digits on the clock and the most consistent background-wall color, closely matching the reference. Upon close inspection of the zoomed-in crops, markedly stronger global noise suppression becomes evident. In Figure 4, ECA-Diff yields, in set 1, cleaner character edges and improved background color; in set 2, it yields clearer restoration of the cloud-like texture, higher wall color fidelity, and more uniform large-scale color across the floor and wall. These gains stem from ECAE’s global context modeling, which stabilizes color, suppresses noise, and preserves fine structures.

4.2. Unpaired Benchmarks (DICM/LIME/MEF/NPE/VV)

Quantitative Results. On real-world unpaired datasets, we compare ECA-Diff against LIME [4], Zero-DCE [59], EnlightenGAN [60], KinD++ [13], SNR-Aware [14], LLFlow [18], CLEDiff [21], and DiffLL [20]. Quantitative results are summarized in Table 2.
Across the five unpaired datasets, ECA-Diff attains the best score on at least one metric for every dataset, including simultaneous best NIQE and BRISQUE on MEF, best BRISQUE on LIME and VV, and best NIQE on DICM, while remaining competitive elsewhere (second-best NIQE on LIME and VV, and second-best BRISQUE on NPE). Aggregating over datasets, it delivers the best average on both NIQE and BRISQUE. Since NIQE quantifies deviation from natural image statistics and BRISQUE penalizes departures from spatial naturalness (e.g., artifacts and over/under-enhancement), jointly lower values indicate improved global illumination correction with fewer perceptual distortions.
It is worth noting that the no-reference datasets are not used for training, and their content and illumination styles differ significantly from both the paired training sets and from each other. Consequently, NIQE and BRISQUE on these datasets mainly reflect the generalization ability of each method under distribution shift, so their per-dataset rankings may naturally deviate from the full-reference PSNR/SSIM/LPIPS rankings obtained on paired benchmarks. This behavior is further explained by the fact that NIQE and BRISQUE are derived from different natural scene statistics and prefer different trade-offs between smoothness, sharpness, and contrast. In our results, occasional cases where another method achieves a slightly better NIQE or BRISQUE but performs worse on full-reference metrics are consistent with the known biases of these measures and dataset-specific statistics, rather than indicating instability of ECA-Diff. Overall, the full-reference improvements on paired datasets together with the favorable no-reference scores on unseen real-world images demonstrate that ECA-Diff generalizes robustly with a balanced trade-off between fidelity and perceptual naturalness.
Qualitative Results. Figure 5 presents qualitative comparisons on LIME, MEF, and NPE, covering scenes with severe underexposure and color casts. Overall, ECA-Diff produces sharper textures, more faithful colors, and fewer artifacts, yielding a more natural and visually comfortable appearance across datasets. In contrast, KinD++ often exhibits over-smoothing together with saturation overshoot; LLFlow tends to break down in dark regions (e.g., tree branches in the NPE example) and shows hue discontinuities in the sky (LIME/MEF); CLEDiff frequently over-exposes bright areas. DiffLL delivers generally pleasing results (e.g., the balloons in MEF) and can rival ours in some views, yet issues remain: oversaturation in green vegetation (LIME/NPE), a cyan/green shift in the MEF sky, and non-smooth artifacts that become apparent upon zoom-in. Together with the no-reference scores in Table 2, these observations substantiate the strong low-light enhancement capability of ECA-Diff on real, unpaired images.

5. Ablation Studies

We dissect ECA-Diff along five axes: (i) where to inject ECAE within the backbone; (ii) which ECAE components contribute most; (iii) the efficiency and computational complexity analysis of ECA-Diff; (iv) the impact of training patch size to probe global-context learning; and (v) the effect of removing LACL objective during training. Unless otherwise specified, all ablation experiments are conducted on the LOL dataset with batch size b s = 12 and patch size p s = 128 . The efficiency comparison is based on fully optimized ECA-Diff, while the patch-size study varies p s with all other settings fixed. We report PSNR and SSIM for quantitative evaluation.

5.1. ECAE Placement Across Backbone Stages

This subsection investigates where to inject ECAE into the diffusion backbone and quantifies the accuracy–efficiency trade-off. We follow the U-Net [25] taxonomy and use the placement notation Down (D) for the encoder/down-sampling stages, Mid for the bottleneck/middle stage, and Up (U) for the decoder/up-sampling stages. Unless stated otherwise, efficiency numbers are reported as per-step averages for DDIM sampling with N = 100 on full-resolution LOL images ( 400 × 600 ).
As illustrated in Table 4, inserting the ECAE-based context guidance at any backbone stage consistently improves quality over the baseline; mid-stage injection offers the best accuracy–efficiency trade-off, adding the up-sampling stage on top of mid yields the largest absolute gain, and enabling all stages delivers the overall best PSNR/SSIM (26.46/0.864). Although the external encoder increases the number of trainable parameters, the amortized per-step latency grows only slightly because the encoder is executed once and reused across diffusion steps, as indicated by the small gap between ms/step (image) and ms/step (sample)—most of the overhead arises from per-step feature injection rather than the one-off encoder. Consequently, we adopt the Full insertion as the default: with 128 × 128 training patches it improves the baseline by approximately + 0 . 90 dB PSNR and + 0 . 035 SSIM with an amortized per-step time increase of ≈22.9% (from 157.5 ms to 193.5 ms), alongside a moderate rise in per-step FLOPs.
Table 4. ECAE insertion study on LOL. Quality is reported with PSNR/SSIM ↑. Efficiency columns are per-step averages for DDIM sampling at N = 100. FLOPs/step and ms/step (image) include the one-off ECAE amortized over steps, whereas ms/step (sample) excludes the one-off encoder. ✓ indicates insertion of ECAE at the corresponding stage; bold values denote the best results.

5.2. ECAE Component Analysis

This subsection conducts a hierarchical ablation under the mid-stage injection configuration, denoted as ECA-Diff (Mid); at the macro level we toggle the VAE latent encoder E and LCN, at the micro level we dissect HTC block by enabling or disabling the Transformer branch, the CNN branch, and Bi-CA. Beyond these switches, we further isolate specific components: the external prior of pretrained E with controlled alternatives (Table 5), and the HTC fusion strategies (Bi-CA versus addition/concatenation, Table 6).
Table 5. Impact of the latent encoder E pretraining under ECA-Diff (Mid) on LOL. All rows except the baseline use ECAE with LCN trained from scratch; only E training states differ. Init: Rand = random init; Pre = pretrained. Train: Frz = frozen; Scratch = train from scratch; FT = fine-tune. Metrics: PSNR/SSIM ↑. Bold values denote the best results.
Table 6. Comparison of fusion strategies in the HTC block under ECA-Diff (Mid) on LOL. Reported Params and Runtime refer only to the ECAE module (external context); runtime is per-image. “Add” denotes element-wise addition; “Concat” denotes channel concatenation. Metrics: PSNR/SSIM ↑. Bold values denote the best results.
(1) Components Ablations. From Table 7, three observations emerge: (i) LCN is indispensable, directly injecting the VAE latent into the mid stage without LCN substantially degrades quality, showing that structured context extraction is necessary; (ii) within HTC block, the Transformer and CNN branches are complementary, removing either branch reduces performance; and (iii) disabling the Bi-CA further erodes accuracy, indicating that cross-branch fusion is an effective mechanism for leveraging heterogeneous features. Overall, the complete ECAE injection achieves the best quality among the tested variants; compared with the baseline, ECA-Diff (Mid) improves performance by + 0 . 42 dB PSNR and + 0 . 015 SSIM.
Table 7. Component ablation of ECAE under ECA-Diff (Mid) on LOL. Columns are organized hierarchically: Macro components (VAE latent encoder E ; LCN) and HTC components. A ✓ indicates that the component is enabled. When only a single branch is present (w/o Trans or w/o CNN), Bi-CA is not applicable. Metrics: PSNR/SSIM ↑. Bold values denote the best results.
(2) Impact of the Pretrained Latent Encoder. To assess the benefit of the external prior and the potential domain gap, we evaluate variants under the mid-stage injection setting with LCN enabled (Table 5). The trends are clear: freezing E leads to noticeable degradation, and freezing a pretrained encoder harms performance even more than freezing a randomly initialized one, indicating a mismatch between the pretraining domain and LLIE when reused without adaptation. Allowing E to learn from scratch recovers and slightly improves over the baseline, while fine-tuning a pretrained E consistently achieves the best PSNR/SSIM among all settings. These results show that large-scale pretraining provides useful global semantics, but light task-specific adaptation is necessary to mitigate the domain gap, motivating our default choice of a pretrained and fine-tuned E .
(3) Effect of Fusion Strategies in the HTC Block. Table 6 compares three fusion strategies in HTC under the same backbone and mid-stage injection setting. Element-wise addition fails to exploit ECAE and even degrades performance, indicating that naive linear fusion cannot properly align the Transformer branch (global semantics) and the CNN branch (local details). Channel concatenation delivers clear quality improvements over the baseline but increases parameters and runtime due to the wider projection layer. The proposed Bi-CA achieves the best PSNR/SSIM with only a moderate overhead relative to concatenation, showing that explicitly modeling bidirectional interactions between global and local features is a more effective and efficient strategy for context refinement in HTC.

5.3. Efficiency and Complexity Analysis

To comprehensively assess the efficiency of ECA-Diff and its design choices, we establish a unified evaluation protocol consisting of: (i) step-count analysis, (ii) overall latency comparison, (iii) comparable-capacity studies, and (iv) condition injection overhead analysis. All measurements are conducted on the LOL dataset with an input resolution of 400 × 600 , a batch size of 1, and a shared hardware/software environment (a single RTX 4090 GPU). We report (i) PSNR/SSIM, (ii) the total number of learnable parameters (M), (iii) the per-step sampling latency of the denoiser T step (s/step), and (iv) the end-to-end inference latency T image (s/image). Unless otherwise specified, all efficiency experiments are based on the fully optimized ECA-Diff (Full) configuration, whose performance is reported in Table 1.
For standard diffusion-based models without additional one-off modules, the inference latency is well approximated by
T image N · T step ,
where N is the number of DDIM sampling steps and T step is the average denoising-step latency. For ECA-Diff, which introduces a once-per-image ECAE, the total latency can be decomposed as
T image T ECAE + N · T step ,
where T ECAE denotes the one-off ECAE cost and T step is measured from the iterative sampler alone. In practice, Equations (16) and (17) closely matches the observed runtimes, with minor deviations due to data transfer, kernel scheduling, and measurement noise. For single-pass Transformer-based baselines (N = 1) without iterative sampling, we directly report the end-to-end latency as T image .
(1) Effect of DDIM Steps. Table 8 reports the behavior of ECA-Diff under different DDIM sampling steps N. With very few steps (N = 5), the model is clearly under-converged, leading to ineffective denoising and significant quality degradation. From N = 10 onward the performance improves substantially, and starting at N = 25 the results become highly stable: for N = 25, 50, 75, 100, both PSNR and SSIM fluctuate only marginally, indicating that ECA-Diff quickly saturates once a moderate number of steps is used. The sampling time T sample grows approximately linearly with N, with a nearly constant per-step sampling cost around 0.18–0.20 s. The one-time ECAE overhead (latent encoder E plus LCN) is about 0.07 s per image, which remains small compared with the cumulative sampling cost in typical multi-step settings and is well captured by Equation (17). In the main quantitative comparisons (Table 1), we adopt a conservative default of N = 100 steps to report a stable high-quality configuration, while the above results show that ECA-Diff maintains competitive performance even with reduced step counts (e.g., N = 25 or 50).
Table 8. Effect of DDIM sampling steps on ECA-Diff on LOL. Metrics are PSNR/SSIM ↑. T sample = T step × N denotes the diffusion sampling time per image excluding the one-time ECAE cost. Bold values denote the best results.
(2) Overall Latency Comparison. Table 3 compares representative transformer-based and diffusion-based LLIE methods on LOL. For transformer-based methods (SNR-Aware [14], RetinexFormer [16]), T image is the direct single-pass latency. For diffusion-based methods (DiffLL [20], PyDiff [19], CLEDiff [21]), we list the default iterative number N, the measured end-to-end latency T image , and the average per-step sampling cost T step . For methods without additional one-off modules, T step closely matches T image / N ; for ECA-Diff, T step is obtained from the isolated sampler runtime and the total latency follows Equation (17).
The transformer baselines achieve low latency thanks to their lightweight architectures and single-step inference, with RetinexFormer reaching 27.18 dB PSNR using only 1.61 M parameters. Diffusion-based LLIE models offer stronger generative capacity, but their absolute latency is heavily influenced by backbone design and step scheduling. DiffLL exploits wavelet-domain diffusion and PyDiff adopts a pyramid formulation with only four sampling steps, both of which explicitly optimize the sampling process for efficiency. Under this setting, our ECA-Diff, built upon a classical DDPM-style framework with standard DDIM sampling, is not directly tailored to the same extreme-latency regime as these specialized designs.
Our focus here is on the efficiency of the proposed ECA-Diff framework itself. CLEDiff serves as a strong and fair baseline, since it follows the same diffusion paradigm and training strategy. Compared with CLEDiff, ECA-Diff removes the self-attention blocks in the iterative backbone and offloads context modeling to an external, once-executed ECAE, which is coupled back through lightweight TCF modules. Although this redesign roughly doubles the total parameter count (from 84.88 M to 105.74 M + 61.68 M), it yields both faster sampling and higher restoration quality. With N = 10 steps, ECA-Diff reduces the per-step latency from 0.224 s to 0.179 s (about 20% reduction) and the end-to-end latency from 2.240 s to 1.857 s (about 17% reduction), while improving performance by +2.13 dB PSNR and +0.056 SSIM. With N = 100 steps, the per-step cost still decreases (0.224 s → 0.192 s, ∼14% reduction), and ECA-Diff achieves gains of +2.21 dB PSNR and +0.072 SSIM over CLEDiff. These results indicate that reallocating capacity into an external encoder and simplifying the recurrent denoising pathway leads to a more efficient diffusion framework, even when all parameters are accounted for. To further disentangle the benefits of the proposed framework from pure model scaling, the next subsection evaluates ECA-Diff under comparable-capacity settings against CLEDiff and related baselines.
It is also worth noting that the proposed ECA-Diff is orthogonal to backbone-lightening and step-reduction techniques used by DiffLL and PyDiff. Since ECAE and TCF are designed as plug-in components on top of a standard DDPM backbone, they can in principle be combined with wavelet-domain parameterization, pyramid formulations, or other advanced sampling schedules, opening up additional room for jointly improving both efficiency and restoration quality.
(3) Efficiency Under Comparable Capacity. In this study, we compare ECAE against backbones that increase capacity without using ECAE so as to attain comparable capacity (i.e., similar parameter count and/or per-step FLOPs). We use ch_mult to denote the stage-wise channel multipliers of the U-Net encoder–decoder, and attn_use to list the stages (1-indexed) where global attention (GA) blocks are inserted. For a fair comparison, all models listed in Table 9, including ECA-Diff (Full), are trained from scratch under the same setting ( b s = 12 , p s = 128 for 8000 rounds). We report PSNR/SSIM together with model size, FLOPs per step, and T image under DDIM sampling with N = 100 on full-resolution LOL images ( 400 × 600 ).
Table 9. Efficiency under comparable capacity. Quality is PSNR/SSIM ↑. Efficiency numbers (Params and FLOPs) are per-step averages for DDIM sampling at N = 100. “ECAE injection” uses ✓ for enabled and ✗ for disabled. [ ] denotes none GA blocks are used. Bold values denote the best results.
From Table 9 it is clear that enlarging the backbone alone does not guarantee higher LLIE quality. The Deeper model approximately doubles the parameter count (83.1 M → 179.1 M; + 115 % ) and slightly increases per-step FLOPs from 3400 G to 3593 G ( + 5.7 % ), yet its accuracy decreases to 24.93/0.823, indicating that the improvement delivered by our method arises from task-aligned design rather than model capacity. Increasing capacity further with Deeper + Wider + GA strengthens long-range modeling, but the quality remains below ECA-Diff (Full) (25.31/0.833 compared with 26.46/0.864) despite similar compute (4286 G and 4417 G FLOPs per step). In addition, adding GA to shallow, large-resolution stages leads to a much higher per-image latency (27.59 s), whereas ECA-Diff (Full) maintains a lower per-image latency (19.35 s) by computing global context once in the latent space and reusing it across diffusion steps. Overall, we conclude that the superior performance-efficiency of ECA-Diff results from external, one-off context extraction with lightweight per-step injection rather than brute-force increases in parameters or FLOPs.
(4) TCF vs. LDM-style Conditioning. Table 10 compares three context injection strategies under incremental multi-stage injection. To focus on the fusion overhead, we report backbone parameters and the per-step smpling latency T step , excluding the one-time ECAE cost. In the LDM-style token cross-attention (CA) scheme, the ECAE output is compressed into a fixed 16 × 16 token map and each selected stage performs cross-attention with these tokens. Adding more injection stages accumulates additional cross-attention blocks over multi-scale features, leading to a steady increase in T step from the mid-only setting to the full M + U configuration. In the spatial CA scheme, ECAE features are resized to match each stage and full-resolution cross-attention is applied in the spatial domain. The complexity grows rapidly with spatial size, causing a sharp rise in T step when multiple decoder stages are enabled and resulting in out-of-memory for deeper multi-stage variants.
Table 10. Comparison of context injection strategies under incremental multi-stage injection on LOL ( 400 × 600 ). “M” denotes mid-level injection; “U [k]” denotes injection at the k-th upsampling stage. T step is the per-step sampling latency (s), excluding the one-time ECAE cost. “OOM” indicates out-of-memory. Bold values denote the best results.
The proposed TCF instead aligns ECAE features with backbone resolutions and fuses them via lightweight time modulation and channel concatenation, without explicit attention. Its cost scales approximately linearly with the number of injection stages, and the ECA-Diff (M + U) configuration maintains a per-step latency of 0.180 s with 99.18 M backbone parameters, substantially lower than the LDM-style and spatial CA counterparts. This contrast is consistent with the original motivation of LDM-style cross-attention, which is efficient when the backbone operates in a low-resolution latent space with fixed-length text tokens, but becomes less suitable for full-resolution, detail-sensitive LLIE; TCF better matches the design characteristics of ECA-Diff by providing scalable external guidance with well-controlled overhead.

5.4. Impact of Patch Size

We study how the training patch size p s influences performance for both the baseline and ECA-Diff. Since ECAE aggregates global, context-aware information, our hypothesis is that larger patches expose longer-range statistics and thus benefit ECA-Diff more than a baseline without explicit global modeling. To test this, we compare the two models at p s { 64 , 128 , 256 } , as summarized in Table 11. To isolate the effect of p s , all other training and architecture settings are kept fixed. The batch size b s is scaled approximately as 1 / p s 2 so that the per-step pixel throughput b s × p s 2 is nearly constant. The number of epochs is adjusted to keep the total optimizer updates unchanged.
Table 11. Impact of training patch size p s on quality. Metrics are PSNR/SSIM ↑. Here p s denotes the training crop size of p s × p s . The last column ( 128 256 ) indicates a two-stage schedule: first train with p s = 128 until convergence, then fine-tune with p s = 256; this setting is used for our final reported results. Bold values denote the best results.
Table 11 shows that enlarging the training patch improves baseline PSNR, while SSIM remains unstable and even drops at p s = 256, indicating that without explicit global context modeling larger crops mainly enhance pixel fidelity but do not consistently preserve structure. For ECA-Diff, the gain at small patches is negligible at p s = 64 because ECAE downsamples five times and the latent grid is only 2 × 2 , which limits the amount of recoverable global context; as p s increases to 128 and 256, the latent space carries richer scene statistics and quality rises rapidly, surpassing the baseline by + 0.90 dB PSNR and + 0.035 SSIM at p s = 128, and by + 1.29 dB PSNR and + 0.079 SSIM at p s = 256. To obtain the best accuracy while keeping training efficient, we adopt the two-stage schedule 128 → 256, which produces the final results reported in this paper.

5.5. Ablation on the LACL

To verify the effectiveness of the LACL term with minimal confounders, we start from the mid-stage configuration ECA-Diff (Mid) and vary only the training objective in Equation (15). We compare three settings: MSE-only, which uses only the noise-prediction objective L ddpm cond ; MSE + RGB, which additionally enables the RGB-space terms L RGB - 1 , L SSIM , and L perc ; and Full loss, which further includes the LACL term L LACL . As reported in Table 12, RGB-space supervision improves upon MSE-only, and adding L LACL yields the best and most consistent PSNR/SSIM, indicating that LACL provides complementary guidance beyond pixel and perceptual matching. Qualitatively, as shown in Figure 6, MSE-only produces severe color bias with noticeable hue shifts and distorted chroma in neutral regions. Enabling the RGB-space losses (MSE + RGB) markedly improves structure and exposure, yet introduces a clear cool color cast that persists across large areas. Incorporating LACL (Full loss) effectively suppresses this cool tint, restoring neutral whites and more faithful hues; the resulting tone and chroma are closest to the reference image.
Table 12. LACL loss ablation on ECA-Diff (Mid). Metrics are PSNR/SSIM ↑. “MSE-only”: L ddpm cond ; “MSE + RGB”: L ddpm cond + RGB-space losses { L RGB - 1 , L SSIM , L perc } ; “Full loss”: L train in Equation (15). Bold values denote the best results.
Figure 6. Qualitative comparison aligned with Table 12. Please focus on the restoration of tone and chroma over large regions.

6. Discussion

Overall Performance. The experimental results demonstrate that the proposed ECA-Diff framework effectively strengthens diffusion-based LLIE by introducing an external, compute-once global context pathway. Compared with recent CNN-, Transformer-, and diffusion-based approaches, ECA-Diff provides consistent gains on paired benchmarks; under similar diffusion backbone and training pipeline as CLEDiff, it yields notable improvements in PSNR, SSIM, and LPIPS on LOL and LOL-v2 Real, and the efficiency and complexity analysis under parameter- and FLOPs-matched settings further indicates that these gains are attributable to external context guidance rather than increased backbone capacity. On real-world unpaired datasets, ECA-Diff achieves the best or highly competitive BRISQUE and NIQE scores while producing visually stable illumination and color, suggesting that the injected context helps align the generative behavior of diffusion models with natural image statistics instead of over-amplifying contrast or saturation.
Model Design. From a design perspective, ECAE decouples global context estimation from the iterative denoising loop: deep context features are computed once in a compact latent space and reused across all steps, which reduces the potential marginal per-step overhead and alleviates conflicts between local denoising and global correction. Ablation results confirm that effective multi-stage injection is essential; naive reuse of a frozen pretrained latent encoder or simple additive fusion is suboptimal, whereas a pretrained-and-fine-tuned latent encoder, an LCN built with hybrid CNN/Transformer branches, and Bi-CA collectively provide stable and complementary gains. Within this framework, LCN captures long-range dependencies in the latent space, and TCF injects the resulting multi-scale context into the backbone through lightweight, timestep-aware modulation, avoiding the quadratic cost of full-resolution cross-attention. The unified efficiency and complexity analysis shows that these benefits are achieved with moderate overhead, indicating that the performance gains stem from targeted, cost-effective architectural choices rather than brute-force model scaling. The LACL objective complements this design by regularizing chromaticity in CIELAB space, mitigating the typical cool, desaturated bias in dark regions and promoting plausible, stable color reproduction.
Limitations and failure cases. Despite these advantages, several limitations remain. First, the overall inference time is still constrained by the iterative nature of the DDPM-style backbone: ECA-Diff keeps the additional cost modest by computing the external context once and injecting it via lightweight TCF modules, but does not match the absolute runtime of one-pass lightweight LLIE networks or diffusion models with aggressively accelerated samplers or distilled backbones. Second, ECA-Diff can be challenged in extremely low SNR scenes, under strong spatially varying color casts, or with dense fine-grained textures. As illustrated in Figure 7, it effectively suppresses noise and corrects global illumination relative to the baseline diffusion model, but may yield slightly over-smoothed textures and mildly over-enhanced saturation, reflecting a bias toward clean, color-consistent solutions when local evidence is weak and revealing an inherent trade-off between robustness and perfect preservation of micro-structures.
Figure 7. Visual comparison on a challenging indoor scene. ECA-Diff suppresses noise and corrects global illumination compared with the baseline, but shows slightly over-smoothed textures and mildly over-enhanced saturation relative to the reference (please zoom in for a clearer view), illustrating the trade-off between strong denoising with stable color and faithful preservation of fine details.
Future Directions. From a broader perspective, this work highlights the potential of combining task-specific external encoders with diffusion models for low-level vision. Although ECA-Diff is instantiated on a vanilla DDPM-style backbone and thus inherits relatively high absolute runtime, our experiments validate the effectiveness of the compute-once external context paradigm and show that it is orthogonal to backbone and sampling accelerations. A primary direction is therefore to couple ECA-Diff with accelerated diffusion samplers, distilled or lightweight backbones, and other efficiency-oriented designs to better match the training and inference efficiency of emerging diffusion architectures while further improving the quality–efficiency trade-off. In parallel, extending the concept of ECAE and the compute-once, multi-step conditioning scheme to other restoration tasks such as deblurring, dehazing, inpainting, and super-resolution is a natural step, as these problems also benefit from robust global structural cues. Another promising avenue is to design adaptive context encoders that explicitly model sensor characteristics and scene semantics for more reliable enhancement in extremely dark or mixed-illumination environments.

7. Conclusions

We presented ECA-Diff, which enhances diffusion-based LLIE by externalizing global-context estimation via a task-specific encoder and reusing it as step-shared conditioning for the denoising backbone. This decoupled integration yields consistent improvements in fidelity and perceptual quality while adding only a controllable, sampling-dependent latency. In addition, we introduce LACL in CIELAB color space that emphasizes darker and more colorful regions while de-emphasizing near-gray pixels, thereby regularizing diffusion training toward accurate hue and natural saturation. The proposed compute-once, multi-step conditioning paradigm is broadly applicable to low-level restoration tasks. Future work will focus on accelerating the ECAE and its injection pipeline and on extending the paradigm to deblurring, dehazing, and super-resolution.

Author Contributions

Conceptualization, P.T. and Y.W.; methodology, P.T.; investigation, P.T.; resources, Y.W.; data curation, P.T.; writing—original draft preparation, P.T.; writing—review and editing, P.T. and Y.W.; visualization, P.T.; supervision, A.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The publicly archived datasets can be found at https://github.com/weichen582/RetinexNet (LOL dataset, accessed on 15 November 2024), https://github.com/flyywh/CVPR-2020-Semi-Low-Light (LOL-v2 dataset, accessed on 15 November 2024), and https://github.com/baidut/BIMEF (unpaired datasets VV, LIME, NPE, DICM and MEF, accessed on 15 November 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
Bi-CABidirectional Cross-Attention
CACross-Attention
DDIMDenoising Diffusion Implicit Model
DDPMDenoising Diffusion Probabilistic Models
ECA-DiffExternal Context-Aware encoder guided Diffusion framework
ECAEExternal Context-Aware Encoder
FiLMFeature-wise Linear Modulation
GAGlobal Attention
HEHistogram Equalization
HTCHybrid Transformer-Convolution block
IQAImage Quality Assessment
LACLLuminance-Adaptive Chromaticity Loss
LCNLatent Context Network
LDMLatent Diffusion Models
LLIELow-Light Image Enhancement
LPIPSLearned Perceptual Image Patch Similarity
MSEMean Squared Error
PSNRPeak Signal-to-Noise Ratio
SNRSignal-to-Noise Ratio
SSIMStructural Similarity Index Measure
TCFTime-modulated Context Fusion
VAEVariational Auto Encoder

References

  1. Cheng, H.; Shi, X. A simple and effective histogram equalization approach to image enhancement. Digit. Signal Process. 2004, 14, 158–170. [Google Scholar] [CrossRef]
  2. Lee, C.; Lee, C.; Kim, C.S. Contrast Enhancement Based on Layered Difference Representation of 2D Histograms. IEEE Trans. Image Process. 2013, 22, 5372–5384. [Google Scholar] [CrossRef]
  3. Land, E.H. The Retinex Theory of Color Vision. Sci. Am. 1977, 237, 108–129. [Google Scholar] [CrossRef] [PubMed]
  4. Guo, X. LIME: A Method for Low-light IMage Enhancement. In Proceedings of the 24th ACM International Conference on Multimedia, Amsterdam, The Netherlands, 15–19 October 2016; pp. 87–91. [Google Scholar] [CrossRef]
  5. Chen, C.; Chen, Q.; Xu, J.; Koltun, V. Learning to See in the Dark. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3291–3300. [Google Scholar]
  6. Zhang, C.; Yan, Q.; Zhu, Y.; Li, X.; Sun, J.; Zhang, Y. Attention-Based Network For Low-Light Image Enhancement. In Proceedings of the IEEE International Conference on Multimedia and Expo, ICME 2020, London, UK, 6–10 July 2020; pp. 1–6. [Google Scholar] [CrossRef]
  7. Zhu, M.; Pan, P.; Chen, W.; Yang, Y. EEMEFN: Low-Light Image Enhancement via Edge-Enhanced Multi-Exposure Fusion Network. In Proceedings of the The Thirty-Fourth AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 13106–13113. [Google Scholar] [CrossRef]
  8. Xu, K.; Yang, X.; Yin, B.; Lau, R.W.H. Learning to Restore Low-Light Images via Decomposition-and-Enhancement. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, 13–19 June 2020; pp. 2278–2287. [Google Scholar] [CrossRef]
  9. Meng, Z.; Xu, R.; Ho, C.M. GIA-Net: Global Information Aware Network for Low-Light Imaging. In Proceedings of the Computer Vision—ECCV 2020 Workshops, Part III, Glasgow, UK, 23–28 August 2020; Volume 12537, pp. 327–342. [Google Scholar] [CrossRef]
  10. Tang, P.; Guo, X.; Ju, G.; Shen, L.; Men, A. Integration-and-Diffusion Network for Low-Light Image Enhancement. In Proceedings of the 2021 IEEE International Conference on Image Processing, ICIP 2021, Anchorage, AK, USA, 19–22 September 2021; pp. 1664–1668. [Google Scholar] [CrossRef]
  11. Wei, C.; Wang, W.; Yang, W.; Liu, J. Deep Retinex Decomposition for Low-Light Enhancement. In Proceedings of the British Machine Vision Conference 2018, BMVC 2018, Newcastle, UK, 3–6 September 2018; p. 155. [Google Scholar]
  12. Zhang, Y.; Zhang, J.; Guo, X. Kindling the Darkness: A Practical Low-light Image Enhancer. In Proceedings of the 27th ACM International Conference on Multimedia, MM 2019, Nice, France, 21–25 October 2019; pp. 1632–1640. [Google Scholar] [CrossRef]
  13. Zhang, Y.; Guo, X.; Ma, J.; Liu, W.; Zhang, J. Beyond Brightening Low-light Images. Int. J. Comput. Vis. 2021, 129, 1013–1037. [Google Scholar] [CrossRef]
  14. Xu, X.; Wang, R.; Fu, C.; Jia, J. SNR-Aware Low-light Image Enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022, New Orleans, LA, USA, 18–24 June 2022; pp. 17693–17703. [Google Scholar] [CrossRef]
  15. Zou, W.; Gao, H.; Yang, W.; Liu, T. Wave-Mamba: Wavelet State Space Model for Ultra-High-Definition Low-Light Image Enhancement. arXiv 2024, arXiv:2408.01276. [Google Scholar] [CrossRef]
  16. Cai, Y.; Bian, H.; Lin, J.; Wang, H.; Timofte, R.; Zhang, Y. Retinexformer: One-stage Retinex-based Transformer for Low-light Image Enhancement. In Proceedings of the IEEE/CVF International Conference on Computer Vision, ICCV 2023, Paris, France, 1–6 October 2023; pp. 12470–12479. [Google Scholar] [CrossRef]
  17. Brateanu, A.; Balmez, R.; Orhei, C.; Ancuti, C.; Ancuti, C. Enhancing Low-Light Images with Kolmogorov–Arnold Networks in Transformer Attention. Sensors 2025, 25, 327. [Google Scholar] [CrossRef] [PubMed]
  18. Wang, Y.; Wan, R.; Yang, W.; Li, H.; Chau, L.; Kot, A.C. Low-Light Image Enhancement with Normalizing Flow. In Proceedings of the Thirty-Sixth AAAI Conference on Artificial Intelligence, AAAI 2022, online, 22 February–1 March 2022; pp. 2604–2612. [Google Scholar] [CrossRef]
  19. Zhou, D.; Yang, Z.; Yang, Y. Pyramid Diffusion Models for Low-light Image Enhancement. In Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence, IJCAI 2023, Macao, China, 19–25 August 2023; pp. 1795–1803. [Google Scholar] [CrossRef]
  20. Jiang, H.; Luo, A.; Fan, H.; Han, S.; Liu, S. Low-Light Image Enhancement with Wavelet-Based Diffusion Models. ACM Trans. Graph. 2023, 42, 238:1–238:14. [Google Scholar] [CrossRef]
  21. Yin, Y.; Xu, D.; Tan, C.; Liu, P.; Zhao, Y.; Wei, Y. CLE Diffusion: Controllable Light Enhancement Diffusion Model. In Proceedings of the 31st ACM International Conference on Multimedia, MM 2023, Ottawa, ON, Canada, 29 October–3 November 2023; pp. 8145–8156. [Google Scholar] [CrossRef]
  22. Chan, C.; Siu, W.; Chan, Y.; Chan, H.A. AnlightenDiff: Anchoring Diffusion Probabilistic Model on Low Light Image Enhancement. IEEE Trans. Image Process. 2024, 33, 6324–6339. [Google Scholar] [CrossRef]
  23. Zhang, F.; Liu, X.; Gao, C.; Sang, N. Color and Luminance Separated Enhancement for Low-Light Images with Brightness Guidance. Sensors 2024, 24, 2711. [Google Scholar] [CrossRef]
  24. Ho, J.; Jain, A.; Abbeel, P. Denoising Diffusion Probabilistic Models. In Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, Virtual, 6–12 December 2020. [Google Scholar]
  25. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015—18th International Conference, Munich, Germany, 5–9 October 2015; Volume 9351, pp. 234–241. [Google Scholar] [CrossRef]
  26. Rombach, R.; Blattmann, A.; Lorenz, D.; Esser, P.; Ommer, B. High-Resolution Image Synthesis with Latent Diffusion Models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022, New Orleans, LA, USA, 18–24 June 2022; pp. 10674–10685. [Google Scholar] [CrossRef]
  27. ISO/CIE 11664-4:2019; Colorimetry—Part 4: CIE 1976 L* a* b* Colour Space. International Organization for Standardization: Geneva, Switzerland; Commission Internationale de l’Éclairage: Vienna, Austria, 2019.
  28. Abdullah-Al-Wadud, M.; Kabir, M.H.; Dewan, M.A.A.; Chae, O. A Dynamic Histogram Equalization for Image Contrast Enhancement. IEEE Trans. Consum. Electron. 2007, 53, 593–600. [Google Scholar] [CrossRef]
  29. Arici, T.; Dikbas, S.; Altunbasak, Y. A Histogram Modification Framework and Its Application for Image Contrast Enhancement. IEEE Trans. Image Process. 2009, 18, 1921–1935. [Google Scholar] [CrossRef]
  30. Nakai, K.; Hoshi, Y.; Taguchi, A. Color image contrast enhacement method based on differential intensity/saturation gray-levels histograms. In Proceedings of the International Symposium on Intelligent Signal Processing and Communication Systems, ISPACS 2013, Naha-shi, Japan, 12–15 November 2013; pp. 445–449. [Google Scholar] [CrossRef]
  31. Ying, Z.; Li, G.; Ren, Y.; Wang, R.; Wang, W. A New Low-Light Image Enhancement Algorithm Using Camera Response Model. In Proceedings of the 2017 IEEE International Conference on Computer Vision Workshops, ICCV Workshops 2017, Venice, Italy, 22–29 October 2017; pp. 3015–3022. [Google Scholar] [CrossRef]
  32. Lee, C.; Shih, J.; Lien, C.; Han, C. Adaptive Multiscale Retinex for Image Contrast Enhancement. In Proceedings of the Ninth International Conference on Signal-Image Technology & Internet-Based Systems, SITIS 2013, Kyoto, Japan, 2–5 December 2013; pp. 43–50. [Google Scholar] [CrossRef]
  33. Jobson, D.J.; Rahman, Z.; Woodell, G.A. A multiscale retinex for bridging the gap between color images and the human observation of scenes. IEEE Trans. Image Process. 1997, 6, 965–976. [Google Scholar] [CrossRef]
  34. Li, M.; Liu, J.; Yang, W.; Sun, X.; Guo, Z. Structure-Revealing Low-Light Image Enhancement Via Robust Retinex Model. IEEE Trans. Image Process. 2018, 27, 2828–2841. [Google Scholar] [CrossRef]
  35. Cai, B.; Xu, X.; Guo, K.; Jia, K.; Hu, B.; Tao, D. A Joint Intrinsic-Extrinsic Prior Model for Retinex. In Proceedings of the IEEE International Conference on Computer Vision, ICCV 2017, Venice, Italy, 22–29 October 2017; pp. 4020–4029. [Google Scholar] [CrossRef]
  36. Lore, K.G.; Akintayo, A.; Sarkar, S. LLNet: A deep autoencoder approach to natural low-light image enhancement. Pattern Recognit. 2017, 61, 650–662. [Google Scholar] [CrossRef]
  37. Cui, Z.; Li, K.; Gu, L.; Su, S.; Gao, P.; Jiang, Z.; Qiao, Y.; Harada, T. Illumination Adaptive Transformer. arXiv 2022, arXiv:2205.14871. [Google Scholar] [CrossRef]
  38. Zamir, S.W.; Arora, A.; Khan, S.H.; Hayat, M.; Khan, F.S.; Yang, M.; Shao, L. Learning Enriched Features for Real Image Restoration and Enhancement. In Proceedings of the Computer Vision—ECCV 2020—16th European Conference, Part XXV, Glasgow, UK, 23–28 August 2020; Volume 12370, pp. 492–511. [Google Scholar] [CrossRef]
  39. Dai, Z.; Liu, H.; Le, Q.V.; Tan, M. CoAtNet: Marrying Convolution and Attention for All Data Sizes. In Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, NeurIPS 2021, Virtual, 6–14 December 2021; pp. 3965–3977. [Google Scholar]
  40. Lin, X.; Yan, Z.; Deng, X.; Zheng, C.; Yu, L. ConvFormer: Plug-and-Play CNN-Style Transformers for Improving Medical Image Segmentation. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2023—26th International Conference, Vancouver, BC, Canada, 8–12 October 2023; Volume 14223, pp. 642–651. [Google Scholar] [CrossRef]
  41. Saharia, C.; Ho, J.; Chan, W.; Salimans, T.; Fleet, D.J.; Norouzi, M. Image Super-Resolution via Iterative Refinement. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 4713–4726. [Google Scholar] [CrossRef] [PubMed]
  42. Hui, C.C.; Siu, W.; Law, N.; Chan, H.A. Intelligent Painter: New Masking Strategy and Self- Referencing with Resampling. In Proceedings of the 24th International Conference on Digital Signal Processing, DSP 2023, Rhodes, Greece, 11–13 June 2023; pp. 1–5. [Google Scholar] [CrossRef]
  43. Xia, B.; Zhang, Y.; Wang, S.; Wang, Y.; Wu, X.; Tian, Y.; Yang, W.; Gool, L.V. DiffIR: Efficient Diffusion Model for Image Restoration. In Proceedings of the IEEE/CVF International Conference on Computer Vision, ICCV 2023, Paris, France, 1–6 October 2023; pp. 13049–13059. [Google Scholar] [CrossRef]
  44. Guo, L.; Wang, C.; Yang, W.; Huang, S.; Wang, Y.; Pfister, H.; Wen, B. ShadowDiffusion: When Degradation Prior Meets Diffusion Model for Shadow Removal. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2023, Vancouver, BC, Canada, 17–24 June 2023; pp. 14049–14058. [Google Scholar] [CrossRef]
  45. Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.; et al. Segment Anything. In Proceedings of the IEEE/CVF International Conference on Computer Vision, ICCV 2023, Paris, France, 1–6 October 2023; pp. 3992–4003. [Google Scholar] [CrossRef]
  46. Song, J.; Meng, C.; Ermon, S. Denoising Diffusion Implicit Models. In Proceedings of the 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, 3–7 May 2021. [Google Scholar]
  47. Deng, J.; Dong, W.; Socher, R.; Li, L.; Li, K.; Fei-Fei, L. ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition CVPR 2009, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar] [CrossRef]
  48. Perez, E.; Strub, F.; de Vries, H.; Dumoulin, V.; Courville, A.C. FiLM: Visual Reasoning with a General Conditioning Layer. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, (AAAI-18), New Orleans, LA, USA, 2–7 February 2018; pp. 3942–3951. [Google Scholar] [CrossRef]
  49. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  50. Zhang, R.; Isola, P.; Efros, A.A.; Shechtman, E.; Wang, O. The Unreasonable Effectiveness of Deep Features as a Perceptual Metric. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18–22 June 2018; pp. 586–595. [Google Scholar] [CrossRef]
  51. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  52. Yang, W.; Wang, W.; Huang, H.; Wang, S.; Liu, J. Sparse Gradient Regularized Deep Retinex Network for Robust Low-Light Image Enhancement. IEEE Trans. Image Process. 2021, 30, 2072–2086. [Google Scholar] [CrossRef] [PubMed]
  53. Ma, K.; Zeng, K.; Wang, Z. Perceptual Quality Assessment for Multi-Exposure Image Fusion. IEEE Trans. Image Process. 2015, 24, 3345–3356. [Google Scholar] [CrossRef]
  54. Wang, S.; Zheng, J.; Hu, H.; Li, B. Naturalness Preserved Enhancement Algorithm for Non-Uniform Illumination Images. IEEE Trans. Image Process. 2013, 22, 3538–3548. [Google Scholar] [CrossRef]
  55. Vonikakis, V.; Kouskouridas, R.; Gasteratos, A. On the evaluation of illumination compensation algorithms. Multim. Tools Appl. 2018, 77, 9211–9231. [Google Scholar] [CrossRef]
  56. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  57. Mittal, A.; Moorthy, A.K.; Bovik, A.C. No-Reference Image Quality Assessment in the Spatial Domain. IEEE Trans. Image Process. 2012, 21, 4695–4708. [Google Scholar] [CrossRef]
  58. Mittal, A.; Soundararajan, R.; Bovik, A.C. Making a “Completely Blind” Image Quality Analyzer. IEEE Signal Process. Lett. 2013, 20, 209–212. [Google Scholar] [CrossRef]
  59. Li, C.; Guo, C.; Loy, C.C. Learning to Enhance Low-Light Image via Zero-Reference Deep Curve Estimation. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 4225–4238. [Google Scholar] [CrossRef] [PubMed]
  60. Jiang, Y.; Gong, X.; Liu, D.; Cheng, Y.; Fang, C.; Shen, X.; Yang, J.; Zhou, P.; Wang, Z. EnlightenGAN: Deep Light Enhancement Without Paired Supervision. IEEE Trans. Image Process. 2021, 30, 2340–2349. [Google Scholar] [CrossRef] [PubMed]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.