Next Article in Journal
Dual-Branch Deep Learning with Dynamic Stage Detection for CT Tube Life Prediction
Previous Article in Journal
Evaluating Insulin Delivery Systems Using Dynamic Glucose Region Plots and Risk Space Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MSF-ACA: Low-Light Image Enhancement Network Based on Multi-Scale Feature Fusion and Adaptive Contrast Adjustment

1
School of Science, Hubei University of Technology, Wuhan 430068, China
2
College of Informatics, Huazhong Agricultural University, Wuhan 430070, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(15), 4789; https://doi.org/10.3390/s25154789 (registering DOI)
Submission received: 4 June 2025 / Revised: 19 July 2025 / Accepted: 28 July 2025 / Published: 4 August 2025
(This article belongs to the Section Sensing and Imaging)

Abstract

To address the issues of loss of important detailed features, insufficient contrast enhancement, and high computational complexity in existing low-light image enhancing methodologies, this paper presents a low-light image enhancement network (MSF-ACA), which uses multi-scale feature fusion and adaptive contrast adjustment. Focus is placed on designing the local–global image feature fusion module (LG-IFFB) and the adaptive image contrast enhancement module (AICEB), in which the LG-IFFB adopts the local–global dual-branching structure to extract multi-scale image features, and utilizes the element-by-element multiplication method to fuse the local details with the global illumination distribution to alleviate the problem of serious loss of image details, while the AICEB incorporates linear contrast enhancement and confidence adaptive stopping mechanism, which dynamically adjusts the computational depth according to the confidence of the feature map, balancing the contrast enhancement and computational efficiency. According to the results of the experiment, the parameter count of MSF-ACA is 0.02 M, and compared with today’s mainstream algorithms, the suggested model attains 21.53 dB in PSNR when evaluated on the LOL-v2-real evaluation dataset, and the BRI is as low as 16.04 on the unpaired dataset DICM, which provides a better detail clarity and color fidelity in visual enhancement, and it is a highly efficient and robust low-light image model.

1. Introduction

Low-light images are generally caused by factors such as insufficient light given by the environment, limitations or malfunctions of the equipment, noisy signals, or improperly set parameters when shooting. The presence of such images can make it difficult to perform complex optical tasks, which include identity recognition and target detection. There is widespread use of low-light image enhancement algorithms in security surveillance, medical imaging, and other fields [1,2,3,4]. How can the luminance of an image be enhanced, its color recovered, and its texture details improved, while also boosting the computational efficiency of enhancing low-light images. Studying this area is essential for enhancing images taken in low light. Low-light image enhancement is one of the hot topics in the field of image processing.
Image enhancement algorithms can be classified into two main groups: conventional methods using manually constructed parameters and deep learning-based methods for improving images.
Development using traditional methods based on artificially constructed parameters began in the mid-to-late 20th century. These methods improve the overall visual effect by adjusting the illumination, contrast, and color of an image using artificial parameters. Typical algorithms include histogram equalization [5,6,7], gamma correction [8,9,10], and Retinex [11,12,13,14]. The histogram equalization method improves image contrast by redistributing values across pixels; the gamma correction method adjusts the luminosity and contrast of an overexposed or underexposed image using a non-linear transformation; and Retinex algorithm, which was proposed in collaboration with McCann and Land, is rooted in the visual system of the human eye and improves the effect of uneven lighting by separating the light and reflection components. Although these traditional algorithms can effectively improve image clarity and visual effect, they have problems such as relying on experts’ experience, poor scene adaptability, and general enhancement effect.
Deep learning-based image enhancement algorithms are gradually becoming mainstream methods as their results are generally better than traditional algorithms. According to the training method, it falls into two categories: supervised methods and unsupervised methods. Methodologies involving supervision necessitate the acquisition of paired datasets under low and normal lighting conditions through synthesis, shooting and data enhancement, and utilize the paired data for supervised training of image enhancement models. Common low-light image enhancement datasets include LOL dataset [15], MIT-Adobe FiveK dataset [16], LIME dataset [17], etc.
Zhang et al. [18] proposed Image Kindling the Darkness (Kindling the Darkness), which uses supervised learning to train the network using paired image datasets with good results. The network first breaks the image down into light and reflection parts. Then, the light component is used for lighting adjustments, while the reflection component is employed for removing degradations. However, it requires multi-stage training and the convolutional neural networks (CNNs) used to decompose color images perform noise reduction on reflectance, and need to be trained independently to adjust illumination and then connected together for end-to-end fine-tuning, which complicates and lengthens the process of training. As an extension of the Retinex framework, Wu et al. [19] put forward a deep unrolling network called URetinex-Net, which makes use of pair-wise data, and implements image enhancement through a manually designed a priori and optimization-driven approach. This method achieves excellent results in noise suppression and feature detail preservation. However, its strong dependence on training data and high computational complexity limit its generalization ability in complex unknown scenes. FMR-Net, proposed by Chen et al. [20], is a fast multiscale residual network which rapidly boosts the image quality in low-light by combining a highly optimized residual block with a multibranch structure, while preserving image details and contrast, maintaining image details and contrast. However, it suffers from serious loss of high-frequency details and insufficient noise suppression.
The above methods are mainly based on supervised learning. In order minimize dependency on paired illumination datasets, numerous scholars put forward image enhancement methods founded on unsupervised learning for low-light conditions. Jiang et al. [21] proposed EnlightenGAN, which uses generative adversarial networks for unsupervised learning and introduces a self-regularization mechanism. But it suffers from limited generalization and unstable training, etc. Guo et al. [22] proposed a zero-reference depth curve estimation method-Zero-DCE, which learns the dynamic range adjustment through image-specific curves, but its problem is that the accuracy is limited under extreme conditions and the non-reference loss function relies on the image-specific curves, but its accuracy is limited under extreme conditions and the non-reference loss function relies on the image-specific curves. The problem lies in the limited accuracy under extreme conditions and the dependence of the non-reference loss function on empirical parameters (e.g., exposure values). Ma et al. [23] proposed a self-calibrating illumination (SCI) learning framework, which achieves fast convergence of multi-stage outputs to a consistent state through a progressive illumination estimation module with shared weights and a self-calibration module, thus requiring only single-stage inference at test time. Combined with unsupervised training loss (fidelity and smoothing constraints), it enhances the model’s adaptability to complex low-light scenes. However, the network still suffers from detail loss or color bias.
Currently, low-light image enhancement algorithms primarily suffer from detail feature degradation, inadequate contrast enhancement, as well as the model has a high computing burden; for this reason, a low-light image enhancing network is put forward in this paper, which is founded upon multi-scale feature fusion and adaptive contrast adjustment, and designs a local–global image feature fusion module (LG-IFFB) and an adaptive image contrast enhancement module (AICEB). LG-IFFB extracts multi-scale image features through a local–global dual-branch structure and uses element-by-element multiplication to fuse the local details with the global light distribution to alleviate the problem of image detail loss; AICEB combines a linear contrast enhancement formula with a confidence stability-driven adaptive stopping mechanism to dynamically adjust the computational depth according to the confidence of the feature map, balancing the contrast enhancement and computational efficiency. As shown in Figure 1, the method proposed in this paper achieves excellent results when compared qualitatively and quantitatively with some of the current mainstream methods. The primary outcomes of this study are summarized below:
1.
A low-light image enhancing network based upon multi-scale feature fusion and contrast adaptive adjustment is proposed to achieve low-light image enhancement through a lightweight architecture synergizing multi-scale image feature fusion with dynamic optimization of image contrast.
2.
A local–global image feature fusion module (LG-IFFB) is designed, which adopts a dual-path structure of local branching and global branching to simultaneously extract local and global information at different scales of the image, realizing a balance between detail preservation and global light optimization, and providing parameter mapping more suitable for complex low-light scenes for the subsequent luminance enhancement network.
3.
An adaptive image contrast enhancement module (AICEB) is proposed, which consists of multiple iterative sub-modules, each of which dynamically generates contrast enhancement factors and luminance parameters through an adaptive attention normalization block (AANBlock). A confidence scoring mechanism is introduced in the module to realize the adaptive contrast enhancement, effectively balancing the contrast enhancement and computational efficiency.

2. Methods

The MSF-ACA model architecture is illustrated in Figure 2. Firstly, the low-light image undergoes a convolutional operation that increases its amount of channels for feature maps, resulting in 24 channels. Then, using the local–global image feature fusion module (LG-IFFB), the multi-scale image features U are extracted and fused. In the luminance enhancement network, the multi-scale image features U are first uniformly partitioned along the channel dimension, and each group of the segmented features serves as a luminance tuning parameter mapping, which is utilized to perform luminance adjustment on the original image m. Specifically, through the optimization process of eight iterations, the output of the Luminance-enhanced feature map Θ, which will then be used as a component of the input to two subsequent cascaded Adaptive Image Contrast Boosting Modules (AICEBs), where the input to the first AICEB is the two Θs, and the output is the intermediate feature map Y. Input to the second AICEB is the intermediate feature maps Y and Θ, while its output represents the intermediate features map Z. AICEB dynamically generates contrast factors and luminance parameters, and adaptively terminates the redundant computation through a confidence scoring mechanism to balance performance and efficiency. The two AICEBs sequentially optimize the features in a recursive form to gradually improve the image contrast and brightness. Finally, MSF-ACA fuses the channels and reconstructs the image by using a 3 × 3 convolutional fusion channels. The main structure of MSF-ACA references the residual join learning mechanism of local combination global [25].

2.1. Local–Global Image Feature Fusion Block

Local–Global Image Feature Fusion Block (LG-IFFB): Figure 2 illustrates its structure; the division of the process is primarily into two steps: first, multi-scale feature extraction; second, image feature fusion.
Inspired by the structure of RFDB [25] as well as RRFB [25], LG-IFFB employs a multi-scale reparameterization technique for feature extraction, which outputs six different scales of image features (denoted as L1…L6) by cascading six kinds of convolutional kernel to maximize the preservation image texture information. In order to reduce memory loss, ADD operation is used instead of Concat operation for feature fusion. In addition, the remaining image feature channels are then directly combined with the fused features using the residual join method. This design is conducive to enhanced computational efficiency as well as reduced feature redundancy within the cascade structure.
Traditional multi-scale feature fusion methods [26] usually perform only simple feature splicing or summing processing, which cannot deeply fuse to utilize the global contextual information of the features. In order to realize the adaptive fusion of image features, LG-IFFB is designed with a two-branch structure: local branch and global branch. The local branch focuses on the optimal fusion of local features, and the usual practice is to extract local features using a single 3 × 3 convolution [27,28,29], which ignores the correlation of multi-scale local features. For this reason, in this paper, the input multiscale image features (L1…L6) undergo a 1 × 1 convolutional process that extends their dimensionality. Next, a deep convolutional process is employed to extract local information, and, in particular, for each scale, the extracted local features are spliced with other local features extracted at different scales to enhance the correlation. The final locally generated optimized features L1LL6L is obtained by processing the spliced features using deep convolutions and activating functions. In the global branch, the input feature channels are segmented, with only one-fourth of the total number of channels selected to extract and fuse features. Specifically, to integrate multi-scale contextual information, the features at all scales (L1…L6) are first aggregated by element summation to generate the combined representation U′ [30]. The global feature descriptor S is then computed by performing global average pooling (GAP) on U′, which compresses the spatial dimensions to preserve channel statistics and ensures that the feature descriptor S encodes the overall illumination and structural patterns of the entire image. A fully connected (FC) network consisting of two linear layers, a ReLU activation layer, and a Sigmoid layer processes the global feature descriptor S to obtain the attentional weights S′. The attention weight S′ is calibrated by multiplying it with each scale feature to obtain the global calibration feature L1GL6G. The global calibration features (L1GL6G) captures long-range contextual correlations, which are then combined with local optimization features (L1LL6L) are multiplied and fused element by element to generate the final fused feature map U. This fusion strategy ensures that the final output U contains both fine local details and globally consistent illumination information, thus addressing the limitation of the single-branch approach that is unable to extract both local and global features simultaneously. Compared with the DCE-Net [22] in Zero-DCE, which relies primarily on single-branch convolution for learning the mapping relating the input image to the best-fit curve, LG-IFFB can address the issue of detail loss or contrast imbalance in extremely low-light or high-dynamic-range scenarios more effectively by integrating the link with multi-scale local details and global light distribution.

2.2. Luminance Enhancement Network

This study proposes a luminance enhancement function for the luminance Enhancement Network based on the illumination adjustment curves framework of Zero-DCE. The developed recurrent mapping function adheres to two fundamental design criteria: (1) the preservation of inter-pixel intensity relationships through enforced monotonicity constraints, and (2) the implementation of computational simplicity and gradient accessibility to enable efficient error backpropagation. The luminance enhancement function is formulated as a parametric quadratic transformation and can be represented mathematically by the following expression:
E p m = E p 1 m + S p m E p 1 m 1 E p 1 m , p 8
where Sp(m) is the pixel-by-pixel parameter matrix (p denotes the number of iterations and m denotes the input image), the luminance-enhanced image resulting from p iterations is known as Ep(m), whereas the enhanced version of this image after p − 1 iterations is referred to as Ep−1(m). Based on (1), this paper presents a design for a luminance enhancement network; refer to Figure 2. Two parts make up the network’s inputs: the original low-light image mRH×W×3 and the fused feature map U ∈ RH×W×24 after LG-IFFB processing. The final output of the luminance feature map Θ ∈ RH×24 with a channel number of 24 is obtained by dividing the LG-IFFB output of the fused feature map. U is partitioned into 8 identical pixel-by-pixel parameter maps (Sp(m), m = 1, …, 8) in order to take part in the iteration of the function. The initial input original image E0(m) = m, after 8 iterations, outputs a 3-channel enhanced image E8(m), which is subsequently extended to 24 channels by 3 × 3 convolution to generate the luminance-enhanced feature map Θ = Conv(E8(m)).

2.3. Adaptive Image Contrast Enhancement Block

Direct contrast enhancement of input features (e.g., linear stretching) introduces noise or distortion due to unstable feature distribution. In a style migration task, Vedaldi et al. [31] argued that instance normalization techniques can eliminate the contrast difference in image features in order to allow the network to focus on learning content structure rather than luminance or color distribution. Drawing on this idea, this paper uses the channel normalization technique to improve the stability of contrast enhancement of image features. The channel normalization operation is to normalize each channel of the input feature map separately so that its mean is 0 and variance is 1, i.e.,
h = h c μ c σ c 2 + ε
where the feature map of channel c is represented by hc, μc refers to the mean values of channel c, and σc refers to the standard deviation value of channel c. ε is a stabilization factor to avoid a zero denominator.
On this basis, this paper proposes an adaptive image contrast enhancement block (AICEB). Figure 3a illustrates the framework of the AICEB. The AICEB consists of several iterative submodules (Iteration Sub-module), and each Iteration Sub-module contains 1 Adaptive Attention Normalization Block (AANBlock) and 1 ReLU Activation layer. Figure 3b demonstrates the detailed architecture of the AANBlock.
The AANBlock input in the kth submodule has two inputs: the luminance enhancement feature map Θ obtained after processing by the luminance enhancement network and the preceding iterative submodule Xk’s output feature map. For the 1st iterative submodule, X1 is the luminance feature map produced from the luminance enhancement network, which is represented by the symbol Θ. Through the learning of the luminance enhancement feature map Θ, it generates the parameters a and b needed for image contrast adjustment. Thus it allows the network to adaptively adjust the image contrast according to the luminance distribution information adaptively. The specific realization process is as follows:
In each iterative sub-module, take the kth iterative sub-module as an example, the features of the luminance feature map Θ are first further extracted using a 3 × 3 convolution and a 5 × 5 convolution to obtain the feature map Θ’. Then, the global information of Θ’ is computed by Global Average Pooling (GAP) and Global Maximum Pooling (GMP), respectively, and the outputs of GAP and GMP are merged in the dimension of the channel to obtain statistical features of global information. The global information statistical features obtained from splicing are processed by a Fully Connected (FC) Layer and Rectified Linear Unit (ReLU) activation functions are applied to generate the Channel Attention Weight T. Multiplying Θ’ with the channel attention weight T produces the feature map α*. The ReLU Activation Function is used for guaranteeing that pixel values α* are non-negative. Based on α*, a 7 × 7 convolutional layer combined with a Sigmoid Activation Function needs to be used in order to generate the feature map β*. In the formula for linearly enhancing the contrast, pixel values from the feature maps α* and β* act as the parameters a and b.
Employing parameters a and b derived from feature maps α* and β*, a linear contrast transform is applied to the features that have been normalized by the channels, and the contrast-enhanced result Xk+1 can be obtained. The output after ReLU activation processing is then output, i.e., it is the output of the kth iterative sub-module, Xk+1. Particularly, in order to ensure the effective enhancement of the low-light image contrast, a bias term of magnitude 1 is added to α* is added with a bias term of size 1, i.e., α* + 1. The specific formulation is set out below:
U ( m , n ) = a × T ( m , n ) + b , 0 < m < H , 0 < n < W
where T ( m , n ) and U ( m , n ) denote the image features to be enhanced and the image features after contrast enhancement, respectively; a and b are the parameters of image contrast enhancement, specifically the pixel values of the feature maps α* and β*.
X k + 1 = Relu ( ( α * + 1 ) X k μ σ + β * )
where σ and µ represent the standard deviation and average of the feature map Xk, respectively.
In order to balance the image contrast enhancement effect and computational efficiency, this paper sets out a proposal for an adaptive stopping mechanism founded on the stability of the confidence level. While iteratively enhancing the image contrast, the contrast confidence level is calculated in real time to evaluate the enhancement results, and the computational depth of the network is adaptively adjusted, and its flow is presented in Figure 3a.
After the iterative sub-module completes the image contrast enhancement processing, the confidence of the current results is generated using the contrast confidence calculation module C. The contrast confidence calculation module C consists of a variance calculation layer (Contrast), a Layer 1 × 1 Convolution (Conv) followed by a Sigmoid Activation layer. Confidence is calculated as follows:
C o n f i d e n c e = S i g m o i d ( C o n v ( Contrast ( x ) ) ) [ 0 , 1 ]
where the variance calculation layer Contrast(x) is
C o n t r a s t ( x ) = 1 h w u , v ( x u , v μ ) 2
w and h refer to the width and height of the feature map x; xu,v represent a pixel’s value at position (u, v) in feature map x; and μ indicates the mean value of all pixels in x. Variance is widely used to measure image pixel dispersion; the higher the variance, the greater the difference between pixels, i.e., the higher the contrast. If the absolute value of the difference in the confidence level for three consecutive times is less than the preset threshold λ, the enhancement effect of the feature map can be considered to be stabilized, and the iteration is terminated and exited:
C o n f i d e n c e k - C o n f i d e n c e k 1 < λ , k { f , f 1 , f 2 }
where Confidencek represents the confidence at the kth layer, f is the index of the current iteration number k, and the value of f is not less than 3. The preset threshold λ is obtained by experiments on the validation set [15] by taking into account the average number of iterations and the PSNR loss, which is specifically taken as 0.0005.

2.4. Loss Function

For the purpose of optimizing MSF-ACA training, this paper uses spatial consistency loss, color consistency loss, mean absolute error loss, gradient-guided structural consistency loss, and perceptual loss to comprehensively evaluate the image enhancement effect.
Spatial Consistency Loss (Lspa): The structural consistency before and after enhancement is maintained by analyzing the luminance difference in the local region of the image. Firstly, the image is divided into M pixel blocks, where each block p is compared with its four neighborhoods Φ(p). Then the luminance differences between the reference image S and the enhanced image T between the corresponding blocks are calculated, respectively, and finally the stability of the spatial relation is constrained by the mean value of the squared error:
L s p a = 1 M M p = 1 M q Φ ( p ) S p S q T p T q 2
Color Consistency Loss (Lcol): It is used to ensure the stability of the color distribution of the content during the image enhancement process. The loss function achieves this goal by constraining the intensity differences between different color channels. For any pair of channels (m, n) in the set of channel pairs ε, the sum of the squared differences in their average intensity values I m and I n is computed:
L c o l = ( m , n ) ε I m I n 2 , ε = ( R , G ) , ( R , B ) , ( G × B )
Mean Absolute Error Loss (L1): It is mainly used to prevent images from localized overexposure or underexposure. This loss function achieves this goal by comparing the difference in pixel luminance between the reference image and the enhanced image. The luminance value of the ith pixel in the reference image is denoted as Yi, and the luminance value of the corresponding pixel in the enhanced image is denoted as Oi, and let the total number of pixels in the image be N:
L 1 = i = 1 N Y i O i N
Gradient-guided Structural Consistency Loss (Lgsc): By establishing structural similarity constraints in the gradient domain, the consistency of luminance, contrast, and structural features during image enhancement is effectively maintained. This paper presents the gradient-based structural similarity loss function (GSSIM). The GSSIM method outperforms the conventional SSIM method in images with low light and blurring [32]. The gradient magnitude of the augmented image, O, is expressed as Go(u, v), whilst the gradient magnitude of the reference images, Y, is expressed as GY(u, v). (u, v) are the row-column coordinates representing the pixels in the image, whereas C acts as a stabilization constant to prevent the denominator from equaling zero:
L g s c = 1 2 u v G o ( u , v ) G Y ( u , v ) + C u v G o ( u , v ) 2 + u v G Y ( u , v ) 2 + C
Perceptual Loss (Lperceptual): Perceptual loss aims to maintain semantic consistency between the augmented image and the reference image through deep feature alignment. To achieve this goal, this study uses an ImageNet-based pre-trained VGG network architecture as a static feature encoder. The layer l features obtained from the pre-trained VGG network are represented by Φl(x) and Φl(y). x refers to the low-light-enhanced output image, y denotes the reference image and λl indicates the weight used for the perceptual loss of layer l. The VGG network is used to minimize the distance between the low-light-enhanced image and reference image. ‖.‖2 denotes the L2 paradigm.
L p e r c e p t u a l ( x , y ) = l λ l Φ l ( x ) Φ l ( y ) 2 2
The total loss function (Ltotal) is the weighted average of the above five loss functions.
L t o t a l = ω 1 L s p a + ω 2 L c o l + ω 3 L 1 + ω 4 L g s c + ω 5 L p e r c e p t u a l
where ω 1 ω 5 is the weight parameter of each loss function.

3. Experiment

3.1. Implementation Details

In this paper, a total of 1385 images are selected from the LOL datasets series (v1 [15] and v2 [33]), and with the objective of speeding up the training process, a 256 × 256 pixel size window is selected, and the input images undergo center clipping. In the course of network training, Adam [34] optimizer was used. Adam optimizer optimized the network for 40 cycles (batch size = 1, learning rate = 10−4). Initialization of all the network weights was performed using a zero-mean, 0.02 standard Gaussian function, while the bias was originally defined as a fixed value. Implementation of this entire process relied on the PyTorch 1.13.0 [35] framework, and using an NVIDIA GeForce RTX 4060 graphics card, MSF-ACA underwent both training and evaluation phases.
To evaluate the proposed low-light image enhancement approach, the proposed method is compared with eight current state-of-the-art low-light image enhancement algorithms, specifically, CLIP-LIT [36], DSLR [37], SCI [23], RUAS [38], URetinex [19], HEP [24], DeepLPF [39], and FMR-Net [20]. It is worth noting that some of these methods (e.g., HEP and RUAS) are unsupervised frameworks. According to [40], PSNR and SSIM are used to assess the effect of contrast enhancement. PSNR focuses on pixel differences, while SSIM evaluates visual structural similarity.

3.2. Comparison and Analysis of Paired Datasets

Quantitative Analysis: The efficacy of different enhancement methods on LOL-v2 subsets (LOL-v2-real and LOL-v2-syn) was assessed through four metrics: PSNR and SSIM [41] for quality evaluation, complemented by FLOPs for computational efficiency analysis, and the number of model parameters (Params) were chosen to analyze the computational complexity of the model. As can be seen from Table 1, excellent results were achieved in regard to the PSNR, FLOPs, and parametric evaluations are conducted across both real-world and synthetic versions of the LOL-v2 benchmark. Larger values of the PSNR and SSIM metrics represent higher image quality and are indicated by ‘↑’ in Table 1. SCI along with RUAS show remarkable model lightweighting performance, but the PSNR and SSIM of the LOL-v2-real and LOL-v2-syn datasets are significantly lower than those of the MSF-ACA. On the LOL-v2-real dataset, MSF-ACA has a PSNR of 21.53 dB, which is better than URetinex’s 21.09 dB, and an SSIM of 0.771, which is slightly lower than that of URetinex’s 0.858. On the LOL-v2-syn dataset, MSF-ACA ranks higher in both PSNR and SSIM and is ranked first in both metrics. With regard to the two important performance metrics, PSNR and SSIM, which characterize an image enhancement effect, both achieved the top two places, and the FLOPs of MSF-ACA of this paper’s method is 29.97 G, which is higher than that of URetinex’s 1.24 G. However, the comprehensive performance of MSF-ACA is better than that of URetinex on both datasets, with the performance gain far outweighing the increase in the computing cost. The size of the parameter of both models is 0.02 M, which fully verifies the balance between performance and efficiency in lightweight design, and achieves SOTA effect in image quality and model complexity.
Visual Analysis: Figure 4 and Figure 5 show the visual comparison results between this paper’s method and other state-of-the-art image enhancement methods on paired datasets. The LOL-v2-real dataset is selected in Figure 4 and the LOL-v2-Synthetic dataset is selected in Figure 5. The results from Figure 4 and Figure 5 show that the FMR-Net method is poor in detail retention, although it improves in brightness, with obvious noise and artifacts in some regions. DeepLPF is deficient in both detail retention and brightness enhancement. On the other hand, the method in this paper not only significantly improves the overall brightness when enhancing the low-light image, but also performs well in detail retention; the details in the image are clearly visible, yielding an effect that matches the ground truth in visual fidelity. In addition, as can be seen from the zoomed-in regions in Figure 4 and Figure 5, FMR-Net still lacks in the processing of details, with obvious noise appearing in some regions. URetinex has the problem of color deviation on LOL-v2-Synthetic, and the color and luminance of the enhanced image produced by DeepLPF deviates significantly from the referenced image, compared to which, the approach outlined presented in this paper effectively boosts local and global contrast, producing sharper details while also performing well in noise suppression for better visual results.
Figure 4. Visual comparison on the LOL-v2 real dataset. There are two red boxes in the figure, the larger red box is an enlargement of the smaller red box. Same as Figure 5, Figure 6 and Figure 7.
Figure 4. Visual comparison on the LOL-v2 real dataset. There are two red boxes in the figure, the larger red box is an enlargement of the smaller red box. Same as Figure 5, Figure 6 and Figure 7.
Sensors 25 04789 g004
Figure 5. Visual comparison on the LOL-v2 Synthetic dataset.
Figure 5. Visual comparison on the LOL-v2 Synthetic dataset.
Sensors 25 04789 g005
Figure 6. Visual comparison on DICM, MEF, and VV non-reference datasets.
Figure 6. Visual comparison on DICM, MEF, and VV non-reference datasets.
Sensors 25 04789 g006
Figure 7. Visual comparison of ablation studies.
Figure 7. Visual comparison of ablation studies.
Sensors 25 04789 g007

3.3. Comparison and Analysis of Unpaired Datasets

Quantitative Analysis: The proposed MSF-ACA was compared with competing methodologies using five unpaired datasets (LIME [7], DICM [10], NPE [42], MEF [43], and VV) to further validate its effectiveness. Two non-reference perceptual metrics, PI [44] and BRI [45], were chosen for the evaluation of the visual quality of the enhancement results. The visual quality is indicated by lower metrics. Smaller values of the BRI and PI metrics represent higher image quality, as indicated by ‘↓’ in Table 2. As demonstrated in Table 2, experimental results demonstrate that our approach obtains the lowest BRI and PI scores on LIME, DICM, and MEF datasets, as well as the lowest PI scores across the NPE and VV datasets. In the context of average scores across the two metrics, the method in this paper is the lowest on all five datasets, indicating the proposed method still achieves satisfactory performance metrics in unseen real-world scenarios.
Visual Analysis: Figure 6 shows the visual comparison results between this paper’s method and the competing methods on the unpaired dataset, and rows 1–3 in the figure correspond to the image enhancement results from the DICM, MEF, and VV datasets, respectively. For images from the DICM dataset, the overall brightness of the house and the details of the sky sunset color can be better recovered; for images from the MEF dataset, the sky and cloud colors can be enhanced more naturally; for images from the VV dataset, the RUAS method and the SCI method underperform in the recovery of the overall brightness, and in comparison with the CLIP-LIT and the HEP methods; the method in this paper can better recover the texture color of the wall. Overall, this paper’s method can still effectively recover the lighting and recreate the image’s details in unknown scenes, demonstrating its ability to generalize.

3.4. Ablation Experiment

The LOL-v2 dataset is used to perform ablation experiments. Using the luminance-only enhancement network as the baseline model, the performance difference is analyzed by adding the corresponding modules to verify the efficacy of the Local–Global Image Feature Fusion Module (LG-IFFB), and Adaptive Image Contrast Enhancement Block (AICEB) proposed in this paper.
Local–Global Image Feature Fusion Block: Model A introduces only the local–global image feature fusion module based on the baseline model, which improves the SSIM and PSNR metrics to 0.01 and 0.74 dB, respectively. From Figure 7, it is clear that model A results in a greater image texture than the baseline model, such as the clouds in the sky, the texture of trees, and so on. This proves that adding the local–global image feature fusion module can learn more expressive features.
Adaptive Image Contrast Enhancement Block: Model B and Model C are based on Model A, further adding Adaptive Image Contrast Enhancement Module (AICEB). A fixed number of iterations is used by Model B. In this paper, the iteration count for each AICEB is pre-set to 10 and there are two AICEB modules, so the overall iteration count is 20. Model C employs the adaptive stopping mechanism proposed in this paper.
From Table 3, it can be observed that there is no improvement in the SSIM metric value and the PSNR metric is improved by 0.39 dB for Model B compared to Model A. In Model C, by introducing the adaptive stopping mechanism, the number of iterations decreases by an average of 9.9 iterations and the average running time decreases by 0.08 s compared to Model B, while the PSNR and SSIM metric values are improved by 0.77 dB and 0.04, respectively. The accuracy and computational efficiency of the model in processing images are improved at the same time. From the visual test in Figure 7, the color and contrast of the image are significantly improved. The effectiveness of the AICEB module and the adaptive stopping mechanism proposed herein has been proven.
Ablation test outcomes for the quantity of AICEB modules, as carried out in this paper, are displayed in Table 4. When two AICEB modules are introduced into the model, compared with the introduction of two AICEBs, there is no improvement in either the PSNR or the SSIM metrics, but average iteration and running times have increased significantly. When the number of AICEB modules is two, a balance between the accuracy and efficiency of the model can be achieved.

3.5. Selection of Iteration Thresholds

According to [46], PSNR is used to evaluate the effect of contrast enhancement. So, the threshold λ of the adaptive stopping mechanism in the AICEB module of this paper is determined based on a combination of the average number of iterations and PSNR loss. From 0.00001 to 0.001, five different thresholds are selected, and the number of iterations and PSNR values in different cases are shown in Table 5. It is not difficult to find that when λ = 0.0005, the model reduces the number of iterations by 14% while the PSNR loss is only 10%, which realizes the balance between accuracy and efficiency, and for this reason, the λ threshold is set to 0.0005 in this paper.

4. Conclusions

In this paper, a lightweight network (MSF-ACA) combining multi-scale feature fusion and contrast adaptive adjustment is proposed to effectively realize high-quality enhancement to images taken in low-light conditions. By designing a local–global two-branch feature fusion module (LG-IFFB), the integration of multi-scale local details and global illumination information effectively alleviates the problem of texture loss; an adaptive image contrast enhancement module (AICEB) is introduced to adaptively balance the computational efficiency and enhancement effect through a confidence scoring mechanism. A large number of experiments show that MSF-ACA not only outperforms existing methods on typical datasets but also has a significant advantage in computational complexity.

Author Contributions

Resources, Z.F. and Y.L.; Writing—original draft, Z.C. and Y.W.; Supervision, F.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research project was supported in part by the Key Research and Development Program of Hubei Province of China (No. 2024BBB055) and the Major Science and Technology Special Project Plan of Yunnan Province of China (No. 202502AE090003).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, X. Infrared image filtering and enhancement processing method based upon image processing technology. J. Electron. Imaging 2022, 31, 051408. [Google Scholar] [CrossRef]
  2. Gao, X.; Liu, S. DAFuse: A fusion for infrared and visible images based on generative adversarial network. J. Electron. Imaging 2022, 31, 043023. [Google Scholar] [CrossRef]
  3. Yue, G.; Li, Z.; Tao, Y.; Jin, T. Low-illumination traffic object detection using the saliency region of infrared image masking on infrared-visible fusion image. J. Electron. Imaging 2022, 31, 033029. [Google Scholar] [CrossRef]
  4. Ye, Y.; Shen, L. Hopc: A novel similarity metric based on geometric structural properties for multi-modal remote sensing image matching. ISPRS Ann. Photogramm. Remote Sens. Spatial Inf. Sci. 2016, 3, 9–16. [Google Scholar] [CrossRef]
  5. Abdullah-Al-Wadud, M.; Kabir, M.H.; Dewan, M.A.A.; Chae, O. A dynamic histogram equalization for image contrast enhancement. IEEE Trans. Consum. Electron. 2007, 53, 593–600. [Google Scholar] [CrossRef]
  6. Ibrahim, H.; Kong, N.S.P. Brightness preserving dynamic histogram equalization for image contrast enhancement. IEEE Trans. Consum. Electron. 2007, 53, 1752–1758. [Google Scholar] [CrossRef]
  7. Lee, C.; Lee, C.; Kim, C.S. Contrast enhancement based on layered difference representation of 2D histograms. IEEE Trans. Image Process. 2013, 22, 5372–5384. [Google Scholar] [CrossRef] [PubMed]
  8. Jeong, I.; Lee, C. An optimization-based approach to gamma correction parameter estimation for low-light image enhancement. Multimed. Tools Appl. 2021, 80, 18027–18042. [Google Scholar] [CrossRef]
  9. Li, C.; Tang, S.; Yan, J.; Zhou, T. Low-light image enhancement via pair of complementary gamma functions by fusion. IEEE Access 2020, 8, 169887–169896. [Google Scholar] [CrossRef]
  10. Wang, Y.; Liu, Z.; Liu, J.; Xu, S.; Liu, S. Low-light image enhancement with illumination-aware gamma correction and complete image modelling network. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 13128–13137. [Google Scholar]
  11. Land, E.H. The retinex theory of color vision. Sci. Am. 1977, 237, 108–129. [Google Scholar] [CrossRef]
  12. Jobson, D.J.; Rahman, Z.U.; Woodell, G.A. Properties and performance of a center/surround retinex. IEEE Trans. Image Process. 1997, 6, 451–462. [Google Scholar] [CrossRef]
  13. Jobson, D.J.; Rahman, Z.U.; Woodell, G.A. A multiscale retinex for bridging the gap between color images and the human observation of scenes. IEEE Trans. Image Process. 1997, 6, 965–976. [Google Scholar] [CrossRef] [PubMed]
  14. Fu, X.; Zeng, D.; Huang, Y.; Zhang, X.P.; Ding, X. A weighted variational model for simultaneous reflectance and illumination estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2782–2790. [Google Scholar]
  15. Wei, C.; Wang, W.; Yang, W.; Liu, J. Deep retinex decomposition for low-light enhancement. arXiv 2018, arXiv:1808.04560. [Google Scholar] [CrossRef]
  16. Bychkovsky, V.; Paris, S.; Chan, E.; Durand, F. Learning photographic global tonal adjustment with a database of input/output image pairs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Colorado Springs, CO, USA, 20–25 June 2011; pp. 97–104. [Google Scholar]
  17. Guo, X.; Li, Y.; Ling, H. LIME: Low-light image enhancement via illumination map estimation. IEEE Trans. Image Process. 2016, 26, 982–993. [Google Scholar] [CrossRef]
  18. Zhang, Y.; Zhang, J.; Guo, X. Kindling the darkness: A practical low-light image enhancer. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 1632–1640. [Google Scholar]
  19. Wu, W.; Weng, J.; Zhang, P.; Wang, X.; Yang, W.; Jiang, J. Uretinex-net: Retinex-based deep unfolding network for low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 5901–5910. [Google Scholar]
  20. Chen, Y.; Zhu, G.; Wang, X.; Shen, Y. FMR-Net: A fast multi-scale residual network for low-light image enhancement. Multimed. Syst. 2024, 30, 73. [Google Scholar] [CrossRef]
  21. Jiang, Y.; Gong, X.; Liu, D.; Cheng, Y.; Fang, C.; Shen, X.; Wang, Z. Enlightengan: Deep light enhancement without paired supervision. IEEE Trans. Image Process. 2021, 30, 2340–2349. [Google Scholar] [CrossRef]
  22. Li, C.; Guo, C.; Loy, C.C. Learning to enhance low-light image via zero-reference deep curve estimation. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 4225–4238. [Google Scholar] [CrossRef]
  23. Ma, L.; Ma, T.; Liu, R.; Fan, X.; Luo, Z. Toward fast, flexible, and robust low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 5637–5646. [Google Scholar]
  24. Zhang, F.; Shao, Y.; Sun, Y.; Zhu, K.; Gao, C.; Sang, N. Unsupervised low-light image enhancement via histogram equalization prior. arXiv 2021, arXiv:2112.01766. [Google Scholar] [CrossRef]
  25. Deng, W.; Yuan, H.; Deng, L.; Lu, Z. Reparameterized residual feature network for lightweight image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 1712–1721. [Google Scholar]
  26. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [PubMed]
  27. Wang, Z.; Cun, X.; Bao, J.; Zhou, W.; Liu, J.; Li, H. Uformer: A general u-shaped transformer for image restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 17683–17693. [Google Scholar]
  28. Xiao, J.; Fu, X.; Liu, A.; Wu, F.; Zha, Z.J. Image de-raining transformer. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 12978–12995. [Google Scholar] [CrossRef]
  29. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.H. Restormer: Efficient transformer for high-resolution image restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 5728–5739. [Google Scholar]
  30. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.H. Learning enriched features for fast image restoration and enhancement. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 1934–1948. [Google Scholar] [CrossRef] [PubMed]
  31. Ulyanov, D.; Vedaldi, A.; Lempitsky, V. Improved texture networks: Maximizing quality and diversity in feed-forward stylization and texture synthesis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6924–6932. [Google Scholar]
  32. Chen, G.H.; Yang, C.L.; Xie, S.L. Gradient-based structural similarity for image quality assessment. In Proceedings of the 2006 International Conference on Image Processing, Atlanta, GA, USA, 8–11 October 2006; pp. 2929–2932. [Google Scholar]
  33. Yang, W.; Wang, W.; Huang, H.; Wang, S.; Liu, J. Sparse gradient regularized deep retinex network for robust low-light image enhancement. IEEE Trans. Image Process. 2021, 30, 2072–2086. [Google Scholar] [CrossRef]
  34. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  35. Paszke, A. PyTorch: An imperative style, high-performance deep learning library. arXiv 2019, arXiv:1912.01703. [Google Scholar]
  36. Liang, Z.; Li, C.; Zhou, S.; Feng, R.; Loy, C.C. Iterative Prompt Learning for Unsupervised Backlit Image Enhancement. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 8094–8103. [Google Scholar]
  37. Lim, S.; Kim, W. DSLR: Deep Stacked Laplacian Restorer for Low-Light Image Enhancement. IEEE Trans. Multimed. 2020, 23, 4272–4284. [Google Scholar] [CrossRef]
  38. Liu, R.; Ma, L.; Zhang, J.; Fan, X.; Luo, Z. Retinex-inspired unrolling with cooperative prior architecture search for low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 10561–10570. [Google Scholar]
  39. Moran, S.; Marza, P.; McDonagh, S.; Parisot, S.; Slabaugh, G. Deeplpf: Deep local parametric filters for image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 12826–12835. [Google Scholar]
  40. Bai, J.; Liu, Y.; Zhang, X.; Li, H. Retinexmamba: Retinex-based mamba for low-light image enhancement. In International Conference on Neural Information Processing, Proceedings of the 31st International Conference, ICONIP 2024, Auckland, New Zealand, 2–6 December 2024; Proceedings, Part VIII; Springer Nature: Singapore, 2024. [Google Scholar]
  41. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
  42. Wang, S.; Zheng, J.; Hu, H.M.; Li, B. Naturalness preserved enhancement algorithm for non-uniform illumination images. IEEE Trans. Image Process. 2013, 22, 3538–3548. [Google Scholar] [CrossRef]
  43. Ma, K.; Zeng, K.; Wang, Z. Perceptual quality assessment for multi-exposure image fusion. IEEE Trans. Image Process. 2015, 24, 3345–3356. [Google Scholar] [CrossRef]
  44. Blau, Y.; Mechrez, R.; Timofte, R.; Michaeli, T.; Zelnik-Manor, L. The 2018 PIRM challenge on perceptual image super-resolution. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Munich, Germany, 8–14 September 2018. [Google Scholar]
  45. Mittal, A.; Moorthy, A.K.; Bovik, A.C. No-reference image quality assessment in the spatial domain. IEEE Trans. Image Process. 2012, 21, 4695–4708. [Google Scholar] [CrossRef] [PubMed]
  46. Talab, A.W.; Younis, N.K.; Ahmed, M.R. Analysis Equalization Images Contrast Enhancement and Performance Measurement. Open Access Libr. J. 2024, 11, 1–11. [Google Scholar] [CrossRef]
Figure 1. A comprehensive investigation into the enhancement of micro-optical images, encompassing state-of-the-art methodologies incorporating HEP [24], URetinex [19], and FMR-Net [20], is undertaken through a meticulous quantitative and qualitative analysis (i.e., PSNR/SSIM/parameter). Compared with other similar methods, the method in this paper achieves better enhancement results with fewer model parameters.
Figure 1. A comprehensive investigation into the enhancement of micro-optical images, encompassing state-of-the-art methodologies incorporating HEP [24], URetinex [19], and FMR-Net [20], is undertaken through a meticulous quantitative and qualitative analysis (i.e., PSNR/SSIM/parameter). Compared with other similar methods, the method in this paper achieves better enhancement results with fewer model parameters.
Sensors 25 04789 g001
Figure 2. MSF-ACA model structure.
Figure 2. MSF-ACA model structure.
Sensors 25 04789 g002
Figure 3. (a) Structure of AICEB; (b) structure of AANBlock.
Figure 3. (a) Structure of AICEB; (b) structure of AANBlock.
Sensors 25 04789 g003
Table 1. The LOL-v2 dataset is used to facilitate quantitative comparisons.
Table 1. The LOL-v2 dataset is used to facilitate quantitative comparisons.
MethodsComplexityLOLV2-RealLOLV2-Syn
FLOPs (G)Params (M)PSNR ↑SSIM ↑PSNR ↑SSIM ↑
CLIP-LIT18.240.2715.260.60116.160.666
DSLR5.8814.9317.000.59613.670.623
SCI0.060.000317.300.54016.540.614
RUAS0.830.00318.370.72316.550.652
URetinex1.240.0221.090.85813.100.642
HEP14.071.3218.290.74716.490.649
DeepLPF5.861.7714.100.48016.020.587
FMR-Net102.770.1920.560.73619.090.657
Ours29.970.0221.530.77120.270.716
Table 2. Comparison of quantitative assessment of unpaired datasets.
Table 2. Comparison of quantitative assessment of unpaired datasets.
MethodsDICMLIMEMEFNPEVV
BRI ↓PI ↓BRI ↓PI ↓BRI ↓PI ↓BRI ↓PI ↓BRI ↓PI ↓
CLIP-LIT24.183.5520.433.0720.673.1119.372.9136.005.40
DSLR25.674.0722.686.0122.496.7433.695.0728.356.64
SCI27.923.7025.173.3726.713.2828.883.5322.803.64
RUAS46.885.7034.884.5842.124.9248.975.6535.884.32
URetinex24.543.5629.023.7134.723.6626.093.1522.452.89
HEP25.743.0131.865.7430.383.2829.732.3639.862.98
DeepLPF19.933.5924.704.4522.404.0417.093.0823.754.28
FMR-Net19.632.9128.963.7721.673.2518.012.7017.562.64
Ours14.452.3616.612.7618.312.7025.441.7828.022.32
Table 3. Results of ablation experiments. ‘↓’ means that the higher the value, the worse the results reflected by the corresponding indicator.
Table 3. Results of ablation experiments. ‘↓’ means that the higher the value, the worse the results reflected by the corresponding indicator.
ModelsLG-IFFBAICEB
(Fixed Iteration)
AICEBPSNRSSIMNumber of Iterations ↓Time(s) ↓
Baseline 18.370.66--
A 19.110.67--
B 19.500.67200.22
C 20.270.7110.10.14
Table 4. Ablation studies of the number of AICEB modules.
Table 4. Ablation studies of the number of AICEB modules.
Number of AICEB123
PSNR/SSIM19.79/0.6820.27 (+2.4%)/0.71 (+4.4%)20.20 (+2.0%)/0.71 (+4.4%)
Number of iterations7.310.1 (+38%)26.4 (+261%)
Average running time(s)0.110.14 (+27%)0.22 (+100%)
Table 5. Comparison of different thresholds and number of iterations and PSNRs.
Table 5. Comparison of different thresholds and number of iterations and PSNRs.
Threshold0.000010.000050.00010.00050.001
Number of iterations2019.519 (−3%)16.3 (−14%)15.1 (−20%)
PSNR7.69.99.4 (−5%)8.9 (−10%)8.4 (−15%)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cheng, Z.; Wu, Y.; Tian, F.; Feng, Z.; Li, Y. MSF-ACA: Low-Light Image Enhancement Network Based on Multi-Scale Feature Fusion and Adaptive Contrast Adjustment. Sensors 2025, 25, 4789. https://doi.org/10.3390/s25154789

AMA Style

Cheng Z, Wu Y, Tian F, Feng Z, Li Y. MSF-ACA: Low-Light Image Enhancement Network Based on Multi-Scale Feature Fusion and Adaptive Contrast Adjustment. Sensors. 2025; 25(15):4789. https://doi.org/10.3390/s25154789

Chicago/Turabian Style

Cheng, Zhesheng, Yingdan Wu, Fang Tian, Zaiwen Feng, and Yan Li. 2025. "MSF-ACA: Low-Light Image Enhancement Network Based on Multi-Scale Feature Fusion and Adaptive Contrast Adjustment" Sensors 25, no. 15: 4789. https://doi.org/10.3390/s25154789

APA Style

Cheng, Z., Wu, Y., Tian, F., Feng, Z., & Li, Y. (2025). MSF-ACA: Low-Light Image Enhancement Network Based on Multi-Scale Feature Fusion and Adaptive Contrast Adjustment. Sensors, 25(15), 4789. https://doi.org/10.3390/s25154789

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop