Next Article in Journal
An Exploratory Study on Workover Scenario Understanding Using Prompt-Enhanced Vision-Language Models
Previous Article in Journal
Multivariate Modified Dugum Distribution and Its Applications
Previous Article in Special Issue
Probability Density Function Distance-Based Augmented CycleGAN for Image Domain Translation with Asymmetric Sample Size
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Denoising Autoencoder and Contrast Enhancement for RGB and GS Images with Gaussian Noise

by
Armando Adrián Miranda-González
1,
Alberto Jorge Rosales-Silva
1,*,
Dante Mújica-Vargas
2,
Edwards Ernesto Sánchez-Ramírez
3,
Juan Pablo Francisco Posadas-Durán
1,
Dilan Uriostegui-Hernandez
1,
Erick Velázquez-Lozada
1 and
Francisco Javier Gallegos-Funes
1
1
Escuela Superior de Ingeniería Mecánica y Eléctrica Unidad Zacatenco Sección de Estudios de Posgrado e Investigación, Instituto Politécnico Nacional, Mexico City 07738, Mexico
2
Departamento de Ciencias Computacionales, Tecnológico Nacional de México, Cuernavaca 62490, Mexico
3
Instituto de Investigación y Desarrollo Tecnológico de la Armada de México, Veracruz 95269, Mexico
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(10), 1621; https://doi.org/10.3390/math13101621
Submission received: 6 February 2025 / Revised: 20 March 2025 / Accepted: 12 May 2025 / Published: 15 May 2025

Abstract

:
Robust image processing systems require input images that closely resemble real-world scenes. However, external factors, such as adverse environmental conditions or errors in data transmission, can alter the captured image, leading to information loss. These factors may include poor lighting conditions at the time of image capture or the presence of noise, necessitating procedures to restore the data to a representation as close as possible to the real scene. This research project proposes an architecture based on an autoencoder capable of handling both poor lighting conditions and noise in digital images simultaneously, rather than processing them separately. The proposed methodology has been demonstrated to outperform competing techniques specialized in noise reduction or contrast enhancement. This is supported by both objective numerical metrics and visual evaluations using a validation set with varying lighting characteristics. The results indicate that the proposed methodology effectively restores images by improving contrast and reducing noise without requiring separate processing steps.

1. Introduction

Gaussian noise is one of the most common types of noise observed in digital images, often introduced during image acquisition due to various factors. These factors include adverse lighting conditions, thermal instabilities in camera sensors, and failures in the electronic structure of the camera [1]. Adverse lighting conditions have been shown to degrade the visual quality of digital images and exacerbate the noise inherent to camera sensors, thereby increasing the presence of Gaussian noise in the captured image.
The term “adverse lighting conditions” refers to two specific scenarios: low illumination and high illumination. In the case of low illumination, the resulting digital image exhibits reduced brightness and low contrast [2]. Additionally, within the electronic structure of the camera, sensors exhibit increased sensitivity to weak signals, amplifying any signal variations and resulting in the generation of Gaussian noise. Conversely, under high illumination conditions, the camera sensors become saturated, leading to a nonlinear response in the captured data [3] and contributing to the presence of Gaussian noise. Moreover, the heat generated, along with other types of noise, such as thermal noise, further affects the image quality.
Several techniques have been developed to mitigate the impact of Gaussian noise in images captured under poor lighting conditions. These include both classical image processing methods and advanced deep learning-based approaches [4]. Some of these techniques focus on separately addressing noise reduction and illumination enhancement [5]. However, it is important to distinguish between illumination enhancement and contrast enhancement as they are two distinct techniques in image processing that address different aspects of a visual quality of the image. While illumination enhancement adjusts the overall brightness to improve visibility and make the image clearer and easier to interpret [6], contrast enhancement increases the difference between light and dark areas to make details more perceptible [7]. This research highlights the importance of simultaneously enhancing image details through contrast adjustment and reducing noise using an autoencoder neural network. This is achieved by analyzing the image and applying trained models based on the features present in the input data.
Section 2 presents and analyzes the relevant theoretical background. The proposed model is described in Section 3, while the experimental setup and results are discussed in Section 4. Finally, the conclusions of this research work are provided in Section 5.

2. Background Work

Inadequate lighting during scene capture can lead to several issues that negatively impact the visual quality of an image. These issues range from underexposure or oversaturation to noise generation, as illustrated in Figure 1 (derived from Smartphone Image Denoising Dataset (SIDD) [8]).
Noise in digital images refers to variations in pixel values that distort the original information. This alteration can occur during image capture, digitization, storage, or transmission. Different types of noise are characterized by how they interact with the image and can be identified based on their source. One of the most common types of noise in images is Gaussian noise. Various techniques, both traditional and deep learning-based, have been developed to reduce this type of noise. Traditional approaches include methods such as the median filter [9], while deep learning techniques have an inherent ability to overcome the limitations of certain conventional algorithms, such as the following:
  • Denoising Convolutional Neural Network (DnCNN): This study introduced the design of a Convolutional Neural Network (CNN) specifically developed to reduce Gaussian noise in digital images. The network consists of multiple convolutional layers, followed by batch normalization and ReLU activation functions. A distinctive feature of this architecture is that it does not directly learn to generate a denoised image; instead, it employs residual learning, where the model learns to predict the difference between the noisy image and its denoised counterpart. This approach facilitates learning and improves model convergence [10].
  • Nonlinear Activation Free Network (NAFNet): This study presented an alternative C N N designed for image denoising. The network was developed with the objective of improving architectural efficiency and simplifying computations by minimizing the use of nonlinear activation functions, such as Rectified Linear Unit (ReLU). This design choice enhances computational efficiency while maintaining optimal model performance [11].
  • Restoration Transformer (Restormer): This network is based on the Transformer architecture and is specifically designed for image restoration tasks, including Gaussian noise reduction. This approach leverages the advantages of Transformer networks to optimize memory usage and computational efficiency while simultaneously capturing long-range dependencies [12].
In the context of digital images, contrast is defined as the measure of the difference between the brightness levels of the lightest and darkest areas in an image. Adequate contrast enhances the visual quality of an image, whereas insufficient contrast can result in a visually flat appearance [7]. To address the issue of poor contrast in images, several algorithms have been developed in recent years to enhance contrast. Examples include An Advanced Whale Optimization Algorithm for Grayscale Image Enhancement [13], Pixel Intensity Optimization and Detail-Preserving Contextual Contrast Enhancement for Underwater Images [14], and Optimal Bezier Curve Modification Function for Contrast-Degraded Images [15]. However, these algorithms primarily focus on enhancing specific image channels or improving images under a single lighting condition.
Some algorithms that operate across all image channels and are capable of functioning under extreme lighting conditions, whether in low or high illumination, include the following:
  • Single-Scale Retinex (SSR) is a technique designed to enhance the contrast and illumination of digital images. It is based on human perception of color and luminance in real-world scenes, simulating how the human eye adapts to different lighting conditions by adjusting color perception and scene illumination [16]. The application process of S S R is outlined in Equations (1)–(6).
    I ( x , y ) = R ( x , y ) · L ( x , y ) ,
    where I ( x , y ) represents the original image pixel, which can be decomposed into two components: R ( x , y ) , the reflectance component; and L ( x , y ) , the illumination component.
    To facilitate the distinction between reflectance and illumination, the following logarithmic transformation is applied:
    log I ( x , y ) = log R ( x , y ) + log L ( x , y ) .
    Clarity enhancement is achieved by applying a Gaussian filter to smooth the original image:
    L ( x , y ) L ^ ( x , y ) = F ( x , y ) I ( x , y ) ,
    where F ( x , y ) is the Gaussian filter, and ∗ denotes two-dimensional convolution, which is formally defined as follows:
    ( f g ) ( x , y ) = F ( x α , y β ) I ( α , β ) d α d β .
    The reflectance component R ( x , y ) is obtained by subtracting the illumination component in the logarithmic domain:
    R ^ ( x , y ) = log I ( x , y ) log L ^ ( x , y ) .
    Finally, an inverse transformation is performed to reconstruct the processed image:
    I S S R = exp ( R ^ ( x , y ) ) .
  • Multiscale Retinex (MSR) is an extension of the S S R algorithm, and it was designed to overcome the limitations of Gaussian filter scale sensitivity. Unlike S S R , this algorithm operates across multiple filter scales, utilizing the results from different scales to achieve a balance between local and global details [17]. The computation of M S R is given by Equation (7).
    R M S R ( x , y ) = i = 1 n W i · log I ( x , y ) F i ( x , y ) I ( x , y ) ,
    where R M S R ( x , y ) represents the output value at coordinates ( x , y ) , n is the number of scales, and W i and F i denote the weight and the Gaussian filter at scale i, respectively.
  • Multiscale Retinex with Color Restoration (MSRCR) is an enhancement of the M S R algorithm that combines the contrast and detail enhancement capabilities of M S R with a function designed to preserve the natural colors of the image, thereby preventing the loss of color fidelity [18]. The computation of M S R C R is described in Equations (8) and (9).
    R M S R C R ( x , y ) = C ( x , y ) · R M S R ( x , y ) ,
    C ( x , y ) = β · log 1 + I ( x , y ) k I k ( x , y ) ,
    where R M S R C R ( x , y ) represents the output value at coordinates ( x , y ) , C ( x , y ) is the color restoration function, β is a gain-related constant, and I k ( x , y ) corresponds to the pixel intensity in the k different channels of the image.
  • Multiscale Retinex with Chromaticity Preservation (MSRCP) is a refinement of M S R , and it was designed to preserve the chromaticity of the image while enhancing its contrast and detail. This approach ensures a more faithful representation of the original colors [18]. The computation of M S R C P is described in Equations (10)–(12).
    I A ( x , y ) = R ( x , y ) + G ( x , y ) + B ( x , y ) 3 ,
    C k ( x , y ) = I k ( x , y ) I A ( x , y ) , k R , G , B ,
    R M S R C P ( x , y ) = C k ( x , y ) · R M S R ( x , y ) ,
    where I A ( x , y ) represents the average intensity of the pixel at coordinates ( x , y ) , C k ( x , y ) denotes the chromatic proportions, and R M S R C P ( x , y ) is the resulting value.
  • Gamma correction is a technique used to enhance the brightness and contrast of a digital image. It adjusts the relationship between intensity levels and their perceived brightness, thereby helping to correct distortions [19]. The computation of Gamma correction is given by Equation (13).
    S = I γ ,
    where S represents the output intensity, I is the input intensity, and γ is the correction factor.
  • Histogram equalization is a widely used algorithm for contrast enhancement. It improves contrast by redistributing intensity levels (Equation (14)) so that the histogram approaches a uniform distribution. This process enhances details in low-contrast images.
    S = ( L 1 ) p ( I ) ,
    where L represents the maximum intensity value in the image, and p ( I ) is the probability of an event occurring at that intensity.

3. Proposed Model

The aforementioned algorithms have been shown to prioritize a single objective, either noise reduction or contrast enhancement. However, a new algorithm is proposed that performs both tasks simultaneously, yielding superior results compared to existing methods. The proposed algorithm, Denoising Vanilla Autoencoder with Contrast Enhancement (DVACE), was designed to simultaneously address the noise reduction and contrast enhancement in images represented mathematically as multidimensional arrays. First, consider an original image X, defined as a two-dimensional matrix (Gray Scale (GS) image) or a three-dimensional tensor (Red, Green, Blue (RGB) image), where each matrix entry corresponds to the pixel intensity at position ( i , j ) .
Then, let the original multidimensional image be the following:
X R M × N × C ,
where M × N is the spatial resolution of the image, C is the number of channels ( C = 1 for GS, and C = 3 for RGB).
Considering a multidimensional Gaussian noise model [20], the observed noisy image is expressed as follows:
Y = X + η , with η N ( X ¯ , Σ ) ,
where Y R M × N × C is the observed noisy image, X R M × N × C is the original noise-free image, η R M × N × C is additive Gaussian noise, X ¯ is the multidimensional mean matrix (local or global) of pixels, and Σ is the covariance matrix representing the multidimensional noise dispersion (typically Σ = σ 2 I for stationary, uncorrelated noise between pixels and channels, where I is the multidimensional identity matrix).
For each pixel at a specific position ( i , j ) with observed value y i j (vector for RGB and scalar for GS), the Gaussian noise probability distribution is as follows:
P ( y i j ) = 1 ( 2 π ) C | Σ | exp 1 2 ( y i j x ¯ i j ) T Σ 1 ( y i j x ¯ i j ) ,
where y i j is the column vector (for RGB and C = 3 ), the scalar (GS, C = 1 ) is observed at spatial position ( i , j ) , x ¯ i j is the original local mean at position ( i , j ) , and Σ is the noise covariance matrix (simplified often to Σ = σ 2 I C , with I C as the C × C identity matrix).
If noise is stationary and isotropic (equal in all directions), the equation simplifies to the following:
P ( y i j ) = 1 ( 2 π σ 2 ) C / 2 exp y i j x ¯ i j 2 2 2 σ 2 .
The joint probability for the entire observed image, assuming independence among pixels and channels, is as follows:
P ( Y | X ) = i = 1 M j = 1 N P ( y i j ) = 1 ( 2 π σ 2 ) M N C / 2 exp 1 2 σ 2 i = 1 M j = 1 N y i j x ¯ i j 2 2 .
This provides the mathematical foundation on which the DVACE model optimizes the estimation of the original image X by minimizing the exponential term that represents the squared error between the observed noisy image Y and the restored image X. By adjusting the variance, the density of noise present in the image can be increased or decreased. Additionally, by modifying the mean, the image can appear underexposed (Figure 2a) or overexposed (Figure 2b). This demonstrates how the illumination of the image changes, either darkening or brightening. Finally, the histogram corresponding to the simulated image is presented.
Figure 3 presents the flowchart of the proposed model architecture for RGB images, while Figure 4 shows the flowchart of the proposed model architecture for GS images.
Each architecture calculates the Signal-to-Noise Ratio (SNR) of the input image. The SNR metric [21] is used to enhance the network’s ability to determine the most suitable model—whether to apply a model that brightens dark images or one that darkens bright images—during the actual processing stage. The SNR is defined for an image X R w × h × c , where w , h , and c represent the spatial and channel dimensions. The SNR quantifies the mean intensity relative to the variance in the image:
SNR = E [ X ] Var [ X ] ,
where the mean and the variance are computed as follows:
E [ X ] = 1 w h c i = 0 w j = 0 h k = 0 c X ( i , j , k ) ,
Var [ X ] = 1 w h c i = 0 w j = 0 h k = 0 c X ( i , j , k ) E [ X ] 2 .
This formulation provides a robust measure of the image intensity relative to its noise distribution. It is evident that the design of Algorithms 1 and 2 was based on the proposed architectures.
The SNR thresholds used in both the RGB and GS algorithms were determined experimentally by calculating the average SNR of the corrupted images used to train the network. Equations (23)–(31) illustrate the DVACE procedure.   
Mathematics 13 01621 i193
Mathematics 13 01621 i194
Given a set of images in different modalities ( X G S for GS images and X R G B for RGB images), the classification process based on the SNR can be rigorously expressed as a decision function, which is defined as follows:
X = f Unimodal ( X ) if X R M × N , and S N R ( X ) τ G S g Unimodal ( X ) if X R M × N , and S N R ( X ) > τ G S f Multimodal ( X ) if X R M × N × C , and S N R ( X ) τ R G B g Multimodal ( X ) if X R M × N × C , and S N R ( X ) > τ R G B ,
where X represents the input image; X is the processed image by the DVACE model; R M × N represents the GS image space; R M × N × C represents the RGB image space with C channels; S N R ( X ) is the function computing the SNR of the image; and τ G S = 2.6 and τ R G B = 1.73 are predefined SNR thresholds for GS and RGB images, respectively.
f Unimodal and g Unimodal are the unimodal enhancement functions for GS images, and the following apply:
  • f Unimodal ( X ) is applied to images with low SNR (dark images).
  • g Unimodal ( X ) is applied to images with high SNR (bright images).
f Multimodal and g Multimodal are the multimodal enhancement functions for RGB images, and the following apply:
  • f Multimodal ( X ) is applied to images with low SNR (dark images).
  • g Multimodal ( X ) is applied to images with high SNR (bright images).
The convolutional operation ∗ between an input tensor X and a kernel W R m × n × c × k is defined as follows:
( X W ) ( i , j , k ) = m = 0 M 1 n = 0 N 1 c = 0 C 1 X ( i + m Δ , j + n Δ , c ) · W ( m , n , c , k ) + b k ,
where Δ = M / 2 accounts for padding in the kernel size, and b k is the bias term for channel k.
A non-linear transformation is applied to the convolutional result:
Y ( i , j , k ) = ϕ ( X W ) ( i , j , k ) ,
where the activation function ϕ is defined as follows:
ϕ ( x ) = max ( 0 , x ) ,
this introduces non-linearity, enabling feature extraction from high-dimensional spaces.
Dimensional reduction is performed through max-pooling:
Z ( i , j , k ) = max 0 p P , 0 q Q Y ( i + p , j + q , k ) ,
where P , Q define the pooling window size. This operation selects the most dominant feature per region.
A secondary convolutional pass refines the extracted features:
Z ( i , j , k ) = ϕ Z W ,
where W represents a new set of learned weights.
To restore spatial resolution, we applied weighted bilinear interpolation:
0 p P , 0 q Q w n · Y ( i + p , j + q , k ) = Z ( i , j , k ) ,
where w n are interpolation weights satisfying the following:
n w n = 1 .
A final convolutional step reconstructs the enhanced image as follows:
X ( i , j , k ) = ( Y W ) ( i , j , k ) ,
where W represents a final learned weight set for output feature mapping.
Following the Denoising Vanilla Autoencoder (DVA) training structure and methodology [22], two databases were created using images from the “1 Million Faces” dataset [23], from which only 7000 images were selected.
The first database contains images with a mean intensity x ¯ of {0.01 to 0.5} and σ 2 of 0.01 for bright images, while the second database contains images with a mean intensity x ¯ ranging from {−0.01 to −0.05} and a variance σ 2 of 0.01 for dark images. Each database includes images in both RGB and GS. The implementation details to ensure reproducibility are provided in Table 1.
The learning curves obtained during the training process are illustrated in Figure 5 and Figure 6.

4. Experimental Results

It is essential to recognize that all algorithms require a validation process to assess their effectiveness in comparison to existing methods. To gain a comprehensive understanding of their performance, it is crucial to employ techniques that quantitatively and/or qualitatively evaluate their outcomes.
Therefore, the following quantitative and qualitative quality criteria were used to assess and validate the results obtained by DVACE in comparison to the other specialized techniques discussed in Section 2.
Quantitative metrics provide a means of evaluating the quality of digital images after processing. These metrics can be categorized into reference-based metrics, which compare the processed image against a ground truth, and non-reference metrics, which assess image quality without requiring a reference. The metrics used in this study are as follows:
  • Erreur Relative Globale Adimensionnelle de Synthèse (ERGAS) [22,24].
  • Mean Square Error (MSE) [22].
  • Normalized Color Difference (NCD) estimates the perceptual error between two color vectors by converting from the R G B space to the CIELuv space. This conversion is necessary because human color perception cannot be accurately represented using the RGB model as it is a non-linear space [25]. The perceptual color error between the two color vectors is defined as the Euclidean distance between them, as given by Equation (32).
    Δ E L u v = ( Δ L ) 2 + ( Δ u ) 2 + ( Δ v ) 2 ,
    where Δ E L u v is the error, and Δ L , Δ u , and y Δ v are the difference between the components L , u , and v , respectively, between the two color vectors under consideration.
    Once Δ E L u v was found for each one of the pixels of the images under consideration, the NCD was estimated according to Equation (33).
    N C D = i = 0 M 1 j = 0 N 1 Δ E L u v i = 0 M 1 j = 0 N 1 E L u v ,
    where E L u v = ( L ) 2 + ( u ) 2 + ( v ) 2 is the norm of magnitude of the vector of the pixel of the original image not corrupted in space L u v , and M and N are the dimensions of the image.
  • Perception-based Image Quality Evaluator (PIQE) is a no-reference image quality assessment method that evaluates perceived image quality based on visible distortion levels [26]. Despite being a numerical metric, it is particularly useful for identifying regions of high activity, artifacts, and noise, as it generates masks that indicate the areas where these distortions occur. Consequently, PIQE is also classified as a qualitative metric as it is based on human perception and assesses visual quality from a non-mathematical perspective [26].
    The activity mask of an image is a tool that quantifies the level of detail or complexity in a specific region based on intensity variations. Its computation is derived from Equations (34) and (35).
    G ( x , y ) = G x ( x , y ) 2 + G y ( x , y ) 2 ,
    where G ( x , y ) is the gradient of the image, and G x ( x , y ) y G x ( x , y ) are the derivatives of the image in the position ( x , y ) .
    σ G B i 2 = 1 M 2 ( x , y ) B i ( G ( x , y ) μ G B i ) 2 ,
    where σ G B i 2 is the variance in each of the blocks, and the B i of size M x M y μ G B i is the average of the gradient in the block.
    The artifact mask in an image indicates distortions, such as irregular edges that degrade visual quality. These distortions are detected by analyzing non-natural patterns in regions with high activity levels, where inconsistent blocks are identified and classified as artifacts.
    The noise mask is evaluated based on variations in undesired activity within low-activity regions, measuring the dispersion of intensity values within a block, as shown in Equation (36). If the dispersion significantly exceeds the expected level, the region is classified as noise.
    σ I B i 2 = 1 M 2 ( x , y ) B i ( I ( x , y ) μ I B i ) 2 .
  • Peak Signal-to-Noise Ratio (PSNR) [22,27].
  • Relative Average Spectral Error (RASE) [22,28].
  • Root Mean Squared Error (RMSE) [22,29].
  • Spectral Angle Mapper (SAM) [22,30].
  • Structural Similarity Index (SSIM) [22,31].
  • Universal Quality Image Index (UQI) [22,32].
The DVACE evaluation was performed using classic benchmark images commonly used for algorithm assessment, including Airplane, Baboon, Barbara, Cablecar, Goldhill, Lenna, Mondrian, and Peppers, in both RGB and GS formats. Each evaluation image was corrupted with Gaussian noise, with a variance σ 2 of 0.01 and a mean intensity x ¯ ranging from 0.5 to 0.5 , in increments of 0.01 . Figure 7 presents a close-up of the original peppers in both RGB and GS formats.
As shown in Table 2 and Table 3, the quantitative results for the peppers RGB image are presented for x ¯ = 0.5 and x ¯ = 0.5 , respectively, with σ 2 = 0.1 . It is evident that, in most cases where the mean was nonzero, DVACE achieved superior image restoration.
Similarly, Table 4 and Table 5 present the quantitative results for the peppers GS image for different mean values and σ 2 = 0.1 . It was observed that, in most cases where x ¯ 0 , DVACE achieved superior image restoration compared to all the other algorithms used for comparison.
As shown in Table 4 and Table 5, the peppers image was evaluated under different noise conditions. DVACE consistently achieves the highest SSIM and PSNR, with the lowest MSE, RMSE, and NCD, ensuring optimal noise reduction and contrast enhancement. It also minimized ERGAS, RASE, and SAM, confirming its superior spectral fidelity. Histogram Equalization and Gamma Correction improved contrast but introduced spectral distortions. The deep learning-based methods (DnCNN, NafNet, and Restormer) showed variability, while the MSR-based techniques and SSR exhibited higher error rates. DVACE maintained the best trade-off between denoising and structural fidelity.
Table 6 presents a visual comparison of the results obtained by DVACE and the aforementioned algorithms for both noise reduction and contrast enhancement on the baboon image in RGB with x ¯ = 0.5 . This table illustrates that, while the proposed algorithm introduces some distortions, it achieves the best noise reduction results alongside the NAFNet network. Additionally, in terms of contrast enhancement, DVACE demonstrated superior restoration (comparable to Histogram Equalization).
Table 7 presents a visual comparison for the peppers image in RGB with x ¯ = 0.5 . Visually, DVACE and the median filter exhibited less noise reduction. However, the contrast enhancement achieved by DVACE was comparable to that of the dedicated algorithms designed for this task.
Table 8 presents a comparison for the peppers image in GS with x ¯ = 0.5 . The results indicate that DVACE achieved the best performance in both noise reduction and contrast enhancement.
Table 9 presents a comparison for the peppers GS image with x ¯ = 0.5 , confirming the trend observed with DVACE, which achieved the best results in both noise reduction and contrast enhancement.
As such, in general, Table 6, Table 7, Table 8 and Table 9 provide a visual assessment of DVACE against alternative methods. DVACE, DnCNN, and NAFNet produced cleaner images with well-preserved details, while Histogram Equalization and Gamma Correction enhanced contrast but amplified artifacts. Activity masks show DVACE retained details with minimal distortions. Artifact masks reveal that DVACE introduced fewer distortions than Median and MSRCP, while noise masks confirmed superior noise suppression compared to MSR-based methods and SSR. Overall, DVACE provided the most balanced restoration.
To comprehensively present the results of the metrics calculated from the images in the validation dataset, which were processed by each of the aforementioned methods, box plots are provided below. Figure 8 presents the ERGAS metric distribution across different methods. The noisy image showed the highest values, with DVACE achieving a low median and minimal variance, confirming its stable performance. Histogram Equalization and Gamma Correction also performed well, whereas MSR and MSRCR exhibited higher ERGAS values, indicating weaker global reconstruction. DVACE maintained a consistent advantage with fewer outliers.
Figure 9 illustrates the MSE distribution. The noisy image exhibits high error and dispersion, while DVACE achieved a lower median MSE with reduced variance, ensuring effective reconstruction. The deep learning models (DnCNN and NafNet) showed greater variability, and the MSR-based methods performed inconsistently. DVACE remained one of the most reliable techniques.
Notably, Gamma Correction and Histogram Equalization, despite not being deep learning techniques or having noise reduction capabilities, achieved the next best results. In contrast, SSR demonstrated the poorest performance as both its dispersion and average error were significantly higher than those of the other methods.
As shown in Figure 10, the NCD metric, which reflects color fidelity, was evaluated. DVACE achieved one of the lowest median NCD values with minimal dispersion, confirming its effectiveness in preserving perceptual color accuracy. While Histogram Equalization and Gamma Correction yielded competitive results, it introduce variability. The deep learning methods performed well but with slightly higher dispersion.
Figure 11 presents the PSNR distribution. The noisy image exhibited the lowest values, while DVACE achieved a high median PSNR with low variance, ensuring effective noise reduction and image fidelity. The deep learning models maintained competitive values but showed dataset-dependent behavior. The MSR-based methods performed worse in key metrics.
As shown in Figure 12, the RASE values, which indicate spectral reconstruction accuracy, were captured. The noisy image had the highest values, whereas DVACE maintained a lower median with reduced variance. Histogram Equalization and Gamma Correction achieved good results but exhibited more variability. The deep learning models and MSR-based methods showed inconsistent performance.
Figure 13 illustrates the RMSE values, reflecting the reconstruction accuracy. The noisy image exhibited the highest RMSE, while DVACE achieved a low median with reduced dispersion, confirming its stability. The deep learning models remained competitive but more variable. The MSR-based methods and SSR showed weaker performance.
Figure 14 presents the SAM values, which measure the spectral fidelity. The noisy image showed significant spectral distortions, while DVACE achieved one of the lowest median SAM values, ensuring improved spectral consistency. Histogram Equalization and Gamma Correction performed well but introduced more variability.
As shown in Figure 15, the SSIM, which reflects the image quality, was evaluated. The noisy image had the lowest values, while DVACE achieved a high median with minimal variance, confirming its structural preservation. The deep learning models showed competitive performance, while the MSR-based methods underperformed.
Finally, as shown in Figure 16, the UQI values, which assess the perceptual quality, were recorded. The noisy image exhibited the lowest UQI, while DVACE achieved one of the highest medians with low dispersion, ensuring strong consistency. The deep learning models performed well but exhibited slightly higher variability.
Another critical factor to consider when evaluating the effectiveness of an image restoration method is its execution speed. Table 10 presents the execution times of DVACE for images with dimensions 100 × 100 , 200 × 200 , 400 × 400 , 800 × 800 , 1600 × 1600 , and 3200 × 3200 , all of which were corrupted via Gaussian noise with σ 2 = 0.01 and x ¯ [ 0.5 ,   0.5 ] .
Table 11 compares DVACE with two versions of DnCNN, showing that DVACE maintained competitive execution times, especially for larger image resolutions. For 512 × 512 and 1024 × 1024, DVACE outperformed DnCNN in efficiency, with processing times of 0.049 s and 0.075 s, respectively, demonstrating its advantage in speed without compromising restoration quality.

5. Conclusions

This research highlights the importance of proper image processing in addressing two distinct yet simultaneous challenges that can arise during image capture: poor lighting and noise. Based on this, a methodology is proposed using an autoencoder capable of processing images of any size and type (RGB or GS) under noisy and low-light conditions.
When analyzing the results presented, it was observed that DVACE effectively reduces Gaussian noise in images and enhances their contrast through deep learning techniques implemented in the proposed algorithm, regardless of the average noise level in the degraded images. The results of DVACE, both visually and across various quantitative metrics, demonstrate superior noise reduction and contrast enhancement compared to classical and deep learning-based specialized techniques.
One limitation observed in this research was that DVACE introduces distortions and reduces image activity. Therefore, we recommend using DVACE as a foundation for further improvements (such as integrating a sharpness enhancement algorithm to mitigate distortions and increase image activity).

Author Contributions

Conceptualization, A.A.M.-G., A.J.R.-S., and D.M.-V.; methodology, A.A.M.-G., A.J.R.-S., and D.M.-V.; software, A.A.M.-G., A.J.R.-S., D.M.-V., E.E.S.-R., and J.P.F.P.-D.; validation, E.E.S.-R., D.U.-H., E.V.-L., and F.J.G.-F.; formal analysis, A.A.M.-G., A.J.R.-S., D.M.-V., E.E.S.-R., J.P.F.P.-D., D.U.-H., E.V.-L., and F.J.G.-F.; investigation, A.A.M.-G., A.J.R.-S., D.M.-V., and E.E.S.-R.; writing—original draft preparation, A.A.M.-G.; writing—review and editing, A.A.M.-G., A.J.R.-S., D.M.-V., E.E.S.-R., J.P.F.P.-D., D.U.-H., E.V.-L., and F.J.G.-F.; supervision, A.A.M.-G., A.J.R.-S., and D.M.-V. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original contributions presented in the study are included in the article; further inquiries can be directed to the corresponding author.

Acknowledgments

The authors wish to thank Instituto Politécnico Nacional and Consejo Nacional de Humanidades, Ciencias y Tecnologías for their support in carrying out this research work.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CNNConvolutional Neural Network
DnCNNDenoising Convolutional Neural Network
DVADenoising Vanilla Autoencoder
DVACEDenoising Vanilla Autoencoder with Contrast Enhancement
ERGASErreur Relative Globale Adimensionnelle de Synthèse
GSGray Scale
MSEMean Square Error
MSRMultiscale Retinex
MSRCPMultiscale Retinex with Chromaticity Preservation
MSRCRMultiscale Retinex with Color Restoration
NAFNetNonlinear Activation Free Network
NCDNormalized Color Difference
PIQEPerception-based Image Quality Evaluator
PSNRPeak Signal-to-Noise Ratio
RASERelative Average Spectral Error
ReLURectified Linear Unit
RestormerRestoration Transformer
RGBRed, Green, Blue
RMSERoot Mean Squared Error
SAMSpectral Angle Mapper
SIDDSmartphone Image Denoising Dataset
SNRSignal-to-Noise Ratio
SSIMStructural Similarity Index
SSRSingle-Scale Retinex
UQIUniversal Quality Image Index

References

  1. Kumar-Boyat, A.; Kumar-Joshi, B. A Review Paper: Noise Models in Digital Image Processing. Signal Image Process. Int. J. (SIPIJ) 2015, 6, 63–75. [Google Scholar] [CrossRef]
  2. He, Z.; Ran, W.; Liu, S.; Li, K.; Lu, J.; Xie, C.; Liu, Y.; Lu, H. Low-Light Image Enhancement With Multi-Scale Attention and Frequency-Domain Optimization. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 2861–2875. [Google Scholar] [CrossRef]
  3. Wang, Z.; Bovik, A.; Evan, B. Blind measurement of blocking artifacts in images. In Proceedings of the 2000 International Conference on Image Processing, Vancouver, BC, Canada, 10–13 September 2000; Volume 3, pp. 981–984. [Google Scholar] [CrossRef]
  4. Akyüz, A.O.; Reinhard, E. Noise reduction in high dynamic range imaging. J. Vis. Commun. Image Represent. 2007, 18, 366–376. [Google Scholar] [CrossRef]
  5. Ren, W.; Liu, S.; Ma, L.; Xu, Q.; Xu, X.; Cao, X.; Du, J.; Yang, M.H. Low-Light Image Enhancement via a Deep Hybrid Network. IEEE Trans. Image Process. 2019, 28, 4364–4375. [Google Scholar] [CrossRef] [PubMed]
  6. Pitas, I. Digital Image Processing Algorithms and Applications; Wiley-Interscience: Hoboken, NJ, USA, 2000; Volume 1, p. 432. [Google Scholar]
  7. Wang, S.; Zheng, J.; Hu, H.M.; Li, B. Naturalness Preserved Enhancement Algorithm for Non-Uniform Illumination Images. IEEE Trans. Image Process. 2013, 22, 3538–3548. [Google Scholar] [CrossRef] [PubMed]
  8. Abdelhamed, A.; Lin, S.; Brown, M.S. A High-Quality Denoising Dataset for Smartphone Cameras. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar] [CrossRef]
  9. Smith, S.; Brady, J.M. Image processing of multiphase images obtained via X-ray microtomography: A review. Int. J. Comput. Vis. 1997, 23, 45–78. [Google Scholar] [CrossRef]
  10. Zhang, K.; Zou, W.; Chen, Y. Beyond a Gaussian Denoiser: Residual Learning of Deep CNN for Image Denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef] [PubMed]
  11. Chen, L.; Chu, X.; Zhang, X. Simple Baselines for Image Restoration. In Computer Vision–ECCV 2022, Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2022; pp. 17–33. [Google Scholar] [CrossRef]
  12. Waqas-Zamir, S.; Arora, A.; Khan, S. Restormer: Efficient Transformer for High-Resolution Image Restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 1–13. [Google Scholar] [CrossRef]
  13. Han, Y.; Hu, P.; Su, Z.; Liu, L.; Panneerselvam, J. An Advanced Whale Optimization Algorithm for Grayscale Image Enhancement. Biomimetics 2024, 9, 760. [Google Scholar] [CrossRef]
  14. Subramani, B.; Veluchamy, M. Pixel intensity optimization and detail-preserving contextual contrast enhancement for underwater images. Opt. Laser Technol. 2025, 180, 111464. [Google Scholar] [CrossRef]
  15. Subramani, B.; Bhandari, A.K.; Veluchamy, M. Optimal Bezier Curve Modification Function for Contrast Degraded Images. IEEE Trans. Instrum. Meas. 2021, 70, 1–10. [Google Scholar] [CrossRef]
  16. Jobson, D.J.; Rahman, Z.; Woodell, G. Properties and performance of a center/surround retinex. IEEE Trans. Image Process. 1997, 6, 451–462. [Google Scholar] [CrossRef] [PubMed]
  17. Jobson, D.J.; Rahman, Z.; Woodell, G.A. A multiscale retinex for bridging the gap between color images and the human observation of scenes. IEEE Trans. Image Process. 1997, 6, 965–976. [Google Scholar] [CrossRef]
  18. Petro, A.; Shert, C.; Morel, J. Multiscale Retinex; Image Processing On Line: Paris, France, 2014; pp. 71–88. [Google Scholar] [CrossRef]
  19. Poynton, C.A. Gamma and Its Disguises: The Nonlinear Mappings of Intensity in Perception, CRTs, Film, and Video. Smpte J. 1993, 102, 1099–1108. [Google Scholar] [CrossRef]
  20. Kaur, S. Noise types and various removal techniques. Int. J. Adv. Res. Electron. Commun. Eng. 2015, 4, 226–230. [Google Scholar]
  21. Collins, C.M. Fundamentals of Signal-to-Noise Ratio (SNR). In Electromagnetics in Magnetic Resonance Imaging; Morgan and Claypool Publishers: Bristol, UK, 2016; pp. 1–9. [Google Scholar] [CrossRef]
  22. Miranda-González, A.A.; Rosales-Silva, A.J.; Mújica-Vargas, D.; Escamilla-Ambrosio, P.J.; Gallegos-Funes, F.J.; Vianney-Kinani, J.M.; Velázquez-Lozada, E.; Pérez-Hernández, L.M.; Lozano-Vázquez, L.V. Denoising Vanilla Autoencoder for RGB and GS Images with Gaussian Noise. Entropy 2023, 25, 1467. [Google Scholar] [CrossRef]
  23. Bojan, T. 1 Million Faces. Kaggle. 2020. Available online: https://www.kaggle.com/competitions/deepfake-detection-challenge/discussion/121173 (accessed on 4 June 2024).
  24. Du, Q.; Younan, N.H.; King, R. On the Performance Evaluation of Pan-Sharpening Techniques. IEEE Geosci. Remote Sens. Lett. 2007, 4, 518–522. [Google Scholar] [CrossRef]
  25. Poynton, C. Poynton’s Color FAQ. Electronic Preprint. 1995, p. 24. Available online: https://poynton.ca/ (accessed on 1 June 2024).
  26. Venkatanath, N.; Praneeth, D.; Bh, M.C.; Channappayya, S.S.; Medasani, S.S. Blind image quality evaluation using perception based features. In Proceedings of the 2015 Twenty First National Conference on Communications (NCC), Mumbai, India, 27 February–1 March 2015; pp. 1–6. [Google Scholar] [CrossRef]
  27. Naidu, V. Discrete Cosine Transform-based Image Fusion. J. Commun. Navig. Signal Process. 2012, 1, 35–45. [Google Scholar]
  28. Panchal, S.; Thakker, R. Implementation and comparative quantitative assessment of different multispectral image pansharpening approaches. Signal Image Process. Int. J. (SIPIJ) 2015, 6, 35–48. [Google Scholar] [CrossRef]
  29. Liviu-Florin, Z. Quality Evaluation of Multiresolution Remote Sensing Image Fusion. UPB Sci. Bull. 2009, 71, 37–52. [Google Scholar]
  30. Alparone, L.; Wald, L.; Chanussot, J. Comparison of Pansharpening Algorithms: Outcome of the 2006 GRS-S Data-Fusion Contest. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3012–3021. [Google Scholar] [CrossRef]
  31. Wang, Z.; Bovik, A.C.; Sheikh, H. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  32. Alparone, L.; Aiazzi, B.; Baronti, S. Multispectral and Panchromatic Data Fusion Assessment Without Reference. Photogramm. Eng. Remote Sens. 2008, 2, 193–200. [Google Scholar] [CrossRef]
Figure 1. The noise present in saturated and underexposed images.
Figure 1. The noise present in saturated and underexposed images.
Mathematics 13 01621 g001
Figure 2. Image simulation: (a) underexposed with its respective Gaussian distribution, with x ¯ = 0.3 and σ 2 = 0.01 , and its resulting histogram when corrupted; (b) saturated with its respective Gaussian distribution, with x ¯ = 0.3 and σ 2 = 0.01 , and its resulting histogram when corrupted.
Figure 2. Image simulation: (a) underexposed with its respective Gaussian distribution, with x ¯ = 0.3 and σ 2 = 0.01 , and its resulting histogram when corrupted; (b) saturated with its respective Gaussian distribution, with x ¯ = 0.3 and σ 2 = 0.01 , and its resulting histogram when corrupted.
Mathematics 13 01621 g002
Figure 3. Flowchart of the proposed DVACE for RGB images.
Figure 3. Flowchart of the proposed DVACE for RGB images.
Mathematics 13 01621 g003
Figure 4. Flowchart of the proposed DVACE for GS images.
Figure 4. Flowchart of the proposed DVACE for GS images.
Mathematics 13 01621 g004
Figure 5. Learning curves of the algorithm DVACE for the GS images, (a) Unimodal model for dark images, (b) Unimodal model for light images.
Figure 5. Learning curves of the algorithm DVACE for the GS images, (a) Unimodal model for dark images, (b) Unimodal model for light images.
Mathematics 13 01621 g005
Figure 6. Learning curves of the algorithm DVACE for the RGB images, (a) Multimodal model R for dark images, (b) Multimodal model G for dark images, (c) Multimodal model B for dark images, (d) Multimodal model R for light images, (e) Multimodal model R for light images, (f) Multimodal model R for light images.
Figure 6. Learning curves of the algorithm DVACE for the RGB images, (a) Multimodal model R for dark images, (b) Multimodal model G for dark images, (c) Multimodal model B for dark images, (d) Multimodal model R for light images, (e) Multimodal model R for light images, (f) Multimodal model R for light images.
Mathematics 13 01621 g006
Figure 7. Close-up image of the original peppers.
Figure 7. Close-up image of the original peppers.
Mathematics 13 01621 g007
Figure 8. Box plots of the quantitative ERGAS results obtained.
Figure 8. Box plots of the quantitative ERGAS results obtained.
Mathematics 13 01621 g008
Figure 9. Box plots of the quantitative MSE results obtained.
Figure 9. Box plots of the quantitative MSE results obtained.
Mathematics 13 01621 g009
Figure 10. Box plots of the quantitative NCD results obtained.
Figure 10. Box plots of the quantitative NCD results obtained.
Mathematics 13 01621 g010
Figure 11. Box plots of the quantitative PSNR results obtained.
Figure 11. Box plots of the quantitative PSNR results obtained.
Mathematics 13 01621 g011
Figure 12. Box plots of the quantitative RASE results obtained.
Figure 12. Box plots of the quantitative RASE results obtained.
Mathematics 13 01621 g012
Figure 13. Box plots of the quantitative RMSE results obtained.
Figure 13. Box plots of the quantitative RMSE results obtained.
Mathematics 13 01621 g013
Figure 14. Box plots of the quantitative SAM results obtained.
Figure 14. Box plots of the quantitative SAM results obtained.
Mathematics 13 01621 g014
Figure 15. Box plots of the quantitative SSIM results obtained.
Figure 15. Box plots of the quantitative SSIM results obtained.
Mathematics 13 01621 g015
Figure 16. Box plots of the quantitative UQI results obtained.
Figure 16. Box plots of the quantitative UQI results obtained.
Mathematics 13 01621 g016
Table 1. Hyperparameter and training setup.
Table 1. Hyperparameter and training setup.
Hyperparameters and Training Setup.
Image size420 × 420
Seed17
Learning rate0.001
Shuffletrue
OtimizerAdam
Loss functionMSE
Epochs100
Batch size50
Validation split0.1
Table 2. Quantitative results for the peppers image in RGB with x ¯ = 0.5 .
Table 2. Quantitative results for the peppers image in RGB with x ¯ = 0.5 .
ImageSSIMNCDMSEPSNRRMSEUQIERGASRASESAM
Noisy0.5360.88911,9117.371109.1370.45525,52537010.457
DVACE0.8320.402180715.56242.5070.63916,47323710.271
Median0.5610.88311,8227.404108.7280.45425,34936760.453
DnCNN0.5560.87711,7417.434108.3540.45625,34936760.454
Nafnet0.5550.87812,0437.323109.7430.45425,43336880.460
Restormer0.5190.89111,9157.370109.1580.45525,64437180.461
SSR0.4180.92413,2386.913115.0570.45526,48238370.497
MSR0.3690.88111,3987.562106.7620.47126,34738150.499
MSRCP0.5290.88811,8647.389108.9200.45525,55737060.459
MSRCR0.4030.90496708.27798.3360.48225,62837080.476
Gamma0.7080.664495211.18370.3740.53321,19630590.327
Histogram0.7580.514253814.08650.3790.61019,43227920.300
Table 3. Quantitative results for the peppers image in RGB with x ¯ = 0.5 .
Table 3. Quantitative results for the peppers image in RGB with x ¯ = 0.5 .
ImageSSIMNCDMSEPSNRRMSEUQIERGASRASESAM
Noisy0.3330.83971449.59284.5190.1951,349,244inf0.602
DVACE0.8070.28799818.13931.5920.72121,44930210.324
Median0.3320.84271769.57284.7140.1941,288,543inf0.545
DnCNN0.3390.84270709.63784.0830.2111,058,213inf0.494
Nafnet0.3510.84369299.72483.2400.240299,24138,8450.592
Restormer0.3440.83469139.73483.1430.2321,145,639inf0.585
SSR0.7410.273150416.35838.7820.64646,22365260.413
MSR0.7410.279149716.37738.6960.65634,83449390.407
MSRCP0.7420.436200715.10444.8050.507376,226inf0.563
MSRCR0.7610.237116617.46234.1540.65349,32373240.388
Gamma0.6350.580293113.46154.1340.505453,785inf0.433
Histogram0.6030.389433711.75965.8570.59120,97830490.438
Table 4. Quantitative results for the peppers image in GS with x ¯ = 0.5 .
Table 4. Quantitative results for the peppers image in GS with x ¯ = 0.5 .
ImageSSIMNCDMSEPSNRRMSEUQIERGASRASESAM
Noisy0.2862.68513,5676.806116.4750.52814,24835620.381
DVACE0.7140.935252614.10750.2560.792940023500.186
Median0.4582.68113,3956.861115.7360.52814,09535240.371
DnCNN0.5842.63313,1226.951114.5510.53214,01335030.370
Nafnet0.5762.73213,8406.720117.6420.52414,14235360.378
Restormer0.2762.66313,4406.847115.9300.53014,24535610.381
SSR0.2362.54113,8606.713117.7290.53313,90234760.433
MSR0.2821.84297938.22198.9620.58812,75131880.434
MSRCP0.2822.66713,4716.837116.0670.52914,22635560.381
MSRCR0.2821.84698218.20999.0990.58712,75831890.435
Gamma0.2521.94281079.04290.0370.61613,41633540.308
Histogram0.2071.018312313.18555.8860.81612,98632470.278
Table 5. Quantitative results for the peppers image in GS with x ¯ = 0.5 .
Table 5. Quantitative results for the peppers image in GS with x ¯ = 0.5 .
ImageSSIMNCDMSEPSNRRMSEUQIERGASRASESAM
Noisy0.0390.9128,6308.77192.8960.0862,489,114inf0.706
DVACE0.5380.329167315.89740.8990.56061,80215,4500.330
Median0.0850.92686798.74693.1590.0822,451,183inf0.654
DnCNN0.1450.92584568.85991.9540.0871,877,656inf0.607
Nafnet0.1620.91781389.02690.2080.099781,272inf0.580
Restormer0.0400.90985438.81592.4270.0891,708,935inf0.701
SSR0.1360.435315013.14756.1290.449126,33331,5830.500
MSR0.1830.429285913.56853.4730.50067,11516,7790.474
MSRCP0.0680.484348012.71558.9930.415268,579inf0.528
MSRCR0.0750.448360612.56160.0490.383417,730inf0.538
Gamma0.0760.698505911.09071.1290.256562,233inf0.565
Histogram0.3911.254554410.69274.4590.68011,32728320.345
Table 6. Qualitative results for the peppers image in RGB.
Table 6. Qualitative results for the peppers image in RGB.
FeatureNoisy ImageDVACEMedianDnCNNNAFNetRestormer
x ¯ = 0.5 Mathematics 13 01621 i001Mathematics 13 01621 i002Mathematics 13 01621 i003Mathematics 13 01621 i004Mathematics 13 01621 i005Mathematics 13 01621 i006
Activity maskMathematics 13 01621 i007Mathematics 13 01621 i008Mathematics 13 01621 i009Mathematics 13 01621 i010Mathematics 13 01621 i011Mathematics 13 01621 i012
Artifact maskMathematics 13 01621 i013Mathematics 13 01621 i014Mathematics 13 01621 i015Mathematics 13 01621 i016Mathematics 13 01621 i017Mathematics 13 01621 i018
Noise MaskMathematics 13 01621 i019Mathematics 13 01621 i020Mathematics 13 01621 i021Mathematics 13 01621 i022Mathematics 13 01621 i023Mathematics 13 01621 i024
FeatureSSRMSRMSRCPMSRCRGammaHistogram Eq.
x ¯ = 0.5 Mathematics 13 01621 i025Mathematics 13 01621 i026Mathematics 13 01621 i027Mathematics 13 01621 i028Mathematics 13 01621 i029Mathematics 13 01621 i030
Activity MaskMathematics 13 01621 i031Mathematics 13 01621 i032Mathematics 13 01621 i033Mathematics 13 01621 i034Mathematics 13 01621 i035Mathematics 13 01621 i036
Artifact MaskMathematics 13 01621 i037Mathematics 13 01621 i038Mathematics 13 01621 i039Mathematics 13 01621 i040Mathematics 13 01621 i041Mathematics 13 01621 i042
Noise MaskMathematics 13 01621 i043Mathematics 13 01621 i044Mathematics 13 01621 i045Mathematics 13 01621 i046Mathematics 13 01621 i047Mathematics 13 01621 i048
Table 7. Qualitative results for the peppers image in RGB.
Table 7. Qualitative results for the peppers image in RGB.
FeatureNoisy ImageDVACEMedianDnCNNNAFNetRestormer
x ¯ = 0.5 Mathematics 13 01621 i049Mathematics 13 01621 i050Mathematics 13 01621 i051Mathematics 13 01621 i052Mathematics 13 01621 i053Mathematics 13 01621 i054
Activity MaskMathematics 13 01621 i055Mathematics 13 01621 i056Mathematics 13 01621 i057Mathematics 13 01621 i058Mathematics 13 01621 i059Mathematics 13 01621 i060
Artifact MaskMathematics 13 01621 i061Mathematics 13 01621 i062Mathematics 13 01621 i063Mathematics 13 01621 i064Mathematics 13 01621 i065Mathematics 13 01621 i066
Noise MaskMathematics 13 01621 i067Mathematics 13 01621 i068Mathematics 13 01621 i069Mathematics 13 01621 i070Mathematics 13 01621 i071Mathematics 13 01621 i072
FeatureSSRMSRMSRCPMSRCRGammaHistogram Eq.
x ¯ = 0.5 Mathematics 13 01621 i073Mathematics 13 01621 i074Mathematics 13 01621 i075Mathematics 13 01621 i076Mathematics 13 01621 i077Mathematics 13 01621 i078
Activity MaskMathematics 13 01621 i079Mathematics 13 01621 i080Mathematics 13 01621 i081Mathematics 13 01621 i082Mathematics 13 01621 i083Mathematics 13 01621 i084
Artifact MaskMathematics 13 01621 i085Mathematics 13 01621 i086Mathematics 13 01621 i087Mathematics 13 01621 i088Mathematics 13 01621 i089Mathematics 13 01621 i090
Noise MaskMathematics 13 01621 i091Mathematics 13 01621 i092Mathematics 13 01621 i093Mathematics 13 01621 i094Mathematics 13 01621 i095Mathematics 13 01621 i096
Table 8. Qualitative results for the peppers image in GS.
Table 8. Qualitative results for the peppers image in GS.
FeatureNoisy ImageDVACEMedianDnCNNNAFNetRestormer
x ¯ = 0.5 Mathematics 13 01621 i097Mathematics 13 01621 i098Mathematics 13 01621 i099Mathematics 13 01621 i100Mathematics 13 01621 i101Mathematics 13 01621 i102
Activity MaskMathematics 13 01621 i103Mathematics 13 01621 i104Mathematics 13 01621 i105Mathematics 13 01621 i106Mathematics 13 01621 i107Mathematics 13 01621 i108
Artifact MaskMathematics 13 01621 i109Mathematics 13 01621 i110Mathematics 13 01621 i111Mathematics 13 01621 i112Mathematics 13 01621 i113Mathematics 13 01621 i114
Noise MaskMathematics 13 01621 i115Mathematics 13 01621 i116Mathematics 13 01621 i117Mathematics 13 01621 i118Mathematics 13 01621 i119Mathematics 13 01621 i120
FeatureSSRMSRMSRCPMSRCRGammaHistogram Eq.
x ¯ = 0.5 Mathematics 13 01621 i121Mathematics 13 01621 i122Mathematics 13 01621 i123Mathematics 13 01621 i124Mathematics 13 01621 i125Mathematics 13 01621 i126
Activity MaskMathematics 13 01621 i127Mathematics 13 01621 i128Mathematics 13 01621 i129Mathematics 13 01621 i130Mathematics 13 01621 i131Mathematics 13 01621 i132
Artifact MaskMathematics 13 01621 i133Mathematics 13 01621 i134Mathematics 13 01621 i135Mathematics 13 01621 i136Mathematics 13 01621 i137Mathematics 13 01621 i138
Noise MaskMathematics 13 01621 i139Mathematics 13 01621 i140Mathematics 13 01621 i141Mathematics 13 01621 i142Mathematics 13 01621 i143Mathematics 13 01621 i144
Table 9. Qualitative results for the peppers image in GS.
Table 9. Qualitative results for the peppers image in GS.
FeatureNoisy ImageDVACEMedianDnCNNNAFNetRestormer
x ¯ = 0.5 Mathematics 13 01621 i145Mathematics 13 01621 i146Mathematics 13 01621 i147Mathematics 13 01621 i148Mathematics 13 01621 i149Mathematics 13 01621 i150
Activity MaskMathematics 13 01621 i151Mathematics 13 01621 i152Mathematics 13 01621 i153Mathematics 13 01621 i154Mathematics 13 01621 i155Mathematics 13 01621 i156
Artifact MaskMathematics 13 01621 i157Mathematics 13 01621 i158Mathematics 13 01621 i159Mathematics 13 01621 i160Mathematics 13 01621 i161Mathematics 13 01621 i162
Noise MaskMathematics 13 01621 i163Mathematics 13 01621 i164Mathematics 13 01621 i165Mathematics 13 01621 i166Mathematics 13 01621 i167Mathematics 13 01621 i168
FeatureSSRMSRMSRCPMSRCRGammaHistogram Eq.
x ¯ = 0.5 Mathematics 13 01621 i169Mathematics 13 01621 i170Mathematics 13 01621 i171Mathematics 13 01621 i172Mathematics 13 01621 i173Mathematics 13 01621 i174
Activity MaskMathematics 13 01621 i175Mathematics 13 01621 i176Mathematics 13 01621 i177Mathematics 13 01621 i178Mathematics 13 01621 i179Mathematics 13 01621 i180
Artifact MaskMathematics 13 01621 i181Mathematics 13 01621 i182Mathematics 13 01621 i183Mathematics 13 01621 i184Mathematics 13 01621 i185Mathematics 13 01621 i186
Noise MaskMathematics 13 01621 i187Mathematics 13 01621 i188Mathematics 13 01621 i189Mathematics 13 01621 i190Mathematics 13 01621 i191Mathematics 13 01621 i192
Table 10. The average processing time for different images sizes, noise density, and image type.
Table 10. The average processing time for different images sizes, noise density, and image type.
Size100 × 100200 × 200400 × 400800 × 8001600 × 16003200 × 3200
RGB0.029 s0.032 s0.044 s0.058 s0.115 s0.292 s
GS0.085 s0.098 s0.140 s0.183 s0.342 s0.964 s
Table 11. Comparison of the processing time between DVACE and DnCNN of the images in GS.
Table 11. Comparison of the processing time between DVACE and DnCNN of the images in GS.
MethodsDnCNN-SDnCNN-BDVACE
256 × 2560.014 s0.016 s0.038 s
512 × 5120.051 s0.060 s0.049 s
1024 × 10240.200 s0.235 s0.075 s
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Miranda-González, A.A.; Rosales-Silva, A.J.; Mújica-Vargas, D.; Sánchez-Ramírez, E.E.; Posadas-Durán, J.P.F.; Uriostegui-Hernandez, D.; Velázquez-Lozada, E.; Gallegos-Funes, F.J. Denoising Autoencoder and Contrast Enhancement for RGB and GS Images with Gaussian Noise. Mathematics 2025, 13, 1621. https://doi.org/10.3390/math13101621

AMA Style

Miranda-González AA, Rosales-Silva AJ, Mújica-Vargas D, Sánchez-Ramírez EE, Posadas-Durán JPF, Uriostegui-Hernandez D, Velázquez-Lozada E, Gallegos-Funes FJ. Denoising Autoencoder and Contrast Enhancement for RGB and GS Images with Gaussian Noise. Mathematics. 2025; 13(10):1621. https://doi.org/10.3390/math13101621

Chicago/Turabian Style

Miranda-González, Armando Adrián, Alberto Jorge Rosales-Silva, Dante Mújica-Vargas, Edwards Ernesto Sánchez-Ramírez, Juan Pablo Francisco Posadas-Durán, Dilan Uriostegui-Hernandez, Erick Velázquez-Lozada, and Francisco Javier Gallegos-Funes. 2025. "Denoising Autoencoder and Contrast Enhancement for RGB and GS Images with Gaussian Noise" Mathematics 13, no. 10: 1621. https://doi.org/10.3390/math13101621

APA Style

Miranda-González, A. A., Rosales-Silva, A. J., Mújica-Vargas, D., Sánchez-Ramírez, E. E., Posadas-Durán, J. P. F., Uriostegui-Hernandez, D., Velázquez-Lozada, E., & Gallegos-Funes, F. J. (2025). Denoising Autoencoder and Contrast Enhancement for RGB and GS Images with Gaussian Noise. Mathematics, 13(10), 1621. https://doi.org/10.3390/math13101621

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop