Next Article in Journal
ScalableDigitalHealth (SDH): An IoT-Based Scalable Framework for Remote Patient Monitoring
Previous Article in Journal
Jumping Motor Skills in Typically Developing Preschool Children Assessed Using a Battery of Tests
Previous Article in Special Issue
Remote Sensing Retrieval of Cloud Top Height Using Neural Networks and Data from Cloud-Aerosol Lidar with Orthogonal Polarization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

VELIE: A Vehicle-Based Efficient Low-Light Image Enhancement Method for Intelligent Vehicles

1
Department of Electrical and Electronic Engineering, University of Liverpool, Liverpool L69 3BX, UK
2
School of Advance Technology, Xi’an Jiaotong-Liverpool University, 111 Ren’ai Road, Suzhou 215123, China
3
Department of Electrical and Computer Engineering, The University of Texas at Austin, Austin, TX 78712, USA
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2024, 24(4), 1345; https://doi.org/10.3390/s24041345
Submission received: 25 January 2024 / Revised: 9 February 2024 / Accepted: 14 February 2024 / Published: 19 February 2024
(This article belongs to the Special Issue Advances in Remote Sensing Image Enhancement and Classification)

Abstract

:
In Advanced Driving Assistance Systems (ADAS), Automated Driving Systems (ADS), and Driver Assistance Systems (DAS), RGB camera sensors are extensively utilized for object detection, semantic segmentation, and object tracking. Despite their popularity due to low costs, RGB cameras exhibit weak robustness in complex environments, particularly underperforming in low-light conditions, which raises a significant concern. To address these challenges, multi-sensor fusion systems or specialized low-light cameras have been proposed, but their high costs render them unsuitable for widespread deployment. On the other hand, improvements in post-processing algorithms offer a more economical and effective solution. However, current research in low-light image enhancement still shows substantial gaps in detail enhancement on nighttime driving datasets and is characterized by high deployment costs, failing to achieve real-time inference and edge deployment. Therefore, this paper leverages the Swin Vision Transformer combined with a gamma transformation integrated U-Net for the decoupled enhancement of initial low-light inputs, proposing a deep learning enhancement network named Vehicle-based Efficient Low-light Image Enhancement (VELIE). VELIE achieves state-of-the-art performance on various driving datasets with a processing time of only 0.19 s, significantly enhancing high-dimensional environmental perception tasks in low-light conditions.

1. Introduction

In Advanced Driving Assistance Systems (ADAS), Automatic Driving Systems (ADS), and Driver Assistance Systems (DAS), RGB camera sensors are extensively employed due to their low cost and high information density. RGB imaging in photos and videos plays a crucial role in various intelligent driving tasks such as depth estimation, object detection, semantic segmentation, object tracking, and 3D high-dimension mapping. Current research in intelligent driving vehicle perception systems has evolved into systems comprising pure RGB sensing, RGB-guided Lidar and Radar perception, and multi-sensor fusion systems incorporating RGB [1,2,3]. Favored for their low manufacturing and deployment costs, RGB cameras, however, face significant limitations in robustness under complex environments, especially in low-light conditions. Nighttime and low-illumination driving scenarios, unavoidable for vehicles, however, diminish human reaction and perception capabilities, both visually and cognitively, necessitating enhanced perception aids. The limited nighttime performance of RGB not only threatens its reliability in intelligent driving but also challenges the accuracy and safety of intelligent driving systems.
The fundamental reason for the poor nighttime performance of RGB cameras lies in their imaging principle. Relying on visible light for image capture, RGB imaging effectiveness is highly susceptible to external light sources. At nighttime, the absence of natural light sources like sunlight results in weak environmental perception, making it challenging for cameras to capture clear signals. The vehicle’s own light sources, such as headlights, cannot compensate for this global illumination deficiency. Additionally, the sensitivity of RGB camera sensors to light diminishes significantly under low-light conditions. This indicates that the signal-to-noise ratio of the camera sensor decreases under dim lighting, capturing weaker effective light signals relative to the noise generated by the sensor and circuits.
Beyond the deficits in image clarity and noise, low-light conditions also degrade the color capture capability of vehicle-mounted RGB cameras. RGB cameras capture colors through a tri-color filter array (red, green, blue), but in poorly lit environments, the sensor’s ability to capture these colors weakens, causing color distortion or insufficient color information. Furthermore, the dynamic range of nighttime environments, which indicates the measuring range from darkest to brightest areas, often exceeds the processing capabilities of standard vehicle-mounted RGB sensors, resulting in the loss of detailed information. Such high-frequency information, like the contours of objects, colors, ground markings, and traffic light signals, serve as essential recognition features for autonomous and assisted driving; the loss of this information in dark areas significantly increases the likelihood of accidents [4].
To counter the deficiencies of RGB cameras in darkness, past research approaches have been divided into two aspects: hardware improving development and software image signal processing. On the hardware side, the use of sensors like Lidar, event cameras, radar, and thermal [5,6,7,8], which do not primarily rely on illumination for signal acquisition, has been proposed to assist RGB in intelligent automotive tasks. However, the fusion of multiple sensors greatly diminishes the low-cost advantage of RGB cameras. Both in terms of deployment and sensor hardware costs, the cost of using multi-sensor fusion is significantly higher than the cost of multiple RGB cameras. On the other hand, professional low-light cameras produced by companies like Sony, Photonis, SiOnyx, and Texas Instruments have also emerged in the market, but due to their high prices, they are not suitable for deployment in driving scenarios. Professional low-light cameras [9,10] utilize high-performance Charge-Coupled Devices (CCD) or Complementary Metal-Oxide-Semiconductor (CMOS) technology, professional low-light circuits, and filters as core components to enhance low-light imaging quality; hence, their manufacturing process is highly stringent, technically complex, and very costly. Conversely, improving software algorithms offers a more cost-effective alternative, namely enhancing images through preprocessing of the original RGB images. Thus, some research preprocesses the raw data from RGB cameras and then utilizes these enhanced data for high-dimensional perception tasks. This task on low-light images is collectively referred to as Low-Light Image Enhancement (LLIE).
Research on Low-Light Image Enhancement (LLIE) can primarily be categorized into three distinct approaches: traditional LLIE methods, deep learning-based LLIE methods, and LLIE methods tailored for driving scenarios. Each of these categories has been the subject of extensive investigation within their respective domains. Upon reviewing past research on LLIE tasks in driving scenarios, we have identified critical limitations in previous work: (1) General LLIE approaches have not overcome issues of poor generalization on nighttime driving datasets: the inferred images exhibit artifacts, noise, and severe color distortions. (2) Only a few DL-based Low-Light Image Enhancement (LLIE) studies have considered deployment costs (computational resources, inference time), which is not conducive to edge resource deployment. In light of these observations, this paper proposes a specialized lightweight enhancement network for night-time scenarios in intelligent driving, named VELIE. This network demonstrates robust generalization capabilities across multiple synthetic and real datasets, while also offering better interpretability, enhanced global semantic information augmentation, and faster inference times.

2. Literature Review

2.1. Traditional Methods

The endeavor to enhance images captured in low-light conditions, known as Low-Light Image Enhancement (LLIE), is a topic that has garnered considerable attention within the realm of digital image processing. Initial approaches to this challenge predominantly hinged on principles derived from signal processing and optical theories, forming the foundation of early methodologies in this field. Early methods were generally based on signal processing and optical theories. Histogram equalization [11] and its derivatives [12,13,14] improve image contrast and brightness by globally or locally reallocating pixel brightness levels. Methods like gamma correction and log correction are based on intensity transformations, modifying the mathematical representation of each pixel in the original image’s grayscale to achieve the desired illumination output. In addition to computational approaches, another category of LLIE methods is founded on the Retinex theory [15], which hypothesizes that a color image can be decomposed into reflectance and illumination components by simulating the human eye’s imaging principle. In the Retinex-based LLIE approaches, Single-Scale Retinex (SSR) [16] uses a single-scale Gaussian filter to compensate for the illumination map. To address the color distortion or excessive contrast enhancement in SSR, Multi-Scale Retinex (MSR) [17] was proposed, employing multi-scale Gaussian filters and introducing a color restoration factor for more natural color correction. Another traditional LLIE method type is transformation-based methods, such as wavelet transforms, gamma correction, and logarithmic transformations [18,19,20]. The core idea of transformation-based methods is to perform nonlinear transformations on the local or global contrast of an image, thereby achieving signal enhancement in the spatial or frequency domain.
One major advantage of traditional methods is their low computational cost and more deployment-friendly algorithmic environment, which is highly beneficial for on-board hardware deployment. Additionally, due to their foundation in signal processing and optical models, traditional methods possess good interpretability. In the processing of a low-light image by traditional methods, whether in the spatial or frequency domain, the transformation process of each pixel or digital signal can be mathematically modeled. This process is observable and tunable, and since each step is predefined, the sequence of execution and the effect of each operation is very clear. Despite their predictability and interpretability due to the typically non-complex nonlinear operations of traditional methods, these simple transformations also have numerous drawbacks. The degradation of illumination is inherently a nonlinear distortion process, with noise often arising in the reflectance component obtained from the low-light image according to advanced Retinex theory [21]. The core idea of traditional LLIE methods often involves amplifying the sensor’s gain, so the noise present in darkness and the imbalance of colors in the original data are also amplified. The loss of details in dark areas makes it challenging for traditional image processing methods, which lack the ability to learn contextual features, to replenish degraded detail information. Consequently, deep learning networks, with better capabilities for contextual semantic information learning and inference, have been proposed to address the limitations of traditional image processing methods.

2.2. Learning-Based Method

The domain of Machine Learning (ML) and Deep Learning (DL)-based LLIE tasks can be primarily categorized into two types. The first type encompasses learning networks based on traditional enhancement theories. RetinexNet [22] was a pioneering model that combined Retinex theory with deep learning networks, achieving learning-based image decoupling and low-light image enhancement. Subsequently, KinD and KinD++ [21,23] adopted a similar decoupling-before-enhancement approach, designing deep networks that allow manual adjustment of the output illumination level. URetinexNet [24] then established an unfolding Retinex theory to solve the optimization problem in the Retinex theory. Besides the Retinex theory, learning-based networks centered on differentiable histogram equalization [25], and those guided by wavelet transform [26] have also been proposed. Deep enhancement networks that integrate traditional low-light theories maintain interpretability while converting the manual parameter optimization process of illumination and detail enhancement into an automated update procedure. This process essentially moves away from manual trial-and-error and subjectivity, with the system learning the mapping relationships between lighting conditions and mathematical expressions from a vast amount of prior data.
The second type of LLIE method views low-light enhancement as a generative model-like task in deep learning. In these studies, low-light images are approximated as an entirely black mask, prompting the model to understand the underlying probability distribution of the samples (low light) to generate new data (synthetic normal light) that resembles real data (normal light). Gong [27] was the first to introduce the classic generative model, Generative Adversarial Network (GAN), into low-light enhancement, designing the first unsupervised low-light enhancement network. This model demonstrated excellent generalizability in real-world image enhancement [28]. Subsequently, LLFlow, based on a more powerful generative model, the flow model [29], was proposed, achieving high scores in image quality assessment metrics. With the outstanding performance of diffusion models across all generative models, networks based on diffusion generative model baselines for LLIE [30,31] have also been validated for their exceptional performance.

2.3. LLIE in Driving Scenarios

Being different from general LLIE applications, LLIE technology in intelligent driving systems serves as a critical safety feature. Its primary objective is not merely to provide clearer and more precise visual information but to aid the driving perception system in better understanding the surrounding environment. The essential requirements of LLIE tasks in driving scenarios are twofold: real-time computation and the enhancement of critical features. This means that the algorithm must restore features such as road signs, pedestrians, and vehicles in darkness to normal illumination levels in an extremely short time, enabling the autonomous driving system to recognize and react more accurately.
Historically, traditional LLIE methods like histogram equalization and gamma correction [11,18], due to their simple computational process and rapid execution, were suitable for scenarios demanding high real-time performance. However, they have not been widely adopted because they perform poorly in processing most low-light driving conditions, struggling with noise control and detail recovery in extremely dark areas and under multiple light sources. To address this, the authors of [32] proposed a method based on the MSR approach, introducing a more complex computation with Retinex-based techniques for better enhancement effects while controlling the processing time to 0.2 s. Similarly, the authors of [33] introduced a biologically-inspired traditional computation method, utilizing a retinal imaging theory similar to the Retinex theory to achieve more natural night-driving data enhancement, and achieving faster processing speeds by reducing image resolution.
While traditional Retinex-based models can provide better enhanced image quality, they still struggle with noise control and cannot achieve millisecond-level real-time computation. Therefore, Deep Learning (DL)-based LLIE methods are crucial, as DL-LLIE can maintain real-time processing speeds while effectively improving image quality, reducing noise, and restoring details. LE-Net [34] achieved efficient DL-based enhancement in driving scenarios and can be used in extremely dark conditions. AugGAN [35] and IA-GAN [36] have both acquired knowledge of simulated lighting conditions, using GAN models for all-day driving environment data simulation, making their enhancement networks adaptable to multiple detectors. Regarding efficient computation, according to the authors of [37], the current Swin Transformer-based methods can perform calculations with as low as 90k parameters and a computation time of 0.004 s, significantly lower than the processing time of traditional methods, thereby creating a faster response time for real-time hazard avoidance in vehicles.

3. Method

In view of the motivations discussed earlier, we propose a lightweight network, VE-LIE (Vehicle-Based Efficient Low-Light Image Enhancement), which exhibits excellent generalization on driving datasets. For any raw RGB input, according to the Retinex theory, it can be considered as a combination of reflected and ambient illumination. Our network employs a qualitative operation, enhancing the reflectance and illumination components independently in two separate branches, as illustrated in Figure 1. The Mask Decomposition Network (MDN), Reflectance Enhancement Transformer (RET), and Gamma U-Net (G-UNet) exist as key components of our main network and will be explained in detail in the following section. Our VELIE adopts a supervised learning approach in that paired datasets allow the network to effectively model the illumination degradation function. Although this approach slightly increases processing speed, it achieves significantly better results compared to mixed enhancement techniques.
Traditional indoor low-light supervised paired datasets do not address the complex lighting and details present in driving environments. The RGB images captured in intelligent driving often feature complex urban lighting and a variety of targets, which is markedly different from indoor scenes or low-light datasets with no light sources or only a single weak light source. Therefore, we chose CNN [38] and Vision Transformer [39] as different baselines to address the varying complexities of degradation in low and high-frequency information. The degradation on the Illumination Map is linear grayscale attenuation, which can be processed with a smaller parameter CNN. In contrast, degradation on the Reflectance Map includes dark area noise, blurriness, and unclear contextual information, necessitating the use of a more detail-oriented and powerful transformer for processing.

3.1. Mask Decomposition Network (MDN)

In the Retinex theory, a raw RGB input can be divided into two same-scale maps: a reflectance map and an illumination map. The reflectance component generally stores structural, texture, and color information, while the illumination component primarily contains pure lighting information and can be considered as a two-channel grayscale illumination map. For a low-light image, the degradation in dark areas is mixed, including low-frequency degradation such as lighting intensity, as well as high-frequency degradations like color distortion, noise, and texture blurring. This presents a challenge: direct enhancement on raw RGB can lead to uneven variable weights in the degradation function at each pixel, making it difficult to balance with a single convolution operation. This often results in local overexposure or under-enhancement, along with the loss of detail.
To circumvent this, some networks opt for a decoupling approach, breaking down the original raw RGB input into two maps for targeted brightness and detail enhancement. We employ a shallow CNN for the decoupling process, as displayed in Figure 1, following an encoder-decoder structure. Specifically, the input image first passes through successive convolutional layers, each with 3 × 3 filters followed by a ReLU [40] activation function. After certain convolutional layers, 2 × 2 max pooling is performed to reduce the size of the feature maps, thus constructing the down-sampling path of the encoder. In the decoder, the network gradually restores the size of the feature maps through convolutional layers and ReLU activations, combined with up-sampling operations. During this process, feature maps from the encoder are directly copied to the corresponding layers in the decoder to preserve and restore spatial information. The network ultimately outputs two components: the reflectance map and the illumination map. These two components are then independently enhanced by two branches, the Reflectance Enhancement Transformer (RET) and the Gamma U-Net (G-UNet).

3.2. Reflectance Enhancement Transformer (RET)

In the foundational Retinex theory, the reflectance map is assumed to be invariant under different lighting conditions, as it reflects only the intrinsic properties of objects (such as texture and color information) and is not altered by lighting. However, as highlighted in KinD [21], noise often occurs in darker regions, and thus the reflectance map derived from low-light dark areas tends to be noisier. Consequently, the information in the reflectance map actually varies under different lighting conditions, making this nonlinear degradation more challenging to model than the degradation caused by lighting itself. As mentioned earlier, the environmental complexity in driving datasets is significant. Vehicles need to perceive not just other vehicles but also pedestrians, cityscapes, roads, road signs, traffic lights, etc. The restoration of varied texture and color information of these complex elements is crucial, as downstream tasks depend on this low-frequency information for semantic scene understanding. Therefore, we employ a transformer-based deep learning network architecture to enhance the lost details and color distortion in the reflectance map of low-light images, leveraging its superior global perception capabilities. Unlike CNNs, which focus on local perception, transformers can capture the entire image’s global dependencies through their self-attention mechanism. This means that transformers can consider the interconnectedness of information across various local patches in an image when processing detailed information. This is particularly vital for processing the reflectance map in low-light driving conditions, where the loss of information due to insufficient lighting necessitates broader contextual support to restore details and ensure the harmony of structure, texture, and color of the same object. Additionally, since the transformer structure is not limited by a local receptive field, it can more effectively model long-distance dependencies within the image.
However, the enhanced performance of transformers comes at the cost of high computational demands. Considering the need for real-time inference in driving scenarios, we have chosen the Swin Transformer [41] structure as our backbone. The Swin Transformer introduces a shifting window mechanism for computing self-attention on the foundational transformer architecture. This mechanism allows the model to perform self-attention calculations within local windows while connecting the entire image through the shifting of these windows. Traditional transformers, in contrast, compute self-attention considering the whole image or sequence, which can lead to excessive computational loads. In addition to its lower computational resource requirements, the shifting window mechanism in the Swin Transformer also offers greater flexibility in processing different image areas, facilitating smooth enhancement in specific regions (such as shadows or highlights), thus preventing artifacts or overexposure.
In our Reflectance Enhancement Transformer (RET) branch, which is depicted in Figure 2, we employed a Swin Transformer structure based on SwinIR [42] as the baseline. This structure primarily consists of four Swin Transformer Components (STCs), each comprising four Swin Transformer Layers (STLs). We judiciously reduced the number of STCs and STLs from the original SwinIR baseline, as our experiments indicated that excessive modules could lead to overfitting. Each residual STL is mainly composed of two parts: a Multi-head Self-Attention (MSA) module and a Multi-Layer Perceptron (MLP). The MSA module specifically employs a shifting window strategy, the so-called Shifted Window MSA, which facilitates the shifting of window positions between consecutive STCs. All the modules mentioned above are displayed in abbreviated form in Figure 2, which also clearly depicts the connections between them. This strategy aids in establishing connections between windows, thereby integrating global information. In the original SwinIR baseline, both MSA and MLP are preceded by a Layer Normalization (LayerNorm) operation. However, LayerNorm segments the computation of the self-attention mechanism into pixel-level operations, reducing the network’s inference speed. Therefore, in our study, Batch Normalization (BatchNorm) is used to replace LayerNorm [43]. In the inference phase, BatchNorm merges with the convolution operation, avoiding additional computational time, which is crucial for real-time enhancement in driving edge cases. The outputs of each MSA and MLP module are added back to their inputs through residual connections to prevent the vanishing gradient problem in deep network training.
After enhancement through RET, to avoid biased errors in the color enhancement of the reflectance due to a lack of constraints during training, we introduced a custom-designed color loss, named the Color Histogram Matching Loss ( L hist   ). This loss aims to ensure that the enhanced image’s color distribution closely resembles that of the original image. The loss function achieves this by comparing the color histograms of the two images across each color channel. Since we chose the RGB color space to represent the reflectance map, it is necessary to consider the red (R), green (G), and blue (B) channels separately.
Firstly, for each color channel, the color histogram of the image needs to be calculated before and after enhancement. Assuming our image is ( H × W ) pixel size, with color values for each pixel ranging from 0 to 255, for each color channel, we divide the color values into (B) buckets, each representing a certain range of color values. The color histogram H i s t can be calculated in the following way:
H i s t c ( i ) = 1 H × W x = 1 H y = 1 W 1 { b i n ( I c ( x , y ) ) = i }
where c represents the color channel ( R , G , B ), I c ( x , y ) is the color value of the channel in the image at coordinates ( x , y ) , b i n ( I c ( x , y ) ) is the bucket number to which the color value belongs, and 1 is the indicator function. When its internal condition is true, the value is 1, otherwise, it is 0. H i s t c ( i ) is the normalized histogram value of the i th bucket on channel c .
Next, we use the Kullback–Leibler (KL) [44] divergence to measure the difference between these two histograms. The KL divergence is a measurement method used to measure the difference between two probability distributions, which directly optimizes the difference in color distribution to help the enhanced image visually maintain color consistency with the original image. For each color channel, the KL divergence is defined as follows:
D K L   Hist   enhanced   c   Hist   original   c = i = 1 B   Hist   enhanced   c i log Hist enhanced   c i   Hist   original   c i
Here,   Hist   enhanced   c and   Hist   original   c are the color histograms of the enhanced and original images on channel c , respectively. Finally, the color histogram matching loss is the sum of the KL divergence on all color channels:
L hist   = c R , G , B D K L   Hist   enhanced   c   Hist   original   c

3.3. Gamma U-Net (G-UNet)

When the image exposure is insufficient or excessive, the gamma transform can adjust the contrast of the image [19]. Through different approaches, the γ value is used to enhance low or high grayscale parts. Through this nonlinear transformation, the gamma transform can expand the low or high grayscale range of the image, thereby correcting darker or brighter images. The gamma transform follows the following formula for operation:
I out   = I in   γ
In practical applications, the gamma transformation is applied independently to each pixel value in an image. Unlike histogram equalization or linear transformations, it operates on a pixel-wise scale. Such pixel-wise processing grants the algorithm enhanced flexibility in augmenting various image contents. It can adapt to the specific needs of different regions within an image, effectively adjusting both dark and bright areas.
In operation, selecting different gamma γ values allows for the enhancement or reduction in the contrast at each pixel. For instance, a gamma value less than 1 (γ < 1) is typically used to brighten low-light images, whereas a gamma value larger than 1 (γ > 1) is utilized to dim overly bright or overexposed images.
Traditionally, the appropriate gamma parameter is manually adjusted to balance the enhancement between noise and illumination. However, this process is unreliable, as the enhancement effect is dependent on the operator’s intuitive assessment of image quality. Since there are no standard enhancement results for low-light enhancement, image quality metrics like the Peak Signal-to-Noise Ratio (PSNR), Structural Similarity Index Measure (SSIM) [45], and Natural Image Quality Evaluator (NIQE) [46] are often used to evaluate the effectiveness of enhancement. The manual tuning of gamma parameters, relying solely on subjective user studies, makes achieving optimal metric scores and visual quality uncontrolled. Moreover, in the context of autonomous and assisted driving, real-world lighting conditions are variable and complex. Using fixed parameters for enhancement cannot avoid issues like overexposure, insufficient enhancement, or localized distortion.
Therefore, in our model, we propose a learning-based method to automatically regularize the gamma parameter on the illumination map. In this automatic regularization process, the gamma parameter is adaptively obtained through propagation, loss computation, and parameter update algorithms. The rationale behind using a learning-based method for parameter update lies not just in automation and efficiency but also in the superiority of representation learning in simulating complex lighting environments. Convolutional layers learn high-level feature representations through stacked nonlinear transformations, an automatic feature extraction mechanism that allows the model to capture more complex and abstract data patterns.
In our approach, a supervised learning model based on loss minimization is ideally suited for updating the Gamma parameter. Beyond learning basic lighting and core semantic features of driving scenes, we aim for the network to learn underlying distributions and patterns, achieved by maximizing posterior probability in a Bayesian framework and minimizing the empirical risk in a frequentist framework. Thus, selecting an appropriate network architecture and embedding the Gamma learning layer at the right position is critical. Considering that gamma parameter adjustment involves nonlinear mapping, the chosen network architecture should be capable of handling high-dimensional and nonlinear feature spaces. It should also capture both first and second-order statistical features of driving illumination, understanding and integrating higher-order statistical features.
The specific implementation location of the gamma transformation also warrants consideration. Rather than being just a preprocessing step in a single block, gamma transformation can be integrated as a trainable layer within certain stages of a deep learning network. This approach aims to enhance the network’s adaptability and performance, especially in response to varying lighting conditions for each data point in the dataset, allowing the network to autonomously learn and optimize gamma values during training. In our model, the gamma transformation layer is embedded within both the encoder and decoder. This layer operates based on a parameterized function, with the gamma value optimized as a trainable parameter. Specifically, the layer operates according to Formula (5):
O ( i , j ) = I ( i , j ) γ
where O ( i , j ) is the brightness value of the output image at position ( i , j ) , I ( i , j ) is the brightness value of the input image at the same position, and γ is the parameter of the Gamma transform, which controls the degree of nonlinearity in brightness adjustment.
Given that gamma transformation operates on a pixel-wise level, selecting a learning network baseline with a matching receptive field is crucial. In traditional gamma transformation, pixel-level enhancement essentially only considers pixel-level fine-tuning without accounting for the understanding of contextual information between pixels. Therefore, in considering a learning-based approach for gamma transformation, it is possible not only to utilize ground truth comparisons to fit precise gamma parameters but also to control the coordination of each pixel based on input information within a small neighborhood around it. Consequently, U-Net [47] has been chosen as the baseline for integrating the gamma layer, due to its exceptional feature extraction and fine pixel-level processing capabilities. Originally proposed for semantic segmentation tasks, U-Net, with its encoder-decoder structure and skip connections, has also been proven to be an excellent baseline for image enhancement tasks [48,49,50]. Most importantly, U-Net is relatively efficient and compact, with a reasonable number of parameters and inference speed, making it suitable for deployment in edge computing architectures at the driving end.
In the traditional U-Net architecture, which is shown in Figure 3, the decoder consists of several upsampling operations connected by transposed convolutional layers and convolutional layers, aiming to gradually restore image resolution and detail. In our network design, we chose to insert the Gamma Layer (GL) after each upsampling operation. This allows the network to adapt to different lighting conditions earlier when processing higher-resolution feature maps, thus effectively improving global brightness rather than local details. Moreover, we avoid adding gamma transformation modules after every convolution or skip connection to reduce the complexity and training difficulty of the network in adjusting the brightness at multiple levels.
Since we set the gamma parameter as a learnable variable during network training, this learning process is determined through backpropagation and gradient descent. Initially, the gamma parameter γ is initialized to a preset value of 1, as 1 in a gamma transformation indicates that the parameter will not alter the image’s brightness. Then, during training, the input image undergoes forward propagation through the network, and as the image passes through the gamma transformation layer, each pixel value is adjusted according to Formula (5). The network’s adjusted output is then compared with the desired output, i.e., the ground truth, to calculate the loss function. Apart from the original U-Net loss function L 1   [51], we designed a gamma transformation-based loss function L Gamma   to control the disparity between the network output and real labels, also regulating the range of gamma values through a regularization term.
In the loss function tailored for G-UNet, a weighted squared difference and gamma value range constraint scheme for optimization is adopted, making it more precise and effective. Firstly, we set the weight proportional to the degree of deviation of the gamma value from the target value using the following formula. This way, the greater the deviation of the gamma value from the target, the larger the penalty. The optimization formula can be expressed as follows:
Optimized   Weighted   Squared   Deviation = λ i = 1 n γ i μ γ i μ 2
To ensure that the gamma value is within a more reasonable and flexible range, a smooth penalty function is introduced, such as using a logistic function instead of a hard maximum function. This can reduce mutations and provide smoother gradients. The formula is represented as follows:
I m p r o v e d   R a n g e   C o n s t r a i n t = κ i = 1 n 1 1 + e a γ i γ m a x + 1 1 + e a γ i γ m i n
The basic loss function, the optimized weighted squared difference, and the improved gamma range constraints are combined to form the total loss function. The formula can be expressed as follows:
L Gamma   = λ i = 1 n γ i μ γ i μ 2 + κ i = 1 n 1 1 + e a γ i γ m a x + 1 1 + e a γ i γ m i n
In Formulas (6)–(8), λ and κ are weight coefficients to balance the loss from different parts; γ i is the gamma value of the i t h pixel; μ is the target gamma value; γ m a x and γ m i n are the maximum and minimum permissible ranges for the gamma value, respectively; and a is a parameter to adjust the slope of the logistic function, controlling the strictness of the range constraint. This loss function design aims to achieve more detailed and efficient optimization in the image enhancement process by applying weighted penalties to gamma value deviations and flexible range constraints on the gamma values. After calculating the L Gamma   loss function, the result of the loss is used for the backpropagation process. In this process, the contribution or gradient of each parameter (including gamma parameter γ) to the loss is calculated using a chain rule to update the gamma parameters. We use the gradient descent variant Adam optimizer to update the gamma parameters. In this update step, the loss function will gradually be reduced; that is, adjusting the gamma value gradually makes the enhanced low-light result close to the ground truth. Throughout the training phase, the network will repeat the aforementioned processes of forward propagation, loss calculation, backpropagation, and parameter updates until the network fits the stopping condition. The optimized gamma parameters will be saved and used in the testing phase.

3.4. Feature Fusion Output Layer

After an input image has been enhanced through the RET and G-UNet branches, it is essential to recombine these two separate components into the final RGB output image. This recombination process is conducted in accordance with the principles of the Retinex theory, where an RGB image is perceived as the point-wise product of the reflectance map and the illumination map. Consequently, the enhanced reflectance and illumination maps can be merged using the same multiplication operation.

4. Experiment

4.1. Dataset

Similar to most LLIE tasks, we have selected the LOL dataset [22] as a part of the training dataset. The LOL dataset consists of 485 paired training images and 15 test images with the ground truth, collected under stepwise illumination conditions in indoor scenarios. Not only does the LOL dataset simulate varying degrees of low-light scenarios but the chosen scenes also contain rich color and structural information, making it suitable as a5high-quality training set. However, the dataset collection scenes are almost exclusively indoors, lacking the complexity of lighting in driving scenarios and missing semantic information on core driving-related targets. Solely training with the LOL dataset possibly diminishes the effectiveness of the network in restoring unique features in driving scene images. In order to generalize the network’s inference in driving scenarios, we have collected our VELIE dataset to enrich the training and testing scenes.
The VELIE dataset is divided into two parts: paired (P-VELIE) and unpaired (UP-VELIE). In the P-VELIE portion, we adopted the method from RENOIR [52], using a fixed Nikon Z8 camera (Nikon Corporation, Tokyo, Japan) to capture the same scene with different apertures, designating the normal lighting images as the ground truth and the others as low-light scenarios. This fixed shooting mode allowed us to collect 500 pairs of driving scene low-light datasets, which were split into 485 training and 15 testing pairs. These pairs were oriented toward driving scenarios and aligned pixel-by-pixel into supervised pairs using the three-step method from RENOIR. In the UP-VELIE part, we used an SG2-AR0231C-0202-GMSL vehicle-mounted RGB camera for real-world road scene capture in Shanghai and Suzhou, China, and London and Liverpool, United Kingdom. The dataset ultimately comprised 1000 frames, from which we selected 200 high-quality images as the UP-VELIE test set.
Thus, in the training phase of our network, we first used the LOL dataset’s 485 training images to compare baseline performance on general LLIE tasks, followed by mixed training with the 485 P-VELIE training images and the LOL training set to validate the different baselines’ generalizability to driving scenes. In the testing phase, we initially employed the LOL dataset’s 15 test images to assess the training effectiveness using only the LOL dataset, with the results presented in Figure 4 and Table 1. The mixed metric results of the 15 LOL test images and 15 P-VELIE dataset images are also shown in Table 1. Since each test image in the LOL and P-VELIE datasets has the ground truth, we employed common reference-based metrics like PSNR and SSIM to evaluate the quality of the enhanced images. To verify our network’s effectiveness in more realistic scenarios, we conducted tests on the UP-VELIE. Moreover, to demonstrate the excellent generalizability of the VELIE enhancement network, we also conducted tests on the real-world dataset LIME [53] and the driving-based dataset DarkZurich [54]. Visual results on the above three test sets are displayed in Figure 5. We calculated the average No-Reference Image Quality Evaluation (NIQE) scores and average inference time for a single image on a random selection of 100 Images from both the DarkZurich, UP-VELIE, and LIME datasets, with the results presented in Table 2. All of the learning-based baselines we compared were ensured to have the same dataset configuration as ours, allowing for a comparative analysis of the superiority among different enhancement method baselines.

4.2. Implementation

In our evaluative study, we strictly followed the training configurations as described in the source documents of the methodologies we assessed, ensuring the integrity of our experiments by performing both the training and testing phases within the identical GPU framework. All deep learning (DL)-based approaches underwent training and testing solely on GPU platforms. The deployment of conventional computational methods for low-light enhancement was also executed on the same GPU, with all performance metrics being analyzed via GPU processing. Regarding the training duration, our model underwent an 11 h training period on a 4090 GPU using PyTorch,(PyTorch v2.1.2) with detailed training specifications tailored for this GPU environment: the input images were resized to a resolution of 600 × 400 for the training process, with the total number of epochs set at 200 and an initial learning rate of 1 × 10−4. Additionally, the batch size was determined to be 16. For the Reflectance Enhancement Transformer (RET) branch specifically, we utilized a patch size of 32 × 32.
Highlighting the economic considerations and deployment complexities of edge computing, it is important to note that theoretically, all DL-based checkpoint inferences could be made compatible with CPU usage. The application of CPU-based computations for traditional methods is also viable. Nonetheless, the transition to CPU processing could lead to a substantial increase in computation times, making the use of onboard GPUs a preferable choice under current technological and practical considerations. Looking forward, investigating the integration of VELIE within automotive edge computing frameworks and the execution of empirical nocturnal driving tests emerges as a critical avenue for future exploration, aiming to substantiate the utility and efficacy of the proposed enhancement framework in real-world scenarios.

4.3. Enhancement Results

The distinctions among the performance metrics of the various illumination enhancement techniques highlighted in the supplied Figure 4 and Figure 5 and Table 1 and Table 2 are pronounced. Our selection of baselines for comparative analysis is predicated on a multifaceted approach to the issue of low-light enhancement, encapsulating traditional methodologies CLAHE [14] and MSRCR [17], DL-based techniques RetinexNet [22], KinD++ [23], and URetinexNet [24] grounded in the Retinex theory, the prior information-based method PairLIE [55], and generative models EnGAN [28] and LLFlow [29], all of which represent the pinnacle of the state-of-the-art in their respective domains. The intent is to assess their efficacy across distinct datasets.
CLAHE, as an archetypical conventional image enhancement strategy, endeavors to ameliorate visual acuity by modulating local contrast within the image. However, its performance, as quantified by the PSNR and SSIM metrics on the LOL dataset, yields values of 16.21 and 0.57, respectively, with a marginal uptick observed on the P-VELIE dataset to 17.76 and 0.63. This increment betrays the inherent limitations of CLAHE in preserving the structural integrity of images, particularly within scenarios characterized by extensive dynamic ranges. Visual inspection of Figure 4a corroborates that, while contrast enhancement is achieved, CLAHE is predisposed to generating visually discordant effects in regions suffering from uneven illumination or overexposure, thereby compromising detail fidelity.
In a similar vein, MSRCR, another traditionalist technique with a focus on color restitution and contrast augmentation, demonstrates suboptimal PSNR and SSIM values across both datasets, with its deficiencies becoming conspicuously egregious on the LOL dataset. The resultant imagery post-MSRCR enhancement, as depicted in Figure 4b, shows a lackluster improvement in luminosity and a marked degradation in color veracity.
Conversely, deep learning methodologies like RetinexNet and KinD++—both derivatives of the Retinex theory—demonstrate a superior command in image quality reconstruction, as evidenced by their PSNR and SSIM indices. Notably, KinD++ attains a PSNR of 21.30 on the LOL dataset, underscoring its robustness in image quality enhancement. Yet, juxtaposed with VELIE, both Retinex-based methods exhibit discernible deficits, not only in terms of visual aesthetics as illustrated in Figure 4c,d but also in terms of quantitative benchmarks. This discrepancy is attributable to the methods’ inability to restore verisimilitude in detailing and coloration under extreme lighting conditions.
Generative models such as EnGAN and LLFlow, employing the generative adversarial network framework, aspire to synthesize images of superior quality through the extrapolation of extensive data corpuses. Their performance, denoted by the PSNR and SSIM metrics, is commendable on both test sets, with LLFlow’s SSIM reaching a notable 0.91 on the LOL dataset, indicating their proficiency in structural information preservation. Nonetheless, the visual samples in Figure 4e,f delineate that despite commendable detail and texture recovery in certain regions, these models grapple with maintaining color accuracy, lighting uniformity, and the overall naturalness of the imagery, which could potentially undermine tasks that hinge on precise color discrimination, such as the identification of traffic lights and road markers.
In summary, while the aforementioned methods each exhibit unique strengths, they collectively falter when paralleled against the VELIE approach, both in quantitative image quality assessments and qualitative visual outcomes. VELIE not only registers superior PSNR and SSIM metrics on both test sets but also visually outperforms the competing methods in luminance, contrast, saturation, and detail preservation, as palpably demonstrated in Figure 4g. This is particularly salient in the context of processing images beset with complex lighting and expansive dynamic ranges, a feat made possible by the tailored Color Histogram Matching Loss ( L hist   ) and the adaptive Gamma Layer (GL) illumination adjustment mechanism intrinsic to VELIE.
The incorporation of the inference time in the analysis, as elucidated by Table 2 and derived from real-world dataset evaluations, introduces a pivotal dimension to the discourse, particularly within the ambit of autonomous driving, where it directly correlates with the alacrity and real-time decision-making process of the system. In the exigent milieu of real-time driving scenarios, the imperative for image enhancement algorithms transcends mere excellence in visual quality; it mandates the expeditious processing of incoming image data to uphold safety and efficiency paradigms.
For conventional image enhancement techniques such as CLAHE and MSRCR, their performance deficits, already evident in experimental datasets, become conspicuously pronounced in real-world datasets. The tabulated data reveal that, while CLAHE boasts a relatively brief inference time of 0.47 s, surpassing the majority of deep learning counterparts, its performance, as gauged by the NIQE, falls short of the benchmark established by VELIE.
The augmentation in visual quality shown in Figure 5 offered by Retinex theory-based deep learning methods and generative models is tempered by the escalation of inference times, thereby positing a significant constraint. For instance, the inference durations for RetinexNet and KinD++ are clocked at 0.84 s and 0.41 s, respectively. Generative models like EnGAN and LLFlow, despite their commendable visual quality outcomes, are encumbered by LLFlow’s protracted inference time of 2.29 s—a duration untenable within the purview of autonomous driving applications.
In contradistinction, VELIE not only garners the lowest NIQE scores on real-world datasets, indicative of an image quality that most closely approximates the natural state, but it also distinguishes itself with an inference time of a mere 0.19 s, substantially outstripping all alternative methodologies. This facet is quintessential for autonomous driving, ensuring that the system can process images swiftly and accurately even under low illumination conditions, thereby enhancing the comprehension of and response to the driving environment. Such a symbiosis of speed and quality underscores the pivotal suitability of VELIE for driving scenarios, presenting a solution that safeguards security without compromising on response velocity.

4.4. Impact on High-Level Perception

To investigate the practical effectiveness of enhancement results in high-dimensional perception tasks in driving scenarios, we utilized an object detection network to explore the impact of results generated by different enhancement models on detection performance. In our original UP-VELIE dataset, a total of 2556 vehicle appearances were recorded across 1000 frames. We computed the accuracy of the same detection network on various enhancement outcomes, as illustrated in Table 3. Additionally, we calculated the average processing time, which encompasses both the time for image enhancement and detection inference.
Table 3 indicates that high-quality enhancement results indeed provide clearer visual information, enabling the object detection algorithm to more accurately identify vehicles on the road. Images processed with the VELIE method, due to their superior restoration of image details and colors in real datasets, minimizing noise and distortion, offer a better data foundation for the object detection algorithm. Conversely, lower-quality enhancement results, such as those from MSRCR and RetinexNet, may degrade the performance of the detection system. As shown in Figure 5, in extremely dark conditions, most models introduce additional artifacts. Such images interfere with the object detection algorithm, leading to erroneous target recognition—mistaking noise for targets or failing to identify real targets due to a lack of detail. In high-speed driving scenarios, even small errors can have serious consequences.
Furthermore, inference time is a critical factor. Rapid image processing capabilities ensure the real-time updating of visual information in dynamic environments, providing the latest data for the object detection algorithm. If the enhancement algorithm is slow, it may cause the detection system to react based on outdated information, which is unacceptable in rapidly changing driving environments.
By comparing different image enhancement methods in dark environments on vehicle target detection, the superior performance of the VELIE method is evident. In terms of detection rate, VELIE leads with a score of 94.3%, outperforming the second-ranked EnGAN by 2.2%. In processing time, VELIE is also impressive, requiring only 0.31 s for a single image to be processed, approximately 0.05 s faster than the second-place MSRCR, with 0.46 s. This significant time advantage is particularly important in applications such as autonomous driving systems, where high real-time demands are crucial. VELIE’s ability to ensure extremely high detection rates while significantly reducing processing time demonstrates its leading position in the field of nighttime driving target detection. This balance of significantly enhancing target detection accuracy while meeting real-time processing needs represents a significant advancement for the practical deployment of time-sensitive applications such as automated driving assistance systems.

5. Conclusions and Future Work

In this manuscript, we delve into a comprehensive analysis of RGB camera sensors, which are prevalently employed in intelligent vehicle perception systems. Our research culminates in the development of a novel deep learning-based enhancement framework, designated as the Vehicle-based Efficient Low-light Image Enhancement (VELIE). This framework integrates the Swin Vision Transformer with a gamma transformation augmented U-Net architecture, delivering unparalleled performance across a spectrum of driving datasets and markedly bolstering high-dimensional environmental perception under low-light conditions. VELIE distinguishes itself by achieving exemplary scores in PSNR, SSIM, and NIQE across diverse test sets. Visually, it supersedes competing methodologies in aspects such as luminance, contrast, saturation, and detail retention. A salient feature of VELIE is its rapid inference time, clocking at merely 0.19 s, which significantly outpaces other existing approaches. This rapid processing capability is pivotal in autonomous driving applications, ensuring efficient and precise image analysis even in scenarios of diminished illumination, thereby augmenting comprehension and interaction with the driving environment. The synergistic integration of expeditious processing and high-quality output underscores the indispensable suitability of VELIE in automotive contexts, offering a solution that concurrently prioritizes safety and response agility. The potential amalgamation of VELIE with an array of advanced high-dimensional driving perception tasks—encompassing semantic segmentation, depth estimation, 3D reconstruction, and object tracking—presents a fertile ground for further exploration and development in this domain.
Nonetheless, VELIE currently does not achieve inference times in the millisecond range. This limitation is attributed to the intricate computational demands of the transformer architecture and the presence of multiple sub-networks within the VELIE framework. Future endeavors should pivot toward exploring methodologies to expedite the inference process. Moreover, VELIE neither accounts for factors such as the image acquisition timing and the blurring effects caused by different weather conditions and vehicle speeds nor delves into deployment experiments across various industrial and driving cameras. These aspects represent critical variables that can significantly influence the performance and applicability of our framework. Future research should aim to address these gaps by conducting extensive testing under varied environmental conditions and speeds as well as by exploring the framework’s adaptability and efficiency across different camera systems with different shutter types. Such investigations will not only enhance the robustness of VELIE but also extend its applicability, ensuring its effectiveness in a broader range of real-world driving scenarios.

Author Contributions

Conceptualization, L.Y.; Methodology, L.Y., D.W., D.Y. and Z.M.; Investigation, D.Y.; Resources, D.Y. and Z.M.; Writing—original draft, L.Y., D.W. and D.Y.; Writing—review and editing, Z.M. and Q.Z.; Supervision, Q.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

This study did not raise any ethical issues.

Informed Consent Statement

Not applicable.

Data Availability Statement

The new dataset used in this study, given its potential value in commercial applications, will not be made open-source at this time. We fully acknowledge the importance of open-source datasets for scientific research; however, considering business confidentiality and future commercial uses, we must retain exclusive rights to these data. We hope our readers can understand this decision and look forward to sharing our findings with the broader research community when appropriate.

Acknowledgments

We would like to extend our sincere gratitude to the reviewers and editors who will be evaluating our manuscript. Your insights and feedback are invaluable to the refinement of our research and the advancement of knowledge in our field. We deeply appreciate the time and effort you will dedicate to reviewing our work, and we look forward to your constructive comments and suggestions. Thank you for your contribution to maintaining the high standards of academic rigor and integrity.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Gruyer, D.; Magnier, V.; Hamdi, K.; Claussmann, L.; Orfila, O.; Rakotonirainy, A. Perception, information processing and modeling: Critical stages for autonomous driving applications. Annu. Rev. Control 2017, 44, 323–341. [Google Scholar] [CrossRef]
  2. Wang, L.; Zhang, X.; Li, J.; Xv, B.; Fu, R.; Chen, H.; Yang, L.; Jin, D.; Zhao, L. Multi-Modal and Multi-Scale Fusion 3D Object Detection of 4D Radar and LiDAR for Autonomous Driving. IEEE Trans. Veh. Technol. 2022, 72, 5628–5641. [Google Scholar] [CrossRef]
  3. Fang, W.; Zhang, G.; Zheng, Y.; Chen, Y. Multi-Task Learning for UAV Aerial Object Detection in Foggy Weather Condition. Remote Sens. 2023, 15, 4617. [Google Scholar] [CrossRef]
  4. Wood, J.M.; Isoardi, G.; Black, A.; Cowling, I. Night-time driving visibility associated with LED streetlight dimming. Accid. Anal. Prev. 2018, 121, 295–300. [Google Scholar] [CrossRef] [PubMed]
  5. Rashed, H.; Ramzy, M.; Vaquero, V.; El Sallab, A.; Sistu, G.; Yogamani, S. FuseMODNet: Real-Time Camera and LiDAR Based Moving Object Detection for Robust Low-Light Autonomous Driving. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Republic of Korea, 27–28 October 2019; pp. 2393–2402. [Google Scholar]
  6. Paek, D.; Kong, S.; Wijaya, K.T. K-Radar: 4D Radar Object Detection for Autonomous Driving in Various Weather Conditions. Adv. Neural Inf. Process. Syst. 2022, 35, 3819–3829. [Google Scholar] [CrossRef]
  7. Zhang, S.; Zhang, Y.; Jiang, Z.; Zou, D.; Ren, J.; Zhou, B. Learning to see in the dark with events. In Proceedings of the 16th European Conference on Computer Vision (ECCV 2020), Glasgow, UK, 23–28 August 2020; Springer International Publishing: Berlin/Heidelberg, Germany, 2020. Part XVIII 16. [Google Scholar] [CrossRef]
  8. Altay, F.; Velipasalar, S. The Use of Thermal Cameras for Pedestrian Detection. IEEE Sens. J. 2022, 22, 11489–11498. [Google Scholar] [CrossRef]
  9. Faramarzpour, N.; Deen, M.J.; Shirani, S.; Fang, Q.; Liu, L.W.C.; Campos, F.d.S.; Swart, J.W. CMOS-Based Active Pixel for Low-Light-Level Detection: Analysis and Measurements. IEEE Trans. Electron Devices 2007, 54, 3229–3237. [Google Scholar] [CrossRef]
  10. Yuantao, Z.; Mengyang, C.; Dexin, S.; Yinnian, L. TDI Technology Based on Global Shutter sCMOS Image Sensor for Low-Light-Level Imaging. Acta Opt. Sin. 2018, 38, 0911001. [Google Scholar] [CrossRef]
  11. Pizer, S.M.; Amburn, E.P.; Austin, J.D.; Cromartie, R.; Geselowitz, A.; Greer, T.; ter Haar Romeny, B.; Zimmerman, J.B.; Zuiderveld, K. Adaptive histogram equalization and its variations. Comput. Vis. Graph. Image Process. 1987, 39, 355–368. [Google Scholar] [CrossRef]
  12. Kim, Y.-T. Contrast enhancement using brightness preserving bi-histogram equalization. IEEE Trans. Consum. Electron. 1997, 43, 1–8. [Google Scholar] [CrossRef]
  13. Stark, J. Adaptive image contrast enhancement using generalizations of histogram equalization. IEEE Trans. Image Process. 2000, 9, 889–896. [Google Scholar] [CrossRef]
  14. Abdullah-Al-Wadud, M.; Kabir; Dewan, M.A.A.; Chae, O. A Dynamic Histogram Equalization for Image Contrast Enhancement. IEEE Trans. Consum. Electron. 2007, 53, 593–600. [Google Scholar] [CrossRef]
  15. Land, E.H.; McCann, J.J. Lightness and Retinex Theory. J. Opt. Soc. Am. 1971, 61, 1–11. [Google Scholar] [CrossRef]
  16. Hines, G.; Rahman, Z.U.; Jobson, D.; Woodell, G. Single-Scale Retinex Using Digital Signal Processors. In Proceedings of the Global Signal Processing Conference, Bangkok, Thailand, 6–9 December 2005; p. 1324. Available online: https://ntrs.nasa.gov/citations/20050091487 (accessed on 8 February 2024).
  17. Rahman, Z.; Jobson, D.J.; Woodell, G.A. Multi-scale retinex for color image enhancement. In Proceedings of the 3rd IEEE International Conference on Image Processing, Lausanne, Switzerland, 19 September 1996; Volume 3, pp. 1003–1006. [Google Scholar]
  18. Antonini, M.; Barlaud, M.; Mathieu, P.; Daubechies, I. Image coding using wavelet transform. IEEE Trans. Image Process. 1992, 1, 205–220. [Google Scholar] [CrossRef]
  19. Huang, S.-C.; Cheng, F.-C.; Chiu, Y.-S. Efficient Contrast Enhancement Using Adaptive Gamma Correction With Weighting Distribution. IEEE Trans. Image Process. 2012, 22, 1032–1041. [Google Scholar] [CrossRef] [PubMed]
  20. Jourlin, M.; Pinoli, J. A model for logarithmic image processing. J. Microsc. 1988, 149, 21–35. [Google Scholar] [CrossRef]
  21. Zhang, Y.; Zhang, J.; Guo, X. Kindling the Darkness: A Practical Low-Light Image Enhancer. In Proceedings of the 27th ACM International Conference on Multimedia, Melbourne, VIC, Australia, 15 October 2019; pp. 1632–1640. [Google Scholar] [CrossRef]
  22. Li, P.; Tian, J.; Tang, Y.; Wang, G.; Wu, C. Deep Retinex Network for Single Image Dehazing. Available online: https://daooshee.github.io/BMVC2018website/ (accessed on 8 February 2024).
  23. Zhang, Y.; Guo, X.; Ma, J.; Liu, W.; Zhang, J. Beyond Brightening Low-light Images. Int. J. Comput. Vis. 2021, 129, 1013–1037. [Google Scholar] [CrossRef]
  24. Wu, W.; Weng, J.; Zhang, P.; Wang, X.; Yang, W.; Jiang, J. URetinex-Net: Retinex-based Deep Unfolding Network for Low-light Image Enhancement. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 5891–5900. [Google Scholar]
  25. Kim, B.; Lee, S.; Kim, N.; Jang, D.; Kim, D.-S. Learning Color Representations for Low-Light Image Enhancement. In Proceedings of the 2022 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2022; pp. 904–912. [Google Scholar]
  26. Xu, J.; Yuan, M.; Yan, D.-M.; Wu, T. Illumination Guided Attentive Wavelet Network for Low-Light Image Enhancement. IEEE Trans. Multimed. 2022, 25, 6258–6271. [Google Scholar] [CrossRef]
  27. Jiang, Y.; Gong, X.; Liu, D.; Cheng, Y.; Fang, C.; Shen, X.; Yang, J.; Zhou, P.; Wang, Z. EnlightenGAN: Deep Light Enhancement Without Paired Supervision. IEEE Trans. Image Process. 2021, 30, 2340–2349. [Google Scholar] [CrossRef]
  28. Gong, Y.; Liao, P.; Zhang, X.; Zhang, L.; Chen, G.; Zhu, K.; Tan, X.; Lv, Z. Enlighten-GAN for Super Resolution Reconstruction in Mid-Resolution Remote Sensing Images. Remote Sens. 2021, 13, 1104. [Google Scholar] [CrossRef]
  29. Wang, Y.; Wan, R.; Yang, W.; Li, H.; Chau, L.-P.; Kot, A. Low-Light Image Enhancement with Normalizing Flow. Proc. AAAI Conf. Artif. Intell. 2022, 36, 2604–2612. [Google Scholar] [CrossRef]
  30. Yi, X.; Xu, H.; Zhang, H.; Tang, L.; Ma, J. Diff-Retinex: Rethinking Low-light Image Enhancement with A Generative Diffusion Model. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–3 October 2023; pp. 12268–12277. [Google Scholar]
  31. Jiang, H.; Luo, A.; Fan, H.; Han, S.; Liu, S. Low-Light Image Enhancement with Wavelet-Based Diffusion Models. ACM Trans. Graph. 2023, 42, 1–14. [Google Scholar] [CrossRef]
  32. Kuang, H.; Zhang, X.; Li, Y.-J.; Chan, L.L.H.; Yan, H. Nighttime Vehicle Detection Based on Bio-Inspired Image Enhancement and Weighted Score-Level Feature Fusion. IEEE Trans. Intell. Transp. Syst. 2016, 18, 927–936. [Google Scholar] [CrossRef]
  33. Kuang, H.; Chen, L.; Gu, F.; Chen, J.; Chan, L.; Yan, H. Combining Region-of-Interest Extraction and Image Enhancement for Nighttime Vehicle Detection. IEEE Intell. Syst. 2016, 31, 57–65. [Google Scholar] [CrossRef]
  34. Li, G.; Yang, Y.; Qu, X.; Cao, D.; Li, K. A deep learning based image enhancement approach for autonomous driving at night. Knowl.-Based Syst. 2020, 213, 106617. [Google Scholar] [CrossRef]
  35. Lin, C.-T.; Huang, S.-W.; Wu, Y.-Y.; Lai, S.-H. GAN-Based Day-to-Night Image Style Transfer for Nighttime Vehicle Detection. IEEE Trans. Intell. Transp. Syst. 2020, 22, 951–963. [Google Scholar] [CrossRef]
  36. Kandula, P.; Suin, M.; Rajagopalan, A.N. Illumination-Adaptive Unpaired Low-Light Enhancement. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 3726–3736. [Google Scholar] [CrossRef]
  37. Cui, Z.; Li, K.; Gu, L.; Su, S.; Gao, P.; Jiang, Z.; Qiao, Y.; Harada, T. You Only Need 90K Parameters to Adapt Light: A Light Weight Transformer for Image Enhancement and Exposure Correction. arXiv 2022, arXiv:2205.14871, 238. [Google Scholar]
  38. Krichen, M. Convolutional Neural Networks: A Survey. Computers 2023, 12, 151. [Google Scholar] [CrossRef]
  39. Ali, A.M.; Benjdira, B.; Koubaa, A.; El-Shafai, W.; Khan, Z.; Boulila, W. Vision Transformers in Image Restoration: A Survey. Sensors 2023, 23, 2385. [Google Scholar] [CrossRef]
  40. Daubechies, I.; DeVore, R.; Foucart, S.; Hanin, B.; Petrova, G. Nonlinear approximation and (deep) ReLU networks. Constr. Approx. 2022, 55, 127–172. [Google Scholar] [CrossRef]
  41. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar] [CrossRef]
  42. Liang, J.; Cao, J.; Sun, G.; Zhang, K.; Van Gool, L.; Timofte, R. SwinIR: Image Restoration Using Swin Transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 1833–1844. [Google Scholar] [CrossRef]
  43. Ba, J.L.; Kiros, J.R.; Hinton, G.E. Layer Normalization. arXiv 2016. Available online: https://arxiv.org/abs/1607.06450 (accessed on 8 February 2024).
  44. Van Erven, T.; Harremoes, P. Rényi Divergence and Kullback-Leibler Divergence. IEEE Trans. Inf. Theory 2014, 60, 3797–3820. [Google Scholar] [CrossRef]
  45. Horé, A.; Ziou, D. Image Quality Metrics: PSNR vs. SSIM. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 2366–2369. [Google Scholar] [CrossRef]
  46. Mittal, A.; Soundararajan, R.; Bovik, A.C. Making a “Completely Blind” Image Quality Analyzer. IEEE Signal Process. Lett. 2012, 20, 209–212. [Google Scholar] [CrossRef]
  47. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Part III 18. pp. 234–241. [Google Scholar] [CrossRef]
  48. Ullah, F.; Ansari, S.U.; Hanif, M.; Ayari, M.A.; Chowdhury, M.E.H.; Khandakar, A.A.; Khan, M.S. Brain MR Image Enhancement for Tumor Segmentation Using 3D U-Net. Sensors 2021, 21, 7528. [Google Scholar] [CrossRef] [PubMed]
  49. Ai, S.; Kwon, J. Extreme Low-Light Image Enhancement for Surveillance Cameras Using Attention U-Net. Sensors 2020, 20, 495. [Google Scholar] [CrossRef] [PubMed]
  50. Jia, F.; Wong, W.H.; Zeng, T. DDUNet: Dense Dense U-net with Applications in Image Denoising. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 354–364. [Google Scholar] [CrossRef]
  51. Bloomfield, P.; Steiger, W.L. Least Absolute Deviations: Theory, Applications, and Algorithms; Birkhäuser: Boston, MA, USA, 1983; pp. 158–166. [Google Scholar] [CrossRef]
  52. Anaya, J.; Barbu, A. RENOIR—A dataset for real low-light image noise reduction. J. Vis. Commun. Image Represent. 2018, 51, 144–154. [Google Scholar] [CrossRef]
  53. Guo, X.; Li, Y.; Ling, H. LIME: Low-Light Image Enhancement via Illumination Map Estimation. Available online: https://github.com/estija/LIME (accessed on 8 February 2024).
  54. Sakaridis, C.; Dai, D.; Gool, L.V. Guided Curriculum Model Adaptation and Uncertainty-Aware Evaluation for Semantic Nighttime Image Segmentation. Available online: https://www.trace.ethz.ch/publications/2019/GCMA_UIoU/ (accessed on 8 February 2024).
  55. Fu, Z.; Yang, Y.; Tu, X.; Huang, Y.; Ding, X.; Ma, K.K. Learning a Simple Low-Light Image Enhancer From Paired Low-Light Instances. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 22252–22261. [Google Scholar] [CrossRef]
Figure 1. The general pipeline of the proposed Vehicle-based Efficient Low-light Image Enhancement (VELIE) Network.
Figure 1. The general pipeline of the proposed Vehicle-based Efficient Low-light Image Enhancement (VELIE) Network.
Sensors 24 01345 g001
Figure 2. The detailed structure of the Reflectance Enhancement Transformer (RET).
Figure 2. The detailed structure of the Reflectance Enhancement Transformer (RET).
Sensors 24 01345 g002
Figure 3. The detailed structure of Gamma U-Net (G-UNet).
Figure 3. The detailed structure of Gamma U-Net (G-UNet).
Sensors 24 01345 g003
Figure 4. Visualization Comparison of VELIE and different baselines on LOL test set [22] and P-VELIE.
Figure 4. Visualization Comparison of VELIE and different baselines on LOL test set [22] and P-VELIE.
Sensors 24 01345 g004
Figure 5. Visualization Comparison of VELIE and different baselines on LIME [53], DarkZurich [54], and UP-VELIE.
Figure 5. Visualization Comparison of VELIE and different baselines on LIME [53], DarkZurich [54], and UP-VELIE.
Sensors 24 01345 g005aSensors 24 01345 g005b
Table 1. Image quality metric comparison on LOL and P-VELIE from different methods.
Table 1. Image quality metric comparison on LOL and P-VELIE from different methods.
MethodLOL Test SetP-VELIE Test Set
PSNR 1SSIMPSNRSSIM
CLAHE16.210.5717.760.63
MSRCR11.240.359.970.24
RetinexNet16.820.4312.440.39
KinD++21.300.8221.070.79
EnGAN17.480.6519.950.70
LLFlow25.720.9120.340.83
PairLIE19.510.7417.720.69
URetinexNet21.330.8320.190.75
VELIE (Ours)26.070.9123.610.85
1 Higher PSNR and SSIM indicate better image quality.
Table 2. Image quality metric and inference time comparison on unpaired real-world dataset LIME, DarkZurich, and UP-VELIE from different methods.
Table 2. Image quality metric and inference time comparison on unpaired real-world dataset LIME, DarkZurich, and UP-VELIE from different methods.
MethodNIQEInference Time (s)
CLAHE4.310.47
MSRCR8.810.22
RetinexNet7.250.84
KinD++4.360.41
EnGAN5.011.77
LLFlow4.572.29
PairLIE4.990.91
URetinexNet4.331.12
VELIE (Ours)3.970.19
Table 3. Detection rate and processing time comparison on DarkZurich and UP-VELIE from different methods.
Table 3. Detection rate and processing time comparison on DarkZurich and UP-VELIE from different methods.
MethodDetection RateProcessing Time (s)
CLAHE88.70.68
MSRCR73.10.46
RetinexNet77.51.02
KinD++86.70.62
EnGAN92.11.97
LLFlow84.22.52
PairLIE92.51.21
URetinexNet92.41.33
VELIE (Ours)94.30.41
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ye, L.; Wang, D.; Yang, D.; Ma, Z.; Zhang, Q. VELIE: A Vehicle-Based Efficient Low-Light Image Enhancement Method for Intelligent Vehicles. Sensors 2024, 24, 1345. https://doi.org/10.3390/s24041345

AMA Style

Ye L, Wang D, Yang D, Ma Z, Zhang Q. VELIE: A Vehicle-Based Efficient Low-Light Image Enhancement Method for Intelligent Vehicles. Sensors. 2024; 24(4):1345. https://doi.org/10.3390/s24041345

Chicago/Turabian Style

Ye, Linwei, Dong Wang, Dongyi Yang, Zhiyuan Ma, and Quan Zhang. 2024. "VELIE: A Vehicle-Based Efficient Low-Light Image Enhancement Method for Intelligent Vehicles" Sensors 24, no. 4: 1345. https://doi.org/10.3390/s24041345

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop