Next Article in Journal
Obstacle Detection Method Based on RSU and Vehicle Camera Fusion
Previous Article in Journal
Direction-of-Arrival Estimation Method Based on Neural Network with Temporal Structure for Underwater Acoustic Vector Sensor Array
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Combining Low-Light Scene Enhancement for Fast and Accurate Lane Detection

1
School of Science, Beijing University of Civil Engineering and Architecture, Beijing 102616, China
2
School of Geomatics and Urban Spatial Informatics, Beijing University of Civil Engineering and Architecture, Beijing 102616, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(10), 4917; https://doi.org/10.3390/s23104917
Submission received: 21 April 2023 / Revised: 12 May 2023 / Accepted: 18 May 2023 / Published: 19 May 2023
(This article belongs to the Section Vehicular Sensing)

Abstract

:
Lane detection is a crucial task in the field of autonomous driving, as it enables vehicles to safely navigate on the road by interpreting the high-level semantics of traffic signs. Unfortunately, lane detection is a challenging problem due to factors such as low-light conditions, occlusions, and lane line blurring. These factors increase the perplexity and indeterminacy of the lane features, making them hard to distinguish and segment. To tackle these challenges, we propose a method called low-light enhancement fast lane detection (LLFLD) that integrates the automatic low-light scene enhancement network (ALLE) with the lane detection network to improve lane detection performance under low-light conditions. Specifically, we first utilize the ALLE network to enhance the input image’s brightness and contrast while reducing excessive noise and color distortion. Then, we introduce symmetric feature flipping module (SFFM) and channel fusion self-attention mechanism (CFSAT) to the model, which refine the low-level features and utilize more abundant global contextual information, respectively. Moreover, we devise a novel structural loss function that leverages the inherent prior geometric constraints of lanes to optimize the detection results. We evaluate our method on the CULane dataset, a public benchmark for lane detection in various lighting conditions. Our experiments show that our approach surpasses other state of the arts in both daytime and nighttime settings, especially in low-light scenarios.

1. Introduction

Lane detection is a critical and hot task for ensuring safe autonomous driving and advanced driver assistance system (ADAS) [1]. Computer vision technology has facilitated significant advancements in this field, making it a widely researched topic in academia and industry alike. The ability to accurately detect lanes plays a crucial role in guiding vehicles to travel safely on the road.
Various deep learning methods have emerged in lane detection in recent years, enhancing the detection performance and adaptability under diverse scenarios. Nevertheless, the task still faces several challenges that need to be overcome. One of the most pressing issues is the robustness of models to various adverse conditions in real-world scenarios, such as extreme light [2] and weather variations. Furthermore, lane markings are often occluded by other objects, such as cars, which frequently occurs in autonomous driving situations. This challenge requires lane detection models to possess high-level semantic understanding of the scene and be able to distinguish between lane markings and other objects accurately. Although the traditional method is simple and fast, the detection results are often not satisfactory. Addressing these challenges is crucial to enable the reliable deployment of autonomous driving systems in real-world environments. To more efficiently utilize visual information, there are specialized convolution operations that facilitate the aggregation of information from various dimensions in the segment model of SCNN [3]. These operations involve the processing of slice features, which are then added one by one to enable the information aggregation. Despite its effectiveness, this method is relatively slow to use in practice. SAD [4] has emerged as a promising approach to reduce the parameters required for lane detection. Despite its potential, SAD still suffers from suboptimal processing speed, which hinders its practical applicability. UFLD [5] transform this task into a simpler classification problem, which has the advantage of significantly reducing the computational overhead and enabling ultra-fast lane detection. However, the accuracy of UFLD still needs improvement, as it does not perform well in low light environments or for capturing small targets and details.
In this paper, we present a novel approach named low-light enhancement fast lane detection (LLFLD), which leverages the automatic low-light scene enhancement network (ALLE) network for image adaptive enhancement prior to detection. Specifically, we introduce a perceptual brightness threshold and apply enhancement detection to images that fall below this threshold, resulting in improved detection performance in low-light scenarios. To further increase accuracy, we leverage the symmetric distribution of lane markings in the driving perspective and design a symmetric feature flipping module (SFFM) that enhances low-level features for more precise lane localization. In addition to the main branch for detection, we present an auxiliary segmentation module that is only activated during training.
In our auxiliary segmentation module, we incorporate the channel fusion self-attention mechanism (CFSAT) to enhance the acquisition of global context information. This is achieved by establishing connections between the lane markings and the wider global feature map. Moreover, our approach involves the introduction of a novel structural loss function that is based on the geometric shapes of lanes. Through this loss function, we optimize the segmentation and classification performance of our model by leveraging the inherent structure of the lane markings. By employing these strategies, we are able to effectively capture important contextual features while simultaneously mitigating unnecessary redundancy, resulting in a more robust and efficient detection algorithm.
Our approach has been tested extensively using the benchmark CULane dataset [3], and we present comprehensive experimental results along with comparisons to other state-of-the-art methods. Additionally, we conducted an ablation study to analyze the impact of our design choices on the performance of the model. In summary, our contributions are:
  • We propose low-light enhancement fast lane detection (LLFLD), a lane detection system that combines a low-light image enhancement network (ALLE) with a lane detection network. Our approach significantly enhances the performance of the network in low-light environments while maintaining an ultra-fast detection speed.
  • We propose a symmetric feature flipping module (SFFM), which refines the low-level features and gains more precise lane localization.
  • We propose a channel fusion self-attention mechanism (CFSAT) in the auxiliary segmentation module, which captures and utilizes more global context information.
  • We propose a novel structural loss function that leverages the inherent geometric constraints of lanes to optimize the detection results.

2. Related Work

2.1. Traditional Methods

Autonomous driving heavily relies on the accurate detection of lanes, which serves as a fundamental task in achieving self-driving capabilities. This task involves identifying the lanes on the road and providing relevant information such as lane ID, direction, curvature, width, length, and visualization. Over the years, a plethora of computer vision technologies have been developed to achieve robust lane detection without making assumptions about the number of lanes present, involving the location and tracking of lane boundaries in road scenes. Traditional approaches typically utilize hand-crafted features, such as edge detection, the Hough transform, and color filtering [6,7,8]. However, these techniques are known to be susceptible to various environmental factors, including illumination changes, occlusions, and complex road layouts, rendering them less effective in challenging driving scenarios.

2.2. Segmentation Methods

Semantic segmentation represents an active research area in computer vision, with the primary aim of assigning a semantic label to each pixel in an image. Two primary categories of methods can be identified, namely region-based and end-to-end approaches. Region-based methods often leverage region proposal networks (RPNs) [9] or sliding windows to extract regions of interest (ROIs) [10] from the input image. The segmented network is then applied to each ROI to generate precise masks. VPGNet [11] exemplifies a region-based approach, which generates ROIs for lanes and other road markings using RPNs and applies a multi-task network for pixel classification within ROIs. End-to-end methods, on the other hand, directly apply a segmentation network to the entire input image without using any ROIs, which are more computationally efficient. In the context of end-to-end methods, the SCNN [3] approach employs a specialized convolution operation in its segmentation module, enabling the effective aggregation of features from different dimensions. This is achieved by processing sliced features and incrementally adding them together to effectively capture multi-scale contextual information. However, the method can be computationally expensive. To deal with the challenge of real-time applications, researchers have proposed lightweight semantic segmentation methods, such as self-attention distillation (SAD) [4]. To enable effective knowledge transfer between high level semantic information and low-level location feature attentions, SAD utilizes an attention distillation module that treats the former as a teacher and the latter as a student. LaneNet [12] employs an instance segmentation pipeline to handle a varying number of lines; however, the generation of line instances necessitates post-inference clustering.

2.3. Anchor-Based Methods

The anchor-based approach is a widely used technique for lane detection, where a set of predefined anchors is employed to localize the lane markings. Among different variants of this approach, the linear anchor-based method utilizes linear anchors as reference points for accurate lane regression. The pioneering work of Line CNN [13] introduced the use of line anchors in lane detection; later, Feng et al. [14] proposed a Bezier curve-based regression model to address the challenge of a large number of polynomial regression parameters. The Bezier curve model is not only computationally efficient but also highly stable. Lane ATT [15] proposed a novel attention mechanism based on anchor points, which aggregates global information and shows outstanding performance accompanied by high efficiency. Some recent methods employ classification based on row selecting [5,16] as a means of custom down-sampling segmentation in units of one pixel, but still require post-processing. Another noteworthy anchor-based method is UFLD [5], which frames the lane detection problem as a row anchor presetting and row selection task that leverages continuity of lane lines for improving detection speed. It exhibits ultra-fast detection speed, albeit with an accuracy that leaves room for further improvement.

2.4. Low Light Image Enhancement

The field of low-light image enhancement has achieved a relatively high level of development [17,18,19]. Histogram equalization and dehazing-based methods are the two primary techniques employed in traditional image enhancement methods. Histogram equalization is a classical method that enhances images by stretching their histogram distribution to achieve uniform brightness across the entire range. Zhang et al. [20] propose the utilization of various image processing techniques, including histogram equalization, as a data normalization method for machine learning. To evaluate the effectiveness of these techniques, the paper compares their performance to that of z-score normalization on a face-based authentication algorithm using SVM and random forest classifiers. FEBD [21] for multiple histograms modification presents a novel fast expansion-bins-determination method for multiple-histogram-modification-based reversible data hiding (RDH), which is a powerful technique that enables the embedding of data into images while maintaining the original information content. Specifically, the proposed method enhances the RDH approach by introducing an efficient algorithm capable of determining optimal expansion bins quickly and accurately, with only a negligible performance loss compared to the original method. Some approaches adopt the dehazing-based [22], which presents a new single image dehazing algorithm based on the dark channel prior. This method features an improved atmospheric light estimation method and a low-complexity morphological reconstruction technique to generate high-quality dehazed images with significantly reduced computational complexity compared to previous approaches. To solve the scattering of sunlight in the atmosphere, Ref. [23] proposes an image enhancement approach based on inverting low light images and applying image dehazing with an atmospheric light scattering model. Unlike traditional methods that rely on hand-crafted features, deep learning methods leverage the power of neural networks to learn a pixel mapping function to transform low-light images to high-quality outputs. Recently, Ref. [24] raised a new dataset of low-light images and videos, along with an online platform that comprises various popular methods for evaluation and comparison. Differing from the methods which need the reference of Retinex theory and labels [25], Zero-DCE [26] is a no labels approach that considers image brightness enhancement mapping in supervising by several no references luminosity and color loss function. Accordingly, the algorithm exhibits enhanced robustness, a wider range for adjusting the dynamic range of images, and reduced computational costs.

3. Methodology

3.1. Overall Pipeline

LLFLD is a lightweight lane detection model that integrates low light image enhancement techniques. Our approach achieves ultra-low computational costs by transforming the segmentation task into the selection and classification task of predetermined line anchors. An overview of the method is shown in Figure 1. Our framework comprises two parts: a primary classification detection branch and an auxiliary segmentation branch, designed to process RGB images I 3 × H × W captured from a forward-facing vehicle camera. The outputs of the system are selected lane boundary points corresponding to row anchors. To generate these outputs, the input images undergo automatic low-light enhancement before being fed to the classification detection branch, which includes a backbone ResNet stage. The low-level feature maps after convolution still retain strong location information, so prior to entering the backbone ResNet stage, we flip the feature maps symmetrically using the symmetric feature flipping module (SFFM) to enhance the visual information that is often lost due to occlusion and blur, thereby more effectively locating the lanes. Then, we obtain F C l a n e × H s l i c e × W s l i c e by passing the feature maps output by the backbone through a fully connected layer and a reshape layer, where C l a n e , H s l i c e , W s l i c e are, respectively, the number of lanes in each image, and the numbers of row demarcation and column demarcation. Finally, we predict the final output channels using F by transforming the segmentation problem into an anchor-based selection classification problem, which improves the detection efficiency significantly.
The auxiliary segmentation branch performs feature map upsampling from different stages of the Resnet network at varying scales, and concatenates and convolves them, similar to the structure of FPN [27] that fuses high-level features on the top and low-level features on the bottom. We update the feature map after hierarchical fusion with global information through the channel fusion self-attention module (CFSAT) and generate segmentation results to assist parameter learning in the backbone classification branch. It is noteworthy that the auxiliary segmentation branch is exclusively active during training but deactivated during the test and detection phases. In this way, this approach not only facilitates parameter learning in the backbone branch but also preserves fast detection speed.

3.2. Automatic Low-Light Scene Enhancement (ALLE)

In practical applications, low brightness conditions can significantly impede visual perception, particularly in situations such as nighttime driving. The traditional method of improving image brightness often introduces more noise and lacks the necessary dynamic range and robustness. As a result, although it increases the visibility of the target lane lines, it also brings unwanted information that interferes with the detection results. To account for the complex and variable driving conditions, we recognize that not every image requires brightness enhancement. Additionally, applying the same enhancements to both normal and high brightness images can result in overexposure, which is also detrimental to the lane detection results. To overcome these challenges, our objective is to automatically enhance low-light images to achieve superior visual effects, with a specific emphasis on the detection performance of the enhancement model under low-light conditions. In this paper, we propose the automatic low-light scene enhancement (ALLE) module, which employs a Zero-DCE network to estimate the best mapping from low light image to normal exposure image, as depicted in Figure 2. We perform adaptive enhancement according to the brightness of different images, that is, enhance low-brightness images only, and maintain the same normal brightness and high-brightness images, rather than uniform or random preprocessing of all images. Initially, we calculate the root mean square (RMS) of RGB values of the input image I 3 × H × W , which enables us to determine the perceived brightness P b [28] of the image I in the RGB space, as represented by the following equation:
P b = f R , G , B = 0.241 R 2 + 0.691 G 2 + 0.068 B 2
where R , G , and B denote to the RMS-normalized R G B channel values of the input image I . Our study reveals that values of perceived brightness below 60, calculated using Equation (1), noticeably affect human visual perception. On the contrast, for values above 80, the perceived image brightness is considered normal. Thus, we apply deep enhancement processing to those images whose perceived brightness falls below the threshold δ , which has been set to 70.
To accomplish the task of low-light image enhancement, we leverage the Zero-DCE network, which formulates the problem as a non-reference multi-level recursive curve estimation task. The network adopts a lightweight encoder-decoder architecture for feature extraction and image enhancement generation. This approach enables end-to-end enhancement of images under low-light conditions. Furthermore, the perceived brightness of different images varies greatly, which hinders the training of a single network. Minimizing this variation as much as possible facilitates network optimization. Specifically, our objective is to train a mapping function Φ n , capable of transforming low-light images into normally exposed ones, as denoted by the following mathematical expression:
L E n x = L E L E n 1 ; Φ n R , G , B = L E n 1 x + Φ n L E n 1 x 1 L E n 1 x
The equation governing automatic low-light image enhancement entails a parameter map Φ that is trainable and shares the same dimensions as the input image. Here, x refers to the RMS-normalized channel values of the input image, while n denotes the number of iterations that regulate the curvature of the curve. The enhanced version of the final input, denoted as L E n x , is obtained using the previous version L E n 1 x . Hence, the expression for L E 1 x can be expressed as:
L E 1 x = L E x ; Φ R , G , B
In order to learn the most-fitting enhance mapping utilizing ALLE network without any reference, we introduce a suite of differentiable luminosity and color loss functions that enable us to assess the quality of enhanced images while supervisory information is not required at all. Our ALLE network is trained using the following three types of loss functions.
Intensity consistency loss. The loss function L I n t s aims to foster intensity consistency of the images after enhancement. This is achieved by minimizing the difference value between the adjacent domains of the dark images and their corresponding enhanced domains, as represented by the following equation:
L I n t s = 1 M i = 1 M j ω i Z i Z j Y i Y j 2
where M represents the number of divided domains, and ω i denotes a set that contains the five domains including the center domain i and the upper, down, left, and right neighboring domains. The average intensity values of the five domains in the enhanced image and dark image are denoted as Z and Y , respectively.
Exposure constrain loss. To prevent the problem of over exposure or under exposure occurring in the enhanced version, we incorporate an exposure constrain loss function L e x p to regulate the exposure level:
L e x p = 1 N k = 1 N Z k E
where N denotes the number of divided domains of size 8 × 8 which do not overlap each other, Z k is the average intensity value of the k t h divided domains in the enhanced version, and E represents the satisfactory exposure degree of an image, which is set to 0.55. The function encourages the network to produce enhanced low-light images with satisfactory exposure domains and penalizes over exposure or under exposure domains.
Color channel loss. The color channel loss function L c o l is formulated as follows:
L c o l = p , q ε J p J q 2 , ε = R , G , R , B , G , B
where J p and J q denote the average intensity value of the p t h and q t h channels in the enhanced image, respectively. p , q represents a pair of channels included R , G , R , B , and G , B . The color constancy loss function encourages the DCE network to produce enhanced images that are color consistent by minimizing the difference between color channels.
Total enhanced loss. The total loss is formulated as follows:
L t o t a l _ e = L s p a + L e x p + w c o l L c o l
where w c o l is the weight of the losses.
The proposed low light enhancement network (ALLE) is only appropriate for dimly lit settings and cannot be applied in other harsh weather conditions that hinder visibility, for instance, fog and torrential rain. To clarify, the developed ALLE optimizes low-light images to normal-light conditions, and is not designed as a mechanism to remove any obstructions that might impede clear sight.

3.3. Symmetric Feature Flipping Module (SFFM)

The global structure of lanes in driving scenes is examined from a first-person perspective, where it is observed that lanes manifest symmetry. Specifically, in non-curved scenarios, the presence of the left lane line implies the existence of the corresponding lane on the right, indicating a symmetric relationship. This relationship allows the structure of the left lane to inform the description of the missing visual information of the right lane, and vice versa. To leverage this symmetry property, a lane structural loss function is introduced to preserve the global structure of lanes in the enhanced image. This approach effectively enhances the visibility of both the left and right lane lines, thereby preserving the overall structure of the lanes.
To effectively utilize this observation, we propose the symmetric feature flipping module (SFFM) in this paper, as illustrated in Figure 1. The SFFM comprises two distinct convolution and normalization layers that transform each feature map independently. The transformed feature maps are then combined and passed through a ReLU activation function. Specifically, the lower-level feature map contains abundant position information. To obtain a new feature map, we perform a symmetric flip of the lower-level feature map along its longitudinal axis. As shown in Figure 1, the blue and orange regions of the feature map are symmetrically reversed and then concatenated together. A significant proportion of the examples in both our training set and test set consist of either straight or nearly straight slow turns. Approximately 70% of these turns are symmetric or nearly symmetric, 28% are not completely symmetric, and a small percentage of 1.2% are sharp turns. It is important to note that the road lanes are not strictly axisymmetric, but this processing step is implemented to improve the attention scores of the blocked and lost lanes. Notably, we employ deformable convolution [29] to replace the standard convolution operation to address problems caused by camera angle changes, such as jitter and rotation. The proposed module enhances the low-level features and gains more precise lane localization.

3.4. Channel Fusion Self-Attention Mechanism (CFSAT)

Lane detection is a critical computer vision task that demands a balance between high level semantic information understanding and precise low level localization feature refining. To attain this balance, we propose a two-stage approach, wherein we first detect lanes using high-level semantic information and then refine the results based on low-level features. By leveraging the complementary nature of these stages, our method produces more accurate lane detections. In particular, we introduced the symmetric feature fusion module (SFFM) in the previous section to refine the low-level features. However, since a simple segmentation network may not be sufficient to gather adequate global context information, we developed the channel fusion self-attention mechanism (CFSAT) in the auxiliary segmentation branch, as demonstrated in Figure 3. This mechanism enhances the representation of advanced semantic features of lanes by incorporating more global context information.
The resulting updated feature map, denoted M u p d a t e C × H × W , can be expressed by the following formula:
M u p d a t e = M + C o n v S A C o n v M
In the equation above, the feature map before updating with channel fusion self-attention mechanism (CFSAT) is denoted as M C × H × W . C o n v represents a convolutional operation. To further enhance the feature map, an attention mechanism referred to as S A is employed to update the map with attention weights. The S A operation utilizes an attention matrix to weigh different features in the map, resulting in improved accuracy in lane detection. Mathematically, the S A operation is expressed as follows:
S A C o n v M = C o n v M s o f t m a x ( ψ s u m 2 G A C o n v M © G M C o n v M
where the symbol © represents the concatenation operation, while G A and G M denote global average pooling and global maximum pooling, respectively. The function ψ s u m 2 computes the square of the sum of corresponding pixels across channels. The choice of the ψ s u m 2 operator for channel fusion is based on the results of SAD. The SAD experiment revealed that using ψ s u m 2 operator for channel fusion yields maximum performance improvement.

3.5. A Novel Lane Structural Loss Function

Most detection models only utilize segmentation loss or classification loss to supervise detection results, disregarding the structure of lane lines. Nevertheless, we recognize that lane lines possess strong geometric prior information. Specifically, lane lines are typically thin, white strips with symmetric properties. Furthermore, due to the perspective principle, the absolute value of slope of the lane lines on both sides of the lane in which the vehicle is travelling is greater than that of adjacent lanes, as illustrated in Figure 4. In this paper, we introduce the probability P i , j , k to indicate the probability that a grid located at j , k in the i t h lane lines belongs to the lane. To ensure differentiability of the function, we use the softmax function to obtain the probability of different locations:
P r o b a b i l i t y i , j , : = s o f t m a x P i , j , 1 : w
where P r o b a b i l i t y i , j , : represents the probability at each location, P i , j , 1 : w is a w-dimension vector and 1: w indicates from column 1 to column w. It is worth noting that P i , j , k is a ( w + 1 ) d i m e n s i o n vector, where the ( w + 1 ) d i m e n s i o n indicates the probability of presence of a lane. The column localization P o s i , j of the i t h lane in the j t h row is transformed into a mathematical expectation, which is expressed as,
P o s i , j = k = 1 w k P r o b a b i l i t y i , j , k
where w represents the number of columns for partition. In practice, the localization points of lane lines between adjacent rows should be as close as possible. However, due to perspective relationships, there should be a geometric difference of α between the lane localization points of adjacent rows. To address this issue, we propose a novel structural loss function, denoted as L s t r u c , which can be expressed mathematically as follows:
L s t r u c = i = 1 C j = 1 h 1 P o s i , j + 1 P o s i , j α · M
where C denotes the number of lanes and h denotes the number of presetting row demarcation. Meanwhile, M is a binary mask with a value of 1 when i = l 1 or l 2 , and 2 otherwise. Here, l 1 and l 2 correspond to the lanes with the maximum and second largest absolute values of slopes, respectively. Their calculation formulas are as follows:
l 1 = argmax i | P o s i , h P o s i , 1 h 1 |
l 2 = argsecmax i | P o s i , h P o s i , 1 h 1 |
To regulate the output results in the primary branch, we utilize the classification loss function denoted L c l s , which can be expressed as follows:
L c l s = i = 1 C j = 1 h L C E P i , j , : , G T i , j , :
where G T i , j , : denotes the one-hot label of the correct locations, and L C E represents the cross-entropy loss function. In addition to L c l s , we also utilize cross-entropy as the auxiliary segmentation loss function, denoted as L s e g , in the auxiliary segmentation branch. Combining these loss functions, we obtain the overall loss of our method can be expressed as:
L t o t a l = L c l s + w s e g L s e g + w s t r u c L s t r u c
where w s e g and w s t r u c are the weights of segmentation loss function L s e g and structural loss function L s t r u c , respectively.

4. Experiments

CULane [3] is one of the most comprehensive and challenging publicly available datasets for lane detection, which covers a wide range of scenarios and conditions. The dataset contains images with a resolution of 1640 × 590 pixels, and categorizes them into nine different classes according to the degree of difficulty and complexity. To evaluate the performance and robustness of our proposed method, we conducted extensive experiments on the CULane dataset, which comprises 133,235 images in total. The dataset is split into three subsets: a training set with 88,880 images, a validation set with 9675 images, and a test set with 34,680 images. In this section, we first introduce the evaluation metrics and some of the implementation details that are used in our experiments. Then, we report and analyze the experimental results on the CULane dataset and compare them with several state-of-the-art methods. Finally, we present an ablation study to investigate the effectiveness of different components and settings in our method.

4.1. Evalutaion Metrics

For the CULane dataset, the official metric used is the F 1 score proposed in [3].
F 1 = 2 P r e c i s i o n R e c a l l P r e c i s i o n + R e c a l l
where P r e c i s i o n = T P T P + F P and R e c a l l = T P T P + F N . The CULane dataset is evaluated according to a specific protocol that involves representing the ground truth and predicted lane lines as 30-pixel-wide curves in the image space. This width of 30 pixels is used to define the lane lines of the dataset. The match was calculated for each whole line. A prediction line and a ground truth line are considered a match if their pixel IoU is over 0.5.

4.2. Implementation Details

In our implementation, we set the row demarcation for the CULane dataset according to the image height, which is 590 pixels. The row demarcation is a set of horizontal lines that divide the image into several regions for lane detection. We set the row demarcation to start from 0 pixels and end at 530 pixels with an interval of 10 pixels, by considering that the lane lines consistently appear at the bottom of the image (0 pixels high), but are not visible in the top portion of the image (around 530 pixels high), which typically features the sky, mountains, or distant objects in the driver’s view. The interval represents the number of divided rows, and each grid cell provides an anchor to be selected for lane prediction. To balance model efficiency and accuracy, we opted for a relatively modest interval and grid cell. Specifically, we set the interval to 10 pixels and the number of grid cells per row to 155 for this dataset.
We resized the images to 288 × 800 following the method in [3] in the optimization process. With a cosine decay learning rate strategy [30], we trained our model using the SDG-momentum optimizer [31], and the momentum was set to 0.9. The learning rate was initialized to 4 × 10−4. We employed ResNet as our backbone network, specifically utilizing ResNet18 and ResNet34. For this architecture, the convolution kernel size of the backbone network is 3 × 3, and both the stride and padding were set to 1. The loss coefficients w s e g and w s t r u c in Equation (15) are both set to 0.8 to balance the segmentation loss and the structural loss. We use a batch size of 16, and the model is trained for 60 epochs on the CULane dataset. All training and testing are performed using PyTorch 1.9 [32], an Intel(R) Core(TM) i7-4790K and an NVIDIA GTX 1080Ti GPU with 11 GB memory.

5. Results

We present the quantitative results of LLFLD on the CULane dataset in Table 1, where we compare our method with other advanced methods in terms of total F1 score and running speed. We also show some qualitative results of nine scenarios randomly selected from the test set of the CULane dataset in Figure 5, where we visualize the predicted lane lines by proposed methods. As can be seen from Table 1, our method achieves the best performance among all the compared methods on the CULane dataset, with an F1 score of 75.2% while running at 177 FPS. Compared to UFLD, which has the same research type as ours, our method has a significant advantage in accuracy while maintaining a comparable speed. For example, our lightweight model based on ResNet-18 achieves an F1 score of 72.6% while running at 330 FPS, while UFLD’s lightweight model based on ResNet-18 achieves an F1 score of 68.7% while running at 322 FPS on the same machine. Moreover, our method also surpasses the advanced method Lane ATT method, which uses an attention mechanism to enhance the feature representation for lane detection. For instance, our model based on ResNet-34 achieves an F1 score of 75.2% while running at 176.3 FPS, while Lane ATT’s model based on ResNet-34 achieves an F1 score of 74.1% while running at 140 FPS on the same machine. Furthermore, our method shows remarkable improvement in some challenging scenarios, such as night and shadow scenes. For example, in the “night” scene, our method achieves an F1 score of 66.0%, which is 3.4% higher than Lane ATT’s F1 score of 62.6%. Similarly, in the “shadow” scene, our method achieves an F1 score of 66.9%, which is 2.3% higher than Lane ATT’s F1 score of 64.6%. These results demonstrate that our framework and structural loss function can effectively handle the low-light and low-contrast situations and produce accurate and robust lane detection results, as our method can enhance images for low light scenes and utilize abundant global information and geometric information to tackle the no-visual-clue problems.
We show some visualizations of our proposed LLFLD method on the CULane dataset in Figure 5, where we display the lane lines predicted by our method on nine different road conditions. As can be seen from the figure, our method performs well under various scenarios and challenges, such as normal, crowded, dazzle and so on. The figure demonstrates that our method can accurately and robustly detect the lane lines under different illumination, occlusion, and curvature situations.

6. Ablation Study

This experiment aims to evaluate the impact of each major part of the proposed design on the lane detection performance. The four major parts are: automatic low-light scene enhancement (ALLE), symmetric feature flipping module (SFFM), channel fusion self-attention mechanism (CFSAT), and the proposed new structural loss function L s t r u c . To evaluate the performance of our proposed modules, we conducted a series of experiments under the same training settings but with different combinations of modules. The quantitative results of our modules, measured by the same metrics, are presented in Table 2.
To demonstrate the effectiveness of the ALLE module, which is a novel adaptive low-light enhancement module that can adjust the brightness and contrast of the input image according to the ambient illumination, we conducted a comparative experiment on the CULane dataset with night scenes. We compared the detection results before and after turning on the ALLE module on the CULane dataset in night scenes, as shown in Figure 6.
To assess the efficacy of the ALLE module in low-light scenarios, we utilized F1 values in shadow and night scenarios as our evaluation indicator. Table 3 illustrates the ablation experiments performed under two states—with ALLE activated and with ALLE deactivated. From Table 3, we can reach the conclusion that ALLE being activated achieves a significant improvement in performance compared to ALLE being deactivated, which provides evidence for the effectiveness of the proposed module in the low-light scenarios, such as shadow and night.
From Table 2, we can reach the conclusion that all proposed designs achieve a significant improvement in performance compared to the baseline, which provides evidence for the effectiveness of the proposed modules.
It can be observed that Figure 6 visually demonstrates that the detection results of the model with ALLE are better than those without ALLE in the same low-light scenes, which provides more intuitive evidence for the effectiveness of ALLE. Specifically, the ALLE module can help the model detect lanes accurately, even in low-light conditions when they may not be visible to the human eyes. This is a significant advantage of the proposed method, as it can improve the safety and reliability of lane detection in real-world scenarios with low-light conditions.

7. Conclusions

In this paper, we proposed LLFLD: a novel lane detection model specifically designed for enhancing low light scenes based on row anchor selection. We introduced the automatic low-light scene enhancement (ALLE), an adaptive low-light enhancement module that can adjust the brightness and contrast of the input image, according to the ambient illumination, to optimize the lane detection results in low-light scenarios. Our approach leverages feature flipping and channel self-attention mechanisms to effectively collect and utilize both low-level location information and high-level semantic information from the feature maps. Additionally, our new structural loss function leverages the geometric priori of the lane to optimize the detection results. Moreover, our design based on row selection ensures that the model is fast and lightweight, which is suitable for real-time applications. Experimental results on the popular CULane dataset demonstrate the favorable performance (measured by F1 score) of our proposed model. Furthermore, our method achieves fast inference speed and is lightweight. Specifically, the ResNet-34 version of our method can achieve 177 FPS while maintaining comparable performance at the same resolution. Overall, our proposed method is effective in enhancing low-light scenes for better lane detection results while maintaining fast inference speeds and low computational complexity.

Author Contributions

Conceptualization, C.K., Z.X., J.Z. and D.Z.; methodology, C.K. and Z.X.; software, C.K. and D.Z.; validation, C.K., Z.X. and J.Z.; formal analysis, C.K.; investigation, C.K. and Z.X.; resources, C.K. and J.Z.; data curation, C.K. and D.Z.; writing—original draft preparation, C.K.; writing—review and editing, C.K. and Z.X.; visualization, C.K.; supervision, C.K. and D.Z.; project administration, C.K.; funding acquisition, C.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Beijing Natural Science Foundation (grant No. 8202013) and the Beijing University of Civil Engineering and Architecture Graduate Student Innovation Project [No. PG2022145].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: https://xingangpan.github.io/projects/CULane.html (accessed on 23 October 2022).

Acknowledgments

In this section, you can acknowledge any support given which is not covered by the author contribution or funding sections. This may include administrative and technical support, or donations in kind (e.g., materials used for experiments).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Forster, M.; Frank, R.; Gerla, M.; Engel, T. A cooperative advanced driver assistance system to mitigate vehicular traffic shock waves. In Proceedings of the IEEE INFOCOM 2014-IEEE Conference on Computer Communications, Toronto, ON, Canada, 27 April–2 May 2014; pp. 1968–1976. [Google Scholar]
  2. Throngnumchai, K.; Nishiuchi, H.; Matsuno, Y.; Satoh, H. Application of Background Light Elimination Technique for Lane Marker Detection; SAE Technical Paper: Warrendale, PA, USA, 2013; ISSN 0148-7191. [Google Scholar]
  3. Pan, X.; Shi, J.; Luo, P.; Wang, X.; Tang, X. Spatial as deep: Spatial cnn for traffic scene understanding. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–3 February 2018. [Google Scholar]
  4. Hou, Y.; Ma, Z.; Liu, C.; Loy, C.C. Learning lightweight lane detection cnns by self attention distillation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1013–1021. [Google Scholar]
  5. Qin, Z.; Wang, H.; Li, X. Ultra fast structure-aware deep lane detection. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; pp. 276–291. [Google Scholar]
  6. Yu, B.; Jain, A.K. Lane boundary detection using a multiresolution hough transform. In Proceedings of the International Conference on Image Processing, Santa Barbara, CA, USA, 26–29 October 1997; pp. 748–751. [Google Scholar]
  7. Canny, J. A computational approach to edge detection. IEEE Trans. Pattern Anal. Mach. Intell. 1986, PAMI-8, 679–698. [Google Scholar] [CrossRef]
  8. Sun, T.-Y.; Tsai, S.-J.; Chan, V. HSI color model based lane-marking detection. In Proceedings of the 2006 IEEE Intelligent Transportation Systems Conference, Toronto, ON, Canada, 17–20 September 2006; pp. 1168–1172. [Google Scholar]
  9. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28, 91–99. [Google Scholar] [CrossRef] [PubMed]
  10. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  11. Lee, S.; Kim, J.; Shin Yoon, J.; Shin, S.; Bailo, O.; Kim, N.; Lee, T.-H.; Seok Hong, H.; Han, S.-H.; So Kweon, I. Vpgnet: Vanishing point guided network for lane and road marking detection and recognition. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1947–1955. [Google Scholar]
  12. Neven, D.; De Brabandere, B.; Georgoulis, S.; Proesmans, M.; Van Gool, L. Towards end-to-end lane detection: An instance segmentation approach. In Proceedings of the 2018 IEEE Intelligent Vehicles Symposium (IV), Suzhou, China, 26–30 June 2018; pp. 286–291. [Google Scholar]
  13. Li, X.; Li, J.; Hu, X.; Yang, J. Line-cnn: End-to-end traffic line detection with line proposal unit. IEEE Trans. Intell. Transp. Syst. 2019, 21, 248–258. [Google Scholar] [CrossRef]
  14. Feng, Z.; Guo, S.; Tan, X.; Xu, K.; Wang, M.; Ma, L. Rethinking efficient lane detection via curve modeling. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 17062–17070. [Google Scholar]
  15. Tabelini, L.; Berriel, R.; Paixao, T.M.; Badue, C.; De Souza, A.F.; Oliveira-Santos, T. Keep your eyes on the lane: Real-time attention-guided lane detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 294–302. [Google Scholar]
  16. Yoo, S.; Lee, H.S.; Myeong, H.; Yun, S.; Park, H.; Cho, J.; Kim, D.H. End-to-end lane marker detection via row-wise classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 1006–1007. [Google Scholar]
  17. Ibrahim, H.; Kong, N.S.P. Brightness preserving dynamic histogram equalization for image contrast enhancement. IEEE Trans. Consum. Electron. 2007, 53, 1752–1758. [Google Scholar] [CrossRef]
  18. Hao, S.; Han, X.; Guo, Y.; Wang, M. Decoupled Low-Light Image Enhancement. ACM Trans. Multimed. Comput. Commun. Appl. (TOMM) 2022, 18, 1–19. [Google Scholar] [CrossRef]
  19. Jiang, H.; Zheng, Y. Learning to see moving objects in the dark. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 7324–7333. [Google Scholar]
  20. Sada, A.; Kinoshita, Y.; Shiota, S.; Kiya, H. Histogram-based image pre-processing for machine learning. In Proceedings of the 2018 IEEE 7th Global Conference on Consumer Electronics (GCCE), Nara, Japan, 9–12 October 2018; pp. 272–275. [Google Scholar]
  21. Ma, S.; Li, X.; Xiao, M.; Ma, B.; Zhao, Y. Fast expansion-bins-determination for multiple histograms modification based reversible data hiding. IEEE Signal Process. Lett. 2022, 29, 662–666. [Google Scholar] [CrossRef]
  22. Yang, G.; Evans, A.N. Improved single image dehazing methods for resource-constrained platforms. J. Real-Time Image Process. 2021, 18, 2511–2525. [Google Scholar] [CrossRef]
  23. Çimtay, Y.; Yilmaz, G.N. Low Light Image Enhancement on Mobile Devices by Using Dehazing. In Proceedings of the Electrical and Computer Engineering: First International Congress, ICECENG 2022, Virtual Event, 9–12 February 2022; pp. 57–67. [Google Scholar]
  24. Li, C.; Guo, C.; Han, L.; Jiang, J.; Cheng, M.-M.; Gu, J.; Loy, C.C. Low-light image and video enhancement using deep learning: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 9396–9416. [Google Scholar] [CrossRef] [PubMed]
  25. Wei, C.; Wang, W.; Yang, W.; Liu, J. Deep retinex decomposition for low-light enhancement. arXiv 2018, arXiv:1808.04560. preprint. [Google Scholar]
  26. Guo, C.; Li, C.; Guo, J.; Loy, C.C.; Hou, J.; Kwong, S.; Cong, R. Zero-reference deep curve estimation for low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 1780–1789. [Google Scholar]
  27. Lin, T.-Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  28. Poynton, C.A. A Technical Introduction to Digital Video; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 1996. [Google Scholar]
  29. Zhu, X.; Hu, H.; Lin, S.; Dai, J. Deformable convnets v2: More deformable, better results. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019; pp. 9308–9316. [Google Scholar]
  30. Mandalia, H.M.; Salvucci, M.D.D. Using support vector machines for lane-change detection. In Proceedings of the Human Factors and Ergonomics Society Annual Meeting, Orlando, FL, USA, 26–30 September 2005; pp. 1965–1969. [Google Scholar]
  31. Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [PubMed]
  32. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L. Pytorch: An imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 2019, 32, 8024–8035. [Google Scholar]
  33. Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef] [PubMed]
  34. Liu, R.; Yuan, Z.; Liu, T.; Xiong, Z. End-to-end lane shape prediction with transformers. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2021; pp. 3694–3702. [Google Scholar]
  35. Philion, J. Fastdraw: Addressing the long tail of lane detection by adapting a sequential prediction network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019; pp. 11582–11591. [Google Scholar]
  36. Xu, H.; Wang, S.; Cai, X.; Zhang, W.; Liang, X.; Li, Z. Curvelane-nas: Unifying lane-sensitive architecture search and adaptive point blending. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; pp. 689–704. [Google Scholar]
Figure 1. Overview of the proposed LLFLD. The pipeline consists of a primary classification detection branch and an auxiliary segmentation branch. Notably, the auxiliary segmentation module is exclusively activated during the training phase.
Figure 1. Overview of the proposed LLFLD. The pipeline consists of a primary classification detection branch and an auxiliary segmentation branch. Notably, the auxiliary segmentation module is exclusively activated during the training phase.
Sensors 23 04917 g001
Figure 2. The framework of automatic low-light scene enhancement (ALLE). A Zero-DCE network is designed to inference a mapping Φ n that corresponds to the optimal light enhancement tactics. The network enhances an input dark image iteratively, with n representing the number of iterations that regulate the extent of image enhancement. For normal exposure images, we recognize the original images as the outputs.
Figure 2. The framework of automatic low-light scene enhancement (ALLE). A Zero-DCE network is designed to inference a mapping Φ n that corresponds to the optimal light enhancement tactics. The network enhances an input dark image iteratively, with n representing the number of iterations that regulate the extent of image enhancement. For normal exposure images, we recognize the original images as the outputs.
Sensors 23 04917 g002
Figure 3. The framework of channel fusion self-attention mechanism (CFSAT).
Figure 3. The framework of channel fusion self-attention mechanism (CFSAT).
Sensors 23 04917 g003
Figure 4. The illustration of design motivation of the structural loss function. The lane lines can be seen as a white strip in most conditions. From the perspective of driving, the absolute value of the slope of the lane lines on both sides of the driving lane (represented by the red lines) is always greater than the absolute value of the slope of the adjacent lane lines (represented by the green lines) due to the perspective principle.
Figure 4. The illustration of design motivation of the structural loss function. The lane lines can be seen as a white strip in most conditions. From the perspective of driving, the absolute value of the slope of the lane lines on both sides of the driving lane (represented by the red lines) is always greater than the absolute value of the slope of the adjacent lane lines (represented by the green lines) due to the perspective principle.
Sensors 23 04917 g004
Figure 5. Qualitative results on the test set of CULane dataset. It shows nine road conditions in the CULane dataset, from left to right and top to bottom as normal, crowded, dazzle, no-line, shadow, arrow, cross, curve, and night.
Figure 5. Qualitative results on the test set of CULane dataset. It shows nine road conditions in the CULane dataset, from left to right and top to bottom as normal, crowded, dazzle, no-line, shadow, arrow, cross, curve, and night.
Sensors 23 04917 g005
Figure 6. Qualitative comparison of detection results with and without automatic low-light scene enhancement (ALLE). The detection results of the night scenario in the same lane are shown. (a) shows the visualization of results without ALLE, while (b) shows the counterpart with ALLE.
Figure 6. Qualitative comparison of detection results with and without automatic low-light scene enhancement (ALLE). The detection results of the night scenario in the same lane are shown. (a) shows the visualization of results without ALLE, while (b) shows the counterpart with ALLE.
Sensors 23 04917 g006
Table 1. Comparison of F1-measure and FPS on CULane testing set with IoU threshold = 0.5. For crossroads, only false positives are shown. The less, the better. ‘-’ means the result is not available. The best and second-best results across methods are in bold and underlined, respectively. All the images are resized to 288 × 800, and all the experiments were computed on a machine with an Intel(R) Core(TM) i7-4790K CPU and an RTX1080Ti GPU.
Table 1. Comparison of F1-measure and FPS on CULane testing set with IoU threshold = 0.5. For crossroads, only false positives are shown. The less, the better. ‘-’ means the result is not available. The best and second-best results across methods are in bold and underlined, respectively. All the images are resized to 288 × 800, and all the experiments were computed on a machine with an Intel(R) Core(TM) i7-4790K CPU and an RTX1080Ti GPU.
MethodTotalNormalCrowdedDazzleShadowNo lineArrowCurveCrossNightFPSMACs(G)
Res50-Seg [33]66.7087.4364.1054.1260.7038.1079.0059.83250560.60--
LSTR(ResNet-18,2×) [34]68.7286.7867.3456.6359.8240.1078.6656.64116659.92--
FastDraw (ResNet-50) [35]67.1385.9063.6057.0059.9040.6079.4065.20701357.8090.3-
SCNN [3]71.6090.6069.758.5066.9043.4084.1064.40199066.107.5328.4
ENet-SAD [4]70.890.1068.8060.2065.9041.6084.0065.70199866.00757.8
UFLD(ResNet-18) [5]68.4087.7066.0058.4062.8040.2081.0057.90174362.10322.5-
UFLD(ResNet-34) [5]72.3090.7070.2059.5069.3044.4085.7069.50203766.70175.0-
CurveLanes-NAS-M [36]73.5090.2070.5065.9069.3048.8085.7067.50235968.20-35.7
Res18-Ours71.3089.2067.2058.5063.3042.5082.8058.00181966.5033017.4
Res34-Ours75.2091.0071.8065.3070.2047.8086.2069.50191370.5017733.2
Table 2. Performance comparison of proposed modules on CULane dataset with ResNet-34 backbone. The baseline indicates that none of the modules are added. The F1-measure is computed on the CULane testing set with an IoU threshold of 0.5.
Table 2. Performance comparison of proposed modules on CULane dataset with ResNet-34 backbone. The baseline indicates that none of the modules are added. The F1-measure is computed on the CULane testing set with an IoU threshold of 0.5.
BaselineLow-Light
Enhancement
Flipping
Module
Attention MechanismStructural
Loss
F1
72.1
73.8 (+1.7)
74.3 (+2.2)
74.9 (+2.8)
75.2 (+3.1)
Table 3. Quantitative comparison of detection results on “shadow” and “night” scenarios with and without automatic low-light scene enhancement (ALLE). The F1-measure is computed on the CULane testing set with an IoU threshold of 0.5.
Table 3. Quantitative comparison of detection results on “shadow” and “night” scenarios with and without automatic low-light scene enhancement (ALLE). The F1-measure is computed on the CULane testing set with an IoU threshold of 0.5.
W/O ALLEW/ALLEShadow-F1Night-F1
68.9065.73
70.2070.50
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ke, C.; Xu, Z.; Zhang, J.; Zhang, D. Combining Low-Light Scene Enhancement for Fast and Accurate Lane Detection. Sensors 2023, 23, 4917. https://doi.org/10.3390/s23104917

AMA Style

Ke C, Xu Z, Zhang J, Zhang D. Combining Low-Light Scene Enhancement for Fast and Accurate Lane Detection. Sensors. 2023; 23(10):4917. https://doi.org/10.3390/s23104917

Chicago/Turabian Style

Ke, Changshuo, Zhijie Xu, Jianqin Zhang, and Dongmei Zhang. 2023. "Combining Low-Light Scene Enhancement for Fast and Accurate Lane Detection" Sensors 23, no. 10: 4917. https://doi.org/10.3390/s23104917

APA Style

Ke, C., Xu, Z., Zhang, J., & Zhang, D. (2023). Combining Low-Light Scene Enhancement for Fast and Accurate Lane Detection. Sensors, 23(10), 4917. https://doi.org/10.3390/s23104917

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop