Next Article in Journal
Decoding Mental Effort in a Quasi-Realistic Scenario: A Feasibility Study on Multimodal Data Fusion and Classification
Previous Article in Journal
A Lateral Flow Assay for the Detection of Leptospira lipL32 Gene Using CRISPR Technology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Interactive Attention Learning on Detection of Lane and Lane Marking on the Road by Monocular Camera Image

1
Tongji University, Shanghai 201804, China
2
Institute of Measurement and Control Systems, Karlsruhe Institute of Technology, 76131 Karlsruhe, Germany
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(14), 6545; https://doi.org/10.3390/s23146545
Submission received: 29 May 2023 / Revised: 9 July 2023 / Accepted: 17 July 2023 / Published: 20 July 2023
(This article belongs to the Section Vehicular Sensing)

Abstract

:
Vision-based identification of lane area and lane marking on the road is an indispensable function for intelligent driving vehicles, especially for localization, mapping and planning tasks. However, due to the increasing complexity of traffic scenes, such as occlusion and discontinuity, detecting lanes and lane markings from an image captured by a monocular camera becomes persistently challenging. The lanes and lane markings have a strong position correlation and are constrained by a spatial geometry prior to the driving scene. Most existing studies only explore a single task, i.e., either lane marking or lane detection, and do not consider the inherent connection or exploit the modeling of this kind of relationship between both elements to improve the detection performance of both tasks. In this paper, we establish a novel multi-task encoder–decoder framework for the simultaneous detection of lanes and lane markings. This approach deploys a dual-branch architecture to extract image information from different scales. By revealing the spatial constraints between lanes and lane markings, we propose an interactive attention learning for their feature information, which involves a Deformable Feature Fusion module for feature encoding, a Cross-Context module as information decoder, a Cross-IoU loss and a Focal-style loss weighting for robust training. Without bells and whistles, our method achieves state-of-the-art results on tasks of lane marking detection (with 32.53% on IoU, 81.61% on accuracy) and lane segmentation (with 91.72% on mIoU) of the BDD100K dataset, which showcases an improvement of 6.33% on IoU, 11.11% on accuracy in lane marking detection and 0.22% on mIoU in lane detection compared to the previous methods.

1. Introduction

Lanes and lane markings are essential road information for intelligent driving vehicles. The lane marking detection aims to accurately locate road elements like lane lines, crosswalks, and stop zones, while the lane detection focuses on segmenting lane-level areas where vehicles can drive on the road. Due to the low cost and the high representability of scene information, optical sensors and instruments, such as the on-board camera, are widely adopted for road information perception. By applying lane and lane marking detection approaches, visual features of road symbols, arrows, lane markings, pedestrian crosswalks, and vehicle drivable areas, etc., are extracted from the image. These features are indispensable for both high-level autonomous driving or for general ADAS-assisted driving systems. They can be considered as elements in the high-definition map construction, or further converted into the information required by the planning and control system, to assist the driving behavior of vehicles, especially in applications such as adaptive cruise control (ACC), driving route navigation, lane keeping assistance (LKA), etc., thus ensuring driving safety and reliability [1,2,3].
Generally, the detection of lanes and lane markings can be classified in two categories: the traditional paradigm [4,5], and the deep learning paradigm [6,7,8]. Traditional methods rely on hand-crafted features and sophistically designed rules to manipulate the information from color space or shape structure to detect lanes and lane markings. Due to their poor feature representability, these methods are only limited scalable to varied scenes. In recent years, the deep learning approaches in computer vision have achieved remarkable progress, especially in object detection and semantic segmentation tasks. Since the lanes and lane lines are normally made with inherent long and thin shapes and even irregular ones, the difficulty lies in the exploration of effective representation learning of their complex structures.
In current studies, the detection of lanes and lane markings are considered as two individual tasks. The lane detection is typically interpreted as a pixel-wise semantic segmentation problem while the lane marking can be predicted with various formulations such as instance segmentation [9,10], point regression [11,12], curve parameter estimation [13,14,15], etc. Although both tasks have witnessed persistent progress in recent years, especially on public benchmarks [9,16,17], one fact that has been neglected is that the information of lanes and lane markings on the road scene are complementary to each other. For instance, on structured roads, the associated lane lines can be used to identify the lane boundaries while in scenarios where lane lines or crosswalks are partially missing or broken (due to occlusion), they can still be inferred by the width of the lane. Thus, the detection of lane and lane marking are inherently correlated due to their spatial connectivity. In real driving scenarios, the detection robustness of a single task is poor, and it can be easily disturbed by the disappearance of visual markings, e.g., due to occlusions. However, leveraging the spatial connectivity between the lane and lane marking, the detection robustness can be improved by modeling this internal connection, which has not been studied in existing methods.
Inspired by this fact, this work proposes a novel multi-task encoder–decoder architecture for simultaneous detection of lane and lane marking based on the spatial relationship between them. Specifically, this architecture interprets both tasks as semantic segmentation problems and adopts multi-scale inputs. In the encoding stage, it employs a feature interactive attention structure, namely the Deformable Feature Fusion (DFF) module, to calculate a relative offset between feature encodings of lane and lane marking, supporting a deformable convolution operation for interaction. The decoding for lanes and lane markings is conducted separately, and a Cross-Context module is used to transfer the decoding information between them. To further exploit their spatial relationship, we add a Cross Intersection-over-Union (CIoU) loss at the output of the lane and lane marking decoders. Thus, the interactive learning has been leveraged at the encoding stage, the decoding stage, and the outputs. Furthermore, we deploy a Focal-style loss weighting to adaptively set loss weights at different pixel locations to alleviate the data imbalance problem in lane and lane marking segmentation. The whole framework is shown in Figure 1. By testing on the BDD100K dataset, our method manifests a state-of-the-art performance on the lane detection (with 32.53% lane IoU, 81.61% lane accuracy) and the drivable area segmentation (with 91.72% drivable mIoU) tasks.
The main contributions of the proposed work are summarized as follows:
We propose a novel multi-task encoder–decoder architecture. It is the first to introduce the concept of interactive attention learning into the joint detection of lane and lane marking.
We propose the DFF module in calculating discriminative encoding features, and employ the Cross-Context module to transfer information between prediction heads, thus shifting the focus of learning on the spatial correlation between lane and lane marking.
We propose an enhanced loss function with a CIoU loss to emphasize the lane and lane marking interaction and an adaptive pixel-level loss weighting to alleviate data imbalance.

2. Related Works

In this section, we give a brief review about related works in terms of the task setup and architecture design of our work, i.e., the lane marking detection approaches, the lane detection approaches, and multi-task approaches.

2.1. Lane Marking Detection

As aforementioned, traditional lane marking detection approaches generally rely on sophisticated model design and hand-crafted features, involving color conversion [4], combination of Kalman and particle filter [18], bar filter [19] and Hough transform [5]. These approaches directly output lane segments, which are further post-processed to remove false positives and grouped to form the lane markings. Aly [20] proposed a robust real-time lane marking detection method, which first generated a top view image by projection transform and then extracted lane markings using a bar filter and a simple Hough transform. Assidiq et al. [21] detected edges with the Canny operator and extracted line features through the Hough transform. The lane marking was obtained by line fitting to selected pixels. However, limited by the poor feature representation, traditional methods show inrobustness in complex scenarios, such as with broken lane markings or occlusion by vehicles and pedestrians.
In recent years, the deep learning technique has significantly boosted the lane marking detection performance. According to the modeling strategy, such approaches can be classified into four categories: segmentation-based, anchor-based, row-wise detection, and parametric prediction methods. The segmentation-based methods commonly adopt the semantic segmentation or instance segmentation to make pixel-wise predictions [6,9,10]. Supervised by a sufficient amount of labeled data, these approaches show advantages in detecting various kinds of lane markings. The aerial LaneNet [22] proposed a fully convolutional neural network in a symmetrical structure, which is enhanced by wavelet transform for lane marking segmentation in aerial imagery. Guan et al. [23] incorporated the attention mechanism into FPN networks to extract better road marking segmentation results from high resolution UAV images. The anchor-based methods leverage the anchor concept from traditional object detection, but differ from them by taking into account the shape characteristics of lane markings. For instance, the PointLaneNet [7] and CurveLane-NAS [24] define anchors with vertical lines, while the Line-CNN [11] and LaneATT [12] adopt the Line Proposal Unit, which resembles the Region Proposal network (RPN) of the Faster-RCNN [25]. The row-wise detection approaches make full use of the prior shape of lane markings as well as their spatial distribution characteristics. They divide the image into grids and make row-wise predictions to locate the lane markings [26,27,28]. In contrast, the parametric prediction methods define lane markings (especially lane lines) as curve functions with a set of parameters, such as polynomials [13,14], and Bézier curves [15]. Their interpretations are significantly different from the above-mentioned methods and the corresponding curve parameters are difficult to learn. In addition, to solve the problem of difficult scenes for lane marking detection such as occlusion and low-visibility, Wang et al. [29] proposed a dynamic data augmentation framework based on imitating real scenes.

2.2. Lane Detection

The task of lane detection is also known as the drivable area detection, which is mainly classified as a segmentation task at present. As a result of the great successes of the deep learning, many methods based on semantic segmentation and instance segmentation can be transferred to the drivable area detection. The FCN [30] is the first work to introduce the fully convolutional network to semantic segmentation, which makes CNN-based methods widely applicable for lane detection. The UNet [31] further constructs an encoder–decoder framework to extract lane semantic information from high-dimensional features. The DeepLabV3 [32] combines the atrous convolutions [33] with different artous rates to fuse the feature pyramid, namely ASPP, obtaining different receptive fields on feature maps. The PSPNet [34] proposes the pyramid pooling module for feature extraction of various scales, which enhances the accuracy of the model. It is also worth noting that both DeepLabV3 and PSPNet leverage the fusion of multi-scale feature information to improve the segmentation performance. He et al. [35] embedded the Swin transformer into the classical network (UNet) to improve the semantic segmentation performance for remote sensing images. Xie et al. [36] presented a segmentation method for RGB-D data and adopted the motion detection to improve the inference accuracy. Meyer et al. [37] expanded the Cityscapes dataset [38] by lane-level annotations and presented a novel lane detection pipeline, which used the stereo system to convert the front-view segmentation results into a form of 3D point cloud and projected it to the top-view. Sun et al. [39] proposed to leverage crowd-sourced GPS data to extract roads from an aerial image, which achieved improved road segmentation compared to previous works. Fontanelli et al. [40] performed lane detection in the front-view image and projected it to the top-view for the construction of the path, which is used to plan the future motion of the robot.

2.3. Multi-Task Approaches

Although previous studies have achieved excellent performance in a single detection or segmentation task, the multi-task architecture to process perception information is more friendly to practical applications. The goal of multi-task approaches is to establish a trade-off between the detection performance and the computational complexity by utilizing the shared feature information and model structure.
The MultiNet [41] first introduces a multi-task architecture into the autonomous driving perception task. The architecture adopts a shared backbone and three decoders to perform tasks of road segmentation, vehicle detection, and scene classification simultaneously. The DLT-Net [42] inherits the encoder–decoder architecture with a shared backbone and multi-task decoders. It transmits the information from the drivable area decoder, namely the context tensor, to both the lane marking decoder and the traffic object decoder, thus sharing the decoder information to a certain extent. The RBNet [43] proposes a multi-task neural network model for unified detection of road and road boundary, which combined the input image, road and road boundary as three nodes into a Bayesian network. Zhang et al. [44] considered the geometric constraint between the road and its boundaries and constructed interlinked sub-networks for overall performance improvement of both detection tasks. The RoadNet [45] develops a multi-task convolutional neural network to simultaneously make predictions of road boundaries, surfaces, and centerline based on the high-resolution images from remote sensing. The HYDRO-3D [46] incorporates object detection features with historical object tracking information to improve the performance of both tasks, which achieves robust object detection. Xia et al. [47] proposes a platform for automated driving system data acquisition and analysis, which presents a holistic pipeline for data processing based on connected automated vehicles. However, the exploration on the interaction between lane and lane marking information is still insufficient in the above-mentioned studies.

3. Methodology

Here, we present the proposed lane and lane marking detection architecture in detail, including a configured encoder, a DFF module, a decoder with a Cross-Context module, and loss functions with adaptive weights. Our code is publicly available at https://github.com/HerrYu123/IALaneNet, accessed on 1 December 2022.

3.1. Architecture Overview

The overall architecture as presented in Figure 1 is divided into two branches for lane and lane marking detection, respectively. Both branches have a similar structure, consisting of a scaled input image, a backbone, a contextual neck, and a decoder. The two branches are connected by three interaction modules, i.e., the DFF module for the encoder, the Cross-Context module for the decoder, and the Cross-IoU loss for the lane and lane marking outputs.
We employ a modified variant of the ConvNeXt [48] as our backbone, which generates lane and lane marking feature maps with an output stride of 8. The feature maps are processed by the neck network, i.e., the REcurrent Feature-Shift Aggregator (RESA) [10], which adopts 1 × 9 convolution kernels for the spatial feature aggregation in which the sliced feature map is shifted in horizontal and vertical directions. The aggregated feature maps from both branches are fused in the DFF module. The lane and lane marking segmentation are predicted by their output heads. In an effort to refine the results, we leverage a Cross-Context module, which consists of deformable convolutions to transfer complementary information between two prediction heads. The Cross-IoU loss is further used to enhance the interaction of segmentation results. Our entire network is end-to-end differentiable and both tasks can be jointly learned.

3.2. Encoder

3.2.1. Input

As aforementioned, the multi-scale feature information plays an crucial role for the segmentation task. Thus, the encoder of our network takes two scales of an image as inputs, i.e., the 0.5× scale and 1× scale. Each branch is made up of a backbone network and a neck network, where only the backbone network is weight-shared. This is different from other popular multi-task methods [41,42,49], which only use one scale image and one shared backbone. Considering that lane markings are relatively smaller than lane areas and thus require representation with higher resolution, we empirically set the branch with a 0.5× scale of input image for lane inference and the other for lane marking inference, which enables the branches to extract features in appropriate scales for both tasks.

3.2.2. Backbone

The ConvNeXt [48] is used as the backbone of our architecture due to its outstanding performance in object detection tasks. However, the 1/32 downsampling layer in the naive ConvNeXt discards too much spatial information. As depicted in Figure 2, we simply replace the ordinary convolution in the downsampling layer of the third stage with atrous convolution and substitute the layers in the fourth stage for a dilated block to configure the backbone with an output stride of 8, which enables the following modules to extract scene features from a higher spatial resolution.

3.2.3. Neck

The neck is used to extract contextual information from the feature maps, which are generated by the backbone. Here, we simply adopt the REcurrent Feature-Shift Aggregator modules for the neck network, as they are more efficient in gathering spatial information horizontally and vertically compared to other mainstream methods.

3.3. Deformable Feature Fusion Module

Since the lane markings are typically in very thin and complex shapes, the learning of their accurate localization becomes more challenging. To address the above problem, here we propose the Deformable Feature Fusion Module, dubbed as DFF, illustrated in Figure 3. Such a module fuses features from the lane branch to assist the spatial information learning of lane markings, which is motivated by their strong spatial correlations. Specifically, the 0.5× scale lane feature map is fed into a 1 × 1 convolution operation and then recovered to the normal scale by a 2× bilinear interpolation upsampling. It is further concatenated with the 1× scale lane marking feature map. The interaction between the lane and lane marking is interpreted in a deformable convolution form, where the concatenated feature map will be leveraged to learn the offsets of the convolution. Thus, the spatial information from both encoders can be interactively learned by this module.
To ensure the spatial information correctly embedded into the input feature map of the DFF module, we simply add an auxiliary semantic segmentation branch to the output of each neck during the training process. These auxiliary segmentation branches are supervised by groundtruth labels of the lane and lane marking. This supervision can also be considered as a generalized form of residual learning. By setting up such supervision for shallow layers, the model could learn the basic semantic features in advance while the subsequent heads could focus on the learning of high-level information.

3.4. Decoder

As mentioned above, we cast the lane and lane marking detection as semantic segmentation tasks performed in two separate heads. However, we empirically found that the lane and lane marking have distinct characteristics. First, compared to the lane area, the slender lane marking is much more complicated to segment and thus it requires a higher processing complexity. Second, we noticed that the lane features are more suitable to decode from the shallower layers of the model than the lane marking features. Hence, the setup of the decoder is as follows.

3.4.1. Lane Marking Prediction Head

Taking into account the slender characteristic of the lane markings, we adopt the Bilateral Up-Sampling Decoder (BUSD) [10] for decoding the lane marking features. The BUSD is able to combine the coarse grained feature and fine detailed feature in upsampling stage, which are extracted by direct bilinear interpolation and transpose convolution, respectively.

3.4.2. Lane Prediction Head

Since the semantic area of the lane is larger and easier to detect compared to that of lane marking, we use a 1 × 1 output convolution and three standard semantic upsample modules, each consisting of a 3 × 3 convolution, a batch norm, and a ReLU function.

3.4.3. Cross-Context Module

This Cross-Context mainly focuses on the pixel information located adjacent to the lane and lane marking, which is motivated by the their tight spatial connectivity. Since no attention is needed for the global features, we choose a convolution block to directly aggregate the information of local pixels. Taking into account the shape characteristics of lane and lane marking, the deformable CNNs [50] are leveraged in the Cross-Context module to implicitly establish a feature mapping for transferring the complementary information between two prediction heads, so that the decoding outputs can be reciprocally refined.
As illustrated in Figure 4, feature maps are first fed into the lane and lane marking prediction heads. The predicted segmentation maps are processed by the Cross-Context module and added to the input feature maps of each branch, respectively. The summed feature maps then pass the lane and lane marking decoder for the second time to obtain the final segmentation results. Thus, the Cross-Context module is able to draw the supplementary information from one decoder to another. Specifically, one Cross-Context module consists of three Cross-Context blocks for lane prediction head to lane marking prediction head, and another three for lane marking prediction head to lane prediction head. The Cross-Context block is made up of deformable convolution, batch normalization and max pooling. The Cross-Context module is expected to enhance the detection performance for both lane and lane marking with such a feature interaction of decoding information. Since the forward computation is performed twice in the decoders, it can be considered as a coarse-to-fine optimization.

3.5. Loss Function

3.5.1. Segmentation Dice Loss

As a general consensus, the Dice loss [51] comes from the Dice coefficient and is proposed in the segmentation for alleviating the data imbalance problem by adjusting the training gradient distributions of positive and negative samples. Given semantic groundtruth labels [ y i 1 , y i 2 , , y i C ] of C classes and the predicted probabilities [ p i 1 , p i 2 , , p i C ] for the i-th pixel, the Dice loss can be formulated as
L i = 1 1 C k = 1 C 2 i N y i k p i k i N ( y i k + p i k + ϵ ) ,
where N represents the number of pixels, ϵ is a small positive number to avoid zero division, and the numerical range is set as y i k { 0 , 1 } and p i k [ 0 , 1 ] . Here, the Dice loss is used for the prediction heads of lane and lane marking segmentation branches.

3.5.2. Focal-Style Loss Weighting

To further alleviate the data imbalance problem in both lane marking and lane segmentation, with the inspiration of the focal loss [52], we propose a weighting process for both the lane and lane marking segmentation loss. The intuition of this loss weighting is to dynamically adjust weights for pixel-wise segmentation results so that hard pixels can be emphasized by assigning them larger back-propagated gradients during training. Reusing the semantic labels and predicted probabilities defined in Equation (1), the weight for the i-th pixel is defined as
W i = 1 + α C k = 1 C ( y i k ( 1 p i k ) γ + ( 1 y i k ) ( p i k ) γ ) ,
where α and γ are hyperparameters.

3.5.3. Cross-IoU Loss

The Intersection over Union (IoU) [53] metric is commonly utilized to evaluate the pixel-level prediction performance in terms of tasks of object detection and segmentation, defined as
I o U = T P F P + T P + F N ,
where T P , F P , and F N denote the true positive, the false positive, and the false negative pixel counts, respectively. Considering the fact that the lane and lane marking are closely connected but do not overlap with each other, we also employ the IoU loss to suppress the overlapping between them and thus obtain the Cross-IoU Loss, dubbed as CIoU and computed by
L C I o U = R l R ^ m R l R ^ m + R m R ^ l R m R ^ l ,
where R l and R m denote the pixel area of predicted lane and lane marking, respectively. The superscript ^ refers to the corresponding groundtruth.

3.5.4. Total Learning Loss

To train the proposed multi-task architecture, the total learning loss is defined as
L t = w m i N W i , m L i , m + w l i N W i , l L i , l + w C I o U L C I o U + w a u x , l L a u x , l + w a u x , m L a u x , m ,
where L a u x , l and L a u x , m denote the auxiliary losses (in the Cross Entropy form) for the lane and lane marking segmentation branches, respectively. L i , l and L i , m denote the pixel-wise Dice loss for the outputs of the lane and lane marking prediction heads, and W i , l and W i , m are the proposed Focal-style loss weightings. Other trade-off factors w l , w m , w C I o U , w a u x , l and w a u s , m are hyperparameters.

4. Experiments and Evaluation

4.1. Implementation Details

4.1.1. Datasets

To train and validate our proposed architecture, it requires annotations for both lane markings and drivable lane areas in the same dataset. Among the public benchmarks, the Berkeley DeepDrive (BDD100K) dataset [16] is the only one that can provide such kinds of annotations, which are also with a high variety in traffic scenes including illumination change and complex weathers. Thus, we choose it for our experiments as it supports the multi-task learning of our approach. The BDD100K consists of 100 K images in a size of 1280 × 720 pixels, in which 70 K images are used as a training set, 10 K images are used as a validation set, and 20 K images are used as a test set. Since the test labels are not publicly accessible and the evaluation of lane and lane marking detection is also unavailable on the server, we opt to evaluate the proposed method on the validation set.

4.1.2. Metric

As mentioned above, we define both lane and lane marking detection as semantic segmentation tasks. Following the common protocol [38], we evaluate the accuracy of segmented drivable area and background using the mean IoU (mIoU) metric for lane detection while the lane marking segmentation is only evaluated by the IoU metric, to exclude the influence of background pixels, which occupy over 90% of the image. Specifically, given the predicted mask M p , i and the groundtruth mask M g , i of image i, the true positives count T P , true negative count T N , false negative count F N and false positive count F P are computed as
T P = i | | M p , i · M g , i | | 0 ,
T N = i | | ( 1 M p , i ) · ( 1 M g , i ) | | 0 ,
F N = i | | ( 1 M p , i ) · M g , i | | 0 ,
F P = i | | M p , i · ( 1 M g , i ) | | 0 .
The IoU metric can thus be calculated by referring to Equation (3). Additionally, we also evaluate the lane marking segmentation by the Pixel Accuracy (PA) metric, which can be calculated as
P A = T P + T N T P + T N + F P + F N .

4.1.3. Training

For a fair comparison, the images of the BDD100K dataset are resized into a resolution of 640 × 384 pixels as the input of our architecture, which is the same size as used in other state-of-the-art approaches. In the experiment, the data augmentation tricks including the random rotation, random cutout, photometric distortions, and random horizontal flip are also adopted in the training. We use the Adam optimizer with a learning rate initialized to 2 × 10 4 and a weight decay set to 1 × 10 5 . Other parameters follow the default settings. For the total loss, we set the weights w l , w m , w i o u , w a u x , l and w a u x , m to 1.0, 0.1, 0.1, 0.01 and 0.01, respectively. The hyperparameters of Focal-style loss weighting are empirically set to α = 0.5 and γ = 1 . The model is trained with a batch-size of 8 in 150 epochs. All the experiments are carried out on a computer platform with a CPU of 2.5 GHz and a GPU of NVIDIA RTX 3090. Furthermore, for comparison with state-of-the-art methods, the ConvNeXt-Small [48] is chosen as our backbone.
Here, we evaluate our proposed architecture on the lane and lane marking detection by comparing it with other state-of-the-art approaches. For evaluation on the lane marking detection, we choose the ENet [54], SCNN [9], and ENet-SAD [6] as comparison sets. For comparison on the lane segmentation, we choose the ERFNet [55], MultiNet [41], DLT-Net [42], and PSPNet [34]. Additionally, we compare our architecture with the efficient feature aggregator RESA [10] and recently proposed multi-task approach YOLOP [49] in both evaluations (also visualized in Figure 5). Note that the RESA is originally designed for lane marking segmentation while we add an additional output head so that it is capable to segment the lane area.

4.2. Comparison with State-of-the-Arts

4.2.1. Lane Marking Detection

The lane marking labels in the BDD100K dataset are annotated with sets of points, which is troublesome to directly use them. Thus, we follow the work [49] to substitute the two-line annotation with one center line and dilate the line width in the training to 8 pixels and keep it in the test as 2 pixels. The lane marking detection results are listed in Table 1.
As depicted in the table, our method outperforms ENet, SCNN, ENet-SAD, RESA, YOLOP by 47.49%, 45.82%, 45.05%, 20.25%, 11.11% in terms of accuracy and 17.89%, 16.69%, 16.51%, 15.82%, 6.33% in terms of IoU. The ENet adopts an early downsampling strategy for the encoding features to obtain a very fast processing of 100 frames per second (fps), yet leading to decreased detection accuracy. Based the shape characteristics of lane marking, the SCNN replaces the ordinary layer-by-layer convolutions with the slice-by-slice convolutions to pass information between rows and columns, resulting in a slow forward computation. The ENet-SAD introduces the paradigm of knowledge distillation on the basis of ENet, through which the intermediate encoding features are enhanced. However, limited to the architecture of ENet, the accuracy of ENet-SAD is not much improved. The RESA adopts an operation named Recurrent Feature-Shift to pass information between rows and columns, which is more accurate and efficient compared to slice-by-slice convolutions in SCNN. The YOLOP sets up multi-task prediction heads based on the pretraining of YOLOv5, which achieves better detection accuracy and real-time performance. However, it does not consider the geometric constraints between lane and lane markings and thus its prediction is still less optimal. Examples of the lane marking detection are visualized in Figure 5a. Note that we only visualize the results of those approaches whose codes are publicly available. As shown in Figure 5b, the RESA achieves a poor performance in the night driving scenario due to the limited visual information. Despite integrated with multi-tasks, there are still many discontinuities in the lane marking detection results of YOLOP, implying an insufficient learning about the interactive information between tasks. In comparison, our method predicts the lane markings more accurately.

4.2.2. Lane Detection

For utility and briefness, the lane segmentation labels {direct, alternative} in the BDD100K dataset are merged into the label {lane}. Here, the lane area segmentation task is simplified to distinguish between the lane areas and the background areas in the image, resulting in two class labels. The lane detection results of compared approaches are listed in Table 2. From the reported results, we can observe that our method outperforms the ERFNet, MultiNet, DLT-Net, PSPNet, RESA and YOLOP by 23.02%, 20.12%, 19.62%, 2.12%, 2.42% and 0.22% in terms of mIoU. The ERFNet runs faster than the MultiNet, DLT-Net and PSPNet due to its residual connections and factorized convolutions for semantic segmentation. But it has only a simple encoder and a decoder, yielding an ordinary detection performance. The MultiNet handles the tasks of vehicle detection, scene classification, and lane detection at the same time and outperforms ERFNet by 2.9%. The DLT-Net adopts context tensors to share the information between subtask decoders, which results in an improved lane detection accuracy than MultiNet. The PSPNet consists of one encoder, one decoder, and one Pyramid Pooling module which incorporates multi-scale information and improves the lane detection performance. The RESA also achieves good results on the lane detection task by aggregating spatial information of intermediate features. Due to the advantages of the multi-task form and the pretrained backbone of YOLOv5, the YOLOP surpasses previous methods on the lane detection task. Several lane detection results are visualized in Figure 5a. Interestingly, although our method has a similar mIoU to YOLOP in the lane detection task, the visualization shows that our proposed method performs a more accurate and robust detection in several driving scenarios.
Moreover, the processing efficiency of our architecture is 26 fps, which is still appropriate for real-time applications.

4.3. Exploration on Interaction Learning Modules

Here, we explore the effectiveness of interaction-learning-related modules utilized in our architecture such as the DFF module, the Cross-Context module and the Focal-style loss weighting. For a qualitative impression, we also adopt the Grad-CAM [56] tool to visualize features extracted from the intermediate layer of those modules.

4.3.1. DFF Module

In an effort to verify the DFF module, we choose the network layer before the DFF and the intermediate convolution layer that generates the offsets for the deformable convolution to output the Grad-CAM activation map. The visualized activation maps “before” and “within” the DFF module are shown in Figure 6. Obviously, the model pays homogeneous attention to each pixel in the image before applying the DFF module, while with the DFF processing, the model focuses mainly on the features located on the lane markings. This fact verifies the capability of the DFF module in guiding the model learning discriminative features with the interaction information.

4.3.2. Cross-Context Module

For the lane and lane marking prediction heads, we select the input layer of decoder and the output layer of Cross-Context module, respectively, to output the Grad-CAM activation map, as visualized in Figure 7a–d. It is worth noting that Figure 7c just depicts the feature map to be added to the input of lane marking prediction head while Figure 7d shows the feature map added to the other one. Thus, it is obviously that the Cross-Context Module can enhance the input features of both lane and lane marking prediction heads by reciprocal information transferring.

4.3.3. Focal-Style Loss Weighting

The weight map generated by the Focal-style loss weighting is shown in Figure 8. From the visualization, we can confirm that the weight map sets larger weights on the lane markings and lane boundaries. Under this circumstance, the model pays more attention to the corresponding pixels during training, which helps the inference of lane and lane markings and improves the detection accuracy, thus verifying the effectiveness of Focal-style loss weighting.
For the quantitative evaluation of all interaction learning modules, we provide an ablation study. For the baseline, we design a model with a single-branch backbone (i.e., 1× scale of input image), a single neck, and two separate task heads. For training efficiency, we adopt the ResNet-18 as backbone. Subsequently, we gradually integrate the encoder with multi-scale inputs, the DFF Module, the Cross-Context module in the decoder, the Cross-IoU loss, and the Focal-style loss weighting into the baseline. We report the performance of above integrated versions in Table 3. As can be seen, the detection accuracy of both lane and lane marking gradually increases with more modules integrated. The performance gain brought by the Cross-Context module, the CIoU loss and the Focal-style loss weighting are relative large, which is over 4% on the lane marking IoU metric and over 2% on the lane mIoU metric, further demonstrating the great advantages by interaction learning.

4.4. Further Exploration on Architecture Design

In an effort to further explore the impact of different backbone paradigms on the detection performance of our architecture, we select a group of backbone networks for comparison. Concretely, we set up the ResNet18, ResNet34, ConvNeXt-tiny, and ConvNeXt-small as the backbone and other modules remain unchanged. It is worth noting that the 1/32 downsampling layers in the naive ResNet and ConvNeXt discard too much information, which degrades the performance of following modules. Thus, for the ResNet, we replace the ordinary convolution in the downsampling layer of the C4 and C5 stage with the atrous convolution, which can expand the receptive field of the network while maintaining a higher spatial resolution. The ConvNeXt can thus be configured referring to Section 3.2.2. The results are listed in Table 4. As to be seen, since the ConvNeXt has a well-designed architecture with more network parameters, its accuracy on both lane and lane marking detection increases compared to the ResNet while the ResNet shows a lower computation amount and a faster real-time performance due to its fewer parameters.

5. Conclusions

In this paper, we put forward a novel multi-task framework for vision-based lane and lane marking detection on the road by introducing the interaction learning of their tight spatial correlation, which is persistently neglected in existing researches. The efficient learning of the interaction between the lane and lane marking information is achieved by three novel modules, i.e., the Deformable Feature Fusion Module for feature encoding, the Cross-Context Module for information decoding, the Cross-IoU loss and the Focal-style loss weighting for robust training. The effectiveness of each module has been validated based on throughout analysis of comprehensive experiments on the challenging BDD100K dataset. Therefore, the neglected spatial correlation between lane and lane marking in previous works has been proven essential to the improvement of detection robustness. Our proposed architecture also surpasses state-of-the-art approaches on both lane and lane detection tasks at a processing speed of 26 fps, which is promising for applications with real-time requirements. In the future work, we will introduce the interactive attention learning into the transformer architecture and further reduce the computation cost of attention estimation. We will also extend our method to various weather conditions and validate it with more new datasets, since the number of public datasets currently supporting both lane and lane marking detection tasks is still limited. Moreover, we will investigate the joint learning of both dynamic elements (e.g., vehicles, pedestrians) and static elements (e.g., lanes, lane markings) to improve the detection accuracy.

Author Contributions

Conceptualization, W.T.; methodology, W.T. and X.Y.; software, X.Y.; validation, X.Y.; formal analysis, W.T. and X.Y.; data curation, X.Y.; writing—original draft preparation, W.T. and X.Y.; writing—review and editing, W.T., X.Y. and H.H.; visualization, X.Y.; supervision, W.T. and H.H.; project administration, W.T.; funding acquisition, W.T. All authors have read and agreed to the published version of the manuscript.

Funding

Project supported by the National Natural Science Foundation of China (No. 52002285), the Shanghai Science and Technology Commission (No. 21ZR1467400), the original research project of Tongji University (No. 22120220593), the National Key R&D Program of China (No. 2021YFB2501104), and the Natural Science Foundation of Chongqing (No. 2023NSCQ-MSX4511).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets generated and analysed during the current study are available in the [BDD100K] repository. [https://bdd-data.berkeley.edu/, accessed on 1 December 2022].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. HERE. HERE HD Live Map: The Most Intelligent Sensor for Autonomous Driving. 2022. Available online: https://www.here.com/platform/automotive-services/hd-maps (accessed on 1 December 2022).
  2. TomTom. HD Maps—Highly Accurate Border-to-Border Model of the Road. 2022. Available online: https://www.tomtom.com/products/hd-map (accessed on 1 December 2022).
  3. Homayounfar, N.; Ma, W.C.; Liang, J.; Wu, X.; Fan, J.; Urtasun, R. DAGmapper: Learning to map by discovering lane topology. In Proceedings of the IEEE/CVF International Conference on Computer Vision (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 2911–2920. [Google Scholar]
  4. Chiu, K.Y.; Lin, S.F. Lane detection using color-based segmentation. In Proceedings of the IEEE Intelligent Vehicles Symposium, Las Vegas, NV, USA, 6–8 June 2005; pp. 706–711. [Google Scholar]
  5. Satzoda, R.K.; Sathyanarayana, S.; Srikanthan, T.; Sathyanarayana, S. Hierarchical additive Hough transform for lane detection. IEEE Embed. Syst. Lett. 2010, 2, 23–26. [Google Scholar] [CrossRef]
  6. Hou, Y.; Ma, Z.; Liu, C.; Loy, C.C. Learning lightweight lane detection cnns by self attention distillation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1013–1021. [Google Scholar]
  7. Chen, Z.; Liu, Q.; Lian, C. Pointlanenet: Efficient end-to-end cnns for accurate real-time lane detection. In Proceedings of the IEEE Intelligent Vehicles Symposium (IV), Paris, France, 9–12 June 2019; pp. 2563–2568. [Google Scholar]
  8. Tian, W.; Ren, X.; Yu, X.; Wu, M.; Zhao, W.; Li, Q. Vision-based mapping of lane semantics and topology for intelligent vehicles. Int. J. Appl. Earth Obs. Geoinf. 2022, 111, 102851. [Google Scholar] [CrossRef]
  9. Pan, X.; Shi, J.; Luo, P.; Wang, X.; Tang, X. Spatial as deep: Spatial cnn for traffic scene understanding. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
  10. Zheng, T.; Fang, H.; Zhang, Y.; Tang, W.; Yang, Z.; Liu, H.; Cai, D. Resa: Recurrent feature-shift aggregator for lane detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 2–9 February 2021; Volume 35, pp. 3547–3554. [Google Scholar]
  11. Li, X.; Li, J.; Hu, X.; Yang, J. Line-cnn: End-to-end traffic line detection with line proposal unit. IEEE Trans. Intell. Transp. Syst. 2019, 21, 248–258. [Google Scholar] [CrossRef]
  12. Tabelini, L.; Berriel, R.; Paixao, T.M.; Badue, C.; De Souza, A.F.; Oliveira-Santos, T. Keep your eyes on the lane: Real-time attention-guided lane detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 294–302. [Google Scholar]
  13. Tabelini, L.; Berriel, R.; Paixao, T.M.; Badue, C.; De Souza, A.F.; Oliveira-Santos, T. Polylanenet: Lane estimation via deep polynomial regression. In Proceedings of the International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2021; pp. 6150–6156. [Google Scholar]
  14. Liu, R.; Yuan, Z.; Liu, T.; Xiong, Z. End-to-end lane shape prediction with transformers. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Online, 5–9 January 2021; pp. 3694–3702. [Google Scholar]
  15. Feng, Z.; Guo, S.; Tan, X.; Xu, K.; Wang, M.; Ma, L. Rethinking Efficient Lane Detection via Curve Modeling. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 17062–17070. [Google Scholar]
  16. Yu, F.; Chen, H.; Wang, X.; Xian, W.; Chen, Y.; Liu, F.; Madhavan, V.; Darrell, T. Bdd100k: A diverse driving dataset for heterogeneous multitask learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2636–2645. [Google Scholar]
  17. TuSimple. 2022. Available online: http://benchmark.tusimple.ai/ (accessed on 1 December 2022).
  18. Loose, H.; Franke, U.; Stiller, C. Kalman particle filter for lane recognition on rural roads. In Proceedings of the IEEE Intelligent Vehicles Symposium, Xi’an, China, 3–5 June 2009; pp. 60–65. [Google Scholar]
  19. Teng, Z.; Kim, J.H.; Kang, D.J. Real-time lane detection by using multiple cues. In Proceedings of the International Conference on Control Automation and Systems (ICCAS 2010), Suwon, Republic of Korea, 27–30 October 2010; pp. 2334–2337. [Google Scholar]
  20. Aly, M. Real time detection of lane markers in urban streets. In Proceedings of the IEEE Intelligent Vehicles Symposium, Eindhoven, The Netherlands, 4–6 June 2008; pp. 7–12. [Google Scholar]
  21. Assidiq, A.A.; Khalifa, O.O.; Islam, M.R.; Khan, S. Real time lane detection for autonomous vehicles. In Proceedings of the International Conference on Computer and Communication Engineering, Kuala Lumpur, Malaysia, 13–15 May 2008; pp. 82–88. [Google Scholar]
  22. Azimi, S.M.; Fischer, P.; Körner, M.; Reinartz, P. Aerial LaneNet: Lane-marking semantic segmentation in aerial imagery using wavelet-enhanced cost-sensitive symmetric fully convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2018, 57, 2920–2938. [Google Scholar] [CrossRef] [Green Version]
  23. Guan, H.; Lei, X.; Yu, Y.; Zhao, H.; Peng, D.; Junior, J.M.; Li, J. Road marking extraction in UAV imagery using attentive capsule feature pyramid network. Int. J. Appl. Earth Obs. Geoinf. 2022, 107, 102677. [Google Scholar] [CrossRef]
  24. Xu, H.; Wang, S.; Cai, X.; Zhang, W.; Liang, X.; Li, Z. Curvelane-nas: Unifying lane-sensitive architecture search and adaptive point blending. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 689–704. [Google Scholar]
  25. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; Volume 28. [Google Scholar]
  26. Qin, Z.; Wang, H.; Li, X. Ultra fast structure-aware deep lane detection. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 276–291. [Google Scholar]
  27. Yoo, S.; Lee, H.S.; Myeong, H.; Yun, S.; Park, H.; Cho, J.; Kim, D.H. End-to-end lane marker detection via row-wise classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 1006–1007. [Google Scholar]
  28. Liu, L.; Chen, X.; Zhu, S.; Tan, P. Condlanenet: A top-to-down lane detection framework based on conditional convolution. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 3773–3782. [Google Scholar]
  29. Wang, Q.; Wang, L.; Chi, Y.; Shen, T.; Song, J.; Gao, J.; Shen, S. Dynamic Data Augmentation Based on Imitating Real Scene for Lane Line Detection. Remote Sens. 2023, 15, 1212. [Google Scholar] [CrossRef]
  30. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  31. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  32. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  33. Yu, F.; Koltun, V. Multi-scale context aggregation by dilated convolutions. arXiv 2015, arXiv:1511.07122. [Google Scholar]
  34. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid scene parsing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2881–2890. [Google Scholar]
  35. He, X.; Zhou, Y.; Zhao, J.; Zhang, D.; Yao, R.; Xue, Y. Swin Transformer Embedding UNet for Remote Sensing Image Semantic Segmentation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  36. Xie, W.; Liu, P.X.; Zheng, M. Moving object segmentation and detection for robust RGBD-SLAM in dynamic environments. IEEE Trans. Instrum. Meas. 2020, 70, 1–8. [Google Scholar] [CrossRef]
  37. Meyer, A.; Salscheider, N.O.; Orzechowski, P.F.; Stiller, C. Deep semantic lane segmentation for mapless driving. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 869–875. [Google Scholar]
  38. Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T.; Enzweiler, M.; Benenson, R.; Franke, U.; Roth, S.; Schiele, B. The cityscapes dataset for semantic urban scene understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3213–3223. [Google Scholar]
  39. Sun, T.; Di, Z.; Che, P.; Liu, C.; Wang, Y. Leveraging crowdsourced GPS data for road extraction from aerial imagery. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7509–7518. [Google Scholar]
  40. Fontanelli, D.; Moro, F.; Rizano, T.; Palopoli, L. Vision-based robust path reconstruction for robot control. IEEE Trans. Instrum. Meas. 2013, 63, 826–837. [Google Scholar] [CrossRef]
  41. Teichmann, M.; Weber, M.; Zoellner, M.; Cipolla, R.; Urtasun, R. Multinet: Real-time joint semantic reasoning for autonomous driving. In Proceedings of the IEEE Intelligent Vehicles Symposium (IV), Suzhou, China, 26–30 June 2018; pp. 1013–1020. [Google Scholar]
  42. Qian, Y.; Dolan, J.M.; Yang, M. DLT-Net: Joint detection of drivable areas, lane lines, and traffic objects. IEEE Trans. Intell. Transp. Syst. 2019, 21, 4670–4679. [Google Scholar] [CrossRef]
  43. Chen, Z.; Chen, Z. Rbnet: A deep neural network for unified road and road boundary detection. In Proceedings of the International Conference on Neural Information Processing, Long Beach, CA, USA, 4–9 December 2017; pp. 677–687. [Google Scholar]
  44. Zhang, J.; Xu, Y.; Ni, B.; Duan, Z. Geometric constrained joint lane segmentation and lane boundary detection. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 486–502. [Google Scholar]
  45. Liu, Y.; Yao, J.; Lu, X.; Xia, M.; Wang, X.; Liu, Y. RoadNet: Learning to comprehensively analyze road networks in complex urban scenes from high-resolution remotely sensed images. IEEE Trans. Geosci. Remote Sens. 2018, 57, 2043–2056. [Google Scholar] [CrossRef]
  46. Meng, Z.; Xia, X.; Xu, R.; Liu, W.; Ma, J. HYDRO-3D: Hybrid Object Detection and Tracking for Cooperative Perception Using 3D LiDAR. IEEE Trans. Intell. Veh. 2023, 1–13. [Google Scholar] [CrossRef]
  47. Xia, X.; Meng, Z.; Han, X.; Li, H.; Tsukiji, T.; Xu, R.; Zheng, Z.; Ma, J. An automated driving systems data acquisition and analytics platform. Transp. Res. Part C Emerg. Technol. 2023, 151, 104120. [Google Scholar] [CrossRef]
  48. Liu, Z.; Mao, H.; Wu, C.Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A convnet for the 2020s. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11976–11986. [Google Scholar]
  49. Wu, D.; Liao, M.; Zhang, W.; Wang, X. Yolop: You only look once for panoptic driving perception. arXiv 2021, arXiv:2108.11250. [Google Scholar]
  50. Dai, J.; Qi, H.; Xiong, Y.; Li, Y.; Zhang, G.; Hu, H.; Wei, Y. Deformable convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 764–773. [Google Scholar]
  51. Milletari, F.; Navab, N.; Ahmadi, S.A. V-net: Fully convolutional neural networks for volumetric medical image segmentation. In Proceedings of the International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar]
  52. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  53. Rahman, M.A.; Wang, Y. Optimizing intersection-over-union in deep neural networks for image segmentation. In Proceedings of the International Symposium on Visual Computing, Las Vegas, NV, USA, 12–14 December 2016; pp. 234–244. [Google Scholar]
  54. Paszke, A.; Chaurasia, A.; Kim, S.; Culurciello, E. Enet: A deep neural network architecture for real-time semantic segmentation. arXiv 2016, arXiv:1606.02147. [Google Scholar]
  55. Romera, E.; Alvarez, J.M.; Bergasa, L.M.; Arroyo, R. Erfnet: Efficient residual factorized convnet for real-time semantic segmentation. IEEE Trans. Intell. Transp. Syst. 2017, 19, 263–272. [Google Scholar] [CrossRef]
  56. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
Figure 1. Overview of the proposed interactive attention learning model on detection of lanes and lane markings. Our model is decomposed into two branches, each consisting of a backbone, a neck and a prediction head with information interactively learned by the DFF, Cross-Context and CIoU loss module.
Figure 1. Overview of the proposed interactive attention learning model on detection of lanes and lane markings. Our model is decomposed into two branches, each consisting of a backbone, a neck and a prediction head with information interactively learned by the DFF, Cross-Context and CIoU loss module.
Sensors 23 06545 g001
Figure 2. The structure of the backbone, which is configured based on the first three stages of the ConvNeXt [48] and one additional dilated block. Only the ConvNeXt-tiny is illustrated here.
Figure 2. The structure of the backbone, which is configured based on the first three stages of the ConvNeXt [48] and one additional dilated block. Only the ConvNeXt-tiny is illustrated here.
Sensors 23 06545 g002
Figure 3. Architecture of the Deformable Feature Fusion Module.
Figure 3. Architecture of the Deformable Feature Fusion Module.
Sensors 23 06545 g003
Figure 4. Architecture of the Cross-Context Module for decoding.
Figure 4. Architecture of the Cross-Context Module for decoding.
Sensors 23 06545 g004
Figure 5. Visualized lane and lane marking detection by compared methods. “GT” is short for the groundtruth. (a) Lane marking detection results; (b) lane detection results.
Figure 5. Visualized lane and lane marking detection by compared methods. “GT” is short for the groundtruth. (a) Lane marking detection results; (b) lane detection results.
Sensors 23 06545 g005
Figure 6. Activation map visualization for the DFF module at the layer before (top) and at the intermediate convolution layer within this module (bottom).
Figure 6. Activation map visualization for the DFF module at the layer before (top) and at the intermediate convolution layer within this module (bottom).
Sensors 23 06545 g006
Figure 7. Activation map visualization for the Cross-Context module: (a) at the input layer of the lane marking prediction head; (b) at the input layer of the lane prediction head; (c) at the output layer of the Cross-Context module, which is added to the lane marking prediction head; (d) at the output layer of the Cross-Context module, which is added to the lane prediction head.
Figure 7. Activation map visualization for the Cross-Context module: (a) at the input layer of the lane marking prediction head; (b) at the input layer of the lane prediction head; (c) at the output layer of the Cross-Context module, which is added to the lane marking prediction head; (d) at the output layer of the Cross-Context module, which is added to the lane prediction head.
Sensors 23 06545 g007
Figure 8. Visualization of the Focal-style loss weighting in heatmap for lane marking detection (top) and lane area segmentation (bottom). The brighter the color of heat map is, the greater the loss weight will be, and vice versa.
Figure 8. Visualization of the Focal-style loss weighting in heatmap for lane marking detection (top) and lane area segmentation (bottom). The brighter the color of heat map is, the greater the loss weight will be, and vice versa.
Sensors 23 06545 g008
Table 1. Comparison results on lane marking detection.
Table 1. Comparison results on lane marking detection.
NetworkAccuracy (%) ↑IoU (%) ↑Speed (fps) ↑
ENet [54]34.1214.64100
SCNN [9]35.7915.8419.8
ENet-SAD [6]36.5616.0250.6
RESA [10]61.2616.7147.4
YOLOP [49]70.5026.2041
Ours81.6132.5326
Table 2. Comparison results on lane detection.
Table 2. Comparison results on lane detection.
NetworkmIoU (%) ↑Speed (fps) ↑
ERFNet [55]68.722.8
MultiNet [41]71.68.6
DLT-Net [42]72.19.3
PSPNet [34]89.611.1
RESA [10]89.347.4
YOLOP [49]91.541
Ours91.7226
Table 3. Performance exploration of proposed modules on the BDD100K dataset. The baseline is the model with a single-scale input encoder and two separate task decoders.
Table 3. Performance exploration of proposed modules on the BDD100K dataset. The baseline is the model with a single-scale input encoder and two separate task decoders.
BaselineMulti-ScaleDFFCross-ContextCIoUFocal-StyleLane Marking IoU (%) ↑Lane mIoU (%) ↑
19.1487.40
21.49 (+2.35)87.47 (+0.07)
21.83 (+2.69)88.85 (+1.45)
23.2 (+4.06)89.47 (+2.07)
23.49 (+4.35)89.54 (+2.14)
23.96 (+4.82)90.07 (+2.67)
Table 4. Comparison of model performance with different backbone paradigms on the BDD100K dataset.
Table 4. Comparison of model performance with different backbone paradigms on the BDD100K dataset.
BackboneLane Marking IoU (%) ↑Lane mIoU (%) ↑Params (M) ↓FLOPs (G) ↓Speed (fps) ↑
ResNet-1830.3990.5417.0589.8358
ResNet-3430.4690.6127.16139.4640
ConvNeXt-tiny31.4891.2918.3596.5239
ConvNeXt-small32.5391.7239.97200.0726
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tian, W.; Yu, X.; Hu, H. Interactive Attention Learning on Detection of Lane and Lane Marking on the Road by Monocular Camera Image. Sensors 2023, 23, 6545. https://doi.org/10.3390/s23146545

AMA Style

Tian W, Yu X, Hu H. Interactive Attention Learning on Detection of Lane and Lane Marking on the Road by Monocular Camera Image. Sensors. 2023; 23(14):6545. https://doi.org/10.3390/s23146545

Chicago/Turabian Style

Tian, Wei, Xianwang Yu, and Haohao Hu. 2023. "Interactive Attention Learning on Detection of Lane and Lane Marking on the Road by Monocular Camera Image" Sensors 23, no. 14: 6545. https://doi.org/10.3390/s23146545

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop