Next Article in Journal
The Synergy between a Humanoid Robot and Whisper: Bridging a Gap in Education
Next Article in Special Issue
Exploring Deep Learning for Adaptive Energy Detection Threshold Determination: A Multistage Approach
Previous Article in Journal
Activeness: A Novel Neural Coding Scheme Integrating the Spike Rate and Temporal Information in the Spiking Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MultiNet-GS: Structured Road Perception Model Based on Multi-Task Convolutional Neural Network

1
School of Information Engineering, Chang’an University, Xi’an 710064, China
2
School of Computer Science and Technology, Xidian University, Xi’an 710126, China
3
National Engineering Laboratory of Robot Visual Perception and Control Technology, Hunan University, Changsha 410082, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(19), 3994; https://doi.org/10.3390/electronics12193994
Submission received: 6 August 2023 / Revised: 17 September 2023 / Accepted: 20 September 2023 / Published: 22 September 2023
(This article belongs to the Special Issue Machine Learning Techniques in Autonomous Driving)

Abstract

:
In order to address the issue of environmental perception in autonomous driving on structured roads, we propose MultiNet-GS, a convolutional neural network model based on an encoder–decoder architecture that tackles multiple tasks simultaneously. We use the main structure of the latest object detection model, the YOLOv8 model, as the encoder structure of our model. We introduce a new dynamic sparse attention mechanism, BiFormer, in the feature extraction part of the model to achieve more flexible computing resource allocation, which can significantly improve the computational efficiency and occupy a small computational overhead. We introduce a lightweight convolution, GSConv, in the feature fusion part of the network, which is used to build the neck part into a new slim-neck structure so as to reduce the computational complexity and inference time of the detector. We also add an additional detector for tiny objects to the conventional three-head detector structure. Finally, we introduce a lane detection method based on guide lines in the lane detection part, which can aggregate the lane feature information into multiple key points, obtain the lane heat map response through conditional convolution, and then describe the lane line through the adaptive decoder, which effectively makes up for the shortcomings of the traditional lane detection method. Our comparative experiments on the BDD100K dataset on the embedded platform NVIDIA Jetson TX2 show that compared with SOTA(YOLOPv2), the [email protected] of the model in traffic object detection reaches 82.1%, which is increased by 2.7%. The accuracy of the model in drivable area detection reaches 93.2%, which is increased by 0.5%. The accuracy of the model in lane detection reaches 85.7%, which is increased by 4.3%. The Params and FLOPs of the model reach 47.5 M and 117.5, which are reduced by 6.6 M and 8.3, respectively. The model achieves 72 FPS, which is increased by 5. Our MultiNet-GS model has the highest detection accuracy among the current mainstream models while maintaining a good detection speed and has certain superiority.

1. Introduction

The fields of computer vision and deep learning have witnessed significant advancements in recent years, and camera-based techniques, such as object detection and semantic segmentation, are widely used in the field of autonomous driving due to their low cost. Environment perception under structured road (as shown in Figure 1) is a key link in autonomous driving technology. The environment perception system extracts visual features from the images captured by the camera, understands the scene information, detects the position of obstacles and lanes, determines whether the road is drivable, and helps the decision-making system control the movement of the vehicle, which is crucial for planning the vehicle’s driving route. The environmental perception system makes accurate and timely decisions in the process of autonomous driving, which is the key to ensuring driving safety. Therefore, ensuring the accuracy and real-time performance of the system are the two most critical and challenging requirements.
Object detection and semantic segmentation are two topics that have been widely studied in the field of computer vision. For object detection, a series of accurate and efficient methods have been proposed, such as CenterNet [1], RetinaNet [2], DETR [3], R-CNN series [4,5,6], SSD [7], YOLO series [8,9,10], DINO [11], and EVA [12], which have achieved good detection results on public datasets. Classic semantic segmentation networks, such as UNet [13], ENet [14], SegNet [15], PSPNet [16], ERFNet [17], PSANet [18], EncNet [19], CFNet [20], and EANet [21] are usually used to perform the segmentation task of road drivable areas. For lane line detection tasks, it is necessary to obtain more refined semantic features, combine high-level features and low-level features, and use context information to detect channels. Networks such as SCNN [22], ENet-SAD [23], LaneNet [24], LaneATT [25], and LaneAF [26] are designed to specifically perform this task.
In order to be able to run multiple detection and segmentation tasks at the same time, considering that the information of many tasks in the traffic scene is interrelated, a multi-task learning network model [27,28,29,30,31] was proposed, which adopted an encoder–decoder structure. Feature extraction and feature fusion are performed in the encoder structure, in which the information of multiple tasks can be shared with each other. The decoder structure can run multiple detection tasks at the same time, and the detection results are output without mutual interference between multiple tasks. Although these methods have achieved excellent performance, there are still many problems in these traditional multi-task network models in the process of environment perception of autonomous driving. First, for traffic objects with small size, the detection accuracy of the object detection part of the model is low and is prone to missed detection and false detection. Second, the segmentation accuracy of drivable area and lane line is low when the image is blurred by extreme weather, such as rain, snow, and fog, or the road is occluded. Third, the segmentation of lane lines is prone to discontinuity in the case of large changes in lane slope and sharp lane turns. Fourth, the structure of the existing multi-task network model is complex, the number of parameters and calculation of the model are generally large, and the performance is poor on embedded devices with limited computing resources, which does not meet the real-time requirements in the process of autonomous driving.
To this end, we propose a lightweight and efficient multi-task network model, MultiNet-GS, to solve the multi-task problem in autonomous driving environment perception, that is, traffic object detection, drivable area detection, and lane line detection in structured road scenes, and improve the shortcomings of traditional multi-task network models. Our contributions are as follows:
  • We introduce a new dynamic sparse attention mechanism, BiFormer [32], in the feature extraction part of the model to achieve more flexible computing resource allocation, which can significantly improve the computational efficiency and occupy a small computational overhead. The introduction of the attention mechanism makes the model pay more attention to the region of interest and improves the accuracy of object detection, drivable area segmentation, and lane line segmentation.
  • We introduce a lightweight convolution, GSConv [33], in the feature fusion part of the network, which is used to build the neck part into a new slim-neck structure, so as to reduce the computational complexity and inference time of the detector. The introduction of slim-neck structure greatly reduces the number of parameters and calculation of the model so that the model can adapt to embedded devices with limited computing power and meet the real-time requirements in autonomous driving.
  • We add an additional detector for tiny objects to the conventional three-head detector structure. This four-head detector structure can effectively alleviate the negative impact of drastic changes in the size of traffic objects and greatly improve the detection performance of tiny objects.
  • In the lane detection part, we introduce a lane detection method based on guide lines, which can aggregate the lane feature information into multiple key points and then obtain the lane heat map response through conditional convolution. Finally, the lane line is described by the adaptive decoder, which effectively makes up for the shortcomings of the traditional lane detection method so that the model can accurately and continuously segment the lane line in the case of drastic changes in lane slope and sharp lane turns.
The rest of this paper is organized as follows: In Section 2, we introduce the current popular single-task network models and multi-task network models, as well as the current popular attention mechanism modules, and introduce their advantages and disadvantages in detail to lead to our work. In Section 3, we elaborate the structure of our proposed model and the adopted method. In Section 4, we design detailed ablation experiments and contrast experiments for our proposed improvements to observe the effect of our proposed improvements. Finally, we show the superiority of our proposed network model over the current mainstream network models and the content of future work.

2. Related Work

In this section, we will introduce the current popular research methods and their shortcomings as well as the current popular multi-task learning models in detail for the three tasks in the structured road environment perception task based on deep learning.

2.1. Traffic Object Detection

A variety of object detectors with excellent performance have been proposed in the field of deep learning, which can be roughly divided into two types: one-stage methods and two-stage methods.
The two-stage method represented by R-CNN series [4,5,6] first generates the region (a pre-selected box that may contain the object to be detected), and then classifies the samples by convolutional neural network. This method has high detection accuracy, but it will have a large time overhead and cannot meet the real-time requirements in the process of automatic driving. The one-stage method represented by YOLO series [8,9,10] will directly extract network features for the purpose of predicting object classification and location, effectively balancing detection accuracy and detection time. YOLOv3 [8] introduces the residual model Darknet-53 for the first time and adopts the FPN architecture to realize multi-scale detection. Based on the original YOLO object detection architecture, YOLOv4 [9] adopts many optimization strategies to further improve the network performance. YOLOv5 introduces functions such as adaptive anchor and adaptive image scaling to make the training of the network more efficient. YOLOv7 [10] incorporates several trainable bags-of-freebies to significantly enhance detection accuracy of the real-time detector without increasing inference cost, among other improvements.
These traditional object detection methods perform well on public datasets because objects on public datasets, such as COCO and VOC, are generally more evenly distributed and have more obvious features. However, these methods have many shortcomings in the detection of traffic targets in traffic scenes. Firstly, the traffic images captured by the camera are limited by the shooting angle and shooting distance, and some traffic targets in them are small in scale, which is prone to missed detection or false detection. Secondly, the background of some road scenes is complex, and it is difficult to accurately detect the traffic targets we need. Finally, limited by the influence of light intensity, such as night or heavy rain, snow, fog, and other extreme weather, the traffic images captured by the camera are blurred and have low resolution and are also prone to missed detection or false detection.

2.2. Drivable Area Detection

Drivable area detection is often used as a regional constraint for lane detection. Currently, road detection is commonly considered a subset of road scene comprehension and remains one of the most formidable challenges for intelligent autonomous driving systems in related fields. Object detection in the environment perception system of autonomous driving provides the accurate location of various objects, such as pedestrians, vehicles, and obstacles in the road scene, which is of great significance for autonomous driving vehicles to find the driving area.
Many CNN-based methods have demonstrated impressive performance in the field of semantic segmentation and are widely used in drivable region detection, providing pixel-wise results. ENet [14] proposes a novel deep neural network architecture specifically for low-latency manipulation tasks, effectively balancing the accuracy and processing time of the network. PSPNet [16] provides an effective solution for pixel-level prediction, which effectively aggregates global context information and achieves good results on scene parsing tasks. ERFNet [17] proposes an end-to-end trainable deep neural network, incorporating residual connections and decomposed convolutions, thereby significantly enhancing computational efficiency compared to alternative architectures.
However, with the increase of network depth, the problem brought by the network structure model also becomes larger and more complex. In the semantic segmentation of road environments, since the application scenario is the vehicle computing platform, in addition to the segmentation accuracy, people pay more attention to whether the network processing speed can achieve real-time performance. Conventional semantic segmentation networks typically entail a substantial computational burden, rendering them unsuitable for meeting the real-time demands of autonomous driving systems.

2.3. Lane Detection

Lane detection plays a crucial role in advanced automated driver assistance systems. Traditional feature-based lane detection methods need to manually find specific lane features, such as geometry, color, gradient, and so on. This method is less robust, susceptible to noise interference, and has low detection efficiency.
The lane detection method based on deep learning leverages the robust feature extraction capability of convolutional neural networks to further extract lane-specific features. The current mainstream lane detection methods utilize semantic segmentation of images to comprehend the road environment semantically by isolating the lane as foreground and other areas and addressing the problem of lane detection from a semantic segmentation perspective. LaneNet [24] proposes an end-to-end fast lane detection algorithm using a novel view transformation method to correctly fit the lane even when the road plane changes greatly. SCNN [22] proposes a space-based convolutional neural network architecture capable of passing messages between pixels across rows and columns in layers, which is well suited for continuous shape object detection with strong spatial relationships. ENet-SAD [23] proposes a novel self-attention distillation method called SAD based on ENet [14], which autonomously learns and improves the model to cope with complex road situations without any additional supervision.
However, despite the fact that lane detection is commonly approached as semantic segmentation, it does not solely entail pixel-level classification. In the actual road environment, the lane often has incomplete marking lines, vehicle occlusion, and rain and snow cover, which makes the lane unvisible. In addition, the drastic change of road slope and angle in 2D images will also make it impossible to fit the lane correctly.

2.4. Multi-Task Learning Model

Now, the multi-task learning method based on CNN has been widely used. Multi-task learning can greatly improve the training efficiency by sharing information between multiple tasks, and multiple tasks can promote each other’s performance. MultiNet [27] proposes an end-to-end training method for classification, detection, and segmentation through a unified architecture, which has achieved good results in road detection tasks. DLT-Net [28] proposes an encoder–decoder-based network architecture that builds context tensors between decoders to share information between multiple tasks, and multiple tasks are able to promote each other’s performance during learning. YOLOP [29] introduces the residual network structure of YOLOv4 [9] into the network backbone for the first time, which can run multiple tasks accurately and quickly and achieves good results on embedded devices. HybridNets [30] introduces the BiFPN [34] structure into the feature fusion part of the network and proposes a more effective training strategy and loss function to achieve better accuracy while maintaining real-time performance.
These current popular multi-task learning networks are widely used in traffic scene detection tasks, which provide us with a good reference.

2.5. Attention Mechanism

The attention mechanism has emerged as a prominent research area in the realm of deep learning, garnering significant interest and attention from scholars worldwide. Compared with the convolution module as a local operator, the attention module can expand the global receptive field, making the model pay more attention to important weights during training. However, adding an attention mechanism has a certain cost, which has high computational complexity and takes up large memory.
In recent years, many attention mechanism modules have been proposed to improve the detection performance of existing CNN models. SENet [35] proposes a novel architecture that focuses on channel information and adaptively calibrates channel feature responses, yielding significant improvements in model performance at a small additional computational cost. CA [36] incorporates location information into channel attention, enabling the retention of precise spatial details and demonstrating exceptional performance across contemporary mainstream models. CBAM [37] takes into account both channel and spatial dimensions and adaptively refines the input feature map from two directions. It can be integrated into most CNN architectures and has a wide range of applicability. NAM [38] proposes a new attention mechanism based on normalization, which applies a weight sparsity penalty to suppress less important feature information and shows high computational efficiency. GAM [39] proposes a global attention mechanism with multi-layer perceptrons to improve the performance of deep neural networks by amplifying the global interaction representation, which achieves promising results on several public datasets.
These traditional attention mechanism modules adopt many methods to reduce computational overhead, such as restricting attention operations to local Windows or introducing unknown sparsity. These operations are often inflexible and have limited effect.

3. Methodology

We propose an efficient and lightweight multi-task network capable of simultaneously performing traffic object detection, drivable area detection, and lane detection tasks. It consists of an encoder and three decoders; three tasks perform feature extraction and feature fusion operations in a common encoder and perform information sharing, and finally, three different decoders solve the specific task. We name this network MultiNet-GS, and it can be trained end to end, incurs minimal computational overhead while maintaining high performance, and satisfies the real-time requirements for autonomous driving.

3.1. Encoder

3.1.1. Backbone

In view of the good performance of YOLOv8 network in the field of object detection, we choose CSPDarknet [9] as our backbone structure and introduce the C2f module in YOLOv8 network to form our feature extraction part. The overall structure of the model is shown in Figure 2. To enhance the feature extraction capability of the backbone and make the network pay more attention to important weights during feature extraction, we introduce a new dynamic sparse attention mechanism in the backbone, BiFormer, to achieve more flexible computing resource allocation. It demonstrates query-adaptive behavior by selectively focusing on relevant feature information while effectively filtering out irrelevant features. It exhibits exceptional performance and computational efficiency in dense prediction tasks.

3.1.2. Neck

We adopt the feature pyramid network FPN [40] + PAN structure in the neck part for fusing the features generated in the backbone part. FPN is a feature pyramid from top to bottom, which passes down the strong semantic features from high levels and enhances the whole pyramid. PAN is introduced as a supplementary component to FPN, incorporating a bottom-up pyramid structure for enhanced feature transmission and integration of low-level localization information. The pyramid formed in this way combines both semantic information and localization information. To fulfill the real-time requirements, we propose the integration of a lightweight convolution, GSConv, in the neck component to establish a novel slim-neck architecture that effectively balances model accuracy and speed.
In the conventional CNN network architecture, each spatial compression and channel expansion during the process of feature extraction and fusion result in the loss of inter-channel connections, leading to a degradation in semantic information. As the number of layers of the model goes deeper, the inference time will also grow significantly. GSConv can preserve these connections as much as possible while occupying a small computational overhead. We preserve the original backbone structure of the network and incorporate the GSConv module in the neck part to construct a slim-neck architecture without compromising the model’s feature extraction capability, thereby reducing computational complexity and inference time for object detection.

3.2. Decoders

3.2.1. Traffic Object Detection Head

In the traffic scene, the size of the traffic object in the image sometimes changes dramatically, and the size of some traffic objects is tiny, so the detector will miss and misdetect. In view of this situation, we add an extra detector for the tiny object on the traditional three-head detector structure. The four-head detector structure effectively mitigates the adverse effects caused by significant variations in traffic object size, thereby enhancing the detection performance of tiny objects with minimal computational overhead.

3.2.2. Drivable Area and Lane Detection Head

We put the semantic segmentation head at the end of the neck layer so that it can make full use of the fused feature information. Our slim-neck structure is able to effectively reduce the computational overhead that increases with the number of network layers. We utilize two different heads to perform the drivable area segmentation and lane segmentation tasks, respectively. The curve fitting method and semantic segmentation method are usually used in traditional lane detection. The curve fitting method uses mathematical methods to model the lane lines and fits the lane lines to a mathematical curve. This method is not ideal in some traffic scenes where the lane lines are occluded. Semantic segmentation methods regard lane detection as a traditional semantic segmentation problem. Because lane pixels and adjacent pixels usually have similar features, it is difficult to segment lane lines, and the effect is not ideal.
We introduce a new detection method in lane detection to distinguish it from the semantic segmentation method in drivable area detection. In the instance extraction branch, we use guide lines to identify lane keypoints, obtain instance features through RIM (recurrent instance module), and then calculate Gaussian masks and offset heatmaps through conditional convolution. Finally, we infer lane coordinates with an adaptive decoder.

3.3. Method Details

3.3.1. Biformer Attention Mechanism

In order to alleviate the scalability problem in the attention mechanism, the sparse attention mechanism is widely adopted in the attention mechanism module. The sparse attention mechanism is proposed to address the scalability problem in the attention mechanism by focusing only on a limited number of key-value pairs instead of all of them in the query. However, existing approaches either utilize hand-crafted static schemas or share a sampled subset of key-value pairs across all queries. In contrast, Biformer proposed a dynamic and query-aware sparse attention mechanism that includes an additional step for identifying regions to focus on, constructing and pruning region-level graphs and gathering key-value pairs from the routed regions.
As shown in Figure 3, the BiFomer module leverages the four-layer pyramid structure of Vision Transformer, incorporating the patch embedding module in the initial stage and integrating the patch merging module in subsequent stages two to four. With this bi-level routing attention (BRA) mechanism, it exploits sparsity to save computation and memory while only involving GPU-friendly dense matrix multiplication. The fundamental concept of BRA is to selectively extract the most pertinent key-value pairs at a macroscopic regional level. This objective is accomplished by initially constructing and refining a directed graph at the regional level, followed by employing meticulous token-to-token attention within the amalgamation of routed regions.
The BiFormer model adopts a query adaptive approach to focus on a limited set of relevant tokens, effectively disregarding irrelevant ones. This design choice not only ensures superior performance but also enhances computational efficiency, particularly in dense prediction tasks. Notably, BiFormer has demonstrated remarkable efficacy across various domains such as image classification, object detection, and semantic segmentation.

3.3.2. Gsconv Module

The requirements of real-time object detection pose a challenge for achieving a large-scale model on vehicular edge computing platforms. Conversely, lightweight models constructed with numerous depthwise separable convolutional layers often fall short in terms of accuracy. In order to address this issue, we propose the integration of GSConv, a lightweight convolution technique, into the neck part to establish a novel slim-neck structure that strikes an optimal balance between model accuracy and speed.
The input image in the CNN backbone typically undergoes a similar transformation process to enhance the computation speed of the final prediction, wherein spatial information is gradually transferred to the channel. However, each compression of spatial dimensions and expansion of channels in the feature map leads to a partial loss of semantic information. While channel-dense convolution maximally preserves hidden connections between each channel, channel-sparse convolution completely severs these connections. GSConv aims to preserve these connections as much as possible with low time complexity. The structure of the GSConv module is shown in Figure 4.
The introduction of slim-neck structure improves the computational efficiency of the model in the feature fusion part; avoids the loss of semantic information in the feature fusion stage; significantly improves the detection accuracy of the three detection head parts of object detection, drivable area detection, and lane detection; and greatly reduces the computational overhead of the model.

3.3.3. Lane Detection Method Based on Guide Line

Guide lines are often used to mark the origin of lane lines, as shown in Figure 5. The black dashed line represents the detected vectorized lane line, and the red dashed line represents the pixel range over which the response was generated on the heatmap. d is the radius of the range, and the grazing angles of the two guide lines are α 1 and α 2 , respectively. As shown in Equation (1), the smaller the grazing angle of the guide line is, the larger the response range is and, conversely, the smaller the response range is. The addition of guidance lines allows the model to focus on the dramatic changes in lane lines around corners.
x i = 2 d s i n ( α i ) ; i = 1 , 2
According to the guiding line principle, we introduce a U-shaped curve with the lower half being a partial ellipse and the upper half being the image side boundary, and the ellipse is defined in Equation (2), where w and h are the width and height of the feature map, respectively, and c x and c y are the hyperparameters that adjust the center of the ellipse. Due to the bending characteristics at the curve, the U-shaped guide line will lead to the increase of the sweep angle of the curve so that the model can be more sensitive to the sharp change of lane lines at the corner of the road. This special curved guide line usually performs better than the normal guide line when locating the lane line.
( x c x · w ) 2 ( c x · w ) + ( y c y · h ) 2 ( c y · h ) = 1
We use the normalized Gaussian kernel of the offset to extract the keypoints in the lane line instance as shown in Equation (3). Where x and y represent the exact coordinates of each keypoint. Finally, by using Gaussian mask and adaptive decoder mechanism to decode the key point set in the supervision stage, the final fitted lane line is less affected by the surrounding pixels, and the detection performance is greatly improved compared with the traditional semantic segmentation method.
Y x y = n o r m a l i z e ( e x p ( x 2 + y 2 2 σ 2 ) )
The improved lane line detection head is shown in Figure 6.

3.4. Loss Function

Since our model is composed of three decoders, we use different loss functions for different decoders. In the object detection part, we adopt the weighted sum of bounding box loss, classification loss, and object loss to form our object detection loss function L d e t , as shown in Equation (4).
L d e t = α 1 L b o x + α 2 L c l a s s + α 2 L o b j
Among them, L b o x adopts L C I o U , which is used to evaluate the distance, overlap rate, and size similarity between the predicted box and the actual traffic object. Focal loss is applied to both L c l a s s and L o b j , where L c l a s s is used to penalize classification and L o b j is used to predict the confidence of an object.
Cross-entropy loss with Logits L c e is used in the drivable area detection partial loss function L d a , while L l l uses L c e and L I o U in the lane line detection partial loss function, as shown in Equations (5) and (6). Both aim to reduce the error between the output pixels of the model and the actual image regions.
L d a = L c e
L l l = L c e + L I o U
L a l l = γ 1 L d e t + γ 2 L d a + γ 3 L l l
Our final loss function is a weighted sum of the three partial loss functions, as shown in Equation (7). α 1 , α 2 , α 3 , γ 1 , γ 2 , γ 3 are adjusted to balance the final loss function.

4. Experiments

4.1. Setting

4.1.1. Dataset Construction

We utilized the BDD100K dataset, extensively employed in the domain of autonomous driving, for training and evaluating our network. The dataset comprises three components, with 70,000 images allocated as the training set, 20,000 images designated as the test set, and 10,000 images assigned to the validation set. There are a total of 100,000 images, covering traffic scenes under various environments, times of day, and weather, which is robust enough. In addition, we collected a large number of traffic surveillance videos of the ring highway in Xi’an, China, and produced a self-made dataset with rich scenes, about 30,000 images, including cars, trucks, motorcycles, pedestrians, roadblocks, signs, aba-performed objects, and other complex traffic targets. Compared with the BDD100K dataset with only cars as traffic targets, we collected a large number of traffic surveillance videos of the ring highway in Xi’an, China. Our dataset contains a large number of traffic targets with tiny size that are difficult to detect in order to verify the detection performance of our model for tiny objects. The dataset is shown in Figure 7.

4.1.2. Implementation Details

We implement MultiNet-GS on Pytorch 1.8.1, CUDA 11.3. All of our models use an NVIDIA RTX3090 GPU for training and testing. In addition, in order to test the running effect of our model on embedded devices, we also adopt the NVIDIA Jetson TX2 (4 GB) embedded platform to comprehensively evaluate the detection accuracy and detection speed of our model. We jointly train the three tasks of the model to share the image features extracted in the encoder. We set the training epochs to 300, the initial learning rate to 0.01, and dynamically reset the learning rate in the first few epochs. We set an early stop mechanism during the training process so that the model stops training if the loss curve of the model cannot continue to converge. We also adopt mixup data augmentation technique in the process of dataset loading to expand the number of dataset samples and enhance the robustness of the model. We resize the images in the BDD100K dataset to 640 × 640 × 3 during the training phase.

4.1.3. Evaluation Index

In the object detection part of the model, mean average precision ( m A P ), precision (P), recall (R), are used to evaluate the performance of the algorithm. Where mean average precision ( m A P ), precision (P), and recall (R) are expressed as
m A P = 1 n n i 1 A P i
P = T P T P + F P
R = T P T P + F N
where average precision ( A P ) indicates the average precision of an object class, while true positives ( T P ) represents the number of correctly detected objects. False positives ( F P ) denotes the number of incorrectly detected objects, and false negatives ( F N ) signifies the number of undetected objects
In the drivable area detection part and lane detection part of the model, a c c u r a c y , I O U , m I O U , are used to evaluate the performance of the algorithm. Where a c c u r a c y , I O U , and m I O U are expressed as
a c c u r a c y = i n i i i t i
I o U = i n i i t i + j n j i n i i
m I o U = 1 n c l s i n i i t i + j n j i n i i
where n i i represents the part where the prediction is correct, that is, the part where the two regions coincide. t i corresponds to the total number of pixels of class i, and j n j i represents the total number of all pixels predicted as class i.
In addition, in order to evaluate the real-time performance of our proposed algorithm, we introduce two metrics, Params and FLOPs, to evaluate the number of parameters and the amount of computation of the algorithm model, respectively. We introduce FPS to evaluate the detection speed when the model is running.

4.2. Result

4.2.1. Comparison Experiments of Model with Improved Four-Head Structure

We add an additional object detection head to the object detection part of the baseline model to detect tiny size objects and verify our improved results on our self-made traffic scene dataset. The experimental results are shown in Table 1, and the loss function curve of the model is shown in Figure 8.
The experimental results show that the additional detection head we added has a significant improvement on the detection effect of the model. The detection effect of the model on tiny size objects, such as pedestrians, roadblocks, reflective clothing, boxes, and motorcycles, in the dataset is significantly improved, and the detection accuracy is increased by 5.4%, 5.6%, 3.2%, 2.6%, 2.0% respectively, which proves the effectiveness of our improvement.

4.2.2. Ablation Experiments after Introducing Different Improved Methods

In order to compare the influence of different improvements on model performance, we will call the model with BiFormer structure as Model-B, the model with GSConv structure as Model-G, the model with four-head structure as Model-F, and the model with improved lane detection structure as Model-L. We design ablation experiments from the following two directions: (1) Based on the baseline model, each improvement method was individually incorporated to assess its impact on the baseline model’s performance. (2) Based on the final model, each improvement method was systematically eliminated to evaluate its influence on the final model’s outcome.
As shown in Table 2, experimental results show that in the object detection part of the model, the introduction of BiFormer has the largest improvement on the model detection effect, with the [email protected] increased by 5.8%, but Params and FLOPs are increased by 1.8 M and 6.8 respectively, the detection speed of the model is reduced by 7 FPS. The introduction of GSConv has the largest improvement on the model detection speed, with an increase of 13 FPS, and Params and FLOPs are reduced by 4 M and 11.4, respectively. Compared with the baseline model, the final improved model has the [email protected] increased by 8.1%, Params and FLOPs are increased by 3.4 M and 5.8, respectively, and the detection speed decreased by 2 FPS. The final improved model maintains a good detection speed while greatly improving the detection accuracy, which proves the effectiveness of our improvement in the object detection part.
As shown in Table 3, experimental results show that in the drivable area detection part of the model, the introduction of BiFormer has the largest improvement on the model detection effect, with the accuracy increased by 5.8%, IOU increased by 2.4, and mIOU increased by 2.1, but Params and FLOPs are increased by 1.8 M and 6.8, respectively, and the detection speed of the model is reduced by 4 FPS. The introduction of GSConv has the largest improvement on the model detection speed, with an increase of 13 FPS, and Params and FLOPs are reduced by 4 M and 11.4, respectively. Compared with the baseline model, the final improved model has the accuracy increased by 6.4%, the IOU increased by 1.7, and the mIOU increased by 1.9, Params and FLOPs are reduced by 1.8 M and 6.7, respectively. The final improved model maintains a good detection speed while greatly improving the detection accuracy, which proves the effectiveness of our improvement in the object detection part.
As shown in Table 4, experimental results show that in the line detection part of the model, the improvement of the lane detection part has the largest improvement on the model detection effect, with the accuracy increased by 5.7%, IOU increased by 4.5, and mIOU increased by 1.9, but Params and FLOPs are increased by 1.4M and 1.4, respectively, and the model detection speed reduced by 3 FPS. The introduction of GSConv has the largest improvement on the model detection speed, with an increase of 11 FPS, and Params and FLOPs are reduced by 4 M and 11.4, respectively. Compared with the baseline model, the final improved model has the accuracy increased by 11.3%, the IOU increased by 10.2, the mIOU increased by 14, Params and FLOPs are increased by 3.1M and 3.7, respectively, and the detection speed increased by 2 FPS. The final improved model maintains a good detection speed while greatly improving the detection accuracy, which proves the effectiveness of our improvement in the object detection part.

4.2.3. Comparison Experiment with the Current Mainstream Network Models

As shown in Table 5, experimental results show that, in the object detection part of the model, our MultiNet-GS model has an increase of 2.3% compared with SOTA (YOLOPv2) on [email protected], Params and FLOPs are reduced by 6.6 M and 8.3, respectively, and the detection speed is reduced by 3 FPS. It has the highest detection accuracy among the current mainstream object detection models, while maintaining a high detection speed.
As shown in Table 6, experimental results show that in the drivable area detection part of the model, compared with SOTA (YOLOPv2), our MultiNet-GS model improves the accuracy by 1.5%, improves the IOU by 2.0, and improves the mIOU by 2.3; Params and FLOPs are reduced by 6.6 M and 8.3, respectively, and the detection speed is reduced by 3 FPS. It has the highest detection accuracy among the current mainstream object detection models, while maintaining a high detection speed.
As shown in Table 7, experimental results show that in the line detection part of the model, compared with SOTA (YOLOPv2), our MultiNet-GS model improves the accuracy by 3.2%, improves the IOU by 4.0, and improves the mIOU by 2.3; Params and FLOPs are reduced by 6.6 M and 8.3, respectively, and the detection speed is reduced by 3 FPS. It has the highest detection accuracy among the current mainstream object detection models, while maintaining a high detection speed.

4.2.4. Comparison Experiment of the Improved Model on Embedded Platform

In order to test the running effect of our proposed model on the embedded platform, we deploy the model on the NVIDIA Jetson TX2 embedded AI platform and compare the experiment with the current mainstream multi-task network model, where [email protected] represents the detection accuracy of the traffic object detection part of the model, accuracy1 represents the accuracy of the drivable area segmentation part of the model, and accuracy2 represents the accuracy of the lane line segmentation part of the model.
As shown in Table 8, experimental results show that although the detection performance of our proposed model slightly decreases on the embedded platform, it still maintains good detection accuracy and detection speed as a whole. Compared with SOTA(YOLOPv2), the [email protected] of the model in traffic object detection reaches 82.1%, which is increased by 2.7%. The accuracy of the model in drivable area detection reaches 93.2%, which is increased by 0.5%. The accuracy of the model in lane detection reaches 85.7%, which is increased by 4.3%. The Params and FLOPs of the model reach 47.5 M and 117.5 M, which are reduced by 6.6 M and 8.3, respectively. The model achieves 72 FPS, which is increased by 5. The detection effect of the model on the BDD100K dataset is shown in Figure 9 and Figure 10.

5. Conclusions

In order to solve the problem of environment perception under structured roads in the process of autonomous driving, a multi-task network model is proposed to simultaneously process three tasks: traffic object detection, drivable area detection, and lane detection. However, the current mainstream multi-task network model has some shortcomings in complex traffic scenarios. First, for traffic objects with small size, the detection accuracy of the object detection part of the model is low, which is prone to missed detection and false detection. Second, the segmentation accuracy of drivable area and lane line is low when the image is blurred by extreme weather, such as rain, snow, and fog, or the road is occluded. Third, the segmentation of lane lines is prone to discontinuity in the case of large changes in lane slope and sharp lane turns. Fourth, most of the multi-task network models have complex structures, occupy large computational overhead, and perform poorly on embedded devices with limited computing power, and their performance cannot meet the real-time requirements in autonomous driving.
To this end, we propose a multi-task convolutional neural network model MultiNet-GS based on encoder–decoder structure. We use the main structure of the latest object detection model, YOLOv8 model, as the encoder structure of our model. We introduce a new dynamic sparse attention mechanism, BiFormer, in the feature extraction part of the model to achieve more flexible computing resource allocation, which can significantly improve the computational efficiency and occupy a small computational overhead. The introduction of the attention mechanism makes the model pay more attention to the region of interest and improves the accuracy of object detection, drivable area segmentation, and lane line segmentation. We introduce a lightweight convolution, GSConv, in the feature fusion part of the network, which is used to build the neck part into a new slim-neck structure so as to reduce the computational complexity and inference time of the detector. The introduction of slim-neck structure greatly reduces the number of parameters and calculation of the model so that the model can adapt to embedded devices with limited computing power and meet the real-time requirements in autonomous driving. We also add an additional detector for tiny objects to the traditional three-head detector structure, and the introduction of the four-head structure leads to a substantial increase in the accuracy of the model for detecting tiny objects in traffic images. Finally, we introduce a lane detection method based on guide line in the lane detection part, which can aggregate the lane feature information into multiple key points, and then obtain the lane heat map response through conditional convolution. Finally, the lane line is described by the adaptive decoder, which effectively makes up for the shortcomings of the traditional lane detection method. So that the model can accurately and continuously segment the lane line in the case of drastic changes in lane slope and sharp lane turns.
Our comparative experiments on the BDD100K dataset on the embedded platform NVIDIA Jetson TX2 show that compared with SOTA(YOLOPv2), the [email protected] of the model in traffic object detection reaches 82.1%, which is increased by 2.7%. The accuracy of the model in drivable area detection reaches 93.2%, which is increased by 0.5%. The accuracy of the model in lane detection reaches 85.7%, which is increased by 4.3%. The Params and FLOPs of the model reach 47.5 M and 117.5, which are reduced by 6.6 M and 8.3, respectively. The model achieves 72 FPS, which is increased by 5. Our MultiNet-GS model has the highest detection accuracy among the current mainstream models, while maintaining a good detection speed. Our proposed model makes up for the shortcomings of the traditional multi-task network model and can maintain good real-time performance on the embedded devices in the process of autonomous driving. At the same time, it can maintain the detection accuracy of traffic objects, drivable areas, and lane lines in complex traffic scenes so that the vehicle can make timely and accurate decisions and control in the process of environment perception. It effectively ensures the safety of automatic driving.
In the future work, we will continue to optimize the number of parameters and calculation of the network model, reduce the time cost of the model operation, and optimize and improve the network model for specific traffic scenarios to enhance the robustness of the model so that it can adapt to more diverse traffic scenarios.

Author Contributions

Methodology, A.L.; Software, A.L.; Validation, A.L.; Writing—original draft, A.L.; Writing—review & editing, Z.Z., S.S., M.F. and C.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by The National Natural Science Foundation of China grant number 62006026 and 62072053 and The Central Universities Basic Research Special Funds grant number 300102241304 and 300102241202. The APC was funded by Zhaoyang Zhang.

Data Availability Statement

The datasets we produced and the experimental data we obtained are not publicly available.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Duan, K.; Bai, S.; Xie, L.; Qi, H.; Huang, Q.; Tian, Q. CenterNet: Object Detection with Keypoint Triplets. arXiv 2019, arXiv:1904.08189. [Google Scholar]
  2. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollar, P. Focal Loss for Dense Object Detection. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar] [CrossRef]
  3. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-End Object Detection with Transformers. In Computer Vision—ECCV 2020, Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2020; pp. 213–229. [Google Scholar] [CrossRef]
  4. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014. [Google Scholar] [CrossRef]
  5. Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015. [Google Scholar] [CrossRef]
  6. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 28, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  7. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Computer Vision–ECCV 2016, Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar] [CrossRef]
  8. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  9. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  10. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv 2022, arXiv:2207.02696. [Google Scholar]
  11. Zhang, H.; Li, F.; Liu, S.; Zhang, L.; Su, H.; Zhu, J.; Ni, L.; Shum, H.Y. DINO: DETR with Improved DeNoising Anchor Boxes for End-to-End Object Detection. arXiv 2022, arXiv:2203.03605. [Google Scholar]
  12. Fang, Y.; Wang, W.; Xie, B.; Sun, Q.; Wu, L.; Wang, X.; Huang, T.; Wang, X.; Cao, Y. EVA: Exploring the Limits of Masked Visual Representation Learning at Scale. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 19358–19369. [Google Scholar]
  13. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Lecture Notes in Computer Science, Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015; Springer International Publishing: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar] [CrossRef]
  14. Paszke, A.; Chaurasia, A.; Kim, S.; Culurciello, E. ENet: A Deep Neural Network Architecture for Real-Time Semantic Segmentation. arXiv 2016, arXiv:1606.02147. [Google Scholar]
  15. Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef] [PubMed]
  16. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid Scene Parsing Network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar] [CrossRef]
  17. Romera, E.; Alvarez, J.M.; Bergasa, L.M.; Arroyo, R. ERFNet: Efficient Residual Factorized ConvNet for Real-Time Semantic Segmentation. IEEE Trans. Intell. Transp. Syst. 2018, 19, 263–272. [Google Scholar] [CrossRef]
  18. Zhao, H.; Zhang, Y.; Liu, S.; Shi, J.; Loy, C.C.; Lin, D.; Jia, J. PSANet: Point-wise Spatial Attention Network for Scene Parsing. In Proceedings of the Computer Vision–ECCV 2018, Lecture Notes in Computer Science, Munich, Germany, 8–14 September 2018; pp. 270–286. [Google Scholar] [CrossRef]
  19. Zhang, H.; Dana, K.; Shi, J.; Zhang, Z.; Wang, X.; Tyagi, A.; Agrawal, A. Context Encoding for Semantic Segmentation. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar] [CrossRef]
  20. Zhang, H.; Zhang, H.; Wang, C.; Xie, J. Co-Occurrent Features in Semantic Segmentation. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar] [CrossRef]
  21. Guo, M.H.; Liu, Z.N.; Mu, T.J.; Hu, S.M. Beyond Self-attention: External Attention using Two Linear Layers for Visual Tasks. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 5436–5447. [Google Scholar] [CrossRef] [PubMed]
  22. Pan, X.; Shi, J.; Luo, P.; Wang, X.; Tang, X. Spatial as deep: Spatial CNN for traffic scene understanding. Proc. AAAI Conf. Artif. Intell. 2022, 32. [Google Scholar] [CrossRef]
  23. Hou, Y.; Ma, Z.; Liu, C.; Loy, C.C. Learning Lightweight Lane Detection CNNs by Self Attention Distillation. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar] [CrossRef]
  24. Neven, D.; Brabandere, B.D.; Georgoulis, S.; Proesmans, M.; Gool, L.V. Towards End-to-End Lane Detection: An Instance Segmentation Approach. In Proceedings of the 2018 IEEE Intelligent Vehicles Symposium (IV), Changshu, China, 26–30 June 2018. [Google Scholar] [CrossRef]
  25. Tabelini, L.; Berriel, R.; Paixao, T.M.; Badue, C.; De Souza, A.F.; Oliveira-Santos, T. Keep your Eyes on the Lane: Real-time Attention-guided Lane Detection. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021. [Google Scholar] [CrossRef]
  26. Abualsaud, H.; Liu, S.; Lu, D.B.; Situ, K.; Rangesh, A.; Trivedi, M.M. LaneAF: Robust Multi-Lane Detection with Affinity Fields. IEEE Robot. Autom. Lett. 2021, 6, 7477–7484. [Google Scholar] [CrossRef]
  27. Teichmann, M.; Weber, M.; Zollner, M.; Cipolla, R.; Urtasun, R. MultiNet: Real-time Joint Semantic Reasoning for Autonomous Driving. In Proceedings of the 2018 IEEE Intelligent Vehicles Symposium (IV), Changshu, China, 26–30 June 2018. [Google Scholar] [CrossRef]
  28. Qian, Y.; Dolan, J.M.; Yang, M. DLT-Net: Joint Detection of Drivable Areas, Lane Lines, and Traffic Objects. IEEE Trans. Intell. Transp. Syst. 2020, 21, 4670–4679. [Google Scholar] [CrossRef]
  29. Wu, D.; Liao, M.W.; Zhang, W.T.; Wang, X.G.; Bai, X.; Cheng, W.Q.; Liu, W.Y. YOLOP: You Only Look Once for Panoptic Driving Perception. Mach. Intell. Res. 2022, 19, 550–562. [Google Scholar] [CrossRef]
  30. Vu, D.; Ngo, B.; Phan, H. HybridNets: End-to-End Perception Network. arXiv 2022, arXiv:2203.09035. [Google Scholar]
  31. Han, C.; Zhao, Q.; Zhang, S.; Chen, Y.; Zhang, Z.; Yuan, J. YOLOPv2: Better, Faster, Stronger for Panoptic Driving Perception. arXiv 2022, arXiv:2208.11434. [Google Scholar]
  32. Zhu, L.; Wang, X.; Ke, Z.; Zhang, W.; Lau, R. BiFormer: Vision Transformer with Bi-Level Routing Attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 10323–10333. [Google Scholar]
  33. Li, H.; Li, J.; Wei, H.; Liu, Z.; Zhan, Z.; Ren, Q. Slim-neck by GSConv: A better design paradigm of detector architectures for autonomous vehicles. arXiv 2022, arXiv:2206.02424. [Google Scholar]
  34. Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and Efficient Object Detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar] [CrossRef]
  35. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-Excitation Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 2011–2023. [Google Scholar] [CrossRef] [PubMed]
  36. Hou, Q.; Zhou, D.; Feng, J. Coordinate Attention for Efficient Mobile Network Design. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021. [Google Scholar] [CrossRef]
  37. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Computer Vision–ECCV 2018, Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2018; pp. 3–19. [Google Scholar] [CrossRef]
  38. Liu, Y.; Shao, Z.; Teng, Y.; Hoffmann, N. NAM: Normalization-based Attention Module. arXiv 2021, arXiv:2111.12419. [Google Scholar]
  39. Liu, Y.; Shao, Z.; Hoffmann, N. Global Attention Mechanism: Retain Information to Enhance Channel-Spatial Interactions. arXiv 2021, arXiv:2112.05561. [Google Scholar]
  40. Lin, T.Y.; Dollar, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 June 2017. [Google Scholar] [CrossRef]
Figure 1. Structured roads refer to highways, urban arterial roads, and other well-structured highways. This kind of roads has clear road marking lines, a relatively single background environment of roads, and obvious geometric characteristics of roads.
Figure 1. Structured roads refer to highways, urban arterial roads, and other well-structured highways. This kind of roads has clear road marking lines, a relatively single background environment of roads, and obvious geometric characteristics of roads.
Electronics 12 03994 g001
Figure 2. The overall structure of the model.
Figure 2. The overall structure of the model.
Electronics 12 03994 g002
Figure 3. The structure of the BiFormer module.
Figure 3. The structure of the BiFormer module.
Electronics 12 03994 g003
Figure 4. The structure of the GSConv module.
Figure 4. The structure of the GSConv module.
Electronics 12 03994 g004
Figure 5. Guide lines used to detect grazing angles of lane lines. The blue and yellow lines are the added guide lines.
Figure 5. Guide lines used to detect grazing angles of lane lines. The blue and yellow lines are the added guide lines.
Electronics 12 03994 g005
Figure 6. The structure of the lane detection head.
Figure 6. The structure of the lane detection head.
Electronics 12 03994 g006
Figure 7. BDD100K dataset and our self-made dataset.
Figure 7. BDD100K dataset and our self-made dataset.
Electronics 12 03994 g007
Figure 8. Loss function curves for the three-head structure model and the four-head structure model.
Figure 8. Loss function curves for the three-head structure model and the four-head structure model.
Electronics 12 03994 g008
Figure 9. Detection results of MultiNet-GS model on different tasks on BDD100K dataset.
Figure 9. Detection results of MultiNet-GS model on different tasks on BDD100K dataset.
Electronics 12 03994 g009
Figure 10. Detection results of MultiNet-GS model under different weather and lighting environments on BDD100K dataset.
Figure 10. Detection results of MultiNet-GS model under different weather and lighting environments on BDD100K dataset.
Electronics 12 03994 g010
Table 1. Experimental results of the four-head structure model on the self-made dataset (the experimental results are given in [email protected](%)).
Table 1. Experimental results of the four-head structure model on the self-made dataset (the experimental results are given in [email protected](%)).
MethodsAllCarTruckBusPersonFireSmokeConeDivSuitBoxMoto
Three-head76.788.185.774.269.790.292.472.765.955.467.881.2
Four-head79.490.287.477.975.190.791.978.369.758.670.483.2
Table 2. Ablation experiments after introducing different improved methods in the traffic object detection part. “+” represents the introduction of this method.
Table 2. Ablation experiments after introducing different improved methods in the traffic object detection part. “+” represents the introduction of this method.
MethodsBGF[email protected] (%)Precision (%)Recall (%)Params (M)FLOPsFPS
Model (Baseline) 77.613.885.442.7108.892
Model-B+ 81.811.991.244.5115.685
Model-G + 79.813.788.738.797.4105
Model-F +81.412.190.243.8112.792
Model-BG++ 82.011.092.140.9102.192
Model-BF+ +85.511.294.047.6121.787
Model-GF ++82.112.391.741.4104.494
Model-BGF+++85.711.294.146.1114.690
Table 3. Ablation experiments after introducing different improved methods in the driving area detection part. “+” represents the introduction of this method.
Table 3. Ablation experiments after introducing different improved methods in the driving area detection part. “+” represents the introduction of this method.
MethodsBGAccuracy (%)IOUmIOUParams (M)FLOPsFPS
Model (Baseline) 90.388.790.242.7108.894
Model-B+ 96.191.192.344.5115.690
Model-G +94.288.291.138.797.4107
Model-BG++96.790.492.140.9102.194
Table 4. Ablation experiments after introducing different improved methods in the lane detection part. “+” represents the introduction of this method.
Table 4. Ablation experiments after introducing different improved methods in the lane detection part. “+” represents the introduction of this method.
MethodsBGLAccuracy (%)IOUmIOUParams (M)FLOPsFPS
Model (Baseline) 76.465.275.442.7108.894
Model-B+ 81.969.180.944.5115.688
Model-G + 79.866.476.338.797.4105
Model-L +82.169.777.344.1110.291
Model-BG++ 86.872.785.240.9102.192
Model-BL+ +87.273.185.146.6120.481
Model-GL ++86.972.785.741.0100.399
Model-BGL+++87.775.489.445.8112.596
Table 5. Comparison experiments of our method and current mainstream methods on object detection.
Table 5. Comparison experiments of our method and current mainstream methods on object detection.
Methods[email protected] (%)Precision(%)Recall (%)Params (M)FLOPsFPS
Faster R-CNN55.622.777.267.9147.28.8
YOLOV5s77.217.186.87.216.5114
YOLOV7s78.414.785.336.5103.577
YOLOV8s81.713.888.410.225.797
MultiNet60.220.781.272.4188.68.6
DLT-Net68.419.789.4108.3211.79.3
YOLOP76.511.889.257.2141.741
HybridNets77.310.792.887.9171.228
YOLOPv283.410.491.154.1125.891
MultiNet-GS85.711.294.147.5117.588
Table 6. Comparison experiments of our method and current mainstream methods on drivable area detection.
Table 6. Comparison experiments of our method and current mainstream methods on drivable area detection.
MethodsAccuracy (%)IOUmIOUParams (M)FLOPsFPS
ENet69.170.477.958.4124.751
SegNet82.476.579.672.1142.444
PSPNet81.677.680.245.287.611.1
MultiNet87.281.485.672.4188.68.6
DLT-Net89.480.782.9108.3211.79.3
YOLOP93.085.489.757.2141.741
HybridNets93.588.191.587.9171.228
YOLOPv295.288.490.454.1125.891
MultiNet-GS96.790.492.747.5117.588
Table 7. Comparison experiments of our method and current mainstream methods on lane detection.
Table 7. Comparison experiments of our method and current mainstream methods on lane detection.
MethodsAccuracy (%)IOUmIOUParams (M)FLOPsFPS
LaneNet60.658.666.444.277.647
ENet-SAD68.765.972.437.662.150.6
SCNN71.660.477.851.4102.119.8
DLT-Net70.261.779.3108.3211.79.3
YOLOP79.665.781.457.2141.741
HybridNets82.169.788.687.9171.228
YOLOPv284.571.487.154.1125.891
MultiNet-GS87.775.489.447.5117.588
Table 8. Comparison experiments of our method and current mainstream methods on embedded platform.
Table 8. Comparison experiments of our method and current mainstream methods on embedded platform.
Methods[email protected] (%)Accuracy1 (%)Accuracy2 (%)Params (M)FLOPsFPS
MultiNet55.682.4072.4188.65.6
DLT-Net66.785.768.9108.3211.77.2
YOLOP70.390.276.557.2141.733
HybridNets69.889.680.287.9171.217
YOLOPv279.492.781.454.1125.867
MultiNet-GS82.193.285.747.5117.572
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, A.; Zhang, Z.; Sun, S.; Feng, M.; Wu, C. MultiNet-GS: Structured Road Perception Model Based on Multi-Task Convolutional Neural Network. Electronics 2023, 12, 3994. https://doi.org/10.3390/electronics12193994

AMA Style

Li A, Zhang Z, Sun S, Feng M, Wu C. MultiNet-GS: Structured Road Perception Model Based on Multi-Task Convolutional Neural Network. Electronics. 2023; 12(19):3994. https://doi.org/10.3390/electronics12193994

Chicago/Turabian Style

Li, Ang, Zhaoyang Zhang, Shijie Sun, Mingtao Feng, and Chengzhong Wu. 2023. "MultiNet-GS: Structured Road Perception Model Based on Multi-Task Convolutional Neural Network" Electronics 12, no. 19: 3994. https://doi.org/10.3390/electronics12193994

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop