Next Article in Journal
Cooperative-Aware Radio Resource Allocation Scheme for 5G Network Slicing in Cloud Radio Access Networks
Previous Article in Journal
Enhancing the Squareness and Bi-Phase Magnetic Switching of Co2FeSi Microwires for Sensing Application
Previous Article in Special Issue
Observer-Based Optimal Control of a Quadplane with Active Wind Disturbance and Actuator Fault Rejection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fault Diagnosis of the Autonomous Driving Perception System Based on Information Fusion

1
School of Reliability and Systems Engineering, Beihang University, Beijing 100191, China
2
General Design Department, Beijing Mechanical and Electrical Engineering, Beijing 100005, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(11), 5110; https://doi.org/10.3390/s23115110
Submission received: 12 April 2023 / Revised: 18 May 2023 / Accepted: 22 May 2023 / Published: 26 May 2023

Abstract

:
The reliability of autonomous driving sensing systems impacts the overall safety of the driving system. However, perception system fault diagnosis is currently a weak area of research, with limited attention and solutions. In this paper, we present an information-fusion-based fault-diagnosis method for autonomous driving perception systems. To begin, we built an autonomous driving simulation scenario using PreScan software, which collects information from a single millimeter wave (MMW) radar and a single camera sensor. The photos are then identified and labeled via the convolutional neural network (CNN). Then, we fused the sensory inputs from a single MMW radar sensor and a single camera sensor in space and time and mapped the MMW radar points onto the camera image to obtain the region of interest (ROI). Lastly, we developed a method to use information from a single MMW radar to aid in diagnosing defects in a single camera sensor. As the simulation results show, for missing row/column pixel failure, the deviation typically falls between 34.11% and 99.84%, with a response time of 0.02 s to 1.6 s; for pixel shift faults, the deviation range is between 0.32% and 9.92%, with a response time of 0 s to 0.16 s; for target color loss, faults have a deviation range of 0.26% to 2.88% and a response time of 0 s to 0.05 s. These results prove the technology is effective in detecting sensor faults and issuing real-time fault alerts, providing a basis for designing and developing simpler and more user-friendly autonomous driving systems. Furthermore, this method illustrates the principles and methods of information fusion between camera and MMW radar sensors, establishing the foundation for creating more complicated autonomous driving systems.

1. Introduction

Autonomous driving systems have emerged as a prominent development direction in the 5G era, particularly in China’s automobile industry, which increasingly focuses on intelligent vehicles [1]. However, the rise of autonomous driving has also been accompanied by notable and frequent accidents. Among the key components of autonomous driving systems, sensors, controllers, and actuators play critical roles. The sensor component, in particular, is essential for accurate positioning and obstacle detection, and a sensor failure can lead to catastrophic accidents. As a result, research on fault diagnosis of autonomous driving sensing systems has become increasingly important and is currently a topic of growing interest [2].
Autonomous vehicles necessitate the deployment of numerous sensors to gather broad information for comprehensive vehicle analysis and decision making in the face of complex road conditions [3]. With the aid of extremely effective and dependable sensor fusion algorithms, the sensors built into autonomous vehicles must deliver prompt and accurate feedback in an array of environments and usage conditions to allow continuous navigation. Due to their limited information-gathering capabilities, poor noise resistance, and low fault tolerance, the deployment of homogeneous sensors is insufficient to meet the requirements of autonomous vehicles for road condition analysis, making it impossible to achieve autonomous driving in complex situations. Research challenges for autonomous driving sensing systems currently center on two areas: (1) information fusion of data collected from multiple sensors, thereby reducing the uncertainty of information and improving the accuracy of characterization of the environment [4]; (2) choosing appropriate sensor fault diagnosis methods, thereby lowering the cost of diagnosis and reducing diagnosis time [5].
Scholars have been actively studying the first problem and exploring multi-sensor fusion algorithms that improve robustness and detection accuracy. Some scholars committed to studying multi-sensor fusion under poor conditions (weather/distance/light) have achieved significant results. For instance, Jiang et al. [6] proposed a millimeter wave (MMW) radar and camera fusion-based technique for vehicle perimeter sensing, which may improve target recognition precision in foggy circumstances. In particular, the method defogs the camera-taken images before mapping the crucial data fragments from the MMW radar to the defogged image to create the region of interest (ROI). Lastly, the detection data are combined using a weighting algorithm. Meanwhile, S. Chadwick et al. [7] developed a feature-level fusion structure to detect small objects at a distance. To extract the reflected area and velocity data from pictures and MMW radar, respectively, the structure initially employs residual networks (ResNet). Then, it stitches together radar point cloud and picture feature data. Finally, using ResNet, three-resolution feature maps are produced from the stitched features for diverse sensing applications. The proposed method balances vehicle detection accuracy and computing efficiency, achieving an optimal performance trade off. Wang et al. [8] introduced a decision-level fusion system to identify cars in wet situations. This method involves ground-calibrating the radar and camera, transmitting the radar detection results onto the camera picture to estimate target size, and leveraging the radar’s precise longitudinal distance and azimuth angle to lessen the vehicle width error introduced by image blurring. It allows for utilizing binocular vision and a traditional beam adjustment optimization synchronized point location and map-building technique. To address the challenge of real-time localization in low-light and dark environments, Zheng et al. [9] presented a novel approach based on the fusion of infrared vision and lidar for target identification and localization.
Some scholars have innovated with regard to the fusion method. Wang et al. [10] presented a synergistic fusion technique that combines a monocular camera with an MMW radar. The visual processing module processes the ROI after the MMW radar has detected the target and generated it to create a bounding box. The vehicle is then identified within the bounding box using the active contour method. Based on binocular vision sensors and the global navigation satellite system (GNSS), Yang [11] developed a multi-sensor fusion approach for attitude-tracking tasks and localization. Nabati R. [12] used the visual cone method to compare the radar detection results with the centroids of the candidate targets and then corrected the a priori detection results by regressing the candidate targets’ depth, rotation angle, and velocity. Zhou et al. [13] proposed a target-tracking technique based on multi-mode switching of onboard frequency-modulated continuous wave (FMCW) and MMW radar data fusion to improve tracking precision. They used real-world measurement data and simulated vehicle scenarios to test and validate the proposed technique, and the results demonstrate a significant improvement in tracking accuracy. Using CNNs to fuse different picture information, Zhou et al. [14] created a vehicle chassis recognition system based on multi-source sensor technologies, considerably boosting detection and classification accuracy. Meanwhile, Jia et al. [15] combined MMW radar and machine vision to detect front vehicles. To increase the system’s real-time performance and environmental flexibility, the authors used two types of cameras to capture and fuse front images: telephoto and short-focus. Additionally altered was the deep-learning algorithm’s candidate frame.
Other scholars have included drones in their research to explore new sensor-fusion techniques. Zhang et al. [16] proposed a quadrotor autonomous obstacle detection model to solve the monocular vision unmanned aerial vehicle (UAV) obstacle detection problem. The monocular estimation model offers depth information about the obstacle, whereas the target detection model delivers its position information. Hou et al. [17] developed and implemented a UAV system capable of autonomously identifying unknown external environments and automatically planning trajectories in real time to solve the problem of autonomous UAV navigation. Yang et al.’s [18] use of a multi-sensor fusion method to incorporate data from sensors such as machine vision, MMW radar, and a GPS navigation system improved the UAV’s capacity to avoid hazards such as cable poles and towers. They completed path planning and obstacle assessment using the virtual force field method (VFF).
For the second problem, some researchers use non-artificial neural network methods for sensor fault diagnosis. For example, Sharifi et al. [19] developed a probabilistic principal component analysis (MPPCA) hybrid model to identify single-sensor failure issues in nonlinear systems. Locally linear segments of the measurement space are coupled with a probabilistic principal component analysis (PPCA) model. The residual vector is built via a parity method using the transform related to each PPCA model, and Bayesian analysis is used to identify and isolate sensor failures. Meng et al. [20] suggested a method for determining whether a sensor is malfunctioning by examining the residuals between the model’s projected voltage and the sensor’s observed voltage. They used the unscented Kalman filter (UKF) algorithm to estimate the battery’s end voltage and monitor the residuals using the cumulative sum (CUSUM) approach to identify potential sensor faults based on their accumulated changes. Zhao et al. [21] introduced a time-series analysis-based fault-diagnosis method for aero-engine sensors. Their approach involves training an autoregressive moving average (ARMA) model with normal data offline and using it for real-time fault diagnosis online. Li et al. [22] offered a sensor fault-diagnosis method for acceleration sensors operating in harsh environments encountered in health monitoring systems. Their approach is based on weighted statistics of principal component analysis (PCA) residual space. Wang [23] et al. presented a generation approach based on integrating information-geometry causal inference (IGCS) and K2 score search strategy to improve the hydraulic condition monitoring system’s problem identification precision.
Others use artificial neural network methods for sensor fault diagnosis. Mariam et al. [24] used automated associative neural networks (AANN) to identify single and many sensor failures in a system. The model could validate sensor measurements via sensor error correction, missing data substitution, and noise filtering. Guo et al. [25] proposed a hybrid feature model combined with deep learning to identify UAV sensor flaws. The method utilizes the short-time Fourier transform (STFT) to transform residual signals of various sensor faults into corresponding time–frequency maps. The features of the images are then recovered using a convolutional neural network (CNN). Zhang et al. [26] developed a low-power multi-sensor vibration signal fault diagnostic technique (MLPC-CNN). To accurately extract grayscale image properties from multi-sensor data, they introduced a single sensor to single channel convolution (STSSC) approach. Next, use a mean pool layer-based bypass branching structure to preserve low-dimensional data while extracting high-level characteristics. Then, adding a multilayer pool classifier decreases the number of network parameters and removes the possibility of overfitting. Li et al. [27] collected time and frequency domain features and morphological information from multi-dimensional aero-engine sensor signals to represent the sensor’s health status and suggested an improved Henry gas solubility optimization approach for feature selection. Ultimately, they used the feature vector as the sensor’s health indicator to perform intelligent defect diagnoses using deep belief networks (DBN). Guo et al. [28] proposed a structural acceleration sensor fault self-diagnosis and fault signal self-recovery algorithm, combining CNN and deep convolutional generative adversarial networks (DCGAN). Zhang et al. [29] suggested a CGA-LSTM-based sensor failure detection technique that first used CNN to extract features from data, then integrated with the Long Short-Term Memory (LSTM) model, and last employed a Genetic algorithm (GA) to optimize the essential hyper-parameters in the LSTM network. Ma et al. [30] developed a fault-detection technique for multi-source sensors that can diagnose fixed deviation and drift deviation faults in complex systems. This method employs a CNN to extract features between different sensors and recurrent networks to describe the temporal properties of the sensors. Lin [31] suggested a hybrid approach for sensor fault diagnosis and fault data reconstruction based on enhanced LSTM and random forest (RF).
In conclusion, this research contributes in the following two ways:
(1)
We propose a space–time fusion algorithm to combine data from a single MMW radar with a single camera sensor;
(2)
To determine the effect of each failure mode on the sensor, we developed an information-fusion-based fault diagnosis method.
The rest of this paper is structured as follows: Section 2 briefly introduces the theory. Section 3 explains the proposed information fusion and fault diagnosis method. Section 4 shows the simulation process and the experimental results. Finally, Section 5 offers the conclusion.

2. The Theory

2.1. The Failure Criteria of the Autonomous Driving Perception System

Fault standards for autonomous driving perception systems mainly involve fault detection and diagnosis, and their standards are usually formulated based on system performance and reliability requirements. Failure criteria usually include the following aspects [32].
  • Fault classification: Classify possible faults for better fault diagnosis and repair.
  • Fault detection accuracy: Require and evaluate the accuracy of fault detection to ensure that the detected faults actually exist.
  • Fault diagnosis accuracy: Require and evaluate the accuracy of fault diagnosis to ensure that the determined cause and location of the fault are correct.
  • Fault response time: Require and evaluate the response time of fault repair to ensure timely repair when a fault occurs.
  • Fault tolerance: Require and evaluate the system’s tolerance to different types of faults to ensure that the system can still operate normally when some faults occur.
In this paper, we evaluate the performance and reliability of the autonomous driving perception system by analyzing the fault classification threshold, fault diagnosis accuracy, and time to fault detection.
The performance evaluation of automotive data fusion often requires summarizing aspects of perceptual performance into a small number of scalar values for comparison [33]. Computing lower-level metrics requires associating the estimated tracks of the System-Under-Test to their corresponding reference tracks, which is realized in the following way: the pairwise distances between all estimated and all reference objects or tracks are computed using an object distance function [34,35]. In this paper, we use the Euclidean distance τ of the observation point of the MMW radar and the center point of the camera sensor to associate them and then use the ratio of τ to the vehicle length W as the threshold for fault classification. The specific operation is given in Section 3.3.

2.2. Information Fusion

The general definition of information fusion can be roughly summarized as follows: information fusion is the study of efficient methods for automatically or semi-automatically transforming information from different sources and different points in time into a representation that provides effective support for human or automated decision making [36,37]. The traditional multi-source information fusion theory includes data-level information fusion, feature-level information fusion, and target-level data fusion [4].
This paper uses target-level data fusion, as shown in Figure 1: we perform feature extraction on the data from different sensors, process them according to the requirements of object detection, and ultimately output the discrimination result. Target-level data fusion is more fault-tolerant and real-time [38].

2.3. Caffe-Based CNN

A CNN is a multilayer network structure designed for image-recognition tasks. The network employs unsupervised learning for training and includes an input layer, an output layer, a convolutional layer, and a pooling layer [39]. The convolutional layer primarily utilizes a sampler to capture essential data content from the input data. The pooling layer’s objective is to reduce the size of the feature map, control overfitting, and shorten the training time by limiting the number of parameters and computations [40]. The fully connected layer transforms the two-dimensional feature map from the convolution output into a one-dimensional vector, then transmits the output values to the classifier for mapping to the sample label space [41]. To overcome the redundant parameters of the fully connected layer, high-performing network models such as ResNet and GoogLeNet employ global average pooling instead, with loss functions such as softmax as the network objective function [42,43], which fuses the learned deep features.
We selected the AlexNet network as the model for Caffe. It consists of eight layers, including five convolutional layers and three fully connected layers. Maximum pooling is applied after the first, second, and fifth convolutional layers. Figure 2 illustrates the network structure and output feature map size of each layer in the network.

2.4. Coordinate Calibration

The purpose of the coordinate calibration is to obtain the absolute position of the target [44]. The following describes the three coordinate systems involved in this paper, as well as the rotation matrix used to represent the relative direction between two space coordinates.

2.4.1. Coordinate System

  • Image coordinate system
There are two different kinds of coordinate systems for images: pixel coordinate systems and physical coordinate systems [45]. The original point of the pixel coordinate system is in the upper left corner of the image, as shown in the O u v of Figure 3, which represents the logical distance between targets in the picture; the original point of the physical coordinate system is in the center of the image, as shown in the O x y of Figure 3, which represents the real distance between targets in space.
Following are the steps involved in converting a physical coordinate system O x y to a pixel coordinate system O u v .
u v 1 = 1 / d x 0 u 0 0 1 / d y v 0 0 0 1 x y 1
2.
Camera coordinate system
The camera coordinate system is a three-dimensional spatial coordinate system, and the object on this coordinate system is inverted on the image coordinate system [46]. As illustrated in Figure 3, O c is the camera optical center; the Z c axis along the camera optical axis and the image plane is perpendicular to the direction of the image and is positive; the X c , Y c axis parallels the image physical coordinate system x , y axis; and O c ,   O x y is its focal length f . O c X c Y c Z c forms the camera coordinate system.
Following are the steps involved in converting a camera coordinate system O c to a physical image coordinate system   O x y .
Z c x y 1 = f 0 0 0 0 f 0 0 0 0 1 0 X c Y c Z c 1
3.
World coordinate system
The camera’s position and any other object in the environment relationship can be represented using the world coordinate system. Figure 3 depicts the link between the world coordinate system and various coordinate systems.
Following are the steps involved in converting a world coordinate system O w to a camera coordinate system O c .
X c Y c Z c 1 = R T 0 1 X w Y w Z w 1
In the above equation, R is a 3 × 3 orthogonal unit matrix (also called a rotation matrix), and T is a three-dimensional translation vector. The vector 0 = 0 , 0 , 0 .

2.4.2. Rotation Matrix

Two coordinate systems with the same origin can be rotated around their axes using a rotation matrix [47,48]. In a right-handed coordinate system, if the angles x ,   y , and z are rotated around the ψ ,   φ , and θ axes in turn, then the total rotation matrix R = R x ψ · R y φ · R z θ .
The active rotation around the x-axis (counterclockwise in the yz plane) is defined as
R x ψ = 1 0 0 0 cos ψ sin ψ 0 sin ψ cos ψ .
The active rotation around the y-axis (counterclockwise in the xz plane) is defined as
R y φ = cos φ 0 sin φ 0 1 0 sin φ 0 cos φ .
The active rotation around the z-axis (counterclockwise in the xy plane) is defined as
R z θ = cos θ sin θ 0 sin θ cos θ 0 0 0 1 .

3. The Proposed Method

The key issues to be addressed in this study are as follows:
(1)
Realize the information fusion of a single camera and a single MMW radar sensor;
(2)
Use information fusion from multiple sensors to diagnose sensor failures and evaluate the accuracy of fault diagnosis and its impact on autonomous driving safety.
As illustrated in Figure 4, the fault diagnosis procedure described in this research uses three major methods: a fault diagnostic method to pinpoint camera sensor failures, an information fusion algorithm to achieve the space–time fusion of two sensors, and CNN for vehicle recognition and labeling.
The fault diagnosis process of the perception system proposed in this paper is as follows:
(1)
Perform Prescan/Matlab co-simulation to obtain data output of a single camera sensor and a single MMW radar sensor;
(2)
Define and simulate the sensor failure. The MMW radar simulation provides a direct output of the target point and has a single failure mode. Therefore, we assumed that there were no failures. Only three failure modes are injected into the camera sensor;
(3)
Use the Caffe-based CNN (AlexNet) to identify and label vehicles in the images obtained in the first step;
(4)
Carry out joint calibration and data fusion of the MMW radar sensor and the camera sensor;
(5)
Study the fusion results under different failure modes;
  • The index: τ —the pixel Euclidean distance between the observation point of the MMW radar and the center point of the camera sensor; W—the width of the target vehicle, W = 1.6 m.
  • The strategy: if fusion fails, or τ 30 %   W , it will be regarded as a camera sensor failure, and an alarm will be issued; if the fusion succeeds and 10 %   W τ < 30 %   W , it will still be regarded as a camera sensor failure, but no alarm will be issued, and the system will only reduce the level of automatic driving; if the fusion succeeds and τ < 10 %   W , it will be regarded as a fusion error, with no alarm, and the system is safe.
(6)
Compare the fusion result with the threshold and draw a conclusion.

3.1. CNN-Based Vehicle Recognition Algorithm

Although R-CNN, Faster R-CNN, and YOLO perform well in detection accuracy, their training processes are multi-stage, and the training cost is relatively expensive in time and space [49,50,51]. This paper focuses on injecting faults into sensor output images and exploring the correlation between information fusion and fault diagnosis based on the diagnostic results, as well as their impact on autonomous driving safety. Additionally, the computational resources in our laboratory are limited. Therefore, we ultimately chose a Caffe-based CNN (AlexNet) with only a one-stage training process to explore a target detection method that saves on training costs.

3.1.1. Image Preprocessing

The data obtained by the camera are in the form of RGB images, which have a high pixel count that can adversely impact real-time processing. To facilitate image recognition, we converted these images into grayscale [52]. Gray image conversion is usually performed by weighting the three primary colors of red, green, and blue with a coefficient of 0.3, 0.59, and 0.11 [53]. These coefficients are determined based on the human luminance perception system’s adjustment.
V G a r y = 0.3 × V Re d + 0.59 × V G r e e n + 0.11 × V B l u e

3.1.2. CNN Based on the Caffe Framework

In this study, we added a classifier for vehicle target detection at the end of the neural network, used the AlxeNet network to perform image recognition and classification on the images obtained by Prescan simulation, and then used the trained classifier to detect vehicle targets. Most of the images generated by the PreScan simulation are used as our training set (“positive samples” containing vehicle images, “negative samples” not containing vehicle images), and the remaining images constitute the test set. After multi-scale training and Fisher feature extraction, we input the test set to the trained vehicle detection classifier for vehicle classification and labeling. The concept of this study is shown in Figure 5.

3.2. Information Fusion Algorithm Based on Joint Sensor Calibration

The purpose of this research is to investigate the information fusion method of a single camera and a single MMW radar in autonomous driving. This approach was chosen for several reasons:
  • Cameras and MMW wave radars are well-established sensing techniques in the field of autonomous driving, with proven success in detecting objects and providing valuable information about the environment [54,55];
  • The camera provides high-resolution texture information of the surrounding environment of the vehicle and exhibits excellent performance in terms of horizontal position, horizontal detection distance, and target classification ability, but the camera performs poorly in terms of the vertical detection distance [12]. The MMW radar is highly robust against various lighting conditions and is less susceptible to adverse weather conditions such as fog and rain compared to cameras [4]. However, classifying objects by radar is very challenging due to the low resolution of radar [56];
  • Integrating data from different sensors can significantly enhance the accuracy and reliability of autonomous driving systems [57,58].
In summary, the use of cameras and MMW radars for information fusion has great potential to enhance the accuracy and reliability of autonomous driving systems, making them an important focus of this study.
Although multiple cameras and MMW radar sensors are commonly used in self-driving cars, investigating autonomous driving technologies that utilize only a single camera and a single MMW radar to fuse information remains of practical interest. This approach can potentially reduce the cost of autonomous driving systems by decreasing the number of sensors, which can lower manufacturing costs and energy consumption. Furthermore, this approach can facilitate the development of simpler and more straightforward autonomous driving systems. Additionally, this approach can offer insight into the principles and methods of information fusion between cameras and MMW radar sensors, serving as a foundation for developing more complex autonomous driving systems.

3.2.1. MMW Radar, Camera Model, and Coordinate Selection

The origin of the vehicle coordinate system, denoted by the letters O W X W Y W Z W in this paper, is the central position of the rear axle at zero height above the ground. As seen in Figure 6, the vehicle coordinate system, which is blue, complies with the right-hand rule.
We utilized the ARS408 MMW radar, which has an actual refresh rate of approximately 20 Hz, a frame rate of 15 FPS, and a cycle period of roughly 67 ms. Figure 6 displays the MMW radar coordinate system in the green area, where the coordinate origin is at the center of the vehicle’s front bumper, and the coordinate plane is parallel to the vehicle’s coordinate plane.
Additionally, we employed the LI-USB30-AR023ZWDR camera, which has an actual refresh rate of 20 Hz, a frame rate of 20 FPS, and a cycle time of 50 ms. Figure 7 shows the camera coordinate system in the yellow area, with the coordinate origin behind the front windshield and the coordinate plane parallel to the vehicle’s coordinate plane.

3.2.2. Joint MMW Radar and Camera Calibration

The distance from the target item to the radar panel, the angle between the line connecting the target object to the radar panel and the normal radar panel, and the relative velocity to the radar are all output by conventional MMW radar, which can identify up to 64 detection targets. The output angle value is positive if the item is on the right side of the radar; otherwise, it is negative. In this study, we state that an MMW radar can produce up to 20 targets, and we represent the MMW radar’s three-dimensional coordinates as O r ( R r , θ r , V r ) , with the radar panel’s center point serving as the origin.
Since the MMW radar detection sweep plane is two-dimensional, this paper assumes that the radar plane normally is parallel to the ground and does not output z-axis direction information. The following is an expression for the conversion relationship between the world coordinate system O W and the radar coordinate system O R :
x r y r z r 1 = R z T 0 1 x w y w z w 1 .
We can substitute into Equation (6) and simplify as follows:
x r y r 1 = cos θ sin θ x 0 sin θ cos θ y 0 0 0 1 x w y w 1 .
θ represents the angle between the radar coordinate system and the world coordinate system, and x 0 , y 0 can be solved by selecting two sets of characteristic points and bringing them into Equation (10).
The conversion from the radar coordinate system O R to the camera coordinate system O C is as follows.
R r c t r c 0 1 = R w c t w c 0 1 R w r t w r 0 1 .
R w c and t w c denote the conversion matrix from the radar coordinate system to the camera coordinate system, and R w r and t w r denote the conversion matrix from the radar coordinate system to the world coordinate system.

3.2.3. Space–Time Fusion of MMW Radar and Camera

First, the spatial integration of the camera and MMW radar is accomplished. The MMW radar detection sweep plane is two-dimensional, so it is possible to think of the transformation of the radar coordinate system O R to the camera coordinate system O C as the transformation of the two-dimensional coordinate system. The procedure for converting radar data x r , y r to image coordinates u , v can be easily determined by consulting the literature [59].
u v 1 = T I R x r y r 1 = t 11 t 12 t 13 t 21 t 22 t 23 t 31 t 32 t 33 x r y r 1
By replacing n joint calibration points (where n < 3), the matrix least-squares method can be used to generate the coefficients for the transformation matrix T I R to be solved. Assuming that ( x r j , y r j ) and ( u j , v j ) are the jth pair of calibration points, the pending parameters can be written as follows [6,59]:
T I R = x r 1 x r n y r 1 y r 1 1 1 x r 1 x r n y r 1 y r 1 1 1 T 1 x r 1 x r n y r 1 y r 1 1 1 u 1 u n v 1 v n 1 1 T .
The MMW radar and camera, which are fused in time, come next. For this, data collected from the radar and camera must be coordinated. Because the frame rates of the radar and camera employed in this study are around 15 FPS and 20 FPS, respectively, we use the measurement time of the lower-frequency MMW radar as a benchmark to make the higher-frequency camera data backward-compatible. Since there is no variation among the images the camera took throughout its 50 ms acquisition period, we may instead utilize the image closest to the reference time to calibrate the MMW radar camera’s timing. This is displayed in Figure 8.

3.3. Fault Diagnosis Method Based on Information Fusion

Since there is only one failure mode for MMW radar, this research exclusively considers the failure mode of the camera sensor. To verify the suggested strategy, three camera sensor failure scenarios and one typical instance are defined, as demonstrated in Table 1 below.
In this study, the LI-USB30-AR023ZWDR camera had a frame rate of 30 FPS, allowing the rendering of 30 images or frames per second. The cycle period of this camera is 50 ms, representing the time between two consecutive photographs. In contrast, the ARS408 MMW radar’s frame rate is approximately 15 FPS, enabling it to process 15 images per second. This radar’s cycle period is 67 ms, denoting the interval between two successive images. Hence, it is apparent that there exists an approximate 25% difference between the camera and MMW radar’s processing time.
We have increased the range of permissible errors to consider the many operational situations that cameras and MMW radars may experience, including changes in lighting and unfavorable weather such as rain and snow. Specifically, we have defined the threshold for “failure” as a deviation equal to or greater than 30%, while the threshold for “no failure” is less than 10%. Deviations that fall between these two thresholds are categorized as “deviation failure”.
The following steps are used in the fault diagnosis based on the information fusion process: first, establish a pixel coordinate system to determine the Euclidean distance τ between observation point b x 2 , y 2 of MMW radar and the centroid a x 1 , y 1 of visual markers [33]. The centers of the two sensors are seen in Figure 9.
τ = d a b = x 1 x 2 2 + y 1 y 2 2
Then, comparing the pixel distance with the target width W of the body (the target width W is estimated to be 1.6 m based on engineering experience), the problem diagnostic process is illustrated in Figure 4.
Since the vision sensor output is set to 320 × 240 in the PreScan autonomous driving scene modeling, the pixel width/physical width should be adjusted to 1:64.

4. Simulations Experiments and Results Analysis

As illustrated in Figure 10, a simulation environment is designed for this paper using Prescan8.5.0/ MATLAB R2014a. The primary components are a driving vehicle module with a control system, straight and curving road conditions, roadside obstacle vehicles, buildings and trees, and weather lighting. All of the vehicles, with the exception of the primary vehicle, are parked on the side of the road and move at a speed of 5 m/s.
The motion parameters and the settings for the two sensors are set as shown in the following Table 2 and Table 3.
Accounting for the potential interference of weather on the output of MMW radar, we introduced Gaussian noise to the radar signal in the simulation environment. The chosen MMW radar sensor can directly output the target information, the frequency is 20 Hz, and the experiment time is 8 s. Figure 11 displays the 160 groups of detection results.
To preserve more information about targets, we configure the camera sensor to output a color image. The output data has time-series characteristics and is displayed as numerical matrixes with three levels, R, G, and B, as illustrated in Figure 12. Using the in-built conversion module of Matlab, we transform the numerical matrix generated by the camera sensor into an RGB image, as shown in Figure 13.

4.1. Vehicle Detection Based on Vision Sensors

According to the method in Section 3.1.1, the RGB images converted from the camera’s numerical matrix are grayed out, and the results are shown in Figure 14.
Figure 15 below displays the outcomes of utilizing a CNN to identify and label the vehicles in the photos.

4.2. Sensor Information Fusion Based on Joint Calibration

Target detection is carried out on this ROI by projecting the target point received by the MMW radar sensor onto the camera image and incorporating a priori knowledge. The exact procedure is as follows: import the output from the MMW radar and camera sensors, dynamically alter the external parameters, and assume this parameter is right if the rectangular box on the image is in the intended location. By fixing the external parameters, sensor information fusion based on joint calibration can be performed.
The output from the MMW radar and camera sensors appears in Figure 16 below, and the joint calibration is completed by modifying the external parameter (i.e., Euler angle), according to Figure 17, where the red dot denotes the center of the vehicle image labeling, and the green dot denotes the target point of radar recognition (default at the center of the vehicle).

4.3. Fault Simulation and Fault Identification Method Examination

Corresponding to Section 3.3, the failure modes of the vision sensor studied in this paper include missing row/column pixels, pixel displacement, and target color loss. The objective of this experiment is to independently evaluate the impact of each failure mode on image quality by administering them individually. Concurrent injection of multiple fault modes can interfere or overlap, posing challenges to determining the isolated contribution of each failure mode.

4.3.1. Missing Row/Column Pixel Failure

It is shown that some or all of the rows/columns are missing pixels, and the output data are missing and replaced by white pixel blocks in the image. The output image is 320 × 240 pixels so that 100 to 140 horizontal rows become white pixel blocks, as shown in Figure 18.
After taking ten photos from each of the three sets of fault samples, Table 4 displays the resulting Euclidean distance.
Figure 19 shows that the τ / W % typically falls between 34.11% and 99.84%, with a response time of 0.02 s to 1.6 s. According to these results, the deviations are between 30% and 100%, which means that our method can effectively and promptly identify the missing row/column pixel failure. However, some of the data fall within the [10,30] interval, contrary to what is expected. Because the fault injection is not ideal, the target vehicle is not adequately covered by the white pixel patches in these data images. The unexpected results indicate that the failure of missing row/column pixels can greatly compromise the safety of autonomous vehicles. This can lead to inaccurate or unrecognizable target recognition, prompting an alarm, reducing the autonomy level, or requiring the vehicle to stop.

4.3.2. Pixel Shift Fault

This issue is manifested as a blurred or distorted image, as shown in Figure 20.
Ten photos from each of the three sets of fault samples were taken, and Table 5 displays the resulting Euclidean distance.
Figure 21 shows that the τ / W % typically falls between 0.32% and 9.92%, with a response time of 0 s to 0.16 s. Based on the results, the deviations are between 0 and 10%, which deviates from the prediction. This is because this pixel shift fault is reproduced by increasing the target vehicle’s pixel noise without altering the target vehicle’s appearance or color, which does not influence picture identification. Hence, when such faults occur, the sensor data fusion remains stable to ensure safe autonomous driving, and the system function remains operational.

4.3.3. Target Color Loss Fault

The performance is a complete loss or assimilation of the color of a class of targets, with the actual output of part of the pixel block data being the same, as shown in Figure 22. The different color indicates objects color loss.
Ten photos from each of the three groups of fault samples were collected, and Table 6 displays the resulting Euclidean distance.
As seen in Figure 23, the τ / W % typically falls between 0.26% to 2.88%, with a response time of 0 s to 0.05 s. Based on the results, the deviations are between 0 and 3%, which is unexpected. This is because this fault is reproduced by changing the pixel color of the target vehicle without altering the target vehicle’s appearance, which does not influence picture identification. Hence, when such faults occur, the sensor data fusion remains stable to ensure safe autonomous driving, and the system function remains operational.

5. Conclusions

Based on the information fusion of the MMW radar and the camera, this research offers a fault diagnosis method for an autonomous driving sensing system. First, we generated an autonomous driving simulation scenario using PreScan and collected sensor data from a single MMW radar and a single camera. Then, we identified and labeled the automobiles in the photos using a CNN. Third, we suggested a sensor target-level information fusion approach based on the MMW radar. Finally, we proposed an information-fusion-based fault diagnosis strategy centered on the Euclidean distance between the MMW radar and the camera centroid. We injected three separate failure modes into the camera sensors in turn; the simulation results indicate that for missing row/column pixel failure, the deviation typically falls between 34.11% and 99.84%, with a response time of 0.02 s to 1.6 s; for pixel shift faults, the deviation range is between 0.32% and 9.92%, with a response time of 0 s to 0.16 s; target color loss faults have a deviation range of 0.26% to 2.88%, and a response time of 0 s to 0.05 s. This demonstrates that the suggested method may identify sensor faults to achieve real-time fault alerts and ensure the safety and dependability of an autonomous driving system.
Further research can expand on this study in several ways. First, to enhance the robustness and efficiency of the system in practical applications, it is crucial to consider incorporating additional sensors besides MMW radar and cameras. The number of different sensors required can vary depending on the task and scenario and must be carefully evaluated. Second, in real-world scenarios, multiple sensors may simultaneously experience various types of faults, which necessitates the development of methods for detecting and isolating mixed sensor faults. This remains a challenging task for future research.

Author Contributions

Conceptualization, W.L. and P.L.; methodology, W.L. and P.L.; software, W.L. and P.L.; validation, W.L. and P.L.; formal analysis, W.L. and P.L.; investigation, W.L. and P.L.; resources, W.L. and P.L.; data curation, W.L. and P.L.; writing—original draft preparation, W.L.; writing—review and editing, W.L.; visualization, W.L. and P.L.; supervision, W.H.; project administration, W.H.; funding acquisition, W.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author.

Acknowledgments

The authors would like to express their sincere gratitude to the academic editors and reviewers for their helpful comments and constructive suggestions. We express our gratitude to the third author of the paper, Pengyu Li, for purchasing the copyrights of Prescan8.5.0 and MATLAB R2014a for the purpose of our research in this study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cai, S. The Development of Automotive Intelligence under the 5G Technology. Automob. Parts 2020, 8, 106–108. [Google Scholar]
  2. Antonante, P.; Spivak, D.I.; Carlone, L. Monitoring and Diagnosability of Perception Systems. In Proceedings of the 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic, 27 September 2021; pp. 168–175. [Google Scholar]
  3. Zhou, W.; Lu, L.; Wang, J. Research Progress on Multi-Sensor Information Fusion in Unmanned Driving. Automot. Dig. 2022, 1, 45–51. [Google Scholar]
  4. Liu, Z.; Cai, Y.; Wang, H.; Chen, L.; Gao, H.; Jia, Y.; Li, Y. Robust Target Recognition and Tracking of Self-Driving Cars with Radar and Camera Information Fusion Under Severe Weather Conditions. IEEE Trans. Intell. Transp. Syst. 2022, 23, 6640–6653. [Google Scholar] [CrossRef]
  5. Goelles, T.; Schlager, B.; Muckenhuber, S. Fault Detection, Isolation, Identification and Recovery (FDIIR) Methods for Automotive Perception Sensors Including a Detailed Literature Survey for Lidar. Sensors 2020, 20, 3662. [Google Scholar] [CrossRef] [PubMed]
  6. Jiang, Q.; Zhang, L.; Meng, D. Target Detection Algorithm Based on MMW Radar and Camera Fusion. In Proceedings of the 2019 IEEE Intelligent Transportation Systems Conference (ITSC), Auckland, New Zealand, 27 October 2019; pp. 1–6. [Google Scholar]
  7. Chadwick, S.; Maddern, W.; Newman, P. Distant Vehicle Detection Using Radar and Vision. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 8311–8317. [Google Scholar]
  8. Wang, J.-G.; Chen, S.J.; Zhou, L.-B.; Wan, K.-W.; Yau, W.-Y. Vehicle Detection and Width Estimation in Rain by Fusing Radar and Vision. In Proceedings of the 2018 15th International Conference on Control, Automation, Robotics and Vision (ICARCV), Singapore, 18 November 2018; pp. 1063–1068. [Google Scholar]
  9. Zheng, X.; Lai, J.; Lyu, P.; Yuan, C.; Fan, W. Object Detection and Positioning Method Based on Infrared Vision/Lidar Fusion. Navig. Position. Timing 2021, 3, 34–41. [Google Scholar]
  10. Wang, X.; Xu, L.; Sun, H.; Xin, J.; Zheng, N. On-Road Vehicle Detection and Tracking Using MMW Radar and Monovision Fusion. IEEE Trans. Intell. Transp. Syst. 2016, 17, 2075–2084. [Google Scholar] [CrossRef]
  11. Yang, J. Pose Tracking and Path Planning for UAV Based on Multi-Sensor Fusion. Master’s Thesis, Zhejiang University, Hangzhou, China, 2019. [Google Scholar]
  12. Nabati, R.; Qi, H. Center Fusion: Center-Based Radar and Camera Fusion for 3D Object Detection. In Proceedings of the 2021 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2021; pp. 1526–1535. [Google Scholar]
  13. Zhou, X.; Zhang, Y. Vehicle radar data fusion target tracking algorithm with multi-mode switching. Laser J. 2021, 11, 53–59. [Google Scholar]
  14. Zhao, G.; Li, J.; Ma, L. Design and implementation of vehicle trajectory perception with multi-sensor information fusion. Electron. Des. Eng. 2022, 1, 1–7. [Google Scholar]
  15. Jia, P.; Liu, Q.; Peng, K.; Li, Z.; Wang, Q.; Hua, Y. Front vehicle detection based on multi-sensor information fusion. Infrared Laser Eng. 2022, 6, 498–505. [Google Scholar]
  16. Zhang, X.; Zhang, L.; Song, Y.; Pei, H. Obstacle Avoidance Algorithm for Unmanned Aerial Vehicle Vision Based on Deep Learning. J. South China Univ. Technol. Nat. Sci. Ed. 2022, 50, 101–108+131. [Google Scholar]
  17. Hou, Y.; Liu, Y.; Lu, H.; Wu, Q.; Zhao, J.; Chen, Y. An Autonomous Navigation Systems of UAVs Based on Binocular Vision. J. Tianjin Univ. Sci. Technol. 2019, 52, 1262–1269. [Google Scholar]
  18. Yang, L.; Chen, F.; Chen, K.; Liu, S. Research and Application of Obstacle Avoidance Method Based on Multi-sensor for UAV. Comput. Meas. Control 2019, 1, 280–283+287. [Google Scholar]
  19. Sharifi, R.; Langari, R. Nonlinear Sensor Fault Diagnosis Using Mixture of Probabilistic PCA Models. Mech. Syst. Signal Process. 2017, 85, 638–650. [Google Scholar] [CrossRef]
  20. Meng, D.; Shu, Q.; Wang, J.; Wang, Y.; Ma, Z. Fault Diagnosis for Sensors in Automotive Li-Ion Power Battery Based on SVD-UKF. Chin. J. Automot. Eng. 2022, 4, 528–537. [Google Scholar]
  21. Zhao, W.; Guo, Y.; Yang, J.; Sun, H. Aero-engine Sensor Fault Diagnosis and Real-time Verification Based on ARMA Model. Aeronaut. Comput. Technol. 2022, 1, 16–20. [Google Scholar]
  22. Li, L.; Liu, G.; Zhang, L.; Li, Q. Accelerometer Fault Diagnosis with Weighted PCA Residual Space. J. Vib. Meas. Diagn. 2021, 5, 1007–1013+1039–1040. [Google Scholar]
  23. Wang, J.; Zhang, Y.; Cen, G. Fault diagnosis method of hydraulic condition monitoring system based on information entropy. Comput. Eng. Des. 2021, 8, 2257–2264. [Google Scholar]
  24. Elnour, M.; Meskin, N.; Al-Naemi, M. Sensor Data Validation and Fault Diagnosis Using Auto-Associative Neural Network for HVAC Systems. J. Build. Eng. 2020, 27, 100935. [Google Scholar] [CrossRef]
  25. Guo, D.; Zhong, M.; Ji, H.; Liu, Y.; Yang, R. A Hybrid Feature Model and Deep Learning Based Fault Diagnosis for Unmanned Aerial Vehicle Sensors. Neurocomputing 2018, 319, 155–163. [Google Scholar] [CrossRef]
  26. Zhang, Y.; He, L.; Cheng, G. MLPC-CNN: A Multi-Sensor Vibration Signal Fault Diagnosis Method under Less Computing Resources. Measurement 2022, 188, 110407. [Google Scholar] [CrossRef]
  27. Li, H.; Gou, L.; Chen, Y.; Li, H. Fault Diagnosis of Aeroengine Control System Sensor Based on Optimized and Fused Multidomain Feature. IEEE Access 2022, 10, 96967–96983. [Google Scholar] [CrossRef]
  28. Guo, X.; Luo, Y.; Wang, L.; Liu, J.; Liao, F.; You, D. Fault self-diagnosis of structural vibration monitoring sensor and monitoring data recovery based on CNN and DCGAN. J. Railw. Sci. Eng. 2022. accepted. [Google Scholar]
  29. Zhang, S.; Zhang, T. Sensor fault diagnosis method based on CGA-LSTM. In Proceedings of the 13th China Satellite Navigation Annual Conference, Beijing, China, 25 May 2022. [Google Scholar]
  30. Ma, L.; Guo, J.; Wang, S.; Wang, J. Multi-Source Sensor Fault Diagnosis Method Based on Improved CNN-GRU Network. Trans. Beijing Inst. Technol. 2021, 12, 1245–1252. [Google Scholar]
  31. Lin, T.; Zhang, D.; Wang, J. Sensor fault diagnosis and data reconstruction based on improved LSTM-RF algorithm. Comput. Eng. Sci. 2021, 5, 845–852. [Google Scholar]
  32. Realpe, M.; Vintimilla, B.X.; Vlacic, L. A Fault Tolerant Perception System for Autonomous Vehicles. In Proceedings of the 2016 35th Chinese Control Conference (CCC), Chengdu, China, 27–29 July 2016; pp. 6531–6536. [Google Scholar]
  33. Tan, F.; Liu, W.; Huang, L.; Zhai, C. Object Re-Identification Algorithm Based on Weighted Euclidean Distance Metric. J. South China Univ. Technol. Nat. Sci. Ed. 2015, 9, 88–94. [Google Scholar]
  34. Hoss, M.; Scholtes, M.; Eckstein, L. A Review of Testing Object-Based Environment Perception for Safe Automated Driving. Automot. Innov. 2022, 5, 223–250. [Google Scholar] [CrossRef]
  35. Delecki, H.; Itkina, M.; Lange, B.; Senanayake, R.; Kochenderfer, M.J. How Do We Fail? Stress Testing Perception in Autonomous Vehicles. In Proceedings of the 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Kyoto, Japan, 23–27 October 2022; pp. 5139–5146. [Google Scholar]
  36. Wang, G.; Hall, D.L.; McMullen, S.A. Mathematical Techniques in Multisensor Data Fusion. Biomed. Eng. 1995, 13, 11–23. [Google Scholar]
  37. Khaleghi, B.; Khamis, A.; Karray, F.O.; Razavi, S.N. Multisensor Data Fusion: A Review of the State-of-the-Art. Inf. Fusion 2013, 14, 28–44. [Google Scholar] [CrossRef]
  38. Wei, Z.; Zhang, F.; Chang, S.; Liu, Y.; Wu, H.; Feng, Z. MmWave Radar and Vision Fusion for Object Detection in Autonomous Driving: A Review. Sensors 2022, 22, 2542. [Google Scholar] [CrossRef]
  39. Tao, X.; Li, K.; Liu, Y. License Plate Recognition Algorithm Based on Deep Learning Model LeNet-5-L. Comput. Meas. Control 2021, 6, 181–187. [Google Scholar]
  40. Wanda, P.; Jie, H. RunPool: A Dynamic Pooling Layer for Convolution Neural Network. Int. J. Comput. Intell. Syst. 2020, 13, 23–30. [Google Scholar]
  41. Buffoni, L.; Civitelli, E.; Giambagli, L.; Chicchi, L.; Fanelli, D. Spectral pruning of fully connected layers. Sci. Rep. 2022, 12, 11201. [Google Scholar] [CrossRef]
  42. Panda, M.K.; Sharma, A.; Bajpai, V.; Subudhi, B.N.; Thangaraj, V.; Jakhetiya, V. Encoder and Decoder Network with ResNet-50 and Global Average Feature Pooling for Local Change Detection. Comput. Vis. Image Underst. 2022, 222, 103501. [Google Scholar] [CrossRef]
  43. Tang, P.; Wang, H.; Kwong, S. G-MS2F: GoogLeNet Based Multi-Stage Feature Fusion of Deep CNN for Scene Recognition. Neurocomputing 2017, 225, 188–197. [Google Scholar] [CrossRef]
  44. Yong, H.; Daqing, F.; Fuliang, T.; Minglu, T.; Daoguang, S.; Yang, S. Research on Vehicle Trajectory Tracking Control in Expressway Maintenance Work Area Based on Coordinate Calibration. IOP Conf. Ser. Earth Environ. Sci. 2020, 619, 012096. [Google Scholar] [CrossRef]
  45. Lyu, D. Research and Implementation of Infrared Binocular Camera Calibration Method. Master’s Thesis, Dalian University of Technology, Dalian, China, 2020. [Google Scholar]
  46. Hua, J.; Zeng, L. Hand–Eye Calibration Algorithm Based on an Optimized Neural Network. Actuators 2021, 10, 85. [Google Scholar] [CrossRef]
  47. Zhu, C.; Jun, L.; Huang, B.; Su, Y.; Zheng, Y. Trajectory Tracking Control for Autonomous Underwater Vehicle Based on Rotation Matrix Attitude Representation. Ocean Eng. 2022, 252, 111206. [Google Scholar] [CrossRef]
  48. Verde-Star, L. Construction of Generalized Rotations and Quasi-Orthogonal Matrices. Spec. Matrices 2019, 7, 107–113. [Google Scholar] [CrossRef]
  49. Girshick, R. Fast R-CNN. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  50. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. In Proceedings of the Advances in Neural Information Processing Systems; Curran Associates, Inc.: Red Hook, NY, USA, 2015; Volume 28. [Google Scholar]
  51. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 779–788. [Google Scholar]
  52. He, T.; Zhang, J. Effect of Color Weight Balance on Visual Aesthetics Based on Gray-Scale Algorithm. In Proceedings of the Advances in Neuroergonomics and Cognitive Engineering; Ayaz, H., Asgher, U., Paletta, L., Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 328–336. [Google Scholar]
  53. Xuan, H.; Liu, H.; Yuan, J.; Li, Q. Robust Lane-Mark Extraction for Autonomous Driving Under Complex Real Conditions. IEEE Access 2018, 6, 5749–5765. [Google Scholar] [CrossRef]
  54. Caesar, H.; Bankiti, V.; Lang, A.H.; Vora, S.; Liong, V.E.; Xu, Q.; Krishnan, A.; Pan, Y.; Baldan, G.; Beijbom, O. NuScenes: A Multimodal Dataset for Autonomous Driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11621–11631. [Google Scholar]
  55. Sun, P.; Kretzschmar, H.; Dotiwalla, X.; Chouard, A.; Patnaik, V.; Tsui, P.; Guo, J.; Zhou, Y.; Chai, Y.; Caine, B.; et al. Scalability in Perception for Autonomous Driving: Waymo Open Dataset. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2446–2454. [Google Scholar]
  56. Feng, D.; Haase-Schütz, C.; Rosenbaum, L.; Hertlein, H.; Gläser, C.; Timm, F.; Wiesbeck, W.; Dietmayer, K. Deep Multi-Modal Object Detection and Semantic Segmentation for Autonomous Driving: Datasets, Methods, and Challenges. IEEE Trans. Intell. Transp. Syst. 2021, 22, 1341–1360. [Google Scholar] [CrossRef]
  57. Fayyad, J.; Jaradat, M.A.; Gruyer, D.; Najjaran, H. Deep Learning Sensor Fusion for Autonomous Vehicle Perception and Localization: A Review. Sensors 2020, 20, 4220. [Google Scholar] [CrossRef] [PubMed]
  58. Xiao, L.; Wang, R.; Dai, B.; Fang, Y.; Liu, D.; Wu, T. Hybrid Conditional Random Field Based Camera-LIDAR Fusion for Road Detection. Inf. Sci. 2018, 432, 543–558. [Google Scholar] [CrossRef]
  59. Wang, T.; Zheng, N.; Xin, J.; Ma, Z. Integrating MMW Radar with a Monocular Vision Sensor for On-Road Obstacle Detection Applications. Sensors 2011, 11, 8992–9008. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Target-level data fusion.
Figure 1. Target-level data fusion.
Sensors 23 05110 g001
Figure 2. AlexNet network structure.
Figure 2. AlexNet network structure.
Sensors 23 05110 g002
Figure 3. Image coordinate systems, camera coordinate systems, and world coordinate systems.
Figure 3. Image coordinate systems, camera coordinate systems, and world coordinate systems.
Sensors 23 05110 g003
Figure 4. Flowchart of fault diagnosis based on information fusion.
Figure 4. Flowchart of fault diagnosis based on information fusion.
Sensors 23 05110 g004
Figure 5. The CNN-based research framework for vehicle recognition.
Figure 5. The CNN-based research framework for vehicle recognition.
Sensors 23 05110 g005
Figure 6. Radar model installation position. (a) The length of the vehicle. (b) The width of the vehicle. (c) The height of the vehicle.
Figure 6. Radar model installation position. (a) The length of the vehicle. (b) The width of the vehicle. (c) The height of the vehicle.
Sensors 23 05110 g006
Figure 7. Camera model installation position. (a) The length of the vehicle. (b) The width of the vehicle. (c) The height of the vehicle.
Figure 7. Camera model installation position. (a) The length of the vehicle. (b) The width of the vehicle. (c) The height of the vehicle.
Sensors 23 05110 g007
Figure 8. Temporal fusion between MMW radar and camera.
Figure 8. Temporal fusion between MMW radar and camera.
Sensors 23 05110 g008
Figure 9. Two types of sensors labeled with center points.
Figure 9. Two types of sensors labeled with center points.
Sensors 23 05110 g009
Figure 10. The simulation environment.
Figure 10. The simulation environment.
Sensors 23 05110 g010
Figure 11. Numerical matrix of the MMW radar.
Figure 11. Numerical matrix of the MMW radar.
Sensors 23 05110 g011
Figure 12. Numerical matrix of the camera output layer (B).
Figure 12. Numerical matrix of the camera output layer (B).
Sensors 23 05110 g012
Figure 13. RGB images converted from camera numerical matrix.
Figure 13. RGB images converted from camera numerical matrix.
Sensors 23 05110 g013
Figure 14. Sensor output results after grayscale processing.
Figure 14. Sensor output results after grayscale processing.
Sensors 23 05110 g014
Figure 15. CNN identifies and marks vehicles. The red boxes indicate the objects marked by CNN.
Figure 15. CNN identifies and marks vehicles. The red boxes indicate the objects marked by CNN.
Sensors 23 05110 g015
Figure 16. Program running interface.
Figure 16. Program running interface.
Sensors 23 05110 g016
Figure 17. Joint calibration result. The red boxes indicate the vehicles marked by CNN. The green circle indicates the millimeter-wave radar observation point (default at the center of the vehicle). The red circle indicates the center point of the camera annotation (the center point of the visual annotation).
Figure 17. Joint calibration result. The red boxes indicate the vehicles marked by CNN. The green circle indicates the millimeter-wave radar observation point (default at the center of the vehicle). The red circle indicates the center point of the camera annotation (the center point of the visual annotation).
Sensors 23 05110 g017
Figure 18. Missing row/column pixel fault characterization.
Figure 18. Missing row/column pixel fault characterization.
Sensors 23 05110 g018
Figure 19. (a) The resulting statistics of missing row/column pixel fault identification. (b) The curve of missing row/column pixel fault response time.
Figure 19. (a) The resulting statistics of missing row/column pixel fault identification. (b) The curve of missing row/column pixel fault response time.
Sensors 23 05110 g019
Figure 20. Pixel displacement fault characterization.
Figure 20. Pixel displacement fault characterization.
Sensors 23 05110 g020
Figure 21. (a) The resulting statistics of pixel displacement fault identification. (b) The curve of pixel displacement fault response time.
Figure 21. (a) The resulting statistics of pixel displacement fault identification. (b) The curve of pixel displacement fault response time.
Sensors 23 05110 g021
Figure 22. Target color loss fault characterization.
Figure 22. Target color loss fault characterization.
Sensors 23 05110 g022
Figure 23. (a) The resulting statistics of target color loss fault identification. (b) The curve of target color loss fault response time.
Figure 23. (a) The resulting statistics of target color loss fault identification. (b) The curve of target color loss fault response time.
Sensors 23 05110 g023
Table 1. Failure modes and characteristics.
Table 1. Failure modes and characteristics.
Failure TypeFailure ModeFault CharacterizationExpected Test Results
No fault τ < 10 %   W
Failure faultMissing row/column pixelsMissing pixels are white τ 30 %   W
Deviation faultPixel displacementBlurred images 10 %   W τ < 30 %   W
Deviation faultTarget color lossSingle image color 10 %   W τ < 30 %   W
τ : the Euclidean distance between the MMW radar’s observation point and the camera’s center point, namely, the error between the center points of the two sensor tags. W: the target width of vehicle, W = 1.6 m.
Table 2. Motion parameter setting table.
Table 2. Motion parameter setting table.
ParameterValue
Speed (m/s)5.00
Roll friction coefficient0.01
Drag coefficient0.31
Mass (kg)1471
Reference area (m3)2.74
Air density (kg/m2)1.28
Gravitation (m/s2)9.81
Max acceleration (g)0.30
Max deceleration (g)0.30
Table 3. Sensor settings table.
Table 3. Sensor settings table.
CameraMMW Radar
LocationXYZLocationXYZ
(m)2.0000.0001.320(m)3.9400.0000.370
OrientationBankPitchHeadingOrientationBankPitchHeading
(deg)0.00.00.0(deg)0.00.00.0
ResolutionHorizontalVerticalLine scan directionLeft to Right/Top to Bottom
(pixel)320240
Frame rate (Hz)20Resulting scan frequency per beam (Hz)20
Intensity factor (RGB)0.300.590.11Beam center line orientation (azimuth)(deg)From −22.5 to 22.5
Operating frequency (GHz)25.000
Max. objects to output7
Table 4. Missing row/column pixel fault identification results.
Table 4. Missing row/column pixel fault identification results.
12345678910
τ 1 m 0.921.390.751.570.100.050.161.431.010.24
τ 1 / W % 57.2586.6447.0397.876.053.2110.2289.2162.8415.30
τ 2 m 1.531.550.931.600.070.040.201.331.110.33
τ 2 / W % 95.4796.8457.9199.844.662.5812.3883.4069.2620.66
τ 3 m 1.531.270.551.520.020.050.130.831.430.29
τ 3 / W % 95.4779.5334.1195.231.232.937.9952.1789.2218.42
Table 5. The outcomes of identifying pixel displacement faults.
Table 5. The outcomes of identifying pixel displacement faults.
12345678910
τ 1 m 0.040.100.130.010.160.140.130.080.010.10
τ 1 / W % 2.776.078.100.329.929.028.395.130.635.97
τ 2 m 0.050.080.040.090.140.030.050.120.160.02
τ 2 / W % 2.824.882.495.448.921.603.297.519.911.39
τ 3 m 0.090.020.020.040.120.100.000.090.120.01
τ 3 / W % 5.341.271.402.437.286.420.025.877.470.83
Table 6. Target color loss fault identification results.
Table 6. Target color loss fault identification results.
12345678910
τ 1 m 0.020.010.030.030.030.020.020.050.020.04
τ 1 / W % 1.230.841.971.622.100.971.452.881.002.55
τ 2 m 0.030.020.040.010.040.040.000.010.020.01
τ 2 / W % 1.771.352.330.712.432.480.260.581.190.37
τ 3 m 0.040.000.030.030.050.020.020.030.030.04
τ 3 / W % 2.370.181.891.862.831.261.421.712.042.25
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hou, W.; Li, W.; Li, P. Fault Diagnosis of the Autonomous Driving Perception System Based on Information Fusion. Sensors 2023, 23, 5110. https://doi.org/10.3390/s23115110

AMA Style

Hou W, Li W, Li P. Fault Diagnosis of the Autonomous Driving Perception System Based on Information Fusion. Sensors. 2023; 23(11):5110. https://doi.org/10.3390/s23115110

Chicago/Turabian Style

Hou, Wenkui, Wanyu Li, and Pengyu Li. 2023. "Fault Diagnosis of the Autonomous Driving Perception System Based on Information Fusion" Sensors 23, no. 11: 5110. https://doi.org/10.3390/s23115110

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop