1. Introduction
As an emerging aircraft, unmanned aerial vehicles (UAVs) have been widely used in military, civilian, and commercial fields due to their flexibility, efficiency, and safety [
1]. As the key components of UAV systems, UAV sensors are mainly used to measure real-time flight state parameters and provide feedback to the flight control system. The accuracy of sensor measurements is an essential prerequisite for ensuring the safety of UAVs [
2]. However, UAVs are susceptible to external interferences, such as wind, magnetic fields, vibrations, etc. Several UAV crashes were related to sensor failures [
3]. The aforementioned aspects can cause severe challenges for UAV sensors during complex flight environments. Natural disasters such as hurricanes or lightning can cause physical damage to the UAV, disabling measurements by particular sensors, and the sudden appearance of obstacles can impair the sensor’s ability to record the data. In addition, a rapid change in altitude can also affect the ability of the sensors to measure different data [
4]. Effective detection of sensor failures can ensure the safety and reliability of UAV flights, avoiding potential losses. Therefore, finding a method for accurately diagnosing UAV sensor faults holds significant practical significance. Conventional statistical threshold-based UAV sensor fault diagnosis suffers from two main limitations. They are:
(1) For the conventional statistical threshold value-based fault diagnosis approach, the threshold value corresponding to a specific fault for a specific sensor is different. Since UAVs are highly coupled electromechanical systems typically equipped with multiple heterogeneous sensors, designing efficient and consistent thresholds for each of the sensors corresponding to any particular type of fault is an inefficient approach during large-scale application.
(2) During real-time application, different UAVs are expected to work under different operating conditions. Hence, it is expected that sensor data collected from different UAVs will vary in terms of different data properties such as magnitude and strength. Therefore, designing the threshold largely depends on historical data properties, and sensor data collected from different UAVs under any particular operating condition are required to be logged properly and stored in a data storage device. This requires extra effort in data management and computer storage.
As a result, intelligent diagnosis of UAV sensors has received much attention in recent years. Existing methods for intelligent UAV sensor fault diagnosis can be categorized into three main approaches: model-based, data-driven, and knowledge-based methods. Model-based methods are commonly used in traditional fault diagnosis, including parameter estimation, state estimation, and the equivalent space method. The main idea is to establish precise mathematical models and employ control theory to obtain residual signals [
5]. In this context, the Kalman filter (KF) and its improvements, such as the extended Kalman filter (EKF), can play significant roles by calculating the best possible output from the measured data with the predicted data through a feedback system [
6,
7]. However, in real-life applications, UAVs operate in nonlinear situations, which significantly affect the performance of the Kalman filter in correcting the noisy measurement output from the related sensors. A multi-sensor navigation unit provides the flight state feedback required. Some researchers apply signal processing techniques to extract signal features and identify system faults. Chen et al. [
8] conducted fault diagnosis using acoustic signals from UAV motors. Features were extracted from the sound signals using statistical equations and then classified using a decision tree (DT), support vector machine (SVM), and K-nearest neighbor (KNN) algorithm. Since this study was based on acoustic signals, the experiments needed to be conducted indoors, making the experimental results susceptible to environmental interference. Some researchers have conducted health monitoring of the UAV by observing the vibration spectrum of the aircraft body and the vibration data were processed utilizing a fast Fourier transform (FFT) [
8]. However, model-based methods rely on extensive domain knowledge and experience, making them less adaptable to different fault scenarios. Moreover, the variety and complexity of sensor types in UAVs pose challenges in obtaining precise system models.
Data-driven methods avoid cumbersome high-fidelity system identification and validation and have strong nonlinear mapping capability. Several scholars have conducted fault diagnosis research on mechanical structures such as bearings and gears, leveraging the advantages of neural networks in feature extraction. Liu et al. [
9] proposed a novel bearing fault diagnosis method based on CNN networks that directly process time-domain raw signals, eliminating the need for time-consuming feature extraction and reducing dependence on expert experience. Mao et al. [
10] performed fault diagnosis based on a short-time Fourier transform and convolutional neural network for rolling bearing vibration signals, achieving end-to-end fault pattern recognition. The fault diagnosis cases above for mechanical structures typically use only a single sensor signal type, such as vibration or current signals. However, UAVs are highly coupled electromechanical systems typically equipped with multiple heterogeneous sensors, exhibiting nonlinearity and time-varying characteristics. Model-based and conventional neural network methods struggle to extract and fuse features from multi-sensor data. Based on the aforementioned limitations of model-based methods and data-driven methods, two main objectives of this research are:
- (1)
Incorporating the prior knowledge regarding UAV sensors under operation for fault diagnosis by utilizing the knowledge graph-based adaptive network.
- (2)
Integrating the information collected by a heterogenous sensing system in a UAV for achieving robust diagnosis performance.
A graph neural network (GNN) can effectively address this challenge as an emerging knowledge-based neural network model. The graph convolutional neural network (GCN) method, in particular, can easily integrate information from multiple sensors by introducing the association graph in non-Euclidean space, possessing advantages such as solid adaptability, robustness, and high accuracy [
11]. Some scholars have researched the GCN and applied it to fault diagnosis and remaining useful life prediction. Tama et al. [
12] converted the time-frequency features of original vibration signals into graphs, fed into the GCN to diagnose faults in the gearbox of wind turbines. Sadid et al. [
13] developed a spatiotemporal graph convolutional network for fault diagnosis of unmanned vehicles, incorporating the mathematical model of the vehicles to construct the adjacency matrix. He et al. [
14] proposed a graph attention network model suitable for fault diagnosis of UAVs. The masked spatial graph attention (Masked-SGAT) module aggregates spatial information, and the gate recurrent unit (GRU) module is employed to extract temporal features.
The key issue in fault diagnosis based on a GCN is establishing the relationship between sensors in UAVs. Conventional GCN methods typically rely on the similarity of variable features or model learning to determine the graph, which fails to ensure the accuracy of the graph. The concept of spatiotemporal fault detection and diagnosis can play a key role in improving the performance of the original GCN algorithm [
15,
16,
17]. To address this issue, prior knowledge related to engineering equipment under maintenance can ensure the accuracy of associations and the robustness of the model [
18]. In this context, the mathematical model of a quadrotor is introduced, and the graph of each sensor variable is established based on the proposed STDGCN model. Then, a spatiotemporal difference graph convolutional network (STDGCN) is constructed, which includes a differential layer to enhance input data features using local differences and introduces a spatiotemporal graph convolution module to extract time and spatial correlations between sensors. Unlike conventional model-driven and data-driven techniques, an STDGCN can utilize the relationship among the data collected from different types of sensors to diagnose UAV sensor faults. Comparative experiments are conducted against several existing fault diagnosis methods, demonstrating the superiority of the STDGCN model. The main contributions of this paper are summarized as follows:
- (1)
A sensor fault diagnosis method for UAVs based on graph neural networks is proposed, addressing the challenge of diverse sensor types and the difficulty in obtaining precise system models.
- (2)
UAV dynamics models are introduced as prior knowledge to construct the association graph, ensuring the accuracy of associations and the robustness of the network model.
- (3)
The STDGCN model is constructed, which extracts both temporal and spatial features from sensor data, overcoming the difficulties involved in traditional methods in extracting and integrating features from multiple sensors.
The remaining chapters of this paper are arranged as follows.
Section 2 introduces the theory of a GCN, providing the formula and derivations for graph convolutional operations.
Section 3 describes the proposed association of the graph construction method and the constructed STDGCN in detail. A quadrotor UAV is employed to generate actual flight data in
Section 4, and six categories of sensor fault data are generated based on the flight data. After that, the effectiveness and superiority of our proposal are validated by ablation studies and comparative experiments in
Section 5. Finally,
Section 6 summarizes this paper and gives directions for future work.
2. Preliminary Theory of a Graph Convolutional Network (GCN)
Construction of the association graph: A graph
can be defined by a set of nodes
and edges
[
18,
19]. The relationship between nodes
and
is represented by an edge
. An adjacency matrix
is constructed to facilitate information aggregation in the graph structure, where
if the edge
exists, and
otherwise.
Forward propagation of the GCN: According to the convolution theorem, the Fourier transform of the convolution of two signals is equivalent to the pointwise multiplication of their respective Fourier transforms. Let
denote the convolution operation in the spatial domain, which
represents a dataset containing n data points and
are the trainable parameters in the neural network. This operation can be transformed into the frequency domain using the Fourier transform [
14,
20].
where
F represents the Fourier transform. By applying the inverse Fourier transform
to both sides of Equation (1), the convolution operation
in the spatial domain can be expressed as:
where
represents the Fourier basis and
denotes element-wise multiplication. The GCN is designed to incorporate the association graph into neural networks. To achieve this, the GCN utilizes the Laplacian matrix of the graph to obtain the Fourier basis. Suppose
is the Laplacian matrix of a graph. It can be normalized as
, where
represents a unit matrix and
is the adjacent matrix [
21].
denotes the degree matrix,
. Then, the Fourier basis,
, along with the eigenvalue matrix
, can be obtained by the eigenvalue decomposition:
Based on the properties of the Laplacian matrix, the set
comprises orthogonal matrices that conform to the mathematical requirements of Fourier transformation. Let
be the diagonal matrix,
. Then, Equation (2) can be simplified as follows:
The eigenvalue decomposition of the Laplacian matrix is a crucial step in the graph convolution process. However, when the graph scale is large, the computational complexity grows quadratically with the number of nodes. The high cost of eigenvalue decomposition restricts the applicability of graph convolution algorithms primarily to small-scale graphs. To address this issue, Krizhevsky et al. [
19] proposed an approximation of
using Chebyshev polynomials
, which can be expressed as follows:
where
is the Chebyshev coefficient and
denotes the
-th term in the Chebyshev polynomial. Specifically, it can be defined as
,
, and
.
is a diagonal matrix of scale eigenvalues. Thus, (4) can be expressed as:
where
and
denote the maximum eigenvalue of the Laplacian matrix. Using Chebyshev polynomial approximation allows aggregating information from the 0-th to the
-th order neighborhood nodes using the convolutional kernel, effectively capturing the local information around the central node.
Xiao et al. [
22] further simplified the Chebyshev polynomials by setting
and
, which means that only the information from the first-order neighborhood of the central node is aggregated. Consequently, (6) can be simplified as follows:
By setting the parameter
, (7) can be further obtained:
Furthermore, to facilitate the network’s training through backpropagation, the parameters “
,
” are usually renormalized by
and
, respectively. Finally, the convolution operation in the spectral domain can be defined as:
A schematic diagram of the original GCN algorithm is shown in
Figure 1 as follows.
5. Discussion
5.1. Comparative Analysis of Models
The performance of the STDGCN method was assessed and evaluated through a cross-validation test of three additional flight experiments to obtain spatiotemporal Dataset-1, Dataset_2, and Dataset_3, with sample sizes of 456 and 276, respectively. The performance of the STDGCN is being tested multiple times with the same parameter configuration on different datasets. The purpose of these tests was to confirm the effectiveness of the STDGCN method in various scenarios. Subsequently, six existing fault diagnosis methods are introduced for comparison, including deep neural network (DNN) [
31], recurrent neural network (RNN) [
24], gated recurrent unit (GRU) [
32], long short-term memory network (LSTM) [
33], convolutional neural network (CNN) [
34], and LeNet architecture [
35]. This paper’s constructed LSTM and GRU models consist of two hidden layers, each with 320 units, time steps, and input dimensions of 100 and 20, respectively, as well as fault classification through fully connected layers. The CNN model constructed in this paper comprises a 2D convolutional layer, a max-pooling layer, and three fully connected layers. The LeNet architecture is also incorporated to validate further convolutional neural networks’ effectiveness in sensor fault diagnosis. The LeNet model includes two 2D convolutional layers, two activation layers, two max-pooling layers, and three fully connected layers. The final test accuracy is shown in
Figure 14.
Overall, the STDGCN performs better than the other models on all three datasets. It achieves outstanding performance on the Dataset_3 with fewer samples, achieving an accuracy of 92.97%. Furthermore, due to more samples, each model performs better on Dataset_2 compared to the other datasets. The performance gap between LSTM and GRU varies significantly across different datasets, as smaller datasets may not provide sufficient information to train complex models for recurrent neural networks.
Specifically, the accuracy of the RNN is consistently lower than that of other models across all three datasets, indicating poor fault diagnosis performance. This is attributed to the simple structure of the RNN, which struggles to handle long time series and suffers from the vanishing gradient problem. In addition, the highest accuracies achieved by LSTM and GRU are 80.22% and 83.52%. Unlike the STDGCN, LSTM and GRU demonstrate the ability to learn short-term dependencies in time series data. However, their capacity to learn long-term dependencies is limited, making it challenging to capture complex features across multiple time points and impossible to obtain spatial information from various sensors.
Finally, the highest accuracies achieved by the CNN and LeNet models are 70.77% and 72.75%. The aforementioned diagnostic performance by the CNN and LeNET models is undesirable, which can be caused by the inability of CNN-based models to integrate sensor data effectively. Traditional convolutional kernels generate new feature maps by aggregating features from all channels, regardless of their relevance. However, 16 sensor-related input variables are not entirely correlated. Due to the dependence of CNN on spatial information, fault information in specific channels may be influenced by irrelevant, redundant information in other channels, thereby affecting the model’s performance.
This experimental result is further validated by metrics such as F1 score and recall, as shown by Equations (33) and (34).
It can be seen from
Table 5,
Table 6 and
Table 7 that F1 score and recall corresponding to dataset 1, dataset 2, and dataset 3 for the proposed STDGCN are highest compared to other methods. The improvement across the F1 score and recall metrics suggests that the STDGCN method effectively leverages the diversity and volume of data distribution across edge devices. As a result, it can be said that the proposed STDGCN performs best in detecting different types of UAV sensor faults compared to other data-driven methods.
It can be seen from
Table 5,
Table 6 and
Table 7 that F1 score and recall corresponding to dataset 1, dataset 2, and dataset 3 for the proposed STDGCN are highest compared to other methods. As a result, it can be said that the proposed STDGCN performs best in detecting different types of UAV sensor faults compared to other data-driven methods.
5.2. Benefits of a Designed Sensor Data Association Graph
In this section, an experimental study is conducted to validate the benefits of the sensor data association graph. Only self-connections are added to the nodes in the association graph without considering the relationships between the nodes; the model’s structure is unchanged. The experimental results are shown in
Table 8. It can be observed that the diagnostic accuracy of the STDGCN outperforms that of the STDGCN-non-A and exhibits better stability. The average value of the STDGCN is approximately 4.48% higher than that of the STDGCN-non-A. The designed sensor data association graph utilizes prior knowledge, ensuring that the model explicitly captures spatial correlations. These observations demonstrate the beneficial role of the designed association graph in fault diagnosis.
5.3. Benefit of the Difference Layer
The benefit of the utilization of the difference layer during the implementation of the STDGCN is studied in this section, as shown in
Table 9. During this process, the diagnostic accuracy of STDGCN application is compared for the three studied datasets with and without the incorporation of a difference layer, as shown in
Figure 10. The rest of the network structure remained unchanged. As a result, data points are directly incorporated into the STGCM layer without being passed through the difference layer.
It can be seen from
Table 5 that the STDGCN with a differential layer demonstrates higher accuracy consistently in comparison to the application of the STDGCN without a differential layer. On average, for the three compared datasets, the proposed STDGCN with a difference layer demonstrates 4.86% higher accuracy compared to the one without a difference layer.
Through the aforementioned comparative analysis, several reasons can be identified for the high accuracy of the STDGCN. Firstly, the introduction of the mathematical model has enabled the acquisition of accurate spatial dependencies among sensor measurements. Secondly, implementing the difference layer has facilitated the extraction of differential features between adjacent time steps in the time series. Finally, the STGCM allows for the simultaneous extraction of temporal and spatial information from sensor data.
6. Conclusions
This paper applies an emerging GCN-based method for fault diagnosis of UAV sensors. The data required for the model are obtained through real UAV flight experiments. Moreover, to better integrate multi-source sensor data, a mathematical model of a quadcopter UAV is introduced to construct the sensor data graph. The proposed STDGCN model in this study explores the spatial dependencies of individual nodes through graph convolutional layers and captures the temporal dependencies of each node through stacked gated convolutional layers. Additionally, a differential layer is incorporated in the STDGCN model to calculate higher-order backward features, enhancing the node features in the graph. Upon comparing the STDGCN model with other state-of-the-art neural network models, the main advantages of the STDGCN model can be listed as follows: (1) The association graph is accurately constructed by utilizing prior knowledge of UAV mathematical models, allowing the STDGCN to capture spatial dependencies among sensor variables effectively. (2) The difference layer enhances the graph nodes’ features, improving the model’s accuracy. (3) The introduction of an STGCM enables the extraction of spatial and temporal dependencies of sensor data, ensuring the stability and robustness of the model. Despite having the advantages mentioned above, not all sensor data on the UAV are integrated into this study. Instead, only 16 sensor variables involved in the mathematical model of the UAV are processed, which can be increased in the future. Furthermore, the performance of the proposed method is dependent on the availability and robustness of the model during the real flight time. To this extent, future research can be conducted regarding the real-time implementation and reliability insurance of the proposed model during UAV flight time. In addition, research can be conducted on enhancing the computational cost of the proposed STDGCN caused by the data association from different sensors for efficient real-time application. Additionally, the quadcopter UAV has limited battery capacity, resulting in limited flight time and insufficient flight data. In the future, more data can be obtained by improving the experimental design of the UAV sensors. Considering the information shared among different sensors during the application of the proposed STDGCN, future research can also be conducted to ensure data privacy, such as with the federated learning framework. The applications of the proposed STDGCN can be expanded into other domains of research with multi-source sensor data requirements, such as in commercial aircraft, satellites, submarines, etc.