You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

16 November 2023

Research Method for Ship Engine Fault Diagnosis Based on Multi-Head Graph Attention Feature Fusion

,
,
,
,
and
1
Marine Engineering College, Dalian Maritime University, Dalian 116026, China
2
Dalian Maritime University Smart Ship Limited·Company, Dalian 116026, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Technical Diagnostics and Predictive Maintenance

Abstract

At present, there are problems such as low fault data, insufficient labeling information, and poor fault diagnosis in the field of ship engine diagnosis. To address the above problems, this paper proposes a fault diagnosis method based on probabilistic similarity and rank-order similarity of multi-head graph attention neural networks (MPGANN) models. Firstly, the ship engine dataset is used to explore the similarity between the data using the probabilistic similarity of T_SNE and the rank order similarity of Spearman’s correlation coefficient to define the neighbor relationship between the samples, and then the appropriate weights are selected for the early fusion of the two graph structures to fuse the feature information of the two scales. Finally, the graph attention neural networks (GANN) incorporating the multi-head attention mechanism are utilized to complete the fault diagnosis. In this paper, comparative experiments such as graph construction and algorithm performance are carried out based on the simulated ship engine dataset, and the experimental results show that the MPGANN outperforms the comparative methods in terms of accuracy, F1 score, and total elapsed time, with an accuracy rate of 97.58%. The experimental results show that the model proposed in this paper can still fulfill the ship engine fault diagnosis task well under unfavorable conditions such as small samples and insufficient label information, which is of practical significance in the field of intelligent ship cabins and fault diagnosis.

1. Introduction

As the main power source for ship navigation and power generation, ship engines occupy an important position in the field of ships. Due to its importance, if the engine operation failure cannot be found in time, it will cause huge economic losses or even casualties. Therefore, timely monitoring and diagnosis of the engine’s working condition can effectively improve the reliability, economy, and safety of normal ship operation.
With the wave of the era of ship intelligence and the continuous development and improvement of artificial intelligence and algorithmic arithmetic power, more and more scholars apply artificial intelligence to the field of condition detection and fault diagnosis, and the fault diagnosis methods of ship engines also progress and develop. Fault diagnosis of ship engines faces challenges such as data scarcity, multiple fault types, data noise interference, and insufficient labeling information. In order to provide safety for normal ship navigation, Zhong et al. [1] introduced deep learning into ship diesel engine fault diagnosis and proposed a fault diagnosis method based on correlation distribution and deep confidence networks, which achieved good results. However, this type of method does not take into account the problem of imbalance in the ship fault data categories, resulting in some defects in the training and application of the model. Aiming at the above problems Hou et al. [2] first used principal component analysis (PCA) to reduce the dimensionality of the data. Then, the sample size optimization algorithm (SSO) was used to solve the problem of data imbalance, and the three-dimensional Arnold mapping to the particle swarm optimization algorithm was introduced to improve the holding vector machine (SVM), which effectively improves the generalization ability of the classification model. Finally, the model is validated by example using the fault data of the fuel supply system, and the experimental results show that the method is able to alleviate the impact of unbalanced data. Zhong et al. [3] proposed a semi-supervised principal component analysis (SSPCA) to improve the diagnostic accuracy of the model by fusing labeled and unlabeled samples as an alternative to unsupervised learning, with better robustness to false alarms.
The above model has achieved good results in fault monitoring, but it can only determine whether the fault occurs or not and cannot diagnose the specific cause of the fault, which makes it difficult to play a guiding role for the subsequent operation and maintenance decision-making work. In this regard, Ren et al. [4] established a fault tree, transformed the fault tree into a Bayesian network, constructed a Bayesian network diagnostic model of the lubrication system, and, in the example validation, the model was able to accurately diagnose the cause of the failure. Han et al. [5] introduced a convolutional neural network to solve the problem of propulsion fault detection and classification of time series. Xu et al. [6] fused an artificial neural network (ANN) model, a belief rule-based reasoning (BRB) model, and an ER rule model and used a genetic algorithm to optimize the importance weights of each model to improve the overall performance of the fusion system, and finally, through the three models, joint decision-making to realize the fault diagnosis of the ship system. In recent years, algorithmic improvements have achieved significant results in the field of industrial fault diagnosis [7,8,9,10,11]. Samet [12] proposed a new kNN-based classifier (PFS-kNN) to find the k-nearest neighbors using Minkowski’s metric of the image fuzzy soft matrix, which was validated on the UCI medical dataset, and the PFS-kNN outperformed the most state-of-the-art kNN-based algorithms. Yang et al. [13] proposed a diagnostic method based on fault sign simulation to solve the problem of a too-small sample size of faults in the fault diagnosis of a particular engine. Agrawal et al. [14] proposed a data-driven model for predictive health maintenance to monitor the health status of the equipment and take action before a fault occurs. The model will alert for any major defects in the system and can be effective in avoiding work interruptions in the production process.
A graph neural network (GNN) is a neural model that captures graph dependencies through message passing between graph nodes. Zhou et al. [15] divided GNN applications into six directions: natural language processing, computational technology, natural science research, knowledge graph, combinatorial optimization, and graph generation, and made a more comprehensive and detailed summary of GNN applications.
In recent years, variants of GNNs such as Graph Convolutional Networks (GCNs) [16] and GANN [17] have achieved breakthroughs in many deep learning tasks. Kipf et al. [18] applied GCNs to fast and scalable semi-supervised classification of nodes in graphs, which demonstrated significant superiority over a large number of datasets. Wang et al. [19] proposed a multi-graph-based GCN fault diagnosis method that fuses two feature embeddings into one combined embedding by introducing a self-attention mechanism to improve classification accuracy and stability in unbalanced datasets. Li et al. [20] proposed a domain adversarial graph convolutional network (DAGCN), which incorporates three types of information—class label, domain label, and data structure—into a deep network for modeling, thus realizing unsupervised domain adaptive (UDA) fault diagnosis. Zheng et al. [21] proposed a new quantum graph convolutional neural network (QGCN) model by drawing on quantum neural networks and graph convolutional neural networks, which utilize quantum parametric circuits and the high computational power of quantum systems to accomplish the traditional task of graph classification in machine learning. In all of the above studies, GCN has achieved good results in the field of graph structure classification due to its advantage of being able to handle non-Euclidean distance data well. However, GCN’s mean-aggregation operation in the metric matrix makes it insufficiently attentive to important nodes. Therefore, Veličković et al. [22] constructed GANN by adding a self-attention layer to traditional GCNs, which are able to set different weights for different nodes in the neighborhood, thus solving the drawback that GCNs cannot consider the importance of neighbors. On this basis, Yang et al. [23] proposed a full graph attention neural network (FGANN), which also considers the influence of nodes other than neighboring nodes and can handle the graph classification task well. Zhang et al. [24] noticed that in sentiment classification tasks, traditional models are not sensitive to syntactic structural information due to the complexity of syntactic analysis relationships, and the models lack the utilization of external sentiment knowledge. In this regard, a new graph attention neural network is proposed that realizes attention to multiple levels of syntax, semantics, and knowledge by introducing the mechanism of multi-head attention and acquires knowledge about sentiment by considering information such as syntax.
The relationship between numerous thermal parameters of ship engines is not just a simple linear relationship but is often very complex and agnostic. In this paper, to address this problem, graph learning theory is introduced to explore the relationship of geometric structure among data, and a MPGANN ship engine fault diagnosis model is proposed. Based on the ship engine failure dataset, the model uses both T_SNE and Spearman to construct the probabilistic graph structure and ordinal graph structure, selects appropriate weights for early fusion of the two graph structures, and finally adopts GANN to transfer and extract the combined features. The core of the model is that it can mine the similarity between the data from two scales and more comprehensively describe the relationship between the data and the graph structure to provide more useful information for the model. The main contributions of this paper can be summarized as follows:
(1)
Transform the ship engine dataset into two graph structures from different scales and make the two graph structures contain similarity relationships from multiple perspectives by extracting the neighbor relationships between samples so as to achieve complementation and extension of the model input information.
(2)
Introducing fusion weights, the two graph structures are structurally fused according to appropriate weights to obtain a fused graph structure that contains deeper information.
(3)
Input the obtained fusion graph structure into the GANN of fused multi-head attention for multi-channel feature extraction, and finally connect the Softmax layer to realize fault diagnosis.
(4)
The ship engine thermal parameter dataset is used to verify that the MPGANN proposed in this paper outperforms other classical algorithms and achieves higher accuracy.
The remaining sections of this paper are organized as follows: In Section 2, the theory related to the MPGANN model is introduced in detail. Section 3 presents experiments and comparisons of the construction of the model proposed in this paper, the setting of hyperparameters, and the performance of the algorithm through case studies. Section 4 then summarizes and outlooks the paper.

3. Case Study

3.1. Data Description

The subject of this experiment is a Wärtsilä 9L34DF marine engine. The real ship failure simulation is costly, dangerous, and destructive; therefore, the engine failure simulation performed by AVL-BOOST (R2019.2) software is used in this paper. The simulation model of the marine engine is shown in Figure 4, where SB is the system boundary; CL is the air filter; TC is the supercharger; CO is the air cooler; PL is the intake and exhaust manifold; C is the cylinder; MP is the arrangement of the detection points; and the black line is the corresponding connecting pipe.
Figure 4. Simulation model of the ship’s engine.
Before the fault, simulation needs to verify the effectiveness of the model. The text for the parameters under different working loads is verified for the comparison of measured values and simulation values. Error analysis and comparison results are shown in Table 1.
Table 1. Comparison results between model simulation data and measured data.
As can be seen from Table 1, the error between the measured and real values of each main performance parameter under different loads is less than 4%. Especially under 100% load, the error of each main performance parameter is within 2%, and the error is within the acceptable range of requirements, so this paper chooses the model under 100% working conditions for fault simulation and fault sample collection and simulates five states, namely, injection timing advance (F1), injection timing delay (F2), supercharger efficiency decline (F3), air cooler efficiency decline (F4), and normal operation (F5). Five states are simulated, and the fault simulation scheme is shown in Table 2.
Table 2. Engine failure simulation scheme.
After the above fault simulation scheme, in order to fully explore and analyze the relationship between internal engine parameters and faults and realize the diagnosis of engine faults, a total of 400 sets of samples of 23-dimensional detection index parameters were selected, of which 80 sets of samples for each state were used to carry out the fault diagnosis study. The dataset is as specific as Table 3, and the collected detection indexes are shown in Table 4.
Table 3. Engine failure dataset.
Table 4. Detection indicator indexes.

3.2. Comparison of Graph Structures

In this section, by comparing and analyzing the commonly used correlation coefficients, including Pearson correlation coefficient, Kendall correlation coefficient, and Spearman correlation coefficient, as well as their constructed correlation graph structure and probability graph structure, respectively, we compare the model accuracy and loss and select the graph structure construction method that is suitable for the proposed model in this paper. The activation function selected for the model is ReLU, the learning rate is 0.01, the number of iterations is 1000, and the number of multi-attention heads is 16. The ratio of the model training set, validation set, and test set is 7:1:2, which is selected randomly. Each graph structure construction method was subjected to 20 experiments, and the average accuracy of the 20 ornaments was used as the basis for judgment.
As can be seen in Figure 5, the average accuracy of the Spearman correlation coefficient in the correlation graph structure is better than that of the Pearson correlation coefficient and Kendall correlation coefficient, and the average accuracy of the fusion graph obtained after the fusion of the graph structures is improved as a model input. This is due to the fact that different graph construction methods extract information from different aspects of the data, and the fused graph structure contains information from both aspects of the data, which can be input into the neural network with more information than a single graph structure. Especially when the probabilistic graph structure and the correlation graph structure constructed by the Spearman correlation coefficient are fused, the average accuracy of the model is greatly improved, reaching 96.36%. It can be seen that the model performance is optimal when the correlation graph structure is constructed with the Spearman correlation coefficient, which is chosen to construct the graphs in all subsequent experiments.
Figure 5. Test accuracy for each graph structure and total elapsed time.
Figure 6 shows the fused graph structure after fusion using the T_SNE probabilistic graph structure and the Spearman rank order graph structure. The edges between the nodes indicate that the nodes are neighbors to each other, as can be seen from the figure that the node relationships are only roughly delineated in a single graph structure, and after fusion of the graph structures, the neighbor relationships between the nodes have been clearly described.
Figure 6. Fusion of probabilistic graph structures with rank-order graph structures.
After constructing the neighbor relationship, the labels of some nodes are removed; nodes with color in Figure 7 indicate that the node has label information, while nodes without color indicate that the node has no label information. During the graph learning process, these unlabeled nodes do not provide labeling information, but they can aggregate the neighbor’s node information to be used as auxiliary data in order to provide the information used to explore the similarity structure between the nodes. The information of the nodes is used to aggregate the neighboring information through multi-head attention, and after the graph learning, the updated new nodes are classified, and all the samples are tagged with labels. The above figure shows the classification results, and it can be seen that although the model is not given all the labeling information, it can still correctly label the unlabeled nodes.
Figure 7. Graph learning process.

3.3. Network Parameter Setting

In this subsection, we will delve into the critical hyperparameters of the neural network model for fault diagnosis. These include the activation function, the number of attention heads, and the model learning rate. We maintain the ratio of the training set, validation set, and test set at 7:1:2, conducting a total of 20 experiments. The results from these experiments will be averaged to determine the average accuracy rate, serving as the basis for the final assessment and guiding the selection of the most suitable network structure.

3.3.1. Activation Function

The activation function is a very important part of the neural network structure, which can nonlinearly transform the input data. In this paper, five typical activation functions, Sigmoid, ReLU, ELU, ReLU6, and Tanh, as well as three novel activation functions, SupEx [36], αSechSig, and αTanhSig [37], are selected for comparison, and the number of iterations is set to be 1000 and the learning rate is set to be 0.01. Figure 8 shows the total time consumed and the average accuracy in the test set for the models with different activation functions selected. of the total elapsed time and the average accuracy of the test set.
Figure 8. Average accuracy and total elapsed time for different activation functions.
In Figure 8, it can be seen that Sigmoid has the lowest accuracy when it is used as an activation function, and the total time consumed by αSechSig and ELU is significantly higher than the other activation functions, and the total time consumed by the rest of the activation functions, even though the total time consumed by SupEx and αTanhSig is less, the total time consumed is not long in the data of the small samples in the first place, so the accuracy is used as the main criterion and the total time consumed is used as the auxiliary criterion, and the model’s classification accuracy for the test set is over 0.97 when Tanh and ReLU6 are used as activation functions, the classification accuracy of the model on the test set exceeds 0.97, reaching 0.9758 and 0.9727, of which Tanh not only has high accuracy but also the total time consumed is less than ReLU6.
In Figure 9, we can intuitively see that Tanh, as the activation function, has the highest average accuracy; the fluctuation range of the experimental results is also the smallest; and most of the experimental results are in the 0.975 or so range. The synthesis of the above two results can be seen in this paper’s model Tanh, as the activation function is the optimal effect.
Figure 9. Boxplot of the accuracy of different activation functions.

3.3.2. Number of Attention Heads

Multi-head attention allows the model to focus on different channels of the input at the same time, thus improving the expressiveness and performance of the model. It should be noted that while multi-head attention increases the complexity of the model, it also increases the computational overhead of training and inference, so the optimal number of heads needs to be determined experimentally to achieve the best performance of the model. The activation function selected in the experiment is Tanh, the value range of H is set to [1, 50], and other settings are the same as in the previous experiment. The final comparison of the average accuracy of the test set under different numbers of heads and the experimental results is shown in the following figure.
Based on Figure 10, it is evident that the model achieves higher accuracy within the range of head counts H = 16, H = 22, and H = 30. The highest accuracy is attained at a learning rate of Lr = 0.01 with H = 16. However, it is worth noting that excessively high head counts lead to longer running times and increased computational costs. Therefore, the optimal configuration of the model is Lr = 0.01 and H = 16.
Figure 10. Heat map of model accuracy at different learning rates and head counts.

3.3.3. Algorithm Performance Comparison

In order to verify the validity of the proposed model in this paper, GCN, CNN, SVM, and BPNN algorithms are selected in the experiments in this subsection to compare the algorithm performance with the MPGANN model proposed in this paper. The parameters of GCN, such as the neighbor graph construction method, the number of hidden layers, the number of neurons in the hidden layer, and the dropout rate, are all the same as those of MPGANN. CNN has a total of four convolutional layers. The size of the convolution kernel is 5-3-2-2, the stride is 1-1-1-1, and each convolution layer is connected to an average pooling layer after each convolution layer. The window size is 2-2-2-2, padding is 0, and stride is 1. The kernel function of SVM is the Gaussian kernel function; set the penalty factor c = 1.0 and set the gamma to auto. In the BPNN hidden layer, the number of neurons is set to 16, the number of iterations for all algorithms is set to 1000, and the learning rate is set to 0.01.
Accuracy is an overall performance indicator; if it is used as a performance evaluation index for diagnostic algorithms, it will likely mask the specific classification of diagnostic algorithms, so in order to solve this type of problem, this paper introduces precision, recall, and the F1 score, calculated as:
R = T P T P + F N ,
P = T P T P + F P ,
F = 2 1 P + 1 R = 2 R P R + P ,
where R is the recall rate; P is the precision rate; F is the F1 score; T P is the number of samples that classify positive samples as positive samples; F N is the number of positive samples classified as negative samples, and F P is the number of negative samples classified as positive samples.
The precision and recall of each algorithm used in this paper are shown in Table 5. From the table, it can be seen that each algorithm can basically correctly classify the normal working conditions, but the complexity of the engine fault data leads to a relatively poor classification effect for the traditional diagnostic algorithms. In this experiment, the precision rate and recall rate of BPNN are lower than other diagnostic algorithms, and although the precision rate and recall rate of GCN and CNN diagnosis for each condition are better than those of SVM and BPNN, the diagnostic effect is still unsatisfactory in F1 and F2 conditions. The MPGANN proposed in this paper achieves 100% accuracy and recall in all working conditions except for 0.89 accuracy and 0.88 recall in F2 and F1 working conditions, which is an obvious advantage over other diagnostic algorithms and is more in line with practical industrial needs.
Table 5. Precision and recall for each algorithm.
From Figure 11, it can be seen that fault F1 is misclassified in all diagnostic algorithms, while after the feature mining of probabilistic graph relationships and rank order graph structure, MPGANN and GCN only have two F1 fault samples misclassified, and the MPGANN proposed in this paper only has two samples misclassified, which shows that the MPGANN model can mine more information between the data to improve the diagnostic algorithm’s classification accuracy and stability and is better than other models.
Figure 11. Categorization of test set failures: (a) MPGAT, (b) GCN, (c) CNN, (d) SVM, and (e) BPNN.
Figure 12 shows the average accuracy, average precision, average recall, and F1 score of each diagnostic algorithm. The performance of each diagnostic algorithm can be seen more clearly by comparing the figures. CNN, as a network model for deep learning, is widely used in the field of mechanical fault diagnosis and has achieved excellent results. However, due to the fact that there is no large amount of data for model training in this experiment, the training effect of CNN is not good, and it cannot make correct classification judgments. GCN is limited by its own information propagation rules, and even though it adopts the same topology structure for fault diagnosis, the diagnostic effect is still not good. However, the MPGANN proposed in this paper uses the multi-head attention mechanism in the aggregation operation, which makes the feature aggregation more reasonable and increases the learning efficiency of the model, and the average precision, average accuracy, average recall, and F1 score reach 0.978, 0.976, 0.975, and 0.976.
Figure 12. Algorithm performance comparison.

4. Conclusions

In this paper, a new MPGANN model is proposed, and in the experiments on the ship engine dataset, firstly, different graph structure construction methods are compared, and the fused graph structure enhances the expressive ability, especially the fusion of probabilistic and ordinal graph structures, which dramatically improves the diagnostic accuracy of the model. Secondly, the accuracy of the model under different activation functions, head counts, and learning rates is greatly different, which shows that the influence of hyperparameters on the model is also not negligible. Finally, when MPGANN is compared with other algorithms, it outperforms the various methods compared in diagnosis in terms of accuracy, precision, recall, and F1 score. Therefore, the MPGANN proposed in this paper has a certain practical significance in the field of intelligent ship fault diagnosis. In summary, the research for this paper is summarized as follows:
(1)
The two graph structure construction methods in the MPGANN model can effectively obtain probabilistic similarity and ordinal similarity between data samples and transform the data into probabilistic graph structure and ordinal graph structure.
(2)
Early fusion is employed to combine the probability map structure and rank-order map structure with the incorporation of feature weights. This integration process effectively amalgamates information from samples at various scales.
(3)
The multi-head attention mechanism is applied to conduct multi-channel feature screening on the fusion graph structure, extracting feature information with higher relevance to enhance the diagnostic performance of the model.
(4)
The model effect was validated using the ship engine fault dataset, and compared with other models in terms of accuracy, precision, recall, and F1 score, MPGANN was the most effective, with a diagnostic accuracy as high as 97.58% and an F1 score of 97.6%.
The proposed diagnostic model in this paper fuses features from different aspects of the data, mining the information of the dataset at a deeper level, but it also has some limitations, such as: the fused features contain only probabilistic similarity and rank-order similarity; the fusion weights need to be set in advance; and so on.
There are three goals for future work: (1) Physical connections in real space will also imply information, and the actual structural connections of the engine will be transformed into graph structures to mine the correlation information between the samples. (2) Choose more graph fusion methods or assign meanings to edge features so that the model has more information for training to improve its diagnostic accuracy. (3) Introduce the attention mechanism into the fusion process to achieve adaptive allocation of fusion weights.

Author Contributions

Conceptualization, Z.A. and H.C.; methodology, Z.A.; validation, Z.A., H.C. and J.W.; formal analysis, H.C.; investigation, Z.C. and J.W.; resources, H.C.; data curation, Z.A., L.W. and K.J.; writing—original draft preparation, Z.A.; writing—review and editing, Z.A., H.C. and L.W.; visualization, Z.A. and Z.C.; supervision, H.C.; project administration, H.C.; funding acquisition, H.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the project: Development of Ship Operation Condition Monitoring and Simulation Platform (funding number: 1638882993269) and the research and application of the Smart Ship Digital Twin Information Platform (fundingt number: 2022JH1/10800097).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

Author Hui Cao was employed by Dalian Maritime University Smart Ship Limited Company. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Zhong, G.Q.; Wang, H.Y.; Zhang, K.Y.; Jia, B.Z. Fault diagnosis of Marine diesel engine based on deep belief network. In Proceedings of the 2019 Chinese Automation Congress (CAC), Hangzhou, China, 22–24 November 2019; pp. 3415–3419. [Google Scholar] [CrossRef]
  2. Hou, L.; Zhang, J.; Du, B. A Fault Diagnosis Model of Marine Diesel Engine Fuel Oil Supply System Using PCA and Optimized SVM. J. Phys. Conf. Ser. 2020, 1576, 012045. [Google Scholar] [CrossRef]
  3. Zhong, K.; Li, J.B.; Wang, J.; Han, M. Fault Detection for Marine Diesel Engine Using Semi-supervised Principal Component Analysis. In Proceedings of the 9th International Conference on Information Science and Technology (ICIST), Hulunbuir, China, 2–5 August 2019; pp. 146–151. [Google Scholar]
  4. Ren, D.P.; Zeng, H.; Wang, X.L.; Pang, S.; Wang, J.D. Fault Diagnosis of Diesel Engine Lubrication System Based on Bayesian Network. In Proceedings of the 6th International Conference on Control, Automation and Robotics (ICCAR), Electr Network, Singapore, 20–23 April 2020; pp. 423–429. [Google Scholar]
  5. Han, P.H.; Li, G.Y.; Skulstad, R.; Skjong, S.; Zhang, H.X. A Deep Learning Approach to Detect and Isolate Thruster Failures for Dynamically Positioned Vessels Using Motion Data. IEEE Trans. Instrum. Meas. 2021, 70, 1–11. [Google Scholar] [CrossRef]
  6. Xu, X.J.; Zhao, Z.Z.; Xu, X.B.; Yang, J.B.; Chang, L.L.; Yan, X.P.; Wang, G.D. Machine learning-based wear fault diagnosis for marine diesel engine by fusing multiple data-driven models. Knowl. Based Syst. 2020, 190, 105324. [Google Scholar] [CrossRef]
  7. Cao, H.; Zhang, J.D.; Cao, X.; Li, R.; Wang, Y.R. Optimized SVM-Driven Multi-Class Approach by Improved ABC to Estimating Ship Systems State. IEEE Access 2020, 8, 206719–206733. [Google Scholar] [CrossRef]
  8. Wang, R.H.; Chen, H.; Guan, C. Random convolutional neural network structure: An intelligent health monitoring scheme for diesel engines. Measurement 2021, 171, 108786. [Google Scholar] [CrossRef]
  9. Memis, S.; Arslan, B.; Aydin, T.; Enginoglu, S.; Camci, C. Distance and Similarity Measures of Intuitionistic Fuzzy Parameterized Intuitionistic Fuzzy Soft Matrices and Their Applications to Data Classification in Supervised Learning. Axioms 2023, 12, 463. [Google Scholar] [CrossRef]
  10. Choo, K.B.; Cho, H.; Park, J.H.; Huang, J.F.; Jung, D.; Lee, J.; Jeong, S.K.; Yoon, J.; Choo, J.; Choi, H.S. A Research on Fault Diagnosis of a USV Thruster Based on PCA and Entropy. Appl. Sci. 2023, 13, 3344. [Google Scholar] [CrossRef]
  11. Zhao, N.Y.; Mao, Z.W.; Wei, D.H.; Zhao, H.P.; Zhang, J.J.; Jiang, Z.N. Fault Diagnosis of Diesel Engine Valve Clearance Based on Variational Mode Decomposition and Random Forest. Appl. Sci. 2020, 10, 1124. [Google Scholar] [CrossRef]
  12. Memis, S. Picture Fuzzy Soft Matrices and Application of Their Distance Measures to Supervised Learning Picture Fuzzy Soft k-Nearest Neighbor (PFS-kNN). Electronics 2023, 12, 4129. [Google Scholar] [CrossRef]
  13. Yang, K.; Fan, H.Y. Research on Fault Diagnosis Method of Diesel Engine Thermal Power Conversion Process. In Proceedings of the 3rd International Forum on Energy, Environment Science and Materials (IFEESM), Shenzhen, China, 25–26 November 2017; pp. 1445–1450. [Google Scholar]
  14. Agrawal, S.K.; Banerjee, S.; Sinha, A.; Das, D. SafeEngine: Fault Detection with Severity Prediction for Diesel Engine. In Proceedings of the 2022 IEEE 10th Region 10 Humanitarian Technology Conference (R10-HTC), Hyderabad, India, 16–18 September 2022; pp. 216–220. [Google Scholar]
  15. Zhou, J.; Cui, G.; Hu, S.; Zhang, Z.; Yang, C.; Liu, Z.; Wang, L.; Li, C.; Sun, M. Graph neural networks: A review of methods and applications. AI Open 2020, 1, 57–81. [Google Scholar] [CrossRef]
  16. Nie, J.; Xu, Y.; Huang, Y.; Li, J. The review of image processing based on graph neural network. In Proceedings of the International Conference on Intelligent Robotics and Applications, Yantai, China, 22–25 October 2021; pp. 534–544. [Google Scholar]
  17. Zhou, A.Z.; Li, Y.F. Structural attention network for graph. Appl. Intell. 2021, 51, 6255–6264. [Google Scholar] [CrossRef]
  18. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  19. Wang, R.; Chen, H.; Guan, C. DPGCN Model: A Novel Fault Diagnosis Method for Marine Diesel Engines Based on Imbalanced Datasets. IEEE Trans. Instrum. Meas. 2022, 72, 1–11. [Google Scholar] [CrossRef]
  20. Li, T.; Zhao, Z.; Sun, C.; Yan, R.; Chen, X. Domain adversarial graph convolutional network for fault diagnosis under variable working conditions. IEEE Trans. Instrum. Meas. 2021, 70, 1–10. [Google Scholar] [CrossRef]
  21. Zheng, J.; Gao, Q.; Lü, Y. Quantum graph convolutional neural networks. In Proceedings of the 2021 40th Chinese Control Conference (CCC), Shanghai, China, 26–28 July 2021; pp. 6335–6340. [Google Scholar]
  22. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. arXiv 2017, arXiv:1710.10903. [Google Scholar]
  23. Yang, F.; Zhang, H.Y.; Tao, S.M. Semi-supervised classification via full-graph attention neural networks. Neurocomputing 2022, 476, 63–74. [Google Scholar] [CrossRef]
  24. Zhang, S.; Gong, H.; She, L. An aspect sentiment classification model for graph attention networks incorporating syntactic, semantic, and knowledge. Knowl. Based Syst. 2023, 275, 110662. [Google Scholar] [CrossRef]
  25. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  26. Zhang, W.-Y.; Wei, Z.-W.; Wang, B.-H.; Han, X.-P. Measuring mixing patterns in complex networks by Spearman rank correlation coefficient. Phys. A Stat. Mech. Appl. 2016, 451, 440–450. [Google Scholar] [CrossRef]
  27. Wang, Y.; Qin, Y.; Guo, J.; Cao, Z.; Jia, L. Multi-point short-term prediction of station passenger flow based on temporal multi-graph convolutional network. Phys. A Stat. Mech. Appl. 2022, 604, 127959. [Google Scholar] [CrossRef]
  28. Wang, L.; Hu, Z.; Kong, Q.; Qi, Q.; Liao, Q. Infrared and Visible Image Fusion via Attention-Based Adaptive Feature Fusion. Entropy 2023, 25, 407. [Google Scholar] [CrossRef] [PubMed]
  29. Luo, Z.; Li, J.; Zhu, Y. A deep feature fusion network based on multiple attention mechanisms for joint iris-periocular biometric recognition. IEEE Signal Process. Lett. 2021, 28, 1060–1064. [Google Scholar] [CrossRef]
  30. Yang, J.; Yang, J.-Y.; Zhang, D.; Lu, J.-F. Feature fusion: Parallel strategy vs. serial strategy. Pattern Recognit. 2003, 36, 1369–1381. [Google Scholar] [CrossRef]
  31. Wang, Y.; Xu, X.; Yu, W.; Xu, R.; Cao, Z.; Shen, H.T. Comprehensive Framework of Early and Late Fusion for Image–Sentence Retrieval. IEEE MultiMedia 2022, 29, 38–47. [Google Scholar] [CrossRef]
  32. Boulahia, S.Y.; Amamra, A.; Madi, M.R.; Daikh, S. Early, intermediate and late fusion strategies for robust deep learning-based multimodal action recognition. Mach. Vis. Appl. 2021, 32, 121. [Google Scholar] [CrossRef]
  33. Ruiz, L.; Gama, F.; Ribeiro, A. Graph neural networks: Architectures, stability, and transferability. Proc. IEEE 2021, 109, 660–682. [Google Scholar] [CrossRef]
  34. Brauwers, G.; Frasincar, F. A general survey on attention mechanisms in deep learning. IEEE Trans. Knowl. Data Eng. 2021, 35, 3279–3298. [Google Scholar] [CrossRef]
  35. Weston, J.; Chopra, S.; Bordes, A. Memory networks. arXiv 2014, arXiv:1410.3916. [Google Scholar]
  36. Kiliçarslan, S.; Közkurt, C.; Bas, S.; Elen, A. Detection and classification of pneumonia using novel Superior Exponential (SupEx) activation function in convolutional neural networks. Expert Syst. Appl. 2023, 217, 119503. [Google Scholar] [CrossRef]
  37. Közkurt, C.; Kiliçarslan, S.; Bas, S.; Elen, A. αSechSig and αTanhSig: Two novel non-monotonic activation functions. Soft Comput. 2023, 27, 18451–18467. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.