You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

24 April 2025

A Robust Multivariate Time Series Classification Approach Based on Topological Data Analysis for Channel Fault Tolerance

and
1
Department of Electrical and Computer Engineering, Inha University, Incheon 22212, Republic of Korea
2
Department of Computer Engineering, Inha University, Incheon 22212, Republic of Korea
*
Author to whom correspondence should be addressed.
This article belongs to the Section Fault Diagnosis & Sensors

Abstract

In this study, we propose a robust artificial intelligence (AI) model for vibration monitoring of rotating equipment to support reliable operation across various industries, including manufacturing, power plants, and aerospace. The reliability and completeness of sensor data are essential for early detection of anomalies in equipment and for performing predictive maintenance. While AI-based predictive maintenance and condition-monitoring technologies have advanced in recent years, the issue of data loss caused by sensor failures remains a significant challenge that leads to performance degradation of AI models. In particular, for equipment utilizing multiple sensors, the complete loss of data from a single sensor significantly diminishes the predictive maintenance capability of AI models, thereby reducing their reliability. To address this issue, this study introduces topological data analysis (TDA) to develop a robust AI model. TDA analyzes the topological structure of sensor data to generate consistent feature vectors that capture the intrinsic characteristics of the data. This enables stable predictions even when certain channels of multi-sensor data are entirely missing. The proposed method demonstrates high performance resilience under conditions of partial sensor data loss, thereby contributing to enhanced reliability of AI-based predictive maintenance systems and the establishment of efficient maintenance strategies in the future.

1. Introduction

Rotating equipment plays a critical role across various industries, including manufacturing, power plants, and aerospace. These systems are susceptible to a range of mechanical and electrical anomalies, which can lead to performance degradation, emergency shutdowns, or, in severe cases, catastrophic failures. As such, continuously monitoring equipment, detecting anomalies early, and performing predictive maintenance are essential tasks for ensuring stable operation. This necessity has driven advancements in the field of Prognostics and Health Management (PHM).
In recent years, research in PHM has increasingly leveraged artificial intelligence (AI) algorithms to detect anomalies in equipment based on sensor data [1,2]. The goals of these studies are to detect performance degradation or faults in advance and maximize equipment uptime through predictive maintenance. Commonly, sensors such as vibration, temperature, and pressure sensors are used to monitor equipment conditions in real time. The collected data are analyzed to detect anomalies and predict the optimal timing for maintenance [3,4]. Therefore, accurate sensor data are critical for ensuring the stability and efficiency of equipment operation.
However, the reliability and completeness of sensor data can be easily compromised by various factors, such as sensor failures, communication errors, and operation in extreme environments. In particular, the complete loss of data from a failed sensor can disrupt the proper functioning of a predictive maintenance system that relies on multiple sensors. This indicates that even the loss of a single sensor’s data can result in incomplete datasets, which may significantly degrade the performance of AI models [5]. Deep learning models, which improve prediction accuracy by learning patterns from large volumes of high-quality training data, are especially vulnerable to such issues. When certain data channels are entirely missing or contain errors, the model’s predictive accuracy can drop drastically. This not only negatively impacts the reliability and safety of the equipment but also increases the risk of incorrect predictions, which may lead to abnormal operations or severe accidents. Therefore, developing robust AI models capable of handling sensor failures is critical.
A robust AI system must provide reliable predictions even in the presence of sensor failures, and research has been conducted to utilize alternative data sources or predictive techniques to achieve this goal [6,7].
To address these challenges, this study proposes a novel approach by applying topological data analysis (TDA) to develop an AI model that is robust against sensor failures. TDA is a technique that analyzes the structural properties of data to enhance robustness to missing data. It is expected to contribute to delivering reliable predictions in multi-sensor predictive maintenance systems [8,9,10].

3. Methodology

The proposed model architecture, depicted in Figure 1, includes two backbone models: a multivariate time series classifier for MTSC and a feature vector generation layer based on topological data analysis (TDA). These are followed by a feature vector classifier that analyzes the generated feature vectors. Both the MTSC Layer and the feature vector classifier employ basic models rather than advanced time series analysis techniques, to eliminate the influence of specific methodologies on the proposed approach.
Figure 1. Architecture of the proposed model. The input data are processed in parallel through two separate layers, and the classification results from each layer are combined using a weighted fusion strategy.
In the MTSC Layer, the backbone model takes multivariate time series data as input, learns their temporal characteristics, and performs classification. The output probabilities for each class are computed through a Softmax layer. On the other hand, the TDA Layer considers the data at each time step t as an m-dimensional point and computes a persistent homology on the m-dimensional point cloud to generate feature vectors. These feature vectors reflect the topological connectivity of data points, providing additional information about the data’s structural form that is difficult to capture using traditional time series analysis. The generated feature vectors are fed into a classifier, which calculates class probabilities similarly to the MTSC Layer. Finally, the probabilities from both layers are combined using weighted summation, allowing the model to incorporate both temporal and topological features of the data.

3.1. Calculate Persistent Homology

3.1.1. Simplex

A simplex is the most fundamental structure in TDA, used to describe the geometric relationships among data points in high-dimensional spaces. A simplex is a geometric object formed by connecting multiple data points, which can extend from being zero dimensional to higher dimensions. In TDA, simplices simplify complex data structures and help analyze their topological characteristics.
Simplices are defined by their dimensions, as illustrated in Figure 2, which shows simplices that range from zero dimensional to three dimensional. A zero-dimensional simplex is the most basic simplex, represented by a single point, which corresponds to individual data points in a dataset. A one-dimensional simplex is a line segment formed by connecting two points, a two-dimensional simplex is a triangle formed by connecting three points, and a three-dimensional simplex is a tetrahedron formed by connecting four points. This concept can be extended to higher dimensions, allowing the representation of complex data in higher-dimensional spaces.
Figure 2. (a) 0 simplex, (b) 1 simplex, (c) 2 simplex, (d) 3 simplex.

3.1.2. Simplicial Complex

A simplicial complex is a mathematical structure formed by combining multiple simplices to represent high-dimensional data and their interconnections. For a set of data points V = { v 1 , v 2 , , v n } and a collection of subsets S , a simplicial complex is defined as ( V , S ) , satisfying the following conditions:
( V , S ) is a simplicial complex if 1 . v V , { v } S 2 . τ σ S τ S
These conditions indicate that every data point v V forms a zero-dimensional simplex within the simplicial complex and that any subset τ of a simplex σ in S also belongs to S .
To quantify the topological characteristics of simplices within a simplicial complex, Betti numbers are calculated. The Betti number β k represents the number of k-dimensional holes or voids. For instance, β 0 denotes the number of connected components, and β 1 denotes the number of holes, as illustrated in Figure 3.
Figure 3. The set { a , b , c } is a connected component, as is { d , e , f , g } . The set { d , e , f , g } constitutes a hole.

3.1.3. Filtration

Persistent homology tracks the changes in Betti numbers during the formation of a simplicial complex as data points are connected under a specific rule, a process known as filtration. One common filtration method is the Vietoris–Rips filtration, which constructs a simplicial complex by connecting data points within a distance 2 ϵ , as shown in Figure 4.
Figure 4. Data points are connected as distance ϵ increases: (a) no connections, (b) connections within distance 1.0, (c) connections within distance 2.0, (d) connections within distance 4.0.
During filtration, the simplicial complexes at each stage form a sequence of inclusions:
Ø = V 0 V 1 V 2 V n = V
Here, V i represents the simplicial complex at stage i, where V 0 denotes the empty set or minimal connectivity, and V n represents the topological space where all data points are connected.

3.1.4. Persistent Homology

Persistent homology tracks the Betti numbers of each simplicial complex throughout filtration, as illustrated in Figure 5.
Figure 5. As ϵ changes following filtration, a connection is created between each data point, and the change in Betti number is tracked accordingly.
As Betti numbers change, the birth and death of each dimension’s components are recorded as birth–death pairs. Persistent homology consists of these pairs for each dimension. A persistence diagram, as shown in Figure 6, visualizes the results.
Figure 6. Two-dimensional visualization of persistence homology, where the x-axis represents birth and the y-axis represents death.
The x-axis of the persistence diagram represents birth, while the y-axis represents death. The diagram in Figure 6 visualizes the birth and death of zero-dimensional points, one-dimensional edges, and two-dimensional faces.

3.2. Calculate Persistent Diagram Heatmap

In this study, we adopt a vectorization approach called the Persistence Diagram Heatmap to process the persistent homology computed in the TDA Layer. This method can be applied without considering dimensions explicitly; however, to observe the characteristics of each dimension more precisely, we constructed the layer by combining feature vectors computed for each dimension.

Persistence Diagram Heatmap

A Persistence Diagram Heatmap visualizes the birth–death pairs from the persistence diagram in a 2D space and applies a heatmap using a kernel. For each birth–death pair ( b i , d i ) from the persistence diagram, the heatmap kernel value at a grid coordinate ( x , y ) is computed as follows:
H ( x , y ; b i , d i ) = exp ( x b i ) 2 + ( y d i ) 2 2 σ 2
Here, σ denotes the standard deviation of the Gaussian kernel. The value decreases exponentially as the distance between the grid coordinate ( x , y ) and the point ( b i , d i ) increases, and increases as the distance decreases. To construct the heatmap, a grid of size r × r within the range [ 0 , 1 ] × [ 0 , 1 ] is defined, with grid coordinates ( x j , y l ) calculated as follows:
x j = j r , y l = l r , for j , l = 0 , 1 , 2 , , r 1
The resulting Persistence Diagram Heatmap H ( x j , y l ) is computed as follows:
H ( x j , y l ) = i = 1 n exp ( x j b i ) 2 + ( y l d i ) 2 2 σ 2
Here, n is the number of birth–death pairs in the persistence diagram. This computation is applied to all dimensions, allowing the visualization of topological structures for each dimension. For example, a zero-dimensional heatmap reflects the birth and death of data points, a one-dimensional heatmap captures the connectivity and loops among data points, and a two-dimensional heatmap visualizes the formation and destruction of surfaces. The heatmap for the k-th dimension H k ( x , y ) is computed as follows:
H k ( x , y ) = i = 1 n k exp ( x b i , k ) 2 + ( y d i , k ) 2 2 σ 2
Here, n k is the number of points in the k-th dimension’s persistence diagram, and b i , k and d i , k are the birth and death coordinates of the k-th dimension’s data points. Finally, the Persistence Diagram Heatmaps for all dimensions are combined into a K-channel tensor:
H = { H 0 ( x , y ) , H 1 ( x , y ) , H 2 ( x , y ) , , H K ( x , y ) }
H represents the Persistence Diagram Heatmaps as a single K-channel image, as shown in Figure 7, enabling the topological features of multivariate time series data to be learned through an image classifier.
Figure 7. Visualization of Persistence Diagram Heatmap calculated based on Betti number.
The Persistence Diagram Heatmap intuitively represents the lifespan of topological features, making it easier to identify key structural characteristics of data without significant information loss. Additionally, by tuning parameters, it is possible to flexibly emphasize either local or global features of the data.

3.3. Weighted Sum of MTSC Layer and TDA Layer

The proposed model integrates the feature vector output from the TDA Layer with the output of the MTSC backbone model to achieve enhanced classification performance. In this process, the probabilities generated by the MTSC Layer and the TDA Layer are linearly combined according to a weighting factor w, as expressed in Equation (4):
P = w · P M T S C + ( 1 w ) · P T D A
Here, the probabilities from each layer are assigned weights, and the weighted sum is normalized to ensure the final result is a probability distribution summing to 1. This linear combination method leverages the complementary strengths of the two layers. The MTSC Layer excels at learning the overall patterns in the time series data, while the TDA Layer effectively captures the topological structures in the data, which reflect abnormal variations or fluctuations that may occur in the time series. By combining the strengths of different models based on the data characteristics and problem at hand, this approach provides more stable and improved performance compared to using a single model.
The proposed weighted summation method optimizes the complementary roles of MTSC and TDA, enabling robust classification performance across various data environments. Although mathematical optimization techniques could have been employed to identify the optimal combination ratio, this study primarily aims to explore the applicability of TDA in addressing the missing data problem in MTSC. Therefore, a heuristic experimental approach is adopted to analyze the effect of varying layer weight ratios.

4. Experiment

4.1. Experiment Setting

4.1.1. Evaluation Method

To evaluate the performance of the proposed model for classifying multivariate time series data, standard performance metrics such as Accuracy, Precision, Recall, and F1 Score were employed. These metrics played a crucial role in comprehensively measuring the model’s classification capability from various perspectives, providing a reliable basis for assessing its performance under different scenarios.
A key focus of this study is to analyze how the model performance changes when information from certain channels is randomly removed. For this purpose, one or two channels were randomly excluded, and the model’s ability to maintain classification performance despite the missing information was evaluated. The performance under these scenarios was compared against the baseline scenario where all channels are utilized. This comparison provided insights into the model’s robustness and resilience against partial data loss. All experiments were conducted under the assumption that the population follows a normal distribution, with statistical tests performed at a 5% significance level.

4.1.2. Dataset

The dataset used in this study was the “Rotating Machinery Type AI Dataset”, provided by the Ministry of SMEs and Startups, Korea AI Manufacturing Platform (KAMP), and the Department of Mechanical Engineering, KAIST (Daejeon, Republic of Korea). The dataset was accessed on 27 December 2021 via the official platform: https://www.kamp-ai.kr.
The data were acquired using the ERA (Educational Rotor Application) Test Station, a registered product of Signalink Co. (Seoul, Republic of Korea), shown in Figure 8 at a rotational speed of 1500 RPM. The detailed specifications of the rotor testbed are presented in Table 1.
Figure 8. Rotor testbed.
Table 1. Rotor testbed specifications.
The dataset was categorized into four states: Normal (type 1), Unbalance (type 2), Mechanical Looseness (type 3), and Unbalance with Mechanical Looseness (type 4). Unbalance refers to a state where the rotor’s mass center does not align with the center of rotation (shaft center), a significant cause of vibration in rotating machinery. Mechanical Looseness occurs when the machinery is not securely fixed to the ground or platform, resulting in horizontal misalignment.
Samples of the training data for each state are shown in Figure 9.
Figure 9. Training data samples.
The raw dataset comprises vibration signals collected over 140 seconds from four different parts of the rotor testbed (Figure 8) using smart vibration sensors with the specifications listed in Table 2.
Table 2. Specifications of the vibration sensor.
Each sensor was attached to a different section of the testbed, and the collected data included a time axis and four vibration channels, resulting in a total of five columns. This configuration produced a dataset with 754,477 rows, amounting to 3,772,385 data points in total. For model training, the data were segmented into windows of size 100 with an equal stride length. The resulting sliding window dataset was then split into training, validation, and test sets in a ratio of 80%, 10%, and 10%, respectively, to ensure reliable performance evaluation.

4.1.3. Configuration

In this study, a simple 1D CNN was employed as the backbone for the MTSC Layer, while ResNet-152 was used as the backbone for the image classifier. These backbone models were combined with the TDA Layer to evaluate the performance of the proposed model under different weighted summation ratios and assess its robustness in scenarios with missing sensor data. The objective was to systematically analyze how the complementary strengths of MTSC and TDA influence the model’s performance and how the model behaves under varying layer weight ratios.
The 1D CNN model used for the MTSC Layer was designed with computational efficiency in mind, consisting of four consecutive convolutional blocks. Each block includes a 1D convolutional layer with a kernel size of 2 and stride of 1, followed by batch normalization, ReLU activation, max pooling with a kernel size of 1, and dropout for regularization. The first three blocks use 100 filters, and the final block projects to the number of output classes. An adaptive average pooling layer reduces the temporal dimension, followed by a fully connected linear layer for classification. The full architecture is summarized in Table 3.
Table 3. Architecture of the 1D CNN used in the MTSC Layer.
For image classification based on topological signatures, the standard ResNet-152 architecture provided by PyTorch (version 2.1.2) was adopted. ResNet-152 is a deep convolutional neural network composed of an initial 7 × 7 convolutional layer, followed by four main stages of bottleneck blocks with identity and projection shortcuts. The model comprises a total of 152 layers, enabling efficient training of very deep architectures via residual learning. A global average pooling layer and a fully connected output layer are used at the final stage to produce class probabilities. An overview of the architecture is provided in Table 4.
Table 4. Overview of the ResNet-152 architecture used for image classification.
The first experiment focused on evaluating performance based on weighted summation ratios. Five different ratios were tested, ranging from using only the MTSC Layer to using only the TDA Layer. This allowed for a quantitative assessment of how the layer weighting impacts model robustness.
Next, the impact of parameter variations in the TDA vectorization method on performance was analyzed. Comparisons were made between scenarios where all channels were used and where one channel was randomly removed, to evaluate how parameter changes in the vectorization method affect performance and how well the model maintains performance under missing channel conditions.
Additionally, the proposed TDA model was compared against statistical imputation methods, such as KNN and Mean Imputation, as well as deep-learning-based imputation models, including VAEI and the MIWAE. Further experiments were conducted on a high-dimensional dataset using the BJTU-RAO Bogie Dataset, which includes nine-channel gearbox vibration data, to explore the scalability of the proposed TDA model to higher-dimensional problems [40].
All training and experiments were conducted under the conditions specified in Table 5. Performance testing under identical conditions (excluding the batch size) showed that processing the entire dataset took approximately 12 minutes. This translates to an average processing time of about 300 ms per single data sample, shaped as a tensor with a ( 1 , 4 , 100 ) shape. Recognizing that data decentralization is an increasingly important issue in real-world industrial environments, recent studies have explored its implications and solutions. For instance, Yang et al. (2024) proposed a targeted transfer learning framework using a distribution barycenter medium to address the challenges of decentralized data in intelligent fault diagnosis [41]. Inspired by such approaches, future extensions of the proposed method could benefit from decentralized processing strategies, which may potentially reduce computation time by enabling parallel or distributed model execution across local nodes.
Table 5. Training and testing conditions.

4.2. Experiment Results

4.2.1. Experimental Results According to Weighted Sum Ratio

The experimental results based on the weighted sum ratio using the Persistence Diagram Heatmap, which demonstrated better performance in previous experiments, are presented in Table 6, Table 7 and Table 8. These experiments measured the model’s performance from cases where all channels were utilized to cases where up to two channels were removed. The weighted sum ratio was adjusted across five levels starting with 1.0 (using only the MTSC Layer), then gradually increasing the weight of the TDA Layer to 0.7, 0.5, 0.3, and, finally, 0.0 (using only the TDA Layer). Parameters for computing the Persistence Diagram Heatmap were fixed at σ = 1.0 and resolution r = 128 .
Table 6. Performance when all channels were utilized.
Table 7. Performance when a single channel was removed.
Table 8. Performance when two channels were removed.
When all channels were utilized, the model achieved the highest classification performance with the MTSC Layer alone (weight ratio 1.0). However, as channels were removed, the performance degraded significantly, indicating that the MTSC Layer alone lacks robustness to channel loss.
In contrast, using only the TDA Layer (weight ratio 0.0) resulted in relatively lower classification performance but exhibited only a slight average performance drop of about 8–9% when channels were removed. This demonstrates that the TDA Layer provides better robustness under channel loss conditions, ensuring more stable classification performance.
The performance based on the weighted sum ratio showed that increasing the MTSC Layer’s weight led to higher basic classification performance. However, similar to the MTSC Layer alone, the performance dropped sharply (up to approximately 70%) when channels were removed. The weighted sum ratio of 0.3 provided the best balance between classification accuracy and robustness. While its classification performance was relatively low at 84% when all channels were used, it maintained a higher performance of approximately 74% and 62% when one or two channels were removed, respectively. This indicates that the weighted sum ratio of 0.3 achieves the highest robustness compared to other ratios.
Therefore, these experimental results highlight that, while the MTSC Layer excels in classification performance, it is vulnerable to channel loss. Conversely, the TDA Layer is robust to channel loss but has lower classification accuracy. Combining these layers ensures stable classification performance under channel loss conditions, enhancing the model’s overall reliability.

4.2.2. Experimental Results According to Parameters

The experimental results evaluating the changes in key parameters—standard deviation σ and resolution r—for computing the Persistence Diagram Heatmap are presented in Table 9. These experiments were conducted based on the weighted sum ratio of 0.3, which demonstrated the highest robustness in previous experiments. The F1 Score was used as the evaluation metric.
Table 9. F1 Score for each parameter.
The results show that the parameter combination σ = 1.0 and r = 512 achieved the highest performance in both scenarios: 92.41% when all channels were utilized, and 81.83% when a single channel was removed. However, this combination exhibited a performance drop of approximately 11% between the two scenarios. In contrast, the combination σ = 0.1 and r = 512 showed a smaller performance drop of around 6%, indicating better robustness. These findings suggest that higher resolutions (r) improve classification performance, while lower standard deviations ( σ ) increase robustness to channel loss.
Higher resolutions enhance classification performance by generating more detailed and sharper heatmaps, which effectively capture complex data patterns. On the other hand, lower standard deviations create localized heatmaps that are more resistant to changes in data structure, thereby maintaining stability under channel loss conditions. Conversely, as the Gaussian Kernel’s standard deviation increases, the heatmap captures features over a wider area, making it more sensitive to global changes in the dataset, which increases vulnerability to channel loss.
Thus, the choice of standard deviation σ and resolution r is a critical factor in balancing classification performance and robustness to channel loss. Proper tuning of these parameters ensures the design of a robust model capable of achieving high classification performance even under missing channel conditions.

4.3. Performance Comparison with Other Models

When a single channel was removed, the performance of the proposed TDA model was compared with existing imputation models using the F1 Score. For data imputation, KNN, Mean, VAEI, and MIWAE methods were applied, and the imputed data were fed into the MTSC Layer for classification tasks. The learning rate, batch size, and epochs for each method were set identically to those used in the TDA model. For the TDA model, performance was evaluated by combining a weight ratio of 0.3, standard deviation σ = 0.5 , and resolutions r = 128 , r = 256 , and r = 512 . The results are summarized in Table 10.
Table 10. F1 Score for each model.
The experimental results show that the F1 Scores of data imputed using KNN and Mean methods were approximately 63% and 50%, respectively. In contrast, the VAEI and MIWAE methods achieved higher F1 Scores of around 71% and 76%, respectively, demonstrating better performance. These results are comparable to the TDA model’s performance with a resolution of r = 128 . However, as the resolution increased to r = 256 and r = 512 , the TDA model outperformed both VAEI and the MIWAE.
This trend indicates that higher resolutions in the TDA model allow it to surpass the performance of VAEI and the MIWAE by capturing complex patterns in the data more effectively, thus improving classification performance.
Meanwhile, the lower F1 Scores observed with KNN and Mean methods can be attributed to their inability to capture the intricate data structures. These methods simply replace missing data without reflecting the underlying patterns, resulting in relatively poor performance. Although VAEI and the MIWAE leverage neural-network-based imputation to better capture the latent structure of the data, they fall short of the detailed feature extraction provided by the TDA model at higher resolutions.

4.4. Experiment with Additional Data

The results of additional experiments conducted using a subset of the BJTU-RAO Bogie Dataset, consisting of nine-channel gearbox vibration data, are presented in Table 11 and Table 12. These experiments were evaluated using the same metrics as the first experiment: Accuracy, Precision, Recall, and F1 Score. The weighted sum ratios were set to three levels (1.0, 0.5, and 0.0), and the performance was compared between cases where all channels were utilized and where three random channels were removed.
Table 11. All channels were utilized with the BJTU-RAO Bogie Dataset.
Table 12. Three channels were removed with the BJTU-RAO Bogie Dataset.
The experimental results indicate that, when all channels were utilized, the highest performance was achieved with a weighted sum ratio of 1.0, showing a similar trend to the first experiment. In contrast, when three channels were removed, the best performance was observed with a weighted sum ratio of 0.5, while the lowest performance degradation occurred with a weighted sum ratio of 0.0.
These results demonstrate that the proposed TDA model is scalable and robust for high-dimensional data. Specifically, the TDA model effectively learns complex patterns and adapts flexibly to missing data, as evidenced by the optimal performance achieved with a weighted sum ratio of 0.5 and the minimal performance degradation observed with a ratio of 0.0.
In this experiment, the nine-channel high-dimensional data yielded performance patterns similar to those observed in the four-channel data experiments. This further validates the proposed model’s suitability for processing high-dimensional data.

5. Conclusions

In conclusion, this paper proposes a robust AI model based on an MTSC model combined with TDA, designed to withstand sensor data loss. The experimental results showed that the model maintained the performance of traditional MTSC while experiencing only about a 10% performance drop in cases where approximately 25% of the sensors used in the data were missing. This demonstrates superior performance retention compared to using the MTSC model alone. In particular, the results confirmed the applicability of TDA in handling data deficiency problems, with observed performance improvements ranging from at least 4% to up to 30% over the baseline MTSC model depending on the missing sensor scenarios.
Future research will explore not only the heatmap kernel but also additional TDA methodologies, aiming to extend the potential of TDA beyond simple classification problems to a wider variety of tasks.

Author Contributions

Conceptualization, S.-Y.J.; Methodology, S.-Y.J.; Validation, S.-Y.J.; Writing—original draft, S.-Y.J.; Writing—review & editing, S.-Y.J. and J.-W.K. All authors have read and agreed to the published version of this manuscript.

Funding

This work was supported by the Korea Institute of Energy Technology Evaluation and Planning (KETEP) and grant funded by the Korean government (MOTIE) (20224B10100060, Development of Artificial Intelligence Vibration Monitoring System for Rotating Machinery). This work was supported by the National Research Foundation of Korea (NRF) and grant funded by the Korean government (MSIT) (no. RS-2022-00144441).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on KAMP (Korea AI Manufacturing Platform) at https://www.kamp-ai.kr, accessed on 27 December 2021.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Li, A.; Yang, X.; Dong, H.; Xie, Z.; Yang, C. Machine Learning-Based Sensor Data Modeling Methods for Power Transformer PHM. Sensors 2018, 18, 4430. [Google Scholar] [CrossRef]
  2. Khan, K.; Sohaib, M.; Rashid, A.; Ali, S.; Akbar, H.; Basit, A.; Ahmad, T. Recent trends and challenges in predictive maintenance of aircraft’s engine and hydraulic system. J. Braz. Soc. Mech. Sci. Eng. 2021, 43, 403. [Google Scholar] [CrossRef]
  3. Kafeel, A.; Aziz, S.; Awais, M.; Khan, M.A.; Afaq, K.; Idris, S.A.; Alshazly, H.; Mostafa, S.M. An Expert System for Rotating Machine Fault Detection Using Vibration Signal Analysis. Sensors 2021, 21, 7587. [Google Scholar] [CrossRef]
  4. Romanssini, M.; de Aguirre, P.C.C.; Compassi-Severo, L.; Girardi, A.G. A Review on Vibration Monitoring Techniques for Predictive Maintenance of Rotating Machinery. Eng 2023, 4, 1797–1817. [Google Scholar] [CrossRef]
  5. Doostmohammadian, M.; Meskin, N. Sensor Fault Detection and Isolation via Networked Estimation: Full-Rank Dynamical Systems. IEEE Trans. Control Netw. Syst. 2021, 8, 987–996. [Google Scholar] [CrossRef]
  6. Kiranyaz, S.; Devecioglu, O.; Alhams, A.; Sassi, S.; Ince, T.; Avci, O.; Gabbouj, M. Exploring Sound vs Vibration for Robust Fault Detection on Rotating Machinery 1. arXiv 2024, arXiv:2312.10742. [Google Scholar] [CrossRef]
  7. Tama, B.A.; Vania, M.; Lee, S.; Lim, S. Recent advances in the application of deep learning for fault diagnosis of rotating machinery using vibration signals. Artif. Intell. Rev. 2023, 56, 4667–4709. [Google Scholar] [CrossRef]
  8. Carlsson, G. Topology and Data. Bull. Amer. Math. Soc. 2009, 46, 255–308. [Google Scholar] [CrossRef]
  9. Zomorodian, A.; Carlsson, G. Computing Persistent Homology. Discrete Comput. Geom. 2005, 33, 249–274. [Google Scholar] [CrossRef]
  10. Otter, N.; Porter, M.A.; Tillmann, U.; Grindrod, P.; Harrington, H.A. A roadmap for the computation of persistent homology. EPJ Data Sci. 2017, 6, 17. [Google Scholar] [CrossRef]
  11. Liu, J.J.; Yao, J.P.; Liu, J.H.; Wang, Z.Y.; Huang, L. Missing data imputation and classification of small sample missing time series data based on gradient penalized adversarial multi-task learning. Appl. Intell. 2024, 54, 2528–2550. [Google Scholar] [CrossRef]
  12. Lepot, M.; Aubin, J.-B.; Clemens, F.H.L.R. Interpolation in Time Series: An Introductive Overview of Existing Methods, Their Performance Criteria and Uncertainty Assessment. Water 2017, 9, 796. [Google Scholar] [CrossRef]
  13. Flores, A.; Tito-Chura, H.; Centty-Villafuerte, D.; Ecos-Espino, A. Pm2.5 Time Series Imputation with Deep Learning and Interpolation. Computers 2023, 12, 165. [Google Scholar] [CrossRef]
  14. Park, J.; Müller, J.; Arora, B.; Faybishenko, B.; Pastorello, G.; Varadharajan, C.; Sahu, R.; Agarwal, D. Long-term missing value imputation for time series data using deep neural networks. Neural Comput. Appl. 2023, 35, 9071–9091. [Google Scholar] [CrossRef]
  15. Wang, J.; Du, W.; Yang, Y.; Qian, L.; Cao, W.; Zhang, K.; Wang, W.; Liang, Y.; Wen, Q. Deep Learning for Multivariate Time Series Imputation: A Survey. arXiv 2024, arXiv:2402.04059. [Google Scholar] [CrossRef]
  16. Bansal, P.; Deshpande, P.; Sarawagi, S. Missing value imputation on multidimensional time series. Proc. VLDB Endow. 2021, 14, 2533–2545. [Google Scholar] [CrossRef]
  17. Cao, W.; Wang, D.; Li, J.; Zhou, H.; Li, L.; Li, Y. BRITS: Bidirectional Recurrent Imputation for Time Series. In Proceedings of the Advances in Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018; Volume 31. [Google Scholar] [CrossRef]
  18. Qiu, Y.L.; Zheng, H.; Gevaert, O. Genomic data imputation with variational auto-encoders. Gigascience 2020, 9, giaa082. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  19. Mattei, P.; Frellsen, J. MIWAE: Deep Generative Modelling and Imputation of Incomplete Data Sets. In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; Volume 97, pp. 4413–4423. Available online: https://proceedings.mlr.press/v97/mattei19a.html (accessed on 4 February 2019).
  20. Rundo, F.; Trenta, F.; di Stallo, A.L.; Battiato, S. Machine Learning for Quantitative Finance Applications: A Survey. Appl. Sci. 2019, 9, 5574. [Google Scholar] [CrossRef]
  21. Morid, M.; Sheng, O.; Dunbar, J. Time Series Prediction using Deep Learning Methods in Healthcare. arXiv 2021, arXiv:2108.13461. [Google Scholar] [CrossRef]
  22. Pinciroli Vago, N.O.; Forbicini, F.; Fraternali, P. Predicting Machine Failures from Multivariate Time Series: An Industrial Case Study. Machines 2024, 12, 357. [Google Scholar] [CrossRef]
  23. Zhang, D.; Gao, J.; Li, X. Multivariate time series classification with crucial timestamps guidance Author links open overlay panel. Exp. Syst. Appl. 2024, 255 Pt B, 124591. [Google Scholar] [CrossRef]
  24. Bagnall, A.; Lines, J.; Bostrom, A.; Large, J.; Keogh, E. The great time series classification bake off: A review and experimental evaluation of recent algorithmic advances. Data Min. Knowl. Disc. 2017, 31, 606–660. [Google Scholar] [CrossRef]
  25. Formisano, E.; De Martino, F.; Valente, G. Multivariate analysis of fMRI time series: Classification and regression of brain responses using machine learning. Magn. Reson. Imaging. 2008, 26, 921–934. [Google Scholar] [CrossRef] [PubMed]
  26. López-Oriona, Á.; Vilar, J.A. F4: An All-Purpose Tool for Multivariate Time Series Classification. Mathematics 2021, 9, 3051. [Google Scholar] [CrossRef]
  27. Wang, Z.; Yan, W.; Oates, T. Time series classification from scratch with deep neural networks: A strong baseline. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 1578–1585. [Google Scholar] [CrossRef]
  28. Ju, J.; Liu, F.-A. Multivariate Time Series Data Prediction Based on ATT-LSTM Network. Appl. Sci. 2021, 11, 9373. [Google Scholar] [CrossRef]
  29. Yazdanbakhsh, O.; Dick, S. Multivariate Time Series Classification using Dilated Convolutional Neural Network. arXiv 2019, arXiv:1905.01697. [Google Scholar] [CrossRef]
  30. Karim, F.; Majumdar, S.; Darabi, H.; Harford, S. Multivariate LSTM-FCNs for time series classification. Neural Networks. arXiv 2018, arXiv:1801.04503. [Google Scholar] [CrossRef]
  31. Le, X.-M.; Luo, L.; Aickelin, U.; Tran, M.-T. ShapeFormer: Shapelet Transformer for Multivariate Time Series Classification. arXiv 2024, arXiv:2405.14608. [Google Scholar] [CrossRef]
  32. Xiao, Z.; Xu, X.; Xing, H.; Zhao, B.; Wang, X.; Song, F.; Qu, R.; Feng, L. DTCM: Deep Transformer Capsule Mutual Distillation for Multivariate Time Series Classification. IEEE Trans. Cogn. Dev. Syst. 2024, 16, 1445–1461. [Google Scholar] [CrossRef]
  33. Foumani, N.M.; Tan, C.W.; Webb, G.I.; Salehi, M. Improving position encoding of transformers for multivariate time series classification. Data Min. Knowl. Disc. 2024, 38, 22–48. [Google Scholar] [CrossRef]
  34. Wasserman, L. Topological Data Analysis. Ann. Rev. Stat. Appl. 2016, 5, 501–532. [Google Scholar] [CrossRef]
  35. Chazal, F.; Michel, B. An Introduction to Topological Data Analysis: Fundamental and Practical Aspects for Data Scientists. Sec. Mach. Learn. Artif. Intell. 2021, 4, 667963. [Google Scholar] [CrossRef] [PubMed]
  36. Motta, F.C. Topological Data Analysis: Developments and Applications. In Advances in Nonlinear Geosciences; Tsonis, A., Ed.; Springer: Berlin/Heidelberg, Germany, 2018. [Google Scholar] [CrossRef]
  37. Patania, A.; Vaccarino, F.; Petri, G. Topological analysis of data. EPJ Data Sci. 2017, 6, 7. [Google Scholar] [CrossRef]
  38. Singh, Y.; Farrelly, C.M.; Hathaway, Q.A.; Leiner, T.; Jagtap, J.; Carlsson, G.E.; Erickson, B.J. Topological data analysis in medical imaging: Current state of the art. Insights Imaging 2023, 14, 58. [Google Scholar] [CrossRef]
  39. Gidea, M. Topological Data Analysis of Critical Transitions in Financial Networks. In Proceedings of the 3rd International Winter School and Conference on Network Science: NetSci-X 2017 3; Springer Proceedings in Complexity; Springer International Publishing AG: Cham, Switzerland, 2017; pp. 47–59. [Google Scholar] [CrossRef]
  40. Wang, B.; Qin, Y.; Ding, A. Introduction to BJTU-RAO Bogie Datasets; technical report; Beijing Jiaotong University: Beijing, China, 2024. [Google Scholar] [CrossRef]
  41. Yang, B.; Lei, Y.; Li, X.; Li, N. Targeted transfer learning through distribution barycenter medium for intelligent fault diagnosis of machines with data decentralization. Exp. Syst. Appl. 2024, 244, 122997. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.