1. Introduction
Artificial intelligence enables data-driven approaches to predict equipment failures, which is called predictive maintenance [
1]. It often helps the industry reduce the risk of equipment failures. Consequently, it mitigates costly and unscheduled downtime. This shift help the industries move away from traditional reactive or preventive maintenance techniques to predictive maintenance. In more industries, traditional defect management focuses on the repair or periodic replacement of equipment after failure, such as replacing worn-out motors [
2]. In many industrial setups, predictive maintenance can be carried out using the information received through the data collected from various sensors. The collected data are then analyzed to track machinery health and performance in real time [
3,
4]. These sensors can measure several important factors, including temperature, vibration, pressure, and flow. Patterns, abnormalities, and possible signs of deterioration or future malfunctions are then analyzed using the collected data through installed sensors. Conventionally, equipment maintenance is performed following the defined schedules using the Industrial Internet of Things (IIoT) [
4]. Due to the predictive strategies, the maintenance structure is rapidly changing with the technological advancements of the fourth industrial revolution [
5]. The IIoT technology is employed in a business setting where digital technology, internet connectivity, and cloud platforms are integrated with real industrial machines, equipment, sensors, and other digital assets [
6,
7]. The main goal of the IIoT is to provide businesses with the tools they need to collect, monitor, analyze, and use data to streamline operations, increase productivity, and make informed decisions, especially for small devices or when a maintenance problem occurs during a device’s operation [
8]. The IIoT also has the potential to support new business models and revenue streams. The data collected from the sensors embedded with the industrial equipment can be used to provide particular services, including healthcare services [
9,
10], performance analysis, and remote monitoring [
11]. This advancement to data-driven services has the potential to increase competitiveness, enhance user experience, and identify opportunities for creative connections and collaboration [
12]. The IIoT continues to change the face of industries in manufacturing, energy and agriculture. It can help the manufacturers achieve greater precision and flexibility by implementing IIoT-enabled smart manufacturing approaches. For example, in the energy industry, it is utilized to optimize energy, strengthen the integration of renewable energy sources, and enhance network performance [
13,
14].
For most industrial environments, the predictive maintenance techniques are employed to reduce costs as well as increase equipment performance and life. As a result, it reduces risk and increases safety. By tracking abnormalities, machine learning can predict maintenance problems such as unusual vibrations or temperature spikes [
15,
16]. Similarly, the IIoT gives opportunities to design new architectures as resource methods [
17] as well as increase system performance and reduce cyber threats [
18]. The IIoT brings connectivity, automation and data-driven decision making, bringing significant changes to manage their operations such as remote monitoring systems. The IIoT enables businesses to enhance their productivity levels, efficiency and creativity by bring the automation and the integration of industrial equipment such as machines, sensors and gadgets into a networked environment. These connected sensor nodes collect real-time data from industrial operations, including manufacturing processes, supply chains, logistics and energy consumption. The data are then sent to a centralized platform where powerful analytics and machine learning algorithms process it into actionable insights [
19]. Machine learning models can be used to predict the remaining useful life (RUL) using advanced deep learning techniques [
20,
21].
Predictive maintenance is an important technique in industrial systems to avoid unplanned equipment breakdown. Compared to preventive or reactive maintenance, where it acts after the failure has occurred or on a time-based schedule, predictive maintenance enables timely action based on actual machine conditions. This is accomplished mainly through technological advancements in the IIoT, where sensor-enabled devices monitor operation parameters such as temperature, pressure, vibration, and flow in real time. Real-time sensor measurements are the foundation of smart decision making and predictive fault prediction. While they are helpful, predictive maintenance models are typically marred by imbalanced datasets, which is primarily because failure occurrences are sparse compared to regular operating information. Machine learning models trained on such skewed data will end up performing badly in identifying minority class instances, which in this scenario refer to the failure states of interest. To counter the problem, synthetic data generation techniques are employed to synthetically oversample the minority (failure) classes [
22].
The proposed GenIIoT is a two-step hybrid framework that solves the tradeoff between the stability of the SMOTE and the realism of GAN-based data synthesis. The SMOTE is first applied to expand the minority class and stabilize GAN training, which would otherwise fail due to the extremely small number of fault samples. In the second stage, the GAN performs nonlinear refinement on SMOTE-generated data and learns the complex IIoT fault distribution. It generates high-fidelity synthetic samples () that are more diverse and realistic than interpolation-based methods, effectively addressing the limitations of SMOTE alone.
Two algorithms have been employed for synthetic data generation: the synthetic minority oversampling technique (SMOTE) and generative adversarial networks (GANs). However, there are some practical differences when applied to IIoT-based predictive maintenance. The SMOTE is light in computation because it creates new instances of the minority class by linear interpolation. It corrects the imbalance but does not exactly follow the variability to replicate the real-world complexity of faults [
23]. Conversely, GANs generate data more close to real data, while the variational autoencoders (VAEs) produce overly smooth data. GANs can better learn the data distribution and thus capture complex fault behaviors. Similarly, while denoising diffusion probabilistic models (DDPMs) have high-quality generation capability, the high training cost and slow generation speed might limit their use in real-time industrial environments [
23].
GANs can create high-fidelity synthetic data that are highly comparable to real IIoT fault patterns, which is critical to train trusted proactive models. Compared to VAEs, which create blurred or averaged samples, or DDPMs, which are too computationally demanding for resource-constrained IIoT edge devices, GANs offer a balance between sample fidelity and efficiency [
12]. The SMOTE extends GANs to address class imbalance in fault datasets, which is a common issue in the IIoT, through local interpolation. Although DDPMs or VAEs can also generate synthetic data, the simplicity and low computational complexity in SMOTE make it more efficient for real-time fault management. The main contribution is the hybridization of GANs and the SMOTE to take advantage of their complementary strengths where GANs produce diverse fault scenarios and the SMOTE fine-tunes minority-class samples for maximizing classifier fairness [
24]. Furthermore, it addresses the dynamic nature of IIoT faults’ non-stationary distributions using the feedback mechanism wherein the output of GAN is recursively validated against real-time fault logs for adaptive sample quality [
19,
25,
26].
Abundant failure data are required for particular equipment for its on-time management using predictive maintenance. A GAN is a technique used to generate data. Similarly, SMOTE is a data-generated technique used in cases of data inadequacy. Machine learning models such as support vector machine (SVMs), k-nearest neighbors (KNN), and decision tree classifiers are used for predictive maintenance. Deep learning models such as convolutional neural networks (CNNs) and long short-term memory (LSTM) are used for predictive maintenance. The proposed method aims to develop a new and useful predictive maintenance system to assist in equipment maintenance. The core contributions of the proposed work are outlined below:
A generative AI-enabled predictive maintenance system framework is proposed to proactively predict equipment faults before they occur within an IIoT network.
The generative AI framework consists of data-generated techniques, including the SMOTE and GANs: the former is skilled at balancing class distribution and producing synthetic failure data, while the latter enables the generation of failure data with unmatched realism and diversity. It mimics the intricate behaviors of real industrial machinery, producing predictive data for maintenance.
Three different data categories, i.e., real, GAN-generated, and SMOTE-generated, are applied to machine learning to evaluate their precision, recall, F1-score, and accuracy performance.
The IIoT network performance is measured using inter-data with a machine learning models comparison as well as an inter-model comparison of machine learning models, such as SVM, KNN, and DTC, and deep learning models, such as CNN and LSTM.
2. Related Work
Class imbalance, where minority groups are underrepresented, creates major challenges in real-world machine learning, leading to biased models and unreliable evaluations. The data augmentation techniques such as the SMOTE and GANs individually help mitigate this problem, but they have notable limitations. This review introduces a unified taxonomy that classifies the causes, types, and effects of class imbalance in various ML tasks. It also explores recent advances in hybrid models combining the SMOTE and GANs and evaluates related datasets, metrics, and methodologies. This study provides practical insights and outlines future research directions to effectively manage class imbalance in machine learning applications [
27]. In the digital age, effective transaction fraud detection (TFD) is essential to ensure financial security. The significant class imbalance, in which the number of legitimate transactions significantly exceeds the number of fraudulent transactions, is a significant challenge for TFD models to accurately identify fraudulent patterns. While existing sample-balancing strategies effectively address class imbalance in many contexts, they often fall short in TFD due to the sophisticated concealment tactics of fraudsters, leading to apparent behavioral overlap between fraudulent and legitimate transactions. In this paper, a novel generative adversarial network-based hybrid sampling method (GANHS) is proposed to effectively address the class imbalance issue. GANHS employs a dual-discriminator generative adversarial network to generate synthetic samples that accurately reflect the characteristics of fraudulent activity, while an adaptive neighborhood-based undersampling technique refines these samples to minimize overlap with legitimate samples. By producing high-quality samples, this hybrid strategy not only increases the model’s capacity to identify fraud tendencies but also strengthens its resistance to highly concealed fraudulent activity. With gains of 0.5–8.7% in mean F1-score and 1.0–7.0% in mean G, GANHS outperforms its rivals in experiments conducted on public and real-world datasets, underscoring its enormous potential to increase the dependability and efficacy of TFD systems in intricate and high-risk financial scenarios [
28].
A significant contribution to predictive maintenance in the oil and gas industry was made by [
29] who used large language models (LLMs) and generative adversarial networks (GANs) to enhance early problem diagnosis and maintenance planning. High sensor noise and inadequately tagged failure data are frequent issues for traditional predictive maintenance systems, the study emphasized. Paroha proposed employing GANs to improve dataset balance, produce realistic-looking artificial sensor signals, and help machine learning models better identify rare failure patterns. Additionally, textual reports and maintenance logs are examined using LLMs, which provided contextual information to support forecasts based on sensors. This coupled approach demonstrated improved accuracy, stability, and operational reliability, highlighting the potential of generative AI to support data-driven and energy-efficient decision making in industrial settings [
30].
Due to Industry 4.0’s rapid development, there are now more opportunities than ever before to apply artificial intelligence (AI) in industrial processes. One of the AI paradigms, generative adversarial networks (GANs), has developed into a powerful tool that can create simulated datasets that appear realistic. Predictive maintenance, which aims to foresee likely equipment breakdowns and optimize maintenance practices, frequently faces issues with data shortages and class imbalance. This is particularly valid for critical but rare failure scenarios. This study explores how GANs can simulate equipment failure scenarios, increase model accuracy, and close gaps in predictive maintenance datasets. GAN-based techniques are investigated for generating synthetic sensor signals, failure patterns, and degradation trajectories to enhance prediction models. The technique includes a detailed analysis of GAN topologies, training strategies, and validation procedures. The discussion evaluates GAN applications across several industries, points out problems like overfitting and domain adaptation, and considers the ethical and legal implications of producing synthetic data. Findings indicate that GANs could transform predictive maintenance by reducing unplanned downtime and enabling sound, data-driven decision making [
28].
To increase the forecast accuracy necessary for predictive maintenance, a number of techniques have been put out and progressively improved. Cutting-edge research is conducted on deep adversarial learning, intelligent manufacturing systems, predictive maintenance strategies, and AI techniques, such as RNNs, that improve the lifespan and functionality of industrial machinery [
31]. Because of the growing number of sensors and functional complexity, intelligent manufacturing systems are more likely to make mistakes that could result in large losses [
32]. Conventional maintenance techniques, like reactive and preventive maintenance, are ineffective because they take up too much time and money and cannot handle forecasting and real-time decision-making issues at the same time [
33]. A potent substitute is predictive maintenance (PdM), which is made possible by machine learning algorithms. For instance, LSTM-GAN, a revolutionary deep adversarial learning technique, has been proposed to address problems like mode collapse and vanishing gradients. While the GAN component aids in the generation of synthetic fault data, hence resolving the issue of data imbalance and lack of error samples, the LSTM component gains from long-term dependency learning, which improves forecasting [
34].
Prediction and maintenance decisions are the two key components that make up the PdM model. These modules employ AI to assess the state of the machine and estimate the likelihood of faults, which are subsequently converted into workable maintenance schedules. With a 99% fault prediction accuracy in equipment diagnostics, LSTM-GAN models have shown great promise for industrial IoT-based settings [
35]. The significance of creating synthetic data with GANs has also been highlighted by recent research. To guarantee that these synthetic data accurately reflect the statistical characteristics of actual data, quality control is necessary. When paired with GANs, data balancing techniques like the synthetic minority oversampling technique (SMOTE) enhance learning by resolving class imbalance, which is a prevalent problem in fault detection datasets. A hybrid SMOTified-GAN method is proposed that improves F1-scores in classification tasks by 9% [
36]. Industrial equipment’s remaining useful life (RUL) is predicted using RNN-based architectures such as LSTM and GRU. Tasks involving temporal dependencies in sequence learning are especially well suited for these models. For instance, multivariate time-series defect data have been synthesized using gated recurrent units (GRUs) in conjunction with conditional GANs (C-GANs), which has improved RUL estimation accuracy by 0.15. Predictive defect management in the context of smart construction includes Building Information Modelling (BIM) in conjunction with GANs and IoT. These technologies enhance fault monitoring systems and enable real-time data flow for facility management. Although BIM and IoT are becoming more popular, there are still sparse yet promising studies integrating these technologies into federated learning systems. Automated generative models and unsupervised learning have also been investigated for steganography and facial recognition applications. Visual inspection systems for manufacturing environments may benefit from the cross-applications of these GAN-based models, which concentrate on feature disentanglement and latent space manipulation [
37].
Intensive errors that may cause extremely large losses are pressured by the intelligent manufacturing system’s increasing number and functional diversification. The disadvantage of passive, or any other conventional maintenance methods, is that the aforementioned processes take a lot of time and are unproductive. In addition, it is almost impossible to predict the forecasting and maintenance requirements of an intelligent manufacturing system simultaneously, whereas the basic preventive maintenance process utilizes solely one model [
9]. Consequently, this paper proposes an innovative predictive maintenance (PdM) method: an LSTM-GAN deep adversarial learning method planned to be more prolonged. Since the LSTM network has a long-term dependency, it can cure vanishing gradient and mode collapse from GAN. The elements of the predictive maintenance model are the two prediction models and the maintenance decision model. Specifically, certain assessments concerning the state of the machine and possible failures can be provided by the parameters created using the above-mentioned prediction models. Later, the maintenance decision model communicates the maintenance plan to the maintenance personnel. LSTM-GAN is an intelligent manufacturing system for equipment diagnostics and prediction. In particular, the error prediction accuracy reaches up to 99%.
By applying IIoT data and RUL evaluations, PdM and AI enhance the industry’s productivity with minimal maintenance downtime [
8]. The majority of printed studies take into account the availability of training data, as well as several normal and erroneous samples, when analyzing different scenarios for machine state. Although there are no error data in the real-world scenarios, it is suggested that a non-uniform training set should be adopted. Since this failure leads to the non-prediction of errors in RUL estimation methodologies, this problem results in erroneous outcomes [
38,
39]. A brand new forecasting paradigm grounded on DGRU networks and C-GANs is introduced. Multivariate errors can be instantiated by the framework, and thus, the data imbalance issue can be solved by the proposed framework, enabling the accurate estimation of RULs of complex systems. As opposed to earlier works using imbalanced data on the C-MAPSS dataset [
40], the authors showed that DGRU training, data augmentation, and learning error samples with a noise basis distribution for RUL prediction have increased accuracy by 0.15 [
11].
Similarly, one of the target areas of predictive fault management is the monitoring and maintenance of buildings. For this purpose, Building Information Modeling (BIM) is linked with GANs [
41]. According to prior research, building maintenance management mainly involves appropriate maintenance parameters and standards. However, these limitations also imply that the purpose of this research is to work with modern technology to introduce a program of predictive maintenance. BIM and the IoT can enhance FMM by exchanging data, facilitating the management of facilities, and exploring efficient solutions [
42,
43]. While BIM and the IoT have been employed in the industry, and multiple concepts employing the mentioned approaches exist, their integration in federated machine learning is still rather limited. A PdM planning framework is introduced, centered on analyzing data for FMM to enhance the maintenance approaches regarding construction equipment. It consists of two layers: the database layer and the information layer. The basis of this technology is in the IoT and BIM. The information layer gathers information from the FM systems, IoT networks, and BIM models used to support the predictive maintenance notations. The application layer can be divided into four submodules [
44].
A new architecture has been developed to identify human faces using the unsupervised and automated learning of various facial features [
45]. It also aids in creating new images of variants using freckles and hair. It makes synthesis controllable. It is important to note that the utilized GAN generator advances the state of the art in conventional distribution quality attributes, and it results in established changed characteristics and a sufficiently accurate extraction of the latent dimensions of variability. It deals with interpolation and disentanglement; for these qualities to be enhanced, a new method was developed, which is better for any generator that is used in GANs. Distortion in steganography is carried out using GANs, whereas a media object that is used in digital communication to accommodate other information refers to steganographic distortion. This is an important constituent of the covert communication method whose name is steganography. The objective of such systems is to conceal the existence of secret communication.
Alhiyari and Domartzaki introduced SMOGAN (synthetic minority oversampling with GAN refinement), which is a two-stage oversampling framework for imbalanced regression problems. In the first step, SMOGN generates initial synthetic samples in sparse regions using interpolation. In the second step, a distribution-aware GAN (DistGAN) refines these samples by reducing MMD losses to better align them with the real data. This blend of SMOTE-based and GAN-based techniques improves the diversity and realism of synthetic data, enhancing model performance on rare target values [
46]. GACNet (Generate Adversarial-Driven Cross-Aware Networks) improves wheat variety identification from hyperspectral images. The framework integrates a semi-supervised GAN (SSGAN) for realistic data augmentation and a cross-aware attention network (CANet) for effective feature extraction using 3D and 2D convolutions with an embedded attention mechanism. Using the Hyperspectral Wheat Variety Dataset (HWVD) of 4560 samples from 19 categories, GACNet achieved superior accuracy over existing methods, demonstrating the strength of GAN-based enhancements and attention-driven learning in hyperspectral classification [
31]. AGANet (Attention-Guided Generative Adversarial Network) is used to overcome data scarcity in hyperspectral corn seed detection. The model integrates attention modules and a classifier within the GAN framework to generate realistic, class-specific hyperspectral images, enhancing spatial feature extraction and reducing the need for large labeled datasets. Experimental results showed that AGANet produces highly realistic synthetic samples, improving data augmentation for deep learning-based agricultural imaging [
32].
In digital twin (DT) systems, generative adversarial networks (GANs) play an important role by generating synthetic sensor data and simulating rare fault scenarios that are difficult to capture in real operation [
47]. The generator creates realistic data reflecting potential system behavior, while the discriminator differentiates between real and synthetic data, thereby continuously improving the quality of generation. This process increases data diversity, addresses data imbalances, and supports more accurate fault detection and predictive maintenance. By integrating GANs, DT systems can model degradation, predict failures in advance, and enable real-time, data-driven decision making in Industry 4.0 and 5.0 environments [
48].
Conditions for applying remote sensing data for land cover include the classification of the diverse and detailed land use and cover classifications and reliable classifications [
24]. In the field of terrestrial remote sensing, the random forest, a strongly developed method of machine learning classifier, has not been globally publicized as the common pattern recognition mechanism. Also, the specific functions incorporated in the RF can perform the basic methods for imputing missing values among several data analysis techniques. Other data analysis tasks performed by the RF include survival, regression, classification, and unsupervised learning. In detail, this paper seeks to discuss the efficiency of the algorithm using factors such as the mapping precision level, the mapping scale, and the level of noise. The Kappa Coefficient values of the RF of and overall accuracy of the result are 92% and 0, respectively. More than just improving the RF model with one decision tree, a full improvement is noticed at
p < 0.00001 McN.
Another challenge that affects classifiers is a severe class imbalance in the dataset. This is due to most of the positive classifications that lead to a poor prediction for the training sets with a high True False Rate (TPR) but low True Negative Rate (TNR) [
36]. In the proposed method, the SMOTE and GANs are combined with the knowledge transfer method, and a two-stage resampling method is introduced. In case higher-order GANs have problems with minority data alone, we transform overgeneralized or unrealistic samples of SMOTE into real data. Hence, the approach involves the use of the above-mentioned small set of pre-sampled data obtained from the SMOTE to train rather than randomly generated data. Experimental works presented for different benchmark datasets reveal that the quality of the samples belonging to the minority class is better. It noted a
improvement to the comparatively conventional algorithm in the F1-score evaluation criterion for the classification of given data using a neural network; the resampled data from the minority class must be in the same proportion as the data in the majority class. Furthermore, to analyze the results on the entire dataset without applying any technique of data augmentation, experiments are performed [
17].
4. Results and Discussion
The similarity between real and synthetic data is evaluated using a distributional comparison. Therefore, the distribution of real and generated data is compared using a histogram. The histogram distribution of key features is illustrated in
Figure 2, where the blue and red represent real and GAN-generated distributions, respectively. The alignment of these curves demonstrates that the GAN-generated data resemble the underlying distribution of the original data. This method is commonly used to assess synthetic data fidelity. By integrating these evaluations, the synthetic fault data generated by the GAN closely mirror the statistical properties of the real dataset, supporting its applicability in training more robust predictive maintenance models. The closer the curves align, the better the quality of the data generated. The GAN model hyperparameters are listed in
Table 4.
Three categories of data are used: AI4I, APS, and SECOM. The models’ performance is evaluated by feeding the three kinds of data to the models. For example, decision tree classifiers are fed with three types of data and categorized as decision tree classifier with normal data (DTC-normal), decision tree classifier with SMOTE-generated data (DTC-SMOTE), and decision tree classifier with GAN-generated data (DTC-GAN). Hyperparameter tuning for all conventional baseline models, including SVM, KNN, decision tree, random forest, and 1D CNN, in addition to the proposed LSTM architecture, is performed by varying key parameters, such as kernel type, regularization strength, gamma settings, neighborhood size, weighting scheme, tree depth, estimator count, learning rate, batch size, hidden units, dropout rate, and loss functions, with optimal configurations summarized in
Table 7. The effectiveness of these models is assessed with key performance metrics as shown in
Figure 6, including the accuracy and recall rate, as shown in
Figure 7, the F1-score as depicted in
Figure 8, and the level of precision as illustrated in
Figure 9 using the hyperparameters listed in
Table 7. In
Figure 9, several models show accuracy and precision values close to 0.999 or 1.0, which is unexpected given the dataset’s strong class imbalance. However, synthetic data generation methods, including GANs and the SMOTE, are used to balance the APS training dataset, which enhanced the model’s ability to correctly identify minority class instances. Consequently, the elevated precision and accuracy scores reflect the effectiveness of the models when trained on a balanced dataset. Mathematically,
The results of the hyperparameter optimization show differences amongst the models that are examined with the LSTM obtaining the maximum accuracy of 98.84%. Its balanced design (0.0005 learning rate, 32 batch size, 64 hidden units, 0.2 dropout, and binary cross-entropy loss) allowed for an efficient modeling of temporal relationships while reducing overfitting, which is responsible for this performance. With 100 estimators and a depth of 20, random forest came in second at 96.3%, capturing intricate feature interactions without introducing undue volatility. Using an RBF kernel with C = 1 and gamma adjusted to scale, the SVM achieved 95.8% accuracy, offering strong nonlinear decision limits. With five neighbors and distance-based weighting, which prioritized nearby, more pertinent samples, KNN fared best with 93.5%. The 1D CNN closely matched this at 93.8%, with 64 filters, a kernel size of 3, and a 0.0005 learning rate, enabling an effective extraction of local temporal–spatial patterns. The decision tree achieved 90.4% accuracy with a depth of 10 and a minimum split of 5, balancing complexity and generalization. Overall, the results confirm that while all models benefited from targeted hyperparameter tuning, deep learning approaches and specifically, LSTM outperformed classical methods, with random forest offering the strongest traditional alternative.
DTC-Normal demonstrates rather low performance. The reasoning behind this is the presence of very scarce failure data affecting the models’ performance, as shown in
Figure 6. Comparatively, DTC-SMOTE and DTC-GAN achieve better results than DTC-Normal in terms of recall, accuracy, and F1-scores, as shown in
Table 8. XGBoost achieves the highest accuracy of 0.9978, which is followed by KNN with an accuracy of 0.9966, and random forest with 0.9957. The deep learning models, such as LSTM and DNN, achieve 0.9884 and 0.9810, respectively, and perform well but were outperformed by the ensemble methods. On the other hand, SVM had the lowest accuracy of 0.9822.
The computational efficiency of the models is compared in terms of the number of parameters, FLOPS used, and training time. The DNN has the fewest parameters of 11,009, while random forest has the most: 169,162 nodes. LSTM had a moderate count of 63,329. Comparatively, only neural models have an approximate FLOPs comparison, where LSTM has 126,560 and DNN has 21,888, as illustrated in
Table 8. Furthermore, the training time of XGBoost is the lowest with 12.42 s, while SVM with 2587.13 s was the highest among them all due to an extensive grid search to achieve the optimal hyperparameters. The deep learning models have moderate training times due to hyperparameter tuning. The models’ comparison on test accuracy, number of parameters, FLOPS, and training time is summarized in
Table 8.
Predictive maintenance (PdM) emerges as a critical application area within the Industrial Internet of Things (IIoT), relying heavily on time-series sensor data to forecast potential equipment failures. To properly analyze the temporal data, deep learning (DL) architectures and traditional machine learning (ML) algorithms were investigated. Long short-term memory (LSTM), convolutional neural networks (CNN), support vector machines (SVMs), K-nearest neighbors (k-NN), and decision trees (DTs) are among the models that have undergone considerable evaluation.
The suggested LSTM model is compared to a number of cutting-edge classification methods documented in recent research in
Table 8. With an accuracy of 98.84% and balanced precision–recall values (99% and 99%, respectively), the results show that the suggested LSTM has significantly improved over the traditional method. In contrast, the random forest model obtained an F1-score of 96.2% and accuracy of 96.3%, while KNN and the 1D CNN obtained accuracies of 93.5% and 93.8%, respectively, and the SVM model also showed impressive results with 95.8% accuracy. On the other hand, the decision tree showed limits in managing the dataset’s complexity, recording the lowest accuracy of 90.4%. While some conventional models are very accurate, the suggested LSTM model has significant advantages when it comes to managing the data’s temporal patterns and sequential dependencies. Because of these features, it is especially well suited for jobs where the sequence and timing of events are crucial to the accuracy of predictions.
Model interpretability is frequently just as crucial in industrial settings as prediction accuracy. Conventional machine learning methods like SVM, KNN, and decision trees are thought to be more interpretable and transparent. Decision trees, for instance, offer easily verifiable decision routes and rules that are accessible by humans as well as domain experts. Similar to this, SVMs provide decision boundaries that are either linear or kernel-based, which facilitates comprehension of the model’s decision-making process. On the other hand, because of their deep architectures and nonlinear transformations, DL models such as CNN and LSTM are frequently regarded as black-box systems. These models are more accurate when dealing with complicated data, but their lack of transparency may make them unsuitable for use in safety-critical or regulated applications [
56]. This has increased interest in Explainable AI (XAI) methods, which try to improve the interpretability of DL models. LIME (Local Interpretable Model-Agnostic Explanation) is a popular XAI technique [
57]. as well as SHAP (SHapley Additive exPlanation), which offer post hoc justifications for model forecasts. These techniques provide a means of striking a compromise between interpretability and performance for both traditional and deep learning models. A thorough taxonomy of XAI approaches was presented by Arrieta et al., who emphasized that interpretability and accuracy should be taken into consideration when choosing models in crucial domains [
56].
The lack of labeled fault data and its imbalance are two of PdM’s biggest problems. Researchers have used generative adversarial networks (GANs) to augment data in order to overcome this. In order to improve model generalization and robustness, conditional GANs (cGANs) are very good at producing synthetic sensor data that correspond to uncommon fault circumstances [
37]. Additionally, it has been demonstrated that cGANs and LSTM networks can predict RUL more accurately in situations with little labeled data. Another method, the SMOTified-GAN framework, improves fault classification in unbalanced industrial datasets by combining ensemble learning approaches with GAN-based oversampling. These techniques emphasize how crucial synthetic data generation and hybrid learning are to the development of PdM systems [
36].
Three distinct datasets, KNN-Normal, KNN-SMOTE, and KNN-GAN, were fed into the KNN model. Among these, the KNN-SMOTE algorithm has better model performance in terms of recall rate, accuracy, and F1-score. In contrast, KNN-GANs perform better in terms of recall but less well in terms of precision. Conversely, KNN-Normal has poor positive classification outcomes. In a similar manner, SVM was fed the three dataset categories, SVM-Normal, SVM-SMOTE, and SVM-GAN. Accuracy, F1-score, and recall measures were used to evaluate the model’s performance. With a marginally higher F1-score, SVM-SMOTE and SVM-GAN outperform SVM-Normal in all three criteria. As illustrated in
Figure 9, the argument goes that raising the failure class aids in lowering the false negative and false positive numbers.
However, because of its sequential structure, an LSTM model can assist in determining the impact in order to further analyze the minority class. Three datasets were utilized to assess the LSTM model’s performance for this purpose. However, the model accuracy approaches 87% on real data, on generated data, and up to 97% on merged data. Furthermore, the F1-scores for every dataset reach at least 0.97, which explains the model’s excellent performance in terms of recall and accuracy parameters. Because of its temporal dependencies and structures, the LSTM model has demonstrated its versatility and usefulness in a variety of data situations, particularly when applied to synthetic and mixed datasets, as illustrated in the
Figure 8. Dense GANs often struggle to capture complex temporal fault behavior in industrial time-series data because they treat inputs as static feature vectors and lack explicit memory for sequential dependencies. Still, research shows that under certain conditions, they can approximate temporal structures indirectly. For example, a recent study on railway-track fault diagnosis used a basic GAN (on one-channel sequential data) and achieved synthetic samples close to real measurements: when those samples were trained on a CNN classifier using GAN-augmented data, test accuracy rose from 89% to 96% [
28].
The CNN model’s performance was assessed using three datasets: produced data, hybrid data, and current data. An F1-score of
, a recall rate of
, and a precision rate of
were obtained on real data. The model performed exceptionally well on generated data, achieving
accuracy, an F1-score of
, and
recall. Additionally, the combined data demonstrated outstanding performance with
recall, an F1-score of
, and
precision. The model exhibits improved predictive power and flexibility across various datasets. The CNN and LSTM models’ accuracy and epochs are examined in
Figure 10.
When applied to real data, the SVM outperforms the other traditional machine learning models in terms of accuracy and recall. This heave and swing immediately indicate that the SVM is capable of accurately and promptly predicting positive events in terms of accuracy, as demonstrated by the accuracy and recall comparative study in
Figure 6. In a comparable manner, it can identify a high proportion of actual positive events in terms of memory without sacrificing any of the characteristics.
Figure 7 illustrates how poorly both the KNN and the DTC perform when tested on actual data in terms of achieving high accuracy and high recall. This is because, as
Table 8 demonstrates, these models frequently either incorrectly anticipate the positively labeled samples or misdiagnose positive cases in terms of low recall. The results illustrated in
Table 8 show that when considering the optimized traditional models, XGBoost achieved 0.9978 and tuned KNN achieved 0.9966 accuracy compared to the tuned LSTM, which achieved 0.9884. This finding aligns with the established literature: when datasets are relatively small, have limited temporal depth, or exhibit strong feature separability, classical machine learning models often outperform deep learning architectures. In these cases, XGBoost effectively leverages gradient-boosted decision rules to handle nonlinear boundaries, while KNN benefits from well-defined cluster structures in the feature space [
58].
Thus, comparing the results of the three models, SVM demonstrates a significantly better accuracy and recall rate than the two other conventional machine learning models when GAN-generated data are applied. Comparatively, the SVM can predict more positive cases with high precision as compared to KNN and DTC, as shown in
Figure 9. Similarly, when data produced through the SMOTE are fed to the models, both KNN and DTC give reasonable precision and recall with up to
and
, respectively, as shown in
Figure 7. The CNN model with GANs achieves up to
,
, and
in precision, F1-score, and recall, respectively. Similarly, the LSTM model achieves up to
,
, and
in precision, F1-score, and recall, respectively, as shown in
Table 8. Using actual data, generated data, and merged data, LSTM and CNN models have been applied successfully. This was most probably true for the case of merged data. These models’ successful generalization on the fused data containing real-world data with artificial data also establishes their anticipations. As it enhances the stability and enlarges the variety of models’ applications, this universality has certain implications for practice in the case of a lack of real data. From the overall assessment of the results, LSTM and CNN models are proven capable of dealing with different types and distributions of data, as shown in
Figure 7.
The comparative analysis described in
Table 9 shows that using data augmentation significantly improves model performance, especially for datasets with imbalanced classes like AI4I. Classical models such as DTC, KNN, and SVM perform poorly on AI4I without augmentation, showing low precision and F1-scores. When the SMOTE or GANs are applied, their performance improves significantly, highlighting the importance of addressing class imbalance. KNN benefits the most from GAN augmentation, achieving nearly perfect precision, recall, and F1-scores, while the SVM also improves noticeably with both the SMOTE and GANs. Deep learning models like LSTM and CNN already capture complex sequential or spatial patterns, so augmentation provides smaller but still meaningful improvements. For example, GAN-augmented LSTM achieves the highest overall performance across all datasets, showing that GAN-generated samples are realistic and helpful for training. In datasets that are more balanced, such as SECOM and APS, all models perform reasonably well even without augmentation, suggesting that the benefits of SMOTE or GANs are strongest for imbalanced or difficult datasets. Overall, GAN-based augmentation provides consistent and reliable improvements across different models, particularly for those that are sensitive to data distribution like KNN and sequential models like LSTM.