You are currently viewing a new version of our website. To view the old version click .
Information
  • Article
  • Open Access

18 December 2025

GenIIoT: Generative Models Aided Proactive Fault Management in Industrial Internet of Things

,
,
,
and
1
School of Computing Sciences, Pak-Austria Fachhochschule: Institute of Applied Sciences and Technology (PAF-IAST), Mang, Haripur 22620, Pakistan
2
Sino-Pak Center for Artificial Intelligence (SPCAI), Pak-Austria Fachhochschule: Institute of Applied Sciences and Technology (PAF-IAST), Mang, Haripur 22620, Pakistan
3
College of Engineering and Computing, American University of Bahrain, Riffa 942, Bahrain
4
College of Computer and Information Sciences, Prince Sultan University, Riyadh 11586, Saudi Arabia
This article belongs to the Special Issue Editorial Board Members’ Collection Series: “Information Processes”, 2nd Edition

Abstract

Detecting active failures is important for the Industrial Internet of Things (IIoT). The IIoT aims to connect devices and machinery across industries. The devices connect via the Internet and provide large amounts of data which, when processed, can generate information and even make automated decisions on the administration of industries. However, traditional active fault management techniques face significant challenges, including highly imbalanced datasets, a limited availability of failure data, and poor generalization to real-world conditions. These issues hinder the effectiveness of prompt and accurate fault detection in real IIoT environments. To overcome these challenges, this work proposes a data augmentation mechanism which integrates generative adversarial networks (GANs) and the synthetic minority oversampling technique (SMOTE). The integrated GAN-SMOTE method increases minority class data by generating failure patterns that closely resemble industrial conditions, increasing model robustness and mitigating data imbalances. Consequently, the dataset is well balanced and suitable for the robust training and validation of learning models. Then, the data are used to train and evaluate a variety of models, including deep learning architectures, such as convolutional neural networks (CNNs) and long short-term memory networks (LSTMs), and conventional machine learning models, such as support vector machines (SVMs), K-nearest neighbors (KNN), and decision trees. The proposed mechanism provides an end-to-end framework that is validated on both generated and real-world industrial datasets. In particular, the evaluation is performed using the AI4I, Secom and APS datasets, which enable comprehensive testing in different fault scenarios. The proposed scheme improves the usability of the model and supports its deployment in a real IIoT environment. The improved detection performance of the integrated GAN-SMOTE framework effectively addresses fault classification challenges. This newly proposed mechanism enhances the classification accuracy up to 0.99. The proposed GAN-SMOTE framework effectively overcomes the major limitations of traditional fault detection approaches and proposes a robust, scalable and practical solution for intelligent maintenance systems in the IIoT environment.

1. Introduction

Artificial intelligence enables data-driven approaches to predict equipment failures, which is called predictive maintenance [1]. It often helps the industry reduce the risk of equipment failures. Consequently, it mitigates costly and unscheduled downtime. This shift help the industries move away from traditional reactive or preventive maintenance techniques to predictive maintenance. In more industries, traditional defect management focuses on the repair or periodic replacement of equipment after failure, such as replacing worn-out motors [2]. In many industrial setups, predictive maintenance can be carried out using the information received through the data collected from various sensors. The collected data are then analyzed to track machinery health and performance in real time [3,4]. These sensors can measure several important factors, including temperature, vibration, pressure, and flow. Patterns, abnormalities, and possible signs of deterioration or future malfunctions are then analyzed using the collected data through installed sensors. Conventionally, equipment maintenance is performed following the defined schedules using the Industrial Internet of Things (IIoT) [4]. Due to the predictive strategies, the maintenance structure is rapidly changing with the technological advancements of the fourth industrial revolution [5]. The IIoT technology is employed in a business setting where digital technology, internet connectivity, and cloud platforms are integrated with real industrial machines, equipment, sensors, and other digital assets [6,7]. The main goal of the IIoT is to provide businesses with the tools they need to collect, monitor, analyze, and use data to streamline operations, increase productivity, and make informed decisions, especially for small devices or when a maintenance problem occurs during a device’s operation [8]. The IIoT also has the potential to support new business models and revenue streams. The data collected from the sensors embedded with the industrial equipment can be used to provide particular services, including healthcare services [9,10], performance analysis, and remote monitoring [11]. This advancement to data-driven services has the potential to increase competitiveness, enhance user experience, and identify opportunities for creative connections and collaboration [12]. The IIoT continues to change the face of industries in manufacturing, energy and agriculture. It can help the manufacturers achieve greater precision and flexibility by implementing IIoT-enabled smart manufacturing approaches. For example, in the energy industry, it is utilized to optimize energy, strengthen the integration of renewable energy sources, and enhance network performance [13,14].
For most industrial environments, the predictive maintenance techniques are employed to reduce costs as well as increase equipment performance and life. As a result, it reduces risk and increases safety. By tracking abnormalities, machine learning can predict maintenance problems such as unusual vibrations or temperature spikes [15,16]. Similarly, the IIoT gives opportunities to design new architectures as resource methods [17] as well as increase system performance and reduce cyber threats [18]. The IIoT brings connectivity, automation and data-driven decision making, bringing significant changes to manage their operations such as remote monitoring systems. The IIoT enables businesses to enhance their productivity levels, efficiency and creativity by bring the automation and the integration of industrial equipment such as machines, sensors and gadgets into a networked environment. These connected sensor nodes collect real-time data from industrial operations, including manufacturing processes, supply chains, logistics and energy consumption. The data are then sent to a centralized platform where powerful analytics and machine learning algorithms process it into actionable insights [19]. Machine learning models can be used to predict the remaining useful life (RUL) using advanced deep learning techniques [20,21].
Predictive maintenance is an important technique in industrial systems to avoid unplanned equipment breakdown. Compared to preventive or reactive maintenance, where it acts after the failure has occurred or on a time-based schedule, predictive maintenance enables timely action based on actual machine conditions. This is accomplished mainly through technological advancements in the IIoT, where sensor-enabled devices monitor operation parameters such as temperature, pressure, vibration, and flow in real time. Real-time sensor measurements are the foundation of smart decision making and predictive fault prediction. While they are helpful, predictive maintenance models are typically marred by imbalanced datasets, which is primarily because failure occurrences are sparse compared to regular operating information. Machine learning models trained on such skewed data will end up performing badly in identifying minority class instances, which in this scenario refer to the failure states of interest. To counter the problem, synthetic data generation techniques are employed to synthetically oversample the minority (failure) classes [22].
The proposed GenIIoT is a two-step hybrid framework that solves the tradeoff between the stability of the SMOTE and the realism of GAN-based data synthesis. The SMOTE is first applied to expand the minority class and stabilize GAN training, which would otherwise fail due to the extremely small number of fault samples. In the second stage, the GAN performs nonlinear refinement on SMOTE-generated data and learns the complex IIoT fault distribution. It generates high-fidelity synthetic samples ( X GenIIoT ) that are more diverse and realistic than interpolation-based methods, effectively addressing the limitations of SMOTE alone.
Two algorithms have been employed for synthetic data generation: the synthetic minority oversampling technique (SMOTE) and generative adversarial networks (GANs). However, there are some practical differences when applied to IIoT-based predictive maintenance. The SMOTE is light in computation because it creates new instances of the minority class by linear interpolation. It corrects the imbalance but does not exactly follow the variability to replicate the real-world complexity of faults [23]. Conversely, GANs generate data more close to real data, while the variational autoencoders (VAEs) produce overly smooth data. GANs can better learn the data distribution and thus capture complex fault behaviors. Similarly, while denoising diffusion probabilistic models (DDPMs) have high-quality generation capability, the high training cost and slow generation speed might limit their use in real-time industrial environments [23].
GANs can create high-fidelity synthetic data that are highly comparable to real IIoT fault patterns, which is critical to train trusted proactive models. Compared to VAEs, which create blurred or averaged samples, or DDPMs, which are too computationally demanding for resource-constrained IIoT edge devices, GANs offer a balance between sample fidelity and efficiency [12]. The SMOTE extends GANs to address class imbalance in fault datasets, which is a common issue in the IIoT, through local interpolation. Although DDPMs or VAEs can also generate synthetic data, the simplicity and low computational complexity in SMOTE make it more efficient for real-time fault management. The main contribution is the hybridization of GANs and the SMOTE to take advantage of their complementary strengths where GANs produce diverse fault scenarios and the SMOTE fine-tunes minority-class samples for maximizing classifier fairness [24]. Furthermore, it addresses the dynamic nature of IIoT faults’ non-stationary distributions using the feedback mechanism wherein the output of GAN is recursively validated against real-time fault logs for adaptive sample quality [19,25,26].
Abundant failure data are required for particular equipment for its on-time management using predictive maintenance. A GAN is a technique used to generate data. Similarly, SMOTE is a data-generated technique used in cases of data inadequacy. Machine learning models such as support vector machine (SVMs), k-nearest neighbors (KNN), and decision tree classifiers are used for predictive maintenance. Deep learning models such as convolutional neural networks (CNNs) and long short-term memory (LSTM) are used for predictive maintenance. The proposed method aims to develop a new and useful predictive maintenance system to assist in equipment maintenance. The core contributions of the proposed work are outlined below:
  • A generative AI-enabled predictive maintenance system framework is proposed to proactively predict equipment faults before they occur within an IIoT network.
  • The generative AI framework consists of data-generated techniques, including the SMOTE and GANs: the former is skilled at balancing class distribution and producing synthetic failure data, while the latter enables the generation of failure data with unmatched realism and diversity. It mimics the intricate behaviors of real industrial machinery, producing predictive data for maintenance.
  • Three different data categories, i.e., real, GAN-generated, and SMOTE-generated, are applied to machine learning to evaluate their precision, recall, F1-score, and accuracy performance.
  • The IIoT network performance is measured using inter-data with a machine learning models comparison as well as an inter-model comparison of machine learning models, such as SVM, KNN, and DTC, and deep learning models, such as CNN and LSTM.

3. Proposed GenIIoT Mechanism

A framework is proposed for predictive maintenance using generative learning to predict the fault in machines prior to its occurrence, as shown in Figure 1. Three different datasets such as APS, SECOM an AI4I with different categories and sample count as illustrated in Table 1 are used.
Figure 1. Proposed framework for predictive maintenance using generative learning.
Table 1. Class distribution across predictive maintenance datasets.

3.1. Artificial Intelligence Dataset for Industry (AI41)

The dataset contains 10,000 samples discretized from the time series, with 14 features, which include seven thermal, three mechanical, one rotary, one cutting, and two fault types, including thermal, feed, surge, torque, and random faults. A binary variable identifying the status of a machine at the time of taking a sample as a machine failure flag (1 for failure, 0 for normal), reveals 96.6% of data for active machines and 3.4% of failed ones, as illustrated in Table 2. Data were obtained from the AI4I 2020 dataset [50], which collects real-time measures through sensors for forecasting servicing needs. Changes in temperature, vibration, and rotational speeds are measured and recorded to capture both healthy and faulty states to inform the evaluation of failure modes. As for the 3.4% failure rate, oversampling techniques are used such as the SMOTE. To maintain the homogeneity of data, the features are standardized in terms of temperature, speed, and torque values. Some of the fault types have been converted into 0 and 1 to ensure they are easily distinguishable. This confirms and corrects for the missing data and also for the duplicity of some data.
Table 2. Data distribution across different sampling methods.

3.2. Semiconductor Manufacturing (SECOM) Dataset

There are 1567 samples in total with 592 features in this dataset. Each sample stores one or more sensor readings collected from the industrial process; parameters may include temperatures, pressures, rotation speed, and various fault signals and symptoms. The dataset also has a Pass/Fail label in which “−1” is given to represent a failure, while “1” is given to describe a successful operation. The data were collected continuously by using various sensors and focused on the operational data of the machinery. The objective was thus to try and establish when the machinery fails to make some useful inputs into the models used in prescriptive maintenance, the dataset information is illustrated in Table 1.

3.3. Air Pressure System (APS) Dataset

This dataset includes 60,000 samples, each containing 171 features related to sensor readings and operational metrics from industrial equipment, as shown in Table 1. The samples are labeled in the class column, where pos indicates a failure and neg indicates normal operation. Each feature captures measurements such as temperature, pressure, and flow rates for various components within the system. Data were collected from APS (Air Pressure System) units in trucks, where sensors monitored conditions that might lead to a failure. These data help identify failures and support predictive maintenance models by capturing conditions before both normal operations and breakdowns.

3.4. Data Preprocessing Techniques

Thorough preprocessing procedures are used to guarantee the robustness and dependability of the suggested predictive maintenance system. Occasional missing readings from gearbox problems or sensor malfunctions make up the dataset. For the numerical properties of the sensors, data imputation methods such as mean imputation are employed. In this method, the missing values are replaced with the feature-wise mean values, aiming to retain the statistical characteristics of each feature and avoid bias. Furthermore, feature normalization enables stable and efficient training of both the GAN and the downstream predictive model. Therefore, the values of the features are scaled into the range [0, 1] employing min–max normalization. It reduces the impact of the significant difference between the feature values that accelerate convergence during training. Prior to synthetic data generation, the dataset is partitioned into training and testing subsets using a stratified split, ensuring that the class distributions, e.g., normal and failure proportions, are preserved. Importantly, synthetic data generation through GAN and SMOTE techniques is applied to the training data to prevent data leakage, as shown in Figure 2. Moreover, data analysis (EDA) is applied to identify and exclude sensor features with near-zero variance or high inter-feature correlation, i.e., Pearson correlation coefficient > 0.95. It reduces redundancy and enhances model generalization. Considering this, explicit outlier removal is avoided to preserve the integrity of real-world operational data, as shown in Figure 1. The data are normalized as
x normalized = x x min x max x min .
Figure 2. Comparison of the distributions of real and generated data.

3.5. SMOTE Technique

This technique solves the class imbalance problem prevalent in target data regarding machine learning models. In particular, the class disparity refers to the condition in which one of the classes has considerably fewer samples in contrast with other classes, as shown in Figure 3 [51]. Task B reduces the distorted contribution of the numerous classes to learning by leveling the contribution of classes in the dataset through placing the minority class samples together through synthesized data, as shown in Figure 4. Before performing the balancing of effects of classes, an individual has to decide which class is the minority one that needs to be sampled to bring it to par with another class. However, there are two classes; the ratio in their actual dataset has to be determined first before applying the SMOTE, aiming to increase the percentage of the minority class. It balances the class distribution of the dataset by using a small percentage of data from the minority classes for the improvement of the model generalization, leading to enhanced model performance [35].
Figure 3. Proposed workflow of machine learning models with the SMOTE.
Figure 4. Illustration of the SMOTE workflow for minority class oversampling.
The workflow of SMOTE analysis is shown in Figure 3. For this purpose, subsequently, the AI4I dataset is used [50], which has a skewed class distribution, since minority class samples are significantly less compared to the majority class samples. Therefore, a string of characters with a random row N is chosen. Then, its nearby numbers are identified to find the nature of the nearby data. One of the nearest points, which is the closest to the chosen one, is randomly selected using K nearest neighbors (KNN). In this case, when synthesizing the sample, the number of closest neighbors is K = 5 . For creating a new synthetic sample for SMOTE, feature vectors of the selected sample and the KNN of the selected sample are used [33]. Average measures of KNN for the classification of the sample of a particular minority class are determined by using the Euclidean distance between the chosen minority class and the majority class. The Euclidean distance between the two points P 1 ( x 1 , y 1 ) and P 2 ( x 2 , y 2 ) can be defined as [52]
d E = ( x 2 x 1 ) 2 + ( y 2 y 1 ) 2 .
It generates synthetic samples by linear interpolation between the feature vectors of the selected sample and some of its nearest neighbors. A new synthetic instance, x new , from an original minority sample x i and a randomly selected neighbor x j , can be generated as
x new = x i + λ × ( x j x i ) ,
where x new represents the newly generated synthetic sample, x i is the randomly selected minority class instance, x j represents a neighbor chosen from the k-nearest neighbors of x i , and λ U ( 0 , 1 ) is a random number drawn from a uniform distribution between 0 and 1. As a preprocessing step, data imputation is carried out to replace the missing values using the mean of the whole feature represented as
c imputed = c ¯ = 1 N i = 1 N c i
where c imputed denotes the value assigned to replace a missing entry in the dataset. The term c ¯ refers to the mean of the corresponding feature column, which was calculated from the available data. Here, N represents the total number of non-missing observations for that feature, and c i indicates the ith observed value within that feature. These parameters are collectively used to compute the imputed value based on the existing observations.
The SMOTE allows the creation of a ratio of up to 20% for the failure class and 80% for the non-failure class. It minimizes the dominance of the majority class through an approach that deals with the generation of new noisy data as described in Algorithm 1. Thus, the data obtained by the SMOTE are noisy and unrealistic, providing the enhancement of the lower impact of the generated non-failure data where the same challenge persists. This could help balance the effect of the major class by using the factors in the contending classes and typing in more classes. Hence, to offset the impact of the minority class, through data augmentation, the weights associated with the non-failure data can be created and augmented [34].
Algorithm 1 SMOTE: Synthetic Minority Oversampling Technique
  1:
Input:  X min —Minority class samples, k—Number of neighbors, N—Number of synthetic samples
  2:
X syn {Initialize the list for synthetic samples}
  3:
for  i = 1  to  N do
  4:
     x min randomly select a sample from X min
  5:
     k nearest find k nearest neighbors of x min in X min
  6:
     x new empty feature vector
  7:
    for  j = 1  to number of features do
  8:
         diff randomly choose a value between 0 and 1
  9:
         feature _ diff x min [ j ] k nearest [ j ] × diff
10:
         x new [ j ] x min [ j ] + feature _ diff
11:
    end for
12:
     X syn X syn { x new }
13:
end for
14:
return  X syn

3.6. GANs Architecture

3.6.1. Proposed GAN Architecture

The proposed generative adversarial network (GAN) consists of two fully connected components: a generator and a discriminator, as shown in Table 3, which are both implemented using fully connected layers tailored to the dimensionality of the input feature space. The generator is composed of four dense layers designed to transform a 100-dimensional latent noise vector into a synthetic feature sample resembling the real data distribution. The generator follows the architecture with an input layer of size 100, which is followed by dense layers with 128, 256, and 128 units, respectively, and finally an output dense layer matching the size of the feature vector. ReLU activation functions are used in all hidden layers. In contrast, a sigmoid activation is applied at the output layer to ensure the generated features remain within a normalized range, as shown in Table 4. The discriminator is composed of three dense layers that process input samples to classify them as real or synthetic. Its architecture includes an input layer matching the feature vector size, followed by hidden dense layers of 128 and 64 units, and a final output layer with one unit for binary classification. LeakyReLU activation is employed in the hidden layers to support better gradient propagation, while the output layer uses a sigmoid function. Both networks are trained using the Adam optimizer with a learning rate of 0.001, β 1 set to 0.5, and β 2 set to 0.999 to ensure training stability and convergence. The loss function represents the standard minimax optimization used to train a GAN,
min G max D V ( D , G ) = E x P data ( x ) log D ( x ) + E z P z ( z ) log ( 1 D ( G ( z ) ) ) ,
where G represents the generator network, which produces synthetic samples from a random noise vector, and D is the discriminator network that attempts to distinguish between real and generated samples. Let P data ( x ) represent the distribution of the real data used to train the discriminator, and let P z ( z ) represent the distribution of the input noise vector, which is typically sampled from a uniform or Gaussian distribution. The discriminator’s output for a real sample x is given by D ( x ) , which represents the probability that x is authentic, while D ( G ( z ) ) is the discriminator’s output probability for a synthetic sample generated by G.
Table 3. Sequential architecture of GAN generator and discriminator.
Table 4. Hyperparameter settings for the proposed GAN model.

3.6.2. Hyperparameter Configuration

A combination of grid search and manual tuning was employed to fine-tune the model’s hyperparameters and mitigate overfitting. The final configuration is summarized in Table 4.

3.7. Machine Learning Models

The prepossessed data is fed to SMOTE to augment the data aiming to enhance the minority class. Then, the the data is fed to GANs to generate the fault data for accurately predicting the faults prior to its occurrence, as illustrated in Figure 5. Machine learning models are trained on the actual data, SMOTE generated data and GANs generated data respectively. The Machine learning models, such as SVM, KNN, and decision tree classifiers, and deep learning models, such as CNN and LSTM, are employed to feed actual data, where 3.4% are in the failure class and 96.6% are non-failure data [50]. The architectures of the LSTM and CNN models used are shown in Table 5 and Table 6, respectively. The performance of these models is low because of the small failure data. Thus, the SMOTE is used to increase the failure dataset. It is conceded that after applying the SMOTE, their findings are ambiguous because it synthesizes noisy data to address the class imbalance. Another employed method in the generation of data is GANs, where after adding newly generated data, the ratios are 20% failure class and 80% non-failure class.
Figure 5. Proposed workflow model using data generating techniques.
Table 5. LSTM architecture.
Table 6. CNN architecture.
Since the failure data are very scarce, it directly affects the accuracy of the models. To have a meaningful impact of machine failure on the model performance, the size of the failures needs to be increased. For this purpose, the SMOTE is used to incorporate meaningful failure data in the dataset for training and testing. Once the data were generated, with 20% failure data and 80% non-failure data, the data were fed to the machine learning models.
In the next step, the failure dataset is generated with the help of the GANs technique, which is then fed to machine learning models for performance analysis [53]. The main difference between the SMOTE and GANs is that the former reduces the class imbalance by the interpolation of available classes, while the latter generates entirely new data considering the existing classes. SVMs can maximize any loss function and can be used both for regression and for classification [54]. The decision trees provide models that can be explained; they show decision-making processes, and KNN is a simple yet effective classification that distinguishes based on the nearest neighbors. Key performance measurement metrics are used to assess the efficacy of a model, such as accuracy, precision, recall, and F1-score [55].
Predictive maintenance deals with equipment state classification using the SVM, which is categorized as normal or failed [13]. The SVM employs different kernel functions, like the polynomial kernel or radial basis function (RBF), for effectively dealing with the nonlinearity of feature relationships. KNN assists in anomaly detection in predictive maintenance, where it works by comparing a data point to the closest points to flag the outlying patterns that require little necessity for fine-tuning of the parameters. Decision trees assist in finding the characteristics and or elements that cause equipment failures. CNNs accommodate the spatial pyramids of data, which helps find anomalous or erroneous regions in machines. Comparatively, LSTM works better on sequential data that would be useful in assisting predictive maintenance, which depends on the data provided with time [16].
The system requirements for this research setup include support for Windows 10/11, Linux, or macOS operating systems. Hardware specifications require a minimum of 8 GB RAM and 4 GB GPU memory, and up to 16 GB RAM and 8 GB or more GPU memory are recommended for optimal performance. Additionally, a minimum of 10 GB of storage space is required to accommodate datasets and models. For improved computational efficiency, an NVIDIA GPU with CUDA support is recommended. The computing resources for the environment utilize Jupyter Notebook 6.5, Python 3.10 with TensorFlow 2.x, including GPU support for CUDA-compatible NVIDIA GPUs. The machine learning and deep learning frameworks rely on TensorFlow/Keras with automatic memory allocation in TensorFlow to ensure the efficient handling of large datasets. The effectiveness of the proposed technique is assessed through a structured evaluation process that combines several complementary strategies. First, hold-out validation is performed using stratified train–test splits (70/30 or 80/20 with fixed random seeds). All preprocessing steps are applied only to the training data, and augmentation methods such as the SMOTE or GANs are restricted to the training portion, while the test set remains entirely real and untouched. In addition, stratified five-fold cross-validation, with optional repetitions, is carried out to provide a more reliable estimate of model performance. Within each fold, preprocessing and class balancing are limited to the training split, and results are reported as the mean with standard deviation, with model selection guided by minority-class metrics such as PR-AUC and F1-score. To evaluate external validity, the pipeline is also applied to independent datasets, including APS and UCI SECOM, and domain-shift experiments are conducted by training on AI4I and testing on APS or SECOM after aligning the features. Further analysis is carried out through ablation studies, in which different class-balancing strategies (no balancing, SMOTE, and GAN-based augmentation), feature subsets, and robustness under controlled label-noise injection are compared. Hyperparameter tuning is performed through grid search within the training data using either nested cross-validation or an inner split, which is followed by retraining on the complete training set. Strict leakage prevention is ensured throughout the evaluation, meaning no synthetic samples or resampling are introduced into validation or test sets, and all preprocessing steps, such as scaling or encoding, are fitted exclusively on training data. Results are primarily reported in terms of PR-AUC and F1-score for the minority class, which are complemented by additional indicators including precision, recall, accuracy, confusion matrices, and threshold analyses derived from precision–recall curves with probability calibration applied when required.

4. Results and Discussion

The similarity between real and synthetic data is evaluated using a distributional comparison. Therefore, the distribution of real and generated data is compared using a histogram. The histogram distribution of key features is illustrated in Figure 2, where the blue and red represent real and GAN-generated distributions, respectively. The alignment of these curves demonstrates that the GAN-generated data resemble the underlying distribution of the original data. This method is commonly used to assess synthetic data fidelity. By integrating these evaluations, the synthetic fault data generated by the GAN closely mirror the statistical properties of the real dataset, supporting its applicability in training more robust predictive maintenance models. The closer the curves align, the better the quality of the data generated. The GAN model hyperparameters are listed in Table 4.
Three categories of data are used: AI4I, APS, and SECOM. The models’ performance is evaluated by feeding the three kinds of data to the models. For example, decision tree classifiers are fed with three types of data and categorized as decision tree classifier with normal data (DTC-normal), decision tree classifier with SMOTE-generated data (DTC-SMOTE), and decision tree classifier with GAN-generated data (DTC-GAN). Hyperparameter tuning for all conventional baseline models, including SVM, KNN, decision tree, random forest, and 1D CNN, in addition to the proposed LSTM architecture, is performed by varying key parameters, such as kernel type, regularization strength, gamma settings, neighborhood size, weighting scheme, tree depth, estimator count, learning rate, batch size, hidden units, dropout rate, and loss functions, with optimal configurations summarized in Table 7. The effectiveness of these models is assessed with key performance metrics as shown in Figure 6, including the accuracy and recall rate, as shown in Figure 7, the F1-score as depicted in Figure 8, and the level of precision as illustrated in Figure 9 using the hyperparameters listed in Table 7. In Figure 9, several models show accuracy and precision values close to 0.999 or 1.0, which is unexpected given the dataset’s strong class imbalance. However, synthetic data generation methods, including GANs and the SMOTE, are used to balance the APS training dataset, which enhanced the model’s ability to correctly identify minority class instances. Consequently, the elevated precision and accuracy scores reflect the effectiveness of the models when trained on a balanced dataset. Mathematically,
Accuracy ( ACC ) = T P + T N T P + T N + F P + F N
Precision ( P ) = T P T P + F P
Recall ( R ) = T P T P + F N
F 1 - Score ( F 1 ) = 2 × Precision × Recall Precision + Recall .
Table 7. Comparison of hyperparameters tested for different models with their optimal values and corresponding classification accuracy.
Figure 6. Representation of machine learning and deep learning models’ results using AI4I dataset.
Figure 7. Representation of machine learning and deep learning models’ results using SECOM dataset.
Figure 8. Accuracy and epochs analysis of CNN and LSTM models.
Figure 9. Representation of machine learning and deep learning models’ results using APS dataset.
The results of the hyperparameter optimization show differences amongst the models that are examined with the LSTM obtaining the maximum accuracy of 98.84%. Its balanced design (0.0005 learning rate, 32 batch size, 64 hidden units, 0.2 dropout, and binary cross-entropy loss) allowed for an efficient modeling of temporal relationships while reducing overfitting, which is responsible for this performance. With 100 estimators and a depth of 20, random forest came in second at 96.3%, capturing intricate feature interactions without introducing undue volatility. Using an RBF kernel with C = 1 and gamma adjusted to scale, the SVM achieved 95.8% accuracy, offering strong nonlinear decision limits. With five neighbors and distance-based weighting, which prioritized nearby, more pertinent samples, KNN fared best with 93.5%. The 1D CNN closely matched this at 93.8%, with 64 filters, a kernel size of 3, and a 0.0005 learning rate, enabling an effective extraction of local temporal–spatial patterns. The decision tree achieved 90.4% accuracy with a depth of 10 and a minimum split of 5, balancing complexity and generalization. Overall, the results confirm that while all models benefited from targeted hyperparameter tuning, deep learning approaches and specifically, LSTM outperformed classical methods, with random forest offering the strongest traditional alternative.
DTC-Normal demonstrates rather low performance. The reasoning behind this is the presence of very scarce failure data affecting the models’ performance, as shown in Figure 6. Comparatively, DTC-SMOTE and DTC-GAN achieve better results than DTC-Normal in terms of recall, accuracy, and F1-scores, as shown in Table 8. XGBoost achieves the highest accuracy of 0.9978, which is followed by KNN with an accuracy of 0.9966, and random forest with 0.9957. The deep learning models, such as LSTM and DNN, achieve 0.9884 and 0.9810, respectively, and perform well but were outperformed by the ensemble methods. On the other hand, SVM had the lowest accuracy of 0.9822.
Table 8. Model comparison on test accuracy, parameters, FLOPs, and training time.
The computational efficiency of the models is compared in terms of the number of parameters, FLOPS used, and training time. The DNN has the fewest parameters of 11,009, while random forest has the most: 169,162 nodes. LSTM had a moderate count of 63,329. Comparatively, only neural models have an approximate FLOPs comparison, where LSTM has 126,560 and DNN has 21,888, as illustrated in Table 8. Furthermore, the training time of XGBoost is the lowest with 12.42 s, while SVM with 2587.13 s was the highest among them all due to an extensive grid search to achieve the optimal hyperparameters. The deep learning models have moderate training times due to hyperparameter tuning. The models’ comparison on test accuracy, number of parameters, FLOPS, and training time is summarized in Table 8.
Predictive maintenance (PdM) emerges as a critical application area within the Industrial Internet of Things (IIoT), relying heavily on time-series sensor data to forecast potential equipment failures. To properly analyze the temporal data, deep learning (DL) architectures and traditional machine learning (ML) algorithms were investigated. Long short-term memory (LSTM), convolutional neural networks (CNN), support vector machines (SVMs), K-nearest neighbors (k-NN), and decision trees (DTs) are among the models that have undergone considerable evaluation.
The suggested LSTM model is compared to a number of cutting-edge classification methods documented in recent research in Table 8. With an accuracy of 98.84% and balanced precision–recall values (99% and 99%, respectively), the results show that the suggested LSTM has significantly improved over the traditional method. In contrast, the random forest model obtained an F1-score of 96.2% and accuracy of 96.3%, while KNN and the 1D CNN obtained accuracies of 93.5% and 93.8%, respectively, and the SVM model also showed impressive results with 95.8% accuracy. On the other hand, the decision tree showed limits in managing the dataset’s complexity, recording the lowest accuracy of 90.4%. While some conventional models are very accurate, the suggested LSTM model has significant advantages when it comes to managing the data’s temporal patterns and sequential dependencies. Because of these features, it is especially well suited for jobs where the sequence and timing of events are crucial to the accuracy of predictions.
Model interpretability is frequently just as crucial in industrial settings as prediction accuracy. Conventional machine learning methods like SVM, KNN, and decision trees are thought to be more interpretable and transparent. Decision trees, for instance, offer easily verifiable decision routes and rules that are accessible by humans as well as domain experts. Similar to this, SVMs provide decision boundaries that are either linear or kernel-based, which facilitates comprehension of the model’s decision-making process. On the other hand, because of their deep architectures and nonlinear transformations, DL models such as CNN and LSTM are frequently regarded as black-box systems. These models are more accurate when dealing with complicated data, but their lack of transparency may make them unsuitable for use in safety-critical or regulated applications [56]. This has increased interest in Explainable AI (XAI) methods, which try to improve the interpretability of DL models. LIME (Local Interpretable Model-Agnostic Explanation) is a popular XAI technique [57]. as well as SHAP (SHapley Additive exPlanation), which offer post hoc justifications for model forecasts. These techniques provide a means of striking a compromise between interpretability and performance for both traditional and deep learning models. A thorough taxonomy of XAI approaches was presented by Arrieta et al., who emphasized that interpretability and accuracy should be taken into consideration when choosing models in crucial domains [56].
The lack of labeled fault data and its imbalance are two of PdM’s biggest problems. Researchers have used generative adversarial networks (GANs) to augment data in order to overcome this. In order to improve model generalization and robustness, conditional GANs (cGANs) are very good at producing synthetic sensor data that correspond to uncommon fault circumstances [37]. Additionally, it has been demonstrated that cGANs and LSTM networks can predict RUL more accurately in situations with little labeled data. Another method, the SMOTified-GAN framework, improves fault classification in unbalanced industrial datasets by combining ensemble learning approaches with GAN-based oversampling. These techniques emphasize how crucial synthetic data generation and hybrid learning are to the development of PdM systems [36].
Three distinct datasets, KNN-Normal, KNN-SMOTE, and KNN-GAN, were fed into the KNN model. Among these, the KNN-SMOTE algorithm has better model performance in terms of recall rate, accuracy, and F1-score. In contrast, KNN-GANs perform better in terms of recall but less well in terms of precision. Conversely, KNN-Normal has poor positive classification outcomes. In a similar manner, SVM was fed the three dataset categories, SVM-Normal, SVM-SMOTE, and SVM-GAN. Accuracy, F1-score, and recall measures were used to evaluate the model’s performance. With a marginally higher F1-score, SVM-SMOTE and SVM-GAN outperform SVM-Normal in all three criteria. As illustrated in Figure 9, the argument goes that raising the failure class aids in lowering the false negative and false positive numbers.
However, because of its sequential structure, an LSTM model can assist in determining the impact in order to further analyze the minority class. Three datasets were utilized to assess the LSTM model’s performance for this purpose. However, the model accuracy approaches 87% on real data, on generated data, and up to 97% on merged data. Furthermore, the F1-scores for every dataset reach at least 0.97, which explains the model’s excellent performance in terms of recall and accuracy parameters. Because of its temporal dependencies and structures, the LSTM model has demonstrated its versatility and usefulness in a variety of data situations, particularly when applied to synthetic and mixed datasets, as illustrated in the Figure 8. Dense GANs often struggle to capture complex temporal fault behavior in industrial time-series data because they treat inputs as static feature vectors and lack explicit memory for sequential dependencies. Still, research shows that under certain conditions, they can approximate temporal structures indirectly. For example, a recent study on railway-track fault diagnosis used a basic GAN (on one-channel sequential data) and achieved synthetic samples close to real measurements: when those samples were trained on a CNN classifier using GAN-augmented data, test accuracy rose from 89% to 96% [28].
The CNN model’s performance was assessed using three datasets: produced data, hybrid data, and current data. An F1-score of 0.95 , a recall rate of 88 % , and a precision rate of 87 % were obtained on real data. The model performed exceptionally well on generated data, achieving 98 % accuracy, an F1-score of 0.99 , and 94 % recall. Additionally, the combined data demonstrated outstanding performance with 98 % recall, an F1-score of 0.99 , and 98 % precision. The model exhibits improved predictive power and flexibility across various datasets. The CNN and LSTM models’ accuracy and epochs are examined in Figure 10.
Figure 10. Loss–epochs curve of CNN and LSTM models.
When applied to real data, the SVM outperforms the other traditional machine learning models in terms of accuracy and recall. This heave and swing immediately indicate that the SVM is capable of accurately and promptly predicting positive events in terms of accuracy, as demonstrated by the accuracy and recall comparative study in Figure 6. In a comparable manner, it can identify a high proportion of actual positive events in terms of memory without sacrificing any of the characteristics. Figure 7 illustrates how poorly both the KNN and the DTC perform when tested on actual data in terms of achieving high accuracy and high recall. This is because, as Table 8 demonstrates, these models frequently either incorrectly anticipate the positively labeled samples or misdiagnose positive cases in terms of low recall. The results illustrated in Table 8 show that when considering the optimized traditional models, XGBoost achieved 0.9978 and tuned KNN achieved 0.9966 accuracy compared to the tuned LSTM, which achieved 0.9884. This finding aligns with the established literature: when datasets are relatively small, have limited temporal depth, or exhibit strong feature separability, classical machine learning models often outperform deep learning architectures. In these cases, XGBoost effectively leverages gradient-boosted decision rules to handle nonlinear boundaries, while KNN benefits from well-defined cluster structures in the feature space [58].
Thus, comparing the results of the three models, SVM demonstrates a significantly better accuracy and recall rate than the two other conventional machine learning models when GAN-generated data are applied. Comparatively, the SVM can predict more positive cases with high precision as compared to KNN and DTC, as shown in Figure 9. Similarly, when data produced through the SMOTE are fed to the models, both KNN and DTC give reasonable precision and recall with up to 0.93 and 0.94 , respectively, as shown in Figure 7. The CNN model with GANs achieves up to 0.90 , 0.99 , and 0.94 in precision, F1-score, and recall, respectively. Similarly, the LSTM model achieves up to 0.97 , 0.99 , and 0.98 in precision, F1-score, and recall, respectively, as shown in Table 8. Using actual data, generated data, and merged data, LSTM and CNN models have been applied successfully. This was most probably true for the case of merged data. These models’ successful generalization on the fused data containing real-world data with artificial data also establishes their anticipations. As it enhances the stability and enlarges the variety of models’ applications, this universality has certain implications for practice in the case of a lack of real data. From the overall assessment of the results, LSTM and CNN models are proven capable of dealing with different types and distributions of data, as shown in Figure 7.
The comparative analysis described in Table 9 shows that using data augmentation significantly improves model performance, especially for datasets with imbalanced classes like AI4I. Classical models such as DTC, KNN, and SVM perform poorly on AI4I without augmentation, showing low precision and F1-scores. When the SMOTE or GANs are applied, their performance improves significantly, highlighting the importance of addressing class imbalance. KNN benefits the most from GAN augmentation, achieving nearly perfect precision, recall, and F1-scores, while the SVM also improves noticeably with both the SMOTE and GANs. Deep learning models like LSTM and CNN already capture complex sequential or spatial patterns, so augmentation provides smaller but still meaningful improvements. For example, GAN-augmented LSTM achieves the highest overall performance across all datasets, showing that GAN-generated samples are realistic and helpful for training. In datasets that are more balanced, such as SECOM and APS, all models perform reasonably well even without augmentation, suggesting that the benefits of SMOTE or GANs are strongest for imbalanced or difficult datasets. Overall, GAN-based augmentation provides consistent and reliable improvements across different models, particularly for those that are sensitive to data distribution like KNN and sequential models like LSTM.
Table 9. Performance metrics of models on AI4I, SECOM, and APS datasets.

5. Conclusions and Future Work

It is inevitable for industries to take proactive measures to identify equipment faults, as such failures may cause component degradation, equipment breakdown, and potentially result in human casualties, downtime, and high on-site maintenance costs. Therefore, providing industries with an AI-based predictive maintenance system to monitor installed equipment is essential. For this purpose, a generative artificial intelligence system is proposed for the IIoT network. The system consists of two data generation models, SMOTE and GANs, used to generate synthetic data with diversity, which is then applied to machine learning models, including SVM, KNN, and DTC, as well as deep learning models such as CNN and LSTM. Three different data categories, real, GAN-generated, and SMOTE-generated, are applied to the machine learning models to evaluate their performance. GAN-enabled AI models achieve significant performance when intra-dataset comparisons are performed, while deep learning models such as LSTM showed notable improvement when cross-model comparisons are considered. However, the proposed method has some limitations: (1) the synthetic data generation process, while effective, may not fully capture the complexity and variability of real industrial fault patterns; (2) the model is not yet adaptive to evolving fault characteristics in streaming IIoT environments; (3) multi-sensor integration for richer fault context is not implemented; (4) model interpretability for industrial operators remains limited; and (5) the system has not yet been tested under real-world deployment constraints, such as on edge devices. Future research will therefore focus on addressing these limitations by exploring more advanced data generation techniques to improve the diversity and realism of synthetic fault data, developing adaptive learning methods capable of updating in real time from continuous sensor inputs, incorporating data from multiple heterogeneous sensors to provide a more complete diagnostic view, enhancing interpretability to increase user trust, and performing extensive real-world testing particularly on resource-limited edge devices to evaluate practical deployment feasibility.

Author Contributions

Conceptualization, I.Z. and A.I.; methodology, I.Z. and A.I.; software, I.Z. and M.K.; validation, I.Z., A.I. and M.A.A.; formal analysis, M.K. and N.A.; investigation, I.Z. and N.A.; resources, A.I.; data curation, I.Z. and A.I.; writing—original draft preparation, I.Z. and A.I.; writing—review and editing, A.I., M.A.A. and N.A.; visualization, I.Z. and A.I.; supervision, A.I. and M.K.; project administration, A.I. and M.K.; funding acquisition, N.A., M.A.A. and A.I. All authors have read and agreed to the published version of the manuscript.

Funding

The APC was funded by Prince Sultan University.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The datasets used in this study can be accessed through the following links: AI4I dataset: https://doi.org/10.24432/C5HS5C; Secom dataset: https://doi.org/10.24432/C54305; Aps dataset: https://doi.org/10.24432/C51S51. All associated code is also available at https://doi.org/10.5281/zenodo.16900443. The repository includes a README markdown file with detailed code usage and reproducibility instructions.

Acknowledgments

The authors would like to thank Prince Sultan University for paying the Article Processing Charges (APC) of this publication. They would also like to thank Prince Sultan University for their support.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
SMOTESynthetic Minority Oversampling Technique
GANGenerative Adversarial Network
CNNConvolutional Neural Network
LSTMLong Short-Term Memory
IIoTIndustrial Internet of Things
PdMPredictive Maintenance
IoTInternet of Things
AIArtificial Intelligence
MLMachine Learning
DLDeep Learning
RULRemaining Useful Life

References

  1. Haseeb, K.; Rehman, A.; Saba, T.; Wang, H.; Alruwaili, F.F. Empowering Real-Time Data Optimizing Framework Using Artificial Intelligence of Things for Sustainable Computing. IEEE Internet Things J. 2024, 11, 39094–39102. [Google Scholar] [CrossRef]
  2. Ong, K.S.H.; Wang, W.; Hieu, N.Q.; Niyato, D.; Friedrichs, T. Predictive Maintenance Model for IIoT-Based Manufacturing: A Transferable Deep Reinforcement Learning Approach. IEEE Internet Things J. 2022, 9, 15725–15741. [Google Scholar] [CrossRef]
  3. Huang, Q.; Wen, R.; Han, Y.; Li, C.; Zhang, Y. Intelligent Fault Identification for Industrial Internet of Things via Prototype-Guided Partial Domain Adaptation with Momentum Weight. IEEE Internet Things J. 2023, 10, 16381–16391. [Google Scholar] [CrossRef]
  4. Wen, P.; Li, Y.; Chen, S.; Zhao, S. Remaining Useful Life Prediction of IIoT-Enabled Complex Industrial Systems with Hybrid Fusion of Multiple Information Sources. IEEE Internet Things J. 2021, 8, 9045–9058. [Google Scholar] [CrossRef]
  5. Compare, M.; Baraldi, P.; Zio, E. Challenges to IoT-Enabled Predictive Maintenance for Industry 4.0. IEEE Internet Things J. 2019, 7, 4585–4597. [Google Scholar] [CrossRef]
  6. Yao, P.; Yan, B.; Yang, T.; Wang, Y.; Yang, Q.; Wang, W. Security-Enhanced Operational Architecture for Decentralized Industrial Internet of Things: A Blockchain-Based Approach. IEEE Internet Things J. 2024, 11, 11073–11086. [Google Scholar] [CrossRef]
  7. Czeczot, G.; Rojek, I.; Mikołajewski, D.; Sangho, B. AI in IIoT Management of Cybersecurity for Industry 4.0 and Industry 5.0 Purposes. Electronics 2023, 12, 3800. [Google Scholar] [CrossRef]
  8. Khalil, R.A.; Saeed, N.; Masood, M.; Fard, Y.M.; Alouini, M.-S.; Al-Naffouri, T.Y. Deep Learning in the Industrial Internet of Things: Potentials, Challenges, and Emerging Applications. IEEE Internet Things J. 2021, 8, 11016–11040. [Google Scholar] [CrossRef]
  9. Liu, C.; Tang, D.; Zhu, H.; Nie, Q. A Novel Predictive Maintenance Method Based on Deep Adversarial Learning in the Intelligent Manufacturing System. IEEE Access 2021, 9, 49557–49575. [Google Scholar] [CrossRef]
  10. Farman, H.; Shahzad, Y.; Jan, B.; Nasralla, M.M.; Sallam, K.M.; Munasinghe, K.; Jamalipour, A. IoT-Aware Real-Time Healthcare Diagnostic Framework for Diabetes Using Wearable Sensors Through Deep Reinforcement Learning. IEEE Internet Things J. 2025, 12, 18183–18194. [Google Scholar] [CrossRef]
  11. Behera, S.; Misra, R. Generative Adversarial Networks Based Remaining Useful Life Estimation for IIoT. Comput. Electr. Eng. 2021, 92, 107195. [Google Scholar] [CrossRef]
  12. Behera, S.; Misra, R. A Multi-Model Data-Fusion Based Deep Transfer Learning for Improved Remaining Useful Life Estimation for IIOT Based Systems. Eng. Appl. Artif. Intell. 2023, 119, 105712. [Google Scholar] [CrossRef]
  13. Wahid, A.; Breslin, J.G.; Intizar, M.A. Prediction of Machine Failure in Industry 4.0: A Hybrid CNN-LSTM Framework. Appl. Sci. 2022, 12, 4221. [Google Scholar] [CrossRef]
  14. Iqbal, A.; Lee, T.-J. Spatiotemporal Medium Access Control for Wireless Powered IoT Networks. IEEE Internet Things J. 2021, 8, 14822–14834. [Google Scholar] [CrossRef]
  15. Iqbal, A.; Lee, T.-J. GWINs: Group-Based Medium Access for Large-Scale Wireless Powered IoT Networks. IEEE Access 2019, 7, 172913–172927. [Google Scholar] [CrossRef]
  16. Nasser, A.; Al-Khazraji, H. A Hybrid of Convolutional Neural Network and Long Short-Term Memory Network Approach to Predictive Maintenance. Int. J. Electr. Comput. Eng. 2022, 12, 721–730. [Google Scholar] [CrossRef]
  17. Xu, S.; Li, Y.; Guo, S.; Lei, C.; Liu, D.; Qiu, X. Cloud–Edge Collaborative SFC Mapping for Industrial IoT Using Deep Reinforcement Learning. IEEE Trans. Ind. Inform. 2021, 18, 4158–4168. [Google Scholar] [CrossRef]
  18. Haseeb, K.; Saba, T.; Rehman, A.; Abbas, N.; Kim, P.W. AI-Driven IoT-Fog Analytics Interactive Smart System with Data Protection. Expert Syst. 2025, 42, e13573. [Google Scholar] [CrossRef]
  19. Lu, J.; Yang, J.; Li, S.; Li, Y.; Jiang, W.; Dai, J.; Hu, J. A2C-DRL: Dynamic Scheduling for Stochastic Edge–Cloud Environments Using A2C and Deep Reinforcement Learning. IEEE Internet Things J. 2024, 11, 16915–16927. [Google Scholar] [CrossRef]
  20. Babayigit, B.; Abubaker, M. Industrial Internet of Things: A Review of Improvements Over Traditional SCADA Systems for Industrial Automation. IEEE Syst. J. 2024, 18, 120–133. [Google Scholar] [CrossRef]
  21. Wang, H.; Zhang, W.; Yang, D.; Xiang, Y. Deep-Learning-Enabled Predictive Maintenance in Industrial Internet of Things: Methods, Applications, and Challenges. IEEE Syst. J. 2022, 16, 2797–2810. [Google Scholar] [CrossRef]
  22. Xie, X.; Wang, H.; Liu, X. Scheduling for Minimizing the Age of Information in Multisensor Multiserver Industrial Internet of Things Systems. IEEE Trans. Ind. Inform. 2024, 20, 573–582. [Google Scholar] [CrossRef]
  23. Rikalovic, A.; Suzic, N.; Bajic, B.; Piuri, V. Industry 4.0 Implementation Challenges and Opportunities: A Technological Perspective. IEEE Syst. J. 2022, 16, 2797–2810. [Google Scholar] [CrossRef]
  24. Arinez, J.F.; Chang, Q.; Gao, R.X.; Xu, C.; Zhang, J. Artificial Intelligence in Advanced Manufacturing: Current Status and Future Outlook. J. Manuf. Sci. Eng. 2020, 142, 110804. [Google Scholar] [CrossRef]
  25. Lee, J.; Ni, J.; Singh, J.; Jiang, B.; Azamfar, M.; Feng, J. Intelligent Maintenance Systems and Predictive Manufacturing. J. Manuf. Sci. Eng. 2020, 142, 110805. [Google Scholar] [CrossRef]
  26. Reddy, B.R.; Sujith, A.V.L.N. A Comprehensive Literature Review on Data Analytics in IIoT (Industrial Internet of Things). Helix 2018, 8, 2757–2764. [Google Scholar]
  27. Udu, A.G.; Salman, M.T.; Ghalati, M.K.; Lecchini-Visintini, A.; Siddle, D.R.; Dong, H. Emerging SMOTE and GAN-Variants for Data Augmentation in Imbalance Machine Learning Tasks: A Review. IEEE Access 2025, 13, 123456–123470. [Google Scholar] [CrossRef]
  28. Fernández-Bobadilla, H.A.; Bouchikhi, Y.; Martin, U. GAN-based data augmentation of time series for fault diagnosis in railway track. Railw. Eng. Sci. 2025, 33, 642–683. [Google Scholar] [CrossRef]
  29. Paroha, A.D. Advancing Predictive Maintenance in the Oil and Gas Industry: A Generative AI Approach with GANs and LLMs for Sustainable Development. Sustain. Dev. Mach. Learn. AI IoT (CCIS) 2025, 2196, 3–15. [Google Scholar] [CrossRef]
  30. Shan, J.; Xie, Y.; Wei, L. A GAN-based Hybrid Sampling Method for Transaction Fraud Detection. In Proceedings of the 2024 International Conference on Networking, Sensing and Control (ICNSC), Hangzhou, China, 18–20 October 2024; pp. 1–7. [Google Scholar] [CrossRef]
  31. Zhang, W.; Li, Z.; Li, G.; Zhuang, P.; Hou, G.; Zhang, Q.; Li, C. GACNet: Generate Adversarial-Driven Cross-Aware Network for Hyperspectral Wheat Variety Identification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5503314. [Google Scholar] [CrossRef]
  32. Zhang, W.; Li, Z.; Li, G.; Zhou, L.; Zhao, W.; Pan, X. AGANet: Attention-Guided Generative Adversarial Network for Corn Hyperspectral Images Augmentation. IEEE Trans. Consum. Electron. 2024, 71, 3683–3694. [Google Scholar] [CrossRef]
  33. Ishaq, A.; Sadiq, S.; Umer, M.; Ullah, S.; Mirjalili, S.; Rupapara, V.; Nappi, M. Improving the Prediction of Heart Failure Patients’ Survival Using SMOTE and Effective Data Mining Techniques. IEEE Access 2021, 9, 39707–39716. [Google Scholar] [CrossRef]
  34. Douzas, G.; Bacao, F. Geometric SMOTE: A Geometrically Enhanced Drop-in Replacement for SMOTE. Inf. Sci. 2019, 501, 118–135. [Google Scholar] [CrossRef]
  35. Zhang, A.; Yu, H.; Huan, Z.; Yang, X.; Zheng, S.; Gao, S. SMOTE-RkNN: A Hybrid Re-sampling Method Based on SMOTE and Reverse k-Nearest Neighbors. Inf. Sci. 2022, 595, 70–88. [Google Scholar] [CrossRef]
  36. Sharma, A.; Singh, P.K.; Chandra, R. SMOTified-GAN for Class Imbalanced Pattern Classification Problems. IEEE Access 2022, 10, 30655–30665. [Google Scholar] [CrossRef]
  37. Li, D.; Chen, D.; Goh, J.; Ng, S.-K. Anomaly Detection with Generative Adversarial Networks for Multivariate Time Series. arXiv 2018, arXiv:1809.04758. [Google Scholar]
  38. Becerra-Rico, J.; Aceves-Fernández, M.A.; Esquivel-Escalante, K.; Pedraza-Ortega, J.C. Airborne Particle Pollution Predictive Model Using Gated Recurrent Unit (GRU) Deep Neural Networks. Earth Sci. Inform. 2020, 13, 821–834. [Google Scholar] [CrossRef]
  39. Hayatbini, N.; Kong, B.; Hsu, K.-L.; Nguyen, P.; Sorooshian, S.; Stephens, G.; Fowlkes, C.; Nemani, R.; Ganguly, S. Conditional Generative Adversarial Networks (cGANs) for Near Real-Time Precipitation Estimation from Multispectral GOES-16 Satellite Imageries—PERSIANN-cGAN. Remote Sens. 2019, 11, 2193. [Google Scholar] [CrossRef]
  40. National Aeronautics and Space Administration (NASA). CMAPSS Jet Engine Simulated Data; NASA Aerospace: Washington, DC, USA, 2023. Available online: https://data.nasa.gov/dataset/cmapss-jet-engine-simulated-data (accessed on 24 October 2025).
  41. Le, A.T.H.; Domingo, N.; Rasheed, E.; Park, K. Maturity Model of Building Maintenance Management for New Zealand’s State Schools. Build. Res. Inf. 2022, 50, 438–451. [Google Scholar] [CrossRef]
  42. Oh, S.H.; Kim, J.G. LSTM-Based PdM Platform for Automobile SCU Inspection Equipment. In Proceedings of the 2023 International Conference on Information Networking (ICOIN), Bangkok, Thailand, 11–14 January 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 363–365. [Google Scholar]
  43. Alavi, H.; Forcada, N.; Fan, S.; San, W. BIM-Based Augmented Reality for Facility Maintenance Management. In Proceedings of the EC3 Conference 2021, Online, 15–16 July 2021; Volume 2, pp. 431–438. [Google Scholar]
  44. Cheng, J.C.; Chen, W.; Chen, K.; Wang, Q. Data-Driven Predictive Maintenance Planning Framework for MEP Components Based on BIM and IoT Using Machine Learning Algorithms. Autom. Constr. 2020, 112, 103087. [Google Scholar] [CrossRef]
  45. Tang, W.; Tan, S.; Li, B.; Huang, J. Automatic Steganographic Distortion Learning Using a Generative Adversarial Network. IEEE Signal Process. Lett. 2017, 24, 1547–1551. [Google Scholar] [CrossRef]
  46. Alahyari, S.; Domaratzki, M. SMOGAN: Synthetic Minority Oversampling with GAN Refinement for Imbalanced Regression. arXiv 2025, arXiv:2504.21152. [Google Scholar] [CrossRef]
  47. Mahmood, T.; Saba, T.; Rehman, A.; Wang, Y. Enhancing Industrial Internet of Things Performance Through Deep Transfer Learning-Based Neural Network Digital Twin Modeling in Data-Scarce Environments. J. Ind. Inf. Integr. 2025, 48, 100956. [Google Scholar] [CrossRef]
  48. Mikołajewska, E.; Mikołajewski, D.; Mikołajczyk, T.; Paczkowski, T. Generative AI in based digital twins for fault diagnosis for predictive maintenance in Industry 4.0/5.0. Appl. Sci. 2025, 15, 3166. [Google Scholar] [CrossRef]
  49. Zhao, R.; Yan, R.; Chen, Z.; Mao, K.; Wang, P.; Gao, R.X. Deep learning and its applications to machine health monitoring. Mech. Syst. Signal Process. 2019, 115, 213–237. [Google Scholar] [CrossRef]
  50. UCI Machine Learning Repository. AI4I 2020 Predictive Maintenance Dataset. 2020. Available online: https://archive.ics.uci.edu/dataset/601/ai4i+2020+predictive+maintenance+dataset (accessed on 24 October 2025).
  51. Feng, S.; Keung, J.; Yu, X.; Xiao, Y.; Zhang, M. Investigation on the Stability of SMOTE-Based Oversampling Techniques in Software Defect Prediction. Inf. Softw. Technol. 2021, 139, 106662. [Google Scholar] [CrossRef]
  52. Li, R.; Li, X.; Heng, P.-A.; Fu, C.-W. Point cloud upsampling via disentangled refinement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 344–353. [Google Scholar]
  53. Vitolo, P.; Licciardo, G.D.; Amendola, A.C.; Di Benedetto, L.; Liguori, R.; Rubino, A.; Pau, D. Quantized ID-CNN for a low-power PDM-to-PCM conversion in TinyML KWS applications. In Proceedings of the 2022 IEEE 4th International Conference on Artificial Intelligence Circuits and Systems (AICAS), Incheon, Republic of Korea, 13–15 June 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 154–157. [Google Scholar]
  54. Vitolo, P.; Liguori, R.; Di Benedetto, L.; Rubino, A.; Pau, D.; Licciardo, G.D. A new nn-based approach to in-sensor pdm-to-pcm conversion for ultra tinyml kws. IEEE Trans. Circuits Syst. II Express Briefs 2022, 70, 1595–1599. [Google Scholar] [CrossRef]
  55. Shen, F.; Zhou, J.; Huang, Z.; Li, L. Going deeper into OSNR estimation with CNN. Photonics 2021, 8, 402. [Google Scholar] [CrossRef]
  56. Arrieta, A.B.; Díaz-Rodríguez, N.; Del Ser, J.; Bennetot, A.; Tabik, S.; Barbado, A.; García, S.; Gil-López, S.; Molina, D.; Benjamins, R.; et al. Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Inf. Fusion 2020, 58, 82–115. [Google Scholar] [CrossRef]
  57. Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why should I trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1135–1144. [Google Scholar]
  58. Shwartz-Ziv, R.; Armon, A. Tabular data: Deep learning is not all you need. Inf. Fusion 2022, 81, 84–90. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.