Hybrid Explainable AI for Machine Predictive Maintenance: From Symbolic Expressions to Meta-Ensembles
Abstract
1. Introduction
- Is it possible to obtain SEs using the GPSC for machine failure detection/classification with a high classification performance?
- Is it possible to generate scaled/normalized balanced dataset variations using scaling/normalization and oversampling techniques, and is there any influence of the classification performance on the obtained SEs?
- Is it possible to find the optimal combination of GPSC hyperparameters with the RHVS method using whichever generated SEs achieved the highest classification performance?
- Is it possible to obtain a robust set of SEs when the GPSC is trained using 5FCV?
- Is it possible to improve the classification performance by developing the TBVE and adjusting the threshold values?
- If the TBVE based on SEs fails, is it possible to create a meta-dataset based on the SEs outputs, then use this meta-dataset and subject it to various preprocessing + oversampling techniques and train the RFC on each balanced dataset variation, and after that combine the best RFC models in the TBVE for each class and achieve the highest classification performance?
2. Materials and Methods
2.1. Research Methodology
- Step 1—Perform the initial dataset statistical analysis and create different variations of the dataset using data scaling/normalization techniques and oversampling techniques, thus creating multiple variations of the initial datasets.
- Step 2—Train the GPSC using the 5-fold cross-validation process and the RHVS method to obtain a set of SEs with a high classification performance on each balanced dataset variation. The RHVS method is used to find the optimal combination of GPSC hyperparameter values using whichever GPSC generates SEs with the highest classification performance on each balanced dataset variation.
- Step 3—Perform the detailed investigation of the best sets of SEs and select only those that have a classification performance higher than 0.95.
- Step 4—Create the Threshold-Based Voting Ensembles (TBVEs) and adjust the threshold value to see if classification performance could be improved when compared to single SEs. If the TBVE based on SEs fails, create the dataset of SEs outputs, apply the preprocessing + oversampling techniques to create balanced dataset variations, and train the RFC using 5FCV and the random hyperparameter search. After the RFCs are trained, create the TBVE based on the RFC and find the optimal threshold value using whichever highest classification performance is achieved.
2.2. Dataset Description and Statistical Analysis
- 1.
- UID—Unique identifier ranging from 1 to 100,000.
- 2.
- Product ID—The product unique identifier.
- 3.
- Air temperature [K]—Generated using a random walk process and later normalized to 300 [K] with a standard deviation of 2 [K].
- 4.
- Process temperature [K]—Also generated using a random walk process and later normalized to a standard deviation of 1 [K] and added to the air temperature plus 10 [K].
- 5.
- Rotational speed [rpm]—Calculated from a power of 2860 [W] with normally distributed noise.
- 6.
- Torque [Nm]—Torque values, normally distributed around 40 [Nm].
- 7.
- Tool wear [min]—The quality variants H/M/L add 5/3/2 min of tool wear to the used tool in the process.
- 8.
- temp_diff—Temperature difference [K].
- 9.
- temp_ratio—Temperature ratio.
- 10.
- rpm_diff—Rotations per minute difference [rpm].
- 11.
- rpm_ratio—Rotations per minute ratio.
- 12.
- Target—No Failure labeled with 0, failure labeled with 1.
- 13.
- Machine failure—Type of machine failure, i.e., No Failure, Power Failure, Tool Wear Failure, Overstrain Failure, Random Failure, and Heat Dissipation Failure.
- No Failure (Class_0)—All the dataset samples that are some type of Failure Type will be labeled as 0 while the No Failure class will be labeled as 1. So, the number of samples for this new class is Failure-348 and No Failure-9652.
- Power Failure (Class_1)—In the new binary classification target variable of the same name, all the dataset samples that do not belong to that class will be named No Power Failure and labeled 0 (9905 samples) while Power Failure class samples will be labeled as 1 (95 samples).
- Tool Wear Failure (Class_2)—In the new binary classification target variable of the same name, all the dataset samples that do not belong to that class will be named No Tool Wear Failure and labeled 0 (9955) while the Tool Wear Failure class will be labeled as 1 (45 samples).
- Overstrain Failure (Class_3)—The new binary classification target variable of the same name will contain two classes, i.e., No Overstrain Failure labeled 0 (9922 samples) while the Overstrain Failure class will be labeled as 1 (78 samples).
- Random Failure (Class_4)—The new binary classification target variable of the same name will contain two classes, i.e., No Random Failure labeled 0 (9982 samples) while Random Failure class samples will be labeled as 1 (18 samples).
- Heat Dissipation Failure (Class_5)—The new binary classification target variable of the same name will contain two classes, i.e., No Heat Dissipation Failure labeled 0 (9888 samples) while Heat Dissipation Failure samples will be labeled 1 (112 samples).
2.3. Description of Dataset Preprocessing Techniques
2.3.1. Scaling/Normalization Techniques
2.3.2. Oversampling Techniques
2.4. The Results of Dataset Preprocessing Techniques
2.5. Genetic Programming Symbolic Classifier with Random Hyperparameter Values Search Method
- PS—The size of the initial population [24].
- GenMax—The maximum number of generations. This hyperparameter is also one of the GPSC termination criteria, and if reached, the GPSC execution is terminated [24].
- DepthInit—The initial depth of the population members. The population members in the GPSC are represented as tree structures, and the size of the population members can be measured with the depth of the population member starting from a root node, which is 0. It should be noted that DepthInit is defined as the range from the minimum to the maximum value. This reason why this hyperparameter is in range is due to the fact that the method used to create the initial population requires the definition of this hyperparameter in this way. The method is ramped half-and-half, which combines two of the oldest methods, the full and grow method, and the depth of the population members is in range, hence the term ramped [24].
- TS—The tournament size hyperparameter value that defines the size of the tournament selection. The size must always be less than the population size [24].
- Stopping criteria—Another GPSC termination criterion, and this hyperparameter is the predefined minimum value of the fitness function. If this value is reached by one of the population members during the GPSC execution, the GPSC execution is terminated. It should be noted that the fitness function in the GPSC is measured with the binary cross-entropy loss function. To measure the binary cross-entropy loss value for one population member in the GPSC, first, input values from the training set are provided to compute the output. Then, this output is used as input in the Sigmoid function:
- -
- is the true label (0 or 1).
- -
- is the predicted probability (the output of the Sigmoid function).
- -
- N is the total number of samples.
- CrossProb—The crossover probability value and one of the genetic operations, which requires two winners of the tournament selection. The random selection of subtrees is performed on both population members, and the subtree of the second tournament selection winner replaces the subtree of the first tournament selection winner [24].
- SubtreeMute—The subtree mutation probability value of the genetic operation that requires only one tournament selection winner. The random subtree is selected on the winner and replaced with a randomly generated subtree from the primitive set [24].
- HoistMute—The probability of the hoist mutation operation, which requires only one population member. The hoist mutation randomly selects a subtreee and randomly selects a node on that subtree. Then, the node replaces the entire subtree [24].
- PointMute—The probability of the point mutation operation; also requires one tournament selection winner, and on that winner, random nodes are selected. The mathematical functions are replaced with other mathematical functions; however, the number of arguments must be the same. The constant values are replaced with randomly selected constant values from the constant range. The input variables are replaced with other randomly selected variables [24].
- MaxSamples—The maximum percentage of samples that will be used for training of the GPSC while the remaining part will be used for testing. Generally, it is good practice to leave at least 1 to 0.1% of the dataset to see how population members perform on the unseen data. In practice, the fitness function of population members on the training data and the test data (out-of-bag fitness) should be similar. If these values largely deviate, this can cause overfitted models [24].
- ParsCoef—The coefficient of the parsimony pressure method. This method penalizes large SEs by increasing the fitness value and thus makes them least favorable for selection. It is one of the most sensitive parameters to tune [24].
- ConstRange—The range of constant values [24].
- Define the lower and upper boundary for each hyperparameter.
- Test the performance of each hyperparameter value, and if the GPSC did not successfully execute, modify the boundaries.
2.6. Evaluation Metrics
- TP—The SE correctly predicts the positive class when the sample from the dataset is also positive. For example, the SE detects machine failure (positive) when the actual value is positive.
- TN—The SE correctly predicts the negative class when the actual class is negative. The SE correctly identifies the machine without failures (negative) when the machine indeed operates normally without failures.
- FP—The SE incorrectly predicts a positive class when the actual class is negative. The SE detects machine failure (positive) while the actual value is a machine with normal operation (negative), resulting in a false alarm.
- FN—The SE incorrectly predicts a negative class when the actual class is positive. In the case of machine failure, the SE predicts that there is No Failure when the actual value is positive, i.e., the failure in the machine occurred.
2.7. Training Testing Procedure
- The balanced dataset variation is divided into the train–test dataset in a ratio of 70:30. Besides the train–test dataset, the random hyperparameter values are selected using the RHVS method.
- The GPSC is trained using 5FCV where, for each split, an SE is generated, so for 5 splits, a total of 5 SEs are obtained. Besides the SEs at this step, the evaluation metric values are obtained on the train and validation folds inside the 5FCV.
- After the 5FCV process is completed and the SEs are obtained, the mean and standard deviation values are obtained. If the values are higher than 0.9, the process proceeds to the test phase; if they are lower than 0.9, the process starts from the beginning with the random selection of GPSC hyperparameter values using the RHVS method.
- If eventually the process reaches the test phase, a test dataset (30%) is provided to obtain SEs to compute the output. After the output is obtained, they are used to calculate the evaluation metric values. If the evaluation metric values are all higher than 0.9, the process is completed, and the best set of SEs is obtained for balanced dataset variation on which the GPSC was trained. If the evaluation metric values are lower than 0.9 then the process starts from the beginning with random selection of GPSC hyperparameter values using the RHVS method.
2.8. Threshold-Based Voting Ensemble
2.9. Alternative Approach with RFC
- n_estimators: The number of trees in the forest. Increasing this value can improve accuracy but may increase computation time.
- max_depth: Controls the maximum depth of each decision tree. Deeper trees can model more complex relationships but may lead to overfitting.
- min_samples_split: Specifies the minimum number of samples required to split an internal node. Higher values help prevent overfitting.
- min_samples_leaf: Determines the minimum number of samples required to be at a leaf node. Larger values can smooth the model, reducing variance.
- bootstrap: Determines whether bootstrap sampling is used. When set to True, it enables bagging, which helps reduce overfitting.
2.10. Computational Resources
- scikit-learn for classification algorithms (Random Forest), model evaluation (cross-validation and performance metrics), hyperparameter optimization (RandomizedSearchCV), and data preprocessing (scaling and normalization).
- gplearn for symbolic regression and the development of the genetic programming symbolic classifier (GPSC), which was extended to suit the classification task.
- imbalanced-learn for oversampling techniques to address class imbalance.
- NumPy and Pandas for efficient numerical computation and data manipulation.
3. Results
- The analysis of the classification performance of the SEs achieved on balanced dataset variations and optimal combinations of GPSC hyperparameter values using the RHVS that was used to obtain SEs.
- The analysis of the SEs with a high classification performance.
- The development of a meta-dataset using the SE output.
- The classification performance of the RFC on balanced dataset variations.
- The development of a TBVE based on the RFC trained models.
3.1. The Analysis of the SEs Obtained on Balanced Dataset Variations for Fault Detection
3.2. The Analysis of the SEs Obtained on Balanced Dataset Variations for Fault Classification
3.3. The Analysis of the Best SEs
3.4. The Development of Meta-Datasets Used for Training of RFC
4. Discussion
- Fault detection (Figure 18) generated a perfect classification performance for threshold values equal to 0.15 or higher.
- Fault classification Class_0 (Figure 19a) generated a perfect classification performance for threshold values equal to 0.85 or higher.
- Fault classification Class_1 (Figure 19b) generated a perfect classification performance for threshold values equal to 0.3 or higher.
- Fault classification Class_2 (Figure 19c) generated a near-perfect classification performance for threshold values equal to 0.1 or higher.
- Fault classification Class_3 (Figure 20a) generated a near-perfect classification performance for threshold values equal to 0.5 or higher.
- Fault classification Class_4 (Figure 20b) generated a near-perfect classification performance for threshold values equal to 0.1 or higher.
- Fault classification Class_5 (Figure 20c) generated a near-perfect classification performance for threshold values equal to 0.5 or higher.
- Inherent interpretability, enabling deeper insight into fault mechanisms;
- High classification accuracy, comparable or superior to existing black-box models;
- Robustness to imbalanced data, achieved through systematic oversampling and one-vs-rest decomposition;
- Modular ensemble design, which can be tuned and extended for real-time monitoring and incremental learning.These features make the proposed method well-suited for industrial adoption, where both performance and explainability are essential.
5. Conclusions
- The GPSC generated SEs with a high classification performance in machine failure.
- The data scaling/normalization techniques and oversampling techniques generated a large number of dataset balancing techniques. Using these techniques introduced a large number of dataset variations that in the end generated a large number of SEs with a high classification performance.
- The 5FCV training process of the GPSC generated a large number of SEs. Using this training process created a robust set of SEs with a high classification performance.
- The RHVS method proved useful in finding an optimal combination of GPSC hyperparameters using whichever obtained set of SEs had a high classification performance.
- The combination of SEs failed to produce a TBVE with a stable performance. However, a dataset created from the outputs of SEs provides a meta-dataset used for training the RFCs. The TBVEs based on the RFCs generated a stable performance with high classification accuracy.
- Unlike ANNs or other deep learning methods, the GPSC generates SEs that can be explainable, which are easier to handle and process.
- The application of preprocessing and oversampling techniques generated a large number of balanced dataset variations. Using this approach, the balanced datasets have synthetic data based on the statistics of the original dataset.
- The RHVS method generally finds the optimal combination of GPSC hyperparameter values very quickly when compared to grid search methods.
- The TBVE based on the RFC proved to be a very useful approach in creating a robust system with a high classification performance with the adjustment of threshold values.
- The GPSC training is computationally expensive, especially with large population sizes, many generations, and multiple balanced dataset variations created by preprocessing and oversampling.
- The Random Hyperparameter Value Search (RHVS) method, although effective, sometimes requires many GPSC runs to meet performance thresholds, which can further increase training time.
- The method’s performance is lower on classes with extremely limited samples, indicating challenges in handling highly imbalanced or rare Failure Types.
- Reducing GPSC complexity by tuning population size and generation count to balance performance and runtime.
- Investigating alternative hyperparameter optimization techniques such as Bayesian optimization to improve efficiency over RHVS.
- Developing advanced oversampling or data augmentation methods tailored to minority classes with very few samples to boost detection accuracy.
- Applying model pruning or distillation methods to simplify symbolic expressions and ensemble models for faster inference.
- Testing the framework on larger and more diverse real-world predictive maintenance datasets to validate scalability and robustness.
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
Appendix A. The Modified Mathematical Functions Used in GPSC
Appendix B. Where to Find SEs Obtained in This Research and How to Use Them
References
- Molęda, M.; Małysiak-Mrozek, B.; Ding, W.; Sunderam, V.; Mrozek, D. From corrective to predictive maintenance—A review of maintenance approaches for the power industry. Sensors 2023, 23, 5970. [Google Scholar] [CrossRef] [PubMed]
- Lee, J.; Ni, J.; Singh, J.; Jiang, B.; Azamfar, M.; Feng, J. Intelligent maintenance systems and predictive manufacturing. J. Manuf. Sci. Eng. 2020, 142, 110805. [Google Scholar] [CrossRef]
- Yazdi, M. Maintenance strategies and optimization techniques. In Advances in Computational Mathematics for Industrial System Reliability and Maintainability; Springer: Berlin/Heidelberg, Germany, 2024; pp. 43–58. [Google Scholar]
- Uptime Institute. Global Data Center Survey, 2022. Available online: https://uptimeinstitute.com (accessed on 2 July 2025).
- Deloitte. Predictive Maintenance: Next-Generation Maintenance Strategy. 2017. Available online: https://www2.deloitte.com (accessed on 2 July 2025).
- Shah, A.; Kumar, R.; Jain, V. Forecasting Remaining Useful Life for Wind Turbine Predictive Maintenance Based on Deep Learning. arXiv 2024, arXiv:2412.17823. [Google Scholar]
- Al Hasib, A.; Islam, M.; Chowdhury, M. An Interpretable Systematic Review of Machine Learning Models for Predictive Maintenance of Aircraft Engines. arXiv 2023, arXiv:2309.13310. [Google Scholar]
- Carvalho, T.P.; Soares, F.A.; Vita, R.; Francisco, R.d.P.; Basto, J.P.; Alcalá, S.G. A systematic literature review of machine learning methods applied to predictive maintenance. Comput. Ind. Eng. 2019, 137, 106024. [Google Scholar] [CrossRef]
- Vuttipittayamongkol, P.; Arreeras, T. Data-driven industrial machine failure detection in imbalanced environments. In Proceedings of the 2022 IEEE International Conference on Industrial Engineering and Engineering Management (IEEM), Kuala Lumpur, Malaysia, 7–10 December 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1224–1227. [Google Scholar]
- Luo, B.; Wang, H.; Liu, H.; Li, B.; Peng, F. Early fault detection of machine tools based on deep learning and dynamic identification. IEEE Trans. Ind. Electron. 2018, 66, 509–518. [Google Scholar] [CrossRef]
- Samanta, B.; Al-Balushi, K.R.; Al-Araimi, S.A. Bearing fault detection using artificial neural networks and genetic algorithm. EURASIP J. Adv. Signal Process. 2004, 2004, 785672. [Google Scholar] [CrossRef]
- Leh, N.A.M.; Zain, F.M.; Muhammad, Z.; Abd Hamid, S.; Rosli, A.D. Fault detection method using ANN for power transmission line. In Proceedings of the 2020 10th IEEE International Conference on Control System, Computing and Engineering (ICCSCE), Penang, Malaysia, 21–22 August 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 79–84. [Google Scholar]
- Mesai Belgacem, A.; Hadef, M.; Djerdir, A. Fault diagnosis of photovoltaic array based on gated residual network with multi-head self attention mechanism. Turk. J. Electr. Eng. Comput. Sci. 2024, 32, 806–828. [Google Scholar] [CrossRef]
- Apeiranthitis, S.; Zacharia, P.; Chatzopoulos, A.; Papoutsidakis, M. Predictive maintenance of machinery with rotating parts using Convolutional Neural Networks. Electronics 2024, 13, 460. [Google Scholar] [CrossRef]
- Pantelis, L.; Papastefanopoulos, V.; Kotsiantis, S. Explainable ai: A review of machine learning interpretability methods. Entropy 2023, 23, 18. [Google Scholar]
- Dereci, U.; Tuzkaya, G. An explainable artificial intelligence model for predictive maintenance and spare parts optimization. Supply Chain Anal. 2024, 8, 100078. [Google Scholar] [CrossRef]
- Kramer, O.; Kramer, O. Scikit-learn. In Machine Learning for Evolution Strategies; Springer: Berlin/Heidelberg, Germany, 2016; pp. 45–53. [Google Scholar]
- Fernández, A.; Garcia, S.; Herrera, F.; Chawla, N.V. SMOTE for learning from imbalanced data: Progress and challenges, marking the 15-year anniversary. J. Artif. Intell. Res. 2018, 61, 863–905. [Google Scholar] [CrossRef]
- He, H.; Bai, Y.; Garcia, E.A.; Li, S. ADASYN: Adaptive synthetic sampling approach for imbalanced learning. In Proceedings of the 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence), Hong Kong, China, 1–6 June 2008; IEEE: Piscataway, NJ, USA, 2008; pp. 1322–1328. [Google Scholar]
- Han, H.; Wang, W.Y.; Mao, B.H. Borderline-SMOTE: A new over-sampling method in imbalanced data sets learning. In Proceedings of the International Conference on Intelligent Computing, Hefei, China, 23–26 August 2005; Springer: Berlin/Heidelberg, Germany, 2005; pp. 878–887. [Google Scholar]
- Fonseca, J.; Douzas, G.; Bacao, F. Improving imbalanced land cover classification with K-Means SMOTE: Detecting and oversampling distinctive minority spectral signatures. Information 2021, 12, 266. [Google Scholar] [CrossRef]
- Bagui, S.S.; Mink, D.; Bagui, S.C.; Subramaniam, S. Determining Resampling Ratios Using BSMOTE and SVM-SMOTE for Identifying Rare Attacks in Imbalanced Cybersecurity Data. Computers 2023, 12, 204. [Google Scholar] [CrossRef]
- Langdon, W.B.; Poli, R. Foundations of Genetic Programming; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2013. [Google Scholar]
- O’Neill, M.; Poli, R.; Langdon, W.B.; McPhee, N.F. A Field Guide to Genetic Programming; Springer: Berlin/Heidelberg, Germany, 2008; 250p, ISBN 978-1-4092-0073-4. [Google Scholar]
- Kyurkchiev, N.; Markov, S. Sigmoid functions: Some approximation and modelling aspects. LAP LAMBERT Acad. Publ. Saarbr. 2015, 4, 34. [Google Scholar]
- Hurtik, P.; Tomasiello, S.; Hula, J.; Hynar, D. Binary cross-entropy with dynamical clipping. Neural Comput. Appl. 2022, 34, 12029–12041. [Google Scholar] [CrossRef]
- Ben-David, A.; Frank, E. Accuracy of machine learning models versus “hand crafted” expert systems–a credit scoring case study. Expert Syst. Appl. 2009, 36, 5264–5271. [Google Scholar] [CrossRef]
- De Hond, A.A.; Steyerberg, E.W.; Van Calster, B. Interpreting area under the receiver operating characteristic curve. Lancet Digit. Health 2022, 4, e853–e855. [Google Scholar] [CrossRef] [PubMed]
- Zhou, Q.M.; Zhe, L.; Brooke, R.J.; Hudson, M.M.; Yuan, Y. A relationship between the incremental values of area under the ROC curve and of area under the precision-recall curve. Diagn. Progn. Res. 2021, 5, 1–15. [Google Scholar] [CrossRef] [PubMed]
- Yacouby, R.; Axman, D. Probabilistic extension of precision, recall, and f1 score for more thorough evaluation of classification models. In Proceedings of the First Workshop on Evaluation and Comparison of NLP Systems, Online, 20 November 2020; pp. 79–91. [Google Scholar]
- Visa, S.; Ramsay, B.; Ralescu, A.L.; Van Der Knaap, E. Confusion matrix-based feature selection. Maics 2011, 710, 120–127. [Google Scholar]
References | Methods | Results (ACC %) |
---|---|---|
[9] | SVM, DTC, KNN, RFC, ANN | 91 |
[10] | DAE | 98.1 |
[11] | GA + ANN, GA + RBF, GA + PNN | 100 |
[12] | ANN | 100 |
[13] | GRN + MHSA | 99.71 |
[14] | CNN | 100% |
[15] | LSTM, RNN, GRU RFC, KNN, NB, and GB | 97.18 |
[16] | ABC | 99.91 |
Dataset Variables | Count | Mean | Std | Min | Max | GPSC Variable Representation |
---|---|---|---|---|---|---|
Type | 10,000 | 0.8006 | 0.60023 | 0 | 2 | |
Air temperature [K] | 10,000 | 300.0049 | 2.000259 | 295.3 | 304.5 | |
Process temperature [K] | 10,000 | 310.0056 | 1.483734 | 305.7 | 313.8 | |
Rotational speed [rpm] | 10,000 | 1538.776 | 179.2841 | 1168 | 2886 | |
Torque [Nm] | 10,000 | 39.98691 | 9.968934 | 3.8 | 76.6 | |
Tool wear [min] | 10,000 | 107.951 | 63.65415 | 0 | 253 | |
temp_diff [K] | 10,000 | 10.00063 | 1.001094 | 7.6 | 12.1 | |
temp_ratio | 10,000 | 1.033352 | 0.003492 | 1.025116 | 1.040672 | |
rpm_diff [rpm] | 10,000 | −1498.79 | 188.0691 | −2882.2 | −1104.6 | |
rpm_ratio | 10,000 | 0.026895 | 0.008838 | 0.001317 | 0.063833 | |
Target | 10,000 | 0.0339 | 0.180981 | 0 | 1 | |
Failure type | 10,000 | 0.1051 | 0.628883 | 0 | 5 |
Hyperparameter Name | Lower Boundary | Upper Boundary |
---|---|---|
PS | 500 | 1000 |
GenMax | 50 | 100 |
DepthInit | (3, 7) | (12, 18) |
TS | 250 | 500 |
Stopping criteria | ||
CrossProb | 0.001 | 0.1 |
SubtreeMute | 0.95 | 1 |
HoistMute | 0.001 | 1 |
PointMute | 0.001 | 1 |
MaxSamples | 0.99 | 1 |
ParsCoeff | ||
ConstRange | −1000 | 1000 |
Predicted Positive | Predicted Negative | |
---|---|---|
Actual Positive | TP | FN |
Actual Negative | FP | TN |
Hyperparameter | Description |
---|---|
n_estimators | Number of trees in the forest (50, 100, 200, 300, 400, 500) |
max_depth | Maximum depth of the trees (None, 10, 20, 30, 40, 50) |
min_samples_split | Minimum number of samples required to split a node (2, 5, 10) |
min_samples_leaf | Minimum number of samples required to be at a leaf node (1, 2, 4) |
bootstrap | Whether bootstrap samples are used when building trees (True, False) |
Parameter | Count | Mean | Std | Min | Max |
---|---|---|---|---|---|
PopSize | 29 | 670.714 | 333.280 | 0 | 996 |
MaxGen | 29 | 70.314 | 34.305 | 0 | 100 |
TourSize | 29 | 298.314 | 155.409 | 0 | 498 |
InitD_Lower | 29 | 4.114 | 2.246 | 0 | 7 |
InitD_Upper | 29 | 12.543 | 6.060 | 0 | 18 |
Crossover | 29 | 0.010 | 0.009 | 0 | 0.033 |
SubMute | 29 | 0.798 | 0.368 | 0 | 0.985 |
HoistMute | 29 | 0.010 | 0.009 | 0 | 0.030 |
PointMute | 29 | 0.010 | 0.009 | 0 | 0.038 |
StopCrit | 29 | 0.000 | 0.000 | 0 | 0.001 |
MaxSamp | 29 | 0.824 | 0.380 | 0 | 1.000 |
ConstRange_Lower | 29 | −475.396 | 314.405 | −962.228 | 0 |
ConstRange_Upper | 29 | 375.563 | 296.147 | 0 | 973.359 |
ParsCoeff | 29 | 0.000 | 0.000 | 0 | 0.001 |
GPSC Hyperparameter | Class _0 | Class_1 | ||||||
---|---|---|---|---|---|---|---|---|
Mean | Std | Min | Max | Mean | Std | Min | Max | |
PopSize | 636.2 | 349.9694 | 0 | 1000 | 645.8 | 323.494 | 0 | 1000 |
MaxGen | 62.57143 | 33.45347 | 0 | 98 | 64.8 | 32.59087 | 0 | 100 |
TourSize | 280.5429 | 153.3927 | 0 | 493 | 315.2857 | 158.9537 | 0 | 492 |
InitD_Lower | 4.114286 | 2.361099 | 0 | 7 | 3.8 | 2.272599 | 0 | 7 |
InitD_Upper | 12 | 6.310589 | 0 | 18 | 12.28571 | 5.853965 | 0 | 18 |
Crossover | 0.012673 | 0.010371 | 0 | 0.032868 | 0.010981 | 0.008713 | 0 | 0.028027 |
SubMute | 0.765721 | 0.388484 | 0 | 0.97257 | 0.797924 | 0.36834 | 0 | 0.983656 |
HoistMute | 0.010961 | 0.011603 | 0 | 0.046202 | 0.009032 | 0.007416 | 0 | 0.031138 |
PointMute | 0.010315 | 0.00948 | 0 | 0.030935 | 0.010296 | 0.008465 | 0 | 0.029677 |
StopCrit | 0.000402 | 0.000343 | 0 | 0.000971 | 0.000359 | 0.000268 | 0 | 0.000886 |
MaxSamp | 0.795732 | 0.403685 | 0 | 0.999575 | 0.823983 | 0.380276 | 0 | 0.998815 |
ConstRange_Lower | −433.744 | 363.0157 | −984.883 | 0 | −420.764 | 325.7381 | −911.98 | 0 |
ConstRange_Upper | 357.5343 | 336.3315 | 0 | 972.5153 | 413.6174 | 336.8456 | 0 | 962.6954 |
ParsCoeff | 0.000373 | 0.000274 | 0 | 0.000892 | 0.000407 | 0.000294 | 0 | 0.000936 |
Class_2 | Class_3 | |||||||
PopSize | 657.5714 | 355.9934 | 0 | 999 | 640.0571 | 344.8543 | 0 | 981 |
MaxGen | 61.68571 | 33.72446 | 0 | 100 | 59.68571 | 32.79954 | 0 | 99 |
TourSize | 305 | 169.1526 | 0 | 493 | 297.3143 | 168.5285 | 0 | 496 |
InitD_Lower | 3.8 | 2.25962 | 0 | 7 | 4.057143 | 2.496384 | 0 | 7 |
InitD_Upper | 11.45714 | 6.065053 | 0 | 18 | 12 | 6.36165 | 0 | 18 |
Crossover | 0.008244 | 0.00695 | 0 | 0.024056 | 0.012976 | 0.011618 | 0 | 0.037957 |
SubMute | 0.772274 | 0.391877 | 0 | 0.991849 | 0.768624 | 0.389982 | 0 | 0.981681 |
HoistMute | 0.008997 | 0.008584 | 0 | 0.027383 | 0.007199 | 0.006724 | 0 | 0.024263 |
PointMute | 0.0101 | 0.009453 | 0 | 0.036996 | 0.010834 | 0.012729 | 0 | 0.037118 |
StopCrit | 0.000487 | 0.000352 | 0 | 0.000994 | 0.000391 | 0.000293 | 0 | 0.000978 |
MaxSamp | 0.796346 | 0.403996 | 0 | 0.999685 | 0.795956 | 0.403798 | 0 | 0.999844 |
ConstRange_Lower | −442.636 | 324.1984 | −957.474 | 0 | −484.726 | 344.8748 | −985.995 | 0 |
ConstRange_Upper | 436.2911 | 337.2926 | 0 | 982.044 | 419.3863 | 339.6644 | 0 | 985.2862 |
ParsCoeff | 0.000359 | 0.000286 | 0 | 0.000963 | 0.000357 | 0.000316 | 0 | 0.001 |
Class_4 | Class_5 | |||||||
PopSize | 581.7714 | 368.8178 | 0 | 999 | 613.4857 | 366.2327 | 0 | 998 |
MaxGen | 58.17143 | 36.54368 | 0 | 100 | 56.08571 | 33.45266 | 0 | 100 |
TourSize | 275.8571 | 173.916 | 0 | 494 | 281.1429 | 168.6158 | 0 | 491 |
InitD_Lower | 4 | 2.634611 | 0 | 7 | 3.885714 | 2.552639 | 0 | 7 |
InitD_Upper | 11.54286 | 7.122222 | 0 | 18 | 11.82857 | 6.758375 | 0 | 18 |
Crossover | 0.01186 | 0.011 | 0 | 0.033938 | 0.009188 | 0.009472 | 0 | 0.030846 |
SubMute | 0.713995 | 0.426275 | 0 | 0.978499 | 0.740102 | 0.408809 | 0 | 0.990287 |
HoistMute | 0.008839 | 0.00874 | 0 | 0.034525 | 0.011464 | 0.010145 | 0 | 0.031073 |
PointMute | 0.00782 | 0.007817 | 0 | 0.027049 | 0.010248 | 0.010502 | 0 | 0.030401 |
StopCrit | 0.000333 | 0.000309 | 0 | 0.000986 | 0.000358 | 0.000311 | 0 | 0.000879 |
MaxSamp | 0.739088 | 0.441198 | 0 | 0.999657 | 0.767757 | 0.424023 | 0 | 0.999962 |
ConstRange_Lower | −391.923 | 351.6592 | −988.622 | 0 | −466.145 | 347.1164 | −991.948 | 0 |
ConstRange_Upper | 384.5892 | 328.1986 | 0 | 920.976 | 356.9264 | 312.7266 | 0 | 989.7269 |
ParsCoeff | 0.000358 | 0.000299 | 0 | 0.000888 | 0.000353 | 0.000296 | 0 | 0.000917 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Anđelić, N.; Baressi Šegota, S.; Mrzljak, V. Hybrid Explainable AI for Machine Predictive Maintenance: From Symbolic Expressions to Meta-Ensembles. Processes 2025, 13, 2180. https://doi.org/10.3390/pr13072180
Anđelić N, Baressi Šegota S, Mrzljak V. Hybrid Explainable AI for Machine Predictive Maintenance: From Symbolic Expressions to Meta-Ensembles. Processes. 2025; 13(7):2180. https://doi.org/10.3390/pr13072180
Chicago/Turabian StyleAnđelić, Nikola, Sandi Baressi Šegota, and Vedran Mrzljak. 2025. "Hybrid Explainable AI for Machine Predictive Maintenance: From Symbolic Expressions to Meta-Ensembles" Processes 13, no. 7: 2180. https://doi.org/10.3390/pr13072180
APA StyleAnđelić, N., Baressi Šegota, S., & Mrzljak, V. (2025). Hybrid Explainable AI for Machine Predictive Maintenance: From Symbolic Expressions to Meta-Ensembles. Processes, 13(7), 2180. https://doi.org/10.3390/pr13072180