Machine Learning Approaches for Predicting the Ablation Performance of Ceramic Matrix Composites

: Materials used in aircraft engines, gas turbines, nuclear reactors, re-entry vehicles, and hypersonic structures are subject to severe environmental conditions that present significant challenges. With their remarkable properties, such as high melting temperatures, strong resistance to oxidation, corrosion, and ablation, minimal creep, and advantageous thermal cycling behavior, ceramic matrix composites (CMCs) show great promise as a material to meet the strict requirements in these kinds of environments. Furthermore, the addition of boron nitride nanoparticles with continuous fibers to the CMCs can offer thermal resistivity in harsh conditions, which will improve the composites’ strength and fracture toughness. Therefore, in extreme situations, it is crucial to understand the thermal resistivity period of composite materials. To forecast the ablation performance of composites, we developed six machine learning regression methods in this study: decision tree, random forest, support vector machine, gradient boosting, extreme gradient boosting, and adaptive boosting. When evaluating model performance using metrics including R2 score, root mean square error, mean absolute error, and mean absolute percentage error, the gradient boosting and extreme gradient boosting machine learning regression models performed better than the others. The effectiveness of machine learning models as a useful tool for forecasting the ablation behavior of ceramic matrix composites was effectively explained by this study.


Introduction
The development of composite materials suitable for ultra-high-temperature (UHT) applications is currently the focus of much attention in the scientific community.HfB 2 , ZrB 2 , HfC, and ZrC are examples of transition metal borides and carbides that are particularly important because of their exceptionally high melting temperatures-which frequently approach or surpass 3000 • C-and their capacity to produce stable, oxidation-resistant compounds [1][2][3].The materials known as ultra-high-temperature ceramics (UHTCs) are being researched in great detail for applications that require resistance to oxidation and erosion at temperatures higher than 2000 • C [4][5][6].The application of UHTCs as a potential thermal protective covering has the potential to improve carbon-based composites' ablation resistance in harsh settings when exposure to oxygen and extremely high temperatures are present [7][8][9].This is explained by the remarkable qualities of the UHTC, which include good modulus, enhanced hardness, high melting point, and remarkable chemical inertness and ablation resistance [10][11][12][13].Their prospective applications range from cutting edge aerospace vehicle components, such as control surfaces, engine inlets and exits, and hot flow route components, to creative thermal protection systems [14][15][16].The shortcomings of single-phase ceramics, on the other hand, like their low resistance to heat shock and defects, have led to research on substitute strategies.These materials do not have the required high-temperature, thermal shock, or fracture toughness properties, even after adding phases like SiC [17].This indicated that a fiber-reinforced composite strategy is a promising path to explore.Carbon fiber (Cf) and silicon carbide (SiC) fiber are potential candidates if they can provide adequate protection at high application temperatures.
Traditionally, the melting infiltration process is used to create ceramic matrix composites (CMCs).The brittle structure and excessive porosity of the fabricated CMCs made using this process make them incapable of withstanding high mechanical and thermal demands.As an alternative, carbon fibers are added to a ceramic matrix generated from polymers to create polymer-derived ceramic composites with exceptional fracture toughness.Carbon fiber may be able to tolerate high temperatures without oxidizing with the use of protective coatings made of metallic or ceramic materials, such as nickel [18].Ultra-high-temperature CMCs have a wide range of industrial applications, including space and aerospace propulsion, turbine engines, and nuclear reactors.They can tolerate high thermal loads in harsh conditions [19].Therefore, it is important to forecast the ultra-high-temperature CMCs' ablation ability to learn more about how long they will last in harsh conditions.
Machine learning (ML) algorithms are very useful for predicting the mechanical properties of composite materials [20].To make accurate predictions or respond appropriately to new and unknown inputs, machine learning algorithms must be able to recognize patterns and relationships within datasets [21,22].One of the key components of machine learning that underpins its many applications in a variety of disciplines is its innate capacity to learn from data, generalize to new cases, and make acceptable outcomes without having explicit instructions [23][24][25].Qi et al. [26] utilized a decision tree model to forecast the mechanical properties of plastic reinforced with carbon fiber.Similarly, to predict the mechanical properties of polymer composites with alumina modifiers, Kosicka et al. [27] utilized a decision tree approach.Daghigh et al. [28] forecasted the fracture toughness of bio-nanocomposites at numerous sizes with different particle fillers using decision trees and adaptive boosting (AdaBoost) machine learning techniques.Using a random forest regression technique, Hegde et al. [29] investigated the mechanical properties, in particular the hardness of vacuum-sintered Ti-6Al-4V reinforced with SiCp composites.Zhang et al. [30] similarly used a random forest model to predict the mechanical properties of composite laminates, showing that the model could produce precise predictions in a shorter amount of time.Guo et al. [31] showed that the XGBoost regression algorithm is superior to artificial neural networks, support vector regression, classification, and regression trees when it comes to precisely forecasting the tensile strength, ductility, and compressive strength of high-performance fiber-reinforced cementitious composites under the same conditions.
Based on molecular dynamics datasets, Liu et al. [32] used the AdaBoost regression technique to predict the ultimate tensile strength and Young's modulus of graphenereinforced aluminum nanocomposites.Using additional material parameters, Karamov et al. [33] predicted the fracture toughness of pultruded composites using an AdaBoost regression algorithm.A gradient-boosting regression approach was used by Pathan et al. [34] to predict the yield strengths and macroscopic elastic stiffness of unidirectional fiber composites.Support vector machine regression (SVR) was used by Bonifácio et al. [35] to forecast the mechanical properties of concrete, such as its compressive strength and static Young's modulus.The results showed that SVR had an acceptable predictive capacity.SVR was used by Hasanzadeh et al. [36] to estimate the compressive, flexural, and tensile strengths of basalt-fiber-reinforced concrete.This conventional machine learning technique was also used to simulate the modulus of elasticity and compressive stress-strain curves.The decision tree, extra tree, XGBoost, and random forest regression models' feature relevance analyses revealed that the magnesium matrix composites' reinforcing particle form had the biggest impact on their mechanical properties [37].
Prediction models based on machine learning are widely used in literature to develop and forecast the mechanical properties of composite materials.The literature does, however, appear to be lacking in information about the prediction of thermal properties, particularly for ceramic matrix composites that are subjected to extremely high temperatures.This is a critical component to gain understanding of how long ultra-high-temperature ceramics last in harsh situations.To fill the research gap and open the doors of new knowledge, this study introduced six machine learning regression models intended to estimate the ablation performance of continuous fiber-reinforced silicon oxy-carbide ceramic matrix composites.The main contributions of this research were developing the decision tree (DT), random forest (RF), support vector machine (SVM), gradient boost (GB), extreme gradient boosting (XGBoost), and adaptive boosting (AdaBoost) regression models using the CMC's experimental burning through time thermal data collected from the oxy-acetylene torch test of ceramic matrix composites.The boosting ensemble techniques (GB, XGBoost, AdaBoost) showed better performance accuracy than the bagging ensemble techniques (random forest, decision trees) and traditional SVM regression modeling for predicting the thermal performance of ceramic matrix composites.This study presented a novel framework for forecasting the ablation performance of ceramic matrix composites under harsh environmental circumstances by using multiple machine learning regression models.
Top of Form

Data Preparation for Predictive Modeling
The continuous-fiber-reinforced silicon oxy-carbide ceramic matrix composite (CMC) sample was heated to a maximum temperature of 2200 • C in the center of the fire nozzle during the torch test.Poly-siloxane (PSX) resin and woven carbon fabrics were used in the polymer infiltration and pyrolysis technique (PIP) to create continuous-fiber-reinforced silicon oxy-carbide composite.During the oxyacetylene torch test, when the slidable panel is pushed, the sample, which is clamped to it, will encounter the heat.To detect the heat flow of the flame at different distances from the panels, a heat flux sensor was calibrated.Panels were intended to be slid a specific distance along a rail.The temperature of the back surface was determined using thermocouples.Regulators were used to manage the varied ratios of oxygen and acetylene, allowing for the optimization of both flame type and temperature.The American Society for Testing and Materials (ASTM) 258-08 standard test was followed for calibration and testing.
Using a centrifuge, boron nitride (BN) nanoparticles were combined with PSX resin at a weight ratio of 5%.After that, the woven carbon fabric was penetrated by the mixed resin, and six layers of lamination were applied.The carbon fiber or silicon carbide (Cf/SiC) CMCs supplemented with BN nanoparticles underwent pyrolysis in a tube furnace after being post-cured and cured in an autoclave.The torch test was conducted in the direction of thickness until the samples burned through, providing ablative qualities such erosion rate and insulating index.A total of three similar torch tests were conducted with a 524 W/cm 2 heat flux and same BN weight ratio was maintained at a 25 mm distance from the nozzle.Then, the CMC's burning through temperature data were collected from each test for developing the machine learning prediction models.

Machine Learning Predictive Modeling
Python 3.9 was used in this study from google colab for executing machine learning (ML) models.All the machine learning algorithms applied in this study followed an standard method described in scikit learn [38] with Python.Figure 1 is the graphical representation of the methodology followed for developing every machine learning model in this study.Figure 1.Methodology followed in developing machine learning regression models.

Decision Tree Regression Modeling
Decision trees (DTs) are predictive models in supervised learning that are wellknown for their robustness, interpretability, and undeniable utility in a variety of applications [39].The decision tree is a tree-based algorithm that follows a top-down approach to split a series of trees.The main reason for using decision tree technique in our study is the randomization that each feature introduces, which helps to lower the variance in the time temperature dataset.During the spli ing process, Friedman mean square error (MSE) was used to calculate variance reduction for every root and leaf node.The principle of each split was to maximize the variance reduction ( ) function (1) where  ,  ,  ,  , and  represent weights in the n th nodes, predicted temperature, average predicted temperature of every feature, and variance in the root and leaf nodes (2), respectively.

Decision Tree Regression Modeling
Decision trees (DTs) are predictive models in supervised learning that are well-known for their robustness, interpretability, and undeniable utility in a variety of applications [39].The decision tree is a tree-based algorithm that follows a top-down approach to split a series of trees.The main reason for using decision tree technique in our study is the randomization that each feature introduces, which helps to lower the variance in the time temperature dataset.During the splitting process, Friedman mean square error (MSE) was used to calculate variance reduction for every root and leaf node.The principle of each split was to maximize the variance reduction (V R ) function (1) where W i , T i , T i , V Root , and V Lea f represent weights in the n th nodes, predicted temperature, average predicted temperature of every feature, and variance in the root and leaf nodes (2), respectively.
In our research, BN4, BN5, and BN6 had 4813, 4429, and 4391 datapoints, respectively.To minimize the overfitting and underfitting problem, maximum depth of the tree was selected as 3. Several hyperparameters including splitter (best or random), minimum samples split (2,3,4), minimum samples leaf (1, 2, 3), minimum weight fraction leaf (0.0, 0.1, 0.2), maximum features (none, sqrt, log2), random state (none), maximum leaf nodes (none, 10, 20, 30), minimum impurity decrease (0.0, 0.1, 0.2), and minimal cost complexity pruning (0.0, 0.1, 0.2) were tuned to optimize the final model.During tree development, the splitter decides how to choose the split at each node.Whereas 'random' selects the best random split, 'best' selects the best split.The minimum number of samples needed to separate an internal node is determined using the minimum samples split calculation.It also regulates node size when forming a tree.Minimum samples leaf establishes the bare minimum of samples needed for a leaf node.It affects how finely detailed the leaves are in the finished tree.Minimal weight fraction leaf is comparable to minimum samples leaf except it is represented as a percentage of the weighted sum of all instances.The maximum features are the highest count of attributes considered while dividing a node.The 'log2' value represents the logarithm base 2 of the total features, 'sqrt' takes the square root of the entire features, and 'none' indicates that all features are considered.Reproducibility is ensured by using a random state as the seed for the random number generator.Maximum leaf nodes restricts the amount of leaf nodes within the tree.A certain number limits the maximum number of leaf nodes, whereas 'none' permits infinite growth.Minimum impurity decrease indicates that if the impurity decrease is greater than this cutoff, the node is split.It restricts the growth of the tree by demanding a minimum improvement in impurity before a split can happen.Minimal cost complexity pruning influences the cost of adding more nodes to the tree during pruning.
The grid search method was used to find the optimized hyperparameters in the decision tree modeling.Figure 2 shows the decision tree modeling for the BN4 sample using 70% training and 30% test data.In this model, 4813 time-temperature datapoints were used for 144.36 s and the value represented the average predicted temperature (T i ) of the selected time-temperature samples.
to separate an internal node is determined using the minimum samples split calculation.It also regulates node size when forming a tree.Minimum samples leaf establishes the bare minimum of samples needed for a leaf node.It affects how finely detailed the leaves are in the finished tree.Minimal weight fraction leaf is comparable to minimum samples leaf except it is represented as a percentage of the weighted sum of all instances.The maximum features are the highest count of a ributes considered while dividing a node.The 'log2' value represents the logarithm base 2 of the total features, 'sqrt' takes the square root of the entire features, and 'none' indicates that all features are considered.Reproducibility is ensured by using a random state as the seed for the random number generator.Maximum leaf nodes restricts the amount of leaf nodes within the tree.A certain number limits the maximum number of leaf nodes, whereas 'none' permits infinite growth.Minimum impurity decrease indicates that if the impurity decrease is greater than this cutoff, the node is split.It restricts the growth of the tree by demanding a minimum improvement in impurity before a split can happen.Minimal cost complexity pruning influences the cost of adding more nodes to the tree during pruning.
The grid search method was used to find the optimized hyperparameters in the decision tree modeling.Figure 2 shows the decision tree modeling for the BN4 sample using 70% training and 30% test data.In this model, 4813 time-temperature datapoints were used for 144.36 s and the value represented the average predicted temperature ( ) of the selected time-temperature samples.

Random Forest Regression Modeling
To develop a more reliable model for regression problems and increasing the performance from decision tree temperature prediction model, random forest (RF) regression was introduced which entailed training an ensemble of decision trees on a random subset of the bootstrap sampled data, and then combining their predictions following bootstrap aggregations [37].In this study, the row sampling with replacement method was utilized to collect a subset (bootstrap sampled) of torch test time-temperature data () from the total dataset , where  >  .Then, the subset of time-temperature data was passed through the five individual base learner trees having a constant depth of 3 in each tree.

Random Forest Regression Modeling
To develop a more reliable model for regression problems and increasing the performance from decision tree temperature prediction model, random forest (RF) regression was introduced which entailed training an ensemble of decision trees on a random subset of the bootstrap sampled data, and then combining their predictions following bootstrap aggregations [37].In this study, the row sampling with replacement method was utilized to collect a subset (bootstrap sampled) of torch test time-temperature data (d) from the total dataset d, where d > d.Then, the subset of time-temperature data was passed through the five individual base learner trees having a constant depth of 3 in each tree.Finally, the bootstrap aggregation method was followed to predict the thermal resistivity of the ceramic matrix composites as shown in Figure 3.
Finally, the bootstrap aggregation method was followed to predict the thermal resistivity of the ceramic matrix composites as shown in Figure 3. Various hyperparameters were adjusted to optimize the final model, such as the minimum samples split (2,3,4), minimum samples leaf (1, 2, 3), minimum weight fraction leaf (0.0, 0.1, 0.2), maximum features (none, sqrt, log2), random state (none), maximum leaf nodes (none, 10, 20, 30), minimum impurity decrease (0.0, 0.1, 0.2), bootstrap (true), out of bag score (false), and minimal cost complexity pruning (0.0, 0.1, 0.2).Finally, the optimal hyperparameters in the random forest modeling were found using the training data in the grid search method.The bootstrap algorithm determines if each decision tree is constructed using bootstrap samples, which are sampled with replacement.Whether or not to use out-of-bag samples for scoring depends on the out-of-bag score.Samples that are not utilized to train a specific decision tree are known as out-of-bag samples.

Support Vector Machine (SVM) Regression Modeling
The SVM regression method was used in this study because of its ability to find complicated or non-linear relationships between the variables.Choosing an appropriate kernel, training the model by maximizing an objective function, and adjusting hyperparameters to obtain the best possible performance in continuous result prediction are the steps involved in SVM regression [40].To predict the thermal data, this method created an optimized hyperplane.Then, a marginal hyperplane was introduced on both sides of the hyperplane with equal distance ε as shown in Figure 4. To deal with the overfi ing problem, we have used 30 datapoints for individual samples and 20 datapoints for all replicates to predict the thermal data outside the marginal plane boundary.The objective function  of the SVM model was calculated by the following Equation (3).
In the above equation, w represents the slope of the hyperplane,  is the number of datapoints outside the marginal planes, and ξ represents their distance from the marginal plane.Various hyperparameters were adjusted to optimize the final model, such as the minimum samples split (2,3,4), minimum samples leaf (1, 2, 3), minimum weight fraction leaf (0.0, 0.1, 0.2), maximum features (none, sqrt, log2), random state (none), maximum leaf nodes (none, 10, 20, 30), minimum impurity decrease (0.0, 0.1, 0.2), bootstrap (true), out of bag score (false), and minimal cost complexity pruning (0.0, 0.1, 0.2).Finally, the optimal hyperparameters in the random forest modeling were found using the training data in the grid search method.The bootstrap algorithm determines if each decision tree is constructed using bootstrap samples, which are sampled with replacement.Whether or not to use out-of-bag samples for scoring depends on the out-of-bag score.Samples that are not utilized to train a specific decision tree are known as out-of-bag samples.

Support Vector Machine (SVM) Regression Modeling
The SVM regression method was used in this study because of its ability to find complicated or non-linear relationships between the variables.Choosing an appropriate kernel, training the model by maximizing an objective function, and adjusting hyperparameters to obtain the best possible performance in continuous result prediction are the steps involved in SVM regression [40].To predict the thermal data, this method created an optimized hyperplane.Then, a marginal hyperplane was introduced on both sides of the hyperplane with equal distance ε as shown in Figure 4. To deal with the overfitting problem, we have used 30 datapoints for individual samples and 20 datapoints for all replicates to predict the thermal data outside the marginal plane boundary.The objective function O f of the SVM model was calculated by the following Equation (3).
In the above equation, w represents the slope of the hyperplane, C i is the number of datapoints outside the marginal planes, and ξ represents their distance from the marginal plane.
To improve the final model, several hyperparameters were adjusted, such as the kernel (linear, polynomial, radial basis function, sigmoid) and epsilon (0.001, 0.01, 0.1, 1).To address issues with non-linearity in the thermal performance data in CMCs, SVM uses a kernel function to map sample data into a high-dimensional space [41].The kind of decision boundary that the SVM uses is decided by the kernel.
'Linear' decision boundaries are used; 'polynomial' kernels are used; 'radial basis function (RBF)' kernels are used; and 'sigmoid' functions are used.The SVM algorithm's optimization margin of tolerance is determined by epsilon.It establishes a range in which misclassification carries no consequences.To improve the final model, several hyperparameters were adjusted, such as the kernel (linear, polynomial, radial basis function, sigmoid) and epsilon (0.001, 0.01, 0.1, 1).To address issues with non-linearity in the thermal performance data in CMCs, SVM uses a kernel function to map sample data into a high-dimensional space [41].The kind of decision boundary that the SVM uses is decided by the kernel.
'Linear' decision boundaries are used; 'polynomial' kernels are used; 'radial basis function (RBF)' kernels are used; and 'sigmoid' functions are used.The SVM algorithm's optimization margin of tolerance is determined by epsilon.It establishes a range in which misclassification carries no consequences.

Gradient Boost Regression (GBR) Modeling
A series of weak learners (decision trees) are constructed in an ensemble using gradient boosted regression, with each model's goal being to capture the residuals of the preceding models [42].Firstly, the average ablation performance data of ceramic matrix composites was predicted through a base model.Then, the residuals or errors of the base model were used as an input in the decision tree (DT) model.Finally, the predicted errors were minimized through five serially connected DT models to obtain the finally predicted temperature data as described in Figure 5.The finally predicted temperature ( ) can be predicted through the following Equation (4).In this equation, ℎ () represents the predicted temperature of the base model,  represents the learning rate, and ℎ () is the prediction output of the residuals from the DT models.

Gradient Boost Regression (GBR) Modeling
A series of weak learners (decision trees) are constructed in an ensemble using gradient boosted regression, with each model's goal being to capture the residuals of the preceding models [42].Firstly, the average ablation performance data of ceramic matrix composites was predicted through a base model.Then, the residuals or errors of the base model were used as an input in the decision tree (DT) model.Finally, the predicted errors were minimized through five serially connected DT models to obtain the finally predicted temperature data as described in Figure 5.The finally predicted temperature (T p can be predicted through the following Equation ( 4).In this equation, h 0 (x) represents the predicted temperature of the base model, α represents the learning rate, and h i (x) is the prediction output of the residuals from the DT models.
J. Compos.Sci.2024, 8, x FOR PEER REVIEW 8 of 23 learning rate regulates the contribution of every tree to the final prediction.Although more trees are needed for modeling with smaller values, generalization can be enhanced.

Extreme Gradient Boost Regression (XGBoost) Modeling
To prevent overfi ing and improve efficiency, the XGBoost method adds a more regularized gradient tree boosting technique than GBR [43].To develop the XGBoost model, similarity weight and gain were calculated in the leaf node to predict the residuals from To identify the GBR modeling's optimum hyperparameters, the grid search approach was applied.The final model was optimized by adjusting a number of hyperparameters, such as the learning rate (0.001, 0.01, 0.1, 0.2, 0.3, 1), minimum samples split (2,3,4), minimum samples leaf (1, 2, 3), minimum weight fraction leaf (0.0, 0.1, 0.2), maximum features (none, sqrt, log2), random state (none), maximum leaf nodes (none, 10, 20, 30), minimum impurity decrease (0.0, 0.1, 0.2), and minimal cost complexity pruning (0.0, 0.1, 0.2).The learning rate regulates the contribution of every tree to the final prediction.Although more trees are needed for modeling with smaller values, generalization can be enhanced.

Extreme Gradient Boost Regression (XGBoost) Modeling
To prevent overfitting and improve efficiency, the XGBoost method adds a more regularized gradient tree boosting technique than GBR [43].To develop the XGBoost model, similarity weight and gain were calculated in the leaf node to predict the residuals from the first tree.Then, residuals or errors were minimized after predicting the temperature through the series of decision trees as depicted in Figure 6.In this study, five DTs were used with the depth of 3 in each tree.In the grid search method, learning rate (0.001, 0.01, 0.1, 0.2, 0.3, 1) was introduced to overcome the overfitting problem.

Extreme Gradient Boost Regression (XGBoost) Modeling
To prevent overfi ing and improve efficiency, the XGBoost method adds a more regularized gradient tree boosting technique than GBR [43].To develop the XGBoost model, similarity weight and gain were calculated in the leaf node to predict the residuals from the first tree.Then, residuals or errors were minimized after predicting the temperature through the series of decision trees as depicted in Figure 6.In this study, five DTs were used with the depth of 3 in each tree.In the grid search method, learning rate (0.001, 0.01, 0.1, 0.2, 0.3, 1) was introduced to overcome the overfi ing problem.

Adaptive Boosting Regression Modeling
Adaptive boosting (AdaBoost) uses decision trees to train a sequence of weak learners.A weak learner is trained with all datapoints having identical weights at first.Weights are modified in the following cycles in accordance with prediction accuracy.Erroneously predicted points are assigned a higher weight, which directs the subsequent learner towards more difficult cases as shown in Figure 7.This procedure is repeated, and a weighted vote is used to aggregate the guesses made by each learner to create the final prediction [44].During iterations, AdaBoost prioritizes hard-to-predict cases for enhancing the overall performance of the model.In this study, five DTs were used with the depth of 3 in each tree.In the grid search method, we used linear, square, and exponential loss in the training dataset and learning rate (0.001, 0.01, 0.1, 0.2, 0.3, 1).The absolute disparities between expected and actual values are minimized using linear loss.By minimizing the squared discrepancies between the actual and predicted values, square loss highlights

Adaptive Boosting Regression Modeling
Adaptive boosting (AdaBoost) uses decision trees to train a sequence of weak learners.A weak learner is trained with all datapoints having identical weights at first.Weights are modified in the following cycles in accordance with prediction accuracy.Erroneously predicted points are assigned a higher weight, which directs the subsequent learner towards more difficult cases as shown in Figure 7.This procedure is repeated, and a weighted vote is used to aggregate the guesses made by each learner to create the final prediction [44].During iterations, AdaBoost prioritizes hard-to-predict cases for enhancing the overall performance of the model.In this study, five DTs were used with the depth of 3 in each tree.In the grid search method, we used linear, square, and exponential loss in the training dataset and learning rate (0.001, 0.01, 0.1, 0.2, 0.3, 1).The absolute disparities between expected and actual values are minimized using linear loss.By minimizing the squared discrepancies between the actual and predicted values, square loss highlights bigger errors.Exponential loss prioritizes mistake reduction by assigning greater weight to misclassified instances.

Performance Accuracy Measuring Parameters
In this study, four evaluation metrics including coefficient of determination (R 2 ), mean absolute error (MAE), mean absolute percentage error (MAPE), and root mean square error (RMSE) are analyzed for the test dataset to assess the performance of the machine learning regression models.They are defined as: e j − p j e j × 100 (7) where K is the number of datasets, e j and p j stand for the j th experimental and predicted burning through temperature data of the ceramic matrix composites, respectively, and e is the average actual temperature data.
mpos.Sci.2024, 8, x FOR PEER REVIEW 9 o bigger errors.Exponential loss prioritizes mistake reduction by assigning greater weig to misclassified instances.

Performance Accuracy Measuring Parameters
In this study, four evaluation metrics including coefficient of determination (R mean absolute error (MAE), mean absolute percentage error (MAPE), and root me square error (RMSE) are analyzed for the test dataset to assess the performance of machine learning regression models.They are defined as: where K is the number of datasets,  and  stand for the  experimental and p dicted burning through temperature data of the ceramic matrix composites, respective and  is the average actual temperature data.

Results
K-fold cross-validation was used in this study to evaluate the model's performance and generalizability.First, 50 subsets (folds) of the 50-90% training dataset were created, and then every machine learning model was trained and assessed based on training data size percentage.Then, the optimized training data size was selected for every machine learning model to achieve the highest prediction accuracy.In the cross-validation process, one of the folds was utilized as the validation set and the other folds were used for training in each iteration.Every fold was utilized as the validation set exactly once during this operation.After optimizing the dataset partitioning through different training data size percentages, testing data were used to test all the prediction models to assess the decision tree, random forest, support vector machine, gradient boosting, extreme gradient boosting, and AdaBoost regression algorithms' accuracy.

Decision Tree Modeling Results
The optimized best hyperparameters of the decision tree algorithms were found using the grid search method as shown in Table 1.The prediction accuracy measurements after utilizing the cross-validation process are listed in Table 2.The high potential of this model is demonstrated by the fact that the mean absolute percentage error is always less than 8%, the R2 scores are higher than 94% for each replicate, and the root mean square error is less than 48 for individual replicates.When the MAPE is continuously less than 5%, it means that the model is accurate since, on average, the predictions are close to the actual values.A high degree of goodness of fit is indicated by R2 values that often exceed 94%, demonstrating the model's capacity to capture and explain the variation in the torch test data.When the RMSE is continuously less than 48, it indicates that the model's average prediction mistakes are small.
The expected versus actual ablation behavior utilizing 30% of test datasets for BN4, BN5, and BN6 samples is depicted in Figure 8.The results are shown in Figure 8, which indicates that the final decision tree has a maximum depth of 3 which resulted in a configuration with eight leaf nodes.Every leaf node predicted the average constant burning through temperature of the ceramic matrix composites.The greatest average temperature that was forecasted was 1066.399°C, while the lowest average temperature that was predicted was 68.4 °C for the BN4 sample using 70% training and 30% test data.The results are shown in Figure 8, which indicates that the final decision tree has a maximum depth of 3 which resulted in a configuration with eight leaf nodes.Every leaf node predicted the average constant burning through temperature of the ceramic matrix composites.The greatest average temperature that was forecasted was 1066.399• C, while the lowest average temperature that was predicted was 68.4 • C for the BN4 sample using 70% training and 30% test data.

Random Forest Modeling Results
The grid search approach was used to find the random forest algorithm's optimal hyperparameters, as Table 3 illustrates.According to Table 4, the mean absolute percentage error of the random forest model is 1% lower than that of decision tree modeling, and the root mean square error is less than 46 for individual replicates, indicating the model's outstanding prospects.Figure 9 shows the expected versus actual thermal performance curve for the BN4, BN5, and BN6 samples using 30% test datasets.According to Table 4, the mean absolute percentage error of the random forest model is 1% lower than that of decision tree modeling, and the root mean square error is less than 46 for individual replicates, indicating the model's outstanding prospects.Figure 9 shows the expected versus actual thermal performance curve for the BN4, BN5, and BN6 samples using 30% test datasets.The findings are displayed in Figure 9, which shows that there are five decision trees in the final random forest (RF) model, each having a maximum depth of three.The final RF model configuration contained forty leaf nodes in total from five trees.Forty leaf nodes corresponded to forty expected constant thermal values of composites.For the BN4 sample, which consisted of 70% training and 30% test data, the highest average temperature predicted was 1067.001• C, while the lowest average temperature predicted was 58.555 • C.

Support Vector Machine Modeling Results
The grid search approach was used to find the support vector machine (SVM) regression modeling's optimal hyperparameters.The SVM model's poor performance is demonstrated by consistently high mean absolute percentage errors, low R2 values, and high root mean square errors in Figure 10.Cross-validation was used to evaluate the model's performance.The epsilon (ε) value was set to 1 for the BN4 test sample (20%) and the BN5 and BN6 test samples (10%).To obtain better prediction results, however, a slightly lower epsilon value of 0.1 was used for all replicate models.Figure 10 shows the expected versus real ablation performance using 30% test datasets for the BN4, BN5, and BN6 samples.
The radial basis function (RBF), which has been found to be the optimal kernel in this work, is used to calculate the ceramic matrix composites' ablation performance for each sample.But this SVM model performs worse than other machine learning models because of the limitations of marginal plane distance from the hyperplane as depicted in Table 5.
strated by consistently high mean absolute percentage errors, low R2 values, and high root mean square errors in Figure 10.Cross-validation was used to evaluate the model's performance.The epsilon (ε) value was set to 1 for the BN4 test sample (20%) and the BN5 and BN6 test samples (10%).To obtain be er prediction results, however, a slightly lower epsilon value of 0.1 was used for all replicate models.Figure 10 shows the expected versus real ablation performance using 30% test datasets for the BN4, BN5, and BN6 samples.The radial basis function (RBF), which has been found to be the optimal kernel in this work, is used to calculate the ceramic matrix composites' ablation performance for each

Gradient Boost Modeling Results
Every decision tree was designed in the gradient boost regression (GBR) model with maximum depth 3 to address the overfitting issue.For BN4, the hyperparameters hold true when 50% of the test data are used and 10% of the test data for BN5 and BN6.With a minimum of two sample splits and one leaf, the learning rate is set to one.Consistently set to 0.0, 'none', 'none', 'none', 0.0, and 0.0, respectively, are the minimum weight fraction leaf, maximum features, random state, maximum leaf nodes, minimum impurity decrease, and cost complexity pruning.The design highlights an optimized set of hyperparameters that are customized for specific test data subsets after the cross-validation process, guaranteeing excellent model performance on a variety of datasets used in the investigation.
Table 6 presents the prediction accuracy measurements.The last decision tree's eight leaf nodes in the GBR model were able to estimate the final residuals or errors in the ceramic matrix composites' burning through temperature.After deducting the final residual value from the base model forecast, the final temperature of the CMC was estimated.The gradient boosting regression (GBR) model's outstanding performance is demonstrated by consistently low mean absolute percentage errors (never more than 2%), very high R2 values (never less than 99%), and root mean square errors (never more than 16) for each individual replicate.Collectively, these indications show how accurately, precisely, and potently the GBR model functions as an explanatory tool and how well it captures and explains the variance in the data.
Figure 11 compares the expected and actual ablation behavior for the BN4, BN5, and BN6 samples using 30% test datasets.As we have shown in our study, a rigorous cross-validation procedure has been used to tune the hyperparameters of the GBR model, which include the maximum depth of every decision tree.Through this procedure, the model is guaranteed to adjust to the subtle differences between various materials, offering a predictive capacity that is material specific.Capturing non-linear interactions in the data is an area where the GBR model excels.It is common for materials to show complex thermal reactions in the event of material-specific ablation, which may not follow straightforward linear pa erns.These nonlinear interactions were navigated and captured by the GBR model, which enables us to identify intricate linkages between ablation behavior and material characteristics.
Though it is a machine learning approach, the GBR model's capacity to identify intricate pa erns and correlations in the data allows it to indirectly represent the system's physics.Together, the ensemble's decision trees provide predictions that help the final As we have shown in our study, a rigorous cross-validation procedure has been used to tune the hyperparameters of the GBR model, which include the maximum depth of every decision tree.Through this procedure, the model is guaranteed to adjust to the subtle differences between various materials, offering a predictive capacity that is material specific.Capturing non-linear interactions in the data is an area where the GBR model excels.It is common for materials to show complex thermal reactions in the event of material-specific ablation, which may not follow straightforward linear patterns.These non-linear interactions were navigated and captured by the GBR model, which enables us to identify intricate linkages between ablation behavior and material characteristics.
Though it is a machine learning approach, the GBR model's capacity to identify intricate patterns and correlations in the data allows it to indirectly represent the system's physics.Together, the ensemble's decision trees provide predictions that help the final model output by learning from the data.The GBR model is highly proficient in detecting and exploiting complex thermal patterns that are not readily apparent using conventional analytical techniques when it comes to non-linear time-temperature behavior in CMCs.Various thermodynamic and heat transport processes that control the thermal response of CMCs are reflected in the model's predictions, even though the model itself does not include explicit physical equations.

Extreme Gradient Boost Modeling Results
The extreme gradient boosting (XGBoost) model consistently shows a mean absolute percentage error (MAPE) below 2%, demonstrating a high degree of accuracy since its predictions are, on average, quite near to the actual values.When the R2 values are near to 1, it means that the model fits the data well and has a remarkable capacity to explain variation.Furthermore, each replicate's mean absolute error is reliably less than 10, indicating accurate predictions, and the root mean square error never exceeds 16, highlighting the model's overall accuracy and dependability in capturing the diversity of the ablation performance data.Based on 30% test datasets, Figure 12 shows the difference between the actual and predicted ablation behavior for the BN4, BN5, and BN6 samples.Most notably, a learning rate of one was established by the XGBoost modeling procedure.Prediction accuracy metrics are summarized in Table 7, which also shows the model's performance with training data sizes ranging from 50% to 90%.

Extreme Gradient Boost Modeling Results
The extreme gradient boosting (XGBoost) model consistently shows a mean absolute percentage error (MAPE) below 2%, demonstrating a high degree of accuracy since its predictions are, on average, quite near to the actual values.When the R2 values are near to 1, it means that the model fits the data well and has a remarkable capacity to explain variation.Furthermore, each replicate's mean absolute error is reliably less than 10, indicating accurate predictions, and the root mean square error never exceeds 16, highlighting the model's overall accuracy and dependability in capturing the diversity of the ablation performance data.Based on 30% test datasets, Figure 12 shows the difference between the actual and predicted ablation behavior for the BN4, BN5, and BN6 samples.Most notably, a learning rate of one was established by the XGBoost modeling procedure.Prediction accuracy metrics are summarized in Table 7, which also shows the model's performance with training data sizes ranging from 50% to 90%.The adaptive boosting (AdaBoost) model consistently keeps the root mean square error for individual replicates below 37, attains R2 values greater than 98%, and maintains a mean absolute percentage error below 4%.Using 30% test datasets, Figure 13 compares the actual and expected ablation behavior for the BN4, BN5, and BN6 samples.The AdaBoost process used square loss for the time-temperature data and established a learning rate of one in most prediction models.The prediction accuracy values are presented in Table 8.
It was found from our study that all the performance parameters including RMSE, MAE, MAPE, and R2 score in the combined prediction ML models using all the replicates performed more poorly than the single-sample prediction models.The main reason for poor performance is the difficulty faced by all the above ML models to find the non-linearity in the torch test temperature data in three replicates at the CMC's burning through time.The RF model took less simulation time than DT models for predicting the thermal performance of composites with total samples of data divided into five trees from one tree.It was also noticeable that XGBoost and AdaBoost regression models took less simulation time than other techniques for temperature prediction.The adaptive boosting (AdaBoost) model consistently keeps the root mean square error for individual replicates below 37, a ains R2 values greater than 98%, and maintains a mean absolute percentage error below 4%.Using 30% test datasets, Figure 13 compares the actual and expected ablation behavior for the BN4, BN5, and BN6 samples.The Ada-Boost process used square loss for the time-temperature data and established a learning rate of one in most prediction models.The prediction accuracy values are presented in Table 8.

Comparing the Performance of ML Models
To the best of our knowledge, this work is the first attempt to forecast ceramic matrix composites' ablation performance data.The coefficient of determination, often known as the R-squared (R2) score, was used as a performance indicator to assess the effectiveness of the regression models analyzed for this study.For each model, the R2 score-which represents the proportion of variance in temperature data explained across time-was estimated.Notably, the R2 scores of the GBR and XGBoost models were nearly 1, indicating an incredible explanation of the consistency between the observed and anticipated torch test thermal data.All the machine learning models developed in this study have an R2 score higher than 0.85, indicating that all the models can accurately predict the torch test data as shown in Figure 14a.The R2 score was considered together with other performance indicators, such as mean absolute error (MAE), mean absolute percentage error (MAPE), and root mean square error, for a thorough evaluation of the model's performance.According to our research, boosting (XGBoost, AdaBoost, and GBR) regression approaches had a mean absolute percentage error that was higher than that of bagging (random forest), SVM, and decision tree regression techniques.The algorithm's ability to sequentially learn from the mistakes of prior learners after first learning from weak learners is responsible for the significant gain in performance observed in boosting ensemble techniques.The average percentage difference between the actual and anticipated temperature data values highlights the fact that GBR and XGBoost were the most accurate in estimating the ablation performance of ceramic matrix composites during the burning phase as described in Figure 14b.On the other hand, the smaller distance between the marginal plane lines and the hyperplane is responsible for the SVM regression model's poorer performance.This results in incorrect data classification and a rise in the mean absolute per- According to our research, boosting (XGBoost, AdaBoost, and GBR) regression approaches had a mean absolute percentage error that was higher than that of bagging (random forest), SVM, and decision tree regression techniques.The algorithm's ability to sequentially learn from the mistakes of prior learners after first learning from weak learners is responsible for the significant gain in performance observed in boosting ensemble techniques.The average percentage difference between the actual and anticipated temperature data values highlights the fact that GBR and XGBoost were the most accurate in estimating the ablation performance of ceramic matrix composites during the burning phase as described in Figure 14b.On the other hand, the smaller distance between the marginal plane lines and the hyperplane is responsible for the SVM regression model's poorer performance.This results in incorrect data classification and a rise in the mean absolute percentage error when predicting the burning through temperature data of the CMCs.
The root mean squared error (RMSE) is used to evaluate the predictive machine learning models' accuracy regarding the thermal behavior of ceramic matrix composites.This measurement is equivalent to taking the square root of the average of the squared discrepancies between the actual and expected temperatures obtained from the torch test.The GBR and XGBoost models beat other machine learning regression models when the RMSE value is less than 19 as shown in Figure 15a.The SVM regression model's reduced performance can be ascribed to a significant quantity of incorrectly categorized datapoints situated beyond the marginal plane lines.Because various datasets have varying levels of complexity, random forest often has more diverse hyperparameters.The best spli er and fewer hyperparameters are often found in simpler decision trees topologies.AdaBoost trains on time-temperature data using square loss and a learning rate of 1, which is a less complex hyperparameter than that of the decision tree and random forest.The adjustment of hyperparameters for superior model performance, particular to a certain material, is highlighted in this comparison.To ensure flexibility and precision in forecasting the ablation behavior of ceramic matrix composites, each model is customized to fit a specific test data subset.
Across a range of test data splits, the XGBoost and AdaBoost models consistently provide the fastest execution speeds.The execution durations of gradient boosting are modest; it takes more than XGBoost and AdaBoost, but less than SVM and decision tree.In general, support vector machines take longer to execute than ensemble models.Predictive performance and computational efficiency may have a role in model selection, particularly in the context of big datasets or real-time applications.In this situation, XGBoost and AdaBoost perform exceptionally well.Because various datasets have varying levels of complexity, random forest often has more diverse hyperparameters.The best splitter and fewer hyperparameters are often found in simpler decision trees topologies.AdaBoost trains on time-temperature data using square loss and a learning rate of 1, which is a less complex hyperparameter than that of the decision tree and random forest.The adjustment of hyperparameters for superior model performance, particular to a certain material, is highlighted in this comparison.To ensure flexibility and precision in forecasting the ablation behavior of ceramic matrix composites, each model is customized to fit a specific test data subset.
Across a range of test data splits, the XGBoost and AdaBoost models consistently provide the fastest execution speeds.The execution durations of gradient boosting are modest; it takes more than XGBoost and AdaBoost, but less than SVM and decision tree.In general, support vector machines take longer to execute than ensemble models.Predictive performance and computational efficiency may have a role in model selection, particularly in the context of big datasets or real-time applications.In this situation, XGBoost and AdaBoost perform exceptionally well.
Although the prediction accuracy of the regression models has been our main emphasis, we recognize the significance of taking computing efficiency into account, particularly in real-world applications.Especially for big datasets or applications where real-time predictions are crucial, computational efficiency is a key component.The XGBoost and GBR models often showed good accuracy in our investigation, but it is vital to remember that these models' computational efficiency might be affected by variables like the quantity and complexity of the dataset.
The XGBoost technique performed efficiently in our study because of its capacity for parallel and distributed computing.Column block encoding and tree pruning are two of its optimization strategies that are responsible for its scalability.However, GBR used sequential boosting which could gradually construct a more intricate and expressive model because of its sequential structure.GBR identified intricate, non-linear patterns in the torch test data since each weak learner adds a component to the overall model.The process of boosting made sure that the model assigned greater weight to hard-topredict cases, which enhanced its capacity to predict complex interactions in the ablation performance data.Throughout the training phase, GBR allows an adaptive modification of the model's complexity.Hyperparameters such as the number of estimators and the learning rate regulate the inclusion of weak learners.The flexibility of the model guarantees its ability to strike the best possible balance between preventing overfitting and capturing the underlying non-linear patterns.

Conclusions
This is the first time in the literature that machine learning approaches have been utilized to predict the thermal performance of boron-nitride-nanoparticle-containing continuous-fiber-reinforced silicon oxy-carbide ceramic matrix composites in the oxyacetylene torch test.Our study revealed that 70% torch test training data showed better performance accuracy in developing the XGBoost model and 80% torch test training data showed better performance accuracy in developing the GBR model to predict the ablation performance of ceramic matrix composites.The primary shortcoming of the machine learning model developed in this study is its inability to explain the fundamental principles of physics governing the ablation behavior of ceramic matrix composites.This limitation can be addressed by creating machine learning techniques that take physics into account.Future researchers could potentially use the models developed in this study for optimizing the design parameters (fiber volume fraction, resins, boron nitride nanoparticles, fibers, etc.) of continuous-fiber-reinforced silicon oxy-carbide ceramic matrix composites.Future researchers also can use the machine learning models developed in this study for predicting the lifecycle of other thermally resistive composite materials.This research marks the initiation of an effective approach for predicting thermal properties by utilizing oxy-acetylene torch test data, underscoring the significance of machine learning in forecasting the operational longevity of ceramic matrix composites in challenging applications like aeroengines and gas turbines.

Figure 9 .Figure 9 .
Figure 9. Random Forest Regression results for (a) BN4, (b) BN5, and (c) BN6 samples.The findings are displayed in Figure9, which shows that there are five decision trees in the final random forest (RF) model, each having a maximum depth of three.The final RF model configuration contained forty leaf nodes in total from five trees.Forty leaf nodes corresponded to forty expected constant thermal values of composites.For the BN4

Figure 14 .
Figure 14.Performance (a) R 2 score and (b) MAPE results of individual replicates with test data percentage.

Figure 14 .
Figure 14.Performance (a) R 2 score and (b) MAPE results of individual replicates with test data percentage.

Figure 15 .
Figure 15.Performance (a) RMSE and (b) MAE results of individual replicates with test data percentage.The mean absolute error (MAE) is used to assess the temperature forecasting model's accuracy over the course of ceramic matrix composite (CMC) burning.In the torch test, this measure shows the average absolute difference between the experimental and projected temperature values.Lower MAE values indicate be er model performance in comparison to other machine learning models for each sample, especially in the XGBoost and GBR models.Because various datasets have varying levels of complexity, random forest often has more diverse hyperparameters.The best spli er and fewer hyperparameters are often found in simpler decision trees topologies.AdaBoost trains on time-temperature data using square loss and a learning rate of 1, which is a less complex hyperparameter than that of the decision tree and random forest.The adjustment of hyperparameters for superior model performance, particular to a certain material, is highlighted in this comparison.To ensure flexibility and precision in forecasting the ablation behavior of ceramic matrix composites, each model is customized to fit a specific test data subset.Across a range of test data splits, the XGBoost and AdaBoost models consistently provide the fastest execution speeds.The execution durations of gradient boosting are modest; it takes more than XGBoost and AdaBoost, but less than SVM and decision tree.In general, support vector machines take longer to execute than ensemble models.Predictive performance and computational efficiency may have a role in model selection, particularly in the context of big datasets or real-time applications.In this situation, XGBoost and AdaBoost perform exceptionally well.

Figure 15 .
Figure 15.Performance (a) RMSE and (b) MAE results of individual replicates with test data percentage.The mean absolute error (MAE) is used to assess the temperature forecasting model's accuracy over the course of ceramic matrix composite (CMC) burning.In the torch test, this measure shows the average absolute difference between the experimental and projected temperature values.Lower MAE values indicate better model performance in comparison to other machine learning models for each sample, especially in the XGBoost and GBR models.Because various datasets have varying levels of complexity, random forest often has more diverse hyperparameters.The best splitter and fewer hyperparameters are often found in simpler decision trees topologies.AdaBoost trains on time-temperature data using square loss and a learning rate of 1, which is a less complex hyperparameter than that of the decision tree and random forest.The adjustment of hyperparameters for superior model performance, particular to a certain material, is highlighted in this comparison.To ensure flexibility and precision in forecasting the ablation behavior of ceramic matrix composites, each model is customized to fit a specific test data subset.Across a range of test data splits, the XGBoost and AdaBoost models consistently provide the fastest execution speeds.The execution durations of gradient boosting are modest; it takes more than XGBoost and AdaBoost, but less than SVM and decision tree.In 1, 0.2) were tuned to optimize the final model.During tree development, the spli er decides how to choose the split at each node.Whereas 'random' selects the best random split, 'best' selects the best split.The minimum number of samples needed

Table 1 .
Optimized hyperparameters in decision tree regression models of BN4, BN5, BN6 and all the replicates.

Table 2 .
Ablation performance results of boron nitride ceramic matrix composites in Decision Tree model.

Table 4 .
Ablation performance results of boron nitride ceramic matrix composites in Random Forest model.

Table 5 .
Ablation performance results of boron nitride ceramic matrix composites in Support Vector Machine model.

Table 6 .
Ablation performance results of boron nitride ceramic matrix composites in Gradient Boosting model.

Table 6 .
Ablation performance results of boron nitride ceramic matrix composites in Gradient Boosting model.

Table 7 .
Ablation performance results of boron nitride ceramic matrix composites in Extreme Gradient Boosting model.

Table 7 .
Ablation performance results of boron nitride ceramic matrix composites in Extreme Gradient Boosting model.

Table 8 .
Ablation performance results of boron nitride ceramic matrix composites in Adaptive Boosting model.