Next Article in Journal
BIM Application in Multi-Criteria Assessment of Suspended Brickwork Façade Production Methods for a Reconstructed Building
Next Article in Special Issue
Research on Multi-Sensor Simultaneous Localization and Mapping Technology for Complex Environment of Construction Machinery
Previous Article in Journal
Teeth Segmentation in Panoramic Dental X-ray Using Mask Regional Convolutional Neural Network
Previous Article in Special Issue
Intelligent Machinery Fault Diagnosis Method Based on Adaptive Deep Convolutional Neural Network: Using Dental Milling Cutter Malfunction Classifications as an Example
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Predicting and Evaluating Decoring Behavior of Inorganically Bound Sand Cores, Using XGBoost and Artificial Neural Networks

1
Fraunhofer Research Institution for Casting, Composite and Processing Technology IGCV, Lichtenbergstrasse 15, 85748 Garching, Germany
2
Chair of Metal Forming and Casting, Technical University of Munich, Walther-Meissner-Strasse 4, 85748 Garching, Germany
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(13), 7948; https://doi.org/10.3390/app13137948
Submission received: 26 May 2023 / Revised: 28 June 2023 / Accepted: 5 July 2023 / Published: 6 July 2023
(This article belongs to the Special Issue AI Applications in the Industrial Technologies)

Abstract

:
Complex casting parts rely on sand cores that are both high-strength and can be easily decored after casting. Previous works have shown the need to understand the influences on the decoring behavior of inorganically bound sand cores. This work uses black box and explainable machine learning methods to determine the significant influences on the decoring behavior of inorganically bound sand cores based on experimental data. The methods comprise artificial neural networks (ANN), extreme gradient boosting (XGBoost), and SHapley Additive exPlanations (SHAP). The work formulates five hypotheses, for which the available data were split and preprocessed accordingly. The hypotheses were evaluated by comparing the model scores of the various sub-datasets and the overall model performance. One sand-binder system was chosen as a validation system, which was not included in the training. Robust models were successfully trained to predict the decoring behavior for the given sand-binder systems of the test system but only partially for the validation system. Conclusions on which parameters are the main influences on the model behavior were drawn and compared to phenomenological–heuristical models of previous works.

1. Introduction

1.1. Motivation and Context

Complex cast parts require lost cores to form undercuts and cavities. Filigree cores need high strength to withstand the casting process and make them dimensionally stable. However, they have to be removable after casting. Inorganically bound sand cores have environmental advantages compared to organically bound sand cores [1]. However, unlike organically bound cores, inorganically bound cores cannot be completely rinsed out or burned out. They need to be broken up by a mechanical impulse on the surrounding cast part and thus are significantly harder to decore than organically bound sand cores [1]. Additionally, if the cores are too strong, the forces required to break them up deform the cast part. These issues can be addressed by changing the sand-binder systems’ chemical composition and manufacturing process. The acting influences on the decoring properties of inorganically bound sand cores must be understood to achieve these changes systematically. Ettemeyer [2] investigated these influences for the decoring behavior, collecting extensive experimental data. These data were used to capture cause–effect relations and to validate aspects of a formulated theory for the decoring behavior. This theory claims that the decoring behavior can be predicted by various key parameters of the sand-binder system of which the core consists and of the decoring process. Ettemeyer used heuristical and phenomenological methods for the prediction of decoring behavior. This work uses purely data-based machine learning models to predict the decoring behavior based on the same experimental data of the mentioned work. Additionally, the importance of various features for the machine learning model prediction of the decoring behavior is evaluated. This evaluation is achieved by varying the data composition for model training and using interpretable machine learning techniques. The importance of the features according to the machine learning models is then compared to the mentioned theory.

1.2. Decoring of Inorganically Bound Sandcores

Casting technology relies on sand cores to achieve complex, hollow structures with undercuts and cavities [3]. The cores are used individually or are assembled into sand-core packages. They are inserted into the mold. They must be held in position by core holders in the surrounding permanent or sand mold. The sand core stands out of these holding positions after casting and forms a connection to the cavity. This connection is used to extract the sand from the formed cast part, which is the decoring process step. The sand-binder system of the sand core determines which methods can be used to decore it. Sand cores are classified into organically and inorganically bound sand cores [1]. Organic sand-binder systems are based on hydrocarbons and can easily be burnt off. Most of the decoring process is achieved during the casting process due to the heat of the melt. It cracks the sand cores, burns parts of the binder, and allows shaking of the remaining sand out of the openings. The downsides are environmental considerations of the toxic waste and fumes produced during the burnup. Inorganic sand-binder systems use sodium-silicate to harden sand cores. It lacks hydrocarbons and, with it, the downsides of toxicity. However, it cannot be burnt off, and additionally, it hardens even more on the surface, which is in contact with the melt [1]. The decoring of inorganically bound sand cores is their main disadvantage. There are multiple approaches to change the properties, for example, adding additives or changing the ratio of binder to sand [4].
At the same time, research focuses on understanding the decoring behavior of inorganically bound sand cores [5]. Ettemeyer [2] describes the essential parameters to predict the decoring behavior of a sand-binder system and builds a matching theory. According to this theory, the flexural and compressive strength of the sand-binder mixture are the main parameters. Another factor for predicting the decoring behavior is the hydrostatic pressure on the sand core [6]. The cast part shrinks onto the sand core after casting, creating pressure on the sand core. This pressure influences the force needed to fracture the sand core. Starting from an initial fracture area, the decoring proceeds. The Drucker–Prager failure model describes this theory, as shown in Figure 1.
The failure straight is determined from a three-point bending test and a uniaxial compression test. The cohesion and the internal friction angle phi describe the experimentally calculated failure straight. If the von Mises equivalent stress for a prevailing hydrostatic pressure is above the failure straight, a fracture occurs in the sand core. The composition of sand and binder defines the parameters of cohesion and phi. If these parameters of a sand-binder system and the hydrostatic pressure state are known, the force needed to fracture can be predicted. The following bullet points summarize our findings:
  • The assumption of a pressure-dependent material model to describe sand-binder systems is valid. It can be seen as essential for an accurate characterization of the mechanical behavior of decoring.
  • The decoring behavior of inorganically bound sand-binder systems is significantly influenced by the molding material and the binder system used. The molding material dominates the influence on the decoring behavior.
  • Larger binder quantities can lead to higher residual strengths and correspondingly poorer decoring behavior for the same molding material.
  • Assuming constant decoring energy, the angle phi and the cohesion of the remaining residual strength after casting are the two main influencing variables for the decoring behavior.

1.3. Machine Learning Models

This work uses machine learning methods to model cause–effect relationships in the context of decoring inorganically bound sand cores. There are various examples of gaining insights from data in many technology areas, for example, material analysis [7], laser beam welding [8], or injection molding [9]. The basic idea is to train machine learning models using annotated data. The trained machine learning models are then evaluated using statistical approaches or explainable model architectures. This work follows the approach shown in [10,11] to use models such as extreme gradient boosting (XGB) with the evaluation and visualization tool SHapley Additive exPlanations (SHAP) to interpret datasets. Additionally, artificial neural networks (ANN), which are black box models, are trained and evaluated using their achieved model scores. These are evaluated based on their model scores. Both models—ANN and XGB—are trained for all datasets.
ANNs are well-suited for complex datasets. Due to their non-linear calculations, they can abstract non-linear links between input and label. With multiple layers of neurons, they can compute intermediate features from input data, performing a type of model-based feature engineering, an example of which can be found in [12]. Their broad abilities render them a standard tool in machine learning tasks. This advantage and their frequent use in machine learning publications are why they are used in this work as well, making it easier to compare the results to other works.
XGB models generate their predictions from ensembles of decision trees with low depth. Decision trees are transparent, as it is possible to read out each node and its decision. All decisions can be summarized and evaluated, for example, how often a feature is used in the model for a decision in all trees or how much the loss is reduced by these splits [13]. XGB has further advantages that make it a popular choice for machine learning tasks [14], for example, as it is an ensemble method that can be parallelized, thus speeding up calculations on multi-processer units, and can avoid overfitting by using different types of regularization [15].
SHAP allows generating even more information from machine learning models such as XGB. It uses game theory to calculate the average marginal contribution of each feature [16]. The general approach of SHAP is to leave one feature out of the model calculation and to determine the remaining model error, which is repeated for every feature in the dataset used by the investigated model. Lundberg and Lee developed the corresponding algorithm [17]. It allows the generation of various figures. This work uses its bee plot, which shows each feature vertically in descending order of its contribution to the model. All datapoints are depicted individually on a horizontal scale for each feature to illustrate the influence of each feature on each datapoint. This allows for a dense overview of feature importance, the identification of outliers, and an interpretation of the direction the features push at each datapoint.

1.4. Aim of This Work

Models can be categorized depending on the approach and the methods used. Figure 2 depicts a possible categorization of models [18]. Ettemeyer [2] used heuristical and phenomenological approaches to build his theory of important parameters for decoring behavior and predict new sand-binder systems’ decoring behavior. He collected an extensive amount of experimental data during the progress. This work aims to replicate these results with a purely statistical approach using machine learning methods and the collected experimental data. The new approach is evaluated by comparing which parameters are important to predict decoring behavior according to the machine learning models and according to Ettemeyer [2]. This results in three corresponding research questions:
  • Is it possible to train robust machine learning models to predict the decoring behavior of sand-binder systems using the given data?
  • Which features are important for the decoring behavior according to the machine learning models, and are those the same as described in the previous work?
  • Is it possible to extract further, possibly new insights from the data using machine learning models?
All three questions are tested by a forwards–backward approach, as shown in Figure 3. Machine learning models are trained and optimized using the whole experimental data or subsets thereof. Depending on the resulting quality and behavior of the trained models, a conclusion is drawn whether the used subset of data includes the necessary parameters or not. A similar approach was used in [19], where insights into the main parameters of an industrial hydrocyclone were determined.
By optimizing the models, the likeliness increases that the differences in the model quality of the prediction for different feature sets result from the different feature sets and not from randomly better- or worse-fitting models. The second question formulates a general comparison. To be able to answer it, it is transformed into five testable specific hypotheses:
  • Categorical features such as the name of the experimental series contain information beyond the collected data, for example, a different room temperature level during the decoring process or slight but systematical differences in the clamping of the specimen. Including these features in the model training will improve the model quality but reduce the interpretability due to the aggregation of information into a single variable.
  • The acceleration data contain important information for predicting the decoring behavior. Including it in the model training will improve the model quality.
  • Different processing methods of the acceleration data are differently suited for machine learning models.
  • Including all features in the model training will reduce the model quality due to a high data complexity without more information than in a reduced dataset. Reducing the data complexity in multiple steps using a feature selection method will improve the model quality for each dataset. At a given threshold, the model quality will drop significantly and abruptly. All features included at that threshold can be considered significantly important for the decoring behavior.

1.5. Approach and Big Picture

The following approach is used to achieve the described aim and test the hypotheses. The basic idea is to use the behavior and quality of trained models to evaluate the feature sets used. The models will predict the decoring behavior with some degree of deviation. Varying the features selected from all the experimental data to train the model will result in varying deviations in the model prediction. This variation can be used to compare different feature sets for their information content. A feature set yielding a model with a comparatively low deviation can be seen as containing more relevant information than a feature set resulting in a higher error. By using a transparent machine learning model such as XGB, further indications of the importance of features can be generated. Hypotheses of which feature sets are more or less important to decoring behavior are formulated, and datasets are composed accordingly. In all cases, the model quality has to be ensured by optimizing the models and evaluating their robustness. Otherwise, an insufficiently trained or non-optimized model is just as likely as an insufficient feature set.
The specific approach is summarized in Figure 4. First, the target feature has to be defined. The possible alternatives are the measured decored mass after each interval and the measured difference in the decored distance from both open sides of the rectangular tube after each interval. Being easier to understand and easier to measure consistently, the decored mass is defined as the target feature for the model training. Thus, the trained models will predict the decored mass after each interval.
In step two, the ram-impact acceleration curves recorded with the laser vibrometer are converted into scalar features. Three approaches are used for this: Fast Fourier transforms (FFTs), mel frequency cepstral coefficients (mfcc), and a statistical evaluation of the signals in the time domain. All three are typical approaches to translating signal curves into features for machine learning methods. These characteristics are calculated individually for all ram impacts and summarized for the respective interval. Section 3.1 explains the exact calculations.
In step three, the division into datasets takes place. The original dataset with all features is split according to the previously described hypotheses. The main reason for this division of the data into feature sets is that the features in the data overlap in their information content. For example, the sand-binder system’s name indirectly contains the sand’s name, the binder’s name, the binder, the weight content of the binder in the mixture, and their specific properties. It makes evaluating specific features more complicated. The division into sub-datasets can reduce this effect but only negate it partially. The second reason is the number of datapoints. Using all features simultaneously results in a dataset with about as many or more features as datapoints. These kinds of sparse data can lead to more overfitting during training. It can also obstruct the model training if the input vector is overly large compared to the number of datapoints. The detailed data composition is explained in Section 3.2.
Step four is the first round of model calculations. As mentioned before, ANN and XGB models are used. The used programming libraries are listed in Used programming libraries of Appendix A. It is essential to try different model parameter sets and optimize the models for each dataset. Without this step, the achievable model quality may depend significantly on the chosen model parameters. Even with dataset-specific optimization, this is always the case to some degree and must be considered in the interpretation. The likelihood of overfitting is reduced by calculating each parameter set ten times with random initializations. The average of these ten calculations is used as the final model score for a given parameter set. The model quality metrics and their interpretation are explained in Section 2.2. The meta code and further details of the calculations can be found in Section 2.3.
In step five, a data complexity reduction is performed. As mentioned, the ratio of features to datapoints can be a crucial influence on model quality. Reducing the number of features can lead to better-performing models despite losing some of the information. There are multiple ways to reduce data complexity to achieve a high performance boost without losing too much information. One way is to keep all features but reduce the dimensionality by transforming the features into a new set of features, for example, using principle component analysis. Another way is to select features based on calculated metrics, for instance, feature importance values of a decision tree algorithm. This work uses feature selection based on the feature importance values of the XGB models. For each dataset, the XGB model with the lowest root mean square error (RMSE) is defined as the best-fitting XGB model. The feature importance values of this model are used to select the features. The percentage of features kept is called filter fraction (ff). A filter fraction of 100% is equal to all features used to train a model. Twelve filter fractions are used, with 10% steps between 100% and 10%. Smaller steps are used between 10% and 1%. The final number of features is rounded up to the next integer. Twelve datasets and twelve filter fractions result in 144 datasets used to train the models. Step six is to train and optimize ANN and XGB models for the reduced datasets in the same way as in step four.
It is essential to test whether the differences in the results of the datasets are significant. Other works used a two-sided Welch’s t-test to compare different models [20] and different datasets [21]. Therefore, this work applies a two-sided Welch’s t-test with a 0.5 significance level to test whether the results are significant. The test is performed for each dataset. The overall best dataset is used as a comparison set. This work focuses on obtaining insights by comparing model results for different datasets, which is why the Welch’s t-test is performed for the datasets and not explicitly between the ANN and XGBoost models.
Finally, the results of the two rounds of computation are visualized and discussed. The corresponding hypothesis is evaluated based on the model quality that can be achieved in each case. Plots of the model scores for varying filter fractions and selected parameters are discussed to assess the robustness of the models and the likelihood that the model score is a result of the feature selection and not of the model parameters.

2. Materials and Methods

2.1. Data Origin

The work of Ettemeyer [2] delivers the experimental basis of this work and the theory for the comparison. Ettemeyer investigated seventeen sand-binder systems. The sand, the binder, and the binder quantity varied for each system. A sequence of experiments was carried out for each system. Figure 5 depicts the tested specimen. It represents a rectangular bar of an inorganically bound sand core cast in a rectangular aluminum tube. It has two open sides, with some of the sand core extending out of the surrounding aluminum tube. The twelve specimens were carefully cut off from the casting system. The sand core was not damaged significantly, and the shape was preserved in most cases. This work focuses only on those kinds of specimens with undamaged or only slightly damaged specimens.
Figure 6 explains the experimental process. The selected sand was molded into a sand core with the selected binder amount and tempered. A varying share of the cores was tested in this raw state, determining their raw flexural strength and raw compressive strength. These tests were conducted at 25 °C, at 400 °C, at 600 °C, and at 750 °C. After a storage period, twelve sand cores were placed in one circular mold in each casting session, with the metal inflow located in the middle of the mold. The casting temperature of the aluminum melt was 750 °C. The number of casting sessions and, thus, the number of specimens varied for the various sand-binder systems. After an additional storage time, the specimens were cut off from the casting system and were tested. Three kinds of tests were conducted. The first test determined the sand cores’ residual flexural and compressive strength after casting. To make the test possible, some specimens were cut open in a very low-impact way to extract the undamaged sand core. These were destructively tested to measure the strengths after casting. The second test yielded the residual stress configuration of the aluminum tubes using strain gauges and cutting the cast part open near the strain gauges. These two tests were only conducted for some sand-binder systems and are referred to as extended features. The third test was the main decoring test. Figure 7 shows the decoring process. The decoring was achieved by a ram impacting the specimen. A laser vibrometer measured the velocity at a specific point of the specimen during each impact. The resulting velocity curve was translated into an acceleration curve. The ram impacted the specimen until it was fully decored. Figure 8 explains the division of the decoring process in intervals of ram impacts. After each interval, the mass and the decored distance on both open ends of the cast part were measured.
The cohesion and the internal friction angle phi were calculated for all experimentally determined compressive and flexural strengths and form part of the data input for the machine learning models.
In summary, the data collected represents a dataset with over 400 intervals measured and over 100 bars tested. The publication [2] describes the calculation of data, the quality of the collected data, and the considerations underlying its selection in greater detail.

2.2. Used Model Quality Criteria and Their Interpretation

Three different model scores describe the model quality in this work. The mean absolute error (MAE) is easily interpreted since it contains no non-linear transformation. It scores how well a model predicts the target feature on average. Being a linear calculation, it does not punish outliers in the prediction as the average levels them. Though always calculated with scaled values, the MAE is always shown rescaled in its corresponding unit in the Section 3. The root mean squared error (RMSE) is a double non-linear model score. Its unit is the same as the target feature, but the scale shifts due to the non-linear calculation. It loses interpretability, but generally, a low RMSE represents a more robust model than a low MAE since outliners are heavily punished. It is the main score for comparing models in the presented work. In this work, a lower RMSE means a better model. All models described as the best model of a series of calculations have the lowest RMSE for the test data of these models. The third model score is the coefficient of determination (R2). It measures how well a model fits the variation of the data. If the data are predicted perfectly, R2 equals 1. If the resulting error of the prediction is the same as when using the mean of the data, R2 equals 0. If the resulting error is worse than simply using the mean of the data, R2 can become negative. Formulas 1 to 3 show the calculation of the scores, in which y i represents the true value of a datapoint, y ^ i the corresponding predicted value, and n the number of datapoints:
M A E = i n | y i y ^ i | n
R M S E = i n ( y i y ^ i ) 2 n
R 2 = 1 i n ( y i y ^ i ) 2 i n ( y i 1 n i n y i ) 2
In the case of multiple randomly initialized calculation runs, the metrics are mean values over all runs. Therefore, mean MAE and mean RMSE refer to these metrics averaged over all runs. If a metric relates only to a single run, an additional marker is added, for example, best RMSE.
Although the MAE is calculated with scaled parameters, it can be rescaled to its original unit. Particularly meaningful is the relative MAE, which puts the MAE in relation to the mean target value, in this case, the decored mass. Its formal calculation is depicted in Formula 4. The results will show a mean, relative MAE, which is the MAE averaged over all ten randomized runs of the calculation relative to the mean target value:
r e l a t i v e   M A E = M A E 1 n i n y i

2.3. Model Training Sequence

Figure 9 shows the training procedure as a pseudo-code. For both types of models, hyperparameter optimization is performed for each dataset and filter fraction. Each parameter combination is computed with a different random seed ten times to avoid local overfitting. This random seed influences the generation of the train–test splits and the model initialization. The mean values of the ten calculations are used as comparison values to determine the best parameter combination. The results of the best model of each parameter combination are recorded as well.
Following the hold-out method, system D is kept from all model training and serves as a validation set. These validation data provide a second check to see whether overfitting has occurred. System D is chosen as a validation system because it represents a good tradeoff between having enough datapoints and a composition between the other sand-binder systems.
Scaling is always fitted to the training set and applied to all three datasets. Min–max scaling is chosen as the scaling method. Table A4 in Appendix A shows the filter fractions and model parameters used. If multiple values are listed for a parameter, this parameter is varied over these values as part of the hyperparameter optimization. For the XGB models, the parameters eta, alpha, max depth, and parallel trees are varied. For the NN models, the number of layers, the number of neurons per layer, and alpha are varied. This approach is inspired by [22].
In the end, the results of 60,840 models were calculated for this work. This number does not include the number of models trained and evaluated to estimate the fixed hyperparameters.

3. Results

3.1. Generated Experimental Database and Time-Series Processing

In this sub-chapter, the exact features are shown and explained in as much detail as required for the scope of this work. For further details, the work of Ettemeyer [2] can be consulted. Following this work, the data are structured into sand-binder systems, specimen, and decoring intervals. System characteristics were determined for each sand-binder system. Each specimen was decored after a finite number of intervals. Each interval comprises between 10 and 80 ram impacts. It is fundamental for the discussion of the feature importance to understand the relation between the different kinds of features and the composition of the sand-binder systems.
This paper uses 8 of the 17 sand-binder systems investigated in work [2] for model training. Nine are discarded due to an insufficient number of datapoints. Table 1 shows the sand-binder systems used for model training and their compositions. R refers to “reference binder”, and R_low to a binder with a lower modulus. The number of intervals and the number of specimens in each system are listed as well. In the last row, it is marked whether the previously mentioned extended features such as strain or temperature- depended strengths were recorded for this system.
Figure 10 summarizes the data used to train the models. Corresponding to the experimental design described in Section 2.1, the parameters are grouped into six sources. Table A1, Table A2 and Table A3 in Appendix A give a complete overview of all the parameters used in the training process. Additionally, they show their abbreviated names used in the SHAP plots. The target value predicted by the models is the decored mass in a single decoring interval, abbreviated as DM.
The original system features include 3 categorical, 73 scalar quantities, and 1 time series. The categorical quantities describe the name of the sand used, the name of the binder used, and the name of the sand-binder system in the experimental setup. Two different binders and three different sands were used in the data used for model training. By using different binder percentages, eleven systems were combined from these components. These four features describe the composition of a sand-binder system and are therefore aggregated features. They are referred to as “composition features” as the measured physical features directly result from this composition. Of the scalar features, 41 are not used because they represent the standard deviation or the number of measurements for a measurement such as flexural strength. Five scalar features are not used because they represent an evaluation of the target feature. They correlate linearly to the target feature. Six other variables have only been measured for individual sand-binder systems and are discarded. Twenty-one scalar features remain and are used to train the models, seven of which are available for all sand-binder systems that are summarized as “basic features” and another fourteen that are referred to as “extended features.”
This division leads to two basic datasets. The first dataset contains a maximum number of datapoints and thus does not use the extended features. The second dataset uses the maximum number of features with all extended features but not all datapoints. Using both extremes makes it possible to evaluate the extended features without leaving the large number of datapoints available for the basic features unused. Further parameters were measured at the beginning of the interval in terms of mass distribution and decoring progress. They describe the initial state of the decoring interval. The value range and variation of the target value are essential to put the resulting model scores into context. Figure 11 shows the evaluation of the target feature for the two base datasets.
The mentioned time series represents the acceleration at a defined point on the surface of a specimen. The acceleration of the specimen is influenced by deviations in the experimental setup, by variations of physical quantities such as rigidity, weight, shape, and Young’s modulus of the specimen. The signal curves thus describe the sequence of the decoring process in an aggregated manner. Three processing paths are carried out to provide the information content for the models as well as possible in scalar characteristics: Fast Fourier transforms (FFT), mel cepstral coefficient analysis (mfcc), and a time domain evaluation. The first two are standard methods of transforming audio or vibration data into scalar features. Analyzing the maxima and minima of the signal in the time domain results in additional, easily interpretable features.
A trigger signal was recorded for the acceleration process, as shown in Figure 7. This trigger signal is used to read out the number and duration of the impacts in the interval. With this information, the signal was split into individual impacts. The length of the impact signals was globally set to 200 ms by superpositioning all impacts. Figure 12 shows this superposition of 10,988 impact signals. With the selected signal length, the signal is cut off after the first post-oscillation. A constant signal duration and a constant sampling rate for all ram impacts allow the summarization of the calculated FFT and mfcc values over the impacts of an interval and to compare the resulting features between all intervals.
Table A3 in Appendix A summarizes the resulting features. The FFT calculated 4800 bins of 4 Hz width between 0 and 16,800 Hz. Each container is a feature used for model training. The mfcc analysis was performed for five coefficients. Each coefficient was summed over all time windows of the mfcc analysis to produce five quantities: Maximum, minimum, mean, median, and standard deviation. The signal processing in the time domains summarizes the maximum positive and negative accelerations of each hammer blow. The maximum positive and maximum negative acceleration in an entire interval, the sum and the mean value of the maximum positive acceleration of each hammer blow in an interval, the sum and the mean value of the maximum negative acceleration of each hammer blow in the interval, and the common absolute sum of all positive and negative maximum accelerations in the interval were stored as features.

3.2. Data Composition of the Twelve Datasets

As described above, using several datasets with different characteristics aims to improve the interpretability of the results and test the defined hypotheses. For this purpose, six subsets are formed for each of the two described base datasets. The first distinction concerns the composition features. The second distinction involves the processing of the signal data. The two base datasets contain all three processing methods. In addition, six other datasets combine one of the two base datasets and one of the processing methods each. These are intended to provide information about which of the signal-processing methods is most useful for model building. Two additional datasets are created using the base datasets with all signal features simultaneously. Table 2 lists all datasets and their composition.

3.3. Basic Model Scores for Unfiltered and Filtered Datasets

This chapter summarizes the model scores for all datasets, randomly initialized runs, and validation data. All shown ranks are calculated according to the mean RMSE over the ten random initializations of one parameter set. The values are calculated based on that dataset’s randomly selected 10% test data or the fixed-selected validation data. A filter fraction of less than 100% means that a reduced feature set achieved the lowest RMSE for that dataset. Due to the volume of results, the tabular overview can be found in Appendix A in Table A6 and Table A7.
As a point graph, Figure 13 shows the results for the three core metrics, mean RMSE, mean MAE rescaled, and relative mean MAE. A linear connection line was added to indicate the trend and connection of the points. The datasets are sorted by the achieved mean RMSE of their best model in ascending order meaning the overall best model is on the very left and the overall worst model on the very right. The integrated table depicts the composition of each dataset. It also includes the filter fraction and model type of the best model for each dataset.
The range of the mean MAE is between 5.9 g and 10 g, corresponding to a relative mean MAE of 19% to 33%. The difference between the mean MAE and the best-achieved MAE of the model depicts the variation and range of the MAE for the ten runs of a parameter set. For the best-ranking models, this variation diminishes, which can be interpreted as a parameter set resulting in a robust model.
The figure also includes the best run value for all three core metrics. The best run out of the ten randomly initialized runs is defined as the one with the lowest RMSE. Furthermore, for each dataset, the lowest achieved RMSE overall runs of all models was added to illustrate the absolute minimum achieved for all datasets. An evident diminishing variation for the best-ranking models can be seen. Additionally, it depicts the less robust behavior of models for datasets with fewer datapoints, DS 2, 4, 6, 8, 10, and 12. Therefore, some models achieved lower RMSE with the extended features of these datasets than the best-ranking models according to the mean RMSE. However, the high difference to the mean value indicates a high variation in the randomly initialized model runs and, therefore, a less robust model. All four datasets that include FFT features yield the worst scores for both kinds of datasets. All datasets except DS 3 profited from the reduction in data complexity. No significant difference can be seen in the use of composition features for both kinds of datasets.
Two more values extend the view of the model performance. The first one is the mean MAE divided by the standard deviation of DM in the corresponding dataset, which allows comparing the model performance to the variation of the target value. The standard deviation and statistical evaluation of DM are depicted in Figure 11. The second one is the coefficient of determination R2, which estimates the models’ level of abstraction regarding the cause–effect correlations toward the target value. Figure 14 summarizes these metrics for all twelve models shown in Figure 13 and adds the highest achieved R2-value by any model for each dataset. DS 9 reaches the best R2-score for the models that performed best on average, which is also the closest to its highest-performing model. This low deviation between the best model on average and the best model overall indicates a robust parameter set. DS 11, which uses time-based features, achieves the overall highest R2-score. However, the difference between its mean best and overall best models is far higher than for DS 9. The highest achieved R2-score for any model is above 0.75, contrary to the diminishing scores shown in Figure 13. The increasing difference can be interpreted as less robust models. The mean MAE of the models depicted in Figure 13 divided by the target value’s standard deviation ranges between 0.34 and 0.57. All model prediction errors variate far less than the target value in the datasets.
Figure 15 illustrates the results for the validation data. Two sets of models are included. The yellow, full line datapoints represent the same models as in Figure 13. The green, separated line datapoints depict the model that achieved the lowest mean RMSE for each dataset for the validation data. These are different models than those in Figure 13. The best achieved RMSE of the ten random initialized runs is also included for both sets of models. An extreme variation in the results can be seen for DS 4, where the difference between the mean and best value is significant compared to the other datapoints. The same datasets achieve the lowest mean RMSE as for the test data, only with slight differences. However, all models perform far worse for the validation data than for the test data. They still show that the same models performing best for the training data also achieved the best values for the validation data. This fact indicates that these robust models can predict some of the outcomes of the validation dataset. However, it can be suspected that they are missing relevant information unique about sand-binder system D in the training process. On the other hand, other models resulted in better predictions for the validation data. These models were less accurate for the test data than the previously discussed models but have generalized more information that allows for better prediction values of the validation data. However, even these models are still significantly worse than the mean model scores for the train data. In summary, the trained models are partially suited for the validation data. Still, not all relevant influences on system D were included in the other systems used to train the models.
Table 3 evaluates the results statistically, as described in the approach chapter. It lists the average, the standard deviation, and the p-value obtained by the two-sided Welch’s t-test for each dataset. DS 9 was used to evaluate the other datasets since it yielded the best mean RMSE. The first test on the left of Table 3 was performed for the mean RMSE calculated over the ten randomly initialized runs of all 507 parameter sets of each dataset. A second Welch’s t-test was applied to gather more insight into the model behavior for the optimized parameter sets by using only the mean RMSE of each dataset’s best 20 parameter sets. The resulting averages show a similar ranking as Figure 13 with only slight changes, like for DS 3 and 11. All obtained p-values are far smaller than the threshold of 5%, with the highest p-value of 6.63 × 10−4 for DS 3. Thus, the different results of the datasets are statistically significant.

3.4. Feature Importance Values

Figure 16 shows SHAP plots of the best XGB model for six datasets. The meaning of the shortened feature names can be found in Table A1, Table A2 and Table A3 in Appendix A. DS 1 in Figure 16a contains only the base features. DS 3 Figure 16b contains the aggregated features describing the composition of the sand-binder system. DS 7 in Figure 16c comprises the base data and FFT features. DS 9 in Figure 16d uses the mfcc features, and DS 11 in Figure 16e uses only the time-based signal features.
Interestingly, the XGB model for DS 11 achieving the lowest mean RMSE filters all signal features and equals DS 1, which explains the same model scores for both datasets a shown in Figure 13. The sieving percentage of the used sand is the most important feature in all four plots. The sand mass, absolute and relative at the beginning of the interval, is also an important factor. The flexural strength, cohesion, phi, and compressive strength are also influential for all four models. The dataset using the mfcc features scores the lowest mean RMSE. Accordingly, 5 of the 25 mfcc features are listed in the most important feature plot. Figure 16b includes the sand-binder system names. In most cases, these aggregated composition features influence only its corresponding datapoints. For example, feature “system_N” has nearly no effect except on the datapoints belonging to system N, colored red in the plot.
A further aspect is whether the most important features change with varying the filter fraction, which is investigated by comparing DS 1, DS 3, and DS 9. The best model of DS 3 in Figure 16b has a filter fraction of 100%, meaning no filtering occurred. It is identical to DS 1 in Figure 16a except for including the composition features. DS 1 has a filter fraction of 90%, meaning that nearly no feature reduction was performed. DS 9 in Figure 16e has the overall best model score and a much lower filtering fraction of 40%. It contains the same features as DS 1 and the mfcc signal features. For all three datasets, the resulting feature importance is very similar. Even more so, the results for the models of DS 1 and DS 9 are identical. The complexity reduction does not affect the resulting feature importance. DS 7 in Figure 16c seems to contradict this statement. However, aside from the many FFT features, five of the same features as before are listed in the top influences. As DS 7 performed significantly worse than the other three datasets, it can be assumed that even at 4% filter fraction, the multitude of remaining FFT features led to a high content of noise in its dataset, and the main information came from the same features as before.

3.5. Model Scores for Varying ANN Complexity

The number of neurons in total and the number of layers in an ANN influence its ability to interpret complex data. The downside is that a larger ANN requires more datapoints; otherwise, overfitting becomes more likely. The RMSEs in Figure 17, Figure 18, Figure 19 and Figure 20 are the mean over all ten random runs for the best parameter set. Twenty-seven different ANN structures were calculated for each parameter set. The layer sets can be grouped into three groups. The first group consists of two layers, where both layers have the same number of neurons. The second group also has two layers, with the second layer having half as many neurons. The third group has three layers; the first two have the same number of neurons, and the third has half as many. Figure 17 shows the absolute number of neurons in the ANNs, which varied between values of 15 and 2000. Depending on the datasets, the RMSE of the ANN models decreases with an increasing number of neurons until a threshold between 40 and 90 total neurons. After this threshold, the model quality variates too much to identify whether the model quality is decreasing slightly or varying along a plateau. Figure 18, Figure 19 and Figure 20 depict the same behavior as Figure 17 showing a threshold after which the RMSE does not decrease significantly. All models trained with DS 5 and DS 7, which contain the FFT features, systematically have a higher mean RMSE. The high number of features introduces disturbance or noise than information. The points of the graphs resemble the best model, using unfiltered and filtered data. Therefore, filtering did not reduce the noise enough to achieve a benefit, even for more complex ANNs.

3.6. Comparing ANN and XGB over Varying Filter Fractions

As mentioned in the explanation of the approach to this work, two different models are used to predict the decored mass. Figure 13 already gives a first hint about their corresponding performance. The XGB models achieved the lowest RMSE and generated the best models for all except two datasets. Figure 21 and Figure 22 allow a second perspective as they show the mean RMSE plotted over the filter fractions for both kinds of models. The values at 100% filter fraction can be used to compare the results of both models. The XGB models demonstrate less variation in the model scores than the ANN models. Furthermore, they indicate a clear preference for the light-colored datasets without extended features. The ANN models achieve the best performance with the datasets without the extended features but show less preference. ANN models demonstrate exceptionally high RMSE for all datasets containing FFT features, especially compared to the XGB models. These score worse values for the FFT features as well but less strongly than the ANN. With increasing feature reduction, the ANN score improves more than the XGB scores. With increasing feature reduction, the ANN score improves more than the XGB scores.
In summary, the best models of each type perform similarly. Nevertheless, the ANN models have more difficulty handling the data and profit more from the complexity reduction. DS 6 and DS 8 profit the most from the data complexity reduction. Both datasets contain a high number of FFT features. Interestingly, DS 5 and DS 7 do not show this behavior to such a high degree. The explanation is the ratio of features to datapoints. DS 6 and 8 are extended-feature datasets with significantly fewer datapoints than DS 5 and DS 7. Reducing the complexity allowed for a similar mean RMSE for all four datasets, despite the difference in the number of datapoints. All models show a stagnating or decreasing RMSE for reducing data complexity up to a given threshold. Going below this filter fraction leads to a significantly higher mean RMSE. The features of the given dataset at this filter threshold are important for predicting the decoring behavior. Dropping them from the dataset reduces the model quality. Figure 16 shows the features in descending order of importance. The features on top are, thus, the last to be dropped when reducing the filter fraction further. The number of features is always an integer and is rounded up as appropriate. Especially for the datasets with few features, only a single feature remains below single-digit filter fractions.

4. Discussion

4.1. Global Model Results

Figure 13 describes the datasets in descending order of their global rank. First, using the rank and the ratio of the mean MAE to the mean target feature value, the five hypotheses described in Section 1.4 are evaluated.
The first hypothesis claims that aggregated features improve the model quality due to their hidden information about the conduction of experiments. DS 1 and DS 2 versus DS 3 and DS 4 differ in the use of the aggregated composition features. DS 1 scores a lower mean RMSE than DS 3, but DS 4 scores better than DS 2. Looking at the exact values, the use of the extended features is far more significant than the use of the aggregated features. They have no significant effect on the model quality, refuting the hypothesis.
The second hypothesis claims that using the acceleration data improve the model quality. The best dataset (DS 9) uses the mfcc features calculated from the acceleration data. The relative difference in mean RMSE is 5.1% to the base dataset (DS 1), which scores rank 2. As seen in Figure 16e, DS 11 scores best when the time-based signals are dropped. DS 5, 6, 7, and 8 contain the FFT features and are reliably the worst for both base datasets. The high number of features was too complex for the number of datapoints, even after complexity reduction. Figure 16e shows the SHAP plot for the best model using FFT features at a 4% filter fraction, which still results in 193 remaining features. It contains the flexural and compressive strengths and the relative masses at the start of the interval. The other listed features are FFT bins without a recognizable pattern. For the datasets containing the extended features, DS 12 scores best using the time-based signal data. Figure 16f shows its most important features, which comprise the base features and two evaluation features of the maximum negative acceleration. In summary, the signal data contain, in the best case, some information and, in the worst case, mostly noise. Once again, this can be explained with constant experimental processes during data collection. Without significant variation, those features contain mostly noise. The 5.1% improvement for DS 9 is too low to be sure whether the improvement comes from more information or model training.
The third hypothesis claims varying model qualities for different signal-processing approaches. As just explained, the differences between the mfcc and time-based features are small. The FFT features yield worse models due to their high count and complexity.
The fourth hypothesis states that including all features will reduce the model quality despite the use of complexity reduction. This statement is true. DS 5 and DS 6 yield for both base datasets the worst model quality. More features without more information merely introduce noise to the data.
The fifth hypothesis states that lower data complexity results in a better model score. This hypothesis can be confirmed. Only DS 3 achieved its lowest mean RMSE with a 100% filter fraction. All other datasets used filter fractions between 1 and 90%. The results and filter fractions in Figure 13 show that all datasets containing extended features profited from a lower filter fraction than those only containing the base features. The remaining most important features for the extended datasets are only the base features. Only the strengths at 400 °C are used in three out of six datasets, but these are in the lower third of important features in all cases. The datasets using the FFT features profited significantly from the data complexity reduction, as seen in Figure 21. The datasets containing the extended features yielded worse models than the less complex datasets. However, the number of datapoints in this last comparison is not constant.
The number of datapoints in the datasets with the extended features was too small to achieve a low mean RMSE. The best model using extended features is DS 12, with 30% filter fraction, and achieves a mean RMSE of 0.127 and a mean MAE of 8.41 g. This is a 37% higher RMSE than the best dataset not using the extended features.

4.2. Comparison to Decoring Theory

The best models achieved a mean MAE of 16 to 20% of the dataset’s mean decored mass. With only a handful of datapoints for each sand-binder system, this can be adjudged a very good fit. The containing features in the data are suited to predict the target feature. The remaining important features after complexity reduction shown in Figure 16 are the flexural strength, compressive strength, phi, and cohesion, all at room temperature, combined with the absolute and relative mass at the start of the decoring process. This collection supports the claim of the previous work that these features describe the decoring behavior of sand-binder systems.
The signal data have no major influence, as discussed above. The theory of decoring explained in [2] suggests that the way of impact is a significant influence. Therefore, the signal data should have a beneficial effect on model scores. This effect cannot be seen, which leads to the explanations already given in the context of the second hypothesis, or that the signal data were processed in a way that is not suitable for extracting the relevant information. Another explanation is that the unintentional variations during the decoring experiments were insignificant for the decoring behavior, which indicates a robust experimental setup of the previous work.
According to the decoring theory described in [2], cohesion and phi are calculated using the measured flexural and compressive strengths. The best models use all four features and deem them equally important. If both sets contain the same information, at least one of the four should be filtered during the complexity reduction. All four are still deemed valuable for the modeling, which leads to two possible explanations. First, the models divided the information equally between the two sets, and they are still the most important compared to other features. Second, the sets are related, but each contains information not transported by the other.
The extended features, such as the strengths of the sand core after casting, were deemed insignificant by the models. A low effect of the extended features contradicts the conclusions of the previous work. These strengths are a major indication of the decoring behavior in the latter. None of the best models used these features. Instead, some of the models used the strengths at 400 °C. Two explanations are possible. First, the strengths at 400 °C already describe the behavior of inorganically bound sand cores for higher temperatures and during casting in a more direct way than the strengths after casting. Second, the low number of datapoints for the extended features did not allow the models to learn the additional information contained in the extended features.

4.3. Discussion of the Model Training and Model Behavior

Each parameter set was calculated with ten different random initializations. The scores are the mean of these ten calculations. The best runs of each parameter set yield 36% lower RMSE on average than the corresponding mean over all ten runs. All models are far less suited for the validation sand-binder system D. Interestingly, models that do not achieve the best RMSE for the test data have better scores for the validation data. This difference suggests that despite the ten randomly initialized runs, there was some overfitting toward the test data. Figure 15 summarizes this behavior. Despite the lower prediction quality, this graph shows a clear learning process by the models. The best models score a mean RMSE for the validation data of 0.166. This is 78% higher than the best model score for the test data. Nevertheless, they score a similar mean RMSE for both test (0.148) and validation data (0.167), suggesting that these models are less suited for the test data but are similar and partially valid for other sand-binder systems.
The models abstracted some information suitable to predict the validation system D. However, they lacked relevant parameters during the training phase to gain similar prediction values as they gained for the test data. Figure 17, Figure 18, Figure 19, Figure 20, Figure 21 and Figure 22 show robust behavior for the test data without significant outliners. The models can be summarized as robust and usable, but only for sand-binder systems known through the training data. New sand-binder systems can contain variations in the features not modeled during the training, which might change by including more sand-binder systems in future works.
Figure 14 summarizes R2 for the models shown in Figure 13. The R2-scores support the previously derived model behavior. DS 9 shows the most reliable scores of all datasets with a low difference between the highest achieved score and the best mean model. All datasets were able to score R2 values above 0.75, indicating that models could learn a significant part of the cause–effect correlations between the input data and the target value DM. Furthermore, it shows the mean MAE achieved by the models in Figure 13 divided by the standard deviation of the target value. The mean MAE is 34% to 57% of the standard deviation of DM. Thus, the prediction is below the datasets’ variations, indicating functioning models.
Interestingly, the XGB models yielded better scores than the ANN in most cases, which can be explained by the features contained in the dataset. These can be used directly for predicting the target features. No complex model building or meta-feature calculations are needed, which is confirmed by Figure 17, Figure 18, Figure 19 and Figure 20. Only the very low number of neurons have slightly worse scores than the more complex ones. Starting with 20 neurons in each layer, the mean RMSE does not improve significantly with more neurons.

5. Conclusions

In summary, experimental data were used to predict the decoring behavior of various sand-binder systems. Robust models were achieved, yielding a mean MAE between 16% and 37% of the mean of the predicted feature and a mean MAE to standard deviation of 34% to 57%. The chosen forward–backward approach using machine learning models to learn about experimental data and physical relations was successful. Varying the data composition demonstrated the relative importance of various feature sets in predicting the decoring behavior. The acceleration data of the decoring process, which was suspected of having a significant influence, was not used significantly by the models for the prediction. The authors conclude that a constant experimental setup led to the reduced importance of the acceleration signal. The same explanation is used for the low importance of the aggregated composition features. One sand-binder system was used as a validation system to test whether the models could describe the decoring behavior for unknown sand-binder systems. The prediction scores are significantly worse for the validation system. However, the models indicate that some generalized correlations were learned that are also valid for the validation system. The authors suspect that the validation system variates in some parameters in ways unique to it. It is suspected that some overfitting occurred in the models that achieved the lowest mean RMSE for the test data, despite averaging over ten runs of randomized initializations. The best models for the validation system scored similar values for the test data and indicated no overfitting. This work focuses on a specific case of decoring for one sample geometry. The machine learning models calculated in this do not claim to predict the decoring behavior for other geometries. However, the behavior of the models and the identified relevant parameters can be interpreted in a generalized way. Determining which features in this dataset are particularly relevant for ML model training is, therefore, a first step towards a more generally valid ML model. This work provided helpful indications as to which data should be ascribed higher priority in future decoring experiments.

Author Contributions

Conceptualization, F.D.; methodology, F.D. and F.E.; software, F.D. and R.L.; investigation, F.E.; writing—original draft preparation, F.D.; writing—review and editing, F.E., D.G., M.M., W.V., and P.L.; supervision, D.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data are not publicly available due to ongoing research with the data. The code presented in this study is available at https://github.com/FDdiss/PredicitingDecoringBehavior (accessed on 16 June 2023).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

The Appendix contains the results for the best models for all datasets, before and after using complexity reduction. The complete results of all models can be obtained by contacting the author.

Used Programming Libraries

The code was written in Python 3.9 [23]. For the XGB and ANN model creation, the scikit-learn library [24] was used. This library uses the original library for the implementation of the XGBoost models [13]. For data processing, pandas was used [25]. For the visual representation of the XGB model results, SHapley Additive exPlanations, (SHAP) was used [17,26] in combination with Matplotlib [27]. The librosa library was used to create the mfcc evaluations [28]. For the FFT evaluations and many smaller calculations, numpy was used [29].
Table A1. Features for each sand-binder system used for model training. “X” means these features are included in the dataset. _name_ refers to the individual name of binder, sand, or system.
Table A1. Features for each sand-binder system used for model training. “X” means these features are included in the dataset. _name_ refers to the individual name of binder, sand, or system.
System Feature NameUnitComposition FeatureExtended FeatureAbbreviated Name
binder name-X binder_name_he
sand name-X sand_name_he
binder content%wtX binder_wt
sand-binder system name-X system_name_he
sieving percentage% sieving_pct
mean grain diameterµm Xmean_grain_d_microm
flexural strength according to the
manufacturer
MPa flex_manuf
flexural strength at 25 °CMPa flex_raw
compressive strength at 25 °CMPa comp_raw
phi at 25 °C° phi_raw_grad
cohesion at 25 °C- cohesion_raw
flexural strength at 400 °CMPa Xflex_400_grC
compressive strength at 400 °CMPa Xcomp_400_grC
phi at 400 °C° Xphi_400_grC_grad
cohesion at 400 °C- Xcohesion_400_grC
flexural strength after castingMPa Xflex_casted
compressive strength after castingMPa Xcomp_casted
phi after casting° Xphi_casted_grad
cohesion after casting- Xcohesion_casted
tension in the middle, perpendicularMPa Xtension_pp_middle
tension in the middle, lengthwiseMPa Xtension_lw_middle
tension near inflow, perpendicularMPa Xtension_pp_inflow
relative flexural strength after casting, according to the manufacturer% Xflex_residual_pct
flexural strength after casting,
according to the manufacturer
MPa Xflex_residual
relative drop of flexural strength after casting, according to the manufacturer% Xflex_residual_drop_pct
Table A2. Features available for each interval.
Table A2. Features available for each interval.
Interval Feature NameInterval PhaseUnitAbbreviated Name
decored mass
target feature “DM”
endg06_i_m_progress
mass of specimenstartg00_i_s_m_total
already decored distancestartcm01_i_s_d_progress
sand mass of specimenstartg02_i_s_m_sand
already decored massstartg03_i_s_m_progress
ratio of sand mass to metal massstart-04_i_s_m_sand_r_metall
ratio of sand mass to total massstart-05_i_s_m_sand_r_total
number of impacts during the interval 20_i_n_hammerblows
Table A3. Signal-processing methods and resulting features.
Table A3. Signal-processing methods and resulting features.
PreprocessingNo. of
Features
Names
Fast Fourier Transformation
4800 bins of 4 Hz each
4800fft_00004_Hz to fft_19200_hz
Mel Frequency Cepstral Coefficient Analysis
mean, median, stdev, max, and min
for five coefficients
25mean_mfcc01…05
median_mfcc01…05
stdev_mfcc01…05
max_mfcc01…05
min_mfcc01…05
Time-Based Analysis
a: absolute acceleration
sum: sum over interval (i) or bar (b),
max: maximum; pos: positive; neg: negative
Example: 21_i_a_sum_max_posneg means “feature 21, absolute sum over all positive and negative maxima over all hammer blows of the examined interval”
721_i_a_sum_max_posneg
22_i_a_sum_max_pos
23_i_a_sum_max_neg
24_i_a_max_pos
25_i_a_max_pos
26_i_i_a_mean_max_pos
27_i_a_mean_max_neg
Table A4. Filter fractions and model parameters with values in the model calculations.
Table A4. Filter fractions and model parameters with values in the model calculations.
Filter Fractions in Percent:100, 90, 80, 70, 60, 50, 40, 30, 20, 10, 7, 4, 1
XGB-parametervalue(s)
n_estimators2000
early_stopping_rounds50
eta0.1
parallel tree1, 3, 5, 7
max depth2, 4, 6
min child weight 1
subsample0.6
colsample_bytree0.7
colsample_bylevel0.7
colsample_bynode0.7
alpha 0.0001
lambda1
gamma0
random seed3
ANN-parametervalue(s)
max iter2000
n iter no change50
layers(10, 10), (20, 20), (30, 30), (40, 40), (50, 50),
(100, 100), (200, 200), (400, 400), (800, 800),
(10, 5), (20, 10), (30, 15), (40, 20), (50, 25),
(100, 50), (200, 100), (400, 200), (800, 400),
(10, 10, 5), (20, 20, 10), (30, 30, 15), (40, 40, 20), (50, 50, 25),
(100, 100, 50), (200, 200, 100), (400, 400, 200), (800, 800, 400)
alpha0.0001
batch size30
learning_rate_init0.0001
tol1 × 10−6
validation fraction0.1
activationrelu
beta 10.9
beta 20.999
epsilon1 × 10−8
max fun15,000
learning rateconstant
shuffletrue
random state3
Table A5. Global rank of all datasets and mean scores of the best corresponding model according to the RMSE for the test data; also listed are the corresponding filter fraction and the composition of the dataset.
Table A5. Global rank of all datasets and mean scores of the best corresponding model according to the RMSE for the test data; also listed are the corresponding filter fraction and the composition of the dataset.
DatasetDS 9DS 1DS 11DS 3DS 7DS 5DS 12DS 10DS 4DS 2DS 8DS 6
Global Rank123456789101112
mean RMSE (-)0.0930.0980.0980.0990.1140.1170.1270.1330.1380.1420.1470.153
best RMSE (-)0.0690.0700.0700.0730.0810.0900.0800.0770.0790.0800.1220.126
mean MAE (g)5.785.835.835.857.267.448.419.079.239.599.2510.1
best MAE (g)4.904.394.394.845.355.745.666.065.775.758.126.73
mean DM (g)
(of 10% test data)
24.724.724.724.824.824.828.828.129.029.029.830.2
mean MAE,
relative (%)
232424303029323232333134
best MAE,
relative (%)
191818202223212221212625
filter fraction (%)409050100743030303041
extended features xxxxxx
composition
features
x x
signal processingmfccnonetime-
based
noneFFTalltime-
based
mfccnonenoneFFTall
Table A6. Mean scores for all best models for all datasets.
Table A6. Mean scores for all best models for all datasets.
Target Feature: Decored Mass in Interval (DM)
Test
Dataset
Best XGBoostBest ANN
R2
(-)
RMSE
(-)
MAE
(g)
RankR2
(-)
RMSE
(-)
MAE
(g)
Rank
DS 10.750.0995.791/20.730.1056.301
DS 20.560.14510.280.570.1489.986
DS 30.750.0995.851/20.710.1086.382
DS 40.530.14910.490.550.1469.605
DS 50.610.1277.855/60.270.17210.99
DS 60.320.17911.312−0.250.24816.112
DS 70.610.1277.915/60.270.17310.910
DS 80.350.17411.311−0.120.22914.611
DS 90.730.1056.7430.660.1187.514
DS 100.520.15110.0100.480.16010.68
DS 110.700.1106.6040.680.1136.853
DS 120.620.1389.2970.540.1499.817
Val
Dataset
Same XGB Model as AboveSame ANN Model as Above
R2
(-)
RMSE
(-)
MAE
(g)
RankR2
(-)
RMSE
(-)
MAE
(g)
Rank
DS 1−1.470.26125.04−2.470.3128.85
DS 2−2.540.33033.47−13806.12100,40912
DS 3−1.540.26625.45−2.830.3330.46
DS 4−3.500.37136.89−8944.16100,23911
DS 5−0.900.23022.31/2−1.520.2625.13
DS 6−4.540.41140.012−7.540.4744.57
DS 7−0.900.23022.31/2−1.190.2523.72
DS 8−4.170.39738.810−29.70.8271.28
DS 9−1.210.24723.83−0.640.2120.91
DS 10−4.340.40339.411−11885.334279
DS 11−1.590.26925.66−1.800.2826.24
DS 12−2.660.33533.88−28037.51400,17910
Table A7. Using filtered data, mean scores for all best models for all datasets.
Table A7. Using filtered data, mean scores for all best models for all datasets.
Target Feature: Decored Mass
Test
Dataset
Best XGBoostBest ANN
R2 (-)RMSE (-)MAE
(g)
Filter Fraction (%)RankR2 (-)RMSE (-)MAE
(g)
Filter Fraction (%)Rank
DS 10.760.0985.83902/30.730.1056.301003/4
DS 20.560.14510.210090.590.1429.59309
DS 30.750.0995.8510040.730.1046.05602
DS 40.550.14610.290100.610.1389.23305
DS 50.670.1177.44460.540.1378.7846
DS 60.470.15310.11120.420.15710.4411
DS 70.680.1147.26750.510.1419.0348
DS 80.500.1479.254110.430.16110.7412
DS 90.780.0935.784010.750.1006.19201
DS 100.640.1339.073080.560.1439.473010
DS 110.760.0985.83502/30.730.1056.37503/4
DS 120.670.1278.413070.620.1399.19707
Val.
Dataset
Same XGB Model as AboveSame ANN Model as Above
R2 (-)RMSE (-)MAE
(g)
Filter Fraction (%)RankR2 (-)RMSE (-)MAE
(g)
Filter Fraction (%)Rank
DS 1−1.510.26425.2904/5−2.50.30828.81006
DS 2−2.540.33033.41008−3.80.38237.5307
DS 3−1.540.26625.41006−2.00.28827.2605
DS 4−3.330.36436.29010−540.93279.13010
DS 5−1.000.23622.941−0.60.21320.943
DS 6−5.010.42841.3112−5.10.42941.449
DS 7−1.120.24323.472−1.00.23622.944
DS 8−3.940.38838.1411−4.40.40439.448
DS 9−1.150.24523.6403−0.30.19019.0201
DS 10−2.660.33533.8309−441.00287.03011
DS 11−1.510.26425.2504/5−0.50.20520.3502
DS 12−2.400.32232.8307−126053092.00 × 1057012

References

  1. Holtzer, M. Mold and Core Sands in Metalcasting. Sustainable Development; Springer International Publishing AG: Cham, Switzerland, 2020; pp. 219–221. [Google Scholar] [CrossRef]
  2. Ettemeyer, F.; Schweinefuß, M.; Lechner, P.; Stahl, J.; Greß, T.; Kaindl, J.; Durach, L.; Volk, W.; Günther, D. Characterisation of the decoring behaviour of inorganically bound cast-in sand cores for light metal casting. J. Mater. Process. Technol. 2021, 296, 117201. [Google Scholar] [CrossRef]
  3. Herfurth, K.; Scharf, S. Casting. In Springer Handbook of Mechanical Engineering; Karl-Heinrich, G., Hamid, H., Eds.; Springer International Publishing: Cham, Switzerland, 2021; Volume 98, pp. 325–356. [Google Scholar] [CrossRef]
  4. Xin, F.H.; Liu, W.H.; Song, L.; Li, Y.M. Modification of inorganic binder used for sand core-making in foundry practice. China Foundry 2020, 17, 341–346. [Google Scholar] [CrossRef]
  5. Stauder, B.J.; Berbic, M.; Schumacher, P. Mohr-Coulomb failure criterion from unidirectional mechanical testing of sand cores after thermal exposure. J. Mater. Process. Technol. 2019, 274, 116274. [Google Scholar] [CrossRef]
  6. Lechner, P.; Stahl, J.; Hartmann, C.; Ettemeyer, F.; Volk, W. Mohr–Coulomb characterisation of inorganically-bound core materials. J. Mater. Process. Technol. 2021, 296, 117214. [Google Scholar] [CrossRef]
  7. Lee, K.; Ayyasamy, M.V.; Ji, Y.; Balachandran, P.V. A comparison of explainable artificial intelligence methods in the phase classification of multi-principal element alloys. Sci. Rep. 2022, 12, 11591. [Google Scholar] [CrossRef]
  8. Stadter, C.; Kick, M.K.; Schmoeller, M.; Zaeh, M.F. Correlation analysis between the beam propagation and the vapor capillary geometry by machine learning. Procedia CIRP 2020, 94, 742–747. [Google Scholar] [CrossRef]
  9. Wang, P.; Fan, Z.; Kazmer, D.O.; Gao, R.X. Orthogonal Analysis of Multisensor Data Fusion for Improved Quality Control. J. Manuf. Sci. Eng. 2017, 139, 5. [Google Scholar] [CrossRef]
  10. Meng, Y.; Yang, N.; Qian, Z.; Zhang, G. What Makes an Online Review More Helpful: An Interpretation Framework Using XGBoost and SHAP Values. J. Theor. Appl. Electron. Commer. Res. 2021, 16, 466–490. [Google Scholar] [CrossRef]
  11. Philine, K.; Johannes, G.; Dierk, H. Analyse von Gießereidaten mit Methoden des Maschinellen Lernens—Teil 2. Giess.-Prax. 2018, 69, 9–15. Available online: https://www.giesserei-praxis.de/news-artikel/artikel/analyse-von-giessereidaten-mit-methoden-des-maschinellen-lernens-teil-2 (accessed on 16 June 2023).
  12. Nasiri, H.; Kheyroddin, G.; Dorrigiv, M.; Esmaeili, M.; Nafchi, A.R.; Ghorbani, M.H.; Zarkesh-Ha, P. Classification of COVID-19 in Chest X-ray Images Using Fusion of Deep Features and LightGBM. In Proceedings of the 2022 IEEE World AI IoT Congress (AIIoT), Seattle, WA, USA, 6–9 June 2022; IEEE: New York, NY, USA, 2022; pp. 201–206. [Google Scholar] [CrossRef]
  13. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; Association for Computing Machinery (KDD ’16): New York, NY, USA, 2016; pp. 785–794. [Google Scholar] [CrossRef] [Green Version]
  14. Huang, J.; Algahtani, M.; Kaewunruen, S. Energy Forecasting in a Public Building: A Benchmarking Analysis on Long Short-Term Memory (LSTM), Support Vector Regression (SVR), and Extreme Gradient Boosting (XGBoost) Networks. Appl. Sci. 2022, 12, 9788. [Google Scholar] [CrossRef]
  15. Chelgani, S.C.; Nasiri, H.; Tohry, A. Modeling of particle sizes for industrial HPGR products by a unique explainable AI tool-A “Conscious Lab” development. Adv. Powder Technol. 2021, 32, 4141–4148. [Google Scholar] [CrossRef]
  16. Fatahi, R.; Nasiri, H.; Homafar, A.; Khosravi, R.; Siavoshi, H.; Chehreh Chelgani, S. Modeling operational cement rotary kiln variables with explainable artificial intelligence methods—A “conscious lab” development. Part. Sci. Technol. 2023, 41, 715–724. [Google Scholar] [CrossRef]
  17. Lundberg, S.M.; Lee, S.-I. A Unified Approach to Interpreting Model Predictions. In Advances in Neural Information Processing Systems 30; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2017; pp. 4765–4774. Available online: http://papers.nips.cc/paper/7062-a-unified-approach-to-interpreting-model-predictions.pdf (accessed on 16 June 2023).
  18. Volk, W.; Groche, P.; Brosius, A.; Ghiotti, A.; Kinsey, B.L.; Liewald, M.; Junying, M.; Jun, Y. Models and modelling for process limits in metal forming. CIRP Annals 2019, 68, 775–798. [Google Scholar] [CrossRef]
  19. Chehreh Chelgani, S.; Nasiri, H.; Tohry, A.; Heidari, H.R. Modeling industrial hydrocyclone operational variables by SHAP-CatBoost—A “conscious lab” approach. Powder Technol. 2023, 420, 118416. [Google Scholar] [CrossRef]
  20. Nasiri, H.; Ebadzadeh, M.M. MFRFNN: Multi-Functional Recurrent Fuzzy Neural Network for Chaotic Time Series Prediction. Neurocomputing 2022, 507, 292–310. [Google Scholar] [CrossRef]
  21. Chang, A.M.; Freeze, J.G.; Batista, V.S. Hammett neural networks: Prediction of frontier orbital energies of tungsten-benzylidyne photoredox complexes. Chem. Sci. 2019, 10, 6844–6854. [Google Scholar] [CrossRef] [Green Version]
  22. Lechner, P.; Heinle, P.; Hartmann, C.; Bauer, C.; Kirchebner, B.; Dobmeier, F.; Volk, W. Feasibility of Acoustic Print Head Monitoring for Binder Jetting Processes with Artificial Neural Networks. Appl. Sci. 2021, 11, 10672. [Google Scholar] [CrossRef]
  23. Rossum, G.; Drake, F.L., Jr. Python reference manual: Centrum voor Wiskunde en Informatica Amsterdam. 1995. Available online: https://docs.python.org/3/reference/index.html (accessed on 16 June 2023).
  24. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. Available online: https://www.jmlr.org/papers/v12/pedregosa11a.html (accessed on 16 June 2023).
  25. Wes, M. Data Structures for Statistical Computing in Python. In Proceedings of the 9th Python in Science Conference, Austin, TX, USA, 28 June–3 July 2010; pp. 56–61. [Google Scholar] [CrossRef] [Green Version]
  26. Lundberg, S.M.; Erion, G.; Chen, H.; DeGrave, A.; Prutkin, J.M.; Nair, B.; Katz, R.; Himmelfarb, J.; Bansal, N.; Lee, S.-I. From local explanations to global understanding with explainable AI for trees. Nat. Mach. Intell. 2020, 2, 2522–5839. [Google Scholar] [CrossRef] [PubMed]
  27. Hunter, J.D. Matplotlib: A 2D graphics environment. Comput. Sci. Eng. 2007, 9, 90–95. [Google Scholar] [CrossRef]
  28. McFee, B.; Raffel, C.; Liang, D.; Ellis, D.P.; McVicar, M.; Battenberg, E.; Nieto, O. Librosa: Audio and music signal analysis in python. In Proceedings of the 14th Python in Science Conference, Austin, TX, USA, 6–12 July 2015; Volume 8. [Google Scholar] [CrossRef] [Green Version]
  29. Harris, C.R.; Millman, K.J.; Van Der Walt, S.J.; Gommers, R.; Virtanen, P.; Cournapeau, D.; Wieser, E.; Taylor, J.; Berg, S.; Smith, N.J.; et al. Array programming with NumPy. Nature 2020, 585, 357–362. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Drucker–Prager model for failure. Adapted with permission from Ref. [2]. 2021, Elsevier.
Figure 1. Drucker–Prager model for failure. Adapted with permission from Ref. [2]. 2021, Elsevier.
Applsci 13 07948 g001
Figure 2. Types of models [2]. Adapted with permission from Ref. [18]. 2019, Elsevier.
Figure 2. Types of models [2]. Adapted with permission from Ref. [18]. 2019, Elsevier.
Applsci 13 07948 g002
Figure 3. Using model quality and behavior to evaluate experimental data.
Figure 3. Using model quality and behavior to evaluate experimental data.
Applsci 13 07948 g003
Figure 4. Big picture of the chosen approach.
Figure 4. Big picture of the chosen approach.
Applsci 13 07948 g004
Figure 5. Form and size of sand core bar and casted hollow rectangular rod used as specimen [2].
Figure 5. Form and size of sand core bar and casted hollow rectangular rod used as specimen [2].
Applsci 13 07948 g005
Figure 6. Sequence of experiments. Adapted with permission from Ref. [2]. 2021, Elsevier.
Figure 6. Sequence of experiments. Adapted with permission from Ref. [2]. 2021, Elsevier.
Applsci 13 07948 g006
Figure 7. Decoring setup. Adapted with permission from Ref. [2]. 2021, Elsevier.
Figure 7. Decoring setup. Adapted with permission from Ref. [2]. 2021, Elsevier.
Applsci 13 07948 g007
Figure 8. Sequence of decoring and measurements of decoring progress.
Figure 8. Sequence of decoring and measurements of decoring progress.
Applsci 13 07948 g008
Figure 9. Pseudo-code of the model training approach.
Figure 9. Pseudo-code of the model training approach.
Applsci 13 07948 g009
Figure 10. Summary of the available input data for the model training and its sources.
Figure 10. Summary of the available input data for the model training and its sources.
Applsci 13 07948 g010
Figure 11. Box plots of the target value, separated for Dataset 1 and Dataset 2. Dataset 1 uses all datapoints but less features and Dataset 2 uses all features but less datapoints. It lists the number of datapoints (n) and the standard deviation (stdv) of each dataset. All numbers except the mean are integer values, so no decimals are given.
Figure 11. Box plots of the target value, separated for Dataset 1 and Dataset 2. Dataset 1 uses all datapoints but less features and Dataset 2 uses all features but less datapoints. It lists the number of datapoints (n) and the standard deviation (stdv) of each dataset. All numbers except the mean are integer values, so no decimals are given.
Applsci 13 07948 g011
Figure 12. Superposition of all 10,988 impacts.
Figure 12. Superposition of all 10,988 impacts.
Applsci 13 07948 g012
Figure 13. Main results of the model training by showing the best model of each dataset (DS). The datasets are sorted by the achieved mean RMSE of their best model in ascending order meaning the overall best model is on the very left and the overall worst model on the very right. The integrated table depicts the composition of each dataset. “X” means these features are included in the dataset. It also includes the filter fraction and model type of the best model for each dataset. The graph shows the main model metric, the mean RMSE (circle). Additionally, it shows the best RMSE achieved for the same parameter (triangle) and the best RMSE over all parameter sets and randomly initialized runs (square). Furthermore, it shows the mean MAE (circle), rescaled and relative, as well as the corresponding best-achieved value in the ten randomly initialized runs (triangle). All values are calculated using the test data.
Figure 13. Main results of the model training by showing the best model of each dataset (DS). The datasets are sorted by the achieved mean RMSE of their best model in ascending order meaning the overall best model is on the very left and the overall worst model on the very right. The integrated table depicts the composition of each dataset. “X” means these features are included in the dataset. It also includes the filter fraction and model type of the best model for each dataset. The graph shows the main model metric, the mean RMSE (circle). Additionally, it shows the best RMSE achieved for the same parameter (triangle) and the best RMSE over all parameter sets and randomly initialized runs (square). Furthermore, it shows the mean MAE (circle), rescaled and relative, as well as the corresponding best-achieved value in the ten randomly initialized runs (triangle). All values are calculated using the test data.
Applsci 13 07948 g013
Figure 14. Evaluation of R2 and the mean MAE compared to the standard deviation (stdv) of the true values of DM. All values refer to the model predictions for the test data.
Figure 14. Evaluation of R2 and the mean MAE compared to the standard deviation (stdv) of the true values of DM. All values refer to the model predictions for the test data.
Applsci 13 07948 g014
Figure 15. Main results for the validation data. The yellow, full line datapoints represent the same models as in Figure 13. The green, separated line datapoints depict for each dataset the model that achieved the lowest mean RMSE for the validation data. These are different models than in Figure 13.
Figure 15. Main results for the validation data. The yellow, full line datapoints represent the same models as in Figure 13. The green, separated line datapoints depict for each dataset the model that achieved the lowest mean RMSE for the validation data. These are different models than in Figure 13.
Applsci 13 07948 g015
Figure 16. SHAP plots of (a) DS 1 90% filter fraction (ff), (b) DS 3 100% ff, (c) DS 7 4% ff, (d) DS 9 40% ff, (e) DS 11 50% ff, (f) DS 12 70% ff, for the single best model of the ten iterations with the lowest mean RMSE. For each dataset, all features or, at most, the 30 most important features are listed in descending order. Each line comprises all the datapoints available as single dots and corresponds to the feature on the left. The position of the dot on the line represents whether the model output for this datapoint and feature was negative (less decored) or positive (more decored), with the middle line resulting in no change of the model output. The color of the dot represents the original feature value for this datapoint relative to all datapoints: red equals high values; blue equals low values. When many points are at the same SHAP value, they are stacked, making the horizontal line wider at this value.
Figure 16. SHAP plots of (a) DS 1 90% filter fraction (ff), (b) DS 3 100% ff, (c) DS 7 4% ff, (d) DS 9 40% ff, (e) DS 11 50% ff, (f) DS 12 70% ff, for the single best model of the ten iterations with the lowest mean RMSE. For each dataset, all features or, at most, the 30 most important features are listed in descending order. Each line comprises all the datapoints available as single dots and corresponds to the feature on the left. The position of the dot on the line represents whether the model output for this datapoint and feature was negative (less decored) or positive (more decored), with the middle line resulting in no change of the model output. The color of the dot represents the original feature value for this datapoint relative to all datapoints: red equals high values; blue equals low values. When many points are at the same SHAP value, they are stacked, making the horizontal line wider at this value.
Applsci 13 07948 g016aApplsci 13 07948 g016b
Figure 17. Mean RMSE of best ANN parameter set vs. total number of neurons in the ANN.
Figure 17. Mean RMSE of best ANN parameter set vs. total number of neurons in the ANN.
Applsci 13 07948 g017
Figure 18. Mean RMSE of the best ANN parameter set. Number of neurons in layer 1 equals number of neurons in layer 2.
Figure 18. Mean RMSE of the best ANN parameter set. Number of neurons in layer 1 equals number of neurons in layer 2.
Applsci 13 07948 g018
Figure 19. Mean RMSE of best ANN parameter set. Number of neurons in layer 1 is twice the number of neurons in layer 2.
Figure 19. Mean RMSE of best ANN parameter set. Number of neurons in layer 1 is twice the number of neurons in layer 2.
Applsci 13 07948 g019
Figure 20. Mean RMSE of best ANN parameter set. Layer 1 equals layer 2. Both are twice that of layer 3.
Figure 20. Mean RMSE of best ANN parameter set. Layer 1 equals layer 2. Both are twice that of layer 3.
Applsci 13 07948 g020
Figure 21. ANN mean RMSE plotted over varying filter fractions. Light line colors represent datasets without extended features; dark line colors represent datasets with extended features.
Figure 21. ANN mean RMSE plotted over varying filter fractions. Light line colors represent datasets without extended features; dark line colors represent datasets with extended features.
Applsci 13 07948 g021
Figure 22. XGB mean RMSE plotted over varying filter fractions. Light line colors represent datasets without extended features; dark line colors represent datasets with extended features.
Figure 22. XGB mean RMSE plotted over varying filter fractions. Light line colors represent datasets without extended features; dark line colors represent datasets with extended features.
Applsci 13 07948 g022
Table 1. Sand-binder systems used in the presented work.
Table 1. Sand-binder systems used in the presented work.
Sand-Binder Systems
FeatureABDLMNPQ
binder nameRR_lowRRR_lowR_lowR_lowR
binder content1.91.52.251.91.91.91.92.25
sand nameH32H32F34F34H32W65F34H32
number of
specimens tested
171221612655
total number of
intervals
conducted
641723485483653
extended featuresyesyesyesnonononono
Table 2. All datasets with their composition, feature, and data count. “X” means these features are included in the dataset.
Table 2. All datasets with their composition, feature, and data count. “X” means these features are included in the dataset.
DatasetsBase
Features
Extended FeaturesAggregated
Composition Features
Signal
Features
No. of
Features
Number of
Datapoints
DS 1X none10280
DS 2XX none2481
DS 3X Xnone24280
DS 4XXXnone3281
DS 5X all4842280
DS 6XX all485681
DS 7X FFT4810280
DS 8XX FFT482481
DS 9X mfcc35280
DS 10XX mfcc4981
DS 11X time-based17280
DS 12XX time-based3181
Table 3. Statistical evaluation of the twelve parameter sets. The p-value of a two-tailed Welch’s t-test with a 0.05 significance level was calculated between each dataset and DS 9, which yielded the overall best RMSE. The first three columns represent the evaluation of the mean RMSE of all 507 parameter sets of each dataset. The last three columns show the evaluation of the mean RMSE of the best 20 parameter sets, which scored the best mean RMSE for each dataset.
Table 3. Statistical evaluation of the twelve parameter sets. The p-value of a two-tailed Welch’s t-test with a 0.05 significance level was calculated between each dataset and DS 9, which yielded the overall best RMSE. The first three columns represent the evaluation of the mean RMSE of all 507 parameter sets of each dataset. The last three columns show the evaluation of the mean RMSE of the best 20 parameter sets, which scored the best mean RMSE for each dataset.
All 507 Parameter Sets20 Best Parameter Sets of Each Dataset
(According to Mean RMSE)
DatasetsAverageStandard
Deviation
p-ValueAverageStandard
Deviation
p-Value
DS 90.1286.07 × 10−4-0.0961.24 × 10−6-
DS 10.1386.12 × 10−43.08 × 10−100.1002.76 × 10−71.97 × 10−14
DS 110.1389.57 × 10−41.82 × 10−80.1014.98 × 10−74.31 × 10−18
DS 30.1348.85 × 10−46.63 × 10−40.1001.49 × 10−72.15 × 10−14
DS 70.1615.84 × 10−42.43 × 10−860.1171.68 × 10−61.43 × 10−37
DS 50.1615.50 × 10−41.67 × 10−840.1191.23 × 10−67.83 × 10−41
DS 120.1748.28 × 10−49.63 × 10−1230.1328.68 × 10−63.11 × 10−26
DS 100.1796.70 × 10−42.97 × 10−1570.1375.05 × 10−68.84 × 10−34
DS 40.1766.40 × 10−44.81 × 10−1470.1476.46 × 10−65.84 × 10−33
DS 20.1775.92 × 10−43.43 × 10−1520.1483.77 × 10−62.65 × 10−40
DS 80.2181.73 × 10−32.57 × 10−2060.1521.20 × 10−53.85 × 10−28
DS 60.2201.56 × 10−31.61 × 10−2220.1563.29 × 10−66.30 × 10−45
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Dobmeier, F.; Li, R.; Ettemeyer, F.; Mariadass, M.; Lechner, P.; Volk, W.; Günther, D. Predicting and Evaluating Decoring Behavior of Inorganically Bound Sand Cores, Using XGBoost and Artificial Neural Networks. Appl. Sci. 2023, 13, 7948. https://doi.org/10.3390/app13137948

AMA Style

Dobmeier F, Li R, Ettemeyer F, Mariadass M, Lechner P, Volk W, Günther D. Predicting and Evaluating Decoring Behavior of Inorganically Bound Sand Cores, Using XGBoost and Artificial Neural Networks. Applied Sciences. 2023; 13(13):7948. https://doi.org/10.3390/app13137948

Chicago/Turabian Style

Dobmeier, Fabian, Rui Li, Florian Ettemeyer, Melvin Mariadass, Philipp Lechner, Wolfram Volk, and Daniel Günther. 2023. "Predicting and Evaluating Decoring Behavior of Inorganically Bound Sand Cores, Using XGBoost and Artificial Neural Networks" Applied Sciences 13, no. 13: 7948. https://doi.org/10.3390/app13137948

APA Style

Dobmeier, F., Li, R., Ettemeyer, F., Mariadass, M., Lechner, P., Volk, W., & Günther, D. (2023). Predicting and Evaluating Decoring Behavior of Inorganically Bound Sand Cores, Using XGBoost and Artificial Neural Networks. Applied Sciences, 13(13), 7948. https://doi.org/10.3390/app13137948

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop