Previous Article in Journal
Two Classes of Restart Algorithms for Solving Pseudomonotone Nonlinear Equations
Previous Article in Special Issue
Modeling Approaches for Digital Plant Phenotyping Under Dynamic Conditions of Natural, Climatic and Anthropogenic Factors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Compressive Strength of Geopolymer Concrete Prediction Using Machine Learning Methods

by
Sergey A. Stel’makh
1,
Alexey N. Beskopylny
2,*,
Evgenii M. Shcherban’
3,
Irina Razveeva
1,
Samson Oganesyan
1,
Diana M. Shakhalieva
4,
Andrei Chernil’nik
1 and
Gleb Onore
5
1
Department of Unique Buildings and Constructions Engineering, Don State Technical University, 344003 Rostov-on-Don, Russia
2
Department of Transport Systems, Faculty of Roads and Transport Systems, Don State Technical University, 344003 Rostov-on-Don, Russia
3
Department of Engineering Geometry and Computer Graphics, Don State Technical University, 344003 Rostov-on-Don, Russia
4
Department of Design, Don State Technical University, 344003 Rostov-on-Don, Russia
5
Institute of Applied Computer Science, University ITMO, Kronverksky Pr. 49, 197101 Saint Petersburg, Russia
*
Author to whom correspondence should be addressed.
Algorithms 2025, 18(12), 744; https://doi.org/10.3390/a18120744
Submission received: 2 November 2025 / Revised: 22 November 2025 / Accepted: 25 November 2025 / Published: 26 November 2025
(This article belongs to the Special Issue AI Applications and Modern Industry)

Abstract

The implementation of machine learning methods as one of the artificial intelligence technologies has allowed bringing the construction process to a new qualitative level. Significant interest in these methods is observed in predictive modeling of the building materials’ properties. In the scientific field of innovative concretes, limitations exist regarding the disclosure of intelligent algorithms’ capabilities to predict material properties when altering specific chemical elements and process parameters. This article focuses on seven machine learning techniques that are used to solve the issue in forecasting geopolymer concrete’s compressive strength: from the simplest, such as Linear Regression, to more complex and modern methods, including the TabPFNv2 generative transformer model. The dataset was formed based on 204 datasets available in the public domain, including the author’s experimental data. The leading machine learning features were selected: blast-furnace granulated slag (kg/m3); NaOH molarity; NaOH content in the alkaline activator (%); Na2SiO3 content in the alkaline activator (%); fiber type; fiber dosage (%); and curing temperature (°C). The MAE, RMSE, MAPE metrics and the R2 determination coefficient were used to evaluate the prediction quality. The kNN method (MAE = 0.37, RMSE = 0.63, MAPE = 1.62%, R2 = 0.9996) and TabPFNv2 (MAE = 0.46, RMSE = 0.64, MAPE = 1.39%, R2 = 0.9996) presented the highest accuracy in predicting compressive strength, as assessed by the chosen parameters. If computing resources are limited and interpretability is required, it is recommended to use the CatBoost or Random Forest algorithms; if a graphics processing unit and a small dataset are available, it is advisable to use TabPFN; if there is no need for manual parameter adjustment, H2O AutoML is suitable.

1. Introduction

Intelligent methods have become widespread across all sectors of the economy. Smart solutions, which provide accessible automation tools, improve operational processes in all areas of human activity. For example, in agriculture in many countries, such solutions help combat crop diseases [1] and harmful insects [2]. In medicine and healthcare, intelligent solutions are used to ensure the integrity of inventory and timely access to vital resources [3], as well as accurate and early diagnosis of various diseases [4,5,6]. In security and disaster monitoring, computer vision systems enable the rapid and accurate detection of fires and burned-out areas [7,8,9], and also assess the consequences of emergencies [10].
On construction sites, such technologies help ensure the safety of construction work [11]. Computer vision (CV) systems ensure seamless monitoring of working conditions and work progress at all stages [12,13,14]. CV systems are also adapted for monitoring construction materials, products, and structures. Accuracy in identifying defects and problem areas reaches 99% for detection, segmentation, and classification, as confirmed by studies [15,16,17,18].
With the advent of AI in the field of predictive modeling of building materials properties, tools have emerged that complement traditional methods [19,20,21]. For example, machine learning algorithms have proven themselves in determining the mechanical properties of concrete, which are among the most important parameters in the design and assessment of its performance properties [22,23,24]. Studies [25,26] examined how aggressive environments (freeze–thaw, chlorides, sulfates, wetting-drying cycles) affected the predicted strength of vibrocentrifuged variotropic concrete. Therefore, the MAPE for the top three models was 1–2%, and the coefficient of determination ranged from 0.99 to 0.999. This supports the idea that intelligent algorithms can enhance the life cycle management of this building material. Machine learning-based prediction of geopolymer composite properties is reported in [27,28,29,30,31,32]. The accuracy of the presented methods reaches 99%. Advanced ensemble models such as LightGBM, GBRT, XGBoost, and AdaBoost are increasingly emerging as robust and efficient alternatives to traditional testing methods, providing significant time and cost savings. They demonstrate high performance metrics, such as a coefficient of determination starting from 0.97 [33,34]. But they make little use of modern methods such as automated machine learning (AutoML) or the generative transform of the tabular prior network (TabPFN), which is a gap in the current scientific literature [35,36,37,38]. This study aims to implement seven different machine learning methods: from the most trivial, such as linear regression, to more complex and modern methods, including the TabPFNv2 generative transformer model. Their implementation and evaluation of the results will allow us to draw conclusions about the feasibility of using machine learning methods as tools for analyzing the mechanical properties of geopolymer concrete. This study is a first step toward understanding the internal relationships and dependencies when analyzing the influence of specific properties on the strength of geopolymer concretes. The results of the study revealed properties that should be prioritized when implementing materials science concepts that contribute to understanding the behavior of geopolymer concretes. Research objectives:
-
Data collection from [39,40,41,42,43,44,45,46,47] and dataset formation;
-
Exploratory Data Analysis;
-
Implementation, training, and quality assessment of AI models;
-
Developing recommendations regarding the practical implementation of the algorithms obtained.
-
Assessment of future prospects for enhanced forecasting performance.
The subsequent organization of this study is as described. Section 1, Introduction, summarizes the subject area and describes the main areas of application of intelligent algorithms, particularly in the construction industry. It identifies scientific gaps and formulates the purpose and objectives of the study. Section 2, Materials and Methods, presents the general framework of the study, describes the dataset acquisition process, describes its statistical characteristics, and substantiates the choice of intelligent algorithms and metrics for assessing the quality of forecasting results. Section 3, Results and Discussion, presents the results of the algorithm implementation and provides a comparative analysis with existing achievements in this subject area. Section 4, Conclusions, contains the main findings of the study and prospects for the development and improvement of forecasting results.

2. Materials and Methods

Figure 1 shows the overall research framework, which consists of six interconnected stages:
  • Dataset collection.
  • Exploratory Data Analysis (EDA).
  • Selecting AI models.
  • Implementation and training of AI models.
  • Evaluation of quality metrics.
  • Formulation of recommendations for practical use.
Figure 1. Research Steps.
Figure 1. Research Steps.
Algorithms 18 00744 g001

2.1. Dataset Collection

The first stage involved collecting data for further analysis. The initial dataset was drawn from open sources, specifically scientific studies [39,40,41,42,43,44,45,46,47]. During the search and data collection process, the most significant formulation and technological factors from a materials science perspective were identified. The key formulation factors primarily include the type of aluminosilicate component and the type of alkaline activator, which handle the formation of the geopolymer matrix. The presence of additives, such as fiber, is also an important formulation factor. Curing temperature is also a key technological factor. Selecting the optimal curing temperature allows for the activation of the dissolution of the aluminosilicate precursor and, consequently, the acceleration of geopolymerization reactions.
The collected dataset represents information on geopolymer concretes for further processing and ease of analysis, converted into a file with the extension (.xlsx). The file contains 204 rows and the following columns, which serve as features for machine learning models:
x1—blast-furnace granulated slag (kg/m3);
x2—NaOH molarity;
x3—NaOH content in the alkaline activator (%);
x4—Na2SiO3 content in the alkaline activator (%);
x5—fiber type;
x6—fiber dosage (%);
x7—curing temperature (°C).
The output (predicted) characteristic is the variable y—compressive strength (MPa).
Each of the characteristics from a compositional standpoint influences strength as follows.
Factor x1—blast-furnace granulated slag is the main binder aluminosilicate component and ensures the synthesis of the geopolymer structure.
Factors x2, x3, and x4 determine the composition of the alkaline activator. The alkaline activator serves as a medium for the dissolution of silicon and aluminum oxides and the synthesis of the geopolymer structure.
Factors x5 and x6—fibers—are a dispersed reinforcing additive and are introduced into the geopolymer composite to improve its strength properties. The effect of dispersed reinforcement can vary significantly depending on the fiber type and dosage.
Factor x7—curing temperature—affects the rate of geopolymer reactions. Increasing the temperature accelerates the curing process, which improves the physical and mechanical properties of geopolymer composites.
Thus, the factors described are key, and their optimal combination determines the future strength of geopolymer concrete.

2.2. Exploratory Data Analysis (EDA)

To understand the relationships within the data and subsequently correctly build artificial intelligence techniques, it is necessary to conduct “Exploratory Data Analysis”. An initial review of the data revealed that the data types of all values correspond to the logic of the subject domain, the measured values are represented numerically, and there are no gaps.
A description of the main statistics is presented in Table 1. The number of rows in the dataset under study is 204. The variable x5 is categorical, so only the “number” characteristic is presented for it.
A data visualization using the pairplot function from the seaborn library is shown in Figure 2. This function, by default, produces a grid of axes, with each numerical variable in the data allocated to the y-axes within a single row and the x-axes within a single column. The diagonal of the graph displays the distribution of values associated with each individual variable.
The graph displays the relationship between two variables. No atypical values (anomalies or outliers) that could subsequently influence the results of machine learning models were detected.
Figure 3a shows a correlation matrix, demonstrating the degree of relationship between different variables. Figure 3b shows the distance correlation (dCor). This matrix reflects a relationship of any shape.
Variable x1—granulated blast furnace slag (kg/m3)—has the greatest impact on the output variable. Granulated blast furnace slag is the main component of geopolymer concrete and acts as a binder. Variable x6—fiber dosage (%)—has the least impact on the output variable. The inclusion of dispersed reinforcing fiber in geopolymer concrete is an additional formulation decision, and depending on its dosage, compressive strength may show a slight increase. However, fiber in composite building materials is most actively involved in bending and tensile loads and demonstrates more significant increases in bending and tensile strength than in compressive strength [48,49,50,51].
Analyzing the matrices in Figure 3a,b, we can conclude that a nonlinear relationship exists in pairs where the value in the dCor matrix (threshold of 0.5) is higher than in the Pearson matrix (threshold of 0.3). This is observed for the pairs x1/x7 and x7/y.

2.3. Selecting AI Models

Seven ML techniques were selected for predicting the compressive strength of geopolymer concrete in this study: from simple ones, such as linear regression, to more complex and modern methods, including the TabPFNv2 generative transformer model. The implementation was performed in Python (v. 3.12.12) using the following libraries: numpy 2.0.2, pandas 2.2.2, seaborn 0.13.2, matplotlib 3.10.0, catboost 1.2.8, optuna 4.5.0, sklearn 1.6.1, torch 2.7.1 + cu126, tabpfn_extensions 0.1.4, joblib 1.5.1, tabpfn = 2.2.1, h2o: 3.46.0.7, tabpfn: 2.1.3, tabpfnp-extension: 0.1.6, OS Linux-6.6.105.
Figure 4 shows the hyperparameter selection and model evaluation with Optuna and three-box cross-validation.
The main dataset was divided into an 80/20 split for training and testing purposes. To select the optimal model hyperparameters, a three-fold cross-validation procedure was used on the training set using the Optuna optimization algorithm, with the data from each fold split 2/3 for training and 1/3 for validation. The final model quality assessment was performed on an independent test set (20% of the original data), minimizing the risk of information leakage and ensuring the objectivity of the results.

2.4. Implementation and Training of AI Models

2.4.1. Linear Regression

The initial model selected was Linear Regression. This statistical methodology is designed to model the correlation between a dependent variable and one or more independent variables, predicated on the assumption of a linear relationship. This method is often used as a baseline for subsequent comparison with other, more complex models.
The linear regression model can be represented as
f x = β 0 + j = 1 p X j   β j ,
where β j are the model coefficients (weights);
β 0 is the mixing coefficient.
This model is sensitive to data dimensionality and does not handle categorical features. Therefore, before feeding the intelligent model, the data was preprocessed using the following:
-
A class from the scikit-learn StandardScaler library, which converts the data to a standard normal distribution; this approach standardizes features by subtracting the mean and dividing by the standard deviation;
-
One-Hot Encoding, a method for transforming categorical data.
One-Hot Encoding was implemented without extracting a baseline. According to this principle, all 8 fiber types (feature categories x5) were converted into 8 separate binary features. This approach differs from skip coding, which creates only N-1 features (N is the number of categories) and treats one fiber type as the baseline.
The Optuna library was used to select optimal hyperparameters. This hyperparameter optimization method is based on Bayesian search using the TPE (Tree-structured Parzen Estimator) algorithm [52].
During training of this model, the following equation was obtained:
y = 1.3969 + 0.1090   x 1 + 2.2709   x 2 + 0.4995   x 3 0.4995   x 4 + 0.8302   x 6 0.6258   x 7 10.1559   x 5 _ 1 + 19.6404   x 5 _ 2 + 25.952   x 5 _ 3 + 4.539   x 5 _ 4 29.731   x 5 _ 5 7.915   x 5 _ 6 28.373   x 5 _ 7 + 26.044   x 5 _ 8
where x5_1—is basalt fiber content; x5_2—ultra-high-molecular-weight polyethylene fiber; x 5 _ 3 —polyvinyl alcohol fiber; x 5 _ 4 —ceramic fiber; x 5 _ 5 —polypropylene fiber; x 5 _ 6 —polyethylene fiber; x 5 _ 7 —sisal fiber; and x 5 _ 8 —steel fiber.

2.4.2. CatBoost

The next model chosen was the gradient boosting model CatBoost [53]. This method is designed for efficient processing of categorical features and robust performance on heterogeneous data. Since the dataset has a categorical feature (x5), this should presumably improve forecasting quality.
Since we initially had a small dataset, special attention should be paid to optimizing the model’s hyperparameters. To optimize the hyperparameters, the Optuna library was used, similar to the first model. Optimization was performed using the following parameters: number of iterations (500–2000), learning rate (0.01–0.3), tree depth (4–10), L2 leaf regularization (1–10), and feature share (rsm, 0.5–1.0). For methods based on a tree structure, the trees should not be too deep, and measures against overfitting must be taken. During the training process, a categorical feature was taken into account, and an early stopping principle (stopping if there is no significant improvement within 50 iterations) was applied to prevent overfitting. Figure 5 shows the model training graph, with iterations plotted on the x-axis and the error plotted on the y-axis. It is worth noting the stability of the training, with the error steadily declining.

2.4.3. Random Forest

Unlike gradient boosting, Random Forest builds trees in parallel using a bagging method, which helps reduce variance and improve the stability of predictions.
The search for optimal values was conducted using key parameters: number of trees (100–500), maximum tree depth (5–20), minimum number of features for splitting a node (2–20), minimum number of features per leaf (1–10), and feature selection strategy at each split (max_features: sqrt, log2, None).

2.4.4. KNN

The next model chosen was the efficient nonparametric k-nearest neighbors (KNN) model, based on the principle of feature similarity in the feature space.
Before training, categorical features were encoded using One-Hot Encoding, and continuous features were standardized using Standard Scaler, as this method is sensitive to feature scale. Euclidean, Manhattan, and Minkowski distance metrics were considered, the latter with optimization of the p parameter. The Optuna library was used to tune the hyperparameters—the number of neighbors (k in the range 1–50), the weighting strategy (uniform, distance), the neighbor search algorithm, and the distance metric.

2.4.5. AutoML

Automated approaches such as AutoML are currently undergoing rapid development. These technologies allow for the automation of the creation, optimization, and deployment of machine learning models. The current investigation utilized an automated ML framework developed to build, optimize, and evaluate diverse models, such as gradient boosting, linear models, ensembles, and random forests [54]. It implements a full machine learning pipeline, including data preprocessing, hyperparameter selection, cross-validation, and ensemble building [55].
For this study, AutoML was run with a 60-s time limit to ensure comparability with other models and efficient use of computational resources. Cross-validation was performed using the 3-Fold CV scheme using scikit-learn, with the data converted to H2OFrame format at each fold for compatibility with the framework. Categorical features were explicitly transformed into factor variables using the asfactor() method, allowing H2O to correctly process them across various algorithms. Ultimately, the H2O GBM (Gradient Boosting Machine) model was selected to achieve the best results.

2.4.6. MLP

A standard deep learning architecture, the fully connected multilayer perceptron, was used as the foundational neural network method, and it is capable of modeling intricate nonlinear relationships between variables. This method is particularly effective when assuming the presence of hidden patterns and interrelations between variables.
Before training, all features were standardized using StandardScaler, and categorical variables were standardized using One-Hot Encoding. The hyperparameter space included the number of hidden layers (1–5), the number of neurons in each layer (10–200), activation functions (ReLU, tanh, logistic), optimizer type (Adam, LBFGS), L2 regularization coefficient ( 10 5   10 1 ), learning rate type (constant, adaptive), and the initial learning rate ( 10 5   10 1 ) for Adam. Early stopping was also used to prevent overfitting.

2.4.7. TabPFN v2

The TabPFN v2 method, a generative transformer specifically designed for solving regression problems on tabular data, was implemented in the study as the most modern and promising model. Unlike traditional machine learning algorithms, this approach does not require manual hyperparameter selection or retraining on a specific dataset. Instead, it uses a pre-trained model trained on millions of synthetic tasks and employs an attention mechanism to generate predictions in a few-shot learning mode.
For integration into the experimental pipeline, the AutoTabPFNRegressor wrapper, which implements post hoc ensembling, was used. This wrapper generates an ensemble of predictions by averaging the results of multiple model runs, improving stability and accuracy without requiring user intervention in parameter tuning.
Unlike traditional algorithms, TabPFN has no external hyperparameters to optimize. All key parameters, including the model architecture, weights, and attention strategy, are fixed in the pre-trained model. Automatic ensemble management and selection of the best predictions are implemented within AutoTabPFNRegressor, making the use of external hyperparameter optimization methods such as Optuna redundant.
Model limitations:
-
Supports up to 50,000 data rows (a technical limitation of the Transformer architecture);
-
Uses a GPU (Graphics Processing Unit) if CUDA (Compute Unified Device Architecture) support is available; otherwise, a CPU (Central Processing Unit) is used.
The model was trained on synthetic data, which may not cover all the specifics of a particular subject area. If the real data differs significantly from the synthetic training data, further training may be required. When applied to construction materials data, which has a unique internal structure, distribution estimation and in-depth analysis are required.

2.5. Evaluation of Quality Metrics

When implementing machine learning models, it is important to evaluate them using various metrics to gain a comprehensive understanding of the algorithm’s predictive capabilities. In this study, the following metrics were evaluated:
M A E = 1 n i = 1 n y i y ^ i
R M S E = 1 n i = 1 n ( y i y ^ i ) 2
M A P E = 1 n i = 1 n y i y ^ i y ^ i × 100
R 2 = 1 i = 1 n y i y ^ i 2 i = 1 n y i y ¯ 2

2.6. Formation of Recommendations for Practical Use

After obtaining stable training results and quality metrics that satisfy the subject area, the corresponding models can be implemented in real systems in the field of predictive modeling of building materials properties.

3. Results and Discussion

The following values were obtained during hyperparameter selection using Optuna (Table 2).
Figure 6 shows scatter plots obtained during model testing, with the x-axis representing actual values and the y-axis representing values predicted by the linear regression model. Points lying on a straight line (y = x) indicate accurate forecasting. Deviations from this line indicate poor forecasting. On the residual graph, the points are distributed randomly, without clusters or trends, and no “fan” is observed.
The metrics obtained from the model validation on the test set are provided in Table 3.
Figure 7 visualizes the data in Table 3. Within the MAE, RMSE and MAPE metrics, the models are distributed in ascending order of error, and R2 in descending order.
In this study, cross-validation was used to evaluate the robustness of the models to different subsets. Figure 8 shows the results of the coefficient of determination of the models for each fold (this graph is not available for the AutoML model, as the architecture requires the output of the average value of this parameter without displaying it across folds). The largest deviations are observed for the LR (±0.0192) and MLP (±0.0195) models. Overall, stable training is observed on each fold, indicating that the models generalize well to the data and are not susceptible to overfitting, the parameters are optimally selected, and the analyzed data is representative.
Significant differences in the accuracy and robustness of the resulting solutions were revealed by a comparative analysis of seven machine learning models applied to the problem of predicting the properties of geopolymer concrete. The observed quality hierarchy in Figure 6 allows us to formulate the following generalizations:
  • The basic linear regression model demonstrated satisfactory quality (R2 = 0.8977); however, this value is significantly inferior to more complex methods across all key metrics. High errors indicate the presence of nonlinear relationships between features not captured by the linear model.
  • The nonparametric KNN method demonstrated high prediction accuracy (MAPE = 1.62%) due to well-chosen parameters and the small dimensionality of the original dataset. In this study, the optimal number of neighbors and a correct distance metric made it possible to solve the problem using the closest data.
  • Ensemble and boosted methods (CatBoost, Random Forest, AutoML) showed comparable and high results (R2 = 0.9995…0.9981). H2O AutoML delivered stability and high quality without manual parameter selection, confirming the effectiveness of automated machine learning.
  • The neural network model (MLP) demonstrated R2 = 0.9993, but the average absolute percentage error reached 2.70%.
  • The generatively trained transformer (TabPFN) demonstrated good results across all metrics—(MAE = 0.46, RMSE = 0.64, MAPE = 1.39%, R2 = 0.9996). TabPFN does not require hyperparameter selection and ensures maximum accuracy with minimal computational costs during the retraining stage, making it a promising tool for practical application. This method is quite new in the line of intelligent models, which allows it to use modern approaches and the advantages of previously studied models.
Since the original dataset was relatively small, it was decided to expand it by 50 points to test the generalization capabilities of the models. The new data was also collected from open sources and differs in some statistical characteristics from the training set. This difference will allow for a more objective assessment of the generalization ability of the models.
The CatBoost algorithm was chosen for predicting robustness. This method is sensitive to changes in distributions and allows one to evaluate the impact of each characteristic on the final forecast. Table 4 shows the resulting metrics, and Figure 9 shows the Actual vs. Predicted plot. Analysis of the metrics and a visual assessment of Figure 9 suggest good generalization ability of the model.
The reliability of the metrics was assessed using the bootstrap method with 1000 iterations. At each iteration, a sample was generated with replacement from the test set, and the metrics were recalculated. Confidence intervals for R2 [0.7995, 0.8863], RMSE [4.5892, 6.0124], and MAE [2.6347, 3.7185] demonstrate acceptable model stability.
For a more comprehensive interpretation of the model, the study is supplemented with a plot from the SHAP method. For the CB method, the contribution of each feature to the model prediction is shown (Figure 9). Feature x1 is the most significant for the model, based on the largest value along the x-axis. Feature x7 was also found to be significant. Understanding the influence of these variables on the final results allows us to conclude that when modeling compositions, it is worth focusing on the main aluminosilicate binder component and curing temperature.
Comparing the obtained results with similar studies in the field of predicting the mechanical properties of geopolymer concretes using machine learning methods, it can be noted that the obtained results are not inferior in terms of quality metrics to existing ones (Table 5).
Thus, in work [27], a dataset of 162 mixture composition variants was used to train and validate the artificial neural network model, the determination coefficient reached a value of 0.92868 at the testing stage. In our study, R2 = 0.9993, which confirms the ability of neural network models to capture complex dependencies in data describing the formation of strength characteristics of geopolymer concretes. In work [29], “gene expression program” (GEP) and “multi-expression programming” (MEP) methods were adapted to optimize the composition of geopolymer mixtures. The methods were implemented on a relatively small database, comprising results of 301 compressive strength tests and 96 tensile strength tests. The MAE metrics for the MEP and GEP-based techniques for compressive strength prediction were 5.09 MPa and 6.78 MPa. And for post-split tensile strength were 0.42 MPa and 0.51 MPa, respectively. The MAE metric values in this study for the best models ranged from 0.37 to 0.51 MPa.
Despite the small size of the dataset, it reflects the true relationships between factors, which are consistent with previous studies [27,28,29,30,39,40,41,42,43,44,45,46,47,55,56]. The obtained statistical characteristics and correlations affect the developed strength properties of geopolymer concrete.
For industrial applications of intelligent solutions, it is assumed that the trained model will adequately handle new data, i.e., be capable of generalization. However, in practice, changes in characteristics (e.g., changes in distribution or correlations) are often observed, which can lead to incorrect results. For correct application in real-world conditions, it is necessary to train the model on a representative and voluminous dataset that includes all influencing factors.
The strength properties of geopolymer concretes are largely determined by the properties of the main raw materials and their ratios. When designing geopolymer composites, ground granulated blast-furnace slag is a common choice as the aluminosilicate binder. The most popular alkaline activators are sodium hydroxide (NaOH) and sodium silicate (Na2SiO3). Proper selection of an aluminosilicate component with a high Si/Al ratio and the composition of the alkaline activator, including the molarity of NaOH and the ratio of the alkaline activators, enables the activation of geopolymer reactions and the production of a composite with superior strength properties. Curing temperature is the most important process parameter and influences the process and activity of geopolymerization reactions, which also impacts the strength properties. Dispersed reinforcement with various fiber types in rational dosages further improves the strength properties of geopolymer composites and ensures a more ductile, ductile failure mode compared to the brittle failure of unreinforced composites [39,40,41,42,43,44,45,46,47,48].
This study examines the x1–x7 characteristic set for geopolymer concrete. These characteristics are most significant for modeling the composition of this building material. If the developed models are to be applied to other building materials, it is worth evaluating the data distributions; if significant discrepancies are detected, further training of the models will be necessary. However, it is worth noting that training does not need to be performed from scratch, as the existing weights can provide a good starting point.
In addition to data limitations, technical challenges may arise. Integrating intelligent models requires computing power, the quantity and quality of which depend on the volume of data being processed. If real-time forecasting is required, the time delays that may occur during the operation of AI systems must be taken into account.
During the study, bottlenecks were identified that open up avenues for future work. The following actions offer opportunities to improve the developed intelligent models:
  • Dataset development:
    -
    Expanding the training data by integrating information from open sources and laboratory test data to improve the prediction accuracy and generalizability of the algorithms;
    -
    Expanding the list of features in the dataset to account for a greater number of factors influencing the strength properties of geopolymers;
    -
    Creating an open, expandable data source for use by interested parties.
  • Conducting additional analyses:
    -
    Examining additional plots from the SHAP method to accurately assess the impact of an expanded list of features on the strength properties of geopolymers;
    -
    Analyzing multicollinearity to fine-tune models sensitive to multicollinearity;
    -
    Adding complex residual analysis plots.
  • Tuning model parameters:
    -
    Using additional methods to optimize models and evaluate performance. 4. Adaptation of developed models:
    -
    Adaptation of intelligent models to predict other physical and mechanical properties of geopolymer concrete in addition to strength;
    -
    Expanding the capabilities of developed models for use in predictive modeling systems for the properties of building materials.

4. Conclusions

This paper presents the implementation of seven machine learning methods (Linear Regression, CatBoost, Random Forest, KNN, AutoML, MLP, TabPFN v2) for predicting the properties of geopolymer concrete, in particular compressive strength. To ensure stable training, the parameters for the models were selected using Optuna. A database containing 204 rows and the following input parameters was compiled from published literature: x1granulated blast furnace slag, kg/m3; x2NaOH molarity; x3NaOH content in the alkali activator, %; x4Na2SiO3 content in the alkali activator, %; x5—fiber type; x6—fiber dosage, %; x7—curing temperature, °C. The assessment of the developed models’ performance was conducted using diverse quality metrics, namely, MAE, RMSE, MAPE, and R2.
Key findings from the above study:
-
The best models in terms of all quality metrics used are the k-nearest neighbor model (MAE = 0.37, RMSE = 0.63, MAPE = 1.62%, R2 = 0.9996) and TabPFNv2 (MAE = 0.46, RMSE = 0.64, MAPE = 1.39%, R2 = 0.9996);
-
When computing resources are limited and interpretability is required, the CatBoost or Random Forest algorithms are recommended;
-
When a GPU and a small dataset are available, TabPFN is advisable;
-
When manual parameter tuning is not required, H2O AutoML is recommended.
Overall, this study makes a valuable contribution to the study and development of alternative methods for analyzing the strength properties of building materials.
Machine learning methods are essential and reliable tools for predictive modeling of building materials. The developed methods can be used as part of predictive systems in laboratory settings to reduce the time required to analyze the physical and mechanical properties of materials when changing specific chemical elements and process parameters.

Author Contributions

Conceptualization, I.R., S.A.S., E.M.S., A.C. and D.M.S.; methodology, S.A.S., E.M.S. and I.R.; software, G.O., I.R. and S.O.; validation, I.R., S.O., G.O., S.A.S., E.M.S. and D.M.S.; formal analysis, I.R. and A.C.; investigation, I.R., S.A.S., E.M.S., A.N.B., A.C. and D.M.S.; resources, I.R. and S.O.; data curation, I.R.; writing—original draft preparation, I.R., S.A.S., E.M.S. and A.N.B.; writing—review and editing, I.R., S.A.S., E.M.S. and A.N.B.; visualization, I.R., S.A.S., E.M.S., A.N.B. and A.C.; supervision, A.N.B.; project administration, A.N.B.; funding acquisition, E.M.S. All authors have read and agreed to the published version of the manuscript.

Funding

The research was supported by the grant of the Russian Science Foundation No. 25-29-00535, https://rscf.ru/en/project/25-29-00535/ (accessed on 31 October 2025).

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Acknowledgments

The authors would like to acknowledge the administration of Don State Technical University for their support and resources.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
KNNk-nearest neighbors
AutoMLAutomated Machine Learning
MLPMultilayer Perceptron
TabPFNTabular Prior-data Fitted Network
EDAExploratory Data Analysis

References

  1. Mookkandi, K.; Nath, M.K.; Dash, S.S.; Mishra, M.; Blange, R. A Robust Lightweight Vision Transformer for Classification of Crop Diseases. AgriEngineering 2025, 7, 268. [Google Scholar] [CrossRef]
  2. Rahman, W.; Hossain, M.M.; Hasan, M.M.; Iqbal, M.S.; Rahman, M.M.; Hasan, K.F.; Moni, M.A. Automated detection of harmful insects in agriculture: A smart framework leveraging IoT, machine learning, and blockchain. IEEE Trans. Artif. Intell. 2024, 5, 4787–4798. [Google Scholar] [CrossRef]
  3. Riaz, W.; Ullah, A.; Ji, J. Multi-Scale Attention Networks with Feature Refinement for Medical Item Classification in Intelligent Healthcare Systems. Sensors 2025, 25, 5305. [Google Scholar] [CrossRef] [PubMed]
  4. Ashraf, M.H.; Mehmood, M.N.; Ahmed, M.; Hussain, D.; Khan, J.; Jung, Y.; Zakariah, M.; AlSekait, D.M. HIRD-Net: An Explainable CNN-Based Framework with Attention Mechanism for Diabetic Retinopathy Diagnosis Using CLAHE-D-DoG Enhanced Fundus Images. Life 2025, 15, 1411. [Google Scholar] [CrossRef] [PubMed]
  5. Sun, X.; Sun, J.; Lu, X.; Dong, Q.; Zhang, L.; Wang, W.; Liu, J.; Ma, Q.; Wang, X.; Wei, D.; et al. Mapping Neurophysiological Subtypes of Major Depressive Disorder Using Normative Models of the Functional Connectome. Biol. Psychiatry 2023, 94, 936–947. [Google Scholar] [CrossRef]
  6. Lin, R.-H.; Kujabi, B.K.; Chuang, C.-L.; Lin, C.-S.; Chiu, C.-J. Application of Deep Learning to Construct Breast Cancer Diagnosis Model. Appl. Sci. 2022, 12, 1957. [Google Scholar] [CrossRef]
  7. Tlig, M.; Bouchouicha, M.; Sayadi, M.; Moreau, E. Fire Segmentation with an Optimized Weighted Image Fusion Method. Electronics 2024, 13, 3175. [Google Scholar] [CrossRef]
  8. Zou, R.; Xin, Z.; Liao, G.; Huang, P.; Wang, R.; Qiao, Y. A Fire Segmentation Method with Flame Detail Enhancement U-Net in Multispectral Remote Sensing Images Under Category Imbalance. Remote Sens. 2025, 17, 2175. [Google Scholar] [CrossRef]
  9. Hong, Z.; Tang, Z.; Pan, H.; Zhang, Y.; Zheng, Z.; Zhou, R.; Ma, Z.; Zhang, Y.; Han, Y.; Wang, J.; et al. Active fire detection using a novel convolutional neural network based on Himawari-8 satellite images. Front. Environ. Sci. 2022, 10, 794028. [Google Scholar] [CrossRef]
  10. Cheng, M.-Y.; Sholeh, M.N.; Kwek, A. Computer vision-based post-earthquake inspections for building safety assessment. J. Build. Eng. 2024, 94, 109909. [Google Scholar] [CrossRef]
  11. Manzhilevskaya, S.E. Implementation of Machine Learning Models in the Construction Site Organization Process. Mod. Trends Constr. Urban Territ. Plan. 2025, 4, 106–116. [Google Scholar] [CrossRef]
  12. Fang, Q.; Li, H.; Luo, X.; Ding, L.; Luo, H.; Rose, T.M.; An, W. Detecting non-hardhat-use by a deep learning method from far-field surveillance videos. Autom. Constr. 2018, 85, 1–9. [Google Scholar] [CrossRef]
  13. Lee, J.; Lee, S. Construction Site Safety Management: A Computer Vision and Deep Learning Approach. Sensors 2023, 23, 944. [Google Scholar] [CrossRef] [PubMed]
  14. Feng, J.; Hamilton-Fletcher, G.; Ballem, N.; Batavia, M.; Wang, Y.; Zhong, J.; Porfiri, M.; Rizzo, J.-R. Robust Computer-Vision based Construction Site Detection for Assistive-Technology Applications. arXiv 2025, arXiv:2503.04139. [Google Scholar] [CrossRef]
  15. Beskopylny, A.N.; Shcherban’, E.M.; Stel’makh, S.A.; Mailyan, L.R.; Meskhi, B.; Razveeva, I.; Kozhakin, A.; El’shaeva, D.; Beskopylny, N.; Onore, G. Detecting Cracks in Aerated Concrete Samples Using a Convolutional Neural Network. Appl. Sci. 2023, 13, 1904. [Google Scholar] [CrossRef]
  16. Tsay, R.-J. Application of Artificial Intelligence in Graphical Discrimination of Structural Cracks. Eng. Proc. 2025, 108, 4. [Google Scholar] [CrossRef]
  17. Lee, C.; Kim, D.; Kim, D. Optimizing Deep Learning-Based Crack Detection Using No-Reference Image Quality Assessment in a Mobile Tunnel Scanning System. Sensors 2025, 25, 5437. [Google Scholar] [CrossRef]
  18. Su, M.; Wan, J.; Zhou, Q.; Wang, R.; Xie, Y.; Peng, H. Utilizing pretrained convolutional neural networks for crack detection and geometric feature recognition in concrete surface images. J. Build. Eng. 2024, 98, 111386. [Google Scholar] [CrossRef]
  19. Kondratieva, T.N.; Tyurina, V.S.; Chepurnenko, A.S. Predicting the risk of early cracking in massive monolithic foundation slabs using artificial intelligence algorithms. Constr. Mater. Prod. 2025, 8, 6. [Google Scholar] [CrossRef]
  20. Borovkov, A.I.; Vafaeva, K.M.; Vatin, N.I.; Ponyaeva, I. Synergistic Integration of Digital Twins and Neural Networks for Advancing Optimization in the Construction Industry: A Comprehensive Review. Constr. Mater. Prod. 2024, 7, 7. [Google Scholar] [CrossRef]
  21. Elshamy, M.M.; Tiraturyan, A.N.; Uglova, E.V.; Elgendy, M.Z. Evaluation of Pavement Condition Deterioration Using Artificial Intelligence Models. Adv. Eng. Res. 2022, 22, 272–284. [Google Scholar] [CrossRef]
  22. Moein, M.M.; Saradar, A.; Rahmati, K.; Mousavinejad, S.H.G.; Bristow, J.; Aramali, V.; Karakouzian, M. Predictive models for concrete properties using machine learning and deep learning approaches: A review. J. Build. Eng. 2023, 63 Pt A, 105444. [Google Scholar] [CrossRef]
  23. Feng, D.-C.; Liu, Z.-T.; Wang, X.-D.; Chen, Y.; Chang, J.-Q.; Wei, D.-F.; Jiang, Z.-M. Machine learning-based compressive strength prediction for concrete: An adaptive boosting approach. Constr. Build. Mater. 2020, 230, 117000. [Google Scholar] [CrossRef]
  24. Luo, D.; Wang, K.; Wang, D.; Sharma, A.; Li, W. Artificial intelligence in the design, optimization, and performance prediction of concrete materials: A comprehensive review. Npj Mater. Sustain. 2025, 3, 14. [Google Scholar] [CrossRef]
  25. Beskopylny, A.N.; Stel’makh, S.A.; Shcherban’, E.M.; Razveeva, I.; Kozhakin, A.; Pembek, A.; Kondratieva, T.N.; Elshaeva, D.; Chernil’nik, A.; Beskopylny, N. Prediction of the Properties of Vibro-Centrifuged Variatropic Concrete in Aggressive Environments Using Machine Learning Methods. Buildings 2024, 14, 1198. [Google Scholar] [CrossRef]
  26. Beskopylny, A.N.; Stel’makh, S.A.; Shcherban’, E.M.; Mailyan, L.R.; Meskhi, B.; Razveeva, I.; Kozhakin, A.; Pembek, A.; Elshaeva, D.; Chernil’nik, A.; et al. Prediction of the Compressive Strength of Vibrocentrifuged Concrete Using Machine Learning Methods. Buildings 2024, 14, 377. [Google Scholar] [CrossRef]
  27. Khosravi, H.; Bahram, M. Prediction of geopolymer concrete compressive strength using artificial neural network and genetic algorithm. Results Eng. 2025, 27, 105537. [Google Scholar] [CrossRef]
  28. Parhi, S.K.; Patro, S.K. Prediction of compressive strength of geopolymer concrete using a hybrid ensemble of grey wolf optimized machine learning estimators. J. Build. Eng. 2023, 71, 106521. [Google Scholar] [CrossRef]
  29. Huang, B.; Bahrami, A.; Javed, M.F.; Azim, I.; Iqbal, M.A. Evolutionary Algorithms for Strength Prediction of Geopolymer Concrete. Buildings 2024, 14, 1347. [Google Scholar] [CrossRef]
  30. Verma, M. Prediction of compressive strength of geopolymer concrete using random forest machine and deep learning. Asian J. Civ. Eng. 2023, 24, 2659–2668. [Google Scholar] [CrossRef]
  31. Ahmed, H.U.; Abdalla, A.A.; Mohammed, A.S.; Mohammed, A.A.; Mosavi, A. Statistical Methods for Modeling the Compressive Strength of Geopolymer Mortar. Materials 2022, 15, 1868. [Google Scholar] [CrossRef] [PubMed]
  32. Abdellatief, M.; Elsafi, M.; Murali, G.; ElNemr, A. Comparative evaluation of hybrid machine learning models for predicting the strength of metakaolin-based geopolymer concrete enhanced with Gaussian noise augmentation. J. Build. Eng. 2025, 111, 113302. [Google Scholar] [CrossRef]
  33. Tipu, R.K.; Bansal, S.; Batra, V.; Suman; Patel, G.A. Ensemble machine learning models for predicting concrete compressive strength incorporating various sand types. Multiscale Multidiscip. Model. Exp. Des. 2025, 8, 222. [Google Scholar] [CrossRef]
  34. Al-Abdaly, N.M.; Seno, M.E.; Thwaini, M.A.; Imran, H.; Ostrowski, K.A.; Furtak, K. Advanced Ensemble Machine-Learning Models for Predicting Splitting Tensile Strength in Silica Fume-Modified Concrete. Buildings 2024, 14, 4054. [Google Scholar] [CrossRef]
  35. Shi, M.; Shen, W. Automatic Modeling for Concrete Compressive Strength Prediction Using Auto-Sklearn. Buildings 2022, 12, 1406. [Google Scholar] [CrossRef]
  36. Hariri-Ardebili, M.A.; Pourkamali-Anaraki, F. An Automated Machine Learning Engine with Inverse Analysis for Seismic Design of Dams. Water 2022, 14, 3898. [Google Scholar] [CrossRef]
  37. Wang, J.; Deng, J.; Li, S.; Du, W.; Zhang, Z.; Liu, X. Explainable Machine Learning for Multicomponent Concrete: Predictive Modeling and Feature Interaction Insights. Materials 2025, 18, 4456. [Google Scholar] [CrossRef]
  38. Mirzaei, S.; Mao, H.; Al-Nima, R.R.O.; Woo, W.L. Explainable AI Evaluation: A Top-Down Approach for Selecting Optimal Explanations for Black Box Models. Information 2024, 15, 4. [Google Scholar] [CrossRef]
  39. Amin, M.N.; Khan, K.; Ahmad, W.; Javed, M.F.; Qureshi, H.J.; Saleem, M.U.; Qadir, M.G.; Faraz, M.I. Compressive Strength Estimation of Geopolymer Composites Through Novel Computational Approaches. Polymers 2022, 14, 2128. [Google Scholar] [CrossRef]
  40. Luo, R.; Liu, R.; Qin, G.; Jiang, M.; Wu, Y.; Guo, Y. Study on High-Ductility Geopolymer Concrete: The Influence of Oven Heat Curing Conditions on Mechanical Properties and Microstructural Development. Materials 2024, 17, 4011. [Google Scholar] [CrossRef]
  41. Zubarev, K.P.; Shcherban’, E.M.; Stel’makh, S.A.; Beskopylny, A.N.; Elshaeva, D.; Chernil’nik, A.; Zakieva, N.I.; Pimenova, E.V.; Shilov, A.A. Structure and Properties Improvement by Recipe Factors of Geopolymer Basalt Fiber Reinforced Concrete for Building Enclosing Structures. Buildings 2024, 14, 743. [Google Scholar] [CrossRef]
  42. Dalğıç, A.; Yılmazer Polat, B. The Behavior of Ceramic Fiber Geopolymer Concrete under the Effect of High Temperature. Appl. Sci. 2024, 14, 1607. [Google Scholar] [CrossRef]
  43. Zhang, Y.; Chen, J.; Xia, J. Compressive Strength and Chloride Resistance of Slag/Metakaolin-Based Ultra-High-Performance Geopolymer Concrete. Materials 2023, 16, 181. [Google Scholar] [CrossRef] [PubMed]
  44. Thakur, G.; Singh, Y.; Singh, R.; Prakash, C.; Saxena, K.K.; Pramanik, A.; Basak, A.; Subramaniam, S. Development of GGBS-Based Geopolymer Concrete Incorporated with Polypropylene Fibers as Sustainable Materials. Sustainability 2022, 14, 10639. [Google Scholar] [CrossRef]
  45. Xiao, S.; Cai, Y.; Guo, Y.; Lin, J.; Liu, G.; Lan, X.; Song, Y. Experimental Study on Axial Compressive Performance of Polyvinyl Alcohol Fibers Reinforced Fly Ash—Slag Geopolymer Composites. Polymers 2022, 14, 142. [Google Scholar] [CrossRef]
  46. Ren, P.; Zhang, W.; Ye, X.; Liu, X. Mechanical Properties and Constitutive Model of Geopolymer Lightweight Aggregate Concrete. Buildings 2025, 15, 98. [Google Scholar] [CrossRef]
  47. Guo, J.; Shi, J.; Wang, L.; Huang, C.; Tao, X.; Li, C.; Chen, Z. Evaluation of Axial Compressive and Tensile Properties of PE/PVA Hybrid Fiber Reinforced Strain-Hardening Geopolymer Composites. Materials 2024, 17, 4356. [Google Scholar] [CrossRef]
  48. Klyuev, S.V.; Klyuev, A.V.; Ayubov, N.A.; Fediuk, R.S.; Levkina, E.V. Finite Element Design and Analysis of Sustainable Mono-Reinforced and Hybrid-Reinforced Fibergeopolymers. Adv. Eng. Res. 2025, 25, 171–185. [Google Scholar] [CrossRef]
  49. Antipas, I.R. Effect of Glass Fiber Reinforcement on the Mechanical Properties of Polyester Composites. Adv. Eng. Res. 2023, 23, 387–397. [Google Scholar] [CrossRef]
  50. Saprykina, T.K.; Zhadanov, V.I. Improving Theoretical Concepts of Composition Design of Dispersedly-Reinforced Concretes. Mod. Trends Constr. Urban Territ. Plan. 2024, 3, 66–73. [Google Scholar] [CrossRef]
  51. Stel’makh, S.A.; Shcherban’, E.M.; Beskopylny, A.N.; Mailyan, L.R.; Meskhi, B.; Razveeva, I.; Kozhakin, A.; Beskopylny, N. Prediction of Mechanical Properties of Highly Functional Lightweight Fiber-Reinforced Concrete Based on Deep Neural Network and Ensemble Regression Trees Methods. Materials 2022, 15, 6740. [Google Scholar] [CrossRef] [PubMed]
  52. Akiba, T.; Sano, S.; Yanase, T.; Ohta, T.; Koyama, M. Optuna: A Next-generation Hyperparameter Optimization Framework. arXiv 2019, arXiv:1907.10902. [Google Scholar] [CrossRef]
  53. Prokhorenkova, L.; Gusev, S.; Vorobev, A.; Dorogush, A.V.; Gulin, A. CatBoost: Unbiased boosting with categorical features. arXiv 2019, arXiv:1706.09516. [Google Scholar] [CrossRef]
  54. H2O AutoML: Automatic Machine Learning. Available online: https://docs.h2o.ai/h2o/latest-stable/h2o-docs/automl.html#h2o-automl-automatic-machine-learning (accessed on 2 November 2025).
  55. LeDell, E.; Poirier, S. H2O AutoML: Scalable Automatic Machine Learning. In Proceedings of the AutoML Conference, 2020. Virtual Workshop, USA, 18 July 2020; Available online: https://www.automl.org/wp-content/uploads/2020/07/AutoML_2020_paper_61.pdf (accessed on 2 November 2025).
  56. Mintsaev, M.S.; Murtazaev, S.-A.Y.; Salamanova, M.S.; Saidumov, M.S.; Murtazaev, I.S.-A. Features of the synthesis of construction geopolymer composites. Constr. Mater. Prod. 2025, 8, 6. [Google Scholar] [CrossRef]
Figure 2. Scatterplots.
Figure 2. Scatterplots.
Algorithms 18 00744 g002
Figure 3. Data Analysis: (a) Correlation Matrix; (b) Distance Correlation.
Figure 3. Data Analysis: (a) Correlation Matrix; (b) Distance Correlation.
Algorithms 18 00744 g003
Figure 4. Hyperparameter Selection and Model Evaluation Flowchart with Optuna and Three-Block Cross-Validation.
Figure 4. Hyperparameter Selection and Model Evaluation Flowchart with Optuna and Three-Block Cross-Validation.
Algorithms 18 00744 g004
Figure 5. CatBoost Training.
Figure 5. CatBoost Training.
Algorithms 18 00744 g005
Figure 6. Actual vs. Predicted and Residuals: (a,b) Linear Regression, (c,d) CatBoost, (e,f) Random Forest, (g,h) KNN, (i,j) AutoML, (k,l) MLP, (m,n) TabPFN v2.
Figure 6. Actual vs. Predicted and Residuals: (a,b) Linear Regression, (c,d) CatBoost, (e,f) Random Forest, (g,h) KNN, (i,j) AutoML, (k,l) MLP, (m,n) TabPFN v2.
Algorithms 18 00744 g006aAlgorithms 18 00744 g006bAlgorithms 18 00744 g006c
Figure 7. Quality Metrics: (a) MAE; (b) RMSE; (c) MAPE; (d) R2.
Figure 7. Quality Metrics: (a) MAE; (b) RMSE; (c) MAPE; (d) R2.
Algorithms 18 00744 g007aAlgorithms 18 00744 g007b
Figure 8. R2 by folds: (a) Linear Regression, (b) CatBoost, (c) Random Forest, (d) KNN, (e) MLP, (f) TabPFN v2.
Figure 8. R2 by folds: (a) Linear Regression, (b) CatBoost, (c) Random Forest, (d) KNN, (e) MLP, (f) TabPFN v2.
Algorithms 18 00744 g008
Figure 9. Review of the influence of each property on compressive strength.
Figure 9. Review of the influence of each property on compressive strength.
Algorithms 18 00744 g009
Table 1. Data Statistics Description.
Table 1. Data Statistics Description.
AttributeNumberAverageStdMin25%50%75%Max
x1204.0453.13192.2774.0390.0408.0563.0924.0
x2204.010.912.324.010.012.012.014.0
x3204.033.2211.6415.025.029.050.050.0
x4204.066.7811.6450.050.071.075.085.0
x5204.0-------
x6204.01.301.390.00.51.12.010.0
x7204.025.265.3420.025.025.025.060.0
y204.038.9232.373.417.823.449.1136.3
Table 2. Best Parameter Values for the Models.
Table 2. Best Parameter Values for the Models.
ModelBest Parameters/Meta-Model
1Linear Regression{‘model_type’: ‘ridge’, ‘alpha’: 0.7982803713496193}
2CatBoost{‘iterations’: 920, ‘learning_rate’: 0.18190538763505498, ‘depth’: 4, ‘l2_leaf_reg’: 7.478453558099906, ‘rsm’: 0.7537872840036406}
3Random Forest{‘n_estimators’: 352, ‘max_depth’: 14, ‘min_samples_split’: 2, ‘min_samples_leaf’: 1, ‘max_features’: None}
4KNN{‘n_neighbors’: 1 (or 3), ‘weights’: ‘uniform’, ‘metric’: ‘minkowski’, ‘algorithm’: ‘kd_tree’, ‘p’: 3}
5AutoML{‘ntrees’: 89, ‘max_depth’: 6, ‘min_rows’: 1, ‘learn_rate’: 0.1, ‘learn_rate_annealing’: 1.0, ‘sample_rate’: 0.5, ‘col_sample_rate’: 0.7, ‘distribution’: ‘gaussian’, ‘stopping_rounds’: 3, ‘stopping_tolerance’: 0.05, ‘histogram_type’: ‘UniformAdaptive’, ‘categorical_encoding’: ‘Enum’}
6MLP{‘n_layers’: ‘2’, ‘n_units_l0’: ‘96’,’ n_units_l1’: ‘55’, ‘activation’: ‘logistic’, ‘solver’: ‘lbfgs’, ‘alpha’: ‘0.008686426576670337’, ‘learning_rate’: ‘adaptive’}
Table 3. Test set model validation.
Table 3. Test set model validation.
Model TypeMAERMSEMAPE, %R2
1Linear Regression7.7510.1631.510.8977
2CatBoost0.931.395.350.9981
3Random Forest2.974.7010.970.9782
4KNN0.370.631.620.9996
5AutoML0.510.682.120.9995
6MLP0.560.822.700.9993
7TabPFN v20.460.641.390.9996
Table 4. Testing the model at new points.
Table 4. Testing the model at new points.
Model TypeMAERMSEMAPE, %R2
1CatBoost3.135.237.450.8480
Table 5. Comparison of results with similar studies.
Table 5. Comparison of results with similar studies.
No.ResearchNumber of SamplesMethodR2
1This study204LR, CB, RF, KNN, AutoML, MLP, TabPFN0.9996
2[27]162ANN0.9993
3[29]397“gene expression program” (GEP) and “multi-expression programming” (MEP) methods0.73–0.89
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Stel’makh, S.A.; Beskopylny, A.N.; Shcherban’, E.M.; Razveeva, I.; Oganesyan, S.; Shakhalieva, D.M.; Chernil’nik, A.; Onore, G. Compressive Strength of Geopolymer Concrete Prediction Using Machine Learning Methods. Algorithms 2025, 18, 744. https://doi.org/10.3390/a18120744

AMA Style

Stel’makh SA, Beskopylny AN, Shcherban’ EM, Razveeva I, Oganesyan S, Shakhalieva DM, Chernil’nik A, Onore G. Compressive Strength of Geopolymer Concrete Prediction Using Machine Learning Methods. Algorithms. 2025; 18(12):744. https://doi.org/10.3390/a18120744

Chicago/Turabian Style

Stel’makh, Sergey A., Alexey N. Beskopylny, Evgenii M. Shcherban’, Irina Razveeva, Samson Oganesyan, Diana M. Shakhalieva, Andrei Chernil’nik, and Gleb Onore. 2025. "Compressive Strength of Geopolymer Concrete Prediction Using Machine Learning Methods" Algorithms 18, no. 12: 744. https://doi.org/10.3390/a18120744

APA Style

Stel’makh, S. A., Beskopylny, A. N., Shcherban’, E. M., Razveeva, I., Oganesyan, S., Shakhalieva, D. M., Chernil’nik, A., & Onore, G. (2025). Compressive Strength of Geopolymer Concrete Prediction Using Machine Learning Methods. Algorithms, 18(12), 744. https://doi.org/10.3390/a18120744

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop