Next Article in Journal
The Coupling Relationship Between Street View Element Comfort Perception and Eye Movement Metrics and Its Sustainable Research
Previous Article in Journal
The Impact of Carbon Information Disclosure on Firm Value: The Mediating Role of Green M&A—Evidence from China
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Interpretable Machine Learning for Compressive Strength Prediction of Fly Ash-Based Geopolymer Concrete

1
Department of Civil Engineering, Istanbul University-Cerrahpasa, 34320 Istanbul, Türkiye
2
Department of Architecture, Mimar Sinan Fine Arts University, 34427 İstanbul, Türkiye
3
GameAbove College of Engineering and Technology, Eastern Michigan University, Ypsilanti, MI 48197, USA
4
College of IT Convergence, Gachon University, Seongnam 13120, Republic of Korea
*
Authors to whom correspondence should be addressed.
Sustainability 2026, 18(5), 2227; https://doi.org/10.3390/su18052227
Submission received: 18 January 2026 / Revised: 20 February 2026 / Accepted: 22 February 2026 / Published: 25 February 2026

Abstract

Fly ash-based geopolymer concrete (GPC) is a sustainable alternative to conventional cementitious materials; however, its compressive strength is governed by complex and highly correlated mixture parameters, making experimental optimization expensive and data-driven modeling challenging. While machine learning (ML) techniques have been widely applied to predict GPC strength, most studies prioritize predictive accuracy without explicitly addressing multicollinearity among input variables, which can distort feature importance, reduce model stability, and limit engineering interpretability. This study proposes a multicollinearity-integrated and interpretable ML framework that systematically embeds correlation diagnostics and structured feature screening within the modeling pipeline rather than treating interpretability as a post-processing step. Multiple conventional and ensemble learning algorithms were comparatively evaluated using cross-validation to ensure generalization robustness. The proposed framework achieved a maximum coefficient of determination (R2) of 0.96 with low prediction error, outperforming baseline regression models while demonstrating improved stability under correlated input conditions. Unlike existing studies that rely solely on black-box optimization, the integrated interpretability analysis revealed physically consistent dominance of curing temperature, alkali content, and water-related parameters in governing strength development. By explicitly coupling predictive performance with multicollinearity mitigation and engineering-oriented interpretability, this work advances beyond accuracy-driven ML applications and provides a robust and transparent decision-support tool for sustainable geopolymer mix design.

1. Introduction

Concrete is the most widely used construction material worldwide, owing to its versatility, durability, and relatively low cost. However, the dominant use of ordinary Portland cement (OPC) as the primary binder in concrete has raised serious environmental concerns. Cement production is an energy-intensive process. It is responsible for approximately 5–8% of global anthropogenic CO2 emissions, mainly due to clinker production and fossil fuel consumption. As global infrastructure demand continues to increase, reducing the environmental footprint of cement-based materials has become a critical challenge for sustainable development in the construction sector [1].
In response to these concerns, geopolymer concrete has emerged as a promising alternative to OPC-based systems. Geopolymers are inorganic binders formed through the alkaline activation of aluminosilicate-rich materials, such as fly ash, ground granulated blast furnace slag (GGBFS) and other industrial by-products. Unlike OPC, geopolymer binders do not rely on high-temperature clink erization, enabling substantial reductions in CO2 emissions while simultaneously valorizing industrial waste streams [2]. From a sustainability perspective, geopolymer technology aligns with circular economy principles by transforming by-products into value-added construction materials and reducing reliance on virgin raw resources.
Geopolymer concrete has gained considerable attention due to its environmental benefits and mechanical performance. In addition to its behavior under static loading, geo-polymer concrete also exhibits notable performance under dynamic loading conditions. Studies have shown that it provides adequate resistance to impact and vibrations, effectively absorbing applied energy. Therefore, investigating its strength and behavior under dynamic conditions is particularly important for both structural and industrial applications.
Among various precursor materials, fly ash has received particular attention due to its wide availability and favorable chemical composition. Large quantities of fly ash are generated annually by coal-fired power plants. Only a limited fraction is effectively utilized in construction applications. When activated under alkaline conditions, fly ash can form a dense aluminosilicate network. It results in geopolymer concretes with mechanical properties comparable to, and in some cases exceeding those of, conventional OPC concrete [3,4,5]. The incorporation of GGBFS has also been shown to enhance early-age strength development, especially under ambient curing conditions, further improving the practical applicability of geopolymer systems [6].
Despite these advantages, the mechanical performance of geopolymer concrete is governed by a complex interaction of multiple parameters. These include precursor composition, fineness of fly ash, alkaline activator type and concentration, activator to binder ratio, curing temperature and duration, and the presence of chemical admixtures [7]. The nonlinear and interdependent nature of these variables makes it challenging to develop generalized mixed design rules using traditional empirical approaches. As a result, reported findings in the literature often exhibit significant variability, limiting the transferability of experimental results to broader engineering practice.
To address this complexity, machine learning (ML) techniques have increasingly been adopted to model and predict the compressive strength of geopolymer concrete. ML models are capable of capturing nonlinear relationships and interactions among multiple variables and have demonstrated high predictive accuracy. However, many ML-based approaches suffer from limited interpretability, particularly when input variables exhibit strong linear interdependencies. In such cases, model predictions may be accurate, yet the extracted feature importance measures become unreliable, effectively turning the models into black-box predictors with limited physical insight.
From a sustainability-oriented engineering perspective, interpretability is as important as predictive accuracy. Understanding which material parameters most strongly influence compressive strength supports more efficient mix design, reduces unnecessary material usage and facilitates the rational optimization of geopolymer systems with lower environmental impact. Therefore, there is a clear need for modeling frameworks that balance predictive performance with transparency and robustness, especially when dealing with high-dimensional datasets characterized by multicollinearity.
Recent studies have extensively explored machine learning (ML) techniques for predicting the compressive strength (CS) of fly ash-based geopolymer concrete (GPC), highlighting the nonlinear dependency between mixture constituents, curing conditions, and strength development. For instance, Ansari et al. compared conventional ML models (Linear Regression, Artificial Neural Networks) with ensemble techniques such as AdaBoost and reported a substantial improvement in predictive performance using ensemble learning (R2 = 0.944), demonstrating the superiority of boosting-based approaches over traditional regression models [8]. Similarly, Gupta et al. showed that Random Forest Regression outperformed KNN and Linear Regression in predicting GPC compressive strength, further confirming the robustness of ensemble-based frameworks for highly nonlinear concrete datasets [9].
More advanced ensemble and boosting strategies have also been proposed. Ahmad et al. employed ANN, AdaBoost, and boosting techniques, concluding that boosting algorithms achieved the highest accuracy (R2 ≈ 0.96), and incorporated k-fold validation to improve generalization reliability [10]. Dash et al. introduced a firefly-optimized hybrid ensemble model that integrates multiple base learners through stacking. Their study combined metaheuristic hyperparameter tuning with Sobol and FAST global sensitivity analyses, demonstrating that curing temperature and water content dominate strength variability. These studies collectively indicate a strong trend toward hybrid and optimized ensemble frameworks to maximize prediction accuracy [11].
Beyond predictive accuracy, recent research has begun addressing model interpretability. Hu et al. applied SHapley Additive exPlanations (SHAP) to interpret eXtreme Gradient Boosting predictions for FRP-confined concrete columns, identifying parameter importance and nonlinear functional relationships [12]. Likewise, Nguyen et al. proposed a structured three-stage framework that transitions from ML prediction (LR, DNN, ResNet) to global sensitivity analysis and finally to empirical formula derivation, thereby attempting to bridge the gap between black-box ML models and practical engineering applicability [13].
Despite these significant advances, two critical issues remain insufficiently addressed in the current literature. First, while ensemble and hybrid models consistently improve predictive accuracy, most studies prioritize performance metrics (R2, RMSE, MAE) without systematically addressing multicollinearity among input variables—a common characteristic of geopolymer concrete datasets due to correlated binder compositions, activator ratios, and curing parameters. Second, interpretability is often treated as a post-processing step (e.g., SHAP or sensitivity analysis) rather than being structurally integrated into the modeling framework to enhance both robustness and engineering transparency.
Therefore, the existing body of research reveals a methodological gap between high-accuracy ensemble modeling and structured interpretability with explicit treatment of correlated input features. The present study aims to address this gap by integrating multicollinearity mitigation strategies within the ML pipeline and systematically evaluating their impact on both predictive performance and interpretability. In doing so, this work moves beyond incremental accuracy improvement and contributes to a more reliable and engineering-oriented modeling framework for geopolymer concrete strength prediction.
The novelty of this study lies in the development of an interpretable and dependency-aware machine learning framework for estimating the compressive strength of sustainable geopolymer concrete. Unlike conventional ML approaches that prioritize prediction accuracy alone, this work introduces a multi-objective NSGA-II–based feature selection strategy that explicitly accounts for linear interdependencies among input variables. By simultaneously optimizing predictive accuracy and feature independence, the proposed framework enhances both model reliability and interpretability.
Using an extensive dataset comprising multiple mix designs and curing parameters, the study systematically evaluates twelve hyperparameter-tuned machine learning models trained on optimized feature subsets. The results not only confirm the dominant influence of industrial by-products such as GGBFS and key activator-related parameters but also demonstrate that high predictive accuracy can be maintained while significantly reducing input redundancy. From a sustainability standpoint, the proposed methodology supports more informed and resource-efficient geopolymer mix design, reinforcing the role of geopolymer concrete as a viable low-carbon alternative to conventional cement-based materials.

2. Materials and Methods

This study aims to enhance the interpretability and robustness of machine-learning models in the domain of compressive strength prediction of geopolymer concrete by reducing linear interdependence among input features while preserving predictive accuracy.
To ensure a fair and reproducible evaluation framework, the dataset was first divided into training and testing subsets using a single 80/20 holdout split. The same partition was consistently preserved across both the baseline models and the feature-selected models to eliminate performance variability arising from data partitioning differences.
Given the high degree of linear interdependence among geopolymer constituents, a multi-objective feature selection strategy based on the NSGA-II algorithm was implemented. Unlike conventional single-objective selection approaches, the proposed framework simultaneously considers four criteria: predictive performance, multicollinearity reduction, correlation control, and feature subset compactness. These objectives were combined using weighted aggregation to balance model accuracy with statistical stability.
The weight coefficients (0.15, 0.30, 0.40, 0.15) were determined empirically through pre-liminary exploratory experiments to achieve a stable trade-off between predictive performance and multicollinearity mitigation. The selected configuration provided consistent convergence behavior across independent runs. The aim was not to impose theoretical optimality, but to establish a practically balanced configuration suited to the characteristics of the dataset.
Feature selection was performed exclusively on the training set to prevent information leakage. After subset identification, twelve machine learning algorithms were trained using identical training–testing partitions to ensure a valid comparison between baseline and optimized feature configurations.
Hyperparameter optimization was conducted using Tree-structured Parzen Estimator (TPE) search with 50 evaluation iterations per model. This budget was selected as a compromise between computational feasibility and convergence stability, and preliminary runs indicated diminishing performance gains beyond this threshold.
Although the dataset consists of 274 experimental samples, it represents a consolidated compilation of geopolymer concrete mixtures with varying binder compositions, activator ratios, and curing conditions. Therefore, the results should be interpreted within the scope of this dataset, and broader generalization requires further cross-dataset validation.

2.1. Data Collection

To achieve robust predictions and reliable feature interpretation, it is essential to curate and pre-process datasets carefully and ensure their credibility. A key dataset in this field is provided with a manuscript titled “Effect of Alkaline Activators and Other Factors on the Properties of Geopolymer Concrete Using Industrial Waste”, developed with a gene expression programming (GEP) approach by Pham and Nguyen [14]. It is publicly available through the Mendeley Data Repository [15]. The dataset focuses on mixtures incorporating industrial by-products such as fly ash and ground granulated blast furnace slag, offering a solid basis for exploring geopolymer concrete properties through machine learning techniques.

2.2. Exploratory Data Analysis

The dataset contains 19 independent and 1 dependent variable(s) (columns) and 274 data points (rows). To provide a comprehensive overview of the dataset employed in this study, the descriptive statistics of the input variables are presented in Table 1.
Fly ash is a byproduct of coal combustion in thermal power plants. It is a fine powder mainly composed of silica, alumina, and iron oxides. With its pozzolanic properties, it reacts with calcium hydroxide in the presence of water to form cementitious compounds, thereby improving workability, reducing heat of hydration, enhancing long-term strength, and supporting sustainability by recycling industrial waste. Ground granulated blast furnace slag (GGBFS) is generated as a by-product of the iron and steel industry. It also exhibits cementitious and pozzolanic characteristics that enhance durability and mechanical performance over time. The incorporation of these materials in concrete is strongly aligned with sustainable construction practices and the objectives of green development [16].
In the present dataset, Concentration (M) NaOH denotes the molar concentration of the sodium hydroxide solution used as the alkaline activator. This parameter describes the amount of NaOH dissolved per liter of solution and is used to characterize the alkalinity level of the activating medium. The concentration of NaOH is known to play an important role in controlling the dissolution of aluminosilicate precursors and the overall reaction environment during geopolymer formation. Defining concentration in terms of molarity is a common and well-established practice in geopolymer and alkali-activated material studies, and the same convention has been adopted in this dataset.
Sodium silicate solution (Na2SiO3) is obtained by combining alkaline compounds with silica sources and water. Owing to its soluble silica content, it plays a key role in the geopolymerization process, stabilizing the structure and supporting the formation of three-dimensional silicate–aluminate networks [17]. Water within this solution reduces viscosity and improves particle dispersion, thereby enhancing the workability of fresh mixtures. The SiO2/Na2O molar ratio is particularly critical for both strength and durability [18]. Sodium hydroxide (NaOH), which is prepared by dissolving solid NaOH in water, acts as a strong alkaline activator that breaks down aluminosilicate structures, releasing reactive ions to accelerate polymerization and strength development [19]. However, its concentration must be carefully controlled to avoid negative effects such as thermal cracking or reduced workability. Sodium oxide (Na2O) further increases sodium ion availability, strengthening silicate–aluminate gels and promoting the dissolution of precursors, though excessive contents may harm durability [20].
Silicon dioxide (SiO2) provides the main source of reactive silica, enabling the formation of sodium aluminosilicate hydrate (N-A-S-H) gels that contribute to strength, durability, and low permeability [21]. Geopolymer systems also reduce CO2 emissions by replacing Portland cement with recycled pozzolanic materials and alkaline activators [21], while dense gel networks improve resistance against carbonation and aggressive ions, contributing to sustainability. Water primarily acts as a reaction medium [17], while its total amount, including any additional water, must be optimized to balance workability and strength. The molar concentration and dry mass of NaOH [22] provide precise control of activator-to-binder interactions. Superplasticizers enhance flow without excess water [23], though their efficiency depends on compatibility with alkaline conditions.
Curing conditions are equally critical. Initial curing time and temperature strongly affect early geopolymer gel formation and strength gain, with moderate heat curing between 60 and 80 °C being particularly beneficial. A short resting period before heat application can reduce thermal stresses. Final curing temperature [3] governs matrix densification, durability, and efflorescence prevention. Additionally, sodium silicate concentration [5] influences network formation, setting behavior, and compressive strength through its control of the Si/Na ratio.
The variables included in the dataset were chosen mainly for their known effects on the mechanical strength of geopolymer concrete. At the same time, materials that contribute to sustainability, such as industrial by-products like fly ash and GGBFS, were also considered when selecting the features.

2.3. The Experimental Process

Following the collection of the data, a 4-stage experiment was conducted to fulfill the research aim and objectives (Figure 1). Appendix A provides a more comprehensive step-by-step tree diagram of all experimental process (pipeline). The stages of the process were Initialization, Feature Selection with Genetic Algorithm, Performance Assessment of ML Models, Visualization and Verification. This section provides the details of each stage.

2.3.1. Stage 1: Initialization

The process starts with loading the dataset from a CSV file. A check is made following this to detect if there is any missing data in the dataset. In the case of missing data, rows that contain null values would be removed. In fact, in our dataset, there were no rows with missing data.
In the following phase, the Standard Scaler (z-score normalization) X_scaled = (X − μ)/σ, where μ is the mean and σ is the standard deviation, is applied to the dataset. In many machine learning tasks, features often have different scales, which can negatively affect model performance. The Standard Scaler addresses this by standardizing each feature to have a mean of zero and a standard deviation of one. This ensures that all variables contribute equally to the learning process, without being dominated by those with larger magnitudes. By preserving the structure of the data while balancing feature scales, Standard Scaler facilitates more stable and efficient model training. This stage ensures data quality and normalizes features for consistent algorithm performance.
The following phase involves parameter configuration. A set of parameters for the feature selection process is configured in this phase. First, the minimum number of variables to retain (default: 5) is provided by the user. Following this, parameters related to the Genetic Algorithm Population size (default: 60), Number of generations (default: 40), Crossover probability (default: 0.7), and Mutation probability (default: 0.2) are set. Finally, the weights for each objective, VIF, RMSE, R2, and Total Multiple R (MultiR) are provided by the user. These weights are used for transforming the 4-objective optimization problem into a single objective optimization problem. Weights also determine the focus of the feature selection, either by prioritizing one of the objectives (i.e., minimizing VIF) or by giving equal importance to each objective. In our study, the objective weights were provided as twelve weight combinations prioritizing both maximizing R2 and minimizing RMSE, while controlling the effects of VIF and Total Multiple R. The weight combinations are provided in Table 2.
The following phase was correlation analysis and the generation of a correlation plot for the entire dataset based on Pearson Correlation Coefficient values (Figure 2). Figure 2 shows the correlation matrix of 19 key variables related to geopolymer concrete. The relationships between the variables are color-coded based on the correlation coefficient, with red indicating a positive relationship and blue indicating a negative relationship. A very strong negative correlation (r = −0.92) is observed between FA and GGBFS, indicating a strong inverse relationship between the two constituents. Na2SiO3 also shows a high positive correlation with NaOH and total water, since they are often used together in alkaline solutions. The values of water (1), added water, and total water also show relatively high correlations. Most variables related to curing time and temperature (e.g., final curing temperature) are less correlated with other variables, indicating that these parameters are determined independently of the composition of the raw materials. Overall, this matrix accurately shows the internal relationships of geopolymer concrete variables.
Following this phase, the dataset with all features is evaluated with a Random Forest Algorithm using 10-fold cross-validation. In this stage, metrics related to linear dependence and model accuracy are provided with several graphs.

2.3.2. Stage 2a: Feature Selection with Genetic Algorithm (Multi-Objective Optimization)

In this study, we have developed a Genetic Algorithm (GA)-based feature selection method to identify the most relevant input variables. The method optimizes a multi-objective fitness function that balances prediction accuracy, linearity/linear dependence, and the number of selected features. Each solution is represented as a binary chromosome, and standard GA operators such as selection, crossover, and mutation are used. The best-performing feature subset is selected based on repeated runs and average model performance. This approach is adaptable to any regression model and improves both efficiency and interpretability compared to traditional methods. This approach can be used to tackle high VIF scores caused by possibly redundant features. The GA based algorithm developed focuses on the following 4 objectives:
(1)
Minimizing VIF
(2)
Maximizing R2
(3)
Minimizing RMSE
(4)
Minimizing Total Multiple R
The following elaborates on each objective:
  • Minimizing VIF (Variance Inflation Factor):
High multicollinearity arises when independent variables are highly correlated, making it difficult to isolate the individual effect of each predictor. This condition increases the standard errors of the regression coefficients and undermines the model’s reliability. The Variance Inflation Factor (VIF) is commonly used to detect such issues; VIF values greater than 10 typically indicate problematic multicollinearity, suggesting the need to remove or combine variables. Rather than merely treating input features as isolated variables, this objective seeks to examine the underlying dependency structure between them. A high VIF implies that a variable carries redundant information already encoded by other predictors, which not only distorts coefficient estimates but also undermines the clarity of model interpretation. By minimizing VIF, we aim to ensure that each selected variable contributes unique and non-overlapping information, thereby promoting a more stable and trustworthy model structure.
  • Maximizing R2 (Coefficient of Determination):
This objective is rooted in the notion that a strong predictive model should be capable of capturing the majority of the variability present in the response variable(s). A high R2 does not merely signal good performance—it indicates alignment between the model’s internal representation and the natural structure of the data. Optimizing for R2 thus reflects an effort to build models that are not just mathematically accurate, but also conceptually coherent with the phenomenon being modeled.
  • Minimizing RMSE (Root Mean Square Error):
While R2 reflects the proportion of variance explained, RMSE directly captures the magnitude of prediction error in the same unit as the response. It penalizes large deviations more severely than smaller ones, making it particularly sensitive to outliers and noise. Minimizing RMSE ensures that the model’s predictions remain not only close to observed values but also consistent and reliable across diverse instances, without major fluctuations in performance.
  • Minimizing Total Multiple R (Custom Metric):
Beyond standard metrics, we introduce Total Multiple R as a tailored indicator of cumulative linear dependence between input features and the target(s). Total Multiple R measures the overall interdependence among selected variables by summing the Multiple Correlation R values for each variable. The calculation method for the Total Multiple R metric is provided in Table 3.
For each independent variable xi, a regression model is constructed using the remaining V-1 independent variables as given in Equation (1):
y i = β 0 + j = 1 , i j V β j x j + ε
where yi denotes the i-th original variable treated as the dependent variable in the corresponding regression model, the remaining xj are used as independent variables (IVs) of the original dataset, V is total number of independent variables, β0 is intercept, βi is regression coefficient and ε is residual error term. The calculation of coefficient of determination (R2) is done according to Equations (2)–(4).
R i 2 = 1 S S r e s , i / S S t o t , i
S S r e s , i = Σ k = 1 n y i , k y ^ i , k 2
S S t o t , i = Σ k = 1 n y i , k y - i 2
yi,k is the observed value of variable i for sample k, y ^ i , k is the predicted value from regression, y - i is the mean of variable i and n is the number of samples. The multiple correlation coefficient Ri is calculated as Equation (5).
R i = R i 2
Total multiple R (TotalR) is calculated as an overall metric as given in Equation (6).
T o t a l R = Σ I = 1 V R i
As the range of Total Multiple R is [0, Number of Independent Variables], high total correlation (Total Multiple R) values may reflect a scenario where certain features dominate the output space, leading to overfitting and sensitivity to minor changes in those dominant predictors. By minimizing this metric, we aim to balance the influence of individual features and reduce the risk of overly deterministic or fragile model behavior.
The objective function, which aims to fulfill the 4 objectives mentioned above, takes independent variables of the ML problem and objective weights as inputs, trains the Random Forest models with different combination of independent variables, then calculates the values of VIF, R2, RMSE and Total Multiple R, and provides the weighted sum of these as the objective function output.
The optimization algorithm with 4 objectives has been transformed into a single objective optimization problem using the Weighted Sum Method, as explained in [24], by forming a convex combination of the objectives with the weights provided by the user. In runtime, the user is given an option to run the feature selection process either focusing on one of the competing objectives, i.e., minimization of VIF, maximization of R2, minimization of RMSE, and minimization of Total Multiple Correlation R or choosing a balanced approach. When the user selects one of these 4 approaches, a predefined set of weights is provided for that specific focus. When the user selects the balanced approach, equal weights of 0.25 are assigned for each objective. In this specific study, twelve weight combinations, provided in Table 2, were used, prioritizing both maximizing R2 and minimizing RMSE, while controlling the effects of VIF and Total Multiple R.
NSGA-II is implemented as the Genetic Algorithm (GA) variant for this multi-objective problem. Random Forest is used inside the genetic algorithm for training as the ML model that is evaluated with different combinations of independent variables. A fixed set of hyperparameters for the Random Forest model is used when evaluating the model with different variable combinations. The minimum number of independent variables to keep was a constraint of the implemented GA (default: 5) and is input by the user before the algorithm initiates. The user is also asked to input hyperparameters of the GA, such as the number of generations [10–100] and population size [20–200].
Following the completion of the multi-objective genetic algorithm optimization process, the algorithm produces a set of non-dominated solutions known as the Pareto front. These solutions represent optimal trade-offs among the four competing objectives: minimization of VIF, maximization of R2, minimization of RMSE, and minimization of Total Multiple Correlation R. The Pareto front analysis phase involves a comprehensive evaluation of all solutions, and the selection of a single optimal solution based on the user-defined focus and preferred weights set. The set of all Pareto optimal solutions forms the Pareto front, which represents the best possible trade-offs achievable among the competing objectives. Solutions in the Pareto front have the property that any improvement in one objective necessarily requires degradation in at least one other objective.

2.3.3. Stage 2b: Pareto Front Analysis for Multi-Objective Variable Selection

Each individual in the Pareto front is represented as a binary vector b ∈ {0,1}^V, where V is the total number of available variables, bi = 1 indicates variable i is selected, and bi = 0 indicates exclusion. The pseudocode for Pareto front analysis is provided in Table 4.
In the Pareto front analysis, following the initial 3 steps, Step 4 identifies the best solutions for each objective, and Step 5 calculates the weighted (balanced) scores for all possible solutions (i.e., feature subsets). Finally, the optimal solution is selected based on these scores in Step 6. The best solutions for each of the 12 scenarios tested were calculated using this analysis.

2.3.4. Stage 3: Performance Assessment of ML Models with Selected Features

Once the feature selection is complete and the optimal subset of features is selected based on commonly selected features in 12 scenarios, a series of ML models is trained with hyperparameter optimization using this optimal subset. In this stage, the hyperparameter optimization is accomplished through the HyperOpt (version 0.2.7) Python library. This section first presents a brief overview of each of the models evaluated in this stage, and following this, an overview of hyperparameter optimization is provided.
XGBoost: XGBoost is a Gradient Boosting framework that builds decision trees sequentially, with each new tree attempting to reduce the residual errors of the previous ones. Its learning mechanism incorporates both first- and second-order gradients, allowing for precise optimization during training. What sets XGBoost apart is its built-in regularization, which helps prevent overfitting, especially in high-dimensional datasets often encountered in structural and materials engineering. This algorithm has shown superior predictive accuracy in complex regression problems, such as modeling the rheological properties of self-compacting concrete.
CatBoost: CatBoost is another Gradient Boosting algorithm specifically designed to handle categorical features more effectively than traditional models. İt introduces a novel technique known as ordered boosting, which reduces prediction shift and leakage that can occur during training. Furthermore, CatBoost internally transforms categorical variables into numerical formats without the need for external encoding. This capability makes it highly suitable for construction datasets that include multiple qualitative inputs. Although its training time is typically longer than that of other boosting methods, it often achieves comparable or better accuracy.
LightGBM: LightGBM is a Gradient Boosting algorithm developed for efficiency and scalability. It uses a leaf-wise growth strategy, which typically results in deeper trees and better accuracy compared to level-wise methods. Additionally, it employs techniques like Gradient-based One-Side Sampling (GOSS) and Exclusive Feature Bunding (EFB) to reduce computational cost and memory usage. İn engineering applications involving large and heterogeneous datasets, LightGBM offers an excellent balance between training speed and prediction performance, making it an attractive option for real-time modeling.
K-Nearest Neighbors: KNN is a non-parametric, instance-based learning algorithm commonly used for regression and classification tasks. It predicts outcomes by identifying the closest data points in the feature space and aggregating their responses. Although not frequently employed. KNN often serves as a benchmark due to its simplicity and interpretability, particularly when transparency in decision-making is required.
Random Forest: Random Forest is an ensemble learning algorithm that constructs multiple decision trees using random subsets of data and features. The final prediction is obtained by aggregating the outputs of individual trees through majority voting or averaging.
Gradient Boosting: In practical engineering problems where data complexity and non-linearity are prevalent, Gradient Boosting has emerged as a reliable modeling strategy. It functions by building a predictive structure layer by layer, correcting prior errors iteratively, which allows for capturing subtle patterns hidden in the data. Rather than relying on a single deep tree, the model constructs a sequence of shallow ones, each fine-tuning the shortcomings of its predecessor. This stepwise refinement not only improves accuracy but also allows engineers to regulate overfitting by adjusting parameters like learning rate and maximum depth. It is particularly valuable in applications involving material strength prediction or load estimation, where precision is essential and training data may be limited.
Decision Tree: For tasks requiring clarity and explainability, Decision Trees serve as an intuitive analytical framework. Their logic mimics human decision-making: sequential, rule-based, and visually traceable. While they may not offer the highest predictive power on their own, they are instrumental in uncovering key variable interactions and segmenting data in a transparent way. This interpretability is especially advantageous in engineering domains, where understanding cause–effect relationships between inputs and outputs can inform both model development and domain-specific decisions.
Extra Trees: In scenarios where model training speed and generalization are crucial, the Extra Trees algorithm offers an efficient yet robust solution. Randomly selecting both features and splitting thresholds, it eliminates the exhaustive search typical of standard decision trees, which drastically reduces training time. Although the model sacrifices fine-tuned decisions at each node, it gains stability and noise tolerance through sheer ensemble diversity. This makes Extra Trees highly practical in settings where data may include measurement uncertainties or where rapid iterations are needed across many design alternatives.
MLP Neural Network: The Multilayer Perceptron (MLP) is a class of artificial neural networks particularly effective for learning nonlinear relationships in engineering systems. With its layered architecture consisting of interconnected neurons distributed across input, hidden, and output layers, MLP can model complex mappings between input features and output responses. Its strength lies in the backpropagation algorithm, which fine-tunes the internal weights based on the discrepancy between predicted and actual values. This adaptability allows MLPs to capture intricate behavioral patterns in domains such as material science, structural health monitoring, or energy consumption forecasting, even when analytical formulations are not feasible.
Ridge Regression: When dealing with multicollinearity among predictors or limited data in high-dimensional spaces, Ridge Regression serves as a disciplined approach to prevent model overfitting. By introducing a penalty term proportional to the square of the coefficients, it constrains the magnitude of parameter estimates, stabilizing the solution. This regularization not only improves predictive generalization but also allows for a more robust interpretation in data-driven engineering studies. Especially in optimization tasks involving multiple correlated inputs, Ridge Regression provides a mathematically sound alternative to standard linear models by managing trade-offs between bias and variance.
Linear Regression: Linear Regression remains a foundational method in predictive modeling, offering clarity and interpretability in understanding relationships between variables. Its principle assumes that the target variable can be approximated as a weighted sum of the input features, with residuals following a normal distribution. Despite its simplicity, linear regression is valuable in preliminary assessments, sensitivity analyses, or when establishing baseline models for comparison. In engineering contexts, it enables practitioners to quickly identify dominant factors affecting system behavior, while also setting the stage for more advanced modeling if required.
Lasso Regression: Lasso Regression, like Ridge, is a type of penalized linear regression used for feature selection and preventing overfitting. The main difference from Ridge is that Lasso’s penalty can drive some coefficients exactly to zero. This feature allows Lasso to remove unimportant features from the model, creating a simpler and more interpretable model.
Elastic Net: Elastic Net is a combination of Ridge and Lasso regression. This algorithm simultaneously uses both L1 penalty (like Lasso) and L2 penalty (like Ridge). This combination allows Elastic Net to perform feature selection (like Lasso) and effectively manage multicollinearity issues (like Ridge), making it suitable for many datasets.
SVR (Linear): Linear SVR is a version of Support Vector Machine designed for regression problems. Instead of trying to find a line that minimizes error, Linear SVR tries to find a line that keeps most of the data within a specific “margin area”. This makes the model less sensitive to outliers.
SVR (RBF): RBF SVR is also used for regression, but it uses a kernel function called the Radial Basis Function (RBF). This function allows SVR to model complex nonlinear relationships between data. In other words, RBF SVR can find a “line” (actually a curve or hyperplane) that best approximates non-linear data, offering more flexibility than Linear SVR.
In machine learning, hyperparameters are external settings defined by the user before the training process begins. Unlike model weights, which are learned from data, hyperparameters control the overall behavior of the algorithm. They influence aspects such as model complexity, training speed, and generalization ability. Examples include the learning rate, the depth of a decision tree, or the number of estimators in ensemble methods. Selecting appropriate hyperparameters is a critical step, as it directly impacts the model’s accuracy and stability. Hyperparameter optimization aims to systematically determine these parameters to maximize the model’s performance. This process helps the model produce more stable and accurate results by preventing potential problems such as over- or under-fitting. Once the feature selection is finalized and the set of optimal features is determined, training with hyperparameter optimization is performed for each ML model (Random Forest, XGBoost, LightGBM, etc.) through the Hyperopt library [25]. The Hyperparameter Optimization was accomplished using the HyperOpt (version 0.2.7) python package’s Tree-structured Parzen Estimator (TPE), which is used as the optimization algorithm in this stage. Following the training stage, the models are evaluated on a separate holdout set.

2.3.5. Stage 4: Visualization and Verification

This stage involves the generation of convergence plots, Pareto front plots, Feature Importance Plot for the most accurate ML model, in addition to a series of verification checks that are also made to ensure the final VIF values are not extreme, and the linear dependency problem has been resolved.

3. Results

The experiment started with providing the initial parameter values to the Genetic Algorithm as indicated in Table 5. The algorithm used the exact same values for all runs of the algorithmic pipeline, which runs 12 times for different weight scenarios given in Table 2.

3.1. Feature Selection via Genetic Algorithm

Table 6 below presents a summary of outputs of 12 genetic algorithm pipeline runs, for the 12 weight combination scenarios that focus on performance optimization while taking into account tackling variable interdependency. N vars indicates the number of features retained on scenario completion, Max VIF of the solution indicates the Max VIF value exhibited along with all retained variables, R2, RMSE, Total Multi-R indicates the R2, RMSE, Total Multi-R values obtained as result of training with baseline (non-fine-tuned) Random Forest Model for that scenario. Naturally as the number of selected features decreases, due to both the decrease in the predictive power of the model and total correlations among independent variables, the Max VIF, R2 and Total Multi-R values have decreased and the RMSE values have increased.
Table 7 provides the selected features (variables) for each scenario. The set of Concentration(M) NaOH, NaOH (Dry), GGBFS, Initial curing temp (C), Coarse Aggregate formed a solid foundation by being selected in all scenarios. The Na2O (Dry) and Initial curing rest time (day) were selected in ¾ of all scenarios, indicating that these variables can also play a significant role in predictions. Additional water, initial curing time (day), FA, Na2SiO3 were selected only in ¼ of all scenarios, indicating that the contribution of these variables to the predictive ability of the model is limited. A total of 11 features were selected as a result of this selection process; the remaining eight features appear to have either zero or a considerably very low impact on the predictive ability of the model.
Every variable selection happens as a result of the Pareto front analysis process. The weighting scheme helps in transforming the solution from a multi-objective to a single-objective optimization problem. The balanced score is calculated first by applying a min-max normalization to all calculated objective values, then the weighted sum (ws) of normalized values of each objective was calculated according to the formula below.
ws = (1 − normVIF) ∗ w1 + (normR2) ∗ w2 + (1 − normRMSE) ∗ w3 + (1 − normMultiR) ∗ w4
Table 8 provides Pareto front optimal solutions, calculated weighted sum scores (Balanced Score) and selected optimal solution for scenario S4 (chosen as an example as it is the first scenario representing the lowest VIF score obtained). In the table, the column N Vars provides the number of variables (features) selected for that optimal solution. The table also provides the maximum VIF value observed for the variables in that selected subset, along with R2 and RMSE. Total Multiple R values observed in the training with the Random Forest model (Random Forest model was the only model used in all generations of the GA).
The algorithm also provides the best solution per objective for each scenario. The best solutions per objective for scenario S4 are shown in Table 9.
The solution providing the Best Max VIF score was Solution 1, the solution providing the Best R2 was Solution 15, the solution providing the Best RMSE was again Solution 15, and the solution providing the Best Total Multiple R was Solution 1. According to the weights provided for this scenario S4 (0.10, 0.40, 0.40, 0.10), the Best Solution (the Balanced Solution) was selected as Solution 9, with a Balanced Score of 0.8673. The feature subset in the selected solution (Solution 4) included Concentration(M)NaOH, NaOH (Dry), GGBFS, Initial curing temp(C), Coarse Aggregate, Na2O (Dry) and Initial curing rest time (day). The Pareto front plot from the perspective of different objectives of the selected solution is provided in Figure 3.
Figure 3 indicates how variables representing objectives correlate with each other in the selected Pareto front. The Pareto front visualization in Figure 3 illustrates the trade-off relationships among multicollinearity (VIF), predictive accuracy (R2), prediction error (RMSE), and feature redundancy (Total MultiR). Each red diamond represents a non-dominated solution where no objective can be improved without degrading at least one other objective. The optimal solution, indicated by the red star, was selected using the predefined weighted aggregation function. The results demonstrate that although slightly higher R2 values can be achieved at the expense of increased multicollinearity and redundancy, the selected optimal solution provides a balanced compromise by maintaining high predictive performance while controlling feature interdependence. This confirms the effectiveness of the proposed multi-objective optimization framework in identifying robust and parsimonious feature subsets without relying solely on predictive accuracy.
As explained previously, the set consisting of Concentration(M)NaOH, NaOH (Dry), GGBFS, Initial curing temp(C) and Coarse Aggregate formed a solid foundation by being selected in all scenarios. Thus, it was decided to use these five features in the further ML training and evaluation stages. Figure 4 presents the correlation plot of selected variable and target variable (28 d cubic compressive strength (MPa)).
The Pearson correlation analysis indicated that the 28-day compressive strength exhibits weak-to-moderate linear relationships with the input variables. Most importantly, the absence of strong individual correlations suggests that compressive strength is governed by complex multivariate interactions rather than a single dominant factor, supporting the need for the use of advanced machine learning approaches capable of capturing nonlinear relationships. Among the predictors, NaOH concentration showed the highest positive correlation (r = 0.33), followed by NaOH dry mass (r = 0.25), indicating that alkaline activator properties play a relatively more important role in strength development. Despite the fact that we had concerns regarding keeping the GGBFS within the selected set, as GGBFS content and initial curing temperature exhibited very weak correlations (r = 0.07), suggesting that their individual linear contributions to strength variation are limited within the studied range and a strong negative correlation was observed between GGBFS and coarse aggregate (r = −0.92), indicating severe multicollinearity between these variables (which would in turn increase VIF), we finally decided to keep this variable as it is selected in all 12 scenarios in GA optimization process, as it might possibly have a high contribution to the predictive power of the models.

3.2. Baseline Model Evaluation

The ML pipeline involved 30 runs of baseline model evaluation and multi-model fine-tuning. In the first stage, by considering the Random Forest as the algorithm for baseline evaluation, the model was first evaluated in each run of the ML pipeline with the full input variable set. The dataset was divided into training (80%) and holdout (20%) subsets using a fixed seed in each of the 30 runs to ensure the baseline model and fine-tuned models use the same data splits. The model training and hyperparameter evaluation were performed using 5-fold cross-validation exclusively on the training data. Appendix B provides the performance metrics of the baseline model. In the metrics, the cross-validated R2 represents the average out-of-fold predictive performance, while the holdout R2 reflects the final generalization performance on an independent, unseen dataset. The best performing run for the fine-tuned models was for seed 29; thus, in this section, we focus on the baseline model evaluation metrics for the baseline model run for the same seed 29. The R2 value for the holdout set evaluation was 0.9306 with RMSE = 3.1516 and MAE = 2.2696. The R2 value for the test fold of 5-fold cross-validation was 0.8933 with RMSE of 3.6735. (Figure 5).
The VIF values of each variable were calculated to control interdependence between features used in the baseline Random Forest model. Figure 6 presents the VIF bar graph in descending order, 10 features such as Superplasticizer, Additional water, Water (2), NaOH, Total water, Water (1), NaOH (Dry), Na2SiO3, SiO2 (Dry), Na2O (Dry) exhibit extremely high VIF values (>1000).
The R2 value for the test set was also high for the Random Forest model, which is known to be a robust model for high VIF values of predictor variables. In fact, ML models become very weak in terms of interpretability when multiple independent variables are strongly correlated (as multiple variables explain the same signal). In such a case, the Random Forest will often pick one of them for splits, ignore the other, or alternate randomly across trees. This leads to three subtle effects: (i) Feature importance becomes unreliable; Gini or impurity-based importance gets “diluted” across correlated features. None of them look important, even though the group is; (ii) Interpretability lowers, and it would not be possible to say “X1 matters more than X2” when they are shadow-boxing the same information. (iii) Model size and noise increase slightly, redundant features mean more candidate splits, which means more randomness leading to marginal inefficiency. In order to ensure robustness in model evaluation, increase efficiency in accuracy and achieve explainability in the next stage, the pipeline involved 30 runs of fine tuning of a set of ML models with the exact seeds and data splits used for baseline model evaluation.

3.3. Evaluation of Hyperparameter Optimized Models

A set of ML models was evaluated with 30 runs of the fine-tuning pipeline using the selected features in the following stage. The dataset is split into train and holdout sets with an 80/20 ratio using the pre-determined seeds, and only the train set is used in the training process with 5-fold cross-validation. The 5-fold cross-validation was conducted on the training dataset to estimate out-of-sample predictive performance, and the trained model was subsequently evaluated on an independent holdout dataset that was not involved in either model fitting or validation. The evaluation metrics for all 30 runs are presented in Appendix C. The cross-validated (CV) R2 represents the mean performance across the out-of-fold test partitions, whereas the holdout R2 (Holdout) reflects performance on a single, unseen realization of the data distribution
Table 10 and Figure 7 and Figure 8 summarizes average performance of the hyperparameter-optimized models after 30 runs of the pipeline, each run with 1000 TPE iterations of the Hyperopt library. The models were trained with the set consisting of five optimal input variables determined as a result of the GA-based feature selection process explained in the previous section.
When the average performance across the out-of-fold test partitions is compared, CatBoost appeared as the best-performing model, followed by XGBoost and Gradient Boosting. This is followed by ExtraTrees, LightGBM and Random Forest. The results indicate the dominance of ensemble models in exhibiting the best performance, where boosting-based models performed slightly better than the bagging-based ones. Among the foundation models, SVR appeared as the best performing model, followed by MLP, Decision Tree and KNN. Linear models, regardless of being the penalized ones or simple linear regression, exhibited the lowest performance among all models. Comparison of the average CV R2 and holdout R2 values revealed no indication of overfitting and no evidence of bias arising from an unbalanced holdout split, as both metrics exhibited consistent and comparable predictive performance.
The best overall holdout performance was exhibited in the 30th run (seed = 29) (Table 11) by the fine-tuned CatBoost model. In this run, the fine-tuned CatBoost model was able to achieve a holdout R2 score of 0.9627 and RMSE of 2.3113 with only five input variables (i.e., set of Concentration(M)NaOH, NaOH (Dry), GGBFS, Initial curing temp(C), Coarse Aggregate). Feature importances plot for CatBoost is given as Figure 9. Figure 10 provides a predicted vs. actual plot for all algorithms for this run of the pipeline for the holdout set.

3.4. Feature Importance Analysis

In the feature importance analysis, the focus is on evaluating how each input variable contributes to the predictive performance of the best-performing model for estimating the output feature. CatBoost was chosen for this examination as it outperformed the other models. We would like to note that reported importance rankings are model-dependent interpretations rather than universal causal claims. The measure of importance for each variable was determined by assessing the change in the model’s error when that specific variable was used in the prediction process. The algorithm uses each model’s built-in feature_importances attribute to get the feature importance values for each feature. The results show that GGBFS has the highest contribution to the predictive ability by a significant margin. This indicates the important role of blast furnace slag in the mechanical properties of geopolymer concrete. In addition, NaOH (Dry) also has a significant contribution to the predictive ability of the model. In addition, Concentration, Initial Curing Time, and the amount of Coarse Aggregate were identified as factors affecting the concrete strength.
Although GGBFS exhibited a very weak linear correlation with compressive strength (r = 0.07), it was identified as the most important feature by the CatBoost model. This apparent discrepancy arises because Pearson correlation measures only the individual linear relationship between a predictor and the target variable, whereas CatBoost feature importance reflects the overall predictive contribution of a variable within a multivariate and nonlinear modeling framework. In particular, GGBFS showed a strong negative correlation with coarse aggregate (r = −0.92), indicating substantial multicollinearity, which can suppress its apparent marginal correlation with strength. However, tree-based models such as CatBoost can effectively capture nonlinear and interaction effects between variables. Therefore, the high importance of GGBFS suggests that it plays a critical role in combination with other parameters, even though its individual linear association with strength appears weak. This finding highlights the limitation of relying solely on linear correlation analysis and demonstrates the ability of the proposed machine learning framework to uncover complex multivariate relationships governing compressive strength.

4. Discussion

In this study, various machine learning models were used to predict the compressive strength of geopolymer concrete. Boosting-based ensemble models, particularly CatBoost, XGBoost, and Gradient Boosting, achieved the best performance, followed by bagging methods, while SVR was the strongest among base models and linear models performed the worst; moreover, the consistency between CV R2 and holdout R2 indicated reliable generalization without overfitting or holdout bias.
For the 30th run (seed = 29), the Random Forest model trained with the reduced feature set (five variables) achieved (holdout) R2 of 0.8775 on the unseen data, while the baseline Random Forest model trained with the full feature set (19 variables) achieved R2 of 0.9306 under the same conditions. Based on the 30th run, although the baseline model exhibited a slightly higher performance (ΔR2 ≈ 0.06), the overall predictive performances remained comparable. This result along with the average Random Forest performance of R2 of 0.8716 when compared with the baseline model performance, indicate that the hyperparameter optimization was able to largely compensate for the reduction in input dimensionality, and that decreasing the number of features from 19 to 5 did not result in a substantial or practically significant deterioration in predictive performance, while greatly enhancing the robustness in model interpretability.
A comparison of our findings with the existing literature reveals both alignment with established trends and notable methodological distinctions. Previous studies have consistently reported that parameters such as NaOH molarity, GGBFS proportion, and curing conditions exert a marked influence on geopolymer compressive strength; however, these investigations were often restricted to specific experimental setups or relatively small datasets. In this study, a larger and more diverse dataset comprising 274 mix designs was analyzed, enabling the confirmation of these known effects within a broader parameter space. By combining machine learning models with a multi-objective genetic algorithm, we were able to optimize predictive accuracy while reducing redundancy among variables, ensuring interpretability and thereby offering a practical tool that complements and extends the scope of earlier experimental research.
A comparison of the existing literature and the present study on geopolymer compressive strength prediction is given in Table 12.
This study introduced a combined approach of machine learning and multi-objective genetic algorithms to achieve both high predictive accuracy and stable variable selection. By modeling the effects of various input parameters across a broad dataset, multicollinearity is reduced through GA-based feature selection, enhancing interpretability. The analysis also considered ambient curing conditions, capturing their interactions with other mix components through correlation and network-based approaches. The models reproduce known trends, quantify optimal ranges, and examine interactions among mix design factors. Feature importance results from CatBoost highlight Na2SiO3 and dry NaOH as key predictors, consistent with experimental observations. Overall, this work reinforces established patterns while providing a data-driven predictive framework that combines feature selection with GA and optimized machine learning models. The study confirms trends at a larger scale and introduces a methodological contribution by simultaneously optimizing feature stability and prediction accuracy within a single workflow.

5. Conclusions

The production of ordinary Portland cement is a major source of CO2 emissions, whereas geopolymer concrete significantly reduces environmental impact by utilizing industrial by-products such as fly ash and slag. At the same time, geopolymer concrete can achieve mechanical properties comparable to conventional concrete when key parameters, including precursor composition, activator type, and curing conditions, are properly optimized. These characteristics demonstrate that geopolymer concrete offers an effective balance between environmental sustainability and structural performance.
This research was focused on reinforcing the reliability and transparency of machine learning frameworks used to estimate geopolymer concrete compressive strength. Through implementing a dependency-aware feature selection protocol, we mitigate the negative effects of multicollinearity in a dataset (with many linearly dependent features) while maintaining high predictive precision of the model. The methodology followed a three-fold objective: first, engineering a specialized multi-objective NSGA-II genetic algorithm that filters features based on their linear independence; second, restructuring a high-dimensional geopolymer dataset into an optimized variable subset; and finally, benchmarking the predictive efficacy of twelve hyperparameter-tuned models trained on this refined input dataset. The results have shown that the number of linearly dependent features can be efficiently reduced without sacrificing the prediction accuracy for this dataset. It should be emphasized that the absence of a substantial loss in predictive accuracy between the baseline model trained on the full feature set and the models trained on the reduced feature subset can be largely attributed to the hyperparameter optimization applied to the latter, which enabled the models to maintain comparable performance despite the lower input dimensionality.
The features that have an influence on compressive strength prediction were observed as GGBFS, NaOH (Dry), Concentration, Initial curing time and Coarse Aggregate. This highlights the potential of industrial by-products like ground granulated blast furnace slag (GGBFS) as sustainable alternatives to conventional aggregates. Higher NaOH concentrations, particularly in the 8–16 M range, generally improve strength due to enhanced dissolution of aluminosilicate species.
Geopolymer concrete is particularly attractive for industries requiring high durability and resistance to aggressive environments, such as infrastructure, marine constructions, and precast concrete elements.
To emphasize the practical relevance of the findings, the best-performing ensemble model achieved an R2 of 0.96 on the holdout dataset, with consistently low prediction error across cross-validation folds, confirming its robustness. The integration of correlation diagnostics enabled the selection of a reduced and physically consistent feature subset, mitigating the instability commonly caused by multicollinearity in geopolymer datasets. This structured integration of interpretability within the modeling pipeline improves confidence in feature importance rankings and enhances engineering transparency compared to conventional black-box ML applications.
Nevertheless, the proposed framework is constrained by the moderate dataset size and reliance on a limited number of experimental sources, which may affect external validity. Model performance may also exhibit sensitivity to data partitioning strategies. Future work should focus on expanding the dataset with independent experimental campaigns, testing transferability across different precursor systems, and validating the framework under varying curing and environmental conditions. From a practical standpoint, the developed approach provides a transparent decision-support tool for geopolymer mix design, enabling more efficient parameter prioritization and reducing experimental trial-and-error efforts.
Future research should extend the proposed framework by incorporating larger and multi-source experimental datasets to enhance external validity and improve generalization capability. The integration of additional geopolymer systems (e.g., blended precursors or alternative activator types) would allow evaluation of model transferability across different material compositions. Moreover, combining the current multicollinearity-aware pipeline with advanced explainable AI techniques and uncertainty quantification methods could further strengthen interpretability and reliability. The development of hybrid physics-informed or mechanistic-ML models may also improve extrapolation performance beyond the training domain. Finally, implementing the framework within an optimization environment for automated mix proportioning would enable direct decision-support applications for sustainable geopolymer concrete design in practical engineering settings.

Author Contributions

Conceptualization, G.B. and F.A.; methodology, F.A. and Ü.I.; software, Ü.I.; validation, Ü.I. and F.A.; formal analysis, Ü.I., F.A. and G.B.; investigation, Ü.I. and F.A.; writing—original draft preparation, F.A., Ü.I., G.B. and C.C.; writing—review and editing, Ü.I., C.C., S.M.N., Z.W.G. and G.B.; visualization, Ü.I., C.C. and F.A.; supervision, G.B. and Z.W.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original data presented in the study are openly available in Mendeley Data Repository at reference [15].

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Step-by-Step Tree Diagram of Experimental Process

1. Load Data
   ├── Read CSV file
   ├── Handle missing values (dropna)
   ├── Apply StandardScaler normalization
   └── Extract dataset info (19 independent variables + 1 target)
2. Initial Correlation Analysis
   ├── Calculate Pearson correlation matrix
   ├── Generate correlation plot (heatmap)
   ├── Create correlation network plot (threshold = 0.7)
   └── Report correlation statistics
3. Genetic Algorithm Configuration
   ├── Set minimum variables (default: 5)
   ├── Configure GA parameters:
   │    ├── Population size (default: 60)
   │    ├── Generations (default: 40)
   │    ├── Crossover probability (default: 0.7)
   │    └── Mutation probability (default: 0.2)
   └── Set objective weights (VIF, R2, RMSE, Total R)

4. Genetic Algorithm(NSGA-II) Optimization (Run 12 Weight Combinations)
   ├── Initialize population (60 individuals, binary encoding)
   ├── For 40 generations:
   │    ├── Evaluate fitness (VIF, R2, RMSE, Total Multiple R)
   │    ├── Select with NSGA-II (Pareto ranking + crowding distance)
   │    ├── Crossover (prob = 0.7, two-point crossover)
   │    ├── Mutate (prob = 0.2, bit flip mutation)
   │    └── Update Pareto Front (Hall of Fame)
   └── Return Pareto optimal solutions

5. Pareto Front Analysis
   ├── Analyze all Pareto solutions
   ├── Find best solution for each objective
   ├── Calculate balanced scores using weighted normalization
   ├── Select optimal solution (maximizes balanced score)
   └── Variable count: Determined by GA results (typically 5–12)

6. Selected Variables Correlation Analysis
   ├── Calculate correlation for selected subset
   ├── Generate heatmap (selected variables only)
   ├── Create network plot (threshold = 0.7)
   └── Compare to baseline correlations
7. Baseline Evaluation (30 runs)
   ├── Calculate VIF for all 19 variables
   ├── Split data 80/20 train/test
   ├── Train Random Forest with k-fold CV
   ├── Evaluate on test set
   ├── Generate outputs:
   │    ├── VIF analysis plot
   │    ├── Predicted vs. true plots
   │    ├── Feature importance plot
   │    └── Baseline report (.txt)
   └── Purpose: Establish reference metrics before optimization

8. Training, Tuning and Evaluation of ML models(30 runs)
   ├── Extract selected features (N variables from GA)
   ├── Scale selected features (StandardScaler)
   ├── New Split: 80/20 train/test (Different from baseline split)
   ├── For each of 12–15 models:
   │    ├── Optimize hyperparameters (TPE, 50 evaluations)
   │    ├── k-fold CV on training set ONLY
   │    ├── Train final model on training set
   │    └── Evaluate on holdout test set
   ├── Create predicted vs. true grid for all models
   ├── Intelligent feature importance plotting:
   │    ├── Try best model first
   │    ├── If unsupported, try 2nd best, 3rd best, etc.
   │    └── Plot from first model that supports it
   ├── Rank models by test set R2
   └── Print performance summary table

9. Final Analysis and Visualization
   ├── Print Final Model Performance and Ranking:
   │    ├── Display All models
   │    ├── Show holdout test metrics: R2_Test, RMSE_Test, MAE_Test
   │    ├── Include CV metrics for comparison: R2_CV, RMSE_CV
   │    ├── Best model selected by R2_Test (holdout performance)
   │    └── Generalization assessment: ΔR2 = |R2_Test − R2_CV|
   ├── Plot GA convergence (4 objectives over 40 generations)
   ├── Plot Pareto front (4 views of trade-offs)
   ├── Plot feature importance for BEST model (by R2_Test)
   └── Compare baseline vs. final performance

10. Final Verification (embedded in Step 8)
   ├── Calculate final VIF values
   ├── Assess multicollinearity reduction
   ├── Report variable reduction percentage
   ├── Compare baseline vs. final metrics (R2, RMSE)
   └── Summary statistics and conclusions

Appendix B

Baseline Model Metrics (30 Runs)

ModelHoldout R2Holdout RMSEHoldout MAECV R2CV RMSESeed
RF0.92263.27842.64290.89633.57450
RF0.93193.03842.2910.89473.55241
RF0.93223.37692.5480.89423.50512
RF0.88254.13812.93610.90483.38083
RF0.91023.07462.29460.90843.36674
RF0.89863.08432.16760.90713.48025
RF0.92273.00162.27830.90793.4266
RF0.82013.562.78770.92223.29657
RF0.93563.06882.20930.89583.51388
RF0.90723.41492.50080.91733.26999
RF0.93422.76862.03120.8783.964510
RF0.87923.88612.9690.90763.44211
RF0.90063.67122.70490.90783.313312
RF0.93383.5962.86760.89743.397913
RF0.9492.54611.8960.8863.662214
RF0.86743.60072.86470.92223.194815
RF0.93832.92172.16540.91773.257516
RF0.93993.41242.55260.89083.512517
RF0.83143.86182.92150.91673.38218
RF0.9382.63771.9410.9143.320719
RF0.95432.55991.95620.89773.529220
RF0.90193.19832.35960.90563.573721
RF0.8564.12592.79640.91293.374522
RF0.94112.94452.0560.90953.372223
RF0.93773.2772.56920.89433.258124
RF0.92262.92112.22040.88993.610725
RF0.9252.86582.13040.9033.562426
RF0.91143.30762.43150.91473.228827
RF0.91382.81612.09890.91023.502728
RF0.93063.15162.26960.89333.673529

Appendix C

Fine-Tuned Model Metrics (30 Runs)

ModelHoldout R2Holdout RMSEHoldout MAECV R2CV RMSESeed
CatBoost0.95012.63282.04130.91663.2080
XGBoost0.91973.33892.35490.89153.65010
LightGBM0.90823.57072.4920.82634.64470
GradientBoosting0.90593.61392.5640.89043.6710
ExtraTrees0.89583.80383.00690.88613.76580
RandomForest0.88264.03733.1820.86464.11540
SVR0.8514.54763.61130.85664.23690
DecisionTree0.83554.77963.90730.75855.44980
MLP0.80135.25284.45360.84694.34040
KNeighbors0.70076.44654.96250.6736.51760
ElasticNet0.64057.06516.04850.61166.98370
Ridge0.64047.0666.04990.61166.98340
Lasso0.64047.0666.04970.61166.98350
LinearRegression0.64037.06656.05050.61166.98330
CatBoost0.93942.86572.05930.91333.21021
XGBoost0.93243.02562.11320.9053.38241
GradientBoosting0.90463.59572.45520.90553.36121
ExtraTrees0.90083.66723.15070.86434.08791
RandomForest0.88873.88323.19770.84874.321
SVR0.86454.28443.24730.84044.38471
LightGBM0.86164.33052.94420.8574.28671
MLP0.85664.40873.67540.85824.12251
DecisionTree0.80135.18863.80990.79095.06121
KNeighbors0.77245.55324.47870.62966.94881
LinearRegression0.67656.62115.3820.5997.0411
Lasso0.67656.62125.3820.5997.04111
Ridge0.67656.62135.38230.5997.0411
ElasticNet0.67646.62185.38280.5997.04121
CatBoost0.95022.89262.17380.90913.24422
XGBoost0.94223.11662.45110.89333.50892
GradientBoosting0.93163.39152.49270.88473.64912
ExtraTrees0.90933.90363.23430.87393.81042
SVR0.89324.23663.3650.81184.682
LightGBM0.89024.29633.07130.83964.38432
RandomForest0.88654.36713.17540.84854.222
DecisionTree0.87864.51643.77980.80084.81982
MLP0.84275.14224.28870.78635.03062
KNeighbors0.72016.85835.05390.60626.81832
LinearRegression0.6887.24135.89340.57866.94412
Ridge0.68797.24235.89360.57866.94412
Lasso0.68797.24255.89360.57866.94422
ElasticNet0.68777.24455.8940.57866.94432
CatBoost0.9193.4362.60970.91413.13673
MLP0.88964.01093.26530.86034.00443
XGBoost0.88254.13792.82130.92143.04953
SVR0.8784.2163.22670.80934.69473
ExtraTrees0.87354.29363.49080.87663.81983
GradientBoosting0.85964.52332.96820.91393.21133
RandomForest0.84254.79133.63860.86224.12163
DecisionTree0.80035.39444.2380.75715.22963
LightGBM0.78415.60923.80230.84224.26033
KNeighbors0.71446.45065.05330.63476.75223
LinearRegression0.676.9345.59340.57777.04273
Lasso0.676.93435.59270.57777.04273
Ridge0.676.93495.58780.57797.0433
ElasticNet0.66996.93575.58430.5787.04323
XGBoost0.89923.25732.51460.90893.34984
GradientBoosting0.89693.29312.5820.90753.37744
CatBoost0.8893.4182.55350.91863.17554
LightGBM0.88523.4762.64560.85344.23144
RandomForest0.86943.70652.95410.86344.18314
ExtraTrees0.86373.78742.94890.88213.89054
SVR0.8464.02523.27930.82154.72774
MLP0.7445.19034.31460.86634.1414
DecisionTree0.7365.27024.46250.76075.46334
KNeighbors0.69965.62214.47310.65966.69974
LinearRegression0.58786.58575.59280.62747.00124
Ridge0.58686.59415.59630.62777.00044
Lasso0.58636.59785.5970.62767.00184
ElasticNet0.58626.59875.5980.62777.00044
CatBoost0.92482.65591.88680.91753.21955
XGBoost0.92152.71331.72320.90463.49685
GradientBoosting0.89913.07622.00020.90333.57345
ExtraTrees0.89243.17672.54680.88134.00735
LightGBM0.88053.34812.55560.86744.20615
RandomForest0.86423.572.77630.86924.21825
SVR0.80844.23973.20750.85344.41865
MLP0.73874.95094.02280.8993.66195
DecisionTree0.7384.95834.0450.83134.73485
KNeighbors0.70025.30343.91120.68546.63245
ElasticNet0.5546.46855.1890.63337.02765
Lasso0.55336.47395.18550.63327.02835
Ridge0.55096.49095.19680.63347.0265
LinearRegression0.5446.54045.21630.63367.02145
XGBoost0.94392.5571.83630.9083.39546
CatBoost0.9432.57731.9860.91463.26326
GradientBoosting0.92792.8992.03940.9053.42226
SVR0.89233.5442.85360.83224.58626
ExtraTrees0.88493.66382.91520.87763.98886
LightGBM0.8773.7873.02020.82844.75316
RandomForest0.87233.85913.10220.85754.33896
MLP0.86553.96013.19530.86744.09416
DecisionTree0.7395.51714.64550.73865.73876
KNeighbors0.71485.76664.61380.6526.89586
LinearRegression0.63766.50065.43150.6247.00166
Lasso0.6376.50595.43450.6247.00276
Ridge0.63626.51395.44180.6247.00276
ElasticNet0.63536.52125.44690.6247.00326
CatBoost0.83673.39212.59450.92913.04997
XGBoost0.80223.73322.92580.93562.9777
ExtraTrees0.80083.74573.04520.89593.76867
LightGBM0.79743.77753.1160.8664.22777
GradientBoosting0.79673.7842.82420.92483.19417
RandomForest0.7883.86433.03170.87784.11617
SVR0.77453.9863.21610.84894.54287
MLP0.73344.33333.63470.84264.6457
KNeighbors0.65324.9433.88960.71036.31767
DecisionTree0.58615.39994.54920.76425.72647
Lasso0.41696.40915.28240.63237.05647
ElasticNet0.41566.41635.28760.63227.05687
Ridge0.41456.42215.29110.63227.0567
LinearRegression0.41226.4355.30040.63227.05437
CatBoost0.94692.78562.10580.90883.26818
XGBoost0.94312.88432.06080.91123.298
GradientBoosting0.92683.27162.31970.90233.41448
LightGBM0.92223.37362.47650.83574.55438
RandomForest0.89163.98123.07470.84814.26758
ExtraTrees0.88644.07553.22770.87623.86088
SVR0.87554.26653.32990.81754.6318
MLP0.85664.57853.59560.85314.16668
DecisionTree0.72866.29854.71670.76295.31998
KNeighbors0.69276.70234.87520.65596.58898
LinearRegression0.67816.86035.69530.60396.95268
Ridge0.67796.86225.69750.60396.95258
Lasso0.67796.8625.69740.60396.95268
ElasticNet0.67766.86565.70130.60396.95268
CatBoost0.92493.07062.09180.92623.11219
XGBoost0.91993.17312.36220.91713.2529
GradientBoosting0.9193.19042.34590.91013.40899
LightGBM0.90493.45562.27410.84114.5219
RandomForest0.90223.50442.640.86584.1519
ExtraTrees0.89223.67962.94560.88473.86319
DecisionTree0.84264.44663.64270.84254.4459
SVR0.84234.45063.39240.83184.61839
KNeighbors0.77275.34334.24270.67046.52299
MLP0.71825.94984.80590.89073.69399
ElasticNet0.51287.82346.42920.65356.5729
Lasso0.51267.82486.42940.65356.57229
LinearRegression0.51267.82536.42950.65356.57229
Ridge0.51267.82466.42940.65356.57219
XGBoost0.91043.23192.23390.90553.488510
ExtraTrees0.90813.27332.71390.85544.379110
CatBoost0.90793.27642.2150.90383.570510
RandomForest0.90473.3332.63340.83774.60810
GradientBoosting0.89783.45172.32680.87763.987110
MLP0.8833.6932.88040.85554.375210
LightGBM0.87543.81082.73630.85744.313810
SVR0.85854.06142.93230.81344.960110
DecisionTree0.8454.25073.42630.78365.297610
KNeighbors0.81254.67453.69890.65216.816710
Lasso0.65166.37295.27260.61367.120410
Ridge0.6516.3785.28660.61397.121610
ElasticNet0.6516.37835.27710.61337.128810
LinearRegression0.64756.41035.32690.61377.111610
XGBoost0.90953.3632.52790.90453.538811
CatBoost0.89823.56712.55020.92982.961811
GradientBoosting0.89043.70112.62520.88383.753311
ExtraTrees0.88243.83413.29040.87793.981711
RandomForest0.88213.83793.08850.86464.285111
LightGBM0.87543.94562.8870.83624.665111
SVR0.83794.50173.48770.81494.681411
MLP0.77145.3454.46740.79984.832411
DecisionTree0.76825.38274.33760.73255.717711
KNeighbors0.76625.40584.08470.60397.352311
LinearRegression0.55777.43526.03540.6147.075211
Ridge0.55767.43576.03580.6147.075411
Lasso0.55767.43576.03570.6147.075511
ElasticNet0.55757.43676.03660.6147.07611
XGBoost0.93922.87112.17120.86773.861912
GradientBoosting0.92923.09762.16580.87973.728112
CatBoost0.92333.22492.39170.89833.402312
ExtraTrees0.86824.22723.43360.86433.973712
RandomForest0.86334.30583.41280.84834.254712
SVR0.86324.30723.43250.82184.490912
LightGBM0.84164.63483.23470.83664.49112
MLP0.83274.76284.0330.84194.294812
LinearRegression0.72236.13635.10910.56587.246112
Lasso0.72226.1375.10940.56587.246212
Ridge0.72076.15425.12010.56657.246412
ElasticNet0.71996.16285.12740.56657.246612
KNeighbors0.69236.45914.80270.59797.179212
DecisionTree0.67116.67785.20140.745.46412
XGBoost0.94953.14012.31510.88643.577113
CatBoost0.94483.28452.53570.90313.298713
GradientBoosting0.94133.3862.55050.88253.629513
LightGBM0.9223.90242.88140.8424.196413
ExtraTrees0.90594.28633.65850.84944.087313
SVR0.87764.88933.910.83634.270813
MLP0.86485.13984.38150.85993.94913
RandomForest0.84515.50094.26290.85024.105713
DecisionTree0.84285.54224.62080.78654.905913
KNeighbors0.64958.27476.28720.59916.663913
LinearRegression0.64278.35466.66510.62266.536113
Ridge0.64148.36986.66480.62286.534913
ElasticNet0.64078.37776.66470.62286.534913
Lasso0.63998.38666.66890.62276.535613
XGBoost0.94862.55661.87710.88713.623314
CatBoost0.94422.66241.9880.8953.39514
GradientBoosting0.93652.84091.95530.88973.548814
ExtraTrees0.92473.09342.56610.84194.300614
LightGBM0.9063.4562.64510.81034.658314
RandomForest0.90473.48052.80580.83114.496714
SVR0.87433.99743.0070.8124.562514
MLP0.85974.22233.56950.84284.167114
DecisionTree0.77325.36934.47920.75255.425214
KNeighbors0.68036.37414.84640.64646.629614
LinearRegression0.67676.41035.32250.57867.090414
Ridge0.67346.44285.32970.58047.0914
Lasso0.67316.44625.32760.57977.093114
ElasticNet0.67166.46065.33640.58067.090914
CatBoost0.88463.362.69980.92633.049215
XGBoost0.87193.53972.6020.90753.412115
GradientBoosting0.86183.67692.77950.88833.728615
ExtraTrees0.85513.76433.02830.89253.747515
RandomForest0.84743.86283.20670.86934.171815
LightGBM0.80214.39953.50070.85294.337215
SVR0.79564.47093.49070.84424.472215
DecisionTree0.76494.79553.58610.80225.063715
MLP0.74435.00074.07990.86224.203715
KNeighbors0.63825.9494.53330.72346.198615
LinearRegression0.59646.28334.7420.62467.005415
Lasso0.59396.30234.73710.62597.00415
Ridge0.59316.30834.74740.62637.001515
ElasticNet0.59156.32124.75630.62657.00215
CatBoost0.94422.77821.79310.9183.222716
XGBoost0.94012.87881.9920.91723.266116
GradientBoosting0.91653.39812.22450.91473.308716
ExtraTrees0.90793.56762.96410.88953.745516
RandomForest0.893.89993.06640.87254.037416
LightGBM0.88863.92552.82820.88643.86316
SVR0.85274.51363.44660.86544.138116
MLP0.80145.23984.1540.89343.646616
DecisionTree0.79075.37974.19560.76455.426816
KNeighbors0.77275.60644.32840.65466.636616
ElasticNet0.59747.46096.00420.63566.775116
Ridge0.59717.46356.00550.63566.775516
Lasso0.59677.46756.00730.63546.776916
LinearRegression0.59667.46816.00770.63546.777116
CatBoost0.94853.15852.41460.89673.370617
GradientBoosting0.93273.6112.53440.89113.482317
XGBoost0.92993.68562.69120.88873.53117
ExtraTrees0.92593.78923.09820.86613.891717
LightGBM0.91823.98242.99910.84634.181317
MLP0.91094.15563.00830.87063.733617
DecisionTree0.90464.29933.32450.7964.726717
RandomForest0.89454.52133.52570.8374.29717
SVR0.88224.77873.65410.81754.434717
KNeighbors0.74357.05015.36030.57916.874317
LinearRegression0.70617.54676.08460.60586.712717
Ridge0.70127.60996.16210.60746.70817
Lasso0.70047.61986.16640.60666.712517
ElasticNet0.69867.64296.19860.60756.708717
ExtraTrees0.84783.66952.9690.90473.642518
XGBoost0.84143.74642.75870.91663.38218
CatBoost0.82813.89972.92940.93562.976418
GradientBoosting0.82223.96612.83060.92053.299818
RandomForest0.81274.07113.31840.88763.96418
LightGBM0.80034.20383.00640.84934.589618
SVR0.74624.73923.63770.86874.300218
KNeighbors0.70285.12814.09290.68996.57418
MLP0.6675.4284.39370.89753.702418
DecisionTree0.57836.10834.87240.78015.512418
Lasso0.49686.67255.30740.64826.971318
ElasticNet0.49656.67435.30960.64826.970818
Ridge0.49546.68185.31270.64826.970618
LinearRegression0.49346.69545.31710.6486.970918
CatBoost0.92222.95472.32380.92723.033719
XGBoost0.91973.00162.31550.93262.961419
GradientBoosting0.91753.04212.30790.91343.341419
RandomForest0.88683.56352.96780.85884.266119
ExtraTrees0.88633.57142.81250.87753.96519
LightGBM0.8783.69982.74970.8544.298419
SVR0.83874.25473.07510.8544.331119
DecisionTree0.80874.63293.62580.79165.148119
MLP0.70555.74934.57970.86884.135319
KNeighbors0.59316.75795.11320.68396.398719
ElasticNet0.54617.1375.80420.65316.75419
Lasso0.54597.1395.80490.65316.753719
Ridge0.54587.13925.80520.65316.753819
LinearRegression0.54577.14015.80550.65316.753619
CatBoost0.95122.64571.97850.90533.385720
XGBoost0.94412.83112.03720.89443.609420
GradientBoosting0.92063.37312.42170.90023.528320
ExtraTrees0.90073.77172.92120.89043.670520
LightGBM0.89413.89682.57870.81574.770520
DecisionTree0.89033.96573.26830.80844.833120
RandomForest0.88794.00763.0380.86214.172920
SVR0.87154.29233.28190.8464.339320
MLP0.78375.56794.52530.84894.240720
KNeighbors0.68056.76694.79240.66526.586320
Ridge0.63137.26995.96440.60396.92920
ElasticNet0.63137.26985.95710.6046.929520
Lasso0.63127.27095.95210.60416.929920
LinearRegression0.63117.27135.9740.60366.928320
CatBoost0.90843.092.15860.92853.115921
GradientBoosting0.90433.15842.23390.89013.785421
XGBoost0.88483.46592.36190.9023.584621
LightGBM0.88453.46972.45310.84554.446921
RandomForest0.86533.74743.03380.8654.282121
ExtraTrees0.83374.16373.27270.88294.001921
SVR0.78214.76543.59450.84874.566721
DecisionTree0.69025.68264.67670.74795.839921
MLP0.68625.71974.72320.87454.180521
KNeighbors0.64946.04524.37150.69566.444621
ElasticNet0.50927.15255.83890.65826.738321
Ridge0.50877.1565.83880.65826.737921
Lasso0.50877.15655.83740.65826.737621
LinearRegression0.50787.1635.83830.65816.736621
CatBoost0.94622.52251.9980.91513.358322
XGBoost0.91933.08892.31590.91513.33822
ExtraTrees0.9193.09562.35660.88023.961622
LightGBM0.90733.31122.22620.85654.343922
GradientBoosting0.88863.62942.59890.90353.555722
RandomForest0.86763.95612.82710.8614.264322
SVR0.85184.18633.35040.83834.603122
MLP0.83844.37113.66080.84074.578722
DecisionTree0.74725.46744.53190.77695.405122
KNeighbors0.69585.9984.32740.69126.376322
ElasticNet0.65556.38235.38630.6466.852822
Lasso0.65526.38555.38230.6466.851922
Ridge0.65496.38855.38650.6466.852322
LinearRegression0.65346.40215.3870.6466.85222
CatBoost0.95072.69541.95220.92023.133523
SVR0.91923.45042.61180.864.188123
ExtraTrees0.91873.46032.73640.87733.916223
GradientBoosting0.91073.62632.42950.91553.279123
XGBoost0.90983.64522.48560.92823.036523
RandomForest0.90873.66822.88440.85494.251123
LightGBM0.86964.38313.0720.83494.532823
MLP0.86214.50773.74320.86354.121123
DecisionTree0.82855.02653.97140.75675.498923
KNeighbors0.78075.68444.21730.62316.914923
LinearRegression0.6896.76865.54310.59787.043323
Lasso0.6896.76925.54380.59787.043423
Ridge0.68886.77055.54530.59787.043423
ElasticNet0.68876.77185.54680.59787.043623
CatBoost0.95512.7832.16120.90233.101724
GradientBoosting0.94723.01792.30910.87863.43324
XGBoost0.94063.19992.50220.88983.307824
LightGBM0.92323.63942.66430.83754.046324
ExtraTrees0.91163.9043.06490.86643.689224
SVR0.87494.64463.64860.81824.248824
RandomForest0.86374.84783.93540.85163.979924
DecisionTree0.85554.99093.86070.8094.377924
MLP0.8465.15284.30060.84043.986124
KNeighbors0.71037.06685.84240.5996.750424
LinearRegression0.64527.82076.60260.5656.817424
Ridge0.64517.82236.60320.5656.817524
Lasso0.64517.82256.60310.5656.817624
ElasticNet0.64487.82586.60440.56516.817724
CatBoost0.93342.71042.16130.91043.36525
GradientBoosting0.89833.34882.22350.90123.53725
XGBoost0.89673.37522.40450.90963.340625
ExtraTrees0.89113.46462.63230.88023.928725
SVR0.87313.74072.84030.81334.890725
RandomForest0.85384.01563.11820.8574.260225
MLP0.83734.23513.56080.85344.315225
LightGBM0.82894.34312.90350.85314.327925
DecisionTree0.72445.51294.73060.74615.645125
KNeighbors0.69885.76314.48830.64426.911325
LinearRegression0.69245.82414.94440.59587.136325
Ridge0.69235.82474.94540.59587.136425
Lasso0.69235.82474.94550.59587.136425
ElasticNet0.69225.82614.94770.59587.136625
XGBoost0.94452.46571.87010.90293.565726
CatBoost0.94432.46871.84110.90423.553726
LightGBM0.92992.77071.93480.83854.553126
GradientBoosting0.92452.87572.1450.89683.679326
ExtraTrees0.90193.27732.62050.8933.771526
SVR0.8853.54762.92760.82374.846226
RandomForest0.85973.91882.95510.87594.067426
KNeighbors0.83334.27233.50090.65966.709426
MLP0.80814.58363.87460.84084.581126
DecisionTree0.7774.94133.93670.74245.779526
ElasticNet0.65666.13175.23440.63597.006126
Lasso0.65636.13435.23190.63587.006726
Ridge0.65596.13765.24440.63597.00626
LinearRegression0.65426.15235.26480.63537.008926
XGBoost0.92832.97491.97570.91783.200127
GradientBoosting0.91353.26772.25090.91413.28227
CatBoost0.89093.67022.46530.92143.096927
ExtraTrees0.88743.72783.05510.89013.686527
RandomForest0.86894.02323.21470.8674.073127
LightGBM0.84944.31143.08140.85724.292327
SVR0.8284.60723.62640.84644.390127
MLP0.82484.65083.72960.9033.47827
DecisionTree0.75915.45354.32570.75465.584727
KNeighbors0.74065.65834.29530.66236.81827
Ridge0.61586.88685.61010.63726.852527
ElasticNet0.61586.88685.61320.63726.852627
Lasso0.61576.88715.6090.63716.852527
LinearRegression0.61576.88695.60740.63716.852527
GradientBoosting0.91882.73431.96170.91123.437628
XGBoost0.9152.79712.08870.91323.348328
LightGBM0.90992.87982.1790.83564.763728
CatBoost0.89833.05952.31840.92813.123928
ExtraTrees0.89283.14162.54030.88563.936828
RandomForest0.87463.39682.5420.87724.101228
SVR0.83983.842.99410.8444.564628
MLP0.79054.39043.67760.86674.218328
DecisionTree0.74524.84233.91690.75335.743428
KNeighbors0.65755.61414.54890.68156.587128
ElasticNet0.5666.31975.39630.64886.897328
Ridge0.56486.32855.40840.64896.896728
Lasso0.5636.34175.42640.6496.895128
LinearRegression0.56266.34425.42960.64916.894928
CatBoost0.96272.31131.74250.91823.222729
XGBoost0.95032.66691.91020.89623.607929
GradientBoosting0.92283.32332.24580.8893.714529
LightGBM0.91373.51472.71740.84394.43929
ExtraTrees0.91363.51632.8570.88143.874829
RandomForest0.87754.18633.44660.85574.277529
SVR0.86184.44673.31730.85134.361229
MLP0.82624.9874.13730.84314.483929
DecisionTree0.80615.26723.96920.81754.792529
KNeighbors0.78765.5134.46980.66586.523229
Lasso0.57467.80246.63560.65676.627329
Ridge0.57457.80336.63620.65696.626229
ElasticNet0.57457.80296.6380.65696.625929
LinearRegression0.57447.80416.63650.65676.627729

References

  1. Meyer, C. The greening of the concrete industry. Cem. Concr. Compos. 2009, 31, 601–605. [Google Scholar] [CrossRef]
  2. Ahmaruzzaman, M. A review on the utilization of fly ash. Prog. Energy Combust. Sci. 2010, 36, 327–363. [Google Scholar] [CrossRef]
  3. Hardjito, D.; Rangan, B.V. Development and Properties of Low-Calcium Fly Ash-Based Geopolymer Concrete; Research Report GC 1; Faculty of Engineering; Curtin University of Technology: Perth, Australia, 2005. [Google Scholar]
  4. Hardjito, D.; Wallah, S.E.; Sumajouw, D.M.J.; Rangan, B.V. On the development of fly ash-based geopolymer concrete. ACI Mater. J. 2004, 101, 467–472. [Google Scholar] [CrossRef] [PubMed]
  5. Khale, D.; Chaudhary, R. Mechanism of geopolymerization and factors influencing its development: A review. J. Mater. Sci. 2007, 42, 729–746. [Google Scholar] [CrossRef]
  6. Nath, P.; Sarker, P.K. Effect of GGBFS on setting, workability and early strength properties of fly ash geopolymer concrete cured in ambient condition. Constr. Build. Mater. 2014, 66, 163–171. [Google Scholar] [CrossRef]
  7. Patankar, S.V.; Ghugal, Y.M.; Jamkar, S.S. Mix design of fly ash-based geopolymer concrete. In Advances in Structural Engineering; Springer: New Delhi, India, 2015. [Google Scholar] [CrossRef]
  8. Ansari, S.S.; Ibrahim, S.M.; Hasan, S.D. Conventional and ensemble machine learning models to predict the compressive strength of fly ash based geopolymer concrete. Mater. Today Proc. 2023, in press. [Google Scholar] [CrossRef]
  9. Gupta, P.; Gupta, N.; Saxena, K.K. Predicting compressive strength of geopolymer concrete using machine learning. Innov. Emerg. Technol. 2023, 10, 2350003. [Google Scholar] [CrossRef]
  10. Ahmad, A.; Ahmad, W.; Chaiyasarn, K.; Ostrowski, K.A.; Aslam, F.; Zajdel, P.; Joyklad, P. Prediction of geopolymer concrete compressive strength using novel machine learning algorithms. Polymers 2021, 13, 3389. [Google Scholar] [CrossRef] [PubMed]
  11. Dash, P.K.; Parhi, S.K.; Patro, S.K.; Panigrahi, R. Influence of chemical constituents of binder and activator in predicting compressive strength of fly ash-based geopolymer concrete using firefly-optimized hybrid ensemble machine learning model. Mater. Today Commun. 2023, 37, 107485. [Google Scholar] [CrossRef]
  12. Hu, T.; Zhang, H.; Cheng, C.; Li, H.; Zhou, J. Explainable machine learning: Compressive strength prediction of FRP-confined concrete column. Mater. Today Commun. 2024, 39, 108883. [Google Scholar] [CrossRef]
  13. Nguyen, T.K.; Huynh, T.A.; Dang, V.H.; Ahmed, A.; Thai, D.-K. From Machine Learning to Empirical Modelling: A Structured Framework for Predicting Compressive Strength of Fly Ash-Based Geopolymer Concrete. Buildings 2025, 16, 123. [Google Scholar] [CrossRef]
  14. Pham, V.N.; Nguyen, V.Q. Effects of alkaline activators and other factors on the properties of geopolymer concrete using industrial wastes based on GEP-based models. Eur. J. Environ. Civ. Eng. 2024, 28, 3770–3792. [Google Scholar] [CrossRef]
  15. Pham, V.N. Compressive strength of geopolymer concrete. Mendeley Data 2023, V1. [Google Scholar] [CrossRef]
  16. Alsalman, A.; Assi, L.N.; Kareem, R.S.; Carter, K.; Ziehl, P. Energy and CO2 emission assessments of alkali-activated concrete and ordinary Portland cement concrete: A comparative analysis of different grades of concrete. Clean. Environ. Syst. 2021, 3, 100047. [Google Scholar] [CrossRef]
  17. Davidovits, J. Geopolymer Chemistry and Applications; Institut Géopolymère: Saint-Quentin, France, 2008. [Google Scholar]
  18. Leonelli, C.; MacKenzie, K.J.; Seo, D.K.; Kriven, W.M. Geopolymer and alkali-activated materials: Chemistry, structure, and properties. Front. Chem. 2022, 10, 929163. [Google Scholar] [CrossRef] [PubMed]
  19. Shi, C.; Krivenko, P.V.; Roy, D. Alkali-Activated Cements and Concretes; Taylor and Francis: London, UK, 2003. [Google Scholar]
  20. Provis, J.L.; Yong, C.Z.; Duxson, P.; van Deventer, J.S.J. Correlating mechanical and thermal properties of sodium silicate–fly ash geopolymers. Colloids Surf. A Physicochem. Eng. Asp. 2009, 336, 57–63. [Google Scholar] [CrossRef]
  21. Habert, G.; d’Espinose de Lacaillerie, J.B.; Roussel, N. An environmental evaluation of geopolymer-based concrete production: Reviewing current research trends. J. Clean. Prod. 2011, 19, 1229–1238. [Google Scholar] [CrossRef]
  22. Temuujin, J.; van Riessen, A.; MacKenzie, K.J.D. Preparation and characterization of fly ash-based geopolymer mortars. Constr. Build. Mater. 2010, 24, 1906–1910. [Google Scholar] [CrossRef]
  23. Pacheco-Torgal, F.; Labrincha, J.; Leonelli, C.; Palomo, A.; Chindaprasit, P. Handbook of Alkali-Activated Cements, Mortars and Concretes; Elsevier: Oxford, UK, 2014. [Google Scholar]
  24. Miettinen, K. Nonlinear Multiobjective Optimization; Springer: New York, NY, USA, 1999. [Google Scholar] [CrossRef]
  25. Archetti, F.; Candelieri, A. Software Resources. In Bayesian Optimization and Data Science; Springer International Publishing: Cham, Switzerland, 2019; pp. 97–109. [Google Scholar]
  26. Wattanasiriwech, S.; Nurgesang, F.A.; Wattanasiriwech, D.; Timakul, P. Characterisation and properties of geopolymer composite part 1: Role of mullite reinforcement. Ceram. Int. 2017, 43, 16055–16062. [Google Scholar] [CrossRef]
  27. Assi, L.N.; Deaver, E.E.; Ziehl, P. Effect of source and particle size distribution on the mechanical and microstructural properties of fly ash-based geopolymer concrete. Constr. Build. Mater. 2018, 167, 372–380. [Google Scholar] [CrossRef]
Figure 1. Experiment Flowchart.
Figure 1. Experiment Flowchart.
Sustainability 18 02227 g001
Figure 2. Correlation of the input variables.
Figure 2. Correlation of the input variables.
Sustainability 18 02227 g002
Figure 3. Pareto front and objectives.
Figure 3. Pareto front and objectives.
Sustainability 18 02227 g003
Figure 4. Correlation plot of selected variables and target variable.
Figure 4. Correlation plot of selected variables and target variable.
Sustainability 18 02227 g004
Figure 5. Predicted vs. actual plot for baseline random forest model.
Figure 5. Predicted vs. actual plot for baseline random forest model.
Sustainability 18 02227 g005
Figure 6. Variance Inflation Factor (VIF) values for input variables.
Figure 6. Variance Inflation Factor (VIF) values for input variables.
Sustainability 18 02227 g006
Figure 7. Mean R2 scores of the model (30 runs).
Figure 7. Mean R2 scores of the model (30 runs).
Sustainability 18 02227 g007
Figure 8. Mean RMSE scores of the model (30 runs).
Figure 8. Mean RMSE scores of the model (30 runs).
Sustainability 18 02227 g008
Figure 9. Feature importances plot.
Figure 9. Feature importances plot.
Sustainability 18 02227 g009
Figure 10. Comparison of predicted and true f c [MPa] values for Holdout Set.
Figure 10. Comparison of predicted and true f c [MPa] values for Holdout Set.
Sustainability 18 02227 g010
Table 1. Descriptive statistics of input features.
Table 1. Descriptive statistics of input features.
VariableMeanSDMinMax
Fly Aah (FA)255.214391.8794120.000530.000
GGBFS164.699699.88520.000360.000
Coarse Aggregate991.2118199.9069525.40001276.8000
Fine Aggregate (Sand)744.304884.6714514.000196.4000
Na2SiO3 Solution149.243526.716076.2500102.8571
NaOH59.887511.716530.500102.8571
Na2O18.42453.911511.210028.8700
SiO244.53427.988622.420057.80600
Water (1)86.28715.551642.6238112.8214
Concentration8.46892.21292.000016.000
Water (2)39.911110.002013.420073.6000
NaOH (Dry)19.97645.23943.496038.4000
Additional water20.629353.13910.0000170.000
Superplasticizer11.72017.37650.000018.000
Total water153.857162.087656.0438303.5336
Initial curing time0.28940.45430.00001.000
Initial curing temp37.304021.254924.000100.000
Initial curing rest time0.90480.31810.00002.0000
Final curing temp28.32972.694224.00030.000
Table 2. Weight combinations for objectives.
Table 2. Weight combinations for objectives.
ScenarioWeights Combinations
(VIF, R2, RMSE, MultiR)
S1(0.05, 0.45, 0.45, 0.05)
S2(0.05, 0.50, 0.40, 0.05)
S3(0.05, 0.40, 0.50, 0.05)
S4(0.10, 0.40, 0.40, 0.10)
S5(0.10, 0.35, 0.45, 0.10)
S6(0.10, 0.45, 0.35, 0.10)
S7(0.15, 0.35, 0.35, 0.15)
S8(0.15, 0.40, 0.30, 0.15)
S9(0.15, 0.30, 0.40, 0.15)
S10(0.20, 0.30, 0.30, 0.20)
S11(0.20, 0.35, 0.25, 0.20)
S12(0.20, 0.25, 0.35, 0.20)
Table 3. Calculation method for Total Multiple R.
Table 3. Calculation method for Total Multiple R.
Single Variable Multiple R: Calculated for each Independent Variable (IV) separately
Step 1: Define regression problem as Equation (1)

Step 2: Calculate R2 (Coefficient of Determination) according to Equations (2)–(4)

Step 3: Calculate Multiple Correlation R according to Equation (5)
  where R_i ∈ [0, 1]
Total Multiple R: Calculated as an overall metric as given in Equation (6)
Properties:
  - Range: [0, V]
  - Minimum: 0 (perfect independence of IVs)
  - Maximum: V (complete redundancy of IVs)
Table 4. Pseudocode for Pareto front analysis.
Table 4. Pseudocode for Pareto front analysis.
ALGORITHM: Pareto Front Analysis and Optimal Selection

INPUT:
  pareto_front: Set of non-dominated solutions from NSGA-II
  X_scaled: Standardized feature matrix
  y: Target variable
  weights: [w1, w2, w3, w4] objective weights

OUTPUT:
  optimal_solution: Selected best compromise solution
  pareto_df: DataFrame with all solutions and metrics
  best_by_objective: Dictionary of best solutions per objective

PROCEDURE:

1. Initialize result storage
  pareto_solutions ← []

2. FOR EACH individual IN pareto_front DO

  (a) Decode binary representation
     variable_mask ← individual.binary_vector
     selected_vars ← DECODE(variable_mask)
     n_vars ← COUNT(selected_vars)

  (b) Extract subset data
     X_subset ← X_scaled[:, selected_vars]

  (c) Calculate VIF
     max_vif ← CALCULATE_MAX_VIF(X_subset, selected_vars)

  (d) Evaluate model performance (5-fold CV)
     r2, rmse ← EVALUATE_RANDOM_FOREST(X_subset, y)

  (e) Calculate Total Multiple R
     total_r ← 0
     individual_r_scores ← []

     FOR EACH var_idx IN selected_vars DO
     r_value ← CALCULATE_SINGLE_R(var_idx, X_subset)
     individual_r_scores.APPEND(r_value)
     total_r ← total_r + r_value
     END FOR

     avg_r ← total_r/n_vars

  (f) Store comprehensive metrics
     solution ← {
     ‘solution_id’: i,
     ‘n_variables’: n_vars,
     ‘max_vif’: max_vif,
     ‘r2’: r2,
     ‘rmse’: rmse,
     ‘total_multiple_r’: total_r,
     ‘avg_multiple_r’: avg_r,
     ‘variables’: selected_vars,
     ‘individual_r_scores’: individual_r_scores,
     ‘fitness’: individual.fitness.values
     }
     pareto_solutions.APPEND(solution)

3. Create DataFrame
  pareto_df ← DATAFRAME(pareto_solutions)

4. Identify best solutions by individual objectives
  best_by_objective ← {
     ‘vif’: pareto_df[pareto_df[‘max_vif’].IDXMIN()],
     ‘r2’: pareto_df[pareto_df[‘r2’].IDXMAX()],
     ‘rmse’: pareto_df[pareto_df[‘rmse’].IDXMIN()],
     ‘total_r’: pareto_df[pareto_df[‘total_multiple_r’].IDXMIN()]
  }

5. Calculate balanced (i.e., weighted) scores for all solutions
  FOR EACH solution IN pareto_df DO
     // Normalize objectives
     norm_vif ← 1/(1 + solution[‘max_vif’]/100)
     norm_r2 ← solution[‘r2’]
     norm_rmse ← 1/(1 + solution[‘rmse’]/10)
     norm_total_r ← 1/(1 + solution[‘total_multiple_r’]/10)

     // Weighted aggregation
     balanced_score ← w1 × norm_vif +
            w2 × norm_r2 +
            w3 × norm_rmse +
            w4 × norm_total_r
     solution[‘balanced_score’] ← balanced_score
  END FOR

6. Select optimal solution
  optimal_solution ← pareto_df[pareto_df[‘balanced_score’].IDXMAX()]

7. RETURN optimal_solution, pareto_df, best_by_objective

END ALGORITHM
Table 5. Initial Parameter Values for the Experiment.
Table 5. Initial Parameter Values for the Experiment.
ParameterValue
GA Population size60
GA Generations40
Objective weights 12 scenarios
Dataset variables 19
Minimum variables to keep5
Cross-validation folds5
Random state42
Table 6. Optimal solutions for genetic algorithm runs with different weight scenarios.
Table 6. Optimal solutions for genetic algorithm runs with different weight scenarios.
ScenarioWeights (VIF, R2, RMSE, MultiR)N VarsMax VIF of SolutionR2RMSETotal Multi-R
S1(0.05, 0.45, 0.45, 0.05)936.010.91563.30448.2536
S2(0.05, 0.50, 0.40, 0.05)936.010.91563.30448.2536
S3(0.05, 0.40, 0.50, 0.05)936.010.91563.30448.2536
S4(0.10, 0.40, 0.40, 0.10)711.760.91423.32945.7507
S5(0.10, 0.35, 0.45, 0.10)711.760.91423.32945.7507
S6(0.10, 0.45, 0.35, 0.10)711.760.91423.32945.7507
S7(0.15, 0.35, 0.35, 0.15)711.760.91423.32945.7507
S8(0.15, 0.40, 0.30, 0.15)711.760.91423.32945.7507
S9(0.15, 0.30, 0.40, 0.15)711.760.91423.32945.7507
S10(0.20, 0.30, 0.30, 0.20)711.760.91423.32945.7507
S11(0.20, 0.35, 0.25, 0.20)711.760.91423.32945.7507
S12(0.20, 0.25, 0.35, 0.20)711.760.91423.32945.7507
Table 7. Variable selection stability across weight scenarios.
Table 7. Variable selection stability across weight scenarios.
VariableTimes SelectedScenarios
Concentration(M) NaOH12/12S1, S2, S3, S4, S5, S6, S7, S8, S9, S10, S11, S12
NaOH (Dry)12/12S1, S2, S3, S4, S5, S6, S7, S8, S9, S10, S11, S12
GGBFS12/12S1, S2, S3, S4, S5, S6, S7, S8, S9, S10, S11, S12
Initial curing temp (C)12/12S1, S2, S3, S4, S5, S6, S7, S8, S9, S10, S11, S12
Coarse Aggregate12/12S1, S2, S3, S4, S5, S6, S7, S8, S9, S10, S11, S12
Na2O (Dry)9/12S4, S5, S6, S7, S8, S9, S10, S11, S12
Initial curing rest time (day)9/12S4, S5, S6, S7, S8, S9, S10, S11, S12
Additional water3/12S1, S2, S3
Initial curing time (day)3/12S1, S2, S3
FA3/12S1, S2, S3
Na2SiO33/12S1, S2, S3
Table 8. S4 Pareto front optimal solutions.
Table 8. S4 Pareto front optimal solutions.
Solution IDN VarsMax VIFR2RMSETotal Multiple RBalanced Score (ws)Optimal
151.78660.89023.76922.97060.2000False
252.50930.89743.64013.19310.4082False
352.69970.89273.7293.17850.2662False
452.82350.90323.53593.40960.576False
552.96390.89053.76883.13130.2005False
663.11360.91093.39434.38890.7919False
773.72540.91223.36985.47450.8129False
8711.61820.91243.36465.65810.8129False
9711.76110.91423.32945.75070.8673True
101026.21260.91583.3049.22270.8487False
11626.87150.91163.38144.68170.7932False
12936.01410.91563.30448.25360.8548False
13937.86160.91423.33087.80090.8189False
1410162.96580.91643.28789.41480.7855False
1510167.63640.9173.27819.34370.8011False
Table 9. Best solution per objective for S4.
Table 9. Best solution per objective for S4.
CriterionSolution IDN VarsVIFR2RMSEMulti-R
Min VIF151.790.89023.76922.9706
Max R21510167.640.9173.27819.3437
Min RMSE1510167.640.9173.27819.3437
Min Multi-R151.790.89023.76922.9706
Table 10. Average metrics for fine-tuned models (30 runs).
Table 10. Average metrics for fine-tuned models (30 runs).
ModelHoldout R2Holdout RMSEHoldout MAECV R2CV RMSE
CatBoost0.92372.9952.2240.91523.2209
XGBoost0.91673.14882.28670.9063.3978
Gradient Boosting0.90543.37222.39040.89963.5108
ExtraTrees0.88943.67992.97010.87823.9005
LightGBM0.87763.85022.78920.84494.406
Random Forest0.87163.99033.13520.85974.2089
SVR0.85134.26113.29990.83544.5054
MLP0.80644.82273.95770.85794.1708
Decision Tree0.77525.17864.15510.7755.2726
KNeighbors0.71425.96844.58520.65326.688
Lasso0.61566.89715.67160.61846.9271
ElasticNet0.61546.89945.67470.61856.9271
Ridge0.61546.89815.6740.61856.9266
Linear Regression0.61536.89765.67430.61836.9264
Table 11. Performance metrics for fine-tuned models (30th run).
Table 11. Performance metrics for fine-tuned models (30th run).
ModelHoldout R2Holdout RMSEHoldout MAECV R2CV RMSESeed
CatBoost0.96272.31131.74250.91823.222729
GradientBoosting0.92283.32332.24580.8893.714529
LightGBM0.91373.51472.71740.84394.43929
ExtraTrees0.91363.51632.8570.88143.874829
RandomForest0.87754.18633.44660.85574.277529
SVR0.86184.44673.31730.85134.361229
MLP0.82624.9874.13730.84314.483929
DecisionTree0.80615.26723.96920.81754.792529
KNeighbors0.78765.5134.46980.66586.523229
Lasso0.57467.80246.63560.65676.627329
Ridge0.57457.80336.63620.65696.626229
ElasticNet0.57457.80296.6380.65696.625929
LinearRegression0.57447.80416.63650.65676.627729
Table 12. Comparative summary of existing literature and the present study on geopolymer compressive strength prediction.
Table 12. Comparative summary of existing literature and the present study on geopolymer compressive strength prediction.
LiteratureData/Experimental ConditionsMethods/AnalysisKey Findings
Pham and Nguyen (2024) [14]Broadly compiled industrial geopolymer mixes (various FA/GGBFS blends).Gene Expression Programming (GEP)/experimental data compilation.Highlighted the strong effect of alkaline activators and raw material composition.
Hardjito and Rangan (2005) [3]Low-calcium FA, controlled curing (high temperature).Experimental mechanical tests; microstructural analysis.Demonstrated that heat curing accelerates early strength gain, with responses varying by mix composition.
Wattanasiriwech et al. (2017) [26]Sequential curing at 90 °C, followed by humid and then dry curing at 40 °C.Experimental strength testing.Showed that multi-stage curing can produce consistently high strengths.
Assi (2016) [27]Experimental tests and microstructural evaluation.Reported that silica fume-based systems are less sensitive to curing conditions than sodium silicate systems.Reported that silica fume-based systems are less sensitive to curing conditions than sodium silicate systems.
Present Study274 experimental records covering 19 variables (FA, GGBFS, Na2SiO3, NaOH, curing parameters, aggregates, water, SP, etc.).Data preprocessing, correlation and network analysis, VIF calculation, multi-objective GA for feature selection, hyperparameter tuning; models: RF, XGBoost, LightGBM, CatBoost, SVR, MLP, etc.Best performance achieved with CatBoost (R2 = 0.9583). GGBFS identified as most influential feature; NaOH (Dry) is also critical for estimation.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ahadian, F.; Işıkdağ, Ü.; Bekdaş, G.; Nigdeli, S.M.; Cakiroglu, C.; Geem, Z.W. Interpretable Machine Learning for Compressive Strength Prediction of Fly Ash-Based Geopolymer Concrete. Sustainability 2026, 18, 2227. https://doi.org/10.3390/su18052227

AMA Style

Ahadian F, Işıkdağ Ü, Bekdaş G, Nigdeli SM, Cakiroglu C, Geem ZW. Interpretable Machine Learning for Compressive Strength Prediction of Fly Ash-Based Geopolymer Concrete. Sustainability. 2026; 18(5):2227. https://doi.org/10.3390/su18052227

Chicago/Turabian Style

Ahadian, Farnaz, Ümit Işıkdağ, Gebrail Bekdaş, Sinan Melih Nigdeli, Celal Cakiroglu, and Zong Woo Geem. 2026. "Interpretable Machine Learning for Compressive Strength Prediction of Fly Ash-Based Geopolymer Concrete" Sustainability 18, no. 5: 2227. https://doi.org/10.3390/su18052227

APA Style

Ahadian, F., Işıkdağ, Ü., Bekdaş, G., Nigdeli, S. M., Cakiroglu, C., & Geem, Z. W. (2026). Interpretable Machine Learning for Compressive Strength Prediction of Fly Ash-Based Geopolymer Concrete. Sustainability, 18(5), 2227. https://doi.org/10.3390/su18052227

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop