1. Introduction
Alum sludge (also known as red mud or bauxite tailings) is a byproduct of the alum used in the drinking water treatment process [
1]. In recent decades, several studies have examined the use of alum sludge as a soil stabilizer [
2]. As a soil stabilizer, alum sludge has the main advantage of improving the soil’s physical properties, such as strength and stability [
3]. The reason for this, is that alum sludge contains high levels of alum, which reacts with the soil to form a solid and stable structure [
4]. As a result of this reaction, soil density and compaction increase, improving soil stability and reducing erosion sensitivity [
4]. In addition, for pure aluminum sludge, the stability gained by adapted scheduling of bauxite tailings, is critical to warrant safe tailings deposition [
5].
Alum sludge, a byproduct of the water treatment process, can be used as a soil stabilizer, providing several environmental benefits. Firstly, it can reduce soil erosion. The application of alum sludge to soil can help to reduce soil erosion by increasing its stability and reducing its susceptibility to erosion from wind or water [
6]. Secondly, by improving soil structure, alum sludge can also help to reduce runoff [
7], which can carry pollutants and sediment into nearby waterways, damaging aquatic ecosystems. Thirdly, alum sludge can reduce water usage in agriculture [
8], by improving soil moisture retention, reducing the need for irrigation and saving water resources. Moreover, by improving soil structure and nutrient availability, alum sludge can reduce the use of fertilizers [
9], which can contribute to water pollution and environmental degradation. Furthermore, using alum sludge can reduce greenhouse gas emissions [
10] by reducing the emissions associated with the production and transportation of synthetic fertilizers and irrigation. Alum sludge offers another advantage from an environmental standpoint, which is a reduction in waste in landfills [
11]. Alum sludge can be diverted from landfills and used as a soil stabilizer, reducing the amount of waste going into landfills and its associated environmental impacts. Finally, the utilization of alum sludge as a soil stabilizer promotes sustainable agriculture [
12], by improving soil health and productivity while reducing environmental impacts.
According to Nguyen et al. [
4], alum sludge as a soil stabilizer reduces the soil’s flexibility and makes it less susceptible to deformation. This reduction in flexibility results in improved soil structure and increased stability [
4]. Moreover, on the geotechnical side, Nguyen et al. [
4] showed that, the addition of sludge to the soil, increases the undrained compression strength (UCS) of the soil. In addition, several studies have demonstrated that alum sludge can be an effective alternative to traditional soil stabilizing agents, such as cement and lime [
13,
14,
15].
Alum sludge as a soil stabilizer can have several environmental advantages [
16]. The first benefit, is that it reduces the amount of waste produced by the alum production process, by recycling the sludge and reusing it for soil stabilization [
16]. Furthermore, alum sludge reduces the need for conventional soil stabilizers, which are often derived from non-renewable sources and may emit large amounts of carbon dioxide [
17]. In addition to improving soil stability, alum sludge can also improve soil chemical properties [
16]. Studies have shown, that alum sludge can neutralize soil acidity and improve soil pH [
18,
19]. The importance of this, is that soil acidity can inhibit plant growth and reduce crop yields [
18]. As a result, better water infiltration and retention occurs in the soil, which is essential for the growth and development of plants. Additionally, improving soil permeability can reduce soil erosion and contribute to improving soil health [
20]. Alum sludge is cost-effective when used as a soil stabilizer, compared to traditional soil stabilizers such as cement, coal, and fly ash [
21].
As a soil stabilizer, alum sludge has been shown to have varying efficiency, depending on the type of soil and the conditions of its application. As an example, clay soils are naturally active, so the use of alum sludge may greatly benefit them [
22]. Similarly, the effectiveness of alum sludge as a soil stabilizer can be affected by environmental factors such as temperature, humidity, and rainfall [
23]. Studies have shown that it is important to consider the specific soil and environmental conditions when using alum sludge as a soil stabilizer [
23,
24].
Various factors have been identified as influencing the behavior of sludge. There are several parameters to consider, including the soil and sludge density, the specific gravity of the soil, the liquid limit and plasticity of the soil, the sludge content, etc. With such a multiplicity of parameters and nonlinearity, it would be necessary to predict performance using an equation with a large number of parameters, that would require quantification, which does not exist. In the past two decades, artificial intelligence (AI) methods have been used to resolve this issue. Using artificial intelligence techniques, it is possible to determine the relationship between different parameters with a high degree of accuracy, without prior knowledge. Various topics in geotechnical engineering, such as slope stability [
25,
26,
27], tunneling [
28,
29,
30], pavement and road construction [
31,
32], soil cracking [
33,
34,
35], rock mechanics [
36,
37], soil dynamics [
38,
39,
40,
41], and soil stabilizers [
42,
43,
44] have been addressed using artificial intelligence methods [
45]. Nevertheless, only two studies have used artificial intelligence to predict the properties arising from mixing sludge with soil [
46,
47]. Aamir et al. [
46] used a small database of 18 datasets to predict the CBR parameter, using an artificial neural network (ANN). A high
R2 of 0.99 was observed, as well as a small RMSE of 0.057. Similarly, Shah et al. [
47] used a 21-set database and artificial neural network (ANN) method to predict the CBR of mixtures of sludge and soil. The results of the study were very promising, with an
R2 of 0.97 and an RMSE of 0.58. However, both studies used only one classic artificial intelligence method, ANN, and grey box methods were not employed. Grey box methods are those methods that have an output such as a tree or an equation. In addition, the used databases were relatively small, which may limit the range of values of the parameters.
In order to fill these gaps, this study uses two black-box AI methods, namely an artificial neural network (ANN) and support vector machines (SVM), as well as a grey-box AI method, namely genetic programming (GP), in order to predict CBR. By using this approach, both black-box and grey-box models will be examined. The ANN method that is currently used as standard, will be compared with another black-box method, SVM. SVM and GP have been widely used in the field of civil engineering for various applications. For example, SVM has been used to predict the remaining service life of aging infrastructure, such as bridges and buildings [
48,
49,
50]. SVM has also been used for damage detection and identification in structures, using vibration data [
51,
52]. SVM has found application in predicting the degradation of pavements and estimating their remaining service life [
53,
54]. Additionally, SVM has been utilized for classifying the types of distresses on pavement surfaces and forecasting their friction coefficient [
55]. GP has been utilized for optimizing the operation and design of various systems in civil engineering, including water supply [
56] and stormwater management systems [
57], and for analyzing flood forecasting and dam failure [
58]. Furthermore, GP has been applied to optimize the design of various structures, including trusses [
59], frames [
60], and bridges [
61], as well as for the placement of sensors to monitor the structural health of buildings [
62].
In this study, a database of 27 CBR test results, on a variety of soil types, is used for this purpose. Initially, there are nine input parameters from laboratory tests, whose values are discussed in detail. The sensitivity of the AI models, as well as the importance of the input parameters, is evaluated.
2. Data-Driven Modeling
After analyzing the collected database, the CBR was predicted using three artificial intelligence methods, including artificial neural network (ANN), genetic programming (GP), and support vector machine (SVM). The hypothesis of the performed study, is that artificial intelligence (AI) methods can be utilized to predict the California bearing ratio (CBR), which is a crucial parameter in geotechnical engineering, that measures the load-bearing capacity of soil, of soil stabilized with alum sludge, a byproduct of water treatment plants.
The California bearing ratio (CBR) is a penetration test performed on a compacted sample of soil or aggregate, to evaluate its strength and bearing capacity. It is a standardized test method used to determine the relative strength of a soil sample by measuring the pressure required to penetrate the sample with a plunger of standard area at a standardized rate. The CBR value is expressed as a percentage of the pressure required to penetrate a standard material, typically crushed rock or limestone, at the same penetration depth and under the same conditions. A higher CBR value indicates greater strength and load-bearing capacity of the soil or aggregate sample. The CBR test is widely used in civil engineering and construction to determine the design parameters of roads, airfields, and other infrastructure projects.
The use of AI algorithms for material characterization and design, has been met with skepticism, due to concerns about the reliability of their complex models. The lack of transparency and knowledge extraction processes in AI-based models, is a major challenge. Mathematical modeling techniques can be categorized as white-box, black-box, and grey-box, depending on their level of transparency (
Figure 1). While white-box models are based on first principles and provide an explanation of the underlying physical relationships of a system, black-box models do not provide any feasible structure of the model. Grey-box models, on the other hand, identify the patterns between the data, and provide a mathematical structure of the model. ANN is a popular black-box modeling technique widely used in engineering, but its weights and bias representation does not provide details about the derived relationships. GP is a newer grey-box modeling technique, that uses an evolutionary process to develop explicit prediction functions, making it more transparent than other ML methods, especially black-box methods such as ANN and SVM. The mathematical structures derived by GP can be used to gain important information about the system’s performance.
In civil engineering, SVM has been applied for predicting soil properties [
65], classifying building materials [
66], and detecting structural damage [
67]. For instance, SVM has been used for predicting the shear strength of soil, by using the geotechnical parameters as input variables [
68]. SVM has also been employed for detecting structural damage, by using the vibration signals from a structure as input variables [
69].
In civil engineering, GP has been used for optimization problems, such as the optimal design of structures and construction scheduling [
70,
71]. For example, GP has been applied for optimizing the design of steel frames in high-rise buildings, by evolving the design variables such as beam size, column size, and connection details [
72].
2.1. Support Vector Machine (SVM)
Support vector machine (SVM) is one of supervised learning models. SVM initially developed by Vapnik [
73], weights are calculated based on input data, known as training data, in order to learn the governing function from a set of inputs. SVM selects a limited number of input sample vectors, which are always a fraction of the total number of samples. These input vectors are referred to as support vectors. Using these input vectors, the parameter values that minimize a cost function are calculated. As a result, SVM requires much less data than similar methods such as ANN, and as a result, it takes much less time and has a lower cost.
The goal of this method, is to find a classifier that separates the data and maximizes the distance between these two classes.
Figure 2 shows a plane cloud, that separates two sets of points.
According to
Figure 3, the closest points used to determine the hyperplane, are called support vectors. As is clear, there are many hypermaps that can divide samples into two categories. The principle of SVM, is to choose something that maximizes the minimum distance between the hypermap and the training samples (that is, the distance between the hypermap and the support vectors), this distance is called the margin.
The kernel function, as an important part of the SVM method, receives data as input and transforms them into the required form. Various functions are provided for this purpose. Some of these functions, including linear, nonlinear, polynomial, radial basis function (RBF), and sigmoid, are given in Equations (1)–(5). According to Equations (1)–(5), the kernel function is a series of mathematical functions that provide a window for data manipulation. With the help of this transformation, a complex and nonlinear level of decision making becomes a linear equation, but in a larger number of dimensional spaces.
where
xi and
xj are vectors in the input space, the
γ and
σ parameters define the distance influence of a single sample,
d is the degree of the polynomial,
c ∈
R is the relative position to the origin,
κ’ is a real-valued positive type function/kernel, and
α,
T, and
c are certain values.
In this study, two important functions, i.e., the radial basis function (RBF) and Gaussian kernel, which have been successful in past research, were evaluated, along with two other functions, i.e., the sigmoid kernel and polynomial kernel functions.
2.2. Artificial Neural Network (ANN)
Artificial neural networks (ANNs) have a history that dates back to the mid-20th century, when the mathematical model for the perceptron was first introduced by Rosenblatt [
74]. The development of the backpropagation algorithm in the 1980s, by Rumelhart et al. [
75], improved the training of multi-layer ANNs and led to their widespread use in various fields. Since then, ANNs have continued to evolve, with the development of new architectures and learning algorithms [
76].
An artificial neural network (ANN) is a machine learning model that is based on the structure and function of the human brain [
76]. Artificial neural networks are composed of interconnected nodes, or artificial neurons, that process information and make predictions in response to inputs. By combining the inputs and the weights assigned to each connection, the interconnected nodes produce an output. There are a number of applications for artificial neural networks, including image classification, speech recognition, and natural language processing, among others. In the field of machine learning and artificial intelligence, they have proven to be effective at solving complex problems.
Due to their ability to learn from data and make accurate predictions or classifications based on that data, artificial neural networks (ANNs) have become one of the most powerful tools in the field of machine learning and artificial intelligence [
77]. As a result, they simulate the structure and function of the human brain, which is composed of many neurons connected by synapses.
Each node or artificial neuron in an ANN, performs simple computations based on the input it receives and the weights assigned to the connections [
45]. One neuron’s output is then passed on as input to other neurons in the network, allowing the information to be processed and transformed multiple times before it is finally produced. Feedforward, refers to this process of processing and transforming information through multiple layers of neurons.
There are various types of artificial neural networks, each designed for a particular task or application. The most popular types of ANNs include feedforward networks, recurrent networks, and convolutional neural networks [
78]. An ANN can be used for a variety of purposes, and each type has its own strengths and weaknesses.
As the most basic type of ANN, feedforward networks are used for simple tasks, such as binary classification and linear regression. On the other hand, recurrent networks are used for sequences of data, such as speech or text, and are capable of capturing temporal relationships. Image or video analysis can be performed using convolutional neural networks, which are specifically designed to capture the spatial relations between pixels in an image or video.
Artificial neural networks (ANNs) can be used to model complex relationships with no established mathematical relationships [
79]. The ANN method consists of artificial neurons, and is commonly used to fit nonlinear statistical data. For this purpose, ANN identifies the relationships between input and output parameters based on the strength of the connection between the two neurons, known as the weight. Next, the network seeks to optimize the matrix of weights, which is achieved through practice and adjustment, which is called the paradigm [
80]. One of the strongest paradigms, is the backpropagation paradigm. In this study, two backpropagation algorithms are used: (i) the Levenberg–Marquardt (LM) algorithm [
81], and (ii) the Bayesian regularization (BR) [
82] algorithm.
Input, hidden, and output layers are the three main components of ANN architecture. In this study, for each algorithm, i.e., LM and BR, one, two, three, four, and five hidden layers are considered. Log-sig and tan-sig are transfer functions within each neuron. In this study, five different architectures of ANN, with one to five hidden layers, were modelled for each algorithm. Furthermore, by trial and error, 45 neurons were considered for each hidden layer and each network was re-trained three times.
2.3. Genetic Programming (GP)
John Koza, a computer scientist and researcher at Stanford University, developed the first genetic programming system in the early 1990s. Using this system, he evolved computer programs capable of solving mathematical problems and controlling robots. Several genetic programming frameworks and toolkits were developed as a result of this development.
As a subfield of artificial intelligence and evolutionary computation, genetic programming (GP) is a method for solving complex problems, using a process inspired by natural selection and genetics. In this approach, potential solutions are represented as trees of operations (also known as computer programs) and then improved using genetic algorithms.
The GP algorithm generates an initial population of computer programs at random and evaluates their fitness in accordance with a problem-specific objective function. Through genetic operators such as crossover and mutation, the best programs are selected, to produce the next generation. The process continues until a satisfactory solution is found or a stopping criterion is met.
A wide range of problems can be solved using GP, including function approximation, symbolic regression, and even game play. There are several strengths of GP, including its ability to find complex solutions that are difficult or impossible to discover manually, as well as its ability to handle high levels of uncertainty and noise in the data. It can, however, be computationally expensive, and may have difficulty finding solutions in a reasonable amount of time, for very complex problems.
4. Results
4.1. Support Vector Machine (SVM)
In order to derive an optimal SVM model, multiple parameters were adjusted, and the resulting model with the most favorable performance is delineated in this section.
Table 4 presents the specifications of the best support vector machine (SVM) model derived through parameter tuning. The model was constructed using the sequential minimal optimization (SMO) algorithm, with a penalty parameter of
C = 2, and a tolerance value of 0.001. The epsilon parameter, which controls the width of the margin, was set to 0.5. The input data was pre-processed using standardization, to ensure that each feature had a mean of zero and a standard deviation of one. The kernel function used in this model was the radial basis function (RBF), which is a popular kernel function for SVMs due to its flexibility in modeling complex, nonlinear relationships between input variables. The gamma parameter of the RBF kernel was set to 0.5, which controls the smoothness of the decision boundary. A lower value of gamma, results in a smoother decision boundary, while a higher value of gamma, leads to a more complex boundary, that can overfit the data. Overall, these specifications provide a well-performing SVM model, that can accurately classify data while avoiding overfitting. These parameters can be used as a starting point for future SVM modeling tasks or as a benchmark for comparing the performance of other SVM models.
The predicted values of CBR against the values obtained from the tests in the laboratory for the training and testing database, are shown in
Figure 7. The results show that the SVM model has been able to predict the CBR values with a good accuracy.
Table 5 presents the performance metrics of the SVM used to predict the California bearing ratio (CBR) values for mixtures of alum sludge and soil. Both training and testing datasets were used to evaluate the performance of the SVM model.
The first performance metric is the mean absolute error (MAE), which measures the average absolute difference between the predicted and actual CBR values. In the training and testing sets, the values are 0.497 and 0.512, respectively, indicating that the model has an average error of approximately 0.5. A second performance metric is the mean squared error (MSE), which measures the average squared difference between the predicted and actual CBR values. The training and testing sets have values of 0.409 and 0.357, respectively, indicating that the model has a relatively low overall error in predicting the CBR values. The third performance metric, is the root mean squared error (RMSE), which is the square root of the mean square error (MSE). The training and testing sets have values of 0.640 and 0.598, respectively, indicating that the model has a relatively low overall error in predicting the CBR values. The fourth performance metric is the mean squared log error (MSLE). The MSLE measures the difference between the logarithm of the predicted CBR values and the logarithm of the actual CBR values. For the training and testing sets, the values are 0.017 and 0.011, respectively, indicating that the model has a relatively low overall error when predicting the CBR values. The fifth performance metric is the root mean squared log error (RMSLE), which is the square root of the MSLE. In the training and testing sets, the values are 0.129 and 0.103, respectively, which indicates that the model has a relatively low overall error in predicting the CBR values, on a logarithmic scale. The final performance metric is the R2, which measures the amount of variance in the CBR values that can be explained by the model. As can be seen from the values for the training and testing sets, the model has a high degree of correlation between the predicted and actual CBR values. The results suggest that the SVM model is capable of predicting CBR values for mixtures of alum sludge and soil, as indicated by the low error values and high R-squared values, for both the training and testing sets.
4.2. Artificial Neural Network (ANN)
In this study, all artificial neural network (ANN) modeling was performed using MATLAB (R2020a: The Math Works Inc., Natick, MA, USA).
Table 6 displays the results of the Bayesian regularization (BR) and Levenberg–Marquardt (LM) algorithms for all networks. The table contains the
R2 and mean absolute error (MAE) values for both test and training datasets. The results are based on the best-performing network among all 45 neurons and three training repetitions, as determined by the highest
R2 and lowest MAE values.
The findings indicate, that the network with two hidden layers achieved the highest performance for both the BR and LM algorithms. Moreover, the average accuracy of the BR algorithm was superior to that of the LM algorithm.
Figure 8 depicts the comparison between the predicted values of the artificial neural network (ANN) model and the actual values of laboratory experiments. The results indicate that the ANN model exhibits a very high level of precision in predicting the California bearing ratio (CBR) values for both the training and testing databases.
Table 7 displays the best outcomes of utilizing the ANN modeling method to predict the California bearing ratio (CBR) for the mixtures of alum sludge and soil. The performance indicators of the ANN model are exhibited for both the training and testing datasets. The ANN model yielded a mean absolute error (MAE) of 0.392 and 0.303, for the training and testing datasets, respectively; the mean squared error (MSE) values for the training and testing datasets were 0.200 and 0.116, respectively; the root mean squared error (RMSE) values obtained for the training and testing datasets were 0.447 and 0.341, respectively; the mean squared log error (MSLE) values were 0.005 and 0.002, for the training and testing datasets, respectively; while the root mean squared log error (RMSLE) values were 0.074 and 0.046, for the training and testing datasets, respectively. The ANN model yielded a high coefficient of determination (
R2), of 0.989 and 0.980, for the training and testing datasets, respectively, implying that the model is highly precise in predicting the CBR values for mixtures of alum sludge and soil.
One of the key determinants of the precision and complexity of an artificial intelligence (AI) model, is the number of neurons present in each hidden layer. To find the optimal number of neurons, a group of analyses is conducted.
Figure 9a shows the accuracy of the ANN model, based on the number of neurons present in each hidden layer. According to the results, the accuracy of the ANN model does not improve significantly beyond a certain number of neurons, which in this case is 10 neurons. Therefore, it is essential to determine the optimal number of neurons, to achieve the best accuracy while minimizing the complexity of the model.
Figure 9b displays the model’s error rate for each neuron. The results show that the error rate remains almost constant after the 10th neuron. Thus, the optimal number of neurons for the ANN model was determined to be 10. This information is crucial for researchers and developers of AI models, as it allows them to minimize the complexity while achieving the best accuracy. By optimizing the number of neurons, the performance of the AI model can be significantly enhanced.
4.3. Genetic Programming (GP)
The genetics of programming involves applying evolutionary algorithms to generate computer programs through the use of genetic operators, such as crossover and mutation. Important parameters that affect the effectiveness of genetic programming, include population size, mutation rate, and fitness evaluation function. In this study, after a series of analyses, the best GP model was selected, based on its ability to produce high-quality programs for a given problem domain.
Figure 10 shows the projected values of CBR by the best GP model, against the actual values obtained in the experiments, for both the training and testing sets. According to the findings presented in
Figure 10, the best genetic programming (GP) model was capable of accurately predicting the values of the California bearing ratio (CBR) in the experiments, as indicated by the close match between the model’s projected values and the experimental values. These results suggest that the GP model has a high level of accuracy in predicting CBR values.
A significant distinction between genetic programming (GP) and black-box models, is that the GP produces an equation as an outcome, which can be directly utilized by the reader. In this study, Equation (13) represents the GP model outcome for predicting California bearing ratio (CBR) values. The reason for the length of this equation, is that it incorporates seven distinct inputs, namely liquid limit, plasticity index, alum sludge content, number of compaction blows, optimum moisture content of soil and mixture, and
Gs of soil. It is worth noting, that traditional methods are incapable of producing an equation such as Equation (13), with the consideration of all seven influential inputs. Consequently, this study represents the first successful attempt to formulate an equation for predicting CBR numbers based on these seven important inputs. It is important to note, that the parameters in Equation (6) are normalized values, and values should be normalized based on
Table 1 before being used in Equation (13).
where
X1,
X2,
X3,
X4,
X5,
X6, and
X7 are LL, PI, AS content, number of compaction blows, OMC of soil, OMC of mixture, and G
s of soil, respectively.
Table 8 presents the results of the GP model in predicting the California bearing ratio (CBR) for the mixtures of alum sludge and soil. The GP model’s performance was evaluated using various performance metrics on both the training and testing datasets. The mean absolute error (MAE) of the model on the training database was 0.459, indicating that, on average, the model’s predictions were off by 0.459 CBR units. The MAE on the testing database was slightly higher, at 0.475, indicating that the GP model’s generalization ability was slightly worse than its performance on the training database. The mean squared error (MSE) on the training database was 0.358, indicating that the GP model’s predictions had a higher variance than its performance on the testing database (0.304). The root mean squared error (RMSE) on the training database was 0.598, indicating that the GP model’s predictions had a higher magnitude of error than its performance on the testing database (0.552). The mean squared logarithmic error (MSLE) on the training set was 0.010, and the corresponding value on the testing set was slightly lower, at 0.009, indicating that the GP model’s predictions were more accurate on the testing set. The root mean squared logarithmic error (RMSLE) on the training set was 0.100, and the corresponding value on the testing set was slightly lower, at 0.097. Finally, the coefficient of determination (
R2) was high for both the training set (0.980) and the testing set (0.948), indicating that the model explained most of the variance in the data, and that it had a good fit to the data. In conclusion, the results suggest that the GP model can be used to accurately predict CBR values for mixtures of alum sludge and soil, and that it generalizes well to unseen data. However, the model’s predictions on the training set had a higher variance and magnitude of error, compared to the testing set.
One benefit of using an equation generated by genetic programming, is that it can potentially provide a more accurate prediction of the CBR value, compared to traditional methods. The equation is generated based on the data and relationships observed in the inputs and outputs, allowing for a more customized and precise model. Another benefit, is that the generated equation can be easily implemented and applied to new data. Once the equation is generated, it can be used to predict the CBR values of new soil samples, without requiring additional testing or experimentation. A further benefit of the generated equation, is that it can reduce testing time and costs. Traditional methods of predicting CBR values require extensive testing and experimentation. Using a generated equation, can significantly reduce the time and cost involved in testing, as the equation can provide accurate predictions without the need for additional experimentation.
5. Discussion
5.1. Comparison of Different Models
Table 9 compares the results of three artificial intelligence (AI) models, namely artificial neural network (ANN), support vector machine (SVM), and genetic programming (GP), in predicting the California bearing ratio (CBR), using different performance metrics, for both the training and testing databases. The performance metrics used to evaluate the models’ performance include, mean absolute error (MAE), mean squared error (MSE), root mean squared error (RMSE), mean squared logarithmic error (MSLE), root mean squared logarithmic error (RMSLE), and coefficient of determination (
R2).
In addition,
Table 10 shows the ranking of the different AI models. The SVM model was the first AI model examined in this study, and is known as a black-box model. The ANN method, another black-box model, was subsequently employed, to explore another black-box model’s performance. The GP method was utilized to examine grey-box AI methods.
The results demonstrate, that the SVM model exhibits the lowest accuracy for both the training and test datasets, but still, it shows a good performance. In contrast, the ANN method performs significantly better, and predicts CBR values with high accuracy, achieving an R2 of 0.989 and 0.980 for the training and test databases, respectively. The GP method, a grey-box model, achieves slightly lower accuracy than the ANN method but still performs well, with an R2 of 0.98 and 0.948 for the training and testing databases, respectively. Notably, the GP method offers the benefit of producing an equation as output, which can be used for other datasets.
The reason for the observed results, is that the SVM and GP models are known to be better suited for problems with a small number of inputs and a large number of data points, whereas ANN models are better suited for problems with a large number of inputs and a small number of data points. These findings contribute to a better understanding of the performance of different AI models for predicting CBR, which may have implications for future research and practical applications in the field of civil engineering and transportation planning.
5.2. The Variable Importance of Input Parameters
Investigating the significance of the input parameters, is a crucial aspect of artificial intelligence modeling. In this study, the impact of the individual input parameters on network error was examined, by altering each parameter by 100%, while maintaining all other inputs at their actual values. The resulting network errors were recorded for each alteration, and are presented in
Figure 11 for each of the three AI models tested, namely artificial neural network (ANN), support vector machines (SVM), and genetic programming (GP). A greater network error resulting from a given parameter alteration, indicates that the network exhibits increased sensitivity to that particular parameter. The variable importance ranking was evaluated based on the input parameters, including liquid limit of soil (LL-Soil), plasticity index of soil (PI-Soil),
Gs of soil, number of compaction blows (Compaction Num. Hammer), optimum moisture content of soil (OMC-Soil), maximum dry density of soil (MDD-Soil), optimum moisture content of mixture (OMC-mixture), maximum dry density of mixture (MDD-Mixture), and sludge content (% Sludge).
Table 11 summarizes the ranking of inputs based on the different AI models.
The results reveal, that the SVM model ranked the LL of soil as the most important input parameter, followed by the number of compaction blows and the PI of soil. While the MDD of soil was the least important parameter. The ANN ranked the number of compaction blows as the most important input parameter, followed by the PI and LL of soil. While the MDD of the mixture was the least important parameter. The GP ranked the number of compaction blows as the most important input parameter, followed by sludge content and LL of soil. For all three models, and according to
Table 11, it can be concluded that the number of compaction blows and LL of soil parameters, are the most important input parameters, while the MDD of soil and mixture, are the least important.
The explanation for the variation in the importance ranking of input parameters across AI models, could be the differences in the underlying algorithms and architectures. SVM is a kernel-based method, that tries to find a hyperplane that maximally separates data points. ANN is a feedforward neural network with hidden layers, and it is trained using a backpropagation algorithm. GP is an evolutionary algorithm, that evolves a population of candidate solutions using genetic operators, such as mutation and crossover. Therefore, these algorithms have different ways of handling the input parameters and constructing models.
Another explanation for the variation in importance ranking, could be the nature of the input parameters themselves. Some parameters may have stronger correlations with the target variable, which is CBR, in certain datasets or applications, while others may be less relevant. Moreover, different input parameters may interact with each other in complex ways, making it challenging to determine their individual contributions to the model’s performance.
The California bearing ratio (CBR) is a measure of the soil’s load-bearing capacity and is widely used in geotechnical engineering to assess the suitability of soils for construction. The CBR test involves measuring the resistance of a soil sample to penetration by a standard plunger, under controlled conditions of moisture and compaction. The stiffness and load-bearing capacity of a soil depend on various factors, including its texture, structure, moisture content, density, and compaction effort. The liquid limit of soil is an important input parameter for predicting CBR, because it reflects the soil’s ability to resist deformation and support loads. A soil with a high liquid limit is more plastic and less stable, which results in lower CBR values. In contrast, a soil with a low liquid limit is more rigid and stable, leading to higher CBR values. The number of compaction blows is another critical input parameter for predicting CBR, because it represents the compaction effort applied to the soil during construction. The more compaction blows applied, the higher the soil density and stiffness, which result in higher CBR values. Conversely, inadequate compaction effort leads to low soil density and stiffness, resulting in lower CBR values.
Regarding the maximum dry density, it is an indicator of the soil’s compactability and weight. However, it does not directly relate to the soil’s stiffness and load-bearing capacity, which are the primary factors affecting CBR. Therefore, while it may affect CBR values to some extent, its influence is relatively weak compared to other parameters, such as liquid limit and compaction effort.
5.3. Sustainability
The performed study is highly significant in terms of sustainability. The use of alum sludge as a soil stabilizer can provide an alternative to traditional stabilizers, which can be expensive and have negative environmental impacts. By utilizing alum sludge, this study offers a sustainable solution for improving soil stability in various engineering applications. Additionally, the application of artificial intelligence (AI) methods to predict the California bearing ratio (CBR) of soils stabilized with alum sludge, provides a more efficient and accurate way to determine the effectiveness of this stabilizer. This can reduce the need for extensive and costly field testing, leading to a more sustainable and cost-effective approach to engineering projects. The results of this study also provide valuable insights into the sensitivity and importance of input parameters, which can guide future research and development in the field of soil stabilization.
5.4. Limitations and Scope for Future Works
Although this study provides valuable insights into the use of AI methods for predicting the CBR of soil–alum sludge mixtures, there are limitations that need to be addressed in future work. Firstly, the study only considered nine input parameters, and there may be other important parameters that were not included. Therefore, future research could explore the effect of additional parameters, such as the pH value of the sludge and the particle size distribution of the soil, on the CBR prediction accuracy.
Secondly, the study only used a limited number of CBR test results, which may not be representative of all possible soil–alum sludge mixtures. Thus, future work could expand the database to include more CBR test results, from a wider range of soil and sludge combinations, to further evaluate the accuracy and robustness of the AI models.
Thirdly, while the study evaluated the sensitivity and importance of the input parameters, it did not consider the interaction effects between the parameters. Therefore, future work could investigate the interaction effects, to provide a more comprehensive understanding of the CBR prediction process.
Fourthly, while the AI models used in this study can accurately predict CBR values, they do not provide physical insights into the underlying mechanisms. Understanding the physical processes that govern the strength and stability of soil–alum sludge mixtures, could help to improve the accuracy and reliability of the AI models. Therefore, future work could combine AI methods with physical modeling approaches, to gain a deeper understanding of the system’s behavior.
Finally, the study focused on predicting the CBR of soil–alum sludge mixtures for geotechnical engineering applications. However, the use of alum sludge has potential applications in other fields, such as agriculture and water treatment. Thus, future work could explore the use of AI methods to predict other parameters relevant to these fields, such as soil pH and nutrient availability.
6. Conclusions
The use of alum sludge in geotechnical engineering, has gained importance due to its cost-effectiveness and environmental benefits, as it has been found to improve the strength and stability of soils. The CBR test is a crucial parameter in geotechnical engineering, which measures the load-bearing capacity of soil. However, predicting the CBR of soil–alum sludge mixtures can be challenging, due to the large number of input variables. This highlights the significance of utilizing artificial intelligence (AI) methods to overcome this challenge, as this approach has not been widely used previously in this field.
This study found that artificial intelligence (AI) methods can be utilized to predict the CBR, which is a crucial parameter in geotechnical engineering, that measures the load-bearing capacity of soil, of soil–alum sludge mixtures. The use of AI methods can overcome the challenge of predicting CBR, due to the large number of input variables. The study compared the performance of three different AI models (two black-box models and one grey-box model), using a database of 27 CBR test results, and found that the artificial neural network (ANN) and genetic programming (GP) methods outperformed the support vector machines (SVM) method. The database consisted of nine parameters, including liquid limit of soil, plasticity index of soil, specific gravity of soil, number of compaction blows, optimum moisture content of soil, maximum dry density of soil, optimum moisture content of mixture, maximum dry density of mixture, and sludge content.
The ANN method achieved high accuracy, with R2 of 0.989 and 0.980 for the training and test databases, respectively. The GP method had slightly lower accuracy, but had the advantage of producing an equation as output, that can be used for other datasets. The study also determined the optimal number of neurons for the ANN model, which is 10 neurons. The examination of parameter sensitivity and importance, indicated that the number of compaction hammer blows, and the soil’s liquid limit were the most significant parameters, while the maximum dry density parameters for soil and mixture were the least significant. This information is important for engineers and researchers, to optimize their soil stabilization process. Overall, the study provides valuable insights into the performance of different AI models for predicting CBR and can be useful for future research and practical applications.