In Situ Skin Friction Capacity Modeling with Advanced Neuro-Fuzzy Optimized by Metaheuristic Algorithms

: Developing new optimization algorithms and data mining has improved traditional engineering structural analysis models (meaning basically swarm-based solutions). Additionally, an accurate quantiﬁcation of in situ friction capacity (ISFC) of driven piles is of paramount importance in design/construction of geotechnical infrastructures. A number of studies have underscored the use of models developed via artiﬁcial neural networks (ANNs) in anticipation of the bearing capacity of driven piles. Nonetheless, the main drawbacks of implementing the techniques relying on artiﬁcial neural networks are their slow convergence rate and reliable testing outputs. The current research focused on establishing an accurate/reliable predictive network of ISFC. Therefore, an adaptive neuro-fuzzy inference system (ANFIS) coupled with Harris hawk optimization (HHO), salp swarm algorithm (SSA), teaching-learning-based optimization (TLBO), and water-cycle algorithm (WCA) is employed. The ﬁndings revealed that the four models could accurately assimilate the correlation of ISFC to the referenced parameters. The values of the root mean square error (RMSE) realized in the prediction phase were 8.2844, 7.4746, 6.6572, and 6.8528 for the HHO-ANFIS, SSA-ANFIS, TLBO-ANFIS, and WCA-ANFIS, respectively. The results depicted WCA-ANFIS as more accurate than the three other algorithms at the testing and training phase, and could probably be utilized as a substitute for laboratory/classical methods.


Introduction
The process of discovering the optimum combination of a set of decision variables to solve a certain issue is referred to as an optimization algorithm. The issue of engineering optimization could be considered a constrained optimization problem, and it is one of the most significant challenges in practical problems [1][2][3]. It is challenging to identify the optimal worldwide solution to complicated optimization problems appearing in diverse domains, including engineering, economics, and medicine, employing traditional mathematical optimization approaches. The strategy has the benefit of being quicker to implement due to its decreased complexity. Nevertheless, its drawback is that it could get stuck in the local region, preventing it from obtaining the optimal global solution. Swarm intelligence optimization algorithms that simulate the behavior of actual organisms and have effectively resolved numerous complicated optimization problems [4][5][6] are also categorized into two main types: evolutionary algorithms and swarm intelligence algorithms [7]. Evolutionary algorithms rely on a biological evolution-inspired system that comprises four operators: random selection, reproduction, recombination, and mutation. The swarm intelligence algorithm is a population-based algorithm that has emerged through social behavior. Meta-heuristic optimization algorithms evolve rapidly [8,9] due to their simple idea, adaptability, and capability to prevent local optima, and are frequently employed to solve various complicated real-world optimization problems [10]. Meta-heuristics could be separated into three primary types depending on the algorithmic inspiration: techniques does the capacity for exploration, and it has a proven outstanding performance in several real-time optimizations compared to other metaheuristic algorithms [38]. Because of such advantages, HHO and its variants have lately been commonly employed for real-world challenges. HHO has greatly succeeded in various domains and applications [39][40][41][42]. Wang et al. [43] suggested an enhanced hybrid Aquila Optimizer (AO) and HHO that outperforms the basic AO and HHO in terms of global search performance and convergence speed compared to standard and CEC2017 benchmark functions. Alabool et al. [44] carried out a comprehensive review on HHO and its variant, and postulated that it is easy to execute, well structured, and flexible. Nevertheless, since there is no one ideal optimizer for all potential computing tasks, population diversity, convergence rate, and the balance between exploration and exploitation of HHO must all be improved in multi-objective, complicated, and composite optimization situations [45]. Chantar et al. [46] suggested another hybrid optimizer solution called "Binary Harris Hawks Optimizer," hybridized with the timevarying scheme (briefly called BHHO-TVS) for the classification process [47]. The proposed technique achieves the greatest accuracy rates on 67 percent of datasets compared to similar feature selection algorithms published in previous studies. In addition, the HHO could be used to resolve problems involving unidentified kinds of search space and discrete and continuous spaces [48,49], to improve solution quality [50][51][52], extract optimum parameters with high precision [53,54], and improve prediction performance [55,56]. In 2017, Mirjalili et al. [57] developed a novel swarm intelligence optimization algorithm whose optimization approach was inspired by the salp swarm chain foraging in the ocean (also called the SSA). Due to its simple design and basic implementation, the essential SSA random search optimization technique attracted several researchers' attention. SSA's searching behavior as a meta-heuristic algorithm is separated into two major phases: exploration and exploitation. During the exploration phase, it can effectively find the search space mostly by randomization, although it may encounter unexpected modifications. In the period of exploitation, it converges on the most promising location. SSA may therefore be used to solve a variety of research difficulties. SSA seems to be an innovative and promising technique previously applied successfully in several situations. It has a distinct benefit in simply having one parameter (r 1 ) for balancing exploration and exploitation. Abualigah et al. [58] carried out a critical review of SSA. They hypothesized that SSA is highly feasible for continuing community employment while highlighting its weaknesses, limitations, and drawbacks. Invoking the no free lunch (NFL) theorem, no optimization algorithm can solve all optimization problems. Because of the limited set F of benchmark functions that must equalize with the objective function in F, SSA may require adjustment and alteration when solving certain real-world optimization issues. Second, SSA has a single objective function, which allows it to handle only single-objective optimization problems. Specific operators must solve binary, discrete, continuous, dynamic, multi-objective, and other issues [59]. The major disadvantage of SSA is its limited capacity to regulate the challenges of multimodal search methods since all three parameters (a, v, and F) appear to converge to the identical solution. Due to its stochastic character and shortage of balance between exploration and exploitation, SSA suffers from the issue of delayed convergence and is frequently trapped in local optima.
Rao et al. [60] popularized the TLBO algorithm, lacking any set of rules and specification data, and applied mechanical design optimization problems to show superiority over other population-based optimization in terms of the best solution, average solution, convergence rate, and computational effort with the potential to be easily extended to other engineering design optimization problems like prediction ISFC. It employs the impact of a teacher's influence on students, separated into the "Teacher Phase" and "Learner Phase." It needs only simple governing parameters like the magnitude of population and number of generations to function. TLBO is a nature-inspired population-based algorithm utilizing a population of solutions to proceed to the global solution and is widely accepted by optimization researchers. TLBO, which appears to follow a similar traditional approach to the instructor's teaching method and the studying or exam process of the students, depending on the physical phenomena of learning, is an essential heuristic algorithm for optimization applications [61][62][63]. The population size with iteration number is the most essential condition of the TLBO. It consistently ignores parameters in its work and has a quick convergence velocity and good searching accuracy. Since the TLBO commenced its work, it has successfully resolved various difficulties because it has attracted the interest of scientists [64] in several fields. Chen et al. [65] combined TLBO with learning enthusiasmbased (LebTLBO) to analyze three excellent control issues in chemical engineering. The results depict its potential capability to achieve the intended purpose. Zhao et al. [66] hybridized MLP with shuffled complex evolution (SCE) as SCE-MLP and TLBO-MLP ensembles to predict CSC; prediction errors were lowered. In addition, SCE is a considerably more time-efficient optimizer concerning computation time.
Eskandar et al. [67] suggested the water cycle algorithm. The inspiration for WCA originated from viewing nature and researching the water cycle and analyzing how streams and rivers flow downhill toward the sea in the natural world. A river or stream is formed when water flows from a higher point to a lower one. As a result, most rivers originate at the summits of mountains, where snow melts. In turn, the rivers continually flow down, replenishing water from rainfall and other streams along the way until ending up in the sea. Even though the innovative water cycle algorithm may solve the issue of entrapment at local optima, its efficiency as a spam classifier is still questionable. Unlike other optimization algorithms, including the Harmony Search algorithm, the WCA has just three control parameters. Employing an evaporation strategy, WCA as a feature selector may handle the problem of quick convergence of local optima entrapment [68]. The analysis and development of innovative ways for dealing with FS and cruse dimensionality difficulties are still active research areas, notably for spam classifiers. As a result, the FS methods are being evaluated for the purposes mentioned above: (a) improved performance (perhaps prediction accuracy or learning speed); (b) data simplification for model selection; and (c) removal of redundant or unnecessary features (dimensionality reduction) [69]. Noting that, feature selection or reduction strategies have been used in several types of research during the last ten years. Such a strategy has proven practical outputs for many published examples.
Furthermore, this study has concentrated on enhancing the effectiveness of the method. This section covers a variety of relevant technologies and methodologies depending on metaheuristic techniques [70], such as local search and population-based methods for both hybrid and heuristics metaheuristics [71][72][73][74][75]. Moayedi and Mosavi [76] employed WCA-MLP to test a finite dataset to estimate the settlement, and the outcomes are correlated with electromagnetic field optimization (EFO) and SCE benchmarks [77]. It was discovered that the training error of WCA is less than that of EFO and SCE, while the accuracy of WCA-MLP is more prominent in the testing phase rather than the learning phase [76]. The WCA is a powerful investigation approach widely utilized in various applications, as demonstrated in Foong et al. [78] and Nasir et al. [79]. WCA gained popularity because of its capacity to offer an optimum solution while swiftly converging [80]. WCA was effectively used for diverse engineering in a short period [81][82][83]. The advantage of WCA over other explored approaches is demonstrated in [84,85], which is why the algorithm continues to attract increasing attention from academics. WCA has been used lately to predict the bearing capacity of shallow foundations [76]. Multiple successful applications of HHO, SSA, TLBO, and WCA inspired the researcher to modify and use hybridized versions of the above algorithms for the first time to solve the problematic issue of ISFC prediction [86][87][88][89]. However, the uncertainty from theoretical, experimental, and numerical methods of friction capacity estimation for piles under axial compressive loading resulting from complexity of soil-pile interaction, prompted utilization of metaheuristic approach to reliably and accurately predict an in situ friction capacity of piles. Goh [36] stated that the length and diameter of the pile, shear strength, and vertical stress are four factors affecting ISFC prediction. These factors are incorporated to predict the ISFC of the pile with an ANFIS hybridized with HHO, SSA, TLBO, and WCA.
The remaining sections of the paper are structured as follows. Using the information from the prior research, Section 2 provides definitions for the various significant terms. The strategies and procedures that we have developed for our work with deep neural networks are detailed in Section 3. The exploratory configuration employed to execute the recommended strategy and the findings produced for categorizing default customers are described in Section 4. Conclusions and prospective research plans are covered in Section 5.

Established Database
Sixty-five in situ tests were conducted to generate the datasets. The data from Goh [90] were used to develop the optimal structure (e.g., GOA, WDO, SHO, and MFO) using MLP models. Additionally, the results of pile load tests are collected, as are data on the nearby soil properties. As a result, the training and testing sets are built on the basis of extensive in situ investigations. Both models are trained on a dataset of 52 field experiments, and then tested on a dataset of 11 tests. Both the input data (for example, the pile length (m) and diameter (cm)) and the output data (for example, the friction capacity of installed shafts) and output data layers are shown graphically in Figure 1.
Geotechnics 2022, 2, FOR PEER REVIEW 5 factors affecting ISFC prediction. These factors are incorporated to predict the ISFC of the pile with an ANFIS hybridized with HHO, SSA, TLBO, and WCA. The remaining sections of the paper are structured as follows. Using the information from the prior research, Section II provides definitions for the various significant terms. The strategies and procedures that we have developed for our work with deep neural networks are detailed in Section III. The exploratory configuration employed to execute the recommended strategy and the findings produced for categorizing default customers are described in Section IV. Conclusions and prospective research plans are covered in Section V.

Established Database
Sixty-five in situ tests were conducted to generate the datasets. The data from Goh [90] were used to develop the optimal structure (e.g., GOA, WDO, SHO, and MFO) using MLP models. Additionally, the results of pile load tests are collected, as are data on the nearby soil properties. As a result, the training and testing sets are built on the basis of extensive in situ investigations. Both models are trained on a dataset of 52 field experiments, and then tested on a dataset of 11 tests. Both the input data (for example, the pile length (m) and diameter (cm)) and the output data (for example, the friction capacity of installed shafts) and output data layers are shown graphically in Figure 1.

Adaptive Neuro-Fuzzy Inference Hybrid (ANFIS)
To develop a set of fuzzy 'if-then' rules with a proper membership function for generating the specified input-output pairs, Jang [91] presented ANFIS. This system can classify approximation highly nonlinear functions, identify discrete control systems online, and predict chaotic time series [92]. The input-output data optimize the membership functions (MFs). ANFIS involves tuning a FIS system using a backpropagation algorithm depending on input collected data. A FIS's basic structure includes three conceptual elements: a rule base that consists of a set of fuzzy rules; a database that describes the membership functions (MFs) utilized in the fuzzy rules; and a reasoning mechanism that performs the inference procedure on the rules and the provided facts to produce a reasonable output or conclusion. Such intelligent systems integrate information, techniques, and procedures from numerous sources. They have human-like skills in a particular domain-they adapt and can perform efficiently in changing environments. In ANFIS, neural networks identify patterns and aid in environmental adaptation. FIS integrates human expertise and executes interface and decision-making [93]. Besides, ANFIS is adjusted via a backpropagation approach utilizing input-output data. Investigators utilized fuzzy theory extensively to illustrate complicated processes by applying ideas and if-then rules. Researchers utilized the fuzzy theory due to its ability to solve complex procedures using the if-then rules approach. The theory of decision-making inspirited it in human life [42,94]. As a result, the artificial neural network (ANN) [92] has been applied to optimize the fuzzy theory as ANFIS to acquire self-learning capabilities [95]. Jang [91] proposes ANFIS as an amalgamation of ANN with a fuzzy method. It has been hypothesized that the ANFIS is preferable to the FIS for analyzing nonlinear issues [96]. According to the ANFIS, a FIS is used in a multilayer feed-forward network for training [97]. FIS membership function (MF) parameters may be learned using the ANFIS's training of the input data by combining least-squares approaches with backpropagation gradient descent. According to Termeh et al. [98], the ANFIS structure consists of five different layers where neurons in each layer contain adaptive nodes expressed by Equations (1) and (2): The input neurons are defined as x and y. A and B signify linguistic variables, whereas µA i (x) and µB i (y) represent the proposed node's MFs.
In the 2nd layer, Equation (3) represents the output of each node, which is the result of all input signals to the suggested node: where W i represents the output of each node. The layer three nodes consist of the normalized outputs of layer two, while for layer four, a node function is utilized to link every node, as shown in Equations (4) and (5): where w l indicates the firepower normalized by layer three p i , q i, and r i specified for the node parameters.
In layer four, the parameters are considered to be result parameters. The total of all input signals is regarded as a single signal to the output. Layer 5 considers the total of all input signals to the output as a single node: HHO method is firstly proposed by Heidari et al. [37] to solve numerous optimization problems through adopting great teamwork. The algorithm addresses numerous optimization issues by simulating the cooperative behavior of Harris's hawks. Hawks catch their prey by tracing, encircling, approaching, and eventually attacking. Exploration and exploitation are the two major stages of the HHO. Waiting, searching, and identifying the prospective prey are part of the initial phase. Let P rabit represent the rabbit position, while the hawks' position is described in the following: P rand is indeed one of the available hawks that are randomly suggested. In addition, r i (i = 1, 2, 3, 4, q) is a random integer in the range [0,1]. Moreover, P m represents the average position. Considering P i and N as the location of the hawks and their size, respectively, the following equation is used to determine P m .
In the second stage, let T and E 0 ∈ (−1.1) be the maximum size of the repetitions, and the initial energy, the escaping energy of the hunt (E) that could alter exploration and exploitation, is expressed as follows: Depending on the magnitude of |E|, it is determined to begin the exploration phase (|E| ≥ 1) or to exploit the solutions' neighborhood (|E|< 1). In the final phase, depending on the value of |E|, the hawks determine whether to apply a soft (|E|≥ 0.5) or hard (|E|< 0.5) besiege to capture the target from multiple directions. Interestingly, the parameter r is used to compute the target's escaping probability, if it is more significant than 0.5, the hunt effectively escapes, and if it is less than 0.5, it fails [37].

Salp Swarm Inspired Algorithm (SSA)
Salp swarm-inspired algorithm (SSA) is a new swarm intelligence algorithm evolved in 2017 by Mirjalili et al. [57]. SSA is a population-based strategy that simulates salp swarms' behavior and social interaction. Salp is a clear marine invertebrate residing in cold water and foraging on plankton. In simulating the salp foraging chain, the salp swarm must be separated into leaders and followers with two distinct update functions. The definitive process of SSA consists of three sections: initialization, the leader's determination, and updating. The definitive process of SSA consists of three sections: (i) initialization, (ii) determination of the leader, and finally, (iii) updating. The initialization of the objective function is defined by Equation (10): where N is the objective function's dimension. Assume that the following variables have constant boundaries: lb i represents the lower boundary of the i-th variable, whereas ub i represents the upper boundary. Assume the salp swarm's population is D, commonly known as the number of search agents. Every searching agent's initial position is specified as: where rand (D, N) is a D × N dimension matrix, of which every element is set to be a random number Rand (D, N) is a D x N matrix. Noting that every member is set by a random number between 0 and 1. Following the initialization of the salp position, each individual's position is inserted into the objective function (1), obtaining D fitness values. After that, the D fitness values are sorted, the least fitness value is identified, and the corresponding person is recognized as the leader's determination. In contrast, the other people are rated as followers from small to large. It is significant to mention that the leader is nearest to the food position (optimal position). As a result, the rows of matrix The leader position is updated as Equation (13): where X i is the turned to i-th variable of the leader position, F i is the individual with the best fitness value in the preceding frame, ub, and lb are the upper and lower bounds of the corresponding dimensional variables, c 1 , c 2 , c 3 are known as three control parameters, and c 2 , c 3 are both random numbers between 0 and 1. It seems essential to mention that the leader's position update is solely connected to the individual position with the optimum fitness value. It has nothing to do with the former positions of other individuals or the leader's status. The updating direction is set by c 3 , which may be adjusted or omitted from the initial basis. The significant parameter that defines the update step is c 1 . It is described as follows: where l and L are the current and maximum iterations and e is the natural constant. The lower the number of repetitions, the higher the value of c 1 . The great update step may accelerate the leader's approach to the optimum global region during the first few iterations. As the number of iterations reaches the maximum number of iterations, the value gradually approaches zero. The locations of the followers are then updated. The followers are attached throughout the foraging process, forming a chain-like combination shape. The chain formation highly influences individuals before and after the followers' aggressive moves. The position update is dependent on itself and the person in front of it. The formula for updating is as follows: where X i j denotes the j-th individual's i-th variable, and X j−i i indicates the j-th position of the adjacent individual in front of the j-th individual. All individuals' fitness values are computed after the leader and follower positions are updated independently. Allowing the followers to reach the leader, the leader's information is updated once again. A series of iterations complete the optimization.

Teaching-Learning-Based Optimization (TLBO)
TLBO consists of the "Teacher phase" and the "Learner phase." Rao [60] explains the operation of both phases. During the first phase of the algorithm, the learner receives instruction from the teacher. In contrast, the second part focuses on the learner's knowledge growth via interaction with peers and acquiring novel information from more experienced learners. At any iteration, i, suppose that there are 'm' number of subjects (i.e., design variables), 'n' number of learners (i.e., population size, k = 1,2, . . . ,n), and M j,i be the mean result of the learners in a specific subject 'j' (j = 1,2, . . . , m). The greatest learner, kbest, is the individual with the best overall score, Xtotal-kbest, when all topics are evaluated together, and the entire learner population is analyzed. Nevertheless, as the teacher is frequently observed as a competent person who instructs learners to achieve higher results, the algorithm determines the best learner to be the teacher. The difference between the current mean result for every topic and the teacher's outcome for every subject is represented by Equation (16): where X j,kbest,i is the result of the best learner in subject j. T F is the teaching factor that determines the new value of the mean, whereas r i is a random integer in the range [0,1]. T F may have a value of 1 or 2. T F s value is determined at random with the same probability as, where T F is not a TLBO algorithm parameter. The value of T F is not presented as an input to the method; instead, the algorithm utilizing Equation (17) determines it randomly. Following a series of studies on a variety of benchmark functions, it was determined that the algorithm performs better when the value of T F is between 1 and 2. Nevertheless, the technique performs considerably superior if T F is either 1 or 2; therefore, to simplify the algorithm, the teaching factor is recommended to be either 1 or 2 based on the rounding up requirements stated in Equation (17). According to the di f f erence_Mean j,k,i , the existing solution is updated in the teacher phase based on the following expression.
where X j,k,i is the updated value of X j,k,i . X j,k,i is accepted if it offers a higher function value. All acceptable function values from the teacher phase are preserved, and such values are used as input in the learner phase. The teacher phase determines the learner phase. The instructor phase influences the learner phase. On the other hand, the learner phase is the second phase of the algorithm, in which learners improve their knowledge by interacting with one another. A learner communicates with other learners at random to enhance their learning. A learner acquires new information if another learner has greater knowledge. Considering a population size of 'n,' randomly select two learners, P, and Q, such that X total−P,i = X total−Q,i (where, X total−P,i , i and X total−Q,i , I are the updated function values of Xtotal-P, i and Xtotal-Q, i of P and Q, respectively, at the end of teacher phase). X j,P,i = X j,P,i + r i X j,P,i − X j,Q,i i f X total−P,i < X total−Q,i X j,P,i = X j,P,i + r i X j,Q,i − X j,P,i i f X total−Q,i < X total−P,i X j,P,i is accepted if it offers a higher function value. Equations (19) and (20) can deal with the minimization issues. Equations (21) and (22) are employed in maximizing problems.
X j,P,i = X j,P,i + r i X j,P,i − X j,Q,i i f X total−Q,i < X total−P,i X j,P,i = X j,P,i + r i X j,Q,i − X j,P,i i f X total−P,i < X total−Q,i More detail could be found in the literature about TLBO in Yan-Kwang et al. [99], Harmandeep et al. [100], and Chen et al. [65].

Water-Cycle Algorithm (WCA)
The water-cycle algorithm (WCA) was designed by Eskandar et al. [67], replicating the water cycle process as an inspiration, imitating how rivers and streams end up in the sea. The steps necessary to execute the WCA could be described as follows: Step 1: Setting the initial parameters of the WCA algorithm, for instance, K sr , K pop , d max , It max.
Step 2: Scattering the initial population and determining sea, streams, and rivers. Assuming K pop as the total size of the population, K sr = 1 + the number of rivers, and K streams (= K pop − K sr ) as the number of streams, Equation (23) illustrates the procedure: where river, stream, and sea are each represented by a 1 K-dimensional array as [x1, x2, . . . , xK].
Step 3: The cost of each current population member is calculated by using the following: Step 4: Equations (3) and (4) yield the flow intensity of the sea and rivers given NS k as the number of streams emptying into the respective rivers or the sea: Step 5: The following relationships depict the streams and rivers that flow into the sea: where G is a variable number between 1 and 2 (close to 2), significantly, once C > 1, streams are permitted to enter rivers from diverse directions.
Step 6: The flow of rivers toward the sea (or downwards) may be expressed as follows: Step 7: The river is replaced with a stream that provides a better-fitting solution.
Step 8: Likewise, the location of a river that provides a more suitable solution substitutes that of the sea.
Step 9: Using d max as a minimal value for managing the intensification level and B U and B L as upper and lower bounds, the preceding algorithm validates the evaporation conditions (for unconstrained issues): Rain based on Equation (31) End if Regarding local challenges, the WCA employs the following code to enhance its capacity: Rain based on Equation (33) End if randn is a random integer, and δ denotes the variance term indicating the search area around the sea. Equation (32) is only used to avoid early convergence in these issues for streams with natural movements toward the sea.
Step 10: The d max is reduced as follows: Step 11: The algorithm is finished if any stopping criterion is met.
Step 12: If any stopping criteria are fulfilled, the algorithm is completed; otherwise, the procedure repeats in step 5.
The research uses ANFIS models combined with HHO, SSA, TLBO, and WCA optimization techniques to estimate ISFC. The description of the ANFIS, HHO, SSA, TLBO, and WCA is explained later, while Figure 2 portrays the procedure used in ISFC prediction.
Geotechnics 2022, 2, FOR PEER REVIEW 11 is a random integer, and δ denotes the variance term indicating the search area around the sea. Equation (32) is only used to avoid early convergence in these issues for streams with natural movements toward the sea.
Step 10: The is reduced as follows: Step 11: The algorithm is finished if any stopping criterion is met.
Step 11: If any stopping criteria are fulfilled, the algorithm is completed; otherwise, the procedure repeats in step 5.
The research uses ANFIS models combined with HHO, SSA, TLBO, and WCA optimization techniques to estimate ISFC. The description of the ANFIS, HHO, SSA, TLBO, and WCA is explained later, while Figure 2 portrays the procedure used in ISFC prediction.

Data Provision
In order to test the findings, the best-fit predictive network (to predict the ISFC), a proper dataset of driven shafts is prepared. The proposed solutions were mainly affected

Data Provision
In order to test the findings, the best-fit predictive network (to predict the ISFC), a proper dataset of driven shafts is prepared. The proposed solutions were mainly affected by four introduced conditional factors, as mentioned in Goh [90]. In this sense, there was a total of 65 samples (e.g., the dataset was collected mainly by Goh [90]), with 52 samples taken for the training (80%) and 13 for testing (20%). The ISFC values are affected by four factors mentioned in Goh [90], playing a vital role as input parameters for the ISFC prediction. Table 1 describes the numerical estimate of the dataset. From Figure 1a-d, the majority of pile length was between 3-33 m, pile diameter was between 10-40 cm, effective vertical stress is between 20-130 kPa, and undrained shear strength was between 10-62 kPa.

Results and Discussion
Optimization aims to discover all potential outcomes in a search space and determine the optimum solution based on conditions and parameters. Optimization was previously used in engineering and scientific fields, which are inherently complicated to optimize. This motivates the development of various meta-heuristic algorithms to discover the optimal solution. The significance of accurately calculating the friction capacity of driven piles embedded in cohesive soils and its complication in engineering projects is well recognized. This section discusses the HHO-ANFIS, SSA-ANFIS, TLBO-ANFIS, and WCA-ANFIS. The performance and accuracy of the estimated output for the proposed models (e.g., in predicting ISFC) are appraised by measuring MAE, RMSE, and R 2 , calculated using Equations (35)-(37), respectively.
where ISFC i is expected to indicate the expected ISFCs, ISFC i simulated stance for the simulated ones, and G presents the sum of ISFCs.

Developing Hybridized Fuzzy Tools
This section discusses model creation in the MATLAB environment; a random selection with a ratio of 80:20 was applied to the dataset first, containing 52 and 13 samples for the training and testing groups, respectively. Table 2 shows the data for the implemented HHO-ANFIS, SSA-ANFIS, TLBO-ANFIS, and WCA-ANFIS. ANFIS training is achieved through assessing the training samples by changing the parameters of membership functions, and Figure 1 presents the process of realizing an excellent hybridized ANFIS. This was done because each complexity of the assemblage accuracy differs from others and Figure 3a-d gives an account of the convergence process of HHO-ANFIS, SSA-ANFIS, TLBO-ANFIS, and WCA-ANFIS, respectively.  A sum of 1000 iterations was designated for the four assemblages. That means that every proposed hybrid model of HHO, SSA, TLBO, and WCA would update the location of their operator 1000 times. The RMSE changes were controlled by the population sizes (Npops) of 10, 25, 50, 100, 200, 300, 400, and 500. Table 3 shows RMSE of the HHO-ANFIS,  SSO-ANFIS, TLBO-ANFIS, and WCA-ANFIS for Npops of 10, 25, 50, 100, 200, 300, 400, and 500, respectively. The best Npops for the HHO-ANFIS, SSO-ANFIS, TLBO-ANFIS, and WCA-ANFIS are 300, 100, 300, and 200, respectively. These population sizes were selected because of their lowest RMSE value in the training phase, as displayed in italic in Table 3. The selected Npop for HHO-ANFIS, SSO-ANFIS, TLBO-ANFIS, and WCA-ANFIS are 300, 100, 300, and 200, respectively. The RMSE starts from 96.9887, 83.9397, 45.4107, and 58.2593, and ends with 5.3473, 6.3191, 5.1124, and 3.9843, respectively. Significant findings could be derived from the graphs in Figure 3, which is their convergence rate when the first 500 iterations of the hybrid algorithm demonstrate the majority of error minimization. The TLBO-ANFIS attains its optimum point much faster than HHO-ANFIS, SSO-ANFIS, and WCA-ANFIS.  Table 3. Additionally, according to this table, WCA-ANFIS has the lowest RMSE (RMSE = 3.984) related to the population size of 200 and the highest RMSE value is 12.483, related to SSA-ANFIS. A sum of 1000 iterations was designated for the four assemblages. That means that every proposed hybrid model of HHO, SSA, TLBO, and WCA would update the location of their operator 1000 times. The RMSE changes were controlled by the population sizes (N pops ) of 10, 25, 50, 100, 200, 300, 400, and 500. Table 3 shows RMSE of the HHO-ANFIS, SSO-ANFIS, TLBO-ANFIS, and WCA-ANFIS for N pops of 10, 25, 50, 100, 200, 300, 400, and 500, respectively. The best N pops for the HHO-ANFIS, SSO-ANFIS, TLBO-ANFIS, and WCA-ANFIS are 300, 100, 300, and 200, respectively. These population sizes were selected because of their lowest RMSE value in the training phase, as displayed in italic in Table 3. The selected N pop for HHO-ANFIS, SSO-ANFIS, TLBO-ANFIS, and WCA-ANFIS are 300, 100, 300, and 200, respectively. The RMSE starts from 96.9887, 83.9397, 45.4107, and 58.2593, and ends with 5.3473, 6.3191, 5.1124, and 3.9843, respectively. Significant findings could be derived from the graphs in Figure 3, which is their convergence rate when the first 500 iterations of the hybrid algorithm demonstrate the majority of error minimization. The TLBO-ANFIS attains its optimum point much faster than HHO-ANFIS, SSO-ANFIS, and WCA-ANFIS. The summary of the results from Figure 3 is given in Table 3. Additionally, according to this table, WCA-ANFIS has the lowest RMSE (RMSE = 3.984) related to the population size of 200 and the highest RMSE value is 12.483, related to SSA-ANFIS.

Simulation and Assessment
The parameters are grouped into training and testing to help in evaluating the model's two main objectives: i.
Pattern recognition: It provides insight into the question, "How accurately could each model understand the link between pile length, pile diameter, effective vertical stress, and undrained shear strength using the ISFC?" As stated earlier, this task is realized through adjusting the parameters of membership functions that are variables of the ANFIS's function. ii.
Pattern generalization: It responds to the question, "How well can each trained model predict the ISFC under unexpected circumstances?", the purpose for which testing data is not similar to those utilized in the training phase.
The HHO-ANFIS, SSO-ANFIS, TLBO-ANFIS, and WCA-ANFIS realized the learning RMSE of 5.3473, 6.3191, 5.1124, and 3.9843 with MAE of 4.1377, 4.5357, 3.9051, and 2.8724, respectively. Moreover, Figure 4 exhibits the correlation of the results of the training phase, its mean, and standard deviation. From the previously two indices mentioned above and taking the statistical details of Table 1 into account, all models realize an adequate level of accuracy. However, the higher training veracity of the WCA-ANFIS can be inferred. After taking a closer look at the graphs in Figure 4a-d shows that the ISFCs forecasted by all models correlate outstandingly with expected values. The R 2 values of 0.98744, 0.98241, 0.98853, and 0.99304 correlate well with the training results. These proposed models were subsequently integrated into the testing inputs, resulting in an assessment of the ISFC for pile-related scenarios that were previously unknown.          [101,102].
Due to the outstanding results given by the TLBO-ANFIS and WCA-ANFIS is deducing and replicating the ISFC pattern. Therefore, these developed methods can be alternatively utilized for practical applications and other related problems. The error of the WCA-ANFIS is lower than that of the other three assemblages, as shown in the accuracy indices obtained in the training phase. Moreover, there are around 0.56, 1.06, and 0.45% differences in R 2 ; therefore, it could be postulated that WCA can train ANFIS more aptly in comparison with HHO, SSA, and TLBO. It is worth noting that WCA-ANFIS featured as an accurate model, followed by the TLBO-ANFIS, SSA-ANFIS, and HHO-ANFIS in the testing phase. Figure 8a-b compares the target ISFC with the model output for training and testing, depicting an almost perfect correlation.  Geotechnics 2022, 2, FOR PEER REVIEW 20 utilization of experimental and traditional simulation approaches is costly and time-consuming in predicting the behavior of driven piles [101,102]. Due to the outstanding results given by the TLBO-ANFIS and WCA-ANFIS is deducing and replicating the ISFC pattern. Therefore, these developed methods can be alternatively utilized for practical applications and other related problems. The error of the WCA-ANFIS is lower than that of the other three assemblages, as shown in the accuracy indices obtained in the training phase. Moreover, there are around 0.56, 1.06, and 0.45% differences in R 2 ; therefore, it could be postulated that WCA can train ANFIS more aptly in comparison with HHO, SSA, and TLBO. It is worth noting that WCA-ANFIS featured as an accurate model, followed by the TLBO-ANFIS, SSA-ANFIS, and HHO-ANFIS in the testing phase. Figure 8a-b compares the target ISFC with the model output for training and testing, depicting an almost perfect correlation.
.   Due to the outstanding results given by the TLBO-ANFIS and WCA-ANFIS is deducing and replicating the ISFC pattern. Therefore, these developed methods can be alternatively utilized for practical applications and other related problems. The error of the WCA-ANFIS is lower than that of the other three assemblages, as shown in the accuracy indices obtained in the training phase. Moreover, there are around 0.56, 1.06, and 0.45% differences in R 2 ; therefore, it could be postulated that WCA can train ANFIS more aptly in comparison with HHO, SSA, and TLBO. It is worth noting that WCA-ANFIS featured as an accurate model, followed by the TLBO-ANFIS, SSA-ANFIS, and HHO-ANFIS in the testing phase. Figure 8a-b compares the target ISFC with the model output for training and testing, depicting an almost perfect correlation. .

Conclusions
The current research aims to evaluate how efficiently several artificial intelligence approaches could measure the friction capacity of driven piles in a cohesive soil environment. Furthermore, as input data, four practical elements that impact the friction capacity of driven piles were studied: length, diameter, effective stress, and undrained cohesion strength of the adjacent soils. The feasibility of the ANFIS system hybridized with HHO, SSA, TLBO, and WCA is used to estimate the ISFC of piles. Driven piles friction capacity prediction cannot be overemphasized in engineering projects. In the same vein, an ANFIS was hybridized with HHO, SSA, TLBO, and WCA to predict the ISFC of driven piles. The lowest value of RMSE and the higher value of R 2 shows the best predictive model. The optimized composition of the four models demonstrated high-quality performances of approximately 97.50, 96.51, 97.72, and 98.61% correlation of training data for the R 2 value. In the case of RMSE, the values of 8.2844, 7.4746, 6.6572, and 6.8528 are obtained for the HHO-ANFIS, SSA-ANFIS, TLBO-ANFIS, and WCA-ANFIS, respectively. This implies proficient dependability for utilizing metaheuristic algorithms in adjusting ANFIS internal parameters. Compared with other models, the solution acquired demonstrated that the WCA-ANFIS is more encouraging in the training and testing phases. The calculated RMSE values for the TLBO-ANFIS-based models were lower than HHO-ANFIS, SSA-ANFIS, and WCA-ANFIS. Irrespective of their level of accuracy, employing the tested models in practical problems is recommended.