Next Article in Journal
Spatiotemporal Variation in Driving Factors of Vegetation Dynamics in the Yellow River Delta Estuarine Wetlands from 2000 to 2020
Previous Article in Journal
Soil Moisture Monitoring at Kilometer Scale: Assimilation of Sentinel-1 Products in ISBA
Previous Article in Special Issue
Deep Learning for Earthquake Disaster Assessment: Objects, Data, Models, Stages, Challenges, and Opportunities
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Landslide Susceptibility Mapping: Analysis of Different Feature Selection Techniques with Artificial Neural Network Tuned by Bayesian and Metaheuristic Algorithms

1
School of Computer Science, China University of Geosciences, Wuhan 430074, China
2
Institute of Soil and Environmental Sciences, University of Agriculture Faisalabad, Faisalabad 38000, Pakistan
3
Department of Computer Science, Karakoram International University, Gilgit 15100, Pakistan
4
School of Environmental Studies, China University of Geosciences, Wuhan 430074, China
5
School of Geography, Karakoram International University, Gilgit 15100, Pakistan
6
Department of Zoology, College of Science, King Saud University, P.O. Box 2455, Riyadh 11451, Saudi Arabia
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(17), 4330; https://doi.org/10.3390/rs15174330
Submission received: 16 July 2023 / Revised: 30 August 2023 / Accepted: 30 August 2023 / Published: 2 September 2023

Abstract

:
The most frequent and noticeable natural calamity in the Karakoram region is landslides. Extreme landslides have occurred frequently along Karakoram Highway, particularly during monsoons, causing a major loss of life and property. Therefore, it is necessary to look for a solution to increase growth and vigilance in order to lessen losses related to landslides caused by natural disasters. By utilizing contemporary technologies, an early warning system might be developed. Artificial neural networks (ANNs) are widely used nowadays across many industries. This paper’s major goal is to provide new integrative models for assessing landslide susceptibility in a prone area in the north of Pakistan. To achieve this, the training of an artificial neural network (ANN) was supervised using metaheuristic and Bayesian techniques: Particle Swarm Optimization (PSO) algorithm, Genetic algorithm (GA), Bayesian Optimization Gaussian Process (BO_GP), and Bayesian Optimization Tree-structured Parzen Estimator (BO_TPE). In total, 304 previous landslides and the eight most prevalent conditioning elements were combined to form a geospatial database. The models were hyperparameter optimized, and the best ones were employed to generate susceptibility maps. The obtained area under the curve (AUC) accuracy index demonstrated that the maps produced by both Bayesian and metaheuristic algorithms are highly accurate. The effectiveness and efficiency of applying ANNs for landslide mapping, susceptibility analysis, and forecasting were studied in this research, and it was observed from experimentation that the performance differences for GA, BO_GP, and PSO compared to BO_TPE were relatively small, ranging from 0.32% to 1.84%. This suggests that these techniques achieved comparable performance to BO_TPE in terms of AUC. However, it is important to note that the significance of these differences can vary depending on the specific context and requirements of the ML task. Additionally, in this study, we explore eight feature selection algorithms to determine the geospatial variable importance for landslide susceptibility mapping along the Karakoram Highway (KKH). The algorithms considered include Information Gain, Variance Inflation Factor, OneR Classifier, Subset Evaluators, principal components, Relief Attribute Evaluator, correlation, and Symmetrical Uncertainty. These algorithms enable us to evaluate the relevance and significance of different geospatial variables in predicting landslide susceptibility. By applying these feature selection algorithms, we aim to identify the most influential geospatial variables that contribute to landslide occurrences along the KKH. The algorithms encompass a diverse range of techniques, such as measuring entropy reduction, accounting for attribute bias, generating single rules, evaluating feature subsets, reducing dimensionality, and assessing correlation and information sharing. The findings of this study will provide valuable insights into the critical geospatial variables associated with landslide susceptibility along the KKH. These insights can aid in the development of effective landslide mitigation strategies, infrastructure planning, and targeted hazard management efforts. Additionally, the study contributes to the field of geospatial analysis by showcasing the applicability and effectiveness of various feature selection algorithms in the context of landslide susceptibility mapping.

Graphical Abstract

1. Introduction

Artificial neural networks are utilized to solve a range of issues in industries, like banking, manufacturing, electronics, and medicine, among others [1]. The backpropagation approach is typically used to train neural networks [2,3]. The backpropagation method can be trained using a variety of algorithms, including gradient descent, conjugate gradient descent, robust, BFGS quasi-Newton, one-step secant, Levenberg–Marquardt, and Bayesian regularization [4]. Some characteristics, such as the number of hidden layers and the number of neurons in each layer, must be defined during the design and training of an ANN; these aspects vary depending on the particular application. There is not a standard, clear approach for selecting these settings. It is utilized in a way that is almost like trial and error, but it requires more computing time and is not very accurate. Therefore, it is required to propose a method to choose the ideal set of parameters that will have the greatest impact on ANN performance.
From the literature, it can be inferred that metaheuristic optimization algorithms have effectively been employed to overcome the computing limitations of traditional landslide predictive models, hence enhancing their performance. The local minimum trap could become a barrier to precise estimations when it comes to ANNs. But recent research has demonstrated that by using metaheuristic techniques, such computational problems can be resolved [5]. Testing novel metaheuristic methods is an important step toward discovering more efficient models, but, as well-known optimizers (such as the PSO and GA) have been sufficiently evaluated. This study’s main objective is to assess the effectiveness of four optimization strategies—Particle Swarm Optimization (PSO), Genetic algorithm (GA), Bayesian Optimization Gaussian Process (BO-GP), Bayesian Optimization Tree-structured Parzen Estimator (BO-TPE) —in combination with artificial neural networks (ANN) to produce a map of the KKH region’s susceptibility to landslides in northern areas of Pakistan [6].
In the context of artificial neural networks (ANNs), optimization algorithms, such as PSO, GA, and Bayesian Optimization, have been widely used to improve the performance of ANNs in various classification tasks [7]. The following are some specific examples of their applications in the context of ANNs for data classification:
Particle Swarm Optimization (PSO) for ANN
  • PSO has been used to optimize the weights and biases of the ANN to improve its classification accuracy. By searching the weight space, PSO helps to find the optimal set of weights that minimize the prediction error and maximize the classification performance.
  • PSO has been applied to optimize the architecture of the ANN, including the number of hidden layers and the number of neurons in each layer. It helps to determine the optimal network structure that suits the complexity of the classification problem.
Genetic Algorithm (GA) for ANN
  • GA has been used to train ANNs by adjusting the weights and biases of the network. By applying genetic operators, such as crossover and mutation, GA explores the search space of weight configurations to find the best set of weights that lead to improved classification accuracy.
  • GA has been employed to optimize the architecture of the ANN, including the number of hidden layers and neurons. By evolving populations of networks with different architectures, GA helps to identify the optimal network structure that achieves better classification performance.
Bayesian Optimization for ANN
  • Bayesian optimization has been used to optimize the hyperparameters of ANNs, such as learning rate, regularization strength, and network architecture. By leveraging Bayesian inference, it explores the hyperparameter space to find the optimal configuration that maximizes the classification accuracy.
  • Bayesian optimization has been applied to tune the activation functions, dropout rates, and other architectural choices of ANNs. It helps to identify the best combinations of these hyperparameters that result in improved classification performance.
In all these cases, the optimization algorithms (PSO, GA, and Bayesian optimization) play a crucial role in training and fine-tuning ANNs for data classification. They help to search for optimal weights, biases, and hyperparameter configurations, allowing ANNs to achieve higher accuracy and better generalization in classification tasks [8]. By tuning the weights and biases, these algorithms supervise the performance of the ANN in predicting the landslide susceptibility index (LSI). In the context of predicting landslide susceptibility using artificial neural networks (ANNs), the weights and biases of the network are adjusted or tuned by the optimization algorithms (such as Genetic Algorithms, Particle Swarm Optimization, or Bayesian Optimization) to improve the performance of the ANN in predicting the landslide susceptibility index (LSI). The weights and biases in an ANN determine the strength and influence of the connections between neurons in different layers. By adjusting these parameters, the ANN can learn and adapt to the patterns and relationships present in the input data, which in this case are the conditioning elements used to assess landslide susceptibility. During the training phase, the optimization algorithms iterate through different combinations of weights and biases, evaluating the performance of the ANN based on a defined metric (e.g., accuracy and area under the curve). The algorithms aim to find the set of weights and biases that minimize the prediction error or maximize the performance metric, ultimately leading to a more accurate prediction of the LSI. The optimization algorithms use various strategies, such as gradient descent, genetic operators (crossover and mutation), or probabilistic modeling, to explore the search space of weight and bias combinations [9,10,11,12,13]. Through iterations and feedback from evaluating the ANN’s predictions, the algorithms progressively adjust the weights and biases to converge towards an optimal solution that yields the best performance in predicting the LSI. By continuously fine-tuning the weights and biases, these optimization algorithms guide the ANN to learn and generalize patterns from the training data, enabling the model to make more accurate predictions of landslide susceptibility based on the conditioning elements. This iterative optimization process helps to improve the effectiveness of the ANN in capturing the complex relationships between the input features and the LSI, leading to more reliable and accurate predictions of landslide susceptibility.
In addition to discussing the fine-tuning of ANNs using metaheuristic and Bayesian algorithms, this study also addresses the challenge of identifying the geospatial variable importance for landslide susceptibility mapping. To achieve this, we employ eight different feature selection algorithms: Information Gain, Variance Inflation Factor (VIF), OneR Classifier, Subset Evaluators, principal components, Relief Attribute Evaluator, correlation, and Symmetrical Uncertainty. Each algorithm brings a unique approach to assessing the relevance and significance of geospatial variables in predicting landslide occurrences. The selection of these feature selection algorithms offers a comprehensive evaluation of different approaches to geospatial variable importance assessment. By leveraging these algorithms, we aim to uncover the most influential variables that contribute to landslide susceptibility mapping in our case study. The outcomes of this research will provide valuable insights for hazard management, infrastructure planning, and proactive decision making to mitigate the risks associated with landslides along this critical transportation corridor. This study not only contributes to the specific context of landslide susceptibility mapping along the KKH, but also has broader implications for geospatial analysis and hazard management. The findings will enhance our understanding of the geospatial variables influencing landslide occurrences and serve as a foundation for future research and the development of effective strategies to reduce landslide risks in mountainous regions. Analyzing feature selection algorithms helps in identifying the most important geospatial variables for landslide susceptibility mapping. This deepens our understanding of the specific variables that contribute to landslide occurrences along the KKH in Pakistan. By uncovering the key variables, we can gain valuable insights into the underlying mechanisms and processes that influence landslide susceptibility. By determining the geospatial variable importance, we can refine landslide prediction models and improve the accuracy of landslide susceptibility maps. Focusing on the most relevant variables allows us to develop more precise and reliable mapping methodologies. This, in turn, enables better hazard assessment, infrastructure planning, and land management decisions along the KKH.
Analyzing geospatial variable importance helps in proactive hazard management along the KKH. With a better understanding of the critical variables, authorities and stakeholders can implement targeted measures to minimize the risks associated with landslides. This includes implementing early warning systems, designing effective slope stabilization measures, and adopting appropriate land use planning strategies. By identifying the most influential geospatial variables, analyzing feature selection algorithms enables optimal resource allocation for landslide mitigation efforts. Instead of allocating resources uniformly across all variables, decision makers can prioritize and allocate resources to monitor, manage, and mitigate the variables that have the greatest impact on landslide susceptibility. This ensures the efficient utilization of limited resources for effective hazard management. Geospatial variable importance analysis contributes to informed infrastructure planning and design along the KKH. By understanding the key variables that influence landslide susceptibility, engineers and planners can incorporate appropriate measures to reduce the vulnerability of infrastructure to landslides. This includes considering geotechnical investigations, slope stabilization techniques, and route optimization to minimize exposure to high-risk areas. The findings from analyzing feature selection algorithms for geospatial variable importance can be applied beyond landslide susceptibility mapping. The knowledge gained can inform similar geospatial analysis tasks in other regions facing similar challenges. The insights obtained can guide researchers and practitioners in assessing the importance of variables for various geospatial phenomena and developing targeted mitigation strategies.
In summary, analyzing feature selection algorithms for geospatial variable importance in the context of landslide susceptibility mapping along the KKH offers several benefits. It enhances our understanding of landslide mechanisms, improves prediction and mapping accuracy, facilitates proactive hazard management, optimizes resource allocation, informs infrastructure planning, and provides valuable insights for broader geospatial analysis tasks. These benefits contribute to safer infrastructure, reduced risks, and more effective decision making for landslide-prone areas. Overall, by combining advanced feature selection algorithms with geospatial analysis, this study aims to provide a comprehensive understanding of the geospatial variable importance for landslide susceptibility mapping along the KKH, ultimately contributing to safer infrastructure and improved hazard management in the region. We also employ advance hyper optimization techniques to optimize the ANN model with the aim of maximizing the predictive capabilities of machine learning algorithms, using both hyperparameter and feature optimization techniques to give insight into complex phenomena, like landslides, and to optimize complex modeling techniques. It is essential to identify advance modeling techniques, such ensemble methods like Random Forests and Gradient Boosting, that are effective in handling the intricate interactions between different variables. They can capture nonlinear relationships, account for missing data, and highlight the significance of particular features. Geospatial data analysis offers a valuable perspective. By integrating Geographic Information System (GIS) techniques, spatial relationships and topographical features can be incorporated into the analysis. Techniques like spatial statistics, including variogram analysis, can identify spatial autocorrelation, providing insight into patterns and correlations [14,15,16]. Advanced models can also be combined using ensemble methods for spatial analysis. Stacking or bagging, for example, allow for the amalgamation of distinct models, thereby leveraging the strengths of individual models to improve predictive accuracy. Deep Learning, specifically Convolutional Neural Networks (CNNs), comes into play when satellite or aerial imagery data are available. CNNs can extract pertinent features from images, capturing spatial patterns and texture information that are indicative of landslide susceptibility. Hybrid approaches that combine diverse data sources—such as remote sensing data, topographical data, and historical landslide records—can yield a comprehensive set of input features for the models being used in this study, enriching the understanding of susceptibility variables [17,18]. For situations where temporal dynamics are influential, models that incorporate both spatial and temporal features can provide a more accurate representation of landslide susceptibility changes over time. Moreover, Bayesian Network Models offer a probabilistic perspective, enabling the capture of intricate relationships among variables and providing a visualization of causal connections that contribute to landslide susceptibility. To strike a balance between predictive accuracy and interpretability, hybrid geostatistical models can merge geostatistical methods (e.g., kriging) with machine learning models, thereby considering spatial autocorrelation in predictions.
The main objective of this paper is to comprehensively investigate and improve the accuracy of landslide susceptibility mapping through the integration of various methodologies. This includes delineating the study area, identifying relevant landslide conditioning variables, generating susceptibility maps using the specified methodology, exploring different feature selection techniques to assess geospatial variables, and employing advanced hyperparameter techniques like Bayesian and Metaheuristic methods. The primary aim is to evaluate the effectiveness of these approaches and techniques in enhancing the accuracy of landslide susceptibility predictions, thereby contributing to a deeper understanding of landslide dynamics and potential mitigation strategies. This paper is structured into various sections. Section 2 delves into elucidating the study area and the variables contributing to landslides. In Section 3, the methodology employed for generating susceptibility maps is expounded upon. Section 4 elucidates the utilization of eight distinct feature selection techniques, each assessing geospatial variables based on criteria elucidated in detail within that section. Section 5 provides insight into the application of Bayesian and Metaheuristic hyperparameter techniques. The evaluation of the results obtained from our experiment is covered in Section 6.

1.1. Architecture of Artificial Neural Network

The architecture of the Artificial Neural Network (ANN) for landslide susceptibility mapping can be described as the input layer of the ANN, which receives the relevant data or features related to landslide susceptibility in the KKH region. These features may include variables such as geological characteristics, slope angles, aspect, geological properties, and land cover. The number of neurons specific to each setting, such as activation function, batch size, and epochs for each technique determined during the design and training of the ANN for BO_GP, BO_TPE, PSO, and GA, are mentioned in Table 1 below.
Table 1 represents the experimental results of different hyperparameter optimization techniques applied to ANN. Each row in the table corresponds to a different optimization technique, and the columns represent the evaluated performance metrics or hyperparameters.
The activation function is a crucial component of ANNs, as it introduces nonlinearity and affects the network’s ability to model complex relationships. The table provides information about the activation functions used for each technique, such as “sgd”, “tanh”, and the value “1.742” for PSO.
Batch Size and Epochs: Batch size refers to the number of training samples used in each iteration of the optimization algorithm, while epochs represent the number of times the entire training dataset is passed through the network.
Neurons: Neurons refer to the number of units or nodes in each layer of the ANN. Table 1 shows the number of neurons used for each technique, such as 64 for GA.
The output layer of the ANN represents the predicted susceptibility to landslides in the KKH region. It generates a susceptibility map indicating the likelihood of landslides occurring at different locations within the study area. The backpropagation approach is commonly used to train ANNs. There are various training algorithms that can be applied, including gradient descent, conjugate gradient descent, robust, BFGS quasi-Newton, one-step secant, and Levenberg–Marquardt, and Bayesian regularization [19,20,21]. The choice of training algorithm may depend on the specific requirements and characteristics of the landslide susceptibility problem in the KKH region. The focus of this study is to evaluate the effectiveness of four optimization strategies in combination with ANNs. These strategies are PSO, GA, BO-GP, and BO-TPE. These optimization techniques are employed to fine-tune the hyperparameters of the ANN and enhance its performance in generating accurate landslide susceptibility maps. Overall, the architecture of the ANN involves an input layer that receives relevant data, hidden layers for processing and feature extraction, an output layer that produces the susceptibility map, training algorithms for optimizing the network’s parameters, and optimization techniques to improve the performance of the ANN in generating accurate landslide susceptibility maps for the KKH region [22].

1.2. Benefits of Artificial Neural Networks

Artificial neural networks (ANNs) offer several benefits in various fields and applications. Here are some of the key benefits of using ANNs. Nonlinear mapping: ANNs excel at capturing and modeling complex, nonlinear relationships between input and output variables. They can learn and represent highly intricate patterns, making them effective in handling complex tasks that may not have clear linear relationships. Adaptive learning: ANNs have the ability to adapt and learn from examples or data. Through a process called training, ANNs can adjust their internal parameters to improve their performance on specific tasks. This adaptability allows ANNs to continuously learn and improve their predictions or classifications as new data becomes available. Parallel processing: ANNs can perform computations in parallel, making them suitable for handling large-scale and computationally intensive tasks [8,23,24]. This parallel processing capability enables ANNs to process multiple inputs simultaneously, leading to faster and more efficient computations. Fault tolerance and redundancy: ANNs have the ability to tolerate faults or errors in the system. Due to their distributed nature and interconnected structure, ANNs can still provide useful outputs even when some of the nodes or connections in the network are damaged or missing. This fault tolerance and redundancy makes ANNs robust in real-world scenarios where there may be noise or incomplete data. Pattern recognition and classification: ANNs are highly effective in pattern recognition and classification tasks [25,26,27,28]. They can learn and recognize complex patterns, allowing them to classify and categorize data into different classes or groups. This capability has numerous applications in fields such as image and speech recognition, natural language processing, and data mining. Generalization: ANNs can generalize learned patterns to new, unseen data. Once trained on a representative dataset, ANNs can make accurate predictions or classifications on similar but previously unseen inputs. This generalization ability makes ANNs valuable in scenarios where new data needs to be processed or classified in real-time. Feature extraction: ANNs can automatically extract relevant features or representations from raw data, reducing the need for manual feature engineering. This capability is particularly useful when dealing with high-dimensional data, as ANNs can automatically learn and extract the most informative features to improve performance. Real-time processing: ANNs can be implemented in real-time systems, allowing for the quick and efficient processing of data. This capability is crucial in applications, such as real-time monitoring, control systems, and decision making, where immediate responses are required. These benefits demonstrate the versatility and power of artificial neural networks in solving complex problems like landslide susceptibility mapping and making accurate predictions or classifications across various domains.

2. Study Area

We conducted a study in the northern part of Pakistan, focusing on an approximately 332 km stretch of the KKH highway. The KKH highway is a major road that spans 1300 km, connecting various provinces of Pakistan, namely Punjab, Khyber Pakhtunkhwa, and Gilgit Baltistan, with Xinjiang, an autonomous region of China. Our study specifically covered the Gilgit, Hunza, and Nagar districts. Along the KKH, there are several villages including Juglot (located between 36°12′147″N latitude and 74°18′772″E longitude), Jutal, Rahimbad, Aliabad, and others, ultimately reaching Khunjarab Top, which serves as the border crossing between China and Pakistan. This region is situated alongside the Indus River, Hunza River, and Gilgit River.
The study area encompassed a length of 332 km and had a radius of 10 km, covering an area of 3320 km2 along the KKH. The majority of the region consists of mountainous terrain, with its highest point reaching an elevation of 5370 m and the lowest point at 1210 m. Common natural hazards in this area include snow avalanches, landslides, and earthquakes Figure 1.
Technological advancements have propelled improvements in landslide modeling methods, encompassing the integration of remote sensing and GIS for comprehensive data analysis. Machine learning and AI algorithms aid in pattern recognition and predictive precision, while diverse datasets are fused to offer a holistic understanding of triggers and behavior [8,29]. Numerical simulation models enhance mechanics comprehension, and probabilistic approaches address uncertainties. Real-time analysis is enabled through big data and cloud computing, complemented by sensor networks for continuous monitoring. Crowdsourced data and early warning systems facilitate real-time data collection and alerts. The integration of climate change, decision support tools, and collaborative efforts elevate accuracy and preparedness, culminating in strides to mitigate landslide risks and impacts. A case study can showcase the application of advanced machine learning algorithms, including ANN, along with hyper optimization techniques, like Bayesian and metaheuristic methods, incorporating sophisticated feature selection methodologies for optimizing the modeling of complex phenomena, such as landslides [6].
The most common types of landslide in our study are debris flow, rockfalls and floods. Notable instances of devastating debris flows are the Attabad and Hunza landslides. Various rockfall events occurred on the Karakoram Highway, such as the rockfall incident that took place in Kohistan’s Barseen area and Kohistan [30,31]. Streams can negatively impact the stability of slopes through two main mechanisms: undercutting due to toe erosion and saturation of the slide toe caused by increased water penetration. In the studied region, both the road and stream networks wield significant influence over the occurrence of landslides. Particularly, the area within a 100 m buffer zone between roads and watercourses demonstrates one of the highest Information Gain ratios, rivaled only by the slope (for further details, see Section 4). This trend is clarified by the fact that uncontrolled blasting and excavation during road construction, as well as the establishment of stream channels for irrigation on these vulnerable slopes, frequently trigger land movements.
As expected, areas covered by woodland, shrubland, and alpine pasture exhibit the lowest tendency of experiencing landslides. In a comparison between the landslide inventory and variables contributing to landslides, it becomes evident that roads, streams, and slope gradients hold the utmost importance as variables dictating the spatial distribution of landslides within the area. Gilgit receives an annual rainfall of approximately 154 mm. The region’s irrigation depends on the flow of streams and rivers fed by the melting snow and glaciers from lofty mountains. The summer season is prolonged and brings higher temperatures. While the average wintertime temperature remains below 10 °C, occurrences of intense sunshine are infrequent, and temperatures rarely reach 40 °C (104 °F). The region’s harsh weather conditions contribute to frequent instances of landslides and avalanches. These events are a result of the area’s geologically unstable terrain and soils, which play a pivotal role in slope instability. Additionally, the steep mountain slopes make them susceptible to landslides as well.

Geospatial Variables

The geospatial variables used in our case study are as follows: slope, aspect, land cover, geology, precipitation, distance to faults, distance to streams, and distance to roads. The sources of the dataset are shown in Table 2 and Table 3.
Selecting and identifying the most influential geospatial variables is a challenging task, requiring significant ground reality understanding as well as understanding of the interdependence between the variables and variance and uncertainty that they hold therefor it’s essential to carefully select the important variables and understand their importance for predicting accurate landslide vulnerabilities in future [16,32,33]. Furthermore Table 3 and Figure 2 represent the data source of geospatial variable used in our case study.
The categorization of influencing variables for landslide occurrences in our research area (Figure 2 and Table 3) involves four main categories: topological, hydrological, geological, and anthropological. Among these, slope and aspect are topological variables, with slope angle being the primary contributor to slope stability, while aspect-related variables, like sunlight exposure, winds, rainfall, soil moisture, and cracks, influence landslide frequency. Aspect and slope are determined using the SRTM DEM with a 30 m resolution and categorized into classes for analysis. Geological variables consider the geological map of Pakistan, including fault lines and various formations, as different geological units impact susceptibility to geomorphological processes. Hydrological variables encompass precipitation and proximity to waterways; this is significant due to rainfall-induced landslides in the study location. Anthropological elements include land usage and distance to highways, which are evaluated using Sentinel 2 images and supervised classification to analyze their effects on landslide movement. The map’s accuracy was validated at 87%. Land use, road development, and construction activity are also considered as influencing variables on landslide stability. In total, eight variables—slope, aspect, land cover, geology, precipitation, distance to faults, distance to streams, and distance to roads—are examined in this case study (Figure 2).

3. Methodology

This study employed HOP techniques to search for the optimal hyperparameter configurations for the ANN models. This is important because selecting appropriate hyperparameters, such as activation function, batch size, epochs, and neuron configurations, can significantly impact the performance and generalization capabilities of ANNs. The four optimization techniques used in the study, namely PSO, GA, BO_GP, and BO_TPE, were applied to fine-tune the hyperparameters of the ANN models. Each technique employs a different approach to explore the hyperparameter space and find the best set of values. The performance of the ANN models was evaluated using AUC. Overall, the methodology involved applying different optimization techniques to search for the optimal hyperparameter configurations for the ANN models. The methodology aims to identify the most effective optimization approach and hyperparameter settings for accurately predicting the landslide susceptibility index. Additionally, our case study incorporated eight different techniques while considering various criteria, like multicollinearity, correlations, information sharing, variance, and uncertainty, to select landslide conditioning variables, as depicted in Figure 3.
In complex tasks like predicting landslide susceptibility, where the relationships between variables can be intricate and data might be limited, it is essential to leverage all available tools to ensure model accuracy and reliability. Integrating feature selection and hyperparameter optimization allows us to build a more robust and effective model that can capture the complexities of the underlying processes. Combining feature selection and hyperparameter optimization in a methodology can provide several benefits for complex modeling tasks like predicting landslide susceptibility. For example, feature selection helps in identifying the most relevant features, which can prevent overfitting by reducing noise and irrelevant information. Hyperparameter optimization ensures that the model’s complexity is controlled, further reducing the risk of overfitting. By selecting important features and optimizing hyperparameters, the model is more likely to generalize well to new, unseen data. This is crucial in predicting landslides, as the model needs to perform well on different terrains and conditions. Optimal hyperparameters and relevant features contribute to the model’s performance. A well-tuned model with the right features can provide accurate predictions, which is vital for landslide susceptibility assessments. Combining these aspects can lead to better insights into the domain itself. Understanding which features are crucial and how hyperparameters affect model behavior can provide valuable information about the processes leading to landslides. Instead of exploring the entire hyperparameter space exhaustively, which can be very time-consuming, focusing on a reduced feature space due to feature selection can make the optimization process more efficient. Landslide susceptibility can vary across different regions and conditions. By optimizing both hyperparameters and feature selection, the model can be adapted and fine-tuned to different contexts. Including both feature selection and hyperparameter optimization can lead to more robust models. If certain features become less relevant due to changes in data distribution, the model’s performance can still be maintained through proper hyperparameter tuning. Thus, the suggested methodology of combining feature selection and hyperparameter optimization in the context of predicting landslide susceptibility offers numerous benefits, particularly when dealing with complex modeling tasks, such as landslides. Overall, the combined approach brings synergy to the optimization process, leading to a more accurate, reliable, and efficient model for predicting landslide susceptibility.

4. Feature Selection Techniques

Feature selection is a vital process in machine learning and data analysis that involves choosing the most relevant features from a dataset while discarding irrelevant or redundant ones. It offers numerous benefits, including improved model performance by reducing overfitting and enhancing generalization; faster computation due to decreased data dimensionality; and increased interpretability, making models more understandable and usable. Additionally, it mitigates the challenges of high-dimensional data, simplifies model maintenance and updates, fosters better feature engineering practices, and supports efficient visualization. Feature selection also contributes to model robustness, cost savings by prioritizing valuable features, and alignment with algorithm assumptions, ultimately leading to more effective and practical applications of machine learning in various domains. Some of the most common and popular variable selection techniques are mentioned in our paper and are described in the subsequent subsections.

4.1. Information Gain

Information Gain (IG) is a measure used in decision tree algorithms and feature selection to assess the importance of a feature in a classification task. It quantifies the amount of information gained about the target variable by including a particular feature in the decision-making process.
Mathematically, Information Gain is calculated using the concept of entropy. Entropy measures the impurity or disorder of a set of instances in the dataset. The entropy E(C) and Information gain (IG) can be computed in relation to the conditioning variables of landslides and is defined as:
E C = i = 1 j P i log 2 P i
In this context, P i denotes the ratio of the ith class within the complete dataset, and “m factor classes” encompass a set of values, i.e., [ v 1 , v 2 , , v m ].
I G Y , X = E Y v     {   F   |   1 , , m } Y v Y E Y v
I G Y , X = E Y E ( Y   |   X )
The IG resulting from conditioning variables is represented as IG(Y,X), while the anticipated weighted entropy is denoted as E(Y, X). Here, E(Y) represents the entropy of a landslide inventory Y containing j classes, while X, also with n classes, represents the conditioning variables of landslides. The value of E(Y) serves as an indicator of the uniformity within landslides in Y, and it is effectively utilized for selecting the most suitable conditioning factor.
The Information Gain value ranges from 0 to 1, with a higher value indicating that the feature provides more valuable information for the classification task. A higher Information Gain suggests that the feature helps in reducing the uncertainty or disorder in the dataset and contributes to better classification.
From Table 4 and Figure 4, “Roads” stand out as the most significant feature, with an Information Gain of 0.352. This indicates that it strongly aids in making accurate predictions. The “Faults” feature follows with an Information Gain of 0.195, making it a moderately influential predictor. “Streams” provide a reasonable level of information, contributing moderately with an Information Gain of 0.091. “Geology” and “Slope” possess modest Information Gain values of 0.039, implying that they offer some useful insights, though not as prominently as the top features. “Precipitation” has a lower Information Gain of 0.012, suggesting it contributes a limited amount of relevant information. “Aspect” and “Land Cover” exhibit no meaningful Information Gain (0.000), implying they have little to no impact on accurate predictions. In essence, the Information Gain values help prioritize the importance of each feature in facilitating accurate predictions or classifications, with “Roads” and “Faults” leading the way.

4.2. Variable Inflation Factor

The Variance Inflation Factor (VIF) is a measure used to assess the severity of multicollinearity among predictor variables in a regression model. It quantifies the extent to which the variance of the estimated regression coefficient for a specific predictor variable is inflated due to its correlation with other predictor variables.
Mathematically, the VIF for a predictor variable i is calculated as follows:
V I F i = 1 1 R i 2
where:
V I F i is the Variance Inflation Factor for predictor variable i;
R i 2 is the coefficient of determination (R-squared) obtained by regressing variable i against all other predictor variables.
The VIF value ranges from 1 upwards, with values greater than 1 indicating the presence of multicollinearity. A VIF value of 1 indicates no multicollinearity, while values greater than 1 suggest increasing levels of multicollinearity. Generally, a VIF value of 5 or higher is often used as a threshold to identify problematic levels of multicollinearity, although the specific threshold may vary depending on the context and the field of study.
In practice, high VIF values indicate that the variance of the regression coefficient estimates for the predictor variable is inflated, which can lead to unstable and unreliable results. Multicollinearity can make it challenging to interpret the individual contributions and effects of predictor variables in the model. Remedial actions, such as excluding highly correlated variables or performing dimensionality reduction techniques, may be necessary to mitigate the issues caused by multicollinearity.
From Table 5 and Figure 5 “Land Cover” stands out with the highest VIF, indicating strong multicollinearity. Other features like “Precipitation”, “Slope”, “Geology”, and “Aspect” also display varying degrees of multicollinearity. “Streams”, “Faults”, and “Roads” show relatively lower multicollinearity, making them potentially more independent in the analysis.

4.3. OneR Classifier

The OneR algorithm is a simple and interpretable classification algorithm. It generates a single rule for each predictor (attribute) in the data and selects the rule with the smallest total error. Each rule is based on the value of a single attribute and predicts the most frequent class for that attribute value. The OneR Classifier provides a straightforward way to assess the predictive power of individual attributes. The total error is calculated by summing the errors made by the rule for each attribute value. The formula for calculating the total error in OneR Classifier is dependent on the evaluation metric chosen, such as misclassification error rate, accuracy, or other relevant metrics.
OneR Algorithm
 1. for each predictor,
 2.  for each value of that predictor, make a rule as follows;
 3.   Count how often each value of target (class) appears
 4.   Find the most frequent class
 5.   Make the rule assign that class to this value of the predictor
 6.  Calculate the total error of the rules of each predictor
 7. Choose the predictor with the smallest total error.
Assume that we have a dataset consisting of “n” instances, each represented by a feature “X” and a corresponding class label “Y”. The One-R Classifier aims to find the best rule by selecting a single feature that minimizes the classification error. The rule is based on the mode (most frequent) class value for each unique value of the selected feature.
The following are the variables:
X i is the ith feature value;
Y i is the corresponding class label;
f is the selected feature (attribute) for the rule;
C is the set of all possible class labels.
R u l e :   Y = a r g m a X P ( Y = c | X = f X i )
For each unique value of the selected feature f(Xi), the classifier calculates the conditional probability of each class label C given that the feature has the value f(Xi). It selects the class label with the highest conditional probability as the prediction for instances with that particular feature value.
Table 6 and Figure 6 summarize the accuracy scores for various features such as Roads, which is the most significant feature, achieving an accuracy score of 0.902 and therefore making it a strong predictor. “Faults” also stands out with an accuracy score of 0.707, indicating a notable predictive capability. “Streams” exhibit a decent accuracy score of 0.659, suggesting a meaningful predictive influence. “Land Cover” contributes with an accuracy score of 0.622, showcasing a moderate level of predictive power. These values emphasize the predictive strength of these four features, namely “Roads”, “Faults”, “Streams”, and “Land Cover”, in the context of the OneR Classifier.

4.4. Subset Evaluators

The principle behind the Correlation Feature Selection (CFS) measure is that it assesses feature subsets by considering the hypothesis that effective subsets should encompass features that are strongly correlated with the classification while maintaining minimal correlations among themselves. The subsequent equation provides the assessment of the quality of a feature subset S comprising k features:
M e r i t S k = k r c f ¯ k + k k 1 r f f ¯
The symbol r c f ¯ represents the average of all correlations between the features and the classification, while r f f ¯ stands for the average of all correlations between the features themselves. The CFS criterion is formulated in the subsequent manner:
C F S = m a x S k r c f 1 + r c f 2 + + r c f k k + 2 r f 1 f 2 + r f i f j + + r f k f k 1
The variables r f i f j , r c f k are termed as correlations; however, they do not specifically represent Pearson’s correlation coefficient or Spearman’s ρ . In Hall’s dissertation [34], these correlations are not based on either of these methods. Instead, three distinct measures of relatedness are employed: minimum description length (MDL), Symmetrical Uncertainty, and relief. Correlation is explained in detail in Section 4.5, Section 4.6 and Section 4.7.
Assuming xi represents the indicator function for the membership of feature fi in a set, the previous statement can be reformulated into an optimization problem as follows:
C F S = m a x x ϵ 0 , 1 n i = 1 n a i x i 2 i = 1 n x i + i j 2 b i j x i x j
Here, x ϵ 0 , 1 n represents a binary vector x of length n, where each element x i is either 0 or 1. This vector encodes the selection of features in a subset, where 1 indicates that the feature is selected, while 0 indicates that it is not. a i is the coefficient associated with each feature. This coefficient quantifies the correlation between the ith feature and the classification. In other words, they represent how strongly the ith feature is related to the classification b i j . This coefficient represents the correlations between pairs of features. The coefficient quantifies the correlations between different pairs of features, indicating how much they are correlated with each other. Subset Evaluators evaluate the performance of feature subsets by training a classifier on either the entire training dataset or a separate hold-out testing set. They assess the accuracy or other performance metrics of different subsets to determine the most informative features. Common Subset Evaluators include the evaluation of classification accuracy, F1 score, or other relevant metrics to measure the effectiveness of feature subsets. Subset Evaluators assess the performance of feature subsets by training a classifier on either the entire training dataset or a separate hold-out testing set.
From Table 7 and Figure 7 among the feature combinations tested, two stand out for their prominent accuracy scores: Geology and Roads. This combination yields the highest accuracy score of 0.927, showcasing strong predictive power. Similarly, the “Land Cover” and “Roads” combination also achieves an accuracy score of 0.927. The pairing of “Roads” and “Streams” maintains the same accuracy score of 0.927. Combining “Faults” and “Roads” results in a notable accuracy score of 0.915. The combination of “Aspect” and “Roads” follows with a substantial accuracy score of 0.902. These combinations exhibit the highest accuracy scores and thus stand out as the most prominent and predictive feature combinations in the context of the classification task.

4.5. Relief Attribute Evaluator

The Relief algorithm evaluates the worth of an attribute by sampling instances and comparing the attribute values of the nearest instances from the same and different classes. It measures the attribute’s ability to distinguish between classes based on the differences in attribute values for nearby instances. The Relief Attribute Evaluator is commonly used in feature selection for classification tasks, particularly for handling imbalanced datasets. The Relief algorithm evaluates the worth of an attribute by sampling instances and comparing the attribute values of the nearest instances from the same and different classes. Relief computes the weight for each attribute based on the differences in attribute values between the nearest instances and Is an algorithm for feature weighting that demonstrates sensitivity to interactions among features. It strives to estimate the ensuing difference in probabilities related to the weight of a feature X, where
W X = P ( different   value   of   X |   nearest   instance   of   different   class ) P ( different   value   of   X |   nearest   instance   of   same   class )
By eliminating the contextual sensitivity introduced by the “nearest instance” condition, attributes are treated as mutually independent.
R e l i e f X = P different   value   of   X |   different   class P ( different   value   of   X |   same   class )
which can be reformulated as
R e l i e f X = Gin ı ̀ × x X p x   2 1 c C   p c   2   c C   p c   2
where C is the class variable, and c C   p c   2 calculates the sum of probabilities squared for each class. It reflects the concentration of instances in each class, and x X p x   2 calculates the sum of probabilities squared for each possible value of attribute X. It reflects how concentrated the distribution of attribute values is.
Gin ı ̀ = c C   p c 1 p c x X p x 2 x X p x   2 c ϵ C p ( c | x ) ( 1 p ( c | x ) )
Gin ı ̀ represents a modification of an alternative attribute quality measure known as the Gini-index2. Both Gin ı ̀ and the Gini-index share similarities with Information Gain, as they both exhibit a bias towards attributes with a greater number of values. For the symmetrical utilization of the relief method with two features, the measurement can be computed twice (with each feature considered as the “class” in turn), and then the results can be averaged.
From Table 8 and Figure 8, the Relief Attribute Evaluator highlights the variable importance of different geospatial features, with “Roads”, “Geology”, and “Faults” being the most prominent contributors with an importance of 0.227, 0.156, 0.118 to the analysis, respectively.

4.6. Correlation

Correlation measures the linear relationship between an attribute and the class variable. Pearson’s correlation coefficient is a common measure used to assess the strength and direction of the linear association. Higher absolute correlation values indicate a stronger relationship between the attribute and the class variable. Correlation analysis helps identify attributes that exhibit a significant relationship with the target variable.
Correlation measures the linear relationship between an attribute and the class variable. Pearson’s correlation coefficient is commonly used to assess the strength and direction of the linear association. The formula for Pearson’s correlation coefficient is:
Correlation = Cov(x, y)/(StdDev(x) * StdDev(y))
where
Cov(x, y) is the covariance between the attribute x and the class variable y;
StdDev(x) is the standard deviation of attribute x;
StdDev(y) is the standard deviation of the class variable y;
Formula for Pearson Correlation Coefficient:
γ x y = i = 1 n x i x ¯ y i y ¯ i = 1 n x i x ¯ i = 1 n y i y ¯ 2
Here, n represents the sample size, while x i and y i denote individual sample points indexed by i, x ¯ = 1 n i = 1 n x i (the sample mean), similarly for y ¯ .
Rearranging leads us to the following formula for the correlation coefficient γ x y :
γ x y = n x i y i x i y i n x i 2 x i 2   n y i 2 y i 2
Rearranging again give us:
γ x y = i x i y i n x ¯ y ¯ x i 2 n x ¯ 2   y i 2 n y ¯ 2
An alternative representation provides the formula for γ x y as the average of the products of the standardized scores, as shown below:
γ x y = 1 n 1 i = 1 n ( x i x ¯ s x ) ( y i y ¯ s y )
where n, x i , y i , x ¯ , and y ¯ are defined above, while s x and s y are defined as s x = 1 n 1 i = 1 n x i x ¯ 2 , the sample standard deviation, similarly for s y .
Table 9 and Figure 9 provides a ranking of geospatial variables based on their correlation with the target variable, with Aspect being the most important and Roads being the least important regarding correlation with the class variable this information can be used to prioritize variables for further investigation or modeling in their geospatial analysis.

4.7. Symmetrical Uncertainty

Symmetrical Uncertainty is an information-theoretic correlation measure based on entropy. It quantifies the amount of shared information between an attribute and the class variable. The measure takes into account both the attribute’s predictability of the class and the class’s predictability of the attribute. Symmetrical Uncertainty allows for the assessment of the mutual dependence between attributes and the target variable, enabling the identification of informative attributes.
The formula for Symmetrical Uncertainty is:
Symmetrical Uncertainty = (2 * Mutual Information)/(Entropy(X) + Entropy(Y))
where
Mutual Information measures the amount of shared information between the attribute X and the class variable Y;
Entropy(X) is the entropy of attribute X;
Entropy(Y) is the entropy of the class variable Y;
Normalized forms of mutual information can be derived using coefficients of constraint [35], uncertainty coefficients [36,37], or proficiency measures:
C X Y = I X ; Y H Y   and   C Y X = I X ; Y H X
The values of the two coefficients lie within the range of 0 to 1, but they might not be identical. In certain situations, there could be a preference for a symmetric measurement, like the subsequent redundancy measure.
R m a x = min H X , H X H X + H Y
When one variable becomes entirely superfluous given the information about the other variable. Refer also to redundancy in information theory. An additional measure that exhibits symmetry is the symmetric uncertainty [38], which is defined as follows:
U X , Y = 2 R = 2   I X ; Y H X + H Y
This corresponds to the harmonic mean of the two uncertainty coefficients, denoted as C X Y and C Y X . If we view mutual information as a specific instance of total correlation or dual total correlation, their respective normalized versions are as follows:
I X ; Y min H X , H X   a n d   I X ; Y H X , Y
This standardized variant is also recognized as the Information Quality Ratio (IQR) [39], which assesses the quantity of information in one variable concerning another variable in relation to the overall uncertainty.
I Q R X , Y = E I X ; Y = I X ; Y H X , Y = x ϵ X y ϵ Y p x , y log p x p y x ϵ X y ϵ Y p x , y log p x , y
A normalization process [40,41] can be derived by initially considering mutual information as an analogy to covariance (where Shannon entropy is comparable to variance). Subsequently, the normalized mutual information is computed in a manner similar to the Pearson correlation coefficient.
I X ; Y H X H Y
Each of these feature selection techniques offers a unique approach to assess the importance and relevance of geospatial variables for landslide susceptibility mapping. They provide different perspectives and criteria to evaluate the impact of variables on the target variable, allowing researchers and practitioners to gain valuable insights into the most influential variables affecting landslide occurrences along the Karakoram Highway.
From Table 10 and Figure 10 we can observed that the Symmetrical Uncertainty values provide insights into the degree of association between each geospatial variables. “Roads” and “Faults” appear as the most associated features, followed by “Streams” and “Geology”, while “Precipitation” and “Aspect” show minimal to no association.

4.8. Principal Components

Principal Components Analysis (PCA) is a dimensionality reduction technique that transforms the original features into a new set of uncorrelated variables called principal components. These components are ordered based on the amount of variance they explain in the data. PCA helps identify the most important features by considering those associated with the highest-variance principal components, thereby reducing the dimensionality of the dataset.
Mathematically, the transformation is defined by a set of size ι of p-dimensional weight vectors or coefficients w k = w 1 ,   w 2 . w p k that correspond to each row vector xi of matrix X. These weight vectors map each xi to a new vector of principal component scores following the form t i = t i , t ι i , as shown below.
t k i = X i   .   W k
To optimize variance maximization, the initial weight vector W 1 must adhere to the condition:
W 1 = arg max W = 1 i t 1 i 2 = arg max W = 1 i X i . w 2
Alternatively, expressing this in matrix notation yields
W 1 = arg max W = 1 X w 2 = arg max W = 1 w T X T X w w T w
The quantity aimed for maximization can be identified as a Rayleigh quotient. A well-known outcome for a positive semidefinite matrix like X T X is having the upper limit of the quotient be the matrix’s largest eigenvalue, which si achieved when w is the corresponding eigenvector. Once W 1 is determined, the initial principal component of a data vector can be expressed either as a score, t 1 i = X i .   W 1 , in the transformed coordinates, or as the corresponding vector in the original variables, { X i .   W 1 } W 1 .
The k-th component can be computed by subtracting the sum of the first k − 1 principal components from the matrix X, such that
X ^ k = X s = 1 k 1 X w s   w s T
and then determining the weight vector that captures the highest variance from this modified data matrix, where
W k = arg max W = 1 X ^ k w 2 = arg   max w T X ^ k T X ^ k w w T w
Hence, the k-th principal component of a data vector X i can be expressed as a score, where t 1 i = X i .   W 1 in the transformed coordinates. Alternatively, it can be represented as the corresponding vector in the original variable space, X i   .   W k W k , where W k signifies the kth eigenvector of the matrix X T X .
Consequently, the complete decomposition of X into principal components can be represented as:
T = X W
Here, W is a p-by-p matrix of weights, with its columns representing the eigenvectors of X T X . The transpose of W is occasionally referred to as the whitening or sphering transformation. The columns of W, scaled by the square root of corresponding eigenvalues—effectively eigenvectors amplified by variances—are known as loadings in PCA or factor analysis.
X T X can be identified as being proportional to the empirical sample covariance matrix of the dataset X T . The sample covariance “Q” between two distinct principal components across the dataset is defined as
Q P C j , P C k α X w j T X w k = w j T X T X w k = w j T λ k w k = λ k w j T w k
Expressed in matrix notation, the empirical covariance matrix for the original variables can be formulated as
Q α   X T X = w Λ w T
The empirical covariance matrix among the principal components takes on the following form:
w T Q w α w T w Λ w T w = Λ
Here, Λ represents the diagonal matrix comprising the eigenvalues λ k of X T X . Each λ k is the sum of squared values across the dataset for component k; specifically, λ k = i t k i 2 = i X i   .   w k 2 .
Mapping T = XW converts a data vector X i from an initial p-variable space to a fresh p-variable space, in which the variables are uncorrelated across the dataset. However, it is not obligatory to retain all the principal components. Selecting solely the first L principal components, derived by employing only the initial L eigenvectors, results in the truncated transformation:
T L = X w L
Here, the matrix T L now possesses n rows while containing only L columns. In simpler terms, PCA grasps a linear transformation, t = w L T x , x ϵ R p , t ϵ   R L , where x belongs to the p-dimensional space of real numbers, and t belongs to the L-dimensional space of real numbers. In this equation, the columns of the p × L matrix w L constitute an orthogonal basis for the L features (the elements of the representation t), ensuring their lack of correlation. Through its design, among all the transformed data matrices containing only L columns, this score matrix maximizes the retained variance from the initial data while simultaneously minimizing the aggregate squared reconstruction error.
T w T T L w L   T 2 2   or   X X L 2 2
Dimensionality reduction like this is an immensely beneficial step for visualizing and handling datasets with high dimensions, all while retaining the maximum possible variance within the dataset. For instance, when L = 2 is chosen and solely the first two principal components are retained, it results in a two-dimensional plane amidst the high-dimensional dataset in which the data’s dispersion is maximized. Consequently, if the data contains clusters, they will also maximally spread out, making them more distinguishable when plotted on a two-dimensional graph. In contrast, if two directions through the data or two original variables are chosen arbitrarily, the clusters might be far less separated and might even substantially overlap, rendering them practically indistinguishable.
Table 11 and Figure 11 shows variance ratios that tell how much of the total variability in your dataset is explained by each principal component. The first principal component (PC1), which corresponds to Aspect, captures the largest portion of variation, indicating that Aspect is the most dominant variable in terms of explaining the data’s variability. The subsequent components represent decreasing amounts of variance, with Streams having the least impact on the overall variation.
In conclusion, we can observe a consistent presence and importance of the “Roads” feature in all feature selection techniques mentioned above. “Roads” demonstrates several favorable characteristics across different techniques, including high Information Gain; low multicollinearity, as indicated by a low VIF; high accuracy scores in the OneR Classifier; inclusion in high-accuracy feature combinations; and relatively high importance scores from the Relief Attribute Evaluator and for Symmetrical Uncertainty. These findings suggest that “Roads” holds potential significance and influence in the analysis, regardless of the specific evaluation technique or metric employed. Additionally, the analysis provides further insights into other features. For example, “Fault” and “Streams” show relatively high Information Gain, “Aspect” has a high variance ratio in the PCA components, and “Aspect” and “Slope” display lower importance scores and correlation with the target variable. However, to make a final determination, it is recommended that a comprehensive analysis be conducted, including assessing multicollinearity; evaluating the contributions of other features; and considering additional factors, such as interpretability and practicality. Such analysis will help determine the optimal set of variables for accurate and reliable landslide susceptibility mapping.
It can also be evident from ground surveys that the roads are constructed in harsh mountainous terrain with improper blasting that makes the slope very vulnerable near the roads, where most of the landslides occur. The second most influential factor that we identified from our analysis is “Fault”. This factor appears in multiple feature selection techniques (Section 4) with notable characteristics. It has relatively high Information Gain, moderate importance scores, and is included in high-accuracy feature combinations in the subset evaluator table. While it may not be as consistently influential as “Roads”, it still demonstrates significance in the analysis. And this fact can also be evident from ground survey, which highlighted the fact that most landslide happened around the region near fault lines. The “Streams” factor also appears in multiple tables, showing relatively high Information Gain and being included in high-accuracy feature combinations. It demonstrates importance in the analysis, although to a lesser extent compared to “Roads” and “Fault”. Again this phenomena can be verified from our case study, in which most of the land is irrigated from the streams the emerge from rivers that are filled with water from the melting snow and glaciers from high mountains, and even the man-made stream channel makes the region more venerable to landslides. Other features, while not as consistently influential as the three variables mentioned above, show importance in specific tables. For example, “Geology” and “Aspect” have relatively high importance scores in the Relief Attribute Evaluator table, and “Land Cover” has notable importance in the subset evaluator table. It is important to note that the level of influence may vary depending on the specific analysis techniques and evaluation measures used eight different techniques used in our paper. Additionally, the interpretation of influence should consider the context and objectives of the analysis. Further analysis, considering the dataset and specific requirements, is recommended to finalize the most influential variables for the specific task at hand. As we understand, diminishing entropy can lead to a loss of information, potentially causing important patterns to be forfeited [42]. Therefore, it is essential to carefully consider every variable.

5. Metaheuristic and Bayesian Algorithms

5.1. Genetic Algorithms

Although it is difficult to guarantee that the ANN can always generalize the testing data successfully, using a genetic method to avoid the model from getting stuck in a local minimum situation could assist in increasing accuracy rates. The three main components of the Genetic Algorithm (GA) are crossover, mutation, and selection. To realize elitism, the process first chooses the elite parents for the gene pool (a list that monitors the best weighting matrix). The crossover is then put into practice. The method chooses two genes at random out of the best genes (weighted matrix) and recombines them according to a specific strategy defined in the accompanying Python code. For instance, we chose a split point for the elite genes 1 and 2 at random in this instance. We then combine the second portion of gene 2 with the first part of gene 1, and for the remaining pieces of the two genes, we do the reverse procedure. We have two possibly elite genes that were recombined as a result. Then, given that it occurs at random, a mutation could happen. After completing the crossover for each generation, the mechanism will produce a random integer between 0 and 1. A specific area of the weighted matrix, which is likewise created randomly, will be multiplied by yet another random number between 2 and 5. If the randomly generated value is less than or equal to 0.05. To facilitate the mutation process and avoid the ANN model from being trained in the wrong direction (resulting in a poorer training accuracy rate), some weighted matrix values can be slightly scaled. An ANN’s performance could be impacted by a wide range of variables. The number of layers, the number of neurons in each layer, the learning rate, the optimization function, the loss function, and other factors are among them [43,44,45,46,47]. In a Genetic Algorithm, the ANN must be constructed while taking into account the population size, the number of generations, crossover rate, mutation rate, and probability. The model should provide a weighted matrix with a higher variance of values than those with a lower mutation rate and lower mutation probability, for example, if the mutation rate and its mutation probability are high. Its major objective is to attempt to address the flaw in the conventional gradient descent learning technique by ensuring that there is a variety of feasible weighted matrices rather than possibly training a model in an incorrect manner without obtaining the best possible solution for any problem.

5.2. PSO

PSO, which stands for Particle Swarm Optimization, is a metaheuristic optimization algorithm inspired by the behavior of birds flocking or fish schooling. It is commonly used in combination with artificial neural networks (ANNs) to optimize their performance. In the context of ANNs, PSO is employed to determine the optimal values of the ANN’s parameters or weights. The algorithm operates by simulating a population of particles that move through a multidimensional search space. Each particle represents a potential solution or set of parameters for the ANN.
Initially, the particles are assigned random positions and velocities within the search space. The positions represent the values of the ANN’s parameters, while the velocities determine the particles’ movement direction and speed. The algorithm then evaluates the fitness or performance of each particle’s solution by training the ANN with the corresponding parameter values and measuring its accuracy or error. During each iteration of the algorithm, the particles adjust their velocities and positions based on their own historical best solution (the best fitness achieved by the particle itself) and the global best solution (the best fitness achieved by any particle in the swarm). This collective learning and sharing of information guide the particles towards promising regions in the search space.
The adjustment of particle velocities is governed by two main factors: a cognitive component and a social component. The cognitive component represents the particle’s own knowledge, which encourages it to move towards its best solution. The social component reflects the influence of the swarm and directs the particle towards the global best solution. As the iterations progress, the particles converge towards the optimal parameter values, which correspond to the weights that result in the best performance of the ANN. The algorithm terminates when a specified stopping criterion is met, such as reaching a maximum number of iterations or achieving a desired level of performance [5,29,48,49,50]. By using PSO, ANN models can be effectively fine-tuned and optimized to improve their accuracy, convergence speed, and generalization capabilities. PSO provides a powerful approach for exploring and exploiting the parameter space of ANNs, leading to enhanced performance in a variety of tasks and applications.

5.3. BO-GP

Bayesian Optimization Gaussian Process (BO-GP) is a metaheuristic optimization algorithm commonly used for fine-tuning and optimizing artificial neural network (ANN) models. It combines Bayesian optimization, which is a sequential model-based optimization approach, with Gaussian processes, which are statistical models used for modeling the behavior of functions. In the context of ANN models, BO-GP aims to find the optimal set of hyperparameters that result in the best performance for the given task [51,52,53]. These hyperparameters include the number of hidden layers, the number of neurons in each layer, the learning rate, regularization parameters, activation functions, and other architectural choices. The BO-GP algorithm works iteratively as follows:
  • Define a search space: Determine the range or values for each hyperparameter that will be explored during the optimization process.
  • Create an initial design: Select a small set of initial hyperparameter configurations to evaluate the ANN model’s performance. This initial design is often chosen using techniques like random sampling.
  • Build a surrogate model: Fit a Gaussian process regression model to the initial design data. The surrogate model approximates the behavior of the ANN model based on the observed hyperparameter performance pairs.
  • Optimize the acquisition function: The acquisition function guides the selection of the next set of hyperparameters to evaluate. It balances exploration (sampling new regions of the search space) and exploitation (focusing on promising regions). Common acquisition functions include Expected Improvement (EI), Upper Confidence Bound (UCB), and Probability of Improvement (PI).
  • Evaluate the ANN model: Select the next set of hyperparameters based on the optimized acquisition function and evaluate the performance of the ANN model with those hyperparameters. This involves training the ANN on a training set and evaluating its performance on a validation set or using cross-validation.
  • Update the surrogate model: Incorporate the new hyperparameter-performance pair into the existing data and update the Gaussian process regression model. This allows the surrogate model to improve its approximation of the ANN model’s behavior.
  • Repeat steps 4 to 6: Iterate the process by optimizing the acquisition function, evaluating the ANN model, and updating the surrogate model until reaching a specified termination criterion (e.g., a maximum number of iterations or a desired level of performance).
The goal of BO-GP is to efficiently explore the hyperparameter search space and find the optimal configuration that maximizes the performance of the ANN model. By leveraging Bayesian optimization and Gaussian process modeling, BO-GP balances exploration and exploitation, enabling effective hyperparameter tuning and improving the overall performance of ANN models.

5.4. BO-TPE

Bayesian Optimization Tree-structured Parzen Estimator (BO-TPE) is another metaheuristic optimization algorithm commonly used for fine-tuning and optimizing artificial neural network (ANN) models. It is a variant of Bayesian optimization that employs a Tree-structured Parzen Estimator to model the performance of different hyperparameter configurations.
The BO-TPE algorithm works in the following steps:
  • Define a search space: Specify the range or values for each hyperparameter that will be explored during the optimization process.
  • Initialize the hyperparameter sampling: Randomly sample a set of hyperparameter configurations from the search space to create an initial design.
  • Evaluate the initial design: Train and evaluate the ANN models according to the initial set of hyperparameter configurations. This typically involves splitting the data into training and validation sets and using cross-validation or hold-out validation.
  • Build a probabilistic model: Based on the observed hyperparameter-performance pairs from the initial design, construct a probabilistic model using a Tree-structured Parzen Estimator. This model captures the relationship between hyperparameter values and the corresponding performance of the ANN models.
  • Update the model and sample new hyperparameters: The Tree-structured Parzen Estimator uses the previous observations and probabilistic model to guide the sampling of new hyperparameters for the next iteration. It balances exploration (sampling new regions) and exploitation (focusing on promising regions) by considering both the Probability of Improvement and the Expected Improvement.
  • Evaluate the new hyperparameter configurations: Train and evaluate the ANN models with the newly sampled hyperparameter configurations. Update the observed hyperparameter-performance pairs.
  • Update the probabilistic model: Incorporate the new observations into the probabilistic model. This allows the model to refine its estimation of the performance landscape and guide the sampling process more effectively.
  • Repeat steps 5 to 7: Iterate the process by sampling new hyperparameters, evaluating the ANN models, and updating the probabilistic model until reaching a termination criterion, such as a maximum number of iterations or a desired level of performance.
BO-TPE leverages the Bayesian optimization framework to efficiently explore the hyperparameter search space and identify the optimal configuration for the ANN model. By using a Tree-structured Parzen Estimator, it captures the complex relationship between hyperparameters and performance, enabling effective hyperparameter tuning and improving the overall performance of ANN models [54,55,56].

6. Results

Using the Keras model in Python, we evaluated ANN models trained using Adam (Adaptive Moment Estimation). The outputs of the various algorithms are shown below. It is evident from our results that both Bayesian optimization and metaheuristic algorithms, such as Particle Swarm Optimization (PSO) and Genetic Algorithm (GA), can both be effective for optimizing artificial neural networks (ANNs) due to their complementary strengths and capabilities. Bayesian optimization and metaheuristic algorithms can perform equally well for ANN optimization (Figure 12). Bayesian optimization is known for its ability to balance exploitation (exploiting known good solutions) and exploration (exploring new regions in the search space). It uses probabilistic models to guide the search towards promising regions, based on observed performance. On the other hand, metaheuristic algorithms like PSO and GA incorporate stochastic search techniques that can explore the search space more extensively, searching for global optima. ANNs often have a large number of hyperparameters, making the optimization problem high-dimensional. Bayesian optimization excels in handling high-dimensional spaces by building surrogate models that capture the relationship between hyperparameters and performance. It effectively guides the search to promising regions based on the surrogate model’s predictions. Metaheuristic algorithms can also handle high-dimensional spaces through their population-based search, which allows for more diverse exploration.
The performance landscape of ANNs can be complex and non-convex, with multiple local optima. Bayesian optimization and metaheuristic algorithms approach the optimization problem from different perspectives. Bayesian optimization focuses on modeling the performance landscape and exploiting its information to guide the search, while metaheuristic algorithms explore the landscape through heuristics and search operators to escape local optima. Both Bayesian optimization and metaheuristic algorithms offer flexibility in terms of the search strategy and problem representation. They can be applied to different types of optimization problems, including ANN optimization, by adapting their operators or objective functions accordingly. This adaptability allows them to be tuned and customized based on the specific requirements of the ANN optimization task.
Ultimately, the effectiveness of Bayesian optimization and metaheuristic algorithms for ANN optimization can depend on various factors, including the problem complexity, the size of the search space, and the availability of computational resources. It is often beneficial to experiment with different optimization techniques and select the one that performs best for a particular ANN optimization task.
The results of the AUC values (Table 12) indicate the performance of each HPO technique in optimizing the hyperparameters of the ANN for the given ML task. A higher AUC value generally indicates better discrimination power and predictive accuracy of the model.
Based on the given data, BO_TPE achieved the highest AUC value of 0.95289, followed by GA with 0.94987, BO_GP with 0.94685, and PSO with 0.93538. This suggests that BO_TPE performed the best among the evaluated HPO techniques for the given ML task.
There could be several reasons why BO_TPE outperformed the other techniques. BO_TPE combines both exploration and exploitation strategies effectively. It intelligently explores the hyperparameter space to identify promising regions and then exploits the information gained to refine the search for optimal hyperparameters. This balance between exploration and exploitation can lead to better performance. BO_TPE utilizes a sequential Bayesian optimization strategy, which leverages the information gathered from previous iterations to guide the search for optimal hyperparameters. This adaptive approach helps in efficiently exploring the hyperparameter space and quickly converging onto promising solutions. BO_TPE employs a Tree-structured Parzen Estimator (TPE) to efficiently explore the hyperparameter space. TPE focuses more on areas that are likely to contain better hyperparameter configurations, allowing for more efficient optimization compared to other techniques. BO_TPE utilizes the observed performance of previous hyperparameter configurations to build a probabilistic model and guide the search towards regions with higher potential. This exploitation of information helps in quickly identifying and refining good hyperparameter settings.
It is important to note that the performance of HPO techniques can vary depending on the specific dataset, ML task, and hyperparameter space being considered. The given results suggest that for the specific ML task and dataset in this study, BO_TPE performed better than GA, BO_GP, and PSO in optimizing the hyperparameters of the ANN model, resulting in higher predictive accuracy and discrimination power, as indicated by the AUC values.
The fact that GA and BO_TPE performed equally well, as shown in Table 12 and Figure 13, might suggest that both algorithms were able to navigate the search space effectively to find good solutions. As discussed earlier, GA uses a population-based approach, in which potential solutions evolve over generations through processes like mutation and crossover. BO_TPE, on the other hand, models the distribution of the objective function and uses Bayesian reasoning to focus the search on promising regions of the hyperparameter space. In this particular case, the characteristics of the problem may have allowed both techniques to reach similar levels of optimization performance.
Bayesian Optimization Gaussian Processes (BO_GP) outperformed Particle Swarm Optimization (PSO) in terms of AUC. BO_GP’s advantage might be attributed to its ability to model the objective function as a probabilistic surrogate, allowing it to make informed decisions about where to explore the hyperparameter space. PSO, while a powerful optimization technique inspired by social behavior of birds or fish, can sometimes struggle in complex or high-dimensional spaces due to its exploration and exploitation balance. It relies on particle movement guided by personal and global best solutions, which may not be as effective as BO_GP’s probabilistic modeling for certain problems. It’s possible that the optimization landscape of this problem favored the probabilistic modeling and informed exploration strategy employed by BO_GP, leading to its better performance compared to PSO.

7. Conclusions

In conclusion, this research endeavors to enhance the accuracy and reliability of predicting landslide susceptibility along the Karakoram Highway (KKH) by employing a comprehensive methodology that combines feature selection and hyperparameter optimization. This study acknowledges the intricate nature of landslide occurrences and the critical role that relevant geospatial variables and well-tuned model parameters play in accurate predictions. Through a rigorous process of experimentation and analysis, the study successfully integrates modern techniques to address the challenges posed by landslides in the Karakoram region. The utilization of artificial neural networks (ANNs) as predictive models, along with four distinct hyperparameter optimization techniques—Particle Swarm Optimization (PSO), Genetic Algorithm (GA), Bayesian Optimization Gaussian Process (BO_GP), and Bayesian Optimization Tree-structured Parzen Estimator (BO_TPE)—illustrates the commitment to refining model configurations for optimal performance. Furthermore, this study recognizes the significance of selecting pertinent features from the dataset. The incorporation of eight feature selection algorithms, each evaluating variables through different lenses, such as multicollinearity, correlations, information sharing, variance, and uncertainty, reflects a comprehensive approach to identifying the most influential geospatial variables contributing to landslide susceptibility. The synergy achieved by combining feature selection and hyperparameter optimization is instrumental in addressing the complexities of predicting landslide occurrences. The advantages encompass the prevention of overfitting, controlled model complexity, generalization to new and unseen data, the efficient utilization of computational resources, adaptation to varying contexts, and robustness in the face of changing data distributions. This approach is not only valuable for accurate predictions but also contributes to a deeper understanding of the underlying processes leading to landslides.
The study’s findings provide a substantial contribution to the field of geospatial analysis, disaster management, and infrastructure planning. By offering insight into optimal feature selection and model parameterization, the research equips decision makers with valuable tools to develop effective mitigation strategies, plan resilient infrastructure, and manage hazards along the KKH. The combined methodology of feature selection and hyperparameter optimization stands as a testament to the potential of integrating advanced techniques to tackle complex challenges. The research not only enhances the accuracy of landslide susceptibility predictions but also underscores the broader applicability of such integrated approaches in various domains of data analysis and decision making. As natural calamities, like landslides, continue to pose threats to communities and infrastructure, this study exemplifies the kind of innovative thinking that can lead to more informed and proactive disaster management efforts.

Author Contributions

Conceptualization, F.A. (Farkhanda Abbas), F.A. (Fazila Abbas) and G.K.; Methodology, F.A. (Farkhanda Abbas) and F.A. (Fazila Abbas); Software, F.A. (Farkhanda Abbas); Validation, D.H.; Formal analysis, M.I., D.H. and M.F.A.; Investigation, M.I.; Resources, A.F.A. and M.F.A.; Data curation, M.I., J.I., G.K., A.F.A. and M.F.A.; Writing—original draft, F.A. (Farkhanda Abbas); Writing—review & editing, F.A. (Fazila Abbas) and J.I.; Supervision, F.Z.; Project administration, G.K. and A.F.A. All authors have read and agreed to the published version of the manuscript.

Funding

We extend our appreciation to the Researchers Supporting Project (no. RSP2023R218), King Saud University, Riyadh, Saudi Arabia.

Data Availability Statement

The data presented in the study are available upon request from the first and corresponding authors. The data are not publicly available due to the thesis that is being prepared using these data.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hagan, M.; Demuth, H.; Beale, M. Neural Network Design; PWS: Boston, MA, USA, 1996. [Google Scholar]
  2. Werbos, P.J. Backpropagation through time: What it does and how to do it. Proc. IEEE 1990, 78, 1550–1560. [Google Scholar] [CrossRef]
  3. Hertz, J.; Krogh, A.; Palmer, R.G.; Horner, H. Introduction to the theory of neural computation. Phys. Today 1991, 44, 70. [Google Scholar] [CrossRef]
  4. Hagan, M.T.; Menhaj, M.B. Training feedforward networks with the Marquardt algorithm. IEEE Trans. Neural Netw. 1994, 5, 989–993. [Google Scholar] [CrossRef]
  5. Moayedi, H.; Mehrabi, M.; Mosallanezhad, M.; Rashid, A.S.A.; Pradhan, B. Modification of landslide susceptibility mapping using optimized PSO-ANN technique. Eng. Comp. 2019, 35, 967–984. [Google Scholar] [CrossRef]
  6. Abbas, F.; Zhang, F.; Ismail, M.; Khan, G.; Iqbal, J.; Alrefaei, A.F.; Albeshr, M.F. Optimizing Machine Learning Algorithms for Landslide Susceptibility Mapping along the Karakoram Highway, Gilgit Baltistan, Pakistan: A Comparative Study of Baseline, Bayesian, and Metaheuristic Hyperparameter Optimization Techniques. Sensors 2023, 23, 6843. [Google Scholar] [CrossRef] [PubMed]
  7. Mamindla, A.K.; Ramadevi, D.; Sedova, O.V.; Alekseev, A.G.; Ashraf, M.T.; Ariffin, D.S.B.B.; Samy, S.; Fouzia, E.; Nabil, E.; Khalid, S. ANN-ABC meta-heuristic hyper parameter tuning for mammogram classification. J. Theor. Appl. Inf. Technol. 2023, 101. Available online: http://jatit.org/volumes/hundredone1.php (accessed on 29 August 2023).
  8. Lisboa, P.J.; Taktak, A.F. The use of artificial neural networks in decision support in cancer: A systematic review. Neural Netw. 2006, 19, 408–415. [Google Scholar] [CrossRef] [PubMed]
  9. Maniezzo, V. Genetic evolution of the topology and weight distribution of neural networks. IEEE Trans. Neural Netw. 1994, 5, 39–53. [Google Scholar] [CrossRef]
  10. Fischer, M.M.; Leung, Y. A genetic-algorithms based evolutionary computational neural network for modelling spatial interaction dataNeural network for modelling spatial interaction data. Ann. Reg. Sci. 1998, 32, 437–458. [Google Scholar] [CrossRef]
  11. Irani, R.; Nasimi, R. Evolving neural network using real coded genetic algorithm for permeability estimation of the reservoir. Expert Syst. Appl. 2011, 38, 9862–9866. [Google Scholar] [CrossRef]
  12. Miller, G.F.; Todd, P.M.; Hegde, S.U. Designing Neural Networks Using Genetic Algorithms. ICGA 1989, 89, 379–384. [Google Scholar]
  13. Liu, Z.; Liu, A.; Wang, C.; Niu, Z. Evolving neural network using real coded genetic algorithm (GA) for multispectral image classification. Future Gener. Comp. Syst. 2004, 20, 1119–1129. [Google Scholar] [CrossRef]
  14. Anselin, L. The Moran scatterplot as an ESDA tool to assess local instability in spatial association. In Spatial Analytical Perspectives on GIS; Routledge: London, UK, 2019; pp. 111–126. [Google Scholar]
  15. Wagner, H.H.; Fortin, M.-J. Spatial analysis of landscapes: Concepts and statistics. Ecology 2005, 86, 1975–1987. [Google Scholar] [CrossRef]
  16. Ma, J.; Scott, N.R.; DeGloria, S.D.; Lembo, A.J. Siting analysis of farm-based centralized anaerobic digester systems for distributed generation using GIS. Biomass Bioenergy 2005, 28, 591–600. [Google Scholar] [CrossRef]
  17. Van Loock, P. Optical hybrid approaches to quantum information. Laser Photon. Rev. 2011, 5, 167–200. [Google Scholar] [CrossRef]
  18. Hjortdal, V.E.; Redington, A.; De Leval, M.; Tsang, V. Hybrid approaches to complex congenital cardiac surgery. Eur. J. Cardio-Thorac. Surg. 2002, 22, 885–890. [Google Scholar] [CrossRef]
  19. Wong, Y.J.; Arumugasamy, S.K.; Jewaratnam, J. Performance comparison of feedforward neural network training algorithms in modeling for synthesis of polycaprolactone via biopolymerization. Clean Technol. Environ. Policy 2018, 20, 1971–1986. [Google Scholar] [CrossRef]
  20. Hessami, M.; Anctil, F.; Viau, A.A. Selection of an artificial neural network model for the post-calibration of weather radar rainfall estimation. J. Data Sci. 2004, 2, 107–124. [Google Scholar] [CrossRef]
  21. Arthur, C.K.; Temeng, V.A.; Ziggah, Y.Y. Performance evaluation of training algorithms in backpropagation neural network approach to blast-induced ground vibration prediction. Ghana Min. J. 2020, 20, 20–33. [Google Scholar] [CrossRef]
  22. Abbas, F.; Zhang, F.; Ismail, M.x.; Khan, G.; Iqbal, J.; Alrefaei, A.F.; Albeshr, M.F. Hyperparameter Optimization for Landslide Susceptibility Mapping: A Comparison between Baseline, Bayesian and Metaheuristic Hyperparameter Optimization Techniques for Machine Learning Algorithms. Preprint 2023. [Google Scholar] [CrossRef]
  23. Gardner, M.W.; Dorling, S. Artificial neural networks (the multilayer perceptron)—A review of applications in the atmospheric sciences. Atmos. Environ. 1998, 32, 2627–2636. [Google Scholar] [CrossRef]
  24. Lisboa, P.J. A review of evidence of health benefit from artificial neural networks in medical intervention. Neural Netw. 2002, 15, 11–39. [Google Scholar] [CrossRef] [PubMed]
  25. Watson, I.; Gurd, J. A practical data flow computer. Computer 1982, 15, 51–57. [Google Scholar] [CrossRef]
  26. Rogers, L.L.; Dowla, F.U. Optimization of groundwater remediation using artificial neural networks with parallel solute transport modeling. Water Resour. Res. 1994, 30, 457–481. [Google Scholar] [CrossRef]
  27. Meireles, M.R.; Almeida, P.E.; Simões, M.G. A comprehensive review for industrial applicability of artificial neural networks. IEEE Trans. Ind. Electron. 2003, 50, 585–601. [Google Scholar] [CrossRef]
  28. Yang, H.; Alphones, A.; Xiong, Z.; Niyato, D.; Zhao, J.; Wu, K. Artificial-intelligence-enabled intelligent 6G networks. IEEE Netw. 2020, 34, 272–280. [Google Scholar] [CrossRef]
  29. Geethanjali, M.; Slochanal, S.M.R.; Bhavani, R. PSO trained ANN-based differential protection scheme for power transformers. Neurocomputing 2008, 71, 904–918. [Google Scholar] [CrossRef]
  30. Sökefeld, M. The Attabad Landslide and the Politics of Disaster in Gojal, Gilgit-Baltistan. 2012. Available online: https://epub.ub.uni-muenchen.de/29608/1/Martin_Soekefeld_The_Attabad_landslide_and_the_politics_of_disaster_in_Gojal,_Gilgit-Baltistan.pdf (accessed on 29 August 2023).
  31. Gardezi, H.; Bilal, M.; Cheng, Q.; Xing, A.; Zhuang, Y.; Masood, T. A comparative analysis of attabad landslide on january 4, 2010, using two numerical models. Nat. Hazards 2021, 107, 519–538. [Google Scholar] [CrossRef]
  32. Gibson, L.; Wilson, B.; Cahill, D.; Hill, J. Spatial prediction of rufous bristlebird habitat in a coastal heathland: A GIS-based approach. J. Appl. Ecol. 2004, 41, 213–223. [Google Scholar] [CrossRef]
  33. Jankowski, P.; Nyerges, T. GIS-supported collaborative decision making: Results of an experiment. Ann. Assoc. Am. Geogr. 2001, 91, 48–70. [Google Scholar] [CrossRef]
  34. Hall, A.M.; Ferreira, P.H.; Maher, C.G.; Latimer, J.; Ferreira, M.L. The influence of the therapist-patient relationship on treatment outcome in physical rehabilitation: A systematic review. Phys. Ther. 2010, 90, 1099–1110. [Google Scholar] [CrossRef] [PubMed]
  35. Fang, S.-C.; Hu, C.-F.; Wang, H.-F.; Wu, S.-Y. Linear programming with fuzzy coefficients in constraints. Comput. Math. Appl. 1999, 37, 63–76. [Google Scholar] [CrossRef]
  36. Imhoff, E.A., Jr.; Lobo, G.J. The effect of ex ante earnings uncertainty on earnings response coefficients. Account. Rev. 1992, 67, 427–439. [Google Scholar]
  37. Feltham, G.A.; Pae, J. Analysis of the impact of accounting accruals on earnings uncertainty and response coefficients. J. Account. Audit. Financ. 2000, 15, 199–220. [Google Scholar] [CrossRef]
  38. Witten, I.H.; Frank, E.; Hall, M.A.; Pal, C.J.; Data, M. Practical machine learning tools and techniques. In Data Mining; Morgan Kaufmann: San Francisco, CA, USA, 2005; pp. 403–413. [Google Scholar]
  39. Wijaya, D.R.; Sarno, R.; Zulaika, E. Information Quality Ratio as a novel metric for mother wavelet selection. Chemom. Intell. Lab. Syst. 2017, 160, 59–71. [Google Scholar] [CrossRef]
  40. Rioul, O. Information theoretic proofs of entropy power inequalities. IEEE Trans. Inf. Theory 2010, 57, 33–55. [Google Scholar] [CrossRef]
  41. Pham, D.-T.; Cardoso, J.-F. Blind separation of instantaneous mixtures of nonstationary sources. IEEE Trans. Signal Proc. 2001, 49, 1837–1848. [Google Scholar] [CrossRef]
  42. Abbas, F.; Zhang, F.; Iqbal, J.; Abbas, F.; Alrefaei, A.F.; Albeshr, M. Assessing the Dimensionality Reduction of the Geospatial Dataset Using Principal Component Analysis (PCA) and Its Impact on the Accuracy and Performance of Ensembled and Non-ensembled Algorithms. Preprints.org 2023, 2023070529. [Google Scholar] [CrossRef]
  43. Idrissi, M.A.J.; Ramchoun, H.; Ghanou, Y.; Ettaouil, M. Genetic algorithm for neural network architecture optimization. In Proceedings of the 2016 3rd International Conference on Logistics Operations Management (GOL), Fez, Morocco, 23–25 May 2016; pp. 1–4. [Google Scholar]
  44. Roy, G.; Lee, H.; Welch, J.L.; Zhao, Y.; Pandey, V.; Thurston, D. A distributed pool architecture for genetic algorithms. In Proceedings of the 2009 IEEE Congress on Evolutionary Computation, Trondheim, Norway, 18–21 May 2009; pp. 1177–1184. [Google Scholar]
  45. Pospichal, P.; Jaros, J.; Schwarz, J. Parallel genetic algorithm on the cuda architecture. In Proceedings of the Applications of Evolutionary Computation: EvoApplicatons 2010: EvoCOMPLEX, EvoGAMES, EvoIASP, EvoINTELLIGENCE, EvoNUM, and EvoSTOC, Part I, Istanbul, Turkey, 7–9 April 2010; pp. 442–451. [Google Scholar]
  46. Yu, T.-L.; Yassine, A.A.; Goldberg, D.E. An information theoretic method for developing modular architectures using genetic algorithms. Res. Eng. Design 2007, 18, 91–109. [Google Scholar] [CrossRef]
  47. DeLanda, M. Deleuze and the Use of the Genetic Algorithm in Architecture. Archit. Design 2002, 71, 9–12. [Google Scholar]
  48. Zhang, C.; Shao, H.; Li, Y. Particle swarm optimisation for evolving artificial neural network. In Proceedings of the 2000 IEEE International Conference on Systems, Man and Cybernetics, Cybernetics Evolving to Systems, Humans, Organizations, and Their Complex Interactions, Nashville, TN, USA, 8–11 October 2000; pp. 2487–2490. [Google Scholar]
  49. Qi, C.; Fourie, A.; Chen, Q. Neural network and particle swarm optimization for predicting the unconfined compressive strength of cemented paste backfill. Constr. Build. Mater. 2018, 159, 473–478. [Google Scholar] [CrossRef]
  50. Fregoso, J.; Gonzalez, C.I.; Martinez, G.E. Optimization of convolutional neural networks architectures using PSO for sign language recognition. Axioms 2021, 10, 139. [Google Scholar] [CrossRef]
  51. Kadhim, Z.S.; Abdullah, H.S.; Ghathwan, K.I. Artificial Neural Network Hyperparameters Optimization: A Survey. Int. J. Online Biomed. Eng. 2022, 18, 59–87. [Google Scholar] [CrossRef]
  52. Masum, M.; Shahriar, H.; Haddad, H.; Faruk, M.J.H.; Valero, M.; Khan, M.A.; Rahman, M.A.; Adnan, M.I.; Cuzzocrea, A.; Wu, F. Bayesian hyperparameter optimization for deep neural network-based network intrusion detection. In Proceedings of the 2021 IEEE International Conference on Big Data (Big Data), Orlando, FL, USA, 15–18 December 2021; pp. 5413–5419. [Google Scholar]
  53. Hosein Ghanemi, A.; Tarighat, A. Use of Different Hyperparameter Optimization Algorithms in ANN for Predicting the Compressive Strength of Concrete Containing Calcined Clay. Pract. Period. Struct. Design Constr. 2022, 27, 04022002. [Google Scholar] [CrossRef]
  54. Tørring, J.O.; Elster, A.C. Analyzing Search Techniques for Autotuning Image-based GPU Kernels: The Impact of Sample Sizes. In Proceedings of the 2022 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW), Lyon, France, 30 May–3 June 2022; pp. 972–981. [Google Scholar]
  55. Janković, R.; Mihajlović, I.; Štrbac, N.; Amelio, A. Machine learning models for ecological footprint prediction based on energy parameters. Neural Comp. Appl. 2021, 33, 7073–7087. [Google Scholar] [CrossRef]
  56. Elola, A.; Aramendi, E.; Irusta, U.; Picón, A.; Alonso, E.; Owens, P.; Idris, A. Deep neural networks for ECG-based pulse detection during out-of-hospital cardiac arrest. Entropy 2019, 21, 305. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Our research focuses on the northern region of Pakistan. The KKH is place precisely where our study area is located.
Figure 1. Our research focuses on the northern region of Pakistan. The KKH is place precisely where our study area is located.
Remotesensing 15 04330 g001
Figure 2. (a) Slope; (b) precipitation; (c) land cover; (d) aspect; (e) geology; (f) proximity to streams; (g) proximity to faults; (h) proximity to roads.
Figure 2. (a) Slope; (b) precipitation; (c) land cover; (d) aspect; (e) geology; (f) proximity to streams; (g) proximity to faults; (h) proximity to roads.
Remotesensing 15 04330 g002
Figure 3. Methodology used for landslide susceptibility mapping along the KHH.
Figure 3. Methodology used for landslide susceptibility mapping along the KHH.
Remotesensing 15 04330 g003
Figure 4. Information Gain IG obtained for the geospatial variables.
Figure 4. Information Gain IG obtained for the geospatial variables.
Remotesensing 15 04330 g004
Figure 5. Variable inflation factor VIF obtained for the geospatial variables.
Figure 5. Variable inflation factor VIF obtained for the geospatial variables.
Remotesensing 15 04330 g005
Figure 6. Feature accuracy score obtained for the geospatial variables using OneR Classifier.
Figure 6. Feature accuracy score obtained for the geospatial variables using OneR Classifier.
Remotesensing 15 04330 g006
Figure 7. Feature combination accuracy score obtained for geospatial variables using subset evaluator.
Figure 7. Feature combination accuracy score obtained for geospatial variables using subset evaluator.
Remotesensing 15 04330 g007
Figure 8. Geospatial variable importance obtained from Relief Attribute Evaluator.
Figure 8. Geospatial variable importance obtained from Relief Attribute Evaluator.
Remotesensing 15 04330 g008
Figure 9. Geospatial variable importance obtained from Correlation.
Figure 9. Geospatial variable importance obtained from Correlation.
Remotesensing 15 04330 g009
Figure 10. Symmetrical Uncertainty obtained for the geospatial variables.
Figure 10. Symmetrical Uncertainty obtained for the geospatial variables.
Remotesensing 15 04330 g010
Figure 11. The covariance matrix is computed to understand the relationships between the features in the dataset. It represents how each feature changes with respect to the others.
Figure 11. The covariance matrix is computed to understand the relationships between the features in the dataset. It represents how each feature changes with respect to the others.
Remotesensing 15 04330 g011
Figure 12. AUC value for GA, BO_GP, BO_TPE, and PSO for ANN model.
Figure 12. AUC value for GA, BO_GP, BO_TPE, and PSO for ANN model.
Remotesensing 15 04330 g012
Figure 13. The landslide susceptibility maps generated by ANN using different optimization techniques, including BO_TPE, BO_GP, PSO and GA.
Figure 13. The landslide susceptibility maps generated by ANN using different optimization techniques, including BO_TPE, BO_GP, PSO and GA.
Remotesensing 15 04330 g013
Table 1. The information given includes the hyperparameters related to the optimization techniques (PSO, GA, BO_GP, and BO_TPE).
Table 1. The information given includes the hyperparameters related to the optimization techniques (PSO, GA, BO_GP, and BO_TPE).
HOP TechniqueActivationBatch SizeEpochsNeurons
PSO1.7420.36740.27395.430
GAsgd165064
BO_GPtanh164754
BO_TPE1165080
Table 2. Detail of data sources for geospatial variables used in our case study.
Table 2. Detail of data sources for geospatial variables used in our case study.
DataVariablesScale/ResolutionSource
Sentinel 2 Satellite ImagesLandslide inventory, LCLU, Road network10 m
DEMSlope
Aspect
Stream
Network
30 mSRTM Shuttle Radar Topography Mission (USGS) United States Geological Survey
Geological MapGeology Units and Fault lines30 mGeological Survey of Pakistan
Google Earth MapsLandslide Inventory Land Cover/Land Use Road Network2–5 m
Field SurveyGPS Points1 m
Table 3. Geospatial variables used in our case study for landslide susceptibility mapping.
Table 3. Geospatial variables used in our case study for landslide susceptibility mapping.
VariablesClassesClass Percentage %Landslide Percentage %Reclassification
Slope (°)Very Gentle Slope < 5°17.3621.11Geometrical interval reclassification
Gentle Slope 5–15°20.8728.37
Moderately Steep Slope 15–30°26.6437.89
Steep Slope 30–45°24.4010.90
Escarpments > 45°10.711.73
AspectFlat (−1)22.867.04Remained unmodified (as is from source data).
North (0–22)21.477.03
Northeast (22–67)14.855.00
East (67–112)8.0011.86
Southeast (112–157)5.2214.3
South (157–202)2.8414.40
Southwest (202–247)6.4612.41
West (247–292)7.1916.03
Northwest (292–337)11.0711.96
Land CoverDense Conifer0.3812.73
Sparse Conifer0.2512.80
Broadleaved, Conifer1.5210.86
Grasses/Shrubs25.5410.3
Agriculture Land5.7810.40
Soil/Rocks56.5514.51
Snow/Glacier8.8912.03
Water1.0616.96
GeologyCretaceous Sandstone13.706.38
Devonian–Carboniferous12.345.80
Chalt Group1.438.43
Hunza Plutonic Unit4.7410.74
Paragneisses11.3811.34
Yasin Group10.8010.70
Gilgit Complex5.809.58
Trondhjemite15.659.32
Permian Massive Limestone6.516.61
Permanent Ice12.613.51
Quaternary Alluvium0.328.65
Triassic Massive Limestone and Dolomite1.587.80
Snow3.082.00
Proximity to Streams (meter)0–10019.3718.52Geometrical interval reclassification
100–20010.2621.63
200–30010.7825.16
300–40013.9526.12
400–50018.696.23
>50026.922.34
Proximity to Roads
(meter)
0–10081.0825.70
100–20010.3425.19
200–3006.7227.09
300–4001.2512.02
400–5000.6010.00
Proximity to Faults (meter)000–1000 29.7627.30
2000–300036.2537.40
>300034.1535.03
Table 4. Information Gain IG obtained for the geospatial variables.
Table 4. Information Gain IG obtained for the geospatial variables.
FeatureInformation Gain
Roads0.352
Faults0.195
Streams0.091
Geology0.039
Slope0.039
Precipitation0.012
Aspect0.000
Land Cover0.000
Table 5. Variable inflation factor VIF obtained for the geospatial variables.
Table 5. Variable inflation factor VIF obtained for the geospatial variables.
FeatureVIF
Land Cover18.090
Precipitation7.167
Slope6.276
Geology4.986
Aspect4.585
Streams3.910
Faults3.532
Roads2.992
Table 6. Feature accuracy score obtained for the geospatial variables using OneR Classifier.
Table 6. Feature accuracy score obtained for the geospatial variables using OneR Classifier.
Feature Scores—OneR Classifier
FeatureAccuracy Score
Roads0.902
Faults0.707
Streams0.659
Land Cover0.622
Aspect0.585
Geology0.561
Slope0.549
Precipitation0.500
Selected Feature: Roads
Accuracy Score: 0.902
Table 7. Feature combination accuracy score obtained for geospatial variables using subset evaluator.
Table 7. Feature combination accuracy score obtained for geospatial variables using subset evaluator.
Feature CombinationAccuracy Score
Geology, Roads0.927
Land Cover, Roads0.927
Roads, Streams0.927
Faults, Roads0.915
Aspect, Roads0.902
Roads, Slope0.878
Precipitation, Roads0.854
Faults, Streams0.793
Faults, Precipitation0.780
Faults, Geology0.744
Faults, Land Cover0.744
Faults, Slope0.732
Aspect, Faults0.720
Geology, Land Cover0.720
Geology, Precipitation0.720
Geology, Streams0.707
Aspect, Streams0.683
Aspect, Precipitation0.671
Aspect, Slope0.671
Land Cover, Streams0.634
Precipitation, Streams0.634
Aspect, Geology0.622
Land Cover, Precipitation0.622
Slope, Streams0.622
Geology, Slope0.610
Land Cover, Slope0.610
Aspect, Land Cover0.598
Precipitation, Slope0.573
Table 8. Geospatial variable importance obtained from Relief Attribute Evaluator.
Table 8. Geospatial variable importance obtained from Relief Attribute Evaluator.
FeatureImportance
Roads0.227
Geology0.156
Faults0.118
Streams0.112
Precipitation0.060
Aspect0.050
Slope0.035
Land Cover0.015
Table 9. Geospatial variable importance obtained from Correlation.
Table 9. Geospatial variable importance obtained from Correlation.
FeatureImportance
Aspect1.000
Geology0.100
Precipitation0.090
Faults0.066
Land Cover0.043
Streams0.022
Slope0.020
Roads0.005
Table 10. Symmetrical Uncertainty obtained for the geospatial variables.
Table 10. Symmetrical Uncertainty obtained for the geospatial variables.
FeatureSymmetrical Uncertainty
Roads0.683
Faults0.435
Streams0.247
Geology0.128
Land Cover0.093
Slope0.053
Precipitation0.001
Aspect0.000
Table 11. Principal Component Analysis (PCA) for geospatial variables used in our experiment.
Table 11. Principal Component Analysis (PCA) for geospatial variables used in our experiment.
ComponentsVariance Ratio
Aspect27.22%
Fault18.17%
Geology13.66%
Land Cover12.50%
Precipitation10.57%
Roads9.70%
Slope4.71%
Streams3.48%
Table 12. AUC value for GA, BO_GP, BO_TPE and PSO for ANN model.
Table 12. AUC value for GA, BO_GP, BO_TPE and PSO for ANN model.
HPO TechniquesMLAUC
GAANN0.950
BO_GPANN0.947
BO_TPEANN0.953
PSOANN0.935
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Abbas, F.; Zhang, F.; Abbas, F.; Ismail, M.; Iqbal, J.; Hussain, D.; Khan, G.; Alrefaei, A.F.; Albeshr, M.F. Landslide Susceptibility Mapping: Analysis of Different Feature Selection Techniques with Artificial Neural Network Tuned by Bayesian and Metaheuristic Algorithms. Remote Sens. 2023, 15, 4330. https://doi.org/10.3390/rs15174330

AMA Style

Abbas F, Zhang F, Abbas F, Ismail M, Iqbal J, Hussain D, Khan G, Alrefaei AF, Albeshr MF. Landslide Susceptibility Mapping: Analysis of Different Feature Selection Techniques with Artificial Neural Network Tuned by Bayesian and Metaheuristic Algorithms. Remote Sensing. 2023; 15(17):4330. https://doi.org/10.3390/rs15174330

Chicago/Turabian Style

Abbas, Farkhanda, Feng Zhang, Fazila Abbas, Muhammad Ismail, Javed Iqbal, Dostdar Hussain, Garee Khan, Abdulwahed Fahad Alrefaei, and Mohammed Fahad Albeshr. 2023. "Landslide Susceptibility Mapping: Analysis of Different Feature Selection Techniques with Artificial Neural Network Tuned by Bayesian and Metaheuristic Algorithms" Remote Sensing 15, no. 17: 4330. https://doi.org/10.3390/rs15174330

APA Style

Abbas, F., Zhang, F., Abbas, F., Ismail, M., Iqbal, J., Hussain, D., Khan, G., Alrefaei, A. F., & Albeshr, M. F. (2023). Landslide Susceptibility Mapping: Analysis of Different Feature Selection Techniques with Artificial Neural Network Tuned by Bayesian and Metaheuristic Algorithms. Remote Sensing, 15(17), 4330. https://doi.org/10.3390/rs15174330

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop