Next Article in Journal
Optimizing Time Series Models for Forecasting Environmental Variables: A Rainfall Case Study
Previous Article in Journal
Artificial Water Bodies in Post-Industrial and Urban Landscapes—A Case Study on Assessing Their Potential in Blue–Green Urban Infrastructure
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Machine Learning Models of the Geospatial Distribution of Groundwater Quality: A Systematic Review

1
Department of Earth and Environmental Sciences, School of Natural Sciences, Faculty of Science and Engineering, The University of Manchester, Manchester M13 9PL, UK
2
Department of Mathematics, School of Natural Sciences, Faculty of Science and Engineering, The University of Manchester, Manchester M13 9PL, UK
*
Author to whom correspondence should be addressed.
Water 2025, 17(19), 2861; https://doi.org/10.3390/w17192861
Submission received: 9 August 2025 / Revised: 22 September 2025 / Accepted: 28 September 2025 / Published: 30 September 2025
(This article belongs to the Section Hydrogeology)

Abstract

Assessing the quality of groundwater, a primary source of water in many sectors, is of paramount importance. To this end, modeling the geospatial distribution of chemical contaminants in groundwater can be of great utility. Machine learning (ML) models are being increasingly used to overcome the shortcomings of conventional predictive techniques. We report here a systematic review of the nature and utility of various supervised and unsupervised ML models during the past two decades of machine learning groundwater hazard mapping (MLGHM). We identified and reviewed 284 relevant MLGHM journal articles that met our inclusion criteria. Firstly, trend analysis showed (i) an exponential increase in the number of MLGHM studies published between 2004 and 2025, with geographical distribution outlining Iran, India, the US, and China as the countries with the most extensively studied areas; (ii) nitrate as the most studied target, and groundwater chemicals as the most frequently considered category of predictive variables; (iii) that tree-based ML was the most popular model for feature selection; (iv) that supervised ML was far more favored than unsupervised ML (94% vs. 6% of models) with tree-based category—mostly random forest (RF)—as the most popular supervised ML. Secondly, compiling accuracy-based comparisons of ML models from the explored literature revealed that RF, deep learning, and ensembles (mostly meta-model ensembles and boosting ensembles) were frequently reported as the most accurate models. Thirdly, a critical evaluation of MLGHM models in terms of predictive accuracy, along with several other factors such as models’ computational efficiency and predictive power—which have often been overlooked in earlier review studies—resulted in considering the relative merits of commonly used MLGHM models. Accordingly, a flowchart was designed by integrating several MLGHM key criteria (i.e., accuracy, transparency, training speed, number of hyperparameters, intended scale of modeling, and required user’s expertise) to assist in informed model selection, recognising that the weighting of criteria for model selection may vary from problem to problem. Lastly, potential challenges that may arise during different stages of MLGHM efforts are discussed along with ideas for optimizing MLGHM models.

1. Introduction

1.1. Groundwater and Contamination Hazard

Groundwater is a crucial source of water for different sectors all over the globe, including both developed and developing countries. Agricultural, domestic, and industrial uses account for 69%, 22%, and 9%, respectively, of groundwater abstractions [1,2]. With the increasing populations and correspondingly increased industrial and agricultural economic activities, there are global concerns about both the quantity [3,4,5,6] and quality [7,8] of these resources. Focusing on the quality, groundwater contamination has emerged as a serious problem arising from both anthropogenic and geogenic (natural) processes [9]. Arsenic [10], nitrate [11], fluoride [12], and iron [13] are among well-known groundwater contaminants threatening millions of lives or well-being or otherwise impacting drinking water quality, especially in Asia [14,15]. Concerns over groundwater quality are reflected in the Sustainable Development Goals (SDGs), notably SDG 6, which includes specific targets with respect to serious inorganic contaminants, notably arsenic [16].
Groundwater quality (GWQ) mapping is a fundamental step toward sustainable management of groundwater resources by identifying spatial variations in contamination levels across the study area [17,18]. A GWQ map—as a general term—can be obtained by modeling the probabilistic distribution of groundwater contaminants with respect to environmental and geochemical conditions, and calibrated by alignment with actual (measured) concentrations of contaminants.
Whilst different terminologies, such as groundwater hazard mapping (GHM), groundwater vulnerability mapping (GVM), and groundwater risk mapping (GRM), have been used to address the process that results in a GWQ map, it is acknowledged that these key terms differ in meaningful ways, owing to their underlying conceptual distinctions. Hazard is viewed as a threat, a pre-existing condition that can turn into a risk depending on the effect of endogenous and exogenous factors, whereas vulnerability is perceived as the interactive effects of the social and physical aspects of a system regarding susceptibility to the hazard increasing in the future, and risk is calculated as a function of hazard, exposure, and receptor-vulnerability [19]. Notwithstanding the widespread use of GVM in the literature [20] and acknowledging its relevance (see Text S1), GVM mostly relies on aquifer properties alone to predict future hazard; hence, this study adopts the term GHM in light of the probabilistic essence of the modeling based on the actual current presence of contamination. Henceforth, GHM is used to refer here more broadly to studies that have conducted groundwater hazard/vulnerability/risk/susceptibility mapping.

1.2. Geospatial Analysis and Models

Geospatial analysis—described as a combination of analytical models and spatial software applied to geographic data—is a useful approach for understanding spatial patterns (i.e., distributions) and discovering relationships between various parameters [21]. In all GHM efforts, the use of geospatial analysis is essential [22,23] to account for spatial variations of groundwater contaminants [15,24]. With the help of remote sensing and tools such as geographic information systems (GISs), it is viable to readily acquire, analyze, and visualize environmental and geochemical data for GHM. Exploring the spatial relationship between groundwater hazard and influencing factors enables experts to create better hazard maps and identify critical areas for remedial action or mitigation. Various methods including statistical techniques (e.g., Dempster–Shafer [25], frequency ratio [26], weights of evidence [27]), decision-making techniques (e.g., technique for order of preference by similarity (TOPSIS) [28] and analytic hierarchy process (AHP) [29]), and artificial intelligence (AI) tools [30] have been suggested for this purpose. This study focuses on machine learning (ML), as a branch of AI tools that has been widely used in the geospatial analysis of GHM. This context is hereafter referred to as MLGHM.

1.3. Machine Learning

ML models have attracted tremendous attention for analysing and predicting high-dimensional parameters, in part due to advantages such as handling non-linear and automatic computation compared to conventional statistical techniques. These models are capable of learning the dependencies between inputs and output(s) parameters to perform regression and classification tasks [31]. In the first phase, the models explore ‘training data’ to learn the pattern, and, in the next phase, the quality of their knowledge is tested using the second set of data called ‘testing data’.
A general classification of ML models is shown in Figure 1 [32]. There are three major categories of ML models, viz. supervised learning, unsupervised learning, and reinforcement learning. The first two categories are employed for the common prediction and classification tasks. In supervised learning, the inputs are paired with known targets (i.e., values/classification labels), and in contrast, unsupervised learning represents a training wherein the inputs are not associated with particular targets [33]. In Figure 1, another category is shown called semi-supervised, which is defined as a learning type in which some of the data is labeled and the rest is unlabeled [31]. Further explanations regarding the subsets of these groups are provided in the rest of the article.

1.4. Scope and Objectives

Considering the importance and popularity of ML models for groundwater quality modeling, it is important to update the current knowledge about the latest developments in this field [34]. While there are many published reviews of the application of ML models in general and more particularly to surface water quality modeling [35,36,37], fewer studies have been specifically dedicated to groundwater (see Table S1). Whilst these reviews provide useful insights to MLGHM researchers, there are a number of gaps in coverage, which indicates the potential value of a further and differently focused review. In detail,
(i)
Most previous evaluations of models have been largely focused on model accuracy without substantive further consideration of other model evaluation criteria, notably predictability, complexity and relevant penalization, and computational efficiency [38,39,40,41].
(ii)
Some recent reviews considered only a small number of studies [38,40,42], or were restricted to just a specific contaminant (e.g., nitrate [43]) or a specific ML (e.g., artificial neural network (ANN) [44]).
(iii)
There is some inconsistency in the use of technical terms in relation to ML [38,41],
Based on these identified gaps and limitations, the key aims of this study were as follows:
  • Provide a systematic and comprehensive review of published MLGHM articles meeting stated inclusion criteria.
  • Provide a descriptive analysis of MLGHM articles with respect to geographic distribution, target and predictive variables, and ML approaches.
  • Consider the criteria to be sensibly considered in selecting the best ML model for a particular purpose.
  • Ultimately, provide optimization-oriented suggestions to improve the effectiveness/efficiency of MLGHM.
Note that the List of Abbreviations can be found in Supplementary Information.

2. Materials and Methods

2.1. Survey and Criteria

The publication databases Google Scholar, Web of Science, and Scopus were searched for relevant publications written in English over the period of 2000–2024 using a combination of environmental, ML, and geospatial-related keywords as outlined in Table S2. These searches were supplemented by relevant publications from other sources (e.g., ScienceDirect, SpringerLink, Copernicus Publications, Nature Portfolio, Science, and Wiley Online Library), as well as a number of previous review articles.
Regarding the inclusion/exclusion criteria, as illustrated in Figure 2:
(i)
Only original research articles that were published in Scopus-Indexed journals (December 2024 list [45]) were considered.
(ii)
With some notable exceptions, only those articles in which ML analysis led to generalizing the prediction to unsampled subareas (i.e., GHM) were included. A few articles that did not satisfy this criterion were retained as they included a clear representation of the contaminant distribution in addition to presenting useful comparisons among controversial ML models (e.g., Ref. [46]). Hence, studies that were bound to merely numerical groundwater quality analysis were not included.
(iii)
Articles in which there were evidence inconsistencies in stated target or predictor variables were excluded (Refs. [47,48]).
(iv)
Articles in which it was not immediately clear whether the used models were ML-based (e.g., Ref. [49]) were excluded, as were articles that, notwithstanding that their models were categorized as unsupervised techniques, nevertheless described techniques that are widely considered not to be unsupervised ML. Such techniques included entropy model averaging (EMA) in Ref. [50], fuzzy-catastrophe framework (FCF) in Ref. [51], fuzzy membership framework (FMF) in Ref. [30], as well as the GA and Wilcoxon test in Ref. [52]. Note that the use of metaheuristic techniques such as the genetic algorithm (GA) for optimizing predefined frameworks (e.g., DRASTIC) is considered a popular, yet non-ML, approach for GHM. Examples of these studies will be discussed later.

2.2. Meta-Data

A total of 284 articles were found that satisfied the inclusion criteria. The following information was collected from each paper: country of study area, year of publication, ML model(s) utilized, modeled target parameter(s), predictive variables, feature selection techniques (if used), and optimization techniques/algorithms (if used). Further considerations regarding the study area of some articles can be found in Text S2.

3. Trend Analysis

An objective of this study was to describe temporal trends and the geographic distribution of MLGHM articles, and the considered target and predictive variables. The results and relevant discussions are presented in Supplementary Material (Section S1). The key takeaways were as follows:
(i)
There has been a consistent rise in the rate of MLGHM publications over the period of 2005–2024.
(ii)
Iran, India, and the United States are the countries with the highest number of MLGHM case studies.
(iii)
N O 3 was the most studied target variable.
(iv)
Chemical characteristics of water and, to a lesser extent, textural and chemical characteristics of overlying soils were the most commonly used predictive variables.
(v)
Tree-based models (mostly RF) were the most favored feature selection technique.

4. Machine Learning—Review of Model Types

A classification of the reviewed MLGHM models is provided here, together with a summary of the frequency of use of each category/single model. A total of 676 (not unique) ML models were used in the surveyed articles, out of which 638 were supervised and 38 were unsupervised. The only application of semi-supervised models appeared in Ref. [53] for identifying the groundwater contaminant source. However, it was excluded from this survey due to not meeting the inclusion criteria.
Classification of ML models into further sub-groups was carried out based on pivotal sources, e.g., Refs. [31,32] and earlier similar studies, supported by expert knowledge. The results are shown in Figure 3, with classification details in Table S7 and relevant articles cited in Table A1. Note that (i) for brevity, acronyms represent models’ names here and full names can be found in the Nomenclature and (ii) for consistency, exemplary ‘Bst (95 times)’ is used to indicate 95 appearances of the Bst model/category in the reviewed articles, preferred over ‘Bst (95 studies)’ as many studies contained more than one model belonging to the Bst group/category.

4.1. Supervised Models

Supervised learning addresses the strategy for discovering and predicting patterns when the model is provided with both predictive and target variables [54]. This ML type consists of the most popular models that follow various linear and non-linear approaches to map the input–target relationship. A notable advantage of supervised learning is that, when sufficient labeled data is provided, the model can be satisfactorily trained and tested to achieve a reliable solution. Moreover, these algorithms are automatic, eliminating the need for the user to perform decision-making and tune too many parameters. Such advantages have driven scientists to choose supervised learning algorithms in 94% of MLGHM cases (Figure 3).

4.1.1. Artificial Neural Networks

ANNs are famous members of the ML family that follow a biological approach (human neural system) for pattern analysis. Reportedly, the idea of ANNs was first presented in Ref. [55]. In ANNs, computational units are neurons that, despite lying in different layers, are highly connected by weights (i.e., synapses) throughout the network [56]. This feature, as well as taking advantage of powerful training strategies (e.g., backpropagation [57]), training algorithms (e.g., Levenberg–Marquardt [58]), and non-linear activation functions, enables ANNs to establish an accurate mathematical relationship between the predictive and target variables. Many studies have applied ANNs to model environmental parameters such as water quality [59] and groundwater level [60].
The earliest applications of ANNs in GHM were presented by Almasri and Kaluarachchi [61] in 2005 and Wang, et al. [62] in 2006; however, scientists are still employing ANNs for this purpose. While the reviewed articles used different terminologies to refer to conventional ANNs, the models’ structures indicated that they are mostly MLP, which is known as a universal approximator [56]. The whole group of ANNs forms 14% (=88 times) of supervised MLGHM models comprising conventional ANNs (70 times), DL (11 times), and Other (7 times), which itself comprises MNN (2 times), ELM (2 times), GMDH (1 time), ELNN (1 time), and BRNN (1 time).

4.1.2. Neuro-Fuzzy Models

A neuro-fuzzy (NF) model represents a combination of two paradigms, namely ANN and fuzzy logic (FL). It results in a strong predictive tool that is able to capture non-linear input–target relationships [63]. Introduced by Zadeh [64], FL is known as a mathematical framework that deals with reasoning and decision-making and addresses uncertainty and vagueness. A strength of FL is that it uses membership functions (MFs) such as triangular, trapezoidal, and Gaussian, allowing an element to have a membership degree (e.g., 0–1) instead of belonging to binary classes (0 or 1). ANFIS [65] is the most popular NF model that is composed of five layers that are responsible for the calculations as follows: (1) Fuzzification Layer: calculating the input MFs, (2) Rule Layer: calculating the firing strength, (3) Normalization Layer: normalizing firing strength, (4) Defuzzification Layer: computing weighted outputs, and (5) Summation Layer: computing the final output [66]. It is worth explaining that the difference between ANFIS and other NFs (e.g., CANFIS, DENFIS, and HyFIS) lies in the improvements, including CANFIS using an MNN, HyFIS using an MLP with gradient descent learning, and DENFIS following continuous evolving after new data entry. Due to these merits, NF models have received significant attention in various environmental domains, including hydrology [67] and groundwater level [68].
Whilst sophisticated, NF is one of the pioneering models that was first used by Dixon [69] in 2005. NF models constitute 3% (=20 times) of supervised MLGHM models, comprising ANFIS (16 times), CANFIS (2 times), DENFIS (1 time), and HyFIS (1 time).

4.1.3. Tree-Based Models

Tree-based models form a large portion of the ML community. These models are built upon binary classification trees, in which the most significant predictive variable is positioned at the root, followed by less significant ones in the subsequent positions. Owing to advantages such as less computational cost [70] and high interpretability [71], these models have received tremendous attention in prediction efforts. RF is the most reputable tree-based model that was devised by Breiman [72] as a combination of randomized DTs. This model has successfully dealt with prediction tasks in water-related sectors [73].
The oldest representative of this group was RF, used by Rodriguez-Galiano, et al. [74] and Nolan, et al. [75], both in 2014. This group accounts for 25% (=160 times) of supervised MLGHM models comprising RF (118 times), DT (12 times), CART (10 times), ERT (8 times), M5 (4 times), and Others (8 times) that itself comprise BART (2 times), LMT (2 times), RotF (1 time), J48 (1 time), RT (1 time), and RndT (1 time).

4.1.4. Support Vector-Based Models

Proposed by Cortes [76], SVM is another competent group of ML that is built on statistical learning theory. The key concept in training SVMs amounts to defining decision boundaries by identifying reproducible hyperplanes that maximize the distance between support vectors of the class labels [77]. Strong function approximation and quick learning are strengths of SVMs that have made them popular LM models for various predictions, particularly in hydrology [78] and other environmental sectors [79,80].
Employed by Ammar, et al. [81] in 2008, RVM represented the first appearance of SVMs within the surveyed studies. This group forms 10% (=63 times) of supervised MLGHM models, consisting of SVM (49 times), SVR (12 times), and RVM (2 times).

4.1.5. Regression Models

Being among the most straightforward ML models, regression models are useful to delineate the relationship between predictive and target variables. This group possesses various linear and non-linear notions such as MLR and LgR [82] that have gained significant interest in environmental studies [83], and notably hydrological research [84,85].
The first MLGHM article that used a regression model (LgR) was published by Twarakavi and Kaluarachchi [86] in 2005. The regression family is the fourth popular ML group, accounting for 12% (=75 times) of supervised MLGHM models comprising LgR (37 times), MLR (16 times), GLMs (7 times), MARS (7 times), and Others (8 times) that itself comprise PolyR (2 times), BRR (2 times), glmnet (1 time), AFR (1 time), LsR (1), and RR (1 time).

4.1.6. Probabilistic Models

Probability theory is the essence of many ML-related rules. Depending on the used interpretation (i.e., frequentist or Bayesian [87]), different approaches can be regarded for pattern identification and uncertainty quantification in these models. In this regard, integration of simple probability distributions over (one or) a few variables forms a more complex model [88]. The ease of use and the high predictive potential of these models have made them popular tools for modeling purposes [89], including environmental predictions [90,91].
Probabilistic models have been mostly used in the last decade of the MLGHM literature, as the first application refers to using a BN model by Nolan, et al. [92] in 2015. This ML family accounts for 5% (=34 times) of supervised MLGHM models encompassing five classes of Bayesian (10 times), Gaussian (10 times), discriminant (10 times), Markovian (3 times), and Others (1 time). Among 15 unique probabilistic models, Gaussian GP (9 times) was the most popular, followed by Bayesian NB (6 times), discriminant MulDA (5 times), and discriminant LDA (3 times). Other models that appeared once include Bayesian models of BN, NBT, BMA, and BME, the Gaussian model of GCM, discriminant models of FDA and QDA, Markovian models of MMRF, CMMRF, and FMRF, as well as MaxEnt (Others).

4.1.7. Ensembles

Ensembles are another promising type of ML models that altogether form 27% (=171 times) of supervised MLGHM models. However, considering the strategies used for creating ensembles, this category is divided into two major parts:
Ensemble 1: The category is dedicated to ML models that incorporate either of bootstrap aggregating (bagging, Bag) [93], boosting (Bst) [94], disjoint aggregating (dagging, Dag) [95], and random subspace (RS) [96] techniques for performance improvement. This improvement is achieved by leveraging multiple base learners and aggregating their outputs. While Dag and Bag perform parallel training of the models toward reducing variance, Bst relies on a sequential approach that results in reducing bias. RS, on the other hand, employs techniques based on feature diversity to enhance decorrelation within base models. These models have extensively served for various purposes [97,98,99], including environmental predictions [100,101,102].
Although most Ensemble 1 models have appeared after 2020, the first application was a BRT model in Ref. [92] that was published in 2015. 19% (=121 times) of supervised MLGHM models belong to Ensemble 1 category, comprising Bst (95 times), Bag (17 times), Dag (7 times), and RS (2 times). Notably, Bst encompassed 12 different models, out of which, BRT (28 times), XGBoost (25 times), AdB (9 times), and LGBM (9 times) were the most favored ones.
Ensemble 2: Unlike Ensemble 1 models that rely on resampling techniques, the second group of ensembles represents models that are founded on synthesizing two (or more) algorithms for creating improved versions. Many scholars have successfully applied this type of ensemble to various water- and groundwater-related problems [103,104].
Considering the underlying strategy for combining models, Ensemble 2 consists of three major classes, here referred to as Groups 1, 2, and 3 (detailed below), all accounting for 8% (=50 times) of supervised MLGHM models comprising Group 1 (21 times), Group 2 (21 times), and Group 3 (8 times). The details of these groups are as follows:
(i)
Group 1: Meta-models (MMs) comprising committee and stacking learning that exhibit two-level predictions, i.e., the outputs of the models in Level 1 are considered inputs for another model in Level 2 (A.K.A. Levels 0 and 1). Fijani, et al. [105] presented the first application of such models (SCMAI) in 2013.
(ii)
Group 2: Metaheuristic optimization algorithm (MOA)-based models in which an MOA optimizes a baseline ML. MOAs are known as global optimizers that can tackle complicated problems [106,107]. These algorithms employ search operators to systematically update the solution in the search space through exploiting and exploring the knowledge. Such strategies ensure the final solution is globally optimum, i.e., protected from getting stuck in local minima [108]. The use of GA for optimizing CMAI by Fijani, Nadiri, Moghaddam, Tsai, and Dixon [105] in 2013 was the first appearance of MOAs in the surveyed literature.
(iii)
Group 3: Dual hybrids (DHs) are created by integrating two baseline ML models. These models usually deliver stronger solutions than single conventional models, due to benefiting from the computational advantages of both models simultaneously. The first representative of this group was Cubist—as a hybrid of DT and LR—used by Band, et al. [109] in 2020.

4.1.8. Others

This category accounts for 4% (=27 times) of supervised MLGHM models, consisting of instance-based learning (IBL, A.K.A. lazy learning) (18 times), evolutionary models (7 times), and physics-informed ML (PIML) (2 times), as follows:
IBL: Unlike other ML models that fit a model/function over training data and use it for prediction, IBL stores the training data and performs prediction based on similarities between the stored and new data. These models rely on distance measures such as Euclidean and Cosine distance to find stored samples that are most similar to the new data [110]. Proposed by Fix and Hodges [111], KNN is the only entity of this category that was used for MLGHM purposes. In KNN, the distance of a new point from all training data is calculated, then, a prediction takes place by focusing on the k nearest neighbour points (with k being user-defined). Earlier studies highlighted KNN’s applicability in various sectors [112], including environmental simulations [113]. Starting from 2019 in Refs. [114,115,116], KNN appeared in 18 MLGHM studies.
Evolutionary learning: Evolutionary models are inspired by the Darwinian natural selection principle, aiming to evolve the solution over generations. In these approaches, mutation, crossover, and selection processes are applied to the parent solutions to generate offspring of higher quality [117,118]. These techniques have generated immense interest in various ML studies [119,120], and more particularly, in environmental modeling [121]. GEP, which was designed by Ferreira [122], is the only algorithm of the evolutionary group found in the surveyed literature. Besides its reliable accuracy, a strength of GEP is that it can present the solution in the form of an efficient and explicit mathematical formula that is less complicated than an ANN-derived formula [123]. Nadiri, et al. [124] first used GEP for GHM in 2017. Following this, six other studies used this model, resulting in a total of 7 times of use.
PIML: The reviewed literature reflected one application of a model named U-FNOB (and its extension, U-FNOB-R) that works on the basis of the Fourier neural operator (FNO). These models are among the rarest ML models; hence, their technical details can be found in the relevant article, i.e., Ref. [125].

4.2. Unsupervised Models

Unsupervised learning addresses models that are used for finding the latent patterns/clusters within unlabeled data [126]. In other words, the model is not provided with target (or response) values attributed to the predictive variables. It rather performs clustering on data based on commonalities. It is considered an advantage of these models, enabling them to be used in semi-supervised scenarios where there is a restricted number of labeled data [127]. Unsupervised ML models have served for various predictive purposes, especially in environmental and water-related fields [128,129,130]. Unsupervised category covers 6% (=38 times) of MLGHM models (Figure 3).

4.2.1. Clustering Models

Identified clustering models comprise FCM, KMC, KMdC, HCA, and APC. Whilst all follow the clustering approach for data analysis, there are differences between them that result in pros and cons for each technique. Due to the use of FL in FCM, it can perform soft clustering, in which data points are assigned membership degrees for different clusters rather than pertaining to one. KMC has a low computational cost, while KMdC is robust to outliers due to considering medoid points (rather than centroids) as the centre of the cluster. HCA uses a tree-structured clustering, making it suitable for visualizing hierarchical relationships. APC has the advantage of automatic determination of the number of clusters.
By using a KMC by Das, et al. [131] in 2010, clustering models started gaining attention for GHM, and include 66% (25 times) of unsupervised MLGHM models comprising KMC + KMdC (12 times), HCA (7 times), FCM (5 times), and APC (1 time).

4.2.2. Self-Organizing Map

Proposed by Kohonen [132], SOM is a popular unsupervised ML that has been widely used for environmental modeling purposes. This model has an ANN-like skeleton that enables it to map high-dimensional data onto a simpler grid (mostly 2D) whilst maintaining the topological relationships [133]. The earliest appearance of SOM in MLGHM literature corresponds to Ref. [134] that was published in 2014. SOM-based accounts for 29% (11 times) of unsupervised MLGHM models, comprising single SOM (8 times), SOM combined with KMC and FCM (2 times), and GNG (1 time).

4.2.3. Unsupervised Deep Learning

Contributing to 5% (2 times) of unsupervised MLGHM models, Faal, et al. [135] used two DL models, namely DBN and DSA, in an unsupervised manner. Technical details regarding these models can be found in the cited article.

5. What Defines a Better Model?

A key aim of this review was to assess the comparative advantages and disadvantages of various MLGHM models with a view to providing guidance on optimal model selection for different users. Model selection ideally requires an understanding of both (i) what are important model selection criteria and (ii) how these various criteria should be weighted—in other words, what defines a ‘better’ model? Clearly, the weighting of various criteria may be quite subjective and may vary from problem to problem. Further, as Fisher, et al. [136] stated “All models are wrong, but many are useful”. With these limitations in mind, this section addresses several characteristics of a ‘good’ model, based on which the surveyed MLGHM models will be assessed in Section 6. Note that most of these criteria are described briefly here with further discussion in Section 8. In addition to the criteria for assessing the selection of ML models, we also consider various aspects of ML model design.

5.1. Model Design and Data Handling

(i)
Role of assumptions: Many data-driven models are able to satisfactorily learn and reproduce intrinsic non-linear patterns. There is overwhelming evidence of the suitability of ML models for exploring the relationship between GWQ and its influencing factors. But an overlooked question regarding many ML-based efforts is ’To what extent can the performance of an ML be influenced by the modeling assumptions?’
(ii)
Feature selection: As shown in Section S1.5, many MLGHM studies have benefited from feature selection to optimize the models in terms of dimensionality. These techniques can help ML models by improving learning performance, preventing overfitting, and/or reducing computational costs [137].
(iii)
Dataset adjustment: The data splitting ratio, which specifies what portions of the dataset are dedicated to the training and testing phases [39], is an important factor that can impact the quality of training, and consequently, the quality of the prediction results. In addition, data-centric cross-validation techniques (e.g., K-Fold) [138] can be regarded as addressing the stability of the model.

5.2. Model’s Performance and Reliability

(iv)
Accuracy and validation: Accuracy assessment is one of the most important steps of any modeling. In MLGHM efforts, scientists have extensively used accuracy metrics (e.g., RMSE, AUC, and R2) to validate and compare the performance of their models.
(v)
Hyperparameters (tuning and presentation): Recognizing and proper determination of models’ hyperparameters is vital as they can impact the quality of the prediction. ML models, therefore, should ideally be hyperparameter-tuned to maximize the model’s potential and to be hyperparameter-transparent to support reproducible research.
(vi)
Stability: Stability of an ML model can be defined as its ability to maintain consistent performance when exposed to small perturbations. In this sense, recognizing the sources of uncertainty, as well as assessing the models’ repeatability and ensuring the results’ reproducibility, can be helpful.
(vii)
Generalizability (transferability): overfitting can be a general threat even to models that exhibit high accuracy and can result in models not being generalizable or transferable to new (unseen) data [139,140], due to the learnt pattern being excessively specific to the training data [141].
(viii)
Explainability (transparency/interpretability): Many of the reviewed ML models lack explainability, i.e., a clear presentation of the internal processes, and so are labeled here as ‘opaque-box’ models. In contrast, explainable artificial intelligence (XAI) is an increasingly used term to describe more explainable ML models that enable human users to better understand and so better trust [142].
(ix)
Optimizability: optimizing an ML model can be conducted with respect to data and dimension, the model’s architecture, hyperparameters, and the training algorithms, and can lead to a more applicable model.

5.3. Computational Efficiency

(x)
Computation time: Particularly for models of large datasets, reducing the computation time to attain a target accuracy may be an important model characteristic [143] necessitating a computationally efficient model. This factor should be evaluated with respect to the operating system’s characteristics (e.g., RAM and CPU frequency) to reflect the models’ computational demand fairly.
(xi)
Penalizing for complexity: Penalty functions aim to address a trade-off between the model’s fitness and its complexity. To this end, the model is penalized (sometimes partly) with respect to the number of its computational parameters (e.g., the number of intercepts, coefficients, and error variance in MLR).

5.4. Practical Application

(xii)
Informed hazard zonation (IHZ): Once a GHM is produced, it is essential to correctly classify the study area into different hazard classes for optimizing decision-making and cost estimations. Popular methods for doing so are based on statistical methods (e.g., Equal Intervals) and/or regulatory standards; however, more recent studies explicitly consider model accuracy and purpose-related model-informed cost-benefit balances.

6. Machine Learning Models—Comparative Assessment

This section provides an extensive comparison of the reviewed MLGHM models utilising ‘better model’ metrics beyond just model accuracy [38,39,40,41]. Due to the shortage of adequate evidence for the relative assessment of unsupervised models, this section focuses on supervised MLs and, in particular, considers those reviewed MLs from articles that were more comprehensive and informative
Table 1 summarizes the information extracted from 46 selected articles. While all studies performed reliable accuracy assessment using several popular metrics (e.g., RMSE, MAE, AUC, R2) that mostly resulted in explicitly distinguishing one model as superior, not all of them provided an explicit ranking of the used models. For conciseness, both in Table 1 and throughout the text, “>” and “<“ are used to indicate that the first model is more accurate or less accurate, respectively, than the second model. Most of the reviewed papers provided a firm scientific background and assumptions to conduct MLGHM. Assessment of these assumptions is discussed in Section 8.
The studies are divided into five categories labeled as follows:
  • Label A focuses on comparing mostly conventional models, including (but not limited to) ANNs, NFs, SVMs, and simple tree-based models,
  • Label B focuses on the competition between tree-based and Ensemble 1 models,
  • Label C addresses the competency of DL, asserting itself as a competitive model to the above two groups,
  • Label D focuses on MM ensembles (Ensemble 2—Group 1),
  • Label E evaluates MOA-based ensembles (Ensemble 2—Group 2) versus relevant baseline models.
Some further features of MLs are outlined below:
(i)
Data Adjustment: Many studies considered feature selection (and/or importance assessment) using popular models (e.g., tree-based MLs, PDP, and PCA) to delineate the role of predictive variables. As an interesting idea, some studies (e.g., Refs [145,180]) created and imposed various combinations of predictive variables that led to identifying the most accurate prediction scenarios. Many works also took advantage of data-centric cross-validation (mostly the K-fold technique) as part of data adjustment. Nevertheless, a few studies (e.g., Refs. [46,116]) paid attention to optimizing the data splitting ratio that reportedly can be an accuracy-determinant factor.
(ii)
Hyperparameters: Reporting the used hyperparameters was inconsistent, i.e., present in many articles but missing in many others. Most studies, however, benefit from automatic methods such as the GS algorithm and manual methods such as T&E to ensure proper assignment of hyperparameters.
(iii)
Computational efficiency: Some studies considered documenting computational details (e.g., computation time [135,155,171,182,183] and the used operating system [171,176,180]). These two factors can reflect important information regarding an informed selection of efficient MLs, due to the complexities of some MLs and occasional limitations in computer facilities. Moreover, utilizing penalty functions to constrain the model’s complexity is another promising direction for proper selection of the model (and predictive variables). However, it was observed in a small portion of the surveyed articles (e.g., Refs. [152,158,163,182]). To exemplify, in Refs. [135,171], the outstanding model not only achieved the highest accuracy, but also was quicker in terms of training time (and convergence of the solution [183]). Further, Ref. [158] indicated that its superior model achieved these merits along with the lowest penalty expressed by AIC and BIC.
(iv)
Others: Many studies leveraged internal and external techniques (e.g., XGBoost’s inherent regularisation [159] and Dropout for DL [171]) to prevent overfitting. In contrast, quantitative evaluation of prediction uncertainty was explored in a few studies (e.g., using statistical indices of PICP [156] and U95% [180]). Models’ explainability, generalizability, and results’ reproducibility likewise received limited attention, as they were often discussed in brief. Last but not least, IHZ remained almost underexplored despite its great importance. The vast majority of the reviewed studies employed well-established methods (e.g., natural breaks, equal hazard interval, and WHO standards) to classify the studied aquifers in terms of hazard. Only a few studies (e.g., Ref. [157] from Table 1 and others [185,186,187]) identified AUC-based cutoff optimization for determining hazardous areas.

6.1. Accuracy Comparison

6.1.1. Label A

This category provides evidence of the high ability of various conventional MLs that outperformed each other and even ensemble models.
ANNs: Various ANNs, including MLP [144], DL [170], and ELM [146], were found to be more powerful than other MLs. The MLP used in Ref. [144] also exhibited lower uncertainty than SVM, MLR, RT, GP, and BRT.
NFs: ANFIS surpassed (non-ML) FLs, including Mamdani and Sugeno [174,177], reflecting the positive contribution of the neural system that is embedded in this model (See Section 4.1.2). In Refs. [175,188], ANFIS stood in a close competition with ANN and SVM, whereas it was outperformed by them in Refs. [124,189,190]. Modified versions of ANFIS, however, mostly provided better accuracy than different MLs. In this regard, CANFIS [191] and DENFIS [145] were more accurate than SOM, ANN, GEP, GMDH, MARS, and M5, while HyFIS [147] showed weaker results than RF and SVM.
Tree-based models: These models—and on top of them, RF—came up with higher accuracy than a wide variety of MLs in many comparative studies. In Ref. [116], RF achieved the most balanced set of accuracy metrics when it was in competition with LgR, ANN, SVM, DT, KNN, LDA, NB, AdB, QDA, GBC, and GP. Similarly, Ref. [192] revealed the greater accuracy of RF over SVM, XGBoost, KNN, ANN, and GCM. Agreeable outcomes can be found in other studies, such as Refs. [148,150,165,193]. As for other tree-based models, several efforts demonstrated the superior accuracy of ERT [151], BART [194], and CART [195] versus diverse ML competitors.
SVMs: Despite possessing several advantages, viz., high speed, reliability, and cost-effectiveness for GHM [196], SVMs appeared frequently as benchmark models. SVMs showed promising performance, providing higher accuracy than popular MLs, viz., MulDA, BRT, and GLM [197], RF and BRT [154], and ANFIS [190]. SVR is a relatively more powerful member of this family that was found to be more accurate than MLs, including RF and SVM [198], RF [199], KNN [200], MARS and GLM [150], and ANN [201].
Regression models: Conventional versions of regression models have rarely been dominant in the MLGHM literature. MARS once appeared to be the most accurate model, outperforming SVM, MDA, GLM, and GBM in Ref. [202]. Also, Ref. [150] gives an example of its higher accuracy compared to GLM. RR is a potent regression model that, in Ref. [152], ranked higher than RF, XGBoost, DT, and KNN accuracy-wise. LgR that was used in Ref. [116]—yielding a comparable performance with RF—was more accurate than SVM, GP, and ANN. Moreover, GLM exhibited higher accuracy compared to FDA in Ref. [203].
Probabilistic models: Multiple advantageous applications of probabilistic models have put them among successful MLs in the surveyed literature. Researchers suggested GHM uncertainty evaluative techniques through Bayesian frameworks of BMA [204] and BME [205]. In Ref. [204], this idea resulted in the superior accuracy of BMA versus ANN, GEP, and SVM. GP is another powerful probabilistic model that showed higher accuracy compared to RF and ELNN [206], as well as (non-ML) kriging models [207]. Another notable application of probabilistic models was carried out in Ref. [149] using CMMRF that surpassed RF in terms of prediction accuracy.
Others: While KNN was mostly used as a comparative benchmark for more sophisticated models [114,208], there are a few studies quantifying the relative merits of this model. For instance, KNN’s accuracy was found to be higher than that of LDA and QDA in Ref. [116] and higher than ANN and GP in Ref. [192]. Regarding GEP, its accuracy-wise excellence was identified in Ref. [145], surpassing that of M5Tree, MARS, GMDH, and DENFIS, and in Ref. [155], outperforming M5 and two MOA-tuned models of GA-PolyR and PSO-MARS.

6.1.2. Label B

The use of resampling techniques, including Bag, Bst, Dag, and RS, has resulted in significant improvements in many MLs. For instance, Ref. [166] showed that the baseline models of LMT and NBT experience higher accuracy once they were incorporated with Bag, Dag, and RS techniques. Similar improvement was attained for RF, XGBoost, LGBM, AdB, and CatBoost by incorporating Bag and Dag in Ref. [161]. Moreover, a subset of literature highlights the greater accuracy of these techniques against conventional MLs, viz., Bag, Bst > RF [164,167], Bag, Bst > RF, SVM, ANN [165], Bag > ANN, KNN [151].
Ensemble 1 models appeared to be more powerful than several MLs, viz., BRT > SVM, MARS, GLM [150], XGBoost > DL, MLR [169], GBDT > SVM, ANN [165]. As far as the comparison among Bst models is concerned, the accuracy ranking varies across the literature. For example,
(i)
XGBoost > LGBM, AdB, CatBoost in Ref. [161],
(ii)
XGBoost > GBM, LGBM in Ref. [159],
(iii)
CatBoost > XGBoost, LGBM in Ref. [158],
(iv)
CatBoost > AdB in Ref. [162],
(v)
AdB > XGBoost, LGBM in Ref. [178],
(vi)
GBR > XGBoost, AdB in Ref. [156].
In addition, there is contradictory evidence suggesting both superior and inferior results of Bst models against RF. More precisely,
(i)
RF < XGBoost [158,159], BRT [209], GBR/GBM [156,159]), AdB [162], CatBoost [158,162],
(ii)
RF > XGBoost [152,153,156,157,160], BRT [150,163], GBC [116], AdB [116,156], LGBM, CatBoost [161].
From these comparisons, it can be said that Bst models show high sensitivity to the problem conditions. It, therefore, calls for assessing various Bst models—along with RF—to disclose a reliable problem-dependent solution. As a useful external (non-GHM) study, Ref. [210] sheds light on the relative performance of these models.
The potential of the discussed resampling techniques (i.e., Bag, Bst, Dag, and RS) has been comparatively evaluated in several studies as follows:
(i)
Bag > Dag > RS (applied to LMT and NBT) in Ref. [166],
(ii)
Bag > Dag (applied to RF, XGBoost, LGBM, AdB, and CatBoost) in Ref. [161],
(iii)
Bst > Bag in Refs. [164,167],
(iv)
Bag > Bst [165].
It is worth highlighting that in Ref. [161], the simultaneous inclusion of Bag and Bst techniques (in the Bag-XGBoost model) resulted in an outstanding accuracy. Although more investigations may be required to be conclusive, this paragraph suggests that Bag and Bst presented stronger predictions than the other two. Further insights regarding the comparative potential of Bag and Bst can be found in external (non-GHM) studies such as Ref. [211].

6.1.3. Label C

While DL’s performance was found to be more promising than various conventional MLs (e.g., SVM [212], DT, and NB [170]), there are still some ambiguities regarding its conclusive robustness against powerful models, including shallow ANN, RF, and ensembles.
Refs. [46,168,212] provided evidence to support a considerably higher accuracy of DL than conventional ANN. In contrast, Ref. [213] proved otherwise, i.e., (GP and) ANN outperforming DL. Moreover, Ref. [168] reported that while the used DL achieves higher accuracy, its stability—expressed as the coefficient of variation of accuracy metrics—is significantly lower than ANN. Hence, the choice between deep and shallow ANNs for MLGHM may remain a challenge that calls for more comparative efforts.
Concerning the performance of DL vs. RF, Refs. [46,168,170] reported the higher accuracy of DL, whereas Ref. [171] showed RF performs more accurately than different DLs (unimodal 2D and 3D versions). Therefore, it can be said that while DL has emerged to be a serious opponent to RF, its superiority is still ambiguous and requires further assessment.
In the surveyed literature, XGBoost and GBM were the only ensemble models that were assessed against DL. In Refs. [46,170], DL exhibited higher accuracy than XGBoost and GBM, while it was outperformed by XGBoost in Ref. [169]. Hence, similar to RF and ANNs, the evaluation of DL against ensemble models requires more evidence. It should be additionally noted that ensemble models were represented by only two Bst models in this comparison. Thus, ideally, they should be accompanied by other ensembles (other Ensemble 1 members and Ensemble 2 MLs) toward reaching reliable concluding remarks.

6.1.4. Label D

MMs that can be built upon both conventional MLs (e.g., KNN [172], NB [179], and ANN [175]) and sophisticated MLs (e.g., AdB [178]), offered improvements in GHM by outperforming several powerful models, as listed below:
(i)
KNN-SEL > GBDT, XGBoost, RF in Ref. [172],
(ii)
LDA-SEL > RF, ERT, DT, XGBoost, AdB, GBDT in Ref. [173],
(iii)
AdB-Ens > AdB, XGBoost, LGBM, RF in Ref. [178],
(iv)
GP-Ens, SVR-Ens > ANN-Ens, supervised and unsupervised DLs in Ref. [135],
(v)
NB-SEL > RF, AdB, XGBoost, LGBM, CatBoost, GBDT, BA-DT in Ref. [179].
Different from the above studies, two ensemble approaches were created by averaging the predictions of several MLs at pixel-scale (normal mean in Ref. [116] and AUC-weighted mean in Ref. [203]). While the ensemble used in Ref. [203] demonstrated superior accuracy compared to FDA, SVM, and GLM, the one in Ref. [116] was considered comparable to LgR, RFC, QDA, ANN, and SVM. Moreover, SCFL—in the position of an ANN-Ens—was used thrice in Refs. [214,215,216]. Notwithstanding this model achieving high-quality results, these three studies only used (non-ML) FL techniques for comparison. Hence, SCFL’s competency relative to other MLs remains unassessed.
Altogether, this section suggests that utilizing ML-based MMs not only provides reliable solutions for GHM but also can result in improving the accuracy of reputable MLs (e.g., Ensemble 1, RF, and DLs).

6.1.5. Label E

Several studies proved the proficiency of MOA-based ensembles mostly in comparison with the corresponding original (standalone) MLs. For instance, these comparisons
(i)
ANFIS-PSO, ANFIS-DE, ANFIS-GA > ANFIS in Ref. [181],
(ii)
ANFIS-AO, ANFIS-SMA, ANFIS-ACO > LsR, ANFIS in Ref. [180],
(iii)
RF-FA, RF-MOE, RF-Ant, RF-GWO, RF-PSO > RF in Ref. [183],
(iv)
RF-GOA, RF-GWO, RF-PSO > RF in Ref. [184],
(v)
ANN-PSO, ANN-GWO > ANN in Ref. [217],
(vi)
PSO-ANN > SVM, MLR in Ref. [182].
This suggests that synthesizing appropriate MOAs with conventional MLs such as ANN, RF, and ANFIS can result in higher accuracy. However, this section gives two important reflections regarding the comparative proficiency of MOA-based models:
(i)
The effectiveness of MOAs varies across studies (e.g., PSO being strongest in Refs. [181,217] and weakest Refs. [183,184]),
(ii)
Since each study considers only one baseline ML (ANN/ANFIS/RF), cross-category comparisons are required to assess the power of a given MOA in optimizing different baseline MLs in the same study, e.g., ANN-PSO vs. ANFIS-PSO vs. RF-PSO.

6.2. Overall Comparison

The above evaluations demonstrate that accuracy-based ranking of MLs (except for MMs) can vary from one study to another. This arises because model accuracy may be a material function of a multitude of factors (e.g., scale of modeling, data representativeness, computational artefacts, and algorithm-specific settings coupled with the operator’s influence) [41].
Further, even for well-known MLGHMs such as RF and Bst models, their computational merits have not always protected them against issues such as overfitting. In this regard, RF and GBDT that were used in Ref. [165] attained strong training and presented comparably poor test results. Supporting this, an external (non-GHM) multi-dataset evaluation of Bst ensembles in Ref. [211] showed that their performance can largely depend on the imposed datasets. Likewise, DL showed a weak performance in Ref. [213] that was attributed to the lack of large-scale balanced data with a fixed time series.
Further distinctions between MLs are considered below:
Performance and reliability: It was observed that most conventional MLs can be outperformed by more sophisticated ones. Based on Table 1, the incorporation of in-built data optimization and regularization strategies for harnessing overfitting can be two major reasons for the better performance of the latter group. RF can handle prediction tasks even when the predictive variables outnumber the samples in a dataset [218]. Besides, this model has a global reputation for in-built feature selection [219] and relative robustness to outliers and noise [72,220]. DLs, on the other hand, can benefit from the dropout strategy to regularize the training and avoid overfitting [221]. Similarly, there is an inherent use of L1 and/or L2 regularization techniques in ensembles such as LightGBM and XGBoost. Stacking and MOA-based ensembles also take advantage of multiple algorithms to protect MLs against undesirable computational phenomena.
All in all, whilst barriers such as insufficient concrete evidence and existing counterexamples precluded this study from introducing specific consistently top-performing MLs, it was witnessed that RF, DLs, and most ensembles tend to provide more accurate results than many standalone MLs, including (but not limited to) SVMs, regression models, probabilistic models, KNN, and shallow ANNs.
Complexity: The complexity of MLs is usually addressed through several criteria encompassing structure, the number of hyperparameters, and explainability, among others. Since these (currently) examinable factors are interrelated, this section discusses them together. Relying on explainability [222], it varies across various groups of MLs. Simple models, such as MLR, LgR, DT, and NB, are often listed as intrinsically explainable, while more complicated ones, including ANNs, RF, SVMs, and ensembles, are referred to as opaque-box. Not surprisingly, these opaque-box models often possess more complicated structures and a larger number of hyperparameters compared to explainable ones [223]. Also, it must be acknowledged that the degree of complexity may vary among the models that pertain to the same group. Take DLs and RF, both being opaque-box, DLs are among the most complex MLs [222], whose performance largely depends on hyperparameters; RF, on the other hand, has a few key hyperparameters leading to a lower dependency on hyperparameters [224]. As for ensembles, it can be different for each group. Whilst the complexity of MM and DH ensembles is contingent upon the applied MLs, the need for hyperparameter tuning is well-established for many MOAs [225] and Bst models [210].
The key takeaway from this section is that most promising MLs are characterized by high degrees of complexity. Referring to earlier accuracy assessments, these more complex models tend to offer a higher accuracy than simpler MLs. However, the use of RF can be considered when a favorable complexity-accuracy trade-off is required.
Computation time: As with the above factors, the computation time of a model can vary based on the problem conditions and the model’s complexity. Albeit, based on the MLGHM articles that included the model’s training time, as well as a number of external studies (e.g., Refs. [226,227,228,229]), a tentative (and prone-to-exceptions) classification of well-known MLs in terms of training time can be expressed as:
(i)
Very Fast to Fast: MLR, LgR, DT, ERT, KNN, and LDA.
(ii)
Moderate to Slow: RF, NFs, shallow ANNs, SVMs, Ensemble 1, and M5.
(iii)
Slow to Very Slow: GP, DLs, and MOA-based models.
In Ref. [210], RF was found to be slower than two ensembles of LightGBM and XGBoost, and in Ref. [171], it appeared to be over 100 times faster than a 3D DL.
In a nutshell, this section outlined several strong MLs (e.g., RF, NFs, shallow ANNs, SVMs, Ensemble 1 models) with a relatively reasonable training time that can be considered more efficient compared to heavy models such as DL and MOA. The advent of fast graphics processing units (GPUs) [230], however, indicates the precedence of accuracy over computational efficiency in assessing the relative merits of MLGHM models.

7. Informed Model Selection

This section provides a framework for selecting better MLGHM models from diverse options. Figure 4 illustrates the suggested flowchart that considers several criteria, including relative accuracy, scale of interest, and complexity (as a function of transparency, computation speed, and the number of hyperparameters, with respect to the operators’ expertise). The framework is based on the general characteristics of models, but can be prone to exceptions.
In Figure 4, the models with less complexity can be labeled as transparent-box and are generally faster. Amongst those with superior accuracy, LgR and KNN have been more used for large-scale modeling, whereas DT may offer greater utility for small-scale modeling (see Table S3 and Figure 3). Concerning MLs with moderate complexity and speed, RF has been the most popular model in both global- and local-scale predictions. In this regard, Figure 3 shows the extensive use of RF for many small-scale modelings, whereas Table S3 shows several applications of this model for supernational scales, e.g., RF by Araya, et al. [231] for salinity prediction in the Horn of Africa and by Podgorski and Berg [7] for global arsenic modeling. Other promising models, including shallow ANN, NF, SVM, and ERT, can be regarded mainly for small case studies. As for the third group, i.e., opaque-box and slower MLs, scientists have used mostly Bst (GBMs) and MM (stacking) for broad GHM, while hybrids, DLs, and MOA-based models can be good candidates for local investigations (see Table S3 and Figure 3). Overall, promising models of the first group (transparent-box and fast) can attract the attention of researchers who aim at employing lighter and simpler-to-interpret MLs. Moreover, due to the more straightforward coding and lower number of hyperparameters of these models, they demand less coding knowledge. The models in the second group (moderate transparency and speed) may require more effort to be implemented optimally, but they usually provide desirable accuracy-complexity trade-offs. Hence, they can be preferred for modeling in average circumstances. The models in the third group (opaque-box and slow), on the other hand, can mostly provide efficient modeling once executed by experienced operators via powerful computers.

8. Discussion, Conceptual Framework, and Future Work

8.1. Conceptual Framework

Optimizing MLGHM depends not only on ML model selection but how the ML is implemented. Based on the identified research gaps, this section leverages deeper discussions of the MLGHM assessment criteria (Section 5) to yield a framework that can support steps toward achieving ‘better’ modeling. Figure 5 shows this framework, suggesting the incorporation of important yet sometimes overlooked steps throughout the modeling. Explanations regarding each stage are presented, combined with the discussion in the following sections.

8.1.1. Stage 0

Role of assumptions: Scorzato [232] states that the concept of assumption-free predictions is a fallacy. Here are some important (from well-studied to partially considered to be largely overlooked) aspects regarding model assumptions that need to be checked during model design, data preparation, and implementation:
(a)
Spatial heterogeneity of data: How does the model deal with the heterogeneous distribution of the relationship between target and predictor variables? For large study areas, a possible solution can be found in Ref. [124] wherein the study area was partitioned into three sub-domains to account for the area’s heterogeneity.
(b)
Temporal variations (upon occurrence): How does the model account for temporal changes in contaminant concentrations and environmental conditions? Performing separate seasonal assessments (as in Ref. [233]) and considering temporal predictive variables (e.g., precipitation and temperature seasonality, as in Ref. [234]) may help address this issue.
(c)
Data quality and representativeness: To what extent can the provided dataset represent the sampling frame in the study area? Proper selection of predictive variables can assist the model in obtaining a deeper understanding of the relationship between contaminant and environmental conditions. For instance, aquifer dynamics can be expressed in terms of both physical and chemical characteristics [235], necessitating the inclusion of predictive variables such as time-dependent flow and water pH. Moreover, testing for multicollinearity in the data can address another significant question: Does each factor effectively (and independently) represent an aspect of the problem? Removing unnecessary factors can improve model stability and interpretability.
(d)
Problem’s non-linearity: How to ensure the model can capture any non-linear dependency of a contaminant target variable on various factors simultaneously? Comparing the accuracy of the candidate ML models to benchmark linear models (e.g., MLR in Ref. [163]) can reflect the relative capability of the models in handling MLGHM non-linearities. Plus, some linear models have pre-implementation conditions to be satisfied (e.g., ensuring the normal distribution of the target variable when using MLR).
(e)
Model verification: How to ensure a satisfactorily trained model has been protected against undesirable phenomena such as local minima and overfitting? (See Table 1).
Based on these questions, it can be said that analysing a dataset via a powerful ML model—even one achieving a high accuracy—does not necessarily guarantee results that may accurately reflect actual physical processes or have a satisfactory predictive power beyond the area for which data are available. Proper consideration of assumptions and recognizing sources of uncertainties and possible biases [236] can be of great help in model evaluation, which may also require integration of geochemical, environmental, and data science expert knowledge.

8.1.2. Stage 1

Feature selection: Most feature selection techniques (e.g., PCA and RF) follow strong statistical principles to assess the importance of predictive variables. However, for a given dataset, different feature selectors can yield different results. For instance, while PCA works based on explaining the variation of the target parameter by the predictive variables [237], the tree-based approaches reflect a decline in accuracy caused by permutations in the predictive variable [219]. For this reason, trying different feature selection techniques to achieve optimal combinations of predictive variables can be useful. However, applied feature selection methods can be unstable. For example, Mahanty, et al. [238] showed that excluding a single water quality factor from the PCA method can result in up to a 60% change in the WQI score. They suggested the use of Monte Carlo resampling and cross-validation to assess the PCA’s stability.
Splitting ratio: Two ratios of 70/30 and 80/20 are widely used for splitting ML datasets into training and testing datasets. However, a more deliberate selection of the splitting ratio is advisable, as the optimal ratio can vary from one problem (and/or one model) to another. An example of splitting ratio-centric sensitivity analysis can be found in Figure 5 and Table 6 of Ref. [116].
Cross-validation: Many ML studies rely upon a single and random split of the training and testing data, which can make the model vulnerable to performance overestimation. More clearly, spatial autocorrelation coupled with the possible unequal distribution of extreme values within the training and testing subsets can reduce the model’s accuracy. This flaw may lead to an unreliable conclusion about the model’s performance in dealing with real-world unseen extreme conditions. Data-centric cross-validation can replace the random splitting policy to secure the model against such risks. K-fold cross-validation [239] is one of the most popular techniques that enables the model to be trained and tested by multiple subsets of data.

8.1.3. Stage 2

Hyperparameters: Hyperparameters can highly influence the robustness of ML models. A list of well-known ML hyperparameters is presented in Table 2 of Ref. [223]. So far, various techniques, including model-free algorithms, gradient-based optimization, Bayesian optimization, multi-fidelity optimization, and MOAs, have been proposed for optimizing hyperparameters, noting that the optimal technique is ML-type dependent [223,240]. While utilizing an optimization algorithm reduces human effort by automating hyperparameter tuning, working with advanced algorithms can be challenging. Many ML models can benefit from simpler methods such as the GS algorithm and manual T&E (A.K.A. babysitting) (Table 1). Moreover, referring to similar earlier studies can supply useful insights to tune a model’s hyperparameters. The sensitivity of the model to a hyperparameter can be monitored through repeatedly computing the accuracy and/or error indicators. For instance, Figure 5 and Figure 6 of Ref. [241] show the variation of accuracy vs. the number of trees for StoGB and variable subsets (and ensemble size) for RotF, respectively. Another example is provided by Ref. [193] (their Figure S5) in which error measurement was used to determine an optimum number of trees for RF, and accuracy measurements were used to determine an optimum tree depth for BRT.
Computational efficiency: The advent of sophisticated methods, such as ensembles and DL, has raised concerns regarding model efficiency [242], especially for complicated environmental simulations such as GHM that demand processing large datasets. Christin, et al. [243] discussed to-be-considered factors (such as programming framework, data provision and augmentation, and required computer power) when applying DL models to ecological studies. Sivakumar, et al. [244] highlighted the following factors to evaluate the efficiency of ML algorithms: (a) training time, (b) prediction time, (c) memory usage, (d) CPU usage, (e) GPU usage, (f) RAM usage, (g) scalability, (h) robustness, and (i) adaptability. They also used an AHP framework to assess the importance of these factors in tackling different problems. The results showed that, depending on the essence of the problem at hand, these factors can be prioritized differently. However, in the absence of powerful computers, training time becomes a critical factor for most DLs and MOA-trained models that require executing deep/or numerous iterations. It is worth mentioning that, depending on the used programming environments, measuring the training and prediction times is readily viable by adding relevant functions (e.g., tic-toc in MATLAB).
Penalty functions can provide valuable insights regarding a model’s complexity. AIC [245] and BIC [246] are among the reputable penalty functions (details in Refs. [247,248]). These metrics can offer two major benefits to ML models, viz. (a) optimizing combinations of predictive variables and (b) performance assessment.

8.1.4. Stage 3

Generalizability and overfitting: In the context of MLGHM, one may ask ’To what extent is the model that is developed for Aquifer A applicable to Aquifer B?’ knowing that the two aquifers may have different characteristics. When these differences are not substantial, area-centric cross-validation can reflect the model’s generalizability. It includes training the model with Aquifer A data and testing it on Aquifer B. However, there are aspects (e.g., bias correction and uncertainty estimation) that require attention [249]. An example of such area-centric cross-validation can be found in Ref. [187], in which Bangladesh data were used to calibrate an LgR model to predict As contamination in South Louisiana, US.
As an impediment to generalizability, overfitting threatens ML models. A general solution to this issue is finding a trade-off between bias (i.e., the error due to oversimplification) and variance (i.e., the model’s sensitivity to small changes in data). Technical details of overfitting and feasible specific solutions are well discussed in relevant studies, such as Refs. [141,250,251,252]. That said, some straightforward methodological hints to deal with overfitting are as follows:
(i)
Possible reasons: Apart from the presence of noise in data and having a limited number of training samples (but with many predictive variables), it is established that the higher a model’s complexity (referred to as high-variance models), the higher the susceptibility to overfitting [250].
(ii)
Indicator: An indicator of overfitting is a considerable difference between the training and testing accuracies (AccuracyTraining >> AccuracyTesting). Moreover, one may estimate the proneness to overfitting by monitoring accuracy during area-centric and/or data-centric cross-validations.
(iii)
Possible solutions: Feasible solutions to overfitting can include leveraging data augmentation techniques in the event of data insufficiency (as in Ref. [253]); applying feature selection in the case of high dimensionality; using regularization methods (See Table 1); trying simpler (shallower) versions if the current models are too complex (deep); and employing models that are known for being more immune to overfitting (e.g., XGBoost, See Table 1).
Explainability: The growing demand of end-users to replace conventional AI models with explainable ones (noting that Interpretability ⊂ Explainability [254]) may lead to the rejection of opaque-box AI and the switch to XAI. Notwithstanding the growing XAI trends, some argue about the necessity of switching from opaque-box to XAI. As Holm [255] stated, “This blanket rejection of … [opaque-box] methods may be hasty”. A potential reason for such arguments, perhaps, can be sought in the relationship between a model’s explainability and its accuracy, which is usually inverse (see Figure 3 in Ref. [256]; and consideration of DL and DT as examples of opaque-box and XIA models, respectively, of various environmental hazards, e.g., GHM [170], landslide [257], and flood [258]).
This raises an important question: ‘Can we move from opaque-box ML to explainable ML without sacrificing accuracy?’ While attaining a trade-off between accuracy and transparency remains a central consideration, a positive answer to this question may lie in increasing the transparency of accurate models instead of switching to naturally explainable ones, and thereby, maintaining the precedence of modeling accuracy. Some guidelines for increasing the explainability of DL—as one of the most complex ML models—are presented in Text S3 and Figure S7.
Optimization algorithms: Implementing an ML model can be associated with a number of computational challenges, such as local minima, the presence of which may require auxiliary algorithms to achieve a reliable prediction. MOAs can assist ML for proper tuning of the computational variables and model’s hyperparameters to protect the prediction from computational drawbacks. In the surveyed MLGHM literature, several studies were found that optimized ML using MOAs (See Text S4 and Tables S8 and S9).
MOAs cope with computational difficulties through an iterative process. For instance, Figure 4 in Ref. [181] demonstrates the continuous efforts of the applied MOAs; an optimization curve with no degradation in quality, even when no improvement is observed. This either-steady-or-upward trend of the solution’s quality is deemed a significant advantage against conventional ML models that may yield performance decline after several epochs due to overfitting or learning rate issues (e.g., consecutive validation fails, which is a well-known stopping criterion for ANNs). This iterative effort, however, demands time and computational power depending on the algorithm’s nature, which can be partially manifested by the number of required iterations and population size. In most reviewed MOA-based studies, these two factors could have been tuned more carefully as MOAs were executed with a small number of iterations (≤300) and small population sizes (≤100). Owing to the MOAs’ behaviour, some would benefit from larger population sizes and may keep converging even up to 1000 iterations (e.g., in Refs. [259,260]). To address the issue of intensive computations, scientists may also use more efficient MOAs—e.g., electromagnetic field optimization (EFO)—that can perform a high-quality optimization in a reasonable time [237].

8.1.5. Stage 4

Informed hazard zonation:
Hazard zonation can influence exposure results (and subsequently, the ultimate risk assessment), leading to substantial health and/or monetary impacts on the stakeholders (e.g., people, government, and NGOs). Therefore, as one of the final steps of GHM, utilizing a proper classification strategy is of utmost importance. There are several classification methods of modeled environmental hazards. Natural Breaks (Jenks’ optimization method [261]), Equal Intervals, and Quantiles are among the widely used methods, each with specific pros and cons. More recently, models leverage the receiver operating characteristic (ROC) curve to find a cutoff point that optimally differentiates hazard levels. For instance, Cantarino, et al. [262] proposed a new ROC-based model for landslide susceptibility mapping. They reported a considerable distinction (>20% of the surface) between different models in detecting the most susceptible parts. As for GHM, Wu and Polya [263] proposed a potentially cost-saving ROC-based objective cost-benefit informed criteria, which led to different cutoff values for identifying high and low groundwater arsenic hazard.

8.2. Future Directions

This section addresses some emerging directions that can hold promise to attain advanced MLGHM in future efforts.
(i)
Although XAI was discussed earlier (e.g., Section 8.1.4), increasing the transparency is still considered a significant emerging direction towards enhancing stakeholders’ trust in modeling outcomes. This study offered an example of moving from opaque-box to transparent-box for DL (see Text S3 and Figure S7). However, other methods, such as SHAP, along with monotonic constraints [264], can be helpful to elucidate models’ calculations.
(ii)
Employing PIML models—embedding physical laws into learning algorithms—is another growing frontier to enhance scientific plausibility and reliability of the model under extrapolation [265,266]. An example of this in MLGHM works can be incorporating hydrogeological rules (e.g., groundwater flow directs contamination, hence, no upgradient contamination movement) into the mathematical computations of the predictive model.
(iii)
The use of IoT-empowered sensing networks can enrich modeling by providing continuous, real-time data [267,268]. It can also lead to more realistic results due to analyzing dynamic conditions of the predicted and predictive factors instead of static data.
(iv)
Transfer learning (e.g., fine-tuning a pre-trained model with limited data from the new to-be-assessed region) [269] and federated learning (e.g., collaborative training of models without sharing raw data) [270] are two approaches that can contribute to tackling challenges stemming from data scarcity and data privacy.
(v)
While many traditional hazard zonation models are easily accessible in software such as GIS, more recent models may require external tools (and more statistical knowledge) to implement, making them challenging to use for some. Hence, future efforts should also focus on developing user-friendly graphical user interfaces (GUIs) to facilitate the use of GHM models and new hazard classification methods.
These directions, together with concepts discussed in Section 5, Section 6, Section 7 and Section 8, suggest that the basis for strengthening the role of next-generation ML in GHM (and generally sustainable water resource management) will go beyond technical accuracy alone, extending to the model’s real-time applicability and scientific consistency.

9. Conclusions

The main conclusions of this systematic survey of 284 ML in groundwater hazard mapping (MLGHM) articles published over the period 2005–2024 are summarized below.
(i)
There has been an exponential increase in published applications of MLGHM models over the past two decades.
(ii)
Iran, India, the US, and China are the countries most modeled in these studies, but there are many other countries that would benefit from (further) MLGHM.
(iii)
Nitrate, water quality index, and arsenic featured as the most studied target groundwater quality indicators. The most frequently considered categories of predictive variables were groundwater chemical characteristics and geomorphological factors.
(iv)
Tree-based ML was the most popular feature selection method.
(v)
Tree-based ML (mostly RF) was the most popular MLGHM model.
(vi)
Published evaluations of model superiority have been largely based on accuracy. On the basis of many articles, resampling-improved ensembles, meta-model ensembles, and tree-based models outperformed a wide variety of models, including (but not limited to) shallow ANNs, SVMs, NFs, and regression models. A more detailed accuracy comparison (based on the existing pieces of evidence) then reflected a close competition among boosting ensembles, random forest, and deep learning as the most promising members of their families.
(vii)
Aside from accuracy, other key factors—notably including model complexity and predictability—have been largely overlooked in assessing ML model quality, notwithstanding their often key importance in determining what are better models for particular types of problems.
(viii)
We present a flowchart that can inform GHM researchers to help better select the most appropriate popular ML models on the basis of requirements for (a) accuracy, (b) transparency, (c) training speed, and (d) number of hyperparameters, as well as (e) the intended modeling scale and (f) required user’s expertise.
(ix)
Notable gaps that were identified in the surveyed studies were discussed in depth, and several ideas were suggested for optimizing MLGHM models by emphasizing the features of a ‘better model’.
Whilst this review has covered a large number of high-quality articles, some articles might have been missed by the search strings or due to the rapid developments in the field of MLGHM. Challenges in both classifying reported variables and models are acknowledged due, in part, to differences in terminology reported. Robust global conclusions were precluded for some model types because of a paucity of studies or reported model quality criteria (other than accuracy) in many other studies.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/w17192861/s1. Text S1: Further explanations regarding GV assessment. Text S2: Further explanations regarding the study area of some articles. Text S3: An example of improving deep learning transparency. Text S4: Further explanations regarding the MOA-based studies. Figure S1: Historical (2005–2024) trends of the number of publications per annum of ML models related to groundwater quality. The red curve is an indicative exponential trend. Figure S2: Geographical distribution of the study-focus of reviewed published MLGHM studies. Boundaries and names were adopted from generalized shapefiles of the world obtained from the United Nations [1] and Environmental Systems Research Institute (ESRI) [2]. Figure S3: The proportion of the target parameters in the surveyed literature. Acronyms can be found in the Nomenclature. (Format of the included texts: Name, Frequency, Percentage of frequency). Figure S4: Frequency of use of major categories of predictive variables. Acronyms can be found in the Nomenclature. (Format of the included texts: Name, Frequency, Percentage of frequency). Figure S5: Sankey Diagram showing predictive variables and their frequency of use. See Tables S4 and S5 for details. (Plotted using Sankeymatic [3]). Figure S6: The frequency of use of feature selection techniques used in the surveyed literature. Acronyms can be found in the Nomenclature. (Format of the included texts: Name, Frequency, Percentage of frequency). Figure S7: A schematic example of increasing the transparency of a DL opaque-box model (Feature Selection image reproduced from Podgorski, et al. [4], licensed under CC BY 4.0, and the graph of Tansig activation function adopted from https://uk.mathworks.com/help/deeplearning/ref/tansig.html (8 August 2025) [5])). Table S1: Recent review articles of the ML-based groundwater quality assessment. Table S2: The keywords (besides prediction, geospatial, mapping, GIS) used during the search for MLGHM studies. Table S3: MLGHM studies with supranational study areas. Table S4: Details of the classification of the inputs used in the MLGHM literature. Table S5: Components of the included predefined datasets. Table S6: Specific predictive variables with the highest frequencies of use. Table S7: Classification of the ML models used in the MLGHM literature. Table S8: Application of optimization algorithms in the MLGHM literature. Table S9: Examples of using MOAs out of the MLGHM literature.

Author Contributions

Conceptualization, M.M., D.A.P. and Y.H.; methodology, M.M., D.A.P. and Y.H.; software, M.M.; validation, M.M., D.A.P. and Y.H.; formal analysis, M.M.; investigation, M.M., D.A.P. and Y.H.; resources, D.A.P. and Y.H.; data curation, M.M.; writing—original draft preparation, M.M.; writing—review and editing, M.M., D.A.P. and Y.H.; visualization, M.M.; supervision, D.A.P. and Y.H.; project administration, D.A.P. All authors have read and agreed to the published version of the manuscript.

Funding

UK Engineering and Physical Sciences Research Council (EPSRC) & Economic and Social Research Council (ESRC) innovation acceleration account (IAA) award (University of Manchester IAA527) to D.A.P., Y.H., and Ilya Strashnov.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Acknowledgments

M.M. acknowledges with thanks the receipt of a University of Manchester Faculty of Science and Engineering Dean’s Doctoral Scholarship. We acknowledge receipt of a UK Engineering and Physical Sciences Research Council (EPSRC) & Economic and Social Research Council (ESRC) innovation acceleration account (IAA) award (University of Manchester IAA527) to D.A.P., Y.H., and Ilya Strashnov.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1. Surveyed studies for the identified ML categories.
Table A1. Surveyed studies for the identified ML categories.
TypeCategoryReference
SupervisedANN[46,50,61,62,92,105,109,116,144,145,146,151,160,165,168,169,170,171,173,174,175,176,177,179,188,189,191,192,194,201,204,206,212,213,217,233,271,272,273,274,275,276,277,278,279,280,281,282,283,284,285,286,287,288,289,290,291,292,293,294,295,296,297,298,299,300,301,302,303,304,305,306,307,308,309,310,311,312,313,314,315]
NF[69,105,124,145,147,174,175,177,180,181,188,189,190,191,315,316,317,318,319,320]
Tree-based[7,8,46,74,75,92,109,115,116,144,145,146,147,148,149,150,151,152,153,154,155,156,157,158,159,160,161,162,163,164,165,166,167,168,170,171,172,173,175,176,178,179,187,192,195,198,183,184,185,193,194,199,200,206,208,209,231,234,241,280,281,282,283,284,286,294,297,298,303,304,306,308,311,313,320,321,322,323,324,325,326,327,328,329,330,331,332,333,334,335,336,337,338,339,340,341,342,343,344,345,346,347,348,349,350,351,352,353,354,355,356,357,358,359,360,361,362,363,364,365,366,367,368,369,370,371,372,373,374,375,376,377,378,379,380,381,382,383]
SVM[50,51,52,81,109,115,116,124,144,146,147,148,150,153,154,160,165,173,174,175,176,179,182,188,190,192,196,197,198,199,200,201,202,203,204,212,233,276,284,280,281,282,286,290,292,296,298,304,306,308,309,322,331,336,343,358,383,384,385,386,387,388,389]
Regression[74,75,86,116,144,145,150,152,155,157,163,169,173,176,179,182,186,187,193,194,195,202,203,213,241,285,296,303,306,308,309,311,318,319,320,332,339,341,343,347,352,356,357,366,376,377,383,390,391,392,393,394,395,396,397,398,399,400,401,402,403,404,405,406,407,408,409]
Probabilistic[92,116,144,148,149,159,166,170,172,173,179,192,197,199,202,203,204,205,206,207,209,213,292,309,331,385,386,410]
Ensemble 1[46,92,114,116,144,146,150,151,152,153,154,155,156,157,158,159,160,161,162,163,164,165,166,167,169,178,179,187,192,193,195,197,202,209,241,291,298,306,308,323,332,336,340,346,347,348,350,352,354,355,360,363,368,372,377,383,385,386,411,412,413,414,415,416,417,418,419,420,421,422,423,424,425,426]
Ensemble 2[105,109,116,124,135,147,172,173,174,175,176,177,179,180,181,182,183,184,188,203,214,215,216,217,302,306,340,342,391,427]
Others[30,51,114,115,116,124,125,145,147,148,151,152,155,172,173,179,192,199,200,204,208,308,341,342,358,428]
UnsupervisedClustering[131,324,427,429,430,431,432,433,434,435,436,437,438,439,440,441,442,443,444,445]
SOM[134,191,439,446,447,448,449,450,451,452,453]
DL[135]

References

  1. Ryder, G. The United Nations World Water Development Report; Wastewater: The Untapped Resource; UNESCO Pub.: London, UK, 2017. [Google Scholar]
  2. Rodella, A.-S.; Zaveri, E.; Bertone, F. The Hidden Wealth of Nations: The Economics of Groundwater in Times of Climate Change; World Bank: Washington, DC, USA, 2023. [Google Scholar]
  3. Hasan, M.F.; Smith, R.; Vajedian, S.; Pommerenke, R.; Majumdar, S. Global land subsidence mapping reveals widespread loss of aquifer storage capacity. Nat. Commun. 2023, 14, 6180. [Google Scholar] [CrossRef]
  4. Jasechko, S.; Perrone, D. Global groundwater wells at risk of running dry. Science 2021, 372, 418–421. [Google Scholar] [CrossRef]
  5. Fan, Y.; Li, H.; Miguez-Macho, G. Global patterns of groundwater table depth. Science 2013, 339, 940–943. [Google Scholar] [CrossRef]
  6. Niazi, H.; Wild, T.B.; Turner, S.W.; Graham, N.T.; Hejazi, M.; Msangi, S.; Kim, S.; Lamontagne, J.R.; Zhao, M. Global peak water limit of future groundwater withdrawals. Nat. Sustain. 2024, 7, 413–422. [Google Scholar] [CrossRef]
  7. Podgorski, J.; Berg, M. Global threat of arsenic in groundwater. Science 2020, 368, 845–850. [Google Scholar] [CrossRef] [PubMed]
  8. Podgorski, J.; Berg, M. Global analysis and prediction of fluoride in groundwater. Nat. Commun. 2022, 13, 4232. [Google Scholar] [CrossRef]
  9. Foster, S.; Garduno, H.; Kemper, K.; Tuinhof, A.; Nanni, M.; Dumars, C. Groundwater Quality Protection: Defining Strategy and Setting Priorities; The World Bank: Washington, DC, USA, 2003. [Google Scholar]
  10. Hassan, M.M. Arsenic in Groundwater: Poisoning and Risk Assessment; CRC Press: Boca Raton, FL, USA, 2018. [Google Scholar]
  11. Muhib, M.I.; Ali, M.M.; Tareq, S.M.; Rahman, M.M. Nitrate Pollution in the Groundwater of Bangladesh: An Emerging Threat. Sustainability 2023, 15, 8188. [Google Scholar] [CrossRef]
  12. Shaji, E.; Sarath, K.; Santosh, M.; Krishnaprasad, P.; Arya, B.; Babu, M.S. Fluoride contamination in groundwater: A global review of the status, processes, challenges, and remedial measures. Geosci. Front. 2024, 15, 101734. [Google Scholar] [CrossRef]
  13. Haldar, D.; Duarah, P.; Purkait, M.K. MOFs for the treatment of arsenic, fluoride and iron contaminated drinking water: A review. Chemosphere 2020, 251, 126388. [Google Scholar] [CrossRef]
  14. Zheng, Y. Global solutions to a silent poison. Science 2020, 368, 818–819. [Google Scholar] [CrossRef]
  15. Fendorf, S.; Michael, H.A.; Van Geen, A. Spatial and temporal variations of groundwater arsenic in South and Southeast Asia. Science 2010, 328, 1123–1127. [Google Scholar] [CrossRef]
  16. Nations, T.U. Goal 6: Ensure Access to Water and Sanitation for All. Available online: https://www.un.org/sustainabledevelopment/water-and-sanitation/#:~:text=Goal%206:%20Ensure%20access%20to%20water%20and%20sanitation%20for%20all&text=Access%20to%20safe%20water%2C%20sanitation,cent%20to%2073%20per%20cent (accessed on 1 June 2025).
  17. Bhattacharya, P.; Polya, D.; Jovanovic, D. Best Practice Guide on the Control of Arsenic in Drinking Water; IWA Publishing: London, UK, 2017. [Google Scholar]
  18. UNICEF. Arsenic Primer: Guidance on the Investigation & Mitigation of Arsenic Contamination; UNICEF: New York, NY, USA, 2018. [Google Scholar]
  19. Dewan, A.; Dewan, A.M. Hazards, Risk, and Vulnerability; Springer: Berlin/Heidelberg, Germany, 2013. [Google Scholar]
  20. Xiong, H.; Wang, Y.; Guo, X.; Han, J.; Ma, C.; Zhang, X. Current status and future challenges of groundwater vulnerability assessment: A bibliometric analysis. J. Hydrol. 2022, 615, 128694. [Google Scholar] [CrossRef]
  21. Abegeja, D.; Nedaw, D. Identification of groundwater potential zones using geospatial technologies in Meki Catchment, Ethiopia. Geol. Ecol. Landsc. 2024, 1–16. [Google Scholar] [CrossRef]
  22. Machiwal, D.; Cloutier, V.; Güler, C.; Kazakis, N. A review of GIS-integrated statistical techniques for groundwater quality evaluation and protection. Environ. Earth Sci. 2018, 77, 681. [Google Scholar] [CrossRef]
  23. Jain, H. Groundwater vulnerability and risk mitigation: A comprehensive review of the techniques and applications. Groundw. Sustain. Dev. 2023, 22, 100968. [Google Scholar] [CrossRef]
  24. Nistor, M.-M. Groundwater vulnerability in Europe under climate change. Quat. Int. 2020, 547, 185–196. [Google Scholar] [CrossRef]
  25. Al-Abadi, A.M. The application of Dempster–Shafer theory of evidence for assessing groundwater vulnerability at Galal Badra basin, Wasit governorate, east of Iraq. Appl. Water Sci. 2017, 7, 1725–1740. [Google Scholar] [CrossRef]
  26. Neshat, A.; Pradhan, B. An integrated DRASTIC model using frequency ratio and two new hybrid methods for groundwater vulnerability assessment. Nat. Hazards 2015, 76, 543–563. [Google Scholar] [CrossRef]
  27. Abbasi, S.; Mohammadi, K.; Kholghi, M.; Howard, K. Aquifer vulnerability assessments using DRASTIC, weights of evidence and the analytic element method. Hydrol. Sci. J. 2013, 58, 186–197. [Google Scholar] [CrossRef]
  28. Goudarzi, S.; Jozi, S.A.; Monavari, S.M.; Karbasi, A.; Hasani, A.H. Assessment of groundwater vulnerability to nitrate pollution caused by agricultural practices. Water Qual. Res. J. 2015, 52, 64–77. [Google Scholar] [CrossRef]
  29. Kang, J.; Zhao, L.; Li, R.; Mo, H.; Li, Y. Groundwater vulnerability assessment based on modified DRASTIC model: A case study in Changli County, China. Geocarto Int. 2017, 32, 749–758. [Google Scholar] [CrossRef]
  30. Msaddek, M.H.; Moumni, Y.; Ayari, A.; El May, M.; Chenini, I. Artificial intelligence modelling framework for mapping groundwater vulnerability of fractured aquifer. Geocarto Int. 2022, 37, 10480–10510. [Google Scholar] [CrossRef]
  31. El Naqa, I.; Murphy, M.J. What Is Machine Learning; Springer: Berlin/Heidelberg, Germany, 2015. [Google Scholar]
  32. Ayodele, T.O. Types of machine learning algorithms. New Adv. Mach. Learn. 2010, 3, 19–48. [Google Scholar]
  33. Morales, E.F.; Escalante, H.J. A brief introduction to supervised, unsupervised, and reinforcement learning. In Biosignal Processing and Classification Using Computational Learning and Intelligence; Elsevier: Amsterdam, The Netherlands, 2022; pp. 111–129. [Google Scholar]
  34. Wachniew, P.; Zurek, A.J.; Stumpp, C.; Gemitzi, A.; Gargini, A.; Filippini, M.; Rozanski, K.; Meeks, J.; Kværner, J.; Witczak, S. Toward operational methods for the assessment of intrinsic groundwater vulnerability: A review. Crit. Rev. Environ. Sci. Technol. 2016, 46, 827–884. [Google Scholar] [CrossRef]
  35. Tung, T.M.; Yaseen, Z.M. A survey on river water quality modelling using artificial intelligence models: 2000–2020. J. Hydrol. 2020, 585, 124670. [Google Scholar] [CrossRef]
  36. Zhu, M.; Wang, J.; Yang, X.; Zhang, Y.; Zhang, L.; Ren, H.; Wu, B.; Ye, L. A review of the application of machine learning in water quality evaluation. Eco-Environ. Health 2022, 1, 107–116. [Google Scholar] [CrossRef]
  37. Ewuzie, U.; Bolade, O.P.; Egbedina, A.O. Application of deep learning and machine learning methods in water quality modeling and prediction: A review. In Current Trends and Advances in Computer-Aided Intelligent Environmental Data Engineering; Academic Press: Cambridge, MA, USA, 2022; pp. 185–218. [Google Scholar] [CrossRef]
  38. Hanoon, M.S.; Ahmed, A.N.; Fai, C.M.; Birima, A.H.; Razzaq, A.; Sherif, M.; Sefelnasr, A.; El-Shafie, A. Application of artificial intelligence models for modeling water quality in groundwater: Comprehensive review, evaluation and future trends. Water Air Soil Pollut. 2021, 232, 411. [Google Scholar] [CrossRef]
  39. Huang, Y.; Wang, C.; Wang, Y.; Lyu, G.; Lin, S.; Liu, W.; Niu, H.; Hu, Q. Application of machine learning models in groundwater quality assessment and prediction: Progress and challenges. Front. Environ. Sci. Eng. 2024, 18, 29. [Google Scholar] [CrossRef]
  40. Chowdhury, T.N.N.; Battamo, A.Y.; Nag, R.; Zekker, I.; Salauddin, M. Impacts of climate change on groundwater quality: A systematic literature review of analytical models and machine learning techniques. Environ. Res. Lett. 2025, 20, 033003. [Google Scholar] [CrossRef]
  41. Haggerty, R.; Sun, J.; Yu, H.; Li, Y. Application of machine learning in groundwater quality modeling-A comprehensive review. Water Res. 2023, 233, 119745. [Google Scholar] [CrossRef]
  42. Pandya, H.; Jaiswal, K.; Shah, M. A comprehensive review of machine learning algorithms and its application in groundwater quality prediction. Arch. Comput. Methods Eng. 2024, 31, 4633–4654. [Google Scholar] [CrossRef]
  43. Haghbin, M.; Sharafati, A.; Dixon, B.; Kumar, V. Application of soft computing models for simulating nitrate contamination in groundwater: Comprehensive review, assessment and future opportunities. Arch. Comput. Methods Eng. 2021, 28, 3569–3591. [Google Scholar] [CrossRef]
  44. Kumari, N.; Pathak, G.; Prakash, O. A review of application of artificial neural network in ground water modeling. In Proceedings of the International Conference on Recent Advances in Mathematics, Statistics and Computer Science 2015, Bihar, India, 29–31 May 2015; pp. 393–402. [Google Scholar] [CrossRef]
  45. Scopus. Scopus-Indexed Journals. Available online: https://www.scopus.com/ (accessed on 1 February 2025).
  46. Singha, S.; Pasupuleti, S.; Singha, S.S.; Singh, R.; Kumar, S. Prediction of groundwater quality using efficient machine learning technique. Chemosphere 2021, 276, 130265. [Google Scholar] [CrossRef]
  47. Masoud, A.A. Groundwater quality assessment of the shallow aquifers west of the Nile Delta (Egypt) using multivariate statistical and geostatistical techniques. J. Afr. Earth Sci. 2014, 95, 123–137. [Google Scholar] [CrossRef]
  48. Sun, Y.; Xu, S.; Wang, Q.; Hu, S.; Qin, G.; Yu, H. Response of a Coastal Groundwater System to Natural and Anthropogenic Factors: Case study on east coast of laizhou bay, china. Int. J. Environ. Res. Public Health 2020, 17, 5204. [Google Scholar] [CrossRef]
  49. Romero, J.M.; Salazar, D.C.; Melo, C.E. Hydrogeological spatial modelling: A comparison between frequentist and Bayesian statistics. J. Geophys. Eng. 2023, 20, 523–537. [Google Scholar] [CrossRef]
  50. Gharekhani, M.; Nadiri, A.A.; Khatibi, R.; Nikoo, M.R.; Barzegar, R.; Sadeghfam, S.; Moghaddam, A.A. Quantifying the groundwater total contamination risk using an inclusive multi-level modelling strategy. J. Environ. Manag. 2023, 332, 117287. [Google Scholar] [CrossRef]
  51. Nadiri, A.A.; Norouzi, H.; Khatibi, R.; Gharekhani, M. Groundwater DRASTIC vulnerability mapping by unsupervised and supervised techniques using a modelling strategy in two levels. J. Hydrol. 2019, 574, 744–759. [Google Scholar] [CrossRef]
  52. Nadiri, A.A.; Gharekhani, M.; Khatibi, R. Mapping aquifer vulnerability indices using artificial intelligence-running multiple frameworks (AIMF) with supervised and unsupervised learning. Water Resour. Manag. 2018, 32, 3023–3040. [Google Scholar] [CrossRef]
  53. Vesselinov, V.V.; Alexandrov, B.S.; O’Malley, D. Contaminant source identification using semi-supervised machine learning. J. Contam. Hydrol. 2018, 212, 134–142. [Google Scholar] [CrossRef]
  54. Tiwari, A. Supervised learning: From theory to applications. In Artificial Intelligence and Machine Learning for EDGE Computing; Elsevier: Amsterdam, The Netherlands, 2022; pp. 23–32. [Google Scholar]
  55. McCulloch, W.S.; Pitts, W. A logical calculus of the ideas immanent in nervous activity. Bull. Math. Biophys. 1943, 5, 115–133. [Google Scholar] [CrossRef]
  56. Hornik, K.; Stinchcombe, M.; White, H. Multilayer feedforward networks are universal approximators. Neural Netw. 1989, 2, 359–366. [Google Scholar] [CrossRef]
  57. Werbos, P.J. Backpropagation through time: What it does and how to do it. Proc. IEEE 1990, 78, 1550–1560. [Google Scholar] [CrossRef]
  58. Moré, J.J. The Levenberg-Marquardt algorithm: Implementation and theory. In Proceedings of the Numerical analysis: Proceedings of the Biennial Conference Held at Berlin, Heidelberg, Germany, 28 June–1 July 1977; Springer: Berlin/Heidelberg, Germany, 2006; pp. 105–116. [Google Scholar]
  59. Chen, Y.; Song, L.; Liu, Y.; Yang, L.; Li, D. A review of the artificial neural network models for water quality prediction. Appl. Sci. 2020, 10, 5776. [Google Scholar] [CrossRef]
  60. Boo, K.B.W.; El-Shafie, A.; Othman, F.; Khan, M.M.H.; Birima, A.H.; Ahmed, A.N. Groundwater level forecasting with machine learning models: A review. Water Res. 2024, 252, 121249. [Google Scholar] [CrossRef]
  61. Almasri, M.N.; Kaluarachchi, J.J. Modular neural networks to predict the nitrate distribution in ground water using the on-ground nitrogen loading and recharge data. Environ. Model. Softw. 2005, 20, 851–871. [Google Scholar] [CrossRef]
  62. Wang, M.; Liu, G.; Wu, W.; Bao, Y.; Liu, W. Prediction of agriculture derived groundwater nitrate distribution in North China Plain with GIS-based BPNN. Environ. Geol. 2006, 50, 637–644. [Google Scholar] [CrossRef]
  63. Ouifak, H.; Idri, A. Application of neuro-fuzzy ensembles across domains: A systematic review of the two last decades (2000–2022). Eng. Appl. Artif. Intell. 2023, 124, 106582. [Google Scholar] [CrossRef]
  64. Zadeh, L.A. Fuzzy sets. Inf. Control 1965, 8, 338–353. [Google Scholar] [CrossRef]
  65. Jang, J.-S. ANFIS: Adaptive-network-based fuzzy inference system. IEEE Trans. Syst. Man Cybern. 1993, 23, 665–685. [Google Scholar] [CrossRef]
  66. Mehrabi, M.; Pradhan, B.; Moayedi, H.; Alamri, A. Optimizing an adaptive neuro-fuzzy inference system for spatial prediction of landslide susceptibility using four state-of-the-art metaheuristic techniques. Sensors 2020, 20, 1723. [Google Scholar]
  67. Ang, Y.K.; Talei, A.; Zahidi, I.; Rashidi, A. Past, present, and future of using neuro-fuzzy systems for hydrological modeling and forecasting. Hydrology 2023, 10, 36. [Google Scholar] [CrossRef]
  68. Tao, H.; Hameed, M.M.; Marhoon, H.A.; Zounemat-Kermani, M.; Heddam, S.; Kim, S.; Sulaiman, S.O.; Tan, M.L.; Sa’adi, Z.; Mehr, A.D. Groundwater level prediction using machine learning models: A comprehensive review. Neurocomputing 2022, 489, 271–308. [Google Scholar] [CrossRef]
  69. Dixon, B. Applicability of neuro-fuzzy techniques in predicting ground-water vulnerability: A GIS-based sensitivity analysis. J. Hydrol. 2005, 309, 17–38. [Google Scholar] [CrossRef]
  70. Grinsztajn, L.; Oyallon, E.; Varoquaux, G. Why do tree-based models still outperform deep learning on typical tabular data? Adv. Neural Inf. Process. Syst. 2022, 35, 507–520. [Google Scholar]
  71. Januschowski, T.; Wang, Y.; Torkkola, K.; Erkkilä, T.; Hasson, H.; Gasthaus, J. Forecasting with trees. Int. J. Forecast. 2022, 38, 1473–1481. [Google Scholar] [CrossRef]
  72. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  73. Tyralis, H.; Papacharalampous, G.; Langousis, A. A brief review of random forests for water scientists and practitioners and their recent history in water resources. Water 2019, 11, 910. [Google Scholar] [CrossRef]
  74. Rodriguez-Galiano, V.; Mendes, M.P.; Garcia-Soldado, M.J.; Chica-Olmo, M.; Ribeiro, L. Predictive modeling of groundwater nitrate pollution using Random Forest and multisource variables related to intrinsic and specific vulnerability: A case study in an agricultural setting (Southern Spain). Sci. Total Environ. 2014, 476, 189–206. [Google Scholar] [CrossRef]
  75. Nolan, B.T.; Gronberg, J.M.; Faunt, C.C.; Eberts, S.M.; Belitz, K. Modeling nitrate at domestic and public-supply well depths in the Central Valley, California. Environ. Sci. Technol. 2014, 48, 5643–5651. [Google Scholar] [CrossRef]
  76. Cortes, C. Support-Vector Networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  77. Wang, L. Support Vector Machines: Theory and Applications; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2005; Volume 177. [Google Scholar]
  78. Deka, P.C. Support vector machine applications in the field of hydrology: A review. Appl. Soft Comput. 2014, 19, 372–386. [Google Scholar] [CrossRef]
  79. Zendehboudi, A.; Baseer, M.A.; Saidur, R. Application of support vector machine models for forecasting solar and wind energy resources: A review. J. Clean. Prod. 2018, 199, 272–285. [Google Scholar] [CrossRef]
  80. Kok, Z.H.; Shariff, A.R.M.; Alfatni, M.S.M.; Khairunniza-Bejo, S. Support vector machine in precision agriculture: A review. Comput. Electron. Agric. 2021, 191, 106546. [Google Scholar] [CrossRef]
  81. Ammar, K.; Khalil, A.; McKee, M.; Kaluarachchi, J. Bayesian deduction for redundancy detection in groundwater quality monitoring networks. Water Resour. Res. 2008, 44, W08412. [Google Scholar] [CrossRef]
  82. Fahrmeir, L.; Kneib, T.; Lang, S.; Marx, B.; Fahrmeir, L.; Kneib, T.; Lang, S.; Marx, B. Regression Models; Springer: Berlin/Heidelberg, Germany, 2013. [Google Scholar]
  83. Ma, X.; Zou, B.; Deng, J.; Gao, J.; Longley, I.; Xiao, S.; Guo, B.; Wu, Y.; Xu, T.; Xu, X. A comprehensive review of the development of land use regression approaches for modeling spatiotemporal variations of ambient air pollution: A perspective from 2011 to 2023. Environ. Int. 2024, 183, 108430. [Google Scholar] [CrossRef]
  84. Rajaee, T.; Ebrahimi, H.; Nourani, V. A review of the artificial intelligence methods in groundwater level modeling. J. Hydrol. 2019, 572, 336–351. [Google Scholar] [CrossRef]
  85. Mosavi, A.; Ozturk, P.; Chau, K.-w. Flood prediction using machine learning models: Literature review. Water 2018, 10, 1536. [Google Scholar] [CrossRef]
  86. Twarakavi, N.K.; Kaluarachchi, J.J. Aquifer vulnerability assessment to heavy metals using ordinal logistic regression. Groundwater 2005, 43, 200–214. [Google Scholar] [CrossRef]
  87. Murphy, K.P. Machine Learning: A Probabilistic Perspective; MIT Press: Cambridge, MA, USA, 2012. [Google Scholar]
  88. Ghahramani, Z. Probabilistic machine learning and artificial intelligence. Nature 2015, 521, 452–459. [Google Scholar] [CrossRef]
  89. Bharadiya, J.P. A review of Bayesian machine learning principles, methods, and applications. Int. J. Innov. Sci. Res. Technol. 2023, 8, 2033–2038. [Google Scholar]
  90. Krapu, C.; Borsuk, M. Probabilistic programming: A review for environmental modellers. Environ. Model. Softw. 2019, 114, 40–48. [Google Scholar] [CrossRef]
  91. Papacharalampous, G.; Tyralis, H. A review of machine learning concepts and methods for addressing challenges in probabilistic hydrological post-processing and forecasting. Front. Water 2022, 4, 961954. [Google Scholar] [CrossRef]
  92. Nolan, B.T.; Fienen, M.N.; Lorenz, D.L. A statistical learning framework for groundwater nitrate models of the Central Valley, California, USA. J. Hydrol. 2015, 531, 902–911. [Google Scholar] [CrossRef]
  93. Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef]
  94. Schapire, R.E. The strength of weak learnability. Mach. Learn. 1990, 5, 197–227. [Google Scholar] [CrossRef]
  95. Ting, K.M.; Witten, I.H. Stacking Bagged and Dagged Models; Working paper 97/09; University of Waikato, Department of Computer Science: Hamilton, New Zealand, 1997; Available online: https://researchcommons.waikato.ac.nz/entities/publication/37f19edd-e874-4b2c-a26b-0b4a202ca555 (accessed on 8 August 2025).
  96. Ho, T.K. The random subspace method for constructing decision forests. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 832–844. [Google Scholar] [CrossRef]
  97. Ren, Y.; Zhang, L.; Suganthan, P.N. Ensemble classification and regression-recent developments, applications and future directions. IEEE Comput. Intell. Mag. 2016, 11, 41–53. [Google Scholar] [CrossRef]
  98. Mahajan, P.; Uddin, S.; Hajati, F.; Moni, M.A. Ensemble learning for disease prediction: A review. Healthcare 2023, 11, 1808. [Google Scholar] [CrossRef]
  99. Ganaie, M.A.; Hu, M.; Malik, A.K.; Tanveer, M.; Suganthan, P.N. Ensemble deep learning: A review. Eng. Appl. Artif. Intell. 2022, 115, 105151. [Google Scholar] [CrossRef]
  100. Sakib, M.; Mustajab, S.; Alam, M. Ensemble deep learning techniques for time series analysis: A comprehensive review, applications, open issues, challenges, and future directions. Clust. Comput. 2025, 28, 73. [Google Scholar] [CrossRef]
  101. Wen, L.; Hughes, M. Coastal wetland mapping using ensemble learning algorithms: A comparative study of bagging, boosting and stacking techniques. Remote Sens. 2020, 12, 1683. [Google Scholar] [CrossRef]
  102. Zhang, Y.; Liu, J.; Shen, W. A review of ensemble learning algorithms used in remote sensing applications. Appl. Sci. 2022, 12, 8654. [Google Scholar] [CrossRef]
  103. Zaresefat, M.; Derakhshani, R. Revolutionizing groundwater management with hybrid AI models: A practical review. Water 2023, 15, 1750. [Google Scholar] [CrossRef]
  104. Kumar, V.; Yadav, S. A state-of-the-Art review of heuristic and metaheuristic optimization techniques for the management of water resources. Water Supply 2022, 22, 3702–3728. [Google Scholar] [CrossRef]
  105. Fijani, E.; Nadiri, A.A.; Moghaddam, A.A.; Tsai, F.T.-C.; Dixon, B. Optimization of DRASTIC method by supervised committee machine artificial intelligence to assess groundwater vulnerability for Maragheh–Bonab plain aquifer, Iran. J. Hydrol. 2013, 503, 89–100. [Google Scholar] [CrossRef]
  106. Abdel-Basset, M.; Abdel-Fatah, L.; Sangaiah, A.K. Metaheuristic algorithms: A comprehensive review. Comput. Intell. Multimed. Big Data Cloud Eng. Appl. 2018, 185–231. [Google Scholar] [CrossRef]
  107. Yang, X.-S. Nature-Inspired Metaheuristic Algorithms; Luniver Press: Bristol, UK, 2010. [Google Scholar]
  108. Akay, B.; Karaboga, D.; Akay, R. A comprehensive survey on optimizing deep learning models by metaheuristics. Artif. Intell. Rev. 2022, 55, 829–894. [Google Scholar] [CrossRef]
  109. Band, S.S.; Janizadeh, S.; Pal, S.C.; Chowdhuri, I.; Siabi, Z.; Norouzi, A.; Melesse, A.M.; Shokri, M.; Mosavi, A. Comparative analysis of artificial intelligence models for accurate estimation of groundwater nitrate concentration. Sensors 2020, 20, 5763. [Google Scholar] [CrossRef]
  110. Aha, D.W.; Kibler, D.; Albert, M.K. Instance-based learning algorithms. Mach. Learn. 1991, 6, 37–66. [Google Scholar] [CrossRef]
  111. Fix, E.; Hodges, J.L., Jr. Discriminatory Analysis-Nonparametric Discrimination: Small Sample Performance; (No. UCB11). 1952. Available online: https://www.semanticscholar.org/paper/Discriminatory-Analysis-Nonparametric-Small-Sample-Fix-Hodges/1fa93fb4aa010d46b22393131bc61a0a0d041da0 (accessed on 8 August 2025).
  112. Halder, R.K.; Uddin, M.N.; Uddin, M.A.; Aryal, S.; Khraisat, A. Enhancing K-nearest neighbor algorithm: A comprehensive review and performance analysis of modifications. J. Big Data 2024, 11, 113. [Google Scholar] [CrossRef]
  113. Chirici, G.; Mura, M.; McInerney, D.; Py, N.; Tomppo, E.O.; Waser, L.T.; Travaglini, D.; McRoberts, R.E. A meta-analysis and review of the literature on the k-Nearest Neighbors technique for forestry applications that use remotely sensed data. Remote Sens. Environ. 2016, 176, 282–294. [Google Scholar] [CrossRef]
  114. Motevalli, A.; Naghibi, S.A.; Hashemi, H.; Berndtsson, R.; Pradhan, B.; Gholami, V. Inverse method using boosted regression tree and k-nearest neighbor to quantify effects of point and non-point source nitrate pollution in groundwater. J. Clean. Prod. 2019, 228, 1248–1263. [Google Scholar] [CrossRef]
  115. Rahmati, O.; Choubin, B.; Fathabadi, A.; Coulon, F.; Soltani, E.; Shahabi, H.; Mollaefar, E.; Tiefenbacher, J.; Cipullo, S.; Ahmad, B.B. Predicting uncertainty of machine learning models for modelling nitrate pollution of groundwater using quantile regression and UNEEC methods. Sci. Total Environ. 2019, 688, 855–866. [Google Scholar] [CrossRef]
  116. Díaz-Alcaide, S.; Martínez-Santos, P. Mapping fecal pollution in rural groundwater supplies by means of artificial intelligence classifiers. J. Hydrol. 2019, 577, 124006. [Google Scholar] [CrossRef]
  117. Al-Sahaf, H.; Bi, Y.; Chen, Q.; Lensen, A.; Mei, Y.; Sun, Y.; Tran, B.; Xue, B.; Zhang, M. A survey on evolutionary machine learning. J. R. Soc. N. Z. 2019, 49, 205–228. [Google Scholar] [CrossRef]
  118. Zhong, J.; Feng, L.; Ong, Y.-S. Gene expression programming: A survey. IEEE Comput. Intell. Mag. 2017, 12, 54–72. [Google Scholar] [CrossRef]
  119. Banzhaf, W.; Machado, P.; Zhang, M. Handbook of Evolutionary Machine Learning; Springer: Berlin/Heidelberg, Germany, 2024. [Google Scholar]
  120. Telikani, A.; Tahmassebi, A.; Banzhaf, W.; Gandomi, A.H. Evolutionary machine learning: A survey. ACM Comput. Surv. (CSUR) 2021, 54, 1–35. [Google Scholar] [CrossRef]
  121. Batista, J.E.; Silva, S. Evolutionary Machine Learning in Environmental Science. Handb. Evol. Mach. Learn. 2023, 563–590. [Google Scholar] [CrossRef]
  122. Ferreira, C. Gene Expression Programming: A New Adaptive Algorithm for Solving Problems. Complex Syst. 2001, 13, 87–129. [Google Scholar]
  123. Gao, W.; Moayedi, H.; Shahsavar, A. The feasibility of genetic programming and ANFIS in prediction energetic performance of a building integrated photovoltaic thermal (BIPVT) system. Sol. Energy 2019, 183, 293–305. [Google Scholar] [CrossRef]
  124. Nadiri, A.A.; Gharekhani, M.; Khatibi, R.; Sadeghfam, S.; Moghaddam, A.A. Groundwater vulnerability indices conditioned by supervised intelligence committee machine (SICM). Sci. Total Environ. 2017, 574, 691–706. [Google Scholar] [CrossRef]
  125. Meray, A.; Wang, L.; Kurihana, T.; Mastilovic, I.; Praveen, S.; Xu, Z.; Memarzadeh, M.; Lavin, A.; Wainwright, H. Physics-informed surrogate modeling for supporting climate resilience at groundwater contamination sites. Comput. Geosci. 2024, 183, 105508. [Google Scholar] [CrossRef]
  126. Ghahramani, Z. Unsupervised learning. In Summer School on Machine Learning; Springer: Berlin/Heidelberg, Germany, 2003; pp. 72–112. [Google Scholar]
  127. Makhzani, A.; Frey, B.J. Winner-take-all autoencoders. Adv. Neural Inf. Process. Syst. 2015, 28, 2791–2799. [Google Scholar]
  128. Narvaez-Montoya, C.; Mahlknecht, J.; Torres-Martínez, J.A.; Mora, A.; Bertrand, G. Seawater intrusion pattern recognition supported by unsupervised learning: A systematic review and application. Sci. Total Environ. 2023, 864, 160933. [Google Scholar] [CrossRef]
  129. Drogkoula, M.; Kokkinos, K.; Samaras, N. A comprehensive survey of machine learning methodologies with emphasis in water resources management. Appl. Sci. 2023, 13, 12147. [Google Scholar] [CrossRef]
  130. Kalteh, A.M.; Hjorth, P.; Berndtsson, R. Review of the self-organizing map (SOM) approach in water resources: Analysis, modelling and application. Environ. Model. Softw. 2008, 23, 835–845. [Google Scholar] [CrossRef]
  131. Das, M.; Kumar, A.; Mohapatra, M.; Muduli, S. Evaluation of drinking quality of groundwater through multivariate techniques in urban area. Environ. Monit. Assess. 2010, 166, 149–157. [Google Scholar] [CrossRef]
  132. Kohonen, T. Self-organized formation of topologically correct feature maps. Biol. Cybern. 1982, 43, 59–69. [Google Scholar] [CrossRef]
  133. Kohonen, T. Essentials of the self-organizing map. Neural Netw. 2013, 37, 52–65. [Google Scholar] [CrossRef]
  134. Choi, B.-Y.; Yun, S.-T.; Kim, K.-H.; Kim, J.-W.; Kim, H.M.; Koh, Y.-K. Hydrogeochemical interpretation of South Korean groundwater monitoring data using Self-Organizing Maps. J. Geochem. Explor. 2014, 137, 73–84. [Google Scholar] [CrossRef]
  135. Faal, F.; Nikoo, M.R.; Ashrafi, S.M.; Šimůnek, J. Advancing aquifer vulnerability mapping through integrated deep learning approaches. J. Clean. Prod. 2024, 481, 144112. [Google Scholar] [CrossRef]
  136. Fisher, A.; Rudin, C.; Dominici, F. All models are wrong, but many are useful: Learning a variable’s importance by studying an entire class of prediction models simultaneously. J. Mach. Learn. Res. 2019, 20, 1–81. [Google Scholar]
  137. Li, J.; Cheng, K.; Wang, S.; Morstatter, F.; Trevino, R.P.; Tang, J.; Liu, H. Feature selection: A data perspective. ACM Comput. Surv. (CSUR) 2017, 50, 94. [Google Scholar] [CrossRef]
  138. Stone, M. Cross-validatory choice and assessment of statistical predictions. J. R. Stat. Soc. Ser. B (Methodol.) 1974, 36, 111–133. [Google Scholar] [CrossRef]
  139. Ludwig, M.; Moreno-Martinez, A.; Hölzel, N.; Pebesma, E.; Meyer, H. Assessing and improving the transferability of current global spatial prediction models. Glob. Ecol. Biogeogr. 2023, 32, 356–368. [Google Scholar] [CrossRef]
  140. Wenger, S.J.; Olden, J.D. Assessing transferability of ecological models: An underappreciated aspect of statistical validation. Methods Ecol. Evol. 2012, 3, 260–267. [Google Scholar] [CrossRef]
  141. Montesinos López, O.A.; Montesinos López, A.; Crossa, J. Overfitting, model tuning, and evaluation of prediction performance. In Multivariate Statistical Machine Learning Methods for Genomic Prediction; Springer: Berlin/Heidelberg, Germany, 2022; pp. 109–139. [Google Scholar]
  142. Dwivedi, R.; Dave, D.; Naik, H.; Singhal, S.; Omer, R.; Patel, P.; Qian, B.; Wen, Z.; Shah, T.; Morgan, G. Explainable AI (XAI): Core ideas, techniques, and solutions. ACM Comput. Surv. 2023, 55, 194. [Google Scholar] [CrossRef]
  143. Al-Jarrah, O.Y.; Yoo, P.D.; Muhaidat, S.; Karagiannidis, G.K.; Taha, K. Efficient machine learning for big data: A review. Big Data Res. 2015, 2, 87–93. [Google Scholar] [CrossRef]
  144. Khan, I.; Ayaz, M. Sensitivity analysis-driven machine learning approach for groundwater quality prediction: Insights from integrating ENTROPY and CRITIC methods. Groundw. Sustain. Dev. 2024, 26, 101309. [Google Scholar] [CrossRef]
  145. Esmaeilbeiki, F.; Nikpour, M.R.; Singh, V.K.; Kisi, O.; Sihag, P.; Sanikhani, H. Exploring the application of soft computing techniques for spatial evaluation of groundwater quality variables. J. Clean. Prod. 2020, 276, 124206. [Google Scholar] [CrossRef]
  146. Kerketta, A.; Kapoor, H.S.; Sahoo, P.K. Groundwater fluoride prediction modeling using physicochemical parameters in Punjab, India: A machine-learning approach. Front. Soil Sci. 2024, 4, 1407502. [Google Scholar] [CrossRef]
  147. Boudibi, S.; Fadlaoui, H.; Hiouani, F.; Bouzidi, N.; Aissaoui, A.; Khomri, Z.-E. Groundwater salinity modeling and mapping using machine learning approaches: A case study in Sidi Okba region, Algeria. Environ. Sci. Pollut. Res. 2024, 31, 48955–48971. [Google Scholar] [CrossRef] [PubMed]
  148. Khiavi, A.N.; Tavoosi, M.; Kuriqi, A. Conjunct application of machine learning and game theory in groundwater quality mapping. Environ. Earth Sci. 2023, 82, 395. [Google Scholar] [CrossRef]
  149. Javidan, R.; Javidan, N. A novel artificial intelligence-based approach for mapping groundwater nitrate pollution in the Andimeshk-Dezful plain, Iran. Geocarto Int. 2022, 37, 10434–10458. [Google Scholar] [CrossRef]
  150. Tavakoli, M.; Motlagh, Z.K.; Sayadi, M.H.; Ibraheem, I.M.; Youssef, Y.M. Sustainable Groundwater management using machine learning-based DRASTIC model in rurbanizing riverine region: A case study of Kerman Province, Iran. Water 2024, 16, 2748. [Google Scholar] [CrossRef]
  151. Chahid, M.; El Messari, J.E.S.; Hilal, I.; Aqnouy, M. Application of the DRASTIC-LU/LC method combined with machine learning models to assess and predict the vulnerability of the Rmel aquifer (Northwest, Morocco). Groundw. Sustain. Dev. 2024, 27, 101345. [Google Scholar] [CrossRef]
  152. El-Rawy, M.; Wahba, M.; Fathi, H.; Alshehri, F.; Abdalla, F.; El Attar, R.M. Assessment of groundwater quality in arid regions utilizing principal component analysis, GIS, and machine learning techniques. Mar. Pollut. Bull. 2024, 205, 116645. [Google Scholar] [CrossRef]
  153. Liang, Y.; Zhang, X.; Sun, Y.; Yao, L.; Gan, L.; Wu, J.; Chen, S.; Li, J.; Wang, J. Enhanced groundwater vulnerability assessment to nitrate contamination in Chongqing, Southwest China: Integrating novel explainable machine learning algorithms with DRASTIC-LU. Hydrol. Res. 2024, 55, 663–682. [Google Scholar] [CrossRef]
  154. Bordbar, M.; Busico, G.; Sirna, M.; Tedesco, D.; Mastrocicco, M. A multi-step approach to evaluate the sustainable use of groundwater resources for human consumption and agriculture. J. Environ. Manag. 2023, 347, 119041. [Google Scholar] [CrossRef]
  155. Bakhtiarizadeh, A.; Najafzadeh, M.; Mohamadi, S. Enhancement of groundwater resources quality prediction by machine learning models on the basis of an improved DRASTIC method. Sci. Rep. 2024, 14, 29933. [Google Scholar] [CrossRef]
  156. Liang, Y.; Zhang, X.; Gan, L.; Chen, S.; Zhao, S.; Ding, J.; Kang, W.; Yang, H. Mapping specific groundwater nitrate concentrations from spatial data using machine learning: A case study of chongqing, China. Heliyon 2024, 10, e27867. [Google Scholar] [CrossRef]
  157. Sarkar, S.; Das, K.; Mukherjee, A. Groundwater salinity across India: Predicting occurrences and controls by field-observations and machine learning modeling. Environ. Sci. Technol. 2024, 58, 3953–3965. [Google Scholar] [CrossRef] [PubMed]
  158. Tran, D.A.; Tsujimura, M.; Ha, N.T.; Van Binh, D.; Dang, T.D.; Doan, Q.-V.; Bui, D.T.; Ngoc, T.A.; Thuc, P.T.B.; Pham, T.D. Evaluating the predictive power of different machine learning algorithms for groundwater salinity prediction of multi-layer coastal aquifers in the Mekong Delta, Vietnam. Ecol. Indic. 2021, 127, 107790. [Google Scholar] [CrossRef]
  159. Huang, P.; Hou, M.; Sun, T.; Xu, H.; Ma, C.; Zhou, A. Sustainable groundwater management in coastal cities: Insights from groundwater potential and vulnerability using ensemble learning and knowledge-driven models. J. Clean. Prod. 2024, 442, 141152. [Google Scholar] [CrossRef]
  160. Aju, C.; Achu, A.; Mohammed, M.P.; Raicy, M.; Gopinath, G.; Reghunath, R. Groundwater quality prediction and risk assessment in Kerala, India: A machine-learning approach. J. Environ. Manag. 2024, 370, 122616. [Google Scholar] [CrossRef]
  161. Barzegar, R.; Razzagh, S.; Quilty, J.; Adamowski, J.; Pour, H.K.; Booij, M.J. Improving GALDIT-based groundwater vulnerability predictive mapping using coupled resampling algorithms and machine learning models. J. Hydrol. 2021, 598, 126370. [Google Scholar] [CrossRef]
  162. Tachi, S.E.; Bouguerra, H.; Djellal, M.; Benaroussi, O.; Belaroui, A.; Łozowski, B.; Augustyniak, M.; Benmamar, S.; Benziada, S.; Woźnica, A. Assessing the Risk of Groundwater Pollution in Northern Algeria through the Evaluation of Influencing Parameters and Ensemble Methods. In Doklady Earth Sciences; Pleiades Publishing: Moscow, Russia, 2023; pp. 1233–1243. [Google Scholar]
  163. Knoll, L.; Breuer, L.; Bach, M. Large scale prediction of groundwater nitrate concentrations from spatial data using machine learning. Sci. Total Environ. 2019, 668, 1317–1327. [Google Scholar] [CrossRef]
  164. Pal, S.C.; Ruidas, D.; Saha, A.; Islam, A.R.M.T.; Chowdhuri, I. Application of novel data-mining technique based nitrate concentration susceptibility prediction approach for coastal aquifers in India. J. Clean. Prod. 2022, 346, 131205. [Google Scholar] [CrossRef]
  165. Sun, X.; Cao, W.; Pan, D.; Li, Y.; Ren, Y.; Nan, T. Assessment of aquifer specific vulnerability to total nitrate contamination using ensemble learning and geochemical evidence. Sci. Total Environ. 2024, 912, 169497. [Google Scholar] [CrossRef]
  166. Bordbar, M.; Khosravi, K.; Murgulet, D.; Tsai, F.T.-C.; Golkarian, A. The use of hybrid machine learning models for improving the GALDIT model for coastal aquifer vulnerability mapping. Environ. Earth Sci. 2022, 81, 402. [Google Scholar] [CrossRef]
  167. Islam, A.R.M.T.; Pal, S.C.; Chowdhuri, I.; Salam, R.; Islam, M.S.; Rahman, M.M.; Zahid, A.; Idris, A.M. Application of novel framework approach for prediction of nitrate concentration susceptibility in coastal multi-aquifers, Bangladesh. Sci. Total Environ. 2021, 801, 149811. [Google Scholar] [CrossRef] [PubMed]
  168. Meng, L.; Yan, Y.; Jing, H.; Baloch, M.Y.J.; Du, S.; Du, S. Large-scale groundwater pollution risk assessment research based on artificial intelligence technology: A case study of Shenyang City in Northeast China. Ecol. Indic. 2024, 169, 112915. [Google Scholar] [CrossRef]
  169. Gholami, V.; Booij, M. Use of machine learning and geographical information system to predict nitrate concentration in an unconfined aquifer in Iran. J. Clean. Prod. 2022, 360, 131847. [Google Scholar] [CrossRef]
  170. Kumar, S.; Pati, J. Machine learning approach for assessment of arsenic levels using physicochemical properties of water, soil, elevation, and land cover. Environ. Monit. Assess. 2023, 195, 641. [Google Scholar] [CrossRef]
  171. Karimanzira, D.; Weis, J.; Wunsch, A.; Ritzau, L.; Liesch, T.; Ohmer, M. Application of machine learning and deep neural networks for spatial prediction of groundwater nitrate concentration to improve land use management practices. Front. Water 2023, 5, 1193142. [Google Scholar] [CrossRef]
  172. Li, X.; Liang, G.; Wang, L.; Yang, Y.; Li, Y.; Li, Z.; He, B.; Wang, G. Identifying the spatial pattern and driving factors of nitrate in groundwater using a novel framework of interpretable stacking ensemble learning. Environ. Geochem. Health 2024, 46, 482. [Google Scholar] [CrossRef]
  173. Cao, W.; Zhang, Z.; Fu, Y.; Zhao, L.; Ren, Y.; Nan, T.; Guo, H. Prediction of arsenic and fluoride in groundwater of the North China Plain using enhanced stacking ensemble learning. Water Res. 2024, 259, 121848. [Google Scholar] [CrossRef]
  174. Moazamnia, M.; Hassanzadeh, Y.; Nadiri, A.A.; Sadeghfam, S. Vulnerability indexing to saltwater intrusion from models at two levels using artificial intelligence multiple model (AIMM). J. Environ. Manag. 2020, 255, 109871. [Google Scholar] [CrossRef]
  175. Bordbar, M.; Neshat, A.; Javadi, S.; Pradhan, B.; Dixon, B.; Paryani, S. Improving the coastal aquifers’ vulnerability assessment using SCMAI ensemble of three machine learning approaches. Nat. Hazards 2022, 110, 1799–1820. [Google Scholar] [CrossRef]
  176. Barzegar, R.; Moghaddam, A.A.; Deo, R.; Fijani, E.; Tziritis, E. Mapping groundwater contamination risk of multiple aquifers using multi-model ensemble of machine learning algorithms. Sci. Total Environ. 2018, 621, 697–712. [Google Scholar] [CrossRef]
  177. Barzegar, R.; Moghaddam, A.A.; Baghban, H. A supervised committee machine artificial intelligent for improving DRASTIC method to assess groundwater contamination risk: A case study from Tabriz plain aquifer, Iran. Stoch. Environ. Res. Risk Assess. 2016, 30, 883–899. [Google Scholar] [CrossRef]
  178. Jafarzadeh, F.; Moghaddam, A.A.; Razzagh, S.; Barzegar, R.; Cloutier, V.; Rosa, E. A meta-ensemble machine learning strategy to assess groundwater holistic vulnerability in coastal aquifers. Groundw. Sustain. Dev. 2024, 26, 101296. [Google Scholar] [CrossRef]
  179. Usman, U.S.; Salh, Y.H.M.; Yan, B.; Namahoro, J.P.; Zeng, Q.; Sallah, I. Fluoride contamination in African groundwater: Predictive modeling using stacking ensemble techniques. Sci. Total Environ. 2024, 957, 177693. [Google Scholar] [CrossRef] [PubMed]
  180. Jamei, M.; Karbasi, M.; Malik, A.; Abualigah, L.; Islam, A.R.M.T.; Yaseen, Z.M. Computational assessment of groundwater salinity distribution within coastal multi-aquifers of Bangladesh. Sci. Rep. 2022, 12, 11165. [Google Scholar] [CrossRef]
  181. Elzain, H.E.; Chung, S.Y.; Park, K.-H.; Senapathi, V.; Sekar, S.; Sabarathinam, C.; Hassan, M. ANFIS-MOA models for the assessment of groundwater contamination vulnerability in a nitrate contaminated area. J. Environ. Manag. 2021, 286, 112162. [Google Scholar] [CrossRef]
  182. De Jesus, K.L.M.; Senoro, D.B.; Dela Cruz, J.C.; Chan, E.B. Neuro-particle swarm optimization based in-situ prediction model for heavy metals concentration in groundwater and surface water. Toxics 2022, 10, 95. [Google Scholar] [CrossRef]
  183. Pham, Q.B.; Tran, D.A.; Ha, N.T.; Islam, A.R.M.T.; Salam, R. Random forest and nature-inspired algorithms for mapping groundwater nitrate concentration in a coastal multi-layer aquifer system. J. Clean. Prod. 2022, 343, 130900. [Google Scholar] [CrossRef]
  184. Saha, A.; Pal, S.C.; Chowdhuri, I.; Roy, P.; Chakrabortty, R. Effect of hydrogeochemical behavior on groundwater resources in Holocene aquifers of moribund Ganges Delta, India: Infusing data-driven algorithms. Environ. Pollut. 2022, 314, 120203. [Google Scholar] [CrossRef]
  185. Wu, R.; Alvareda, E.M.; Polya, D.A.; Blanco, G.; Gamazo, P. Distribution of groundwater arsenic in Uruguay using hybrid machine learning and expert system approaches. Water 2021, 13, 527. [Google Scholar] [CrossRef]
  186. Wu, R.; Podgorski, J.; Berg, M.; Polya, D.A. Geostatistical model of the spatial distribution of arsenic in groundwaters in Gujarat State, India. Environ. Geochem. Health 2021, 43, 2649–2664. [Google Scholar] [CrossRef]
  187. Yang, N.; Winkel, L.H.; Johannesson, K.H. Predicting geogenic arsenic contamination in shallow groundwater of South Louisiana, United States. Environ. Sci. Technol. 2014, 48, 5660–5666. [Google Scholar] [CrossRef] [PubMed]
  188. Nourani, V.; Ghaffari, A.; Behfar, N.; Foroumandi, E.; Zeinali, A.; Ke, C.-Q.; Sankaran, A. Spatiotemporal assessment of groundwater quality and quantity using geostatistical and ensemble artificial intelligence tools. J. Environ. Manag. 2024, 355, 120495. [Google Scholar] [CrossRef] [PubMed]
  189. Taşan, S. Estimation of groundwater quality using an integration of water quality index, artificial intelligence methods and GIS: Case study, Central Mediterranean Region of Turkey. Appl. Water Sci. 2023, 13, 15. [Google Scholar] [CrossRef]
  190. Karandish, F.; Darzi-Naftchali, A.; Asgari, A. Application of machine-learning models for diagnosing health hazard of nitrate toxicity in shallow aquifers. Paddy Water Environ. 2017, 15, 201–215. [Google Scholar] [CrossRef]
  191. Gholami, V.; Khaleghi, M.; Pirasteh, S.; Booij, M.J. Comparison of self-organizing map, artificial neural network, and co-active neuro-fuzzy inference system methods in simulating groundwater quality: Geospatial artificial intelligence. Water Resour. Manag. 2022, 36, 451–469. [Google Scholar] [CrossRef]
  192. Sahour, S.; Khanbeyki, M.; Gholami, V.; Sahour, H.; Kahvazade, I.; Karimi, H. Evaluation of machine learning algorithms for groundwater quality modeling. Environ. Sci. Pollut. Res. 2023, 30, 46004–46021. [Google Scholar] [CrossRef]
  193. Chakraborty, M.; Sarkar, S.; Mukherjee, A.; Shamsudduha, M.; Ahmed, K.M.; Bhattacharya, A.; Mitra, A. Modeling regional-scale groundwater arsenic hazard in the transboundary Ganges River Delta, India and Bangladesh: Infusing physically-based model with machine learning. Sci. Total Environ. 2020, 748, 141107. [Google Scholar] [CrossRef]
  194. Alkindi, K.M.; Mukherjee, K.; Pandey, M.; Arora, A.; Janizadeh, S.; Pham, Q.B.; Anh, D.T.; Ahmadi, K. Prediction of groundwater nitrate concentration in a semiarid region using hybrid Bayesian artificial intelligence approaches. Environ. Sci. Pollut. Res. 2022, 29, 20421–20436. [Google Scholar] [CrossRef]
  195. Uddameri, V.; Silva, A.L.B.; Singaraju, S.; Mohammadi, G.; Hernandez, E.A. Tree-based modeling methods to predict nitrate exceedances in the Ogallala aquifer in Texas. Water 2020, 12, 1023. [Google Scholar] [CrossRef]
  196. Arabgol, R.; Sartaj, M.; Asghari, K. Predicting nitrate concentration and its spatial distribution in groundwater resources using support vector machines (SVMs) model. Environ. Model. Assess. 2016, 21, 71–82. [Google Scholar] [CrossRef]
  197. Rokhshad, A.M.; Khashei Siuki, A.; Yaghoobzadeh, M. Evaluation of a machine-based learning method to estimate the rate of nitrate penetration and groundwater contamination. Arab. J. Geosci. 2021, 14, 40. [Google Scholar] [CrossRef]
  198. Mishra, D.; Chakrabortty, R.; Sen, K.; Pal, S.C.; Mondal, N.K. Groundwater vulnerability assessment of elevated arsenic in Gangetic plain of West Bengal, India; Using primary information, lithological transport, state-of-the-art approaches. J. Contam. Hydrol. 2023, 256, 104195. [Google Scholar] [CrossRef] [PubMed]
  199. Mohammed, M.A.; Kaya, F.; Mohamed, A.; Alarifi, S.S.; Abdelrady, A.; Keshavarzi, A.; Szabó, N.P.; Szűcs, P. Application of GIS-based machine learning algorithms for prediction of irrigational groundwater quality indices. Front. Earth Sci. 2023, 11, 1274142. [Google Scholar] [CrossRef]
  200. Mahboobi, H.; Shakiba, A.; Mirbagheri, B. Improving groundwater nitrate concentration prediction using local ensemble of machine learning models. J. Environ. Manag. 2023, 345, 118782. [Google Scholar] [CrossRef]
  201. Mousavi, M.; Qaderi, F.; Ahmadi, A. Spatial prediction of temporary and permanent hardness concentrations in groundwater based on chemistry parameters by artificial intelligence. Int. J. Environ. Sci. Technol. 2023, 20, 6665–6684. [Google Scholar] [CrossRef]
  202. Mirchooli, F.; Motevalli, A.; Pourghasemi, H.R.; Mohammadi, M.; Bhattacharya, P.; Maghsood, F.F.; Tiefenbacher, J.P. How do data-mining models consider arsenic contamination in sediments and variables importance? Environ. Monit. Assess. 2019, 191, 777. [Google Scholar] [CrossRef]
  203. Rafiei-Sardooi, E.; Azareh, A.; Ghazanfarpour, H.; Parteli, E.J.R.; Faryabi, M.; Barkhori, S. An integrated modeling framework for groundwater contamination risk assessment in arid, data-scarce environments. Acta Geophys. 2024, 73, 1865–1889. [Google Scholar] [CrossRef]
  204. Gharekhani, M.; Nadiri, A.A.; Khatibi, R.; Sadeghfam, S.; Moghaddam, A.A. A study of uncertainties in groundwater vulnerability modelling using Bayesian model averaging (BMA). J. Environ. Manag. 2022, 303, 114168. [Google Scholar] [CrossRef]
  205. Alizadeh, Z.; Mahjouri, N. A spatiotemporal Bayesian maximum entropy-based methodology for dealing with sparse data in revising groundwater quality monitoring networks: The Tehran region experience. Environ. Earth Sci. 2017, 76, 436. [Google Scholar] [CrossRef]
  206. Abba, S.; Yassin, M.A.; Jibril, M.M.; Tawabini, B.; Soupios, P.; Khogali, A.; Shah, S.M.H.; Usman, J.; Aljundi, I.H. Nitrate concentrations tracking from multi-aquifer groundwater vulnerability zones: Insight from machine learning and spatial mapping. Process Saf. Environ. Prot. 2024, 184, 1143–1157. [Google Scholar] [CrossRef]
  207. Shadrin, D.; Nikitin, A.; Tregubova, P.; Terekhova, V.; Jana, R.; Matveev, S.; Pukalchik, M. An automated approach to groundwater quality monitoring—Geospatial mapping based on combined application of Gaussian Process Regression and Bayesian Information Criterion. Water 2021, 13, 400. [Google Scholar] [CrossRef]
  208. Pal, S.C.; Islam, A.R.M.T.; Chakrabortty, R.; Islam, M.S.; Saha, A.; Shit, M. Application of data-mining technique and hydro-chemical data for evaluating vulnerability of groundwater in Indo-Gangetic Plain. J. Environ. Manag. 2022, 318, 115582. [Google Scholar]
  209. Mosavi, A.; Hosseini, F.S.; Choubin, B.; Abdolshahnejad, M.; Gharechaee, H.; Lahijanzadeh, A.; Dineva, A.A. Susceptibility prediction of groundwater hardness using ensemble machine learning models. Water 2020, 12, 2770. [Google Scholar] [CrossRef]
  210. Bentéjac, C.; Csörgő, A.; Martínez-Muñoz, G. A comparative analysis of gradient boosting algorithms. Artif. Intell. Rev. 2021, 54, 1937–1967. [Google Scholar] [CrossRef]
  211. Opitz, D.; Maclin, R. Popular ensemble methods: An empirical study. J. Artif. Intell. Res. 1999, 11, 169–198. [Google Scholar] [CrossRef]
  212. Perović, M.; Šenk, I.; Tarjan, L.; Obradović, V.; Dimkić, M. Machine learning models for predicting the ammonium concentration in alluvial groundwaters. Environ. Model. Assess. 2021, 26, 187–203. [Google Scholar] [CrossRef]
  213. Nosair, A.M.; Shams, M.Y.; AbouElmagd, L.M.; Hassanein, A.E.; Fryar, A.E.; Abu Salem, H.S. Predictive model for progressive salinization in a coastal aquifer using artificial intelligence and hydrogeochemical techniques: A case study of the Nile Delta aquifer, Egypt. Environ. Sci. Pollut. Res. 2022, 29, 9318–9340. [Google Scholar] [CrossRef]
  214. Nadiri, A.A.; Sedghi, Z.; Khatibi, R.; Gharekhani, M. Mapping vulnerability of multiple aquifers using multiple models and fuzzy logic to objectively derive model structures. Sci. Total Environ. 2017, 593, 75–90. [Google Scholar] [CrossRef]
  215. Nadiri, A.A.; Gharekhani, M.; Khatibi, R.; Moghaddam, A.A. Assessment of groundwater vulnerability using supervised committee to combine fuzzy logic models. Environ. Sci. Pollut. Res. 2017, 24, 8562–8577. [Google Scholar] [CrossRef]
  216. Hamamin, D.F.; Nadiri, A.A. Supervised committee fuzzy logic model to assess groundwater intrinsic vulnerability in multiple aquifer systems. Arab. J. Geosci. 2018, 11, 176. [Google Scholar] [CrossRef]
  217. Sahour, S.; Khanbeyki, M.; Gholami, V.; Sahour, H.; Karimi, H.; Mohammadi, M. Particle swarm and grey wolf optimization: Enhancing groundwater quality models through artificial neural networks. Stoch. Environ. Res. Risk Assess. 2024, 38, 993–1007. [Google Scholar] [CrossRef]
  218. Biau, G.; Scornet, E. A random forest guided tour. Test 2016, 25, 197–227. [Google Scholar] [CrossRef]
  219. Antoniadis, A.; Lambert-Lacroix, S.; Poggi, J.-M. Random forests for global sensitivity analysis: A selective review. Reliab. Eng. Syst. Saf. 2021, 206, 107312. [Google Scholar] [CrossRef]
  220. Rodriguez-Galiano, V.F.; Ghimire, B.; Rogan, J.; Chica-Olmo, M.; Rigol-Sanchez, J.P. An assessment of the effectiveness of a random forest classifier for land-cover classification. ISPRS J. Photogramm. Remote Sens. 2012, 67, 93–104. [Google Scholar] [CrossRef]
  221. Wu, L.; Li, J.; Wang, Y.; Meng, Q.; Qin, T.; Chen, W.; Zhang, M.; Liu, T.-Y. R-drop: Regularized dropout for neural networks. Adv. Neural Inf. Process. Syst. 2021, 34, 10890–10905. [Google Scholar]
  222. Molnar, C. Interpretable Machine Learning; Lulu. com: Durham, NC, USA, 2020. [Google Scholar]
  223. Yang, L.; Shami, A. On hyperparameter optimization of machine learning algorithms: Theory and practice. Neurocomputing 2020, 415, 295–316. [Google Scholar] [CrossRef]
  224. Fernández-Delgado, M.; Cernadas, E.; Barro, S.; Amorim, D. Do we need hundreds of classifiers to solve real world classification problems? J. Mach. Learn. Res. 2014, 15, 3133–3181. [Google Scholar]
  225. Li, Q.; Liu, S.-Y.; Yang, X.-S. Influence of initialization on the performance of metaheuristic optimizers. Appl. Soft Comput. 2020, 91, 106193. [Google Scholar] [CrossRef]
  226. Malakar, P.; Balaprakash, P.; Vishwanath, V.; Morozov, V.; Kumaran, K. Benchmarking machine learning methods for performance modeling of scientific applications. In Proceedings of the 2018 IEEE/ACM Performance Modeling, Benchmarking and Simulation of High Performance Computer Systems (PMBS), Dallas, TX, USA, 12 November 2018; pp. 33–44. [Google Scholar]
  227. Akinola, S.O.; Oyabugbe, O.J. Accuracies and training times of data mining classification algorithms: An empirical comparative study. J. Softw. Eng. Appl. 2015, 8, 470–477. [Google Scholar] [CrossRef]
  228. Caruana, R.; Karampatziakis, N.; Yessenalina, A. An empirical evaluation of supervised learning in high dimensions. In Proceedings of the 25th international conference on Machine learning, Helsinki, Finland, 5–9 July 2008; pp. 96–103. [Google Scholar]
  229. Wang, P.; Fan, E.; Wang, P. Comparative analysis of image classification algorithms based on traditional machine learning and deep learning. Pattern Recognit. Lett. 2021, 141, 61–67. [Google Scholar] [CrossRef]
  230. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  231. Araya, D.; Podgorski, J.; Berg, M. Groundwater salinity in the Horn of Africa: Spatial prediction modeling and estimated people at risk. Environ. Int. 2023, 176, 107925. [Google Scholar] [CrossRef] [PubMed]
  232. Scorzato, L. Reliability and Interpretability in Science and Deep Learning. Minds Mach. 2024, 34, 27. [Google Scholar] [CrossRef]
  233. Berhanu, K.G.; Lohani, T.K.; Hatiye, S.D. Spatial and seasonal groundwater quality assessment for drinking suitability using index and machine learning approach. Heliyon 2024, 10, e30362. [Google Scholar] [CrossRef] [PubMed]
  234. Podgorski, J.; Kracht, O.; Araguas-Araguas, L.; Terzer-Wassmuth, S.; Miller, J.; Straub, R.; Kipfer, R.; Berg, M. Groundwater vulnerability to pollution in Africa’s Sahel region. Nat. Sustain. 2024, 7, 558–567. [Google Scholar] [CrossRef]
  235. Zhang, C.; Li, P.; Lun, Z.; Gu, Z.; Li, Z. Unveiling the beneficial effects of N2 as a CO2 impurity on fluid-rock reactions during carbon sequestration in carbonate reservoir aquifers: Challenging the notion of purer is always better. Environ. Sci. Technol. 2024, 58, 22980–22991. [Google Scholar] [CrossRef]
  236. Mehrabi, N.; Morstatter, F.; Saxena, N.; Lerman, K.; Galstyan, A. A survey on bias and fairness in machine learning. ACM Comput. Surv. (CSUR) 2021, 54, 115. [Google Scholar] [CrossRef]
  237. Mehrabi, M.; Nalivan, O.A.; Scaioni, M.; Karvarinasab, M.; Kornejady, A.; Moayedi, H. Spatial mapping of gully erosion susceptibility using an efficient metaheuristic neural network. Environ. Earth Sci. 2023, 82, 459. [Google Scholar] [CrossRef]
  238. Mahanty, B.; Lhamo, P.; Sahoo, N.K. Inconsistency of PCA-based water quality index–Does it reflect the quality? Sci. Total Environ. 2023, 866, 161353. [Google Scholar] [CrossRef]
  239. Kohavi, R. A study of cross-validation and bootstrap for accuracy estimation and model selection. In Proceedings of the IJCAI, Montreal, QC, Canada, 20–25 August 1995; pp. 1137–1145. [Google Scholar]
  240. Agrawal, T. Hyperparameter Optimization in Machine Learning: Make Your Machine Learning and Deep Learning Models More Efficient; Springer: Berlin/Heidelberg, Germany, 2021. [Google Scholar]
  241. Mosavi, A.; Hosseini, F.S.; Choubin, B.; Goodarzi, M.; Dineva, A.A. Groundwater salinity susceptibility mapping using classifier ensemble and Bayesian machine learning models. IEEE Access 2020, 8, 145564–145576. [Google Scholar] [CrossRef]
  242. Sze, V.; Chen, Y.-H.; Yang, T.-J.; Emer, J.S. Efficient processing of deep neural networks: A tutorial and survey. Proc. IEEE 2017, 105, 2295–2329. [Google Scholar] [CrossRef]
  243. Christin, S.; Hervet, É.; Lecomte, N. Applications for deep learning in ecology. Methods Ecol. Evol. 2019, 10, 1632–1644. [Google Scholar] [CrossRef]
  244. Sivakumar, M.; Parthasarathy, S.; Padmapriya, T. A simplified approach for efficiency analysis of machine learning algorithms. PeerJ Comput. Sci. 2024, 10, e2418. [Google Scholar] [CrossRef] [PubMed]
  245. Akaike, H. A new look at the statistical model identification. IEEE Trans. Autom. Control 1974, 19, 716–723. [Google Scholar] [CrossRef]
  246. Schwarz, G. Estimating the dimension of a model. Ann. Stat. 1978, 6, 461–464. [Google Scholar] [CrossRef]
  247. Kuha, J. AIC and BIC: Comparisons of assumptions and performance. Sociol. Methods Res. 2004, 33, 188–229. [Google Scholar] [CrossRef]
  248. Zhang, J.; Yang, Y.; Ding, J. Information criteria for model selection. Wiley Interdiscip. Rev. Comput. Stat. 2023, 15, e1607. [Google Scholar] [CrossRef]
  249. Yates, L.A.; Aandahl, Z.; Richards, S.A.; Brook, B.W. Cross validation for model selection: A review with examples from ecology. Ecol. Monogr. 2023, 93, e1557. [Google Scholar] [CrossRef]
  250. Ying, X. An overview of overfitting and its solutions. In Journal of Physics: Conference Series; IOP Publishing: Bristol, UK, 2019; p. 022022. [Google Scholar]
  251. Rice, L.; Wong, E.; Kolter, Z. Overfitting in adversarially robust deep learning. In Proceedings of the International Conference on Machine Learning, Virtual, 13–18 July 2020; pp. 8093–8104. [Google Scholar]
  252. Jabbar, H.; Khan, R.Z. Methods to avoid over-fitting and under-fitting in supervised machine learning (comparative study). Comput. Sci. Commun. Instrum. Devices 2015, 70, 978–981. [Google Scholar]
  253. Qi, S.; He, M.; Hoang, R.; Zhou, Y.; Namadi, P.; Tom, B.; Sandhu, P.; Bai, Z.; Chung, F.; Ding, Z. Salinity modeling using deep learning with data augmentation and transfer learning. Water 2023, 15, 2482. [Google Scholar] [CrossRef]
  254. Gilpin, L.H.; Bau, D.; Yuan, B.Z.; Bajwa, A.; Specter, M.; Kagal, L. Explaining explanations: An overview of interpretability of machine learning. In Proceedings of the 2018 IEEE 5th International Conference on Data Science and Advanced Analytics (DSAA), Virtual, 13–18 July 2018; pp. 80–89. [Google Scholar]
  255. Holm, E.A. In defense of the black box. Science 2019, 364, 26–27. [Google Scholar] [CrossRef] [PubMed]
  256. Xu, F.; Uszkoreit, H.; Du, Y.; Fan, W.; Zhao, D.; Zhu, J. Explainable AI: A brief survey on history, research areas, approaches and challenges. In Proceedings of the Natural Language Processing and Chinese Computing: 8th CCF International Conference, NLPCC 2019, Dunhuang, China, 9–14 October 2019; pp. 563–574. [Google Scholar]
  257. Saha, S.; Roy, J.; Hembram, T.K.; Pradhan, B.; Dikshit, A.; Abdul Maulud, K.N.; Alamri, A.M. Comparison between deep learning and tree-based machine learning approaches for landslide susceptibility mapping. Water 2021, 13, 2664. [Google Scholar] [CrossRef]
  258. Costache, R.; Arabameri, A.; Blaschke, T.; Pham, Q.B.; Pham, B.T.; Pandey, M.; Arora, A.; Linh, N.T.T.; Costache, I. Flash-flood potential mapping using deep learning, alternating decision trees and data provided by remote sensing sensors. Sensors 2021, 21, 280. [Google Scholar] [CrossRef] [PubMed]
  259. Moayedi, H.; Mehrabi, M.; Bui, D.T.; Pradhan, B.; Foong, L.K. Fuzzy-metaheuristic ensembles for spatial assessment of forest fire susceptibility. J. Environ. Manag. 2020, 260, 109867. [Google Scholar] [CrossRef]
  260. Asadi Nalivan, O.; Mousavi Tayebi, S.A.; Mehrabi, M.; Ghasemieh, H.; Scaioni, M. A hybrid intelligent model for spatial analysis of groundwater potential around Urmia Lake, Iran. Stoch. Environ. Res. Risk Assess. 2023, 37, 1821–1838. [Google Scholar] [CrossRef]
  261. Jenks, G.F. The data model concept in statistical mapping. Int. Yearb. Cartogr. 1967, 7, 186–190. [Google Scholar]
  262. Cantarino, I.; Carrion, M.A.; Goerlich, F.; Martinez Ibañez, V. A ROC analysis-based classification method for landslide susceptibility maps. Landslides 2019, 16, 265–282. [Google Scholar] [CrossRef]
  263. Wu, R.; Polya, D.A. Objective cost-informed cutoff criteria improve the utility of machine learning models of environmental hazards: A case study of groundwater arsenic distribution in India. Expo. Health 2024, 16, 619–631. [Google Scholar] [CrossRef]
  264. Bramm, A.M.; Matrenin, P.V.; Khalyasmaa, A.I. A Review of XAI Methods Applications in Forecasting Runoff and Water Level Hydrological Tasks. Mathematics 2025, 13, 2830. [Google Scholar] [CrossRef]
  265. Schuster, G.T.; Chen, Y.; Feng, S. Review of physics-informed machine-learning inversion of geophysical data. Geophysics 2024, 89, T337–T356. [Google Scholar] [CrossRef]
  266. Bhasme, P.; Vagadiya, J.; Bhatia, U. Enhancing predictive skills in physically-consistent way: Physics informed machine learning for hydrological processes. J. Hydrol. 2022, 615, 128618. [Google Scholar] [CrossRef]
  267. Jayaraman, P.; Nagarajan, K.K.; Partheeban, P.; Krishnamurthy, V. Critical review on water quality analysis using IoT and machine learning models. Int. J. Inf. Manag. Data Insights 2024, 4, 100210. [Google Scholar] [CrossRef]
  268. Singh, Y.; Walingo, T. Smart water quality monitoring with IOT wireless sensor networks. Sensors 2024, 24, 2871. [Google Scholar] [CrossRef]
  269. Weiss, K.; Khoshgoftaar, T.M.; Wang, D. A survey of transfer learning. J. Big Data 2016, 3, 9. [Google Scholar] [CrossRef]
  270. Yang, Q.; Liu, Y.; Chen, T.; Tong, Y. Federated machine learning: Concept and applications. ACM Trans. Intell. Syst. Technol. (TIST) 2019, 10, 12. [Google Scholar] [CrossRef]
  271. Chang, F.-J.; Kao, L.-s.; Kuo, Y.-M.; Liu, C.-W. Artificial neural networks for estimating regional arsenic concentrations in a blackfoot disease area in Taiwan. J. Hydrol. 2010, 388, 65–76. [Google Scholar] [CrossRef]
  272. Al-Mahallawi, K.; Mania, J.; Hani, A.; Shahrour, I. Using of neural networks for the prediction of nitrate groundwater contamination in rural and agricultural areas. Environ. Earth Sci. 2012, 65, 917–928. [Google Scholar] [CrossRef]
  273. Nourani, V.; Alami, M.T.; Vousoughi, F.D. Self-organizing map clustering technique for ANN-based spatiotemporal modeling of groundwater quality parameters. J. Hydroinform. 2016, 18, 288–309. [Google Scholar] [CrossRef]
  274. Baghapour, M.A.; Fadaei Nobandegani, A.; Talebbeydokhti, N.; Bagherzadeh, S.; Nadiri, A.A.; Gharekhani, M.; Chitsazan, N. Optimization of DRASTIC method by artificial neural network, nitrate vulnerability index, and composite DRASTIC models to assess groundwater vulnerability for unconfined aquifer of Shiraz Plain, Iran. J. Environ. Health Sci. Eng. 2016, 14, 13. [Google Scholar] [CrossRef]
  275. Wagh, V.; Panaskar, D.; Muley, A.; Mukate, S.; Gaikwad, S. Neural network modelling for nitrate concentration in groundwater of Kadava River basin, Nashik, Maharashtra, India. Groundw. Sustain. Dev. 2018, 7, 436–445. [Google Scholar] [CrossRef]
  276. Nadiri, A.A.; Sedghi, Z.; Khatibi, R.; Sadeghfam, S. Mapping specific vulnerability of multiple confined and unconfined aquifers by using artificial intelligence to learn from multiple DRASTIC frameworks. J. Environ. Manag. 2018, 227, 415–428. [Google Scholar] [CrossRef]
  277. Jang, W.S.; Engel, B.; Yeum, C.M. Integrated environmental modeling for efficient aquifer vulnerability assessment using machine learning. Environ. Model. Softw. 2020, 124, 104602. [Google Scholar] [CrossRef]
  278. Elzain, H.E.; Chung, S.Y.; Senapathi, V.; Sekar, S.; Park, N.; Mahmoud, A.A. Modeling of aquifer vulnerability index using deep learning neural networks coupling with optimization algorithms. Environ. Sci. Pollut. Res. 2021, 28, 57030–57045. [Google Scholar] [CrossRef]
  279. Roostaei, J.; Colley, S.; Mulhern, R.; May, A.A.; Gibson, J.M. Predicting the risk of GenX contamination in private well water using a machine-learned Bayesian network model. J. Hazard. Mater. 2021, 411, 125075. [Google Scholar] [CrossRef]
  280. Elzain, H.E.; Chung, S.Y.; Senapathi, V.; Sekar, S.; Lee, S.Y.; Roy, P.D.; Hassan, A.; Sabarathinam, C. Comparative study of machine learning models for evaluating groundwater vulnerability to nitrate contamination. Ecotoxicol. Environ. Saf. 2022, 229, 113061. [Google Scholar] [CrossRef] [PubMed]
  281. Nadiri, A.A.; Moazamnia, M.; Sadeghfam, S.; Gnanachandrasamy, G.; Venkatramanan, S. Formulating Convolutional Neural Network for mapping total aquifer vulnerability to pollution. Environ. Pollut. 2022, 304, 119208. [Google Scholar] [CrossRef] [PubMed]
  282. Ijlil, S.; Essahlaoui, A.; Mohajane, M.; Essahlaoui, N.; Mili, E.M.; Van Rompaey, A. Machine learning algorithms for modeling and mapping of groundwater pollution risk: A study to reach water security and sustainable development (Sdg) goals in a mediterranean aquifer system. Remote Sens. 2022, 14, 2379. [Google Scholar] [CrossRef]
  283. Ruidas, D.; Pal, S.C.; Saha, A.; Chowdhuri, I.; Shit, M. Hydrogeochemical characterization based water resources vulnerability assessment in India’s first Ramsar site of Chilka lake. Mar. Pollut. Bull. 2022, 184, 114107. [Google Scholar] [CrossRef] [PubMed]
  284. Nadiri, A.A.; Barzegar, R.; Sadeghfam, S.; Rostami, A.A. Developing a Data-Fused Water Quality Index Based on Artificial Intelligence Models to Mitigate Conflicts between GQI and GWQI. Water 2022, 14, 3185. [Google Scholar] [CrossRef]
  285. Akakuru, O.C.; Akaolisa, C.C.Z.; Aigbadon, G.O.; Eyankware, M.O.; Opara, A.I.; Obasi, P.N.; Ofoh, I.J.; Njoku, A.O.; Akudinobi, B.E.B. Integrating machine learning and multi-linear regression modeling approaches in groundwater quality assessment around Obosi, SE Nigeria. Environ. Dev. Sustain. 2023, 25, 14567–14606. [Google Scholar] [CrossRef]
  286. Gani, A.; Singh, M.; Pathak, S.; Hussain, A. Groundwater quality index development using the ANN model of Delhi Metropolitan City, India. Environ. Sci. Pollut. Res. 2023, 32, 7269–7284. [Google Scholar] [CrossRef]
  287. El Yousfi, Y.; Himi, M.; El Ouarghi, H.; Aqnouy, M.; Benyoussef, S.; Gueddari, H.; Ait Hmeid, H.; Alitane, A.; Chaibi, M.; Zahid, M. Assessment and Prediction of the Water Quality Index for the Groundwater of the Ghiss-Nekkor (Al Hoceima, Northeastern Morocco). Sustainability 2022, 15, 402. [Google Scholar] [CrossRef]
  288. Smida, H.; Tarki, M.; Gammoudi, N.; Dassi, L. GIS-based multicriteria and artificial neural network (ANN) investigation for the assessment of groundwater vulnerability and pollution hazard in the Braga shallow aquifer (Central Tunisia): A critical review of generic and modified DRASTIC models. J. Contam. Hydrol. 2023, 259, 104245. [Google Scholar] [CrossRef]
  289. Nadiri, A.A.; Bordbar, M.; Nikoo, M.R.; Silabi, L.S.S.; Senapathi, V.; Xiao, Y. Assessing vulnerability of coastal aquifer to seawater intrusion using Convolutional Neural Network. Mar. Pollut. Bull. 2023, 197, 115669. [Google Scholar] [CrossRef]
  290. Ahn, S.H.; Jeong, D.H.; Kim, M.; Lee, T.K.; Kim, H.-K. Prediction of groundwater quality index to assess suitability for drinking purpose using averaged neural network and geospatial analysis. Ecotoxicol. Environ. Saf. 2023, 265, 115485. [Google Scholar] [CrossRef] [PubMed]
  291. Gaagai, A.; Aouissi, H.A.; Bencedira, S.; Hinge, G.; Athamena, A.; Heddam, S.; Gad, M.; Elsherbiny, O.; Elsayed, S.; Eid, M.H. Application of water quality indices, machine learning approaches, and GIS to identify groundwater quality for irrigation purposes: A case study of Sahara Aquifer, Doucen Plain, Algeria. Water 2023, 15, 289. [Google Scholar] [CrossRef]
  292. Sajib, A.M.; Diganta, M.T.M.; Rahman, A.; Dabrowski, T.; Olbert, A.I.; Uddin, M.G. Developing a novel tool for assessing the groundwater incorporating water quality index and machine learning approach. Groundw. Sustain. Dev. 2023, 23, 101049. [Google Scholar] [CrossRef]
  293. Mohammed, M.; Mohamed, A.; Szabó, N.; Szűcs, P. Development of machine-learning-based models for identifying the sources of nitrate and fluoride in groundwater and predicting their human health risks. Int. J. Energy Water Resour. 2023, 8, 161–180. [Google Scholar] [CrossRef]
  294. Anjum, R.; Ali, S.A.; Siddiqui, M.A. Assessing the impact of land cover on groundwater quality in a smart city using gis and machine learning algorithms. Water Air Soil Pollut. 2023, 234, 182. [Google Scholar] [CrossRef]
  295. Omeka, M.E. Evaluation and prediction of irrigation water quality of an agricultural district, SE Nigeria: An integrated heuristic GIS-based and machine learning approach. Environ. Sci. Pollut. Res. 2023, 31, 54178–54203. [Google Scholar] [CrossRef]
  296. Dhaoui, O.; Antunes, I.M.; Benhenda, I.; Agoubi, B.; Kharroubi, A. Groundwater salinization risk assessment using combined artificial intelligence models. Environ. Sci. Pollut. Res. 2024, 31, 33398–33413. [Google Scholar] [CrossRef]
  297. Thanh, N.N.; Chotpantarat, S.; Ngu, N.H.; Thunyawatcharakul, P.; Kaewdum, N. Integrating machine learning models with cross-validation and bootstrapping for evaluating groundwater quality in Kanchanaburi province, Thailand. Environ. Res. 2024, 252, 118952. [Google Scholar] [CrossRef]
  298. Zhao, Y.; Yang, L.; Pan, H.; Li, Y.; Shao, Y.; Li, J.; Xie, X. Spatio-temporal prediction of groundwater vulnerability based on CNN-LSTM model with self-attention mechanism: A case study in Hetao Plain, northern China. J. Environ. Sci. 2024, 153, 128–142. [Google Scholar] [CrossRef]
  299. Dasgupta, R.; Banerjee, G.; Hidayetullah, S.M.; Saha, N.; Das, S.; Mazumdar, A. A comparative analysis of statistical, MCDM and machine learning based modification strategies to reduce subjective errors of DRASTIC models. Environ. Earth Sci. 2024, 83, 211. [Google Scholar] [CrossRef]
  300. Bordbar, M.; Heggy, E.; Jun, C.; Bateni, S.M.; Kim, D.; Moghaddam, H.K.; Rezaie, F. Comparative study for coastal aquifer vulnerability assessment using deep learning and metaheuristic algorithms. Environ. Sci. Pollut. Res. 2024, 31, 24235–24249. [Google Scholar] [CrossRef]
  301. Mohammadpour, A.; Keshtkar, M.; Samaei, M.R.; Isazadeh, S.; Khaneghah, A.M. Assessing water quality index and health risk using deterministic and probabilistic approaches in Darab County, Iran; A machine learning for fluoride prediction. Chemosphere 2024, 352, 141284. [Google Scholar] [CrossRef]
  302. Gholami, V.; Sahour, H.; Khaleghi, M.R.; Ghajari, Y.E.; Sahour, S. Combining artificial neural networks and genetic algorithms to model nitrate contamination in groundwater. Nat. Hazards 2024, 120, 4789–4809. [Google Scholar] [CrossRef]
  303. Iqbal, J.; Su, C.; Ahmad, M.; Baloch, M.Y.J.; Rashid, A.; Ullah, Z.; Abbas, H.; Nigar, A.; Ali, A.; Ullah, A. Hydrogeochemistry and prediction of arsenic contamination in groundwater of Vehari, Pakistan: Comparison of artificial neural network, random forest and logistic regression models. Environ. Geochem. Health 2024, 46, 14. [Google Scholar] [CrossRef]
  304. Tian, L.; Hu, L.; Wang, D.; Cao, X. Site-scale groundwater pollution risk assessment using surrogate models and statistical analysis. J. Contam. Hydrol. 2024, 261, 104288. [Google Scholar] [CrossRef]
  305. Das, C.R.; Das, S. Coastal groundwater quality prediction using objective-weighted WQI and machine learning approach. Environ. Sci. Pollut. Res. 2024, 31, 19439–19457. [Google Scholar] [CrossRef]
  306. Sakizadeh, M.; Zhang, C.; Milewski, A. Spatial distribution pattern and health risk of groundwater contamination by cadmium, manganese, lead and nitrate in groundwater of an arid area. Environ. Geochem. Health 2024, 46, 80. [Google Scholar] [CrossRef]
  307. Eid, M.H.; Mikita, V.; Eissa, M.; Ramadan, H.S.; Mohamed, E.A.; Abukhadra, M.R.; El-Sherbeeny, A.M.; Kovács, A.; Szűcs, P. An advanced approach for drinking water quality indexing and health risk assessment supported by machine learning modelling in Siwa Oasis, Egypt. J. Hydrol. Reg. Stud. 2024, 56, 101967. [Google Scholar] [CrossRef]
  308. Jacob, C.; Masilamani, U.S. Performance evaluation of different machine learning algorithms for prediction of nitrate in groundwater in Thiruvannamalai District. CLEAN–Soil Air Water 2025, 43, 2400060. [Google Scholar] [CrossRef]
  309. Khan, I.; Umar, R. Machine Learning-driven Optimization of Water Quality Index: A Synergistic ENTROPY-CRITIC Approach Using Spatio-Temporal Data. Earth Syst. Environ. 2024, 8, 1453–1475. [Google Scholar] [CrossRef]
  310. Gad, M.; Gaagai, A.; Agrama, A.A.; El-Fiqy, W.F.; Eid, M.H.; Szűcs, P.; Elsayed, S.; Elsherbiny, O.; Khadr, M.; Abukhadra, M.R. Comprehensive evaluation and prediction of groundwater quality and risk indices using quantitative approaches, multivariate analysis, and machine learning models: An exploratory study. Heliyon 2024, 10, e36606. [Google Scholar] [CrossRef]
  311. Saha, A.; Pal, S.C.; Islam, A.R.M.T.; Islam, A.; Alam, E.; Islam, M.K. Hydro-chemical based assessment of groundwater vulnerability in the Holocene multi-aquifers of Ganges delta. Sci. Rep. 2024, 14, 1265. [Google Scholar] [CrossRef]
  312. Egbueri, J.C.; Abu, M.; Agbasi, J.C. An integrated appraisal of the hydrogeochemistry and the potential public health risks of groundwater nitrate and fluoride in eastern Ghana. Groundw. Sustain. Dev. 2024, 26, 101264. [Google Scholar] [CrossRef]
  313. Best, K.B.; Miro, M.E.; Kirpes, R.M.; Kaynar, N.; Chesler, A.N. Data-driven decision support tools for assessing the vulnerability of community water systems to groundwater contamination in Los Angeles County. Environ. Sci. Policy 2021, 124, 393–400. [Google Scholar] [CrossRef]
  314. Nolan, B.T.; Malone, R.W.; Gronberg, J.A.; Thorp, K.R.; Ma, L. Verifiable metamodels for nitrate losses to drains and groundwater in the Corn Belt, USA. Environ. Sci. Technol. 2012, 46, 901–908. [Google Scholar] [CrossRef]
  315. Gholami, V.; Khaleghi, M.; Sebghati, M. A method of groundwater quality assessment based on fuzzy network-CANFIS and geographic information system (GIS). Appl. Water Sci. 2017, 7, 3633–3647. [Google Scholar] [CrossRef]
  316. Jebastina, N.; Prince Arulraj, G. Spatial prediction of nitrate concentration using GIS and ANFIS modelling in groundwater. Bull. Environ. Contam. Toxicol. 2018, 101, 403–409. [Google Scholar] [CrossRef]
  317. Asghari Moghaddam, A.; Nouri Sangarab, S.; Kadkhodaie Ilkhchi, A. Assessing groundwater vulnerability potential using modified DRASTIC in Ajabshir Plain, NW of Iran. Environ. Monit. Assess. 2023, 195, 497. [Google Scholar] [CrossRef]
  318. Amini, M.; Abbaspour, K.C.; Berg, M.; Winkel, L.; Hug, S.J.; Hoehn, E.; Yang, H.; Johnson, C.A. Statistical modeling of global geogenic arsenic contamination in groundwater. Environ. Sci. Technol. 2008, 42, 3669–3675. [Google Scholar] [CrossRef]
  319. Amini, M.; Mueller, K.; Abbaspour, K.C.; Rosenberg, T.; Afyuni, M.; Møller, K.N.; Sarr, M.; Johnson, C.A. Statistical modeling of global geogenic fluoride contamination in groundwaters. Environ. Sci. Technol. 2008, 42, 3662–3668. [Google Scholar] [CrossRef]
  320. Bindal, S.; Singh, C.K. Predicting groundwater arsenic contamination: Regions at risk in highest populated state of India. Water Res. 2019, 159, 65–76. [Google Scholar] [CrossRef]
  321. Wheeler, D.C.; Nolan, B.T.; Flory, A.R.; DellaValle, C.T.; Ward, M.H. Modeling groundwater nitrate concentrations in private wells in Iowa. Sci. Total Environ. 2015, 536, 481–488. [Google Scholar] [CrossRef]
  322. Rodriguez-Galiano, V.F.; Luque-Espinar, J.A.; Chica-Olmo, M.; Mendes, M.P. Feature selection approaches for predictive modelling of groundwater nitrate pollution: An evaluation of filters, embedded and wrapper methods. Sci. Total Environ. 2018, 624, 661–672. [Google Scholar] [CrossRef]
  323. Khosravi, K.; Sartaj, M.; Tsai, F.T.-C.; Singh, V.P.; Kazakis, N.; Melesse, A.M.; Prakash, I.; Bui, D.T.; Pham, B.T. A comparison study of DRASTIC methods with various objective methods for groundwater vulnerability assessment. Sci. Total Environ. 2018, 642, 1032–1049. [Google Scholar] [CrossRef]
  324. Adnan, S.; Iqbal, J.; Maltamo, M.; Bacha, M.S.; Shahab, A.; Valbuena, R. A simple approach of groundwater quality analysis, classification, and mapping in Peshawar, Pakistan. Environments 2019, 6, 123. [Google Scholar] [CrossRef]
  325. Podgorski, J.; Wu, R.; Chakravorty, B.; Polya, D.A. Groundwater arsenic distribution in India by machine learning geospatial modeling. Int. J. Environ. Res. Public Health 2020, 17, 7119. [Google Scholar] [CrossRef] [PubMed]
  326. Norouzi, H.; Moghaddam, A.A. Groundwater quality assessment using random forest method based on groundwater quality indices (case study: Miandoab plain aquifer, NW of Iran). Arab. J. Geosci. 2020, 13, 912. [Google Scholar] [CrossRef]
  327. Lahjouj, A.; El Hmaidi, A.; Bouhafa, K.; Boufala, M.h. Mapping specific groundwater vulnerability to nitrate using random forest: Case of Sais basin, Morocco. Model. Earth Syst. Environ. 2020, 6, 1451–1466. [Google Scholar] [CrossRef]
  328. Chakrabortty, R.; Roy, P.; Chowdhuri, I.; Pal, S.C. Groundwater vulnerability assessment using random forest approach in a water-stressed paddy cultivated region of West Bengal, India. Groundw. Geochem. Pollut. Remediat. Methods 2021, 392–410. [Google Scholar] [CrossRef]
  329. Norouzi, H.; Moghaddam, A.A.; Celico, F.; Shiri, J. Assessment of groundwater vulnerability using genetic algorithm and random forest methods (case study: Miandoab plain, NW of Iran). Environ. Sci. Pollut. Res. 2021, 28, 39598–39613. [Google Scholar] [CrossRef]
  330. Cardenas-Martinez, A.; Rodriguez-Galiano, V.; Luque-Espinar, J.A.; Mendes, M.P. Predictive modelling benchmark of nitrate Vulnerable Zones at a regional scale based on Machine learning and remote sensing. J. Hydrol. 2021, 603, 127092. [Google Scholar] [CrossRef]
  331. Khan, Q.; Liaqat, M.U.; Mohamed, M.M. A comparative assessment of modeling groundwater vulnerability using DRASTIC method from GIS and a novel classification method using machine learning classifiers. Geocarto Int. 2022, 37, 5832–5850. [Google Scholar] [CrossRef]
  332. Sarkar, S.; Mukherjee, A.; Gupta, S.D.; Bhanja, S.N.; Bhattacharya, A. Predicting regional-scale elevated groundwater nitrate contamination risk using machine learning on natural and human-induced factors. Acs EsT Eng. 2022, 2, 689–702. [Google Scholar] [CrossRef]
  333. Judeh, T.; Almasri, M.N.; Shadeed, S.M.; Bian, H.; Shahrour, I. Use of GIS, statistics and machine learning for groundwater quality management: Application to nitrate contamination. Water Resour. 2022, 49, 503–514. [Google Scholar] [CrossRef]
  334. Sarkar, S.; Mukherjee, A.; Senapati, B.; Duttagupta, S. Predicting potential climate change impacts on groundwater nitrate pollution and risk in an intensely cultivated area of South Asia. ACS Environ. Au 2022, 2, 556–576. [Google Scholar] [CrossRef]
  335. Singha, S.S.; Singha, S.; Pasupuleti, S.; Venkatesh, A. Knowledge-driven and machine learning decision tree-based approach for assessment of geospatial variation of groundwater quality around coal mining regions, Korba district, Central India. Environ. Earth Sci. 2022, 81, 36. [Google Scholar] [CrossRef]
  336. Islam, A.R.M.T.; Pal, S.C.; Chakrabortty, R.; Idris, A.M.; Salam, R.; Islam, M.S.; Zahid, A.; Shahid, S.; Ismail, Z.B. A coupled novel framework for assessing vulnerability of water resources using hydrochemical analysis and data-driven models. J. Clean. Prod. 2022, 336, 130407. [Google Scholar] [CrossRef]
  337. He, S.; Wu, J.; Wang, D.; He, X. Predictive modeling of groundwater nitrate pollution and evaluating its main impact factors using random forest. Chemosphere 2022, 290, 133388. [Google Scholar] [CrossRef]
  338. Aind, D.A.; Malakar, P.; Sarkar, S.; Mukherjee, A. Controls on groundwater fluoride contamination in eastern parts of India: Insights from Unsaturated zone fluoride profiles and AI-based modeling. Water 2022, 14, 3220. [Google Scholar] [CrossRef]
  339. Madani, A.; Hagage, M.; Elbeih, S.F. Random Forest and Logistic Regression algorithms for prediction of groundwater contamination using ammonia concentration. Arab. J. Geosci. 2022, 15, 1619. [Google Scholar] [CrossRef]
  340. Biswas, T.; Pal, S.C.; Chowdhuri, I.; Ruidas, D.; Saha, A.; Islam, A.R.M.T.; Shit, M. Effects of elevated arsenic and nitrate concentrations on groundwater resources in deltaic region of Sundarban Ramsar site, Indo-Bangladesh region. Mar. Pollut. Bull. 2023, 188, 114618. [Google Scholar] [CrossRef]
  341. Goodarzi, M.R.; Niknam, A.R.R.; Barzkar, A.; Niazkar, M.; Zare Mehrjerdi, Y.; Abedi, M.J.; Heydari Pour, M. Water quality index estimations using machine learning algorithms: A case study of Yazd-Ardakan Plain, Iran. Water 2023, 15, 1876. [Google Scholar] [CrossRef]
  342. Elzain, H.E.; Chung, S.Y.; Venkatramanan, S.; Selvam, S.; Ahemd, H.A.; Seo, Y.K.; Bhuyan, M.S.; Yassin, M.A. Novel machine learning algorithms to predict the groundwater vulnerability index to nitrate pollution at two levels of modeling. Chemosphere 2023, 314, 137671. [Google Scholar] [CrossRef]
  343. Motlagh, Z.K.; Derakhshani, R.; Sayadi, M.H. Groundwater vulnerability assessment in central Iran: Integration of GIS-based DRASTIC model and a machine learning approach. Groundw. Sustain. Dev. 2023, 23, 101037. [Google Scholar] [CrossRef]
  344. Thanh, N.N.; Chotpantarat, S.; Ha, N.-T.; Trung, N.H. Determination of conditioning factors for mapping nickel contamination susceptibility in groundwater in Kanchanaburi Province, Thailand, using random forest and maximum entropy. Environ. Geochem. Health 2023, 45, 4583–4602. [Google Scholar] [CrossRef]
  345. Wu, R.; Richards, L.A.; Roshan, A.; Polya, D.A. Artificial Intelligence Modelling to Support the Groundwater Chemistry-Dependent Selection of Groundwater Arsenic Remediation Approaches in Bangladesh. Water 2023, 15, 3539. [Google Scholar] [CrossRef]
  346. Subbarayan, S.; Thiyagarajan, S.; Karuppannan, S.; Panneerselvam, B. Enhancing groundwater vulnerability assessment: Comparative study of three machine learning models and five classification schemes for Cuddalore district. Environ. Res. 2024, 242, 117769. [Google Scholar] [CrossRef]
  347. Chu, Y.; He, B.; He, J.; Zou, H.; Sun, J.; Wen, D. Revealing the drivers and genesis of NO3-N pollution classification in shallow groundwater of the Shaying River Basin by explainable machine learning and pathway analysis method. Sci. Total Environ. 2024, 918, 170742. [Google Scholar] [CrossRef]
  348. Chen, K.; Liu, Q.; Yang, T.; Ju, Q.; Zhu, M. Risk assessment of nitrate groundwater contamination using GIS-based machine learning methods: A case study in the northern Anhui plain, China. J. Contam. Hydrol. 2024, 261, 104300. [Google Scholar] [CrossRef] [PubMed]
  349. Apogba, J.N.; Anornu, G.K.; Koon, A.B.; Dekongmen, B.W.; Sunkari, E.D.; Fynn, O.F.; Kpiebaya, P. Application of machine learning techniques to predict groundwater quality in the Nabogo Basin, Northern Ghana. Heliyon 2024, 10, e28527. [Google Scholar] [CrossRef] [PubMed]
  350. Elzain, H.E.; Abdalla, O.; Ahmed, H.A.; Kacimov, A.; Al-Maktoumi, A.; Al-Higgi, K.; Abdallah, M.; Yassin, M.A.; Senapathi, V. An innovative approach for predicting groundwater TDS using optimized ensemble machine learning algorithms at two levels of modeling strategy. J. Environ. Manag. 2024, 351, 119896. [Google Scholar] [CrossRef]
  351. Raj, A.; Sinha, A.; Singh, A.; Pasupuleti, S. Assessment and prediction of hexavalent chromium vulnerability in groundwater by Geochemical modelling, NOBLES Index and Random Forest Model. Sci. Total Environ. 2024, 906, 167570. [Google Scholar]
  352. Saha, A.; Pal, S.C. Modelling groundwater vulnerability in a vulnerable deltaic coastal region of Sundarban Biosphere Reserve, India. Environ. Geochem. Health 2024, 46, 8. [Google Scholar] [CrossRef]
  353. Khusulio, K.; Sharma, N.R.; Das, I.C.; Setia, R.; Pathak, A.; Kumar, R. Hydrochemical assessment of groundwater with special emphasis on fluoride in parts of Punjab and fluoride prediction using GIS and ML. Environ. Earth Sci. 2024, 83, 570. [Google Scholar] [CrossRef]
  354. Gómez-Escalonilla, V.; Martínez-Santos, P. A Machine Learning Approach to Map the Vulnerability of Groundwater Resources to Agricultural Contamination. Hydrology 2024, 11, 153. [Google Scholar] [CrossRef]
  355. Gómez-Escalonilla, V.; Montero-González, E.; Díaz-Alcaide, S.; Martín-Loeches, M.; del Rosario, M.R.; Martínez-Santos, P. A machine learning approach to site groundwater contamination monitoring wells. Appl. Water Sci. 2024, 14, 250. [Google Scholar] [CrossRef]
  356. Arafa, N.A.; El-Said Salem, Z.; Abdeldayem, A.L.; Ghorab, M.A.; Moustafa, Y.M.; Soliman, S.A.; Farag, M.H.; Purohit, S.; Elhag, M.; Youssef, Y.M. Advancing Deltaic Aquifer Vulnerability Mapping to Seawater Intrusion and Human Impacts in Eastern Nile Delta: Insights from Machine Learning and Hydrochemical Perspective. Earth Syst. Environ. 2024, 9, 2481–2506. [Google Scholar] [CrossRef]
  357. Nandi, R.; Mondal, S.; Mandal, J.; Bhattacharyya, P. From fuzzy-TOPSIS to machine learning: A holistic approach to understanding groundwater fluoride contamination. Sci. Total Environ. 2024, 912, 169323. [Google Scholar] [CrossRef] [PubMed]
  358. Jalali, R.; Tishehzan, P.; Hashemi, H. A machine learning framework for spatio-temporal vulnerability mapping of groundwaters to nitrate in a data scarce region in Lenjanat Plain, Iran. Environ. Sci. Pollut. Res. 2024, 31, 42088–42110. [Google Scholar] [CrossRef]
  359. Raisa, S.S.; Sarkar, S.K.; Sadiq, M.A. Advancing groundwater vulnerability assessment in Bangladesh: A comprehensive machine learning approach. Groundw. Sustain. Dev. 2024, 25, 101128. [Google Scholar] [CrossRef]
  360. Lombard, M.A.; Bryan, M.S.; Jones, D.K.; Bulka, C.; Bradley, P.M.; Backer, L.C.; Focazio, M.J.; Silverman, D.T.; Toccalino, P.; Argos, M. Machine learning models of arsenic in private wells throughout the conterminous United States as a tool for exposure assessment in human health studies. Environ. Sci. Technol. 2021, 55, 5012–5023. [Google Scholar] [CrossRef]
  361. Soriano, M.A., Jr.; Deziel, N.C.; Saiers, J.E. Regional scale assessment of shallow groundwater vulnerability to contamination from unconventional hydrocarbon extraction. Environ. Sci. Technol. 2022, 56, 12126–12136. [Google Scholar] [CrossRef]
  362. Hamlin, Q.; Martin, S.; Kendall, A.; Hyndman, D. Examining relationships between groundwater nitrate concentrations in drinking water and landscape characteristics to understand health risks. GeoHealth 2022, 6, e2021GH000524. [Google Scholar] [CrossRef]
  363. Bulka, C.M.; Bryan, M.S.; Lombard, M.A.; Bartell, S.M.; Jones, D.K.; Bradley, P.M.; Vieira, V.M.; Silverman, D.T.; Focazio, M.; Toccalino, P.L. Arsenic in private well water and birth outcomes in the United States. Environ. Int. 2022, 163, 107176. [Google Scholar] [CrossRef]
  364. Liu, C.; Bartelt-Hunt, S.; Li, Y. Precipitation, temperature, and landcovers drive spatiotemporal variability of groundwater nitrate concentration across the Continental United States. Sci. Total Environ. 2024, 945, 174040. [Google Scholar] [CrossRef]
  365. Tesoriero, A.J.; Wherry, S.A.; Dupuy, D.I.; Johnson, T.D. Predicting redox conditions in groundwater at a national scale using random forest classification. Environ. Sci. Technol. 2024, 58, 5079–5092. [Google Scholar] [CrossRef] [PubMed]
  366. Tesoriero, A.J.; Gronberg, J.A.; Juckem, P.F.; Miller, M.P.; Austin, B.P. Predicting redox-sensitive contaminant concentrations in groundwater using random forest classification. Water Resour. Res. 2017, 53, 7316–7331. [Google Scholar] [CrossRef]
  367. Gharechaee, H.; Samani, A.N.; Sigaroodi, S.K.; Sadeghi, S.M.M.; Sharifitabesh, S.; Mousavi, M.S.; Marcu, M.V.; Hubbart, J.A. Introducing a novel approach for assessment of groundwater salinity hazard, vulnerability, and risk in a semiarid region. Ecol. Inform. 2024, 81, 102647. [Google Scholar] [CrossRef]
  368. Rosecrans, C.Z.; Nolan, B.T.; Gronberg, J.M. Prediction and visualization of redox conditions in the groundwater of Central Valley, California. J. Hydrol. 2017, 546, 341–356. [Google Scholar] [CrossRef]
  369. Canion, A.; McCloud, L.; Dobberfuhl, D. Predictive modeling of elevated groundwater nitrate in a karstic spring-contributing area using random forests and regression-kriging. Environ. Earth Sci. 2019, 78, 271. [Google Scholar] [CrossRef]
  370. Bentekhici, N.; Benkesmia, Y.; Bouhlala, M.A.; Saad, A.; Ghabi, M. Mapping and assessment of groundwater pollution risks in the main aquifer of the Mostaganem plateau (Northwest Algeria): Utilizing the novel vulnerability index and decision tree model. Environ. Sci. Pollut. Res. 2024, 31, 45074–45104. [Google Scholar] [CrossRef]
  371. Jeihouni, M.; Toomanian, A.; Mansourian, A. Decision tree-based data mining and rule induction for identifying high quality groundwater zones to water supply management: A novel hybrid use of data mining and GIS. Water Resour. Manag. 2020, 34, 139–154. [Google Scholar] [CrossRef]
  372. Podgorski, J.; Araya, D.; Berg, M. Geogenic manganese and iron in groundwater of Southeast Asia and Bangladesh–machine learning spatial prediction modeling and comparison with arsenic. Sci. Total Environ. 2022, 833, 155131. [Google Scholar] [CrossRef]
  373. Araya, D.; Podgorski, J.; Kumi, M.; Mainoo, P.A.; Berg, M. Fluoride contamination of groundwater resources in Ghana: Country-wide hazard modeling and estimated population at risk. Water Res. 2022, 212, 118083. [Google Scholar] [CrossRef]
  374. Ling, Y.; Podgorski, J.; Sadiq, M.; Rasheed, H.; Eqani, S.A.M.A.S.; Berg, M. Monitoring and prediction of high fluoride concentrations in groundwater in Pakistan. Sci. Total Environ. 2022, 839, 156058. [Google Scholar] [CrossRef]
  375. Joodavi, A.; Aghlmand, R.; Podgorski, J.; Dehbandi, R.; Abbasi, A. Characterization, geostatistical modeling and health risk assessment of potentially toxic elements in groundwater resources of northeastern Iran. J. Hydrol. Reg. Stud. 2021, 37, 100885. [Google Scholar] [CrossRef]
  376. Podgorski, J.E.; Labhasetwar, P.; Saha, D.; Berg, M. Prediction modeling and mapping of groundwater fluoride contamination throughout India. Environ. Sci. Technol. 2018, 52, 9889–9898. [Google Scholar] [CrossRef]
  377. Sarkar, S.; Mukherjee, A.; Chakraborty, M.; Quamar, M.T.; Duttagupta, S.; Bhattacharya, A. Prediction of elevated groundwater fluoride across India using multi-model approach: Insights on the influence of geologic and environmental factors. Environ. Sci. Pollut. Res. 2023, 30, 31998–32013. [Google Scholar] [CrossRef] [PubMed]
  378. Affum, A.O.; Kwaansa-Ansah, E.E.; Osae, S.D. Estimating groundwater geogenic arsenic contamination and the affected population of river basins underlain mostly with crystalline rocks in Ghana. Environ. Chall. 2024, 15, 100898. [Google Scholar]
  379. Nath, B.; Chowdhury, R.; Ni-Meister, W.; Mahanta, C. Predicting the distribution of arsenic in groundwater by a geospatial machine learning technique in the two most affected districts of Assam, India: The public health implications. GeoHealth 2022, 6, e2021GH000585. [Google Scholar] [CrossRef] [PubMed]
  380. Xiangcao, Z.; Su, C.; Xianjun, X.; Ge, W.; Xiao, Z.; Yang, L.; Pan, H. Employing machine learning to predict the occurrence and spatial variability of high fluoride groundwater in intensively irrigated areas. Appl. Geochem. 2024, 167, 106000. [Google Scholar] [CrossRef]
  381. Knoll, L.; Breuer, L.; Bach, M. Nation-wide estimation of groundwater redox conditions and nitrate concentrations through machine learning. Environ. Res. Lett. 2020, 15, 064004. [Google Scholar] [CrossRef]
  382. Hájek, M.; Jiménez-Alfaro, B.; Hájek, O.; Brancaleoni, L.; Cantonati, M.; Carbognani, M.; Dedić, A.; Dítě, D.; Gerdol, R.; Hájková, P. A European map of groundwater pH and calcium. Earth Syst. Sci. Data Discuss. 2020, 2020, 1–41. [Google Scholar]
  383. Abba, S.I.; Yassin, M.A.; Mubarak, A.S.; Shah, S.M.H.; Usman, J.; Oudah, A.Y.; Naganna, S.R.; Aljundi, I.H. Drinking water resources suitability assessment based on pollution index of groundwater using improved explainable artificial intelligence. Sustainability 2023, 15, 15655. [Google Scholar] [CrossRef]
  384. Ammar, K.; McKee, M.; Kaluarachchi, J. Bayesian method for groundwater quality monitoring network analysis. J. Water Resour. Plan. Manag. 2011, 137, 51–61. [Google Scholar] [CrossRef]
  385. Sajedi-Hosseini, F.; Malekian, A.; Choubin, B.; Rahmati, O.; Cipullo, S.; Coulon, F.; Pradhan, B. A novel machine learning-based approach for the risk assessment of nitrate groundwater contamination. Sci. Total Environ. 2018, 644, 954–962. [Google Scholar] [CrossRef] [PubMed]
  386. Awais, M.; Aslam, B.; Maqsoom, A.; Khalil, U.; Ullah, F.; Azam, S.; Imran, M. Assessing nitrate contamination risks in groundwater: A machine learning approach. Appl. Sci. 2021, 11, 10034. [Google Scholar] [CrossRef]
  387. Wei, A.; Bi, P.; Guo, J.; Lu, S.; Li, D. Modified DRASTIC model for groundwater vulnerability to nitrate contamination in the Dagujia river basin, China. Water Supply 2021, 21, 1793–1805. [Google Scholar] [CrossRef]
  388. Agrawal, P.; Sinha, A.; Kumar, S.; Agarwal, A.; Banerjee, A.; Villuri, V.G.K.; Annavarapu, C.S.R.; Dwivedi, R.; Dera, V.V.R.; Sinha, J. Exploring artificial intelligence techniques for groundwater quality assessment. Water 2021, 13, 1172. [Google Scholar] [CrossRef]
  389. Abu El-Magd, S.A.; Ismael, I.S.; El-Sabri, M.A.S.; Abdo, M.S.; Farhat, H.I. Integrated machine learning–based model and WQI for groundwater quality assessment: ML, geospatial, and hydro-index approaches. Environ. Sci. Pollut. Res. 2023, 30, 53862–53875. [Google Scholar] [CrossRef]
  390. Twarakavi, N.K.; Kaluarachchi, J.J. Sustainability of ground water quality considering land use changes and public health risks. J. Environ. Manag. 2006, 81, 405–419. [Google Scholar] [CrossRef]
  391. Asadi, P.; Hosseini, S.M.; Ataie-Ashtiani, B.; Simmons, C.T. Fuzzy vulnerability mapping of urban groundwater systems to nitrate contamination. Environ. Model. Softw. 2017, 96, 146–157. [Google Scholar] [CrossRef]
  392. Jmal, I.; Ayed, B.; Boughariou, E.; Allouche, N.; Saidi, S.; Hamdi, M.; Bouri, S. Assessing groundwater vulnerability to nitrate pollution using statistical approaches: A case study of Sidi Bouzid shallow aquifer, Central Tunisia. Arab. J. Geosci. 2017, 10, 364. [Google Scholar] [CrossRef]
  393. Rizeei, H.M.; Azeez, O.S.; Pradhan, B.; Khamees, H.H. Assessment of groundwater nitrate contamination hazard in a semi-arid region by using integrated parametric IPNOA and data-driven logistic regression models. Environ. Monit. Assess. 2018, 190, 633. [Google Scholar] [CrossRef]
  394. El Baba, M.; Kayastha, P.; Huysmans, M.; De Smedt, F. Groundwater vulnerability and nitrate contamination assessment and mapping using DRASTIC and geostatistical analysis. Water 2020, 12, 2022. [Google Scholar] [CrossRef]
  395. Adiat, K.; Akeredolu, B.; Akinlalu, A.; Olayanju, G. Application of logistic regression analysis in prediction of groundwater vulnerability in gold mining environment: A case of Ilesa gold mining area, southwestern, Nigeria. Environ. Monit. Assess. 2020, 192, 577. [Google Scholar] [CrossRef]
  396. Jaydhar, A.K.; Pal, S.C.; Saha, A.; Islam, A.R.M.T.; Ruidas, D. Hydrogeochemical evaluation and corresponding health risk from elevated arsenic and fluoride contamination in recurrent coastal multi-aquifers of eastern India. J. Clean. Prod. 2022, 369, 133150. [Google Scholar] [CrossRef]
  397. Mohammaddost, A.; Mohammadi, Z.; Rezaei, M.; Pourghasemi, H.R.; Farahmand, A. Assessment of groundwater vulnerability in an urban area: A comparative study based on DRASTIC, EBF, and LR models. Environ. Sci. Pollut. Res. 2022, 29, 72908–72928. [Google Scholar] [CrossRef] [PubMed]
  398. Jang, C.-S. Geostatistical estimates of groundwater nitrate-nitrogen concentrations with spatial auxiliary information on DRASTIC-LU-based aquifer contamination vulnerability. Environ. Sci. Pollut. Res. 2023, 30, 81113–81130. [Google Scholar] [CrossRef] [PubMed]
  399. Twarakavi, N.K.C.; Kaluarachchi, J.J. Arsenic in the shallow ground waters of conterminous United States: Assessment, health risks, and costs for mcl compliance 1. JAWRA J. Am. Water Resour. Assoc. 2006, 42, 275–294. [Google Scholar] [CrossRef]
  400. Gurdak, J.J.; Qi, S.L. Vulnerability of recently recharged groundwater in principle aquifers of the United States to nitrate contamination. Environ. Sci. Technol. 2012, 46, 6004–6012. [Google Scholar] [CrossRef]
  401. Nolan, B.T.; Hitt, K.J. Vulnerability of shallow groundwater and drinking-water wells to nitrate in the United States. Environ. Sci. Technol. 2006, 40, 7834–7840. [Google Scholar] [CrossRef]
  402. Podgorski, J.E.; Eqani, S.A.M.A.S.; Khanam, T.; Ullah, R.; Shen, H.; Berg, M. Extensive arsenic contamination in high-pH unconfined aquifers in the Indus Valley. Sci. Adv. 2017, 3, e1700935. [Google Scholar] [CrossRef]
  403. LaMotte, A.; Greene, E. Spatial analysis of land use and shallow groundwater vulnerability in the watershed adjacent to Assateague Island National Seashore, Maryland and Virginia, USA. Environ. Geol. 2007, 52, 1413–1421. [Google Scholar] [CrossRef]
  404. Bretzler, A.; Lalanne, F.; Nikiema, J.; Podgorski, J.; Pfenninger, N.; Berg, M.; Schirmer, M. Groundwater arsenic contamination in Burkina Faso, West Africa: Predicting and verifying regions at risk. Sci. Total Environ. 2017, 584, 958–970. [Google Scholar] [CrossRef]
  405. Ayotte, J.D.; Medalie, L.; Qi, S.L.; Backer, L.C.; Nolan, B.T. Estimating the high-arsenic domestic-well population in the conterminous United States. Environ. Sci. Technol. 2017, 51, 12443–12454. [Google Scholar] [CrossRef]
  406. Winkel, L.; Berg, M.; Amini, M.; Hug, S.J.; Annette Johnson, C. Predicting groundwater arsenic contamination in Southeast Asia from surface parameters. Nat. Geosci. 2008, 1, 536–542. [Google Scholar] [CrossRef]
  407. Rodríguez-Lado, L.; Sun, G.; Berg, M.; Zhang, Q.; Xue, H.; Zheng, Q.; Johnson, C.A. Groundwater arsenic contamination throughout China. Science 2013, 341, 866–868. [Google Scholar] [CrossRef]
  408. Pal, S.C.; Biswas, T.; Jaydhar, A.K.; Ruidas, D.; Saha, A.; Chowdhuri, I.; Mandal, S.; Islam, A.; Islam, A.R.M.T.; Pande, C.B. Source identification and potential health risks from elevated groundwater nitrate contamination in Sundarbans coastal aquifers, India. Sci. Rep. 2024, 14, 4153. [Google Scholar] [CrossRef] [PubMed]
  409. Zhang, Q.; Rodriguez-Lado, L.; Liu, J.; Johnson, C.A.; Zheng, Q.; Sun, G. Coupling predicted model of arsenic in groundwater with endemic arsenism occurrence in Shanxi Province, Northern China. J. Hazard. Mater. 2013, 262, 1147–1153. [Google Scholar] [CrossRef] [PubMed]
  410. Atashi Yazdi, S.S.; Motamedvaziri, B.; Hosseini, S.Z.; Ahmadi, H. Reciprocal analysis of groundwater potentiality and vulnerability modeling in the Bahabad Plain, Iran. Environ. Sci. Pollut. Res. 2023, 30, 39586–39604. [Google Scholar] [CrossRef] [PubMed]
  411. Nolan, B.T.; Green, C.T.; Juckem, P.F.; Liao, L.; Reddy, J.E. Metamodeling and mapping of nitrate flux in the unsaturated zone and groundwater, Wisconsin, USA. J. Hydrol. 2018, 559, 428–441. [Google Scholar] [CrossRef]
  412. Ghosh, T. Groundwater hydro-geochemical inferences and eXplainable Artificial Intelligence augmented groundwater quality prediction in arid and semi-arid segment of Rajasthan, India. Groundw. Sustain. Dev. 2024, 26, 101272. [Google Scholar]
  413. Uddin, M.G.; Rana, M.S.P.; Diganta, M.T.M.; Bamal, A.; Sajib, A.M.; Abioui, M.; Shaibur, M.R.; Ashekuzzaman, S.; Nikoo, M.R.; Rahman, A. Enhancing Groundwater Quality Assessment in Coastal Area: A Hybrid Modeling Approach. Heliyon 2024, 10, e33082. [Google Scholar] [CrossRef]
  414. Guo, X.; Xiong, H.; Li, H.; Gui, X.; Hu, X.; Li, Y.; Cui, H.; Qiu, Y.; Zhang, F.; Ma, C. Designing dynamic groundwater management strategies through a composite groundwater vulnerability model: Integrating human-related parameters into the DRASTIC model using LightGBM regression and SHAP analysis. Environ. Res. 2023, 236, 116871. [Google Scholar] [CrossRef]
  415. Yang, S.; Luo, D.; Tan, J.; Li, S.; Song, X.; Xiong, R.; Wang, J.; Ma, C.; Xiong, H. Spatial Mapping and Prediction of Groundwater Quality Using Ensemble Learning Models and SHapley Additive exPlanations with Spatial Uncertainty Analysis. Water 2024, 16, 2375. [Google Scholar] [CrossRef]
  416. Erickson, M.L.; Elliott, S.M.; Brown, C.J.; Stackelberg, P.; Ransom, K.M.; Reddy, J.E. Machine learning predicted redox conditions in the glacial aquifer system, northern continental United States. Water Resour. Res. 2021, 57, e2020WR028207. [Google Scholar] [CrossRef]
  417. Ransom, K.M.; Nolan, B.T.; Stackelberg, P.; Belitz, K.; Fram, M.S. Machine learning predictions of nitrate in groundwater used for drinking supply in the conterminous United States. Sci. Total Environ. 2022, 807, 151065. [Google Scholar] [CrossRef]
  418. Ransom, K.M.; Nolan, B.T.; Traum, J.A.; Faunt, C.C.; Bell, A.M.; Gronberg, J.A.M.; Wheeler, D.C.; Rosecrans, C.Z.; Jurgens, B.; Schwarz, G.E. A hybrid machine learning model to predict and visualize nitrate concentration throughout the Central Valley aquifer, California, USA. Sci. Total Environ. 2017, 601, 1160–1172. [Google Scholar] [CrossRef]
  419. DeSimone, L.A.; Pope, J.P.; Ransom, K.M. Machine-learning models to map pH and redox conditions in groundwater in a layered aquifer system, Northern Atlantic Coastal Plain, eastern USA. J. Hydrol. Reg. Stud. 2020, 30, 100697. [Google Scholar] [CrossRef]
  420. Erickson, M.L.; Elliott, S.M.; Brown, C.J.; Stackelberg, P.E.; Ransom, K.M.; Reddy, J.E.; Cravotta, C.A., III. Machine-learning predictions of high arsenic and high manganese at drinking water depths of the glacial aquifer system, northern continental United States. Environ. Sci. Technol. 2021, 55, 5791–5805. [Google Scholar] [CrossRef] [PubMed]
  421. Knierim, K.J.; Kingsbury, J.A.; Belitz, K.; Stackelberg, P.E.; Minsley, B.J.; Rigby, J.R. Mapped predictions of manganese and arsenic in an alluvial aquifer using boosted regression trees. Groundwater 2022, 60, 362–376. [Google Scholar] [CrossRef] [PubMed]
  422. Stackelberg, P.E.; Belitz, K.; Brown, C.J.; Erickson, M.L.; Elliott, S.M.; Kauffman, L.J.; Ransom, K.M.; Reddy, J.E. Machine learning predictions of pH in the glacial aquifer system, Northern USA. Groundwater 2021, 59, 352–368. [Google Scholar] [CrossRef] [PubMed]
  423. Erickson, M.L.; Elliott, S.M.; Christenson, C.; Krall, A.L. Predicting geogenic arsenic in drinking water wells in glacial aquifers, north-central USA: Accounting for depth-dependent features. Water Resour. Res. 2018, 54, 10172–10187. [Google Scholar] [CrossRef]
  424. DeSimone, L.A.; Ransom, K.M. Manganese in the Northern Atlantic Coastal Plain aquifer system, eastern USA—Modeling regional occurrence with pH, redox, and machine learning. J. Hydrol. Reg. Stud. 2021, 37, 100925. [Google Scholar] [CrossRef]
  425. Attia, M.; Tsai, F.T.-C. Successive bootstrapping deep learning approach and airborne EM-borehole data fusion to understand salt water in the Mississippi River Valley Alluvial Aquifer. Sci. Total Environ. 2024, 932, 172950. [Google Scholar] [CrossRef]
  426. Tokranov, A.K.; Ransom, K.M.; Bexfield, L.M.; Lindsey, B.D.; Watson, E.; Dupuy, D.I.; Stackelberg, P.E.; Fram, M.S.; Voss, S.A.; Kingsbury, J.A. Predictions of groundwater PFAS occurrence at drinking water supply depths in the United States. Science 2024, 386, 748–755. [Google Scholar] [CrossRef]
  427. Meray, A.O.; Sturla, S.; Siddiquee, M.R.; Serata, R.; Uhlemann, S.; Gonzalez-Raymat, H.; Denham, M.; Upadhyay, H.; Lagos, L.E.; Eddy-Dilek, C. PyLEnM: A machine learning framework for long-term groundwater contamination monitoring strategies. Environ. Sci. Technol. 2022, 56, 5973–5983. [Google Scholar] [CrossRef] [PubMed]
  428. Wood, W.W.; Smedley, P.L.; Lindsey, B.D.; Wood, W.T.; Kirchheim, R.E.; Cherry, J.A. Global groundwater solute composition and concentrations. Groundwater 2022, 60, 714–720. [Google Scholar] [CrossRef] [PubMed]
  429. Güler, C.; Kurt, M.A.; Alpaslan, M.; Akbulut, C. Assessment of the impact of anthropogenic activities on the groundwater hydrology and chemistry in Tarsus coastal plain (Mersin, SE Turkey) using fuzzy clustering, multivariate statistics and GIS techniques. J. Hydrol. 2012, 414, 435–451. [Google Scholar] [CrossRef]
  430. Zghibi, A.; Merzougui, A.; Zouhri, L.; Tarhouni, J. Understanding groundwater chemistry using multivariate statistics techniques to the study of contamination in the Korba unconfined aquifer system of Cap-Bon (North-east of Tunisia). J. Afr. Earth Sci. 2014, 89, 1–15. [Google Scholar] [CrossRef]
  431. Machiwal, D.; Jha, M.K. Identifying sources of groundwater contamination in a hard-rock aquifer system using multivariate statistical analyses and GIS-based geostatistical modeling techniques. J. Hydrol. Reg. Stud. 2015, 4, 80–110. [Google Scholar] [CrossRef]
  432. Javadi, S.; Hashemy, S.; Mohammadi, K.; Howard, K.; Neshat, A. Classification of aquifer vulnerability using K-means cluster analysis. J. Hydrol. 2017, 549, 27–37. [Google Scholar] [CrossRef]
  433. Mohammadrezapour, O.; Kisi, O.; Pourahmad, F. Fuzzy c-means and K-means clustering with genetic algorithm for identification of homogeneous regions of groundwater quality. Neural Comput. Appl. 2020, 32, 3763–3775. [Google Scholar] [CrossRef]
  434. Marín Celestino, A.E.; Ramos Leal, J.A.; Martínez Cruz, D.A.; Tuxpan Vargas, J.; De Lara Bashulto, J.; Morán Ramírez, J. Identification of the hydrogeochemical processes and assessment of groundwater quality, using multivariate statistical approaches and water quality index in a wastewater irrigated region. Water 2019, 11, 1702. [Google Scholar] [CrossRef]
  435. Yang, J.; Ye, M.; Tang, Z.; Jiao, T.; Song, X.; Pei, Y.; Liu, H. Using cluster analysis for understanding spatial and temporal patterns and controlling factors of groundwater geochemistry in a regional aquifer. J. Hydrol. 2020, 583, 124594. [Google Scholar] [CrossRef]
  436. Rahmani, B.; Javadi, S.; Shahdany, S.M.H. Evaluation of aquifer vulnerability using PCA technique and various clustering methods. Geocarto Int. 2021, 36, 2117–2140. [Google Scholar] [CrossRef]
  437. Javadi, S.; Hashemy Shahdany, S.M.; Neshat, A.; Chambel, A. Multi-parameter risk mapping of Qazvin aquifer by classic and fuzzy clustering techniques. Geocarto Int. 2022, 37, 1160–1182. [Google Scholar] [CrossRef]
  438. Fashina, L.; Luffman, I.; Joyner, T.A.; Nandi, A. Geospatial assessment of karst spring water quality in Northeast Tennessee, USA. Geosciences 2022, 12, 303. [Google Scholar] [CrossRef]
  439. Nourani, V.; Ghaneei, P.; Kantoush, S.A. Robust clustering for assessing the spatiotemporal variability of groundwater quantity and quality. J. Hydrol. 2022, 604, 127272. [Google Scholar] [CrossRef]
  440. Uzcategui-Salazar, M.; Lillo, J. A new approach to pollution vulnerability assessment in aquifers using K-means analysis. Environ. Earth Sci. 2022, 81, 495. [Google Scholar] [CrossRef]
  441. Wang, W.; Li, W.; Xue, M.; Gu, X.; Ye, C.; Jiao, Y.; Liu, B.; Han, Y.; Tong, Y.; Zhang, X. Spatial-temporal characteristics and influencing factors of lake water and groundwater chemistry in Hulun Lake, Northeast China. Water 2023, 15, 937. [Google Scholar] [CrossRef]
  442. Guo, X.; Yang, Z.; Li, C.; Xiong, H.; Ma, C. Combining the classic vulnerability index and affinity propagation clustering algorithm to assess the intrinsic aquifer vulnerability of coastal aquifers on an integrated scale. Environ. Res. 2023, 217, 114877. [Google Scholar]
  443. Halder, S.; Bhattacharya, S.; Roy, M.B.; Roy, P.K. Application of fuzzy C-means clustering and fuzzy EDAS to assess groundwater irrigation suitability and prioritization for agricultural development in a complex hydrogeological basin. Environ. Sci. Pollut. Res. 2023, 30, 57529–57557. [Google Scholar] [CrossRef]
  444. Eid, M.H.; Eissa, M.; Mohamed, E.A.; Ramadan, H.S.; Czuppon, G.; Kovács, A.; Szűcs, P. Application of stable isotopes, mixing models, and K-means cluster analysis to detect recharge and salinity origins in siwa oasis, Egypt. Groundw. Sustain. Dev. 2024, 25, 101124. [Google Scholar] [CrossRef]
  445. Panagiotou, C.F.; Chekirbane, A.; Eliades, M.; Papoutsa, C.; Akylas, E.; Stylianou, M.; Stathopoulos, N. Assessing the groundwater quality of El Fahs aquifer (NE Tunisia) using multivariate statistical techniques and geostatistical modeling. Appl. Water Sci. 2024, 14, 170. [Google Scholar] [CrossRef]
  446. Page, R.M.; Huggenberger, P.; Lischeid, G. Multivariate analysis of groundwater-quality time-series using self-organizing maps and sammon’s mapping. Water Resour. Manag. 2015, 29, 3957–3970. [Google Scholar]
  447. Yang, X.; Jia, C.; Yang, F.; Yang, H.; Yao, Y. Spatio-temporal variation of groundwater pollution in urban wetlands and management strategies for zoning. J. Environ. Manag. 2023, 342, 118318. [Google Scholar] [CrossRef] [PubMed]
  448. Wu, Z.; Wu, Y.; Yu, Y.; Wang, L.; Qi, P.; Sun, Y.; Fu, Q.; Zhang, G. Assessment of groundwater quality variation characteristics and influencing factors in an intensified agricultural area: An integrated hydrochemical and machine learning approach. J. Environ. Manag. 2024, 371, 123233. [Google Scholar] [CrossRef]
  449. Gholami, V.; Sahour, H. A Hybrid Approach of Supervised Self-organizing Maps and Genetic Algorithms for Predictive Mapping of Arsenic Pollution in Groundwater Resources. Expo. Health 2023, 16, 775–790. [Google Scholar] [CrossRef]
  450. Gholami, V.; Khaleghi, M.R.; Taghvaye Salimi, E. Groundwater quality modeling using self-organizing map (SOM) and geographic information system (GIS) on the Caspian southern coasts. J. Mt. Sci. 2020, 17, 1724–1734. [Google Scholar] [CrossRef]
  451. Rezaei, F.; Ahmadzadeh, M.R.; Safavi, H.R. SOM-DRASTIC: Using self-organizing map for evaluating groundwater potential to pollution. Stoch. Environ. Res. Risk Assess. 2017, 31, 1941–1956. [Google Scholar] [CrossRef]
  452. Lee, K.-J.; Yun, S.-T.; Yu, S.; Kim, K.-H.; Lee, J.-H.; Lee, S.-H. The combined use of self-organizing map technique and fuzzy c-means clustering to evaluate urban groundwater quality in Seoul metropolitan city, South Korea. J. Hydrol. 2019, 569, 685–697. [Google Scholar] [CrossRef]
  453. Nandi, R.; Mondal, S.; Mandal, J.; Bhattacharyya, P. Deciphering geochemical fingerprints and health implications of groundwater fluoride contamination in mica mining regions using machine learning tactics. Environ. Geochem. Health 2024, 46, 400. [Google Scholar] [CrossRef]
Figure 1. A general classification of ML models.
Figure 1. A general classification of ML models.
Water 17 02861 g001
Figure 2. PRISMA diagram illustrating the identification and screening processes to create the dataset of the study (n represents the number of studies).
Figure 2. PRISMA diagram illustrating the identification and screening processes to create the dataset of the study (n represents the number of studies).
Water 17 02861 g002
Figure 3. Classification and frequency of use of ML models in the surveyed literature. Acronyms can be found in the nomenclature. (Format of the included texts: name, frequency).
Figure 3. Classification and frequency of use of ML models in the surveyed literature. Acronyms can be found in the nomenclature. (Format of the included texts: name, frequency).
Water 17 02861 g003
Figure 4. Flowchart for the selection of MLGHM models based on MLs’ relative complexity (here determined as a function of transparency, speed, hyperparameters, and required user’s expertise), MLs’ relative accuracy, and scale of area to be modeled.
Figure 4. Flowchart for the selection of MLGHM models based on MLs’ relative complexity (here determined as a function of transparency, speed, hyperparameters, and required user’s expertise), MLs’ relative accuracy, and scale of area to be modeled.
Water 17 02861 g004
Figure 5. Flowchart for suggested stages in which MLGHM modeling can be improved. Acronyms can be found in the nomenclature.
Figure 5. Flowchart for suggested stages in which MLGHM modeling can be improved. Acronyms can be found in the nomenclature.
Water 17 02861 g005
Table 1. Multi-criteria comparison of selected reviewed MLGHM models.
Table 1. Multi-criteria comparison of selected reviewed MLGHM models.
L.ReferenceTarget ParameterCountryModels
(Most Accurate > Others)
Criteria Considered
Data AdjustmentHyperpa.Comp. EfficiencyOthers/Comments
TSRCVFS/IAOptRepCTOSPF
AKhan and Ayaz [144] WQIIndiaANN > MLR, GP, SVM, RT, BRT ★✖ - Addressing uncertainty
Esmaeilbeiki, et al. [145]H, ECIndiaDENFIS, GEP > GMDH, MARS, M5 - Considering input scenarios
Kerketta, et al. [146]FIndiaELM > XGBoost, SVM, RF, ANN ⃞  ✔✖ - POF by various inherent features
- Using MOAs (GA) for model optimization
Boudibi, et al. [147]Salinity (EC)AlgeriaRF > HyFIS, KNN, Cubist, SVM
Khiavi, et al. [148]WQIIranRF > NB, SVM, KNN ★✔ - Using game theory for model assessment
Javidan and Javidan [149] N O 3 IranCMMRF > RF, MMRF, FMRF - Using cutoff-dependent accuracy metrics
- Briefly discussing transferability
Tavakoli, et al. [150]RI (TDS)IranRF > BRT, SVM, MARS, GLM - Briefly discussing transferability
Chahid, et al. [151]VI ( N O 3 )MoroccoERT > BA, ANN, KNN - Using evolutionary algorithms for model optimization
- POF by various inherent features
El-Rawy, et al. [152]WQIEgyptRR > RF, XGBoost, DT, KNN ★✖❄ - Considering input scenarios
- POF by various inherent features
Liang, et al. [153] N O 3 ChinaRF > XGBoost, SVM O ⃞  - Using SHAP to address explainability
- POF by XGBoost’s inherent features
- Addressing uncertainty by statistical analysis
Bordbar, et al. [154]SWI, WQIItalyRF, SVM > BRT ✖✔ - Briefly discussing transferability
Bakhtiarizadeh, et al. [155] N O 3 IranGEP, GA-PolyR > PSO-MARS, M5 - POF by various techniques
- Discussing explainability
BDíaz-Alcaide and Martínez-Santos [116]Fecal cont. MaliRF > LgR, ANN, SVM, DT, KNN, LDA, NB, AdB, QDA, GBC, GP ⃞  ⃟  - Ensembling the elite models
- Addressing uncertainty
- Briefly discussing transferability and explainability
Liang, et al. [156] N O 3 ChinaGBR > RF, XGBoost, AdB ✔★ ⃞  ✖ - POF by XGBoost’s inherent features
- Addressing uncertainty using PICP and MPI
Sarkar, et al. [157]Salinity (EC)IndiaRF > GBoost, LgR ▲✖
✔★
- Using DSRM to address the issue of imbalanced data
- IHZ using cutoff point
- Addressing transferability by large-scale modeling
Tran, et al. [158]Salinity (Cl−) VietnamCatBoost > RF, XGBoost, LGB ⃞  ⃟  - GHM classified based on WHO standards
- Superior model achieved lowest penalty and fastest training
Huang, et al. [159] N O 3 ChinaXGBoost > CART, GBM, RF, LGBM O✔ - Using SHAP to address explainability
- POF by XGBoost’s inherent features
Aju, et al. [160]WQIIndiaRF > XGBoost, SVR, ANN O - POF by XGBoost’s inherent features
- Using SHAP to address explainability
Barzegar, et al. [161]TDSIranXGBoost > R, LGBM, AdB, CatBoost ⃞  - POF by XGBoost’s inherent regularization
- Addressing reproducibility
Tachi, et al. [162] N O 3 AlgeriaCatBoost > AdB, RF ★✔ - POF by various inherent features
Knoll, et al. [163] N O 3 GermanyRF > BRT, CART, MLR - Using stratified random sampling for data division
- POF by various inherent features
- Briefly discussing transferability
- Addressing transferability by large-scale modeling
Pal, et al. [164] N O 3 IndiaBst > Bag, RF - POF by various inherent features
Sun, et al. [165] N O 3 ChinaBag > Bst, GBDT, RF, SVM, ANN, ★O✔ - POF by various inherent features
- Using SHAP to address explainability
Bordbar, et al. [166]VI (TDS)IranBA-LMT > DA-LMT, BA-NBT, DA-NBT, RS-NBT, RS-LMT, LMT, NBT - POF by various inherent features
Islam, et al. [167] N O 3 BangladeshBst > Bag, RF - POF by various inherent features
CSingha, Pasupuleti, Singha, Singh, and Kumar [46]WQIIndiaDL > XGBoost, ANN, RF ⃞  ✔✖ - POF by Dropout, mini-batch size, and L2 regularization
Meng, et al. [168]RI ( N O 3 )ChinaDL > RF, ANN - Stability analysis by coefficient of variation
- Assessing prediction symmetry by OEE/UEE
Gholami and Booij [169] N O 3 IranXGBoost > DL, MLR ★✔ - POF by XGBoost’s inherent features
Kumar and Pati [170]AsIndiaDL > XGBoost, RF, GBM, DT, NB ⃞ 
Karimanzira, et al. [171] N O 3 GermanyRF > various DLs ▲✔ - Using leave-one-out encoding technique
- POF in DL using Dropout technique
- Addressing uncertainty using PICP and MPIW
- Briefly discussing transferability
- Addressing explainability
- Superior model achieved fastest training
DLi, et al. [172] N O 3 EnglandKNN-SEL > GBDT, XGBoost, RF, ERT, KNN ★O - Ensembling using KNN
- Using SHAP to address explainability
- Addressing stability and generalizability through CV
- Superior model achieved better generalizability and stability
Cao, et al. [173]As, FChinaLDA-SEL > RF, ERT, DT, XGBoost, AdB, GBDT, SVM, KNN, LDA, LgR, ANN - Ensembling using LDA
- Using optimal entropy-weighted to select SEL components
Moazamnia, et al. [174]VI (Salinity)IranSVM-AIMM > ANN, NF - Ensembling using SVM
Nadiri, Gharekhani, Khatibi, Sadeghfam, and Moghaddam [124] N O 3 IranANN-SICM > SVM, NF, GEP - Ensembling using ANN
- Yielding an explicit formula from GEP
- Dividing study area into three parts due to heterogeneity
Bordbar, et al. [175]VI (SWI)IranANN-SCMAI > NF, SVM, ANN - Ensembling using ANN
- Assessing statistical significance of models’ differences by Friedman’s and WSR tests
Fijani, Nadiri, Moghaddam, Tsai, and Dixon [105]VI ( N O 3 )IranANN-SCMAI > ANN, ANFIS, - Ensembling using ANN
- Using MOA (GA) to train the ensemble model
Barzegar, et al. [176]RI ( N O 3 )IranANN-Committee > SVR, M5 Tree, ELM, MARS ⃞ ✔✖ - Ensembling using ANN
- POF by various inherent features
- Briefly discussing explainability
Barzegar, et al. [177]VI ( N O 3 )IranANN-SCMAI > ANN, NF - Ensembling using ANN
- Using MOA (GA) to train the ensemble model
Jafarzadeh, et al. [178] N O 3 , TDSIranAdB-Ens > AdB, XGBoost, LGBM, RF ⃞  - Ensembling using AdB
- POF by various inherent features
- Performing holistic GVM by integrating DRASTIC and GALDIT
Faal, Nikoo, Ashrafi, and Šimůnek [135]Salinity (TDS)IranGP-Ens > SVR-Ens, ANN-Ens, DNN, CNN, DBN, DSA - Ensembling using GP, SVR, ANN
- POF by various inherent features
- Superior model achieved fastest training
Usman, et al. [179]FAfricaNB-SEL > LgR, DT, RF, SVM, NB, KNN, AdB, XGBoost, LGBM, CatBoost, GBDT, BA-DT, ANN, ERT, LDA - Ensembling using NB
- Briefly discussing transferability
- POF by various inherent features
- Addressing transferability by large-scale modeling
EJamei, et al. [180]SalinityBangladeshANFIS-AO > ANFIS-SMA, ANFIS-ACO, LsR, ANFIS, ✔ ⃞  - Addressing uncertainty by U95%
- POF by Monte Carlo technique
- Using MOAs (AO, ACO, SMA) for model optimization
- Considering input scenarios
Elzain, et al. [181]VI ( N O 3 )S. KoreaANFIS-PSO > ANFIS-DE, ANFIS-GA, ANFIS - Using various MOAs for model optimization
- Superior model achieved fastest training (amid MOAs)
- Briefly discussing transferability
De Jesus, et al. [182]Cr, Cd, Fe, Mn, Zn, Ni, Pb, CuPhilippinesPSO-ANN > SVM, MLR ★✖ - Using MOAs (PSO) for model optimization
- Using KGE for model assessment
- POF using early stopping approach
Pham, et al. [183] N O 3 VietnamRF-FA > RF-MOE, RF-Ant, RF-GWO, RF-PSO, RF - Using various MOAs for model optimization
- Briefly discussing transferability
- Superior model achieved fastest convergence (amid MOAs)
Saha, et al. [184]VIIndiaRF-GOA > RF-GWO, RF-PSO, RF ❄✔★ - Using various MOAs for model optimization
- Briefly discussing transferability
Notes: Legend: Heads => L: Label, Ref: Reference, Loc: Location. Data Adjustment => TSR: Testing Splitting Ratios, CV: Data Cross-Validation, FS/IA: Feature Selection/Importance Assessment. Hyperparameters => Hyperpa: Hyperparameters, Opt: Optimization, Rep: Report. Comp. Efficiency => CT: Computation time, OS: Operating System, PF: Penalty Function. Ensemble Models => X-Ens: Ensembled using X model. CV => ✔ K-fold, Not Stated. FS/IA => ✔ Tree-based, ★ Correlation, ✚ Multiple, O SHAP ❄ PCA, ▲ PDP, Others. Opt => ✔ T&E, ⃞  GS, Others/Not Stated. CT => ✔ Quantitatively, ⃟  Qualitatively. PF => ✔ AIC and/BIC, Others.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mehrabi, M.; Polya, D.A.; Han, Y. Machine Learning Models of the Geospatial Distribution of Groundwater Quality: A Systematic Review. Water 2025, 17, 2861. https://doi.org/10.3390/w17192861

AMA Style

Mehrabi M, Polya DA, Han Y. Machine Learning Models of the Geospatial Distribution of Groundwater Quality: A Systematic Review. Water. 2025; 17(19):2861. https://doi.org/10.3390/w17192861

Chicago/Turabian Style

Mehrabi, Mohammad, David A. Polya, and Yang Han. 2025. "Machine Learning Models of the Geospatial Distribution of Groundwater Quality: A Systematic Review" Water 17, no. 19: 2861. https://doi.org/10.3390/w17192861

APA Style

Mehrabi, M., Polya, D. A., & Han, Y. (2025). Machine Learning Models of the Geospatial Distribution of Groundwater Quality: A Systematic Review. Water, 17(19), 2861. https://doi.org/10.3390/w17192861

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop