Next Article in Journal
A Decision-Making Model for Green and Sustainable Remediation of Contaminated Sites Based on CRITIC–Entropy–TOPSIS
Previous Article in Journal
Numerical Investigation of Flow Division at Lateral Diversions
 
 
Due to scheduled maintenance work on our servers, there may be short service disruptions on this website between 11:00 and 12:00 CEST on March 28th.
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Systematic Review

Machine Learning in Surface Mining—A Systematic Review

Associated Laboratory for Energy, Transports and Aeronautics (LAETA)—PROA, Faculty of Engineering, University of Porto, 4200-465 Porto, Portugal
*
Author to whom correspondence should be addressed.
Appl. Sci. 2026, 16(7), 3246; https://doi.org/10.3390/app16073246
Submission received: 12 January 2026 / Revised: 28 February 2026 / Accepted: 17 March 2026 / Published: 27 March 2026
(This article belongs to the Topic Advances in Mining and Geotechnical Engineering)

Abstract

Objective: The objective of this study was to map and critically synthesize empirical evidence on ML/AI applications across surface mining unit operations, and to characterize models, validation practices, and evidence gaps. Eligibility criteria: Our eligibility criteria comprised peer-reviewed studies (2020–2025) applying ML/AI to surface mining activities, training/validating models on empirical datasets, and reporting quantitative performance metrics. Information sources: Scopus, ScienceDirect, Dimensions, and Web of Science were our information sources, last searched December 2025 and supplemented by website and citation snowballing. Risk of bias: Risk of bias was assessed using an adapted domain-based approach based on PROBAST, used to interpret findings without excluding studies. Synthesis method: Our research employed a narrative synthesis (no meta-analysis due to heterogeneity in datasets, algorithms, contexts, and metrics), grouped by application domain. Results: From 5317 records, 57 studies were included, concentrated in blasting (43), followed by load and haul (6), post-dismantling management (4), extraction (2), and overall exploitation (2). Studies predominantly reported statistical metrics (e.g., R2, RMSE, and MAE), with limited operational performance indicators; validation was frequently site-specific. Dataset sizes were not reported consistently across studies. Limitations: This study’s limitations were database coverage, restricted timeframe, and incomplete reporting (e.g., software/tooling). Conclusions: ML/AI shows strong potential, especially in blasting, but scalable deployment is constrained by site specificity, inconsistent reporting, and heterogeneous validation; standardized reporting and operational indicators are priorities. Registration: The systematic review protocol was registered in OSF with DOI 10.17605/OSF.IO/5UMKB. Funding: EU Erasmus+ STRIM project (1010832727).

1. Introduction

The extractive industry plays a crucial role in supplying raw materials and is closely linked to geopolitical dynamics [1], influencing supply chains, industrial development, and resource security.
The extraction of raw materials is essential to modern society [2]. These raw materials are used in a wide range of industries, including automotive [3], steel [4], petrochemical [1], energy [5], construction [6], and agriculture [7]. Raw materials extraction is usually classified into surface and underground mining. There is a greater number of surface mining operations than underground mining operations worldwide [2]. For example, currently, underground mining contributes to 12 to 17% of metal ore production, while surface mining accounts for the remaining 83 to 88% [8].
Mining operations can be divided into key stages, such as mine planning and design, drilling and blasting, and haulage and loading [9], all regulated by environmental regulations, occupational safety requirements, and energy consumption, significant challenges for the extractive industry [10].
In underground mining, specific unit operations vary depending on the mining method employed. The most common and general ones are drilling, charging, blasting, loading, bolting, and cleaning [11].
In surface mining, the primary unit operations consist of drilling, blasting, loading, and hauling [2]. Surface mining is also associated with several environmental issues specific to this type of exploitation, including landscape changes, air pollution, water depletion, rock weathering caused by exposed rocks during excavations, soil vibration, and changes in air pressure [12]. These environmental and operational challenges are expected to intensify as global demand for raw materials continues to rise. This will directly influence the number of exploitations (surface or underground) that continue to increase. As demand grows, the availability of shallow deposits has decreased significantly, forcing the extractive industry to exploit ore bodies located at increased depths. This combination of rising extraction needs and increasingly challenging geological conditions makes the extractive sector highly complex. In dynamic workplaces with diverse hazards, there is an increasing need for tools that enhance prediction, monitoring, and decision-making.
In response to the increasing challenges, the sector is gradually relying on digital technologies and advanced technology methods [13].
Within the broad spectrum of digital and advanced technologies being adopted in the extractive industry, including sensor-based monitoring systems [13], automation, and digital twins [14], machine learning (ML) has emerged as an up-and-coming tool. This technological evolution is connected to Mining 4.0, defined as the integration of Cyber–Physical Systems (CPSs) and the Internet of Things (IoT) to create interconnected, autonomous mining operations. The central pillar of this paradigm is the digital twin, which goes beyond static 3D modeling to a dynamic virtual replica that updates in real time in order to mirror the physical status of the exploitation. In this context, ML algorithms act as the analytical engine, enabling these to transition from passive monitoring to predictive decision-making.
It is used to optimize different types of exploitation processes, such as drilling [15], haulage [16], and geotechnical monitoring, to develop predictive models for high precision and to assist in decision-making.
In recent years, there has been a substantial increase in studies focusing on ML prediction of rock fragmentation [17], blast-induced ground vibration [18], blasting air pressure [19], equipment allocation [20], and geotechnical monitoring [21]. To achieve various objectives, different algorithms are employed, including Support Vector Machines (SVMs), Decision Trees (DTs), Random Forests (RFs), Extra Trees, Gradient Boosting (GB), and hybrid approaches that integrate principal component analysis (PCA) or neural networks. These algorithms have demonstrated the ability to reduce uncertainties and improve operational performance [22].
Although ML applications in surface mining have evolved rapidly, important questions remain unanswered regarding the consistency of reported results, the robustness of the applied methodology, and the availability of empirical validation.
Existing reviews of ML in mining often aggregate surface and underground contexts [22], or focus on narrow tasks (e.g., blasting only) [23], limiting the transferability of conclusions across unit operations. A focused synthesis restricted to surface mining unit operations is needed to characterize validation practices, reporting completeness, and evidence gaps relevant to deployment. This characterization is fundamental to understanding the transition from theoretical models to practical deployment, ensuring that ML tools provide reliable decision support in complex, real-time operational environments
This review aims to map and critically synthesize empirical evidence on ML/AI in surface mining unit operations and to characterize validation practices and evidence gaps.
The guiding research questions addressed by this review are as follows:
(1)
Which ML algorithms demonstrate the highest and most effective performance across unit operations in surface mining?
(2)
How do existing studies evaluate and validate ML models, and how do validation methods affect the reliability of reported results in specific task types?
(3)
What methodological limitations, biases, and evidence gaps create a challenge for the practical use of ML-based decision-making support systems in the mining industry unit operation?
In this review, “surface mining” refers to unit operations conducted from drilling/blasting through loading/hauling and associated on-site operational monitoring/decision support.

2. Methodology

To ensure methodological rigor, transparency, and reproducibility, this study follows the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines [24]. This systematic review is registered in OSF with DOI 10.17605/OSF.IO/5UMKB.

2.1. Search Strategy

This research was carried out in December 2025. Four databases were used in this study: Scopus, ScienceDirect, Dimensions, and Web of Science. Keywords related to the subject were identified and grouped into four keyword groups: technological and geospatial modeling terms; artificial intelligence and learning algorithms; mining operations and application context; and sustainability and environmental dimensions.
While the search strategy aimed for consistency across all platforms, queries were adaptively structured to accommodate the specified syntax, Boolean operator limits, and field availability for each database. For instance, ScienceDirect imposes a limit on the number of Boolean connectors, requiring a simplified string that preserves the core conceptual groups without compromising search sensitivity.
For the Scopus database, the search was conducted through the following query, inserted in Title/Abstract/Keywords:
Title/Abstract/Keywords (“LiDAR” OR “laser scanning” OR “point cloud” OR “simulation” OR “model” OR “planning “) AND (“machine learning” OR “artificial intelligence” OR “deep learning”) AND (“fleet optimization” OR “haul road” OR “equipment scheduling” OR “mining” OR “quarry” OR “open pit”) AND (“sustainability” OR “emissions” OR “environmental”).
For the ScienceDirect database, the search was conducted using the following query, entered in the Title, Abstract, or Author-Specified Keywords fields. Due to the platform’s Boolean operator limit, a simplified query was used, ensuring that at least one representative keyword from each group was used:
Title, abstract or author-specified keywords: (“LiDAR” OR “point cloud” OR “simulation”) AND (“machine learning” OR “artificial intelligence”) AND (“mining” OR “open pit”) AND (“sustainability” OR “environmental”).
For the Dimensions database, the search was conducted through the following query, inserted in Title and Abstract:
Title and Abstract (“LiDAR” OR “laser scanning” OR “point cloud” OR “simulation” OR “model” OR “planning”) AND (“machine learning” OR “artificial intelligence” OR “deep learning”) AND (“fleet optimization” OR “haul road” OR “equipment scheduling” OR “mining” OR “quarry” OR “open pit”) AND (“sustainability” OR “emissions” OR “environmental”).
For the Web of Science database, the search was conducted through the following query, inserted in Topic:
Topic (“LiDAR” OR “laser scanning” OR “point cloud” OR “simulation” OR “model” OR “planning”) AND (“machine learning” OR “artificial intelligence” OR “deep learning”) AND (“fleet optimization” OR “haul road” OR “equipment scheduling” OR “mining” OR “quarry” OR “open pit”) AND (“sustainability” OR “emissions” OR “environmental”).
During the preliminary identification phase, automated computational tools were deployed to screen the dataset effectively, excluding ineligible records prior to the subsequent stages of analysis with the following screening stages: (1) date—only articles published between 2020 and 2025 were considered in the first phase; (2) the type of document was limited to research articles (no systematic reviews nor gray literature was considered); (3) only peer-review journals were considered; and (4) only studies written in English were considered.
All the identified reports were exported from the databases to Zotero (7.0.32) and then imported into Rayyan (https://www.rayyan.ai/ (accessed on 10 December 2025)), facilitating the removal of duplicates and initial screening.
This process was carried out independently by two reviewers, who screened titles/abstracts and full texts; a third reviewer resolved disagreements.

2.2. Eligibility and Exclusion Criteria

To be eligible, each study had to meet strict requirements:
Only peer-reviewed scientific articles that applied ML or AI techniques to surface mining activities published between 2020 and 2025, written in English, were considered.
To be considered eligible, studies were required to develop, train, and validate AI or ML models using empirical datasets.
Studies had to focus directly on exploitation-related activities. Research addressing dust control, land-change dynamics, landscape impacts, or sustainability aspects was included only when these topics were analyzed, action was taken to change the exploitation activities process, and the research contributed evidence on how AI/ML improved operational decision-making or operational outcomes in surface mining.
To ensure comparability across studies, only articles reporting at least one quantitative evaluation metric were included. These metrics are essential for assessing model robustness and validity, as well as the practical impact on mining performance. Where metrics were not directly comparable, results were summarized descriptively, and cross-metric pooling was avoided.
Articles were not excluded from the systematic review if they did not have one of the following types of information: “Software”, “Equipment”, or “Company/Site”.
Studies were excluded if they did not apply ML/AI methods, lacked quantitative performance reporting, focused on underground mining without separate data, relied solely on qualitative descriptions, or provided insufficient methodological detail.
Two reviewers independently screened titles/abstracts and subsequently assessed full texts against the eligibility criteria described above. Discrepancies were resolved by consensus; when unresolved, a third reviewer adjudicated. Screening decisions were recorded in a structured log.

2.3. Data Extraction and Synthesis

Data was extracted using a piloted form. One reviewer extracted data, and a second reviewer verified all extracted fields; disagreements were resolved through discussion.
A Microsoft Excel sheet was used to create detailed tables with the key information from all the articles. The information was divided into four categories:
(1)
General information—author, publication year, and country;
(2)
Site specifications—if it is a quarry or a mine, the type of commodity being exploited, type of unitary operation the study addresses, and company/site;
(3)
Model characteristics—input data, ML model, validation approach, equipment, software, and application scale;
(4)
Methodology and results—implementation protocol, findings, and limitations.
Primary outcomes were defined as R2 (coefficient of determination) and RMSE (root mean square error), or their task-specific equivalents. Secondary outcomes included MAE (mean absolute error), MAPE (mean absolute percentage error), or similar statistical measures.
In the final stage (December 2025), snowballing techniques were applied, including citation tracking and searching, to identify additional records that could be considered eligible.
A comparative synthesis approach was applied to integrate all the extracted data. The studies were grouped according to their application domain, defined as blasting phase, load and haul, post-dismantling management, extraction, and overall exploitation, and were synthesized within each group. Quantitative performance indicators were compared across studies to identify performance patterns and methodological consistency. To ensure consistency, metric labels were standardized, and the best-reported performance under the primary validation setting was extracted. If any information was missing, it was recorded as “not reported” or “NaN”.
Due to the significant heterogeneity of datasets, algorithms, operational contexts, and reported metrics, no meta-analyses were conducted.
Heterogeneity was studied by data size, validation type (cross-validation vs. hold-out vs. external), and model/machine learning usage.
A narrative synthesis was employed to structure the interpretation of results. Study characteristics and performance metrics were tabulated, and distributions were summarized by domain. To systematize all the information using visual representations, the PRISMA flow, MapChart tools (https://www.mapchart.net/world.html (accessed 20 December 2025)), and Microsoft Excel (2602) were used.
The narrative approach allowed the findings to be organized by thematic, methodological trends; application contexts; and comparative performance across different studies, ensuring structured grouping and transparent reporting of the included results.

2.4. Bias Assessment

To ensure methodological integrity, the risk of bias (RoB) was assessed using an adapted version of the Prediction Model Risk of Bias Assessment Tool (PROBAST) [25]. The assessment was performed independently by two reviewers; a third reviewer resolved any disagreements to ensure objectivity and consensus.
Risk of bias was judged per PROBAST domain, using Low Risk (LR), Medium Risk (MR), and High Risk (HR) ratings tailored to the specificities of ML applications in surface mining, evaluating four essential domains: (1) participants (data sources)—representatives of the data, including sample size, data quality, and sampling locations with selection bias; (2) predictors (input variables)—assessed if input variables, such as geomechanical properties, atmospheric conditions, and blasting characteristics, were consistently measured and available at the time; (3) outcome (targets)—verified the definition and measurement of the target variables to ensure they were determined without knowledge of the predictor data; and (4) analysis (algorithmic rigor)—evaluated the transparency of the algorithms, protocols, handling of missing data, and robustness of performance metrics. Additionally, this domain also assessed the risk of selective reporting by checking consistency between the study objectives and the reported results.
An overall RoB judgement followed predefined rules: a study was classified as HR if at least two domains were rated as high; Low Risk if at least three domains were rated as LR and one as MR; and Medium Risk if at least two domains were rated as MR, even in the absence of a HR domain, or if the study presented a mix of risks not meeting the criteria for Low or High Risk.
Studies assessed as HR were not excluded; instead, they were included to ensure a comprehensive overview of the current state of research. Their findings were weighted with caution and integrated through a narrative synthesis. Studies classified as MR were monitored for consistency and influence, while those classified as LR were treated as robust evidence.
Regarding reporting bias, because quantitative pooling was not performed and protocols were rarely available, a formal assessment was not feasible; instead, selective reporting was qualitatively assessed in the analysis domain during the RoB appraisal.
Finally, the certainty of evidence was assessed qualitatively per domain using a structured framework (considering RoB, inconsistency, indirectness, imprecision, and publication bias), and summarized as high, medium, and low.

3. Results

3.1. Research Results

During the first stage of the PRISMA methodology, 5317 articles were identified and filtered using automated tools from the databases. The reasons for exclusion were as follows: (1) 928 were outside the period of reference, (2) 1199 were excluded due to document type, and (3) 179 were excluded for the publishing language. After reading the article’s title and abstract, an additional 2518 articles were excluded because they did not align with the research topic. After this whole process, the articles were uploaded to Rayyan (https://www.rayyan.ai/ (accessed on 10 December 2025)), where 301 duplicates were identified and removed, and 192 articles were moved to the eligibility assessment stage. Six of the 192 articles could not be retrieved after contacting the authors. After a full-text analysis, 56 were removed because they were about waste management and did not demonstrate a direct influence on the exploitation activities of mineral resources, 21 because they were only dust emission predictions without a direct impact on exploitation, and 20 because they were connected to the processing plant. Others were excluded for not being about the type of exploitation we were looking for: 20 for being underground mining, six for being only related to water management, and six for being a mix of underground mining and surface mining without separation of data. Moreover, five were excluded because they presented solely theoretical discussion without practical application or the use of site data, and three were excluded for being systematic reviews.
A total of 49 articles were incorporated into the qualitative synthesis. Different snowballing techniques were employed, allowing for the addition of eight more articles through website searches (three) and citation searches (five). A total of 57 articles covering different uses of ML in unit operations in surface mining were included in this systematic review. The whole process is summarized in Figure 1.

3.2. Studies’ Content Analysis

In Figure 2, it is possible to observe the articles’ geographical distribution; in addition, there is information in the caption about the number of articles by country. The countries with the highest number of articles are China [26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43], Iran [44,45,46,47,48,49,50,51], and India [52,53,54,55,56,57]. The other represented nations were as follows: Vietnam [58,59,60], Japan [61,62,63], Nigeria [64,65,66], Turkey [67,68], Ghana [69,70], the United States of America (USA) [71], the Republic of Korea [72], Greece [73], Germany [74], Ethiopia [18], Egypt [75], Spain [76], Canada [77], Botswana [78], and Australia [79].
Overall, 19 countries are represented in studies on the application of ML or AI in surface exploitation, underscoring the importance of these technologies in the sector.
In terms of publication dates, the number of articles published over time has increased, with these tools being more widely used in recent years.
Of the 57 articles, 27 were related to mines [27,28,30,37,38,39,40,43,44,45,46,48,51,53,54,55,56,57,59,60,62,68,69,74,78,79,80], 23 were related to quarries [18,31,32,33,34,35,36,42,47,49,50,52,58,61,63,65,66,67,70,72,73,75,81], seven did not specify whether the site was a mine or a quarry (being described as “Surface Mining”) [26,29,41,64,76,77], and one involved study data from both a mine and a quarry [71].
The 57 articles were categorized into five groups: blasting phase (43 articles), load and haul (six), post-dismantling management (four), extraction (two), and overall exploitation (two). All the main categories are further divided into subcategories, as shown in Figure 3.
By grouping the literature in this way, it is possible to compare validation strategies and dataset scales across equivalent areas, thereby explaining variability and avoiding inconsistencies arising from different mining phases.
To provide a scientometric context, a keyword co-occurrence network analysis was assessed and is illustrated in Figure 4. The high nodal density of “prediction”, “blasting phase”, “neural network”, and “machine learning” in the central cluster confirms their dominance in the literature. In contrast, the position of “haul” corroborates the lower research volume in this domain compared to blasting and safety monitoring.

3.2.1. Blasting Phase

The blasting phase had the most articles, totaling 43. They were categorized into ML/AI used in the article and the software used.
In total, 165 individual model mentions were identified across the 43 articles in this category. Of those 165, despite a wide variety of hybrid algorithms, four base models had the highest usage: Support Vector Machine (SVM), with 24; Artificial Neural Network (ANN), with 20; Random Forest (RF), with 19; and Extreme Gradient Boosting (XGBoost), with 10. Together, these four account for 73 model applications, representing 44.2% of all the models used in the blasting phase.
Regarding computational tools, 11 different software/programming languages were identified. In total, the articles mentioned 31 software applications, with one article using one or more of these applications. Of the 43 analyzed articles, 30 (69.8%) specified the software used; in 13 (30.2%), the information was not specified. The most widely used software, MATLAB, appeared in 14 articles, followed by Python, with 12.
All information about the models and software used in each article can be found in Table 1.

3.2.2. Load and Haul

The load and haul class was the second class with the most articles, totaling six articles. Focusing mainly on fuel consumption [45,54], driving assistance [43,55], ore blending [30], and truck assistance [72].
Unlike the blasting phase, which is dominated by regression of physical impacts, the load and haul sector shows a significant interest in reinforcement learning and Computer Vision to address the dynamics of truck movements.
A total of 20 different models were used. The four architectures (RF, ANN/MLP, and SVM) continue to have a significant impact, appearing in nine mentions (45% of the models used in this category). Specifically, neural network variants (ANN, MLP, and RBF) are the most frequent, with five applications, followed by RF, with three applications, and SVM, with two applications.
The articles used nine different software packages [33,49,50,76], with two of the articles not specifying the type of software used [46,67]. The most-used software was Python, being used in three different articles, followed by TensorFlow, used in two articles.
All the models and the software used in each article are presented in Table 2.

3.2.3. Post-Dismantling Management

Post-dismantling management included four articles that focused on dust mitigation [37,53,80] and land changes [74].
Across these four articles, diverse models were utilized, with RF [37,53,80] and neural networks [74,80] as the primary architectures. Random Forest variants were applied in three of the four articles, showing a high usage for dust prediction.
For software, four different ones were used across two studies, with the other two not specifying which software they used.
All the detailed information is shown in Table 3.

3.2.4. Extraction

In the “extraction” category, two studies were analyzed, focusing on hydraulic rock-drill fault [26] and continuous excavation [41]. These studies focused on machine monitoring and multi-objective decision-making.
Across these two articles, a total of seven different machine learning models were utilized. While one article used DenseNet, BiLSTM, SVM, and GBDT [26], the other article used DoppelGANger (DG), Contrastive Language-Image Pre-Training (CLIP), and DT [41].
For software, five different software programs were identified and used across the two articles. Python was used in both instances. One of the few Python [26] implementations, E-GCDT, also used Python, in addition to Unity and Visual Studio, for coding within it, and ROS2 Humble Hawksbill [41].

3.2.5. Overall Exploitation

There are two articles classified as overall exploitation that focus on monitoring [29] and decision-making [39].
In the monitoring article, Spatio-Temporal Models (STMs) and Gated Recurrent Unit (GRU) were used. In the decision-making article, the machine learning models used were the Huber Regressor (HR), RF, GBR, SVR, XGBoost Regressor, and CatBoost Regressor (CB-GWO).
As for the software used, the monitoring article used MATLAB, Python, and Unity, while the decision-making software only mentioned Python.

3.3. Training Validation and Results

While the previous section categorized ML applications and software, this section aims to synthesize the validation approaches and training methods used across all reviewed studies.
In the blasting phase, a total of 11 types of training were employed, with one article failing to specify the type of training used. The most-used split for training the ML model was 80% training and 20% testing in 23 articles, and 70% training and 30% testing in 10 articles. These two split types alone account for 76% of the trains.
Dataset sizes in this category vary widely, ranging from small-scale experimental collections of blastings to large-scale operational databases with approximately 3740 observations.
As for the validation, the blasting phase used error magnitude metrics as its evaluation metrics, which can be categorized into four groups: mean absolute error (MAE), mean squared error (MSE), root mean squared error (RMSE), and mean absolute percentage error (MAPE), used for quantifying the deviation between predicted and observed values. Correlations such as Pearson’s Correlation Coefficient (R), Coefficient of Determination (R2), and Adjusted R2 were also used. To evaluate the predictive power, Nash–Sutcliffe Efficiency (NSE), Index of Agreement (IoA), and Variance Accounted For (VAF) were used. Finally, accuracy was scrutinized through the Coefficient of Residual Mass (CRM), Bias Factor, and Scatter Index (SI). The most commonly used metrics were R2 (37 articles), RMSE (33 articles), and MAE (21 articles).
The results obtained reveal a shift toward high-fidelity machine learning architectures that consistently outperform traditional empirical modeling. In Peak Particulate Velocity (PPV) testing, hybrid models such as ELM-HHo and ELM-GOA achieved remarkable statistical reliability, with R2 values of 0.941 for training and 0.9105 for testing in ELM-GOA [32].
In Flyrock and Air Overpressure, substantial improvements were achieved, with models such as WOA or SVM delivering R2 values exceeding 0.98, effectively minimizing the MAE to levels that support critical safety decision-making [31,42,78]. In rock fragmentation, methods such as XGBoost and Extra Trees Regression were highly effective, achieving R2 values over 0.93 [28,57,68].
The complete model’s information can be found in Appendix A.
For the load and haul category, three types of training were applied; one article did not specify any training. This category relies on substantial datasets, including records of up to 400,000 observations for fuel-consumption monitoring. Two of the six articles used an 80% train–20% validation split, while others used reinforcement learning training (RLT), with up to 90,000 iterations, or supervised machine learning (SML).
As for validation, load and haul articles can be categorized into four different groups: statistical reliability, where models such as R2, NSE, MSE, MAE, and RMSE were used; signal and image evaluation, using Peak Signal-to-Noise Ratio (PSNR), Structural Similarity Index (SSIM), Visual Information Fidelity (VIF), and the Universal Quality Index (UQI); real-world applicability, where the Convergence Rate and computation time were assessed; and deviation control, where Distance Class Deviation was also used.
The results show a superior predictive capability of neural networks and reinforcement learning frameworks within the load and haul category. As for fuel consumption, models such as ANN achieve R2 values of 0.989. In ore-blending scheduling, analysis showed that the MADDPG algorithm significantly outperformed DDPG in convergence and precision, minimizing operational deviations. Regarding production and safety, ore production was predicted with high accuracy using SVM, while anticollision systems used RBF networks to provide a precise distance estimation via signal processing. All of this detailed information is in Table 4.
In the post-dismantling management category, three different train splits were used across four articles. Data included high-volume datasets featuring 70,000 paired image patches, as well as sensor-based studies with over 265 h. The most-used splits were 70% train and 30% test; 80% train and 20% test; and 60% train, 30% test, and 10% validation.
As for the validation process, two categories were identified: statistical reliability, including MSE, RMSE, MAE, R, R2, and MAPE; and model performance, evaluated using F1-score and IoU.
In predicting dust mitigation, models such as the probability-based deep learning algorithm and RF achieved high statistical reliability in managing particulate matter and dust dispersion. As for land-change analysis, deep learning architectures utilizing high-volume paired-image datasets and RF models significantly improved the precision of environmental monitoring and estimation in surface mining. All the detailed information can be consulted in Table 5.
In the extraction category, the two articles used two different training models. One of the models utilized cross-validation protocols during training, specifically a “leave-one-operator-out” approach [29], while the other applied offline reinforcement learning, leveraging GAN-enhanced data to augment 18 real trajectories into 155 synthetic samples [41].
Regarding the validation process in these articles, the articles used three different categories: operational efficiency, using indicators such as Full Bucket Rate (FBR); Digging Efficacy (DE); and Digging Time. The stability of the learning process was monitored via Training Loss (MSE) and SD metrics [41].
In the other study, accuracy measures were used, such as Average Accuracy and Average Weighted Accuracy [26].
In predicting hydraulic rock-drill faults, the model based on x-vectors utilizing the Focal Loss function achieved the highest performance across key metrics, such as AA and AWA [26]. In continuous excavation systems, analysis showed that the E-GCDT (DGAN + CLIP + DT) algorithm significantly outperformed both human operators and alternative tested algorithms, demonstrating superior efficiency in autonomous digging tasks [41].
In the overall exploitation, two articles used different types of model learning. Data collection involved high-resolution signals, with 20,000 points per sample for one case, using a 70% train, 15% validation, and 15% test split [32], whereas the other article used an 8-supervised machine learning algorithm [39].
To evaluate predictive accuracy, R2, MAE, and RMSE were used. To ensure the measurement precision, the NMSE was used [39]. In the other articles, metrics such as accuracy and measurement precision were applied [29].
In machine monitoring prediction using digital twin integration, the Gated Recurrent Unit (GRU) model achieved the best performance, achieving a test accuracy of 97.42% and correctly identifying vibrational signs with a precision of around 1 mm [29]. In the sustainable gold mining analysis, the CatBoost algorithm, when optimized with GWo, achieved an R2 of 0.978 and an MAE of 3.361, making it the best-performing model [39].

3.4. Bias Analysis

The PROBAST bias assessment for the blasting phase shows a distribution mainly between “Low Risk” and “Medium Risk”. Most articles in the “Predictors” and “Outcome” domains consistently showed “Low Risk” due to well-defined variables and targets. The most frequent bias was identified in the “Participants” and “Analysis” categories. The “Medium Risk” in “Participants” is often due to the use of single-site datasets. However, recent studies have mitigated this by using data enrichment techniques such as CTGANs and Monte Carlo Simulations. As for “Analysis”, a critical point arises: some studies are categorized as “High Risk”, typically due to limited external validation or a lack of transparency in algorithmic processing. Overall, the category maintains statistical integrity through robust performance metrics despite challenges with site-specific data representativeness. All the details of the blasting phase are shown in Table 6.
The bias assessment in the load and haul category shows a global MR assessment, with four out of the six articles being considered as such. While the “Outcome” and the “Predictors” domains demonstrate a “Low Risk”, indicating well-defined targets and reliable variables, the overall risk is elevated due to the “Participants” and “Analysis” domains. The reliance on single-site datasets led to a “High Risk” rating in the “Participants” domain for 17% of the articles. At the same time, deficiencies in validation protocols resulted in a “High Risk” rating in the Analysis domain for one study. All the detailed information is shown in Table 7.
As for the post-dismantling category, results indicate that environmental monitoring risks are concentrated in specific domains. Global change detection showcases “Low Risk” due to its robust dataset scale and transparency. However, dust monitoring models exhibit a “Medium Risk” trend, with one study classified as “High Risk” in the “Participants” domain, while “Analysis” showed MR to LR ratings. All the detailed information is shown in Table 8.
The “extraction” articles present distributions of “Low Risk” [29] and “Medium Risk” [44], reflecting a high level of methodological rigor across both studies. This moderated risk in the latter is primarily linked to the “Participants” and “Analysis” domains, which are common in reinforcement learning applications. Nonetheless, both articles successfully implemented advanced data enrichment (e.g., GAN-enhanced data) and robust validation techniques to address small-sample constraints, providing a reliable foundation for real-world mining operations [26,41].
In the exploitation category, both analyzed articles are classified as “Medium Risk” overall. Despite this, they demonstrate a high level of methodological rigor, particularly in the “Predictors” and “Outcome” domains, which were consistently rated as “Low Risk”. Prioritizing data-driven characteristics and virtual visualization, through digital twin frameworks and multi-objective optimization, these studies establish a foundation for future practical applications in self-sustaining and sustainable mining operations [29,39].
Overall, the “High Risk” studies in specific domains did not alter the study’s conclusions; instead, they necessitated a more cautious interpretation of certain findings. Confidence remains partially constrained, primarily due to the prevalent reliance on single-site data and the scarcity of independent external validation across the analyzed studies.

3.5. Results Synthesis

The synthesis of the 57 studies included in the present review reveals a significant concentration during the “blasting phase”, accounting for approximately 75% of the total sample.
Methodologically, the literature demonstrates a high degree of standardization in data treatment, as 76% of the “blasting phase” studies employed training/testing splits of 80/20 or 70/30. Across all four categories, the primary machine learning architectures identified were SVM, ANN, RF, and XGBoost. These models consistently delivered high-fidelity performance, frequently reporting R2 values exceeding 0.94 and outperforming traditional empirical formulas.
In contrast, there is heterogeneity in dataset sizes, ranging from small experimental collections of fewer than 100 blasts to industrial databases with over 400.000 records for fuel-consumption monitoring. The RoB assessment indicates that while statistical integrity remains high, the generalizability of findings is limited. Most studies are classified as “Medium Risk” in the “Participants” and “Analysis” domains due to a heavy reliance on site-specific, single-quarry/mine datasets and a lack of independent external validation.

3.6. Reporting Bias

Due to the heterogeneity of the data and metrics, quantitative assessment methods were not feasible. However, a qualitative assessment within the PROBAST “Analysis” domain suggests a risk of reporting bias across the included studies. Several studies focused solely on successful model implementations without reporting negative or null results. Furthermore, the lack of pre-registered protocols in the majority of primary studies prevents a definitive comparison between planned and reported outcomes, suggesting a potential selective reporting bias favoring high-accuracy metrics.

3.7. Certainty of Evidence

According to the qualitative framework defined in the methodology, the certainty of the evidence varies across application domains.
In the “blasting phase”, there is a Moderate Certainty. While the volume of studies is high (43 articles), and statistical metrics are robust, the evidence is downgraded due to the dominance of single-site data and deficiencies in the validation protocol.
In “post-dismantling management”, there is a Low Certainty. Evidence is drawn from only a few studies (four articles). While land change detection benefits from large-scale datasets and a low risk of bias, the certainty for dust mitigation is reduced by an HR rating in the “Participants” section of one study and the prevalence of MR in the “Analysis” domains.
As for the “load and haul” and “extraction”, there is a Low Certainty. The evidence is limited by a small number of studies (six and two articles, respectively) and by imprecision, with datasets varying significantly in size and consistency, ranging from small experimental setups to large operational databases.

4. Discussion

4.1. Analysis by Category

The geographic concertation of research in China, Iran, and India reflects the massive scale of their extractive industries and specific national strategic priorities. These countries are among the world’s leading producers of raw materials, such as coal and iron ore, necessitating high-volume operational optimization to meet global demand. Driving substantial investment in digital technologies and Industry 4.0 is necessary to maintain competitiveness within global supply chains and geopolitical dynamics. However, this concentration suggests that the specific regulatory frameworks, labor costs, and geological conditions of these regions may heavily influence current findings. Consequently, the generalizability of these results to Western or smaller-scale mining contexts should be approached with caution.
The dominance of the “blasting phase” in current research underscores its critical role in the mining value chain. The direct impact of blasting on operational safety and subsequent downstream costs in the production chain can be understood through this phenomenon. In surface mining, the quality of fragmentation directly determines the efficiency of subsequent loading, hauling, and crushing.
Additionally, the high volume of literature in this category can be attributed to the nature of blasting data, which consists of measurable geological parameters and explosive characteristics, leading to high-quality experimental modeling and regression. Some of the articles used data from other authors to test their own models, thereby facilitating the acquisition of new data. The consistently high predictive performance reported in these studies, where R2 values frequently exceed 0.94 for PPV, environmental issues, and rock fragmentation [31,32,57,68,78], validates the transition from traditional empirical formulas to ML architecture as the new gold standard.
While base models, such as SVM, ANN, and RF, remain the most common, accounting for 44.2% of all applications in the blasting category, there is a distinct shift toward hybrid model architecture. Hybrid approaches, such as HHo or GOA, integrated with ELM have demonstrated great statistical reliability [32]. Similarly, techniques such as XGBoost and Extra Trees Regression have proven highly effective for predicting rock fragmentation, demonstrating that tree-based methods are remarkably robust against the noise inherent in site-specific blasting data [28,57,68]. This shift toward ensemble and meta-heuristic optimization reflects the industry’s need for models that can better handle the nonlinear complexities of rock–explosive interactions than standalone algorithms.
The contrast between blasting phase (43 articles) and load and haul (6 articles) appears to reflect a combination of operational prioritization and data complexity rather than lack of industrial need. While blasting is the primary lever in “Mine-to-Mill” [82], the scarcity of load and haul studies highlights a significant research gap. In this context, load and haul requires handling dynamic, real-time telemetry, and multi-agent interactions. Consequently, the limited number of studies suggests that while the industry prioritizes the root cause of efficiency, the complexity of modeling dynamic fleet behavior remains an under-addressed challenge.
Unlike blasting, which focuses on physical impact regression, this sector relies on RL and Computer Vision to manage dynamic movements. The success of the MADDPG algorithm over conventional DDPG in ore blending illustrates the necessity for multi-agent frameworks to manage the dynamic operational complexity of surface mining [30]. This focus on RL suggests that while regression is suitable for static physical predictions, the nature of equipment coordination requires algorithms capable of adaptive learning in real-time environments.
In the post-dismantling management category, environmental issues are addressed. In these articles, deep learning [74] and RF models [37,53,80] have provided high statistical reliability in managing particulate matter and environmental estimation, showcasing the role of AI in sustainable reclamation. However, the identified risk of bias in some environmental monitoring studies highlights a critical need for more rigorous data extraction protocols to ensure the long-term reliability of results [80].
As for the extraction category, research focuses on machine monitoring [26] and multi-objective decision-making [41]. Advancement is the use of offline reinforcement learning supported by GAN-enhanced data [41], which allowed algorithms like E-GCDT to outperform human operators in autonomous digging tasks. One of the two articles uses Unity, showcasing the importance of not only planning but also visualizing the tasks [41]. The integration of GAN-enhanced data to augment real trajectories is an example of how the sector is successfully overcoming data scarcity to train robust autonomous systems.
Finally, overall exploitation integrates digital twin technology for self-sustaining machine monitoring [29]. The use of Gated Recurrent Units (GRUs) allowed researchers to achieve high accuracy in identifying vibrational signs while using Unity as a visualization software [32]. At the same time, the CatBoost algorithm proved effective for multi-objective optimization in sustainable low-emission gold mining [39]. These applications of high-resolution signal processing and visualization provide a foundation for real-time decision-making, effectively bridging the gap between digital simulation and physical reality in a smart mine context.
While blasting studies predominantly utilized ML for static regression to optimize single events, the research in overall exploitation and extraction demonstrates the true operational potential of digital twins as dynamic virtual replicas. The integration of GRU with visualization platforms and the use of GAN-enhanced reinforcement learning moves beyond the prediction to autonomous action. Consequently, for ML to fully enable the Mining 4.0 paradigm, research must shift from isolated predictions to integrated systems that can visualize, simulate, and act withing a digital twin environment.
These findings directly address Research Question 1 (RQ1), identifying SVM, ANN, and RF as the most effective algorithms across unit operations, while highlighting a transition toward hybrid architectures in blasting and reinforcement learning in dynamic tasks. However, unlike the construction and civil infrastructure sectors, where ML is already widely integrated for predictive maintenance and structural health monitoring [83], surface mining appears to be in a transitional phase. While the blasting domain has reached a maturity level comparable to structural analysis, overall exploitation and digital twin integration remain emerging areas, lagging behind the fully automated process control seen in advanced manufacturing [84].

4.2. Methodological Rigor and Validation Reliability

The adoption of standardized data split ratios of 80/20 and 70/30 suggests a mature methodological framework in ML/AI research in mining, although its application to heterogeneous datasets raises concerns about model reliability. However, the reliability of the reported results is strongly influenced by the extreme heterogeneity in data sizes, ranging from small collections of 100 blasts [61] to industrial databases containing 400,000 observations for fuel monitoring [45]. This disparity implies that models trained on more databases likely possess significantly greater capabilities [48] than those derived from localized experimental trials, which often require data enrichment techniques such as Monte Carlo Simulation [50,51].
Additionally, the data input during the blasting phase was not entirely homogeneous; in some articles, important information was omitted, which could have affected the results [33,61,70]. Additionally, in most articles, data was used from only one region or piece of equipment, which could affect the model’s scalability [30,43,54,55,72,80].
While statistical metrics such as R2, RMSE, and MAE are dominant across all categories, the extraction area introduces specific operational efficiency indicators, including Full Bucket Rate and Digging Efficiency [41]. The inclusion of measures such as AWA ensures that the learning process is not only theoretical/mathematical but also operationally stable [26].
This transition from purely statistical to physical performance metrics provides robust validation of the models and demonstrates their utility in a real-world mining environment. By validating specific tasks with human operators and using tools like Unity for visualization, the literature is effectively shifting toward a more pragmatic, evidence-based approach that is essential for Mining 4.0 technologies.
Regarding Research Question 2 (RQ2), this review highlights that validation practices significantly influence the reliability of results. The methodology predominantly relies on random hold-out and k-fold cross-validation (typically 5- or 10-fold), with a notable scarcity of external validation on independent mining sites. Consequently, no meta-analysis was conducted due to significant heterogeneity in datasets, algorithms, and reported metrics. This reliance on internal statistical validation often overshadows operational performance indicators. While statistical metrics like R2 are ubiquitous, the limited reporting of operational KPIs, such as Full Bucket Rate or Digging Efficiency, constrains the assessment of practical deployability.

4.3. Methodological Limitations and Evidence Gap

The assessment of bias and limitations identifies critical barriers to the practical implementation of ML-based systems. Most investigations are site-specific, relying on empirical datasets from a single mine or quarry, which limits the scalability of the models to different geological/operational contexts [30,33,37,43,54,55,61,70,72,75,80].
Another significant gap is the lack of transparency regarding computational tools. In the blasting category, 13 articles did not specify the software utilized [36,38,40,42,46,59,60,62,68,73,76,78,79], and in load and haul, two did not specify it either [43,72].
The high predictive performance frequently reported in the reviewed literature, with R2 often exceeding 0.94, demonstrates the strong potential of these technologies but warrants a cautious interpretation regarding generalizability. In studies utilizing complex nonlinear architectures such as ANN or SVM on limited datasets, such metrics likely indicate overfitting. Furthermore, the prevalence of different train–test splits introduces significant data-leakage risks in spatially dependent mining environments, where random splits fail to ensure independence between the training and testing sets.
Addressing these gaps requires more rigorous data extraction protocols and standardized reporting to facilitate the transition to Mining 4.0.
A foundation for digital twin operations and autonomous systems can only be achieved if the industry moves toward open-access datasets and standardized performance metrics that go beyond statistical accuracy. However, widely adopting such frameworks creates substantial barriers. The mining industry’s competitive nature fosters a culture of data secrecy, where high-resolution operational data is treated as proprietary intellectual property to protect strategic advantages regarding production rates and reserve characteristics [85]. Also, the lack of interoperability between the equipment manufacturers’ standards creates technical bottlenecks, making the anonymization and standardization required for public sharing both technically challenging and resource intensive.
There is a need to ensure methodological transparency and cross-site validation to enable the extractive sector to successfully integrate ML into its core decision-making process.
Addressing Research Question 3 (RQ3), several barriers persist for practical ML employment. In terms of evidence limitations, across domains, certainty is limited by frequent reliance on site-specific databases, incomplete reporting of feature engineering and software, and a high or medium risk of bias in analysis/validation domains. Regarding the review process itself, this review is limited by the predefined publication window (2020–2025), database coverage, and potential missed non-indexed engineering reports. These constraints may underrepresent emerging applications. Finally, concerning actionable implications, standardized reporting and operational indicators are priorities. For practice, studies should report operational KPIs (e.g., ton/h, cycle time, fuel consumption, and bucket fill), alongside statistical metrics, and provide deployable decision-support validation. For research, minimum reporting checklists for datasets, preprocessing, and validation should be adopted.

5. Conclusions

This systematic review included 57 empirical studies (2020–2025) on machine learning (ML) in applications across surface mining unit operations.
Due to substantial heterogeneity across datasets, algorithms, and metrics, a narrative synthesis was conducted rather than a meta-analysis. Results indicate a significant concentration of research in the blasting phase (43 of the 57 articles), where hybrid architectures such as XGBoost, RF, and SVM consistently achieve high statistical reliability, with R2 values frequently exceeding 0.94. In contrast, domains such as “load and haul”, “extraction”, and “post-dismantling management” remain less explored but offer promising applications of reinforcement learning and Computer Vision for dynamic tasks, such as digital twin integration. The application of multi-agent frameworks, such as MADDPG algorithms, proves more effective than traditional models for managing complex tasks, such as ore-blending scheduling. However, despite this potential, the robustness of the current evidence is constrained by its reliance on single-site datasets; independent external validation is required before industrial-scale deployment. The start of a digital twin is being achieved by utilizing models such as Gated Recurrent Unit (GRU) and visualization software, such as Unity, which can bridge the gap between digital simulation and physical reality, providing a more robust foundation for real-time decision-making in an exploitation context.
Despite the strong statistical performance observed, the transition to scalable deployment is likely constrained by the context-dependent nature of single-site datasets and the scarcity of external validation. Furthermore, evidence quality is affected by inconsistent reporting of dataset sizes, participant details, and software tools, which limits reproducibility. Across domains, performance is commonly reported using statistical metrics (e.g., R2, RMSE, and MAE), but operational efficiency and safety-relevant indicators are inconsistently incorporated, creating a gap between model accuracy and practical utility.
To advance toward robust Mining 4.0 applications, future research must prioritize standardized reporting protocols, specifically regarding data provenance, preprocessing pipelines, and algorithm parameters. Additionally, validation strategies should expand beyond internal cross-validation to include multi-site testing and operationally meaningful outcomes to ensure these tools provide reliable, deployable decision support.

Author Contributions

Conceptualization, J.D.; methodology, V.B.R. and J.D.; data extraction, V.B.R.; formal analysis, V.B.R., J.S.B. and J.D.; original draft preparation, V.B.R.; review and editing, J.S.B. and J.D. All authors have read and agreed to the published version of the manuscript.

Funding

This publication was financed by the EU Erasmus+ under the STRIM (Safety Training with Real Immersivity for Mining) project 101083272.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

During the preparation of this manuscript/study, the authors used SciSpace (https://scispace.com/ (accessed on 15 December 2025)) software to extract data from the articles in the systematic review and Grammarly (v1.2.150.1644) to enhance the English throughout the entire article. The authors have reviewed and edited the output and take full responsibility for the content of this publication.

Conflicts of Interest

The funders had no role in this study’s design; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results. The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MLMachine learning
CPSCyber–Physical System
IoTInternet of Things
PRISMAPreferred Reporting Items for Systematic Reviews and Meta-Analyses
PROBASTPrediction model Risk of Bias Assessment Tool
XGBoostExtreme Gradient Boosting
R2Coefficient of Determination
AIArtificial intelligence
SVMSupport Vector Machine
DTDecision Tree
RFRandom Forest
GBGradient Boosting
PCAPrincipal Component Analysis
LiDARLight Detection and Ranging
RMSERoot mean square error
MAEMean absolute error
MAPEMean absolute percentage error
HRHigh Risk
MRMedium Risk
LRLow Risk
USAUnited States of America
ANNArtificial Neural Network
ETsExtra Trees
SVRSupport Vector Regression
ANFISAdaptive Neuro-Fuzzy Inference System
GEPEGene Expression Programming
MARSMultivariate Adaptive Regression Splines
LSTMLong Short-Term Memory
LightGMBLight Gradient Boosting Machine
AHAArtificial Hummingbird Algorithm
GPRGaussian Process Regression
MADDPGMulti-Agent Deep Deterministic Policy Gradient
DDPGDeep Deterministic Policy Gradient
MLRMultiple Linear Regression
CNNConvolutional Neural Network
SIFTScale-Invariant Feature Transform
MLPMulti-Layer Perceptron
RBFRadial Basis Function
STMSpatio-Temporal Models
GRUGated Recurrent Unit
MSEMean Squared Error
RPearson’s Correlation Coefficient
NSENash–Sutcliffe Efficiency
IoAIndex of Agreement
VAFVariance Accounted For
CRMCoefficient of Residual Mass
SIScatter Index
PPVPeak Particle Velocity
PSNRPeak Signal-to-Noise Ratio
SSIMStructural Similarity Index
VIFVisual Information Fidelity
UQIUniversal Quality Index
IoUIntersection over Union
FBRFull Bucket Rate
DEDigging Efficacy
AAAverage Accuracy
AWAAverage Weighted Accuracy
GANGenerative Adversarial Network
CLIPContrastive Language-Image Pre-Training
NMSENormalized Mean Squared Error
KPIKey Performance Indicator

Appendix A

Table A1. Training validation and results—blasting phase.
Table A1. Training validation and results—blasting phase.
ArticleDatasetSplit/TrainEvaluation MethodBest Model
[52]12570/30R2, RMSE, MAE, Adjusted R2, Performance Index (PI)PCA-RF (R2: 0.995, RMSE: 0.011)
[28]374080/20R2, MAE, RMSE, Max ErrorTPE-ET (R2: 0.93, RMSE: 0.04)
[79]23485/15, 200 train 34 testRMSE, MAE, Variance Accounted For (VAF)AW-MKL (VAF: 99.92, MAE: 0.98, RMSE: 2.05)
[27]11180/20
5-fold cross-validation
R, R2, IoA, RMSE, MAPE, NSECB-BOA (R2: 0.989)
[44]20580/20R2, RMSE, MAEICA-ANN (R2: 0.89, RMSE: 5.66 m)
[69]32480/20
10-fold cross-validation
RMSE, SI, Coefficient of Residual Mass (CRM)NCA-BPNN (R: 0.912, RMSE 1.558 dB)
[61]10080/20R, MSE, RMSEBNN (R: 0.94, RMSE: 0.17)
[70]101CV 80/20R2, RMSEGPR (R2: 0.997, MSE: 0.09)
[31]7680/20
10-fold cross-validation
R2, RMSE, MAE, VAFSVM-MFO (R2 train: 0.9939; test: 0.9941)
[58]10270/30R2, RMSE, NSE, CRM, CPAGPSO3-ELM (R: 0.95, RMSE: 0.08, NSE: 0.9, MAE: 0.07, Cp: 0.94)
[32]16680/20
using a trial-and-error approach
R2, RMSE, MAPE, MAE, NSEGOA-ELM (R2: 0.9410 (Train) and 0.9105 (Test))
[33]13610-fold CV, 80/20R2, RMSE, MAEFFA-GBM (R: 0.996)
[34]6280/20R, MSE, MAEXGBTree = 0.929 and MSE = 2.205.
[81]10070/30
SCA_ANN used Levenberg–Marquard
R2, RMSESCA-ANN: 0.9995.
[35]12070/30
Feature Selection (FS)
R2, RMSE, MAE, VAFFS-RF (R2: 0.83)
[36]10270/30
Hyperparameter adjust
R2FS-RF
[67]220414 data for training and 74 for validationAbsolute Error of PPV, Percentage Error of PPVGA
[64]100180/20 MARS
GCV
R2, RMSEMARS (R2: 0.951, RMSE: 0.227)
[46]16280/20
Stacked Generalization
R2, RMSE, MAE, VAFEXGBoosts (R2: 0.968)
[59]21670/30
5-fold cross-validation with 3 repetitions
R2, RMSE, MAERF Enhanced (R2: 0.938)
[60]18370/30
10-fold cross-validation
R2, RMSE, MAERandom Forest (RF)
(R2: 0.874 (train) and 0.826 (validation))
[65]4880/10/10
Levenberg–Marquardt (trainlm),
Bayesian Regularization (trainbr)
Optimization with ICA
R2, RMSE, MAPE, ADJUSTED R2, PI, VAFICA-ANN (R2: 0.962, error: 2.7%)
[62]7280/20R2, MSE, MAPERF (R2: 0.924, MSE: 3.40)
[38]262 + 10980/20
Optimization:
LSO and POA
R2, RMSE, MAPE, SI(LSO-RF e POA-RF) (R2 > 0.95)
[47]7670/15/15R2, RMSE, MAE, CPZ-BRCWNN R2 of 0.999, 0.988 and 0.983
[48]25280/20
Cross-validations
R2, RMSE, MAEJSO-CatBoost has the highest predictive performance
[66]25880/20R2, MSE, RMSE, MAE, SILSTM (R2: 0.999)
[63]7580/20
5-fold cross-validation
R2, RMSE, SENSITIVITY ANALYSISCapSA-MLP (R2: 0.904)
[49]26280/20
5-fold cross-validation
R2, MSE, COEFFICIENT OF VARIATION (COV)SSM-Bagging (R2: 0.974)
[73]10970, 20, 10TAYLOR DIAGRAM, R2, RMSE, MAPEDF-EDML (R2: 0.835 (train) and 0.820 (validation))
[56]1000+70/20/10
10-fold cross-validation
R2, RMSE, MAEPINNS + XGBoost (R2: 0.92)
[68]45775/25 K-Fold cross-validation
normalized Min–Max
BIAS FACTOR FOR EVALUATION, R2, RMSE, MAE, VAFThe Voting 8 (LightGBM-GBM-DT-ET-RF-CatBoost-CART-AdaBoost-XGBoost) model has the highest R2 (0.9876, 0.9726)
[75]1438
Isolation Forest reduced to 992
70/30
GridSearchCV 10-fold cross-validation
R2, RMSEDecision Tree Regressor (DT) optimized (R2: 0.997)
[40]103 + 11480/20
10-fold cross-validation
TAYLOR DIAGRAM, R2, RMSEAHA-GPR (R2: 0.978)
[78]10480/10/10R2, RMSE, MAEANN model with an architecture of 8-10-1 RMSE (0.273), MAE (0.189), R2 (0.988)
[50]118
MCs expanded to 10.000
70/30
trial and error
R2, RMSE, VAFPDNN’s
[51]1032
MCs 10,000
70/30
range of 2–11 for the number of hidden nodes in BRNN
ACCURACY, R2, RMSE, MAE, VAFGEP (R2: 0.97)
[71]10290 Train, 12 test, 20% train set aside for validation
hyper-parameter tuning via the grid search using
the 5-fold cross-validation
R2, RMSEANN (R2: 0.87, MSE: 0.0031)
[77]63,116 sample images were produced.
Sample images contain a total of 23,125,486
61,853 samples for training, 631 for validation,
and 632 for testing
PERCENTAGE ERROR OF PPV, RESIDUAL ERROR, MSEResNet50
[57]10280% training and 20% testing
10-fold cross-validation
R2, MAPE, RMSEXGBoost (R2: 0.952).
Fragmentation Prediction (R2: 0.94, RMSE: 1.82, MAE: 1.4518)
PPV (R2: 0.92, RMSE: 1.15, MAE: 0.8819)
[18]219199 data points were considered for training the network,
9 data points for cross-validation,
11 data points for testing the model
R2, MSE, RMSE, MAPE, MAEANN (architecture 5-64-32-16-1)
[76]7680/20
5-fold cross-validation
R2, MSE, VAFSVR-GWO (R2: 0.8353)
[42]24080/20
Greedy layer-wise with RBM
WOA for optimization
R2, RMSEDNN

References

  1. Al-Shwaf, L.; Bell, J.E. Raw Material Supply Risks: Examining Extraction and Geopolitical Conflict. Transp. J. 2025, 64, e70006. [Google Scholar] [CrossRef]
  2. Ramani, R.V. Surface Mining Technology: Progress and Prospects. Procedia Eng. 2012, 46, 9–21. [Google Scholar] [CrossRef]
  3. Petavratzi, E.; Gunn, G. Decarbonizing the Automotive Sector: A Primary Raw Material Perspective on Targets and Timescales. Min. Econ. 2023, 36, 545–561. [Google Scholar] [CrossRef]
  4. Cheng, C.; Chu, H.; Zhang, L.; Tang, L. Green Supply Chain for Steel Raw Materials under Price and Demand Uncertainty. J. Clean. Prod. 2024, 462, 142621. [Google Scholar] [CrossRef]
  5. Berthet, E.; Lavalley, J.; Anquetil-Deck, C.; Ballesteros, F.; Stadler, K.; Soytas, U.; Hauschild, M.; Laurent, A. Assessing the Social and Environmental Impacts of Critical Mineral Supply Chains for the Energy Transition in Europe. Glob. Environ. Change 2024, 86, 102841. [Google Scholar] [CrossRef]
  6. Valentini, L. Sustainable Sourcing of Raw Materials for the Built Environment. Mater. Today Proc. 2023; in press. [CrossRef]
  7. Anlauf, A. An Extractive Bioeconomy? Phosphate Mining, Fertilizer Commodity Chains, and Alternative Technologies. Sustain. Sci. 2023, 18, 633–644. [Google Scholar] [CrossRef]
  8. Ghorbani, Y.; Nwaila, G.T.; Zhang, S.E.; Bourdeau, J.E.; Cánovas, M.; Arzua, J.; Nikadat, N. Moving towards Deep Underground Mineral Resources: Drivers, Challenges and Potential Solutions. Resour. Policy 2023, 80, 103222. [Google Scholar] [CrossRef]
  9. Nikkhah, A.; Vakylabad, A.B.; Hassanzadeh, A.; Niedoba, T.; Surowiak, A. An Evaluation on the Impact of Ore Fragmented by Blasting on Mining Performance. Minerals 2022, 12, 258. [Google Scholar] [CrossRef]
  10. Rakhmangulov, A.; Burmistrov, K.; Osintsev, N. Selection of Open-Pit Mining and Technical System’s Sustainable Development Strategies Based on MCDM. Sustainability 2022, 14, 8003. [Google Scholar] [CrossRef]
  11. Aalian, Y.; Gamache, M.; Pesant, G. Short-Term Underground Mine Planning with Uncertain Activity Durations Using Constraint Programming. J Sched 2024, 27, 423–439. [Google Scholar] [CrossRef]
  12. Koščová, M.; Hellmer, M.; Anyona, S.; Gvozdkova, T. Geo-Environmental Problems of Open PitMining: Classification and Solutions. E3S Web Conf. 2018, 41, 01034. [Google Scholar] [CrossRef]
  13. Onifade, M.; Adebisi, J.A.; Shivute, A.P.; Genc, B. Challenges and Applications of Digital Technology in the Mineral Industry. Resour. Policy 2023, 85, 103978. [Google Scholar] [CrossRef]
  14. Duarte, J.; Baptista, J.S. Digital Twin Applications in the Extractive Industry—A Short Review. In Occupational and Environmental Safety and Health V; Arezes, P.M., Melo, R.B., Carneiro, P., Castelo Branco, J., Colim, A., Costa, N., Costa, S., Duarte, J., Guedes, J.C., Perestrelo, G., et al., Eds.; Springer Nature: Cham, Switzerland, 2024; pp. 771–781. [Google Scholar]
  15. Noshi, C.I.; Schubert, J.J. The Role of Machine Learning in Drilling Operations; A Review|Request PDF. In Proceedings of the SPE/AAPG Eastern Regional Meeting, Pittsburgh, PA, USA, 7 October 2018. [Google Scholar]
  16. Baek, J.; Choi, Y. Deep Neural Network for Predicting Ore Production by Truck-Haulage Systems in Open-Pit Mines. Appl. Sci. 2020, 10, 1657. [Google Scholar] [CrossRef]
  17. Gladious, J.; Paul, P.S.; Mukhopadhyay, M. Machine Learning Based Prediction of Geotechnical Parameters Affecting Slope Stability in Open-Pit Iron Ore Mines in High Precipitation Zone. Sci. Rep. 2025, 15, 21868. [Google Scholar] [CrossRef]
  18. Gebretsadik, A.; Kumar, R.; Fissha, Y.; Kide, Y.; Okada, N.; Ikeda, H.; Mishra, A.K.; Armaghani, D.J.; Ohtomo, Y.; Kawamura, Y. Enhancing Rock Fragmentation Assessment in Mine Blasting through Machine Learning Algorithms: A Practical Approach. Discov. Appl. Sci. 2024, 6, 223. [Google Scholar] [CrossRef]
  19. Bonagiri, D.; Ragam, P. Ensemble Machine Learning Models for Blast-Induced Air Noise: A Review of Transformative Innovations in Minerals. J. Mines Met. Fuels 2025, 73, 2051–2082. [Google Scholar] [CrossRef]
  20. Senses, S.; Kumral, M. An Optimization-Based Approach to Fleet Reliability and Allocation in Open-Pit Mining. Decis. Anal. J. 2025, 15, 100583. [Google Scholar] [CrossRef]
  21. Elwahab, A.A.; Topal, E.; Jang, H.D. Review of Machine Learning Application in Mine Blasting. Arab. J. Geosci. 2023, 16, 133. [Google Scholar] [CrossRef]
  22. Jung, D.; Choi, Y. Systematic Review of Machine Learning Applications in Mining: Exploration, Exploitation, and Reclamation. Minerals 2021, 11, 148. [Google Scholar] [CrossRef]
  23. Arthur, C.K.; Bhatawdekar, R.M.; Temeng, V.A.; Agyei, G.; Ziggah, Y.Y. Application of Artificial Intelligence in Predicting Blast-Induced Ground Vibration. In Applications of Artificial Intelligence in Mining and Geotechnical Engineering; Elsevier: Amsterdam, The Netherlands, 2024; pp. 251–267. [Google Scholar]
  24. Page, M.J.; McKenzie, J.E.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. The PRISMA 2020 Statement: An Updated Guideline for Reporting Systematic Reviews. BMJ 2021, 372, n71. [Google Scholar] [CrossRef] [PubMed]
  25. Wolff, R.F.; Moons, K.G.M.; Riley, R.D.; Whiting, P.F.; Westwood, M.; Collins, G.S.; Reitsma, J.B.; Kleijnen, J.; Mallett, S.; PROBAST Group. PROBAST: A Tool to Assess the Risk of Bias and Applicability of Prediction Model Studies. Ann. Intern. Med. 2019, 170, 51–58. [Google Scholar] [CrossRef]
  26. Ling, H.; Gao, T.; Gong, T.; Wu, J.; Zou, L. Hydraulic Rock Drill Fault Classification Using X−Vectors. Mathematics 2023, 11, 1724. [Google Scholar] [CrossRef]
  27. Chen, L.; Fissha, Y.; Hasanipanah, M.; Ghodhbani, R.; Dehghani, H.; Khatti, J. Accurate Prediction of Blast-Induced Ground Vibration Intensity Using Optimized Machine Learning Models. Def. Technol. 2025, 52, 32–46. [Google Scholar] [CrossRef]
  28. Mame, M.; Huang, S.; Li, C.; Zhou, J. Application of Extra-Trees Regression and Tree-Structured Parzen Estimators Optimization Algorithm to Predict Blast-Induced Mean Fragmentation Size in Open-Pit Mines. Appl. Sci. 2025, 15, 8363. [Google Scholar] [CrossRef]
  29. Jiang, J.; Fan, C.; Chen, H.; Wu, F.; Feng, X.; Xiao, C.; Pan, H.; Wu, X.; Zhang, Z. A Self-Powered Triboelectric Nano-Sensor Enabled Digital Twin for Self-Sustained Machine Monitoring in Smart Mine. Nano Res. 2025, 18, 94907287. [Google Scholar] [CrossRef]
  30. Feng, Z.; Liu, G.; Wang, L.; Gu, Q.; Chen, L. Research on the Multiobjective and Efficient Ore-Blending Scheduling of Open-Pit Mines Based on Multiagent Deep Reinforcement Learning. Sustainability 2023, 15, 5279. [Google Scholar] [CrossRef]
  31. Chen, L.; Jahed Armaghani, D.J.; Fakharian, P.; Bhatawdekar, R.M.; Samui, P.; Khandelwal, M.; Khedher, K.M. A Study on Environmental Issues of Blasting Using Advanced Support Vector Machine Algorithms. Int. J. Environ. Sci. Technol. 2022, 19, 6221–6240. [Google Scholar] [CrossRef]
  32. Yu, C.; Koopialipoor, M.; Murlidhar, B.; Mohammed, A.; Armaghani, D.; Mohamad, E.; Wang, Z. Optimal ELM-Harris Hawks Optimization and ELM-Grasshopper Optimization Models to Forecast Peak Particle Velocity Resulting from Mine Blasting. Nat. Resour. Res. 2021, 30, 2647–2662. [Google Scholar] [CrossRef]
  33. Xie, C.; Nguyen, H.; Xuan Nam, X.-N.; Choi, Y.; Zhou, J.; Nguyen-Trang, T. Predicting Rock Size Distribution in Mine Blasting Using Various Novel Soft Computing Models Based on Meta-Heuristics and Machine Learning Algorithms. Geosci. Front. 2021, 12, 101108. [Google Scholar] [CrossRef]
  34. He, Z.; Jahed Armaghani, D.J.; Masoumnezhad, M.; Khandelwal, M.; Zhou, J.; Bhatawdekar, B.R. A Combination of Expert-Based System and Advanced Decision-Tree Algorithms to Predict Air-Overpressure Resulting from Quarry Blasting. Nat. Resour. Res. 2021, 30, 1889–1903. [Google Scholar] [CrossRef]
  35. Zhang, H.; Zhou, J.; Armaghani, D.; Tahir, M.; Pham, B.; Huynh, V. A Combination of Feature Selection and Random Forest Techniques to Solve a Problem Related to Blast-Induced Ground Vibration. Appl. Sci. 2020, 10, 869. [Google Scholar] [CrossRef]
  36. Zhou, J.; Asteris, P.; Armaghani, D.; Pham, B. Prediction of Ground Vibration Induced by Blasting Operations through the Use of the Bayesian Network and Random Forest Models. SOIL Dyn. Earthq. Eng. 2020, 139, 106390. [Google Scholar] [CrossRef]
  37. Luan, B.; Zhou, W.; Jiskani, I.M.; Wang, Z. An Improved Machine Learning Approach for Optimizing Dust Concentration Estimation in Open-Pit Mines. Int. J. Environ. Res. Public Health 2023, 20, 1353. [Google Scholar] [CrossRef]
  38. Zhang, Y.; Qiu, Y.; Du, K.; Nguyen, H.; Armaghani, D.J.; Zhou, J. Optimizing Flyrock Forecasting in Open-Pit Blasting Using Hybrid Machine Learning Models. Rock Mech. Rock Eng. 2025, 58, 12523–12550. [Google Scholar] [CrossRef]
  39. Qiu, L.; Yang, X.; Tang, J.; Fan, L. Machine Learning-Driven Multi-Objective Optimization for Sustainable, Cost-Effective, and Low-Emission Gold Mining. J. Clean. Prod. 2025, 511, 145621. [Google Scholar] [CrossRef]
  40. Yu, Z.; Du, L.-F.; Liu, J.-X.; Zhou, J.; Li, C.-Q. Feasibility of a Hybrid AHA-GPR Model for Predicting Blasting Fragmention in Surface Mines. Earth Sci. Inform. 2025, 18, 278. [Google Scholar] [CrossRef]
  41. Zhao, Q.; Gao, L.; Wu, D.; Lei, Y.; Wang, L.; Qi, J.; Hu, J. E-GCDT: Advanced Reinforcement Learning with GAN-Enhanced Data for Continuous Excavation System. Appl. Intell. 2025, 55, 413. [Google Scholar] [CrossRef]
  42. Guo, H.; Zhou, J.; Koopialipoor, M.; Jahed Armaghani, D.; Tahir, M.M. Deep Neural Network and Whale Optimization Algorithm to Assess Flyrock Induced by Blasting. Eng. Comput. 2021, 37, 173–186. [Google Scholar] [CrossRef]
  43. Xiao, D.; Li, H.; Ji, Z.; Xu, E.; Luo, B.; Chen, J. An Anti-Collision Early Warning System for Mine Trucks Based on RBF Network and WIFI. J. Phys. Conf. Ser. 2020, 1631, 012157. [Google Scholar] [CrossRef]
  44. Hanifehnia, J.; Esmaeilzadeh, A.; Mikaeil, R.; Atalou, S. Prediction of Blast-Induced Flyrock by Using Neural-Imperialist Competitive Method (Case Study: Sungun Copper Mine). Rud. Geol. Naft. Zb. 2024, 39, 109–120. [Google Scholar] [CrossRef]
  45. Alamdari, S.; Basiri, M.; Mousavi, A.; Soofastaei, A. Application of Machine Learning Techniques to Predict Haul Truck Fuel Consumption in Open-Pit Mines. J. Min. Environ. 2022, 13, 69–85. [Google Scholar] [CrossRef]
  46. Hosseini, S.; Poormirzaee, R.; Jahed Armaghani, D.J.; Sabri, M.M. Prediction of Ground Vibration Due to Mine Blasting in a Surface Lead–Zinc Mine Using Machine Learning Ensemble Techniques. Sci. Rep. 2023, 13, 6591. [Google Scholar] [CrossRef]
  47. Hosseini, S.; Lawal, A.I.; Mulenga, F. Prediction of Blast-Induced Ground Vibration in Dolomitic Marble Quarry Using Z-Number Information and Fuzzy Cognitive Map Based Neural Network Models. Rock Mech. Bull. 2025, 4, 100217. [Google Scholar] [CrossRef]
  48. Rouhani, M.M.; Hasanipanah, M.; Yin, X.; Ahmadianfar, I.; Dehghani, H. Intelligent Prediction of Flyrock Hazards in Surface Mining Using Optimized Gradient Boosting Models. Nat. Resour. Res. 2025, 35, 629–651. [Google Scholar] [CrossRef]
  49. Barkhordari, M.S.; Jahed Armaghani, D.J.; Fakharian, P. Ensemble Machine Learning Models for Prediction of Flyrock Due to Quarry Blasting. Int. J. Environ. Sci. Technol. 2022, 19, 8661–8676. [Google Scholar] [CrossRef]
  50. Hosseini, S.; Poormirzaee, R. Green Policy for Managing Blasting Induced Dust Dispersion in Open-Pit Mines Using Probability-Based Deep Learning Algorithm. Expert Syst. Appl. 2024, 240, 122469. [Google Scholar] [CrossRef]
  51. Hosseini, S.; Mousavi, A.; Monjezi, M.; Khandelwal, M. Mine-to-Crusher Policy: Planning of Mine Blasting Patterns for Environmentally Friendly and Optimum Fragmentation Using Monte Carlo Simulation-Based Multi-Objective Grey Wolf Optimization Approach. Resour. Policy 2022, 79, 103087. [Google Scholar] [CrossRef]
  52. Dukuly, L.P.; Gupta, M.; Ghani, S.; Akram, W. PCA-Integrated Machine Learning Framework for Predicting Rock Fragmentation in Blasting Operations. Multiscale Multidiscip. Model. Exp. Des. 2025, 8, 409. [Google Scholar] [CrossRef]
  53. Podicheti, R.K.; Karra, R.C. Analysis of Concentration of Ambient Particulate Matter in the Surrounding Area of an Opencast Coal Mine Using Machine Learning Techniques. J. Min. Environ. 2024, 15, 961–976. [Google Scholar] [CrossRef]
  54. Jha, S.; Agrawal, H.; Rai, P. AI-Powered Prediction for Estimating Specific Fuel Consumption in Heavy-Duty Dumpers in Coal Mines. J. Inst. Eng. (India) Ser. D 2025. [Google Scholar] [CrossRef]
  55. Chaulya, S.K.; Choudhary, M.; Kumar, N.; Kumar, V.; Chowdhury, A. Smart Driving Assistance System for Mining Operations in Foggy Environments. Discov. Electron. 2025, 2, 13. [Google Scholar] [CrossRef]
  56. Ala, C.K.; Mayaluri, Z.L.; Kaushik, A.; Nikhat, N.; Saxena, S.; Zamani, A.T.; Muduli, D. An Explainable AI-Based Framework for Predicting and Optimizing Blast-Induced Ground Vibrations in Surface Mining. Results Eng. 2025, 27, 106046. [Google Scholar] [CrossRef]
  57. Chandrahas, N.S.; Choudhary, B.S.; Teja, M.V.; Venkataramayya, M.S.; Prasad, N.S.R.K. XG Boost Algorithm to Simultaneous Prediction of Rock Fragmentation and Induced Ground Vibration Using Unique Blast Data. Appl. Sci. 2022, 12, 5269. [Google Scholar] [CrossRef]
  58. Armaghani, D.; Kumar, D.; Samui, P.; Hasanipanah, M.; Roy, B. A Novel Approach for Forecasting of Ground Vibrations Resulting from Blasting: Modified Particle Swarm Optimization Coupled Extreme Learning Machine. Eng. Comput. 2021, 37, 3221–3235. [Google Scholar] [CrossRef]
  59. Nguyen, H.; Xuan Nam, X.-N.; Drebenstedt, C. Machine Learning Algorithms for Data Enrichment: A Promising Solution for Enhancing Accuracy in Predicting Blast-Induced Ground Vibration in Open-Pit Mines. Inz. Miner. 2023, 1, 79–88. [Google Scholar] [CrossRef]
  60. Nguyen, H.; Xuan Nam, X.-N.; Drebenstedt, C.; Choi, Y. Improving PPV Prediction in Open-Pit Blasting through Cubist-Based Feature Enrichment and Machine Learning Models. Int. J. Min. Reclam. Environ. 2025, 40, 234–264. [Google Scholar] [CrossRef]
  61. Fissha, Y.; Ikeda, H.; Toriya, H.; Adachi, T.; Kawamura, Y. Application of Bayesian Neural Network (BNN) for the Prediction of Blast-Induced Ground Vibration. Appl. Sci. 2023, 13, 3128. [Google Scholar] [CrossRef]
  62. Ezatullah, R.; Bassir, E.; Akihiro, H.; Takashi, S.; Hideki, S. A Comparative Study of Two Tree-Based Models for Predicting Flyrock Velocity at Open Pit Bench Mining. Open J. Appl. Sci. 2024, 14, 267–287. [Google Scholar] [CrossRef]
  63. Gaopale, K.; Sasaoka, T.; Hamanaka, A.; Shimada, H. Integrated Capuchin Search Algorithm-Optimized Multilayer Perceptron for Robust and Precise Prediction of Blast-Induced Airblast in a Blasting Mining Operation. Geosciences 2025, 15, 306. [Google Scholar] [CrossRef]
  64. Komadja, G.; Rana, A.; Glodji, L.; Anye, V.; Jadaun, G.; Onwualu, P.; Sawmliana, C. Assessing Ground Vibration Caused by Rock Blasting in Surface Mines Using Machine-Learning Approaches: A Comparison of CART, SVR and MARS. Sustainability 2022, 14, 11060. [Google Scholar] [CrossRef]
  65. Taiwo, B.O.; Ajibona, A.I.; Gebrestsadik, A.; amobuwa, O.V.F.; Thomas, O.A.; Omosebi, A.O. Artificial Intelligence Based Smart Blasting Using ICA Optimized Neural Network for Oversize Prediction in a Small Scale Dolomite Quarry in Nigeria. Rock Mech. Lett. 2025, 2, 132–140. [Google Scholar] [CrossRef]
  66. Taiwo, B.O.; Fissha, Y.; Hosseini, S.; Khishe, M.; Kahraman, E.; Adebayo, B.; Sazid, M.; Adesida, P.A.; Famobuwa, O.V.; Faluyi, J.O.; et al. Machine Learning Based Prediction of Flyrock Distance in Rock Blasting: A Safe and Sustainable Mining Approach. Green Smart Min. Eng. 2024, 1, 346–361. [Google Scholar] [CrossRef]
  67. Yardimci, A.; Erkayaoglu, M. Simulation of Blast-Induced Ground Vibrations Using a Machine Learning-Assisted Mechanical Framework. Environ. EARTH Sci. 2023, 82, 508. [Google Scholar] [CrossRef]
  68. Kahraman, E.; Hosseini, S.; Taiwo, B.O.; Fissha, Y.; Jebutu, V.A.; Akinlabi, A.A.; Adachi, T. Fostering Sustainable Mining Practices in Rock Blasting: Assessment of Blast Toe Volume Prediction Using Comparative Analysis of Hybrid Ensemble Machine Learning Techniques. J. Saf. Sustain. 2024, 1, 75–88. [Google Scholar] [CrossRef]
  69. Ziggah, Y.Y.; Temeng, V.A.; Arthur, C.K. A New Synergetic Model of Neighbourhood Component Analysis and Artificial Intelligence Method for Blast-Induced Noise Prediction. Model. Earth Syst. Environ. 2023, 9, 3483–3502. [Google Scholar] [CrossRef]
  70. Arthur, C.K.; Bhatawdekar, R.M.; Mohamad, E.T.; Sabri, M.M.S.; Bohra, M.; Khandelwal, M.; Kwon, S. Prediction of Blast-Induced Ground Vibration at a Limestone Quarry: An Artificial Intelligence Approach. Appl. Sci. 2022, 12, 9189. [Google Scholar] [CrossRef]
  71. Amoako, R.; Jha, A.; Zhong, S. Rock Fragmentation Prediction Using an Artificial Neural Network and Support Vector Regression Hybrid Approach. Mining 2022, 2, 233–247. [Google Scholar] [CrossRef]
  72. Choi, Y.; Nguyen, H.; Bui, X.-N.; Nguyen-Thoi, T.; Park, S. Estimating Ore Production in Open-Pit Mines Using Various Machine Learning Algorithms Based on a Truck-Haulage System and Support of Internet of Things. Nat. Resour. Res. 2021, 30, 1141–1173. [Google Scholar] [CrossRef]
  73. Asteris, P.; Armaghani, D. An Empirical-Driven Machine Learning (EDML) Approach to Predict PPV Caused by Quarry Blasting. Bull. Eng. Geol. Environ. 2025, 84, 200. [Google Scholar] [CrossRef]
  74. Yu, W.; Zhang, X.; Gloaguen, R.; Zhu, X.; Ghamisi, P. MineNetCD: A Benchmark for Global Mining Change Detection on Remote Sensing Imagery. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–16. [Google Scholar] [CrossRef]
  75. Moustafa, S.S.R.; Abdalzaher, M.S.; Yassien, M.H.; Wang, T.; Elwekeil, M.; Mossa, H.E.A. Development of an Optimized Regression Model to Predict Blast-Driven Ground Vibrations. IEEE Access 2021, 9, 31826–31841. [Google Scholar] [CrossRef]
  76. Li, E.; Yang, F.; Ren, M.; Zhang, X.; Zhou, J.; Khandelwal, M. Prediction of Blasting Mean Fragment Size Using Support Vector Regression Combined with Five Optimization Algorithms. J. Rock Mech. Geotech. Eng. 2021, 13, 1380–1397. [Google Scholar] [CrossRef]
  77. Bamford, T.; Esmaeili, K.; Schoellig, A.P. A Deep Learning Approach for Rock Fragmentation Analysis. Int. J. Rock Mech. Min. Sci. 2021, 145, 104839. [Google Scholar] [CrossRef]
  78. Saubi, O.; Jamisola, R.S., Jr.; Suglo, R.S.; Matsebe, O. Machine Learning Tool to Minimise and Predict Airblast during Blasting and to Optimize the Design of Blasting Operations. Int. J. Min. Miner. Eng. 2025, 16, 148–167. [Google Scholar] [CrossRef]
  79. Zhang, R.; Li, Y.; Gui, Y.; Jahed Armaghani, D.J.; Yari, M. Adaptive Weighted Multi-Kernel Learning for Blast-Induced Flyrock Distance Prediction. Rock Mech. Rock Eng. 2025, 58, 679–695. [Google Scholar] [CrossRef]
  80. Liu, Z.; Zhang, R.; Ma, J.; Zhang, W.; Li, L. Analysis and Prediction of the Meteorological Characteristics of Dust Concentrations in Open-Pit Mines. Sustainability 2023, 15, 4837. [Google Scholar] [CrossRef]
  81. Lawal, A.I.; Kwon, S.; Hammed, O.S.; Idris, M.A. Blast-Induced Ground Vibration Prediction in Granite Quarries: An Application of Gene Expression Programming, ANFIS, and Sine Cosine Algorithm Optimized ANN. Int. J. Min. Sci. Technol. 2021, 31, 265–277. [Google Scholar] [CrossRef]
  82. Saldana, M.; Gallegos, S.; Arias, D.; Salazar, I.; Castillo, J.; Salinas-Rodríguez, E.; Navarra, A.; Toro, N.; Cisternas, L.A. Applications of Kuz–Ram Models in Mine-to-Mill Integration and Optimization—A Review. Minerals 2024, 14, 1162. [Google Scholar] [CrossRef]
  83. Carvalho, T.P.; Soares, F.A.A.M.N.; Vita, R.; Francisco, R.d.P.; Basto, J.P.; Alcalá, S.G.S. A Systematic Literature Review of Machine Learning Methods Applied to Predictive Maintenance. Comput. Ind. Eng. 2019, 137, 106024. [Google Scholar] [CrossRef]
  84. Inayathullah, S.; Buddala, R. Review of Machine Learning Applications in Additive Manufacturing. Results Eng. 2025, 25, 103676. [Google Scholar] [CrossRef]
  85. Nemala, P.; Chen, B.; Cui, H. A Privacy Preserving Attribute-Based Access Control Model for the Tokenization of Mineral Resources via Blockchain. Appl. Sci. 2025, 15, 8290. [Google Scholar] [CrossRef]
Figure 1. PRISMA flow diagram, adapted from Page et al. [24].
Figure 1. PRISMA flow diagram, adapted from Page et al. [24].
Applsci 16 03246 g001
Figure 2. Study distribution by country.
Figure 2. Study distribution by country.
Applsci 16 03246 g002
Figure 3. Area of ML/AI application.
Figure 3. Area of ML/AI application.
Applsci 16 03246 g003
Figure 4. Scientometric analysis of keyword co-occurrence network.
Figure 4. Scientometric analysis of keyword co-occurrence network.
Applsci 16 03246 g004
Table 1. ML/AI and software used—blasting phase.
Table 1. ML/AI and software used—blasting phase.
ArticleML/AISoftware
[52]PCA-RF, PCA-XGB, PCA-ANN, PCA-SVMSplit-Desktop (N/S), WipFrag (N/S),
FragScan (N/S)
[28]TPE-ET, TPE-GB, TPE-RFPython (N/S)
[79]AW-MKL, KRR, SVRNaN
[27]MAIN ONE CB,
BAT, BOA, GOA, SSA
Python (N/S), Excel (N/S)
[44]ICA-ANN, MLP-ANNMATLAB (N/S)
[69]NCA-SVM, NCA-BPNN, NCA-GRNN and NCA-RBFNNMATLAB (N/S)
[61]BNN, GB, RF, KNN, DTMATLAB (N/S)
[70]GPR, ELM, BPNNMATLAB (N/S)
[31]SVM, SVM-MFO, SVM-PSO, SVM-GWO, SVM-COA, SVM, WOAMATLAB (N/S), Python (N/S)
[58]PSO-ELM, AGPSO-ELM, ELM, GPR, MPMR and LS–SVMMATLAB (N/S)
[32]GOA-ELM, HHO-ELM, ELMMATLAB (N/S)
[33]FFA-GBM, FFA-SVM, FFA-ANN, FFA-GPSplit-Desktop (N/S)
[34]FDM-XGBoost-tree, FDM-RF. XGBoost-tree, RFMATLAB (N/S)
[81]SCA-ANN, ANFIS, GEPMATLAB (N/S), Excel (N/S), GEneXpro (tools 5.0)
[35]RF (Random Forest).
CART (Classification and Regression Trees).
CHAID (Chi-squared Automatic Interaction Detection).
ANN (Artificial Neural Network).
SVM (Support Vector Machine
IBM SPSS Modeler (18.2.1)
[36]FS-RF, FS-BNNan
[67]Mechanical Simulation Framework
(calibrated using (GA)),
ANFIS
Python (N/S)
[64]MARS, CART, SVRPython (Anaconda3)
[46]EXGBoosts, ANNsNaN
[59]RF, SVM, KNN, CARTNaN
[60]SVM, RF, k-NN, and GBMNaN
[65]ICA-ANN, ANNMATLAB (N/S)
[62]RF, DTNaN
[38]RF-LSO, RF-POA, RFNaN
[47]BRNN, BRCWNN, Z-BRCWNNMATLAB (N/S), Minitab (N/S)
[48]AOA-LightGBM, JSO-LightGBM, HHO-LightGBM,
GMO-LightGBM, AOA-CatBoost, JSO-CatBoost,
HHO-CatBoost and GMO-CatBoost
Python (N/S)
[66]SVR, ANN, MLP, RF, BRNN, LSTMMATLAB (Version 2021)
[63]CapSA-MLP, PSO-ANNMATLAB (R2024a)
[49]SAE, WAE, ISM, SSM, BXGBoostGridSearchCV(N/S), SHAP (N/S),
Python (N/S)
[73]EDML, DF, XGBOOSTNaN
[56]*INNS, XGBOOST, LSTM, RF, SVM, ANNPython (N/S)
[68]XGBoost, AdaBoost, CART, CAtBoost, RF, DT,
ET, GBM, LightGBM, LGBM combination with all
NaN
[75]Extra Trees Regressor, Random Forest Regressor,
Bagging Regressor, Gradient Boosting Regressor,
HistGradient Boosting Regressor, XGBRegressor,
AdaBoost Regressor
Python (N/S)
[40]AHA-GPR, GPR, ANN, SVRNaN
[78]ANN, SVM, k-NN, RFNaN
[50]PDNN, PANN, DNN, ANN MCsPython (N/S), Matlab (N/S)
[51]GEP, BRNN, MNLR, MOGWOSplit-Desktop (N/S),
GeneXPro Tools (5.0)
[71]ANN-SVR, ANNPython (N/S), Keras (N/S)
[77]DNN with ResNet50, Pixel Classifier,Split-Desktop (N/S)
[57]XGBoost, RF, KNN, SVR, ANNStrayos (N/S), O-PitBlast (N/S)
[18]RFR, SVR, XGBoost, and ANNPython (N/S)
[76]SVR-GWO, SVR-PSONaN
[42]WOA, DNN, RBMNaN
N/S: Software version not specified in the original study.
Table 2. ML/AI and software used—load and haul.
Table 2. ML/AI and software used—load and haul.
ArticleML/AISoftware
[30]MADDPG, DDPGTensorFlow (N/S), Python (3.6), PyCharm (2019.1.1),
MATLAB (N/S)
[45]MLR, RF, ANN, SVM, K-NNPython (N/S)
[54]ANN, RF, ANFISWEKA (N/S)
[55]CISAAC, CNN, SIFTTensorFlow (N/S), Python (N/S), OpenCV (N/S), KERAS (N/S),
PIX4DMpper (N/S), LabellMG (N/S)
[72]Random Forest (RF),
Support Vector Machine (SVM),
Multi-Layer Perceptron (MLP),
Classification and Regression Tree (CART),
k-Nearest Neighbors (kNNs) and M5Tree
NaN
[43]RBFNaN
N/S: Software version not specified in the original study.
Table 3. ML/AI and software used—post-dismantling management.
Table 3. ML/AI and software used—post-dismantling management.
ArticleML/AI Software
[53]Bagging, RF, DTNaN
[74]ChangeFFT, CNN, Transformers, VMamba,
A2Net, BIT, ChangeFormer, DMINet, FC-EF,
FCNPP, ICIFNet, RDPNet, ResUnet,
SiamUnet-Conc, SiamUnet-Diff, and SNUNet
PyTorch (N/S), HuggingFace (N/S),
eCognition (N/S)
[80]LSTM, RFR, SVR, BBNMATLAB (N/S)
[37]RF-MC, RF-PSONaN
N/S: Software version not specified in the original study.
Table 4. Training validation and results—load and haul.
Table 4. Training validation and results—load and haul.
ArticleDatasetSplit/TrainEvaluation MethodBest Model
[30]90.000Reinforcement learning (RL)Deviation Control, Convergence Rate, Computation TimeMADDPG
[45]400.000 registers80/20
K-fold cross-validation
R2, MSE, MAEANN
[54]66 dump trucks:
27 with 190-ton capacity
16 with 120 tons
23 with 100 tons
RF: 20-fold cross-validation
ANFIS: split (training) and (testing)
ANN: 10 hidden neurons, tanh activation,
max 10.000 epochs or 10−5 error threshold
R2, MAE, RMSE, NSEANN:
(R2: 0.989, RMSE: 0.195, MAE: 0.142)
[55]7 pieces of equipment with 1550 to 1600 by class
Total 11.000 imagens
80/20
K-Fold cross-validation
UQI, SSIM, VIF, PSNRCISACC for image and architecture
SSD–MobileNet for detection
[72]16,005 datasets were collected, using the downscaling method
was applied to downscale the size of the
dataset into 3.000 observations
Models were validated using three downscaled.
observational datasets, evaluated via standard
engineering performance metrics
R2, MAE, RMSESVM
[43]NaNRBF learning: unsupervised clustering for
hidden units, followed by supervised
output layer training
Average Error, Accuracy, Distance Class
Deviations
RBF
Table 5. Training validation and results—post-dismantling management.
Table 5. Training validation and results—post-dismantling management.
ArticleDatasetSplit/TrainEvaluation MethodBest Model
[53]24080/20
Recursive Feature Elimination (RFE)
prioritized independent variables to
optimize model performance
MSE, RNSE, R2Bagging with higher precision for PM10
Decision Tree with higher precision for PM2.5
[74]70.000 paired patches of bi-temporal high-resolution remote-sensing images and pixel-level
annotations from 100 mining
sites worldwide
60% train, 30% test and 10% validationF1-Score, IoUMineNetCD outperformed 12 baselines, with the Swin-T variant achieving optimal performance.
[80]265 h of valid data was
collected
70/30
The LSTM was configured with a 7-feature
input layer, a 32-neuron hidden layer, and a
single output layer
RMSE, R2, MAE, MAPESTM
[37]41.381 measured datasets70/30
A 30% test split was implemented, reserving the final 300 data points as a hold-out set for
Markov Chain validation
RMSE, R, MAERF-MC
Table 6. Bias analysis—blasting phase.
Table 6. Bias analysis—blasting phase.
ArticlesParticipantsPredictorsOutcomeAnalysisGeneral
[52]MRLRLRMRMR
[28]MRLRLRMRMR
[79]MRMRLRMRMR
[27]LRLRLRLRLR
[44]MRLRLRMRMR
[69]MRLRLRLRMR
[61]MRMRLRHRMR
[70]MRMRLRMRMR
[31]MRMRLRMRMR
[58]MRLRLRMRMR
[32]MRLRLRMRMR
[33]MRMRLRLRMR
[34]MRLRLRMRMR
[81]MRLRLRHRMR
[35]MRLRLRMRMR
[36]MRLRLRMRMR
[67]LRMRLRLRLR
[64]LRLRLRLRLR
[46]MRLRLRMRMR
[59]MRLRLRMRMR
[60]MRLRLRMRMR
[65]HRLRLRMRMR
[62]HRLRLRMRMR
[38]LRLRLRLRLR
[47]MRLRLRMRMR
[48]MRLRLRMRMR
[66]LRLRLRMRLR
[63]MRLRLRLRLR
[49]MRLRLRMRMR
[73]MRLRLRMRMR
[56]LRLRLRLRLR
[68]MRLRLRMRMR
[75]MRMRLRMRMR
[40]MRLRLRMRMR
[78]MRMRLRMRMR
[50]MRLRLRMRMR
[51]LRLRLRMRLR
[71]LRLRLRMRLR
[77]MRLRLRLRLR
[57]MRLRLRLRLR
[18]MRLRLRMRMR
[76]MRLRLRMRMR
[42]MRLRLRMRMR
Table 7. Bias analysis—load and haul.
Table 7. Bias analysis—load and haul.
ArticleParticipantsPredictorsOutcomeAnalysisGeneral
[30]MRLRLRMRMR
[45]LRLRLRMRLR
[54]HRLRLRMRMR
[55]MRLRLRMRMR
[72]LRLRLRLRLR
[43]MRMRLRHRMR
Table 8. Bias analysis—post-dismantling management.
Table 8. Bias analysis—post-dismantling management.
ArticleParticipantsPredictorsOutcomeAnalysisGeneral
[53]MRLRLRMRMR
[74]LRLRLRLRLR
[80]HRLRLRMRMR
[37]MRLRLRMRMR
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Reis, V.B.; Baptista, J.S.; Duarte, J. Machine Learning in Surface Mining—A Systematic Review. Appl. Sci. 2026, 16, 3246. https://doi.org/10.3390/app16073246

AMA Style

Reis VB, Baptista JS, Duarte J. Machine Learning in Surface Mining—A Systematic Review. Applied Sciences. 2026; 16(7):3246. https://doi.org/10.3390/app16073246

Chicago/Turabian Style

Reis, Vasco Belo, João Santos Baptista, and Joana Duarte. 2026. "Machine Learning in Surface Mining—A Systematic Review" Applied Sciences 16, no. 7: 3246. https://doi.org/10.3390/app16073246

APA Style

Reis, V. B., Baptista, J. S., & Duarte, J. (2026). Machine Learning in Surface Mining—A Systematic Review. Applied Sciences, 16(7), 3246. https://doi.org/10.3390/app16073246

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop