Next Article in Journal
Exploring the Effects of Acute Digital Sports Dance Intervention on Children’s Gross Motor Development, Executive Function, and Muscle Coordination Using Electromyography Sensors: A Randomized Repeated-Measures Study
Previous Article in Journal
Dual-Head Pix2Pix Network for Material Decomposition of Conventional CT Projections with Photon-Counting Guidance
Previous Article in Special Issue
Comprehensive Review of Deep Learning Approaches for Single-Image Super-Resolution
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Sensor Based Waste Rock Detection Method in Copper Mining Under Low Light Environment

1
Institute of Minerals Research, University of Science and Technology Beijing, Beijing 100083, China
2
School of Resources and Safety Engineering, University of Science and Technology Beijing, Beijing 100083, China
3
School of Economics and Management, University of Science and Technology Beijing, Beijing 100083, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(19), 5961; https://doi.org/10.3390/s25195961
Submission received: 30 August 2025 / Revised: 23 September 2025 / Accepted: 23 September 2025 / Published: 25 September 2025

Abstract

During production, copper mining could generate substantial waste rock that impacts land use and the environment. Advances in deep learning have enabled efficient, cost-effective intelligent sorting, where vision sensor performance critically determines sorting accuracy and efficiency. However, the sorting environment of copper mine waste rock is inherently complex, particularly within the conveyor belt section of the sorting machine, where insufficient and uneven lighting significantly impairs the performance of vision-based detection systems. To address the above challenges, a deep-learning-based copper mine waste rock detection algorithm under low-light environments is proposed. Firstly, an Illumination Adaptive Transformer (IAT) module is added as a preprocessing layer at the beginning of the Backbone to enhance the brightness of the images acquired by the vision sensor. Secondly, a Local Enhancement-Global Modulation (LEGM) module is integrated after the A2C2f and C3k2 modules in the Neck to enhance the detection accuracy. Finally, to further improve the model performance, MPDIoU is introduced to optimize the original loss function CIoU. As a result, the proposed algorithm achieved an mAP@0.5 of 0.957 and an mAP@0.5:0.95 of 0.689, outperforming advanced methods by 1.9% and 8.6%, respectively.

1. Introduction

Copper plays a crucial role in supporting the development of modern society, technology and infrastructure [1]. It is widely used in various industries, such as electric power, machinery, construction, transportation, etc. [2], making the efficient and sustainable utilization of copper ore resources particularly important. During the production process, waste rock should be separated from copper minerals as early as possible to conserve energy and time in subsequent processing steps. Therefore, sorting of copper ore waste rock is not only an urgent need for environmental protection but also an important strategy to achieve sustainable development of the mining industry [3,4,5].
With the rapid development of the global technological revolution and the advancement of industrialization, technologies in waste rock sorting are also evolving continuously. The traditional manual sorting model, characterized by high labor intensity, low efficiency, and poor precision under harsh working conditions, can no longer meet the demands of high-quality industry development. With the widespread use of deep learning in object detection, computer-vision-based intelligent sorting technology has emerged, providing an efficient solution for the automated sorting of copper mine waste rock [6,7]. However, practical implementation still faces significant technical challenges, the most significant being low and uneven lighting inside sorting machines. Such conditions make it difficult for image sensors to capture ore features accurately, leading to lower image quality, loss of detail, and frequent detection errors. Therefore, developing a real-time detection method for waste rocks on conveyor belts under low-light environments is critical for mineral processing development [8,9].
Current research on mineral waste rock sorting is primarily based on computer vision classification algorithms. With various image-based learning methods [10], it can preprocess input mineral images and construct classification models to achieve mineral sorting. For example, Liu Y et al. [11] designed a lightweight ore sorting network (LOSN), which integrates lightweight structures, attention mechanisms, and multi-scale feature fusion techniques to achieve end-to-end localization and classification of ore particles. Shatwell et al. [12] developed an algorithm based on image processing and machine learning, specifically designed for classifying high-grade ore rocks and low-grade waste rocks in gold and silver mines, enabling real-time ore sorting. Tsangaratos et al. [13] proposed a real-time mineral classification system that integrates classical computer vision techniques with advanced deep learning algorithms. This system can identify multiple minerals within a single frame and output probabilities for multiple mineral types, facilitating real-time mineral classification. However, existing methods heavily rely on the quality of input image. This dependence not only makes image acquisition challenging under harsh working conditions but also significantly increases recognition time and reduces sorting efficiency. Low-light or uneven-lighting conditions in actual sorting environments further degrade image quality, which makes feature extraction more difficult. In recent years, numerous scholars from various fields have developed methods to enhance real-time target detection in low-light environments by improving the input image quality, focusing primarily on image enhancement, small-target detection, noise suppression, and target deblurring.
Regarding image enhancement, Hui Y et al. [14] proposed the WSA-YOLO algorithm, a weakly supervised adaptive object detection algorithm for low-light environments based on YOLOv7. This algorithm decomposes images into reflectance and illumination maps for separate enhancement through a decomposition network, addressing performance degradation in object detection caused by the loss of fundamental image features under low-light conditions and the imbalance in low-level semantic information between low-light and normal images. Zhou Y et al. [15] introduced an improved spatial object detection method based on YOLO11, which enhances image quality by applying bilateral filtering to remove noise while preserving edge texture details and using Contrast Limited Adaptive Histogram Equalization (CLAHE) to improve image contrast. Zhang Q et al. [16] proposed the multi-module network LLD-YOLO for robust vehicle detection in low-light conditions. It enhances low-light images by integrating a DarkNet module based on self-calibrated illumination learning for adaptive illumination adjustment.
In the small target detection field, Zhang M et al. [17] proposed a Gaussian Curvature-Inspired Network (GCI-Net) for single-frame infrared small target detection, addressing the difficulty of separating small infrared targets from backgrounds with high brightness and strong edges in complex environments where they share similar physical characteristics. Wang S et al. [18] introduced a lightweight infrared small target detection algorithm called PHSI-RTDETR for drone aerial imagery, tackling issues of low model accuracy and high computational complexity caused by complex ground environments and uneven target scales in infrared image target detection. Wang Z et al. [19] proposed the PC-YOLO11s model to overcome detection difficulties in small target scenarios arising from factors such as small size, low resolution, weak contrast, and significant background interference. Fang, W. et al. [20] proposed a lightweight TBF-YOLOv8n tea bud detection model based on an improved YOLOv8n to address issues such as the high computational complexity of deep learning detection models and the difficulty of detecting tea buds due to their small size, complex background, and high density.
Regarding noise suppression and target blurring issues. Peng D et al. [21] proposed NLE-YOLO, a novel low-light object detection method based on YOLOv5 with fused feature enhancement, to overcome accuracy issues caused by insufficient illumination and noise interference in low-light environments. Das P P et al. [22] introduced YOLO-D, a domain-adaptive low-light object detection framework based on YOLOv8, designed to mitigate performance degradation in conventional detection systems due to low visibility, poor contrast, and high noise in low-light conditions. Liu Q et al. [23] developed an efficient object detection network architecture suitable for complex low-light environments, tackling low accuracy issues caused by blurring and low contrast during target detection under weak light. Han Z et al. [24] proposed 3L-YOLO, a lightweight low-light object detection algorithm based on YOLOv8n, addressing performance degradation and high computational resource consumption resulting from low contrast, high noise, and blurred boundaries in low-light detection. Ye S et al. [25] presented YES, a novel low-light object detection framework, to resolve the “target-background confusion” issue caused by uneven illumination in dim environments. Qi Z et al. [26] proposed the HSG-Net model, based on the lightweight YOLOv8, aiming to enhance detection accuracy and reduce computational costs in variable lighting conditions and complex backgrounds.
Additionally, some researchers have focused on improving real-time performance and computational efficiency from a methodological perspective rather than relying solely on input quality. Zhao Y et al. [27] proposed a Real-time Efficient Enhancement Network (RTEE-Net) for low-light images, addressing the trade-off between model complexity and performance in low-light image enhancement. Qiu Z et al. [28] introduced GS-YOLO-Seg, an improved YOLO11-Seg-based method, to resolve issues of low efficiency in traditional approaches and performance bottlenecks on edge devices during low-grade graphite ore sorting. Li J et al. [29] developed an enhanced model based on YOLOv5, featuring a parallel convolutional architecture, which improves the accuracy and efficiency of photovoltaic panel detection by addressing challenges arising from low resolution, small target sizes, and complex backgrounds in remote sensing imagery.
Based on a comprehensive analysis of existing research, various algorithms demonstrate certain advantages within their respective low-light application contexts. However, some inadequately extract comprehensive features when sorting copper minerals in complex and dim environments. Others are tailored to particular scenarios and struggle to adapt to specialized operational conditions such as conveyor belts, while certain methods impose demanding hardware requirements due to high model complexity, failing to meet real-time sorting needs.
This study proposes a novel method for identifying waste rock in copper ore within conveyor systems, specifically designed for operation under low-light conditions. Industrial experiments were conducted in environments free from overexposure or significant noise interference. Furthermore, the use of high-frame-rate industrial cameras minimized image distortion caused by abnormal exposure and noise. Consequently, the core objective of this research is to overcome recognition challenges in low-light environments. To address issues such as low detection accuracy, high false negative rates, and high false positive rates caused by blurred images resulting from dim lighting and uneven illumination, we introduce the YOLO-ILM algorithm. This algorithm enables real-time detection within low-light conveyor belt environments. Building upon the advanced YOLOv12 framework [30], YOLO-ILM leverages its efficient feature extraction and target localization capabilities. The algorithm further integrates a series of image enhancement and feature optimization techniques specifically designed for low-light and blurry scenarios. Through synergistic module collaboration, it significantly enhances adaptability to complex lighting conditions while ensuring detection speed meets the industrial requirements. The model introduces the following innovations:
(1)
To address image quality degradation caused by low-light conditions, an Illumination-Adaptive Transformer module is proposed and integrated as a preprocessing layer at the network front-end to enhance the brightness of subsequent input images.
(2)
To improve the detection accuracy of copper ore waste rocks, a method is introduced that integrates local feature embedding into global feature extraction modules following the A2C2f and C3k2 modules.
(3)
To further enhance the detection accuracy of copper ore waste rocks, the original loss function has been refined, optimizing the network.

2. Materials and Methods

The proposed YOLO-ILM algorithm for low-light copper mine waste rock detection is depicted in Figure 1, which consists of three core components: the Backbone, the Neck, and the Head.
In the Backbone section, to address the issue of poor input quality under low-light conditions, the Illumination Adaptive Transformer module (IAT) [31] is incorporated as a preprocessing layer at the front-end of the Backbone. This module effectively mitigates the adverse effects of uneven illumination, thereby providing high-quality input for subsequent copper mine waste rock detection tasks. Its lightweight design also ensures efficient operation under computational resource constraints.
In the Neck section, the preprocessed waste rock images from copper mines are input into the Backbone for feature extraction after IAT preprocessing. The extracted features subsequently enter the Neck for feature fusion. To more comprehensively integrate multi-scale features and enhance the model’s discriminative capability for waste rock characteristics, the Local Feature-Embedded Global Feature Extraction Module (LEGM) [32] is integrated after both the A2C2f and C3k2 modules within the Neck. By combining self-attention and convolution, it integrates local and global features; by aggregating multi-source inputs and depth information, it achieves multi-scale feature fusion. This enables the model to adaptively learn dependencies between different features, enhances focus on critical regions, and suppresses noise interference. The fused features are then forwarded to the Head for final prediction.
In the Head section, the Minimum Points Distance IoU (MPDIoU) loss function [33] is introduced, which places greater emphasis on the positional relationships between bounding boxes and their overlapping regions. This further improves the detection accuracy for copper mine waste rock.

2.1. Illumination Adaptive Transformer Module

Copper mine waste rock detection in real-world environments suffers from low-light and insufficient-light conditions, resulting in degraded visual quality that adversely affects downstream computer vision tasks. Images captured under weak or insufficient lighting exhibit darkness due to limited photon counts and in-camera noise. Therefore, it is necessary to pre-integrate an illumination-adaptive preprocessing module. The method of “optimizing input quality first and then improving model performance” has been verified in many fields [34]. Numerous solutions have been proposed in existing research, such as low-light enhancement or methods improving robustness in low-light and over-exposed conditions with oversaturation. However, these approaches have limitations and shortcomings. Low-light enhancement methods, while recovering details and suppressing accompanying noise, may inadvertently obscure fine details. Furthermore, when local image areas are severely over-exposed (pure white) or under-exposed (pure black), the original scene’s brightness information is irretrievably lost. Methods enhancing robustness to oversaturation under low-light and over-exposure conditions suffer from high parameter complexity and significant computational cost.
In contrast, the Illumination Adaptive Transformer (IAT) addresses both low-light enhancement and exposure correction through its designed lightweight dual-branch structure. IAT simulates the camera Image Signal Processor (ISP) pipeline, as illustrated in Figure 2, dividing image enhancement into two complementary components: local pixel-level illumination adjustment and global dynamic adjustment of ISP parameters.
For an input sRGB image Ii captured under lighting condition Li, illumination-induced degradation is addressed through the Illumination Adaptive Transformer (IAT) module via its two core components: local pixel-level adjustment and global ISP parameter adjustment.
First, illumination effects are corrected while preserving image details via the Local Branch. This branch employs a transformer-style architecture, processing the image while maintaining the input resolution. Its core consists of stacked Pixel Enhancement Modules (PEMs). Each PEM first encodes positional information using a 3 × 3 depth-wise convolution. Subsequently, a Pointwise Convolution (PWConv)—Depth-wise Convolution (DWConv)—Pointwise Convolution (PWConv) structure enhances local details. Two separate 1 × 1 convolutions then strengthen feature representations. Innovatively, standard Layer Normalization is replaced with Illumination Normalization (I Norm), which learns scaling factor a and bias term b applied before channel fusion. Three PEM modules are stacked within the branch. Their outputs are fused with the input features via skip connections (element-wise addition), effectively preserving original details. Finally, a 3 × 3 convolution reduces dimensionality, followed by the application of ReLU (yielding a multiplicative map M) and Tanh (yielding an additive map A) activation functions. This design achieves a balance between lightweight implementation and detail enhancement.
Second, parameter modulation is performed by the Global Branch. This branch forms the core of the IAT model for handling ISP pipeline operations and generating the target sRGB image. It focuses on capturing global interactions while maintaining lightweight efficiency. The process is as follows: First, a lightweight encoder, composed of two stacked convolutional layers, encodes the input image. This design reduces feature map resolution to minimize computational cost while extracting high-dimensional global features. The encoded features are then fed into the Global Prediction Module (GPM). Within the GPM, a globally initialized component query (Q) attends to keys (K) and values (V) generated from the encoded features. Notably, positional encoding for K and V is implemented via depth-wise convolution, enabling flexible adaptation to varying input resolutions. The features are subsequently processed by a Feed-Forward Network (FFN) containing two linear layers. Finally, two specially initialized parameters are appended, outputting the joint color transformation matrix (W) and the gamma correction parameter (γ). The initialization strategy (setting the color matrix to the identity matrix and γ to 1) effectively ensures training stability. This entire structure dynamically controls global ISP-related parameters (color correction, gamma correction) through a transformer-style query mechanism. This enables adaptation to diverse lighting conditions while rigorously maintaining the model’s lightweight nature and computational efficiency.
When the IAT module is not incorporated into the model, the image input is expressed as shown in Equation (1):
I n p u t = I i   = I r a w
After incorporating the IAT module into the model, the preprocessed image output by the IAT module is expressed as shown in Equation (2):
I t   = m a x ( c j W c i , c j   ( I i   M + A ) , 0 ) γ
Functioning as a preprocessing layer, the IAT module transforms the raw input Iraw into an illumination-enhanced image, facilitating subsequent feature extraction by convolutional layers. The enhancement effect is visualized in Figure 3. Figure 3a,b depicts the original image and the IAT-processed result (denoted as IAT Output), respectively. Across diverse scenarios—including Isolated (single waste rock), Plural (multiple waste rocks), Small (smaller waste rocks), Large (larger waste rocks), and notably Extremely Dark (severe low-light conditions commonly encountered in areas like conveyor belt edges where illumination is insufficient and ambient brightness is critically low)—the comparison between original and processed images demonstrates the IAT module’s pronounced enhancement effect on low-illumination imagery. Additionally, this study employs Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index (SSIM) as primary metrics for evaluating image enhancement performance. Higher PSNR values indicate smaller differences between the enhanced image and the reference image, while SSIM values closer to 1 signify stronger consistency in image structure and texture. Table 1 presents experimental comparisons between IAT and representative enhancement methods on this dataset. These results further validate IAT’s superior image enhancement performance in this study.

2.2. LEGM Attention Mechanism

Building upon the base model, the LEGM attention mechanism is introduced to enhance both detail preservation and global context modeling. The fundamental principle of self-attention mechanisms is that when processing sequential data, each element can establish relationships with any other element in the sequence, rather than relying solely on adjacent elements. It adaptively captures long-range dependencies by calculating the relative importance between elements. Consequently, the self-attention module computes association weights between all positions in the feature map, dynamically aggregating global information.
However, conventional self-attention suffers from a key limitation: it focuses exclusively on global dependencies while neglecting local details, often resulting in structurally blurred restored images. To address this, the LEGM module integrates convolutional layers for local feature embedding. This design combines the strengths of convolutional layers—excellent at local feature extraction—with the self-attention module’s ability to model global context. By fusing these capabilities, LEGM effectively models global haze distribution patterns while preserving fine details, thereby significantly enhancing the visual realism of the dehazed results.
The LEGM module integrates a self-attention mechanism with a convolutional module. It concurrently receives and concatenates shallow, mid-level, and deep features. During the feature processing stage, the concatenated features first pass through convolutional layers and normalization layers to explicitly extract local structures, thereby overcoming the inherent local receptive field limitations of traditional self-attention. These features subsequently transform a Multilayer Perceptron (MLP), are further processed by a linear layer, and finally pass through a Softmax function to generate the attention weight matrix. In the feature fusion stage, this attention weight matrix is element-wise multiplied with the input features. The resulting features are then processed through convolutional and linear layers. To enhance the precision of attention focus—particularly on distant, high-haze-concentration areas—these processed features are additively combined with the original input features via a residual connection. Following this, a filtering operation is applied, and the features are finally output through another Multilayer Perceptron (MLP). By synergistically fusing convolution-based local feature embedding with depth-prior-guided self-attention, LEGM effectively achieves a balance between preserving local details and capturing global structural information for the image dehazing task, as illustrated in Figure 4.
Fusion of input features:
F i n   = C o n c a t ( C o n v 1 × 1 ( F u n e t   ) , C o n v 3 × 3 ( F u n e t   ) , D R D B ( F d e p t h   ) )
Self-attentive module:
Q = C o n v ( F i n   ) , K = C o n v ( F i n   ) , V = C o n v ( F i n   ) , d K   = d i m ( K )
F a t t = Softmax Q K T d K V
MLP transformation:
F m l p = M L P ( F a t t )
Output Equation:
F o u t = C o n v L a y e r N o r m F i n + F m l p
Following LEGM integration, the network output is:
F o u t 1 = L E G M A 2 C 2 f ( F i n 1 ) F o u t 2 = L E G M A 2 C 2 f ( F i n 2 ) F o u t 3 = L E G M C 3 K 2 ( F i n 3 )
After incorporating the LEGM module into the original model, the self- attentive module integrates the locally extracted convolutional features with global contextual information, enhancing feature representation capabilities and improving the utilization efficiency of image structure and depth information.

2.3. MPDIoU Loss Function

The original CIoU [38] loss function in the model considers both center-point distance and aspect ratio. However, it focuses solely on relative aspect ratio deviation rather than absolute size differences, severely constraining regression accuracy and convergence efficiency when predicted and ground-truth bounding boxes share identical aspect ratios but differ significantly in scale. In contrast, the MPDIoU loss function currently employed in our model fundamentally resolves this limitation by innovatively minimizing the distances between corresponding top-left and bottom-right corner points of predicted and ground-truth bounding boxes. This approach unifies the geometric properties of bounding boxes (defined by four corner coordinates) into a concise metric that comprehensively incorporates critical factors, including overlap area, center-point distance, and width-height deviations, while directly optimizing absolute size differences. Simultaneously, it streamlines the computational process, delivering an efficient and robust solution for bounding box regression.
The final CIoU loss is computed as follows:
L C I O U = L b a s e + α × 4 π 2 arctan w 1 h 1 arctan w 2 h 2 2
where L b a s e denotes the fundamental geometric error, α represents the dynamic weighting parameter, and their computational formulations are given, respectively, as follows:
L b a s e = ( 1 I o U ) + ρ 2 ( M 2 , M 1 ) c 2
α = 0 , if   I o U < 0.5 , V ( 1 I o U ) + V , if   I o U 0.5 .
where M1 = (xA, yA) and M2 = (xB, yB) denote the center coordinates of ground-truth box A and predicted box B, respectively; C represents the minimum enclosing rectangle covering both boxes, ρ corresponds to the Euclidean distance between M1 and M2. w1 and w2 indicate the widths of boxes A and B; h1 and h2 signify their corresponding heights.
The MPDIoU loss function is formulated as follows:
L M P D I o U = 1 M P D I o U
The core computational formulation of MPDIoU is defined as follows:
M P D I o U = I o U d 1 2 w 2 + h 2 + d 2 2 w 2 + h 2
where IoU measures the overlap, and d12 and d22 represent the distances from the top-left corner and the bottom-right corner, respectively. These are calculated as follows:
I o U = I U  
d 1 2 = ( x 1 B x 1 A ) 2 + ( y 1 B y 1 A ) 2  
d 2 2 = ( x 2 B x 2 A ) 2 + ( y 2 B y 2 A ) 2
Consequently, when employed in object detection, MPDIoU offers significant advantages. By comprehensively accounting for overlapping and non-overlapping regions, center point distance, and width-height deviations, it effectively resolves optimization inefficiencies that occur when bounding boxes share identical aspect ratios but differ in absolute dimensions, enabling precise discrimination between predicted boxes located inside and outside the ground truth. Furthermore, its computational efficiency stems from determining relevant factors directly from the four bounding box coordinates, minimizing computational overhead, and facilitating straightforward integration into detection algorithms. Finally, as shown in the Figure 5, MPDIoU accelerates training convergence, thus achieving higher accuracy in shorter timeframes while also delivering superior final detection precision.

2.4. Evaluation Metrics

To better evaluate the comprehensive performance of each algorithm, based on the specific characteristics of the copper ore waste rock detection scenario, this paper selects three primary evaluation metrics: Precision (P), Recall (R), and Mean Average Precision (mAP), where:
(1)
Precision (P) refers to the proportion of true positive samples among all samples predicted as positive by the model. A higher precision indicates a lower probability of misjudgment by the model. Its calculation formula is:
Precision = TP TP + FP
In the formula, TP (True Positive) represents the number of samples correctly predicted as positive by the model, i.e., samples that are positive and predicted as positive; FP (False Positive) represents the number of samples incorrectly predicted as positive by the model, i.e., samples that are negative but predicted as positive.
(2)
Recall (R) refers to the proportion of actual positive samples correctly predicted as positive by the model among all actual positive samples. A higher recall indicates a lower probability of missed detections by the model. Its calculation formula is:
Recall = TP TP + FN
In the formula, FN (False Negative) represents the number of samples incorrectly predicted as negative by the model, that is, samples that are positive but predicted as negative.
(3)
Average Precision (AP) refers to the area under the Precision-Recall curve. A higher AP indicates better model precision, recall, and overall performance. Mean Average Precision (mAP) is the average of AP values across all categories. The calculation formulas for AP and mAP are, respectively, as follows:
AP = 0 1 P ( R ) d R
mAP = 1 M i = 1 M AP i
In the formula, APi represents the average detection accuracy of the model for a specific category; M represents the total number of categories.

3. Results

3.1. Experimental Dataset

The dataset used in this experiment is a copper ore waste rock dataset collected from the sorting site, which contains a total of 1200 images of copper ore waste rocks on the conveyor belt. To fully evaluate the performance of the algorithm and model, the dataset is divided into a training set, a validation set, and a test set in an 8:1:1 ratio. Specifically, the training set includes 960 images, the validation set contains 120 images, and the test set comprises 120 images. Some samples of the dataset are shown in Figure 6.

3.2. Experimental Environment and Parameters

In this experiment, four mainstream object detection algorithms were selected: Faster R-CNN [39], YOLOv8 [40], YOLOv11 [41], and YOLOv12. During the experiment, each algorithm was loaded onto the copper ore waste rock dataset for training, ultimately resulting in four distinct object detection models.
The hardware and software environment for this experiment are detailed in Table 2. In terms of hardware, the core computing device featured a central processing unit (CPU) of Intel Core Ultra 9-275HX (Intel Corporation, Santa Clara, CA, USA) and a graphics processing unit (GPU) of NVIDIA GeForce RTX 5080 (NVIDIA Corporation, Santa Clara, CA, USA) discrete graphics card, equipped with the CUDA platform and its corresponding cuDNN library. On the software side, the experiment was conducted on the Windows 11 operating system, utilizing the Python programming language and the PyTorch deep learning framework. PyCharm was chosen as the integrated development environment for writing, debugging, and executing the code.
To fully leverage the computational capabilities of the graphics processor used in the experiment and balance model training speed, this experiment sets the input image size to 640 × 640, the number of training epochs to 200, and the batch size to 32, ensuring efficient model training.
To optimize the training process, the stochastic gradient descent (SGD) algorithm was selected as the optimizer. To enhance the convergence speed and stability of SGD, the momentum mechanism was employed, with the momentum parameter set to 0.937. To prevent the computer from allocating excessive computational resources at once—which could lead to exceeding the upper limit of the graphics processor’s memory and cause Out of Memory (OOM) errors—the total number of GPU worker threads was set to 4. This configuration balances computational resource allocation and ensures stable model training.

3.3. Comparative Experiments

To fully validate the superiority of the YOLO-ILM model algorithm, the experiment selected Faster R-CNN, YOLOv8, YOLOv11, and YOLOv12 as comparative models. The experimental results of training and prediction on the copper ore waste rock dataset were used for comparison. The experimental outcomes from training different models are presented in Table 3 and Figure 7. Table 3 displays the Precision (P), Recall (R), and mAP evaluation metrics for each model, while Figure 7 illustrates the mAP@0.5 and mAP@0.5:0.95 results curves across the models.
Combining Table 3 and Figure 7, it is evident that among all evaluated models, the YOLO series models collectively outperform Faster R-CNN in precision, recall, and mean average precision (mAP). Specifically, YOLO-ILM achieved optimal performance across four key evaluation metrics: precision (p = 0.948), recall (R = 0.941), mAP@0.5 (0.957), and mAP@0.5:0.95(0.689). Compared to the original model, it improved mAP@0.5 and mAP@0.5:0.95 by 1.9% and 8.6%, respectively. Additionally, as evidenced by inference time, although the YOLO-ILM model is 2.7 ms slower than the baseline model, our model can meet real-time detection requirements.
To further validate the effectiveness of the YOLO-ILM algorithm, the trained model is deployed to an on-site sorting machine and conducted field tests in complex low-light environments. The experimental results are shown in Figure 8, Figure 9, Figure 10, Figure 11 and Figure 12, where “rock” denotes detected copper ore waste. The results demonstrate that for multiple targets, Faster R-CNN and YOLOv8 exhibited partial missed detections (Figure 9) with low confidence, while YOLOv11 and YOLOv12 detected all targets, their confidence was lower than YOLO-ILM. Through comparison with other algorithm models, this fully demonstrates YOLO-ILM’s advanced performance in low-light target detection and its suitability for this task.

3.4. Ablation Experiments

To further validate the effectiveness of each module (IAT, LEGM, MPDIoU) in the proposed algorithm, ablation experiments were conducted on the added modules. The experimental results are shown in Table 4 and Figure 13. In Table 4, using the YOLOv12 model as the baseline, the effectiveness of each improvement method was evaluated by incrementally adding one enhancement at a time. The symbol “√” indicates that the corresponding improvement module has been incorporated into the model. Figure 13 presents a comparative analysis of mAP@0.5 and mAP@0.5:0.95 results between various model combinations in the ablation experiments and the baseline model.
Combining Table 4 and Table 5, and Figure 9, initially, the baseline model YOLOv12 was tested, achieving mAP@0.5 of 0.938 and mAP@0.5:0.95 of 0.603. Separately introducing the IAT module, LEGM attention mechanism, and MPDIoU loss function each increased both mAP@0.5 and mAP@0.5:0.95 compared to the baseline model, indicating individual contributions from each module. Next, pairwise combinations of modules demonstrated further improvements in mAP@0.5 and mAP@0.5:0.95, with a greater magnitude than single modules alone. This demonstrates synergistic effects when modules are combined pairwise, significantly enhancing the model’s object detection performance. Finally, integrating all three improved modules achieved peak results: precision (p = 0.948), recall (R = 0.941), mAP@0.5(0.957), and mAP@0.5:0.95(0.689), representing improvements of 1.9% and 8.6% in mAP@0.5 and mAP@0.5:0.95, respectively, over the baseline model. These results confirm maximal synergistic gains when all three modules operate jointly, delivering comprehensive optimal performance across all evaluation metrics.

4. Conclusions

This study addresses the issue of low target detection accuracy caused by poor image quality in low-light environments during copper ore waste rock processing. YOLO-ILM, a real-time detection algorithm for low-light copper ore waste, is proposed with core contributions including:
(1)
Integrating the YOLOv12 network with the Illumination Adaptive Transformer module IAT to enhance model adaptability to low-light conditions;
(2)
Introducing the Locally Embedded Global-feature attention mechanism LEGM to improve target feature capture capability under complex environmental interference;
(3)
Introducing an efficient and accurate bounding box regression loss function MPDIoU to reduce computational load while optimizing localization accuracy.
This method satisfies real-time detection requirements for copper ore waste in low-light scenarios. The experimental results demonstrate that through comparison with different detection algorithms, the proposed YOLO-ILM model shows significant advantages over Faster R-CNN, YOLOv8, YOLOv11, and YOLOv12 in both subjective visual detection effects and objective evaluation metrics. Compared to the baseline model, YOLO-ILM improves mAP@0.5 and mAP@0.5:0.95 by 1.9% and 8.6%, respectively, confirming the method’s superiority for real-time copper ore waste detection under complex conditions.

Author Contributions

Conceptualization, J.D. and F.Q.; methodology, W.Z., J.X., L.Z. and Y.J.; validation, J.D. and W.Z.; investigation, J.X., L.Z. and Y.J.; data curation, J.D., W.Z. and J.X.; writing—original draft preparation, J.D., W.Z. and J.X.; writing—review and editing, F.Q., W.Z. and L.Z.; supervision, J.D. and F.Q.; project administration, J.D. and F.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Key Research and Development Program of China under Grant 2023YFC3904203.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in this article; further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Moreau, K.; Laamanen, C.; Bose, R.; Shang, H.; Scott, J.A. Environmental impact improvements due to introducing automation into underground copper mines. Int. J. Min. Sci. Technol. 2021, 31, 1159–1167. [Google Scholar] [CrossRef]
  2. Tang, J.; Yang, H.; Chen, H.; Li, F.; Wang, C.; Liu, Q.; Zhang, Q.; Zhang, R.; Yu, L. Potential and future direction for copper resource exploration in China. Green Smart Min. Eng. 2024, 1, 127–131. [Google Scholar] [CrossRef]
  3. Wang, Y.; Chen, Q.; Dai, B.; Wang, D. Guidance and review: Advancing mining technology for enhanced production and supply of strategic minerals in China. Green Smart Min. Eng. 2024, 1, 2–11. [Google Scholar] [CrossRef]
  4. Liang, G.; Liang, Y.; Niu, D.; Shaheen, M. Balancing sustainability and innovation: The role of artificial intelligence in shaping mining practices for sustainable mining development. Resour. Policy 2024, 90, 104793. [Google Scholar] [CrossRef]
  5. Shiryayeva, O.; Suleimenov, B.; Kulakova, Y. Sustainable Mineral Processing Technologies Using Hybrid Intelligent Algorithms. Technologies 2025, 13, 269. [Google Scholar] [CrossRef]
  6. Quan, Z.; Sun, J. A Feature-Enhanced Small Object Detection Algorithm Based on Attention Mechanism. Sensors 2025, 25, 589. [Google Scholar] [CrossRef]
  7. Korie, J.; Gabriela, C.-A.; Ezeonyema, C.; Oshim, F. Machine Learning Innovations for Improving Mineral Recovery and Processing: A Comprehensive Review. Int. J. Econ. Environ. Geol. 2024, 15, 26–31. [Google Scholar]
  8. Corrigan, C.C.; Ikonnikova, S.A. A review of the use of AI in the mining industry: Insights and ethical considerations for multi-objective optimization. Extr. Ind. Soc. 2024, 17, 101440. [Google Scholar] [CrossRef]
  9. Tao, M.; Lv, S.; Feng, S. Study on the Evaluation of the Development Efficiency of Smart Mine Construction and the Influencing Factors Based on the US-SBM Model. Sustainability 2023, 15, 5183. [Google Scholar] [CrossRef]
  10. Shen, X.; Li, L.; Ma, Y.; Xu, S.; Liu, J.; Yang, Z.; Shi, Y. VLCIM: A vision-language cyclic interaction model for industrial defect detection. IEEE Trans. Instrum. Meas. 2025, 74, 1–13. [Google Scholar] [CrossRef]
  11. Liu, Y.; Wang, X.; Zhang, Z.; Deng, F. LOSN: Lightweight ore sorting networks for edge device environment. Eng. Appl. Artif. Intell. 2023, 123, 106191. [Google Scholar] [CrossRef]
  12. Shatwell, D.G.; Murray, V.; Barton, A. Real-time ore sorting using color and texture analysis. Int. J. Min. Sci. Technol. 2023, 33, 659–674. [Google Scholar] [CrossRef]
  13. Tsangaratos, P.; Ilia, I.; Spanoudakis, N.; Karageorgiou, G.; Perraki, M. Machine learning approaches for real-time mineral classification and educational applications. Appl. Sci. 2025, 15, 1871. [Google Scholar] [CrossRef]
  14. Hui, Y.; Wang, J.; Li, B. WSA-YOLO: Weak-supervised and Adaptive object detection in the low-light environment for YOLOV7. IEEE Trans. Instrum. Meas. 2024, 73, 1–12. [Google Scholar] [CrossRef]
  15. Zhou, Y.; Zhang, T.; Li, Z.; Qiu, J. Improved Space Object Detection Based on YOLO11. Aerospace 2025, 12, 568. [Google Scholar] [CrossRef]
  16. Zhang, Q.; Guo, W.; Lin, M. LLD-YOLO: A multi-module network for robust vehicle detection in low-light conditions. Signal Image Video Process. 2025, 19, 271. [Google Scholar] [CrossRef]
  17. Zhang, M.; Yue, K.; Li, B.; Guo, J.; Li, Y.; Gao, X. Single-Frame Infrared Small Target Detection via Gaussian Curvature Inspired Network. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–13. [Google Scholar] [CrossRef]
  18. Wang, S.; Jiang, H.; Li, Z.; Yang, J.; Ma, X.; Chen, J.; Tang, X. PHSI-RTDETR: A Lightweight Infrared Small Target Detection Algorithm Based on UAV Aerial Photography. Drones 2024, 8, 240. [Google Scholar] [CrossRef]
  19. Wang, Z.; Su, Y.; Kang, F.; Wang, L.; Lin, Y.; Wu, Q.; Li, H.; Cai, Z. PC-YOLO11s: A Lightweight and Effective Feature Extraction Method for Small Target Image Detection. Sensors 2025, 25, 348. [Google Scholar] [CrossRef] [PubMed]
  20. Fang, W.; Chen, W. TBF-YOLOv8n: A Lightweight Tea Bud Detection Model Based on YOLOv8n Improvements. Sensors 2025, 25, 547. [Google Scholar] [CrossRef]
  21. Peng, D.; Ding, W.; Zhen, T. A novel low light object detection method based on the YOLOv5 fusion feature enhancement. Sci. Rep. 2024, 14, 4486. [Google Scholar] [CrossRef]
  22. Das, P.P.; Ganguly, T.; Chaudhuri, R.; Deb, S. YOLO-D: A Domain Adaptive approach towards low light object detection. Procedia Comput. Sci. 2025, 258, 3042–3051. [Google Scholar] [CrossRef]
  23. Liu, Q.; Huang, W.; Hu, T.; Duan, X.; Yu, J.; Huang, J.; Wei, J. Efficient network architecture for target detection in challenging low-light environments. Eng. Appl. Artif. Intell. 2025, 142, 109967. [Google Scholar] [CrossRef]
  24. Han, Z.; Yue, Z.; Liu, L. 3L-YOLO: A Lightweight Low-Light Object Detection Algorithm. Appl. Sci. 2025, 15, 90. [Google Scholar] [CrossRef]
  25. Ye, S.; Huang, W.; Liu, W.; Chen, L.; Wang, X.; Zhong, X. YES: You should Examine Suspect cues for low-light object detection. Comput. Vis. Image Underst. 2025, 251, 104271. [Google Scholar] [CrossRef]
  26. Qi, Z.; Wang, J.; Yang, G.; Wang, Y. Lightweight YOLOv8-Based Model for Weed Detection in Dryland Spring Wheat Fields. Sustainability 2025, 17, 6150. [Google Scholar] [CrossRef]
  27. Zhao, Y.; Li, W.; Yang, R.; Liu, Y. Real-time efficient image enhancement in low-light condition with novel supervised deep learning pipeline. Digit. Signal Process. 2025, 165, 105342. [Google Scholar] [CrossRef]
  28. Qiu, Z.; Huang, X.; Sun, Z.; Li, S.; Wang, J. GS-YOLO-Seg: A Lightweight Instance Segmentation Method for Low-Grade Graphite Ore Sorting Based on Improved YOLO11-Seg. Sustainability 2025, 17, 5663. [Google Scholar] [CrossRef]
  29. Li, J.; Meng, X.; Wang, S.; Lu, Z.; Yu, H.; Qu, Z.; Wang, J. Enhanced Parallel Convolution Architecture YOLO Photovoltaic Panel Detection Model for Remote Sensing Images. Sustainability 2025, 17, 6476. [Google Scholar] [CrossRef]
  30. Tian, Y.; Ye, Q.; Doermann, D. Yolov12: Attention-centric real-time object detectors. arXiv 2025, arXiv:2502.12524. [Google Scholar]
  31. Cui, Z.; Li, K.; Gu, L.; Su, S.; Gao, P.; Jiang, Z.; Qiao, Y.; Harada, T. You only need 90k parameters to adapt light: A light weight transformer for image enhancement and exposure correction. arXiv 2022, arXiv:2205.14871. [Google Scholar] [CrossRef]
  32. Zhang, Y.; Zhou, S.; Li, H. Depth information assisted collaborative mutual promotion network for single image dehazing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 2846–2855. [Google Scholar]
  33. Ma, S.; Xu, Y. Mpdiou: A loss for efficient and accurate bounding box regression. arXiv 2023, arXiv:2307.07662. [Google Scholar] [CrossRef]
  34. Li, Y.; Fu, C.; Liu, T.; Hu, Z.; Song, G. Two-stage remaining useful life prediction method across operating conditions based on small samples and novel health indicators. Reliab. Eng. Syst. Saf. 2025, 264, 111290. [Google Scholar] [CrossRef]
  35. Guo, C.; Li, C.; Guo, J.; Loy, C.C.; Hou, J.; Kwong, S.; Cong, R. Zero-reference deep curve estimation for low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 1780–1789. [Google Scholar]
  36. Wei, C.; Wang, W.; Yang, W.; Liu, J. Deep retinex decomposition for low-light enhancement. arXiv 2018, arXiv:1808.04560. [Google Scholar] [CrossRef]
  37. Jha, R.; Lenka, A.; Ramanagopal, M.; Sankaranarayanan, A.; Mitra, K. RT-X Net: RGB-Thermal cross attention network for Low-Light Image Enhancement. arXiv 2025, arXiv:2505.24705. [Google Scholar]
  38. Zheng, Z.; Wang, P.; Ren, D.; Liu, W.; Ye, R.; Hu, Q.; Zuo, W. Enhancing Geometric Factors in Model Learning and Inference for Object Detection and Instance Segmentation. IEEE Trans. Cybern. 2022, 52, 8574–8586. [Google Scholar] [CrossRef]
  39. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  40. Varghese, R.; Sambath, M. Yolov8: A novel object detection algorithm with enhanced performance and robustness. In Proceedings of the 2024 International Conference on Advances in Data Engineering and Intelligent Computing Systems (ADICS), Chennai, India, 18–19 April 2024; pp. 1–6. [Google Scholar]
  41. Khanam, R.; Hussain, M. Yolov11: An overview of the key architectural enhancements. arXiv 2024, arXiv:2410.17725. [Google Scholar] [CrossRef]
Figure 1. YOLO-ILM Model Architecture Diagram (The red arrow represents the newly added module, whereas the remaining arrows indicate connections between the existing workflow steps.).
Figure 1. YOLO-ILM Model Architecture Diagram (The red arrow represents the newly added module, whereas the remaining arrows indicate connections between the existing workflow steps.).
Sensors 25 05961 g001
Figure 2. IAT Structure Diagram.
Figure 2. IAT Structure Diagram.
Sensors 25 05961 g002
Figure 3. Effect Diagram Before and After IAT Processing.
Figure 3. Effect Diagram Before and After IAT Processing.
Sensors 25 05961 g003
Figure 4. LEGM Structure Diagram.
Figure 4. LEGM Structure Diagram.
Sensors 25 05961 g004
Figure 5. Prediction Box Comparisons.
Figure 5. Prediction Box Comparisons.
Sensors 25 05961 g005
Figure 6. Instance Diagram.
Figure 6. Instance Diagram.
Sensors 25 05961 g006
Figure 7. Experimental Plots under Different Models.
Figure 7. Experimental Plots under Different Models.
Sensors 25 05961 g007
Figure 8. Isolated Detection Results.
Figure 8. Isolated Detection Results.
Sensors 25 05961 g008
Figure 9. Plural Detection Results.
Figure 9. Plural Detection Results.
Sensors 25 05961 g009
Figure 10. Small Detection Results.
Figure 10. Small Detection Results.
Sensors 25 05961 g010
Figure 11. Large Detection Results.
Figure 11. Large Detection Results.
Sensors 25 05961 g011
Figure 12. Extremely Dark Detection Results.
Figure 12. Extremely Dark Detection Results.
Sensors 25 05961 g012
Figure 13. Ablation Test Results.
Figure 13. Ablation Test Results.
Sensors 25 05961 g013aSensors 25 05961 g013bSensors 25 05961 g013c
Table 1. Quantitative Comparison of IAT and Other Mainstream Methods.
Table 1. Quantitative Comparison of IAT and Other Mainstream Methods.
MethodTest 1Test 2Test 3Test 4Test 5
PSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIM
Zero-DCE [35]9.2070.5159.4290.5289.9690.6029.3720.57310.4490.225
RetinexNet [36]6.7050.3246.7140.4507.0090.3116.8050.4546.5840.197
RT-X Net [37]10.2350.41510.5210.44810.8740.43210.2360.52911.1450.375
IAT15.1630.67915.4850.63315.4870.70515.4070.68616.0210.669
Table 2. Experimental Environment.
Table 2. Experimental Environment.
Hardware/Software UsedDescription
CPUIntel Core Ultra 9-275HX
GPUNVIDIA GeForce RTX 5080
GPU Parallel Processing PlatformCUDA 12.8
CUDA Deep Neural Network LibrarycuDNN v 8.9.7
Operating SystemWindows 11
Programming LanguagePython 3.12.7
Deep Learning FrameworkPyTorch 2.7.1
Integrated Development EnvironmentPyCharm 2025.1.3.1-Windows
Table 3. Experimental Results under Different Models.
Table 3. Experimental Results under Different Models.
ModelPrecision (P)Recall (R)mAP@0.5mAP@0.5:0.95GFLOPsMemory (G)Inference Times (ms)
Faster R-CNN0.8520.8760.9150.574181.436.229.56
YOLOv80.8740.8810.9230.58228.8011.711.41
YOLOv110.8800.8840.9350.59421.707.399.85
YOLOv120.9030.9100.9380.60319.607.026.73
YOLO-ILM0.9480.9410.9570.68945.29.929.43
Table 4. Ablation Test Results.
Table 4. Ablation Test Results.
TestIATLEGMMPDIoUPrecision (P)Recall (R)mAP@0.5mAP@0.5:0.95GFLOPsInference Times (ms)
1 0.9030.9100.9380.60319.606.73
2 0.9090.9120.9420.63224.206.88
3 0.9160.9140.9440.64927.308.87
4 0.9140.9150.9400.64119.607.13
5 0.9290.9250.9580.65745.209.31
6 0.9300.9340.9520.66324.208.92
7 0.9360.9290.9490.67627.309.34
80.9480.9410.9570.68945.209.43
Table 5. Improvement in Each Module of the Ablation Experiments.
Table 5. Improvement in Each Module of the Ablation Experiments.
Module CombinationΔPrecision(%)ΔRecall (%)ΔmAP@0.5 (%)ΔmAP@0.5:0.95 (%)
Baseline
IAT0.600.200.402.90
LEGM1.300.400.604.60
IAT + LEGM + MPDIoU1.100.500.203.80
IAT + LEGM2.601.502.005.40
IAT + MPDIoU2.702.401.406.00
LEGM + MPDIoU3.301.901.107.30
IAT + LEGM + MPDIoU4.503.101.908.60
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ding, J.; Qu, F.; Zhou, W.; Xu, J.; Zhao, L.; Ji, Y. A Sensor Based Waste Rock Detection Method in Copper Mining Under Low Light Environment. Sensors 2025, 25, 5961. https://doi.org/10.3390/s25195961

AMA Style

Ding J, Qu F, Zhou W, Xu J, Zhao L, Ji Y. A Sensor Based Waste Rock Detection Method in Copper Mining Under Low Light Environment. Sensors. 2025; 25(19):5961. https://doi.org/10.3390/s25195961

Chicago/Turabian Style

Ding, Jianing, Fuming Qu, Weihua Zhou, Jiajun Xu, Lingyu Zhao, and Yaming Ji. 2025. "A Sensor Based Waste Rock Detection Method in Copper Mining Under Low Light Environment" Sensors 25, no. 19: 5961. https://doi.org/10.3390/s25195961

APA Style

Ding, J., Qu, F., Zhou, W., Xu, J., Zhao, L., & Ji, Y. (2025). A Sensor Based Waste Rock Detection Method in Copper Mining Under Low Light Environment. Sensors, 25(19), 5961. https://doi.org/10.3390/s25195961

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop