MSS-YOLO: Multi-Scale Edge-Enhanced Lightweight Network for Personnel Detection and Location in Coal Mines
Abstract
:1. Introduction
- To address the omission and misdetection issues of traditional deep learning algorithms in miner localization for an integrated mining face, this paper proposes the MSS-YOLO model. It achieves real-time and stable miner detection under uneven illumination, feature blurring, and varying dust concentrations.
- The proposed MSS-YOLO model integrates a Multi-Scale Edge Enhancement (MSEE) module into the C2f module of YOLOv8 to strengthen edge representation and reduce missed or false detections caused by high dust conditions and jitter. We also construct a lightweight Shared Convolutional Detection Head (SCDH) based on group-normalized shared-weight convolution, which significantly reduces the computational overhead while maintaining accuracy. Furthermore, replacing the SPPF module with a Spatial Pyramid Shared Conv (SPSC) module reduces long-range target feature loss and decreases model redundancy.
- A UCMD coal mine underground miner localization image dataset is constructed by simulating various conditions in a fully mechanized mining face. Based on this dataset, the paper integrates MSS-YOLO with the SGBM binocular stereo vision matching algorithm to solve the 3D spatial positions of miners and verifies its performance in a simulated roadway. Experimental results show that MSS-YOLO detects miners in real time and with high accuracy under different dust concentrations and distances, while the binocular-based localization achieves excellent precision and stability.
2. Related Work
3. Methods
3.1. MSS-YOLO Network Structure
3.1.1. The Multi-Scale Edge Enhance Module
3.1.2. The Shared Convolutional Detection Head Module
3.1.3. Spatial Pyramid Shared Conv
3.2. Underground Coal Miner Dataset
3.3. Personnel Localization Method Based on SGBM Stereo Vision Matching Algorithm and MSS-YOLO
4. Experiments and Performance
4.1. Training Environment and Parameter Settings
4.2. Experimental Results and Analysis
4.2.1. Experimental Validation of MSS-YOLO-Based Miner Identification
4.2.2. Experimental Validation of Personnel Localization Method Based on SGBM Stereo Vision Matching Algorithm + MSS-YOLO
5. Discussion and Conclusions
- Aiming to address the problems that traditional deep learning algorithms face, such as being prone to target omission and misdetection under the complex working environment of coal mines and the inability to guarantee good real-time performance under limited computational resources, the MSS-YOLO model, suitable for the identification and detection of miners in a fully mechanized mining face, is proposed. Combining part of the miners’ data from the DsLMF+ dataset and miners’ data from different distances and different dust concentrations in a simulated roadway environment of a generalized mining face, we constructed the personnel localization dataset UCMD, which contains 7675 images, and manually labeled the miners. Compared with Faster-RCNN, SSD, and other mainstream versions of the YOLO series, the performance of the proposed MSS-YOLO model is significantly improved, and it is able to accomplish the personnel detection and identification tasks in real time and stably under different lighting, different distances, and different cutting dust concentrations.
- The proposed underground coal mine miner detection network MSS-YOLO was utilized for validation on the UCMD dataset. Compared with the benchmark model YOLOv8n, MSS-YOLO showed improvements of 3.655% on the AP50 metric, 0.494% on AP50-95, 1.336% on Recall, and 1.26% on accuracy and reduced the inference time by 1.2 ms. Although its Rec value was slightly lower than that of YOLOv11n, it outperformed YOLOv11n in all other metrics.
- In order to realize miner localization in the fully mechanized mining face, this paper combined MSS-YOLO and the SGBM binocular stereo vision matching algorithm. Based on the detection results of MSS-YOLO and the binocular vision disparity solution, the three-dimensional spatial coordinates of the center area of the anchor frame, where the miner was located, were located, and experimental validation of this miner localization method was carried out. The experimental results showed that in the measurement range of 3–10 m in 20 groups, the measurement errors of the positions in the x-, y-, and z-directions were within 0.170 m, 0.160 m, and 0.200 m, with an average error of 0.014 m, 0.008 m, and 0.043 m, respectively. Meanwhile, in terms of the measurement distances, the measurement errors of the distances were within 0.185 m, which meets the maximum permitted positioning error for personnel in the working face of the coal mine.
- From the above conclusions, it can be seen that the MSS-YOLO model proposed in this paper is capable of obtaining excellent results under the premise of ensuring good real-time performance in recognizing and detecting miners under the complex background of integrated mining faces in coal mines. However, when the vision sensor operates under extreme conditions such as artificial over-obscuration and severe exposure, it can lead to the loss of most of the features of the target miner, resulting in unstable detection results from MSS-YOLO and problems such as missed and false detections. At the same time, the SGBM binocular stereo vision matching algorithm is also prone to problems such as reduced matching accuracy, discontinuity, and differential leakage under such working conditions, which affects the accuracy of miner positioning. Therefore, in future work, we will focus on discussing and researching image target feature extraction and recovery under extreme coal mine conditions, such as the study of dust degradation modeling based on underground coal mine images and the study of quaternion-based [39] image brightness overexposure and underexposure correction, etc. Meanwhile, for problems such as the insufficient parallax matching solution accuracy of the SGBM binocular stereo vision matching algorithm under extreme conditions, we will study a fast and efficient parallax solving method to optimize and improve the stereo vision matching effect based on the combination of octonion-based transform moments with deep learning [40], in order to further improve the effectiveness of miner identification and localization.
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Liu, Z.; Li, C.; Wu, D.; Dai, W.; Geng, S.; Ding, Q. A Wireless Sensor Network Based Personnel Positioning Scheme in Coal Mines with Blind Areas. Sensors 2010, 10, 9891–9918. [Google Scholar] [CrossRef] [PubMed]
- Chen, W. Personnel precise positioning system of coal mine underground based on UWB. J. Phys. Conf. Ser. 2021, 1920, 012115. [Google Scholar] [CrossRef]
- Li, X.; Wang, S.; Liu, B.; Chen, W.; Fan, W.; Tian, Z. Improved YOLOv4 network using infrared images for personnel detection in coal mines. J. Electron. Imaging 2022, 31, 013017. [Google Scholar] [CrossRef]
- Yang, W.; Zhang, X.; Ma, B.; Wang, Y.; Wu, Y.; Yan, J.; Zhang, C.; Wan, J.; Wang, Y.; Huang, M.; et al. An open dataset for intelligent recognition and classification of abnormal condition in longwall mining. Sci. Data 2023, 10, 416. [Google Scholar] [CrossRef] [PubMed]
- Li, X.; Wang, S.; Chen, W.; Weng, Z.; Fan, W.; Tian, Z. An Intelligent Vision-Based Tracking Method for Underground Human Using Infrared Videos. Symmetry 2022, 14, 1750. [Google Scholar] [CrossRef]
- Zhang, H.; Han, W.; Xu, Y.; Wang, Z. Analysis on the Development Status of Coal Mine Dust Disaster Prevention Technology in China. J. Healthc. Eng. 2021, 2021, 574579. [Google Scholar] [CrossRef]
- Wang, Z.; Liu, Y.; Duan, S.; Pan, H. An efficient detection of non-standard miner behavior using improved YOLOv8. Comput. Electr. Eng. 2023, 112, 109021. [Google Scholar] [CrossRef]
- Hirschmuller, H. Stereo processing by semiglobal matching and mutual information. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 30, 328–341. [Google Scholar] [CrossRef]
- Zheng, C.; Pei, W.; Yan, Q.; Chong, Y. Pedestrian detection based on gradient and texture feature integration. Neurocomputing 2017, 228, 71–78. [Google Scholar] [CrossRef]
- Viola, P.; Jones, M. Rapid object detection using a boosted cascade of simple features. In Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), Kauai, HI, USA, 8–14 December 2001. [Google Scholar]
- Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005. [Google Scholar]
- Forsyth, D. Object Detection with Discriminatively Trained Part-Based Models. Computer 2014, 47, 6–7. [Google Scholar] [CrossRef]
- Bodla, N.; Singh, B.; Chellappa, R.; Davis, L.S. Soft-NMS—Improving Object Detection with One Line of Code. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
- Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. arXiv 2014, arXiv:1311.2524. [Google Scholar]
- Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [PubMed]
- Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Los Alamitos, CA, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
- Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. SSD: Single shot multibox detector. In Computer Vision–ECCV 2016, Proceedings of the 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part I 14; Springer International Publishing: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
- Tan, M.; Pang, R.; Le, Q. EfficientDet: Scalable and Efficient Object Detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2019), Seattle, WA, USA, 27 July 2020. [Google Scholar]
- Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-End Object Detection with Transformers. In Proceedings of the Computer Vision—ECCV 2020, Glasgow, UK, 23–28 August 2020; pp. 213–229. [Google Scholar]
- Jin, Y.; Zhang, Y.; Cen, Y.; Li, Y.; Mladenovic, V.; Voronin, V. Pedestrian detection with super-resolution reconstruction for low-quality image. Pattern Recognit. 2021, 115, 107846. [Google Scholar] [CrossRef]
- Esmaeil Abbasi, A.; Mangini, A.M.; Fanti, M.P. Object and Pedestrian Detection on Road in Foggy Weather Conditions by Hyperparameterized YOLOv8 Model. Electronics 2024, 13, 3661. [Google Scholar] [CrossRef]
- Liang, Z.; Wang, W.; Meng, R.; Yang, H.; Wang, J.; Gao, H.; Li, B.; Fan, J. Vehicle and Pedestrian Detection Based on Improved YOLOv7-Tiny. Electronics 2024, 13, 4010. [Google Scholar] [CrossRef]
- Liu, S.; Cao, L.; Li, Y. Lightweight Pedestrian Detection Network for UAV Remote Sensing Images Based on Strideless Pooling. Remote Sens. 2024, 16, 2331. [Google Scholar] [CrossRef]
- Hsu, W.-Y.; Lin, W.-Y. Ratio-and-Scale-Aware YOLO for Pedestrian Detection. IEEE Trans. Image Process. 2021, 30, 934–947. [Google Scholar] [CrossRef]
- Li, X.; Chen, S.; Tian, C.; Zhou, H.; Zhang, Z. M2FNet: Mask-Guided Multi-Level Fusion for RGB-T Pedestrian Detection. IEEE Trans. Multimed. 2024, 26, 8678–8690. [Google Scholar] [CrossRef]
- Zhang, H.; Mu, C.; Ma, X.; Guo, X.; Hu, C. MEAG-YOLO: A Novel Approach for the Accurate Detection of Personal Protective Equipment in Substations. Appl. Sci. 2024, 14, 4766. [Google Scholar] [CrossRef]
- Ying, Y.; Wang, H.; Zhou, H. Research on Abnormal Behavior Monitoring in University Laboratories Based on Video Analysis Technology. Appl. Sci. 2024, 14, 9374. [Google Scholar] [CrossRef]
- Wang, Y.; Zou, X.; Shi, J. YOLOv5-based Dense Small Target Detection Algorithm for Aerial Images Using DIOU-NMS. Radioengineering 2024, 33, 12–22. [Google Scholar] [CrossRef]
- Chen, D.; Xiong, S.; Guo, L. Research on Detection Method for Tunnel Lining Defects Based on DCAM-YOLOv5 in GPR B-Scan. Radioengineering 2023, 32, 299–311. [Google Scholar] [CrossRef]
- Varghese, R.; Sambath, M. YOLOv8: A Novel Object Detection Algorithm with Enhanced Performance and Robustness. In Proceedings of the 2024 International Conference on Advances in Data Engineering and Intelligent Computing Systems (ADICS), Chennai, India, 18-19 April 2024; pp. 1–6. [Google Scholar]
- Tian, Z.; Shen, C.; Chen, H.; He, T. FCOS: Fully convolutional one-stage object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9627–9636. [Google Scholar]
- Zhang, Z. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar] [CrossRef]
- Scharstein, D.; Szeliski, R.; Zabih, R. A Taxonomy And Evaluation Of Dense Two-Frame Stereo Correspondence Algorithms. Int. J. Comput. Vis. IJCV 2000, 47, 7–42. [Google Scholar] [CrossRef]
- Everingham, M.; Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The Pascal Visual Object Classes (VOC) Challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef]
- Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef]
- Zhao, H.; Zhang, H.; Zhao, Y. Yolov7-sea: Object detection of maritime uav images based on improved yolov7. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2023; pp. 233–238. [Google Scholar]
- Khanam, R.; Hussain, M. YOLOv11: An Overview of the Key Architectural Enhancements. arXiv 2024, arXiv:2410.17725. [Google Scholar]
- Tahiri, M.A.; Amakdouf, H.; El mallahi, M.; Qjidaa, H. Optimized Quaternion Radial Hahn Moments Application to Deep Learning for the Classification of Diabetic Retinopathy. Multimed. Tools Appl. 2023, 82, 46217–46240. [Google Scholar] [CrossRef]
- Tahiri, M.A.; Boudaaoua, B.; Karmouni, H.; Tahiri, H.; Oufettoul, H.; Amakdouf, H.; Qjidaa, H.; Abouhawwash, M.; Askar, S.S.; Sayyouri, M. Octonion-based transform moments for innovative stereo image classification with deep learning. Complex Intell. Syst. 2024, 10, 3493–3511. [Google Scholar] [CrossRef]
Model | MSEE | SCDH | SPSC | AP/50% | AP50-95/% | Rec/% | Pre/% | Latency (ms) |
---|---|---|---|---|---|---|---|---|
YOLOv8n | × | × | × | 0.9326 | 0.69922 | 0.92015 | 0.94908 | 7.8 |
YOLOv8n-M | √ | × | × | 0.96062 | 0.70291 | 0.93138 | 0.96012 | 7.5 |
YOLOv8n-S | × | √ | × | 0.95783 | 0.69933 | 0.92909 | 0.94992 | 5.6 |
YOLOv8-MS | √ | √ | × | 0.95891 | 0.70180 | 0.93000 | 0.95919 | 6.9 |
MSS-YOLO (ours) | √ | √ | √ | 0.96915 | 0.70416 | 0.93351 | 0.96168 | 6.6 |
Model | AP50/% | AP50-95/% | Rec/% | Pre/% | Latency (ms) |
---|---|---|---|---|---|
Faster-RCNN [36] | 0.92451 | 0.54902 | 0.90636 | 0.92210 | 57.0 |
SSD [18] | 0.86431 | 0.49625 | 0.85257 | 0.85332 | 18.5 |
YOLOv5s6 [17] | 0.93164 | 0.68271 | 0.91323 | 0.92800 | 10.3 |
YOLOv7-tiny [37] | 0.92482 | 0.67733 | 0.91912 | 0.91823 | 8.6 |
YOLOv8n [31] | 0.93260 | 0.69922 | 0.92015 | 0.94908 | 7.8 |
YOLOv11n [38] | 0.94727 | 0.70224 | 0.93598 | 0.95077 | 8.9 |
MSS-YOLO (ours) | 0.96915 | 0.70416 | 0.93351 | 0.96168 | 6.6 |
Parameter | minDisparity | numDisparities | blockSize | speckleWindowSize | Mode |
---|---|---|---|---|---|
Value | 1 | 512 | 9 | 100 | STEREO_SGBM_MODE_HH4 |
Groups | Total Station Data (xyz) | Camera Data (xyz) | X-Direction Error (m) | Y-Direction Error (m) | Z-Direction Error (m) | Practical Distance (m) | Measuring Distance (m) | Distance Error (m) |
---|---|---|---|---|---|---|---|---|
1 | (−0.77, −3.31, 1.69) | (−0.79, −3.32, 1.59) | 0.02 | 0.01 | 0.1 | 3.7954 | 3.7649 | 0.0305 |
2 | (−0.68, −3.30, 1.68) | (−0.65, −3.31, 1.58) | −0.03 | 0.01 | 0.1 | 3.7649 | 3.7249 | 0.0400 |
3 | (−0.98, −3.82, 1.65) | (−0.81, −3.77, 1.70) | −0.17 | −0.05 | −0.05 | 4.2749 | 4.2141 | 0.0608 |
4 | (−1.93, −4.11, 1.60) | (−1.90, −4.05, 1.58) | −0.1 | −0.1 | 0.02 | 4.8142 | 4.7443 | 0.0699 |
5 | (−2.59, −4.45, 1.61) | (−2.46, −4.51, 1.46) | −0.13 | 0.06 | 0.15 | 5.3947 | 5.3407 | 0.0540 |
6 | (−3.44, −4.33, 1.63) | (−3.38, −4.22, 1.57) | −0.06 | −0.11 | 0.06 | 5.7653 | 5.6301 | 0.1352 |
7 | (−1.51, −5.37, 1.62) | (−1.60, −5.52, 1.67) | 0.09 | 0.15 | −0.05 | 5.8087 | 5.9849 | 0.1762 |
8 | (−4.32, −4.29, 1.64) | (−4.31, −4.30, 1.44) | −0.01 | 0.01 | 0.2 | 6.3052 | 6.2562 | 0.0490 |
9 | (−3.77, −5.60, 1.66) | (−3.75, −5.49, 1.63) | −0.02 | −0.11 | 0.03 | 6.9519 | 6.8453 | 0.1066 |
10 | (−3.92, −5.91, 1.60) | (−4.01, −5.82, 1.55) | 0.09 | −0.09 | 0.05 | 7.2701 | 7.2357 | 0.0344 |
11 | (−4.73, −6.65, 1.67) | (−4.76, −6.60, 1.79) | 0.03 | −0.05 | −0.12 | 8.3297 | 8.3320 | 0.0023 |
12 | (−5.06, −7.28, 1.66) | (−5.13, −7.33, 1.80) | 0.07 | 0.05 | −0.14 | 9.0198 | 9.1261 | 0.1063 |
13 | (−1.29, −3.78, 1.68) | (−1.32, −3.87, 1.77) | 0.03 | 0.1 | 0.09 | 4.3330 | 4.4556 | 0.1226 |
14 | (−2.74, −4.94, 1.63) | (−2.68, −4.88, 1.46) | −0.06 | −0.06 | 0.17 | 5.8795 | 5.7557 | 0.1238 |
15 | (−3.04, −5.24, 1.68) | (−3.14, −5.40, 1.58) | 0.1 | 0.16 | 0.1 | 6.2866 | 6.4433 | 0.1567 |
16 | (−4.42, −6.23, 1.59) | (−4.33, −6.09, 1.48) | −0.09 | −0.14 | 0.11 | 7.8024 | 7.6176 | 0.1848 |
17 | (−4.85, −6.95, 1.64) | (−5.00, −7.02, 1.66) | 0.15 | 0.07 | −0.02 | 8.6321 | 8.7770 | 0.1449 |
18 | (−3.37, −5.99, 1.66) | (−3.32, −6.04, 1.61) | −0.05 | 0.05 | 0.05 | 7.0705 | 7.0779 | 0.0074 |
19 | (−3.82, −6.34, 1.60) | (−3.68, −6.51, 1.62) | 0.14 | 0.17 | −0.02 | 7.5728 | 7.6516 | 0.0788 |
20 | (−5.56, −7.62, 1.57) | (−5.49, −7.61, 1.37) | −0.07 | −0.01 | −0.2 | 9.5626 | 9.7020 | 0.1394 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Yang, W.; Wang, Y.; Zhang, X.; Zhu, L.; Wang, T.; Chi, Y.; Jiang, J. MSS-YOLO: Multi-Scale Edge-Enhanced Lightweight Network for Personnel Detection and Location in Coal Mines. Appl. Sci. 2025, 15, 3238. https://doi.org/10.3390/app15063238
Yang W, Wang Y, Zhang X, Zhu L, Wang T, Chi Y, Jiang J. MSS-YOLO: Multi-Scale Edge-Enhanced Lightweight Network for Personnel Detection and Location in Coal Mines. Applied Sciences. 2025; 15(6):3238. https://doi.org/10.3390/app15063238
Chicago/Turabian StyleYang, Wenjuan, Yanqun Wang, Xuhui Zhang, Le Zhu, Tenghui Wang, Yunkai Chi, and Jie Jiang. 2025. "MSS-YOLO: Multi-Scale Edge-Enhanced Lightweight Network for Personnel Detection and Location in Coal Mines" Applied Sciences 15, no. 6: 3238. https://doi.org/10.3390/app15063238
APA StyleYang, W., Wang, Y., Zhang, X., Zhu, L., Wang, T., Chi, Y., & Jiang, J. (2025). MSS-YOLO: Multi-Scale Edge-Enhanced Lightweight Network for Personnel Detection and Location in Coal Mines. Applied Sciences, 15(6), 3238. https://doi.org/10.3390/app15063238