Research on Inspection Method of Intelligent Factory Inspection Robot for Personnel Safety Protection
Abstract
:1. Introduction
- (a)
- This study constructs a self-built safety attire detection dataset for detecting safety attire in intelligent factory environments.
- (b)
- This research proposes an SFA module to enhance the detection speed and feature representation capability of the network.
- (c)
- Comparative and ablation experiments are conducted to verify the effectiveness of the proposed method.
- (d)
- Generalization experiments demonstrated that the proposed method exhibits strong generalization ability in the field of industrial inspection and can be extended to other detection tasks.
2. Related Works
3. Inspection Robot Platform and Data Resources
3.1. Inspection Robot Platform
3.2. Safety Attire Dataset
3.3. Loss Function and Evaluation Metrics
- Localization Loss
- 2.
- Confidence Loss
- 3.
- Classification Loss
- Precision Curve
- 2.
- Recall Curve
- 3.
- F1 Score Curve
- 4.
- Precision–Recall (PR) Curve:
- 5.
- Mean Average Precision (mAP):
- 6.
- Frames Per Second (FPS):
4. SFA-YOLO Network
4.1. SFA Method
- (1)
- More effective handling of missing data
- (2)
- Enhanced feature selectivity
- (3)
- Improved robustness
- (4)
- Stronger learning capability
4.2. SFA-YOLO Network Architecture
- (1)
- The SFA block serves as the core component within the network and offers clear advantages. It balances lightweight design, computational efficiency, and memory usage. The SFA block strengthens the model’s feature selection ability and suits deep learning tasks with high real-time requirements. The model adjusts channel weights dynamically to focus on key regions and suppress irrelevant information. This improves detection precision and robustness. It runs efficiently on devices with limited computational resources and adapts well to complex environments and diverse tasks. The structure of the SFA module is shown in Figure 4.
- (2)
- SPPF (Spatial Pyramid Pooling Fusion): SPPF strengthens the model’s multi-scale feature extraction ability by incorporating spatial pyramid pooling. It performs pooling operations at different scales, enabling the model to capture hierarchical and multi-size features [19]. This fusion enhances the model’s adaptability to object size and location variations while improving recognition in complex scenes [20,21,22,23,24]. Additionally, SPPF reduces feature map dimensions while preserving essential spatial information, improving computational efficiency and memory usage, which is especially useful for real-time object detection.
- (3)
- C2PSA (Cross-Stage Partial with Self-Attention) enhances feature extraction with a combination of the CSP (Cross Stage Partial) structure and the PSA (Pyramid Squeeze Attention) mechanism. It improves multi-scale feature representation and strengthens both channel and spatial attention. This design strengthens the effectiveness of feature extraction and information fusion. C2PSA uses channel attention to select important feature channels and suppress irrelevant information. This process improves the quality of feature representation. Meanwhile, spatial attention focuses on key regions in the image and strengthens target localization precision. This fused attention design increases model precision and robustness. In complex backgrounds, it helps the model recognize targets more effectively. The efficient structure of the C2PSA module maintains a strong performance while optimizing computational efficiency. It suits various deep learning tasks, especially in scenarios with fine feature extraction and precise localization requirements.
- (4)
- The upsampling module offers significant advantages in deep learning by effectively increasing the resolution of feature maps and enhancing the model’s ability to capture fine-grained details. The upsampling module restores spatial information by performing upsampling on low-resolution feature maps. It improves the model’s capacity for fine-grained processing in tasks such as object detection and semantic segmentation.
- (5)
- The Detect module processes the features extracted by the backbone network and generates final predictions, including class labels, bounding boxes, and object confidence scores. The head consists of upsampling, concatenation, and convolution operations, producing outputs at multiple resolutions.
5. Experimental Results and Analysis
5.1. Personal Protective Equipment Image Detection
5.2. Performance Comparison of Different Detection Models
5.3. Ablation Experiment
5.4. The Generalizability of the SFA-YOLO Method
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Vijayakumar, A.; Vairavasundaram, S. Yolo-based object detection models: A review and its applications. Multimed. Tools Appl. 2024, 83, 83535–83574. [Google Scholar] [CrossRef]
- Liu, L.; Ouyang, W.; Wang, X. Deep learning for generic object detection: A survey. Int. J. Comput. Vis. 2020, 128, 261–318. [Google Scholar] [CrossRef]
- Zhao, B.; Chang, L.; Liu, Z. Fast-YOLO Network Model for X-Ray Image Detection of Pneumonia. Electronics 2025, 14, 903. [Google Scholar] [CrossRef]
- Ansah, P.A.K.; Appati, J.K.; Owusu, E.; Boahen, E.K.; Boakye-Sekyerehene, P.; Dwumfour, A. SB-YOLO-V8: A Multilayered Deep Learning Approach for Real-Time Human Detection. Eng. Rep. 2025, 7, e70033. [Google Scholar] [CrossRef]
- Fan, Q.S. LUD-YOLO: A novel lightweight object detection network for unmanned aerial vehicle. Inf. Sci. 2025, 686, 121366. [Google Scholar] [CrossRef]
- Xu, S.; Jiang, W.; Liu, Q.; Wang, H.; Zhang, J.; Li, J.; Huang, X.; Bo, Y. Coal-rock interface real-time recognition based on the improved YOLO detection and bilateral segmentation network. Undergr. Space 2025, 21, 22–43. [Google Scholar] [CrossRef]
- Abudukelimu, A.; Ye, J.; Halidanmu, A.; Guo, W. DM-YOLO: Improved YOLOv9 model for tomato leaf disease detection. Front. Plant Sci. 2025, 15, 1473928. [Google Scholar]
- Pei, S.; Wang, W.; Wu, P.; Hu, C.; Sun, H.; Li, K.; Wu, M.; Lan, B. Detection of hydrophobicity grade of insulators based on AHC-YOLO algorithm. Sci. Rep. 2025, 15, 9673. [Google Scholar] [CrossRef] [PubMed]
- Aishwarya, N.; Kannaa, G.Y.; Seemakurthy, K. YOLOSkin: A fusion framework for improved skin cancer diagnosis using YOLO detectors on Nvidia Jetson Nano. Biomed. Signal Process. Control. 2025, 100, 107093. [Google Scholar] [CrossRef]
- He, L.; Wei, H. CBAM-YOLOv5: A promising network model for wear particle recognition. Wirel. Commun. Mob. Comput. 2023, 2023, 2520933. [Google Scholar] [CrossRef]
- Gong, X.; Zhang, Y.; Hu, S. ASAFormer: Visual tracking with convolutional vision transformer and asymmetric selective attention. Knowl.-Based Syst. 2024, 291, 111562. [Google Scholar] [CrossRef]
- Flores-Calero, M.; Astudillo, C.A.; Guevara, D.; Maza, J.; Lita, B.S.; Defaz, B.; Ante, J.S.; Zabala-Blanco, D.; Armingol Moreno, J.M. Traffic Sign Detection and Recognition Using YOLO Object Detection Algorithm: A Systematic Review. Mathematics 2024, 12, 297. [Google Scholar] [CrossRef]
- Sangaiah, A.K.; Yu, F.N.; Lin, Y.B.; Shen, W.C.; Sharma, A. UAV T-YOLO-rice: An enhanced tiny YOLO networks for rice leaves diseases detection in paddy agronomy. IEEE Trans. Netw. Sci. Eng. 2024, 11, 5201–5216. [Google Scholar] [CrossRef]
- Xiao, G.; Hou, S.; Zhou, H. PCB defect detection algorithm based on CDI-YOLO. Sci. Rep. 2024, 14, 7351. [Google Scholar] [CrossRef]
- Liu, Z.; Abeyrathna, R.M.R.D.; Sampurno, R.M.; Victor, M.N.; Tofael, A. Faster-YOLO-AP: A lightweight apple detection algorithm based on improved YOLOv8 with a new efficient PDWConv in orchard. Comput. Electron. Agric. 2024, 223, 109118. [Google Scholar] [CrossRef]
- Prinzi, F.; Insalaco, M.; Orlando, A.; Gagilo, S.; Vitable, S. A yolo-based model for breast cancer detection in mammograms. Cogn. Comput. 2024, 16, 107–120. [Google Scholar] [CrossRef]
- Lalinia, M.; Sahafi, A. Colorectal polyp detection in colonoscopy images using YOLO-V8 network. Signal Image Video Process. 2024, 18, 2047–2058. [Google Scholar] [CrossRef]
- Paul, A.; Machavaram, R.; Ambuj; Kumar, D.; Nagar, H. Smart solutions for capsicum Harvesting: Unleashing the power of YOLO for Detection, Segmentation, growth stage Classification, Counting, and real-time mobile identification. Comput. Electron. Agric. 2024, 219, 108832. [Google Scholar] [CrossRef]
- Sun, Q.; Li, P.; He, C.; Song, Q.; Chen, J.; Kong, X.; Luo, Z. A Lightweight and High-Precision Passion Fruit YOLO Detection Model for Deployment in Embedded Devices. Sensors 2024, 24, 4942. [Google Scholar] [CrossRef]
- Bhavana, N.; Kodabagi, M.M.; Kumar, B.M.; Ajay, P.; Muthukumaran, N.; Ahilan, A. POT-YOLO: Real-Time Road Potholes Detection using Edge Segmentation based Yolo V8 Network. IEEE Sens. J. 2024, 24, 24802–24809. [Google Scholar] [CrossRef]
- Azevedo, P.; Santos, V. Comparative analysis of multiple YOLO-based target detectors and trackers for ADAS in edge devices. Robot. Auton. Syst. 2024, 171, 104558. [Google Scholar] [CrossRef]
- Moksyakov, A.; Wu, Y.; Gadsden, S.A.; Yawney, J.; AlShabi, M. Object Detection and Tracking with YOLO and the Sliding Innovation Filter. Sensors 2024, 24, 2107. [Google Scholar] [CrossRef] [PubMed]
- Panigrahy, S.; Karmakar, S. Real-time condition monitoring of transmission line insulators using the YOLO object detection model with a UAV. IEEE Trans. Instrum. Meas. 2024, 73, 1–9. [Google Scholar] [CrossRef]
- Zhao, B.; Wu, C.; Chang, L.; Jiang, Y.; Sun, R. Research on Zero-Force control and collision detection of deep learning methods in collaborative robots. Displays 2025, 87, 102969. [Google Scholar] [CrossRef]
Attire Classes | Number of Labels |
---|---|
vest | 3619 |
no-helmet | 4678 |
helmet | 4204 |
no-vest | 4723 |
Name | Version |
---|---|
OS | Ubuntu 16.04LTS |
CPU | Intel(R) Core(TM) i7-11800H @ 3.20GHz |
RAM | 32 GB |
GPU | GeForce RTX 3060 |
Driver | 456.19.01 |
CUDA | 11.3.1 |
python | 3.9.18 |
torch | 1.10.1+cu111 |
torchvision | 0.11.2++cu111 |
Parameters | Value | Parameters | Value |
---|---|---|---|
epochs | 500 | max_det | 300 |
patience | 100 | lr0 | 0.01 |
batch | 32 | lrf | 0.01 |
imgsz | 640 | dropout | 0.0 |
workers | 4 | optimizer | SGD |
Algorithms | FPS | Precision | Recall | mAP@0.5 | mAP@0.5:0.95 |
---|---|---|---|---|---|
YOLOv5n | 256 | 84.1% | 72.2% | 76.3% | 37.8% |
YOLOv5s | 156 | 85.6% | 78.1% | 81.2% | 42.1% |
YOLOv7-Tiny | 172 | 87.2% | 72.4% | 78.2% | 40.8% |
CBAM-YOLO | 123 | 86.3% | 77.8% | 81.7% | 47.6% |
YOLOv8n | 259 | 84.5% | 72.6% | 76.9% | 38.1% |
YOLOX-Nano | 271 | 81.2% | 69.6% | 73.1% | 34.9% |
ViT | 87 | 82.9% | 81.0% | 83.4% | 48.4% |
SFA-YOLO | 149 | 89.3% | 83.7% | 89.4% | 50.3% |
Algorithms | FPS | Precision | Recall | mAP@0.5 | mAP@0.5:0.95 |
---|---|---|---|---|---|
Without SFA | 130 | 85.8% | 77.1% | 84.2% | 45.3% |
Without SPPF | 145 | 88.4% | 84.2% | 89.1% | 50.4% |
Without C2PSA | 154 | 89.7% | 81.2% | 88.4% | 49.7% |
Algorithms | FPS | Precision | Recall | mAP@0.5 | mAP@0.5:0.95 |
---|---|---|---|---|---|
YOLOv5n | 88 | 96.7% | 94.7% | 97.5% | 77.0% |
YOLOv5s | 75 | 95.1% | 94.7% | 97.4% | 82.7% |
YOLOv7-Tiny | 103 | 96.8% | 94.3% | 97.7% | 76.6% |
CBAM-YOLO | 68 | 96.6% | 95.1% | 95.8% | 84.2% |
YOLOv8n | 94 | 97.2% | 94.9% | 97.6% | 77.5% |
YOLOX-nano | 101 | 95.3% | 93.6% | 95.4% | 76.3% |
ViT | 55 | 93.4% | 91.8% | 96.6% | 89.6% |
SFA-YOLO | 143 | 95.8% | 95.9% | 97.7% | 97.6% |
Algorithms | FPS | Precision | Recall | mAP@0.5 | mAP@0.5:0.95 |
---|---|---|---|---|---|
YOLOv5n | 213 | 87.8% | 83.3% | 89.8% | 64.3% |
YOLOv5s | 187 | 88.5% | 83.6% | 90.1% | 66.4% |
YOLOv7-Tiny | 195 | 90.3% | 84.2% | 89.6% | 65.0% |
CBAM-YOLO | 166 | 89.2% | 82.3% | 90.6% | 66.8% |
YOLOv8n | 224 | 88.4% | 84.1% | 90.3% | 64.9% |
YOLOX-nano | 235 | 85.1% | 82.4% | 88.5% | 63.2% |
ViT | 82 | 82.6% | 84.5% | 86.6% | 68.1% |
SFA-YOLO | 135 | 85.0% | 90.4% | 92.4% | 71.5% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Sun, R.; Zhao, B.; Wu, C.; Qin, X. Research on Inspection Method of Intelligent Factory Inspection Robot for Personnel Safety Protection. Appl. Sci. 2025, 15, 5750. https://doi.org/10.3390/app15105750
Sun R, Zhao B, Wu C, Qin X. Research on Inspection Method of Intelligent Factory Inspection Robot for Personnel Safety Protection. Applied Sciences. 2025; 15(10):5750. https://doi.org/10.3390/app15105750
Chicago/Turabian StyleSun, Ruohuai, Bin Zhao, Chengdong Wu, and Xiaohong Qin. 2025. "Research on Inspection Method of Intelligent Factory Inspection Robot for Personnel Safety Protection" Applied Sciences 15, no. 10: 5750. https://doi.org/10.3390/app15105750
APA StyleSun, R., Zhao, B., Wu, C., & Qin, X. (2025). Research on Inspection Method of Intelligent Factory Inspection Robot for Personnel Safety Protection. Applied Sciences, 15(10), 5750. https://doi.org/10.3390/app15105750