UAV Low-Altitude Remote Sensing Inspection System Using a Small Target Detection Network for Helmet Wear Detection
Abstract
:1. Introduction
- Firstly, most of the helmet-wearing detection work relies on ground-mounted surveillance cameras. However, such a method has many limitations, including fixed camera view, limited set-up density, installation location restricted by the site environment, and wired power supply and data transmission. Therefore, it is challenging to automate supervision on construction sites.
- Secondly, we believe that it is also necessary to identify the specific color of the helmet, such as red, blue, white, or yellow, when it comes to intelligent supervision and detection. This is because although helmet colors represent different meanings in different regions, it is common to use them to distinguish, for example, the skilled trades of workers, managers, and supervisors. This is necessary to supervise whether a specific group of people is in the corresponding area during construction. However, most of the proposed methods in this work only detect whether workers wear helmets [23,24,25,26,27,28]. Even though a small number of studies classify helmet colors in more detail [29,30], it is essential to improve the accuracy of the color classification of helmets, limited by the performance of their proposed models.
- In addition, due to the limitations of ground surveillance, workers who are further away appear smaller in size in the images, and most of the recognition algorithms proposed in the state-of-the-art work cannot directly meet the needs of this type of detection.
- This paper proposes a novel scheme for inspecting construction sites with UAVs. The flexible maneuverability of UAVs will be used to perform fast and efficient helmet-wearing inspections periodically on construction sites. We validate the system in real construction sites using the deep learning model proposed in this paper.
- For UAV remote sensing images, our proposed single-stage end-to-end helmet detection network is based on the highest accuracy Swin Transformer module as the backbone network, which ensures efficient feature extraction. At the same time, the model complexity is low, and we propose an attention-weighted fusion network so that a deeper network with more powerful semantic information can improve the efficiency of the network for small target detection.
- The experimental results show that our proposed method can accurately classify the actual construction site in the patrol according to whether or not wearing helmet and helmet color type and reach the mAP of 92.87% on the open-source dataset GDUT-Hardhat Wearing Detection (GDUT-HWD) and improved the AP to 88.7% for small-size target detection.
2. Materials and Methods
2.1. UAV Inspection Program
2.2. Helmet Detection Network for UAV Inspection Systems
2.2.1. Design of Backbone Network (Step 1)
2.2.2. Attention-Weighted Fusion Network (Step 2)
2.2.3. Decoupled Headers (Step 3)
2.3. Experiments
2.3.1. Dataset and Experimental Environment
2.3.2. Evaluation Index and Hyperparameter Setting
3. Results
3.1. Comparison of Model Training
3.2. Comparison of Attention Modules
3.3. Comparison of Model Performance
3.4. Visualization of UAV Inspections
4. Discussion
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Roldán, J.J.; Joossen, G.; Sanz, D.; Del Cerro, J.; Barrientos, A. Mini-UAV based sensory system for measuring environmental variables in greenhouses. Sensors 2015, 15, 3334–3350. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Gomez, C.; Purdie, H. UAV-based Photogrammetry and Geocomputing for Hazards and Disaster Risk Monitoring—A Review. Geoenviron. Disasters 2016, 3, 23. [Google Scholar] [CrossRef] [Green Version]
- Karaca, Y.; Cicek, M.; Tatli, O.; Sahin, A.; Pasli, S.; Beser, M.F.; Turedi, S. The potential use of unmanned aircraft systems (drones) in mountain search and rescue operations. Am. J. Emerg. Med. 2018, 36, 583–588. [Google Scholar] [CrossRef] [PubMed]
- Burke, C.; McWhirter, P.R.; Veitch-Michaelis, J.; McAree, O.; Pointon, H.A.; Wich, S.; Longmore, S. Requirements and limitations of thermal drones for effective search and rescue in marine and coastal areas. Drones 2019, 3, 78. [Google Scholar] [CrossRef] [Green Version]
- Oudjehane, A.; Baker, T.; Moeini, S. The role and value of integrating AI, drones and associate technologies in construction projects. In Proceedings of the Canadian Society for Civil Engineering Annual Conference, Laval (Greater Montreal), QC, Canada, 12–15 June 2019. [Google Scholar]
- Falorca, J.F.; Miraldes, J.P.N.D.; Lanzinha, J.C.G. New trends in visual inspection of buildings and structures: Study for the use of drones. Open Eng. 2021, 11, 734–743. [Google Scholar] [CrossRef]
- Martinez, J.G.; Gheisari, M.; Alarcón, L.F. UAV integration in current construction safety planning and monitoring processes: Case study of a high-rise building construction project in Chile. J. Manag. Eng. 2020, 36, 05020005. [Google Scholar] [CrossRef]
- Jiang, W.; Zhou, Y.; Ding, L.; Zhou, C.; Ning, X. UAV-based 3D reconstruction for hoist site mapping and layout planning in petrochemical construction. Autom. Constr. 2020, 113, 103137. [Google Scholar] [CrossRef]
- Elghaish, F.; Matarneh, S.; Talebi, S.; Kagioglou, M.; Hosseini, M.R.; Abrishami, S. Toward digitalization in the construction industry with immersive and drones technologies: A critical literature review. Smart Sustain. Built Environ. 2020, 10, 345–363. [Google Scholar] [CrossRef]
- Li, Y.; Liu, C. Applications of multirotor drone technologies in construction management. Int. J. Constr. Manag. 2018, 19, 401–412. [Google Scholar] [CrossRef]
- Liu, J.; Jennesse, M., Jr.; Holley, P. Utilizing light unmanned aerial vehicles for the inspection of curtain walls: A case study. In Proceedings of the Construction Research Congress 2016, San Juan, Puerto Rico, 31 May–2 June 2016; pp. 2651–2659. [Google Scholar]
- Choi, S.-S.; Kim, E.-K.; Sung-Suk, C. Design and implementation of vision-based structural safety inspection system using small unmanned aircraft. In Proceedings of the 2015 17th International Conference on Advanced Communication Technology (ICACT), Pyeongchang, Republic of Korea, 1–3 July 2015; pp. 562–567. [Google Scholar] [CrossRef]
- Qi, J.; Issa, R.R.A.; Hinze, J.; Olbina, S. Integration of safety in design through the use of building information modeling. In Proceedings of the International Workshop on Computing in Civil Engineering 2011, Miami, FL, USA, 19–22 June 2011; pp. 698–705. [Google Scholar] [CrossRef]
- Suderman, B.L.; Hoover, R.W.; Ching, R.P.; Scher, I.S. The effect of hardhats on head and neck response to vertical impacts from large construction objects. Accid. Anal. Prev. 2014, 73, 116–124. [Google Scholar] [CrossRef]
- Liang, H.; Zhang, S.; Su, Y. The structure and emerging trends of construction safety management research: A bibliometric review. Int. J. Occup. Saf. Ergon. 2018, 26, 469–488. [Google Scholar] [CrossRef] [PubMed]
- Irizarry, J.; Gheisari, M.; Walker, B.N. Usability assessment of drone technology as safety inspection tools. J. Inf. Technol. Constr. 2012, 17, 194–212. [Google Scholar]
- Patrick, O.O.; Nnadi, E.O.; Ajaelu, H.C. Effective use of quadcopter drones for safety and security monitoring in a building construction sites: Case study Enugu Metropolis Nigeria. J. Eng. Technol. Res. 2020, 12, 37–46. [Google Scholar]
- Rubaiyat, A.H.; Toma, T.T.; Kalantari-Khandani, M.; Rahman, S.A.; Chen, L.; Ye, Y.; Pan, C.S. Automatic detection of helmet uses for construction safety. In Proceedings of the 2016 IEEE/WIC/ACM International Conference on Web Intelligence Workshops (WIW), Omaha, NE, USA, 13–16 October 2016; pp. 135–142. [Google Scholar]
- Jin, M.; Zhang, J.; Chen, X.; Wang, Q.; Lu, B.; Zhou, W.; Nie, G.; Wang, X. Safety helmet detection algorithm based on color and hog features. In Proceedings of the 2020 IEEE 19th International Conference on Cognitive Informatics and Cognitive Computing (ICCI*CC), Beijing, China, 26–28 September 2020; pp. 215–219. [Google Scholar]
- Long, X.; Cui, W.; Zheng, Z. Safety helmet wearing detection based on deep learning. In Proceedings of the 2019 IEEE 3rd Information Technology, Networking, Electronic and Automation Control Conference (ITNEC), Chengdu, China, 15–17 March 2019; pp. 2495–2499. [Google Scholar]
- Dai, B.; Nie, Y.; Cui, W.; Liu, R.; Zheng, Z. Real-time safety helmet detection system based on improved SSD. In Proceedings of the 2nd International Conference on Artificial Intelligence and Advanced Manufacture, Manchester, UK, 15–17 October 2020; pp. 95–99. [Google Scholar]
- Han, G.; Zhu, M.; Zhao, X.; Gao, H. Method based on the cross-layer attention mechanism and multiscale perception for safety helmet-wearing detection. Comput. Electr. Eng. 2021, 95, 107458. [Google Scholar] [CrossRef]
- Deng, L.; Li, H.; Liu, H.; Gu, J. A lightweight YOLOv3 algorithm used for safety helmet detection. Sci. Rep. 2022, 12, 10981. [Google Scholar] [CrossRef]
- Hayat, A.; Morgado-Dias, F. Deep learning-based automatic safety helmet detection system for construction safety. Appl. Sci. 2022, 12, 8268. [Google Scholar] [CrossRef]
- Sun, C.; Zhang, S.; Qu, P.; Wu, X.; Feng, P.; Tao, Z.; Zhang, J.; Wang, Y. MCA-YOLOV5-Light: A faster, stronger and lighter algorithm for helmet-wearing detection. Appl. Sci. 2022, 12, 9697. [Google Scholar] [CrossRef]
- Zhang, B.; Sun, C.-F.; Fang, S.-Q.; Zhao, Y.-H.; Su, S. Workshop safety helmet wearing detection model based on SCM-YOLO. Sensors 2022, 22, 6702. [Google Scholar] [CrossRef]
- Chen, M.; Kong, R.; Zhu, J.; Wang, L.; Qi, J. Application research of safety helmet detection based on low computing power platform using YOLO v5. In Proceedings of the International Conference on Adaptive and Intelligent Systems, Larnaca, Cyprus, 25–26 May 2022; Springer: Cham, Switzerland, 2022; pp. 107–117. [Google Scholar] [CrossRef]
- Xia, X.; Zhai, J. Study on small target detection of construction site helmet based on YOLOv5. In Proceedings of the Chinese Intelligent Systems Conference, Changsha, China, 10–13 May 2022; Springer: Singapore, 2022; pp. 18–27. [Google Scholar] [CrossRef]
- Tran, V.T.; To, T.S.; Nguyen, T.-N.; Tran, T.D. Safety helmet detection at construction sites using YOLOv5 and YOLOR. In Proceedings of the International Conference on Intelligence of Things, Delft, The Netherlands, 7–10 November 2022; Springer: Cham, Switzerland, 2022; pp. 339–347. [Google Scholar] [CrossRef]
- Zhang, C.; Tian, Z.; Song, J.; Zheng, Y.; Xu, B. Construction worker hardhat-wearing detection based on an improved BiFPN. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 8600–8607. [Google Scholar] [CrossRef]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Proceedings of the Annual Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
- Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
- Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; Springer: Cham, Switzerland, 2014; pp. 740–755. [Google Scholar]
- Liang, H.; Seo, S. Lightweight deep learning for road environment recognition. Appl. Sci. 2022, 12, 3168. [Google Scholar] [CrossRef]
- Liang, H.; Seo, S. Automatic detection of construction workers’ helmet wear based on lightweight deep learning. Appl. Sci. 2022, 12, 10369. [Google Scholar] [CrossRef]
- Qin, Z.; Zhang, P.; Wu, F.; Li, X. FCANet: Frequency channel attention networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021. [Google Scholar]
- Wu, J.; Cai, N.; Chen, W.; Wang, H.; Wang, G. Automatic detection of hardhats worn by construction personnel: A deep learning approach and benchmark dataset. Autom. Constr. 2019, 106, 102894. [Google Scholar] [CrossRef]
- Loshchilov, I.; Hutter, F. Sgdr: Stochastic gradient descent with warm restarts. arXiv 2016, arXiv:1608.03983. [Google Scholar]
- Zheng, Z.; Wang, P.; Liu, W.; Li, J.; Ye, R.; Ren, D. Distance-IoU loss: Faster and better learning for bounding box regression. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12993–13000. [Google Scholar]
- Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
- Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient channel attention for deep convolutional neural networks. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar] [CrossRef]
- Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
- Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; Springer: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
- Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef]
- Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. YoloX: Exceeding Yolo series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
- Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and efficient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10781–10790. [Google Scholar]
UAV Parameters | Experimental Parameters | ||
---|---|---|---|
Takeoff weight | 0.4 kg | Inspection height (H) | 5 m, 10 m, 15 m |
Dimensions (L × W × H) | 180 × 180 × 80 mm | Shooting angles (R) | 30°~70° |
Max resolution | 4 K/60 fps | The total duration of a single inspection | 10 min |
Max video transmission bitrate | 50 Mbps | Inspection cycle (T) | 1.5 h |
Field of view (FOV) | 155° | Back-and-forth spacing | 10 m |
Propeller guard | Built-in | Wind speed | 0~3 m/s |
Communication frequency band | Up to 40 MHz | Flight speed | 8 m/s |
Anchor Layer | Anchor Size (Width, Height) |
---|---|
Anchor 1 | (48, 87); (69, 123); (114,202) |
Anchor 2 | (20, 36); (26, 48); (35, 64) |
Anchor 3 | (6, 9); (9, 15); (14, 25) |
Items | Description | |
---|---|---|
Hardware | Central Processing Unit | Intel(R) Core (TM) i5-11400F |
Random Access Memory | 16 GB | |
Solid State Drive | Samsung SSD 500 GB | |
Graphics Card | NVIDIA GeForce RTX 3050 | |
Software | Operating System | Windows 11 Pro, 64 bit |
Programming Language | Python 3.7 | |
Learning Framework | Pytorch 1.9.0 |
Input Settings | Loss Calculation | Data Enhancement | ||||||
---|---|---|---|---|---|---|---|---|
Input shape | Batch size | Total Epoch | Loss Function | Max_lr | Min_lr | Decay Type | Mosaic | Mixup |
640 × 640 | 8 | 500 | CIoU | 0.01 | 0.0001 | Cosine Annealing | True | True |
Backbone | Baseline | √ | √ | √ | √ | √ | √ |
Step-scale feature fusion | √ | √ | √ | √ | √ | √ | |
SENet [40] | √ | ||||||
ECA-Net [41] | √ | ||||||
CBAM [42] | √ | ||||||
LRCA-Netv2 [35] | √ | ||||||
LRCA-Netv3 | √ | ||||||
Parameters (millions) | 85.3 | 86.7 | 86.7 | 86.7 | 86.7 | 86.7 | 86.7 |
mAP | 84.07% | 89.14% | 89.57% | 89.71% | 91.52% | 92.01% | 92.87% |
Method | Input Size | Sizes-AP(%) | Categories-AP(%) | mAP(%) | F1(%) | Parameters (Millions) | G-FLOPs(G) | Inference Speed (s) | |||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Small | Medium | Large | Blue | Red | White | Yellow | None | ||||||||
SSD | 600 × 600 | 38.2 | 90.1 | 96.5 | 79.41 | 75.63 | 78.43 | 77.81 | 74.26 | 77.11 | 63.1 | 26.29 | 247.01 | 0.17648 | |
Faster-RCNN | 600 × 600 | 37.5 | 92.5 | 96.1 | 80.15 | 69.83 | 77.45 | 78.46 | 71.51 | 75.48 | 64.2 | 136.78 | 370.01 | 0.28335 | |
600 × 600 | 38.9 | 94.1 | 97.4 | 83.74 | 78.07 | 80.87 | 79.30 | 77.44 | 79.88 | 65.8 | 28.316 | 939.36 | 0.32351 | ||
YOLOv5 | L | 640 × 640 | 78.7 | 95.3 | 97.5 | 84.34 | 81.97 | 83.64 | 83.17 | 78.55 | 82.33 | 81.7 | 46.65 | 114.31 | 0.06981 |
X | 640 × 640 | 77.9 | 95.1 | 98.2 | 86.42 | 83.58 | 84.80 | 85.74 | 80.99 | 84.31 | 82.2 | 87.27 | 217.41 | 0.07008 | |
YOLOX | L | 640 × 640 | 78.1 | 94.0 | 97.4 | 85.24 | 83.09 | 83.72 | 84.37 | 80.98 | 83.48 | 79.44 | 37.62 | 106.12 | 0.06996 |
X | 640 × 640 | 79.5 | 95.5 | 98.7 | 86.45 | 83.12 | 85.18 | 85.86 | 81.84 | 84.51 | 81.2 | 70.84 | 188.51 | 0.07326 | |
EfficientDet | D4 | 1024 × 1024 | 68.2 | 74.5 | 85.9 | 77.08 | 70.14 | 74.74 | 72.05 | 69.19 | 72.64 | 73.8 | 20.70 | 110.26 | 0.07014 |
D5 | 1280 × 1280 | 70.3 | 73.8 | 87.2 | 78.41 | 72.77 | 75.68 | 73.18 | 70.07 | 74.02 | 73.1 | 33.63 | 270.65 | 0.09178 | |
D6 | 1280 × 1280 | 70.2 | 75.0 | 90.8 | 77.91 | 73.54 | 76.93 | 75.14 | 74.11 | 75.53 | 75.4 | 51.84 | 546.31 | 0.14435 | |
D7 | 1536 × 1536 | 64.4 | 80.4 | 94.5 | 78.49 | 73.31 | 77.89 | 75.91 | 75.89 | 76.30 | 78.7 | 57.57 | 650.17 | 0.19472 | |
Our Approach | 640 × 640 | 88.7 | 92.88 | 96.74 | 91.81 | 90.89 | 95.86 | 92.69 | 93.09 | 92.87 | 90.8 | 86.70 | 249.41 | 0.07762 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Liang, H.; Seo, S. UAV Low-Altitude Remote Sensing Inspection System Using a Small Target Detection Network for Helmet Wear Detection. Remote Sens. 2023, 15, 196. https://doi.org/10.3390/rs15010196
Liang H, Seo S. UAV Low-Altitude Remote Sensing Inspection System Using a Small Target Detection Network for Helmet Wear Detection. Remote Sensing. 2023; 15(1):196. https://doi.org/10.3390/rs15010196
Chicago/Turabian StyleLiang, Han, and Suyoung Seo. 2023. "UAV Low-Altitude Remote Sensing Inspection System Using a Small Target Detection Network for Helmet Wear Detection" Remote Sensing 15, no. 1: 196. https://doi.org/10.3390/rs15010196
APA StyleLiang, H., & Seo, S. (2023). UAV Low-Altitude Remote Sensing Inspection System Using a Small Target Detection Network for Helmet Wear Detection. Remote Sensing, 15(1), 196. https://doi.org/10.3390/rs15010196