Next Article in Journal
Efficient Human Activity Recognition Using Machine Learning and Wearable Sensor Data
Previous Article in Journal
Enhancing Far-Field Speech Recognition with Mixer: A Novel Data Augmentation Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Intelligent Monitoring System for the Driving Environment of Explosives Transport Vehicles Based on Consumer-Grade Cameras

1
State Key Laboratory of Precision Blasting, Jianghan University, Wuhan 430056, China
2
Hubei Key Laboratory of Blasting Engineering, Jianghan University, Wuhan 430056, China
3
School of Civil Engineering, Tianjin University, Tianjin 300350, China
4
Key Laboratory of Coast Civil Structure Safety of the Ministry of Education, Tianjin University, Tianjin 300350, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(7), 4072; https://doi.org/10.3390/app15074072
Submission received: 27 December 2024 / Revised: 18 March 2025 / Accepted: 4 April 2025 / Published: 7 April 2025

Abstract

:
With the development of industry and society, explosives are widely used in social production as an important industrial product and require transportation. Explosives transport vehicles are susceptible to various objective factors during driving, increasing the risk of transportation. At present, new transport vehicles are generally equipped with intelligent driving monitoring systems. However, for old transport vehicles, the cost of installing such systems is relatively high. To enhance the safety of older explosives transport vehicles, this study proposes a cost-effective intelligent monitoring system using consumer-grade IP cameras and edge computing. The system integrates YOLOv8 for real-time vehicle detection and a novel hybrid ranging strategy combining monocular (fast) and binocular (accurate) techniques to measure distances, ensuring rapid warnings and precise proximity monitoring. An optimized stereo matching workflow reduces processing latency by 23.5%, enabling real-time performance on low-cost devices. Experimental results confirm that the system meets safety requirements, offering a practical, application-specific solution for improving driving safety in resource-limited explosive transport environments.

1. Introduction

With the development of industry and society, explosives are widely used in social production and require transportation. Explosives are classified as a first-category dangerous good, which can produce gases with a certain temperature, pressure, and speed that can cause serious damage to surrounding objects through chemical reactions under external effects (such as heat, pressure, impact, etc.) [1]. Due to the fast destruction speed and wide range of explosion, explosives transport vehicles have high potential danger. If an accident occurs, it will cause serious casualties and property damage [2,3]. The Regulations on the Management of Road Transport of Dangerous Goods of China clearly emphasizes the importance of road specialized vehicles transporting dangerous chemicals, fireworks, and civilian explosives [4]. Due to the uniqueness of explosives transport vehicles, reducing the probability of traffic accidents involving explosives transport vehicles is a top priority in ensuring road transportation safety. For example, on 13 June 2020, a liquefied petroleum gas tank truck overturned and exploded due to the high speed during a bend. The resulting impact caused the tank truck to crash into a factory. The accident resulted in 20 deaths and 172 injuries, and the surrounding buildings at the explosion site were damaged to varying degrees [5]. Explosives transport vehicles may face various risks due to objective factors during the transportation of explosives. For example, when transporting explosives under adverse weather conditions, the visibility of the road and the driving stability of the transport vehicle can be affected by the driver, which can easily cause traffic accidents [6]. Therefore, the driving speed should be adjusted appropriately to maintain a safe front and rear distance. In crowded road sections, if the distance between vehicles is too close, accidents may occur and lead to additional casualties and property losses. Therefore, it is necessary to maintain a safe distance from the surrounding vehicles. If the transport vehicle is too close to other vehicles, measures must be taken in time to prevent an accident.
To ensure the safety of explosives transport vehicles during operation and provide early warnings of environmental risks, an intelligent vehicle driving monitoring system has been developed. It is a core component of modern automotive safety technology, utilizing advanced sensors, controllers, actuators, and communication modules to monitor and analyze driving states in real time, thereby enhancing safety and efficiency. Key technologies include vehicle-mounted cameras, radars, LiDAR, and ultrasonic sensors, which collectively enable comprehensive environmental perception around the vehicle [7]. While intelligent monitoring systems are widely implemented in newer transport vehicles, older explosives transport vehicles may encounter significant cost challenges in adopting such systems. While LiDAR demonstrates excellent performance in terms of ranging accuracy and real-time capability, its high cost and sensitivity to lighting and weather conditions limit its applicability in complex environments [8]. Ultrasonic radar has certain advantages in short-range detection but performs poorly in dynamic scenarios and is highly sensitive to changes in temperature and humidity [9]. In contrast, IP cameras, though less accurate in range than LiDAR, offer a wide detection range, high real-time performance, and low cost. Moreover, their strong environmental adaptability makes them better suited for real-time monitoring in the complex driving environments of explosives transport vehicles [10]. The comparison of each sensor is shown in Table 1.
In order to develop an economical and efficient intelligent monitoring system for the driving environment of explosives transport vehicles, integrating machine learning and computer vision technology has been planned. The system will focus on real-time monitoring of the front and rear distance of the vehicle and surrounding distances to enhance the safety of road driving. In the research of vehicle object detection, deep learning-based object detection algorithms have been widely developed and applied, including two-stage object and single-stage object detection algorithms. The two-stage detection algorithm needs to first generate candidate boxes and then perform classification regression, which has a high accuracy but slow speed. The commonly used algorithm is R-CNN [11,12,13]. The single-stage detection algorithm directly performs classification regression with low accuracy but fast speed. Commonly used algorithms include SSD [14,15,16] and YOLO [17,18,19]. Compared with two-stage detection algorithms, single-stage detection algorithms are more suitable for real-time detection during driving. In the research of binocular ranging technology based on computer vision (CV), Zhao et al. proposed a distance measurement method suitable for low-brightness environments based on the principle of binocular vision and the reconstruction theory of triangles [20]. In order to achieve dynamic ranging, Zaarane first uses a single camera to capture images of vehicles and detect targets and then uses template matching technology to detect the same vehicle captured by the second camera [21]. Shen first used MATLAB (R2013b) software to calibrate the camera and then used OpenCV (2.4.7) to perform stereo matching and correction on the image. The disparity of the image was used to calculate the distance between the target vehicle and the camera, thus achieving the goal of distance measurement [22].
The analysis of existing research results indicates that intelligent driving technology can already provide early warning and monitoring of hazardous road conditions and obstacles during the driving process of explosives transport vehicles. However, there remains a significant issue in the current application process: intelligent driving currently relies on embedded development, which necessitates industrial control computers and supporting sensors capable of stable long-term operation. The demand for high-end software and hardware facilities results in excessively high costs [23]. To tackle this issue, the YOLOv8 object detection algorithm, along with monocular and binocular camera ranging, is applied to establish an intelligent monitoring system for the driving environment of explosives transport vehicles. The system uses IP cameras that are flexibly arranged around the vehicle body to capture road images in real time and quickly transmits these images to an edge computing device for fast processing. The processed data will be displayed to the driver in an intuitive visual format, enabling the driver to grasp the dynamics of the surrounding environment of the vehicle in real time. With the help of this intelligent monitoring system, drivers can participate in traffic operation more accurately, so as to effectively monitor and prevent possible accidents. The application process of the system is shown in Figure 1.
The main contributions of this article are as follows:
(1)
A low-cost monitoring solution based on consumer-grade IP cameras and edge computing is proposed, addressing the applicability gap of traditional high-cost systems in older explosives transport vehicles.
(2)
A novel integration of YOLOv8 with hybrid ranging techniques is introduced. The system optimizes binocular ranging by performing stereo matching only within YOLOv8-detected bounding boxes, reducing processing latency by 23.5% and enabling real-time performance on low-cost devices. Additionally, the combination of monocular (fast) and binocular (accurate) ranging ensures rapid warnings for distant obstacles and precise measurements in high-risk proximity zones, tailored to the unique demands of explosive transport environments.
(3)
This study delivers an efficient, reliable, and application-specific monitoring system that meets the stringent safety and efficiency requirements of explosives transport vehicles. The system’s performance is rigorously validated under real-world conditions, demonstrating its ability to enhance driving safety while maintaining cost-effectiveness.

2. System Architecture and Hardware Layout

2.1. Selection and Layout of Consumer Cameras

Explosives transport vehicles belong to a type of dangerous goods transport vehicles. According to Road Transport Safety Management Measures for Dangerous Goods of China, the speed regulations for dangerous goods transport vehicles are as follows: dangerous goods transport vehicles cannot exceed 80 km per hour on highways and cannot exceed 60 km per hour on other roads. If the speed indicated on road speed limit signs or markings is lower than the prescribed speed, the vehicle shall not exceed the indicated speed limit. Additionally, Article 80 of the Implementation Regulations of the Road Traffic Safety Law of the People’s Republic of China stipulates that when a motor vehicle is driving on a highway and its speed exceeds 100 km per hour, vehicles must maintain a distance of more than 100 m from the vehicle in front of the same lane. When the speed is less than 100 km per hour, and the distance from the vehicle in front of the same lane can be appropriately shortened, the minimum distance should not be less than 50 m. Therefore, while meeting the driving standards of explosives transport vehicles, taking into account the safety requirements of explosives transport vehicles in actual explosives transport and the error of consumer-grade camera ranging, the monitoring standards are set at a speed of 60 km per hour, a safe following distance of 50 m for front and rear vehicles, and a 2 m distance from surrounding vehicles.
Focal length is the distance from the optical center of the lens to the imaging plane. The smaller the focal length of the camera, the stronger its ability to concentrate light, the wider the viewing range, and the closer the viewing distance. Conversely, the larger the focal length of the camera, the weaker its ability to concentrate light, the narrower the viewing range, and the farther the viewing distance. The system monitors two types of distances: the distance between the front and rear vehicles during transportation and the surrounding vehicles. Therefore, based on the optical characteristics of the camera and the functional requirements of the system, the appropriate camera (Tianjin, China) and its parameters are selected, as shown in Table 2.
The schematic diagram of the monitoring range of the binocular camera is shown in Figure 2, where the sector depicts the field of view of a single camera, and the gray part represents the overlap of the two sectors (i.e., the monitoring range of the binocular camera). θ is the horizontal field of view angle of the camera, d is the baseline distance between two cameras, and x is the length of the monitoring blind zone. The length of the monitoring blind zone x is proportional to the baseline distance d, and d can be determined as Equation (1).
d = 2 x 1 cos θ 1 + cos θ
The blind zone distance x of binocular monitoring must be less than the safety threshold of the system. Based on Equation (1) and the monitoring standard of a 2 m surrounding safety distance, the maximum baseline distance dmax of binocular cameras II listed in Table 1 is calculated using Equation (2):
θ = 70 ° , x = 2   m , d m a x = 2 x 1 cos θ 1 + cos θ = 2 × 2 × 1 0.34 1 + 0.34 2.81   m
A box truck (6 m long, 2.2 m wide, and 2.3 m high) was selected to evaluate the camera layout. It is appropriate to specify a monitoring blind zone of 1.5 m for monitoring, so the baseline is 2 m. The resulting camera layout and detection scale, based on the above analysis, are detailed in Table 3 and Figure 3.

2.2. Hardware Architecture of the System

The hardware architecture of the system is divided into three core components, as shown in Figure 4:
(1)
Management side: The YOLO target detection model, which is specially designed to identify and locate objects in the image, is trained on the PC. After the training is completed, the model will be deployed to the edge computing device.
(2)
Data acquisition: IP cameras installed in the transport vehicle body are responsible for capturing the environmental image of the vehicle in real time. Through real-time transport protocol (RTSP) and power over Ethernet (PoE), these image data are efficiently transmitted to network switches and then routed to edge computing devices.
(3)
Data processing and display: Edge computing devices quickly and accurately calculate and process the transmitted data. The processed information is converted into an intuitive visual format and displayed on a dedicated monitoring platform for real-time viewing and analysis by drivers or managers.

3. System Function Realization

The driving environment intelligent monitoring system has two core functions, and the architecture of function realization is shown in Figure 5.
(1)
Front and rear distance monitoring:
The monocular ranging algorithm provides a faster ranging capability because it processes a single image stream, which reduces the computational complexity compared to the binocular ranging algorithm that requires the analysis of two image streams. Nevertheless, binocular ranging offers higher precision than monocular ranging. The system captures image data in real time through the IP camera and uses the RTSP to transmit the image to the edge computing device. With the help of the YOLOv8 target monitoring algorithm, the system can effectively identify the vehicles in the image. Once the vehicle is detected, the system will input the pixel coordinates of the target to the monocular ranging algorithm to initially determine the vehicle distance. The algorithm uses the monocular camera installed before and after the vehicle body for rough measurement. If the measurement shows that the vehicle distance is less than the preset safety threshold (50 m), the system will start the binocular ranging algorithm and carry out a more precise vehicle distance measurement through the binocular camera before and after the vehicle body and issue an early warning.
(2)
Surrounding distance monitoring:
This function focuses on monitoring vehicles on the side of the vehicle and uses binocular ranging technology for measurement. When the system recognizes the surrounding target, the pixel coordinates of the target are immediately input into the binocular ranging algorithm. Using binocular cameras on both sides of the car body, the system can accurately measure the surrounding distance. Once the measurement result is lower than the set safe distance (2 m), the system will issue an early warning.

3.1. YOLOV8 Target Detection

The comparison of different target detection algorithms is shown in Table 4. Compared with other algorithms, YOLOv8 has significant advantages in terms of accuracy, real-time performance, and robustness [13,14,25].
YOLOv8, an object detection model known for its accuracy and fast inference, consists of three key components: Backbone, Neck, and Head. The Backbone extracts features using convolutional and deconvolutional layers, with residual connections and bottleneck structures to optimize network size and performance. It uses the C2f module, which has fewer parameters and better feature extraction than YOLOv5’s C3 module [26]. The Neck module enhances image representation by merging features from different stages of the Backbone, including an SPPF module, a PAA module, and two PAN modules. The Head module handles object identification and classification, with a detection head made up of convolutional and deconvolutional layers using global average pooling to produce the classification results. The network’s structure is illustrated in Figure 6.
Dataset acquisition is crucial for object detection. The intelligent detection system mainly identifies vehicles using a dataset obtained from the open-source UA-DETRAC. To enhance training accuracy, data augmentation techniques are essential. Initially, the training data undergo custom augmentation. Subsequently, the input data are enhanced with color jitter, random horizontal flipping, and random scaling by 10%.
The image, regarded as a collection of pixels, can be expressed as a matrix. YOLOv8 analyzes the input image, capturing its features through feature extraction technology. After calculations, the program identifies objects and provides their locations and labels. This process is realized by the trained weights, which are the core of the algorithm. These weights form the basis for image reasoning and target identification, equivalent to the criteria and standards used in the recognition process. Figure 7a illustrates the model’s performance in target monitoring in terms of accuracy and recall at various thresholds.
Assuming that there are only two types of targets to be classified, positive and negative, the following four metrics are used: (1) TP (true positive): the number of positive examples that are correctly identified as positive; (2) FP (false positive): the number of negative examples that are incorrectly identified as positive; (3) TN (true negative): the number of negative examples that are correctly identified as negative; (4) FN (false negative): the number of positive examples that are incorrectly identified as negative.
Precision measures the accuracy of model predictions, indicating the proportion of true positives within the predicted positives.
P r e c i s i o n = T P T P + F P
Recall indicates the proportion of actual positives correctly identified by the model, reflecting its comprehensiveness.
R e c a l l = T P T P + F N
Average precision (AP), the area under the precision–recall curve, measures the model’s overall performance across all thresholds and is a key indicator of its effectiveness. A high AP value indicates that the model maintains high accuracy and recall across different thresholds, enhancing its application effectiveness.
To enhance the model’s cross-platform deployment capability, we convert the YOLOv8-trained weights into the versatile ONNX format. This conversion significantly improves the model’s compatibility and applicability across various platforms. We then deploy the model on the edge device, enabling the YOLOv8 target detection algorithm to operate effectively. The detection performance is depicted in Figure 7b.

3.2. Monocular Camera Rough Distance Measurement

According to the principle of camera imaging, the size of the object in the camera image is proportional to its actual size and related to the focal length. Thus, a similar triangle method can be used for the monocular camera coarse distance measurement. W is the actual width of the object, W′ is the pixel width of the object on the imaging plane, d is the distance between the object and the camera lens, and f is the focal length of the camera. The proportional relationship between the object width and distance is derived from the similarity of triangles. Using this relationship, the distance between the object and the camera lens can be calculated as shown in Equation (5).
d = W × f W
Firstly, the camera is calibrated to determine its pixel focal length. Subsequently, this focal length and the true width of the vehicle (W) are input into the system. The image is then imported, and the system is provided with the target detection box width (W′) calculated by YOLOv8, as well as the pixel coordinates of the target. Using these values, the target’s real-world distance is calculated using Equation (5). The final ranging result is then output. The schematic diagram and flow chart of the monocular ranging algorithm are shown in Figure 8.

3.3. Binocular Camera Accurate Distance Measurement

A binocular camera consists of two cameras separated by a baseline distance, capturing the object in 2D images. The object’s distance is calculated by measuring the pixel disparities. The Bouguet polar line correction method is used to ensure that the imaging origin coordinates of the left and right camera views are consistent, the optical axes of the two cameras are parallel, the left and right imaging planes are coplanar, and the polar lines are aligned. Using the internal parameters, rotation, and translation matrices from camera calibration, the camera views are adjusted and aligned through rotation and translation.
By observing the object through a binocular lens and calculating the parallax d, combined with parameters such as the baseline distance b and angle of view of two cameras, the three-dimensional coordinates of the object can be calculated using the principle of triangulation, and then the distance D of the object can be calculated using Equation (6). (f is the focal length of the cameras; X, Y, and Z are the global coordinates of the object; and xl, xr, yl, and yr are the projection coordinates of the object in the left and right cameras.) The schematic diagram of the binocular ranging principle is shown in Figure 9a.
D f = X x l = X b x r = Y y l = Y y r , D = f × b x l x r = f × b d
Initially, binocular camera calibration is conducted to derive the internal and external parameter matrices essential for rectifying the binocular camera system. Subsequently, the image captured by the binocular camera and the target’s bounding box pixel coordinates, as ascertained by the YOLOv8 algorithm, are input into the system. The image undergoes conversion to grayscale, and distortion is rectified utilizing the calibrated internal and external parameter matrices. Thereafter, the Semi-Global Block Matching (SGBM) algorithm is engaged to perform stereo matching, which ascertains just the pixel parallax at the target’s bounding boxes. Armed with these data, the target’s real-world distance is computed employing Equation (6). Finally, the calculated world distance of the target is output. The flow chart of the binocular ranging algorithm is shown in Figure 9b and Figure 10.

4. Results and Discussion

To verify the feasibility and scalability of the hardware and software solutions proposed in this system, as shown in Figure 11, an experiment was conducted on a transport vehicle measuring 6 m in length, 2.2 m in width, and 2.3 m in height. The IP cameras and camera layout utilized in the experiment are detailed in Table 2 and Table 3 of Section 2.1. The cameras, switches, and edge computing device are interconnected via network cables. The switches and edge computing device are powered by vehicle power supply. The configuration of the edge computing device [27] and the algorithmic environment of the system during the test are presented in Table 5.
To ensure the camera’s safety during the experiment and establish a fixed distance between the binocular cameras, an adjustable bracket, specifically designed for the binocular camera and fabricated using 3D printing technology with photosensitive resin material, was created. In this test, we bolted the bracket 2 m above the ground surface to the vehicle’s body. However, in actual transportation scenarios, the bracket’s fixed position should be carefully planned based on the shape and structure of the transport vehicle to further ensure the camera’s safety during the transportation of explosives.

4.1. Monocular and Binocular Camera Calibrations

The checkerboard calibration method is used to calibrate the monocular camera and the binocular camera with a baseline distance of 2 m. The calibration board is a Zhang Zhengyou checkerboard with a grid side length of 28 mm and a black and white grid count of 10 × 7. The reprojection error is obtained by calibration, and the points with a large error are selected for elimination. Finally, the pixel focal length of the monocular camera and the internal and external parameter matrices of the binocular camera are calculated. As shown in Figure 12 and Figure 13, the reprojection errors of the monocular and binocular camera calibration meet the requirements.

4.2. Front and Rear Distance Monitoring Test

With reference to the 50 m safe distance standard specified in Section 2.1, in the range of 50 m, we used the monocular cameras (A, B) and the binocular camera (a1, a2, b1, b2) to perform monocular and binocular ranging tests on 20 groups of different distances. The calculated distances (Cd) are recorded, and the maximum running time of the algorithm in the experiment is 0.13 s (the minimum FPS for image reasoning and calculation in the experiment is 7.7 f/s). The distance in the experiment is calculated by monocular and binocular ranging algorithms, while the real distance (Rd) is measured by the laser range finder. The test arrangement is shown in Figure 14. The error (E) calculation is based on Equation (7), and the results are shown in Figure 15.
E = | C d R d | R d × 100 %
As shown in Figure 15, the measurement error of the front and rear distance of the system is controlled within 5% in the range of 50 m, which meets the requirements of error in actual driving. Compared with monocular ranging, binocular ranging shows higher accuracy. In addition, the accuracy of monocular and binocular ranging is proportional to the target distance.

4.3. Surrounding Distance Monitoring Test

With reference to the 2 m safe surrounding distance standard specified in Section 2.1, in the range of 15 m, we used the binocular cameras (c1, c2, d1, d2) to perform binocular ranging tests on 20 groups of different distances. The same as in Section 4.2, the distance in the experiment is calculated by binocular ranging algorithms, while the real distance (Rd) is measured by the laser range finder. The test arrangement is shown in Figure 16. The results are shown in Figure 17.
As shown in Figure 17, the measurement error of the surrounding distance of the system is controlled within 2% in the range of 5 m, which meets the requirements of error in actual driving. And the accuracy of the binocular ranging is proportional to the target distance too.

4.4. FPS Comparative Analysis Test

The FPS (Frames Per Second) of an algorithm refers to the number of image frames that the algorithm can process within a unit of time. It is an important metric for measuring the real-time performance and efficiency of an algorithm. The higher the FPS, the better the real-time performance of the algorithm. The FPS can be calculated using Equation (8):
F P S = 1 P e r f r a m e   P r o c e s s i n g   T i m e
In order to verify the real-time performance of the system, we compared the monocular ranging method and binocular ranging method based on local matching used in the system with the binocular ranging method based on full image matching without optimization and conducted FPS tests on 16 groups of test data. The results are shown in Figure 18.
As shown in Figure 18, compared with the binocular ranging method based on local matching, the FPS of the monocular ranging method is improved by 55% on average; compared with the non-optimized binocular ranging method based on full matching, the FPS of the binocular ranging method based on local matching used in this system is improved by 23.5% on average. Therefore, the system has significant advantages in the real-time performance of early warning and ranging.

4.5. System Visual Display

In order to verify the feasibility of the system in practical application, we designed a front-end visual interface for the system. The design of the interface is shown in Figure 19. The interface displays the actual distance between the front and back of the vehicle and the distance between the surrounding objects in real time in a visual form on the corresponding position around the vehicle. If the detected distance falls below the preset safety threshold, the interface will display the value with a striking red warning sign to immediately attract the driver’s attention. Additionally, the driver can observe the vehicle’s surroundings in real time on the right side of the interface, enabling timely corresponding driving actions and effectively preventing potential risks. In order to ensure driving safety, an audible and visual alarm system will be developed in the future: early warning can be achieved through the linkage of in-vehicle lights and sounds, and when danger is detected, lights and voice prompts will be triggered synchronously so as to improve the driver’s response speed and reduce the risk of accidents [28].

5. Conclusions

This article proposes an intelligent monitoring system for the driving environment of explosives transport vehicles based on consumer-grade cameras. The system uses a consumer-grade IP camera to capture images of road conditions in real time and quickly transmits these images to edge computing devices. On the edge device, combined with deep learning and computer vision technology, target detection and ranging calculation are performed. Subsequently, the processed data are presented to the driver in a visual form so that the driver can grasp the traffic conditions and participate in the traffic operation. The system significantly improves driving safety and helps to effectively prevent traffic accidents.
(1)
A low-cost monitoring solution based on IP cameras and edge computing is proposed, filling the applicability gap of traditional high-cost systems in old transport vehicles. The system captures real-time data via IP cameras and processes them on edge devices, enabling efficient monitoring of the driving environment at a fraction of the cost of traditional systems.
(2)
The system integrates YOLOv8 for target detection with a hybrid ranging approach, combining the speed of monocular ranging and the precision of binocular ranging. This dual-mode strategy ensures rapid warnings for distant obstacles and accurate measurements in high-risk proximity zones, tailored to the unique demands of explosive transport.
(3)
A key innovation is the optimized binocular ranging algorithm, which performs stereo matching only within YOLOv8-detected bounding boxes. This optimization reduces processing latency by 23.5%, enabling real-time performance on low-cost edge devices while maintaining high accuracy.
(4)
Tests show that the system works well in real-world conditions. It keeps distance measurement errors below 5% within a 50 m range, meeting the safety needs of explosive transport. This proves that the system is reliable and robust for actual driving scenarios.
Research shortcomings and prospects: This study found that the current system, relying on consumer-grade cameras, is limited in performance under extreme lighting conditions. Additionally, its practical effectiveness in complex traffic scenarios such as rain, snow, and nighttime has not been fully validated. In future research, we will explore more advanced camera technologies to enhance system performance under extreme lighting conditions and further validate its application in complex traffic scenarios, aiming to provide more reliable outcomes for relevant research and practice.

Author Contributions

Conceptualization, J.S.; methodology, J.S., J.T., R.Z., X.L., W.J. and J.X.; validation, J.T.; formal analysis, W.J.; investigation, R.Z., X.L. and J.T.; resources, J.S.; data curation, J.T.; writing—original draft preparation, J.S. and J.T.; writing—review and editing, X.L., W.J. and J.X.; visualization, R.Z. and J.T.; supervision, J.S.; funding acquisition, J.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the State Key Laboratory of Precision Blasting and Hubei Key Laboratory of Blasting Engineering, Jianghan University (Grant No. PBSKL2022B03).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on reasonable request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yu, Q. United Nations Classification of Hazardous Goods and Hazardous Chemicals. China Pet. Chem. Stand. Qual. 2004, 5, 70–81. [Google Scholar]
  2. Guo, M. Study on Risk Assessment of Explosives in Highway Transportation. Sci. Informationization 2018, 27, 135. [Google Scholar]
  3. Sun, J.; Zheng, R.; Huang, N. Driver behavior monitoring system for explosive transport. In Proceedings of the ITSSC 2023, Xi’an, China, 10–12 November 2023. [Google Scholar]
  4. Wang, J. How to improve the supervision of “two customers and one danger”. Econ. Res. Guide 2016, 17, 172–173. [Google Scholar]
  5. Han, G.; Li, Y.; Tan, Z. Danger analysis and emergency disposal of liquefied petroleum gas highway transportation-Zhejiang Wenling ‘6.13’ tank truck explosion accident warning. Guangdong Chem. 2020, 47, 104–105+110. [Google Scholar]
  6. Liu, Y. Analysis on Safety Problems in the Transportation of Dangerous Chemicals. Chem. Ind. Manag. 2018, 20, 50–51. [Google Scholar]
  7. Zhu, X.; Wang, H.; You, H.; Zhang, W. Survey on Testing of Intelligent Systems in Autonomous Vehicles. J. Softw. 2021, 32, 2056–2077. Available online: http://www.jos.org.cn/1000-9825/6266.htm (accessed on 5 December 2023). (In Chinese).
  8. Fang, J.; Zhou, X.; Mao, X. Doppler lidar is used to achieve simultaneous distance and velocity measurements. Optoelectron. Eng. 2016, 43, 212–218. [Google Scholar]
  9. Wei, X.; Che, C.; Song, C. A review of ultrasonic sensor applications. Ind. Control. Comput. 2014, 27, 135–136+139. [Google Scholar] [CrossRef]
  10. Sun, J.; Zheng, R.; Liu, X. A Driving Warning System for Explosive Transport Vehicles Based on Object Detection Algorithm. Sensors 2024, 24, 6339. [Google Scholar] [CrossRef]
  11. Xu, X.; Zhao, M.; Shi, P. Crack Detection and Comparison Study Based on Faster R-CNN and Mask R-CNN. Sensors 2022, 22, 3. [Google Scholar] [CrossRef]
  12. Li, J.; Liang, X.; Shen, S. Scale-Aware Fast R-CNN for Pedestrian Detection. IEEE Trans. Multimed. 2018, 20, 985–996. [Google Scholar] [CrossRef]
  13. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  14. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.; Berg, A. SSD: Single shot MultiBox detector. In Computer Vision—ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016, Proceedings, Part I; Springer: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar] [CrossRef]
  15. Yun, J.; Jiang, D.; Liu, Y. Real-Time Target Detection Method Based on Lightweight Convolutional Neural Network. Front. Bioeng. Biotechnol. 2022, 10, 861286. [Google Scholar]
  16. Biswas, D.; Su, H.; Wang, C. An automatic traffic density estimation using Single Shot Detection (SSD) and MobileNet-SSD. Phys. Chem. Earth 2019, 110, 176–184. [Google Scholar] [CrossRef]
  17. Zhang, Z.; Wang, J.; Fu, X. DC-SPP-YOLO: Dense connection and spatial pyramid pooling based YOLO for object detection. Inf. Sci. 2020, 522, 241–258. [Google Scholar]
  18. Liu, Z.; Gao, Y.; Du, Q. YOLO-Extract: Improved YOLOv5 for Aircraft Object Detection in Remote Sensing Images. IEEE Access 2023, 11, 1742–1751. [Google Scholar]
  19. Li, P.; Zhao, W. Image fire detection algorithms based on convolutional neural networks. Case Stud. Therm. Eng. 2020, 19, 100625. [Google Scholar]
  20. Zhao, C.; Zhao, Z.; Chen, W. Low Luminance Measurement Based on Binocular Ranging. In Proceedings of the 2018 5th International Conference on Systems and Informatics(ICSAI), Nanjing, China, 10–12 November 2018; pp. 983–987. [Google Scholar]
  21. Zaarane, A.; Slimani, I.; Al Okaishi, W. Distance measurement system for autonomous vehicles using stereo camera. Array 2020, 5, 100016. [Google Scholar] [CrossRef]
  22. Shen, T.; Liu, W.; Wang, J. Target ranging system based on binocular stereo vision. Electron. Meas. Tech. 2015, 38, 52–54. [Google Scholar]
  23. Hao, Y.; Wen, S.; Wu, Y. Design and implementation of model-based driving assistance system. Comput. Simul. 2021, 38, 108–111+449. [Google Scholar]
  24. Hikvision-Leading the New Future of Intelligent Internet of Things. Available online: https://www.hikvision.com/en/ (accessed on 6 May 2024).
  25. Han, Q.; Liu, X.; Xu, J. Detection and Location of Steel Structure Surface Cracks Based on Unmanned Aerial Vehicle Images. J. Build. Eng. 2022, 50, 104098. [Google Scholar]
  26. Terven, J. A Comprehensive Review of YOLO Architectures in Computer Vision: From YOLOv1 to YOLOv8 and YOLO-NAS. Mach. Learn. Knowl. Extr. 2023, 5, 1680–1716. [Google Scholar] [CrossRef]
  27. TWOWIN Technology -AI Edge Computer_5G Mobile BOX Jetson Nano Orin nx AGX Xavier. Available online: https://twowinit.com/ (accessed on 21 May 2024).
  28. Zhang, Z. Acousto-Optic Double Control Automatic Alarm and Lighting Circuit Design. Adv. Mater. Res. 2014, 1079–1080, 1053–1056. [Google Scholar]
Figure 1. The application architecture of the system.
Figure 1. The application architecture of the system.
Applsci 15 04072 g001
Figure 2. Schematic diagram of binocular camera monitoring range. L and R are the locations of the cameras; θ is the horizontal field of view angle; d is the baseline distance; x is the length of the monitoring blind zone.
Figure 2. Schematic diagram of binocular camera monitoring range. L and R are the locations of the cameras; θ is the horizontal field of view angle; d is the baseline distance; x is the length of the monitoring blind zone.
Applsci 15 04072 g002
Figure 3. Camera layout of this system. A, B, a, b, c, and d are the locations of the cameras.
Figure 3. Camera layout of this system. A, B, a, b, c, and d are the locations of the cameras.
Applsci 15 04072 g003
Figure 4. The hardware architecture of the system.
Figure 4. The hardware architecture of the system.
Applsci 15 04072 g004
Figure 5. Architecture of function realization.
Figure 5. Architecture of function realization.
Applsci 15 04072 g005
Figure 6. The YOLOv8 network.
Figure 6. The YOLOv8 network.
Applsci 15 04072 g006
Figure 7. The performance of the trained model. (a) P-R curve; (b) detection performance.
Figure 7. The performance of the trained model. (a) P-R curve; (b) detection performance.
Applsci 15 04072 g007
Figure 8. The schematic diagram and flow chart of the monocular ranging algorithm. (a) Algorithm schematic diagram; (b) algorithm flow chart.
Figure 8. The schematic diagram and flow chart of the monocular ranging algorithm. (a) Algorithm schematic diagram; (b) algorithm flow chart.
Applsci 15 04072 g008
Figure 9. The schematic diagram and flow chart of the monocular ranging algorithm. (a) Algorithm schematic diagram (f is the focal length of the cameras; X, Y, and Z are the global coordinates of the object; xl, xr, yl, and yr are the projection coordinates of the object in the left and right cameras; D is the distance of the object; and b is the baseline distance); (b) algorithm flow chart.
Figure 9. The schematic diagram and flow chart of the monocular ranging algorithm. (a) Algorithm schematic diagram (f is the focal length of the cameras; X, Y, and Z are the global coordinates of the object; xl, xr, yl, and yr are the projection coordinates of the object in the left and right cameras; D is the distance of the object; and b is the baseline distance); (b) algorithm flow chart.
Applsci 15 04072 g009
Figure 10. Local matching only for the target box area.
Figure 10. Local matching only for the target box area.
Applsci 15 04072 g010
Figure 11. Test vehicle and camera arrangement.
Figure 11. Test vehicle and camera arrangement.
Applsci 15 04072 g011
Figure 12. The performance of monocular camera calibration errors.
Figure 12. The performance of monocular camera calibration errors.
Applsci 15 04072 g012
Figure 13. The performance of binocular camera calibration errors.
Figure 13. The performance of binocular camera calibration errors.
Applsci 15 04072 g013
Figure 14. Camera arrangement.
Figure 14. Camera arrangement.
Applsci 15 04072 g014
Figure 15. Analysis of front and rear distance monitoring results. (a) Monitoring results; (b) error analysis histogram.
Figure 15. Analysis of front and rear distance monitoring results. (a) Monitoring results; (b) error analysis histogram.
Applsci 15 04072 g015
Figure 16. Camera arrangement.
Figure 16. Camera arrangement.
Applsci 15 04072 g016
Figure 17. Analysis of surrounding distance monitoring results. (a) Monitoring results; (b) error analysis histogram.
Figure 17. Analysis of surrounding distance monitoring results. (a) Monitoring results; (b) error analysis histogram.
Applsci 15 04072 g017
Figure 18. Analysis of FPS comparison. (a) Monocular ranging method and binocular ranging method based on local matching used in the system; (b) binocular ranging method based on local matching used in the system and the binocular ranging method based on full image matching without optimization.
Figure 18. Analysis of FPS comparison. (a) Monocular ranging method and binocular ranging method based on local matching used in the system; (b) binocular ranging method based on local matching used in the system and the binocular ranging method based on full image matching without optimization.
Applsci 15 04072 g018
Figure 19. Visual display of this system.
Figure 19. Visual display of this system.
Applsci 15 04072 g019
Table 1. Comparison of distance measuring sensors.
Table 1. Comparison of distance measuring sensors.
SensorsPrecisionSpeedRanging RangeEnvironmental AdaptabilityCost
LidarHighFast50–300 mLight-sensitiveHigh
Ultrasonic radarLowSlow≤5 mTemperature- and humidity-sensitivelow
IP camerasMedium–highFast5–100 mAdaptablelow
Table 2. Camera model selection [24].
Table 2. Camera model selection [24].
FunctionCamera ModelFocal LengthHorizontal Field of View AngleDetection Distance
IMonocular cameraDS-2CD1245(D)-LA12 mm27°80 m
IIBinocular camera4 mm70°50 m
Table 3. Camera layout.
Table 3. Camera layout.
BaselineLocation
I A, B
II2 ma1, a2, b1, b2, c1, c2, d1, d2
Table 4. Comparison of target detection algorithms.
Table 4. Comparison of target detection algorithms.
CharacteristicYOLOv8YOLOv5Faster R-CNNDETRSSD
FPS40 FPS33.3 FPS5–10 FPS<1 FPS22.22 FPS
mAP@0.571.5%50.5%75–80%94%74%
RobustnessExcellent performance in complex scenes and small targetsLimitation in small target detectionExcellent performance in small targets but slowHigh requirements for hardwareExcellent performance in complex scenes but general adaptability to small targets
Table 5. Hardware and software configuration.
Table 5. Hardware and software configuration.
Edge Computing Device: TW-T506SSystem’s Algorithmic Environment
ProcessorNVIDIA Jetson NXPackageVersion
colored logs15.0.1
AI21TOPSflat buffers24.3.25
humanfriendly10.0
CPU6-core NVIDIA Carmel ARM® v8.2 64-bit CPU
6MB L2 + 4MB L3
mpmath1.3.0
numpy1.24.4
GPU384-core NVIDIA Volta™ GPU
with 48 Tensor Cores
onnxruntime1.17.3
opencv-python4.9.0.80
Memory8 GB 128-bit LPDDR4x 51.2GB/spackaging24.0
pillow10.3.0
Storage1× 16 GB eMMC 5.1
1× M.2 Key M NVMe 2280
pip24.0
protobuf5.26.1
DL Accelerator2× NVDLA Enginespyreadline33.4.1
setuptools68.2.2
Video output1× HDMI 2.0 @ 4Kp60sympy1.12
wheel0.41.2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sun, J.; Tang, J.; Zheng, R.; Liu, X.; Jiang, W.; Xu, J. An Intelligent Monitoring System for the Driving Environment of Explosives Transport Vehicles Based on Consumer-Grade Cameras. Appl. Sci. 2025, 15, 4072. https://doi.org/10.3390/app15074072

AMA Style

Sun J, Tang J, Zheng R, Liu X, Jiang W, Xu J. An Intelligent Monitoring System for the Driving Environment of Explosives Transport Vehicles Based on Consumer-Grade Cameras. Applied Sciences. 2025; 15(7):4072. https://doi.org/10.3390/app15074072

Chicago/Turabian Style

Sun, Jinshan, Jianhui Tang, Ronghuan Zheng, Xuan Liu, Weitao Jiang, and Jie Xu. 2025. "An Intelligent Monitoring System for the Driving Environment of Explosives Transport Vehicles Based on Consumer-Grade Cameras" Applied Sciences 15, no. 7: 4072. https://doi.org/10.3390/app15074072

APA Style

Sun, J., Tang, J., Zheng, R., Liu, X., Jiang, W., & Xu, J. (2025). An Intelligent Monitoring System for the Driving Environment of Explosives Transport Vehicles Based on Consumer-Grade Cameras. Applied Sciences, 15(7), 4072. https://doi.org/10.3390/app15074072

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop