Next Article in Journal
Predicting Tacit Coordination Success Using Electroencephalogram Trajectories: The Impact of Task Difficulty
Previous Article in Journal
Design and Implementation of an Embedded Data Acquisition System for Vehicle Vertical Dynamics Analysis
Previous Article in Special Issue
A Multi-Scale Recursive Attention Feature Fusion Network for Image Super-Resolution Reconstruction Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Lightweight Model for Real-Time Detection of Vehicle Black Smoke

1
College of Surveying and Land Information Engineering, Henan Polytechnic University, Jiaozuo 454000, China
2
School of Environment and Spatial Informatics, China University of Mining and Technology, Xuzhou 221116, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(23), 9492; https://doi.org/10.3390/s23239492
Submission received: 31 October 2023 / Revised: 21 November 2023 / Accepted: 26 November 2023 / Published: 29 November 2023
(This article belongs to the Special Issue Computer Vision Sensing and Pattern Recognition)

Abstract

:
This paper discusses the application of deep learning technology in recognizing vehicle black smoke in road traffic monitoring videos. The use of massive surveillance video data imposes higher demands on the real-time performance of vehicle black smoke detection models. The YOLOv5s model, known for its excellent single-stage object detection performance, has a complex network structure. Therefore, this study proposes a lightweight real-time detection model for vehicle black smoke, named MGSNet, based on the YOLOv5s framework. The research involved collecting road traffic monitoring video data and creating a custom dataset for vehicle black smoke detection by applying data augmentation techniques such as changing image brightness and contrast. The experiment explored three different lightweight networks, namely ShuffleNetv2, MobileNetv3 and GhostNetv1, to reconstruct the CSPDarknet53 backbone feature extraction network of YOLOv5s. Comparative experimental results indicate that reconstructing the backbone network with MobileNetv3 achieved a better balance between detection accuracy and speed. The introduction of the squeeze excitation attention mechanism and inverted residual structure from MobileNetv3 effectively reduced the complexity of black smoke feature fusion. Simultaneously, a novel convolution module, GSConv, was introduced to enhance the expression capability of black smoke features in the neck network. The combination of depthwise separable convolution and standard convolution in the module further reduced the model’s parameter count. After the improvement, the parameter count of the model is compressed to 1/6 of the YOLOv5s model. The lightweight vehicle black smoke real-time detection network, MGSNet, achieved a detection speed of 44.6 frames per second on the test set, an increase of 18.9 frames per second compared with the YOLOv5s model. The [email protected] still exceeded 95%, meeting the application requirements for real-time and accurate detection of vehicle black smoke.

1. Introduction

The urgency of controlling exhaust pollution from motor vehicles is becoming increasingly apparent. With the development of computer vision technology, automatic detection of black smoke emissions from vehicles can be achieved using road traffic surveillance videos. Extracting features related to black smoke emissions from vehicles and combining them with classifiers can enable automatic detection of black smoke. Among the various methods, deep neural networks have been used to build object detection models that are categorized into two-stage and single-stage models [1,2,3,4]. Cao et al. [5] utilized the Inceptionv3 convolutional neural network to capture spatial information in surveillance videos with suspected black smoke frames, while a long short-term memory network learned the temporal dependencies between frames. Xia et al. [6] proposed using the LeNet-5 model, based on convolutional neural networks, for vehicle black smoke recognition, achieving an accuracy of over 80% for both positive and negative samples. By introducing the Inception module, the improved LeNet-5 model achieved a 5.38% increase in accuracy for positive sample recognition. Guo et al. [7] presented a dual-branch black smoke and vehicle detection network based on the CenterNet framework. They used vehicle feature maps to guide the training of black smoke exhaust feature maps. Test results indicated that the model achieved a detection speed of up to 25.46 FPS (frames per second) and mAP (mean average precision) of 92.5% at an intersection-over-union threshold of 0.5. Zhang et al. [8,9] used both 2D and 3D convolutions to extract spatial information and spatiotemporal features from black smoke emissions. Their approach, based on the 2D-3D multi-frame classification network, achieved a recognition accuracy of 90.3% for vehicle black smoke detection, with a single-frame image inference time of 45.9 milliseconds.
Convolutional neural networks often generate a large number of redundant feature maps when extracting image features, which is caused by repetitive calculations in traditional convolution. As the number of convolutional kernels and channels continues to increase, the model’s parameter count and computational load also grow, posing a challenge for deploying vehicle black smoke automatic detection models on mobile platforms [10,11,12]. More efficient network structure designs can effectively reduce the model’s parameter count. Wang et al. [13] proposed a lightweight network, the Yolo-Light model, for deployment on the embedded platform Jetson Nano. The recognition accuracy for the vehicle black smoke test set was 91.57%, with a detection speed of 16 FPS. Zhang et al. [14] designed lightweight networks, YOLOv3-M3-CBAM and YOLOv4-GhostNet, based on YOLOv3 and YOLOv4 models; the improved models achieved a detection speed of 20 FPS. Du et al. [15] introduced a lightweight network, the RA-YOLOv5s model, which incorporates spatial and channel attention mechanisms in CSP modules to learn useful image features. Zhao et al. [16] addressed the issue of poor real-time performance in table-tennis-ball object detection by proposing a lightweight improved network, the SYOLO5 model, which compresses the model’s parameter count to one-fifth of the original model. The improved lightweight network, ShuffleNetv2, restructured the YOLOv5 model’s backbone network and introduced an efficient attention mechanism in the feature fusion process to enhance detection performance. Peng et al. [17] presented a lightweight network, the Ghost-YOLOv5-Shuffle model, which uses the GhostBottleneckCSP module and ShuffleConv module with fewer parameters to replace the C3 module in the YOLOv5 model’s backbone network, reducing the computational load in the feature channel fusion process.
Existing vehicle black smoke automatic detection models face challenges related to high parameter counts and computational load, making their deployment on mobile platforms difficult. However, lightweight model improvement solutions have not effectively balanced the accuracy of vehicle black smoke recognition with the demands of real-time detection applications. Therefore, this study, based on the excellent object detection performance of the YOLO network framework, proposes a lightweight real-time vehicle black smoke detection network called MGSNet. The experiments involved the reconstruction of the YOLOv5s model’s backbone feature extraction network using three different lightweight networks: ShuffleNetv2, MobileNetv3, and GhostNetv1. The accuracy and speed evaluation metrics of the models before and after lightweight improvements were compared to determine the best lightweight backbone network improvement solution. A new convolution module called GSConv is introduced into the neck network of the YOLOv5s model to enhance black smoke feature expression. This allows the lightweight improved vehicle black smoke real-time detection network, MGSNet, to be suitable for small-capacity embedded devices.

2. YOLOv5 Network Framework

The YOLO series algorithms for single-stage object detection can significantly improve detection speed while sacrificing only a small amount of accuracy. Therefore, this paper selects the YOLOv5 model, known for its excellent object detection performance, as the foundational network framework for lightweight improvement. Different versions of the YOLOv5 model are available, varying in terms of network depth and width, including sizes such as “s,” “m,” “l,” and “x,” and also come in versions like 1.0, 5.0, and 6.0. Taking into consideration the relatively small size of the homemade vehicle black smoke dataset used in this study and the real-time detection requirements, the decision was made to choose YOLOv5s version 6.0 as the base network for lightweight improvements [18,19]. The YOLOv5s model is structured in four main parts: the input layer, the backbone network, the neck network, and the output layer. The input layer involves image preprocessing, including mosaic data augmentation, adaptive image scaling, and adaptive anchor box adjustments. The output layer is responsible for detecting the categories and positions of target objects and is primarily composed of loss functions and non-maximum suppression techniques [20,21].
The backbone network is primarily composed of modules including the Conv (convolution) module, C3 module, and SPPF (spatial pyramid pooling fast) module. In version 6.0, the original focus module is replaced with a convolutional layer with a convolution kernel of 6, a stride of 2, and a padding of 2. This change was made to improve efficiency, especially for GPUs with limited performance. Using a convolutional layer in this context is more efficient than the focus module [22]. Version 6.0 employs the C3 module, which serves a similar role as the CSP module but differs in that it eliminates the ordinary convolution module after removing the residual output. The neck network utilizes a feature pyramid network combined with a path aggregation network structure to further process features extracted at different stages. The feature pyramid network employs upsampling to convey stronger semantic information from deeper feature maps to shallower ones, while the path aggregation network uses downsampling to transmit positional information from shallower feature maps to deeper ones. This simultaneous up and down sampling achieves multi-scale feature fusion [23,24].

3. Lightweight Model Improvements

3.1. Backbone Network Lightweight Optimization

The lightweight network ShuffleNetv2’s structural design takes into account the memory access cost of different network structures and proposes a more efficient shuffle module. It introduces the channel split operation, which divides the input feature map into two branches along the channel dimension. The results of these two branches are then connected, followed by a channel shuffle operation to merge the channel information [25,26]. In this study, the YOLOv5s model’s backbone feature extraction network, CSPDarkNet53, has been restructured using the lightweight ShuffleNetv2 network. A total of six shuffle modules are employed. The first shuffle module is preceded by a CBRM consisting of Conv, BN, ReLU, and Maxpool. The CBS module, which reconstructs the original SPPF module, is placed before the output of the backbone network and consists of Conv2d, BN, and SiLU. Figure 1b corresponds to the shuffle module with a stride of 2, directly inputting the feature map into two branches, each containing depthwise separable convolutions with a stride of 2. The feature maps are then connected, followed by the channel shuffle operation.
The lightweight MobileNetv3 not only continues the depthwise separable convolution from the v1 version but also maintains the inverted residual structure introduced in the v2 version. The novel aspect of its design is the incorporation of the SE attention mechanism, which consists of compression and excitation parts [27,28]. After global average pooling, features are sequentially input into two fully connected layers with ReLU6 and h-swish activation functions. The h-swish activation function effectively addresses the issue of complex derivative computation [29]. Taking into account the relatively small custom vehicle black smoke dataset used in this study and the requirement for real-time detection, the study chooses the lightweight MobileNetv3-small to restructure the CSPDarkNet53 backbone feature extraction network of the YOLOv5s model. A total of eleven MB modules have been used for this reconstruction. The first MB module is preceded by a CBH module. Depending on the stride, the inverted residual structure can be categorized into the two cases illustrated in Figure 2.
The novelty in the structural design of the lightweight GhostNetv1 network lies in the introduction of a new Ghost module, which condenses essential features from the input by using a small number of regular convolutions. Initially, necessary features are condensed through a series of simple linear operations, generating redundant feature layers. Finally, the feature layers generated by convolutions are concatenated with the redundant feature layers generated by linear operations [30]. The Ghost module combines the redundant features it extracts with the features obtained through simple linear operations, resulting in a large number of black smoke feature maps with lower computational load for the simple linear operations. In this study, the lightweight GhostNetv1 network has been used to restructure the CSPDarkNet53 backbone feature extraction network of the YOLOv5s model while retaining the original SPPF module. The SPPF module can fuse multiple receptive fields and separate the most significant contextual features. This allows it to extract spatial feature information of different sizes, enhancing the model’s robustness [31]. When the stride is 1, Ghost bottlenecks consist of two directly stacked Ghost modules, and the second Ghost module does not use an activation function when producing its output. When the stride is 2, depthwise separable convolution layers are introduced to compress the width and height dimensions, as shown in Figure 3.

3.2. Neck Network Lightweight Optimization

The neck network introduces a new convolution module called GSConv to restructure the standard convolution module Conv, as shown in Figure 4. The GSConv module strives to retain these connections with lower computational complexity. It combines regular convolution with depthwise separable convolution followed by channel shuffling to maintain accuracy while reducing the model’s computational load to the maximum extent [32]. The channel shuffling operation merges information generated by regular convolution with information produced by depthwise separable convolution, achieving a better balance between model accuracy and speed. Channel-dense convolutions maximize the interconnections between each channel, whereas channel-sparse convolutions completely sever these connections [33]. The GSConv module enhances the network’s non-linear expressive capability by adding depthwise separable convolution layers and channel shuffling operations. Experiments have shown that introducing the GSConv module to replace regular Conv in various stages of the model results in a deeper network [34]. When the feature maps extracted by the YOLOv5s model’s backbone network are input to the neck network, their channel size has reached its maximum value, while the width and height dimensions are at their minimum. Therefore, the neck network introduces the GSConv module to process the input feature maps, reducing redundant and repetitive information in the network.

3.3. MGSNet Model

This study is based on the YOLOv5s network framework and focuses on lightweight improvements to the backbone feature extraction network and the neck network to increase network inference speed by reducing the model’s parameter count and computational load. Various existing lightweight networks have their strengths in network structure design. The experiment involved the reconstruction of the CSPDarknet53 backbone feature extraction network using three different lightweight networks: ShuffleNetv2 [35], MobileNetv3 [36], and GhostNetv1 [37]. A comprehensive comparison of accuracy and speed evaluation metrics was conducted. Meanwhile, a new convolution module called GSConv (group shuffle convolution) [38] was introduced in the neck network to restructure the standard convolution module Conv. Deconstruction experiments were designed to validate the lightweight improvement effects of the GSConv module. Figure 5 illustrates the structural framework of the lightweight real-time vehicle black smoke detection network, MGSNet, proposed in this study based on the YOLOv5s network framework. It first employs MobileNetv3 to reconstruct the CSPDarkNet53 backbone network of the YOLOv5s model. Additionally, the neck network introduces a new convolution module, GSConv, to replace the standard Conv. The lightweight MobileNetv3 network introduces a lightweight SE (squeeze and excitation) attention module that better extracts black smoke feature information. The inverted residual structure facilitates the flow of black smoke feature information between layers, and the h-swish activation function accelerates computation [39,40]. The GSConv module combines depthwise separable convolution with standard convolution. The depthwise separable convolution reduces the model’s computational complexity, while standard convolution alleviates issues related to weak feature fusion capabilities in depthwise separable convolution, which can result in lower recognition accuracy [41,42,43].

4. Experimental Preparation

4.1. Experimental Environment and Parameter Configuration

The experimental environment configuration parameters, as shown in Table 1, were set based on previous relevant research and comparative experiments. The resolution of input images was uniformly scaled to 640 × 640. Prior to training, initial anchor boxes were clustered using the k-means algorithm, resulting in (10, 13, 16, 30, 33, 23), (30, 61, 62, 45, 59, 119), and (116, 90, 156, 198, 373, 326). The model was trained for a total of 200 epochs with a batch size of eight. The optimizer chosen was the Adam algorithm, with an initial learning rate of 1 × 10−3 and an initial decay rate of 1 × 10−5. Learning rate reduction was achieved through the cosine annealing strategy. Random dropout techniques were applied to the fully connected layers to prevent overfitting during model training, with a dropout rate of 0.5 [44,45,46].

4.2. Custom Dataset

The study aim is to perform real-time detection of whether motor vehicles on the road are emitting black smoke emissions, with research data sourced from road traffic monitoring videos. The color characteristics of the target black smoke emissions can be further explored by enriching the custom vehicle black smoke dataset through data augmentation methods such as adjusting image brightness and contrast. In the experiment, the LabelImg software (version 4.5.13) was used to annotate black smoke emissions. The dataset consisted of 2362 images of vehicle black smoke data, which were divided into a training set, a test set, and a validation set in an 8:1:1 ratio. Sample datasets before and after data augmentation are illustrated in Figure 6; Figure 6a represents the original image from the road monitoring video, while Figure 6b–d represent images processed with contrast enhancement, brightness enhancement, and color enhancement, respectively. The contrast enhancement can expand the contrast of the features of interest, while brightness and color enhancement play a role in image clarity and color distinction. Using data augmentation techniques related to color attributes not only enriches the custom vehicle black smoke dataset but also extends the image features of the areas of interest where black smoke emissions are present.

4.3. Evaluation Metrics

The confusion matrix is a metric for assessing the results of a model, helping to determine the quality of model classification. It is also known as an error matrix. Based on the results of the confusion matrix, various secondary evaluation metrics can be calculated, such as Precision, Recall, F1_Score, and mAP (mean average precision), where [email protected] [47,48] refers to the mean average precision at an intersection over union threshold of 0.5.
P = T P T P + F P
R = T P T P + F N
A P = 0 1 P R dr
m A P = 1 n i = 0 1 A P i
F 1 _ s c o r e = 2 P R P + R × 100 %
In the equation, T P represents the number of true positive samples, F P represents the number of false positive samples, and F N represents the number of false negative samples. The index i denotes different classes, and in this study, it specifically refers to black smoke emissions. The value of model parameters indicates the amount of memory space required for model computation, with smaller values suggesting less memory usage. The FPS represents the number of images the model can process per second, and a higher value indicates better real-time performance for model computation.

5. Experimental Results and Analysis

5.1. Performance Comparison of Lightweight Networks

In this study, three different lightweight networks, ShuffleNetv2, MobileNetv3, and GhostNetv1, were used, respectively, to restructure the backbone feature extraction network (CSPDarknet53) of the YOLOv5s model. Experimental comparisons were conducted to assess the performance differences before and after the improvements. Table 2 presents the test results for the models after the improvements using different lightweight networks. YOLOv5s-S represents the lightweight model with the main feature extraction network restructured using ShuffleNetv2; YOLOv5s-M represents the lightweight model with the main network replaced by MobileNetv3; YOLOv5s-G represents the lightweight model with the main network restructured using GhostNetv1.
Comparing the precision evaluation metrics of the models, YOLOv5s-G shows a slight improvement in all precision metrics, while YOLOv5s-S exhibits a decreasing trend in all metrics. YOLOv5s-M has a 1.1% increase in precision but a 3.3% decrease in recall compared with YOLOv5s. In terms of speed evaluation metrics, all three different lightweight network improvements reduce the model’s parameter count and computational complexity. YOLOv5s-S, in particular, compresses the parameter count to well below one million, representing only 1/10 of the original model’s parameters. YOLOv5s-M also reduces the parameter count to 1/6 of the original model. YOLOv5s-S exhibits a notable advantage in inference speed, while YOLOv5s-G surpasses the original in detection precision. However, YOLOv5s-M strikes a better balance between model precision and speed, ensuring both accuracy and lightweight model design. Its inference speed meets the real-time requirements for vehicle black smoke detection. As a result, this study ultimately chose to restructure the backbone feature extraction network CSPDarknet53 of YOLOv5s using MobileNetv3 for lightweight improvement.

5.2. Ablation Experiments

The neck network of the YOLOv5s model is primarily composed of two upsampling modules, four CBS modules, and four C3 modules, where both the CBS and C3 modules contain standard Conv convolution layers. In this study, a new convolution module, GSConv, was introduced to replace the standard Conv convolution modules, and a set of ablation experiments were conducted to verify the lightweight improvement of the new GSConv module. The results of the ablation experiments are presented in Table 3. In the table, “YOLOv5s-GS-CBS” denotes the introduction of GSConv to restructure the Conv in the CBS modules of the neck network, “YOLOv5s-GS-C3” represents the introduction of GSConv to restructure the Conv in the C3 modules, “YOLOv5s-GS” means the introduction of GSConv to restructure all the Conv modules in the neck network of the YOLOv5s model, and “MGSNet” signifies the introduction of GSConv to restructure all the Conv modules in the neck network of the YOLOv5s-M lightweight model. Introducing GSConv in the CBS and C3 modules in the neck network effectively reduced the parameter count of the YOLOv5s model. Particularly, replacing Conv with GSConv in the C3 modules yielded better results, reducing the model’s parameter count by 2.7%. Therefore, the introduction of the new convolution module GSConv restructures all the regular convolution modules in the neck network of the lightweight YOLOv5s-M model. The real-time vehicle smoke detection network after lightweight improvement is referred to as MGSNet.
Comparing the evaluation metrics before and after the lightweight model improvement, it is evident that the precision of MGSNet slightly decreases after lightweight improvement. Specifically, precision, recall, [email protected], and F1_score decrease by 1%, 2.1%, 2.9%, and 2% respectively, in comparison to the YOLOv5s model. However, the [email protected] value still reaches 95.4%. After lightweight improvement, MGSNet experienced a significant reduction in model parameters and computational complexity. The parameter count has been compressed to 1/6 of the original model, and the number of images processed per second increased from 25 images to 44 images. The results of the ablation experiments demonstrate the feasibility of the lightweight improvement proposed in this study, as it can significantly reduce model parameters and computational complexity.
The change curve of [email protected] before and after lightweight improvements is shown in Figure 7; the red curve represents the MGSNet model, and the black curve represents the YOLOv5s model. The model was trained for 200 epochs, and after about 50 epochs of training, the [email protected] value of the MGSNet model reached around 0.95 with minimal fluctuations. The model exhibited faster learning efficiency, and its accuracy remained almost unchanged before and after the lightweight improvements. The change in the P-R curve during the training of the MGSNet model is shown in Figure 8, with the horizontal axis representing the recall rate and the vertical axis representing the precision rate. Precision decreases as recall rate increases, and the area enclosed by the P-R curve corresponds to the [email protected] value.

5.3. Comparison of Existing Algorithms

To demonstrate the real-time detection advantage of the lightweight improvement model MGSNet proposed in this study, experiments compared the MGSNet model with the YOLOv5s model and other lightweight improvement models based on the YOLOv5s network framework using our custom vehicle smoke dataset. The test results of different models are shown in Table 4. The RA-YOLOv5s model focuses on lightweight improvements to the residual units of the YOLOv5s backbone CSP module. The test results indicate that the model’s parameter count was reduced by only 5%, while there was a significant decrease in detection accuracy. The SYOLO5 model involves the reconstruction of the YOLOv5 model’s backbone CSPDarkNet53 using the lightweight network ShuffleNetv2, compressing the model’s parameter count to 1/5 of the YOLOv5s model. The inference time for a single image is 23.6 ms, and the lightweight improvement effect is more noticeable. The Ghost-YOLOv5-Shuffle model is capable of significantly reducing the model’s parameter count, thanks to the replacement of original modules with lightweight Ghost and shuffle modules. Figure 9 displays real-time detection results of the lightweight improvement model MGSNet proposed in this study, accurately identifying vehicle smoke in the video.

6. Conclusions

This study proposes a lightweight real-time detection model for black-smoke-emitting vehicles, named MGSNet, based on the YOLOv5s framework. The goal is to meet the requirements of real-time detection of black-smoke-emitting vehicles and reduce the deployment complexity for mobile applications. The approach involves restructuring the main feature extraction network CSPDarkNet53 using the lightweight MobileNetv3 for YOLOv5s, and introducing a novel convolution module, GSConv, to restructure all standard Conv modules in the neck network. The experimental results demonstrate that the lightweight MGSNet model achieved significant reductions in parameter count, down to 1/6 of the YOLOv5s model’s parameters. The inference time for a single image has been reduced by 16.5 ms. The model can process 44 images per second, meeting the real-time requirements for automatic detection of black-smoke-emitting vehicles. MGSNet exhibited excellent performance on the test set, with a precision of 95.7%, recall of 94.7%, [email protected] of 95.4%, and an F1 score of 95%. During testing, it was observed that black smoke emissions occupy a small portion of the entire image. Future research will focus on optimizing the network structure to ensure detection accuracy for small objects, while maintaining the improved inference speed of the MGSNet model.

Author Contributions

Conceptualization, K.C.; methodology, H.W. and Y.Z.; software, K.C.; validation, K.C.; formal analysis, Y.Z.; investigation, H.W.; resources, K.C. and Y.Z.; data curation, K.C. and Y.Z.; writing—original draft preparation, K.C.; writing—review and editing, H.W.; visualization, H.W.; supervision, H.W. and K.C.; project administration, K.C. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the Fundamental Research Funds for the Central Universities (No. 2023KYJD1003), National Natural Science Foundation of China (No. 42075132; No. 41975036), Natural Science Foundation of Jiangsu Province Basic Research Program (No. BK20231502).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are not publicly available due to privacy restrictions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liao, X.; Zeng, X. Review of target detection algorithm based on deep learning. In Proceedings of the 2020 International Conference on Artificial Intelligence and Communication Technology (AICT), Chongqing, China, 28–29 March 2020; pp. 55–59. [Google Scholar]
  2. Ge, H.; Dai, Y.; Zhu, Z.; Zang, X. Single-stage underwater target detection based on feature anchor frame double optimization network. Sensors 2022, 22, 7875. [Google Scholar] [CrossRef]
  3. Ellouze, A.; Ksantini, M.; Delmotte, F.; Karray, M. Multiple object tracking: Case of aircraft detection and tracking. In Proceedings of the 2019 16th International Multi-Conference on Systems, Signals & Devices (SSD 2019), Istanbul, Turkey, 21–24 March 2019; pp. 473–478. [Google Scholar]
  4. Ellouze, A.; Ksantini, M.; Delmotte, F.; Karray, M. Single object tracking applied to an aircraft. In Proceedings of the 2018 15th International Multi-Conference on Systems, Signals & Devices (SSD 2018), Hammamet, Tunisia, 19–22 March 2018; pp. 1441–1446. [Google Scholar]
  5. Cao, Y.; Lu, X. Learning spatial-temporal representation for smoke vehicle detection. Multimed. Tools Appl. 2019, 78, 27871–27889. [Google Scholar] [CrossRef]
  6. Xia, X. Research on Smoke Vehicle Detection Technology Based on Video Image. Master’s Thesis, Southeast University, Nanjing, China, 2019. [Google Scholar]
  7. Guo, T.; Ren, M. Dual branch network for black smoke and vehicle detection based on attention mechanism. Comput. Digit. Eng. 2022, 50, 147–151. [Google Scholar]
  8. Zhang, G.; Zhang, D.; LU, X.; Cao, Y. Smoky vehicle detection algorithm based on improved transfer learning. In Proceedings of the 2019 6th International Conference on Systems and Informatics (ICSAI), Shanghai, China, 2–4 November 2019; pp. 155–159. [Google Scholar]
  9. Zhang, G. Research on Smoke Vehicle Detection Technology Based on Deep Learning. Master’s Thesis, Southeast University, Nanjing, China, 2021. [Google Scholar]
  10. Isana, F.; Taichi, Y.; Xi, Z.; Masahiro, I. Image adjustment for multi-exposure images based on convolutional neural networks. IEICE Trans. Inf. Syst. 2022, 1, 123–133. [Google Scholar]
  11. Pelletier, C.; Webb, G.I.; Petitjean, F. Temporal convolutional neural network for the classification of satellite image time series. Remote Sens. 2019, 11, 523. [Google Scholar] [CrossRef]
  12. Anowar, F.; Sadaoui, S.; Selim, B. Conceptual and empirical comparison of dimensionality reduction algorithms (PCA, KPCA, LDA, MDS, SVD, LLE, ISOMAP, LE, ICA, t-SNE). Comput. Sci. Rev. 2021, 40, 100378. [Google Scholar] [CrossRef]
  13. Wang, P.; Wang, H.; Yu, F.; Xia, W. A high-precision fast smoky vehicle detection method based on improved Yolov5 network. In Proceedings of the 2021 IEEE International Conference on Artificial Intelligence and Industrial Design (AIID), Guangzhou, China, 28–30 May 2021; pp. 255–259. [Google Scholar]
  14. Zhang, Q. Research on Smoky Vehicle Detection Technology Based on Computer Vision. Master’s Thesis, Hebei University of Science and Technology, Shijiazhuang, China, 2021. [Google Scholar]
  15. Du, C.; Wang, Y.; Liu, S.; Song, X. Granary pest detection model based on RA-YOLOv5s. Comput. Simul. 2023, 40, 486–491. [Google Scholar]
  16. Zhao, Y.; Wang, Q.; Sha, J.; Guo, Q. Lightweight object detection algorithm for table tennis based on YOLOv5s. J. Comput. Appl. 2023, 43, 229–234. [Google Scholar]
  17. Peng, C.; Zhang, Q.; Tang, Z.; Gui, W. Research on mask wearing detection method based on YOLOv5 enhancement model. Comput. Eng. 2022, 48, 39–49. [Google Scholar]
  18. Lindenheim-Locher, W.; Świtoński, A.; Krzeszowski, T.; Paleta, G.; Hasiec, P.; Josiński, H.; Paszkuta, M.; Wojciechowski, K.; Rosner, J. YOLOv5 drone detection using multimodal data registered by the vicon system. Sensors 2023, 23, 6396. [Google Scholar] [CrossRef]
  19. Tang, H.; Liang, S.; Yao, D.; Qiao, Y. A visual defect detection for optics lens based on the YOLOv5-C3CA-SPPF network model. Opt. Express 2023, 31, 2628–2643. [Google Scholar] [CrossRef]
  20. Yi, X.; Qian, C.; Wu, P.; Maponde, B.T.; Jiang, T.; Ge, W. Research on fine-grained image recognition of birds based on improved YOLOv5. Sensors 2023, 23, 8204. [Google Scholar] [CrossRef]
  21. Dong, W.; Pan, L.; Zhang, Q.; Zhang, W. Athlete target detection method in dynamic scenario based on nonlinear filtering and YOLOv5. Alex. Eng. J. 2023, 82, 208–217. [Google Scholar] [CrossRef]
  22. Yar, H.; Khan, Z.; Ullah, F.; Ullah, W.; Baik, S. A modified YOLOv5 architecture for efficient fire detection in smart cities. Expert Syst. Appl. 2023, 231, 120465. [Google Scholar] [CrossRef]
  23. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [PubMed]
  24. Yang, J.; Fu, X.; Hu, Y.; Huang, Y.; Ding, X.; Paisley, J. PanNet: A deep network architecture for pan-sharpening. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Honolulu, HI, USA, 21–26 July 2017; pp. 1753–1761. [Google Scholar]
  25. Liu, M.; Wan, L.; Wang, B.; Wang, T. SE-YOLOv4: Shuffle expansion YOLOv4 for pedestrian detection based on PixelShuffle. Appl. Intell. 2023, 53, 18171–18188. [Google Scholar] [CrossRef]
  26. Philippe, M.; Isaac, R. Shuffle polygraphic resolutions for operads. J. Lond. Math. Soc. 2022, 107, 61–122. [Google Scholar]
  27. Guha, R.; Nassir, N.; Christian, W. Recalibrating fully convolutional networks with spatial and channel “squeeze and excitation” blocks. IEEE Trans. Med. Imaging 2019, 38, 540–549. [Google Scholar]
  28. Jiang, H.; Ma, L.; Wang, X.; Zhang, J.; Liu, Y.; Wang, D.; Wu, P.; Han, W. Focus prediction of medical microscopic images based on lightweight densely connected with squeeze-and-excitation network. Front. Neurosci. 2023, 17, 1213176. [Google Scholar] [CrossRef]
  29. Liu, J.; Song, S.; Wang, J.; Balaiti, M.; Song, N.; Li, S. Flatness prediction of cold rolled strip based on deep neural network with improved activation function. Sensors 2022, 22, 656. [Google Scholar] [CrossRef]
  30. Zheng, H.; Liu, H.; Qi, W.; Xie, H. Little-YOLOv4: A Lightweight pedestrian detection network based on YOLOv4 and GhostNet. Wirel. Commun. Mob. Comput. 2022, 2022, 5155970. [Google Scholar] [CrossRef]
  31. Yang, Y.; Wang, L.; Huang, M.; Zhu, Q.; Wang, R. Polarization imaging based bruise detection of nectarine by using ResNet-18 and ghost bottleneck. Postharvest Biol. Technol. 2022, 189, 111916. [Google Scholar] [CrossRef]
  32. Yu, M.; Wan, Q.; Tian, S.; Hou, Y.; Wang, Y.; Zhao, J. Equipment identification and localization method based on improved YOLOv5s model for production line. Sensors 2022, 22, 10011. [Google Scholar] [CrossRef] [PubMed]
  33. Liu, P.; Wang, Q.; Zhang, H.; Mi, J.; Liu, Y. A lightweight object detection algorithm for remote sensing images based on attention mechanism and YOLOv5s. Remote Sens. 2023, 15, 2429. [Google Scholar] [CrossRef]
  34. Wang, J.; Li, Q.; Fang, Z.; Zhou, X.; Tang, Z.; Han, Y.; Ma, Z. YOLOv6-ESG: A lightweight seafood detection method. J. Mar. Sci. Eng. 2023, 11, 1623. [Google Scholar] [CrossRef]
  35. Cengil, E.; Çınar, A.; Yıldırım, M. An efficient and fast lightweight-model with ShuffleNetv2 based on YOLOv5 for detection of hardhat-wearing. Rev. Comput. Eng. Stud. 2022, 9, 116–123. [Google Scholar] [CrossRef]
  36. Howard, A.; Sandler, M.; Chu, G.; Chen, L.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudenvan, V.; et al. Searching for MobileNetV3. IEICE Trans. Fundam. Electron. Commun. Comput. Sci. 2019, 1905, 02244. [Google Scholar]
  37. Hai, K.; Wang, Y.; Tian, Q.; Guo, J.; Xu, C.; Xu, C. GhostNet: More feature from cheap operations. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020; pp. 1577–1586. [Google Scholar]
  38. Zhang, M.; Gao, F.; Yang, W.; Zhang, H. Wildlife object detection method applying segmentation gradient flow and feature dimensionality reduction. Electronics 2023, 12, 377. [Google Scholar] [CrossRef]
  39. Zheng, H.; Duan, J.; Dong, Y.; Liu, Y. Real-time fire detection algorithms running on small embedded devices based on MobileNetV3 and YOLOv4. Fire Ecol. 2023, 19, 31. [Google Scholar] [CrossRef]
  40. Abd Elaziz, M.; Dahou, A.; Alsaleh, N.A.; Elsheikh, A.H.; Saba, A.I.; Ahmadein, M. Boosting COVID-19 image classification using MobileNetV3 and aquila optimizer algorithm. Entropy 2021, 23, 1383. [Google Scholar] [CrossRef]
  41. Howard, A.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  42. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L. MobileNetV2: Inverted residuals and linear bottlenecks. In Proceedings of the 2018 Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–21 June 2018; pp. 4510–4520. [Google Scholar]
  43. Lang, J.; Liu, Y. LCCF-Net: Lightweight contextual and channel fusion network for medical image segmentation. Biomed. Signal Process. Control 2023, 86, 105134. [Google Scholar] [CrossRef]
  44. Wei, J.; Zhang, X.; Zhuo, Z.; Ji, Z.; Wei, Z.; Li, J.; Li, Q. Leader population learning rate schedule. Inf. Sci. 2023, 623, 455–468. [Google Scholar] [CrossRef]
  45. Yang, Y.; Ahmadipour, P.; Shanechi, M. Adaptive latent state modeling of brain network dynamics with real-time learning rate optimization. J. Neural Eng. 2021, 18, 036013. [Google Scholar] [CrossRef]
  46. Fan, L.; Zhang, T.; Zhao, X.; Wang, H.; Zheng, M. Deep topology network: A framework based on feedback adjustment learning rate for image classification. Adv. Eng. Inform. 2019, 42, 100935. [Google Scholar] [CrossRef]
  47. Kaavya, K.; Lakshmi, G. A new 3D convolutional neural network (3D-CNN) framework for multimedia event detection. Signal Image Video Process. 2020, 15, 779–787. [Google Scholar]
  48. Arora, A.; Grover, A.; Chugh, R.; Reka, S. Real time multi object detection for blind using single shot multibox detector. Wirel. Pers. Commun. 2019, 107, 651–661. [Google Scholar] [CrossRef]
Figure 1. Shuffle module structure diagram: (a) stride = 1; (b) stride = 2.
Figure 1. Shuffle module structure diagram: (a) stride = 1; (b) stride = 2.
Sensors 23 09492 g001
Figure 2. Inverted residual structure diagram: (a) stride = 1; (b) stride = 2.
Figure 2. Inverted residual structure diagram: (a) stride = 1; (b) stride = 2.
Sensors 23 09492 g002
Figure 3. Ghost bottlenecks structure diagram: (a) stride = 1; (b) stride = 2.
Figure 3. Ghost bottlenecks structure diagram: (a) stride = 1; (b) stride = 2.
Sensors 23 09492 g003
Figure 4. Convolution module GSConv structure diagram.
Figure 4. Convolution module GSConv structure diagram.
Sensors 23 09492 g004
Figure 5. The network architecture diagram of the MGSNet model.
Figure 5. The network architecture diagram of the MGSNet model.
Sensors 23 09492 g005
Figure 6. Effect images of different data augmentation methods: (a) original image; (b) contrast enhancement; (c) brightness enhancement; (d) color enhancement.
Figure 6. Effect images of different data augmentation methods: (a) original image; (b) contrast enhancement; (c) brightness enhancement; (d) color enhancement.
Sensors 23 09492 g006
Figure 7. [email protected] change curve graph before and after lightweight improvements.
Figure 7. [email protected] change curve graph before and after lightweight improvements.
Sensors 23 09492 g007
Figure 8. P-R curve graph for the MGSNet model training.
Figure 8. P-R curve graph for the MGSNet model training.
Sensors 23 09492 g008
Figure 9. Test results of the MGSNet model.
Figure 9. Test results of the MGSNet model.
Sensors 23 09492 g009
Table 1. Experimental environment configuration.
Table 1. Experimental environment configuration.
NameVersion Model
Operating systemWindows10
CPUIntel(R) Core (TM) i5-11400F @2.60 GHz
GPUNVIDIA GeForce GTX 1650
Programming languagePython3.8.13
Deep learning frameworkPytorch1.13.0, CUDA11.7
Table 2. Test results of different lightweight network improvements.
Table 2. Test results of different lightweight network improvements.
ModelP (%)R (%)mAP (%)F1_Score (%)Parameters (M)FPS
YOLOv5s0.9670.9680.9830.97701282225.7
YOLOv5s-S0.9550.9470.9550.9484235852.0
YOLOv5s-M0.9780.9350.9670.96135445437.6
YOLOv5s-G0.9690.9880.9870.98507897431.5
Table 3. Results of ablation experiments.
Table 3. Results of ablation experiments.
ModelP (%)R (%)mAP (%)F1_Score (%)Parameters (M)FPS
YOLOv5s-GS-CBS0.9630.9620.9810.97690584627.0
YOLOv5s-GS-C30.9580.9600.9800.96682589627.6
YOLOv5s-GS0.9610.9590.9780.96676432228.5
MGSNet0.9570.9470.9540.95128020444.6
Table 4. Comparison of different lightweight improvement models.
Table 4. Comparison of different lightweight improvement models.
ModelP (%)mAP (%)Parameters (M)FPS
YOLOv5s0.9670.983701282225.7
RA-YOLOv5s [15]0.9070.916665866027.5
SYOLO5 [16]0.9420.932138094742.4
Ghost-YOLOv5-Shuffle [17]0.9360.938241919134.6
MGSNet0.9570.954128020444.6
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, K.; Wang, H.; Zhai, Y. A Lightweight Model for Real-Time Detection of Vehicle Black Smoke. Sensors 2023, 23, 9492. https://doi.org/10.3390/s23239492

AMA Style

Chen K, Wang H, Zhai Y. A Lightweight Model for Real-Time Detection of Vehicle Black Smoke. Sensors. 2023; 23(23):9492. https://doi.org/10.3390/s23239492

Chicago/Turabian Style

Chen, Ke, Han Wang, and Yingchao Zhai. 2023. "A Lightweight Model for Real-Time Detection of Vehicle Black Smoke" Sensors 23, no. 23: 9492. https://doi.org/10.3390/s23239492

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop