Next Article in Journal
Network Thermodynamics-Based Scalable Compartmental Model for Multi-Strain Epidemics
Next Article in Special Issue
Resolving Cross-Site Scripting Attacks through Fusion Verification and Machine Learning
Previous Article in Journal
Anti-Forcing Spectra of Convex Hexagonal Systems
Previous Article in Special Issue
Syntactically Enhanced Dependency-POS Weighted Graph Convolutional Network for Aspect-Based Sentiment Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Joint Semantic Intelligent Detection of Vehicle Color under Rainy Conditions

1
School of Communications and Information Engineering, Xi’an University of Posts and Telecommunications, Xi’an 710121, China
2
Department of Statistics & Data Science, Southern University of Science and Technology, Shenzhen 518055, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2022, 10(19), 3512; https://doi.org/10.3390/math10193512
Submission received: 9 August 2022 / Revised: 15 September 2022 / Accepted: 20 September 2022 / Published: 26 September 2022

Abstract

:
Color is an important feature of vehicles, and it plays a key role in intelligent traffic management and criminal investigation. Existing algorithms for vehicle color recognition are typically trained on data under good weather conditions and have poor robustness for outdoor visual tasks. Fine vehicle color recognition under rainy conditions is still a challenging problem. In this paper, an algorithm for jointly deraining and recognizing vehicle color, ( J A D A R ), is proposed, where three layers of U N e t are embedded into R e t i n a N e t - 50 to obtain joint semantic fusion information. More precisely, the U N e t subnet is used for deraining, and the feature maps of the recovered clean image and the extracted feature maps of the input image are cascaded into the Feature Pyramid Net ( F P N ) module to achieve joint semantic learning. The joint feature maps are then fed into the class and box subnets to classify and locate objects. The R a i n V e h i c l e C o l o r - 24 dataset is used to train the J A D A R for vehicle color recognition under rainy conditions, and extensive experiments are conducted. Since the deraining and detecting modules share the feature extraction layers, our algorithm maintains the test time of R e t i n a N e t - 50 while improving its robustness. Testing on self-built and public real datasets, the mean average precision ( m A P ) of vehicle color recognition reaches 72.07 % , which beats both sate-of-the-art algorithms for vehicle color recognition and popular target detection algorithms.

1. Introduction

Vehicle information recognition has been applied in the field of intelligent traffic management and criminal investigation. License plate, model, and vehicle color comprise the main vehicle information. Although license plate recognition is a commonly used vehicle information recognition technology [1], it also faces many challenges in criminal investigation and for intelligent traffic law enforcement, as license plates can be easily obscured (partially or fully) or faked/duplicated by criminals. As it can still be identified despite partial occlusion or viewpoint changes, vehicle color recognition is widely applied in video surveillance [2], vehicle detection [3], vehicle tracking [4], automatic driving [5,6], criminal investigation [7], etc. All the above-mentioned tasks inevitably encounter adverse weather conditions, especially rain. This, in turn, adversely affects the performance of object recognition/retrieval, because rain can significantly reduce the contrast of the scene and reduces visibility, compromising image quality. Many scholars have conducted research on how to improve the performance of object detection under rainy conditions.
Vehicle color recognition methods are typically classified as traditional model-driven [8,9,10,11] or data-driven deep learning [12,13,14,15,16,17,18]. Traditional methods usually use handcrafted feature descriptors to extract visual features and train a classifier to recognize vehicle color. For example, Chen et al. [8] select the region of interest (ROI) of the vehicle to recognize its dominant color and than train a linear support vector machine to classify it. Jeong et al. [9] adopt A d a B o o s t to classify an H S V histogram of the vehicle’s homogeneity patches into seven color categories.
Deep neural networks have been employed to learn effective feature representations from raw pixels, which has proven to be more powerful than traditional methods. To be more specific, these deep learning methods fall into two groups: general object detection algorithms [19,20,21,22,23,24] applied to vehicle color recognition [16,17,18] and algorithms specially designed for color recognition. All these algorithms are trained on datasets with 7–24 colors [8,9,10,18] obtained under normal weather conditions. Of course, there exists some research to address object recognition under rainy conditions; basically, these proposed methods adopt two-stage instead of end-to-end procedures, which inevitably increases the running time of the entire task.
On the other hand, a number of scholars have paid attention to joint processing of low-level and high-level tasks. Generally, they improve the robustness of object detection [25,26,27,28] by embedding domain adaptation, image restoration, style transfer, or other modules into the object detection framework, or by a few-shot transfer learning mechanism [29,30,31,32]. These methods have explored the robustness of performance for downstream tasks in many harsh environments except under rainy weather conditions, which motivated our work in the present paper.
In this paper, a Joint Algorithm for Deraining And Recognition (JADAR) is proposed for fine recognition of vehicle color under rainy conditions. The network architecture is shown in Figure 1. To be more specific, we embed the three-layer decoder of U N e t - 3 into the last three layers of the feature extraction submodule of R e t i n a N e t - 50 . The main contributions are as follows:
  • J A D A R contains far fewer parameters than previous two-stage methods since its subnets, U N E T and R e t i n a N e t share the same feature extracting layers. This is of high practical value as the size of outdoor mobile equipment can be substantially reduced.
  • In J A D A R , the multi-scale fusion information obtained by cascading feature maps of original rainy images and recovered images are fed into the subsequent class-box subnet. In so doing, multi-scale information across domains is crucially beneficial for fine vehicle color recognition.
  • The joint processing of low-level and high-level tasks can be mutually beneficial. Embedding the image restoration module can help improve the performance of subsequent high-level tasks under severe weather; conversely, the performance of subsequent high-level tasks as evaluation metrics can, in turn, fine-tune the image restoration algorithms.
  • Comprehensive experiments show that our proposed methods outperform basic detection networks and two-stage network and transfer learning methods for the task of color recognition under rainy conditions. Further, our training and testing times are shortened.
Next, related work is introduced in Section 2. J A D A R is constructed in Section 3. Section 4 shows that our method is superior to the state-of-the-art quantitatively and qualitatively. Section 5 concludes the main content.

2. Related Work

There exists some research on vehicle color recognition under normal weather and object detection under adverse weather conditions, which is reviewed below.

2.1. Vehicle Color Recognition under Normal Weather Conditions

Vehicle color recognition methods generally fall into traditional model-based methods [8,9,10,11] and data-based deep learning methods [12,13,14,15,16,17,18]. Regarding traditional model-based methods, Chen et al. [8] train a linear support vector machine classifier on the region of interest R O I in vehicle images based on eight color types; Jeong et al. [9] adopt the multi-class AdaBoost algorithm to classify the color of front-of-vehicle images into seven types; Dule et al. [11] train three classifiers (KNN, ANN, and SVM) for two R O I s (smooth hood section and semi-front of vehicle).
Data-based methods have been receiving increasing attention for vehicle color recognition. Hu et al. [12] were the first to apply a convolutional neural network (CNN) with a spatial pyramid strategy to boost the accuracy of vehicle color recognition. Zhang et al. [15] proposed a lightweight C N N for vehicle color recognition. Fu et al. [16] designed M C F F - C N N (Multiscale Comprehensive Feature Fusion Convolutional Neural Network) to recognize eight vehicle colors. Hu et al. [18] proposed vehicle color recognition based on a Smooth Modulation Neural Network with Multi-Scale Feature Fusion ( S M N N - M S F F ).
It is worth mentioning that there has been no research on vehicle color detection under bad weather conditions, which is the focus of this paper.

2.2. Object Detection under Adverse Weather Conditions

Bad weather includes rain, snow, haze, etc. The quality of outdoor images or videos collected in these weathers is severely degraded, so target-detection models trained on high-quality images have difficulty handling bad weather. This challenge has been investigated by many scholars with many solutions provided, such as embedding a domain-adaptation module [25,26,27,28,31,33,34] into the object detection backbone, such as Y O L O , F a s t e r R C N N , R e t i n a N e t , etc., two-stages methods consisting of preprocessing and object detection [35,36,37,38,39,40], or using a few-shot transfer learning mechanism [29,30,31,32].
For example, Chen et al. [33] embedded two domain adaptation modules into F a s t e r R C N N to reduce the domain discrepancy on image level and instance level. Sindagi et al. [31] proposed an unsupervised domain-adapting method to improve generalization of object detection under hazy and rainy conditions. Style transfer is considered in [27], in which the authors construct a cross-domain representation learning method including domain diversification and a multi-domain invariant. Huang et al. [41] combine dual subnet frameworks for object detection under foggy conditions.
Except for the two-stage methods, the above-mentioned methods do not pay special attention to the rainy conditions. However, two-stage methods such as [35,36,37,38,39,42] do pay attention to image deraining instead of object detection. in other words, the joint tasks of deraining and object detection are not taken seriously. Motivated by the above considerations, we propose J A D A R for joint semantic intelligent detection of vehicle color in rainy scenes.

3. JADAR Algorithm

3.1. Fusion Network Design

In this paper, a Joint Algorithm for Deraining And Recognition (JADAR) is designed for vehicle color recognition in inclement weather conditions; it is based on R e t i n a N e t - 50 , as shown in Figure 1. In Figure 1, O is the rainy image input, B is the corresponding clean background image, B ^ and y o , respectively, are the outputs of the deraining and detecting results. To see results clearly, we zoom in on the recognition results of each car in picture y i o ( i = 1 , 2 , 3 ) ; y 1 o is the enlarged result of the first car in the picture—the recognition color is silver-gray with a confidence level of 0.91 ; y 2 o is the enlarged result of the second car in the picture—the recognition color is black with a confidence level of 0.58; y 3 o is the enlarged result of the third car in the picture—the recognition color is dark gray with a confidence level of 0.81 . The green/blue/purple/orange boxes represent the feature extraction module/UNet-3/ information fusion module/class+bbox subnets, respectively. L r e g is the regression loss using the smooth L1 loss. L c l s is the classification loss using the focal loss. The loss function for deraining is MSE loss. J A D A R is trained by the weighted sum of these three losses (see Equation (7)).
The J A D A R framework is designed by embedding the three-layer decoder of UNet-3 [43] into the last three sub-blocks of the feature extraction module, as illustrated by the green-tinted box in Figure 1. The whole framework consists of four main modules: image feature extraction module, deraining module, information fusion module, and c l a s s + b o x subnets. The rain removal and feature extraction modules share three layers, avoiding extra computational burden. In fact, Section 4.5 shows that J A D A R has the same testing time as R e t i n a N e t - 50 . The last three feature maps and the corresponding recovered feature maps are cascaded together and then fed into their respective c l a s s + b o x subnets, which can learn multi-scale joint semantic representations to improve object detection accuracy under rainy conditions. The feature fusion sub-module setting is illustrated in Figure 2.
The overall object function is back-propagated to train the deraining module and to improve rainy image deraining performance recursively. The object detection backbone network uses three-scale c l a s s + b o x subnets to leverage multi-scale fusion color feature maps to classify 24 car color types and locate the bounding-box.

3.2. Model Formulation and Model Optimization

Let the physical mechanism of rainy image corruption be
x = y + z
where x , y , z denote rainy image O, recovered clean background image B, and rain layer R, respectively. To tackle the problem of supervised vehicle object detection by color in inclement weather conditions, a joint network is proposed to learn joint semantic representation from an input rainy image x. Let y denote the corresponding label of rainy image x.
As demonstrated by the green box in Figure 1, the last three feature maps f 1 ( x ) , f 2 ( x ) , f 3 ( x ) are taken from the feature extraction sub-blocks of R e t i n a N e t . Then, f 1 ( x ) is fed into the corresponding last layer of the decoder of U N e t -3, and g 1 ( x ) is output. Next, g 1 ( x ) and f 2 ( x ) are cascaded into the penultimate layer U N e t -3, and g 2 ( x ) is output; then g 2 ( x ) and f 3 ( x ) are cascaded into the last decoder layer of U N e t -3, and g 3 ( x ) is output. The output of the deraining module y ^ is denoted by g 3 ( x ) . Thus, the mean square error ( M S E loss) is used as deraining object function L d e r as follows:
L d e r = 1 n i = 1 n ( y ^ y ) 2
where n is the number of rainy images. Finally, f 1 ( x ) and g 1 ( x ) , f 2 ( x ) and g 2 ( x ) , and f 3 ( x ) and g 3 ( x ) are cascaded and input into differently scaled c l a s s + b o x subnets, where joint semantic information is fused, 24 vehicle colors are classified, and box-bounded regressions are achieved; the last cascading output image is denoted y o .
The classification loss function is
L c l s ( p i t ) = i = 1 C ( α t ( 1 p i t ) γ log ( p i t ) + ( 1 α t ) ( p i t ) γ log ( p i t ) )
where α t is a balancing factor to balance the uneven proportion of positive and negative examples of every vehicle color category, C = 24 denotes the number of all vehicle color categories, γ 0 is a tunable focusing parameter (we take γ = 2.0 in Section 4 following [24]), t is equal to 0 or 1, which denotes the positive or negative sample, p i 1 [ 0 , 1 ] denotes the prediction probability of the positive sample of the i - th vehicle color class, and 1 p i 1 indicates the prediction probability of negative examples of every vehicle color category i { 1 , 2 , , 24 } ; i.e.,
p i t = { r c 1 p i if t = 1 1 p i if o t h e r w i s e .
The loss function of the box bounding regression is
L r e g = 1 n i = 1 n L r e g ( i ) ,
with
L r e g ( i ) = { 0.5 a 2 if | a | < 1 | a | 0.5 if o t h e r w i s e ,
where a = t i t i * , and t i = { t x , t y , t w , t h } , t i * = { t x * , t y * , t w * , t h * } . Here ( x , y ) denotes the center coordinates of the bounding box, w / h denotes the width / height, and t i ( t i * ) represents the offset of the prediction box ( the ground truth box).
Now, L r e g ( i ) represents the regression loss for the i-th image, and L r e g represents the total regression loss for all images. The total loss function is then given by
L t o l = L c l s ( p i t ) + L r e g + λ L d e r ,
where λ [ 0 , 1 ] is a hyperparameter controlling the strength of the image deraining module’s adjustment to the rainy weather target detection performance. In this context, for λ = 0.5 , m A P of the proposed network detection is optimal from many ablation experiments. See Section 4.3 for details.

4. Experiments

4.1. Experimental Setup

Implementation Details. J A D A R is trained end-to-end on the R a i n V e h i c l e C o l o r - 24 image set using the A d a m optimizer [44] to simultaneously learn image deraining, color classification, and object localization on the P y T o r c h platform. All experiments are implemented on the A u t o D L platform with a T e s l a P 40 . The hyper-parameters α and γ of the classification loss function L c l s are set to 0.25 and 2, respectively. We divide R a i n V e h i c l e C o l o r - 24 into a training set, a validation set and a testing set at a ratio of 8:1:1. The batch size is 4, the epoch is 100, and the confidence threshold is 0.5 . The learning rate is 10 4 for the first 50 epochs, 10 5 for the next 30 epochs, and 10 6 for the last 20 epochs.
Evaluation Metric. Generally, object detection uses I o U (Intersection over Union) [21], P r e c i s i o n (accuracy) [45], R e c a l l [45], A P (Average Precision) [18], m A P (mean Average Precision) [41], or other evaluation metrics; these concepts are well-known, so we list the formulas in brief:
I o U = A B A B
where A / B denotes G T (bounding box of the object) / the prediction bounding box.
Mathematical definitions of P r e c i s i o n and R e c a l l are as follows:
p r e c i s i o n = T P T P + F P
r e c a l l = T P T P + F N
where T P is true positives (correctly predicted as positive), F P is false positives (incorrectly predicted as positive), and F N is false negatives (failed to predict a positive).
A P is calculated by
A P = 0 1 p ( r ) d r
where p is Precision, and r is Recall.
The m A P (mean Average Precision) is the average of A P , so m A P is calculated by
m A P = 1 N A P N
where N is the number of categories.

4.2. Datasets

4.2.1. Synthetic Dataset Rain Vehicle Color-24

Few datasets are available for vehicle color recognition under rainy weather conditions. All our experiments are conducted on enhanced R a i n V e h i c l e C o l o r - 24 [46], from which some examples are illustrated in Figure 3.

4.2.2. Real Rain Vehicle Datasets: R I D and R I S

Li et al. collected two real rainy image vehicle datasets, R I D and R I S [38], for testing object detection. R I D is rainy images collected from in-vehicle cameras while driving on rainy days, and R I S is surveillance rainy images collected from network traffic surveillance cameras during rainy weather conditions. The two datasets differ in many aspects: rainfall type, image quality, target size and angle, etc. They represent real-world application scenarios where deraining may be required. R I D includes 2495 images, and its rainy image effect is closest to “raindrops” on the camera lens. R I S includes 2048 images, and its rainy image effect is closest to “rain and fog” (many cameras have fog condensation when it rains, and lower resolutions also cause more fog effects) [47]. Due to the highly complex scenes of these two rainy image datasets, it is a challenging dataset, and we choose these two datasets for testing to better illustrate the effectiveness of our proposed algorithm. Examples of these two datasets are given in Figure 4.

4.3. Ablation Study

To determine the optimal design of our proposed framework, we train four combinations on the R a i n V e h i c l e C o l o r - 24 dataset: R e t i n a N e t , J A D A R 1 , J A D A R 2 , and J A D A R . All these models are trained and tested on R a i n V e h i c l e C o l o r - 24 using different loss functions: λ = 0, 0.1, 1.0, and 0.5, respectively. Figure 5 shows that the testing m A P values of the J A D A R 1 , J A D A R 2 , and J A D A R models are 2.92%, −3.99%, and 4.3% higher, respectively, than the R e t i n a N e t model, which clarifies that joint semantic feature extraction is beneficial to improve vehicle color recognition performance under rainy weather conditions. Referring to Table 1, when the hyper-parameter λ is 0.1 , the rain removal module provides a weak assisting effect on vehicle color recognition under rainy weather conditions. When λ is 1.0 , it plays the opposite effect. When λ is 0.5 , J A D A R performs best; so we choose this value in our method.

4.4. Experiments and Analysis

4.4.1. Results on Synthetic Datasets

In this section, our proposed algorithm, the vehicle color recognition method, the target detection method, the two-stage method combining rain removal with target detection, and the transfer learning method are compared.
To discuss vehicle color recognition performance, J A D A R and S M N N - M S F F [18] are compared. Both are trained on R a i n V e h i c l e C o l o r - 24 training subset and tested on its test subset. The quantitative results are shown in the second column of Table 2. These quantitative results confirm that the m A P of our method reaches 72.07 % , which is 23.49 % higher than S M N N - M S F F . The qualitative results are shown in Figure 6. J A D A R outperforms S M N N - M S F F under rainy conditions; for example, there are five vehicles recognized by J A D A R , while three vehicles are recognized by S M N N - M S F F . A white vehicle is recognized by J A D A R with a confidence score of 0.79 , while S M N N - M S F F recognizes it with a confidence score of 0.62 .
To compare object detection performance, J A D A R , R e t i n a N e t [24], F a s t e r R C N N [19], S S D [20], and Y O L O V 3 [21] are compared qualitatively and evaluated by m A P quantitatively. In our experiments, the loss function and settings (i.e., scale, anchor or default box, backbone network, classifier, etc.) of each compared method remains unchanged from the original work. Furthermore, all methods are trained on the R a i n V e h i c l e C o l o r - 24 dataset and tested on its test set. The qualitative results of J A D A R , F a s t e r R C N N , Y O L O V 3 , S S D , and R e t i n a N e t for vehicle color recognition in rain are shown in Figure 7 and Figure 8. As can be seen from the figures, our proposed J A D A R outperforms other models for fine vehicle color recognition. The quantitative results show that the proposed J A D A R is 11.42 % , 22.19 % , 5.74 % , and 4.3 % better than F a s t e r R C N N , Y O L O V 3 , S S D and R e t i n a N e t , respectively, from Table 2.
To compare recognition performances of different joint methods, three state-of-the-art rain removal methods, i.e., L P N e t [35], P R e N e t [48], and R C D N e t [49]), are chosen to first derain the images, and then R e t i n a N e t is leveraged to recognize vehicle colors. These methods are denoted L R , P R , and R R . Figure 9 and Figure 10 give qualitative comparisons of our J A D A R and three two-stage methods for vehicle color recognition under rainy weather conditions. J A D A R performs better than other models. From Table 3, our J A D A R is 15.56 % , 20.37 % , and 2.06 % higher than L R , P R , and R R , respectively.
To compare with transfer learning methods, two domain-adaptation methods, D a - f a s t e r [33] and A T F [50], are compared with J A D A R . Here, the V C - 24 is the source domain, and R a i n V e h i c l e C o l o r - 24 is the target domain; they are leveraged to train the above algorithms. From the 5 - th and 6 - th columns of Table 3, our method is 25.95% and 9.14% better than D a - f a s t e r and A T F , respectively. The qualitative results in Figure 9 and Figure 10 show that JADAR identifies more vehicles with higher confidence than the other two methods.

4.4.2. Results on Real Datasets

We train J A D A R , R e t i n a N e t , F a s t e r R C N N , S S D , Y O L O V 3 , L R , P R , R R , D a - f a s t e r , and A T F on R a i n V e h i c l e C o l o r - 24 and test them on real rainy image vehicle datasets, R I D and R I S . The test results are shown in Figure 11, Figure 12, Figure 13 and Figure 14. As can be seen from these figures, the test results of J A D A R on the real datasets, R I D and R I S , are generally better than those of other methods. As can be seen from Figure 11, the J A D A R and S S D algorithms can correctly identify the two cars in the picture; Y o l o V 3 can also identify the two cars in the picture, but the black color is mistakenly identified as silver-gray; while the other three algorithms can hardly identify any vehicles in the picture. Referring to Figure 12, because the recognition effects of F a s t e r R C N N and S S D are better than others’, we find a limitation of J A D A R in recognizing small targets. Referring to Figure 13, all algorithms can identify the color of the vehicle in the image but with different confidence values; specifically, A T F has the highest confidence value for blue vehicle, with 0.98 . However, Figure 14 shows that only J A D A R and A T F can identify a certain white vehicle.

4.5. Inference Time

In order to compare the test time of all methods, all network models are tested on a testing subset with an input of 1920 × 1080 images. The test times are shown in Table 4. JADAR takes 1.7 s per image on a single Tesla P 40 GPU, which is the same as for R e t i n a N e t , but JADAR is 21.8, 1.1, 4.4, 0.8, and 0.9 seconds faster than L R , P R and R R , D a - f a s t e r , and A T F , respectively. Therefore, although J A D A R has one more decoder module than R e t i n a N e t , it still maintains its original high detection speed.

5. Conclusions

In this paper, we study vehicle color recognition under rainy conditions and propose a joint semantics learning method J A D A R , which is designed by embedding U N e t - 3 into R e t i n a N e t . The U N e t module achieves rainy image removal and restores the clean background image. The recovered background image and the rainy image are input together into the c l a s s + b b o x sub-module of R e t i n a N e t network to accurately extract the joint semantic of the vehicle color features maps. J A D A R outperforms other methods under rainy as well as normal conditions for fine vehicle color recognition. Extensive experimental results show that the m A P of the proposed method reaches 72.07 % in identifying 24 colors under rainy conditions. Because our algorithm is trained on synthetic datasets, its generalization is not guaranteed. In the future, semi-supervised or few-shot learning is planned to further improve the generalization and realizability of the algorithm. As a further research topic, one can consider fusing overlap functions and fuzzy (rough) sets (see [51,52,53,54,55]) to develop the method of this paper.

Author Contributions

Writing—original draft preparation, M.H.; Experiments and editing, Y.W.; review, J.F., writing—review, B.J. All authors have read and agreed to the published version of the manuscript.

Funding

This study was funded by the National Natural Science Foundation of China (no. 62071378), the Shaanxi Province International Science and Technology Cooperation Program (no. 2022KW-04), and the Xi’an Science and Technology Plan Project (no. 21XJZZ0072).

Data Availability Statement

The data that support the findings of this study are openly available at [email protected].

Conflicts of Interest

The authors declare that they have no conflict of interest.

References

  1. Tang, J.; Zeng, J. Spatiotemporal gated graph attention network for urban traffic flow prediction based on license plate recognition data. Comput.-Aided Civ. Infrastruct. Eng. 2022, 37, 3–23. [Google Scholar] [CrossRef]
  2. Rajavel, R.; Ravichandran, S.K.; Harimoorthy, K.; Nagappan, P.; Gobichettipalayam, K.R. IoT-based smart healthcare video surveillance system using edge computing. J. Ambient. Intell. Humaniz. Comput. 2022, 13, 3195–3207. [Google Scholar] [CrossRef]
  3. Wang, Z.; Zhan, J.; Duan, C.; Guan, X.; Lu, P. A review of vehicle detection techniques for intelligent vehicles. IEEE Trans. Neural Netw. Learn. Syst. 2022, 23, 1–21. [Google Scholar] [CrossRef]
  4. Wang, J.; Gao, J.; Zhao, S.; Zhu, R.; Jiang, Z. From Model to Algorithms: Distributed Magnetic Sensor System for Vehicle Tracking. IEEE Trans. Ind. Inform. 2022, 33, 1. [Google Scholar] [CrossRef]
  5. Lee, S.; Park, S.H. Concept drift modeling for robust autonomous vehicle control systems in time-varying traffic environments. Expert Syst. Appl. 2022, 190, 116–206. [Google Scholar] [CrossRef]
  6. Khan, M.A.; Sayed, H.E.; Malik, S.; Zia, T.; Khan, J. Level-5 Autonomous Driving—Are We There Yet? A Review of Research Literature. ACM Comput. Surv. 2022, 55, 1–38. [Google Scholar] [CrossRef]
  7. Custers, B. AI in Criminal Law: An Overview of AI Applications in Substantive and Procedural Criminal Law. In Law and Artificial Intelligence; Asser Press: The Hague, The Netherlands, 2022; pp. 205–223. [Google Scholar]
  8. Chen, P.; Bai, X.; Liu, W. Vehicle color recognition on urban road by feature context. IEEE Trans. Intell. Transp. Syst. 2014, 15, 2340–2346. [Google Scholar] [CrossRef]
  9. Jeong, Y.; Park, K.H.; Park, D. Homogeneity patch search method for voting-based efficient vehicle color classification using front-of-vehicle image. Multimed. Tools Appl. 2019, 78, 28633–28648. [Google Scholar] [CrossRef]
  10. Tilakaratna, D.S.; Watchareeruetai, U.; Siddhichai, S.; Natcharapinchai, N. Image analysis algorithms for vehicle color recognition. In Proceedings of the 2017 International Electrical Engineering Congress, Pattaya, Thailand, 8–10 March 2017; pp. 1–4. [Google Scholar]
  11. Dule, E.; Gokmen, M.; Beratoglu, M.S. A convenient feature vector construction for vehicle color recognition. In Proceedings of the 11th WSEAS International Conference on Nural Networks and 11th WSEAS International Conference on Evolutionary Computing and 11th WSEAS International Conference on Fuzzy Systems, Iasi Romania, 13–15 June 2010; pp. 250–255. [Google Scholar]
  12. Hu, C.; Bai, X.; Qi, L.; Chen, P.; Xue, G. Vehicle color recognition with spatial pyramid deep learning. IEEE Trans. Intell. Transp. Syst. 2015, 16, 2925–2934. [Google Scholar] [CrossRef]
  13. Rachmadi, R.F.; Purnama, I.K. Vehicle color recognition using convolutional neural network. In Proceedings of the IEEE International Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–10 June 2015. [Google Scholar]
  14. Zhuo, L.; Zhang, Q.; Li, J.; Zhang, J.; Li, X. High-accuracy vehicle color recognition using hierarchical fine-tuning strategy for urban surveillance videos. J. Electron. Imaging 2018, 27, 1–9. [Google Scholar] [CrossRef]
  15. Zhang, Q.; Zhuo, L.; Li, J.; Zhang, J.; Zhang, H. Vehicle color recognition using Multiple-Layer Feature Representations of lightweight convolutional neural network. Signal Process. 2018, 147, 146–153. [Google Scholar] [CrossRef]
  16. Fu, H.; Ma, H.; Wang, G.; Zhang, X.; Zhang, Y. Mcff-cnn: Multiscale comprehensive feature fusion convolutional neural network for vehicle color recognition based on residual learning. Neurocomputing 2020, 395, 178–187. [Google Scholar] [CrossRef]
  17. Nafzi, M.; Brauckmann, M.; Glasmachers, T. Vehicle shape and color classification using convolutional neural network. In Proceedings of the IEEE International Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–21 June 2019. [Google Scholar]
  18. Hu, M.; Bai, L.; Li, Y.; Zhao, S.R.; Chen, E.H. Vehicle Color Recognition Based on Smooth Modulation Neural Network with Multi-Scale Feature Fusion. arXiv 2021, arXiv:2107.09944. [Google Scholar]
  19. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef]
  20. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S. Ssd: Single shot multi-box detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 21–37. [Google Scholar]
  21. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  22. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y. Yolov4: Optimal speed and accuracy of object detection. In Proceedings of the 2020 IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  23. Tan, M.; Pang, R.; Le, Q.V. Efficientdet: Scalable and efficient object detection. In Proceedings of the 2020 IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10781–10790. [Google Scholar]
  24. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2980–2988. [Google Scholar]
  25. Xu, M.; Wang, H.; Ni, B.; Tian, Q.; Zhang, W. Cross-domain detection via graph-induced prototype alignment. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 12355–12364. [Google Scholar]
  26. Zhang, Y.; Wang, Z.; Mao, Y. Rpn prototype alignment for domain adaptive object detector. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 12425–12434. [Google Scholar]
  27. Kim, T.; Jeong, M.; Kim, S.; Choi, S.; Kim, C. Diversify and match: A domain adaptive representation learning paradigm for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 12456–12465. [Google Scholar]
  28. Dong, H.; Pan, J.; Xiang, L.; Hu, Z.; Zhang, X.; Wang, F.; Yang, M.H. Multi-scale boosted dehazing network with dense feature fusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2157–2167. [Google Scholar]
  29. Xu, C.D.; Zhao, X.R.; Jin, X.; Wei, X.S. Exploring categorical regularization for domain adaptive object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11724–11733. [Google Scholar]
  30. Vs, V.; Gupta, V.; Oza, P.; Sindagi, V.A.; Patel, V.M. Mega-cda: Memory guided attention for category-aware unsupervised domain adaptive object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 4516–4526. [Google Scholar]
  31. Sindagi, V.A.; Oza, P.; Yasarla, R.; Patel, V.M. Prior-based domain adaptive object detection for hazy and rainy conditions. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 763–780. [Google Scholar]
  32. Wang, T.; Zhang, X.; Yuan, L.; Feng, J. Few-shot adaptive faster r-cnn. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7173–7182. [Google Scholar]
  33. Chen, Y.; Li, W.; Sakaridis, C.; Dai, D.; Van, G.L. Domain adaptive faster rcnn for object detection in the wild. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 3339–3348. [Google Scholar]
  34. Liu, W.; Ren, G.; Yu, R.; Guo, S.; Zhu, J. Image-adaptive yolo for object detection in adverse weather conditions. In Proceedings of the AAAI Conference on Artificial Intelligence, Arlington, VA, USA, 17–19 November 2022; pp. 1792–1800. [Google Scholar]
  35. Fu, X.; Liang, B.; Huang, Y.; Ding, X.; Paisley, J. Lightweight pyramid networks for image deraining. IEEE Trans. Neural Netw. Learn. Syst. 2019, 31, 1794–1807. [Google Scholar] [CrossRef]
  36. Fan, Z.; Wu, H.; Fu, X.; Hunag, Y.; Ding, X. Residual-guide feature fusion network for single image deraining. arXiv 2018, arXiv:1804.07493. [Google Scholar]
  37. Jiang, K.; Wang, Z.; Yi, P.; Chen, C.; Huang, B. Multi-scale progressive fusion network for single image deraining. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8346–8355. [Google Scholar]
  38. Li, S.Y.; Araujo, I.B.; Ren, W.Q.; Wang, Z.Y.; Tokuda, E.K. Single image deraining: A comprehensive benchmark analysis. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3833–3842. [Google Scholar]
  39. Tremblay, M.; Halder, S.S.; De Charette, R.; Lalonde, J.F. Rain rendering for evaluating and improving robustness to bad weather. Int. J. Comput. Vis. 2021, 129, 341–360. [Google Scholar] [CrossRef]
  40. Zhang, H.; Sindagi, V.; Patel, V.M. Image de-raining using a conditional generative adversarial network. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 3943–3956. [Google Scholar] [CrossRef]
  41. Huang, S.C.; Le, T.H.; Jaw, D.W. DSNet: Joint semantic learning for object detection in inclement weather conditions. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 2623–2633. [Google Scholar] [CrossRef]
  42. Hu, M.; Yang, J.; Lin, N.; Liu, Y.; Fan, J. Lightweight single image deraining algorithm incorporating visual saliency. IET Image Process. 2022, 16, 3190–3200. [Google Scholar] [CrossRef]
  43. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  44. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  45. Jose, H.; Vadivukarasi, T.; Devakumar, J. Extraction of protein interaction data: A comparative analysis of methods in use. EURASIP J. Bioinform. Syst. Biol. 2007, 43, 53096. [Google Scholar] [CrossRef] [PubMed]
  46. Hu, M.; Wang, C.; Yang, J.; Wu, Y.; Fan, J.; Jing, B. Rain Rendering and Construction of Rain Vehicle Color-24 Dataset. Mathematics 2022, 10, 3210. [Google Scholar] [CrossRef]
  47. Hu, M.; Wu, Y.; Song, Y.; Yang, J.B.; Zhang, R.F. The integrated evaluation and review of single image rain removal based datasets and deep methods. J. Image Graph. 2022, 27, 1359–1391. [Google Scholar]
  48. Ren, D.; Zuo, W.; Hu, Q.; Zhu, P.; Meng, D.Y. Progressive image deraining networks: A better and simpler baseline. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3937–3946. [Google Scholar]
  49. Wang, H.; Xie, Q.; Zhao, Q.; Meng, D.Y. A model-driven deep neural network for single image rain removal. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 3100–3109. [Google Scholar]
  50. He, Z.; Zhang, L. Domain adaptive object detection via asymmetric tri-way faster-rcnn. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Cham, Switzerland, 2020; pp. 309–324. [Google Scholar]
  51. Liang, R.; Zhang, X. Interval-valued pseudo overlap functions and application. Axioms 2022, 11, 216. [Google Scholar] [CrossRef]
  52. Wang, J.; Zhang, X. A novel multi-criteria decision-making method based on rough sets and fuzzy measures. Axioms 2022, 11, 275. [Google Scholar] [CrossRef]
  53. Zhang, X.; Wang, J.; Zhan, J.; Dai, J. Fuzzy measures and Choquet integrals based on fuzzy covering rough sets. IEEE Trans. Fuzzy Syst. 2021, 30, 2360–2374. [Google Scholar] [CrossRef]
  54. Sheng, N.; Zhang, X. Regular partial residuated lattices and their filters. Mathematics 2022, 10, 2429. [Google Scholar] [CrossRef]
  55. Wang, J.; Zhang, X.; Hu, Q. Three-way fuzzy sets and their applications (II). Axioms 2022. under review of the second version. [Google Scholar]
Figure 1. Framework overview for our J A D A R ; detailed explanations are in the text.
Figure 1. Framework overview for our J A D A R ; detailed explanations are in the text.
Mathematics 10 03512 g001
Figure 2. Architecture and weights of the proposed network in detail.
Figure 2. Architecture and weights of the proposed network in detail.
Mathematics 10 03512 g002
Figure 3. Examples from R a i n V e h i c l e C o l o r - 24 .
Figure 3. Examples from R a i n V e h i c l e C o l o r - 24 .
Mathematics 10 03512 g003
Figure 4. Examples of RID and RIS images [38].
Figure 4. Examples of RID and RIS images [38].
Mathematics 10 03512 g004
Figure 5. A P s of different models on the Rain Vehicle Color-24 test set. The x-axis represents the average precision, and the y - axis represents the color categories (24 categories in total).
Figure 5. A P s of different models on the Rain Vehicle Color-24 test set. The x-axis represents the average precision, and the y - axis represents the color categories (24 categories in total).
Mathematics 10 03512 g005
Figure 6. Test results of J A D A R and S M N N - M S F F on the Rain Vehicle Color-24 test set. Each subtitle gives object detection method with the corresponding confidence value in parentheses.
Figure 6. Test results of J A D A R and S M N N - M S F F on the Rain Vehicle Color-24 test set. Each subtitle gives object detection method with the corresponding confidence value in parentheses.
Mathematics 10 03512 g006
Figure 7. Example 1 of test results of JADAR and object detection methods on the Rain Vehicle Color-24 test set.
Figure 7. Example 1 of test results of JADAR and object detection methods on the Rain Vehicle Color-24 test set.
Mathematics 10 03512 g007
Figure 8. Example 2 of test results of JADAR and object detection methods and domain adaptation methods on the Rain Vehicle Color-24 test set.
Figure 8. Example 2 of test results of JADAR and object detection methods and domain adaptation methods on the Rain Vehicle Color-24 test set.
Mathematics 10 03512 g008
Figure 9. Example 1 of test results of JADAR and two-stage methods on the Rain Vehicle Color-24 test set.
Figure 9. Example 1 of test results of JADAR and two-stage methods on the Rain Vehicle Color-24 test set.
Mathematics 10 03512 g009
Figure 10. Example 2 of test results of JADAR and two-stage methods and domain-adaptation methods on the Rain Vehicle Color-24 test set.
Figure 10. Example 2 of test results of JADAR and two-stage methods and domain-adaptation methods on the Rain Vehicle Color-24 test set.
Mathematics 10 03512 g010
Figure 11. Test results of JADAR and color recognition and object detection methods on the RID.
Figure 11. Test results of JADAR and color recognition and object detection methods on the RID.
Mathematics 10 03512 g011
Figure 12. Test results of JADAR and color recognition and object detection methods on the RIS.
Figure 12. Test results of JADAR and color recognition and object detection methods on the RIS.
Mathematics 10 03512 g012
Figure 13. Test results of JADAR and two-stage and domain-adaptation methods on the RID.
Figure 13. Test results of JADAR and two-stage and domain-adaptation methods on the RID.
Mathematics 10 03512 g013
Figure 14. Test results of J A D A R and two-stage and domain-adaptation methods on the R I S .
Figure 14. Test results of J A D A R and two-stage and domain-adaptation methods on the R I S .
Mathematics 10 03512 g014
Table 1. The mAPs using different values for the loss function coefficient λ of the deraining module in J A D A R on the R V C - 24 test set.
Table 1. The mAPs using different values for the loss function coefficient λ of the deraining module in J A D A R on the R V C - 24 test set.
λ ModelmAP
0.1 J A D A R 1 70.69%
0.5 J A D A R 72.07%
1.0 J A D A R 2 63.78%
Table 2. Comparison of recognition accuracy of 24 colors for different network classifications: SM, SMNN-MSFF; FR, Faster RCNN; Yolo, Yolo V3; RN, RetinaNet.
Table 2. Comparison of recognition accuracy of 24 colors for different network classifications: SM, SMNN-MSFF; FR, Faster RCNN; Yolo, Yolo V3; RN, RetinaNet.
Category (R, G, B)SMFRYoloSSDRNJADAR
white (255, 255, 255)0.600.960.950.950.970.97
black (0, 0, 0)0.690.690.920.930.930.95
orange (237, 145, 33)0.770.900.930.950.970.94
silver-gray (128, 138, 135)0.310.810.860.860.860.85
grass-green (0, 255, 0)0.820.930.950.960.960.96
dark-gray (128, 128, 105)0.430.690.630.670.640.73
dark-red (156, 102, 31)0.480.790.750.880.910.91
gray (192, 192, 192)0.310.410.150.310.270.36
red (255, 0, 0)0.440.560.600.760.770.83
cyan (0, 255, 255)0.620.820.810.870.870.93
champagne (255, 227, 132)0.280.740.580.730.660.69
dark-blue (25, 25, 112)0.520.790.770.750.860.81
blue (0, 0, 255)0.560.690.550.690.670.69
dark-brown (94, 38, 18)0.350.180.400.470.370.50
brown (128, 128, 42)0.350.100.220.340.300.21
yellow (255, 255, 0)0.350.830.740.920.860.81
lemon-yellow (255, 215, 0)0.570.990.730.750.750.94
dark-orange (210, 105, 30)0.320.280.440.470.900.53
dark-green (48, 128, 20)0.590.180.000.340.440.55
red-orange (255, 97, 0)0.380.070.000.520.000.05
earthy-yellow (184, 134, 11)0.680.450.000.280.931.00
green (0, 255, 0)0.180.850.000.970.600.33
pink (255, 192, 203)0.840.540.000.520.770.75
purple (160, 32, 240)0.220.030.000.060.001.00
mAP48.58%60.65%49.88%66.33%67.77%72.07%
Table 3. Comparison of recognition accuracy of 24 colors for different network classifications.
Table 3. Comparison of recognition accuracy of 24 colors for different network classifications.
Category (R, G, B)LRPRRRDa-FasterATFJADAR
white (255, 255, 255)0.880.860.950.740.930.97
black (0, 0, 0)0.750.750.920.610.870.95
orange (237, 145, 33)0.910.920.960.730.900.94
silver-gray (128, 138, 135)0.710.660.820.630.630.85
grass-green (0, 255, 0)0.840.880.950.730.920.96
dark-gray (128, 128, 105)0.550.440.730.380.650.73
dark-red (156, 102, 31)0.670.750.870.540.660.91
gray (192, 192, 192)0.270.200.370.190.270.36
red (255, 0, 0)0.740.700.870.460.630.83
cyan (0, 255, 255)0.810.780.920.750.930.93
champagne (255, 227, 132)0.420.430.620.180.670.69
dark-blue (25, 25, 112)0.580.300.670.160.890.81
blue (0, 0, 255)0.330.480.530.440.920.69
dark-brown (94, 38, 18)0.440.320.560.370.750.50
brown (128, 128, 42)0.330.250.450.060.000.21
yellow (255, 255, 0)0.790.770.890.391.000.81
lemon-yellow(255, 215, 0)0.770.780.970.060.000.94
dark-orange (210, 105, 30)0.560.180.901.000.000.53
dark-green (48, 128, 20)0.650.500.950.220.070.55
red-orange (255, 97, 0)0.000.000.000.391.000.05
earthy-yellow (184, 134, 11)0.050.670.250.010.401.00
green (0, 255, 0)0.060.130.001.001.000.33
pink (255, 192, 203)0.440.670.670.920.930.75
purple (160, 32,240)1.000.001.000.000.001.00
mAP56.51%51.70%70.01%46.12%62.93%72.07%
Table 4. Comparison of different network recognition speeds (GPU).
Table 4. Comparison of different network recognition speeds (GPU).
AlgorithmRNLRPRRRDa-FasterATFJADAR
Speed (s)1.723.52.86.12.52.61.7
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hu, M.; Wu, Y.; Fan, J.; Jing, B. Joint Semantic Intelligent Detection of Vehicle Color under Rainy Conditions. Mathematics 2022, 10, 3512. https://doi.org/10.3390/math10193512

AMA Style

Hu M, Wu Y, Fan J, Jing B. Joint Semantic Intelligent Detection of Vehicle Color under Rainy Conditions. Mathematics. 2022; 10(19):3512. https://doi.org/10.3390/math10193512

Chicago/Turabian Style

Hu, Mingdi, Yi Wu, Jiulun Fan, and Bingyi Jing. 2022. "Joint Semantic Intelligent Detection of Vehicle Color under Rainy Conditions" Mathematics 10, no. 19: 3512. https://doi.org/10.3390/math10193512

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop