Next Article in Journal
A Validation Study of Freezing of Gait (FoG) Detection and Machine-Learning-Based FoG Prediction Using Estimated Gait Characteristics with a Wearable Accelerometer
Next Article in Special Issue
A Bimodal Model to Estimate Dynamic Metropolitan Population by Mobile Phone Data
Previous Article in Journal
Birefringent Bragg Grating in C-Shaped Optical Fiber as a Temperature-Insensitive Refractometer
Previous Article in Special Issue
Reconstructed Order Analysis-Based Vibration Monitoring under Variable Rotation Speed by Using Multiple Blade Tip-Timing Sensors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Domain Adaptation and Adaptive Information Fusion for Object Detection on Foggy Days

1
College of Computer and Information, Hohai University, Nanjing 210098, China
2
Jiangsu Collaborative Innovation Center for Cultural Creativity, Changzhou 213000, China
3
School and Computer Information and Engineering, Changzhou Institute of Technology, Changzhou 213022, China
4
College of Information Engineering, Nanjing Xiaozhuang University, Nanjing 210017, China
*
Authors to whom correspondence should be addressed.
Sensors 2018, 18(10), 3286; https://doi.org/10.3390/s18103286
Submission received: 13 July 2018 / Revised: 20 September 2018 / Accepted: 28 September 2018 / Published: 30 September 2018
(This article belongs to the Special Issue Multi-Sensor Fusion and Data Analysis)

Abstract

:
Foggy days pose many difficulties for outdoor camera surveillance systems. On foggy days, the optical attenuation and scattering effects of the medium significantly distort and degenerate the scene radiation, making it noisy and indistinguishable. Aiming to solve this problem, in this paper we propose a novel object detection method that has the ability to exploit the information in the color and depth domains. To prevent the error propagation problem, we clean the depth information before the training process and remove false samples from the database. A domain adaptation strategy is employed to adaptively fuse the decisions obtained in the color and depth domains. In the experiments, we evaluate the contribution of the depth information for object detection on foggy days. Moreover, the advantages of the multiple-domain adaptation strategy are experimentally demonstrated via comparison with other methods.

1. Introduction

Outdoor camera surveillance systems are widely used in urban areas, and play an important role in traffic management [1] and security maintenance [2]. It is necessary for these systems to operate in all weather conditions. However, foggy days pose many difficulties for vision-based systems. The attenuated scene appearance and strong noise are the two main factors that degenerate object detection results [3,4].
Many efforts have been made to acquire clear images/videos on foggy days, and excellent results have been obtained [5,6]. However, the state-of-the-art image enhancement methods cannot significantly improve the object detection performance. The reasons are twofold. First, object detection entails segmenting objects of interest from the background. Hence, the key problem for object detection on foggy days is how to identify the deviation between the object and background. This is difficult for these enhanced images, which include many textures. Second, error propagation prevents the use of any image preprocessors. The initial errors with image preprocessors will propagate to the subsequent detection process, causing errors in the final object detection results. Hence, the preprocessor-based object detection strategy is questionable in some cases.
Despite its disadvantages, the haze effect on foggy days provides a novel cue for object detection. According to the optical imaging model, the haze concentration changes with depth [7,8]. Hence, we can present the unscaled depth through haze concentration estimation, and the depth contrast between the object and background can be presented according to the point-to-point difference in haze concentration. In addition to the RGB information in the color domain, this information provides a novel feature for object detection. For haze concentration estimation on foggy days, the most efficient method is the dark channel prior model [9]. The advantage of the dark channel prior model is that it can estimate the haze concentration using monocular images. However, its drawback is that it is quite sensitive to image noise, and therefore image outliers cause serious errors in haze concentration estimation results. To solve this problem, a novel data cleaning method is used here to filter the depth data. This can ensure the correctness of the background model, but leads to inequality in the amount of data between the depth and color domains. This problem is solved by employing the domain adaptation learning strategy. Two detectors are trained separately with the color and depth information, and the final domain-adapted detection is performed by combining these two detectors. The novelty of our method is threefold:
(i)
Depth-information-based object detection on foggy days. Aiming to conquer the challenges posed by foggy days, our method exploits the depth information for object detection.
(ii)
Domain-adaptation-learning-based background modeling on foggy days. Our method trains the background models with the color and depth information separately, and they are jointly trained via the domain adaptation learning strategy.
(iii)
Exploring depth and color features in images on foggy days. Our method explores the features in both the color and depth domains, and fuses them for object detection on foggy days.
The paper is structured as follows. In Section 2, we present state-of-the-art research for image processing and object detection on foggy days. Section 3 introduces our proposed method. The experimental results are presented in Section 4, and our conclusions are presented in Section 5.

2. Related Works

Most works related to object detection on foggy days involve a combination of image dehazing and object detection methods. The former is commonly used as a preprocessor for enhancing the object appearance, whereas the object–background transition is identified by the object detection postprocessor.

2.1. Image Processing

Aiming to remove the haze effect of images on foggy days, various image processing methods have been proposed. Generally, image processing on foggy days is achieved by transforming the atmospheric scattering model [10,11,12,13], which can be simplified as follows:
E ( d , λ ) = E d t ( d , λ ) + E a ( d , λ ) ,
where E ( d , λ ) is the acquired image, E d t ( d , λ ) is the term originating from the object radiation, E a ( d , λ ) is the haze term, and d and λ are the transmission distance and wavelength of the light, respectively. In this model, the key issue is the transmission distance estimation. Narasimhan et al. found that the point-to-point contrast in haze environments is related to the depth of the points. This principle is used to remove the haze and restore the original scene radiation [11]. Results obtained using this method can provide a visually desirable presentation. Schechner et al. [14] proposed a depth estimation method based on the fact that the scene depth varies with the degree of optical polarization. This method has a robust physical basis, but requires complicated polarization imaging devices. Liang et al. found that the light polarization state changes with the scene depth, and the scene contrast can be enhanced by altering the angle of light polarization [15,16]. There are also many scene depth and ambient light estimations using high-level image features. Kopf et al. introduced a three-dimensional georeferenced terrain model to estimate the real scene depth [17]. In this method, multi-source information such as the depth, texture, and geographic information were jointly used for image relighting and dehazing. Nishino et al. proposed a Bayesian probabilistic method to estimate the scene albedo and depth from a single foggy image [18]. By combining the contextual regularized L1 norm and boundary constraint, Meng proposed a method for optimally estimating the light transmission parameter [19]. Compared with the aforementioned methods, a more efficient strategy is achieved by the dark channel prior model. According to the dark channel prior, in clear images, an extremely low value of some pixels exists on at least one color channel. Hence, the dark-channel intensity in foggy images indicates the depth-dependent haze concentration [20]. The dark channel model works well in the case of parallel ambient light, whereas its performance is seriously degenerated when a skylight is included in the image. The blocking effect and flickering artifacts pose problems, as they reduce the accuracy of the depth estimation. In order to solve these problems, Li proposed a Markov random field with an intensity value prior to update the dark channel prior model [21]. Wang et al. used a constant intensity threshold to segment the skylight regions and estimate the ambient light in other regions [22]. Qing et al. proposed a mixture-of-Gaussian (MoG) model to estimate the skylight distribution [23]. Recently, Zhu et al. fused the luminance model with the dark channel prior model to remove the haze in images [24].

2.2. Object Detection

Owing to the appearance degeneration and haze effect on foggy days, the deviation between the object and background is seriously degenerated. To solve this problem, most existing methods—including those mentioned above—depend on a two-phase structure: an image preprocessor followed by a detection postprocessor. The advantages and disadvantages of this strategy are discussed in the Introduction (Section 1). Additionally, there are also methods based on optimal mathematical models. Oreifej proposed a three-term low-rank matrix decomposition method to decompose the image data into three components: the scene background, turbulence of the medium, and object of interest. Then, moving objects are segmented with the L1 norm [25]. Gilles adopted a geometric spatiotemporal viewpoint to solve the atmospheric turbulence problem, and developed a model that distinguishes the movement of moving objects in the case of turbulence [26].

3. Proposed Method

A domain adaptation strategy underlies the novel object detection method proposed in this paper [27,28]. Information in two domains—color and depth—is explored in our method. The depth information is estimated using the dark channel prior model, in which the skylight is initially removed. Moreover, we propose a data cleaning method to eliminate false depth information, ensuring the correctness of the training data.
After the data cleaning process, an inequality effect arises between the two sources. This problem is handled by the domain adaptation framework, and the final object detection results are generated by adaptively combining the results obtained separately using the color and depth information. The framework of our proposed method is shown in Figure 1.

3.1. Depth Estimate and Data Cleaning

Among existing depth estimation methods available for haze environments, the most efficient method is the dark channel prior model. Although this model allows only unscaled measurement, the estimated results can present the contrast between the object and background [22]. Aiming to remove the skylight areas, an optical feature correlation method is proposed here to recognize the light component. Moreover, to eliminate the errors in the training data, the depth information is cleaned according to the inter-frame correspondence.

3.1.1. Skylight Area Recognition and Removal

In the dark channel prior model, skylight areas cause errors in depth estimation. In contrast to the ambient light transmitted through the haze medium, the intensities of all color channels of the skylight are homogeneous and significantly higher than surrounding areas [29,30]. The content of skylight areas is easily mistaken as the representation of ambient light when using the dark channel model, causing serious errors in depth estimation. The skylight can be recognized based on two aspects:
(i)
Low channel variation. In contrast to other optical components, the channel variation is relatively low for the skylight.
(ii)
Distance-dependent intensity. Owing to the light scattering factor in haze environments, in skylight areas, the intensity of any point is related to its distance from the optical collimation.
These two principles are mathematically modeled and combined to recognize skylight areas. For the channel variation, it can be mathematically modeled as follows:
V x = V ( I x c , I x o ) = ( I x r I x o ) 2 + ( I x g I x o ) 2 + ( I x b I x o ) 2 ,
where V ( I x c , I x o ) is the variance of the point x in the RGB color space, I x c is the intensity in channels (red, I x r ; green, I x g ; and blue, I x b ), and I x o is the average value of color channels.
For the distance-dependent intensity effect, the skylight area can be mathematically modeled using the intensity–position relation, which is scaled by an exponential distance from the highest intensity in the whole image, as follows:
x = exp ( D x , m ) = exp ( ( x 1 m 1 ) 2 + ( x 2 m 2 ) 2 ) ,
where D x , m is the Euclidean distance between points x and m , which have the highest intensity in the whole image. x = [ x 1 , x 2 ] and m = [ m 1 , m 2 ] are the spatial coordinates of points x and m . Combining these two principles with the correlation calculation, the discriminative function for the skylight can be modeled as follows:
S = c o r r 2 ( x , V x ) .
The corresponding threshold for S is presented as T:
L S k y l i g h t = { 1 i f S > T 0 o t h e r w i s e ,
where c o r r 2 ( ) is the two-dimensional correlation calculation, and T is the threshold for removing skylight areas. The ambient light estimation and the dark channel calculation are performed in the region without the skylight, where L S k y l i g h t = 0 . Samples for skylight recognition and depth estimation on three foggy days are shown in Figure 2. From these samples, we can see that the correlation between the intensity position and channel variation can correctly describe the distribution of the skylight, as their values all fall into the minimum around the region of the skylight. After removing the skylight, the depth estimation results can present the sight–depth contrast between objects and the background.

3.1.2. Dark Channel Prior Model-Based Depth Estimation

According to the dark channel prior model, in most of the haze-free image, an extremely low intensity value indicates at least one color channel, as follows:
I x d a r k = min y Ω x ( min c { r , g , b } ( I y c ) ) 0 ,
where I y c is the channel of point y in the neighborhood and Ω x is the local patch centered at point x . The intensity in the dark channel is a representation of the depth-dependent haze concentration, which is called the transmission in [20]. Hence, depth-dependent haze concentration/transmission can be expressed as follows:
Γ x = 1 w min y Ω x ( min c { r , g , b } ( I y c A c ) ) ,
where w is a coefficient describing the degree of dehazing to present the depth and A c is the ambient light that corresponds to the largest value of the dark channel over the entire images, as follows:
A c = max x ( I x d a r k = c ) .
As previously mentioned, the transmission Γ x strictly depends on the depth d x at point x ; thus, the point-to-point depth difference can be correctly presented by the point-to-point transmission difference. Identifying the deviation between objects and the background is a desired property for object detection. Hence, in this paper, we present the depth d x with the transmission Γ x , as d x Γ x .

3.1.3. Data Cleaning for Depth Information

Although skylight areas can be removed by the process shown in Section 3.1.1, random errors in the depth estimation—possibly caused by burr points—cannot be avoided. To solve this problem, we propose a data cleaning method for depth maps. In a video sequence, the variations between frames in a short interval are minor, and changes appear only in limited patches, whereas most pixels remain the same, as shown in the first row of Figure 3. This indicates that the correlations between frames in a short time interval are strong, which should be the case for corresponding depth maps—otherwise, random errors occur, as shown in the second row of Figure 3.
A pair of correlations between frames in a short interval and between depth maps can be mathematically calculated as follows:
R t = c o r r 2 ( I t , I t k ) ,
R t = c o r r 2 ( ψ t , ψ t k ) ,
where I t and I t k are the frames in time steps t and t k , and ψ t and ψ t k are the corresponding depth maps. Here, the parameter k is designed such that, for depth map ψ t in time step t , ψ t k is the temporally nearest depth map that is correct in previous time steps. For instance, if ψ t 1 is identified as an error in time step t 1 while ψ t 2 is correct, k = 2 , R t = c o r r 2 ( I t , I t 2 ) , and R t = c o r r 2 ( ψ t , ψ t 2 ) .
Assuming that the depth map for the first frame is correct, an error is recognized under the following condition:
R t < λ R t ,
where λ is the moderation parameter. As previously mentioned, this type of error is caused by random noise, such as burr points. Thus, errors do not continually happen in a long time interval. As a result, in practice, the parameter k is not large (typically 1 k 3 ) and the temporal interval for the correlation calculation is limited, which maintains the correspondence between R t and R t for correct samples. This data cleaning process can be illustrated using samples in Figure 3. In theory, during the ambient light estimation, the dark channel prior model extracts the brightest point in the dark channel to present its neighborhood, and the brightest patch in the whole image is extracted to present the ambient light [20]. This strategy is sensitive to image burrs, as even a single noisy spot can cause errors in the ambient light estimation and depth estimation. This is a common case in real-world practice due to imaging noises. For example, the random noises located around [257, 129] in the fourth frame (first row in Figure 3) are mistaken as the representation of the ambient light due to their large intensity in the dark channel. This causes errors in depth estimation (fourth frame in the second row). As the result, the depth estimation output of the fourth frame in Figure 3 is significantly different from the previous frames, although their inputs appear to be similar. This is the reason for the low value of the R t obtained by the fourth frame. According to our proposed data cleaning principle, the depth estimation result of the fourth frame is eliminated, and the fifth sample is compared with the third one to continue the cleaning process.

3.2. Domain Adaptation Learning and Fusion Method

Taking advantage of the depth estimation method, for any scene on foggy days, two sources for object detection can be separately obtained in the color and depth domains. After data cleaning, the amount of data in these two domains is unequal. This problem is solved by a cross-source domain adaptation method, which can improve the background model with little training data (depth information) by using another model with more training data (color information). This is based on the principle that the changes in the depth scale correspond to the variations in the color information, as the depth deviation likely exists at edges where contrasts occur in the color domain. We employ the kernel density estimation (KDE) to establish the background models in the color and depth domains [31].

3.2.1. KDE Model

The reason for using the KDE to establish the background models lies in its good adaptability to short-term changes of complicated scenes. Theoretically, the KDE is a typical nonparametric model that presents the background by training samples rather than any previous assumption regarding the data distribution.
The classic KDE model can be mathematically formulated by comparing the testing samples with the selected training samples, as follows:
P ( x t | B t ) = 1 N i = 1 N K ( x t x i ) ,
where N samples xi are selected in the training data used as the representation of the background, x t represents the testing samples, Bt is the background model at the time step t, and K(x) is the kernel function, which satisfies the conditions K ( x ) d x = 1 , x K ( x ) d x = 0 , and K(x) > 0.

3.2.2. Color–Depth Cross-Source Domain Adaptation

The amounts of color images and depth maps are represented as κ c and κ d . After data cleaning, κ c κ d . We define the color as the source domain and the depth as the target domain. In these two domains, we calculate the histogram of oriented gradients (HOG) to describe the local change [32]. The HOG feature counts the number of occurrences of gradient orientation in the local regions of an image. This gradient information is useful for object detection because it changes significantly at the transition between objects and the background. For the HOG feature, every pixel has two characteristics: magnitude G and direction θ . These characteristics can be numerically presented using the LL sub-band ( ϕ L L ) of the discrete wavelet transform (DWT).
G = ϕ L L 2 ( x ) + ϕ L L 2 ( y )
and
θ = tan 1 ϕ L L ( y ) ϕ L L ( x ) ,
where ϕ L L ( x ) and ϕ L L ( y ) are the derivatives of the LL sub-band in the x and y directions, respectively. A straightforward method for establishing the object detection feature is combining the HOG features in different domains. However, this idea is likely not appropriate for this study, owing to the inequality between the features in the source domain and the target domain.
Aiming to solve this problem, we introduce a domain adaptation learning strategy based on two parallel streams [27]. This method exploits the ample availability of training data from the color domain to learn a model that works effectively in the depth domain, for which fewer examples are available. Specifically, two separate background models B t c and B t d at time step t are trained for each domain. Let P ( x t c | B t c ) and P ( x t d | B t d ) be the distribution of the color and depth features, respectively. We can see that the distributions of the input features differ in the two domains, that is, P ( B t c ) P ( B t d ) . Note that without the adaptation mechanism, this may lead to a poor detection result in the target domain, as a model in the color domain which includes a larger source training set will be trained to perform well in the dense source regions.
We now present the specific domain adaptation KDE algorithms. One of the simplest possible strategies for domain adaptation consists of a convex combination of two KDEs learned independently from the color and depth domains. Despite its simplicity, this framework has been demonstrated to yield good empirical results. As a result, the final domain-adapted background model B t can be generated by a weighted linear combination of two background models in two different domains, as follows:
P ( x t | B t ) = w c P ( x t c | B t c ) + w d P ( x t d | B t d ) = w c 1 N c j = 1 N c K ( x t c x j c ) + w d 1 N d k = 1 N d K ( x t d x k d )
where the weight parameters w c and w d are determined by minimizing the detection errors on the target (depth) domain; w c + w d = 1 , x t c , and x t d are the testing samples at time step t; x j c and x k d are the background samples learned by the KDE model; and N c and N d are the numbers of background samples.
The parameters w c ∈ [0, 1], w d ∈ [0, 1] are determined via grid search by minimizing multiclass errors on the color training set. We avoid biased estimates resulting from learning the hypothesis Bt, w c , and w d on the same training set by applying a two-stage procedure. First, we learn distinct hypotheses using cross-validation (with the hyperparameter value found for KDE) and compute a prediction at each training sample using the cross-validation hypothesis that is not trained on that example. Second, we use these predicted outputs to determine the optimal weights. Finally, we learn the background model using the entire target training set. In general, our depth–color feature learning and fusion process can be expressed in Algorithm 1.
Algorithm 1. Color–Depth Domain Adaptation Learning
Input: color images x c (source domain), depth maps x d (target domain)
Output: optimal weight parameters w c and w d
Initialization: weight parameters w c = w d = 0.5
  • Extract the HOG features in source and target domains // Equations (13) and (14)
  • Establish the background models separately in the source domain P ( x t c | B t c ) and target domain P ( x t d | B t d ) // Equation (12)
  • Calculate the combined detection result P ( x t | B t ) // Equation (15)
  • Optimize the weight parameters w c and w d by minimizing the detection error on the target (depth) domain

4. Experimental Results

For experimental evaluation of our method, we selected public videos from YouTube that include diverse scenes from foggy days [33,34,35,36]. Fifty video sequences acquired on foggy days were included in this evaluation. For each sequence, we collected only one video slice, maintaining the diversity of the testing data. Hence, 50 diverse video slices with 1257 frames were tested in our experiment. For a video slice, the variations between consecutive frames were extremely low. As a result, the redundancy was high in the training dataset. If the background model was trained with all the frames, the training process would be extremely high in time cost. Aiming to remove the redundancy between training samples, we randomly selected one frame within every five-frame interval. In each of the experiments, we kept the resolution of the input frames as the original resolution of the frames. For a video sequence, we selected 250 frames as training samples to model the background. First, we demonstrate the contribution of the depth information by presenting the object detection results with and without depth information. Then, our method is experimentally compared with the existing background modeling methods, that is, spatiotemporal MoG (ST-MoG) [37], Vibe [38], and DECOLOR [39]. Moreover, the phase spectrum of the quaternion Fourier transform (PQFT) method was also selected as a typical preprocessor-/saliency-based object detection method [40]. For these compared methods, features extracted in the color domain were employed, and depth features were not included. Thus, the performance of the domain adaptation strategy can be fairly demonstrated. The excellent deep learning methods were not included in our experiments because they require a large amount of training data, exceeding that collected in this study. Currently, there is no database that contains enough data acquired on foggy days for training a deep network. If we implemented the deep learning methods with databases similar to that of the KDE models, it would be difficult to achieve the desired object detection results, yielding an unfair evaluation of these deep learning methods. Hence, the methods selected for comparison in this study are relatively low in model complexity and have been proven to be efficient for object detection in diverse scenes. The parameters T and λ were set as T = 0.8 and λ = 0.9 , and the window size of the dark channel prior was 3 × 3 in our experiment.

4.1. Evaluation Criterion

The ground truth in our experimental evaluations was obtained by the average of the labels provided by 10 volunteers. Because the purpose of our method is to detect regions of moving objects, moving objects in our ground truth were identified according to a hypothesis: moving objects are identified if their displacement is larger than 10 pixels in 5 consecutive frames. This can prevent the influence of static objects and dynamic noise. According to the PASCAL criterion [41], C was used to evaluate the overlap of the detection results and the ground truth:
C = Ω Ω Ω Ω ,
where Ω represents the detected results and Ω represents the ground truth. The performance of our method was evaluated with respect to six criteria [42]: precision (Pr), similarity (Sim), true positive rate (TPR), F-score (FS), false positive rate (FPR), and percentage of wrong classifications (PWC), as follows:
Pr = t p t p + f t ,   TPR = t p t p + f n ,   Fs = 2 × Pr × TPR Pr + TPR , Sim = t p t p + f p + f n ,   FPR = f p f p + t n ,   PWC = 100 × f n + f p t p + t n + f p + f n
Here, tp, tn, fp, and fn denote the numbers of true positives, true negatives, false positives, and false negatives, respectively.

4.2. Qualitative Evaluation

The performance in two aspects is demonstrated here. First, the object detection performance using the depth information is presented, from which we can qualitatively understand the motivation for combining the information in the color and depth domains. Second, the performance is qualitatively evaluated via comparison with other methods.
Figure 4 shows maps of the depth, color information, and corresponding object detection results for three scenes on foggy days. We observed a complementary relationship between the depth and color information. Generally, the depth information was more sensitive to nearby objects and had a better capability to remove the background noise. However, it was difficult to detect distant objects using the depth information. In contrast to the depth information, the color information performed better for detecting objects far from the camera. However, there were point-like noises in the results obtained with the color information. As a result, we observed that some of the distant objects in Scenes I and II were missed when the depth information was used. However, the depth information performed better for Scene III, as it identified the pedestrian object that was missed in the results obtained using the color information.
A qualitative performance comparison is shown in Figure 5. These results indicate diverse properties among the different methods. The ST-MoG model had a good ability to identify objects, but the drawback of this method can be demonstrated by the noisy points, as well as by the holes in the results. The Vibe and DECOLOR methods presented a better performance for block-like objects. However, the performance of DECOLOR was relatively degenerated when structural objects were close to the camera, such as the samples in the seventh and last rows. The PQFT-based method could only provide rough results, depicting the object areas rather than their accurate contours. Generally, the best results were provided by the proposed method, as it could correctly detect objects, especially nearby ones. However, our method missed distant objects in some cases (e.g., the results in the third and fourth rows). The reason for this error is that the depth information used in our approach is given by the simple dark channel prior model-based unscaled measurement, which has high efficiency but low resolution. Hence, the objects that were far from the camera were likely mistaken as the background, as the unscaled depth information is not sensitive enough to distinguish movements occurring far from the camera. Morphological schemes such as the erosion and dilation operator [43] can be further introduced here to remove the burrs and noise in the results. However, these methods cannot run automatically, and we should carefully configure the controlling parameters, such as the bandwidth of the kernel, on a case-by-case basis. This may lead to a bias for performance comparisons because we can hardly determine whether these parameters are globally optimal for any special results. This is the reason for maintaining the original object detection results in the experimental evaluations.

4.3. Quantitative Evaluation

Using the aforementioned criteria, we provide a quantitative evaluation of the compared methods and our method. As indicated by Table 1, our method exhibited the best performance for four criteria and the second-best performance for two criteria. The most comparable performance was achieved by the Vibe method, as it was the best for two criteria and provided the second-best results for three criteria. According to the scores in Table 1, although the performance of our method could not reach the performance achieved in good weather, our method could be applied in most cases, as an average score of the PASCAL criterion C ¯   >   0.5 indicates successful detection and tracking.

5. Conclusions

In order to solve problems with object detection on foggy days, this study explores and fuses the color and depth information from image data. A series of tricks, such as skylight removal and data cleaning, are proposed to prevent errors in the training dataset. We separately trained and established background models using the features from the color and depth domains. These two background models were combined under a unified domain adaptation framework, which introduces the model in the source domain (color) to the target domain (depth). In our experiments using public data on foggy days, we achieved the desired object detection results. A potential disadvantage of our method, as indicated by the experimental results, is difficulty in detecting distant objects. This problem can probably be solved by updating the depth estimation method.
Our method is the first to investigate depth-feature-based object detection on foggy days. The strategy of our method can be generalized to other object detection tasks where depth information is available, such as RGB-D data-based object detection. Moreover, a fog detection model is included in our future work, which underlies an all-weather system in real-life conditions.

Author Contributions

Conceptualization, Z.C. and H.Z.; Methodology, Z.C. and X.L.; Software, H.G. and Z.C.; Resources, H.W.; Writing—Original Draft Preparation, Z.C.; Supervision, H.W.; Project Administration, Z.C.; Funding Acquisition, Z.C., H.W. and H.G.

Funding

This work is supported in part by the National Natural Science Foundation of China (Nos. 61501173, 61671202, 61701166), the Fundamental Research Funds for the Central Universities (No. 2017B01914), and the Jiangsu Overseas Scholar Program for University Prominent Young & Middle-aged Teachers and Presidents.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. He, H.; Shao, Z.; Tan, J. Recognition of car makes and models from a single traffic-camera image. IEEE Trans. Intell. Transp. Syst. 2015, 16, 3182–3192. [Google Scholar] [CrossRef]
  2. Wu, L.; Du, X.; Fu, X. Security threats to mobile multimedia applications: Camera-based attacks on mobile phones. IEEE Commun. Mag. 2014, 52, 80–87. [Google Scholar] [CrossRef]
  3. Li, Y.; You, S.; Brown, M.S.; Tan, R.T. Haze visibility enhancement: A survey and quantitative benchmarking. Comput. Vis. Image Underst. 2017, 1, 1–6. [Google Scholar] [CrossRef]
  4. Wang, X.; Chen, N.; Xu, L. Spatiotemporal difference-of-Gaussians filters for robust infrared small target tracking in various complex scenes. Appl. Opt. 2015, 54, 1573–1586. [Google Scholar] [CrossRef]
  5. Almaadeed, N.; Asim, M.; Al-Maadeed, S.; Bouridane, A.; Beghdadi, A. Automatic Detection and Classification of Audio Events for Road Surveillance Applications. Sensors 2018, 18, 1858. [Google Scholar] [CrossRef] [PubMed]
  6. Lee, C.; Moon, J.H. Robust Lane Detection and Tracking for Real-Time Applications. IEEE Trans. Intell. Transp. Syst. 2018, 1–6. [Google Scholar] [CrossRef]
  7. Zhu, Q.; Mai, J.; Shao, L. A fast single image haze removal algorithm using color attenuation prior. IEEE Trans. Image Process. 2015, 24, 3522–3533. [Google Scholar] [PubMed]
  8. Pan, J.; Sun, D.; Pfister, H.; Yang, M.H. Blind image deblurring using dark channel prior. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  9. Wang, J.B.; He, N.; Zhang, L.L.; Lu, K. Single image dehazing with a physical model and dark channel prior. Neurocomputing 2015, 149, 718–728. [Google Scholar] [CrossRef]
  10. Nayar, S.K.; Narasimhan, S.G. Vision in bad weather. In Proceedings of the IEEE International Conference on Computer Vision, Kerkyra, Greece, 20–27 September 1999. [Google Scholar]
  11. Narasimhan, S.G.; Nayar, S.K. Contrast restoration of weather degraded images. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 713–724. [Google Scholar] [CrossRef]
  12. Shwartz, S.; Namer, E.; Schechner, Y.Y. Blind haze separation. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, New York, NY, USA, 17–22 June 2006. [Google Scholar]
  13. Ma, T.; Hu, X.; Zhang, L.; Lian, J.; He, X.; Wang, Y.; Xian, Z. An evaluation of skylight polarization patterns for navigation. Sensors 2015, 15, 5895–5913. [Google Scholar] [CrossRef] [PubMed]
  14. Schechner, Y.Y.; Narasimhan, S.G.; Nayar, S.K. Polarization-based vision through haze. Appl. Opt. 2003, 20, 511–525. [Google Scholar] [CrossRef]
  15. Liang, J.; Ren, L.; Ju, H.; Zhang, W.; Qu, E. Polarimetric dehazing method for dense haze removal based on distribution analysis of angle of polarization. Opt. Express 2015, 23, 46–57. [Google Scholar] [CrossRef] [PubMed]
  16. Huang, J.; Yang, J.; Chen, D.; He, X.; Han, Y.; Zhang, J.; Zhang, Z. Ultra-compact broadband polarization beam splitter with strong expansibility. Photonics Res. 2018, 6, 574–578. [Google Scholar] [CrossRef]
  17. Kopf, J.; Neubert, B.; Chen, B.; Cohen, M.; Cohen-Or, D.; Deussen, O.; Uyttendaele, M.; Lischinski, D. Deep Photo: Model-Based Photograph Enhancement and Viewing; ACM: New York, NY, USA, 2008. [Google Scholar]
  18. Nishino, K.; Kratz, L.; Lombardi, S. Bayesian defogging. Int. J. Comput. Vis. 2012, 98, 263–278. [Google Scholar] [CrossRef]
  19. Meng, G.; Wang, Y.; Duan, J.; Xiang, S.; Pan, C. Efficient image dehazing with boundary constraint and contextual regularization. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, NSW, Australia, 1–8 December 2013. [Google Scholar]
  20. He, K.; Sun, J.; Tang, X. Single image haze removal using dark channel prior. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 2341–2353. [Google Scholar] [PubMed]
  21. Cai, B.; Xu, X.; Tao, D. Real-time video dehazing based on spatio-temporal mrf. In Proceedings of the Pacific Rim Conference on Multimedia, Xian, China, 15–16 September 2016. [Google Scholar]
  22. Wang, G.; Ren, G.; Jiang, L.; Quan, T. Single image dehazing algorithm based on sky region segmentation. Inf. Technol. J. 2013, 12, 1168–1175. [Google Scholar] [CrossRef]
  23. Yu, F.; Qing, C.; Xu, X.; Cai, B. Image and video dehazing using view-based cluster segmentation. In Proceedings of the IEEE International Conference on Visual Communications and Image Processing, Chengdu, China, 27–30 November 2016. [Google Scholar]
  24. Zhu, Y.; Tang, G.; Zhang, X.; Jiang, J.; Tian, Q. Haze removal method for natural restoration of images with sky. Neurocomputing 2018, 275, 499–510. [Google Scholar] [CrossRef]
  25. Oreifej, O.; Li, X.; Shah, M. Simultaneous video stabilization and moving object detection in turbulence. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 450–462. [Google Scholar] [CrossRef] [PubMed]
  26. Gilles, J.; Alvarez, F.; Ferrante, N.; Fortman, M.; Tahir, L.; Tarter, A.; von Seeger, A. Detection of moving objects through turbulent media. Decomposition of Oscillatory vs Non-Oscillatory spatio-temporal vector fields. Image Vis. Comput. 2018, 73, 40–55. [Google Scholar] [CrossRef]
  27. Patel, V.M.; Gopalan, R.; Li, R.; Chellappa, R. Visual domain adaptation: A survey of recent advances. IEEE Signal Process. Mag. 2015, 32, 53–69. [Google Scholar] [CrossRef]
  28. Li, W.; Duan, L.; Xu, D.; Tsang, I.W. Learning with augmented features for supervised and semi-supervised heterogeneous domain adaptation. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 1134–1148. [Google Scholar] [CrossRef] [PubMed]
  29. Li, G.; Lu, Z.; Zhou, Y.; Sun, J.; Xu, Q.; Lao, C.; He, H.; Zhang, G.; Liu, L. Far-field outdoor experimental demonstration of down-looking synthetic aperture ladar. Chin. Opt. Lett. 2017, 15, 082801. [Google Scholar]
  30. Zhou, W.; Jin, N.; Jia, M.; Yang, H.; Cai, X. Three-dimensional positioning method for moving particles based on defocused imaging using single-lens dual-camera system. Chin. Opt. Lett. 2016, 10, 031201. [Google Scholar] [CrossRef]
  31. Wang, S.; Wang, J.; Chung, F.L. Kernel density estimation, kernel methods, and fast learning in large data sets. IEEE Trans. Cybern. 2014, 44, 1–20. [Google Scholar] [CrossRef] [PubMed]
  32. Omid-Zohoor, A.; Young, C.; Ta, D.; Murmann, B. Toward Always-On Mobile Object Detection: Energy Versus Performance Tradeoffs for Embedded HOG Feature Extraction. IEEE Trans. Circuits Syst. Video Technol. 2018, 28, 1102–1115. [Google Scholar] [CrossRef]
  33. Foggy Morning with Traffic. Available online: https://www.youtube.com/watch?v=ekh-BaoCLPU (accessed on 17 December 2016).
  34. Heavy Fog Disrupts Traffic. Available online: https://www.youtube.com/watch?v=jde2I1PSW4Y (accessed on 5 February 2017).
  35. Traffic Congestion as Heavy Fog. Available online: https://www.youtube.com/watch?v=wwxhlFo_Nqw (accessed on 6 February 2017).
  36. Static Shot of Street as People Are Walking and Fog Blows Through. Available online: https://www.youtube.com/watch?v=CWNaPcbc1hE (accessed on 30 June 2015).
  37. Zhang, S.; Yao, H.; Liu, S. Dynamic background modeling and subtraction using spatio-temporal local binary patterns. In Proceedings of the IEEE International Conference on Image Processing, San Diego, CA, USA, 12–15 October 2008. [Google Scholar]
  38. Barnich, O.; Van Droogenbroeck, M. ViBe: A universal background subtraction algorithm for video sequences. IEEE Trans. Image Process. 2011, 20, 1709–1724. [Google Scholar] [CrossRef] [PubMed]
  39. Zhou, X.; Yang, C.; Yu, W. Moving object detection by detecting contiguous outliers in the low-rank representation. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 597–610. [Google Scholar] [CrossRef] [PubMed]
  40. Guo, C.; Ma, Q.; Zhang, L. Spatio-temporal saliency detection using phase spectrum of quaternion fourier transform. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008. [Google Scholar]
  41. Everingham, M.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The PASCAL Visual Object Classes (VOC) Challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef]
  42. Smeulders, A.W.; Chu, D.M.; Cucchiara, R.; Calderara, S.; Dehghan, A.; Shah, M. Visual tracking: An experimental survey. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 1442–1468. [Google Scholar] [PubMed]
  43. Haralick, R.M.; Sternberg, S.R.; Zhuang, X. Image analysis using mathematical morphology. IEEE Trans. Pattern Anal. Mach. Intell. 1987, 4, 532–550. [Google Scholar] [CrossRef]
Figure 1. Framework of the proposed method.
Figure 1. Framework of the proposed method.
Sensors 18 03286 g001
Figure 2. Skylight recognition and depth estimation.
Figure 2. Skylight recognition and depth estimation.
Sensors 18 03286 g002
Figure 3. Cleaning of depth information.
Figure 3. Cleaning of depth information.
Sensors 18 03286 g003
Figure 4. Object detection results obtained using the depth (gray) and color (RGB) information. For each scene, the right panel is the result given by the depth information and the left given by the color information.
Figure 4. Object detection results obtained using the depth (gray) and color (RGB) information. For each scene, the right panel is the result given by the depth information and the left given by the color information.
Sensors 18 03286 g004
Figure 5. Performance comparison. GT: ground truth; PQFT: phase spectrum of the quaternion Fourier transform; ST-MoG: spatiotemporal mixture-of-Gaussian.
Figure 5. Performance comparison. GT: ground truth; PQFT: phase spectrum of the quaternion Fourier transform; ST-MoG: spatiotemporal mixture-of-Gaussian.
Sensors 18 03286 g005
Table 1. Quantitative performance comparison of ST-MoG, Vibe, DECOLOR, PQFT, and our method. FPR: false positive rate; FS: F-score; Pr: precision; PWC: percentage of wrong classifications; Sim: similarity TPR: true positive rate.
Table 1. Quantitative performance comparison of ST-MoG, Vibe, DECOLOR, PQFT, and our method. FPR: false positive rate; FS: F-score; Pr: precision; PWC: percentage of wrong classifications; Sim: similarity TPR: true positive rate.
Method C ¯ Pr TPR FS Sim FPR PWC
ST-MoG0.40150.46860.48260.53410.53650.10623.8571
Vibe0.59510.76900.59520.42980.60010.05514.1294
DECOLOR0.52540.45180.57190.55260.56360.06872.7658
PQFT0.42170.50410.41240.42170.46140.15476.3871
Our method0.61470.62150.67540.59550.59970.04253.9200

Share and Cite

MDPI and ACS Style

Chen, Z.; Li, X.; Zheng, H.; Gao, H.; Wang, H. Domain Adaptation and Adaptive Information Fusion for Object Detection on Foggy Days. Sensors 2018, 18, 3286. https://doi.org/10.3390/s18103286

AMA Style

Chen Z, Li X, Zheng H, Gao H, Wang H. Domain Adaptation and Adaptive Information Fusion for Object Detection on Foggy Days. Sensors. 2018; 18(10):3286. https://doi.org/10.3390/s18103286

Chicago/Turabian Style

Chen, Zhe, Xiaofang Li, Hao Zheng, Hongmin Gao, and Huibin Wang. 2018. "Domain Adaptation and Adaptive Information Fusion for Object Detection on Foggy Days" Sensors 18, no. 10: 3286. https://doi.org/10.3390/s18103286

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop