Next Article in Journal
Low Elevation Angle Estimation with Range Super-Resolution in Wideband Radar
Previous Article in Journal
Processing Fluorescence Spectra for Pollutants Detection Systems in Inland Waters
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Moving Object Detection under a Moving Camera via Background Orientation Reconstruction

College of Aerospace Science and Engineering, National University of Defense Technology, Changsha 410073, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(11), 3103; https://doi.org/10.3390/s20113103
Submission received: 27 April 2020 / Revised: 17 May 2020 / Accepted: 28 May 2020 / Published: 30 May 2020
(This article belongs to the Section Physical Sensors)

Abstract

:
Moving object detection under a moving camera is a challenging question, especially in a complex background. This paper proposes a background orientation field reconstruction method based on Poisson fusion for detecting moving objects under a moving camera. As enlightening by the optical flow orientation of a background is not dependent on the scene depth, this paper reconstructs the background orientation through Poisson fusion based on the modified gradient. Then, the motion saliency map is calculated by the difference between the original and the reconstructed orientation field. Based on the similarity in appearance and motion, the paper also proposes a weighted accumulation enhancement method. It can highlight the motion saliency of the moving objects and improve the consistency within the object and background region simultaneously. Furthermore, the proposed method incorporates the motion continuity to reject the false positives. The experimental results obtained by employing publicly available datasets indicate that the proposed method can achieve excellent performance compared with current state-of-the-art methods.

1. Introduction

Moving object detection is incorporated in numerous applications, such as monitoring systems, unmanned aerial vehicles (UAVs), and automatic pilots [1]. The accurate position and shape information of a moving object are of great significance for subsequent tracking and recognition. Based on the camera platform being mobile or immobile, moving object detection can be classified into two categories: static camera and moving camera. Detecting moving objects under a static camera has been well studied. Background modelling methods [2,3,4] can achieve excellent performance in detecting moving object detection under a static camera. Thus, some methods [5,6,7,8] used motion compensation to adapt the background modeling to detect moving object under a moving camera. By assuming the background can be approximated by one or more dominant planes, the motion compensation methods usually use 2D parametric transformation [9,10] to register the current image with the background image. To eliminate the influence of the feature points on the object, [6] proposed a two-layer optimization method to estimate the affine transformation model. The methods [7,8,11] used two models for background and object to reduce the errors introduced in the motion compensation. Kim [12] used a spatial–temporal distributed Gaussian model to eliminate the false positives by registration error and background adaptation problem. Because of the complex background, the motion compensation by a 2D parametric transformation may be invalid for a scene with great depth variation. Detecting moving objects under a moving camera is much more difficult than that under a static camera. However, detecting moving objects under an unconstrained camera with high accuracy and robustness is very helpful for applications such as UAVs and automatic pilots. Thus, this study focuses on detecting moving object under an unconstrained camera accurately and reliably. Although many different methods have been proposed, the accuracy in detecting moving objects under an unconstrained camera is still unsatisfactory, especially for moving objects in a complex background. Usually, the moving object is detected based on the motion difference with the background, where the motion of the background and object are estimated by calculating the optical flow between adjacent frames. Therefore, the difference in optical flow is used to distinguish the object from the background. The motion projected on the image of the background depends on the distance to the camera, i.e., the depth of the scene. The magnitude of the optical flow for the pixels in the background region may be different, although they share the same real-world motion. The orientation of the optical flow is not dependent on the depth of the scene [13]. The variation of optical flow orientation in the background region is continuous in the spatial domain. According to this characteristic of background optical flow, this paper detects the moving object in the orientation field.
This paper proposes a background orientation reconstruction method to detect moving objects in the orientation field. The proposed method modifies the gradient of the original orientation field by removing the object orientation. Then, the background orientation field is reconstructed through Poisson fusion [14]. According to the difference between the original and reconstructed orientation field, this paper calculates the motion saliency map. For detecting all moving objects completely, the motion saliency map is enhanced through spatially weighted accumulation. Fundamentally, pixels belonging to the same category (moving object or background) are similar in appearance or motion in the corresponding neighborhood, and the weight can be defined for measuring the similarities. Based on this fundamental observation, this study proposes a spatially weighted accumulation method for enhancing the difference between the moving object and the background. Furthermore, this study incorporates the continuity of an object’s motion in the temporal domain for rejecting false positives.

2. Related Works

The conventional methods for moving object detection under a moving camera are generally based on the motion difference of the moving object and the background. The motion of the moving object comprises its own motion and the camera platform’s motion, while the motion of the background is solely caused by the camera platform. The motion cues are generally obtained based on the optical flow estimation between the adjacent frames.
Yazdi et al. [15] classified the methods into four categories: background subtraction, trajectory classification, low rank and sparse matrix decomposition, and object tracking. Chapel et al. [16] made a comprehensive review of methods for detecting moving objects under a moving camera, which were categorized into eight different approach groups. Depending on the density of the optical flow, this paper categorizes conventional methods for moving object detection under an unconstrained camera into two types: sparse optical flow methods [17,18,19,20,21,22,23] and dense optical flow methods [13,24,25,26,27,28,29,30,31,32,33,34].
The methods [17,23] clustered sparse sample pixels based on the optical flows and other spatial features as the moving object and background. Considering the classified pixels as seeds, they segmented the moving objects from the background. Nonaka et al. [21] used three distances to cluster the trajectories, and the cluster is classified according to the shape and size. To enhance the robustness of the extracted motion information, long trajectories have been used in some methods to detect the moving objects [18,19,20,22]. Sheikh et al. [18] calculated the basis vectors of the background trajectories based on the rank constraint to extract the trajectories of the moving objects. Spectral clustering was used to segment the moving object in [19,20] based on the affinities between the long trajectories. The spatial–temporal information was used to segment the moving objects densely based on the segmented sparse trajectories. However, the sparse optical flow methods can achieve satisfactory detection only if there are enough sampling pixels in the moving object region. This requirement can be easily met for large-sized moving objects but not for small-sized ones. Hence, such methods cannot easily detect small-sized moving objects.
Dense optical flow methods estimate the optical flow for each pixel between adjacent frames. The motion characteristics of the moving object and the background are subsequently extracted from the dense optical flow. Then, these methods calculate the motion saliency map to highlight the moving objects for detection. Gao et al. [24] measured the saliency by counting the histogram of the motion features. Huang et al. [25] adopted the homograph transformation to model the background’s motion based on the estimated optical flow. Based on the difference between the estimated optical flow and constructed optical flow by employing the homograph transformation, the motion saliency can be calculated for each pixel. Sajid et al. [26] reconstructed the background motion by a low-rank approximation. The probability of the moving object is estimated by the error between the reconstructed background motion and actual motion. The multi-view geometry constraint was used to distinguish between the object and background. Zhou et al. [27] detected moving targets under moving stereo cameras. The motion difference map, defined as the Residual Image Motion Flow, is calculated by the difference between the Measured Optical Flow and Global Image Motion Flow. The Global Image Motion Flow is obtained through geometric constraints of the moving camera. Finally, the motion likelihood, color, and depth cues are combined in the Markov Random Field (MRF) framework for moving object segmentation by graph-cut. Namdev et al. [28] used motion vectors from dense optical flow and motion potentials based on multi-view geometry to form a graph model. Then, a graph-based segmentation algorithm clustered nodes of similar potentials to create the eventual motion segments. The magnitude of the optical flow is dependent on the scene depth, whereas the orientation of the optical flow is independent of the scene depth. Narayana et al. [13] employed the orientation of the optical flow to calculate the motion saliency. Bideau et al. [29] estimated and eliminated the interferences caused by camera rotation for obtaining the orientation field of the optical flow, which is independent of the scene depth. In another study [30], the local spatial difference of the optical flow was used for calculating the saliency of the moving object’s contour, which was then employed for approximately selecting the object’s pixels. Chen et al. [31] proposed a context-aware motion descriptor based on the histogram of the orientation of the optical flow in a certain neighborhood, and the descriptor was employed for detecting the moving object’s contour. Wu et al. [32] constructed the dense particle trajectories based on the optical flow of multiple frames. Then, the motion saliency is calculated by comparing the trajectories with the extracted dominant motion components through the algorithm Reduced Singular Value Decomposition (RSVD) [34]. Zhu et al. [33] formulated the problem as a multi-label segmentation problem by modeling moving objects in different layers. An independent processing layer was assigned to each moving object and background.
To the best of our knowledge, the methods based on dense optical flow can achieve better performance in detecting and segmenting moving objects under an unconstrained camera. However, the two types of methods all have trouble in distinguishing the moving objects (relative the background) in a complex background. Hence, in this study, a background orientation reconstruction method is proposed for detecting the moving objects completely under a moving camera in a complex background. This paper also proposes an enhancement algorithm to highlight the motion saliency map for the moving objects and to smooth the region within the object and background simultaneously. Furthermore, the false positives are rejected based on the continuity of object motion in the temporal domain in this study.

3. Methodology

This paper proposes a moving object detection algorithm under a moving camera based on background orientation reconstruction. Firstly, the orientation of the optical flow between adjacent frames was calculated. Then, the original orientation field was modified in the gradient domain to remove the object region. Finally, the background orientation field was reconstructed based on the modified gradient through Poisson fusion [14]. The motion saliency can be obtained by the difference between the reconstructed background orientation and original orientation. Additionally, the motion saliency is enhanced through the spatially weighted accumulation of the neighborhood pixels. The detected result can be obtained by thresholding the enhanced motion saliency map. Furthermore, the continuity of the object’s motion in the temporal domain is incorporated for rejecting false detections.

3.1. Poisson Fusion

As shown in Figure 1, the Poisson fusion [14] solves the problem of seamlessly fusing the source image into the target image, meanwhile reserving the gradient information of the source image as far as possible.
To solve the problem, Pérez et al. [14] took the gradient field g of the source image as the guidance; the source image boundary Ω of fusion region Ω served as a hard constraint for the desired image f . Then, the problem can be modeled as the following mathematical problem:
min f Ω | f g | 2     w i t h     f | Ω = f | Ω
where f denotes the gradient field of the desired image f .
Solving the above problem, Equation (1) can obtain the following Poisson equation with Dirichlet boundary condition as follows:
Δ f | Ω = d i v ( g ) | Ω     w i t h     f | Ω = f | Ω
where d i v ( g ) = g x x + g y y is the divergence of the gradient field g = ( g x , g y ) .
The variables in Equation (2) are discrete in the image domain. Equation (2) can be solved discretely by the five-point interpolation method in the numerical method of partial differential equations [35]. The calculation process is shown in the Appendix A, and the result of the Poisson fusion is shown in Figure 1c.

3.2. Motion Saliency through Background Orientation Reconstructed

This study uses the optical flow orientation to distinguish the background and moving object because it does not depend on the scene depth for the background region. The orientation field of the background varies continuously in the spatial domain. Thus, the value of the gradient of the background orientation field tends to be continuous. According to this property, the gradient field of the background orientation can be obtained by smoothing the mutant in the original gradient of the orientation field.
The angle of the optical flow between the adjacent frames is used to describe the orientation field. Then, the gradient of the orientation field can be obtained as follows:
g i , j x = θ i , j + 1 θ i , j g i , j y = θ i + 1 , j θ i , j
where θ represents the angle of the original optical flow orientation. g i , j x and g i , j y denote the gradient of the orientation field in the position ( i , j ) along the horizontal and vertical directions, respectively.
This study defines the mutant as the local maximum in the gradient of the orientation field. The smaller value of the gradient in the neighborhood is used to substitute the local maximum to eliminate the mutant. This paper eliminates the mutant of g i , j x along the horizontal direction as follows:
g ^ i , j x = { g i , j 1     x     | g i , j x | > | g i , j 1     x |   a n d   | g i , j x | > | g i , j + 1     x |   a n d   | g i , j + 1 x | > | g i , j 1     x | g i , j + 1     x     | g i , j x | > | g i , j 1     x |   a n d   | g i , j x | > | g i , j + 1     x |   a n d   | g i , j + 1 x | < | g i , j 1     x | g i , j     x     o t h e r w i s e
where g ^ i , j = ( g ^ i , j x , g ^ i , j y ) denotes the modified gradient in the position ( i , j ) .
The vertical gradient field g i , j y is conducted in a similar manner.
Now, the gradient field g ^ i , j = ( g ^ i , j x , g ^ i , j y ) with no mutant is obtained, which serves as the gradient of the background orientation angle.
Assuming that the pixels in the image boundaries are background, similar to Equation (1), the minimization problem can be set up with respect to the angle of the background orientation as follows:
min θ b Ω | θ b g ^ | 2     w i t h     θ b | Ω = θ | Ω
where θ b denotes the angle of the reconstructed background orientation; and Ω and Ω represent the image boundaries and interior region, respectively.
The motion saliency M i , j is defined as the absolute value of the angle difference between the reconstructed background orientation angle and the original orientation angle.
M i , j = | θ i , j θ i , j b |
As shown in Figure 2, the algorithm proposed in this section can reconstruct the orientation field of the background with relative accuracy (see the middle rows). The moving objects can be highlighted in the motion saliency map calculated by Equation (6).

3.3. Enhancement Algorithm Based on Weighted Spatial Accumulation

As shown in Figure 2 (the first and fourth columns), the moving object may be indistinguishable in that the motion saliency calculated by Equation (6) may be low. There also may be inconsistencies within the background and object region. This section’s objective is to enhance the motion saliency map for detecting moving objects as completely as possible. The motion saliency of a pixel can be enhanced by the accumulation of its surrounding pixels in the spatial domain. The increased amount for the moving object is much larger than the background, and the motion saliency within the object or background would become much more uniform. In order to find the pixels belonging to the same target for more exact accumulation, this study enhances the motion saliency map by weighted accumulation. The pixels that belong to the same object usually are similar in terms of appearance and motion. In this study, the weight of each pixel was determined by measuring the appearance and motion similarity between each pixel pair in a certain neighborhood.
Appearance similarity: The appearance similarity between a pixel p α and its neighborhood pixel p β can be calculated by incorporating the color difference S i , j c , as follows:
S α , β c = exp ( ( R α R β ) 2 + ( G α G β ) 2 + ( B α B β ) 2 σ 1 )                   p β N α
where ( R α , G α , B α ) , respectively, denote the red, blue, and green value of p α ; N α represents the n × n neighborhoods of p α ; p β denotes the neighborhood pixel of p α ; and σ 1 is a positive constant parameter and set to 25 in this study.
Motion similarity: The motion similarity S α , β m between p α and p β can be determined by incorporating the optical flow difference, as follows:
S α , β m = exp ( ( u α u β ) 2 + ( v α v β ) 2 σ 2 )
where ( u α , v α ) and ( u β , v β ) , respectively, denote the optical flow vector of p α and p β . Furthermore, σ 2 represents the variance of the Gaussian function, and its value is set to 5 in this study.
The similarity between p α and p β can be obtained from the product of S α , β c and S α , β m , as shown in Equation (9).
S α , β = S α , β c × S α , β m
If the similarity is directly applied as the cumulative weight, the pixels near the object’s contour and those inside the object will be enhanced by different extents owing to varying numbers of object pixels in their corresponding neighborhoods. To avoid the problem of enhancement being dependent on the pixel position, the summation of the similarities of the neighborhood pixels is normalized to the number of the neighborhoods in this study. Thus, the accumulation of weight can be defined as
W α , β = ( n × n 1 ) × S ^ α , β j S ^ α , β ,   where   S ^ α , β = S α , β min j S α , β max β S α , β min β S α , β .
where the value of n is set to 9 in this study.
The enhanced motion saliency M ^ α of pixel p α can be obtained through the spatial accumulation of its neighborhoods with the similarity weight, as shown in Equation (11).
M ^ α = M α + p β N α , p β p α W α , β × M β
where M α and M β donate the motion saliency of pixel p α and p β , respectively.
As shown in Figure 3, the motion saliency of a moving object can be enhanced by Equation (11). To detect all moving objects as completely as possible, this paper adopts a relatively small threshold. The value of the threshold T is determined by incorporating the mean value m and standard deviation σ of the enhanced motion saliency, as follows:
T = m + σ

3.4. False Positives Rejection Based on Motion Continuity in the Temporal Domain

False positives are usually caused by a cluttered background, inaccuracy in the estimated optical flow, and a flawed threshold algorithm, which may exist in the detected results. This study utilizes motion continuity in the temporal domain to address the issue. The trajectory of a real moving object is continuous in the temporal domain, but a false alarm does not have a continuous trajectory. In this study, false detection in the current frame is rejected according to the detection in the next frame, and the correspondence between the current frame and the next frame is established through the optical flow. If a detected region in the current frame is consistent with the next frame’s detection, it is determined as a true moving object; otherwise, it is determined as a false positive. Considering the variation of the moving object shape caused by changes in viewing and the object’s movement, the consistency between the adjacent frames is measured by the area overlap ratio. If the overlap ratio is greater than the threshold λ , the region is identified as a true moving object, and vice versa.
O t , m = { 1           | O ^ t + 1 , m O t + 1 | | O t , m | > λ 0               o t h e r w i s e
where O t , m represents the mth detected object region in the current frame, and O ^ t + 1 , m denotes the projected region in the next frame from the object O t , m based on the optical flow. Furthermore, O t + 1 represents the detected object region in the next frame, and | · | represents the number of total pixels within the region. In this study, the value of the threshold λ is set to 0.3.
As shown in Figure 4, there may exist some false positives in the background regions. The proposed false positives can effectively eliminate the false positives.

4. Experiment

To evaluate the performance of the proposed method, this study conducted experiments by employing the publicly available datasets moseg_dataset [20], BMS (Background Motion Subtraction) [32], and DAVIS (Densely Annotated Video Segmentation) [36]. The publicly available datasets comprise multiple image sequences, which contain moving objects in different complex backgrounds captured by moving cameras with different movements. The quantitative and qualitative comparisons were made with compared methods [5,25,29,30]. The codes of algorithms [5,25,29,30] were all downloaded from the authors’ homepage. The parameters suggested by the respective authors were incorporated into the compared methods. This paper calculated the optical flow in [25,29,30] and the proposed method by employing the same algorithm [37]. Furthermore, all the objects detected by the proposed algorithm were obtained by the same parameters, as described earlier.

4.1. Qualitative Comparison

Figure 5 illustrates some detected results generated from different sequences of the dataset moseg_dataset [20]. As shown in Figure 5c, the detected objects by [5] lost the shape information, which can only be used to determine the approximate position of the objects. The detected results by [25] have some false positives in the background region because of the assumption that it is approximating the background by a plane. Some parts of the moving objects are missed in the results by [30], which loses the complete shape of the objects. The proposed method detects the moving objects much more completely and with less false positives.
Figure 6 illustrates some sample results on the dataset BMS [32] obtained by methods [5,25,29,30] and the proposed method. Seq2 was divided into three subsequences because the algorithm in [29,30] cannot process the entire sequence at once due to the memory constraint. As seen from Figure 6, the proposed method can detect the moving objects completely with less false positives.
Some sample results on the dataset DAVIS [36] obtained by methods [5,25,29,30] and the proposed method are shown in Figure 7. As seen from Figure 7, the proposed method can detect the moving objects completely with less false positives. The detected results by [25] tend to have false positives in the background region. A similar problem appeared in the results by [29], such as car-shadow, car-turn, and motorbike. This problem does not occur in the detected result by the proposed method because of good adaptability to the change in scene depth.

4.2. Quantitative Comparison

In this study, the overlap rate was employed for quantitatively evaluating the accuracy of the detection algorithms. Given the detected object D O and the ground truth G O , the overlap ratio γ is defined as
γ = | D O G O | | D O G O |
The obtained results are summarized in Table 1, wherein figures in boldface font represent the highest overlap rate, and figures in italic font represent the second-highest overlap rate. It can be observed from Table 1 that the proposed method achieves the best performance for all the datasets. The proposed method is comparable to the state-of-the-art algorithms in [25,29,30]. Comparing to the algorithm described in [25], the proposed background orientation reconstruction algorithm does not assume that the background is approximated by planes. Therefore, the proposed method is able to adapt to more complex scenarios. The algorithm [29] has poor robustness in that the moving object may not be detected for a long period of time, such as Seq1 and Cars2. However, the proposed algorithm is more robust because of good adaptability to the variation of scene depth. Thus, our algorithm usually obtained a relatively high score for the sequences difficult to detect, such as Seq1, Seq2(3), Cars4, Cars9, and motorbike. Furthermore, it should be noted that the algorithm in [30] employs many frames in the future to segment the moving object in the current frame. The algorithm in [29] also uses complex optimization to estimate the background orientation. The complexity of the algorithm in [29] is much higher than that of the proposed algorithm.
There is one case in which the proposed method cannot work: Because only the orientation of the optical flow is used to calculate the motion difference between the object and background, the object moving in the same direction as the camera cannot be detected by the proposed method. The amplitude of the optical flow can be used as Supplementary Information to improve the proposed method.

4.3. Computational Efficiency

The proposed background orientation reconstruction method can be very efficient due to the coefficient matrix only depending on the resolution ratio of the image; so, it only needs to be solved once for one sequence because of the same size of the images within the same sequence. However, it needs a large amount of memory. Thus, this paper solves the Poisson equation every time for each frame, and the motion saliency enhancement method is also efficient owing to its high parallelism; it takes about 8 ms for processing each frame at a resolution of 480 × 640 by employing an NVIDIA GTX 1080 from Asus, Chongqing, china. The procedure of false positives rejection takes very little time. Table 2 shows the average computation time per frame with a resolution of 480 × 640 measured by an Intel Core i5-6200U, 2.4GHz PC from Lenovo, Beijing, china. The method [5] was implemented in C++ and the other four by MATLAB. The time spent on optical flow computation was excluded, which is required for all methods except the one in [5]. The computational efficiency of the proposed method is superior to that of the algorithms in [29,30], and the computational efficiency of the proposed method can be improved greatly through adopting a more advanced sparse equation solving method.

5. Conclusions

This paper proposes a novel method for detecting moving objects under a moving camera. This paper reconstructs the background orientation field through Poisson fusion based on the modified gradient and the orientation in image boundaries. The motion saliency map can be obtained by the difference between the original and reconstructed background orientation field. Based on the appearance and motion similarity, the proposed method enhances the motion saliency map through weighted accumulation in the spatial domain. Furthermore, the proposed method incorporates motion continuity in the temporal domain for rejecting false positives. Experimental results based on the publicly available datasets indicate that the proposed method can achieve an excellent performance through qualitative and quantitative comparison.
Since our paper only uses the optical flow orientation, the object moving in the same direction as the camera cannot be detected. In this situation, the amplitude of the optical flow between the moving object and background must be different, or the object is static. In the next work, we will use the amplitude as a check and supplement the scheme, to verify if there are missed objects, and to detect them. On the other hand, the thresholding method for motion saliency mapping is simple, which influences the accuracy of the proposed method. We will look at getting a more precise object from the motion saliency map.

Author Contributions

W.Z., X.S. and Q.Y. designed the research and co-wrote the paper; W.Z. designed the algorithm and programmed the code; W.Z. and X.S. conducted the experiments and analysis of the data; W.Z. and Q.Y. wrote the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

The research was supported by the National Natural Science Foundation of China (11902349).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Five-point interpolation method for solving the Poisson equation in reconstructing the background orientation.
The problem is to solve following optimizing formula:
min θ b Ω | θ b g ^ | 2     w i t h     θ b | Ω = θ | Ω
Solving the above problem, we can obtain the following Poisson equation with Dirichlet boundary conditions:
Δ θ b | Ω = d i v ( g ^ ) | Ω     w i t h     θ b | Ω = θ | Ω
The discrete second-order differential of θ b along the x and y direction is as follows:
( 2 θ i , j b x 2 ) = 1 h 2 ( θ i , j + 1 b + θ i , j 1 b 2 θ i , j b ) + O ( h 2 ) ( 2 θ i , j b y 2 ) = 1 h 2 ( θ i + 1 , j b + θ i 1 , j b 2 θ i , j b ) + O ( h 2 )
Here, h denotes the step of the discretization, which is usually set to 1 in image process.
If we ignore the error term O ( h 2 ) and set the boundary size to 1, the discrete Poisson equation is as follows:
θ i , j + 1 b + θ i , j 1 b + θ i + 1 , j b + θ i 1 , j b 4 θ i , j b = d i v ( g ^ i , j )           2 i r 1 ; 2 j c 1
Here, r and c respectively denote the height and width of the orientation field.
Vectorizing θ b by column priority as
θ T = [ θ 2 , 2   θ 3 , 2 θ r 1 , 2   θ 2 , 3   θ r 1 , c 1 ]
Now, the following equation respect to θ T can be obtained:
A [ ( r 2 ) ( c 2 ) ] × [ ( r 2 ) ( c 2 ) ] · θ = b ( r 2 ) ( c 2 )
where the vector b is depended on the gradient field g ^ and the orientation angle of the pixels at the boundaries; and A is a positive definite symmetric matrix of order ( r 2 ) ( c 2 ) , as follows:
A = [ B I I B I I B I I B ]
Here, I is unit matrix of order ( r 2 ) ; and B is a symmetric matrix of order ( r 2 ) , as follows:
B ( r 2 ) × ( r 2 ) = [ 4 1 1 4 1 1 4 1 1 4 ]
The coefficient matrix A in Equation (A5) is a large sparse matrix, depending only on the size of the image.

References

  1. Royden, C.S.; Moore, K.D. Use of speed cues in the detection of moving objects by moving observers. Vis. Res. 2012, 59, 17–24. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Barnich, O.; Van Droogenbroeck, M. ViBe: A universal background subtraction algorithm for video sequences. IEEE Trans. Image Process. 2010, 20, 1709–1724. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Elgammal, A.; Harwood, D.; Davis, L. Non-parametric model for background subtraction. In Proceedings of the 6th European Conference on Computer Vision, Dublin, Ireland, 26 June–1 July 2000; Springer: Berlin/Heidelberg, Germany, 2000. [Google Scholar]
  4. Yong, H.; Meng, D.; Zuo, W.; Zhang, K. Robust online matrix factorization for dynamic background subtraction. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 1726–1740. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Yi, K.M.; Yun, K.; Kim, S.W.; Chang, H.J.; Choi, J.Y.; Jeong, H. Detection of moving objects with non-stationary cameras in 5.8ms: Bringing motion detection to your mobile device. In Computer Vision & Pattern Recognition Workshops; IEEE: Portland, Oregon, 2013. [Google Scholar]
  6. Wan, Y.; Wang, X.; Hu, H. Automatic moving object segmentation for freely moving cameras. Math. Probl. Eng. 2014, 2014, 1–11. [Google Scholar] [CrossRef]
  7. Wu, M.; Peng, X.; Zhang, Q. Segmenting moving objects from a freely moving camera with an effective segmentation cue. Meas. Sci. Technol. 2011, 22, 25108. [Google Scholar] [CrossRef]
  8. Kurnianggoro, L.; Yu, Y.; Hernandez, D.; Jo, K.-H. Online Background-Subtraction with Motion Compensation for Freely Moving Camera. In International Conference on Intelligent Computing; IEEE: Lanzhou, China, 2016; pp. 569–578. [Google Scholar] [CrossRef]
  9. Odobez, J.M.; Bouthemy, P. Separation of moving regions from background in an image sequence acquired with a mobile camera. In Video Data Compression for Multimedia Computing: Statistically Based and Biologically Inspired Techniques; Springer: Boston, MA, USA, 1997. [Google Scholar] [CrossRef]
  10. Hartley, R.; Zisserman, A. Multiple View Geometry in Computer Vision; Cambridge University Press: Cambridge, UK, 2003. [Google Scholar] [CrossRef]
  11. Zhang, X.; Sain, A.; Qu, Y.; Ge, Y.; Hu, H. Background subtraction based on integration of alternative cues in freely moving camera. IEEE Trans. Circuits Syst. Video Technol. 2019, 29, 1933–1945. [Google Scholar] [CrossRef]
  12. Kim, S.W.; Yun, K.; Yi, K.M.; Kim, S.J.; Choi, J.Y. Detection of moving objects with a moving camera using non-panoramic background model. Mach. Vis. Appl. 2012, 24, 1015–1028. [Google Scholar] [CrossRef]
  13. Narayana, M.; Hanson, A.; Learned-Miller, E. Coherent motion segmentation in moving camera videos using optical flow orientations. In Proceedings of the 2013 IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 1577–1584. [Google Scholar]
  14. Pérez, P.; Gangnet, M.; Blake, A. Poisson image editing. ACM Trans. Graphics 2003, 22, 313. [Google Scholar] [CrossRef]
  15. Yazdi, M.; Bouwmans, T. New trends on moving object detection in video images captured by a moving camera: A survey. Comput. Sci. Rev. 2018, 28, 157–177. [Google Scholar] [CrossRef]
  16. Chapel, M.-N.; Bouwmans, T. Moving objects detection with a moving camera: A comprehensive review. arXiv 2020, arXiv:2001.05238. Available online: https://arxiv.org/abs/2001.05238 (accessed on 30 May 2020).
  17. Kim, J.; Wang, X.; Wang, H.; Zhu, C.; Kim, D. Fast moving object detection with non-stationary background. Multimedia Tools Appl. 2012, 67, 311–335. [Google Scholar] [CrossRef]
  18. Sheikh, Y.; Javed, O.; Kanade, T. Background subtraction for freely moving cameras. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009; pp. 1219–1225. [Google Scholar]
  19. Brox, T.; Malik, J. Object segmentation by long term analysis of point trajectories. In Proceedings of the 11th European Conference on Computer Vision, Heraklion, Greece, 5–11 September 2010; Springer: Heraklion Greece, 2010; pp. 282–295. [Google Scholar]
  20. Ochs, P.; Malik, J.; Brox, T. Segmentation of moving objects by long term video analysis. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 36, 1187–1200. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Nonaka, Y.; Shimada, A.; Nagahara, H.; Taniguchi, R.-I. Real-time foreground segmentation from moving camera based on case-based trajectory classification. In Proceedings of the 2013 2nd IAPR Asian Conference on Pattern Recognition, Okinawa, Japan, 5–8 November 2013; pp. 808–812. [Google Scholar]
  22. Elqursh, A.; Elgammal, A. Online moving camera background subtraction. Appl. Evol. Comput. 2012, 7577, 228–241. [Google Scholar]
  23. Bugeau, A.; Perez, P. Detection and segmentation of moving objects in complex scenes. Comput. Vis. Image Underst. 2009, 113, 459–476. [Google Scholar] [CrossRef] [Green Version]
  24. Gao, Z.; Tang, W.; He, L. Moving object detection with moving camera based on motion saliency. J. Comput. Appl. 2016, 36, 1692–1698. [Google Scholar]
  25. Huang, J.; Zou, W.; Zhu, J.; Zhu, Z. Optical flow based real-time moving object detection in unconstrained scenes. arXiv 2018, arXiv:1807.04890. Available online: https://arxiv.org/abs/1807.04890 (accessed on 30 May 2020).
  26. Sajid, H.; Cheung, S.-C.S.; Jacobs, N. Motion and appearance based background subtraction for freely moving cameras. Signal. Process. Image Commun. 2019, 75, 11–21. [Google Scholar] [CrossRef]
  27. Zhou, D.; Frémont, V.; Quost, B.; Dai, Y.; Li, H. Moving object detection and segmentation in urban environments from a moving platform. Image Vis. Comput. 2017, 68, 76–87. [Google Scholar] [CrossRef]
  28. Namdev, R.K.; Kundu, A.; Krishna, K.M.; Jawahar, C.V. Motion segmentation of multiple objects from a freely moving monocular camera. In Proceedings of the 2012 IEEE International Conference on Robotics and Automation, Saint Paul, MN, USA, 14–18 May 2012; pp. 4092–4099. [Google Scholar]
  29. Bideau, P.; Learned-Miller, E. It’s Moving! A probabilistic model for causal motion segmentation in moving camera videos. In Proceedings of the 14th European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016. [Google Scholar]
  30. Papazoglou, A.; Ferrari, V. Fast object segmentation in unconstrained video. In Proceedings of the 2013 IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 1777–1784. [Google Scholar]
  31. Chen, T.; Lu, S. Object-level motion detection from moving cameras. IEEE Trans. Circuits Syst. Video Technol. 2016, 27, 2333–2343. [Google Scholar] [CrossRef]
  32. Wu, Y.; He, X.; Nguyen, T.Q. Moving object detection with a freely moving camera via background motion subtraction. IEEE Trans. Circuits Syst. Video Technol. 2015, 27, 236–248. [Google Scholar] [CrossRef]
  33. Zhu, Y.; Elgammal, A. A multilayer-based framework for online background subtraction with freely moving cameras. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 5142–5151. [Google Scholar]
  34. Trefethen, L.N.; Bau, D. Numerical Linear Algebra; SIAM: Philadelphia, PA, USA, 1997. [Google Scholar]
  35. Hu, J.; Tang, H. Numerical Method of Differential Equation; Science Press: Beijing, China, 2007. [Google Scholar]
  36. Perazzi, F.; Pont-Tuset, J.; McWilliams, B.; Van Gool, L.; Gross, M.; Sorkine-Hornung, A. A benchmark dataset and evaluation methodology for video object segmentation. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 724–732. [Google Scholar]
  37. Sun, D.; Roth, S.; Black, M. Secrets of optical flow estimation and their principles. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2010, San Francisco, CA, USA, 13–18 June 2010. [Google Scholar]
Figure 1. The sketch map of a Poisson fusion process: (a) is the source image; (b) is the fusion region Ω in the target image; and (c) is the desired image f fused into the target image f .
Figure 1. The sketch map of a Poisson fusion process: (a) is the source image; (b) is the fusion region Ω in the target image; and (c) is the desired image f fused into the target image f .
Sensors 20 03103 g001
Figure 2. Some examples of reconstructed background orientation angle (middle row) and motion saliency map (bottom row). The maps in the (top row) are the original orientation angle.
Figure 2. Some examples of reconstructed background orientation angle (middle row) and motion saliency map (bottom row). The maps in the (top row) are the original orientation angle.
Sensors 20 03103 g002
Figure 3. Motion saliency maps before (top row) and after (bottom row) enhancement (the color map is the same as in Figure 2).
Figure 3. Motion saliency maps before (top row) and after (bottom row) enhancement (the color map is the same as in Figure 2).
Sensors 20 03103 g003
Figure 4. The detected results before (top row) and after (bottom row) false positives rejection (the red color denotes the detected objects).
Figure 4. The detected results before (top row) and after (bottom row) false positives rejection (the red color denotes the detected objects).
Sensors 20 03103 g004
Figure 5. Samples of the detection results using the public dataset moseg_dataset [20]. (a) are the original images and (b) are the corresponding ground-truthed images; (cf) are the results of methods in [5,25,29,30], respectively; and (g) images from the proposed method.
Figure 5. Samples of the detection results using the public dataset moseg_dataset [20]. (a) are the original images and (b) are the corresponding ground-truthed images; (cf) are the results of methods in [5,25,29,30], respectively; and (g) images from the proposed method.
Sensors 20 03103 g005
Figure 6. Some detected results on the public dataset BMS [32] of [5,25,29,30] and the proposed method.
Figure 6. Some detected results on the public dataset BMS [32] of [5,25,29,30] and the proposed method.
Sensors 20 03103 g006
Figure 7. Some detected results from DAVIS [36] by the proposed and compared methods [5,25,29,30].
Figure 7. Some detected results from DAVIS [36] by the proposed and compared methods [5,25,29,30].
Sensors 20 03103 g007
Table 1. The overlap rate (%) obtained after applying the methods in [5,25,29,30] and the proposed method.
Table 1. The overlap rate (%) obtained after applying the methods in [5,25,29,30] and the proposed method.
Method[5][25][29][30]Proposed Method
Sequence
Seq126.4121.8235.088.4936.40
Seq2(1)34.5550.4963.6633.8854.95
Seq2(2)39.5425.8958.1642.2449.43
Seq2(3)24.4033.2724.8524.2449.27
Seq322.4658.5037.8419.9841.54
Seq739.1587.6321.2660.4066.70
Cars127.0464.2459.2468.0374.12
Cars210.4519.9865.351.0748.42
Cars419.9445.2717.2937.2262.32
Cars618.3288.0489.1780.5480.30
Cars722.7064.8091.2442.2781.05
Cars827.4583.2080.6971.6981.41
Cars912.5131.2040.7341.0544.82
People137.9775.6476.5755.4470.34
People236.7369.0577.4675.5273.31
bear11.6137.3479.7683.6870.18
bus39.1981.4279.1970.9080.49
car-shadow29.1779.0330.8869.7974.50
car-turn48.0264.1065.3368.6367.05
motorbike36.4949.9242.3033.5359.91
train16.4775.8390.2272.7075.03
Average27.6557.4658.3950.5463.88
Table 2. Time consumption of the proposed and other methods.
Table 2. Time consumption of the proposed and other methods.
Method[5][25][29][30]Proposed Method
Time (ms)46.2128.941075.356042.88854.96

Share and Cite

MDPI and ACS Style

Zhang, W.; Sun, X.; Yu, Q. Moving Object Detection under a Moving Camera via Background Orientation Reconstruction. Sensors 2020, 20, 3103. https://doi.org/10.3390/s20113103

AMA Style

Zhang W, Sun X, Yu Q. Moving Object Detection under a Moving Camera via Background Orientation Reconstruction. Sensors. 2020; 20(11):3103. https://doi.org/10.3390/s20113103

Chicago/Turabian Style

Zhang, Wenlong, Xiaoliang Sun, and Qifeng Yu. 2020. "Moving Object Detection under a Moving Camera via Background Orientation Reconstruction" Sensors 20, no. 11: 3103. https://doi.org/10.3390/s20113103

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop