You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

30 May 2020

Moving Object Detection under a Moving Camera via Background Orientation Reconstruction

,
and
College of Aerospace Science and Engineering, National University of Defense Technology, Changsha 410073, China
*
Author to whom correspondence should be addressed.
This article belongs to the Section Physical Sensors

Abstract

Moving object detection under a moving camera is a challenging question, especially in a complex background. This paper proposes a background orientation field reconstruction method based on Poisson fusion for detecting moving objects under a moving camera. As enlightening by the optical flow orientation of a background is not dependent on the scene depth, this paper reconstructs the background orientation through Poisson fusion based on the modified gradient. Then, the motion saliency map is calculated by the difference between the original and the reconstructed orientation field. Based on the similarity in appearance and motion, the paper also proposes a weighted accumulation enhancement method. It can highlight the motion saliency of the moving objects and improve the consistency within the object and background region simultaneously. Furthermore, the proposed method incorporates the motion continuity to reject the false positives. The experimental results obtained by employing publicly available datasets indicate that the proposed method can achieve excellent performance compared with current state-of-the-art methods.

1. Introduction

Moving object detection is incorporated in numerous applications, such as monitoring systems, unmanned aerial vehicles (UAVs), and automatic pilots []. The accurate position and shape information of a moving object are of great significance for subsequent tracking and recognition. Based on the camera platform being mobile or immobile, moving object detection can be classified into two categories: static camera and moving camera. Detecting moving objects under a static camera has been well studied. Background modelling methods [,,] can achieve excellent performance in detecting moving object detection under a static camera. Thus, some methods [,,,] used motion compensation to adapt the background modeling to detect moving object under a moving camera. By assuming the background can be approximated by one or more dominant planes, the motion compensation methods usually use 2D parametric transformation [,] to register the current image with the background image. To eliminate the influence of the feature points on the object, [] proposed a two-layer optimization method to estimate the affine transformation model. The methods [,,] used two models for background and object to reduce the errors introduced in the motion compensation. Kim [] used a spatial–temporal distributed Gaussian model to eliminate the false positives by registration error and background adaptation problem. Because of the complex background, the motion compensation by a 2D parametric transformation may be invalid for a scene with great depth variation. Detecting moving objects under a moving camera is much more difficult than that under a static camera. However, detecting moving objects under an unconstrained camera with high accuracy and robustness is very helpful for applications such as UAVs and automatic pilots. Thus, this study focuses on detecting moving object under an unconstrained camera accurately and reliably. Although many different methods have been proposed, the accuracy in detecting moving objects under an unconstrained camera is still unsatisfactory, especially for moving objects in a complex background. Usually, the moving object is detected based on the motion difference with the background, where the motion of the background and object are estimated by calculating the optical flow between adjacent frames. Therefore, the difference in optical flow is used to distinguish the object from the background. The motion projected on the image of the background depends on the distance to the camera, i.e., the depth of the scene. The magnitude of the optical flow for the pixels in the background region may be different, although they share the same real-world motion. The orientation of the optical flow is not dependent on the depth of the scene []. The variation of optical flow orientation in the background region is continuous in the spatial domain. According to this characteristic of background optical flow, this paper detects the moving object in the orientation field.
This paper proposes a background orientation reconstruction method to detect moving objects in the orientation field. The proposed method modifies the gradient of the original orientation field by removing the object orientation. Then, the background orientation field is reconstructed through Poisson fusion []. According to the difference between the original and reconstructed orientation field, this paper calculates the motion saliency map. For detecting all moving objects completely, the motion saliency map is enhanced through spatially weighted accumulation. Fundamentally, pixels belonging to the same category (moving object or background) are similar in appearance or motion in the corresponding neighborhood, and the weight can be defined for measuring the similarities. Based on this fundamental observation, this study proposes a spatially weighted accumulation method for enhancing the difference between the moving object and the background. Furthermore, this study incorporates the continuity of an object’s motion in the temporal domain for rejecting false positives.

3. Methodology

This paper proposes a moving object detection algorithm under a moving camera based on background orientation reconstruction. Firstly, the orientation of the optical flow between adjacent frames was calculated. Then, the original orientation field was modified in the gradient domain to remove the object region. Finally, the background orientation field was reconstructed based on the modified gradient through Poisson fusion []. The motion saliency can be obtained by the difference between the reconstructed background orientation and original orientation. Additionally, the motion saliency is enhanced through the spatially weighted accumulation of the neighborhood pixels. The detected result can be obtained by thresholding the enhanced motion saliency map. Furthermore, the continuity of the object’s motion in the temporal domain is incorporated for rejecting false detections.

3.1. Poisson Fusion

As shown in Figure 1, the Poisson fusion [] solves the problem of seamlessly fusing the source image into the target image, meanwhile reserving the gradient information of the source image as far as possible.
Figure 1. The sketch map of a Poisson fusion process: (a) is the source image; (b) is the fusion region Ω in the target image; and (c) is the desired image f fused into the target image f .
To solve the problem, Pérez et al. [] took the gradient field g of the source image as the guidance; the source image boundary Ω of fusion region Ω served as a hard constraint for the desired image f . Then, the problem can be modeled as the following mathematical problem:
min f Ω | f g | 2     w i t h     f | Ω = f | Ω
where f denotes the gradient field of the desired image f .
Solving the above problem, Equation (1) can obtain the following Poisson equation with Dirichlet boundary condition as follows:
Δ f | Ω = d i v ( g ) | Ω     w i t h     f | Ω = f | Ω
where d i v ( g ) = g x x + g y y is the divergence of the gradient field g = ( g x , g y ) .
The variables in Equation (2) are discrete in the image domain. Equation (2) can be solved discretely by the five-point interpolation method in the numerical method of partial differential equations []. The calculation process is shown in the Appendix A, and the result of the Poisson fusion is shown in Figure 1c.

3.2. Motion Saliency through Background Orientation Reconstructed

This study uses the optical flow orientation to distinguish the background and moving object because it does not depend on the scene depth for the background region. The orientation field of the background varies continuously in the spatial domain. Thus, the value of the gradient of the background orientation field tends to be continuous. According to this property, the gradient field of the background orientation can be obtained by smoothing the mutant in the original gradient of the orientation field.
The angle of the optical flow between the adjacent frames is used to describe the orientation field. Then, the gradient of the orientation field can be obtained as follows:
g i , j x = θ i , j + 1 θ i , j g i , j y = θ i + 1 , j θ i , j
where θ represents the angle of the original optical flow orientation. g i , j x and g i , j y denote the gradient of the orientation field in the position ( i , j ) along the horizontal and vertical directions, respectively.
This study defines the mutant as the local maximum in the gradient of the orientation field. The smaller value of the gradient in the neighborhood is used to substitute the local maximum to eliminate the mutant. This paper eliminates the mutant of g i , j x along the horizontal direction as follows:
g ^ i , j x = { g i , j 1     x     | g i , j x | > | g i , j 1     x |   a n d   | g i , j x | > | g i , j + 1     x |   a n d   | g i , j + 1 x | > | g i , j 1     x | g i , j + 1     x     | g i , j x | > | g i , j 1     x |   a n d   | g i , j x | > | g i , j + 1     x |   a n d   | g i , j + 1 x | < | g i , j 1     x | g i , j     x     o t h e r w i s e
where g ^ i , j = ( g ^ i , j x , g ^ i , j y ) denotes the modified gradient in the position ( i , j ) .
The vertical gradient field g i , j y is conducted in a similar manner.
Now, the gradient field g ^ i , j = ( g ^ i , j x , g ^ i , j y ) with no mutant is obtained, which serves as the gradient of the background orientation angle.
Assuming that the pixels in the image boundaries are background, similar to Equation (1), the minimization problem can be set up with respect to the angle of the background orientation as follows:
min θ b Ω | θ b g ^ | 2     w i t h     θ b | Ω = θ | Ω
where θ b denotes the angle of the reconstructed background orientation; and Ω and Ω represent the image boundaries and interior region, respectively.
The motion saliency M i , j is defined as the absolute value of the angle difference between the reconstructed background orientation angle and the original orientation angle.
M i , j = | θ i , j θ i , j b |
As shown in Figure 2, the algorithm proposed in this section can reconstruct the orientation field of the background with relative accuracy (see the middle rows). The moving objects can be highlighted in the motion saliency map calculated by Equation (6).
Figure 2. Some examples of reconstructed background orientation angle (middle row) and motion saliency map (bottom row). The maps in the (top row) are the original orientation angle.

3.3. Enhancement Algorithm Based on Weighted Spatial Accumulation

As shown in Figure 2 (the first and fourth columns), the moving object may be indistinguishable in that the motion saliency calculated by Equation (6) may be low. There also may be inconsistencies within the background and object region. This section’s objective is to enhance the motion saliency map for detecting moving objects as completely as possible. The motion saliency of a pixel can be enhanced by the accumulation of its surrounding pixels in the spatial domain. The increased amount for the moving object is much larger than the background, and the motion saliency within the object or background would become much more uniform. In order to find the pixels belonging to the same target for more exact accumulation, this study enhances the motion saliency map by weighted accumulation. The pixels that belong to the same object usually are similar in terms of appearance and motion. In this study, the weight of each pixel was determined by measuring the appearance and motion similarity between each pixel pair in a certain neighborhood.
Appearance similarity: The appearance similarity between a pixel p α and its neighborhood pixel p β can be calculated by incorporating the color difference S i , j c , as follows:
S α , β c = exp ( ( R α R β ) 2 + ( G α G β ) 2 + ( B α B β ) 2 σ 1 )                   p β N α
where ( R α , G α , B α ) , respectively, denote the red, blue, and green value of p α ; N α represents the n × n neighborhoods of p α ; p β denotes the neighborhood pixel of p α ; and σ 1 is a positive constant parameter and set to 25 in this study.
Motion similarity: The motion similarity S α , β m between p α and p β can be determined by incorporating the optical flow difference, as follows:
S α , β m = exp ( ( u α u β ) 2 + ( v α v β ) 2 σ 2 )
where ( u α , v α ) and ( u β , v β ) , respectively, denote the optical flow vector of p α and p β . Furthermore, σ 2 represents the variance of the Gaussian function, and its value is set to 5 in this study.
The similarity between p α and p β can be obtained from the product of S α , β c and S α , β m , as shown in Equation (9).
S α , β = S α , β c × S α , β m
If the similarity is directly applied as the cumulative weight, the pixels near the object’s contour and those inside the object will be enhanced by different extents owing to varying numbers of object pixels in their corresponding neighborhoods. To avoid the problem of enhancement being dependent on the pixel position, the summation of the similarities of the neighborhood pixels is normalized to the number of the neighborhoods in this study. Thus, the accumulation of weight can be defined as
W α , β = ( n × n 1 ) × S ^ α , β j S ^ α , β ,   where   S ^ α , β = S α , β min j S α , β max β S α , β min β S α , β .
where the value of n is set to 9 in this study.
The enhanced motion saliency M ^ α of pixel p α can be obtained through the spatial accumulation of its neighborhoods with the similarity weight, as shown in Equation (11).
M ^ α = M α + p β N α , p β p α W α , β × M β
where M α and M β donate the motion saliency of pixel p α and p β , respectively.
As shown in Figure 3, the motion saliency of a moving object can be enhanced by Equation (11). To detect all moving objects as completely as possible, this paper adopts a relatively small threshold. The value of the threshold T is determined by incorporating the mean value m and standard deviation σ of the enhanced motion saliency, as follows:
T = m + σ
Figure 3. Motion saliency maps before (top row) and after (bottom row) enhancement (the color map is the same as in Figure 2).

3.4. False Positives Rejection Based on Motion Continuity in the Temporal Domain

False positives are usually caused by a cluttered background, inaccuracy in the estimated optical flow, and a flawed threshold algorithm, which may exist in the detected results. This study utilizes motion continuity in the temporal domain to address the issue. The trajectory of a real moving object is continuous in the temporal domain, but a false alarm does not have a continuous trajectory. In this study, false detection in the current frame is rejected according to the detection in the next frame, and the correspondence between the current frame and the next frame is established through the optical flow. If a detected region in the current frame is consistent with the next frame’s detection, it is determined as a true moving object; otherwise, it is determined as a false positive. Considering the variation of the moving object shape caused by changes in viewing and the object’s movement, the consistency between the adjacent frames is measured by the area overlap ratio. If the overlap ratio is greater than the threshold λ , the region is identified as a true moving object, and vice versa.
O t , m = { 1           | O ^ t + 1 , m O t + 1 | | O t , m | > λ 0               o t h e r w i s e
where O t , m represents the mth detected object region in the current frame, and O ^ t + 1 , m denotes the projected region in the next frame from the object O t , m based on the optical flow. Furthermore, O t + 1 represents the detected object region in the next frame, and | · | represents the number of total pixels within the region. In this study, the value of the threshold λ is set to 0.3.
As shown in Figure 4, there may exist some false positives in the background regions. The proposed false positives can effectively eliminate the false positives.
Figure 4. The detected results before (top row) and after (bottom row) false positives rejection (the red color denotes the detected objects).

4. Experiment

To evaluate the performance of the proposed method, this study conducted experiments by employing the publicly available datasets moseg_dataset [], BMS (Background Motion Subtraction) [], and DAVIS (Densely Annotated Video Segmentation) []. The publicly available datasets comprise multiple image sequences, which contain moving objects in different complex backgrounds captured by moving cameras with different movements. The quantitative and qualitative comparisons were made with compared methods [,,,]. The codes of algorithms [,,,] were all downloaded from the authors’ homepage. The parameters suggested by the respective authors were incorporated into the compared methods. This paper calculated the optical flow in [,,] and the proposed method by employing the same algorithm []. Furthermore, all the objects detected by the proposed algorithm were obtained by the same parameters, as described earlier.

4.1. Qualitative Comparison

Figure 5 illustrates some detected results generated from different sequences of the dataset moseg_dataset []. As shown in Figure 5c, the detected objects by [] lost the shape information, which can only be used to determine the approximate position of the objects. The detected results by [] have some false positives in the background region because of the assumption that it is approximating the background by a plane. Some parts of the moving objects are missed in the results by [], which loses the complete shape of the objects. The proposed method detects the moving objects much more completely and with less false positives.
Figure 5. Samples of the detection results using the public dataset moseg_dataset []. (a) are the original images and (b) are the corresponding ground-truthed images; (cf) are the results of methods in [,,,], respectively; and (g) images from the proposed method.
Figure 6 illustrates some sample results on the dataset BMS [] obtained by methods [,,,] and the proposed method. Seq2 was divided into three subsequences because the algorithm in [,] cannot process the entire sequence at once due to the memory constraint. As seen from Figure 6, the proposed method can detect the moving objects completely with less false positives.
Figure 6. Some detected results on the public dataset BMS [] of [,,,] and the proposed method.
Some sample results on the dataset DAVIS [] obtained by methods [,,,] and the proposed method are shown in Figure 7. As seen from Figure 7, the proposed method can detect the moving objects completely with less false positives. The detected results by [] tend to have false positives in the background region. A similar problem appeared in the results by [], such as car-shadow, car-turn, and motorbike. This problem does not occur in the detected result by the proposed method because of good adaptability to the change in scene depth.
Figure 7. Some detected results from DAVIS [] by the proposed and compared methods [,,,].

4.2. Quantitative Comparison

In this study, the overlap rate was employed for quantitatively evaluating the accuracy of the detection algorithms. Given the detected object D O and the ground truth G O , the overlap ratio γ is defined as
γ = | D O G O | | D O G O |
The obtained results are summarized in Table 1, wherein figures in boldface font represent the highest overlap rate, and figures in italic font represent the second-highest overlap rate. It can be observed from Table 1 that the proposed method achieves the best performance for all the datasets. The proposed method is comparable to the state-of-the-art algorithms in [,,]. Comparing to the algorithm described in [], the proposed background orientation reconstruction algorithm does not assume that the background is approximated by planes. Therefore, the proposed method is able to adapt to more complex scenarios. The algorithm [] has poor robustness in that the moving object may not be detected for a long period of time, such as Seq1 and Cars2. However, the proposed algorithm is more robust because of good adaptability to the variation of scene depth. Thus, our algorithm usually obtained a relatively high score for the sequences difficult to detect, such as Seq1, Seq2(3), Cars4, Cars9, and motorbike. Furthermore, it should be noted that the algorithm in [] employs many frames in the future to segment the moving object in the current frame. The algorithm in [] also uses complex optimization to estimate the background orientation. The complexity of the algorithm in [] is much higher than that of the proposed algorithm.
Table 1. The overlap rate (%) obtained after applying the methods in [,,,] and the proposed method.
There is one case in which the proposed method cannot work: Because only the orientation of the optical flow is used to calculate the motion difference between the object and background, the object moving in the same direction as the camera cannot be detected by the proposed method. The amplitude of the optical flow can be used as Supplementary Information to improve the proposed method.

4.3. Computational Efficiency

The proposed background orientation reconstruction method can be very efficient due to the coefficient matrix only depending on the resolution ratio of the image; so, it only needs to be solved once for one sequence because of the same size of the images within the same sequence. However, it needs a large amount of memory. Thus, this paper solves the Poisson equation every time for each frame, and the motion saliency enhancement method is also efficient owing to its high parallelism; it takes about 8 ms for processing each frame at a resolution of 480 × 640 by employing an NVIDIA GTX 1080 from Asus, Chongqing, china. The procedure of false positives rejection takes very little time. Table 2 shows the average computation time per frame with a resolution of 480 × 640 measured by an Intel Core i5-6200U, 2.4GHz PC from Lenovo, Beijing, china. The method [] was implemented in C++ and the other four by MATLAB. The time spent on optical flow computation was excluded, which is required for all methods except the one in []. The computational efficiency of the proposed method is superior to that of the algorithms in [,], and the computational efficiency of the proposed method can be improved greatly through adopting a more advanced sparse equation solving method.
Table 2. Time consumption of the proposed and other methods.

5. Conclusions

This paper proposes a novel method for detecting moving objects under a moving camera. This paper reconstructs the background orientation field through Poisson fusion based on the modified gradient and the orientation in image boundaries. The motion saliency map can be obtained by the difference between the original and reconstructed background orientation field. Based on the appearance and motion similarity, the proposed method enhances the motion saliency map through weighted accumulation in the spatial domain. Furthermore, the proposed method incorporates motion continuity in the temporal domain for rejecting false positives. Experimental results based on the publicly available datasets indicate that the proposed method can achieve an excellent performance through qualitative and quantitative comparison.
Since our paper only uses the optical flow orientation, the object moving in the same direction as the camera cannot be detected. In this situation, the amplitude of the optical flow between the moving object and background must be different, or the object is static. In the next work, we will use the amplitude as a check and supplement the scheme, to verify if there are missed objects, and to detect them. On the other hand, the thresholding method for motion saliency mapping is simple, which influences the accuracy of the proposed method. We will look at getting a more precise object from the motion saliency map.

Author Contributions

W.Z., X.S. and Q.Y. designed the research and co-wrote the paper; W.Z. designed the algorithm and programmed the code; W.Z. and X.S. conducted the experiments and analysis of the data; W.Z. and Q.Y. wrote the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

The research was supported by the National Natural Science Foundation of China (11902349).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Five-point interpolation method for solving the Poisson equation in reconstructing the background orientation.
The problem is to solve following optimizing formula:
min θ b Ω | θ b g ^ | 2     w i t h     θ b | Ω = θ | Ω
Solving the above problem, we can obtain the following Poisson equation with Dirichlet boundary conditions:
Δ θ b | Ω = d i v ( g ^ ) | Ω     w i t h     θ b | Ω = θ | Ω
The discrete second-order differential of θ b along the x and y direction is as follows:
( 2 θ i , j b x 2 ) = 1 h 2 ( θ i , j + 1 b + θ i , j 1 b 2 θ i , j b ) + O ( h 2 ) ( 2 θ i , j b y 2 ) = 1 h 2 ( θ i + 1 , j b + θ i 1 , j b 2 θ i , j b ) + O ( h 2 )
Here, h denotes the step of the discretization, which is usually set to 1 in image process.
If we ignore the error term O ( h 2 ) and set the boundary size to 1, the discrete Poisson equation is as follows:
θ i , j + 1 b + θ i , j 1 b + θ i + 1 , j b + θ i 1 , j b 4 θ i , j b = d i v ( g ^ i , j )           2 i r 1 ; 2 j c 1
Here, r and c respectively denote the height and width of the orientation field.
Vectorizing θ b by column priority as
θ T = [ θ 2 , 2   θ 3 , 2 θ r 1 , 2   θ 2 , 3   θ r 1 , c 1 ]
Now, the following equation respect to θ T can be obtained:
A [ ( r 2 ) ( c 2 ) ] × [ ( r 2 ) ( c 2 ) ] · θ = b ( r 2 ) ( c 2 )
where the vector b is depended on the gradient field g ^ and the orientation angle of the pixels at the boundaries; and A is a positive definite symmetric matrix of order ( r 2 ) ( c 2 ) , as follows:
A = [ B I I B I I B I I B ]
Here, I is unit matrix of order ( r 2 ) ; and B is a symmetric matrix of order ( r 2 ) , as follows:
B ( r 2 ) × ( r 2 ) = [ 4 1 1 4 1 1 4 1 1 4 ]
The coefficient matrix A in Equation (A5) is a large sparse matrix, depending only on the size of the image.

References

  1. Royden, C.S.; Moore, K.D. Use of speed cues in the detection of moving objects by moving observers. Vis. Res. 2012, 59, 17–24. [Google Scholar] [CrossRef] [PubMed]
  2. Barnich, O.; Van Droogenbroeck, M. ViBe: A universal background subtraction algorithm for video sequences. IEEE Trans. Image Process. 2010, 20, 1709–1724. [Google Scholar] [CrossRef] [PubMed]
  3. Elgammal, A.; Harwood, D.; Davis, L. Non-parametric model for background subtraction. In Proceedings of the 6th European Conference on Computer Vision, Dublin, Ireland, 26 June–1 July 2000; Springer: Berlin/Heidelberg, Germany, 2000. [Google Scholar]
  4. Yong, H.; Meng, D.; Zuo, W.; Zhang, K. Robust online matrix factorization for dynamic background subtraction. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 1726–1740. [Google Scholar] [CrossRef] [PubMed]
  5. Yi, K.M.; Yun, K.; Kim, S.W.; Chang, H.J.; Choi, J.Y.; Jeong, H. Detection of moving objects with non-stationary cameras in 5.8ms: Bringing motion detection to your mobile device. In Computer Vision & Pattern Recognition Workshops; IEEE: Portland, Oregon, 2013. [Google Scholar]
  6. Wan, Y.; Wang, X.; Hu, H. Automatic moving object segmentation for freely moving cameras. Math. Probl. Eng. 2014, 2014, 1–11. [Google Scholar] [CrossRef]
  7. Wu, M.; Peng, X.; Zhang, Q. Segmenting moving objects from a freely moving camera with an effective segmentation cue. Meas. Sci. Technol. 2011, 22, 25108. [Google Scholar] [CrossRef]
  8. Kurnianggoro, L.; Yu, Y.; Hernandez, D.; Jo, K.-H. Online Background-Subtraction with Motion Compensation for Freely Moving Camera. In International Conference on Intelligent Computing; IEEE: Lanzhou, China, 2016; pp. 569–578. [Google Scholar] [CrossRef]
  9. Odobez, J.M.; Bouthemy, P. Separation of moving regions from background in an image sequence acquired with a mobile camera. In Video Data Compression for Multimedia Computing: Statistically Based and Biologically Inspired Techniques; Springer: Boston, MA, USA, 1997. [Google Scholar] [CrossRef]
  10. Hartley, R.; Zisserman, A. Multiple View Geometry in Computer Vision; Cambridge University Press: Cambridge, UK, 2003. [Google Scholar] [CrossRef]
  11. Zhang, X.; Sain, A.; Qu, Y.; Ge, Y.; Hu, H. Background subtraction based on integration of alternative cues in freely moving camera. IEEE Trans. Circuits Syst. Video Technol. 2019, 29, 1933–1945. [Google Scholar] [CrossRef]
  12. Kim, S.W.; Yun, K.; Yi, K.M.; Kim, S.J.; Choi, J.Y. Detection of moving objects with a moving camera using non-panoramic background model. Mach. Vis. Appl. 2012, 24, 1015–1028. [Google Scholar] [CrossRef]
  13. Narayana, M.; Hanson, A.; Learned-Miller, E. Coherent motion segmentation in moving camera videos using optical flow orientations. In Proceedings of the 2013 IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 1577–1584. [Google Scholar]
  14. Pérez, P.; Gangnet, M.; Blake, A. Poisson image editing. ACM Trans. Graphics 2003, 22, 313. [Google Scholar] [CrossRef]
  15. Yazdi, M.; Bouwmans, T. New trends on moving object detection in video images captured by a moving camera: A survey. Comput. Sci. Rev. 2018, 28, 157–177. [Google Scholar] [CrossRef]
  16. Chapel, M.-N.; Bouwmans, T. Moving objects detection with a moving camera: A comprehensive review. arXiv 2020, arXiv:2001.05238. Available online: https://arxiv.org/abs/2001.05238 (accessed on 30 May 2020).
  17. Kim, J.; Wang, X.; Wang, H.; Zhu, C.; Kim, D. Fast moving object detection with non-stationary background. Multimedia Tools Appl. 2012, 67, 311–335. [Google Scholar] [CrossRef]
  18. Sheikh, Y.; Javed, O.; Kanade, T. Background subtraction for freely moving cameras. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009; pp. 1219–1225. [Google Scholar]
  19. Brox, T.; Malik, J. Object segmentation by long term analysis of point trajectories. In Proceedings of the 11th European Conference on Computer Vision, Heraklion, Greece, 5–11 September 2010; Springer: Heraklion Greece, 2010; pp. 282–295. [Google Scholar]
  20. Ochs, P.; Malik, J.; Brox, T. Segmentation of moving objects by long term video analysis. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 36, 1187–1200. [Google Scholar] [CrossRef] [PubMed]
  21. Nonaka, Y.; Shimada, A.; Nagahara, H.; Taniguchi, R.-I. Real-time foreground segmentation from moving camera based on case-based trajectory classification. In Proceedings of the 2013 2nd IAPR Asian Conference on Pattern Recognition, Okinawa, Japan, 5–8 November 2013; pp. 808–812. [Google Scholar]
  22. Elqursh, A.; Elgammal, A. Online moving camera background subtraction. Appl. Evol. Comput. 2012, 7577, 228–241. [Google Scholar]
  23. Bugeau, A.; Perez, P. Detection and segmentation of moving objects in complex scenes. Comput. Vis. Image Underst. 2009, 113, 459–476. [Google Scholar] [CrossRef]
  24. Gao, Z.; Tang, W.; He, L. Moving object detection with moving camera based on motion saliency. J. Comput. Appl. 2016, 36, 1692–1698. [Google Scholar]
  25. Huang, J.; Zou, W.; Zhu, J.; Zhu, Z. Optical flow based real-time moving object detection in unconstrained scenes. arXiv 2018, arXiv:1807.04890. Available online: https://arxiv.org/abs/1807.04890 (accessed on 30 May 2020).
  26. Sajid, H.; Cheung, S.-C.S.; Jacobs, N. Motion and appearance based background subtraction for freely moving cameras. Signal. Process. Image Commun. 2019, 75, 11–21. [Google Scholar] [CrossRef]
  27. Zhou, D.; Frémont, V.; Quost, B.; Dai, Y.; Li, H. Moving object detection and segmentation in urban environments from a moving platform. Image Vis. Comput. 2017, 68, 76–87. [Google Scholar] [CrossRef]
  28. Namdev, R.K.; Kundu, A.; Krishna, K.M.; Jawahar, C.V. Motion segmentation of multiple objects from a freely moving monocular camera. In Proceedings of the 2012 IEEE International Conference on Robotics and Automation, Saint Paul, MN, USA, 14–18 May 2012; pp. 4092–4099. [Google Scholar]
  29. Bideau, P.; Learned-Miller, E. It’s Moving! A probabilistic model for causal motion segmentation in moving camera videos. In Proceedings of the 14th European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016. [Google Scholar]
  30. Papazoglou, A.; Ferrari, V. Fast object segmentation in unconstrained video. In Proceedings of the 2013 IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 1777–1784. [Google Scholar]
  31. Chen, T.; Lu, S. Object-level motion detection from moving cameras. IEEE Trans. Circuits Syst. Video Technol. 2016, 27, 2333–2343. [Google Scholar] [CrossRef]
  32. Wu, Y.; He, X.; Nguyen, T.Q. Moving object detection with a freely moving camera via background motion subtraction. IEEE Trans. Circuits Syst. Video Technol. 2015, 27, 236–248. [Google Scholar] [CrossRef]
  33. Zhu, Y.; Elgammal, A. A multilayer-based framework for online background subtraction with freely moving cameras. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 5142–5151. [Google Scholar]
  34. Trefethen, L.N.; Bau, D. Numerical Linear Algebra; SIAM: Philadelphia, PA, USA, 1997. [Google Scholar]
  35. Hu, J.; Tang, H. Numerical Method of Differential Equation; Science Press: Beijing, China, 2007. [Google Scholar]
  36. Perazzi, F.; Pont-Tuset, J.; McWilliams, B.; Van Gool, L.; Gross, M.; Sorkine-Hornung, A. A benchmark dataset and evaluation methodology for video object segmentation. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 724–732. [Google Scholar]
  37. Sun, D.; Roth, S.; Black, M. Secrets of optical flow estimation and their principles. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2010, San Francisco, CA, USA, 13–18 June 2010. [Google Scholar]

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.