FreeViBe+: An Enhanced Method for Moving Target Separation
Abstract
1. Introduction
2. Related Work
3. FreeViBe+: An Enhanced Foreground–Background S-Measure Algorithm
3.1. Principles
3.1.1. Ghost Elimination Based on Multi-Frame Background Modeling
3.1.2. Shadow Detection and Removal Based on HSV and Luminance
3.1.3. Void Filling by Integrating the GrabCut Algorithm
3.2. Implementation
- Moving Target Detection Based on ViBe: First, shadow detection based on HSV and luminance is executed to remove potential shadows, as shown in Step in Figure 1: HSV color space identification. The shadow removal module successfully eliminates the elongated shadows attached to the moving target, which were initially misclassified as foreground. Then, foreground target recognition using the ViBe algorithm is carried out.
- Target Detection Based on GrabCut Image Segmentation: Taking the moving target as the region of interest, the Gaussian Mixture Model (GMM) and K-means clustering are iteratively applied to update the results, yielding a relatively complete set of foreground region pixels. The hole-filling module effectively reconstructs the complete silhouette of the object, which was fragmented due to uniform texture.
- Finally, the S-measure is used as the criterion for weight adjustment, and the results obtained from the ViBe algorithm and the GrabCut algorithm are fused to produce the final foreground moving target. This step corresponds to Step in Figure 1: result image fusion.
| Algorithm 1 FreeViBe+: Video-based Moving Object Segmentation Algorithm | |
| 1: | Input: video_file, params |
| 2: | Output: Final segmented frames |
| 3: | |
| 4: | function FreeViBe+() |
| 5: | |
| 6: | while video_file has next frame do |
| 7: | |
| 8: | |
| 9: | append to |
| 10: | end while |
| 11: | |
| 12: | |
| 13 | |
| 14: | |
| 15: | |
| 16: | |
| 17: | for to k do |
| 18: | |
| 19: | append to M |
| 20: | end for |
| 21: | |
| 22: | |
| 23: | |
| 24: | for in P do |
| 25: | |
| 26: | |
| 27: | |
| 28: | |
| 29: | append to G |
| 30: | end for |
| 31: | |
| 32: | |
| 33: | for in P do |
| 34: | |
| 35: | |
| 36: | |
| 37: | |
| 38: | |
| 39: | append to C |
| 40: | end for |
| 41: | |
| 42: | |
| 43: | for to length(P) do |
| 44: | |
| 45: | |
| 46: | |
| 47: | |
| 48: | |
| 49: | append to R |
| 50: | end for |
| 51: | return R |
| 52: | end function |
- initialize_background (P, t) uses frame differencing on the first t frames to initialize the background .
- generate_background_variant(, R, , ) generates a variant of the background with random perturbations (controlled by R, , and ).
- detect_shadows(, ) identifies shadow regions in using HSV color space thresholds.
- initialize_ViBe(M) initializes the ViBe model with background images M.
- ViBe_detect(ViBe_model, -no-shadow) applies ViBe to detect moving foreground in -no-shadow.
- GrabCut_segment(, initial_rect) performs GrabCut segmentation on with an initial bounding box.
- fill_holes(binary_foreground) fills holes in the binary foreground mask.
- calculate_S_metric(, ) computes a similarity metric S between the ViBe foreground and GrabCut result .
- fuse_images(, , w) fuses and using weight w (e.g., weighted average or pixel-wise selection).
4. Experiments and Analysis
4.1. Dataset Information
4.2. Experimental Setup and Results
4.3. Analysis and Discussion
4.3.1. Ablation Analysis
4.3.2. Sensitivity Analysis
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Bouwmans, T.; Jayed, S.; Sultana, M.; Jung, S.K. Deep Neural Network Concepts for Background Subtraction: A Systematic Review and Comparative Evaluation. Neural Netw. 2019, 117, 8–66. [Google Scholar] [CrossRef]
- Tokmakov, P.; Schmid, C.; Alahari, K. Learning to Segment Moving Objects. Int. J. Comput. Vis. 2019, 127, 282–301. [Google Scholar] [CrossRef]
- Bou, X.; Ehret, T.; Facciolo, G.; Morel, J.M.; von Gioi, R.G. Reviewing ViBe, a Popular Background Subtraction Algorithm for Real-Time Applications. Image Process. Line 2022, 12, 527–549. [Google Scholar] [CrossRef]
- Panda, D.K.; Meher, S. Detection of Moving Objects Using Fuzzy Color Difference Histogram Based Background Subtraction. IEEE Signal Process. Lett. 2016, 23, 45–49. [Google Scholar] [CrossRef]
- Barnich, O.; Droogenbroeck, M.V. ViBe: A Universal Background Subtraction Algorithm for Video Sequences. IEEE Trans. Image Process. 2011, 20, 1709–1724. [Google Scholar] [CrossRef] [PubMed]
- Chen, Z.; Wang, R.L.; Zhang, Z.; Wang, H.B.; Xu, L.Z. Background-foreground Interaction for Moving Object Detection in Dynamic Scenes. Inf. Sci. 2019, 483, 65–81. [Google Scholar] [CrossRef]
- Aliouat, A.; Kouadria, N.; Maimour, M.; Harize, S. EVBS-CAT: Enhanced Video Background Subtraction with a Controlled Adaptive Threshold for Constrained Wireless Video Surveillance. J. Real-Time Image Process. 2024, 21, 9. [Google Scholar] [CrossRef]
- Xie, Y.; Gu, S.H.; Liu, Y.; Zuo, W.M.; Zhang, W.S.; Zhang, L. Weighted Schatten p-Norm Minimization for Image Denoising and Background Subtraction. IEEE Trans. Image Process. 2016, 25, 4842–4857. [Google Scholar] [CrossRef]
- Javed, S.; Mahmood, A.; Al-Maadeed, S.; Bouwmans, T.; Jung, S.K. Moving Object Detection in Complex Scene Using Spatiotemporal Structured-Sparse RPCA. IEEE Trans. Image Process. 2019, 28, 1007–1022. [Google Scholar] [CrossRef]
- Dewan, P.; Nivedita, N.; Kumar, R. A Novel Approach for Detection of Moving Objects in Complex Scenes Using Fuzzy Colour Difference Histogram. Int. J. Softw. Innov. 2021, 9, 81–101. [Google Scholar] [CrossRef]
- Zhang, Z.X.; Li, X.P.; Li, Q. SAEN-BGS: Energy-efficient Spiking Autoencoder Network for Background Subtraction. Pattern Recognit. 2026, 169, 111792. [Google Scholar] [CrossRef]
- Babaee, M.; Dinh, D.T.; Rigoll, G. A Deep Convolutional Neural Network for Video Sequence Background Subtraction. Pattern Recognit. 2018, 76, 635–649. [Google Scholar] [CrossRef]
- Houhou, I.; Zitouni, A.; Ruichek, Y.; Bekhouche, S.E.; Kas, M.; Taleb-Ahmed, A. RGBD Deep Multi-scale Network for Background Subtraction. Int. J. Multimed. Inf. Retr. 2022, 11, 395–407. [Google Scholar] [CrossRef]
- Yang, Y.Z.; Ruan, J.H.; Zhang, Y.Q.; Cheng, X.; Zhang, Z.; Xie, G.J. STPNet: A Spatial-Temporal Propagation Network for Background Subtraction. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 2145–2157. [Google Scholar] [CrossRef]
- Zhao, C.Q.; Hu, K.K.; Basu, A. Universal Background Subtraction Based on Arithmetic Distribution Neural Network. IEEE Trans. Image Process. 2022, 31, 2934–2949. [Google Scholar] [CrossRef]
- Dai, Y.; Yang, L. Background Subtraction for Video Sequence Using Deep Neural Network. Multimed. Tools Appl. 2024, 83, 82281–82302. [Google Scholar] [CrossRef]
- Xiong, J.; Wu, J.; Tang, M.; Xiong, P.W.; Huang, Y.S.; Guo, H. Combining YOLO and Background Subtraction for Small Dynamic Target Detection. Vis. Comput. 2025, 41, 481–490. [Google Scholar] [CrossRef]
- Cucchiara, R.; Grana, C.; Piccardi, M.; Prati, A. Detecting Moving Objects, Ghosts, and Shadows in Video Streams. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 1337–1342. [Google Scholar] [CrossRef]
- Barnich, O.; Droogenbroeck, M.V. ViBe: A Powerful Random Technique to Estimate the Background in Video Sequences. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Taipei, Taiwan, 19–24 April 2009; International Conference on Acoustics Speech and Signal Processing ICASSP. pp. 945–948. [Google Scholar]
- Cheng, K.Y.; Hui, K.F.; Zhan, Y.Z.; Qi, M. A Novel Improved ViBe Algorithm to Accelerate the Ghost Suppression. In Proceedings of the 12th International Conference on Natural Computation, Fuzzy Systems and Knowledge Discovery (ICNC-FSKD 2016), Changsha, China, 13–15 August 2016; pp. 1692–1698. [Google Scholar]
- Ye, Y.; Mingwei, C.; Feng, Y. EVibe: An Improved Vibe Algorithm for Detecting Moving Objects. Chin. J. Sci. Instrum. 2014, 35, 924–931. [Google Scholar]
- Ma, Y.J.; Chen, M.L.; Liu, P.P.; Duan, R.G.; Ma, Y.T. ViBe Algorithm-Based Ghost Suppression Method. Laser Optoelectron. Prog. 2020, 57, 105–112. [Google Scholar]
- Rother, C.; Kolmogorov, V.; Blake, A. GrabCut - Interactive Foreground Extraction Using Iterated Graph Cuts. Acm Trans. Graph. 2004, 23, 309–314. [Google Scholar] [CrossRef]
- Fan, D.P.; Cheng, M.M.; Liu, Y.; Li, T.; Borji, A. Structure-measure: A New Way to Evaluate Foreground Maps. In Proceedings of the 16th IEEE International Conference on Computer Vision (ICCV 2017), Venice, Italy, 22–29 October 2017; IEEE International Conference on Computer Vision. pp. 4558–4567. [Google Scholar]
- Cheng, M.M.; Fan, D.P. Structure-Measure: A New Way to Evaluate Foreground Maps. Int. J. Comput. Vis. 2021, 129, 2622–2638. [Google Scholar] [CrossRef]
- Soomro, K.; Zamir, A.; Shah, M. UCF101: A Dataset of 101 Human Actions Classes From Videos in The Wild. arXiv 2012, arXiv:1212.0402. [Google Scholar] [CrossRef]
- Blank, M.; Gorelick, L.; Shechtman, E.; Irani, M.; Basri, R. Actions as Space-Time Shapes. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 2247–2253. [Google Scholar] [CrossRef] [PubMed]
- Sobral, A.; Vacavant, A. A comprehensive review of background subtraction algorithms evaluated with synthetic and real videos. Comput. Vis. Image Underst. 2014, 122, 4–21. [Google Scholar] [CrossRef]




| Dataset | Methods | FD | GMM | ViBe | FreeViBe+ | DeepLabv3+ | ST-Former |
|---|---|---|---|---|---|---|---|
| Weizmann | Precision | 82 ± 3 | 88 ± 2 | 94 ± 1 | 96 ± 1 | 98.5 ± 0.4 | 99 ± 0.3 |
| Weizmann | Recall | 78 ± 4 | 85 ± 3 | 92 ± 2 | 95 ± 1 | 97.5 ± 0.7 | 98 ± 0.5 |
| Weizmann | F-measure | 80 ± 3 | 86 ± 2 | 93 ± 1 | 97 ± 1 | 98 ± 0.5 | 98.5 ± 0.4 |
| UCF101 | Precision | 65 ± 5 | 72 ± 4 | 85 ± 3 | 89 ± 3 | 96 ± 0.8 | 97 ± 0.6 |
| UCF101 | Recall | 58 ± 6 | 68 ± 5 | 81 ± 4 | 85 ± 3 | 95 ± 1 | 96 ± 0.8 |
| UCF101 | F-measure | 61 ± 5 | 70 ± 4 | 83 ± 3 | 88 ± 3 | 95.5 ± 0.9 | 96.5 ± 0.7 |
| Dataset | Finite Difference | GMM | ViBe | FreeViBe+ (Ours) |
|---|---|---|---|---|
| UCF101 Jump1 | 66.62 | 74.06 | 71.18 | 82.21 |
| UCF sports video1 | 68.93 | 71.29 | 73.18 | 80.91 |
| Weizmann Walk1 | 72.28 | 83.69 | 66.32 | 87.75 |
| ViBe original data | 81.90 | 82.11 | 71.43 | 89.51 |
| Method Configuration | Weizmann (F-Measure) | UCF101 (F-Measure) |
|---|---|---|
| Baseline: ViBe | 93 ± 1 | 83 ± 3 |
| +Multi-frame Modeling (t = 3) | 94.5 ± 1.5 (+1.5) | 84.5 ± 2 (+1.5) |
| +Shadow Removal | 95.5 ± 1.5 (+1.0) | 86 ± 2 (+1.5) |
| +GrabCut Refinement: FreeViBe+ | 97 ± 1 (+1.5) | 88 ± 3 (+2.0) |
| Parameter | Tested Range | F-Measure Range | Observation |
|---|---|---|---|
| K | 10–30 | 85.5–89.3% | Performance is stable. K = 18 provides a good balance between model complexity and accuracy. Smaller models (K < 15) lead to a slight drop, while larger models (K > 22) offer negligible improvement. |
| 0.001–0.1 | 86.0–89.5% | The method is relatively sensitive. Very low values ( < 0.005) cause slow adaptation, while high values ( > 0.05) lead to increased noise. The chosen = 0.01 lies within a stable, high-performance plateau. | |
| 3–20 | 87.8–89.2% | Performance is robust across a wide range. A lower threshold ( < 5) is too strict, misclassifying some background as foreground. A higher threshold ( > 15) makes the matching too lenient. = 5 is optimal for our setup. | |
| 0.3–0.8 | 87.5–90.2% | The performance is very stable, confirming the robustness of our fusion strategy. The chosen value of = 0.5 gives equal weight to both feature streams, which yields the best and most balanced result. | |
| t | 1–10 | 87.2–89.4% | Using too few frames (t < 3) results in an insufficiently trained model. Performance plateaus after t = 3, indicating that a small number of frames is sufficient for effective initialization. |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Wu, J.; Zhang, K.; Shen, Y.; Lin, J. FreeViBe+: An Enhanced Method for Moving Target Separation. Information 2025, 16, 1052. https://doi.org/10.3390/info16121052
Wu J, Zhang K, Shen Y, Lin J. FreeViBe+: An Enhanced Method for Moving Target Separation. Information. 2025; 16(12):1052. https://doi.org/10.3390/info16121052
Chicago/Turabian StyleWu, Jianwei, Keju Zhang, Yuhan Shen, and Jiaxiang Lin. 2025. "FreeViBe+: An Enhanced Method for Moving Target Separation" Information 16, no. 12: 1052. https://doi.org/10.3390/info16121052
APA StyleWu, J., Zhang, K., Shen, Y., & Lin, J. (2025). FreeViBe+: An Enhanced Method for Moving Target Separation. Information, 16(12), 1052. https://doi.org/10.3390/info16121052

