Rolling 3D Laplacian Pyramid Video Fusion
AbstractIn this paper, we present a novel algorithm for video fusion of multi-sensor sequences applicable to real-time night vision systems. We employ the Laplacian pyramid fusion of a block of successive frames to add temporal robustness to the fused result. For the fusion rule, we first group high and low frequency levels of the decomposed frames in the block from both input sensor sequences. Then, we define local space-time energy measure to guide the selection based fusion process in a manner that achieves spatio-temporal stability. We demonstrate our approach on several well-known multi-sensor video fusion examples with varying contents and target appearance and show its advantage over conventional video fusion approaches. Computational complexity of the proposed methods is kept low by the use of simple linear filtering that can be easily parallelised for implementation on general-purpose graphics processing units (GPUs). View Full-Text
- Supplementary File 1:
ZIP-Document (ZIP, 1028 KB)
Share & Cite This Article
Pavlović, R.; Petrović, V. Rolling 3D Laplacian Pyramid Video Fusion. Electronics 2019, 8, 447.
Pavlović R, Petrović V. Rolling 3D Laplacian Pyramid Video Fusion. Electronics. 2019; 8(4):447.Chicago/Turabian Style
Pavlović, Rade; Petrović, Vladimir. 2019. "Rolling 3D Laplacian Pyramid Video Fusion." Electronics 8, no. 4: 447.
Note that from the first issue of 2016, MDPI journals use article numbers instead of page numbers. See further details here.