PRNU-Based Video Source Attribution: Which Frames Are You Using?
Abstract
:1. Introduction
2. Basics of Video Compression
3. Prnu Estimation
Algorithm 1 Fingerprint Generation. |
|
Algorithm 2 Frame Testing. |
|
4. Experiments
4.1. Experimental Setup
- whether the frame is an intra or a predicted coded one;
- whether the video is acquired in presence of EIS or not;
- whether the frame belongs or not to the first GOP.
4.2. Evaluation at Frame Level
4.3. Performance at Video Level
5. Findings and Insights
- Finding #1
- a strategy that could be optimal both in presence of EIS and not is difficult to be achieved. In fact, the literature [21,31] shows that for non-stabilized videos, the optimal solution is to aggregate PRNUs extracted from all frames, or at least all the I frames, according to Equation (1). At the same time, this approach can hardly be applied to the case of stabilized videos. Conversely, working at frame level, [18] improves results for stabilized videos, but it is sub-optimal for non-stabilized videos.
- Finding #2
- from the previous finding, we can derive that a system able to classify whether a video is stabilized or not reliably could lessen the problem. Some techniques exist, such as [20], but if used, the final accuracy is clearly affected by the uncertainty that this tool introduces. Moreover, when this solution is adopted in large-scale scenarios, the complexity that the solution introduces becomes not negligible.
- Finding #3
- the overall performance of source attribution based on PRNU heavily depends on the device model and/or brand.
- Finding #4
- I frames indeed convey a less attenuated and distorted PRNU independently from the presence/absence of EIS and the source brand.
- Finding #5
- In presence of EIS, the first I frame usually provides the most significant PRNU information. Depending on the device, it can be affected or not by EIS.
- Finding #6
- P frames could contain a weakly distorted-attenuated PRNU in both cases. In presence of EIS, we found that P frames within the first GOP generally provide the most reliable PRNU information.
- Finding #7
- When possible, P frames that follow an I frame should be preferred. Except for this, the contribution of P frames does not depend on their position within the GOP.
- Insight #1
- If a reliable system for classifying a video as stabilized or not is provided, it should be adopted in order to perform two different matching strategies (global accumulation vs. frame-level analysis). The final performance should take into consideration the accuracy of the whole system.
- Insight #2
- If the video duration is long enough and with a sufficient number of I frames (a 1-min length video, encoded at 30 fps and with a GOP size of 30 frames, contains at least 60 I frames), the choice of using only I frames is a good trade-off between performance and complexity. The first I frames should be privileged in the analysis.
- Insight #3
- If the video is short and the number of I frames is limited, P frames could be exploited to improve performance. In particular, the first P frames following an I frame are proved to be more suitable than the others.
- Insight #4
- When deciding a threshold for the binary decision, a training with a control dataset should be adopted. The dataset should be as close as possible to the target devices in terms of EIS and video encoding to avoid unexpected classification errors.
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
B frame | Bidirectional predicted frame |
DCT | Discrete Cosine Transform |
DET | Detection Error Trade-off |
EER | Equal Error Rate |
EIS | Electronic Image Stabilization |
GOP | Group of Pictures |
HVS | Human Vision System |
I frame | Intra-coded frame |
MB | Macro Block |
MDR | Missed Detection Rate |
P frame | Predicted frame |
PCE | Peak-to-Correlation Energy |
PRNU | Photo Response Non-Uniformity |
LD | Linear dichroism |
References
- Lukas, J.; Fridrich, J.; Goljan, M. Digital camera identification from sensor pattern noise. IEEE Trans. Inf. Forensics Secur. 2006, 1, 205–214. [Google Scholar] [CrossRef]
- Goljan, M.; Fridrich, J. Camera identification from scaled and cropped images. Secur. Forensics Steganography Watermark. Multimed. Contents X 2008, 6819, 68190E. [Google Scholar]
- Caldelli, R.; Amerini, I.; Picchioni, F.; Innocenti, M. Fast image clustering of unknown source images. In Proceedings of the 2010 IEEE International Workshop on Information Forensics and Security, Seattle, WA, USA, 12–15 December 2010; IEEE: New York, NY, USA, 2010; pp. 1–5. [Google Scholar]
- Marra, F.; Poggi, G.; Sansone, C.; Verdoliva, L. Blind PRNU-based image clustering for source identification. IEEE Trans. Inf. For. Secur. 2017, 12, 2197–2211. [Google Scholar] [CrossRef]
- Goljan, M.; Fridrich, J.; Filler, T. Large scale test of sensor fingerprint camera identification. In Media Forensics and Security; International Society for Optics and Photonics: San Jose, CA, USA, 2009; Volume 7254, p. 72540I. [Google Scholar]
- Amerini, I.; Caldelli, R.; Del Mastio, A.; Di Fuccia, A.; Molinari, C.; Rizzo, A.P. Dealing with video source identification in social networks. Signal Process. Image Commun. 2017, 57, 1–7. [Google Scholar] [CrossRef]
- Rouhi, R.; Bertini, F.; Montesi, D.; Li, C.T. Social network forensics through smartphones and shared images. In Proceedings of the 2019 7th International Workshop on Biometrics and Forensics (IWBF), Cancun, Mexico, 2–3 May 2019; IEEE: New York, NY, USA, 2019; pp. 1–6. [Google Scholar]
- Chen, M.; Fridrich, J.; Goljan, M.; Lukáš, J. Source digital camcorder identification using sensor photo response non-uniformity. In Security, Steganography, and Watermarking of Multimedia Contents IX; International Society for Optics and Photonics: San Jose, CA, USA, 2007; Volume 6505, p. 65051G. [Google Scholar]
- Van Houten, W.; Geradts, Z. Using sensor noise to identify low resolution compressed videos from youtube. In International Workshop on Computational Forensics; Springer: Berlin, Germany, 2009; pp. 104–115. [Google Scholar]
- McCloskey, S. Confidence weighting for sensor fingerprinting. In Proceedings of the 2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Anchorage, AK, USA, 23–28 June 2008; IEEE: New York, NY, USA, 2008; pp. 1–6. [Google Scholar]
- López, R.R.; Orozco, A.L.S.; Villalba, L.J.G. Compression effects and scene details on the source camera identification of digital videos. Expert Syst. Appl. 2021, 170, 114515. [Google Scholar] [CrossRef]
- Chuang, W.H.; Su, H.; Wu, M. Exploring compression effects for improved source camera identification using strongly compressed video. In Proceedings of the 2011 18th IEEE International Conference on Image Processing, Brussels, Belgium, 11–14 September 2011; IEEE: New York, NY, USA, 2011; pp. 1953–1956. [Google Scholar]
- Yang, W.C.; Jiang, J.; Chen, C.H. A fast source camera identification and verification method based on PRNU analysis for use in video forensic investigations. Multimed. Tools Appl. 2021, 80, 6617–6638. [Google Scholar] [CrossRef]
- Kouokam, E.K.; Dirik, A.E. PRNU-based source device attribution for YouTube videos. Dig. Investig. 2019, 29, 91–100. [Google Scholar] [CrossRef] [Green Version]
- Altinisik, E.; Tasdemir, K.; Sencar, H.T. Mitigation of H. 264 and H. 265 video compression for reliable PRNU estimation. IEEE Trans. Inf. For. Secur. 2019, 15, 1557–1571. [Google Scholar] [CrossRef] [Green Version]
- Bellavia, F.; Fanfani, M.; Colombo, C.; Piva, A. Experiencing with electronic image stabilization and PRNU through scene content image registration. Pattern Recognit. Lett. 2021, 145, 8–15. [Google Scholar] [CrossRef]
- Zitova, B.; Flusser, J. Image registration methods: A survey. Image Vis. Comput. 2003, 21, 977–1000. [Google Scholar] [CrossRef] [Green Version]
- Mandelli, S.; Bestagini, P.; Verdoliva, L.; Tubaro, S. Facing device attribution problem for stabilized video sequences. IEEE Trans. Inf. For. Secur. 2019, 15, 14–27. [Google Scholar] [CrossRef] [Green Version]
- Fanfani, M.; Piva, A.; Colombo, C. PRNU registration under scale and rotation transform based on Convolutional Neural Networks. Pattern Recognit. 2021, 124, 108413. [Google Scholar] [CrossRef]
- Taspinar, S.; Mohanty, M.; Memon, N. Source camera attribution using stabilized video. In Proceedings of the 2016 IEEE International Workshop on Information Forensics and Security (WIFS), Abu Dhabi, United Arab Emirates, 4–7 December 2016; IEEE: New York, NY, USA, 2016; pp. 1–6. [Google Scholar]
- Shullani, D.; Fontani, M.; Iuliani, M.; Al Shaya, O.; Piva, A. Vision: A video and image dataset for source identification. EURASIP J. Inf. Secur. 2017, 2017, 15. [Google Scholar] [CrossRef]
- Ferrara, P.; Beslay, L. Robust video source recognition in presence of motion stabilization. In Proceedings of the 2020 8th International Workshop on Biometrics and Forensics (IWBF), Porto, Portugal, 29–30 April 2020; IEEE: New York, NY, USA, 2020; pp. 1–6. [Google Scholar]
- Altinisik, E.; Sencar, H.T. Source camera verification for strongly stabilized videos. IEEE Trans. Inf. For. Secur. 2020, 16, 643–657. [Google Scholar] [CrossRef]
- Marpe, D.; Wiegand, T.; Sullivan, G.J. The H. 264/MPEG4 advanced video coding standard and its applications. IEEE Commun. Mag. 2006, 44, 134–143. [Google Scholar] [CrossRef]
- Wiegand, T.; Sullivan, G.J.; Bjontegaard, G.; Luthra, A. Overview of the H. 264/AVC video coding standard. IEEE Trans. Circ. Syst. Video Technol. 2003, 13, 560–576. [Google Scholar] [CrossRef] [Green Version]
- Sze, V.; Budagavi, M.; Sullivan, G.J. High efficiency video coding (HEVC). In Integrated Circuit and Systems, Algorithms and Architectures; Springer: Berlin, Germany, 2014; Volume 39, p. 40. [Google Scholar]
- Iuliani, M.; Fontani, M.; Shullani, D.; Piva, A. Hybrid reference-based video source identification. Sensors 2019, 19, 649. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Zhang, J.; Jia, J.; Sheng, A.; Hirakawa, K. Pixel binning for high dynamic range color image sensor using square sampling lattice. IEEE Trans. Image Process. 2018, 27, 2229–2241. [Google Scholar] [CrossRef] [PubMed]
- Mihcak, M.K.; Kozintsev, I.; Ramchandran, K. Spatially adaptive statistical modeling of wavelet image coefficients and its application to denoising. In Proceedings of the 1999 IEEE International Conference on Acoustics, Speech, and Signal Processing—Proceedings ICASSP99 (Cat. No. 99CH36258), Phoenix, AZ, USA, 15–19 March 1999; IEEE: New York, NY, USA, 1999; Volume 6, pp. 3253–3256. [Google Scholar]
- Chen, M.; Fridrich, J.; Goljan, M.; Lukáš, J. Determining image origin and integrity using sensor noise. IEEE Trans. Inf. For. Secur. 2008, 3, 74–90. [Google Scholar] [CrossRef] [Green Version]
- Ferrara, P.; Haraksim, R.; Beslay, L. Performance Evaluation of Source Camera Attribution by Using Likelihood Ratio Methods. J. Imag. 2021, 7, 116. [Google Scholar] [CrossRef]
Brand | Model | Device | (%) I Frames | (%) P Frames | on I Frames | on P Frames |
---|---|---|---|---|---|---|
Apple | iPhone 4s | D02 | 50.74 | 12.63 | (6.16, 9.87) | (6.50, 8.68) |
iPhone 5c | D05 | 11.76 | 13.21 | (3.48, 0.92) | (1.11, 1.06) | |
iPhone 6 | D06 | 35.69 | 7.97 | (3.80, 4.45) | (2.95, 3.53) | |
iPhone 4 | D09 | 85.22 | 51.37 | (0.44, 0.00) | (0.55, 0.05) | |
iPhone 4s | D10 | 5.37 | 0.06 | (6.55, 6.47) | (3.69, 4.21) | |
iPhone 5c | D14 | 7.36 | 8.64 | (4.72, 4.92) | (5.92, 9.91) | |
iPhone 6 | D15 | 37.73 | 23.52 | (3.89, 4.35) | (2.60, 3.04) | |
iPhone 5c | D18 | 25.65 | 26.07 | (0.42, 0.61) | (1.19, 1.18) | |
iPhone 6 Plus | D19 | 52.50 | 25.54 | (4.88, 5.42) | (3.70, 4.09) | |
iPad Mini | D20 | 36.48 | 19.85 | (5.93, 7.34) | (3.92, 4.27) | |
iPhone 5 | D29 | 9.62 | 6.20 | (2.01, 2.10) | (2.43, 2.07) | |
iPhone 5 | D34 | 10.46 | 13.20 | (0.92, 2.94) | (1.08, 2.04) | |
Huawei | P9 | D03 | 1.84 | 0.03 | (0.00, 0.38) | (0.00, 0.00) |
P9 Lite | D16 | 0.00 | 0.00 | - | - | |
P8 | D28 | 2.68 | 0.00 | (0.00, 0.23) | - | |
Honor 5c | D30 | 0.00 | 0.00 | - | - | |
Ascend | D33 | 2.64 | 0.04 | (0.00, 0.00) | (0.00, 0.00) | |
Samsung | Galaxy S3 Mini | D01 | 1.20 | 0.04 | (0.28, 0.00) | (0.00, 0.00) |
Galaxy Tab 3 | D08 | 0.00 | 0.00 | - | - | |
Galaxy S3 | D11 | 18.61 | 1.64 | (0.00, 0.00) | (0.15, 0.15) | |
Galaxy Trend Plus | D22 | 0.97 | 0.00 | (0.00, 0.00) | - | |
Galaxy S3 Mini | D26 | 0.00 | 0.00 | - | - | |
Galaxy S5 | D27 | 3.33 | 0.10 | (0.46, 0.50) | (0.50, 0.51) | |
Galaxy S4 Mini | D31 | 1.94 | 0.00 | (0.00, 0.00) | - | |
Galaxy Tab A | D35 | 18.19 | 0.83 | (0.00, 0.46) | (0.52, 0.49) | |
OnePlus | A3000 | D25 | 17.50 | 7.84 | (3.74, 2.69) | (0.79, 0.91) |
A3003 | D32 | 22.50 | 17.34 | (4.78, 4.64) | (2.30, 2.43) | |
LG | D290 | D04 | 0.00 | 0.00 | - | - |
Lenovo | P70A | D07 | 0.14 | 0.03 | (0.00, 0.00) | (0.00, 0.00) |
Sony | Xperia Z1 Compact | D12 | 22.08 | 9.40 | (0.77, 0.76) | (0.75, 0.69) |
Microsoft | Lumia 640 LTE | D17 | 33.33 | 12.08 | (0.45, 0.00) | (0.48, 0.08) |
Wiko | Ridge 4G | D21 | 26.51 | 11.77 | (0.00, 0.00) | (0.04, 0.00) |
Asus | Zenfone2 Laser | D23 | 0.00 | 0.00 | - | - |
Xiaomi | Redmi Note 3 | D24 | 0.97 | 0.02 | (0.53, 0.38) | (0.55, 0.45) |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Ferrara, P.; Iuliani, M.; Piva, A. PRNU-Based Video Source Attribution: Which Frames Are You Using? J. Imaging 2022, 8, 57. https://doi.org/10.3390/jimaging8030057
Ferrara P, Iuliani M, Piva A. PRNU-Based Video Source Attribution: Which Frames Are You Using? Journal of Imaging. 2022; 8(3):57. https://doi.org/10.3390/jimaging8030057
Chicago/Turabian StyleFerrara, Pasquale, Massimo Iuliani, and Alessandro Piva. 2022. "PRNU-Based Video Source Attribution: Which Frames Are You Using?" Journal of Imaging 8, no. 3: 57. https://doi.org/10.3390/jimaging8030057
APA StyleFerrara, P., Iuliani, M., & Piva, A. (2022). PRNU-Based Video Source Attribution: Which Frames Are You Using? Journal of Imaging, 8(3), 57. https://doi.org/10.3390/jimaging8030057