A Mixed Reality Design System for Interior Renovation: Inpainting with 360-Degree Live Streaming and Generative Adversarial Networks after Removal
Abstract
:1. Introduction
2. Literature Review
2.1. Interior Renovation with Mixed Reality
2.2. Generate Virtual Environment Using GANs
3. Proposed MR System with GANs Method
3.1. Overview of the Proposed Methodology
3.2. Real-Time Data Collection
3.3. Background Reconstruction
3.3.1. Panorama Conversion
3.3.2. Texture in the Dynamic Mask Model
3.4. GAN Generation
3.5. Occlusion
3.6. System Integration
4. Implementation
4.1. System Data Flow
4.2. Renovation Target Room
4.3. Operating Environment
4.4. Simulation of Renovation Proposal
4.5. Verification of GANs Generation Quality
5. Results
5.1. Numerical Results
- -
- Under 15 dB: Unacceptable.
- -
- 15–25 dB: The quality might be considered poor, with possible noticeable distortions or artifacts.
- -
- 25–30 dB: Medium quality. Acceptable for some applications but might not be for high-quality needs.
- -
- 30–35 dB: Good quality; acceptable for most applications.
- -
- 35–40 dB: Very good quality.
- -
- 40 dB and above: Excellent quality; almost indistinguishable differences from the original image.
- -
- SSIM = 1: The test image is identical to the reference. Perfect structural similarity.
- -
- 0.8 < SSIM < 1: High similarity between the two images.
- -
- 0.5 < SSIM ≤ 0.8: Moderate similarity. There might be some noticeable distortions, but the overall structure remains somewhat consistent with the reference image.
- -
- 0 < SSIM ≤ 0.5: Low similarity. Significant structural differences or distortions are present.
- -
- SSIM = 0: No structural information is shared between the two images.
5.2. Visualization Results
6. Discussion
7. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Klepeis, N.E.; Nelson, W.C.; Ott, W.R.; Robinson, J.P.; Tsang, A.M.; Switzer, P.; Behar, J.V.; Hern, S.C.; Engelmann, W.H. The National Human Activity Pattern Survey (NHAPS): A resource for assessing exposure to environmental pollutants. J. Expo. Sci. Environ. Epidemiol. 2001, 11, 231–252. [Google Scholar] [CrossRef]
- Kovacic, I.; Summer, M.; Achammer, C. Strategies of building stock renovation for ageing society. J. Clean. Prod. 2015, 88, 349–357. [Google Scholar] [CrossRef]
- Zhu, Y.; Fukuda, T.; Yabuki, N. Slam-based MR with animated CFD for building design simulation. In Proceedings of the 23rd International Conference of the Association for Computed-Aided Architectural Design Research in Asia (CAADRIA), Tsinghua University, School of Architecture, Beijing, China, 17–19 May 2018; pp. 17–19. [Google Scholar] [CrossRef]
- Milgram, P.; Kishino, F. A taxonomy of mixed reality visual displays. IEICE Trans. Inf. Syst. 1994, 77, 1321–1329. [Google Scholar]
- Nakajima, Y.; Mori, S.; Saito, H. Semantic object selection and detection for diminished reality based on slam with viewpoint class. In Proceedings of the 2017 IEEE International Symposium on Mixed and Augmented Reality (ISMAR-Adjunct), Nantes, France, 9–13 October 2017; pp. 338–343. [Google Scholar] [CrossRef]
- Herling, J.; Broll, W. Pixmix: A real-time approach to high-quality diminished reality. In Proceedings of the 2012 IEEE International Symposium on Mixed and Augmented Reality (ISMAR), Atlanta, GA, USA, 5–8 November 2012; pp. 141–150. [Google Scholar] [CrossRef]
- Kawai, N.; Sato, T.; Yokoya, N. Diminished reality considering background structures. In Proceedings of the 2013 IEEE International Symposium on Mixed and Augmented Reality (ISMAR), Adelaide, SA, Australia, 1–4 October 2013; pp. 259–260. [Google Scholar] [CrossRef]
- Sasanuma, H.; Manabe, Y.; Yata, N. Diminishing real objects and adding virtual objects using a RGB-D camera. In Proceedings of the 2016 IEEE International Symposium on Mixed and Augmented Reality (ISMAR-Adjunct), Merida, Mexico, 19–23 September 2016; pp. 117–120. [Google Scholar] [CrossRef]
- Cosco, F.I.; Garre, C.; Bruno, F.; Muzzupappa, M.; Otaduy, M.A. Augmented touch without visual obtrusion. In Proceedings of the 2009 8th IEEE International Symposium on Mixed and Augmented Reality, Orlando, FL, USA, 19–22 October 2009; pp. 99–102. [Google Scholar] [CrossRef]
- Enomoto, A.; Saito, H. Diminished reality using multiple handheld cameras. In Proceedings of the 2007 First ACM/IEEE International Conference on Distributed Smart Cameras, Vienna, Austria, 25–28 September 2007; Volume 7, pp. 130–135. [Google Scholar] [CrossRef]
- Meerits, S.; Saito, H. Real-time diminished reality for dynamic scenes. In Proceedings of the 2015 IEEE International Symposium on Mixed and Augmented Reality Workshops, Fukuoka, Japan, 29 September–3 October 2015; pp. 53–59. [Google Scholar] [CrossRef]
- Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27, 784–794. [Google Scholar] [CrossRef]
- Radford, A.; Metz, L.; Chintala, S. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv 2015, arXiv:1511.06434. [Google Scholar] [CrossRef]
- Zhu, J.; Shen, Y.; Zhao, D.; Zhou, B. In-domain gan inversion for real image editing. In European Conference on Computer Vision; Springer International Publishing: Cham, Switzerland, 2020; pp. 592–608. [Google Scholar] [CrossRef]
- Yang, C.; Shen, Y.; Zhou, B. Semantic hierarchy emerges in deep generative representations for scene synthesis. Int. J. Comput. Vis. 2021, 129, 1451–1466. [Google Scholar] [CrossRef]
- Zhu, Y.; Fukuda, T.; Yabuki, N. Synthesizing 360-degree live streaming for an erased background to study renovation using mixed reality. In Proceedings of the 24th International Conference of the Association for Computed-Aided Architectural Design Research in Asia (CAADRIA), Victoria University of, Wellington, Faculty of Architecture & Design, Wellington, New Zealand, 15–18 April 2019; pp. 71–80. [Google Scholar] [CrossRef]
- Chan, D.W.; Kumaraswamy, M.M. A comparative study of causes of time overruns in Hong Kong construction projects. Int. J. Proj. Manag. 1997, 15, 55–63. [Google Scholar] [CrossRef]
- Watson, A. Digital buildings–Challenges and opportunities. Adv. Eng. Inform. 2011, 25, 573–581. [Google Scholar] [CrossRef]
- Ozdemir, M.A. Virtual reality (VR) and augmented reality (AR) technologies for accessibility and marketing in the tourism industry. In ICT Tools and Applications for Accessible Tourism; IGI Global: Hershey, PA, USA, 2021; pp. 277–301. [Google Scholar] [CrossRef]
- Woksepp, S.; Olofsson, T. Credibility and applicability of virtual reality models in design and construction. Adv. Eng. Inform. 2008, 22, 520–528. [Google Scholar] [CrossRef]
- Lee, J.; Miri, M.; Newberry, M. Immersive Virtual Reality, Tool for Accessible Design: Perceived Usability in an Interior Design Studio Setting. J. Inter. Des. 2023, 48, 242–258. [Google Scholar] [CrossRef]
- Chandler, T.; Cordeil, M.; Czauderna, T.; Dwyer, T.; Glowacki, J.; Goncu, C.; Klapperstueck, M.; Klein, K.; Marriott, K.; Schreiber, E.; et al. Immersive analytics. In Proceedings of the 2015 Big Data Visual Analytics (BDVA), Hobart, TAS, Australia, 22–25 September 2015; pp. 1–8. [Google Scholar] [CrossRef]
- Ohno, N.; Kageyama, A. Scientific visualization of geophysical simulation data by the CAVE VR system with volume rendering. Phys. Earth Planet. Inter. 2007, 163, 305–311. [Google Scholar] [CrossRef]
- Zhang, W.; Han, B.; Hui, P. Jaguar: Low latency mobile augmented reality with flexible tracking. In Proceedings of the 26th ACM International Conference on Multimedia, Seoul, Republic of Korea, 22–26 October 2018; pp. 355–363. [Google Scholar]
- Kan, P.; Kurtic, A.; Radwan, M.; Rodriguez, J.M.L. Automatic interior Design in Augmented Reality Based on hierarchical tree of procedural rules. Electronics 2021, 10, 245. [Google Scholar] [CrossRef]
- Wang, X.; Love, P.E.; Kim, M.J.; Park, C.S.; Sing, C.P.; Hou, L. A conceptual framework for integrating building information modeling with augmented reality. Autom. Constr. 2013, 34, 37–44. [Google Scholar] [CrossRef]
- Mori, S.; Ikeda, S.; Saito, H. A survey of diminished reality: Techniques for visually concealing, eliminating, and seeing through real objects. IPSJ Trans. Comput. Vis. Appl. 2017, 9, 1–14. [Google Scholar] [CrossRef]
- Eskandari, R.; Motamedi, A. Diminished reality in architectural and environmental design: Literature review of techniques, applications, and challenges. In ISARC: Proceedings of the International Symposium on Automation and Robotics in Construction, Dubai, United Arab Emirates, 2–4 November 2021; Dubai, UAE, IAARC Publications 2021; Volume 38, pp. 995–1001. [CrossRef]
- Kido, D.; Fukuda, T.; Yabuki, N. Diminished reality system with real-time object detection using deep learning for onsite landscape simulation during redevelopment. Environ. Model. Softw. 2020, 131, 104759. [Google Scholar] [CrossRef]
- Siltanen, S. Diminished reality for augmented reality interior design. Vis. Comput. 2017, 33, 193–208. [Google Scholar] [CrossRef]
- Liu, M.Y.; Huang, X.; Yu, J.; Wang, T.C.; Mallya, A. Generative adversarial networks for image and video synthesis: Algorithms and applications. Proc. IEEE 2021, 109, 839–862. [Google Scholar] [CrossRef]
- Yun, K.; Lu, T.; Chow, E. Occluded object reconstruction for first responders with augmented reality glasses using conditional generative adversarial networks. In Pattern Recognition and Tracking XXIX; SPIE: Philadelphia, PA, USA, 2018; Volume 10649, pp. 225–231. [Google Scholar] [CrossRef]
- Teo, T.; Lawrence, L.; Lee, G.A.; Billinghurst, M.; Adcock, M. Mixed reality remote collaboration combining 360 video and 3d reconstruction. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, Scotland, UK, 4–9 May 2019; pp. 1–14. [Google Scholar] [CrossRef]
- Li, Q.; Takeuchi, O.; Shishido, H.; Kameda, Y.; Kim, H.; Kitahara, I. Generative image quality improvement in omnidirectional free-viewpoint images and assessments. IIEEJ Trans. Image Electron. Vis. Comput. 2022, 10, 107–119. [Google Scholar] [CrossRef]
- Lin, C.H.; Chang, C.C.; Chen, Y.S.; Juan, D.C.; Wei, W.; Chen, H.T. Coco-GAN: Generation by parts via conditional coordinating. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019. 2019; pp. 4512–4521. [Google Scholar] [CrossRef]
- Herling, J.; Broll, W. Advanced self-contained object removal for realizing real-time diminished reality in unconstrained environments. In Proceedings of the 2010 IEEE International Symposium on Mixed and Augmented Reality, Seoul, Republic of Korea, 13–16 October 2010; pp. 207–212. [Google Scholar] [CrossRef]
- Inoue, K.; Fukuda, T.; Cao, R.; Yabuki, N. Tracking Robustness and Green View Index Estimation of Augmented and Diminished Reality for Environmental Design. In Proceedings of the CAADRIA, Beijing, China, 17–19 May 2018; pp. 339–348. [Google Scholar] [CrossRef]
- Meerits, S.; Saito, H. Visualization of dynamic hidden areas by real-time 3d structure acquistion using RGB-D camera. In Proceedings of the 3D Systems and Applications Conference, Marrakech, Morocco, 17–20 November 2015. [Google Scholar] [CrossRef]
- Bourke, P. Converting to/from Cubemaps. 2020. Available online: https://paulbourke.net/panorama/cubemaps// (accessed on 5 October 2023).
- Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1125–1134. [Google Scholar] [CrossRef]
- THETA Developers’ API. Available online: https://github.com/ricohapi/theta-api-specs (accessed on 5 October 2023).
- HDRIHAVEN. HDRIHaven. Available online: https://polyhaven.com/hdris/indoor (accessed on 5 October 2023).
- Shutterstock. Shutterstock. Available online: https://www.shutterstock.com/ja/search/indoor-panorama (accessed on 5 October 2023).
- Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
- Huynh-Thu, Q.; Ghanbari, M. Scope of validity of PSNR in image/video quality assessment. Electron. Lett. 2008, 44, 800–801. [Google Scholar] [CrossRef]
- Horé, A.; Ziou, D. Is there a relationship between peak-signal-to-noise ratio and structural similarity index measure? IET Image Process. 2013, 7, 12–24. [Google Scholar] [CrossRef]
- Jing, J.; Deng, X.; Xu, M.; Wang, J.; Guan, Z. HiNet: Deep image hiding by invertible network. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 4733–4742. [Google Scholar] [CrossRef]
- Cavigelli, L.; Hager, P.; Benini, L. CAS-CNN: A deep convolutional neural network for image compression artifact suppression. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 752–759. [Google Scholar] [CrossRef]
- Franchi, V.; Ntagiou, E. Augmentation of a virtual reality environment using generative adversarial networks. In Proceedings of the 2021 IEEE International Conference on Artificial Intelligence and Virtual Reality (AIVR), Taichung, Taiwan, 15–17 November 2021; pp. 219–223. [Google Scholar] [CrossRef]
ITEM | PERFORMANCE |
---|---|
OS | Windows 10 Enterprise 64-bit |
CPU | Intel Core i5 7500 @ 3.40 GHz |
RAM | 16.0 GB Dual-Channel DDR4 @ 2400 MHz |
MOTHERBOARD | H270-PLUS |
GPU | NVIDIA GeForce GTX 1060 6 G |
ITEM | PERFORMANCE |
---|---|
OS | Ubuntu 16.04 |
CPU | Intel Core i7 7700K @ 4.2 GHz |
RAM | 16.0 GB Dual-Channel DDR4 @ 2400 MHz |
MOTHERBOARD | Z270-K |
GPU | NVIDIA GeForce GTX 2070s |
PACKAGE | VERSION |
---|---|
CUDA TOOLKIT VERSION | CUDA 10.0.130_410.48 |
LINUX X86_64 DRIVER VERSION | NVIDIA driver 410.78 |
CUDNN | 7.4.2.24 |
ANACONDA3 | 2021.05 |
PYTORCH | 1.2.0 |
PYTORCH TORCHVISION | 0.4.0 |
IMAGE NUMBER | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
---|---|---|---|---|---|---|---|---|---|---|
PSNR (dB) | 15.12 | 22.42 | 21.52 | 20.95 | 21.62 | 23.41 | 24.29 | 20.12 | 18.57 | 21.49 |
SSIM | 0.7332 | 0.9265 | 0.8546 | 0.8956 | 0.8103 | 0.9187 | 0.9036 | 0.8452 | 0.8419 | 0.8516 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zhu, Y.; Fukuda, T.; Yabuki, N. A Mixed Reality Design System for Interior Renovation: Inpainting with 360-Degree Live Streaming and Generative Adversarial Networks after Removal. Technologies 2024, 12, 9. https://doi.org/10.3390/technologies12010009
Zhu Y, Fukuda T, Yabuki N. A Mixed Reality Design System for Interior Renovation: Inpainting with 360-Degree Live Streaming and Generative Adversarial Networks after Removal. Technologies. 2024; 12(1):9. https://doi.org/10.3390/technologies12010009
Chicago/Turabian StyleZhu, Yuehan, Tomohiro Fukuda, and Nobuyoshi Yabuki. 2024. "A Mixed Reality Design System for Interior Renovation: Inpainting with 360-Degree Live Streaming and Generative Adversarial Networks after Removal" Technologies 12, no. 1: 9. https://doi.org/10.3390/technologies12010009
APA StyleZhu, Y., Fukuda, T., & Yabuki, N. (2024). A Mixed Reality Design System for Interior Renovation: Inpainting with 360-Degree Live Streaming and Generative Adversarial Networks after Removal. Technologies, 12(1), 9. https://doi.org/10.3390/technologies12010009