Next Article in Journal
Historical Vltava River Valley–Various Historical Sources within Web Mapping Environment
Next Article in Special Issue
FC-RRT*: An Improved Path Planning Algorithm for UAV in 3D Complex Environment
Previous Article in Journal
Encoding Conversion Algorithm of Quaternary Triangular Mesh
Previous Article in Special Issue
Efficient and High Path Quality Autonomous Exploration and Trajectory Planning of UAV in an Unknown Environment
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Potential Use of Drone Ultra-High-Definition Videos for Detailed 3D City Modeling

by
Bashar Alsadik
1,* and
Yousif Hussein Khalaf
2
1
ITC Faculty, University of Twente, 7514 AE Enschede, The Netherlands
2
Department of Surveying, College of Engineering, University of Baghdad, Baghdad 10071, Iraq
*
Author to whom correspondence should be addressed.
ISPRS Int. J. Geo-Inf. 2022, 11(1), 34; https://doi.org/10.3390/ijgi11010034
Submission received: 28 October 2021 / Revised: 15 December 2021 / Accepted: 26 December 2021 / Published: 3 January 2022
(This article belongs to the Special Issue Unmanned Aerial Systems and Geoinformatics)

Abstract

:
Ongoing developments in video resolution either using consumer-grade or professional cameras has opened opportunities for different applications such as in sports events broadcasting and digital cinematography. In the field of geoinformation science and photogrammetry, image-based 3D city modeling is expected to benefit from this technology development. Highly detailed 3D point clouds with low noise are expected to be produced when using ultra high definition UHD videos (e.g., 4K, 8K). Furthermore, a greater benefit is expected when the UHD videos are captured from the air by consumer-grade or professional drones. To the best of our knowledge, no studies have been published to quantify the expected outputs when using UHD cameras in terms of 3D modeling and point cloud density. In this paper, a quantification is shown about the expected point clouds and orthophotos qualities when using UHD videos from consumer-grade drones and a review of which applications they can be applied in. The results show that an improvement in 3D models of ≅65% relative accuracy and ≅90% in point density can be attained when using 8K video frames compared with HD video frames which will open a wide range of applications and business cases in the near future.

1. Introduction

Currently, most consumer-grade drones (unmanned aerial vehicles, UAVs) are equipped with a camera capable of capturing HD videos of 1920 × 1280 pixels (2K). As shown in different research articles, this image resolution enables the creation of 3D models of degraded quality that is more suitable for virtual reality or virtual tourism but still not enough for 3D documentation applications [1]. Accordingly, this research is aimed to answer the following question: what is the improvement in the derived 3D models we can gain if a drone is equipped with an ultra-high definition UHD 6K and 8K video resolution camera?
Increasingly, the 4K cameras either professional or compact are mounted on the drones to offer the customer high-resolution images with higher quality and details. As an example, DJI [2] and Skydio [3] are equipped with 4K cameras. Examples of current 4K cameras are Panasonic Lumix DMC-GH5 [4], Nikon D850, Canon PowerShot G7X [5], Canon EOS R5 [6], and GoPro 9 Black 4K @ 60 fps 5K@30 fps [7] (Figure 1).
Few researchers have shown the use of 4K videos for 3D modeling applications such as in [1] and still, no research has shown the comparison between the quality of the 3D models when using drone HD videos compared with 4K videos or UHD.
On the other hand, 6K and 8K video resolution cameras are starting to be found in the market but on the professional camera level and are expected to popularize 6K and 8K consumer-grade cameras in the near future. As an example of the UHD cameras, the Blackmagic Pocket Cinema camera has a 6K video image resolution of 6144 × 3456 at 60 fps [2] (Figure 1). Increasingly, consumer-grade drones will be equipped with these UHD cameras and this motivates us to introduce this research to quantify the benefit expected when using such UHD video cameras mounted on drones in terms of the quality of the created 3D models and orthophotos.
Noticeably, the UHD video imaging resolutions (e.g., 4K, 6K, and 8K) have several benefits because they show less noise due to bigger sensor size, are effective for low light conditions, have realistic image quality, capture more details, etc. [3].
8K video frame resolution is currently the highest in the industry of digital television and digital cinematography. As it implies from its name, it is equivalent to two times the resolution (pixels) of 4K images and sixteen times the number of pixels of HD images. This means that it is possible to capture recordings from a farther distance while maintaining the same imaging scale with high-quality results. Currently, few companies produced cameras capable of 8K resolution video capturing (8192 × 4320 pixels) which will have a great improvement in the imaging world and filmmaking [4]. The technical challenge to capture 8K videos is the large memory it needs, for example, 40 min of footage can consume up to 2 terabytes of storage memory [5]. Still, having an 8K camera onboard a drone is unreachable due to the high cost of such cameras and the memory required. Big companies such as Canon and Nikon are working on releasing their first 8k video-capable cameras but are still under development. Filmmakers currently can use 8K high-end heavy cameras such as the RED Helium mounted on drones as shown in Figure 1. However, in a few years, low-cost 8K cameras will be available in the market and can then be mounted on consumer-grade drones. Lately, Autel Robotics released their EVO II drone with a camera capable of recording 8K video at 25 fps and 48 MP still shots [6] which is considered the first drone in the world to have this camera’s high-resolution ability.
To illustrate the pictorial improvement gained when using UHD videos, two objects of a coded target of one squared meter and a nearby parked car are shown in Figure 2 graduating from HD (2K) to 8K. Worth mentioning is the ground sampling distance (GSD) at the 8K image resolution is improved four times compared with the 2K image resolution.

1.1. Video-Based 3D Modeling

Mostly, 3D image-based modeling is based on taking still shot images where an overlap percentage should be preserved. Research shows that 80% for both end lap and side lap is sufficient to create a 3D model and orthophotos out of the drone images [7,8]. This high overlap percentage implies a short baseline configuration which is preferred in the dense reconstruction [7,9].
On the other hand, 3D image-based models can be created using videos in what is sometimes called videogrammetry, which refers to making measurements from video images taken using a camcorder [10]. Basically, a video movie comprises a sequence of image frames captured at a certain recording speed. As an example, if a camera is used to capture a one-minute video at a speed of 30 frames per second (fps), it means that a total of 1800 video frames are recorded.
Video images represent a very short baseline imaging configuration where the point of correspondences between the video frames can be calculated by the so-called feature tracking like by using the Kanade–Lucas–Tomasi (KLT) method [11]. However, sampling the required frames can be applied either at fixed-time intervals (Figure 3a) or using more advanced methods such as the 2D features blurry image filtering [12]. In Figure 3b, two adjacent video frames are shown where one is blurred and one is sharp. In Figure 3c, the blurred image has a reduced number of SIFT keypoints [13] compared with the unblurred image. Filtering and sampling are logical to avoid the redundancy of the data, reduce the processing time, filter out the blurry images for a better 3D model [14], and have a geometrically stronger configuration.
Until now, video-based 3D modeling has not been preferred because of the insufficient resolution and very short baseline. As is known, a short baseline can lead to a small base/height (B/H) ratio which is unwanted because it implies a bad intersection angle and a large depth uncertainty compared with the wide baseline imaging configuration. As mentioned, sampling the video frames at longer time intervals can help to have a wider baseline configuration and decrease data redundancy.
Remarkably, taking videos is more flexible to record than still shots where the streaming continues to the target object without paying attention to the camera shutter speed, optimal waypoint along the flight trajectories, etc.
What can increase or popularize the use of video-based 3D modeling is the increase in the video resolution since the other technical details are already solved such as the structure from motion SfM or image matching, and most state-of-the-art software tools can process such data. Accordingly, UHD videos may replace static imagery as they gather the positive aspects of being high resolution, easy to capture and record, offer a wealth of data, and require less effort for planning. However, sampling and filtering are necessary to ensure cost-effective processing and good quality results. Currently, studies show that an accuracy of ≅1/400 or ≅5 cm can be achieved when using video frames of 640 × 480 pixels which can be improved to 1 cm with a higher resolution in the best case [14,15,16,17]. However, no studies have been applied to quantify the accuracies that can be achieved when capturing UHD videos ranging from 4K up to 8K. Accordingly, in this paper, a study will be conducted (Section 3) through two experiments using a drone equipped with a 2K, 4K, 6K, and 8K camera in a simulated environment.

1.2. UAV Flight Planning

Flight planning is an important step to accomplish a successful UAV flight mission and achieve the mapping project goals and requirements. Several flight planning patterns can be applied depending on the task of the flight and the area of interest. For example, the flight plan for road and powerline mapping is different than the flight plans required for mapping an area of land or a tower. Accordingly, several flight plan patterns are found as shown in Figure 4.
To apply a flight plan [18,19], several imaging parameters should be selected. Overlap percentages between successive images and strips in the forward and side directions should be fixed. Camera parameters such as the focal length and flying height are used to set up a required scale and GSD. Camera shutter speed and the drone flying speed are also carefully set up to avoid imaging motion blur.
Worth mentioning is that more advanced flight planning is under continuous development and has started to be used to accomplish production of more autonomous drones where collision avoidance [20] and simultaneous localization and mapping (SLAM) [21,22] are applied. Skydio [23] and Anafi AI [24] are examples of such semi-autonomous consumer-grade drones currently available in the market.

2. Methodology

The methodology followed in this paper (Figure 5) uses a simulated environment using the Blender tool [25] where it is possible to test the four different video resolutions captured from the same drone at exactly the same flight trajectory. This is motivated by having a fair comparison between the produced models. Accordingly, two open-access 3D models are used: an urban scene [26] and a multi-story building [27] where a drone flight trajectory will be simulated. It is worth mentioning that ground control points (GCPs) will also be placed on the models represented by one squared-meter coded target. Then, after applying the drone missions, the video frames will be rendered and exported to the Metashape software tool [28] after sampling and filtering. Since the frames will be captured at a high rate of 20–30 frames/sec and as mentioned in Section 1, we will apply the frame sampling at fixed-time intervals to ensure the overlap percentage in the range of 80–90% for adequate 3D modeling. However, the blur effect will not be considered in the simulated video frames.
The designed GCPs fixed XYZ coordinates in both experiments will be assigned to the detected coded targets in the Metashape tool. Then, the image orientation will be applied using SfM. The dense reconstruction will be followed to create the point clouds and the point density will be estimated by counting the number of neighbors for each point inside a half-meter radius sphere using Cloud Compare [29].
To check the internal (relative) accuracy, planar patches out of the point clouds are extracted and the root mean squared error (RMSE) is calculated for every drone video capture. To continue the assessment, external (absolute) accuracy is investigated by calculating RMSE to GCPs and checkpoints.
Furthermore, the point cloud is turned into a surface mesh and finally, an orthomosaic is created. Then, the three results of the relative accuracy, density, and orthomosaic quality are evaluated for a final conclusion.

3. Results

Two experiments were applied in an urban environment using advanced simulations of the Blender tool. The four video frame resolutions of 2K, 4K, 6K, and 8K were tested in terms of point cloud density and relative accuracy. It is worth mentioning that both tests are applied using a laptop Dell Intel Core i7-9750H, CPU @ 2.60GHz, GPU Intel UHD Graphics 630 with 16 GB RAM.
1st experiment: urban model.
The first experiment is applied in an urban environment of Launceston city using its freely published model [26] where a flight plan is simulated assuming four different video resolutions as mentioned. The flight plan is selected in a grid trajectory (Figure 4a) around the area of interest (Figure 6).
The flight plan is applied using the following flight parameters:
Focal length = 2.77 mm
Pixel size = 1.6 µm
Sensor dimensions =6.16 × 4.6 mm
Flying height = 80 m
Forward overlap = 60% and side overlap = 40%
As mentioned, the simulations are applied using the Blender tool and the images of every video resolution are rendered at the same designed flight path at 80 m average height and assuming a recording speed of 20 frames per second. The video frames are sampled at regular intervals to end up with 63 video frames and ensure an 80% overlap.
The four video resolutions are processed using the Metashape tool and a dense point cloud is acquired at every image resolution. Figure 7 shows two histograms of the relation between the video resolution and the time consumption and density of points.
To clarify by how much the point cloud densities differ among the four video resolutions, color-coded figures are created showing the point densities at each resolution (Figure 8) where point densities above 180 pts/m2 are colored in red. We considered density average > 180 pts/m2 is sufficient for urban 3D city modeling applications from drones where fine structural details can be modeled [30].
Furthermore, orthomosaics are created for the area using the four video resolutions to test the pictorial quality of these orthoimages and to compare visually the differences between them. Figure 9 shows differences in orthomosaic resolutions and clarity. A summary is shown in Table 1 using the four video resolutions and the pixel size of every created orthomosaic.
Moreover, to evaluate the achieved relative accuracy between the four video resolutions, a planar roof slab of a church building is selected as shown in Figure 10a. Every created point cloud is cropped and then the best plane fitting is applied and residuals are computed. Then, the standard deviation σ to the best-fit plane is computed assuming a Gaussian distribution as shown in Figure 10b where the distance between the points and the best fit planes are also visualized.
For external accuracy assessment, we used nine reference points represented by one squared-meter coded target distributed all over the area where five are used as GCPs and four as checkpoints. Table 2 and Figure 11 illustrate the RMSE achieved in the checkpoints. Noticeably, the coded targets are not detected at the 2K drone video, only two at the 4K video, most of the targets at the 6K video, and the full target set at the 8K video. Reasonably, this is highly related to the video resolution and GSD values.
2nd experiment: tall city building.
The second test is applied for a city building of 70 m height above the ground [27] as shown in Figure 12a. Similar to the first experiment, four drone videos are captured around the building of 2K, 4K, 6K, and 8K, respectively, by simulating a flight of DJI Phantom 4 Pro using the following flight parameters:
Focal length = 8.8 mm
Pixel size = 2.37 µm
Sensor dimensions =13.2×8.8 mm
Imaging distance = 30 m
Flight speed = 5 m/s
Forward overlap = 80% and side overlap = 60%
Accordingly, 277 video frames are used along the flight strips around the building facades as shown in Figure 12b. A video illustrating the flight mission around the building is shared in [31] as a Supplementary Materials. Several GCP targets are placed on the building facades and on the ground shown in Figure 12c to have a correct scaling and orientation (Figure 12d).
The rendered drone images are processed in the Metashape tool for each captured video resolution by applying the image orientation followed by the dense reconstruction to create a dense point cloud of the building (Figure 13a). It is worth mentioning that the acquired point clouds of the building are created using only ¼ full resolution of the frames (medium setup in Metashape tool) to reduce the processing time.
In Figure 14, two histograms are shown to clarify the relationship between the time consumed in creating point clouds and point densities with respect to the recorded video resolution.
As in the first test, an evaluation of the achieved relative accuracy between the four video resolutions is applied. A planar façade patch is selected as shown in Figure 15. Every created point cloud is cropped and then the best plane fitting is applied and residuals are computed. Then, the standard deviation σ to the best-fit plane is computed assuming a Gaussian distribution as shown in Figure 15 where the distances (residuals) between the points and the best fit planes are also visualized where the red and blue colors indicate > ±1 cm errors.
For external accuracy assessment, we used the same GCPs placed around the building facades and saved others as checkpoints. Table 3 and Figure 16 illustrate the RMSE achieved in the checkpoints.
For one facade of the building, orthomosaics are created using the four video resolutions and the size and pixel size of every created orthomosaic is summarized in Table 4.

4. Discussion

Based on the results of the two experiments shown in the previous Section 3 and as summarized in Table 5, several observations are made and clear comparisons between the different video frame derived models are found as follows. Logically, the point density increase was related to the video frame resolution increase which is also connected to the achieved GSD values as shown in Figure 17. This is also observed in the produced orthomosaics (Table 2 and Table 4) where the pixel sizes much decreased and then more details are expected to be seen (large-scale) on the orthomosaic.
What was found to be interesting is the improved internal relative accuracy whenever the video frame resolution increased. The improvement is recorded with up to 50% when using 8K videos compared with the HD videos. This means that the derived point clouds at the UHD videos are less noisy compared with the point clouds derived from HD videos.
The external accuracy in both experiments indicated improvements whenever the video frame resolution increased. More than ten times improvement was found in the first experiment and three times in the second test when using 8K videos compared with the HD videos.
It is worth mentioning that measuring the GCP targets either automatically or manually on the HD images is challenging due to the low resolution or the larger GSD which might cause the degraded external accuracy. A significant improvement in point densities is recorded in both experiments with around a 90% increase when using 8K videos compared with the HD videos (Table 5). It should be noted that the acquired point clouds were created using half and ¼ of the full image resolution in the two experiments, respectively, to reduce the processing time.
UHD video-based 3D modeling requires larger computer memory and entails time consumption as shown in Figure 7 and Figure 14. This is expected to be reduced with the continuous developments in the capabilities of the computers and the increase in offered cloud services and solutions. 3D models created using HD videos are of low quality and cannot be used for projects that require highly detailed and highly accurate results.
Table 5 shows the summary of results achieved in both experiments for better comparison.
The achieved point densities and accuracies are shown when using UHD drone videos to enable several applications. According to [32], applications include but are not limited to the following: engineering surveying, road pavement monitoring, cultural heritage documentation, digital terrain modeling, as-built surveying, and quality control. Furthermore, it is also suitable for building information models (BIM) and CAD, power line clearance, GIS applications, slope stability and landslides, virtual tours, etc.

5. Conclusions

In this paper, the impact of using UHD video cameras (6K and 8K) onboard drones was investigated on the 3D reconstructed city models. Furthermore, these UHD video-based models are compared with the same 3D models produced from the currently used HD and 4K cameras. The results were investigated in two simulated flights applied in urban environments following grid flight paths around or above city buildings as shown in Figure 6 and Figure 12. In both experiments, it was shown that increasing the video resolution not only improved the density but also the internal and external accuracies of the created 3D models. As shown in Table 5, the point density and the reconstruction accuracy were improved up to 90% when using 8K videos compared with the HD videos taken from the same drone. Noticeably, the GSD was improved around four times when the 8K image resolution was used compared with the HD resolution while maintaining the same flying height. This improvement will guarantee high details of the reconstructed 3D models and hence opens a wide range of applications for using drones equipped with 8K video cameras for roadway condition assessment, powerline clearance, cultural heritage restoration and documentation, as-built surveying, etc.
However, it is still a challenge when using the UHD videos where the memory required, the processing power needed for the computations, and the time consumption could be increased by more than 20 times on average. Therefore it is recommended to continue the research to find a solution for big data handling and finding on-the-fly or cloud-based solutions to speed up the data handling and the geoinformation data extraction.
Worth mentioning, the simulated video frames were blur-free since the videos are rendered in typical flight stabilization. This is to be considered for future simulation experiments to mimic reality

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/ijgi11010034/s1, A video illustrating the flight mission of the second experiment and the created dense point clouds of the first experiment are available to download online at https://www.researchgate.net/ (accessed on 28 October 2021) with a DOI: 10.13140/RG.2.2.26138.16324. Video: experiment 2.mp4, Point clouds: 2k.ply, 4k.laz.

Author Contributions

Conceptualization, Bashar Alsadik and Yousif Hussein Khalaf; methodology, Bashar Alsadik; software, Yousif Hussein Khalaf; analysis, Bashar Alsadik; investigation, Bashar Alsadik; resources, Bashar Alsadik; writing—original draft preparation, Bashar Alsadik; writing—review and editing, Bashar Alsadik and Yousif Hussein Khalaf; visualization, Bashar Alsadik; supervision, Bashar Alsadik. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data that support the findings of this study are available on request from the corresponding author.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Teo, T. Video-based point cloud generation using multiple action cameras. Int. Arch. Photogramm. Remote Sens. Spatial Inf. Sci. 2015, XL-4/W5, 55–60. [Google Scholar] [CrossRef] [Green Version]
  2. Blackmagic Pocket Cinema Camera. Available online: https://www.blackmagicdesign.com/nl/products/blackmagicpocketcinemacamera (accessed on 27 February 2021).
  3. Thorpe, L. The Ongoing Evolution in Video-Capture Resolution. Tvtech. Available online: https://www.tvtechnology.com/opinions/the-ongoing-evolution-in-video-capture-resolution (accessed on 27 February 2021).
  4. Panasonic Lumix DC-GH5 Review. Available online: https://www.dpreview.com/reviews/panasonic-lumix-dc-gh5 (accessed on 27 February 2021).
  5. Bennett, R. Best 8K Cameras in 2021. Wondershare Filmora. Available online: https://filmora.wondershare.com/8k/best-8k-cameras.html (accessed on 27 February 2021).
  6. EVO II Series. Autel Robotics. Available online: https://auteldrones.com/pages/evo-ii-collections (accessed on 27 February 2021).
  7. Wenzel, K.; Rothermel, M.; Haala, N.; Fritsch, D. SURE—The IFP Software for Dense Image Matching. University of Stuttgart: Stuttgart, Germany, 2013. [Google Scholar]
  8. Wenzel, K. Flight Planning for City Acquisition at High Resolution. EuroSDR Workshop. Available online: http://www.eurosdr.net/sites/default/files/images/inline/04_nframes-eurosdr_isprs-southampton2015.pdf (accessed on 27 February 2021).
  9. Hairuddin, A.; Azri, S.; Ujang, U.; Cuétara, M.G.; Retortillo, G.M.; Mohd Salleh, S. Development of 3D city model using videogrammetry technique. Int. Arch. Photogramm. Remote Sens. Spatial Inf. Sci. 2019, XLII-4/W16, 221–228. [Google Scholar] [CrossRef] [Green Version]
  10. Singh, S.; Jain, K.; Mandla, V.R. 3D Scene Reconstruction from Video Camera for Virtual 3D City Modeling. Am. J. Eng. Res. 2014, 3, 140–148. [Google Scholar]
  11. Tomasi, C.; Kanade, T. Shape and Motion from Image Streams under Orthography: A Factorization Method. Int. J. Comput. Vis. 1992, 9, 137–154. [Google Scholar] [CrossRef]
  12. Nocerino, E.; Lago, F.; Morabito, D.; Remondino, F.; Porzi, L.; Poiesi, F.; Bulò, S.R.; Chippendale, P.; Locher, A.; Havlena, M.; et al. A smartphone-based 3d pipeline for the creative industry the replicate eu project. ISPRS-Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2017, 42, 535–541. [Google Scholar] [CrossRef] [Green Version]
  13. Lowe, D.G. Distinctive Image Features from Scale-Invariant Keypoints. Int. J. Comput. Vision 2004, 60, 91–110. [Google Scholar] [CrossRef]
  14. Torresani, A.; Remondino, F. Videogrammetry Vs Photogrammetry for Heritage 3D Reconstruction. Int. Arch. Photogramm. Remote Sens. Spatial Inf. Sci. 2019, XLII-2/W15, 1157–1162. [Google Scholar] [CrossRef] [Green Version]
  15. PhotoModeler Online Help. Available online: www.photomodeler.com (accessed on 27 February 2021).
  16. Alsadik, B.; Gerke, M.; Vosselman, G. Efficient use of video for 3d modelling of cultural heritage objects. ISPRS Ann. Photogramm. Remote Sens. Spatial Inf. Sci. 2015, II-3/W4, 1–8. [Google Scholar] [CrossRef] [Green Version]
  17. Ballarin, M.; Balletti, C.; Guerra, F. Action cameras and low-cost aerial vehicles in archaeology. Proc. SPIE-Int. Soc. Opt. Eng. 2015, 9528, 952813. [Google Scholar] [CrossRef]
  18. Federman, A.; Santana Quintero, M.; Kretz, S.; Gregg, J.; Lengies, M.; Ouimet, C.; Laliberte, J. Uav photgrammetric workflows: A best practice guideline. Int. Arch. Photogramm. Remote Sens. Spatial Inf. Sci. 2017, XLII-2/W5, 237–244. [Google Scholar] [CrossRef] [Green Version]
  19. Wolf, P.; DeWitt, B. Elements of Photogrammetry with Applications in GIS, 3rd ed.; McGraw-Hill: New York, NY, USA, 2000; p. 624. [Google Scholar]
  20. Yasin, J.N.; Mohamed, S.A.S.; Haghbayan, M.-H.; Heikkonen, J.; Tenhunen, H.; PLoSila, J. Unmanned Aerial Vehicles (UAVs): Collision Avoidance Systems and Approaches. IEEE Access 2020, 8, 105139–105155. [Google Scholar] [CrossRef]
  21. Karam, S.; Alsadik, B. The Simultaneous Localization and Mapping (SLAM)-An Overview. Surv. Geospat. Eng. J. 2021, 1, 34–45. [Google Scholar] [CrossRef]
  22. Cadena, C.; Carlone, L.; Carrillo, H.; Latif, Y.; Scaramuzza, D.; Neira, J.; Reid, I.; Leonard, J. Simultaneous Localization And Mapping: Present, Future, and the Robust-Perception Age. IEEE Trans. Robot. 2016, 32, 1309–1332. [Google Scholar] [CrossRef] [Green Version]
  23. Skydio Inc. Available online: https://www.skydio.com/ (accessed on 20 April 2021).
  24. Parrot. Anafi Ai—The 4g Robotic Uav. Available online: https://www.parrot.com/en/drones/anafi-ai (accessed on 10 March 2021).
  25. Blender. Available online: http://www.blender.org (accessed on 27 February 2021).
  26. Launceston City 3D Model. Available online: http://s3-ap-southeast-2.amazonaws.com/launceston/atlas/index.html (accessed on 15 January 2021).
  27. Wormgoor, D. Modern Modular City Blender Cycles. Artstation. Available online: https://www.artstation.com/marketplace/p/lXM3/modern-modular-city-blender-cycles?utm_source=artstation&utm_medium=referral&utm_campaign=homepage&utm_term=marketplace (accessed on 5 March 2021).
  28. Agisoft. AgiSoft Metashape. Available online: http://www.agisoft.com/downloads/installer/ (accessed on 22 July 2020).
  29. CloudCompare. CloudCompare: 3D Point Cloud and mesh Processing Software. Available online: https://www.danielgm.net/cc/ (accessed on 20 April 2020).
  30. TOPODOT. Establishing Requirements, Extracting Metrics, and Evaluating Quality of LiDAR Data. Available online: https://blog.topodot.com/establishing-requirements-extracting-metrics-and-evaluating-quality-of-lidar-data/#s4.2 (accessed on 10 October 2021).
  31. Alsadik, B.; Khalaf, Y.H. Potential Use of Drone Ultra-High-Definition Videos for Detailed 3D City Modeling—Data. ResearchGate. Available online: https://www.researchgate.net/publication/357419825_Potential_Use_of_Drone_Ultra-High-Definition_Videos_for_Detailed_3D_City_Modeling_-_data/citations (accessed on 30 January 2021).
  32. Olsen, M.J.; Roe, G.V.; Glennie, C.; Persi, F.; Reedy, M.; Hurwitz, D.; Williams, K.; Tuss, H.; Squellati, A.; Knodler, M. Nchrp 15-44 Guidelines for the Use of Mobile Lidar in Transportation Applications; NCHRP Transportation Research Board of The National Academies: Washington, DC, USA, 2013. [Google Scholar]
Figure 1. Different camera types are capable of taking UHD videos and the EVO II drone capable of taking 8K videos.
Figure 1. Different camera types are capable of taking UHD videos and the EVO II drone capable of taking 8K videos.
Ijgi 11 00034 g001
Figure 2. Illustration of the relation between the increasing drone video resolution and the GSD at the same flying height of 85 m.
Figure 2. Illustration of the relation between the increasing drone video resolution and the GSD at the same flying height of 85 m.
Ijgi 11 00034 g002
Figure 3. (a) The principle of sampling the redundant video frames and blurry images. (b) Blurry image caused by shaking will be filtered out (left) and the adjacent image without blur will be kept (right). (c) Blurry image has a significant reduction in keypoints (left) compared with the sharp image (right).
Figure 3. (a) The principle of sampling the redundant video frames and blurry images. (b) Blurry image caused by shaking will be filtered out (left) and the adjacent image without blur will be kept (right). (c) Blurry image has a significant reduction in keypoints (left) compared with the sharp image (right).
Ijgi 11 00034 g003
Figure 4. Different types of drone flight missions. (a) Grid type. (b) Circular type. (c) Linear type.
Figure 4. Different types of drone flight missions. (a) Grid type. (b) Circular type. (c) Linear type.
Ijgi 11 00034 g004
Figure 5. The workflow of the methodology.
Figure 5. The workflow of the methodology.
Ijgi 11 00034 g005
Figure 6. The Launceston 3D city model shows the area of interest inside the yellow boundary (left) and the grid flight trajectory (right).
Figure 6. The Launceston 3D city model shows the area of interest inside the yellow boundary (left) and the grid flight trajectory (right).
Ijgi 11 00034 g006
Figure 7. First experiment results: (left) the relation between the video frame resolutions and the consumed time for processing at a logarithmic scale; (right) the relation between the video resolution and the point densities.
Figure 7. First experiment results: (left) the relation between the video frame resolutions and the consumed time for processing at a logarithmic scale; (right) the relation between the video resolution and the point densities.
Ijgi 11 00034 g007
Figure 8. (a) The created point clouds using the four video resolutions. (b) The achieved point density using the four video resolutions, red-colored when density > 180 pts/m2.
Figure 8. (a) The created point clouds using the four video resolutions. (b) The achieved point density using the four video resolutions, red-colored when density > 180 pts/m2.
Ijgi 11 00034 g008
Figure 9. The road mark writing at a garage entrance in the four video resolutions 2K, 4K, 6K, and 8K created orthoimages.
Figure 9. The road mark writing at a garage entrance in the four video resolutions 2K, 4K, 6K, and 8K created orthoimages.
Ijgi 11 00034 g009
Figure 10. Checking the internal accuracy of the first test. (a) The selected planar roof (dashed area) for testing the relative accuracy. (b) Planar fitting relative error evaluation at the four video resolutions (red > 1 cm, blue < −1 cm).
Figure 10. Checking the internal accuracy of the first test. (a) The selected planar roof (dashed area) for testing the relative accuracy. (b) Planar fitting relative error evaluation at the four video resolutions (red > 1 cm, blue < −1 cm).
Ijgi 11 00034 g010aIjgi 11 00034 g010b
Figure 11. The top view of the checkpoints and GCPs exaggerated error ellipses visualization of the area in the first test at the four drone video resolutions. The error in Z is represented by ellipse color while XY errors are represented by ellipse shape using the Metashape tool.
Figure 11. The top view of the checkpoints and GCPs exaggerated error ellipses visualization of the area in the first test at the four drone video resolutions. The error in Z is represented by ellipse color while XY errors are represented by ellipse shape using the Metashape tool.
Ijgi 11 00034 g011
Figure 12. (a) The building model. (b) Drone flight trajectory (blue line), (c) GCPs placed on the building facades, and (d) oriented video frames around the building.
Figure 12. (a) The building model. (b) Drone flight trajectory (blue line), (c) GCPs placed on the building facades, and (d) oriented video frames around the building.
Ijgi 11 00034 g012
Figure 13. (a) The point clouds created from the video frames graduating from low to high resolution (left to right). (b) The estimated point densities where the red color indicates > 5000 pts/m2.
Figure 13. (a) The point clouds created from the video frames graduating from low to high resolution (left to right). (b) The estimated point densities where the red color indicates > 5000 pts/m2.
Ijgi 11 00034 g013
Figure 14. (left) The relation between the video frame resolutions and the consumed time for processing using a logarithmic scale. (right) The relation between the video resolution and the point densities of the second experiment.
Figure 14. (left) The relation between the video frame resolutions and the consumed time for processing using a logarithmic scale. (right) The relation between the video resolution and the point densities of the second experiment.
Ijgi 11 00034 g014
Figure 15. Checking the internal accuracy of the second test. (a) The selected planar façade (blue patch) for testing the relative accuracy. (b) Planar fitting error evaluation at the four video resolutions.
Figure 15. Checking the internal accuracy of the second test. (a) The selected planar façade (blue patch) for testing the relative accuracy. (b) Planar fitting error evaluation at the four video resolutions.
Ijgi 11 00034 g015
Figure 16. Top view of the checkpoints and GCPs exaggerated error ellipses visualization of the building facades at the four drone video resolutions. The error in Z is represented by ellipse color while XY errors are represented by ellipse shape using the Metashape tool.
Figure 16. Top view of the checkpoints and GCPs exaggerated error ellipses visualization of the building facades at the four drone video resolutions. The error in Z is represented by ellipse color while XY errors are represented by ellipse shape using the Metashape tool.
Ijgi 11 00034 g016
Figure 17. Sample point cloud (using middle ¼ dense reconstruction) of the second test building at the four indicated video resolutions.
Figure 17. Sample point cloud (using middle ¼ dense reconstruction) of the second test building at the four indicated video resolutions.
Ijgi 11 00034 g017
Table 1. Orthomosaic resolution and pixel size of the first experiment.
Table 1. Orthomosaic resolution and pixel size of the first experiment.
Ijgi 11 00034 i001 Ijgi 11 00034 i002 Ijgi 11 00034 i003 Ijgi 11 00034 i004
2K4K6K8K
Orthomosaic resolution4367 × 37229348 × 737016,805 × 12,80521,327 × 17,659
Orthomosaic pixel size (GSD)7.3 cm3.7 cm2.5 cm1.8 cm
Table 2. Estimated RMSE at the four checkpoints distributed over the area at the four drone video resolutions.
Table 2. Estimated RMSE at the four checkpoints distributed over the area at the four drone video resolutions.
Ijgi 11 00034 i005 Using 2K VideoUsing 4K VideoUsing 6K VideoUsing 8K Video
RMSE X [mm]35.88.16.03.7
RMSE Y [mm]21.94.34.72.1
RMSE Z [mm]34.79.02.70.8
RMSE Total [mm]54.412.98.24.3
Table 3. RMSE estimated at the checkpoints distributed over the building facades at the four drone video resolutions.
Table 3. RMSE estimated at the checkpoints distributed over the building facades at the four drone video resolutions.
Ijgi 11 00034 i006 Using 2K VideoUsing 4K VideoUsing 6K VideoUsing 8K Video
RMSE X [mm]24.83.64.85.6
RMSE Y [mm]6.84.31.61.6
RMSE Z [mm]9.72.71.71.1
RMSE Total [mm]27.56.34.75.9
Table 4. Orthomosaic resolution and pixel size of the second experiment.
Table 4. Orthomosaic resolution and pixel size of the second experiment.
2K4K6K8K
Orthomosaic resolution3588 × 63207288 × 12,33412,638 × 23,36613,525 × 23,988
Orthomosaic pixel size12 mm6 mm4.4 mm3.2 mm
Table 5. Summary of results.
Table 5. Summary of results.
Video ResolutionNumber of Points (million)Density pts/m2Orientation Time
(minute)
Dense Reconstruction Time (minute)Average GSD(cm)
First experiment2K1.442 ± 8≅1.0≅37.3
4K22.0168 ± 33≅1.0≅134.5
6K49.0378 ± 76≅1.5≅282.5
8K88.0674 ± 139≅2.5≅711.8
Second experiment2K3.8318 ± 46≅4≅3.21.24
4K14.51216 ± 142≅4≅120.64
6K30.02555 ± 318≅7≅300.44
8K57.84866 ± 670≅11≅480.32
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Alsadik, B.; Khalaf, Y.H. Potential Use of Drone Ultra-High-Definition Videos for Detailed 3D City Modeling. ISPRS Int. J. Geo-Inf. 2022, 11, 34. https://doi.org/10.3390/ijgi11010034

AMA Style

Alsadik B, Khalaf YH. Potential Use of Drone Ultra-High-Definition Videos for Detailed 3D City Modeling. ISPRS International Journal of Geo-Information. 2022; 11(1):34. https://doi.org/10.3390/ijgi11010034

Chicago/Turabian Style

Alsadik, Bashar, and Yousif Hussein Khalaf. 2022. "Potential Use of Drone Ultra-High-Definition Videos for Detailed 3D City Modeling" ISPRS International Journal of Geo-Information 11, no. 1: 34. https://doi.org/10.3390/ijgi11010034

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop