Next Article in Journal
Influence of Spatial Inhomogeneity of Detector Temporal Responses on the Spectral Fidelity in Continuous Wave Cavity Ringdown Spectroscopy
Next Article in Special Issue
Assessment of DSMs Using Backpack-Mounted Systems and Drone Techniques to Characterise Ancient Underground Cellars in the Duero Basin (Spain)
Previous Article in Journal
Cooperative Localization Improvement Using Distance Information in Vehicular Ad Hoc Networks
Previous Article in Special Issue
A Monocular SLAM-based Controller for Multirotors with Sensor Faults under Ground Effect
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Variant Accuracy Evaluation of UAV Imaging Surveys: A Case Study on Investment Area

Institute of Geodesy, University of Warmia and Mazury in Olsztyn, 10-719 Olsztyn, Poland
*
Authors to whom correspondence should be addressed.
Sensors 2019, 19(23), 5229; https://doi.org/10.3390/s19235229
Submission received: 28 October 2019 / Revised: 21 November 2019 / Accepted: 26 November 2019 / Published: 28 November 2019

Abstract

:
The main focus of the presented study is a multi-variant accuracy assessment of a photogrammetric 2D and 3D data collection, whose accuracy meets the appropriate technical requirements, based on the block of 858 digital images (4.6 cm ground sample distance) acquired by Trimble® UX5 unmanned aircraft system equipped with Sony NEX-5T compact system camera. All 1418 well-defined ground control and check points were a posteriori measured applying Global Navigation Satellite Systems (GNSS) using the real-time network method. High accuracy of photogrammetric products was obtained by the computations performed according to the proposed methodology, which assumes multi-variant images processing and extended error analysis. The detection of blurred images was preprocessed applying Laplacian operator and Fourier transform implemented in Python using the Open Source Computer Vision library. The data collection was performed in Pix4Dmapper suite supported by additional software: in the bundle block adjustment (results verified using RealityCapure and PhotoScan applications), on the digital surface model (CloudCompare), and georeferenced orthomosaic in GeoTIFF format (AutoCAD Civil 3D). The study proved the high accuracy and significant statistical reliability of unmanned aerial vehicle (UAV) imaging 2D and 3D surveys. The accuracy fulfills Polish and US technical requirements of planimetric and vertical accuracy (root mean square error less than or equal to 0.10 m and 0.05 m).

1. Introduction

Due to significant advances in the research field of digital photogrammetry, computer vision, and unmanned aerial vehicles/unmanned aerial systems/remotely piloted aircraft systems (UAVs/UASs/RPASs) [1,2] construction, there has been an important change in geoinformation acquisition and the workflow of photogrammetric data collection [3]. This has been achieved by essential developments in the UASs components [4], in sensors production [5], and in the image-based surface reconstruction, e.g., structure-from-motion (SfM) [6,7], multi-view stereo (MVS) pipeline [8] and dense image-matching techniques [9].
The commercial off-the-shelf (COTS) software packages, focusing on UAV imagery applications in geomatics, are based on a large number of variously oriented digital images and work in the composed and often an autonomous processing chain [10,11,12]. They allow an automatic computation of spatial orientation of photos and the parameters of camera interior orientation with self-calibration [13] using the bundle-block adjustment (BBA) method, generation of dense point clouds [11], a 3D mesh model and, finally, the digital surface model (DSM) [10,14] and orthomosaic [15].
The spectrum of problems and research tasks in the new area of photogrammetry, the so-called “UAV Photogrammetry”, is very wide [5,9,16,17,18,19,20,21]. The research and development mainly concerns, among other areas: small UAS [22], low-cost hardware solution [23], onboard sensors integration [24], sensors calibration [13,25,26], images orientation and direct georeferencing [27], data fusion obtained from optical sensors in various spectral ranges [24] and LiDAR (Light Detection and Ranging) [28], automation of flight planning [29], automation of digital images processing [30], testing and comparison of software dedicated to UASs [10,11]. An important research area is also testing the accuracy potential [3,31,32,33,34] of new measurement technologies based on UAV imagery, searching for new areas of its practical implementation [35,36,37], civil law regulations regarding the use of RPASs [18,38].
However, in our opinion, the operational use of UAV imagery acquired by means of single-lens reflex or compact system camera instead of medium or large format aerial digital cameras for data collection in order to undertake large-scale basic, thematic [39,40,41,42] and inventory mapping is one of the fundamental issues for the application of UAVs in geomatics and in particular in geodesy. For example, in Germany, the USA and also in Poland, there are no detailed technical guidelines regarding the use of UAV imagery in geodetic surveys. The only criterion for taking photogrammetric data into a geodetic database is the accuracy requirements [43,44,45,46] for the final products.
The use of RPASs in national mapping was described previously in state-of-art by the EuroSDR organization [38], in orthophoto quality assessment performed by Mesas-Carrascosa et al. [15] and Martínez-Carricondo et al. [47], in photogrammetric mapping evaluation by Agüera-Vega et al. [48] and in an investigation of the optimal number of ground control points (GCPs) for DSM by Tonkin and Midgley [49]. This current situation consequently limits the use of RPAS in geodesy. Researchers reported before that the regulation is the most limiting factor that hinders the thorough application of RPASs [18,38,50].
According to the authors’ knowledge, the complex accuracy assessment of the computing pipeline of particular photogrammetric products (bundle-block adjustment with camera self-calibration, dense point cloud, digital surface model and orthomosaic) which we present in this paper, do not exist at present. Current studies concerning the accuracy assessment of UAV photogrammetric products are realized in incomparable technical conditions of projects and with the use of various statistical parameters of results assessment [39,40] relating to the particular stages of the processing [36,41,42,51]. The most commonly used metrical information of accuracy were root mean square errors RMSE (XYZ) on ground control and check points, which characterized bundle block adjustment results. The additional parameters of sensor interior orientation estimated in the self-calibration procedure were described by standard deviations. In addition, the quality and accuracy of DSMs and orthomosaics were analyzed using RMS (Z) and RMS (XY) deviations, respectively.
The issues mentioned above were the inspiration for defining the following main aims of this study:
  • Definition of the computation pipeline methodology for measurement accuracy assessment under the real conditions of a posteriori ground control and check points measurements
  • A multi-variant accuracy assessment of UAV imaging surveys performed in Pix4Dmapper suite [32,52]—a case of small investment area.
  • Investigation of the impact of photos overlap and georeferences from UAV Global Navigation Satellite Systems (GNSS) receiver on the BBA results.
  • Examination of the usefulness of UAV imagery acquired with the use of a medium spatial resolution [1] system camera for the development of full-featured photogrammetric products whose accuracy meets the appropriate technical requirements.
  • Checking the possibilities of replacing geodetic field surveys by the UAV photogrammetric measurements.
The initial approach to the UAV mapping was signaled and presented by us in the publication [53], in which we found the possibility of using UAV imagery for inventory mapping in the case of the industrial estate. In the presented paper, we extend our considerations to the investment area and we discuss a new testing approach and other research results. The tests were conducted using a new input data set that contains a large number of low-altitude images and ground control points, additional ground check points and height ground check points. Also, we used another study methodology, which led us to pose new final results, and extended accuracy assessment (tabular summary of findings, histograms, screens) of respective stages of computation and image processing (multi-variant BBA, DSM and georeferenced orthomosaic generation).
The discussion of the results and the accuracy assessment takes into account the Polish [43], technical requirements (accuracy defined with errors mXY ≤ 0.10 m and mZ ≤ 0.05 m) for the land surveys of planar-elevation measurements of well-defined terrain details and other practical applications.

2. Materials and Methods

2.1. Study Area and Data Acquisition

The study site was an industrial area in the center of Poland in the Stryków municipality (51°52′38″ N, 19°35′21″ E), around 16 km northeast of the city of Łódź. This study site is representative for newly created industrial zones in the agricultural area because it contains different topographic objects, i.e., industrial buildings and their surroundings, access roads, a roundabout, and agricultural lands. In other study cases, the different localization and higher number of terrain objects may appear.
The flight mission design and photo capturing were made by the Geoplan Polish company in Zgierz. A block of archival low-altitude uncalibrated photos acquired on 15 April 2016, at 8:35–9:08 local time as photographic documentation of the industrial investment [53], was used in the case study. The images were acquired by Trimble® UX5 UAS equipped with Sony NEX-5T 16 megapixels digital compact system (mirrorless) camera with a CMOS (complementary metal-oxide semiconductor) sensor (size 23.5 × 15.6 mm, pixel count 4912 × 3264, pixel pxy = 4.78 µm) and focal length f = 15.5 mm. Initial camera preferences were defined by shutter priority (1/2000 sec). The lens focus was set to infinity. The light source and white balance were set to Auto. The prevailing weather conditions on the day of flight mission were: mostly sunny sky with some clouds and the west wind speed was equal approximately to 5 m/sec and in gusts to 9 m/sec. Due to this fact, there were apparent drift and rotation (ω, φ, κ) of some images.
The image block was characterized by the following photogrammetric parameters: flight mission area was P = 1.0253 km2, 858 photos in 41 photo strips oriented in direction N-S, crosswise to the long axis of the object, forward and side overlap p% = q% = 85%, projected flight height h = 150 m, ground sample distance (GSD) = 4.6 cm and base-to-height ratio ν = 0.15. The GNSS (single frequency receiver C/A, L1) of Trimble® UX5 UAS flight data were saved in LOG file. The flight path is shown in supplementary files (Figure S2 in Appendix 1–6).
The test field includes 5 artificially signalized GCPs, placed before the flight mission. For study purposes in 2019 the 295 natural, accurately defined ground points (water gate valves, manholes, curbstones) were identified on images, mainly used as check points (ChPs). The distribution of the GCPs and ChPs was a consequence of the field details placement, i.e., industrial buildings and elements of the utilities network. The GCPs and ChPs were distributed more densely in the area which was critical for matching (Figure 1a,b) and more sparsely in the remaining interest area (Figure 1c). Due to the thin sheet metal of the hall’s roofing it was impossible to place GCPs there. For tests we used the previous researchers’ experiences described in [27,49,54,55].
Furthermore, for the verification of the photogrammetric-generated DSM, the 1118 height ground ChPs in the form of dense cross-sections localized on scarps and terrain slopes were surveyed. The direct measurement of ground and height points was realized by applying the GNSS using the real-time network (RTN) method executed by virtual reference stations from the Polish Active Geodetic Network ASG EUPOS (virtual reference station solution). At every point, 10 fixed epochs were averaged in the RTN mode.
We decided to use the RTN solution because according to the ASG EUPOS (http://www.asgeupos.pl/index.php?wpg_type=serv&sub=gen) the advantage of RTN over solutions based on a single reference station RTK (real-time kinematic), is the possibility of better modeling of systematic errors related to, e.g., the work of satellite clocks and delays related to the signal propagation in atmosphere. Additionally, RTN solutions enable to achieve the coordinates determination repeatability regardless of the distance between the receiver and physical station. In the RTK mode the error in coordinate determination increases along with the distance to the physical station from which the corrections originate. Also, ASG EUPOS recommends using the RTN in precise positioning wherever possible. Another argument for using the RTN was the distance of apploximately 16 km to the nearest physical station of ASG EUPOS (LODZ). According to the Geospatial Positioning Accuracy Standards, the independent source of higher accuracy should be the highest accuracy feasible and practicable to evaluate the accuracy of the dataset. The RTN solution fulfills these guidelines.
The coordinates of the points were determined using a Trimble SPS882 survey-grade GNSS receiver in the PL-2000 zone 6 coordinate system (EPSG: 2177), and in the PL-KRON86-NH elevation system, with the estimated accuracy mXY = 0.03 m and mZ = 0.05 m. The measured GNSS points fell within the range:
  • Position Dilution of Precision (PDOP) 1.0 ÷ 4.8 (<5), mean PDOP = 1.6
  • Horizontal Dilution of Precision (HDOP) 0.6 ÷ 3.3 (<4), mean HDOP = 0.9
  • Vertical Dilution of Precision (VDOP) 0.8 ÷ 3.4 (<4), mean VDOP = 1.3
which proves a good configuration of satellite geometry with am effect on positioning accuracy of GNSS points: mean σN = 0.012 m, mean σE = 0.008 m, mean σH = 0.020 m and root mean square value RMS = 0.025 m.
Because of the surface properties and a variety of factors that may impact the accuracy and quality of digital photogrammetric processing, the test field was divided into the following sections (Figure 1):
  • The major part which includes the area between industrial buildings (a);
  • The terrain around the industrial buildings (b);
  • The roundabout, access road, gas station (c).

2.2. Used Methods and Software

Due to the use of an archival image block in the study with the potential blur of images caused by wind gusts during Trimble® UX5 UAS flight, their radiometric quality verification was carried out. The graphical contents and a large number of images indicate that the detection of blur on images should be realized using automatic methods [56]. In this case, for preprocessing, the implementation of the Open Source Computer Vision Library (OpenCV) in Python was used. In the beginning, the variance of the absolute values using the Laplacian operator (linear high-pass filter) by Pech-Pacheco et al. [57] (Equation (1)) was calculated:
L A P _ V A R ( I ) = m M n N [ | L ( m , n ) | L ¯ ] 2
where L ¯ is the mean of absolute values; L(m, n) is the convolution of the image Img(m, n) with the mask L; M, N is the function size.
Since some of the images were tagged as potentially blurred, the decision was made to use the second approach for blur detection, which involves the frequency domain. The greyscale intensity images were transformed using Fourier transform F(u, v) accordingly to Equation (2):
F ( u , v ) = 1 M N x = 0 M 1 y = 0 N 1 f ( x , y ) e j 2 π ( u x M + v y N )
where M, N is the number of columns and rows, f(x, y) is the pixel value [27]. The zero-frequency component has to be shifted to the center of the spectrum, and then the normalization from 0 to 255 has to be done (Equation (3)).
F F ( u , v ) = 255 log ( 1 + | F ( u , v ) | ) max ( log ( 1 + | F ( u , v ) | ) )
where FF(u, v) is Fourier transform component [27]. To determine the level of blur in the images, the skewness of the normalized 2D Fourier transform was computed according to Ribeiro-Gomes et al. [58] (Equation (4)):
s k e w = 1 M · N u = 0 M 1 v = 0 N 1 ( F F ( u , v ) F F ( u , v ) ¯ ) 3 ( 1 M · N u = 0 M 1 v = 0 N 1 ( F F ( u , v ) F F ( u , v ) ¯ ) 2 ) 3
where FF(u, v) is each Fourier transform component shifted to the center of the spectrum and normalized and F F ( u , v ) ¯ is the mean value.
The photogrammetric measurements and advanced digital images processing were performed in the following software (in the order in which they were used):
  • Matching (own application), Poland [59]
  • IrfanView v. 4.41 (application by Irfan Skiljan), Vienna Austria (https://www.irfanview.com/).
  • Pix4Dmapper Pro v. 4.3.31 (full license) of Pix4D S.A., Prilly Switzerland [32,52].
  • RealityCapture v. 1.0.3.6310 (commercial license) of Capturing Reality s.r.o, Bratislava Slovakia [60,61].
  • PhotoScan v. 1.5.1 build 7618 (commercial license) of Agisoft LLC, St. Petersburg Russia [10,62].
  • CloudCompare v. 2.10.2 (GNU General Public License) of Telecom ParisTech and the research and development (R&D) division of EDF, Paris France [40,63].
  • AutoCAD Civil 3D 2019 v.13.0.613.0 (academic license) of Autodesk, Inc., San Rafael USA (https://www.autodesk.pl).
Figure 2 demonstrates the proposed workflow of this study for the accuracy evaluation of data collection based on low-altitude images. This scheme is an application of the original methodology for complex photogrammetric data collection using Pix4Dmapper as the main software for processing.
The signalized ground control points and natural ground points were measured on images using the center-weighted method (centroid operator) by means of Matching, the original software [59] with mean subpixel accuracy sx’y’ = 0.15 pixel. On the images, where points could not be measured automatically, the edge filter detection using IrfanView application was applied, and in some cases, the measurement was performed manually. Each ground control point (GCP) was identified on average in 38 images.
At the first stage of computation in Pix4Dmapper, the combined BBA with camera self-calibration, including 218 ground points (5 signalized GCPs and 213 ChPs in three interest areas, Figure 1) were carried out with an automatic calculation of interior (standard designation: c, x’0, y’0, K1, K2, K3, P1, P2) and exterior (X0, Y0, Z0, ω, φ, κ) orientation parameters [64]. In the adjustment, the approximated initial camera positions in the WGS84 coordinate system (*.LOG file) were used, and the output coordinate system was PL-2000 zone 6 (EPSG: 2177).
Due to the limited possibilities of changing settings in Pix4Dmapper, it is not discussed in this paper. The individual settings could be checked in reports from computations, which are in supplementary information files (Appendix 1–6).
The Pix4Dmapper calculates the global projection error for points measured on images without signalizing blunders. To avoid manual inspection of 218 points (each one separately) on mean 38 images in Pix4Dmapper, the RealityCapture software with its fast signalization of blunders was used. This step allows the identification of 4 faulty points numbering on images and their correcting.
To find the optimal solution of the BBA, 44 computing variants were executed in Pix4Dmapper with various numbers and configuration of GCPs and ChPs. In research, it was assumed that for each variant a new project was set up. This approach led to computing each input dataset independently. The 6 selected configuration variants of GCPs (no. 100 ÷ 103—artificially signalized, no. 201 ÷ 211—natural signalized points), optimal in terms of the number and adjustment accuracy, are presented below in Figure 3.
To prepare the accuracy assessment methodology, i.e., statistical test parameters, number and distribution of ground points, the guidelines described in [65] were used. As a statistical parameter the RMSE values are described, which are interpreted as a “square root of the average of the set of squared differences between dataset coordinate values and coordinate values from an independent source of higher accuracy for identical points”. In this paper the 95% confidence level was not applied for accuracy assessment, because used values of RMSE are lower.
In the photogrammetry, various statistical tests are applied for accuracy evaluation. The statistical tests like Student’s t-distribution and Fisher-Snedecor’s F-distribution are used to verify the significance of camera interior orientation parameters computed by the bundle block adjustment with self-calibration or simultaneous (on-the-job) calibration [66]. The Cochran test is used to check the error significance of the 3D coordinates estimation (homogeneity of many standard deviations of independent samples). The blunder detection in the reference dataset, e.g., GCPs coordinates is possible using statistical tests, the Student t-test (used to assess the presence of bias) and chi-square test (used to determine whether random errors are adequate) [15,67]. In our study, for the measurement of GCPs and ChPs, we have used the Polish Active Geodetic Network ASG EUPOS, which is regularly tested and the correctness of its precision was proved before. According to commonly known adjustment computation principles [68] using statistical tests like Student t-test or chi-square was no justified in our case study.
To define reference values of limiting errors for planar-elevation measurements of well-defined terrain details, the ASPRS Positional Accuracy Standards for Digital Geospatial Data [44] and Polish [43] technical requirements were taken into account. The first document recommends the horizontal accuracy class RMSEx and RMSEy for orthoimagery purposes, and the vertical accuracy class RMSEZ for digital elevation data. The second regulation describes the point position errors in the cartesian and elevation coordinate systems. Polish technical requirements were adopted in the paper as more rigorous. To analyze the accuracy of photogrammetric products in this study, the horizontal accuracy RMSE(X) and RMSE(Y) and vertical accuracy RMSE(Z) respectively to 0.10 m and 0.05 m were used.
The results of the BBA in Pix4Dmapper were analyzed in terms of GCPs and ChPs number, RMSE(XYZ) on GCPs and RMSE(XYZ) on ChPs. Based on the accuracy assessment of the BBA for further works, variant 6 (Figure 3f) was chosen as optimal. The use of one GCP (210) in section b (Figure 1) and two GCPs (206, 207) in section a (Figure 1) was caused by significant matching errors on roof surfaces and in their immediate surroundings, and as a consequence, the significant errors of photogrammetrically calculated ChPs coordinates. Since it is possible to obtain different results in different “black-box” applications [10,11,69], for an independent verification of the results, variant 6 was computed in the PhotoScan application. It is a well-known suite and uses other computation algorithms for photogrammetric product generation.
The next aspect was the evaluation of the impact of less forward and side overlap, and reduction of image number on the significant decrease of the BBA accuracy. Based on the BBA optimal variant, the computations were carried out in two options: with and without approximated UAV GNSS coordinates of images’ projection center (position accuracy of several meters). The forward overlap p% = 65% was obtained by removing every second photo in each strip. The side overlap q% = 65% and q% = 45% were received by removing from the initial image block every second, and every second and third strips, respectively.
Advanced images processing in Pix4Dmapper included the generation of dense point clouds, the DSM and finally a georeferenced orthomosaic in GeoTIFF format.
The 1118 height points and dense point cloud generated in Pix4Dmapper were imported to the CloudCompare application [63] for comparison of Z component using the least squares planar fitting of 3D points [40]. For fitting the mathematical model on the nearest point and its 4 neighbors the function (Equation (5)) was applied, which assumes that the sum of squared errors between the Zi and the plane values Axi + Byi + C is minimized.
E ( A , B , C ) = i = 1 m [ ( A x i + B y i + C ) Z i ] 2
As a result, the vertical distance from the compared point cloud to the computed DSM is determined. The assessment of vertical distances distribution was analyzed.
For the orthorectification in Pix4Dmapper, the same GSD size of the orthomosaic and acquired images (GSDortho = GSD = 4.6 cm) was assumed. The received georeferenced orthomosaic and XY coordinates of 11 GCPs and 289 ChPs (incl. 207 ChPs from the BBA) were imported to AutoCAD Civil 3D. The 2D data were compared, and the values and directions of deviation vectors were determined.

3. Results

The data processing in all of the used software was carried out on a workstation with the processor Intel® Core™ i9-7940X, 128 GB RAM DDR4-3200 MHz memory, MSI 1080 Ti graphic card and Samsung 960 Pro SSD hard drive.

3.1. Blur Detection on the Images

At the first step, Laplacian operator (Equation (1)) shows 196 images that were tagged as potentially blurred. The Fourier Transform (Equations (2)–(4)) indicated that 124 images could be blurred. The manual inspection was performed, to check if the blur occurred on the images. It was found that the images pointed out by algorithms included the roofs of industrial buildings, which were characterized by the same color and amorphous texture. There were no images that contain motion and optical blur, so all of the images were allowed for further computation.

3.2. Bundle Block Adjustment and Camera Self-Calibration

For our tests, the in-flight self-calibration, which describes the real acquisition condition and gives more accurate camera calibration [34], was used. The results of the 6 BBA solutions with reports of the camera standard parameter calibration are included in supplementary information files (Appendix 1–6). The main difference obtained in the 6 BBA variants were the values of reprojection error, which are understood as the distance between the marked and the reprojected point on one image. Due to very similar results (max. differences in variants 1 ÷ 6 are Δc = 1.3 pix, Δx’0 = 0.3 pix, Δy’0 = 0.2 pix, ΔK1 = 0.000, ΔK2 = 0.001, ΔK3 = 0.000, ΔP1 = 0.000, ΔP2 = 0.000) of interior orientation parameters calculated in Pix4Dmapper suite, which does not allow the interference in computation process, and because camera self-calibration is well described in the literature [13,25,26], the parameters are not discussed here.
Table 1 presents only the results of the final 6 BBA solutions (218 GCPs and ChPs in various combinations) obtained in the Pix4Dmapper suite. The maximum differences between root mean square errors on GCPs obtained in 6 variants of the BBA in horizontal and vertical planes are respectively to RMSE(X) = 0.002 m, RMSE(Y) = 0.007 m, RMSE(Z) = 0.013 m. In the case of ChPs, the RMSE are following RMSE(X) = 0.003 m, RMSE(Y) = 0.001 m and RMSE(Z) = 0.013 m.
The distribution of deviations ΔX, ΔY, ΔZ (differences between coordinates measured using GNSS and obtained in the BBA) on ChPs for 6 variants are presented in Figure 4.
Table 2 includes the analysis of ΔZ deviations in each computation BBA variants in more detail. In the analysis, the 5 intervals of the ΔZ deviations were assumed. The ΔZ thresholds are the result of the current Polish survey technical regulations, where the NΔZ is the number of absolute deviations values [ΔZ] for defined intervals.
The localization of the points (GCPs, ChPs), which belong to defined ΔZ intervals, is shown in Figure 5.
Upon the analysis of particular histograms (Figure 4) and the values of [ΔZ] deviation (Table 2) can be concluded that the most advantageous distribution of absolute values deviation, especially for [ΔZ] values occurred in variant 6. Furthermore, this variant poses the least percentage (9.7%) of [ΔZ] values higher than 0.05 m. The use of one GCP (210) in section b (Figure 1) and two GCPs (206, 207) in section a (Figure 1) enhanced the accuracy of Z coordinate determination. The differences of [ΔZ] deviation between variants 4 and 6 in individual sections a, b, c are, respectively, to 0.024 m, 0.008 m, 0.002 m.
To check the correctness and reliability of the optimal variant results (V6), the BBA was performed in an additional research and computation tool PhotoScan. This approach allows us to do an independent verification of digital processing results and to receive some additional information about applications’ reliability and functionalities. Supplementary files (Appendix 7) contains the report of computations.
Table 3 contains the comparison of variant 6 of the BBA results obtained in Pix4Dmapper and PhotoScan. The RMSE differences received on GCPs are respectively to RMSE(XGCP) = 0.009 m, RMSE(YGCP) = 0.005 m and on ChPs amount to RMSE(XChP) = 0.002 m, RMSE(YChP) = 0.002 m, RMSE(ZChP) = 0.004 m.
Table 4 presents the comparison of variant 6 of the BBA results with an extended analysis of the ΔZ deviations on ChPs. The main difference between the obtained outcomes is the number NΔZ of [ΔZ] deviations in the range of 0.05 ÷ 0.08 m to the detriment of the PhotoScan. The ChPs localization determined in the PhotoScan with the information about the [ΔZ] value is displayed in Figure 6. Figure 7 shows the histogram of ΔZ deviations on ChPs computed by the PhotoScan software.
The difference between number NΔZ of [ΔZ] deviations above threshold 0.05 m is 16 (ca. 8%). It is a consequence of a deviations number (6—Pix4Dmapper and 17—PhotoScan), which belongs in the 0.05 ÷ 0.06 m range. Therefore, the solutions in both applications are comparable.
In the authors’ opinion, the differences in values and in the number NΔZ of [ΔZ] deviations which are higher than 0.05 m are probably caused by another stochastic (weight) model used in computation algorithms in both types of software, e.g., for tie points measured automatically.

3.3. Influence of Image Overlap and Global Navigation Satellite Systems (GNSS) Exterior Orientation

The assessment of the influence of image overlap value and the use of UAV GNSS external orientation in the bundle adjustment was carried out in multi-variant (7 new variants) computation using the optimal BBA variant 6 as reference. The results of the reference and additional solutions with reports of the camera standard parameter calibration are included in supplementary information files (Appendix 6, 8–14). Table 5 contains the calculation results for four variants (V6p%-q%) of image overlap, respectively to p% = q% = 85% (V685-85), p% = 85% and q% = 45% (V685-45), p% = q% = 65% (V665-65), p% = 65% and q% = 45% (V665-45) with additional observations of exterior orientation parameters included in the adjustment. Table 6 presents the same adjustment variants, but with observations of projection centers coordinates excluded in the adjustment.
In the case (V665-45) of minimal forward p% = 65% and side q% = 45% overlap (ca. 6.2-fold reduction of image number) the RMSE(ZChP) is a little bit over the critical value 0.05 m and the percentage of points with the absolute deviations values [ΔZ] > 0.05 m rise 3.5-fold. In two other variants (V685-65 and V665-65), the absolute deviations values [ΔZ] are lower than 0.05 m. The percentage of points with the absolute deviations values [ΔZ] > 0.05 m rise 2.5-fold. The exclusion of approximated coordinates of projection centers in the BBA process is not causing significant changes in obtained accuracy.
In regard to variant V685-85 (incl. UAV GNSS) the max. differences of interior orientation parameters calculated for 3 variants of the BBA with UAV GNSS data and reduced overlap are the following: Δc = −7.3 pix, Δx’0 = 1.3 pix, Δy’0 = 3.0 pix, ΔK1 = 0.000, ΔK2 = 0.000, ΔK3 = 0.001, ΔP1 = 0.001, ΔP2 = 0.000. Max. differences of parameters between variant V685-85 (excl. UAV GNSS) and 3 variants of the BBA without UAV GNSS data and reduced overlap are respectively: Δc = 9.3 pix, Δx’0 = 1.4 pix, Δy’0 = 3.0 pix, ΔK1 = 0.000, ΔK2 = 0.000, ΔK3 = 0.000, ΔP1 = 0.001, ΔP2 = 0.000.
The value differences of parameters calculated between relevant variants with and without UAV GNSS data are very similar except for variant V665-45 (p% = 65% and q% = 45% overlap), respectively: Δc = 3.3 pix, Δx’0 = 0.1 pix, Δy’0 = 2.4 pix, ΔK1 = 0.000, ΔK2 = 0.000, ΔK3 = 0.001, ΔP1 = 0.001, ΔP2 = 0.000. The standard deviation of determined parameters are in range sc ∈ <0.5; 0.8> pix, sx’o = sy’o ∈ <0.07; 0.14> pix.

3.4. Comparison of Photogrammetric-Generated Digital Surface Model (DSM) and GNSS Point Cloud

Due to the requirements of height accuracy (RMSE(Z) ≤ 0.05 m), the generated DSM in Pix4Dmapper (variant 6) was compared to the point cloud obtained using GNSS. The determination of DSM undulation is realized by Equation (5) and is presented in Figure 8.
For most of the 1418 points (300 ChPs and 1118 height ground ChPs), the values of the vertical distance VD (height differences) belong to 0 ÷ 0.05 m range. Only for some random points, the estimated vertical distances are higher than 0.05 m. The vertical distances outliers comprise 4.3% (61) of the total number of points (1418). This fact confirms that DSM is correctly generated even in the area with the discontinuity of the terrain surface (scarps and terrain slope).

3.5. Comparison of the Orthomosaic and GNSS Data

The photogrammetric planar data collection was performed using orthomosaic. The outcome GeoTIFF ortomosaic (Figure 1) of Pix4Dmapper is 37058 × 15284 pixels (855 MB). To verify if the georeferenced orthomosaic comply with the mapping requirement (RMSE(XY) ≤ 0.10 m) of the point position in the cartesian coordinate system, the 2D reference coordinates of 11 GCPs, 207 ChPs and 82 supplementary ChPs measured by the GNNS method were applied. The 82 additional ChPs were the densification of the ChPs set and allowed an independent accuracy verification of the orthomosaic.
The photogrammetric 2D data collection (object vectorization) on the georeferenced orthomosaic can be treated as detailed land surveys. On the basis of the manual measurement on the GeoTIFF raster in AutoCAD Civil 3D, the deviation vectors DGNSS-BBA and DGNSS-OM were defined (GNSS—coordinates from GNSS measurement, BBA—coordinates computed in the bundle block adjustment, OM—coordinates from orthomosaic measurement). The a priori accuracy of manual pixel coordinates measurement was set at the level of 0.33 pixel size, which corresponds to 0.015 m in the terrain. Table 7 describes the accuracy evaluation of georeferenced orthomosaic. The length distributions of deviation vectors D in the plane are presented in Figure 9.
The measurement on the orthomosaic founds that 61 (20.3%) vectors exceed the threshold value of 0.10 m, where the calculated measurement accuracy for 300 ChPs was RMS(DGNSS-OM) = 0.082 m. In comparison with the reference data (2D coordinates) obtained from the BBA, the number of length vectors outliers increases up to 20-fold. The analysis of the histograms confirms that trend.
To check if the constant shift on the generated orthomosaic exists, the directions of deviation vectors were studied. The vectors’ directions were assigned to 4 quarters of the coordinate system. Table 8 contains the results of the verification.
Figure 10 presents the localization of ChPs with the information about vectors’ length and directions, where each direction is marked with the following colors: N–E: green, N–W: red, S–W: blue, S–E: purple.
The analysis of the analytical BBA data presented a similar direction distribution of the vectors DGNSS-BBA. On this distribution, the systematic errors are not found. However, directions distribution of vectors DGNSS-OM measured on georeferenced orthomosaic has a systematic shift in the N–W () direction. The analysis of vectors (DGNSS-BBA and DGNSS-OM) localization in both cases shows that the similarity of distributions does not exist. The values of deviation vectors on the orthomosaic were affected by the orthorectification process. In the authors’ opinion, the direction of shadow projection and different lighting conditions on image sequences in flight strips may influence on the length and the direction of error vectors.

3.6. Inspection of the Existing Digital Maps

The orthomosaic generated from UAV images is a relatively low-cost and quickly obtained photogrammetric product. Its actual photographic content allows the easy and fast visual detection of outdated information (Figure 11) and incompleteness (Figure 12) of existing vector digital maps, which are the result of geodetic detailed field surveys.

4. Discussion

For accuracy analysis, the current Polish technical standards for the land surveys (planar-elevation measurements of well-defined terrain details) in order to undertake large-scale mapping, included in the Regulation of the Minister of the Interior and Administration in Poland, No. 263 of 9 November 2011 [43], were used as the reference. They precisely describe the accuracy of field geodetic surveys of the ground details, which are uniquely identifiable, retaining the long-term invariability of shape and position. The measurements should achieve the point position in the cartesian coordinate system with error mXY ≤ 0.10 m and point height position in the elevation system with error mZ ≤ 0.05 m. The geodetic control network points used as a reference for dedicated field surveys are treated as errorless.
The planimetric mapping (land survey) is performed on the assumption that the accuracy (error mXY) of a point location of terrain details relative to the nearest horizontal geodetic control network is not less than:
  • 0.10 m: Ist survey group, i.e., survey marks of the control network, boundary points, building objects, and construction equipment, including elements of the utilities network, directly available for measurement;
  • 0.30 m: IInd survey group, i.e., buildings and earth devices in the form of embankments, excavations, dikes, dams, ditches, canals and artificial water reservoirs; invisible parts of buildings and construction equipment, including objects of covered infrastructural networks; land use objects, in particular: parks, green areas, lawns, playgrounds and rest, squares, single trees as well sport fields;
  • 0.50 m: IIIrd survey group, i.e., land use contours and soil outcrops for the needs of the land soil classification; watercourses and water reservoirs with natural boundaries; section boundary in the forests and national parks.
The geodetic height measurement of terrain elements is performed on the assumption that the accuracy of point height (error mZ) in relation to the nearest height (vertical) geodetic control network is not less than:
  • 0.02 m: for underground pipes and sewage devices;
  • 0.05 m: for building objects and construction equipment as well as height spots marked in the field;
  • 0.10 m: for terrain structures and flexible lines or electromagnetically measured underground objects of the infrastructural network; and non-marked height spots of characteristic points of the ground elevation.
The specification of terrain detail types and their planar-elevation survey accuracy contained in the regulation [43] mentioned above in principle applies to the direct geodetic survey and does not strictly concern the photogrammetric measurement. Due to the lack of the technical guidelines for practical usage of UAVs images in geodesy, the listed max. errors for surveys of well-defined over ground points were adopted in tests for the accuracy assessment of 2D and 3D photogrammetric measurement.
In this paper, we have compared conventional survey quality considerations with a photogrammetric approach to evaluate the accuracy of UAV imagery data collection. Our case study was performed according to Federal Geographic Data Committee mapping accuracy standards [65]. For testing purposes, features with known horizontal position (high degree of accuracy) and position with respect to the geodetic datum were used. Also, the independent source of higher accuracy was acquired separately from data used in the aerotriangulation solution. According to [65], check points were distributed more densely in the area of interest and more sparsely in that of little or no interest.
In the study, the set of images used for processing was radiometric proper; they did not contain motion and optical blur. The process of the complex photogrammetric data collection was performed in Pix4Dmapper software in three stages. The first computation stage was the bundle block adjustment. Empirical tests have shown the need to apply 11 GCPs (variant 6 of the BBA) in the configuration (Figure 3f) related to the buildings localization. The use of 3 additional GCPs was necessary to minimize the reprojection errors on the stage of the BBA and to eliminate the influence of amorphous surfaces on the image matching. The interior orientation parameters computed in self-calibration depend significantly on the configuration of the image block (image number, block and strip geometry, forward and side overlap, drift, shift), GCPs (point signalize type, shape, size, color, number, distribution, resolution, accuracy) and UAV GNSS accuracy. The accuracy of camera calibration parameters determined using BBA affects the accuracy and reliability of estimated 3D ground point coordinates. The simultaneous self-calibration method most accurately corresponds to the real conditions of images acquisition and computation, and provides the optimal functional model of the solution by bundle method.
The 3D ground coordinates of all the pass points (natural signalized terrain details) estimated using the BBA can be the background for a 3D vector digital map production. The external accuracy measure, i.e., the RSME on 207 ground ChPs (XYZ) were respectively to RMSE(X) = 0.023 m, RMSE(Y) = 0.026 m and RMSE(Z) = 0.036 m. In respect of theoretical inhomogeneous horizontal and vertical accuracy of the estimated coordinates in the bundle adjustment, the ΔZ deviations were analyzed. It turned out that ca 90% of ΔZ deviations on 207 ChPs were to ΔZ ≤ 0.05 m and only ca. 2% to ΔZ > 0.10 m. Barry and Coakley [39] used in their study 10 GCPs and 45 ChPs on the small field area (0.3250 ha). They acquired 95% reliably within 0.041 mm horizontally and 0.068 m vertically with the 1.17 cm size of GSD. Benassi et al. [70] in their tests used eBee RPAS with direct georeferencing on the test field (20 ha) with 12 GCPs, 14 ChPs. They declare standard deviations of planar camera positions between 0.01 ÷ 0.02 m and 0.03 ÷ 0.05 m in vertical direction, and acquired average horizontal RMSE on ChPs of 0.022 m and 0.055 m in elevation. In our opinion, the small sample of ChPs, their localization far from high objects and no independent points render them inadequate to do the accuracy assessment of RTK RPAS measurements. Besides with fewer GCPs used as a control, error magnitudes on control and check points diverged, which was confirmed by James et al. [41].
Another study aspect relative to the initial dataset was the assessment of the impact of less overlap and, as a consequence, a reduction of the images number in the bundle adjustment. We found that the optimal variant in terms of accuracy and computing time using standard grade workstation is that with reduced photos overlap at the level of p% = q% = 65% (ca. 4-fold decrease of image number). However, the optimal configuration of the GCPs is a condition for reducing the photos number, which is confirmed by the comparison of BBA results for variants V1 and V4 (Table 1), and V665-65 (Table 5). Tests shown that for block of nadir images with decreased overlap to p% = q% = 65%, it is possible to achieve the horizontal and vertical accuracy of ground objects at the level of RMSE(XChP,YChP) = 0.5 GSD and RMSE(ZChP) = 1 GSD (0.046 m), but in this case the number NΔZ of [ΔZ] deviations higher than 0.05 m increases ca. 2.5-fold. As expected due to the low position accuracy (horizontal 5 m and vertical 10 m) of the GNSS single-frequency receiver (C/A, L1) located on the Trimble® UX5 UAS platform, the inclusion of approximated coordinates of projection centers in the bundle adjustment in all tested variants did not affect on the computation results. For the BBA results evaluation, the number and distribution of ChPs, which are treated as tie points in computation should be taken into consideration. Mostafa in his research [71] state that changing the forward and side overlap from p% = q% = 80% to p% = 80% and q% = 40% resulted in the same accuracy within the measurement noise. The investigations were conducted with GSD = 0.008 m and GSD = 0.016 m. Furthermore, it was indicated [71] that ground object positioning accuracy is about 2 ÷ 3 GSD value, and height accuracy is about 4 ÷ 5 GSD value.
The DSM which was generated in the second processing stage has been validated by 1418 GNSS height check points in CloudCompare application. The approach was realized successfully before [53]. The ca. 96% of the height differences are related to ΔZ ≤ 0.05 m and only ca. 4% are outliers ΔZ > 0.05 m. Oniga et al. [55] compared results with terrestrial laser scanning (TLS) data and obtained a standard deviation equal to 0.098 m. Wierzbicki and Nienaltowski [42] created the triangulated irregular network (TIN) model based on points measured using the GNSS RTK technique, and they achieved a standard deviation of 0.090 m for height differences. To avoid the interpolation of the irregular shape of the ground, our study contains height differences only for points measured in the field.
Finally, the manual measurement of the well-defined points on the GeoTIFF orthomosaic (GSDortho = 4.6 cm) was carried out in AutoCAD Civil 3D. The photogrammetric 2D data collection (pointwise object vectorization) on the georeferenced orthomosaic can be treated as detailed land surveys. To take into account the limiting error (mL = 2 m0) of the GeoTIFF manual measurement, it could be considered that ca. 90% of deviation vectors D on the set of 300 ChPs fulfill the planar accuracy requirements RMSE(XY) ≤ 0.10 m. According to the authors’ knowledge, this procedure has not been used before. For describing the accuracy of orthomosaics, researchers used the BBA results [36,72]. Hung et al. [12] and James et al. [41] analyzed the directions and length of deviations vectors for GCPs and ChPs from the BBA. In our study, we also checked them on the orthomosaic, which showed that the constant shift occurs despite the random distribution of vectors from the BBA.
It is worth noting that in a complex process of orthomosaic generation, error propagation occurs. The final geometric accuracy and radiometric quality of orthomosaic is influenced by the many factors caused by, e.g., target marking (GCPs) and coordinate measurement errors, incorrect stochastic (weighted) model in the BBA, image-matching errors, estimated parameters errors of interior and exterior orientation, insufficient correction of image distortions, densified point cloud noise, final density and quality of DSM, applied orthorectification and digital interpolation method during the resampling, radiometric correction of respective orthoimages, quality of mosaicking.
Analyzing the real, not ideal conditions of the input dataset, it can be assumed that the following factors could have a significant impact on the accuracy of the final results: a relatively large size of GSD, a small number and incorrect artificial signalization of GCPs, inappropriate orientation of the strips relative to the shape of the photo block, and difficult conditions of dense matching as a consequence of the amorphous texture of the industrial object surface.
The positive experimental results of the presented study encourage the authors to continue research on the accuracy evaluation of photogrammetric products based on UAV imagery. Therefore, further work will be carried out on a topographically diverse test field, a representative for the general case of study, and the negative factors will be eliminated, which decreases the accuracy of the UAVs survey.

5. Conclusions

The use of unmanned aerial vehicles (UAV) for image-based surveys is becoming increasingly widespread across the geosciences and in industries such as engineering and surveying. However, the increasing availability of such data has not been paralleled by an equivalent increase in research about the data quality. The aim of the study is to maximize the accuracy and reliability of UAV survey results, whilst optimizing the input dataset and digital-processing requirements.
The presented study evaluates the accuracy of photogrammetric 2D and 3D data collection based on processed medium resolution images acquired from Trimble®UX5 UAS production flight mission, in the case of a small investment area with a focus on the influence of ground control points configuration. The study site is representative for other areas with similar characteristic and location of terrain objects.
The initial dataset was characterized by a block of 858 digital images with 16 megapixel resolution, 4.6 cm ground sample distance, 5 artificially signalized ground control points. For the verification of the results, the 295 ground check points and 1118 height check points were measured applying the GNSS using the real-time network method with the mean sigma (NEH) respectively to 0.012 m, 0.008 m, 0.020 m.
The photogrammetric data collection for well-defined points was carried out using the Pix4Dmapper suite in three processing stages: in the bundle block adjustment, on the generated digital surface model and the georeferenced orthomosaic. Furthermore, the bundle block adjustment was performed using RealityCapure and PhotoScan packages. In addition, the extended accuracy analyses were achieved on the digital surface model using CloudCompare application and on the orthomosaic supported by AutoCAD Civil 3D.
The main original research contributions of this paper are the following:
  • Creating a new methodology for complex accuracy assessment of computing pipeline of particular photogrammetric products;
  • Proving high accuracy and significant statistical reliability of photogrammetric 2D and 3D data collection based on UAV medium spatial resolution imagery with lower than 5 cm ground sample distance;
  • Stating the absolute need for a multi-variant bundle block adjustment and error assessment to search for optimal ground control points configuration for further advanced digital processing chain, which allows minimizing image-matching errors caused by the amorphous texture;
  • Proving the high horizontal and vertical accuracy (root mean square error less than or equal to 0.10 m and 0.05 m) for well-defined terrain points of photogrammetric products based on UAV imagery, which fulfills Polish and US technical requirements;
  • Finding the optimal photos overlap in terms of accuracy and computing time;
  • Proving the lack of influence of low-accuracy image georeferences on the bundle block adjustment results;
  • Confirming the appropriateness of UAV imagery and dedicated photogrammetric software packages for 2D and 3D data collection in vector, raster and hybrid form in order to undertake large-scale basic, thematic and inventory mapping;
  • Proving the possibilities of replacing direct geodetic field measurements by the UAVs photogrammetric measurements with the accuracy for large-scale mapping;
  • Verifying the usefulness of UAV imagery and their products for regular updates existing digital base maps and geodetic database.
The accuracy of the achieved products shows the full practical utility of a digital compact system camera for geodata collection. High accuracy could be obtained only under the conditions of multi-variant image processing and extended error analysis, according to the proposed workflow. In the case of a smaller ground sample distance, the accuracy of data collection will be significantly higher.
When the Pix4Dmapper package is applied for the bundle-block adjustment, additional software should be used in the post processing to find the image coordinates measurement blunders and to verify primary adjustment results. For the accuracy assessment of the georeferenced orthomosaic, we propose to analyze the distribution and length of deviation vectors.
The more frequent application of UAV imagery in geodesy and geomatics is, unfortunately, limited by the lack of dedicated national technical regulations and guidelines.
Further study will be carried out on the specially designed test field characterized by a high number of artificially signalized and evenly distributed ground control and check points which have been measured with high accuracy using static GNSS method and precise leveling. In addition, the research area should have a block of nadir and oblique images with a small ground sample distance (1 ÷ 2 cm level), acquired with a medium or large-format digital camera with high-resolution and a digital surface model based on dense point clouds obtained from airborne laser scanning, which will be used as a reference. A new test field will allow for testing the quality and accuracy of standard products (bundle-block adjustment, digital surface model, georeferenced orthomosaic) and other products derived from UAV imagery that were not the subject of the described study, i.e., true orthophoto maps, 3D object reconstruction, and 3D city models in the two-level of detail (LOD3 and LOD4).

Supplementary Materials

The following are available online at https://www.mdpi.com/1424-8220/19/23/5229/s1.

Author Contributions

Conceptualization, G.G. and P.S.; methodology, G.G. and P.S.; software, G.G. and P.S.; validation, G.G. and P.S; formal analysis, G.G. and P.S. investigation G.G. and P.S.; resources, G.G. and P.S.; data curation, G.G. and P.S.; writing—original draft preparation, G.G. and P.S.; writing—review and editing, G.G. and P.S.; visualization, G.G. and P.S.

Funding

This research received no external funding, and the APC was funded by the University of Warmia and Mazury in Olsztyn.

Acknowledgments

We gratefully thank the Geoplan company who provided data used in this study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Granshaw, S.I. Photogrammetric Terminology: Third Edition. Photogramm. Rec. 2016, 31, 210–252. [Google Scholar] [CrossRef]
  2. Granshaw, S.I. RPV, UAV, UAS, RPAS … or just drone? Photogramm. Rec. 2018, 33, 160–170. [Google Scholar] [CrossRef]
  3. Colomina, I.; Molina, P. Unmanned aerial systems for photogrammetry and remote sensing: A review. ISPRS J. Photogramm. Remote Sens. 2014, 92, 79–97. [Google Scholar] [CrossRef]
  4. Austin, R. Unmanned Aircraft Systems: UAVS Design, Development and Deployment; Wiley: Chippenham, UK, 2010; ISBN 9780470058190. [Google Scholar]
  5. Lin, Z.-J. UAV for Mapping—Low Altitude Photogrammetric Survey. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2008, 37, 1183–1186. [Google Scholar]
  6. Hirschmüller, H. Accurate and Efficient Stereo Processing by Semi-Global Matching and Mutual Information. Comput. Vis. Pattern Recognition, 2005. CVPR 2005. IEEE Comput. Soc. Conf. 2005, 2, 807–814. [Google Scholar] [CrossRef]
  7. Hirschmüller, H. Stereo processing by semiglobal matching and mutual information. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 30, 328–341. [Google Scholar] [CrossRef] [PubMed]
  8. Vu, H.H.; Labatut, P.; Pons, J.P.; Keriven, R. High accuracy and visibility-consistent dense multiview stereo. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 889–901. [Google Scholar] [CrossRef] [PubMed]
  9. Remondino, F.; Spera, M.G.; Nocerino, E.; Menna, F.; Nex, F. State of the art in high density image matching. Photogramm. Rec. 2014, 29, 144–166. [Google Scholar] [CrossRef]
  10. Jaud, M.; Passot, S.; Le Bivic, R.; Delacourt, C.; Grandjean, P.; Le Dantec, N. Assessing the accuracy of high resolution digital surface models computed by PhotoScan® and MicMac® in sub-optimal survey conditions. Remote Sens. 2016, 8, 465. [Google Scholar] [CrossRef]
  11. Remondino, F.; Nocerino, E.; Toschi, I.; Menna, F. A critical review of automated photogrammetric processing of large datasets. In Proceedings of the International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Ottawa, ON, Canada, 28 August–1 September 2017. [Google Scholar] [CrossRef]
  12. Hung, I.-K.; Unger, D.; Kulhavy, D.; Zhang, Y. Positional Precision Analysis of Orthomosaics Derived from Drone Captured Aerial Imagery. Drones 2019, 3, 46. [Google Scholar] [CrossRef]
  13. Oniga, V.E.; Pfeifer, N.; Loghin, A.M. 3D calibration test-field for digital cameras mounted on unmanned aerial systems (UAS). Remote Sens. 2018, 10, 2017. [Google Scholar] [CrossRef]
  14. Uysal, M.; Toprak, A.S.; Polat, N. DEM generation with UAV Photogrammetry and accuracy analysis in Sahitler hill. Meas. J. Int. Meas. Confed. 2015, 73, 539–543. [Google Scholar] [CrossRef]
  15. Mesas-Carrascosa, F.J.; Rumbao, I.C.; Berrocal, J.A.B.; Porras, A.G.F. Positional quality assessment of orthophotos obtained from sensors onboard multi-rotor UAV platforms. Sensors 2014, 14, 22394–22407. [Google Scholar] [CrossRef] [PubMed]
  16. Nex, F.; Remondino, F. UAV for 3D mapping applications: A review. Appl. Geomat. 2014, 6, 1–15. [Google Scholar] [CrossRef]
  17. Pajares, G. Overview and current status of remote sensing applications based on unmanned aerial vehicles (UAVs). Photogramm. Eng. Remote Sens. 2015, 81, 281–330. [Google Scholar] [CrossRef]
  18. Crommelinck, S.; Bennett, R.; Gerke, M.; Nex, F.; Yang, M.Y.; Vosselman, G. Review of automatic feature extraction from high-resolution optical sensor data for UAV-based cadastral mapping. Remote Sens. 2016, 8, 689. [Google Scholar] [CrossRef]
  19. Gevaert, C.; Sliuzas, R.; Persello, C.; Vosselman, G. Opportunities for UAV mapping to support unplanned settlement upgrading. Rwanda J. 2017, 1. [Google Scholar] [CrossRef]
  20. D’Oleire-Oltmanns, S.; Marzolff, I.; Peter, K.D.; Ries, J.B. Unmanned aerial vehicle (UAV) for monitoring soil erosion in Morocco. Remote Sens. 2012, 4, 3390–3416. [Google Scholar] [CrossRef]
  21. Yao, H.; Qin, R.; Chen, X. Unmanned Aerial Vehicle for Remote Sensing Applications—A Review. Remote Sens. 2019, 11, 1443. [Google Scholar] [CrossRef]
  22. Lippitt, C.D.; Zhang, S. The impact of small unmanned airborne platforms on passive optical remote sensing: A conceptual perspective. Int. J. Remote Sens. 2018, 39, 4852–4868. [Google Scholar] [CrossRef]
  23. Shahbazi, M.; Sohn, G.; Théau, J.; Menard, P. Development and evaluation of a UAV-photogrammetry system for precise 3D environmental modeling. Sensors 2015, 15, 27493–27524. [Google Scholar] [CrossRef] [PubMed]
  24. Lehmann, J.R.K.; Prinz, T.; Ziller, S.R.; Thiele, J.; Heringer, G.; Meira-Neto, J.A.A.; Buttschardt, T.K. Open-source processing and analysis of aerial imagery acquired with a low-cost Unmanned Aerial System to support invasive plant management. Front. Environ. Sci. 2017, 5, 44. [Google Scholar] [CrossRef]
  25. Harwin, S.; Lucieer, A.; Osborn, J. The impact of the calibration method on the accuracy of point clouds derived using unmanned aerial vehicle multi-view stereopsis. Remote Sens. 2015, 7, 11933–11953. [Google Scholar] [CrossRef]
  26. Remondino, F.; Fraser, C. Digital camera calibration methods: Considerations and comparisons. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2006, 36, 266–272. [Google Scholar]
  27. Forlani, G.; Diotri, F.; Cella, U.M.; Roncella, R. Indirect UAV Strip Georeferencing by On-Board GNSS Data under Poor Satellite Coverage. Remote Sens. 2019, 15, 1765. [Google Scholar] [CrossRef]
  28. Vo, A.V.; Truong-Hong, L.; Laefer, D.F.; Tiede, D.; Doleire-Oltmanns, S.; Baraldi, A.; Shimoni, M.; Moser, G.; Tuia, D. Processing of Extremely High Resolution LiDAR and RGB Data: Outcome of the 2015 IEEE GRSS Data Fusion Contest-Part B: 3-D Contest. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 5560–5575. [Google Scholar] [CrossRef]
  29. Koch, T.; Körner, M.; Fraundorfer, F. Automatic and Semantically-Aware 3D UAV Flight Planning for Image-Based 3D Reconstruction. Remote Sens. 2019, 11, 1550. [Google Scholar] [CrossRef]
  30. Agarwal, S.; Snavely, N.; Simon, I.; Seitz, S.M.; Szeliski, R. Building Rome in a day. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009; pp. 72–79. [Google Scholar] [CrossRef]
  31. Haala, N.; Cramer, M.; Weimer, F.; Trittler, M. Performance Test On Uav-Based Photogrammetric Data Collection. ISPRS Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2012, 38. [Google Scholar] [CrossRef]
  32. Remondino, F.; Barazzetti, L.; Nex, F.; Scaioni, M.; Sarazzi, D. UAV Photogrammetry for Mapping and 3D Modeling—Current Status and Future Perspectives. ISPRS Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2012, 38, C22. [Google Scholar] [CrossRef]
  33. Harwin, S.; Lucieer, A. Assessing the accuracy of georeferenced point clouds produced via multi-view stereopsis from Unmanned Aerial Vehicle (UAV) imagery. Remote Sens. 2012, 4, 1573–1599. [Google Scholar] [CrossRef]
  34. Zhou, Y.; Rupnik, E.; Faure, P.H.; Pierrot-Deseilligny, M. GNSS-assisted integrated sensor orientation with sensor pre-calibration for accurate corridor mapping. Sensors 2018, 18, 2783. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Gindraux, S.; Boesch, R.; Farinotti, D. Accuracy assessment of digital surface models from Unmanned Aerial Vehicles’ imagery on glaciers. Remote Sens. 2017, 9, 186. [Google Scholar] [CrossRef] [Green Version]
  36. Mesas-Carrascosa, F.J.; García, M.D.N.; De Larriva, J.E.M.; García-Ferrer, A. An analysis of the influence of flight parameters in the generation of unmanned aerial vehicle (UAV) orthomosaicks to survey archaeological areas. Sensors 2016, 16, 1838. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Zhang, S.; Lippitt, C.D.; Bogus, S.M.; Loerch, A.C.; Sturm, J.O. The accuracy of aerial triangulation products automatically generated from hyper-spatial resolution digital aerial photography. Remote Sens. Lett. 2016, 7, 160–169. [Google Scholar] [CrossRef]
  38. Cramer, M.; Bovet, S.; Gültlinger, M.; Honkavaara, E.; McGill, A.; Rijsdijk, M.; Tabor, M.; Tournadre, V. On the Use of RPAS in National Mapping—The EuroSDR Point of View. ISPRS Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2013, XL-1/W2, 93–99. [Google Scholar] [CrossRef] [Green Version]
  39. Barry, P.; Coakley, R. Field Accuracy Test of RPAS Photogrammetry. ISPRS Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2013, 40. [Google Scholar] [CrossRef] [Green Version]
  40. James, M.R.; Robson, S.; Smith, M.W. 3-D uncertainty-based topographic change detection with structure-from-motion photogrammetry: Precision maps for ground control and directly georeferenced surveys. Earth Surf. Process. Landf. 2017, 42, 1769–1788. [Google Scholar] [CrossRef]
  41. James, M.R.; Robson, S.; d’Oleire-Oltmanns, S.; Niethammer, U. Optimising UAV topographic surveys processed with structure-from-motion: Ground control quality, quantity and bundle adjustment. Geomorphology 2017, 281, 51–66. [Google Scholar] [CrossRef] [Green Version]
  42. Wierzbicki, D.; Nienaltowski, M. Accuracy Analysis of a 3D Model of Excavation, Created from Images Acquired with an Action Camera from Low Altitudes. ISPRS Int. J. Geo Inf. 2019, 8, 83. [Google Scholar] [CrossRef] [Green Version]
  43. Regulation of the Minister of the Interior and Administration in Poland, No. 263 of 9 November 2011 Dz.U.2011.263.1572 2011. Available online: http://prawo.sejm.gov.pl/isap.nsf/download.xsp/WDU20112631572/O/D20111572.pdf (accessed on 20 October 2019).
  44. ASPRS. Positional Accuracy Standards for Digital Geospatial Data. Photogramm. Eng. Remote Sens. 2015, 3, 1–26. [Google Scholar] [CrossRef]
  45. DIN 18740-3 Photogrammetrische Produkte—Teil 3: Anforderungen an das Orthobild 2015. Available online: https://www.din.de/de/mitwirken/normenausschuesse/nabau/normen/wdc-beuth:din21:235160913 (accessed on 20 October 2019).
  46. DIN 18740-6 Photogrammetrische Produkte—Teil 6: Anforderungen an digitale Höhenmodelle 2014. Available online: https://www.din.de/de/mitwirken/normenausschuesse/nabau/normen/wdc-beuth:din21:222023697 (accessed on 20 October 2019).
  47. Martínez-Carricondo, P.; Agüera-Vega, F.; Carvajal-Ramírez, F.; Mesas-Carrascosa, F.J.; García-Ferrer, A.; Pérez-Porras, F.J. Assessment of UAV-photogrammetric mapping accuracy based on variation of ground control points. Int. J. Appl. Earth Obs. Geoinf. 2018, 72, 1–10. [Google Scholar] [CrossRef]
  48. Agüera-Vega, F.; Carvajal-Ramírez, F.; Martínez-Carricondo, P. Assessment of photogrammetric mapping accuracy based on variation ground control points number using unmanned aerial vehicle. Meas. J. Int. Meas. Confed. 2017, 98, 221–227. [Google Scholar] [CrossRef]
  49. Tonkin, T.N.; Midgley, N.G. Ground-control networks for image based surface reconstruction: An investigation of optimum survey designs using UAV derived imagery and structure-from-motion photogrammetry. Remote Sens. 2016, 8, 786. [Google Scholar] [CrossRef] [Green Version]
  50. Stöcker, C.; Bennett, R.; Nex, F.; Gerke, M.; Zevenbergen, J. Review of the current state of UAV regulations. Remote Sens. 2017, 9, 459. [Google Scholar] [CrossRef] [Green Version]
  51. Murtiyoso, A.; Grussenmeyer, P.; Börlin, N.; Vandermeerschen, J.; Freville, T. Open Source and Independent Methods for Bundle Adjustment Assessment in Close-Range UAV Photogrammetry. Drones 2018, 2, 3. [Google Scholar] [CrossRef] [Green Version]
  52. Cramer, M. The UAV@LGL BW Project—A NMCA Case Study. Photogr. Week 2013, 2011, 165–179. [Google Scholar]
  53. Gabara, G.; Sawicki, P. Application of UAV Imagery for Inventory Mapping—A Case of Industrial Estate. In Proceedings of the 2018 Baltic Geodetic Congress (BGC Geomatics), Olsztyn, Poland, 21–23 June 2018. [Google Scholar] [CrossRef]
  54. Sanz-Ablanedo, E.; Chandler, J.H.; Rodríguez-Pérez, J.R.; Ordóñez, C. Accuracy of Unmanned Aerial Vehicle (UAV) and SfM photogrammetry survey as a function of the number and location of ground control points used. Remote Sens. 2018, 10, 1606. [Google Scholar] [CrossRef] [Green Version]
  55. Oniga, E.; Breaban, A.; Statescu, F. Determining the optimum number of ground control points for obtaining high precision results based on UAS images. Proceedings 2018, 2, 352. [Google Scholar] [CrossRef] [Green Version]
  56. Sieberth, T.; Wackrow, R.; Chandler, J.H. Automatic detection of blurred images in UAV image sets. ISPRS J. Photogramm. Remote Sens. 2016, 122, 1–16. [Google Scholar] [CrossRef] [Green Version]
  57. Pech-Pacheco, J.L.; Cristobal, G.; Chamorro-Martinez, J.; Fernandez-Valdivia, J. Diatom autofocusing in brightfield microscopy: A comparative study. In Proceedings of the 15th International Conference on Pattern Recognition, Barcelona, Spain, 3–7 September 2000. [Google Scholar] [CrossRef]
  58. Ribeiro-Gomes, K.; Hernandez-Lopez, D.; Ballesteros, R.; Moreno, M.A. Approximate georeferencing and automatic blurred image detection to reduce the costs of UAV use in environmental and agricultural applications. Biosyst. Eng. 2016, 151, 308–327. [Google Scholar] [CrossRef]
  59. Sawicki, P.; Ostrowski, B. Research on chosen matching methods for the measurement of points on digital close range images. Ann. Geomat. 2005, 135–144. [Google Scholar]
  60. Jancosek, M.; Shekhovtsov, A.; Pajdla, T. Scalable multi-view stereo. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision Workshops, ICCV Workshops 2009, Kyoto, Japan, 27 September–4 October 2009. [Google Scholar] [CrossRef]
  61. Heller, J.; Havlena, M.; Jancosek, M.; Torii, A.; Pajdla, T. 3D reconstruction from photographs by CMP SfM web service. In Proceedings of the 14th IAPR International Conference on Machine Vision Applications, MVA 2015, Tokyo, Japan, 18–22 May 2015; pp. 30–34. [Google Scholar] [CrossRef]
  62. Gabara, G.; Sawicki, P. Accuracy Study of Close Range 3D Object Reconstruction Based on Point Clouds. In Proceedings of the 2017 Baltic Geodetic Congress (BGC Geomatics), Gdansk, Poland, 22–25 June 2017; pp. 25–29. [Google Scholar] [CrossRef]
  63. Girardeau-Montaut, D. CloudCompare: 3D Point Cloud and Mesh Processing Software. Available online: http://www.cloudcompare.org (accessed on 3 March 2018).
  64. Luhmann, T.; Robson, S.; Kyle, S.; Boehm, J. Close Range Photogrammetry and 3D Imaging; De Gruyter: Berlin, Germany, 2013; ISBN 978-3110302691. [Google Scholar]
  65. Authority, T.V. Federal Geographical Data Committee Geospatial Positioning Accuracy Standards Part 3: National Standard for Spatial Data Accuracy. National Aeronautics and Space Administration: Virginia, NV, USA, 1998. [Google Scholar]
  66. Sawicki, P. Simultaneous Calibration of a Digital Camera Kodak DC4800 in Close Range Photogrammetric Point Measurement. Arch. Photogramm. Cartogr. Remote Sens. 2003, 13, 457–466. (In Polish) [Google Scholar]
  67. American Society of Civil Engineers. Committee on Cartographic Surveying. In Map Uses, Scales and Accuracies for Engineering and Associated Purposes; ASCE: New York, NY, USA, 1983. [Google Scholar]
  68. Wiśniewski, Z. Adjustment Computation in Geodesy (with Examples); Springer: Berlin/Heidelberg, Germany, 2016; ISBN 978-83-8100-057-4. (In Polish) [Google Scholar]
  69. Nikolov, I.; Madsen, C. Benchmarking close-range structure from motion 3D reconstruction software under varying capturing conditions. In Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2016. [Google Scholar] [CrossRef] [Green Version]
  70. Benassi, F.; Dall’Asta, E.; Diotri, F.; Forlani, G.; di Cella, U.M.; Roncella, R.; Santise, M. Testing accuracy and repeatability of UAV blocks oriented with GNSS-supported aerial triangulation. Remote Sens. 2017, 9, 172. [Google Scholar] [CrossRef] [Green Version]
  71. Mostafa, M.M.R. Accuracy Assessment of Professional Grade Unmanned Systems for High Precision Airborne Mapping. ISPRS Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2017, 42, 257–261. [Google Scholar] [CrossRef] [Green Version]
  72. Manfreda, S.; Dvorak, P.; Mullerova, J.; Herban, S.; Vuono, P.; Arranz Justel, J.; Perks, M. Assessing the Accuracy of Digital Surface Models Derived from Optical Imagery Acquired with Unmanned Aerial Systems. Drones 2019, 3, 15. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The test object Stryków with the defined study sections (a), (b), (c) on the orthomosaic generated in Pix4Dmapper.
Figure 1. The test object Stryków with the defined study sections (a), (b), (c) on the orthomosaic generated in Pix4Dmapper.
Sensors 19 05229 g001
Figure 2. The workflow of measurement and processing proposed for unmanned aerial vehicle (UAV) surveys.
Figure 2. The workflow of measurement and processing proposed for unmanned aerial vehicle (UAV) surveys.
Sensors 19 05229 g002
Figure 3. Configuration of ground control points (GCPs) and check points (ChPs) on the test object Stryków in the 6 BBA variants: (a) Variant 1 which contains 8 GCPs (4 signalized) and 210 ChPs; (b) Variant 2 which includes 10 GCPs and 208 ChPs; (c) Variant 3 consists of 10 GCPs (4 signalized) and 208 ChPs; (d) Variant 4 defined as 8 GCPs and 210 ChPs; (e) Variant 5 which is the extension of (d) by 2 GCPs between industrial buildings; (f) Variant 6 (it is characterized by 11 GCPs and 207 ChPs) which is the extension of (e) by 1 GCP at the right side of industrial building and close to it.
Figure 3. Configuration of ground control points (GCPs) and check points (ChPs) on the test object Stryków in the 6 BBA variants: (a) Variant 1 which contains 8 GCPs (4 signalized) and 210 ChPs; (b) Variant 2 which includes 10 GCPs and 208 ChPs; (c) Variant 3 consists of 10 GCPs (4 signalized) and 208 ChPs; (d) Variant 4 defined as 8 GCPs and 210 ChPs; (e) Variant 5 which is the extension of (d) by 2 GCPs between industrial buildings; (f) Variant 6 (it is characterized by 11 GCPs and 207 ChPs) which is the extension of (e) by 1 GCP at the right side of industrial building and close to it.
Sensors 19 05229 g003
Figure 4. Histograms of deviations ΔX, ΔY, ΔZ on ChPs in computed variants (V1 ÷ V6) of the bundle-block adjustment (BBA): (a) ΔX deviation on ChPs; (b) ΔY deviation on ChPs; (c) ΔZ deviation on ChPs.
Figure 4. Histograms of deviations ΔX, ΔY, ΔZ on ChPs in computed variants (V1 ÷ V6) of the bundle-block adjustment (BBA): (a) ΔX deviation on ChPs; (b) ΔY deviation on ChPs; (c) ΔZ deviation on ChPs.
Sensors 19 05229 g004aSensors 19 05229 g004b
Figure 5. Localization of GCPs and ChPs in the BBA with representation of [ΔZ] in deviation intervals: (a) Variant 1; (b) Variant 2; (c) Variant 3; (d) Variant 4; (e) Variant 5; (f) Variant 6.
Figure 5. Localization of GCPs and ChPs in the BBA with representation of [ΔZ] in deviation intervals: (a) Variant 1; (b) Variant 2; (c) Variant 3; (d) Variant 4; (e) Variant 5; (f) Variant 6.
Sensors 19 05229 g005
Figure 6. Localization of GCPs and ChPs in the variant 6 of the BBA performed in PhotoScan with the representation of number NΔZ in [ΔZ] deviation intervals.
Figure 6. Localization of GCPs and ChPs in the variant 6 of the BBA performed in PhotoScan with the representation of number NΔZ in [ΔZ] deviation intervals.
Sensors 19 05229 g006
Figure 7. Histogram of ΔZ deviations on ChPs calculated in the BBA in PhotoScan application.
Figure 7. Histogram of ΔZ deviations on ChPs calculated in the BBA in PhotoScan application.
Sensors 19 05229 g007
Figure 8. Localization and histogram of the vertical distances (VD) between computed digital surface model (DSM) in Pix4Dmapper and GNSS point cloud.
Figure 8. Localization and histogram of the vertical distances (VD) between computed digital surface model (DSM) in Pix4Dmapper and GNSS point cloud.
Sensors 19 05229 g008
Figure 9. Length distribution of deviation vectors D: (a) DGNSS-BBA (218 ChPs); (b) DGNSS-OM (218 ChPs); (c) DGNSS-OM (82 ChPs); (d) DGNSS-OM (300 ChPs).
Figure 9. Length distribution of deviation vectors D: (a) DGNSS-BBA (218 ChPs); (b) DGNSS-OM (218 ChPs); (c) DGNSS-OM (82 ChPs); (d) DGNSS-OM (300 ChPs).
Sensors 19 05229 g009
Figure 10. Localization of vectors D which include length and direction: (a) DGNSS-BBA (218 ChPs); (b) DGNSS-OM (300 ChPs).
Figure 10. Localization of vectors D which include length and direction: (a) DGNSS-BBA (218 ChPs); (b) DGNSS-OM (300 ChPs).
Sensors 19 05229 g010
Figure 11. Visualization of the outdated information on the vector base map using orthomosaic in AutoCAD.
Figure 11. Visualization of the outdated information on the vector base map using orthomosaic in AutoCAD.
Sensors 19 05229 g011
Figure 12. Visualization of the deficiencies on the vector base map using orthomosaic in AutoCAD.
Figure 12. Visualization of the deficiencies on the vector base map using orthomosaic in AutoCAD.
Sensors 19 05229 g012
Table 1. Result of 6 variants of bundle block adjustment in Pix4Dmapper.
Table 1. Result of 6 variants of bundle block adjustment in Pix4Dmapper.
ParametersV1V2V3V4V5V6
No. of GCPs 8101081011
No. of ChPs210208208210208207
RMSE(X) on GCPs [m]0.0110.0110.0110.0090.0100.010
RMSE(Y) on GCPs [m]0.0160.0220.0180.0230.0230.022
RMSE(Z) on GCPs [m]0.0220.0350.0300.0250.0320.033
RMSE(X) on ChPs [m]0.0250.0240.0250.0220.0220.023
RMSE(Y) on ChPs [m]0.0250.0260.0250.0260.0260.026
RMSE(Z) on ChPs [m]0.0480.0370.0400.0490.0370.036
Table 2. Analysis of NΔZ in the ΔZ deviation intervals obtained from the BBA.
Table 2. Analysis of NΔZ in the ΔZ deviation intervals obtained from the BBA.
ParametersNΔZ V1NΔZ V2NΔZ V3NΔZ V4NΔZ V5NΔZ V6
No. of ChPs210208208210208207
[ΔZ] ≤ 0.05 m163185176164182187
0.05 < [ΔZ] ≤ 0.08 m331622331914
0.08 < [ΔZ] ≤ 0.10 m534432
[ΔZ] > 0.10 m946944
[%] of [ΔZ] > 0.05 m22.4%11.1%15.4%21.9%12.5%9.7%
Table 3. Results of the variant 6 of the BBA in Pix4Dmapper and PhotoScan.
Table 3. Results of the variant 6 of the BBA in Pix4Dmapper and PhotoScan.
ParametersV6—Pix4DmapperV6—PhotoScan
No. of GCPs1111
No. of ChPs207207
RMSE(X) on GCPs [m]0.0100.019
RMSE(Y) on GCPs [m]0.0220.027
RMSE(Z) on GCPs [m]0.0330.033
RMSE(X) on ChPs [m]0.0230.025
RMSE(Y) on ChPs [m]0.0260.024
RMSE(Z) on ChPs [m]0.0360.040
Table 4. Analysis of NΔZ (variant 6) calculated in the BBA in Pix4Dmapper and PhotoScan.
Table 4. Analysis of NΔZ (variant 6) calculated in the BBA in Pix4Dmapper and PhotoScan.
ParametersNΔZ V6—Pix4Dmapper NΔZ V6—PhotoScan
No. of ChPs207207
[ΔZ] ≤ 0.05 m187169
0.050 < [ΔZ] ≤ 0.08 m1428
0.080 < [ΔZ] ≤ 0.10 m24
[ΔZ] > 0.10 m44
[%] of [ΔZ] > 0.05 m9.7%17.4%
Table 5. Results of the BBA with different image overlap and exterior orientation parameters of unmanned aerial vehicle position from Global Navigation Satellite Systems receiver (UAV GNSS) included in the adjustment.
Table 5. Results of the BBA with different image overlap and exterior orientation parameters of unmanned aerial vehicle position from Global Navigation Satellite Systems receiver (UAV GNSS) included in the adjustment.
ParametersV685-85 incl. UAV GNSS V685-45 incl. UAV GNSSV665-65 incl. UAV GNSSV665-45 incl. UAV GNSS
No. of images858293205138
No. of GCPs11111111
No. of ChPs207207207207
RMSE(X) on GCPs [m]0.0100.0160.0100.012
RMSE(Y) on GCPs [m]0.0220.0220.0190.014
RMSE(Z) on GCPs [m]0.0330.0280.0200.024
RMSE(X) on ChPs [m]0.0230.0290.0240.037
RMSE(Y) on ChPs [m]0.0260.0310.0290.034
RMSE(Z) on ChPs [m]0.0360.0480.0460.055
[ΔZ] ≤ 0.05 m187155159134
0.050 < [ΔZ] ≤ 0.08 m14333149
0.080 < [ΔZ] ≤ 0.10 m210910
[ΔZ] > 0.10 m49814
[%] of [ΔZ] > 0.05 m9.725.123.235.3
Table 6. Results of the BBA with different image overlap and exterior orientation parameters excluded in the adjustment.
Table 6. Results of the BBA with different image overlap and exterior orientation parameters excluded in the adjustment.
ParametersV685-85 excl. UAV GNSS V685-45 excl. UAV GNSSV665-65 excl. UAV GNSSV665-45 excl. UAV GNSS
No. of images858293205138
No. of GCPs11111111
No. of ChPs207207207207
RMSE(X) on GCPs [m]0.0100.0150.0100.012
RMSE(Y) on GCPs [m]0.0220.0220.0190.016
RMSE(Z) on GCPs [m]0.0290.0270.0190.023
RMSE(X) on ChPs [m]0.0230.0290.0240.038
RMSE(Y) on ChPs [m]0.0260.0310.0290.035
RMSE(Z) on ChPs [m]0.0360.0470.0470.056
[ΔZ] ≤ 0.05 m182156154133
0.050 < [ΔZ] ≤ 0.08 m18323242
0.080 < [ΔZ] ≤ 0.10 m2101318
[ΔZ] > 0.10 m59814
[%] of [ΔZ] > 0.05 m12.124.625.635.7
Table 7. Length analysis of deviation vectors D.
Table 7. Length analysis of deviation vectors D.
ParametersV6—Pix4Dmapper
No. of vectors DGNSS-BBA > 0.10 m (218 ChPs)3
No. of vectors DGNSS-OM > 0.10 m (218 ChPs)50
No. of vectors DGNSS-OM > 0.10 m (82 ChPs)11
No. of vectors DGNSS-OM > 0.10 m (300 ChPs)61
[%] of vectors DGNSS-BBA > 0.10 m (218 ChPs)1.4
[%] of vectors DGNSS-OM > 0.10 m (218 ChPs)22.9
[%] of vectors DGNSS-OM > 0.10 m (82 ChPs)13.4
[%] of vectors DGNSS-OM > 0.10 m (300 ChPs)20.3
RMS(DGNSS-BBA) (218 ChPs) [m]0.034
RMS(DGNSS-OM) (218 ChPs) [m]0.084
RMS(DGNSS-OM) (82 ChPs) [m]0.076
RMS(DGNSS-OM) (300 ChPs) [m]0.082
Table 8. Analysis of direction of deviation vectors D.
Table 8. Analysis of direction of deviation vectors D.
Vector Direction
No. of vectors DGNSS-BBA (218 ChPs)69535442
No. of vectors DGNSS-OM (218 ChPs)6195152
No. of vectors DGNSS-OM (82 ChPs)367102
No. of vectors DGNSS-OM (300 ChPs)9262254

Share and Cite

MDPI and ACS Style

Gabara, G.; Sawicki, P. Multi-Variant Accuracy Evaluation of UAV Imaging Surveys: A Case Study on Investment Area. Sensors 2019, 19, 5229. https://doi.org/10.3390/s19235229

AMA Style

Gabara G, Sawicki P. Multi-Variant Accuracy Evaluation of UAV Imaging Surveys: A Case Study on Investment Area. Sensors. 2019; 19(23):5229. https://doi.org/10.3390/s19235229

Chicago/Turabian Style

Gabara, Grzegorz, and Piotr Sawicki. 2019. "Multi-Variant Accuracy Evaluation of UAV Imaging Surveys: A Case Study on Investment Area" Sensors 19, no. 23: 5229. https://doi.org/10.3390/s19235229

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop