Next Article in Journal
Refined Extraction of Sugarcane Planting Areas in Guangxi Using an Improved U-Net Model
Previous Article in Journal
Enabling Cooperative Autonomy in UUV Clusters: A Survey of Robust State Estimation and Information Fusion Techniques
Previous Article in Special Issue
Towards Safer UAV Operations in Urban Air Mobility: 3D Automated Modelling for CFD-Based Microweather Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimization of UAV Flight Parameters for Urban Photogrammetric Surveys: Balancing Orthomosaic Visual Quality and Operational Efficiency

1
Escuela de Construcción Civil, Pontificia Universidad Católica de Chile, Avenida Vicuña Mackenna 4860, Macul, Santiago 7820436, Chile
2
Escuela de Ingeniería Informática, Pontificia Universidad Católica de Valparaíso, Avenida Brasil 2241, Valparaíso, Valparaíso 2362807, Chile
3
Escuela de Negocios Internacionales, Universidad de Valparaíso, Alcalde Prieto Nieto 452, Viña del Mar 2572048, Chile
*
Author to whom correspondence should be addressed.
Drones 2025, 9(11), 753; https://doi.org/10.3390/drones9110753
Submission received: 28 September 2025 / Revised: 15 October 2025 / Accepted: 22 October 2025 / Published: 30 October 2025

Highlights

What are the main findings?
  • Flying at lower altitudes increases visual detail but reduces operational efficiency in urban environments.
  • Calibrated image overlap improves spatial accuracy without generating redundant data.
What are the implications of the main findings?
  • The results provide practical guidelines to balance image quality, flight time, and resources in photogrammetric missions.
  • The findings support standardized UAV protocols for reliable and cost-effective urban mapping.

Abstract

Unmanned Aerial Vehicles are increasingly used for urban photogrammetry, yet best-practice guidance on flight-planning parameters in dense city blocks remains limited. However, the quality of generated orthomosaics is strongly dependent on the proper configuration of flight parameters, highlighting the need for evidence-based guidance in consolidated urban environments. This study evaluated the impact of various flight configurations on orthomosaic visual quality and operational efficiency. A total of 96 automated flights were conducted over a 1.5-hectare urban area, systematically varying height, frontal overlap, lateral overlap, camera angle, and flight pattern. Orthomosaic photogrammetric reconstructions were generated and assessed using a multi-criteria scoring system based on the image processing time and the visual clarity of control targets. Results show that a flight height of 60 m, 70% frontal overlap, 80% lateral overlap, nadir 90 ° camera angle, and a grid flight pattern provide the best balance between image quality and operational efficiency. Lower heights improved visual detail but increased processing time, while excessive overlaps did not necessarily enhance final image quality.

1. Introduction

In recent years, the technological revolution in geospatial data acquisition has been significantly driven by the adoption of Unmanned Aerial Vehicles (UAVs), also known as drones [1]. These tools have transformed the way topographic and photogrammetric surveying is conducted, particularly in urban environments, by allowing the capture of high-resolution images quickly and at considerably lower costs than traditional methods, such as total stations or the Global Navigation Satellite System (GNSS) [2].
In architecture and construction, UAV integration has improved workflows that previously relied on costly equipment and lengthy survey times. Although traditional surveying systems are still in use, they have some notable disadvantages [3]. These include security risks and limited accessibility in hard-to-reach areas, longer turnaround times and higher operating costs, as well as difficulties in surveying complex geometries [4]. By contrast, UAVs provide a safer and more cost-effective solution, enabling faster and less expensive data collection [2].
The ability to conduct scheduled autonomous flights, together with the use of specialized software for image processing, has facilitated the acquisition of accurate and up-to-date information for decision making [5]. Among the main products generated from these images are orthomosaics, Digital Surface Models (DSMs), Digital Terrain Models (DTMs), and dense point clouds (DPCs), all of which are valuable for urban planning, design, and management [6]. Compared with satellite imagery, UAV-based photogrammetry offers near-real-time data with higher spatial detail.
The quality of the photogrammetric products generated by UAVs is highly dependent on proper flight planning. Parameters such as flight height, image overlap, camera tilt, and flight pattern directly influence the spatial resolution of the images and the accuracy of the generated models [7]. The flight height influences the level of detail captured; overlap ensures the continuity necessary for Three-dimensional (3D) reconstruction; camera tilt and flight pattern determine the coverage and quality of the information obtained [8].
In urban contexts, for instance, in a medium-sized Latin American city as a representative case, the correct selection and optimization of these parameters is essential to maximize the visual quality of orthomosaics and reduce processing times. Recent studies such as those by Jara Valdebenito [5] and Cruz Toribio [6] demonstrate that an appropriate flight configuration not only improves the quality of the generated products but also reduces the demand for computational resources.
In this context, this study aims to analyze the impact of different flight parameters on the quality of photogrammetric products generated with UAVs in a consolidated urban environment with typical lots. To this end, a comparative experiment is conducted in the city of Cuenca (Ecuador), evaluating different configurations of flight height, frontal overlap, lateral overlap, camera tilt, and flight pattern, to establish guidelines for more efficient flight planning in terms of both time and quality. Highlights of this work are listed as follows:
  • Investigating how different UAV fight configurations influence the visual quality and efficiency of urban orthomosaic generation.
  • Analyzing various flight parameters such as flight height, image overlap, camera angle, and flight pattern.
  • Providing guidelines for optimizing photogrammetric missions in dense urban environments.
The subsequent sections of this paper are organized as follows: Section 2 summarizes state-of-the-art works related to UAV flights, digital photogrammetry, flight parameters, and orthomosaic visual quality and processing time. Steps of research methodology, evaluation metrics, and experiments are presented in Section 3. Section 4 presents the results of experiments, statistical analysis of relationships between different flight parameters, and discussions of the optimization of UAV flight parameters. Conclusions are shown in Section 5.

2. Related Work

2.1. UAV in Construction and Urban Planning

UAVs transitioned from their initial military applications to become indispensable tools across diverse civilian, industrial, and commercial sectors. Their remarkable ability to capture high-resolution geospatial data has significantly improved workflows in architecture, civil engineering, and urban planning [9,10,11]. This technological shift addresses long-standing challenges in data acquisition, offering unprecedented levels of efficiency and detail.
UAV are particularly well-suited for urban surveys, especially in challenging or extensive terrains, due to their capacity for autonomous, high-precision flights and detailed imagery acquisition [2]. Beyond simple image capture, UAVs enable the creation of comprehensive 3D models, digital twins, and site analysis crucial for every stage of a construction project, from initial planning and design to progress monitoring and final inspection. The integration of UAV technology streamlines data collection, leading to improved efficiency, accuracy, and safety in project planning and execution, significantly reducing the need for dangerous manual inspections [12,13].
Multi-rotor UAVs, in particular, offer significant advantages in urban environments due to their superior maneuverability, stability, and Vertical Takeoff and Landing (VTOL) capabilities. These characteristics enable precise hovering and accurate trajectory control, which are beneficial for capturing thorough information in areas characterized by complex urban canyons, low-rise buildings, and relatively flat topography [14,15]. Their ability to navigate confined spaces and perform close-range inspections makes them ideal for tasks such as facade inspection, structural assessment, and interior mapping when equipped with appropriate sensors [16]. This adaptability facilitates the acquisition of homogeneous, high-quality data, crucial for generating reliable cartographic products and precise 3D models [17].
Meanwhile, fixed-wing UAVs are primarily employed for vast areas due to their higher operational speeds and longer endurance [18]. Emerging research highlights their growing role in urban mapping, regional infrastructure monitoring, and environmental surveys, particularly when integrated with advanced sensors for thermal, multispectral, or Light Detection and Ranging (LiDAR) data acquisition [19,20,21]. Their ability makes them suitable for large-scale projects, complementing the capabilities of multi-rotor systems by providing a broader contextual overview [18].
Moreover, the legislation regarding UAV flights within urban environments is also studied in Europe and the USA. For example, the European Union Aviation Safety Agency (EASA) adopted a harmonized and risk-based regulatory approach, classifying operations into three categories based on low, medium, and high risks and regulating flight region, time, and altitude to ensure safe and efficient access to urban airspace in U-space concept [22]. Meanwhile, the Federal Aviation Administration (FAA) proposed UAV Traffic Management (UTM) regulation to integrate UAVs into existing airspace safely, and UAVs must be managed in real-time to avoid collisions and ensure safe operations in urban areas [23]. Furthermore, Both EASA and FAA legislations regulate how UAV operators must handle personal data collected during flights for privacy and data protection [24].
Recent advancements in multi-rotor technology include enhanced obstacle avoidance systems, Real-Time Kinematic (RTK) and Post-Processed Kinematic (PPK) GPS integration for improved absolute positioning precision without extensive Ground Control Points (GCPs), and extended flight times, further improving their applicability in complex urban settings [25,26,27]. The growing sophistication of multi-rotor platforms allows for rapid deployment and data acquisition, making them ideal for time-sensitive construction progress monitoring, volumetric calculations, and post-disaster assessments [1,28].
The choice between multi-rotor and fixed-wing UAV often depends on the project’s scale, required detail, and operational environment, with hybrid systems emerging to combine the advantages of both [29]. Therefore, multi-rotor UAVs are selected for urban photogrammetric surveys in this study, considering the cost-effective and high-quality requirements of urban mapping.

2.2. Digital Photogrammetry Applied to UAVs

Digital photogrammetry is a well-established geomatic technique that leverages principles of geometry and computer vision, deriving accurate 3D information about objects and surfaces from 2D images [30]. In the context of UAV operations, photogrammetry relies on the acquisition of multiple overlapping images, which are then processed using sophisticated Structure-from-Motion and Multi-View Stereo algorithms [31,32]. These algorithms are fundamental to reconstructing scenes from unordered image collections, making them highly suitable for UAV-based data.
Structure-from-Motion (SfM) algorithms are designed to automatically identify common feature points across a series of overlapping images. The robustness of SfM allows for the processing of images captured from various angles and heights, such as complex urban environments and high-relief terrains [33]. Through a robust bundle adjustment process, SfM simultaneously determines the camera’s precise position and orientation for each image, as well as the coordinates of the observed features [34,35]. This initial output, typically a sparse point cloud, serves as the geometric foundation for subsequent detailed processing steps.
Following SfM, Multi-View Stereo (MVS) algorithms perform dense image matching and depth estimation, leveraging the known camera positions and orientations to reconstruct a high-quality 3D model of the scene. This process typically involves projecting pixels from multiple images into 3D space and identifying corresponding points to build a dense representation [36]. These dense point clouds serve as the foundation for subsequent processing steps, leading to the generation of a variety of accurate digital models, such as Digital Surface Models (DSMs), Digital Terrain Models (DTMs), and high-resolution orthomosaics [37,38]. For example, SfM and MVS can be used to create detailed 3D models of historical monuments, statues, and facades, documenting sites that are undergoing restoration or are at risk of environmental degradation [39].
The ultimate quality and performance of these photogrammetric products is contingent upon several critical factors. These include the resolution of the acquired images, the spatial distribution of the viewpoints during the flight, the radiometric quality of the imagery, and the meticulous calibration of the UAV’s onboard sensors [40,41]. The strategic use of GCPs remains a crucial aspect for achieving high absolute precision and to georeference the derived models to a specific coordinate system. While advancements in direct georeferencing reduce their number, GCPs provide an independent position check and anchor the model to real-world coordinates, minimizing error propagation [42,43].
Recent studies emphasize the importance of robust camera calibration, including focal length, principal point, lens distortions, and the camera’s position and orientation, as well as self-calibration techniques [44,45]. Therefore, it is necessary to calibrate the optimal parameters for the UAV flights in order to improve the quality of output products.

2.3. Flight Parameters

Effective planning of photogrammetric UAV flights requires careful selection of parameters that directly affect both the quality of the final products and the efficiency of the data acquisition and processing workflow. These parameters must be optimized to achieve the desired level of detail and accuracy while maintaining operational feasibility.
-
Flight Height: Flight height, defined as the vertical distance between the UAV and the ground, is a primary determinant of spatial resolution. Lower altitudes provide higher-resolution images and finer Ground Sample Distance (GSD), yielding more detailed geometric information. However, they also increase the number of images required to cover a given area, enlarging dataset size and processing time [12]. For flat urban areas, a balance between resolution and efficiency is recommended [14]. Best practices suggest selecting height according to the specific mapping objectives, the sensor characteristics, and the target GSD [46]. For instance, in dense urban areas, flights at lower altitudes help preserve the desired GSD and minimize occlusions, while combining nadir and oblique imagery can capture facades and courtyard walls effectively [47].
-
Image Overlap: Image overlap refers to the percentage of scene coverage common between adjacent photographs. High overlap ensures continuity and redundancy, which are essential for robust feature matching and depth calculation in SfM/MVS workflows [48]. Recommended values generally exceed 75% for forward overlap and 70% for lateral overlap [31]. Zhong [49] further highlights that optimal overlap ranges between 60% and 80% depending on terrain complexity and urban morphology. Recent developments explore adaptive overlap strategies, dynamically adjusting to real-time variations in topography or building heights to prevent redundant data collection in flat areas while ensuring sufficient coverage for vertical surfaces [33,50].
-
Camera Tilt: Camera tilt, defined as the orientation angle of the sensor relative to the ground, significantly influences the type of information captured. Nadir images are optimal for horizontal surfaces, yielding high planimetric accuracy in orthomosaics and precise mapping of rooftops and ground features [15]. Conversely, oblique imagery captures vertical structures, enhancing 3D city models, architectural documentation, and building inspections [51]. While their contribution is less critical in uniformly low-rise areas, oblique views still enrich model completeness and visualization by integrating facade and contextual details [14]. Increasingly, UAVs integrate multi-camera systems, enabling simultaneous nadir and oblique acquisitions to produce more geometrically consistent urban models [52,53]. For dense environments or street-canyon contexts, oblique captures with tilt angles between 30° and 45° significantly improve vertical reconstruction and mitigate incomplete facades [54].
-
Flight Pattern: The flight pattern, or trajectory followed by the UAV, determines coverage uniformity, acquisition speed, and data processing effectiveness. Grid patterns—comprising parallel flight lines with planned overlap—are the most widely used, providing homogeneous coverage and facilitating accurate orthomosaic and dense point cloud generation [55]. In contrast, linear or unidirectional paths without systematic overlap often result in gaps, making them unsuitable for precise mapping [56]. Advanced planning tools now propose adaptive flight paths that account for terrain relief, building heights, and environmental conditions (e.g., wind), thereby minimizing redundancy and maximizing coverage [57,58]. For specialized applications, such as cultural heritage documentation or detailed architectural surveys, circular or spiral trajectories around specific structures enable complete multi-angle coverage [59].
Since flight parameters such as altitude and overlap directly determine GSD, redundancy, and spatial completeness of the imagery, they are decisive for both the visual quality of orthomosaics and the computational efficiency of the workflow. Their optimization is therefore critical for balancing accuracy and processing costs in urban photogrammetric surveys.

2.4. Orthomosaic Visual Quality and Processing Time

In architecture and construction projects, both the visual quality of orthomosaics and the efficiency of the processing workflow are fundamental to informed decision making [9]. High-quality orthomosaics—characterized by sharp details, accurate color representation, minimal distortions, and precise geometric fidelity—provide reliable visual and metric information for site analysis, progress monitoring, as-built documentation, and stakeholder communication. Optimizing flight parameters to capture sharp, high-resolution imagery while minimizing processing times significantly enhances productivity and the overall quality of derived geospatial products [14].
This approach differs from studies that prioritize absolute georeferencing accuracy primarily through the extensive use of Ground Control Points (GCPs). Instead, emphasis is placed on achieving high visual quality and operational efficiency, thereby simplifying and accelerating both data acquisition and processing, often with reduced reliance on GCPs [31]. While GCPs undoubtedly improve absolute accuracy by anchoring the photogrammetric model to a known coordinate system and reducing systematic errors, their establishment and surveying are time- and labor-intensive.
Recent advances in direct georeferencing using high-precision Global Navigation Satellite System/Inertial Measurement Unit (GNSS/IMU) technology onboard UAV are reducing the need for dense GCP networks, particularly in applications where relative precision, visual clarity, and rapid deployment are paramount [60]. For instance, mapping a 2 km2 construction site at 2 cm GSD might traditionally require 12–15 well-distributed GCPs, demanding approximately 3–4 h of fieldwork for placement and surveying. With RTK/PPK GNSS systems integrated with high-grade IMUs, this requirement may be reduced to only 2–3 checkpoints, cutting field time by more than 70% [61]. Processing efficiency is also enhanced: datasets that might require around 4 h for aerial triangulation without GNSS/IMU support can often be processed in about 90 min when precise exterior orientation is available, as the need for extensive tie-point searches and full bundle adjustments is diminished [62].
Processing time for orthomosaic generation is inherently dependent on factors such as survey area, number of images, target resolution, and available computational resources. The computational demands of SfM/MVS workflows—particularly during dense point cloud generation and orthomosaic creation—are considerable, often requiring substantial RAM, CPU/GPU power, and storage capacity [38]. To address these challenges in large-scale projects, cloud-based platforms and parallel computing techniques are increasingly employed. These solutions accelerate turnaround times by distributing computational loads across multiple servers, while granting access to high-performance infrastructure without the need for significant hardware investments [63,64].
As shown in Figure 1, a triangular trade-off exists between resolution, processing time, and computational resources in urban photogrammetric surveys. Flight parameters mediate this trade-off: higher overlaps, complex flight patterns, and oblique imagery enhance resolution and model completeness but increase processing demands, whereas higher altitudes and simple grid missions reduce time and computational load at the expense of resolution and facade detail.
Continuous advances in photogrammetric software, such as optimized image matching algorithms, faster bundle adjustment routines, and hardware-accelerated workflows, continue to improve the quality of orthomosaics and reduce processing times. These advances are making UAV photogrammetry increasingly accessible and practical, enabling applications ranging from rapid response mapping to daily construction monitoring.

3. Methodology

3.1. Definition of the Study Area and Selection of the Typical Lot

The study was conducted in a consolidated urban area of Cuenca, Ecuador, specifically in the southern sector of the El Batán parish, in the Ex CREA area. As shown in Figure 2a, the selected polygon corresponds to a complete block bounded by Avenida México, Avenida Unidad Nacional, Calle Colombia, Calle Argentina, and Calle Padua, covering an area of approximately 1.5 hectares. This area was selected for its representative urban morphology, predominantly flat topography, and the coexistence of residential and commercial uses.
Figure 2b shows the interior of the study block, where 33 properties were identified with varying typologies in terms of shape, size, and land use. To characterize a “typical lot”, the Cuenca Municipal Decentralized Autonomous Government (GAD) geovisor was used. This geovisor provides detailed cadastral information, such as the property code, frontage dimensions, land area, and built-up area for each lot.
The GAD Cuenca GeoViewer was used to define the typical lot. This tool provides data such as cadastral key, frontage, depth, and area of each property. These data were tabulated and statistically analyzed to determine the average values representative of the block’s properties.
The results indicate that the typical lot has an average frontage of 13.60 m, a depth of 20.90 m, and an area of 301.88 m2;. These dimensions coincide with the predominant land subdivision and occupation typologies described by Hermida [65], who analyzes the morphotypological patterns of urbanization in intermediate Andean cities such as Cuenca (Ecuador).
The choice of this typical lot is justified by its representativeness of the urban configuration in consolidated sectors of the city. This approach allows the results derived from the photogrammetric analysis to be extrapolated to a significant number of similar urban situations, favoring their application in planning, urban design, and architectural survey projects in comparable contexts.

3.2. Parametric Sweep

Flights were carried out using a DJI Phantom 4 PRO V2, selected to enhance orthomosaic visual quality and operational efficiency. An aerial optimized F2.8 wide-angle lens with a 24 mm equivalent focal length was used to capture sharp and vivid photos at high resolutions. A 6-camera navigation system was used to constantly calculate the relative speed and distance between the aircraft and the objects. A flight autonomy made up an infrared sensing system, dual-band satellite positioning of GPS and GLONASS, and redundant IMUs and compasses, enhancing the reliability and efficiency of flight. An infrared sensing system was used to measure the distance between the aircraft and any obstacles using infrared 3D scanning. Therefore, GCPs and other georeferencing factors were not considered for evaluating flight efficiency due to these built-in functions.
In order to identify the optimal parameters for photogrammetric surveys with UAVs in medium-sized urban areas of approximately 1.5 ha, an experiment based on a parametric sweep was designed, comprising 96 flight combinations. The methodology included systematic flight planning, standardized image processing, and multi-criteria evaluation of the products, emphasizing orthomosaic visual quality and operational efficiency.
Each flight plan was identified by a structured nomenclature that encodes the parameters used: flight height (H60, H90), frontal overlap (TF70, TF75, TF80), lateral overlap (TL65, TL70, TL75, TL80), camera angle (A45, A90), flight pattern type (TN: normal; TG: grid), and number of photographs obtained (F35). For example, the code “01_H60_TF70_TL65_A90_TN_F35” represents the first flight at 60 m altitude, with 70% frontal overlap, 65% lateral overlap, 90 ° camera angle, normal flight pattern, and 35 images captured. All parameters of 96 UAV flights are shown in Table 1 and Table 2.
All flights were planned using Dronelink software with the Growth subscription, which defined the polygon of the study area and automatically configured the parameter combinations. During planning, Dronelink provided estimates of flight time and number of photographs, which were later compared with actual field values. Environment conditions were confirmed before the take-off of UAV flights to ensure success, and thus, it was chosen to be the windless and sunny days.
The information from each flight was systematized into a structured database that includes operational variables such as flight time and distance traveled, processing times such as aerial triangulation and orthomosaic generation, and three visual quality indicators derived from the generated orthomosaic. The latter were evaluated using a control target located in the study area as shown in Figure 3a, with a 1 m × 1 m black-and-white grid used to evaluate geometric precision and sharpness in each orthomosaic. Total operating times were obtained by summing the flight, aerial triangulation, and orthomosaic generation phases, allowing the overall performance to be evaluated for each configuration.
For each orthomosaic, the sharpness with which the internal subdivisions of the 1 × 1 square meter target, composed of black and white patterns of decreasing sizes (50 cm, 25 cm, 12.5 cm, and 6.25 cm), could be distinguished, was evaluated, as shown in Figure 3b. This analysis provided an objective measure of the level of visual resolution achieved in each configuration.

3.3. Image Processing

After completing the flights, the image set from each experimental configuration was organized. Images were stored in individual folders using the standardized nomenclature described previously, ensuring traceability and consistency.
Each set of images was aerially triangulated using iTwin Capture Modeler Master software in the version 24.01.07.4760, because it has higher integration with Geovisor and UAV system in this investigation than other platforms, such as Pix4D and Agisoft Metashape. This process involved automatically estimating camera interior and exterior orientation parameters and establishing geometric correspondences between consecutive images. Processing consistency was ensured by maintaining the software’s default settings.
From the aerial triangulation results, orthomosaics were generated for each of the 96 flight configurations. Orthorectification was also performed using iTwin Capture Modeler Master, preserving the original capture resolution and exporting the products in the georeferenced .TIF format. Each orthomosaic was stored in its corresponding folder using the flight nomenclature, enabling direct association with experimental parameters and evaluation metrics.
Moreover, all UAV image processing and orthomosaic generation were carried out on a workstation running Windows 10, equipped with an Intel Core i7-12700 CPU (20 threads), 128 GB of DDR4 RAM at 3600 MHz, a 2 TB storage unit, and an NVIDIA RTX A2000 graphics card (6 GB VRAM). This configuration provided the computational capacity required for efficient image processing and ensured reproducible conditions across all experiments.

3.4. Evaluation of the Quality of Orthomosaics

The evaluation of the generated orthomosaics focused on three key criteria: (i) accuracy in locating the control target within the orthomosaic, (ii) level of detail visible on the target, and (iii) overall orthomosaic visual quality. These criteria were applied both to compare different flight configurations and to quantify the impact of each parameter combination on the photogrammetric results.

3.4.1. Target Location Accuracy

Each orthomosaic generated in .TIF format was loaded to CivilCAD 2025 software, where the target previously installed in the field was located. It is important to note that this process was not intended to metrically georeference the product, but only to locate the image on the print layout and evaluate the target’s relative location consistency. All orthomosaics were exported to PDF under consistent conditions: fixed scale and letter format, which allowed for standardized visual comparison.
Based on the relative location of the target within the layout, a qualitative rating between 1 and 3 was assigned:
  • The target is not visible within the print area.
  • The target appears partially visible or cut off. Example as shown in Figure 4a:
  • The target is completely visible and centered in the layout. Example as shown in Figure 4b:
    Figure 4. Examples of target location accuracy. (a) Target partially visible. (b) Target completely visible.
    Figure 4. Examples of target location accuracy. (a) Target partially visible. (b) Target completely visible.
    Drones 09 00753 g004
This evaluation identified variations attributable to errors in the orthomosaic’s georeferencing due to different parameter combinations. Although this study does not address georeferencing or ground control points, the shifting of the obtained images is evident.

3.4.2. Level of Visual Detail on the Target

A rating scale of 1 to 5 was applied, based on the smallest clearly distinguishable subdivision size:
  • Only the general outline of the target is visible, as shown in Figure 5.
    Figure 5. Only the general outline visible.
    Figure 5. Only the general outline visible.
    Drones 09 00753 g005
  • The 50 cm × 50 cm subdivision is clearly visible, as shown in Figure 6a.
  • The 25 cm × 25 cm subdivision is clearly visible, as shown in Figure 6b.
  • The 12.5 cm × 12.5 cm subdivision is clearly visible, as shown in Figure 6c.
  • The 6.25 cm × 6.25 cm subdivision is clearly visible, as shown in Figure 6d.
    Figure 6. Examples of visual detail on the target. (a) 50 × 50 cm subdivision visible. (b) 25 × 25 cm subdivision visible. (c) 12.5 × 12.5 cm subdivision visible. (d) 6.25 × 6.25 cm subdivision visible.
    Figure 6. Examples of visual detail on the target. (a) 50 × 50 cm subdivision visible. (b) 25 × 25 cm subdivision visible. (c) 12.5 × 12.5 cm subdivision visible. (d) 6.25 × 6.25 cm subdivision visible.
    Drones 09 00753 g006
This scale allowed us to quantify the degree of effective resolution achieved by each combination of flight height, overlap, camera angle, and flight pattern.

3.4.3. General Quality of the Orthomosaic

Finally, a comprehensive visual evaluation of each orthomosaic was performed, considering the sharpness, homogeneity, absence of aberrations, and overall consistency of the product. This analysis considered typical effects such as geometric distortions, blurred areas, or inconsistencies due to poor overlapping.
A qualitative rating of 1 to 5 was assigned using the following criteria. 1: Significant visual deficiencies, although the area is distinguished as of interest, as shown in Figure 7a. 2: Severe aberrations affecting readability, as shown in Figure 7b. 3: Moderate aberrations, as shown in Figure 7c. 4: Small aberrations, as shown in Figure 7d. 5: High quality without visible aberrations, as shown in Figure 7e.
These three metrics were subsequently used in a multi-criteria analysis to determine the optimal flight configurations, in terms of balance between visual quality and efficiency in orthomosaic processing time.

4. Results

Once the qualification stage was completed, a comparative analysis of the results obtained from the 96 flights was conducted, considering not only the scores assigned to each orthomosaic but also the associated operational variables, such as flight time, number of photographs captured, and processing time. Table 3 and Table 4 summarize the results obtained from the 96 UAV flights. The first column corresponds to the flight identifier, which is directly related to the classification presented in Table 1 and Table 2. The second column reports the number of photographs captured during each mission, while the third column provides the flight time as recorded by Dronelink. The fourth column shows the total image processing time obtained in the Bentley iTwin platform, including both aerotriangulation and orthophoto generation. The fifth, sixth, and seventh columns correspond to the Likert-scale evaluations of location accuracy, target quality, and overall orthophoto quality, respectively, (as defined in Section 3.4). Finally, the last column presents the overall rating, computed as the average of the three individual quality measures: Equation (1)
Overall Rating = Location Accuracy + Target Quality + Orthophoto Quality 3
As an introduction to the analysis of the results, two flights are explained, one with a score of 5/5 and a second flight with a score of 1/5.
Flight orthomosaic rated 5/5, as shown in Figure 7e. This image shows the entire study area in good quality, with quality measures of 5, 5, 5, respectively.
Flight orthomosaic with a 1/5 rating, as shown in Figure 7f. This image does not show the entire study area, as the camera angle was 45 degrees, causing the camera to capture a larger area outside the study area with the quality measures of 1, 1, 1, respectively.
Relationships between flight parameters, orthomosaic visual quality, and operational efficiency are then analyzed and discussed using statistical analysis on the results in table.

4.1. Relationship Between Flight Parameters and Orthomosaic Visual Quality

4.1.1. Flight Height

Figure 8 shows that the highest scores were concentrated in flights taken at a height of 60 m, although with greater dispersion. In contrast, at 90 m, more homogeneous but generally lower scores were observed. This suggests that capturing at lower heights allows for greater detail and precision, resulting in better-quality orthomosaics, albeit at a higher computational cost. The Correlation Coefficient of statistical analysis on test results was −0.469, and the p-value of it was less than 0.001, indicating that the negative trend between flight height and overall rating was fair and the significance of it was strong.

4.1.2. Frontal Overlap

Contrary to the expectation that greater frontal overlap implies better results, the data in Figure 9a show that the highest-scoring flights were concentrated in flights with overlaps of 70% and 75%. In contrast, values of 80% did not necessarily offer better results and even presented lower scores. This indicates the existence of a threshold beyond which increasing the overlap does not improve the visual quality of the product and, instead, may generate unnecessary redundancies. The Correlation Coefficient between frontal overlap and overall rating was −0.259, and the p-value was 0.011 according to the statistical test, which means that the negative trend was weak with a strong significance.

4.1.3. Lateral Overlap

Figure 9b reveals that lateral overlap has a less direct correlation with orthomosaic visual quality. Multiple flights with overlaps of 65%, 70%, 75%, and 80% were identified as achieving the highest rating. These results suggest that greater lateral overlap does not contribute more consistently to improving the quality of the generated product. The Correlation Coefficient of lateral overlap vs. overall rating was 0.126, and the p-value in statistical analysis was 0.218, showing that there was almost no correlation between these two factors, and the significance of their trend was also weak.

4.1.4. Camera Angle

Figure 9c shows that the best scores were obtained with perpendicular camera angles ( 90 ° ). This configuration ensures that the images are centered and aligned with the target surface, minimizing geometric distortions. Conversely, the use of oblique angles ( 45 ° ) led to the capture of images outside the defined area, generating distortions and decreasing the quality of the orthomosaics obtained. The Correlation Coefficient of camera angle and overall rating was 0.706 and the p-value was less than 0.001, and the statistical analysis indicates that the positive trend between the camera angle and overall rating was in a relatively high level and the significance of this correlation was quite strong.

4.1.5. Number of Photographs

Figure 9d shows that a greater number of photographs can guarantee better orthomosaic visual quality. While a low number of images is generally associated with low ratings, there is a high dispersion in cases with many photographs, suggesting that other factors, such as the quality of the overlap or the camera angle, have a more significant impact on the final product quality. Meanwhile, the Correlation Coefficient and p-value between the number of photographs and overall rating were 0.609 and less than 0.001, respectively, and thus, the positive trend and significance of their relationship were both quite strong.

4.1.6. Flight Pattern

Figure 10 shows that flights performed with a grid-type pattern obtained slightly higher scores than flights carried out with a normal pattern. This is due to the greater number of photographs, which translates into a greater amount of detail in the orthomosaics. Statistical analysis results show that the Correlation Coefficient between flight pattern and overall rating was 0.148, while the p-value was 0.149, which means that there was almost no trend between these two factors, and the significance of their correlation was weak as well.

4.2. Relationship Between Operational Efficiency and Orthomosaic Visual Quality

Figure 11a shows that longer processing times do not necessarily translate into better scores in a certain range. The highest average score was achieved in the processing range of 1041 to 2798 s and approximately 17 to 46 min, while times greater than 3700 s and more than 1 h were associated with moderate or low scores. This reinforces the idea that efficiency in data capture and processing should not be sacrificed in favor of excessive data capture. Additionally, the Correlation Coefficient and p-value of processing time and overall rating were 0.481 and less than 0.001, which indicate that there is a positive trend between the processing time and overall rating with a strong significance.

4.3. Relationship Between Flight Parameters and Operational Efficiency

4.3.1. Flight Time

Finally, Figure 11b shows that a longer flight time is related to a longer processing time, as other parameters also play a role. This is because the longer flight time is related to a longer distance traveled and thus a larger number of photos, resulting in a longer processing time. In addition, the Correlation Coefficient between flight time and processing time was 0.811, and its p-value was less than 0.001, indicating that the positive trend between these two was very high with a very strong significance on this correlation.

4.3.2. Flight Height

The data presented in Figure 12 confirm that processing time tends to increase with lower flight height, such as at 60 m. This is because flying at lower heights requires capturing a greater number of photographs to cover the same area, which increases the software’s workload. In contrast, at 90 m, shorter and more consistent processing times were recorded due to the smaller number of photographs generated. Moreover, statistical analysis suggests that the Correlation Coefficient of flight height and processing time was −0.478, and the p-value was less than 0.001, so that the negative trend between them was fairly high and their significance was strong.

4.3.3. Flight Pattern

The analysis of Figure 13 shows that normal flights tend to require less processing time than grid flights. This difference is primarily due to the greater number of images captured in grid flights, which increases the computational effort. Although in certain sections of the graph both types present similar times, at the highest values, a clear trend toward longer processing times is observed for grid flights. Furthermore, the Correlation Coefficient between flight pattern and processing time was 0.531 and the p-value of it was less than 0.001, meaning that there was a strong positive trend between flight pattern and processing time and the significance of it was quite strong.

5. Conclusions

The statistical analysis of the results provides insights into balancing image quality and operational efficiency when defining UAV flight parameters for urban photogrammetric surveys. Flight parameter configurations should be defined according to the type of cartographic product required and the specific environmental conditions. Based on the experimental results and comparative analysis, the following conclusions can be drawn:
When the objective is to cover large areas and the level of detail in the orthomosaic is not a priority, higher flight heights are recommended, as they reduce the number of captured images and shorten processing times. However, maximum height is constrained by national aviation regulations. In contrast, lower flight heights contribute to greater detail and accuracy, resulting in higher-quality orthomosaics, albeit at the cost of increased computational requirements.
The statistical analysis shows little correlation between frontal and lateral overlaps and the overall quality rating. Similarly, the influence of flight pattern on orthomosaic visual quality is not statistically significant. Therefore, it is advisable to adjust these parameters according to the specific conditions of each UAV survey.
In contrast, three factors show strong correlations with orthomosaic visual quality: camera angle, processing time, and number of photographs. Quality improves significantly when using a nadir (90°) camera angle, because orthomosaic generation benefits from a vertical view aligned with the ground surface. There are also clear positive trends between the number of photographs, processing time, and product quality within a certain range. Thus, it is recommended to capture between 60 and 100 photographs to maintain quality while limiting processing demands.
The analysis also confirms that flight time is strongly related to processing time: longer flights generate larger datasets, which increase computational requirements. Grid flight patterns demand more processing time than normal patterns, although they produce slightly higher-quality results. As expected, greater flight time leads to longer processing times due to the larger number of images required for reconstruction.
For the specific case of surveying a typical lot of approximately 300 m2 in the city of Cuenca (Ecuador), the optimal combination of parameters was a flight height of 60 m, 70% frontal overlap, 80% lateral overlap, a nadir 90° camera angle, and a grid flight pattern. These parameters can also be applied in other cases involving plots with similar characteristics.

Author Contributions

Conceptualization, J.L.-R. and E.J.R.; methodology, J.L.-R.; software, J.L.-R. and C.C.; validation, C.C. and J.L.-R.; formal analysis, J.L., M.B.-R. and G.A.; investigation, C.C. and J.L.-R.; resources, C.C. and J.L.-R.; data curation, C.C. and J.L.-R.; writing—original draft preparation, C.C. and J.L.; writing—review and editing, E.J.R., J.L., M.B.-R. and G.A.; visualization, J.L. and C.C.; supervision, J.L.-R.; project administration, J.L.-R.; funding acquisition, M.B.-R. and G.A. All authors have read and agreed to the published version of the manuscript.

Funding

Marcelo Becerra-Rozas is supported by DI Iniciación PUCV 2025/PUCV/039.725/2025.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author.

Acknowledgments

Jingwei Liu is supported by Concurso Postdoctorado UC 2024/UC/Pr-211-24. The authors are grateful to the editor and anonymous reviewers for their constructive comments and valuable suggestions to improve the quality of the article.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
UAVUnmanned Aerial Vehicles
GNSSGlobal Navigation Satellite System
DSMDigital Surface Model
DTMDigital Terrain Models
DPCDense Point Clouds
3DThree-dimensional
VTOLVertical Takeoff and Landing
RTKReal-Time Kinematic
PPKPost-Processed Kinematic
SfMStructure-from-Motion
MVSMulti-View Stereo
GSDGround Sample Distance

References

  1. Molina, A.A.; Huang, Y.; Jiang, Y. A Review of Unmanned Aerial Vehicle Applications in Construction Management: 2016–2021. Standards 2023, 3, 95–109. [Google Scholar] [CrossRef]
  2. Tucci, G.; Gebbia, A.; Conti, A.; Fiorini, L.; Lubello, C. Monitoring and Computation of the Volumes of Stockpiles of Bulk Material by Means of UAV Photogrammetric Surveying. Remote Sens. 2019, 11, 1471. [Google Scholar] [CrossRef]
  3. AgroTop SL. Cómo los Drones Están Revolucionando la Topografía en 2025. 2025. Available online: https://agrotopsl.es/como-los-drones-estan-revolucionando-la-topografia-en-2025/ (accessed on 14 October 2025).
  4. Verdugo, J.A.Y.; Mocha, J.B.C. Análisis Comparativo Entre Levantamientos Topográficos Con Estatión Total y Vehículos Aéreos No Tripulados 534 (UAVS). Ph.D. Thesis, Universidad Politécnica Salesiana location, Cuenca, Ecuador, 2024. [Google Scholar]
  5. Valdebenito, E.J. Aplicación de Técnicas de Fotogrametría Con Drones Para la Detección y Cuantificación de Deterioro de Pavimentos y su Posible Aplicación en Evaluación de Pavimentos Aeroportuarios. Ph.D. Thesis, Universidad de Concepción, Concepción, Chile, 2025. [Google Scholar]
  6. Toribio, J.C. Cálculo del Índice de Regularidad Internacional (IRI) a través de imágenes obtenidas de un Vehículo Aéreo No Tripulado. Ph.D. Thesis, Universidad de Costa Rica, San Pedro Montes de Oca, Costa Rica, 2022. [Google Scholar] [CrossRef]
  7. Chodura, N.; Greeff, M.; Woods, J. Evaluation of Flight Parameters in UAV-based 3D Reconstruction for Rooftop Infrastructure Assessment. arXiv 2025, arXiv:2504.02084. [Google Scholar] [CrossRef]
  8. Pargieła, K. Optimising UAV Data Acquisition and Processing for Photogrammetry: A Review. Geomat. Environ. Eng. 2023, 17, 29–59. [Google Scholar] [CrossRef]
  9. Jacob-Loyola, N.; Rivera, F.M.L.; Herrera, R.F.; Atencio, E. Unmanned Aerial Vehicles (UAVs) for Physical Progress Monitoring of Construction. Sensors 2021, 21, 4227. [Google Scholar] [CrossRef]
  10. Jofré-Briceño, C.; Rivera, F.M.L.; Atencio, E.; Herrera, R.F. Implementation of Facility Management for Port Infrastructure through the Use of UAVs, Photogrammetry and BIM. Sensors 2021, 21, 6686. [Google Scholar] [CrossRef]
  11. Nwaogu, J.M.; Yang, Y.; Chan, A.P.C.; lin Chi, H. Application of drones in the architecture, engineering, and construction (AEC) industry. Autom. Constr. 2023, 150, 104827. [Google Scholar] [CrossRef]
  12. Siebert, S.; Teizer, J. Mobile 3D mapping for surveying earthwork projects using an Unmanned Aerial Vehicle (UAV) system. Autom. Constr. 2014, 41, 1–14. [Google Scholar] [CrossRef]
  13. Molina, A.L.A.; Huang, Y. Using Unmanned Aircraft Systems for Construction Verification, Volume Calculations, and Field Inspection. Ph.D. Thesis, East Carolina University, Greenville, NC, USA, 2023. [Google Scholar]
  14. Agüera-Vega, F.; Ferrer-González, E.; Martínez-Carricondo, P.; Sánchez-Hermosilla, J.; Carvajal-Ramírez, F. Influence of the Inclusion of Off-Nadir Images on UAV-Photogrammetry Projects from Nadir Images and AGL (Above Ground Level) or AMSL (Above Mean Sea Level) Flights. Drones 2024, 8, 662. [Google Scholar] [CrossRef]
  15. Nex, F.; Remondino, F. UAV for 3D mapping applications: A review. Appl. Geomat. 2014, 6, 1–15. [Google Scholar] [CrossRef]
  16. Lyu, C.; Lin, S.; Lynch, A.; Zou, Y.; Liarokapis, M. UAV-based deep learning applications for automated inspection of civil infrastructure. Autom. Constr. 2025, 177, 106285. [Google Scholar] [CrossRef]
  17. Vangu, G.M.; Miluț, M.; Croitoru, A.C. Use of Modern Technologies Such as UAV, Photogrammetry, 3D Modeling, and GIS for Urban Planning. J. Appl. Eng. Sci. 2025, 15, 165–174. [Google Scholar] [CrossRef]
  18. Parente, D.C.; Felix, N.C.; Picanço, A.P. Use of unmanned aerial vehicle (UAV) in the identification of surface pathology in asphalt pavement. Rev. ALCONPAT 2017, 7, 160–171. [Google Scholar] [CrossRef]
  19. Du, Z.; Yang, Y.; Zhu, J.; Lyu, Y. Online Parameter Estimation for Fixed-Wing UAV Based on DREM Method and Adaptive Control. IEEE Trans. Aerosp. Electron. Syst. 2025, 61, 1363–1376. [Google Scholar] [CrossRef]
  20. Cui, X.; Wang, Y.; Yang, S.; Liu, H.; Mou, C. UAV path planning method for data collection of fixed-point equipment in complex forest environment. Front. Neurorobot. 2022, 16, 1105177. [Google Scholar] [CrossRef] [PubMed]
  21. Yuan, Q. Large-scale mapping using UAV-based oblique photogrammetry. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2025, 48, 183–188. [Google Scholar] [CrossRef]
  22. Agapiou, A. Drones in Construction: A Comparative International Review of the Legal and Regulatory Landscape. Proc. Inst. Civ. Eng. Manag. Procure. Law 2021, 174, 118–125. [Google Scholar] [CrossRef]
  23. Mandourah, A.; Hochmair, H. Analyzing the violation of drone regulations in three VGI drone portals across the US, the UK, and France. Geo-Spat. Inf. Sci. 2024, 27, 364–383. [Google Scholar] [CrossRef]
  24. Lee, D.; Hess, D.J.; Heldeweg, M.A. Safety and privacy regulations for unmanned aerial vehicles: A multiple comparative analysis. Technol. Soc. 2022, 71, 102079. [Google Scholar] [CrossRef]
  25. Ito, T.; Funada, R.; Mochida, S.; Kawagoe, T.; Ibuki, T.; Sampei, M. Design and experimental verification of a hoverable quadrotor composed of only clockwise rotors. Adv. Robot. 2023, 37, 667–678. [Google Scholar] [CrossRef]
  26. Mokhtar, M.; Matori, A.; Yusof, K.W.; Chandio, I.; Duong, V.; Dano, D.U.L. A Study of Unmanned Aerial Vehicle Photogrammetry for Environment Mapping: Preliminary Observation. Adv. Mater. Res. 2012, 626, 440–444. [Google Scholar] [CrossRef]
  27. Kim, S.S.; Suk, J.W.; Lim, E.T.; Jung, Y.H.; Koo, S. Quality Asessment of Three Dimensional Topographic and Building Modelling using High-performance RTK Drone. J. Korea-Acad.-Ind. Coop. Soc. 2023, 24, 11–19. [Google Scholar] [CrossRef]
  28. Yıldız, S.; Kivrak, S.; Arslan, G. Using drone technologies for construction project management: A narrative review. J. Constr. Eng. Manag. Innov. 2021, 4, 229–244. [Google Scholar] [CrossRef]
  29. Guan, S.; Zhu, Z.; Wang, G. A Review on UAV-Based Remote Sensing Technologies for Construction and Civil Applications. Drones 2022, 6, 117. [Google Scholar] [CrossRef]
  30. Atencio, E.; Plaza-Muñoz, F.; Rivera, F.M.L.; Lozano-Galant, J.A. Calibration of UAV flight parameters for pavement pothole detection using orthogonal arrays. Autom. Constr. 2022, 143, 104545. [Google Scholar] [CrossRef]
  31. Sanz-Ablanedo, E.; Chandler, J.H.; Rodríguez-Pérez, J.R.; Ordóñez, C. Accuracy of Unmanned Aerial Vehicle (UAV) and SfM Photogrammetry Survey as a Function of the Number and Location of Ground Control Points Used. Remote Sens. 2018, 10, 1606. [Google Scholar] [CrossRef]
  32. Westoby, M.J.; Brasington, J.; Glasser, N.F.; Hambrey, M.J.; Reynolds, J.M. ‘Structure-from-Motion’ photogrammetry: A low-cost, effective tool for geoscience applications. Geomorphology 2012, 179, 300–314. [Google Scholar] [CrossRef]
  33. Dai, W.; Zheng, G.; Antoniazza, G.; Zhao, F.; Chen, K.; Lu, W.; Lane, S.N. Improving UAV-SfM photogrammetry for modelling high-relief terrain: Image collection strategies and ground control quantity. Earth Surf. Process. Landf. 2023, 48, 2884–2899. [Google Scholar] [CrossRef]
  34. Ullman, S. The Interpretation of Structure from Motion. Proc. R. Soc. London Ser. Biol. Sci. 1979, 203, 405–426. [Google Scholar]
  35. Snavely, N.; Seitz, S.M.; Szeliski, R. Modeling the World from Internet Photo Collections. Int. J. Comput. Vis. 2008, 80, 189–210. [Google Scholar] [CrossRef]
  36. Furukawa, Y.; Ponce, J. Accurate, Dense, and Robust Multiview Stereopsis. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 1362–1376. [Google Scholar] [CrossRef] [PubMed]
  37. Li, J.; Peng, Y.; Tang, Z.; Li, Z. Three-Dimensional Reconstruction of Railway Bridges Based on Unmanned Aerial Vehicle–Terrestrial Laser Scanner Point Cloud Fusion. Buildings 2023, 13, 2841. [Google Scholar] [CrossRef]
  38. Remondino, F.; Barazzetti, L.; Nex, F.; Scaioni, M.; Sarazzi, D. UAV photogrammetry for mapping and 3D modeling-Current status and future perspectives. In Proceedings of the International Conference on Unmanned Aerial Vehicle in Geomatics (UAV-g), Zurich, Switzerland, 14–16 September 2011; Volume XXXVIII-1/C22. [Google Scholar] [CrossRef]
  39. Pepe, M.; Alfio, V.S.; Costantino, D. UAV Platforms and the SfM-MVS Approach in the 3D Surveys and Modelling: A Review in the Cultural Heritage Field. Appl. Sci. 2022, 12, 12886. [Google Scholar] [CrossRef]
  40. Gómez-Zurdo, R.S.; Martín, D.G.; González-Rodrigo, B.; Sacristán, M.M.; Marín, R.M. Aplicación de la fotogrametría con drones al control deformacional de estructuras y terreno. Inf. Construcción 2021, 73, e379. [Google Scholar] [CrossRef]
  41. Luhmann, T.; Robson, S.; Kyle, S.; Boehm, J. Close Range Photogrammetry and 3D Imaging; De Gruyter Brill: Berlin, Germany, 2013. [Google Scholar] [CrossRef]
  42. Martínez-Carricondo, P.; Agüera-Vega, F.; Carvajal-Ramírez, F.; Mesas-Carrascosa, F.J.; García-Ferrer, A.; Pérez-Porras, F.J. Assessment of UAV-photogrammetric mapping accuracy based on variation of ground control points. Int. J. Appl. Earth Obs. Geoinf. 2018, 72, 1–10. [Google Scholar] [CrossRef]
  43. Ferrer-González, E.; Agüera-Vega, F.; Carvajal-Ramírez, F.; Martínez-Carricondo, P. UAV Photogrammetry Accuracy Assessment for Corridor Mapping Based on the Number and Distribution of Ground Control Points. Remote Sens. 2020, 12, 2447. [Google Scholar] [CrossRef]
  44. Avola, D.; Cinque, L.; Foresti, G.L.; Lanzino, R.; Marini, M.R.; Mecca, A.; Scarcello, F. A Novel Transformer-Based IMU Self-Calibration Approach through On-Board RGB Camera for UAV Flight Stabilization. Sensors 2023, 23, 2655. [Google Scholar] [CrossRef]
  45. Kelly, J.; Sukhatme, G.S. Visual-Inertial Sensor Fusion: Localization, Mapping and Sensor-to-Sensor Self-calibration. Int. J. Robot. Res. 2010, 30, 56–79. [Google Scholar] [CrossRef]
  46. Becerra-Rozas, M.; Lemus-Romani, J.; Cisternas-Caneo, F.; Crawford, B.; Soto, R.; Astorga, G.; Castro, C.; García, J. Continuous Metaheuristics for Binary Optimization Problems: An Updated Systematic Literature Review. Mathematics 2023, 11, 129. [Google Scholar] [CrossRef]
  47. Mayer, Z.; Epperlein, A.; Vollmer, E.; Volk, R.; Schultmann, F. Investigating the Quality of UAV-Based Images for the Thermographic Analysis of Buildings. Remote Sens. 2023, 15, 301. [Google Scholar] [CrossRef]
  48. Verhoeven, G. Providing an Archaeological Bird’s-Eye View—An Overall Picture of Ground-Based Means to Execute Low-Altitude Aerial Photography (LAAP) in Archaeology. Archaeol. Prospect. 2009, 16, 233–249. [Google Scholar] [CrossRef]
  49. Zhong, J.; Zhou, Q.; Li, M.; Gruen, A.; Liao, X. A Novel Solution for Drone Photogrammetry with Low-overlap Aerial Images using Monocular Depth Estimation. arXiv 2025, arXiv:2503.04513. [Google Scholar] [CrossRef]
  50. de Luis-Ruiz, J.M.; Sedano-Cibrián, J.; Pereda-García, R.; Pérez-Álvarez, R.; Malagón-Picón, B. Optimization of Photogrammetric Flights with UAVs for the Metric Virtualization of Archaeological Sites. Application to Juliobriga (Cantabria, Spain). Appl. Sci. 2021, 11, 1204. [Google Scholar] [CrossRef]
  51. Wei, Y.; Chu, H.; Guo, Q.; Liao, Y. Three-Dimensional Modeling and Accuracy Evaluation Based on Unmanned Aerial Vehicle Oblique Photogrammetry. In Proceedings of the 3rd International Conference on Computer, Artificial Intelligence and Control Engineering, Xi’an China, 26–28 January 2024; Association for Computing Machinery: New York, NY, USA, 2024; pp. 441–445. [Google Scholar] [CrossRef]
  52. Eckenhoff, K.; Geneva, P.; Huang, G. MIMC-VINS: A Versatile and Resilient Multi-IMU Multi-Camera Visual-Inertial Navigation System. arXiv 2020, arXiv:2006.15699. [Google Scholar] [CrossRef]
  53. Chiabrando, F.; Lingua, A.; Maschio, P.; Losè, L.T. The influence of flight planning and camera orientation in UAVs photogrammetry. A test in the area of Rocca San Silvestro (LI), TUSCANY. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2017, 42, 163–170. [Google Scholar] [CrossRef]
  54. Vacca, G.; Dessì, A.; Sacco, A. The Use of Nadir and Oblique UAV Images for Building Knowledge. ISPRS Int. J.-Geo-Inf. 2017, 6, 393. [Google Scholar] [CrossRef]
  55. Eisenbeiss, H.; Sauerbier, M. Investigation of uav systems and flight modes for photogrammetric applications. Photogramm. Rec. 2011, 26, 400–421. [Google Scholar] [CrossRef]
  56. Maes, W.H. Practical Guidelines for Performing UAV Mapping Flights with Snapshot Sensors. Remote Sens. 2025, 17, 606. [Google Scholar] [CrossRef]
  57. Meng, W.; Zhang, X.; Zhou, L.; Guo, H.; Hu, X. Advances in UAV Path Planning: A Comprehensive Review of Methods, Challenges, and Future Directions. Drones 2025, 9, 376. [Google Scholar] [CrossRef]
  58. de Moraes, F.R.; da Silva, I. Assessment of Submillimeter Precision via Structure from Motion Technique in Close-Range Capture Environments. arXiv 2024, arXiv:2409.15602. [Google Scholar] [CrossRef]
  59. Zhao, R.; Huang, Y.; Luo, H.; Huang, X.; Zheng, Y. A Framework for Using UAVs to Detect Pavement Damage Based on Optimal Path Planning and Image Splicing. Sustainability 2023, 15, 2182. [Google Scholar] [CrossRef]
  60. He, Y.; Yang, W.; Su, Q.; He, Q.; Li, H.; Lin, S.; Zhu, S. Three-Dimensional Real-Scene-Enhanced GNSS/Intelligent Vision Surface Deformation Monitoring System. Appl. Sci. 2025, 15, 4983. [Google Scholar] [CrossRef]
  61. Nasrallah, A.A.; Abdelfatah, M.A.; Attia, M.I.E.; El-Fiky, G.S. Positioning and detection of rigid pavement cracks using GNSS data and image processing. Earth Sci. Inform. 2024, 17, 1799–1807. [Google Scholar] [CrossRef]
  62. Zeybek, M.; Taşkaya, S.; Elkhrachy, I.; Tarolli, P. Improving the Spatial Accuracy of UAV Platforms Using Direct Georeferencing Methods: An Application for Steep Slopes. Remote Sens. 2023, 15, 2700. [Google Scholar] [CrossRef]
  63. Han, S.; Han, D. Enhancing Direct Georeferencing Using Real-Time Kinematic UAVs and Structure from Motion-Based Photogrammetry for Large-Scale Infrastructure. Drones 2024, 8, 736. [Google Scholar] [CrossRef]
  64. Sun, C.; Che, G.; Dong, X.; Zou, R.; Feng, L.; Ding, X. Review on Algorithm for Fusion of Oblique Data and Radar Point Cloud. In Communications, Signal Processing, and Systems, Proceedings of the 12th International Conference on Communications, Signal Processing, and Systems: Volume 2; Wang, W., Liu, X., Na, Z., Zhang, B., Eds.; Springer: Singapore, 2024; pp. 527–535. [Google Scholar]
  65. Hermida, M.A.; Hermida, C.; Cabrera, N.; Calle, C. Urban density as variable of city analysis. The case of Cuenca, Ecuador. EURE Rev. Latinoam. Estud. Urbano Reg. 2015, 41, 25–44. [Google Scholar]
Figure 1. Relationship between “resolution–time–resources”.
Figure 1. Relationship between “resolution–time–resources”.
Drones 09 00753 g001
Figure 2. Study area and typical lots. (a) Study area from Google Earth. (b) Lots according to Geovisor.
Figure 2. Study area and typical lots. (a) Study area from Google Earth. (b) Lots according to Geovisor.
Drones 09 00753 g002
Figure 3. Target used for experimentation. (a) Target of geometric precision. (b) Target of sharpnes.
Figure 3. Target used for experimentation. (a) Target of geometric precision. (b) Target of sharpnes.
Drones 09 00753 g003
Figure 7. Examples of orthomosaic general quality. (a) Significant visual deficiencies. (b) Orthomosaic with severe aberrations. (c) Orthomosaic with moderate aberrations. (d) Orthomosaic with small aberrations. (e) High quality with a 5/5 rating. (f) Low quality with a 1/5 rating.
Figure 7. Examples of orthomosaic general quality. (a) Significant visual deficiencies. (b) Orthomosaic with severe aberrations. (c) Orthomosaic with moderate aberrations. (d) Orthomosaic with small aberrations. (e) High quality with a 5/5 rating. (f) Low quality with a 1/5 rating.
Drones 09 00753 g007
Figure 8. Relationships between flight heights and overall quality. (a) Flight height 60 m vs. overall rating. (b) Flight height 90 m vs. overall rating.
Figure 8. Relationships between flight heights and overall quality. (a) Flight height 60 m vs. overall rating. (b) Flight height 90 m vs. overall rating.
Drones 09 00753 g008
Figure 9. Relationships between flight parameters and overall quality. (a) Frontal overlap vs. overall rating. (b) Lateral overlap vs. overall rating. (c) Camera angle vs. overall rating. (d) Number of photos vs. overall ratin.
Figure 9. Relationships between flight parameters and overall quality. (a) Frontal overlap vs. overall rating. (b) Lateral overlap vs. overall rating. (c) Camera angle vs. overall rating. (d) Number of photos vs. overall ratin.
Drones 09 00753 g009
Figure 10. Relationship between flight pattern and overall quality. (a) Normal type vs. overall rating. (b) Grid type vs. overall rating.
Figure 10. Relationship between flight pattern and overall quality. (a) Normal type vs. overall rating. (b) Grid type vs. overall rating.
Drones 09 00753 g010
Figure 11. Relationships between operational efficiency and overall quality. (a) Processing time vs. overall rating. (b) Flight time vs. processing time.
Figure 11. Relationships between operational efficiency and overall quality. (a) Processing time vs. overall rating. (b) Flight time vs. processing time.
Drones 09 00753 g011
Figure 12. Relationships between flight heights and operational efficiency. (a) Flight height 60 m vs. processing time. (b) Flight height 90 m vs. processing time.
Figure 12. Relationships between flight heights and operational efficiency. (a) Flight height 60 m vs. processing time. (b) Flight height 90 m vs. processing time.
Drones 09 00753 g012
Figure 13. Relationships between flight patterns and operational efficiency. (a) Normal type vs. processing time. (b) Grid type vs. processing Time.
Figure 13. Relationships between flight patterns and operational efficiency. (a) Normal type vs. processing time. (b) Grid type vs. processing Time.
Drones 09 00753 g013
Table 1. Parameters of 96 UAV flights—Part 1.
Table 1. Parameters of 96 UAV flights—Part 1.
Flight
Height (m)
Frontal
Overlap (%)
Lateral
Overlap (%)
Camera
Angle (°)
Flight
Pattern
Number of
Photos
160706590TN31
260707090TN32
360707590TN44
460708090TN51
560706590TG56
660707090TG70
760707590TG84
860708090TG105
960706545TN13
1060707045TN17
1160707545TN18
1260708045TN24
1360706545TG28
1460707045TG35
1560707545TG44
1660708045TG51
1760756590TN37
1860757090TN38
1960757590TN48
2060758090TN60
2160756590TG71
2260757090TG78
2360757590TG97
2460758090TG119
2560756545TN15
2660757045TN15
2760757545TN26
2860758045TN28
2960756545TG34
3060757045TG34
3160757545TG46
3260758045TG59
3360806590TN46
3460807090TN46
3560807590TN65
3660808090TN74
3760806590TG93
3860807090TG101
3960807590TG119
4060808090TG149
4160806545TN26
4260807045TN25
4360807545TN27
4460808045TN36
4560806545TG42
4660807045TG50
4760807545TG69
4860808045TG37
Table 2. Parameters of 96 UAV flights—Part 2.
Table 2. Parameters of 96 UAV flights—Part 2.
Flight
Height (m)
Frontal
Overlap (%)
Lateral
Overlap (%)
Camera
Angle (°)
Flight
Pattern
Number of
Photos
4990706590TN13
5090707090TN14
5190707590TN21
5290708090TN25
5390706590TG30
5490707090TG29
5590707590TG36
5690708090TG42
5790706545TN5
5890707045TN5
5990707545TN12
6090708045TN12
6190706545TG15
6290707045TG15
6390707545TG23
6490708045TG21
6590756590TN15
6690757090TN16
6790757590TN24
6890758090TN28
6990756590TG30
7090757090TG33
7190757590TG45
7290758090TG53
7390756545TN6
7490757045TN6
7590757545TN14
7690758045TN13
7790756545TG20
7890757045TG20
7990757545TG24
8090758045TG31
8190806590TN19
8290807090TN19
8390807590TN32
8490808090TN38
8590806590TG38
8690807090TG50
8790807590TG57
8890808090TG66
8990806545TN7
9090807045TN12
9190807545TN17
9290808045TN12
9390806545TG24
9490807045TG24
9590807545TG27
9690808045TG31
Table 3. Results of 96 UAV flights—Part 1.
Table 3. Results of 96 UAV flights—Part 1.
Number
of Photos
Flight
Time (s)
Processing
Time (s)
Location
Accuracy
Target
Quality
Orthophoto
Quality
Overall
Rating
1312399693454.5
23223410413554.5
34429213783455
45131615843554.5
55633416593555
67040718803555
78446522543555
810554527983555
9131823141115
10171845063411
11181887913322.5
122422110783432.5
132828612673243.5
143533215373233
154439619993442.5
165145322073344
173724211953553.5
183823711003555
194828514963555
206031618083555
217138118893355
227841220653454
239749125413554.5
2411954630343555
25151813421115
26151807162321
272621811783332.5
282826312623233
293430416003232.5
303427915743332.5
314638422173333
325941725243343
334625118193353.5
344626320013254
356531226483353.5
367434429663354
379344134353254
3810145935763253.5
3911950541753253.5
4014959451203353.5
412619211001114
422519115213221
432719215113122
443622510471111.5
454230623293231
465035327243232.5
476939837053232.5
483734116653132.5
Table 4. Results of 96 UAV flights—Part 2.
Table 4. Results of 96 UAV flights—Part 2.
Number
of Photos
Flight
Time (s)
Processing
Time (s)
Location
Accuracy
Target
Quality
Orthophoto
Quality
Overall
Rating
49131985782152
50142245722153
51212678933153
52252929753153
533032911512153
542932810853153
553641013393353
564243815373454
5751662141114.5
5851672131111
59122136751121
60122207232231.5
61152798432232.5
62152739863132.5
632333615812132
642132112473142
65151986513152.5
66162236123153
67242609753153
682829310932153
693029211192153
703330212713153
714540416332153
725343718763153
7361672891113
7461673261111
75142108621121
76132188201121.5
772028012533121.5
782027012163121.5
792433514623131.5
803135219182132
81191978153152
82191978153153
833226212663153
843828815263153
853830014663153
865034018013153
875740321393153
88664614201113
8971733953111
90121834001111
91172156073111
92121763561111
932428214763121
942427513733121.5
952730216122131.5
963134118522132
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lemus-Romani, J.; Rueda, E.J.; Becerra-Rozas, M.; Cabrera, C.; Liu, J.; Astorga, G. Optimization of UAV Flight Parameters for Urban Photogrammetric Surveys: Balancing Orthomosaic Visual Quality and Operational Efficiency. Drones 2025, 9, 753. https://doi.org/10.3390/drones9110753

AMA Style

Lemus-Romani J, Rueda EJ, Becerra-Rozas M, Cabrera C, Liu J, Astorga G. Optimization of UAV Flight Parameters for Urban Photogrammetric Surveys: Balancing Orthomosaic Visual Quality and Operational Efficiency. Drones. 2025; 9(11):753. https://doi.org/10.3390/drones9110753

Chicago/Turabian Style

Lemus-Romani, José, Eduardo J. Rueda, Marcelo Becerra-Rozas, Carlos Cabrera, Jingwei Liu, and Gino Astorga. 2025. "Optimization of UAV Flight Parameters for Urban Photogrammetric Surveys: Balancing Orthomosaic Visual Quality and Operational Efficiency" Drones 9, no. 11: 753. https://doi.org/10.3390/drones9110753

APA Style

Lemus-Romani, J., Rueda, E. J., Becerra-Rozas, M., Cabrera, C., Liu, J., & Astorga, G. (2025). Optimization of UAV Flight Parameters for Urban Photogrammetric Surveys: Balancing Orthomosaic Visual Quality and Operational Efficiency. Drones, 9(11), 753. https://doi.org/10.3390/drones9110753

Article Metrics

Back to TopTop