Next Article in Journal
Automatic Electric Tricycles Trajectory Tracking and Multi-Violation Detection
Previous Article in Journal
Detection and Recognition of Bilingual Urdu and English Text in Natural Scene Images Using a Convolutional Neural Network–Recurrent Neural Network Combination with a Connectionist Temporal Classification Decoder
Previous Article in Special Issue
Structural Fault Detection and Diagnosis for Combine Harvesters: A Critical Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

The State of the Art and Potentialities of UAV-Based 3D Measurement Solutions in the Monitoring and Fault Diagnosis of Quasi-Brittle Structures

1
Department of Civil and Environmental Engineering, Politecnico di Milano, Piazza Leonardo da Vinci 32, 20133 Milan, Italy
2
Department of Mechanical Engineering, Politecnico di Milano, Via Privata Giuseppe La Masa 1, 20156 Milan, Italy
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(16), 5134; https://doi.org/10.3390/s25165134
Submission received: 19 June 2025 / Revised: 31 July 2025 / Accepted: 8 August 2025 / Published: 19 August 2025
(This article belongs to the Special Issue Feature Review Papers in Fault Diagnosis & Sensors)

Abstract

The structural health monitoring (SHM) of existing infrastructure and heritage buildings is essential for their preservation and safety. This is a review paper which focuses on modern three-dimensional (3D) measurement techniques, particularly those that enable the assessment of the structural response to environmental actions and operational conditions. The emphasis is on the detection of fractures and the identification of the crack geometry. While traditional monitoring systems—such as pendula, callipers, and strain gauges—have been widely used in massive, quasi-brittle structures like dams and masonry buildings, advancements in non-contact and computer-vision-based methods are increasingly offering flexible and efficient alternatives. The integration of drone-mounted systems facilitates access to challenging inspection zones, enabling the acquisition of quantitative data from full-field surface measurements. Among the reviewed techniques, digital image correlation (DIC) stands out for its superior displacement accuracy, while photogrammetry and time-of-flight (ToF) technologies offer greater operational flexibility but require additional processing to extract displacement data. The collected information contributes to the calibration of digital twins, supporting predictive simulations and real-time anomaly detection. Emerging tools based on machine learning and digital technologies further enhance damage detection capabilities and inform retrofitting strategies. Overall, vision-based methods show strong potential for outdoor SHM applications, though practical constraints such as drone payload and calibration requirements must be carefully managed.

1. Introduction

The safety assessment and preservation of long existing structures, as part of infrastructure networks or historical heritage, relies on the continuous monitoring of their response to external actions [1,2,3,4,5,6]. The purpose of monitoring is to detect anomalies in the response and identify potential damage. The information collected from slender structures like bridges and bell towers is often related to ambient induced vibrations, e.g., due to traffic [4,7,8,9]. Displacements are mostly recovered in the case of massive structures like dams, which are subjected to seasonal changes in temperature and water levels [2,3]. In this context, particularly significant are relative displacements at joints that constitute the main non-linearity source of these facilities [3,10,11]. Relative displacements accompany the evolution of cracks in masonry buildings, a consequence of foundation settlements or non-destructive seismic events [12,13]. In structural assessments, the presence of fractures is a primary concern, as it typically indicates strength degradation. This is often accompanied by a reduction in the overall stiffness, which may lead to excessive deformations. For example, the alkali–aggregate reaction in concrete has been identified as a cause of strength deterioration in dams in [14,15]. In some cases, the source of unrecoverable displacements cannot be easily explained [16]. This highlights the importance of effective and continuous monitoring to obtain deeper insight into the damage state of the examined structure. This review focuses on UAV-based (unmanned aerial vehicle) 3D measurement solutions for structural monitoring. The main goal of these monitoring techniques is twofold: detecting cracks and quantifying their geometry, including crack opening and length, and estimating the relative displacements between adjacent elements of a structure. This manuscript reviews the most recent UAV-based measurement applications, where the emphasis is on the safety assessment of structures undergoing fracture and other degradation phenomena characterized by relative displacements. The search primarily considered recent papers (from the 2000s) on structural health monitoring, with a focus on concrete and fracture-related problems.
Traditionally, structural monitoring of massive quasi-brittle structures relies on instruments such as pendula, collimators, and callipers installed in dams [2] or strain gauges and Linear Variable Differential Transformers (LVDTs) placed in masonry walls [6,12]. In newly built or recently retrofitted facilities, a large network of sensors allows for the recovery of a large amount of data, providing a comprehensive overview of the health status and possible damages during construction. Otherwise, this information can be complemented by measurements performed by non-contact methodologies, such as mounting the relevant equipment on drones [17,18,19,20,21].
Computer-vision methods support the inspection of civil structures and infrastructures. Available techniques allow us to detect cracks, missing parts, and deteriorations and evaluate possible collapse mechanisms in zones of difficult access and in potentially critical situations [22,23,24]. Quantitative information at different accuracy levels, for different applications, can be recovered from full-field measurements depending on the equipment, measurement system, and surface preparation, as illustrated in this review paper. With technological developments, drones have gained widespread popularity as versatile tools for a variety of applications across research and industry. A key advantage of imaging techniques is that the optical sensor can be mounted on a UAV, enabling data collection from various positions.
The collected data can be used to calibrate simulation models of the structure, which typically employ numerical analyses performed mostly within a non-linear finite element context. The availability of displacement maps permits us to overcome the limitations due to uncertain boundary conditions and possible degradation of material properties over time, providing data suitable for implementing sub-structuring and model update strategies [25]. The digital twins of real constructions or part thereof can support the interpretation of the measurements and evidence any anomaly with respect to the expected trend [3,26].
In all cases, emerging procedures based on digital technology and machine learning tools allow us to improve the accuracy and reliability of predictive modelling and real-time monitoring, as well as to plan optimal retrofitting interventions in the event of progressively damaging phenomena [27,28,29,30,31,32].
Different vision-based techniques are reviewed, and their performance is compared in the context of structural assessments. The contribution of this review lies in identifying the strengths of each technique, summarizing the achievable accuracy of each method, and offering insights into potential future developments. In addition, critical factors that significantly affect the performance of each technique are highlighted.

2. Non-Contact Monitoring Techniques

With technological developments, non-contact measurements using vision-based techniques became popular, providing full-field monitoring of the region of interest. Examples of promising methods include 3D digital image correlation (3D-DIC), photogrammetry, and time-of-flight (ToF) sensors. DIC compares a sequence of images to determine the relative motion that have occurred. The setup for a DIC application consists of an image acquisition system, a speckled surface, and suitable lighting [33,34]. Photogrammetry uses a series of images acquired with a single camera from several locations to reconstruct a 3D model of the observed scene [35]. ToF sensors emit several infrared pulses and collect the return times along several directions in the field of view. The distance from the sensor to each point on the surface is then obtained based on the time needed for the pulse to be reflected back from the target [36]. A key advantage of vision-based techniques is that the optical sensor can be mounted on an UAV and can record information about the target from different locations. It should be noted that many optical sensors currently available on the market are capable of capturing frames at very high rates, often exceeding 100 frames per second. Therefore, the sampling rate should not present a challenge for monitoring civil engineering structures, which are often characterized by relatively low natural frequencies.
The ability of drones to move a sensor along the desired trajectory has the disadvantage that the data is not expressed in a single reference system. Furthermore, the rototranslations of drones can only be measured with non-negligible uncertainty. Some review articles discussed drone performance and payload. The authors in [37] mentioned that the UAV motion is sensitive to wind due to its low weight. Control and navigation algorithms can mitigate this issue under constant wind speeds, but fluctuations in wind speed and direction lead to a correction lag and thus affect image quality.
The usage of UAVs in civil engineering was reviewed in [38], dividing its applications into seven main categories: site safety management, city traffic management, disaster damage assessments, wireless communication, power distribution system monitoring, fluvial remote sensing, and structure health monitoring. Another review was performed in [39], which considered papers until 2019.
Various drone models could be employed for monitoring applications, ranging from the simplest and cheapest ones, which can be operated without any professional training or license, to the largest ones, which can be off-the-shelf or even custom devices. Of course, as the size of the UAV increases, so does the allowed payload, leading to the possibility of mounting not only simple digital cameras but also 3D scanners, infrared cameras, lidar sensors, and processing units. On the other hand, the use of larger drones imposes increasingly demanding operator training, as well as more complex flight permission procedures and costs. Sample drones that could be used for UAV measurements are shown in Figure 1; the left-hand side shows a small and cheap drone with an integrated camera, while the right-hand side depicts a larger and custom drone equipped with cameras and 3D scanning devices, and it is controlled by an on-board single-board PC. The review in [40] discusses the criteria for selecting a suitable drone, such as the remote range, the ability to fly without a GPS signal, the presence of source illumination on the drone and payload, and the capability of mounting an additional camera. While the most appropriate drones cost about USD 45,000, other drones with prices ranging from USD 2000 to USD 4000 were deemed acceptable. All these drones have a comparable flight time (20 to 30 min), support an additional flashlight to improve illumination, and have a remote-control range larger than 1000 m. According to the authors in [41], future developments are expected to enhance the reliability of vision-based applications under adverse weather conditions. It was reported in [42] that the main limitations for UAV uptake in civil engineering include flight endurance and payload, legislations and regulations (in some cases a special permission is needed for flying a UAV in public areas), lens distortion, weather conditions, and the volume of data to be processed.
The following applications highlight the importance of each of the aforementioned vision-based techniques in view of their combined use with UAVs.

2.1. Digital Image Correlation (DIC)

The purpose of DIC is to find the displacement and strain fields of the target surface by processing the digital images acquired before and after the deformation occurs. The first image is frequently selected as the reference image that corresponds to the undeformed configuration. The basic concept works by tracking the location of image subsets between the reference and deformed images, as illustrated in Figure 2. The difference between the position of each reference subset and that of the corresponding deformed one yields one displacement vector.
The initial applications of DIC worked in two dimensions, computing pixel displacements along a planar surface [43]. Over the past two decades, research on DIC has grown substantially, driven by improvements in imaging technology and computational power [44]. DIC applications could be either two-dimensional (2D) or three-dimensional (3D). One camera is used in 2D DIC as the in-plane displacements are obtained. Three-dimensional DIC combines the concepts of 2D DIC with stereovision principles, requiring a minimum of two cameras to determine the 3D shape and deformations. DIC is highly effective in laboratory applications under controlled conditions, where cameras are mounted on fixed supports (Figure 3) and additional lighting can be used as needed [45,46,47,48,49,50,51,52,53]. Small crack opening displacements of 0.05 mm and 0.1 mm were measured in [47,53], respectively. These studies demonstrate the high performance of DIC in controlled laboratory conditions; however, when applied using a UAV in field, the accuracy is expected to decrease. This point is discussed in the next paragraphs.
Shear tests were performed on prestressed concrete girders in the laboratory in [48], and it was mentioned that DIC provided more detailed kinematics of the crack compared to the used laser sensors and displacement transducers. The monitored region in Figure 4 corresponds to an extruded polystyrene panel loaded at its midsection. As a result, a crack develops and propagates. A key advantage of DIC is its ability to provide full-field deformation measurements, where both displacements and strains can be used to identify cracks, as illustrated in Figure 4. These displacement fields constitute an important step toward obtaining a digital twin. In [52], for instance, a reduced model of the monitored structure was proposed, with boundary conditions derived from the experimental data obtained using DIC. This approach was further enhanced in [54], where the crack opening displacement values were incorporated into an optimization procedure to determine the fracture parameters.
DIC features many applications for measuring displacements in large specimens. Target-tracking 3D DIC was proposed to examine the vibration of 21 m span bridges subjected to bidirectional loading in the laboratory [55], as well as the vibration of a pedestrian bridge [56]. The vibration responses measured with DIC were used to determine the modal parameters of the bridges. Three-dimensional DIC was also used in the laboratory to detect damage in steel beams using inverse analysis and topology optimization with a finite element model [57]. A 3D DIC system was developed to measure the vibrations of a lighting pole and a wind turbine [58]. When monitoring large structures, the camera’s field of view may not fully cover the region of interest. This limitation can be addressed through image stitching, as achieved by combining DIC with geodetic surveying in [59] and laser tracking in [60].
Cameras can be mounted on a UAV to capture images of the target area, providing a practical solution for large infrastructure that is often difficult to access. In [61], the behaviour of a repaired steel beam in a bridge was examined during and after the removal of the jacking support. In that application, the strains calculated by DIC were larger than those obtained with the strain gauges. The reason for this difference was attributed to an anomaly in the foil gauge measurement. In another study, two cameras mounted on a UAV were used to measure displacements in two bridges over a period of 10 months [19]. A discrepancy of 15% was found between the measurements obtained with DIC and a dial calliper, referring to an absolute error of 0.34 mm. In [62], two cameras mounted on a UAV were used to obtain strains of a rotating wind turbine. In [63], the dust on the bridge beam acted as a speckle pattern. A series of studies was conducted on the use of DIC with UAVs in recent years [64,65,66]. The authors first implemented a vision-based position estimation framework to keep the UAV steady in the desired position [64]. The proposed approach utilizes fiducial markers attached to the surface in order to precisely locate the target area. This system was later validated in a laboratory study [65], as it was used to measure deformations of a prestressed railroad tie loaded in four-point bending. In [66], the system was employed to monitor steel girders in a bridge. The results were not accurate as the obtained strain was larger than the maximum expected value, and it was not possible to distinguish between the loaded and unloaded bridge conditions. This inaccuracy was due to the noise mainly caused by camera vibrations, despite the fact that isolation pads were used. The studies cited above demonstrate that DIC can achieve good accuracy even with cameras mounted on a UAV, even if it is not possible to reach the same accuracy granted by the same DIC system mounted on a tripod, as shown in Table 1.
Some papers presented a literature review about the usage of UAV-aided vision-based measurement techniques on civil engineering structures [67,68,69,70,71]. DIC-based inspection of suspension, masonry, concrete, and steel bridges is considered in [67]. The cost and time saved using vision-based techniques compared to traditional monitoring techniques are analyzed in [69].

2.2. Photogrammetry

In photogrammetry, a series of photos is acquired by one camera from different positions and orientations with respect to the monitored structure. The image processing required to recover the 3D shape of the target involves the following:
  • The identification of a set of features present in each image, typically performed with the scale-invariant feature transform (SIFT) algorithm [72].
  • Robust features identification and tracking, normally obtained with the random sample consensus (RANSAC) algorithm [73], to iteratively compute the rigid motion (rotation and translation) of one camera with respect to another, taking corresponding keypoints in the images as a starting point.
  • Camera alignment and automatic calibration, obtained by finding correspondences between the features in several images [74].
  • Three-dimensional reconstruction, obtained by means of stereo triangulation between pairs of images sharing a set of features [75].
Due to the robustness in the process of feature extraction and matching, no issues are faced when the distance from the camera to the structure slightly varies during acquisition. Photogrammetry, also referred to as structure from motion (SfM) reconstructs scene geometry, camera positions, and orientations without requiring a predefined 3D point network [75], making it particularly advantageous for UAV-based image acquisition [76]. Figure 5 presents a 3D model obtained with photogrammetry, depicting an ancient building in an archeological park. On the left side (top view), the target points that are used for referencing and scaling can be seen. On the right side, images are displayed in 3D at their estimated camera positions, illustrating the multi-view approach in photogrammetry.
Some studies have utilized UAV photogrammetry, primarily focusing on 3D model generation. UAV photogrammetry was employed in various fields, including studies on historical buildings [77,78,79], archeological sites [80,81], and façade structures [82]. In [83], a 3D model of a building was constructed by applying photogrammetric techniques, achieving an accuracy of up to 80 cm. This large error is due to the low resolution of the camera used (1.15 MP) and the large size of the surveyed structure (64 m by 58 m). UAV photogrammetry was used to create 3D models of specific segments in two bridges [84,85]. Additionally, image processing algorithms were implemented to detect cracks in the images and locate them in the 3D models. Under a working distance of 1 m in [84], crack openings varying between 2 mm and 4 mm were identified with a mean error of 13%, while openings less than 1.5 mm could not be accurately detected. The errors dramatically increased when the working distance was doubled. In [85], the generated point cloud had an average accuracy of 0.7 mm, and crack openings extending from 3 mm to 20 mm were identified with a mean error of 1 mm.
In [75], the results obtained with UAV photogrammetry were compared to those obtained with laser scans acquired from fixed station points. Reference points on the surface, known as ground control points (GCPs), were used for orienting and scaling the model. The error between terrestrial laser scanning and photogrammetry was 30 mm. UAV photogrammetry was used to build a 3D model of the Ridracoli dam in Italy [86]. An accuracy of 15 mm was found with respect to measurements by laser scanners. The vertical contraction joints could not be observed by analyzing the horizontal sections of the 3D scene. However, they could be identified in the red–green–blue (RGB) information of the point cloud, being darker than the surrounding region. The accuracy of the main photogrammetry applications is summarized in Table 2. This table highlights the importance of additional processing steps in enhancing the accuracy of photogrammetric measurements. Another study on the Ridracoli dam was performed [87], with the purpose this time on finding the optimal number and location of GCPs which provide the best 3D scene reconstruction. Square markers with a width of 40 cm were attached to the surface and used as GCPs. A mean vertical error of 10 cm (which is less than 1% of the structure height) was considered acceptable. The results showed that the error is sensitive to the number of GCPs along the vertical direction, so placing the GCPs at different elevations increases the accuracy. A qualitative assessment was performed based on a photographic survey of a portion of a dam in Portugal [88]. Image classification was used to detect anomalies, such as recognizing calcium carbonate deposits caused by water leakage below horizontal cracks.

2.3. Tof and Laser Scanners

Time-of-flight (ToF) devices include an infrared (IR) light emitter that illuminates the scene and an imaging sensor that detects the reflected light [89]. The typical setup of a ToF device is illustrated in Figure 6, showing a sketch of the light source illuminating the scene along with the sensor scanning the object in one shot. The integrated depth camera contains a ToF chip, in which the pixels capture the phase shift between the emitted signal and the reflected IR signal [90], which is linked to the time taken by the emitted radiation to reach the target and be reflected back. The distance from the IR emitter and the object is then calculated based on the recorded phase shift [91]. A 2D depth map is produced, where x and y represent the coordinates of each point in pixels, and the depth for each pixel is stored. The measured depth value corresponds to the Z coordinate (in mm) of the sampled point, and it differs from the Euclidean distance [92]. The X and Y coordinates (in mm) of each point can be determined using the camera’s intrinsic parameters, enabling the generation of the 3D point cloud. Some ToF devices feature both a colour camera and a depth camera. By utilizing the intrinsic and extrinsic camera parameters, it is possible to generate a coloured point cloud, in which the colour information is added to the 3D cloud. The advantage of using a coloured point cloud is displayed in Figure 7, where the surface texture is clearly visible and can be used for displacement calculation.
The study in [93] thoroughly evaluated the performance of four ToF devices. The parameters studied included the measurement error, the standard deviation, and the edge noise. Table 3 reports the depth error as a function of the increasing measurement range. The values represent the average error obtained with the four ToF devices. As also observed in [93], the error increases with the operating range.
ToF sensors have been used in a wide range of applications other than civil engineering, including robotics, and clinical purposes such as human motion detection and posture recognition [94,95,96,97,98]. This methodology was also used as a part of a low-cost crack monitoring system [99]. ToF cameras can be mounted on drones to perform measurements of the 3D shape of portions of the target structure. Accurate results were obtained by employing techniques apt at mitigating the sensor vibration and displacement caused by the UAV motion [100]. These approaches allow for the identification of the geometry of defects in concrete, such as spalling and swelling, with an accuracy of less than 4 mm in comparison to a high-resolution 3D scanner [101]. The use of UAVs for the inspection of bridges and tunnels was reviewed in [102], noting that IR-based inspections remain underexplored in this field.
The purpose of this table is to illustrate the variations in maximum achievable accuracy across the different techniques. The second column of this table shows the smallest detected crack or defect reported among all references for each technique. For the corresponding reference, the average crack size is then calculated and reported. For example, the smallest measured crack with UAV-based DIC applications is found in reference [19]; the value reported in the second column represents the average size of the crack measured in that reference. As seen in Table 4, it is evident that DIC possesses the finest measurement accuracy compared to photogrammetry and ToF. Additionally, it is important to highlight that further processing steps are essential in improving the accuracy of the displacements derived with photogrammetry and ToF. Table 4 presents the smallest detectable crack or defect size for each technique.
Laser scanning technology is similar to ToF in terms of basic working principle. Nonetheless, ToF sensors emit a laser pulse that can survey the entire scene at once, while traditional laser scanners emit a signal beam to explore the surface point by point with a scanning mechanism. Due to this reason, laser scanning is not always suitable for UAV-based measurements. However, laser scanning allows for measurements over large distances (hundreds of metres), while ToF sensors have measuring ranges limited to a few metres. Scans from multiple locations are integrated together in laser scanning, and the point clouds from each local coordinate system are registered into a global coordinate system. The errors resulting from the registration process could reach 14 mm [103]. In addition, laser scanning could be integrated with photogrammetry [104,105]. A study on two bridges was conducted in [106], and a discrepancy ranging from 7 mm to 30 mm was found between the point clouds produced by photogrammetry and laser scanners. Laser scanning was applied to monitor the surface of a concrete beam, which was tested in the laboratory [107]. Loading was applied in steps, and a scan of the surface was completed for each step. In another experimental study [108], a fixed laser scanner was used to monitor the deformations of an arch structure. The error in the point cloud data was 0.8 mm. Laser scanning is used in the construction industry as well [109]. The usage of traditional instruments (including Linear Variable Differential Transformers (LVDTs), laser sensors, accelerometers, and strain gauges) and non-contact techniques (such as DIC and laser scanning) during load tests of bridges is recalled in [110]. It was noted that non-contact methods are not commonly used in this context.
LiDAR (Light Detection and Ranging) systems emit a large number of laser pulses toward the target area [111]. UAV-based LiDAR systems can merge multiple scans acquired from different positions to generate a 3D point cloud. Merging these scans requires estimating the UAV’s trajectory with six degrees of freedom. Trajectory estimation is usually named “odometry”, and it is a fundamental chain in UAV-based LiDAR systems. The fidelity of the odometry affects the accuracy of the obtained 3D point cloud. Different approaches could be performed for odometry; a LiDAR-only odometry determines the UAV’s position by analyzing consecutive LiDAR scans [112]. Alternative methods either integrate an IMU (inertial measurement unit) with LiDAR or employ multiple LiDAR systems. However, using multiple LiDAR systems or additional sensors increases computational demands.
LiDAR measurements are not reliable in environments with sparse geometric features or repetitive topographical characteristics, such as long tunnels [112]. Similar challenges may also arise in dams. In some cases, fusion of LiDAR with RGB cameras might be the solution [112], as the RGB cameras provide colour and texture information. This is similar to the process used with ToF devices, where depth and colour frames are merged to create a coloured point cloud. However, the latter is a much simpler procedure and requires less computational effort. Another limitation of LiDAR systems is the large amount of data that needs to be processed [112].
It was mentioned in [113] that the use of UAV-based LiDAR systems is occasional, as such devices are not considered suitable for high-precision monitoring applications. This is in the agreement with the findings in [114], where the performance of some of the available low-cost LiDAR scanners was evaluated. In this study, experiments were conducted to measure the distance between the fixed scanner and a flat target surface. In the outdoor experiment, the error in the distance measurements ranged from 1 cm to 3.7 cm. In [115], 3D models of a bridge were created using two distinct methods: UAV photogrammetry and UAV-based LiDAR. The error in the 3D point cloud generated with photogrammetric techniques was around 2.5 cm, while that obtained with the LiDAR ranged from 5 to 7 cm. Moreover, a solid-state (fixed) LiDAR system was used in [116] to survey the geometry of a ballast system in the laboratory. Each scan was acquired with the LiDAR system mounted on a trolley attached to the track. The error in the measurements ranged from 5 to 7 mm. According to the previously cited studies and the errors found with the LiDAR measurements, it could be said that such systems are not the best choice for high-precision monitoring applications in structural engineering. For this reason, UAV-based LiDAR applications are not thoroughly surveyed in this manuscript.

3. Lighting Issues

Many issues are present in the application of imaging techniques, such as the image noise that is associated with the sensor’s resolution, illumination variability, computational load, and the integration and calibration of multiple sensors. In field applications, vision-based techniques might face challenges under specific environmental conditions, with sunlight disruptions being a major factor. This section reviews the main issues caused by illumination variation, their impact on measurement accuracy, and the possible mitigation techniques. For instance, DIC relies on grey intensity conservation, assuming that a pixel’s grey value distribution remains consistent between the reference and deformed images. However, in field applications, variations in lighting could occur due to moving clouds, camera shadow, changing weather conditions, a large field of view, etc., thus affecting the accuracy of DIC measurements. This issue was reported in [56] during the monitoring of displacements in a pedestrian bridge, where the camera exposure time had to be adjusted during image acquisition due to moving clouds. A similar problem was encountered during bridge monitoring in [19], resulting in overexposed or blurry images, which increased the absolute error from 0.05 mm to 0.34 mm.
The impact of illumination variations on DIC measurements was investigated in some studies. In [117], images were digitally modified to introduce changes in lighting conditions. The deformation scenarios consisted of undeformed samples and deformed samples with homogeneous and non-homogeneous strains. Acceptable measurements were achieved when the brightness in the target image was uniformly modified. However, large errors were detected when lighting variations were introduced across different areas of the image. In a similar study [118], images corresponding to rigid body translation were considered. The image’s brightness was adjusted by a constant offset, and the measurements were acceptable for a mean grey level between 32 and 250. However, a sharp increase in the errors was observed at lower grey levels. These findings align with the formulation-based statements in [34].
Some studies were performed to address the issue of light variation in DIC applications. A method based on the Fourier transform was used to filter illumination noise, and its accuracy was validated both indoors and outdoors by monitoring an aircraft panel subjected to an impact load [119]. Brightness correction was used in monitoring cracks in a cube made of refractory castable, with global DIC being used instead [120]. Grey-level disturbances are caused by the experimental conditions (temperature and humidity) and by the development of cracks. The brightness value at every pixel in the reference image was adjusted by means of brightness and contrast correction functions. Such techniques are useful when illumination changes between images, but they do not solve the issue of nonuniform illumination within the same image. It should be noted that the most advanced DIC tools can accommodate the drift in lighting conditions [33], provided that no abrupt brightness distribution arise from one image to another.
In [121], correction for irregular intensity was achieved by subtracting the background from the original image. The study involved laboratory tests on an aluminum plate under uneven illumination, including rigid body rotation and three-point bending. The correction method reduced the measurement error by an order of magnitude. Another brightness correction method was developed in [122], based on the assumption that the image is composed of an illumination component and a reflection component. The needed information, such as the texture or speckle pattern, is represented by the reflection component. This study did not include a DIC application; however, the developed method was tested on images of various landscapes. Another correction method based on Retinex theory [123] was applied in [124], where first a numerical simulation was performed on digitally modified images to represent the nonuniform intensity distribution. The mean error and the standard deviation were calculated, and a reduction in these quantities was observed after correcting for nonuniform illumination. Laboratory experiments were then conducted including rigid motion and uniaxial testing. A cold lighting source was placed in front of the lighting device to vary the brightness levels. The results of the study in [124] showed that the brightness correction approach reduced the errors and provided smoother displacement fields. In another study [54], outdoor experiments were conducted to simulate the presence of shadows in the monitored region. Image brightness adjustment was then applied, reducing the disturbances in the measured displacements. Figure 8 presents a sample image from this study [54] along with the corresponding vertical displacement fields obtained using DIC. In this test, the monitored region was fixed, and a uniform displacement field was therefore expected. The objective was to assess the noise level in the DIC measurements, both before and after applying image brightness correction. It can be seen that the displacement field on the bottom right side of Figure 8 is smoother than that of the bottom left side. This shows the improvement obtained using the brightness correction algorithms. As a quantitative assessment of the brightness enhancement technique, the standard deviation of the displacement field was calculated in [54], and the values are shown in Table 5. A reduction in the standard deviation can be seen after applying the brightness modification approach, indicating a smoother displacement field.
Similar issues could occur with ToF measurements, as the sunlight IR radiations may create disturbances to the signal detected by the sensor. This could be mitigated by using ToF devices that operate at a wavelength of 940 nm [101], where sunlight is substantially diminished by the filtering effect of the atmosphere [125]. Other sensors used in [101] had wavelengths smaller than 900 nm, and their accuracy remained unchanged when exposed to daylight without direct sunlight. Tests were performed in [94] under direct sunlight, and acceptable results were produced for a measuring range not exceeding 1.5 m.

4. Learning Tools

The information gathered by vision-based sensors can be improved by the use of machine learning and deep learning algorithms, allowing damage detection and quantification. Post-processing techniques can also overcome limitations connected with on-site acquisition of images under variable environmental conditions. Such tools were used to automatically identify cracks or defects in the images of a concrete surface [126,127,128]. Pictures of a laboratory frame and of an elevator tower were acquired by a camera mounted on a UAV [129]. Features identified with a deep learning neural network were used to compensate for camera movement and to calculate the relative displacements. Similarly in [130], features were detected in each recorded frame in order to track displacements between consecutive frames. A bridge inspection was carried out in [131] and defects such as efflorescence, water leakage, spalling, and discoloration were identified. A machine learning framework was applied to images acquired with a UAV in order to measure the crack lengths in a bridge [132]. In [133], an image processing algorithm was implemented to detect cracks in a concrete bridge from images collected by a UAV. In a similar study [134], deep learning was applied to detect and quantify crack openings. Summary data of these two studies is reported in Table 6. In [135], deep learning was applied to detect cracks in the images captured with a UAV. The deep learning model was previously trained on a dataset of images containing dam cracks. Moreover, deep learning was used in [136] to obtain vibration information in tall buildings. In [137], reference targets with specific patterns were installed on the tested structure. These targets were necessary, as they served as identifiable features whose displacements were tracked. Errors of several millimetres (from around 4 to 10 mm) were detected at the reference target locations.
Several review articles on the application of learning techniques for monitoring civil engineering structures have been published recently [138,139,140,141]. It was also mentioned in [141] that only a few studies focused on obtaining the crack geometry (its length and opening). This is mainly due to the fact that, to obtain metric information about the crack, it is necessary to estimate the instantaneous position and orientation of the camera with respect to the target surface, which are hardly available in UAV-based measurements. An additional advantage of learning tools is that they can be integrated with other image processing techniques. This combined approach has been shown to improve accuracy, as previously noted in Table 2.

5. Conclusions

In general, the conducted literature review confirms that vision-based techniques are suitable for outdoor structural monitoring and fault diagnosis. However, factors such as drone size and payload capacity should be considered. A large UAV is required for 3D DIC applications to mount the rigid beam, which supports the two cameras, ensuring that their relative position and orientation remain fixed, to guarantee the validity of the calibration. In addition, the drone should be equipped with a power supply and a computer to control the image acquisition process. A medium-sized drone with a computer on board can be used also for ToF recordings. The photogrammetry technique provides the greatest flexibility in terms of the type of drone that can be used, since only one camera is needed, and it could be remotely controlled. Thus, a drone costing around USD 1000 may suffice for photogrammetry applications, whereas ToF and 3D-DIC techniques typically require a more advanced drone, priced at approximately USD 4000–USD 5000 or higher. However, drone technology is rapidly evolving, and according to information from UAV manufacturers, more efficient platforms are expected to become available in the near future, enabling the integration of ToF devices on lower-cost drones (approximately USD 2000). One point to be highlighted concerning photogrammetry applications is that, if the UAV is not equipped with a GPS device to tag each acquired image with the GPS coordinates, the distance between at least two reference points on the surface must be available in order to adequately scale the model.
In terms of accuracy, DIC performs best, achieving submillimetre precision, whereas photogrammetry and ToF typically reach accuracies within a few millimetres. In particular, DIC measured crack openings as small as 0.05 mm using a 2 MP camera, whereas photogrammetry, even with a 15.6 MP camera, captured a minimum crack size of approximately 2 mm (Table 4). Learning-based tools are also of great importance, with promising results reported when integrated with other techniques, such as photogrammetry. Different techniques with various hardware are compared in this review. The experimental setup (configuration, sensor placement, data acquisition, and processing procedures) has an influence on the final results; however, this information is not always available, and the comparison values given in the tables are therefore meant to be indicative. A main difference between DIC and the other two methods (ToF and photogrammetry) is that the displacements are originally provided in the DIC output, while the other techniques only produce a 3D reconstruction of the surface. Thus, an additional processing step, with associated uncertainty, is required to obtain the displacements. This could be performed either by point cloud registration or by identification of features on the surface. The main issues with ToF devices are depth measurement uncertainty (around 3 mm for sensors suitable for UAV-based structural monitoring) and a low camera resolution, as the most common ones do not exceed 1 Megapixel. New generations of ToF devices are expected in the near future, with improvements that will greatly impact the measurement accuracy. UAV-based LiDAR systems were also considered in this review; relatively large errors ranging from 1 cm to 7 cm were found [114,115]. In addition, these systems present a challenge due to their increased computational cost.
To provide a broad view on the current challenges and future developments, the following comments are provided to improve the new research in this field. Regarding light variations, for example, currently compensation techniques are already present and provide an acceptable result. In addition, mitigation techniques for the UAV motion and drifting are present and could be used. However, significant improvements are needed in the following areas: integrating multiple sensors and improving the UAV position and trajectory estimation (by exploiting global positioning accuracy, as well as magnetometer and IMU data). These developments would require a higher computational effort, and further advancements in data processing are therefore anticipated. Moreover, it should be noted that the use of multiple sensors necessitates careful calibration and synchronization.
All the techniques mentioned in this review retrieve full-field information useful for calibrating continuously updated numerical models (digital twins) of the investigated systems and capturing the evolution of their structural response over time. The reliability of these predictions depends on the accuracy of the measurements, as well as on uncertainties about the actual geometry, element connections, and out-of-sight details. Therefore, no acquisition method outperforms the others in general, but the optimal choice comes from a balance of different and sometimes conflicting requirements.

Author Contributions

Conceptualization, G.B. and E.Z.; methodology, G.B., E.Z. and M.H.; investigation, E.Z. and M.H.; resources, G.B. and E.Z.; writing—original draft preparation, G.B. and M.H.; writing—review and editing, G.B., E.Z. and M.H.; visualization, M.H.; supervision, G.B. and E.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Farrar, C.R.; Dervilis, N.; Worden, K. The past, present and future of structural health monitoring: An overview of three ages. Strain 2025, 61, e12495. [Google Scholar] [CrossRef]
  2. Prakash, G.; Revanth Dugalam, R.; Barbosh, M.; Sadhu, A. Recent advancement of concrete dam health monitoring technology: A systematic literature review. Structures 2022, 44, 766–784. [Google Scholar] [CrossRef]
  3. Bolzon, G.; Frigerio, A.; Hajjar, M.; Nogara, C.; Zappa, E. Structural health assessment of existing dams based on non-destructive testing, physics-based models and machine learning tools. NDTE Int. 2025, 150, 103271. [Google Scholar] [CrossRef]
  4. Figueiredo, E.; Brownjohn, J. Three decades of statistical pattern recognition paradigm for SHM of bridges. Struct. Health Monit. 2022, 21, 3018–3054. [Google Scholar] [CrossRef]
  5. Zhang, J.; Cenci, J.; Becue, V.; Koutra, S. The overview of the conservation and renewal of the industrial Belgian heritage as a vector for cultural regeneration. Information 2021, 12, 27. [Google Scholar] [CrossRef]
  6. Soleymani, A.; Jahangir, H.; Moncef, L.; Nehdi, M.L. Damage detection and monitoring in heritage masonry structures: Systematic review. Constr. Build. Mater. 2023, 397, 132402. [Google Scholar] [CrossRef]
  7. Ubertini, F.; Comanducci, G.; Cavalagli, N. Vibration-based structural health monitoring of a historic bell-tower using output-only measurements and multivariate statistical analysis. Struct. Health Monit. 2016, 15, 438–457. [Google Scholar] [CrossRef]
  8. Chisari, C.; Mattia Zizi, M.; De Matteis, G. Dynamic model identification of the medieval bell tower of Casertavecchia (Italy). Eng. Fail. Anal. 2025, 167, 109055. [Google Scholar] [CrossRef]
  9. Masciotta, M.G.; Pellegrini, D. Tracking the variation of complex mode shapes for damage quantification and localization in structural systems. Mech. Syst. Signal Process. 2022, 169, 108731. [Google Scholar] [CrossRef]
  10. Azmi, M.; Paultre, P. Three-dimensional analysis of concrete dams including contraction joint non-linearity. Eng. Struct. 2002, 24, 757–771. [Google Scholar] [CrossRef]
  11. Mata, J.; Miranda, F.; Lorvão Antunes, A.; Romão, X.; Santos, J. Characterization of relative movements between blocks observed in a concrete dam and definition of thresholds for novelty identification based on machine learning models. Water 2023, 15, 297. [Google Scholar] [CrossRef]
  12. Verstrynge, E.; De Wilder, K.; Drougkas, A.; Voet, E.; Koen Van Balen, K.; Wevers, M. Crack monitoring in historical masonry with distributed strain and acoustic emission sensing techniques. Constr. Build. Mater. 2018, 162, 898–907. [Google Scholar] [CrossRef]
  13. Bamonte, P.; Cardani, G.; Condoleo, P.; Taliercio, A. Crack patterns in double-wall industrial masonry chimneys: Possible causes and numerical modelling. J. Cult. Herit. 2021, 47, 133–142. [Google Scholar] [CrossRef]
  14. Giuseppetti, G.; Donghi, G.; Marcello, A. Experimental investigations and numerical modeling for the analysis of AAR process related to Poglia dam: Evolutive scenarios and design solutions. In Dam Maintenance and Rehabilitation; Taylor & Francis: London, UK, 2003; pp. 183–191. [Google Scholar]
  15. Human, O.; Oosthuizen, C. Notes on the behaviour of a 65 year old concrete arch dam affected by AAR (mainly based on visual observations). In Proceedings of the 84th Annual ICOLD Meeting, Johannesburg, South Africa, 15–20 May 2016. [Google Scholar]
  16. Campos, A.; López, C.M.; Blanco, A.; Aguado, A. Structural diagnosis of a concrete dam with cracking and high nonrecoverable displacements. J. Perform. Constr. Facil. 2016, 30, 04016021. [Google Scholar] [CrossRef]
  17. Sangirardi, M.; Altomare, V.; De Santis, S.; de Felice, G. Detecting damage evolution of masonry structures through computer-vision-based monitoring methods. Buildings 2022, 12, 831. [Google Scholar] [CrossRef]
  18. Germanese, D.; Leone, G.R.; Moroni, D.; Pascali, M.A.; Tampucci, M. Long-term monitoring of crack patterns in historic structures using UAVs and planar markers: A preliminary study. J. Imaging 2018, 4, 99. [Google Scholar] [CrossRef]
  19. Reagan, D.; Sabato, A.; Niezrecki, C. Feasibility of using digital image correlation for unmanned aerial vehicle structural health monitoring of bridges. Struct. Health Monit. 2018, 17, 1056–1072. [Google Scholar] [CrossRef]
  20. Liu, B.; Collier, J.; Sarhosis, V. Digital image correlation based crack monitoring on masonry arch bridges. Eng. Fail. Anal. 2025, 169, 109185. [Google Scholar] [CrossRef]
  21. Belloni, V.; Sjölander, A.; Ravanelli, R.; Crespi, M.; Nascetti, A. Crack Monitoring from Motion (CMfM): Crack detection and measurement using cameras with non-fixed positions. Autom. Constr. 2023, 156, 105072. [Google Scholar] [CrossRef]
  22. Hatir, M.E.; Ince, I.; Korkanç, M. Intelligent detection of deterioration in cultural stone heritage. J. Build. Eng. 2021, 44, 102690. [Google Scholar] [CrossRef]
  23. Yu, R.; Li, P.; Shan, J.; Zhu, H. Structural state estimation of earthquake-damaged building structures by using UAV photogrammetry and point cloud segmentation. Measurement 2022, 202, 111858. [Google Scholar] [CrossRef]
  24. Qing, Y.; Ming, D.; Wen, Q.; Weng, Q.; Xu, L.; Chen, Y.; Zhang, Y.; Zeng, B. Operational earthquake-induced building damage assessment using CNN-based direct remote sensing change detection on superpixel level. Int. J. Appl. Earth Obs. Geoinf. 2022, 112, 102899. [Google Scholar] [CrossRef]
  25. Sengupta, P.; Chakraborty, S. A state-of-the-art review on model reduction and substructuring techniques in finite element model updating for structural health monitoring applications. Arch. Comput. Methods Eng. 2025, 32, 3031–3062. [Google Scholar] [CrossRef]
  26. Li, X.; Zhang, F.-L.; Xiang, W.; Liu, W.-X.; Fu, S.-J. Structural health monitoring system based on digital twins and real-time data-driven methods. Structures 2024, 70, 107739. [Google Scholar] [CrossRef]
  27. Parente, L.; Falvo, E.; Castagnetti, C.; Grassi, F.; Mancini, F.; Rossi, P.; Capra, A. Image-based monitoring of cracks: Effectiveness analysis of an open-source machine learning-assisted procedure. J. Imaging 2022, 8, 22. [Google Scholar] [CrossRef]
  28. Malekloo, A.; Ozer, E.; Alhamaydeh, M.; Girolami, M. Machine learning and structural health monitoring overview with emerging technology and high-dimensional data source highlights. Struct. Health Monit. 2023, 21, 1906–1955. [Google Scholar] [CrossRef]
  29. Flah, M.; Nunez, I.; Ben Chaabene, W.; Nehdi, M.F. Machine learning algorithms in civil structural health monitoring: A systematic review. Arch. Comput. Methods Eng. 2021, 28, 2621–2643. [Google Scholar] [CrossRef]
  30. Hariri-Ardebili, A.; Mahdavi, G.; Nuss, L.K.; Lall, U. The role of artificial intelligence and digital technologies in dam engineering: Narrative review and outlook. Eng. Appl. Artif. Intell. 2023, 126, 106813. [Google Scholar] [CrossRef]
  31. Mishra, M. Machine learning techniques for structural health monitoring of heritage buildings: A state-of-the-art review and case studies. J. Cult. Herit. 2021, 47, 227–245. [Google Scholar] [CrossRef]
  32. Reis, H.C.; Turk, V.; Ustuner, M.; Yildiz, C.M.K.; Tatli, R. Post-seismic structural assessment: Advanced crack detection through complex feature extraction using pre-trained deep learning and machine learning integration. Earth Sci. Inform. 2025, 18, 133. [Google Scholar] [CrossRef]
  33. Sutton, M.; Orteu, J.; Schreier, H. Image Correlation for Shape, Motion and Deformation Measurements: Basic Concepts, Theory and Applications; Springer Science & Business Media: New York, NY, USA, 2009. [Google Scholar] [CrossRef]
  34. Pan, B.; Qian, K.; Xie, H.; Asundi, A. Two-dimensional digital image correlation for in-plane displacement and strain measurement: A review. Meas. Sci. Tech. 2009, 20, 062001. [Google Scholar] [CrossRef]
  35. Hanke, K.; Grussenmeyer, P. Architectural photogrammetry: Basic theory, procedures, tools. ISPRS Comm. V Symp. 2002, 5, 300–339. [Google Scholar] [CrossRef]
  36. Hansard, M.; Lee, S.; Choi, O.; Horaud, R. Time-of-Flight Cameras: Principles, Methods and Applications; Springer Science & Business Media: New York, NY, USA, 2012; Available online: https://link.springer.com/book/10.1007/978-1-4471-4658-2 (accessed on 22 January 2025).
  37. Morgenthal, G.; Hallermann, N. Quality assessment of unmanned aerial vehicle (UAV) based visual inspection of structures. Adv. Struct. Eng. 2014, 17, 289–302. [Google Scholar] [CrossRef]
  38. Hu, X.; Assaad, R. The use of unmanned ground vehicles (mobile robots) and unmanned aerial vehicles (drones) in the civil infrastructure asset management sector: Applications, robotic platforms, sensors, and algorithms. Expert. Syst. Appl. 2023, 232, 120897. [Google Scholar] [CrossRef]
  39. Kapoor, M.; Katsanos, E.; Nalpantidis, L.; Winkler, J.; Thöns, S. Structural Health Monitoring and Management with Unmanned Aerial Vehicles: Review and Potentials; BYG-R-454; Technical University of Denmark: Lyngby, Denmark, 2021; Available online: https://www.byg.dtu.dk/forskning/publikationer/byg_rapporter (accessed on 25 July 2025).
  40. Duque, L.; Seo, J.; Wacker, J. Synthesis of unmanned aerial vehicle applications for infrastructures. J. Perform. Constr. Facil. 2018, 32, 04018046. [Google Scholar] [CrossRef]
  41. Sony, S.; Laventure, S.; Sadhu, A. A literature review of next-generation smart sensing technology in structural health monitoring. Struct. Control Health Monit. 2019, 26, e2321. [Google Scholar] [CrossRef]
  42. Freeman, M.; Kashani, M.; Vardanega, P. Aerial robotic technologies for civil engineering: Established and emerging practice. J. Unmanned Veh. Syst. 2021, 9, 75–91. [Google Scholar] [CrossRef]
  43. Peters, W.; Ranson, W. Digital imaging techniques in experimental stress analysis. Opt. Eng. 1982, 21, 427–431. [Google Scholar] [CrossRef]
  44. Pan, B. Digital image correlation for surface deformation measurement: Historical developments, recent advances and future goals. Meas. Sci. Tech. 2018, 29, 082001. [Google Scholar] [CrossRef]
  45. Hoult, N.; Dutton, M.; Hoag, A.; Take, W. Measuring crack movement in reinforced concrete using digital image correlation: Overview and application to shear slip measurements. Proc. IEEE 2016, 104, 1561–1574. [Google Scholar] [CrossRef]
  46. Fayyad, T.; Lees, J. Experimental investigation of crack propagation and crack branching in lightly reinforced concrete beams using digital image correlation. Eng. Fract. Mech. 2017, 182, 487–505. [Google Scholar] [CrossRef]
  47. Liu, R.; Zappa, E.; Collina, A. Vision-based measurement of crack generation and evolution during static testing of concrete sleepers. Eng. Fract. Mech. 2020, 224, 106715. [Google Scholar] [CrossRef]
  48. Zhang, F.; Zarate Garnica, G.; Yang, Y.; Lantsoght, E.; Sliedrecht, H. Monitoring shear behavior of prestressed concrete bridge girders using acoustic emission and digital image correlation. Sensors 2020, 20, 5622. [Google Scholar] [CrossRef] [PubMed]
  49. Gencturk, B.; Hossain, K.; Kapadia, A.; Labib, E.; Mo, Y. Use of digital image correlation technique in full-scale testing of prestressed concrete structures. Measurement 2014, 47, 505–515. [Google Scholar] [CrossRef]
  50. De Wilder, K.; Lava, P.; Debruyne, D.; Wang, Y.; De Roeck, G.; Vandewalle, L. Experimental investigation on the shear capacity of prestressed concrete beams using digital image correlation. Eng. Struct. 2015, 82, 82–92. [Google Scholar] [CrossRef]
  51. Omondi, B.; Aggelis, B.; Sol, H.; Sitters, C. Improved crack monitoring in structural concrete by combined acoustic emission and digital image correlation techniques. Struct. Health Monit. 2016, 15, 359–378. [Google Scholar] [CrossRef]
  52. Hajjar, M.; Bolzon, G.; Zappa, E. Experimental and numerical analysis of fracture in prestressed concrete. Procedia Struct. Integr. 2023, 47, 354–358. [Google Scholar] [CrossRef]
  53. Thériault, F.; Noël, M.; Sanchez, L. Simplified approach for quantitative inspections of concrete structures using digital image correlation. Eng. Struct. 2022, 252, 113725. [Google Scholar] [CrossRef]
  54. Hajjar, M. UAV Monitoring of Dam Joints: A Feasibility Study. Ph.D. Thesis, Politenico di Milano, Milan, Italy, 2025. Available online: https://www.politesi.polimi.it/handle/10589/233315 (accessed on 1 March 2025).
  55. Ngeljaratan, L.; Moustafa, M. System identification of large-scale bridges using target-tracking digital image correlation. Front. Built Environ. 2019, 5, 85. [Google Scholar] [CrossRef]
  56. Ngeljaratan, L.; Moustafa, M. Structural health monitoring and seismic response assessment of bridge structures using target-tracking digital image correlation. Eng. Struct. 2020, 213, 110551. [Google Scholar] [CrossRef]
  57. Dizaji, M.; Harris, D.; Alipour, M. Integrating visual sensing and structural identification using 3D-digital image correlation and topology optimization to detect and reconstruct the 3D geometry of structural damage. Struct. Health Monit. 2022, 21, 2804–2833. [Google Scholar] [CrossRef]
  58. Kumar, D.; Chiang, C.; Lin, Y. Experimental vibration analysis of large structures using 3D DIC technique with a novel calibration method. J. Civ. Struct. Health Monit. 2022, 12, 391–409. [Google Scholar] [CrossRef]
  59. Malesa, M.; Malowany, K.; Pawlicki, J.; Kujawinska, M.; Skrzypczak, P.; Piekarczuk, A.; Tomasz, L.; Zagorski, A. Non-destructive testing of industrial structures with the use of multi-camera Digital Image Correlation method. Eng. Fail. Anal. 2016, 69, 122–134. [Google Scholar] [CrossRef]
  60. Malowany, K.; Malesa, M.; Kowaluk, T.; Kujawinska, M. Multi-camera digital image correlation method with distributed fields of view. Opt. Lasers Eng. 2017, 98, 198–204. [Google Scholar] [CrossRef]
  61. Dizaji, M.; Harris, D.; Kassner, B.; Hill, J. Full-field non-destructive image-based diagnostics of a structure using 3D digital image correlation and laser scanner techniques. J. Civ. Struct. Health Monit. 2021, 11, 1415–1428. [Google Scholar] [CrossRef]
  62. Khadka, A.; Afshar, A.; Zadeh, M.; Baqersad, J. Strain monitoring of wind turbines using a semi-autonomous drone. Wind. Eng. 2022, 46, 296–307. [Google Scholar] [CrossRef]
  63. Kumarapu, K.; Mesapam, S.; Keesara, V.; Shukla, A.; Manapragada, N.; Javed, B. RCC structural deformation and damage quantification using unmanned aerial vehicle image correlation technique. Appl. Sci. 2022, 12, 6574. [Google Scholar] [CrossRef]
  64. Kalaitzakis, M.; Kattil, S.; Vitzilaios, N.; Rizos, D.; Sutton, M. Dynamic structural health monitoring using a DIC-enabled drone. In Proceedings of the 2019 International Conference on Unmanned Aircraft Systems, Atlanta, GA, USA, 11–14 June 2019; pp. 321–327. [Google Scholar] [CrossRef]
  65. Kalaitzakis, M.; Vitzilaios, N.; Rizos, D.; Sutton, M. Drone-based StereoDIC: System development, experimental validation and infrastructure application. Exp. Mech. 2021, 61, 981–996. [Google Scholar] [CrossRef]
  66. Stuart, C.; Vitzilaios, N.; Rizos, D.; Sutton, M. Railroad Bridge Inspection Using Drone-Based Digital Image Correlation. United States-Department of Transportation-Federal Railroad Administration-Office of Research, Development, and Technology, 2023, RR 23-02. Available online: https://railroads.dot.gov/sites/fra.dot.gov/files/2023-02/Railroad%20Bridge%20Inspection%20Drone-Based.pdf (accessed on 18 February 2025).
  67. Mousa, M.; Yussof, M.; Udi, U.; Nazri, F.; Kamarudin, M.; Parke, G.; Ghahari, S. Application of digital image correlation in structural health monitoring of bridge infrastructures: A review. Infrastructures 2021, 6, 176. [Google Scholar] [CrossRef]
  68. Ye, X.; Dong, C.; Liu, T. A review of machine vision-based structural health monitoring: Methodologies and applications. J. Sens. 2016, 1, 7103039-10. [Google Scholar] [CrossRef]
  69. Feng, D.; Feng, M. Computer vision for SHM of civil infrastructure: From dynamic response measurement to damage detection-A review. Eng. Struct. 2018, 156, 105–117. [Google Scholar] [CrossRef]
  70. Niezrecki, C.; Baqersad, J.; Sabato, A. Digital image correlation techniques for non-destructive evaluation and structural health monitoring. Handb. Adv. Non-Destr. Eval. 2018, 46, 1545–1590. [Google Scholar] [CrossRef]
  71. Blikharskyy, Y.; Kopiika, N.; Khmil, R.; Selejdak, J.; Blikharskyy, Z. Review of development and application of digital image correlation method for study of stress-strain state of RC structures. Appl. Sci. 2022, 12, 10157. [Google Scholar] [CrossRef]
  72. Lowe, D. Object recognition from local scale-invariant features. In Proceedings of the Seventh IEEE International Conference on Computer Vision, Kerkyra, Greece, 20–27 September 1999; pp. 1150–1157. [Google Scholar] [CrossRef]
  73. Fischler, M.; Bolles, R. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  74. Schonberger, J.; Frahm, J. Structure-from-motion revisited. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4104–4113. [Google Scholar] [CrossRef]
  75. Martínez-Carricondo, P.; Agüera-Vega, F.; Carvajal-Ramírez, F. Use of UAV-photogrammetry for quasi-vertical wall surveying. Remote Sens. 2020, 12, 2221. [Google Scholar] [CrossRef]
  76. Fonstad, M.; Dietrich, J.; Courville, B.; Jensen, J.; Carbonneau, P. Topographic structure from motion: A new development in photogrammetric measurement. Earth Surf. Process Landf. 2013, 38, 421–430. [Google Scholar] [CrossRef]
  77. Murtiyoso, A.; Grussenmeyer, P.; Koehl, M.; Freville, T. Acquisition and Processing Experiences of Close Range UAV Images for the 3D Modeling of Heritage Buildings. In Digital Heritage Progress in Cultural Heritage: Documentation, Preservation, and Protection; EuroMed: Limassol, Cyprus, 2016; Volume 10058, pp. 420–431. [Google Scholar] [CrossRef]
  78. Murtiyoso, A.; Koehl, M.; Grussenmeyer, P.; Freville, T. Acquisition and processing protocols for UAV images: 3D modeling of historical buildings using photogrammetry. In Proceedings of the 26th CIPA Int Symposium ISPRS, Ottawa, ON, Canada, 28 August–1 September 2017; Volume 4, pp. 163–170. [Google Scholar] [CrossRef]
  79. Aicardi, I.; Chiabrando, F.; Grasso, N.; Lingua, A.; Noardo, F.; Spanò, A. UAV photogrammetry with oblique images: First analysis on data acquisition and processing. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2016, 41, 835–842. [Google Scholar] [CrossRef]
  80. Ulvi, A. Using UAV photogrammetric technique for monitoring, change detection, and analysis of archeological excavation sites. J. Comp. Cult. Herit. 2022, 15, 1–19. [Google Scholar] [CrossRef]
  81. Fiz, J.; Martín, P.; Cuesta, R.; Subías, E.; Codina, D.; Cartes, A. Examples and results of aerial photogrammetry in archeology with UAV: Geometric documentation, high resolution multispectral analysis, models and 3D printing. Drones 2022, 6, 59. [Google Scholar] [CrossRef]
  82. Fritsch, D.; Becker, S.; Rothermel, M. Modeling facade structures using point clouds from dense image matching. In Proceedings of the International Conference on Advances in Civil, Structural and Mechanical Engineering, Cape Town, South Africa, 2–4 September 2013; Institute of Research Engineers and Doctors: New York, NY, USA, 2013; pp. 57–64. [Google Scholar]
  83. Bori, M.; Hussein, Z. Integration the low cost camera images with the google earth dataset to create a 3D model. Civ. Eng. J. 2022, 6, 446–458. [Google Scholar] [CrossRef]
  84. Liu, Y.; Nie, X.; Fan, J.; Liu, X. Image-based crack assessment of bridge piers using unmanned aerial vehicles and three-dimensional scene reconstruction. Comput. Aided Civ. Infrastruct. Eng. 2020, 35, 511–529. [Google Scholar] [CrossRef]
  85. Ioli, F.; Pinto, A.; Pinto, L. UAV photogrammetry for metric evaluation of concrete bridge cracks. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2022, 43, 1025–1032. [Google Scholar] [CrossRef]
  86. Buffi, G.; Manciola, P.; Grassi, S.; Barberini, M.; Gambi, A. Survey of the Ridracoli dam: UAV-based photogrammetry and traditional topographic techniques in the inspection of vertical structures. Geomat. Nat. Hazards Risk 2017, 8, 1562–1579. [Google Scholar] [CrossRef]
  87. Ridolfi, E.; Buffi, G.; Venturi, S.; Manciola, P. Accuracy analysis of a dam model from drone surveys. Sensors 2017, 17, 1777. [Google Scholar] [CrossRef]
  88. Henriques, M.; Roque, D. Unmanned aerial vehicles (UAV) as a support to visual inspections of concrete dams. In Proceedings of the Second International Dam World Conference 2015, Lisbon, Portugal, 21–24 April 2015; National Laboratory for Civil Engineering: Lisbon, Portugal, 2015; pp. 1–12. [Google Scholar]
  89. Texas Instruments. Time-of-Flight Camera—An Introduction. Available online: https://www.ti.com/lit/wp/sloa190b/sloa190b.pdf (accessed on 10 May 2025).
  90. He, Y.; Chen, S. Recent advances in 3D data acquisition and processing by time-of-flight camera. IEEE Access 2019, 7, 12495–12510. [Google Scholar] [CrossRef]
  91. Lumileds. Infrared Illumination for Time-of-Flight Applications. Available online: https://lumileds.com/wp-content/uploads/files/WP35.pdf. (accessed on 17 May 2025).
  92. Andersen, M.; Jensen, T.; Lisouski, P.; Mortensen, A.; Hansen, M.; Gregersen, T.; Ahrendt, P. Kinect Depth Sensor Evaluation for Computer Vision Applications; Technical Report ECE-TR-6; Aarhus University: Aarhus, Denmark, 2012; pp. 1–37. [Google Scholar]
  93. Paredes, A.; Song, Q.; Conde, M. Performance evaluation of state-of-the-art high-resolution time-of-flight cameras. IEEE Sens. J. 2023, 23, 13711–13727. [Google Scholar] [CrossRef]
  94. Tölgyessy, M.; Dekan, M.; Chovanec, Ľ.; Hubinský, P. Evaluation of the azure kinect and its comparison to kinect v1 and kinect v2. Sensors 2021, 21, 413. [Google Scholar] [CrossRef]
  95. Albert, J.; Owolabi, V.; Gebel, A.; Brahms, C.; Granacher, U.; Arnrich, B. Evaluation of the pose tracking performance of the azure kinect and kinect v2 for gait analysis in comparison with a gold standard: A pilot study. Sensors 2020, 20, 5104. [Google Scholar] [CrossRef]
  96. Li, L.; Liu, X.; Patel, M.; Zhang, L. Depth camera-based model for studying the effects of muscle loading on distal radius fracture healing. Comput. Biol. Med. 2023, 164, 107292. [Google Scholar] [CrossRef] [PubMed]
  97. Brambilla, C.; Marani, R.; Romeo, L.; Nicora, M.; Storm, F.; Reni, G.; Malosio, M.; D’Orazio, T.; Scano, A. Azure Kinect performance evaluation for human motion and upper limb biomechanical analysis. Heliyon 2023, 9, e21606. [Google Scholar] [CrossRef]
  98. Ramasubramanian, A.; Kazasidis, M.; Fay, B.; Papakostas, N. On the evaluation of diverse vision systems towards detecting human pose in collaborative robot applications. Sensors 2024, 24, 578. [Google Scholar] [CrossRef]
  99. Khoa, T.; Lam, P.; Nam, N. An efficient energy measurement system based on the TOF sensor for structural crack monitoring in architecture. J. Inf. Telecommun. 2023, 7, 56–72. [Google Scholar] [CrossRef]
  100. Marchisotti, D.; Zappa, E. Uncertainty mitigation in drone-based 3D scanning of defects in concrete structures. In Proceedings of the 2022 IEEE International Instrumentation and Measurement Technology Conference, Ottawa, ON, Canada, 16–19 May 2022; pp. 1–6. [Google Scholar] [CrossRef]
  101. Marchisotti, D.; Zappa, E. Feasibility study of drone-based 3-D measurement of defects in concrete structures. IEEE Trans. Instrum. Meas. 2022, 71, 1–11. [Google Scholar] [CrossRef]
  102. Toriumi, F.; Bittencourt, T.; Futai, M. UAV-based inspection of bridge and tunnel structures: An application review. Rev. IBRACON De. Estrut. E Mater. 2022, 16, e16103. [Google Scholar] [CrossRef]
  103. Fryskowska, A. Accuracy assessment of point clouds geo-referencing in surveying and documentation of historical complexes. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2017, 42, 161–165. [Google Scholar] [CrossRef]
  104. Hassan, A.; Fritsch, D. Integration of laser scanning and photogrammetry in 3D/4D cultural heritage preservation—A review. Int. J. Appl. Sci. Tech. 2019, 9, 16. [Google Scholar] [CrossRef]
  105. Ulvi, A. Documentation, three-dimensional (3D) modelling and visualization of cultural heritage by using Unmanned Aerial Vehicle (UAV) photogrammetry and terrestrial laser scanners. Int. J. Remote Sens. 2021, 42, 1994–2021. [Google Scholar] [CrossRef]
  106. Pinto, L.; Bianchini, F.; Nova, V.; Passoni, D. Low-cost UAS photogrammetry for road infrastructures inspection. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2020, 43, 1145–1150. [Google Scholar] [CrossRef]
  107. Yang, H.; Xu, X.; Neumann, I. The benefit of 3D laser scanning technology in the generation and calibration of FEM models for health assessment of concrete structures. Sensors 2014, 14, 21889–21904. [Google Scholar] [CrossRef]
  108. Xu, X.; Bureick, J.; Yang, H.; Neumann, I. TLS-based composite structure deformation analysis validated with laser tracker. Compos. Struct. 2018, 202, 60–65. [Google Scholar] [CrossRef]
  109. Li, X.; Deng, E.; Wang, Y.; Ni, Y. 3D laser scanning for predicting the alignment of large-span segmental precast assembled concrete cable-stayed bridges. Autom. Constr. 2023, 155, 105056. [Google Scholar] [CrossRef]
  110. Zarate Garnica, G.; Lantsoght, E.; Yang, Y. Monitoring structural responses during load testing of reinforced concrete bridges: A review. Struct. and Infrastruct. Eng. 2022, 18, 1558–1580. [Google Scholar] [CrossRef]
  111. Wang, C.; Yang, X.; Xi, X.; Nie, S.; Dong, P. LiDAR Remote Sensing Principles. In Introduction to LiDAR Remote Sensing; Taylor & Francis: Abingdon, UK, 2024. [Google Scholar] [CrossRef]
  112. Lee, D.; Jung, M.; Yang, W.; Kim, A. Lidar odometry survey: Recent advancements and remaining challenges. Intell. Serv. Robot. 2024, 17, 95–118. [Google Scholar] [CrossRef]
  113. Lenda, G.; Marmol, U. Integration of high-precision UAV laser scanning and terrestrial scanning measurements for determining the shape of a water tower. Measurement 2023, 218, 113178. [Google Scholar] [CrossRef]
  114. Kelly, C.; Wilkinson, B.; Abd-Elrahman, A.; Cordero, O.; Lassiter, H. Accuracy assessment of low-cost lidar scanners: An analysis of the Velodyne HDL–32E and Livox Mid–40′s temporal stability. Remote Sens. 2022, 14, 4220. [Google Scholar] [CrossRef]
  115. Gaspari, F.; Ioli, F.; Barbieri, F.; Belcore, E.; Pinto, L. Integration of UAV-lidar and UAV-photogrammetry for infrastructure monitoring and bridge assessment. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2022, 43, 995–1002. [Google Scholar] [CrossRef]
  116. Aldao, E.; González-Jorge, H.; González-deSantos, L.; Fontenla-Carrera, G.; Martínez-Sánchez, J. Validation of solid-state LiDAR measurement system for ballast geometry monitoring in rail tracks. Infrastructures 2023, 8, 63. [Google Scholar] [CrossRef]
  117. Tong, W. An evaluation of digital image correlation criteria for strain mapping applications. Strain 2005, 41, 167–175. [Google Scholar] [CrossRef]
  118. Haddadi, H.; Belhabib, S. Use of rigid-body motion for the investigation and estimation of the measurement errors related to digital image correlation technique. Opt. Lasers Eng. 2008, 46, 185–196. [Google Scholar] [CrossRef]
  119. Guo, X.; Li, Y.; Suo, T.; Liang, J. De-noising of digital image correlation based on stationary wavelet transform. Opt. Lasers Eng. 2017, 90, 161–172. [Google Scholar] [CrossRef]
  120. Sciuti, V.; Canto, R.; Neggers, J.; Hild, F. On the benefits of correcting brightness and contrast in global digital image correlation: Monitoring cracks during curing and drying of a refractory castable. Opt. Lasers Eng. 2021, 136, 106316. [Google Scholar] [CrossRef]
  121. Gu, G.; Wang, Y. Uneven intensity change correction of speckle images using morphological top-hat transform in digital image correlation. Imaging Sci. J. 2015, 63, 488–494. [Google Scholar] [CrossRef]
  122. Zhou, H.; Shu, D.; Wu, C.; Wang, Q.; Wang, Q. Image illumination adaptive correction algorithm based on a combined model of bottom-hat and improved gamma transformation. Arab. J. Sci. Eng. 2023, 48, 3947–3960. [Google Scholar] [CrossRef]
  123. Jobson, D.; Rahman, Z.; Woodell, G. Properties and performance of a center/surround Retinex. IEEE Trans. Image Process. 1997, 6, 451–462. [Google Scholar] [CrossRef]
  124. Sun, L.; Tang, C.; Xu, M.; Lei, Z. Non-uniform illumination correction based on multi-scale Retinex in digital image correlation. Appl. Opt. 2021, 60, 5599–5609. [Google Scholar] [CrossRef]
  125. Texas Instruments. Introduction to Time-of-Flight Long Range Proximity and Distance Sensor System Design. Available online: https://www.ti.com/lit/an/sbau305b/sbau305b.pdf?ts=1732838735208 (accessed on 22 May 2025).
  126. Dung, C.; Anh, L.D. Autonomous concrete crack detection using deep fully convolutional neural network. Autom. Constr. 2019, 99, 52–58. [Google Scholar] [CrossRef]
  127. Yu, C.; Du, J.; Li, M.; Li, Y.; Li, W. An improved U-Net model for concrete crack detection. Mach. Learn. Appl. 2022, 10, 100436. [Google Scholar] [CrossRef]
  128. Chen, Y.; Wu, R.; Puranam, A. Multi-task deep learning for crack segmentation and quantification in RC structures. Autom. Constr. 2024, 166, 105599. [Google Scholar] [CrossRef]
  129. Weng, Y.; Shan, J.; Lu, Z.; Lu, X.; Spencer, B., Jr. Homography-based structural displacement measurement for large structures using unmanned aerial vehicles. Comput. Aided Civ. Infrastruct. Eng. 2021, 36, 1114–1128. [Google Scholar] [CrossRef]
  130. Yoon, H.; Shin, J.; Spencer, B., Jr. Structural displacement measurement using an unmanned aerial system. Comput. Aided Civ. Infrastruct. Eng. 2018, 33, 183–192. [Google Scholar] [CrossRef]
  131. Jeong, E.; Seo, J.; Wacker, J. Grayscale drone inspection image enhancement framework for advanced bridge defect measurement. Transp. Res. Rec. 2021, 2675, 603–612. [Google Scholar] [CrossRef]
  132. Jeong, E.; Seo, J.; Wacker, J. UAV-aided bridge inspection protocol through machine learning with improved visibility images. Expert. Syst. Appl. 2022, 197, 116791. [Google Scholar] [CrossRef]
  133. Lei, B.; Wang, N.; Xu, P.; Song, G. New crack detection method for bridge inspection using UAV incorporating image processing. J. Aerosp. Eng. 2018, 31, 04018058. [Google Scholar] [CrossRef]
  134. Bhowmick, S.; Nagarajaiah, S.; Veeraraghavan, A. Vision and deep learning-based algorithms to detect and quantify cracks on concrete surfaces from UAV videos. Sensors 2020, 20, 6299. [Google Scholar] [CrossRef] [PubMed]
  135. Li, X.; Li, L.; Liu, Z.; Peng, Z.; Liu, S.; Zhou, S.; Chai, X.; Jiang, K. Dam crack detection studies by UAV based on YOLO algorithm. In Proceedings of the 2023 2nd International Conference on Robotics, Artificial Intelligence and Intelligent Control (RAIIC), Mianyang, China, 11–13 August 2023; pp. 104–108. [Google Scholar] [CrossRef]
  136. Shan, J.; Huang, P.; Loong, C.; Liu, M. Rapid full-field deformation measurements of tall buildings using UAV videos and deep learning. Eng. Struct. 2024, 305, 117741. [Google Scholar] [CrossRef]
  137. Ji, R.; Sorosh, S.; Lo, E.; Norton, T.; Driscoll, J.; Kuester, F.; Barbosa, A.; Simpson, B.; Hutchinson, T. Application framework and optimal features for UAV-based earthquake-induced structural displacement monitoring. Algorithms 2025, 18, 66. [Google Scholar] [CrossRef]
  138. Koch, C.; Georgieva, K.; Kasireddy, V.; Akinci, B.; Fieguth, P. A review on computer vision based defect detection and condition assessment of concrete and asphalt civil infrastructure. Adv. Eng. Inform. 2015, 29, 196–210. [Google Scholar] [CrossRef]
  139. Ai, D.; Jiang, G.; Lam, S.; He, P.; Li, C. Computer vision framework for crack detection of civil infrastructure-A review. Eng. Appl. Artif. Intell. 2023, 117, 105478. [Google Scholar] [CrossRef]
  140. Luo, K.; Kong, X.; Zhang, J.; Hu, J.; Li, J.; Tang, H. Computer vision-based bridge inspection and monitoring: A review. Sensors 2023, 23, 7863. [Google Scholar] [CrossRef]
  141. Khan, M.; Kee, S.; Pathan, A.; Nahid, A. Image Processing Techniques for Concrete Crack Detection: A Scientometrics Literature Review. Remote Sens. 2023, 15, 2400. [Google Scholar] [CrossRef]
Figure 1. Example of UAVs equipped with optical sensors.
Figure 1. Example of UAVs equipped with optical sensors.
Sensors 25 05134 g001
Figure 2. Illustration of subset tracking between the reference and deformed configurations [3].
Figure 2. Illustration of subset tracking between the reference and deformed configurations [3].
Sensors 25 05134 g002
Figure 3. Cameras mounted on tripods (left) and fixed support (right).
Figure 3. Cameras mounted on tripods (left) and fixed support (right).
Sensors 25 05134 g003
Figure 4. Sample results from an outdoor DIC application [54].
Figure 4. Sample results from an outdoor DIC application [54].
Sensors 25 05134 g004
Figure 5. Sample model obtained with photogrammetry: Top view (left) and side view (right).
Figure 5. Sample model obtained with photogrammetry: Top view (left) and side view (right).
Sensors 25 05134 g005
Figure 6. Typical setup of a ToF device.
Figure 6. Typical setup of a ToF device.
Sensors 25 05134 g006
Figure 7. Raw point cloud (left) and the corresponding coloured cloud (right).
Figure 7. Raw point cloud (left) and the corresponding coloured cloud (right).
Sensors 25 05134 g007
Figure 8. DIC results for a sample image containing a shadow and its corresponding compensated one [54].
Figure 8. DIC results for a sample image containing a shadow and its corresponding compensated one [54].
Sensors 25 05134 g008
Table 1. Precision of DIC measurements: cameras on tripods vs. a UAV [19].
Table 1. Precision of DIC measurements: cameras on tripods vs. a UAV [19].
Average Noise (mm)Standard Deviation (mm)
Cameras on tripods0.010.02
Cameras on a UAV0.050.05
Table 2. Accuracy of photogrammetry measurements in different applications.
Table 2. Accuracy of photogrammetry measurements in different applications.
Reference
Number
Application DetailsQuantity Measured
[75]UAV photogrammetry3D model: accuracy of 30 mm
[83]UAV photogrammetry3D model: accuracy up to 800 mm
[84]UAV photogrammetry +
image processing +
back-projection
Measured crack openings:
2 to 4 mm (mean error 13%)
[85]UAV photogrammetry +
image processing +
back-projection
Measured crack openings:
3 to 20 mm (mean error 1 mm)
[86]UAV photogrammetry3D model: accuracy of 15 mm
Table 3. Average depth error in the ToF measurements in [93].
Table 3. Average depth error in the ToF measurements in [93].
Measurement RangeDepth Error
1 m1 mm
1.5 m2.4 mm
2.5 m3.4 mm
3 m3.7 mm
Table 4. Comparison of the accuracy of the different vision-based techniques in UAV-assisted measurements.
Table 4. Comparison of the accuracy of the different vision-based techniques in UAV-assisted measurements.
TechniqueSmallest Measured Crack/Defect (mm) Camera ResolutionReference Number
DIC0.052 MP19
Photogrammetry215.6 MP84
ToF3.50.37 MP101
Table 5. Standard deviation of the DIC displacements obtained before and after image brightness correction [54].
Table 5. Standard deviation of the DIC displacements obtained before and after image brightness correction [54].
Before CorrectionAfter Correction
Horizontal displacement (mm)0.01450.0128
Vertical displacement (mm)0.01640.0121
Table 6. Summary data of the inspections in [133,134].
Table 6. Summary data of the inspections in [133,134].
Reference NumberMethodMeasured Crack OpeningError
[133]Multiple image
processing steps
8 mm5%
[134]Deep learning tools2.7 mmNot reported
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hajjar, M.; Zappa, E.; Bolzon, G. The State of the Art and Potentialities of UAV-Based 3D Measurement Solutions in the Monitoring and Fault Diagnosis of Quasi-Brittle Structures. Sensors 2025, 25, 5134. https://doi.org/10.3390/s25165134

AMA Style

Hajjar M, Zappa E, Bolzon G. The State of the Art and Potentialities of UAV-Based 3D Measurement Solutions in the Monitoring and Fault Diagnosis of Quasi-Brittle Structures. Sensors. 2025; 25(16):5134. https://doi.org/10.3390/s25165134

Chicago/Turabian Style

Hajjar, Mohammad, Emanuele Zappa, and Gabriella Bolzon. 2025. "The State of the Art and Potentialities of UAV-Based 3D Measurement Solutions in the Monitoring and Fault Diagnosis of Quasi-Brittle Structures" Sensors 25, no. 16: 5134. https://doi.org/10.3390/s25165134

APA Style

Hajjar, M., Zappa, E., & Bolzon, G. (2025). The State of the Art and Potentialities of UAV-Based 3D Measurement Solutions in the Monitoring and Fault Diagnosis of Quasi-Brittle Structures. Sensors, 25(16), 5134. https://doi.org/10.3390/s25165134

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop