Next Article in Journal
Experimental Investigation on the Fracture Behavior of PET-Modified Engineered High-Ductility Concrete: Effects of PET Powder and Precursor Composition
Previous Article in Journal
Evolution of Phase Transformations in the Mg-Ni-Ce System After Mechanical Synthesis and Spark Plasma Sintering
Previous Article in Special Issue
Quantitative Analysis of Carbonation Degree of Mortar Based on Multiple Interactive Influences
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

3D Reconstruction and Large-Scale Detection of Roads Based on UAV Imagery

1
Yunnan Transportation Science Research Institute Co., Ltd., Kunming 650011, China
2
School of Highway, Chang’an University, Xi’an 710064, China
3
School of Future Transportation, Chang’an University, Xi’an 710064, China
*
Author to whom correspondence should be addressed.
Materials 2025, 18(9), 2133; https://doi.org/10.3390/ma18092133
Submission received: 21 December 2024 / Revised: 3 February 2025 / Accepted: 25 February 2025 / Published: 6 May 2025
(This article belongs to the Special Issue Materials, Structures and Designs for Durable Roads)

Abstract

:
Accurate and efficient detection of road damage is crucial in traffic safety and maintenance management. Traditional road detection methods have problems such as low efficiency and insufficient accuracy, making it difficult to meet the needs of large-scale road health assessments. With the development of drone technology and computer vision, new ideas have been provided for the automatic detection of road diseases. The existing drone-based road detection methods have poor performance in dealing with complex road scenes such as vehicle occlusion, and there is still room for improvement in 3D modeling accuracy and disease detection accuracy, lacking a comprehensive and efficient solution. This paper proposes a UAV (Unmanned Aerial Vehicle)-based 3D reconstruction and large-scale disease detection method for roads. By capturing aerial images with UAVs and utilizing an improved YOLOv8 model, vehicles in the images are identified and removed. Apply MVSNet (Multi-View Stereo Network) 3D reconstruction algorithm for road surface modeling, and finally use point cloud processing and DBSCAN (Density-Based Spatial Clustering of Applications with Noise) clustering for disease detection. The experimental results show that this method performs excellently in terms of 3D modeling accuracy and speed. Compared with the traditional colmap method, the reconstruction speed is greatly improved, and the reconstruction density is three times that of colmap. Meanwhile, the reconstructed point cloud can effectively detect road smoothness and settlement. This study provides a new method for effective disease detection under complex road conditions, suitable for large-scale road health assessment tasks.

1. Introduction

1.1. Research Background and Significance

Road facilities are the core components of the transportation infrastructure system, and the stable operation of road infrastructure depends on timely and effective disease diagnosis and maintenance [1]. Road diseases (such as cracks, potholes, etc.) are often caused by natural factors and traffic load. If not detected and repaired in time, it may lead to serious safety hazards and increase maintenance costs [2]. Therefore, it is necessary to automatically identify all kinds of diseases and extract information efficiently to judge maintenance opportunities and select maintenance strategies [3].
The manual field investigation method used for a long time in the past has problems such as low efficiency, strong subjectivity, and poor safety, and it can no longer meet the needs of current pavement disease detection [4]. At present, the commonly used road disease detection methods mainly focus on the automatic recording and analysis of disease information by using specially equipped sensor detection vehicles driving along the road [5]. However, its driving speed is still restricted by road conditions and traffic conditions, and it cannot quickly cover a large area [6]. In addition, large-scale diseases such as rutting and settlement of asphalt pavement cannot be well recognized [7].
In recent years, UAV-based detection methods have attracted wide attention, with advantages such as high coverage and strong flexibility, and can be combined with deep learning technology to improve detection efficiency and accuracy [8], providing strong support for modern transportation infrastructure management. Rapid deployment of UAVs can achieve efficient coverage of a large area of road network and shorten the response time of disease detection [9]. High-resolution sensors collect detailed road image data and combine deep learning algorithms to identify and classify various types of road disorders and facilities (such as rutting, flatness, large-scale settlement, marking, signage, etc.) [10].
With the development of 3D measurement technology, various 3D imaging systems have been used for pavement data acquisition, including time-of-flight [11,12], projection laser system [13], and stereo vision system [14,15,16]. However, the performance of these systems varies in terms of imaging accuracy, resolution, speed, and field of view (FOV) [17,18].
To sum up, it is of great significance to apply deep learning and 3D imaging technology to UAV road detection. This study aims to develop an efficient road 3D modeling and large-scale disease detection system by combining UAV and deep learning technology to improve the accuracy and real-time road disease detection. With the help of the improved YOLOv8 model, the road surface image is recognized quickly, and MVSNet technology is used to complete the three-dimensional reconstruction of the road. Combined with point cloud processing and the DBSCAN clustering algorithm, this study can effectively detect a wide range of road surface diseases, such as rutting and uneven settlement. The system is not only suitable for disease detection under complex road conditions but also provides an innovative solution for road health assessment, meeting the needs of modern road maintenance management for intelligence and automation [19].

1.2. Research Status at Home and Abroad

Pavement distress detection has consistently been a focal point of research within the field of road engineering. The automated detection of visible distresses, such as pavement cracks, potholes, and deformations, has attracted considerable research interest. Due to the maintenance and management demands arising from early large-scale roadway network construction, developed countries such as the United States, Canada, Australia, and Japan have pioneered advancements in automated pavement distress detection, leading to the development of numerous pavement data collection devices and automated recognition methods [19,20,21,22]. As early as 2004, the NCHRP-334 report summarized methods for the automated collection and processing of pavement condition data, encompassing pavement damage identification, rut depth measurement, and ride quality assessment. This report comprehensively encapsulated the research progress and practical cases of early pavement distress detection [23]. With the continuous advancement of sensor technology, various commercial pavement inspection systems are now capable of multidimensional data collection concerning pavement surface characteristics, and intelligent recognition of pavement distresses based on different data types has emerged as a new research direction.
In recent years, pavement inspection systems based on three-dimensional laser scanning technology have increasingly been employed for pavement damage and texture assessment, producing high-quality 3D pavement images. The team led by Kelvin Wang [24] at Oklahoma State University has developed a pavement data collection vehicle equipped with the PaveVision3D high-resolution 3D imaging system, which enables the rapid collection of millimeter-level three-dimensional data for lanes. Commercial multifunctional inspection systems, such as PSI’s PathRunner, Canada’s ARNA 9000 [25], and ZOYON-RTM from Wuhan University of Technology [7], have all integrated laser 3D imaging modules, which can generate high-precision laser point cloud depth maps, achieving collection speeds of up to 100 km/h.
Two-dimensional color images and three-dimensional depth images serve as the primary data sources for current pavement distress identification. Two-dimensional image recognition techniques primarily rely on visual color differences to distinguish distressed areas. Typically, damaged regions, such as cracks and potholes, appear darker compared to normal pavement, enabling distress identification through pixel variations. In contrast, three-dimensional image recognition methods predominantly employ laser scanning and stereo imaging to generate depth images, identifying distresses based on the height characteristics of the distress areas, which are generally significantly higher or lower than the pavement surface. Currently, most research efforts focus on the automated identification of pavement distresses based on two-dimensional images, which can be broadly categorized into image processing-based recognition methods and deep learning-based recognition methods. While two-dimensional images offer the advantage of ease of acquisition, they are susceptible to interference from factors such as lighting, color, and pavement markings, posing substantial analytical challenges in complex environments. Xu Zhigang [26], in summarizing the latest developments in pavement damage detection systems and image recognition algorithms, pointed out that although deep learning methods can enhance the generalization of distress recognition, algorithms based on two-dimensional distress images struggle to achieve an optimal balance among accuracy, robustness, and real-time performance. Furthermore, the advancement of industrial three-dimensional imaging technology has provided various pathways for the digitization of pavement distresses. Mathavan [17] reviewed the applicability of various three-dimensional imaging technologies across different contexts, highlighting their significant potential for application in pavement distress detection and measurement. Researchers such as Cao [16] and Yang [27] have summarized the progress in the automatic recognition of three-dimensional distress images, emphasizing that the integration of three-dimensional imaging and deep learning can notably enhance the robustness of distress identification. It is anticipated that the convergence of three-dimensional imaging technology and deep learning methodologies for the intelligent diagnosis of pavement distress at different scales will become a key focus in future research.
In light of the current research trends both domestically and internationally, as well as a preliminary assessment of future research focuses, it has been decided to utilize Unmanned Aerial Vehicles to assist in the identification and detection of large-scale pavement distresses. Unmanned Aerial Vehicle photography platforms possess the capability to efficiently capture pavement images, enabling the three-dimensional reconstruction of large-scale scenes. To evaluate the spatial accuracy of pavement models generated by the UAV Structure-from-Motion (UAV-SfM) stereo vision system, some studies have utilized Ground Control Points (GCP) and terrestrial laser scanning to estimate the error distribution of the generated point clouds [28,29]. Based on the pavement point cloud models generated by UAV-SfM, Inzerillo et al. [30] measured the depth and width of cracks in asphalt pavements, while Roberts et al. [5] analyzed the deformation severity of the pavement surface using digital models. Furthermore, similar UAV-SfM workflows have been utilized to efficiently measure the International Roughness Index (IRI) of pavements and potholes [31,32]. To determine the optimal flight altitude for drones, Saad [10] and Romero-Chambi [6] quantified the impact of flight height on the three-dimensional reconstruction of potholes, discovering that low-altitude flights could achieve high-precision distress measurements. Although the UAV-SfM stereo vision system can efficiently reconstruct entire pavement scenes, the ongoing traffic from vehicles on the road has limited current research and applications, primarily to closed rosad environments with minimal traffic [10,30,31].

2. UAV Data Acquisition

2.1. UAV Flight Platform and Flight Parameters

The UAV used in this experiment is the DJI Phantom4 Pro V2.0 (Chinese DJI brand) (in Figure 1). Equipped with a flight remote control, the maximum control distance is 6 km, the maximum flight height is 500 m, and the duration of a single flight can reach 25 to 30 min. At the same time, it is equipped with a high-resolution camera, which has high spatial accuracy and image clarity. Table 1 shows the detailed equipment parameters and flight parameters.
It should be noted that the quality of 3D reconstruction based on UAV photography is affected by many factors, such as flight height, image overlap rate, and environmental noise, and photography parameters should be reasonably set according to different scenes. In dynamic traffic scenes, vehicles running on the road will block the road surface, so improving the image overlap rate is necessary to avoid completely blocking the road surface. The image overlap rate will also affect the quality of 3D model reconstruction. Typically, the image overlap rate is set to 75%, the flight height of the UAV along the road is set to 15 m, and the flight speed to 2 m/s to enhance the accuracy of point cloud reconstruction.

2.2. Data Acquisition

The data collection area is on Fuyao Road, Fuping County, Shaanxi Province. The road runs across Fuping County from east to west, with a total length of 41.4 km. The two ends of Fuyao Road are connected to the expressway, and the freight demand of the surrounding industries leads to the daily traffic of cargo vehicles on the test road, which are mainly medium and heavy trucks, resulting in frequent road diseases. Based on this, we selected the road sections with large traffic flow and serious diseases for collection through field survey. To encompass a variety of road diseases., this test road was collected in five sections, with a total length of about 5 km. The specific division is shown in Table 2.
In the process of acquisition, the waypoint control software DJI GS Pro (GS Pro2.0) is used to plan the mapping task of UAVs. According to the specified flight and photography parameters, different routes and waypoints are automatically generated in the detection area. Figure 2 shows the route planning diagram of the detection section. After the route planning is completed, the UAV takes off from both sides of the road. The lens angle is perpendicular to the lane, flies according to the predetermined flight path and waypoint, automatically collects the road surface image at the same time, and adjusts the flight height and speed in real time according to the actual situation of the road.

3. Methods

3.1. Three-Dimensional Road Modeling Based on UAV Imagery

3.1.1. Vehicle Positioning and Removal Based on YOLOV8 Network Algorithm

Since the vehicle will block the road surface in the photos collected by UAV, it needs to be identified and removed from the aerial image to improve the accuracy and efficiency of 3D reconstruction. Compared to various object detection deep learning algorithms, the YOLO network is widely used because it offers the best trade-off between accuracy and speed.
YOLOv8 inherits the consistent feature of the YOLO series—an end-to-end object detection framework that generates target bounding boxes and category labels directly from the input image, so the YOLOV8 detector is used to identify vehicles in UAV images. Figure 3 shows the network architecture of YOLOv8.
In this experiment, the Fuyao Road data set collected by UAV was selected for self-collected datasets. Firstly, the image was processed into 640 × 640 size, and the data set was further divided into training set, test set, and verification set, with a ratio of 8:1:1, to support the effective training and performance evaluation of the model.
In the training process, the validation set is used to evaluate the model performance, and the convergence of the model’s loss function and other indicators such as Precision, Recall, and mean average accuracy (mAP) are checked. These basic assessment measures can be estimated based on True Positive (TP), True Negative (TN), False Positive (FP), and False Negative (FN). mAP is used to evaluate multi-classification target recognition tasks, while AP is used to evaluate the accuracy of single target recognition tasks. Therefore, AP is used in this paper to evaluate the vehicle recognition accuracy of aerial photography images. After the training is completed, the test set is used to conduct independent tests to evaluate the generalization ability and performance of the model, especially the accuracy of vehicle detection under different flight heights, angles, and lighting conditions of the UAV. The Precision and Recall can be defined as follows:
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
where TP is the number of correctly detected vehicles, FP is other objects incorrectly detected as vehicles, and FN is the number of undetected vehicles. The AP and mAP can be defined as follows:
A P = 0 1 P r e c i s i o n ( R ) d R
m A P = i = 0 n A P i n
In the evaluation of the target recognition model, the TP value is affected by the IoU (Intersection over Union) threshold, and the accuracy of the prediction boundary box is determined by whether the IoU value between the prediction box and the truth box is greater than the preset threshold. Therefore, Pr, Re, and AP also depend on the choice of IoU thresholds. In this experiment, the IoU threshold was set at 0.5 to evaluate the vehicle recognition results.
The location of the vehicle in the image was identified by the YOLOv8 detector, and the detection box of the occluded area was replaced with a white block with pixel 0. In this experiment, to eliminate the occluded area of the road surface, UAV shooting technology was used to take multiple shots of the overlapping area. Even if the road surface in one image is partially obscured by vehicles, the obscured area can still be seen in other images. The overlap rate of UAV photography is set at 75%. Once the occluded area in each image is identified, the pavement occluded area can be subtracted from the overlapping area of the two adjacent images to estimate the actual situation of the road surface, as shown in Figure 4.

3.1.2. Three-Dimensional Road Surface Reconstruction Based on MVSNet

MVSNet is a depth estimation network based on deep learning. This network performs convolution learning on input 2D images from multiple perspectives through convolutional layers, extracts image features, and obtains feature maps containing image texture, shape, depth, and other information. Then, based on the generated feature maps and input camera position and pose information, The 3D cost body is constructed by homography warping, and then the obtained cost body is regularized to obtain the final probability body. Depth estimation and optimization are carried out from the probability body to generate the depth map of the image. Finally, the depth map is merged to generate a point cloud model to complete the reconstruction of the target scene. The network architecture of MVSNet is shown in Figure 5.
MVSNet 3D reconstruction requires the resolution of the input image to be 1200 × 1600. Therefore, the 3:4 ratio is used to take photos when collecting the pavement data set to avoid image distortion caused by subsequent image resolution processing. After the image with the appropriate pixel size is processed, colmap is used for preliminary preprocessing of both the training data set and test data set to obtain the camera pose and internal parameter information corresponding to each image.
For the training data set, after preliminary processing, colmap is used for depth estimation and the corresponding depth map of each image of the target scene is obtained.
Compared with the test data set, the processing of the training data set mainly involves the process of generating depth map files, which is mainly used for deep learning prediction of images in the model training process.
Under different environmental conditions, images can affect the quality of reconstruction. For instance, uneven lighting may cause local overexposure or underexposure in the image, which can affect the extraction and matching of feature points and, ultimately, the accuracy of 3D reconstruction. In low-light conditions, the image quality of the camera deteriorates, and noise increases. This makes it more difficult to accurately identify feature points during image matching, thereby affecting the accuracy and completeness of the reconstruction results. Therefore, in complex environmental conditions such as cloudy or rainy days, the model needs to undergo data augmentation to ensure its stability and accuracy under different weather and lighting conditions, thereby improving its generalization ability in these conditions.
By adjusting the brightness of the pictures in the training data set, the picture scenes under different weather conditions, such as sunny and cloudy days, are simulated. This simulation method can improve the generalization of the training model for scene reconstruction under different light scenes so that the model can more accurately reconstruct the real scene when facing various light conditions in the actual application scene. Figure 6 displays the target scene image and its simulated images under different weather conditions.
After the training data set is made, the model is trained in the deployed MVSNet environment. When the loss value reaches the threshold, the training stops automatically. Through further analysis and testing, the performance of MVSNet in practical pavement application scenarios can be more deeply understood and the model can be further optimized.

3.2. Disease Detection Technology Based on Multi-Scale and Full-View

3.2.1. Data Processing

Based on the above 3D reconstruction, the road surface point cloud plane is not parallel to the axis x and y plane, which affects the subsequent extraction of disease information based on coordinates. Therefore, it is necessary to carry out the coordinate rotation for the reconstructed 3D model. The reconstructed three-dimensional model of the road table contains the disease information of multiple lanes. For each lane, the vehicle load is concentrated on the wheel path. As a result, lateral and longitudinal surface deformations are symmetrically distributed on the lane. To measure rutting and flatness, lane division is necessary.
Before the rotation operation is performed, the plane coordinates of the road surface are determined first. In this study, the plane coordinates of the road surface are iterated by the RANSAC (Random Sample Consensus) algorithm, and then the normal vector of the road surface is determined. Figure 7 shows the coordinate rotation process.
As a basic road sign, lane lines are often used to define the trajectory of vehicles. In this study, road markings are extracted by intelligent extraction of lane demarcation lines, and then lane division is carried out.
First, the point cloud data of the multi-vehicle road table are loaded into the processing system, which contains the spatial coordinates (X, Y, Z) of each point, as well as color information. According to the color information, set the color threshold and screen out the points close to white, which is the extracted lane line point cloud. The noise removal process based on a statistical filter is carried out to reduce the subsequent fitting error. After the denoised mark line point cloud is obtained, the linear fitting based on the least square method is carried out to determine the mark line linear equation. According to the linear equation obtained, the threshold value is set for the whole road surface point cloud data for segmentation, and the lane point cloud is determined for lane division. Figure 8 shows the point cloud image after lane segmentation.

3.2.2. Calculation of Road Roughness Based on Vehicle Road Model

Firstly, the lane area is located through lane line segmentation and fitting, and the profile plane extraction line is set according to the curvature of the lane line in the lane according to a certain transverse interval distance. To cover the profile data of the whole lane and the smoothness evaluation, the longitudinal profiles and elevation values of 1/4 and 3/4 transverse positions of each lane were extracted at fixed intervals for a single lane, which was used to calculate the international smoothness values of multiple transverse positions in subsequent models.
The evaluation index of road roughness is mainly based on the road-suspension response model. The road roughness is characterized by the vertical vibration response generated by the vehicle suspension moving on the road. The vehicle suspension model is the core of flatness evaluation. The degree of freedom of the suspension model can be divided into the two-degrees-of-freedom quarter car model, a four-degrees-of-freedom half-car model, and a seven-degrees-of-freedom whole-vehicle model. Among them, the International Roughness Index (IRI) is the most commonly used pavement roughness evaluation index. This index is a response evaluation index proposed by the World Bank to measure the driving quality of infrastructure in a unified way. Its core is to estimate the dynamic response of vehicle suspension relative to road profile by using a two-degrees-of-freedom quarter car model.
As shown in Figure 9, the vehicle suspension system includes auxiliary mechanisms such as springs, damping shock absorbers, swing arms, and tires. The mathematical model of a quarter car suspension is composed of a tire, a spring with stiffness K s   and a shock absorber with damping C s , in which the unsprung mass is m s and the unsprung mass is m u . The system assumes that the tire is always in contact with the road, the tire stiffness is K t l, and the vehicle speed is fixed at 80 km/h. According to the one-quarter car model, IRI can be calculated by the following formula:
I R I = 1 L 0 L V Z ˙ s Z ˙ u d t
I R I = 1 L 0 L Z s Z u d x
where L is the length of the road profile; V is the speed of the vehicle; Z ˙ s is the vertical velocity of the sprung mass part; Z s is the vertical movement distance of the spring-mass part; Z ˙ u is the vertical velocity of the unsprung mass part; Z u is the vertical movement distance of the unsprung mass part; d t is the time differential; and d x is the distance differential.
The essence of the IRI calculation formula above is that when the vehicle travels along the road profile curve at a certain speed, the suspension will vibrate up and down under the excitation of the uneven road surface. Therefore, the cumulative motion trajectory of the vehicle suspension relative to the vehicle body is calculated to reflect the road surface flatness. For ease of solution, the formula can be simplified as follows:
I R I = 1 L 0 L Z s Z u d x
where n is the sample number of test points.
According to the above discrete calculation formula, IRI value calculation must obtain the vertical distance difference between each measuring point’s upper and lower parts. In the field physical measurement process, the motion distance of the two parts can be measured by installing sensors on the bottom of the suspension and the body of the test car. In the process of digital simulation, these two parameters need to be solved using the road profile curve and suspension dynamic equation. With the upper part of the suspension spring and the lower part of the spring divided into the center, the following second-order vibration differential equation can be established.
m s Z ¨ s = K s Z u Z s + C s Z ˙ u Z ˙ s
m u Z ¨ u = K s Z u Z s C s Z ˙ u Z ˙ s + K t Z r Z u
where Z ¨ s is the vertical acceleration of the spring part; Z ¨ u is the vertical acceleration of the unsprung part; and Z r is the vertical change distance of the pavement profile.
According to the standard suspension parameters of the quarter car, u = mu/ms = 0.15, C = Cs/ms = 6.00 s−1, K1 = Ks/ms = 653 s−2, K2 = Kt/ms = 63.3 s−2. The above Formulas (10) and (11) can be simplified as follows:
Z ¨ s = K 1 Z u Z s + C Z ˙ u Z ˙ s
Z ¨ u = K 1 μ Z s Z u + C μ Z ˙ s Z ˙ u + K 2 μ Z r Z u
The solution of the above equation depends on calculating Z ˙ s Z ¨ s Z ˙ u Z ¨ u these four state variables, Z t = Z s Z ˙ s Z u Z ˙ u T , which can be converted to the following:
d Z t d t = A Z t + B Z r t
A = 0 1 K 1 C         0                     0         K 1                   C 0 0 K 1 μ C μ         0 1 K 1 + K 2 μ C μ        
B = 0 0 0 K 2 μ
Formula (11) is a non-homogeneous differential equation. According to the initial value of Z t and the gradient of the section curve, Z t at any time can be solved by recursion, and the motion of Z s and Z u at each measuring point can be obtained to calculate IRI.

3.2.3. Pavement Abnormal Deformation Recognition Based on DBSCAN Clustering

In addition to the two common deformation forms of transverse rutting deformation and longitudinal pavement deformation, large-scale abnormal deformation also exists on some pavement, which is mainly caused by factors such as uneven settlement of subgrade, insufficient compaction degree, and shrinkage and expansion of materials. It is difficult to identify large-scale abnormal deformation of pavement because of its random distribution, fuzzy boundary, and complex shape. Limited by the spatial distribution range and complex change characteristics of large-scale abnormal deformation, most current studies reflect the abnormal deformation through the multi-angle one-dimensional cross-section. However, to fully measure the three-dimensional shape of large-scale abnormal deformation, it is necessary to extract the two-dimensional boundary of the disease from the three-dimensional digital elevation map. Therefore, considering the complex form and fuzzy boundary of large-scale abnormal deformation, this paper adopts unsupervised three-dimensional elevation clustering to extract spatial elevation demarcation points and identify abnormal deformation regions.
The DBSCAN algorithm identifies clusters by density. Its core principle is to divide data into core points, boundary points, and noise points according to the density of data points. First, the algorithm selects a distance threshold ( ε ) and a minimum number of neighbor points (MinPts) and then starts from any unvisited core point and identifies points in its neighborhood. If the number of points in the neighborhood exceeds MinPts, the points are grouped into the same cluster, and the cluster is recursively expanded until no more points can be added. Finally, the resulting clusters have arbitrary shapes and can effectively identify noisy data. The overall search process is shown in Figure 10.
The input to the DBSCAN cluster is multidimensional data and is suitable for point cloud data with different densities. Its main goal is to cluster data points based on their density characteristics, which is particularly suitable for discrete, non-spherical spatial distributions. In the clustering process of DBSCAN, you first need to define two important parameters: ε (neighborhood radius) and MinPts (minimum number of points in the neighborhood).
Initial step: In three-dimensional space, each observed data point is first classified according to the given ε and MinPts conditions. For each point p, calculate the number of points in its ε neighborhood, defined as follows: N ϵ ( p ) , if N ϵ ( p ) M i n P t s , then p is labeled as the core point. If p is in the neighborhood of a core point but does not satisfy the core point condition, then p is labeled as a boundary point.
Clustering procedure: Select a core point p from any unvisited point and create a new cluster C. All points in the neighborhood of p and its ε are added to cluster C. This process can be expressed as follows: C = p N ϵ ( p ) . For each newly added point q if q is also the core point, the neighborhood of q is repeatedly checked, and the cluster C is continued to expand. This process continues until no new core can be found.
Clustering complete: At the end of the process, all the core points and the points in their neighborhood form a cluster, and those that are neither core nor boundary points are labeled as noise.
DBSCAN depends on the distance to define a point’s neighborhood. If the horizontal change is obvious but the vertical change is small, the clustering effect of the points is not ideal. In this study, a 3D digital elevation model is used for clustering. Compared with horizontal distribution information, elevation change information is more effective in accurately identifying deformation regions. Therefore, in the process of three-dimensional spatial clustering, weighted factors should be set to scale the data features of different dimensions and strengthen the attention of spatial clustering.
For the original 3D elevation data, the vertical deformation direction data are weighted based on an unchanged horizontal data scale to enhance the difference of this dimension data. Formula (15) is used to solve the weighting factor w of the vertical data so that the variance of the vertical weighted data is not less than that of the horizontal data.
i = 1 m w 2 · z i z ¯ 2 m i = 1 m x i x ¯ 2 + y i y ¯ 2 m
where x ¯ , y ¯ , z ¯ are the average values of x axis, y axis, and z axis direction data; m is the total number of sample points of observation data.

4. Results and Discussion

4.1. Comparison of Three-Dimensional Reconstruction Results

After the exit point cloud is reconstructed from the multi-view two-dimensional images of the target pavement scene through MVSNet, we will conduct a comparison experiment between the pavement point cloud obtained in this study and the point cloud reconstructed by traditional colmap and photoscan: the point cloud reconstructed by colmap has higher quality but is slower in terms of reconstruction speed. Although Photoscan is used for three-dimensional pavement reconstruction, the reconstruction speed is fast, but the reconstruction effect is poor, so the detailed comparison experiment between MVSNet and Photoscan reconstruction is not carried out. The evaluation of MVSNet 3D reconstruction is mainly completed through detailed comparison experiments with colmap 3D reconstruction with higher reconstruction accuracy. MVSNet can achieve the reconstruction of the target road scene in ten minutes. In this chapter, the road surface point cloud model reconstructed by MVSNet and the road surface point cloud model reconstructed by the traditional three-dimensional reconstruction method colmap will be compared and tested from two aspects of speed and accuracy under the same hardware environment and the same road surface scene. The results of comparative experiments are analyzed and discussed. Figure 11 illustrates the results of reconstruction using different methods.

4.1.1. Comparison Experiment of 3D Reconstruction Accuracy

In this experiment, the high-precision point cloud model reconstructed by colmap is used as a reference to compare the error between the point cloud reconstructed by MVSNet and colmap. The two groups of point clouds are registered by the ICP (Iterative Closest Point) algorithm. ICP is an iterative process in which registration errors are gradually reduced. To greatly improve the calculation speed, an optimization scheme is used to randomly sub-sample the data clouds at each iteration. The parameter is set to 50,000, which is the maximum number of subsampling points. After registration, the average distance between the two clouds is calculated. The average distance of the point cloud is 0.002849, and the standard deviation is 0.025598 (dimensionless).
Since the calculated average distance of the point cloud is dimensionless, the real-world dimensions corresponding to the point cloud are required for dimensional recovery. Take the coordinates of different feature points in the point cloud, calculate their distance d (dimensionless) in the computer world, and then measure the distance D between the corresponding feature points in the corresponding real world. Multiple groups of d and D can be taken for calculation, and the ratio r between the real scene and the reconstructed point cloud size is 73.1 mm:1, as shown in Formula (16).
r = d D = 1 73.1   m m
Therefore, it can be calculated that the average error of point cloud accuracy reconstructed by MVSNet and colmap is 0.57 mm, and the standard deviation is 0.33 mm (as shown in the following Table 3), which indicates that the point cloud accuracy reconstructed by the two is very close.
From the histogram of the distance between two clouds (see Figure 12), it can be seen that the total number of point clouds reconstructed by MVSNet is 6,582,878 points, most of which are distributed in the range of 0 to 0.05 in absolute value.
The results show that the error of the reconstructed point cloud in this study is very small, and the accuracy of the reconstructed point cloud is very similar to that of the reconstructed point cloud in the traditional colmap method.

4.1.2. Comparison of Point Cloud Reconstruction Speed and Density

The evaluation of 3D reconstruction not only needs to evaluate the reconstruction accuracy but also needs to evaluate the time spent on reconstruction.
As for the statistics of the time spent on reconstruction, since the data preprocessing environment of MVSNet reconstruction requires colmap, and MVSNet and colmap have similar processes for data preprocessing, the calculation of the time spent on reconstruction of the two methods does not include the data preprocessing process. Because the resource response allocated by the cloud server may have a certain delay error, the average time of three times of the same reconstruction process is used to calculate the time.
After completing the sparse reconstruction of 178 view pictures of the target road scene and estimating the pose of the corresponding camera, colmap began to carry out the subsequent reconstruction work. This stage included dedistortion, depth estimation, and deep fusion to generate a high-precision point cloud, and the entire reconstruction process took 91 min.
The reconstruction time of MVSNet starts from the input of the processed camera pose and picture and takes 8 min to generate and complete the reconstruction by inputting the picture of the target road scene and the corresponding camera pose information.
The point clouds reconstructed by MVSNet and colmap were respectively intercepted in areas of the same size to calculate the point cloud density. The results are shown in Table 4. The density of the road surface point clouds reconstructed by MVSNet and colmap was 412/cm2, and the density of the road surface point clouds reconstructed by colmap was 138/cm2. The former was 3.32 times the latter.
This result shows that under the same data set or experimental conditions, the MVSNet model trained in this study has higher reconstruction speed and better accuracy than the traditional reconstruction method colmap and has a significant improvement in the three-dimensional reconstruction of road tables. This technology can quickly complete high-quality 3D reconstruction of pavement, and the reconstruction efficiency is much higher than that of traditional 3D reconstruction technologies such as laser scanning and radar scanning. Compared with these techniques, this study only needs to use multi-view two-dimensional images obtained by ordinary cameras for reconstruction, and the cost is lower. But for complex urban environments, better flight path planning and target recognition algorithms may need to be designed to handle occlusion and other complicating factors.

4.2. Test Results of Multi-Section Smoothness and Cause Analysis

In this study, the three-dimensional model of K18+000–K19+000, K19+000–K20+000, K31+000–K32+000, K37+800–K39+000 4 km road section was randomly selected for longitudinal profile observation and flatness analysis. Among them, the longitudinal profiles and elevation values of 1/4 and 3/4 transverse positions of each lane are extracted at fixed intervals and brought into the vehicle model to calculate the smoothness value. According to the Field Test Method of Roadbed and Pavement of Highway Engineering (JTG 3450-2019), the sampling interval of the vehicle laser IRI instrument should be less than 50 cm. In this study, a sampling interval of 25 cm was used to calculate IRI. The IRI of the two profiles on each lane was averaged to indicate the flatness of the entire lane. The following Table 5 shows the smoothness calculation results of the K31+000–K32+000 road section.
According to the test results of the roughness of the four-kilometer road section, this study draws the scatter plot of the average IRI value of the six-lane road of the four-kilometer road, as shown in Figure 13, to analyze the road roughness in this area. In K18+000–K19+000, nearly half of the IRI values of the down-middle lane and up-right lane are greater than 4, and the road roughness is poor. In K19+000–K20+000, more than 90% of the IRI values are above 4, but most of them are below 8. The upgoing left lane smoothness is relatively good, and the entire smoothness of the kilometer road deviates. In K31+000–K32+000, the overall IRI value of the ascending right lane is larger, and the smoothness is worse than that of other lanes. The ascending left lane and descending left lane have better smoothness, and the overall smoothness of the road in this kilometer is incorrect. In K37+800–K39+000, the trend of IRI values of all lanes is uniform, and most of them are above 4. The upgoing left lane has good smoothness, but according to the trend, it is also gradually deteriorating, and the overall smoothness of the road in this kilometer deviates. Four kilometers of roads were randomly selected for intercept profile observation and roughness analysis, and the results were poor, indicating that the overall roughness of Fuyao Road should be deviated and need to be repaired. Among them, the IRI values of most lanes and road sections are greater than 4. It is recommended to carry out comprehensive road surface repairs, including filling cracks, repairing potholes, and polishing the road surface, and regularly monitor the lanes to ensure that their flatness is maintained in good condition.

4.3. Settlement Detection Results and Cluster Analysis

In this study, the digital elevation model of four connected road sections was extracted to identify abnormal deformation of pavement based on DBSCAN clustering, and the clustering accuracy was improved through the data features of weighted pavement elevation dimensions. Table 6 shows the images before and after the clustering of this section. To display the road settlement information more clearly, image processing is carried out on the images after clustering to enhance the settlement characteristics and generate the disease analysis diagram.
According to the road disease analysis chart, the road deformation diseases of this section mainly include rutting, hugging, and local uneven subsidence. Road rutting and hugging are mainly concentrated in road intersections with traffic lights and at the bottom of the downslope, and their distribution locations are mainly located in the middle lane and the side lane near the central divider. The reason for the formation is that there are more heavy-duty vehicles, mainly driving in the middle lane and the side lane of the central divider. Under the repeated action of heavy-duty traffic, rutting and hugging are formed at the wheel track belt. The high-temperature deformation resistance of surface material is insufficient, which leads to rutting more easily. Due to poor bonding between layers, uplift occurs under the lateral action of load.
The irregular settlement of the road section is mainly in the inner two lanes of the road and is often accompanied by mesh cracks. The main causes of the irregular settlement are analyzed as there are more heavy vehicles in the middle lane. The roadbed and soil foundation in the local area of the road are weak, resulting in local subsidence under the repeated action of heavy load traffic, and fatigue cracks caused by the repeated load are not handled in time, resulting in mesh cracks. Local failure occurs in the base layer, resulting in local emptying, and the asphalt surface deformation occurs under the action of load.

5. Conclusions and Future Work

This study proposes and validates a 3D reconstruction and large-scale detection method of road surface based on UAV and deep learning technology to improve the efficiency and accuracy of road maintenance. By integrating the improved YOLOv8 model and the MVSNet 3D modeling algorithm, the fast and accurate 3D reconstruction of the road was successfully realized. By introducing point cloud processing extraction technology and the DBSCAN clustering algorithm, the system can effectively identify settlement areas and detect flatness. This method is especially suitable for detecting complex pavement deformation and provides an innovative solution for pavement health assessment.
Compared with traditional vision-based detection methods, the method proposed in this article can effectively improve the accuracy and efficiency of 3D reconstruction through the improved YOLOv8 model and MVSNet 3D reconstruction algorithm. Compared with detection methods based on LiDAR, this method has a lower cost and is suitable for large-scale detection tasks. However, in strong light scenes, the image quality captured by drones may be affected, leading to a decrease in the accuracy of object detection and 3D reconstruction. Meanwhile, when dealing with large-scale data, 3D reconstruction and disease detection still require high computational resources, which will limit the application of this method in real-time detection.
Although the method in this study has achieved good results in reconstruction accuracy and large-scale detection range, there is still room for further optimization. Future research may consider the following directions:
  • In practical applications, the accuracy of detection using Unmanned Aerial Vehicles needs to be verified by comparing the detection with the current standard specifications;
  • In complex urban environments, especially where there are many buildings and heavy traffic, the difficulty of 3D reconstruction can increase significantly. To improve adaptability in these environments, improved flight path planning, improved image resolution, and enhanced recognition of occluding objects (such as buildings, traffic signs, pedestrians, etc.) may be required. To improve the robustness and generalization of the system, a more diverse set of training data should be used, covering different weather conditions, light environments, and different types of roads;
  • Based on the existing DBSCAN clustering, the unsupervised learning algorithm is further optimized to make the system more accurate in the boundary detection of irregular diseases, especially in scenes of complex disease types and fuzzy boundaries;
  • Research more efficient algorithm optimization techniques, such as model compression and lightweight network structures, to reduce computational resource consumption and enable detection tasks in complex road scenes.
Through further research and optimization, this method will show higher practicability in the field of road detection and maintenance and provide strong technical support for intelligent road management and maintenance.

Author Contributions

Conceptualization, X.Z., S.C. and P.W.; methodology, X.Z., S.C. and P.W.; software, X.Z. and S.C.; validation, X.Z. and P.W.; formal analysis, S.C.; investigation, P.W.; resources, X.Y.; data curation, H.Z. and Y.G.; writing—original draft preparation, X.Y., H.Z. and Y.G.; writing—review and editing, X.Z.; visualization, S.C. and P.W.; supervision, X.Z.; project administration, S.C.; funding acquisition, X.Z., S.C. and P.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article. Further inquiries can be directed to the corresponding author.

Acknowledgments

The success of this project is thanks to the data and technical support provided by Yunnan Provincial Traffic Science Research Institute Co., Ltd., as well as the professional advice and guidance provided by the Yang Xu research group from Chang’an University.

Conflicts of Interest

Author Xiang Zhang was employed by the company Yunnan Transportation Science Research Institute Co., Ltd. Author Shuwei Cheng was employed by the company Yunnan Transportation Science Research Institute Co., Ltd. Author Pu’an Wang was employed by the company Yunnan Transportation Science Research Institute Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Abbreviations

1. Abbreviations
UAVUnmanned Aerial Vehicle
MVSNetMulti-View Stereo Network
DBSCANDensity-Based Spatial Clustering of Applications with Noise
IRIInternational Roughness Index
TPTrue Positive
TNTrue Negative
FPFalse Positive
FNFalse Negative
APAverage Precision
mAPmean Average Precision
IoUIntersection over Union
RANSACRandom Sample Consensus
ICPIterative Closest Point
CPUCentral Processing Unit
GPUGraphics Processing Unit
FOVField of View
CMOSComplementary Metal-Oxide-Semiconductor
2. Lists of symbols
SymbolUnitMean
vm/sSpeed
tsTime
dmDistance
am/s2Acceleration
mkgQuality
FNPower
π Pi
e Base of natural logarithm
fsHzSampling Frequency
A Matrix
I Identity matrix
Infinity
d Differential symbol
Integral symbol
AT Transpose matrix
µ Coefficient of friction
k Scale factor

References

  1. Liu, C.L.; Wu, D.F.; Li, Y.S.; Du, Y.C. Large-scale pavement roughness measurements with vehicle crowdsourced data using semi-supervised learning. Transp. Res. Part C-Emerg. Technol. 2021, 125, 103048. [Google Scholar] [CrossRef]
  2. Loprencipe, G.; Zoccali, P. Comparison of methods for evaluating airport pavement roughness. Int. J. Pavement Eng. 2019, 20, 782–791. [Google Scholar] [CrossRef]
  3. Luo, W.T.; Liu, L.X.; Li, L. Measuring rutting dimension and lateral position using 3D line scanning laser and inertial measuring unit. Autom. Constr. 2020, 111, 103056. [Google Scholar] [CrossRef]
  4. Peraka, N.S.P.; Biligiri, K.P. Pavement asset management systems and technologies: A review. Autom. Constr. 2020, 119, 103336. [Google Scholar] [CrossRef]
  5. Roberts, R.; Inzerillo, L.; Di Mino, G. Using UAV Based 3D Modelling to Provide Smart Monitoring of Road Pavement Conditions. Information 2020, 11, 568. [Google Scholar] [CrossRef]
  6. Romero-Chambi, E.; Villarroel-Quezada, S.; Atencio, E.; Muñoz-La Rivera, F. Analysis of Optimal Flight Parameters of Unmanned Aerial Vehicles (UAVs) for Detecting Potholes in Pavements. Appl. Sci. 2020, 10, 4157. [Google Scholar] [CrossRef]
  7. Zhang, D.J.; Zou, Q.; Lin, H.; Xu, X.; He, L.; Gui, R.; Li, Q.Q. Automatic pavement defect detection using 3D laser profiling technology. Autom. Constr. 2018, 96, 350–365. [Google Scholar] [CrossRef]
  8. Liu, L.; Zhu, P.; Guan, J.X.; Jiang, R.; Zhou, X.L. A binocular reconstruction method fused with Laplacian image information for pavement texture evaluation. Measurement 2021, 185, 110039. [Google Scholar] [CrossRef]
  9. Ibragimov, E.; Lee, H.J.; Lee, J.J.; Kim, N. Automated pavement distress detection using region based convolutional neural networks. Int. J. Pavement Eng. 2022, 23, 1981–1992. [Google Scholar] [CrossRef]
  10. Saad, A.M.; Tahar, K.N. Identification of rut and pothole by using multirotor unmanned aerial vehicle (UAV). Measurement 2019, 137, 647–654. [Google Scholar] [CrossRef]
  11. Barbarella, M.; De Blasiis, M.R.; Fiani, M. Terrestrial laser scanner for the analysis of airport pavement geometry. Int. J. Pavement Eng. 2019, 20, 466–480. [Google Scholar] [CrossRef]
  12. Van Geem, C.; Bellen, M.; Bogaerts, B.; Beusen, B.; Berlémont, B.; Denys, T.; De Meulenaere, P.; Mertens, L.; Hellinckx, P. Sensors on vehicles (SENSOVO)–Proof-of-concept for road surface distress detection with wheel accelerations and ToF camera data collected by a fleet of ordinary vehicles. In Proceedings of the 6th Transport Research Arena (TRA), Warsaw, Poland, 18–21 April 2016; pp. 2966–2975. [Google Scholar]
  13. Fu, M.Y.; Liang, H.; Zhu, C.H.; Dong, Z.P.; Sun, R.D.; Yue, Y.F.; Yang, Y. Image Stitching Techniques Applied to Plane or 3-D Models: A Review. IEEE Sens. J. 2023, 23, 8060–8079. [Google Scholar] [CrossRef]
  14. Guan, J.C.; Yang, X.; Ding, L.; Cheng, X.Y.; Lee, V.C.S.; Jin, C. Automated pixel-level pavement distress detection based on stereo vision and deep learning. Autom. Constr. 2021, 129, 103788. [Google Scholar] [CrossRef]
  15. Puzzo, L.; Loprencipe, G.; Tozzo, C.; D’Andrea, A. Three-dimensional survey method of pavement texture using photographic equipment. Measurement 2017, 111, 146–157. [Google Scholar] [CrossRef]
  16. Cao, W.M.; Liu, Q.F.; He, Z.Q. Review of Pavement Defect Detection Methods. IEEE Access 2020, 8, 14531–14544. [Google Scholar] [CrossRef]
  17. Mathavan, S.; Kamal, K.; Rahman, M. A Review of Three-Dimensional Imaging Technologies for Pavement Distress Detection and Measurements. IEEE Trans. Intell. Transp. Syst. 2015, 16, 2353–2362. [Google Scholar] [CrossRef]
  18. Zhang, A.; Wang, K.C.P.; Li, B.X.; Yang, E.H.; Dai, X.X.; Peng, Y.; Fei, Y.; Liu, Y.; Li, J.Q.; Chen, C. Automated Pixel-Level Pavement Crack Detection on 3D Asphalt Surfaces Using a Deep-Learning Network. Comput. Aided Civ. Infrastruct. Eng. 2017, 32, 805–819. [Google Scholar] [CrossRef]
  19. Bursanescu, L.; Blais, F.; Nat Res Council, C. Automated pavement distress data collection and analysis: A 3-D approach. In Proceedings of the International Conference on Recent Advances in 3-D Digital Imaging and Modeling, Ottawa, ON, Canada, 12–15 May 1997; pp. 311–317. [Google Scholar]
  20. Fukuhara, T.; Terada, K.; Nagao, M.; Kasahara, A.; Ichihashi, S. Automatic Pavement-Distress-Survey System. J. Transp. Eng. 1990, 116, 280–286. [Google Scholar] [CrossRef]
  21. Laurent, J.; Lefebvre, D.; Samson, E. Development of a new 3D transverse laser profiling system for the automatic measurement of road cracks. In Proceedings of the Symposium on Pavement Surface Characteristics, Portoroz, Slovenia, 20–23 October 2008. [Google Scholar]
  22. Groeger, J.L.; Stephanos, P.; Dorsey, P.; Chapman, M. Implementation of automated network-level crack detection processes in Maryland. In Pavement Assessment, Monitoring, and Evaluation 2003: Pavement Design, Management, and Performance; Transportation Research Record; Transportation Research Board Natl Research Council: Washington, DC, USA, 2003; pp. 109–116. [Google Scholar]
  23. Mcghee, K.H. Automated Pavement Distress Collection Techniques. NCHRP Synthesis 334; Transportation Research Board: Washington, DC, USA, 2004. [Google Scholar]
  24. Li, B.X.; Wang, K.C.P.; Zhang, A.; Fei, Y.; Sollazzo, G. Automatic Segmentation and Enhancement of Pavement Cracks Based on 3D Pavement Images. J. Adv. Transp. 2019, 9, 1813763. [Google Scholar] [CrossRef]
  25. Cafiso, S.; D’Agostino, C.; Delfino, E.; Montella, A. From manual to automatic pavement distress detection and classification. In Proceedings of the 5th IEEE International Conference on Models and Technologies for Intelligent Transportation Systems (MT-ITS), Napoli, Italy, 26–28 June 2017; pp. 433–438. [Google Scholar]
  26. Fang, Y.K.; Min, H.G.; Wang, W.Q.; Xu, Z.G.; Zhao, X.M. A fault detection and diagnosis system for autonomous vehicles based on hybrid approaches. IEEE Sens. J. 2020, 20, 9359–9371. [Google Scholar] [CrossRef]
  27. Yang, X.; Guan, J.C.; Ding, L.; You, Z.P.; Lee, V.C.S.; Hasan, M.R.M.; Cheng, X.Y. Research and applications of artificial neural network in pavement engineering: A state-of-the-art review. J. Traffic Transp. Eng. Engl. Ed. 2021, 8, 1000–1021. [Google Scholar] [CrossRef]
  28. Kubota, S.; Ho, C.; Nishi, K. Construction and Usage of Three-dimensional Data for Road Structures Using Terrestrial Laser Scanning and UAV with Photogrammetry. In Proceedings of the International Symposium on Automation and Robotics in Construction, Banff, AB, Canada, 21–24 May 2019; IAARC Publications: Montreal, QC, Canada, 2019. [Google Scholar]
  29. Liao, Y.J.; Wood, R.L. Discrete and Distributed Error Assessment of UAS-SfM Point Clouds of Roadways. Infrastructures 2020, 5, 87. [Google Scholar] [CrossRef]
  30. Inzerillo, L.; Di Mino, G.; Roberts, R. Image-based 3D reconstruction using traditional and UAV datasets for analysis of road pavement distress. Autom. Constr. 2018, 96, 457–469. [Google Scholar] [CrossRef]
  31. Prosser-Contreras, M.; Atencio, E.; La Rivera, F.M.; Herrera, R.F. Use of Unmanned Aerial Vehicles (UAVs) and Photogrammetry to Obtain the International Roughness Index (IRI) on Roads. Appl. Sci. 2020, 10, 8788. [Google Scholar] [CrossRef]
  32. Tan, Y.M.; Li, Y.X. UAV Photogrammetry-Based 3D Road Distress Detection. ISPRS Int. J. Geo-Inf. 2019, 8, 409. [Google Scholar] [CrossRef]
Figure 1. UAV flight acquisition platform.
Figure 1. UAV flight acquisition platform.
Materials 18 02133 g001
Figure 2. Route planning map for different detection sections.
Figure 2. Route planning map for different detection sections.
Materials 18 02133 g002
Figure 3. YOLOv8 detection network architecture diagram.
Figure 3. YOLOv8 detection network architecture diagram.
Materials 18 02133 g003
Figure 4. Road occluded area elimination.
Figure 4. Road occluded area elimination.
Materials 18 02133 g004
Figure 5. MVSNet network architecture.
Figure 5. MVSNet network architecture.
Materials 18 02133 g005
Figure 6. Data enhancement: (a) master drawing, (b) simulated overcast, (c) simulated sunny day.
Figure 6. Data enhancement: (a) master drawing, (b) simulated overcast, (c) simulated sunny day.
Materials 18 02133 g006
Figure 7. Coordinate rotation process.
Figure 7. Coordinate rotation process.
Materials 18 02133 g007
Figure 8. Extracted lane point cloud image.
Figure 8. Extracted lane point cloud image.
Materials 18 02133 g008
Figure 9. Quarter car model: (a) vehicle suspension structure, (b) quarter car model.
Figure 9. Quarter car model: (a) vehicle suspension structure, (b) quarter car model.
Materials 18 02133 g009
Figure 10. DBSCAN algorithm clustering principle diagram.
Figure 10. DBSCAN algorithm clustering principle diagram.
Materials 18 02133 g010
Figure 11. Comparison of reconstruction effect: (a) photoscan, (b) MVSNet, (c) colmap.
Figure 11. Comparison of reconstruction effect: (a) photoscan, (b) MVSNet, (c) colmap.
Materials 18 02133 g011
Figure 12. MVSNet reconstructed point cloud compared to colmap reconstructed point cloud distance map.
Figure 12. MVSNet reconstructed point cloud compared to colmap reconstructed point cloud distance map.
Materials 18 02133 g012
Figure 13. Scatterplot of average IRI values of six lanes on a four-kilometer road: (a) K18+000–K19+000 Scatter plot of six-lane average IRI values, (b) the K19+000–K20+000 Scatter plot of six-lane average IRI values, (c) The K31+000–K32+000 Scatter plot of six-lane average IRI values, (d) the K37+000–K39+000 Scatter plot of six-lane average IRI values.
Figure 13. Scatterplot of average IRI values of six lanes on a four-kilometer road: (a) K18+000–K19+000 Scatter plot of six-lane average IRI values, (b) the K19+000–K20+000 Scatter plot of six-lane average IRI values, (c) The K31+000–K32+000 Scatter plot of six-lane average IRI values, (d) the K37+000–K39+000 Scatter plot of six-lane average IRI values.
Materials 18 02133 g013
Table 1. Technical parameters of UAV flight platform.
Table 1. Technical parameters of UAV flight platform.
Camera ParameterSpecificationsProperties
Sensor1 in. CMOS
Resolution5472 px × 3648 px
Pixel size2.4 μ m   × 2.4 μ m
LensFOV 84° 8.8 mm/24 mm
Aerial flight parametersSpecificationsProperties
Flight altitude15 m
Flight speed2 m/s
Course overlap ratio75%
Side overlap ratio75%
Lens angle−90°
Vertical hover accuracy10 cm
Horizontal hover accuracy30 cm
Ground resolution0.4 cm/pix
Table 2. The division of the road detection area.
Table 2. The division of the road detection area.
NumberRange of Road PilesLength of Detention
1K18+000–K19+0001010 m
2K19+000–K20+0001010 m
3K25+900–K26+700800 m
4K31+000–K32+0001050 m
5K37+800–K39+0001212 m
Table 3. Error of MVSNet relative to colmap point cloud reconstruction.
Table 3. Error of MVSNet relative to colmap point cloud reconstruction.
ScaleMeanDistance Standard Deviation
1:73.1 mm0.0002840.025598
0.208261 mm1.871214 mm
Table 4. Comparison of reconstruction effect between the two methods.
Table 4. Comparison of reconstruction effect between the two methods.
Reconstruction MethodTimePoint Cloud Density
MVSNet8 min412 units/cm2
Colmap91 min138 units/cm2
Table 5. Test results of K31+000-K32+000 road section smoothness.
Table 5. Test results of K31+000-K32+000 road section smoothness.
K31+000–K32+000Upgoing IRI Value (m/km)Downgoing IRI Value (m/km)
Lane LocationLeft LaneMiddle LaneLane LocationLeft LaneMiddle LaneLane Location
0–1004.7435.87850–1004.7435.87850–100
100–2004.2815.824100–2004.2815.824100–200
200–3004.6459.2235200–3004.6459.2235200–300
300–4002.68055.859300–4002.68055.859300–400
400–5004.94255.362400–5004.94255.362400–500
500–6003.5274.203500–6003.5274.203500–600
600–7003.3335.4305600–7003.3335.4305600–700
700–8003.3526.068700–8003.3526.068700–800
800–9004.1915.0265800–9004.1915.0265800–900
900–10004.687.3665900–10004.687.3665900–1000
Table 6. Images before and after clustering.
Table 6. Images before and after clustering.
NumberAbsolute Elevation ChartDBSCAN Cluster DiagramDisease Analysis Chart
Case 1Materials 18 02133 i001Materials 18 02133 i002Materials 18 02133 i003
Case 2Materials 18 02133 i004Materials 18 02133 i005Materials 18 02133 i006
Case 3Materials 18 02133 i007Materials 18 02133 i008Materials 18 02133 i009
Case 4Materials 18 02133 i010Materials 18 02133 i011Materials 18 02133 i012
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, X.; Cheng, S.; Wang, P.; Zheng, H.; Yang, X.; Guo, Y. 3D Reconstruction and Large-Scale Detection of Roads Based on UAV Imagery. Materials 2025, 18, 2133. https://doi.org/10.3390/ma18092133

AMA Style

Zhang X, Cheng S, Wang P, Zheng H, Yang X, Guo Y. 3D Reconstruction and Large-Scale Detection of Roads Based on UAV Imagery. Materials. 2025; 18(9):2133. https://doi.org/10.3390/ma18092133

Chicago/Turabian Style

Zhang, Xiang, Shuwei Cheng, Pu’an Wang, Hao Zheng, Xu Yang, and Yaolin Guo. 2025. "3D Reconstruction and Large-Scale Detection of Roads Based on UAV Imagery" Materials 18, no. 9: 2133. https://doi.org/10.3390/ma18092133

APA Style

Zhang, X., Cheng, S., Wang, P., Zheng, H., Yang, X., & Guo, Y. (2025). 3D Reconstruction and Large-Scale Detection of Roads Based on UAV Imagery. Materials, 18(9), 2133. https://doi.org/10.3390/ma18092133

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop