Next Article in Journal
Monitoring of Plastic Islands in River Environment Using Sentinel-1 SAR Data
Previous Article in Journal
SVASeg: Sparse Voxel-Based Attention for 3D LiDAR Point Cloud Semantic Segmentation
Previous Article in Special Issue
Extraction of Submarine Gas Plume Based on Multibeam Water Column Point Cloud Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MV-GPRNet: Multi-View Subsurface Defect Detection Network for Airport Runway Inspection Based on GPR

1
Tianjin Key Lab for Advanced Signal Processing, Civil Aviation University of China, Tianjin 300300, China
2
Key Laboratory of Smart Airport Theory and System, Civil Aviation University of China, Tianjin 300300, China
3
Shanghai Guimu Robot Co. Ltd., Shanghai 200092, China
4
CSE Department, Texas A&M University, College Station, TX 77843, USA
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(18), 4472; https://doi.org/10.3390/rs14184472
Submission received: 2 August 2022 / Revised: 3 September 2022 / Accepted: 6 September 2022 / Published: 7 September 2022
(This article belongs to the Special Issue Radar and Sonar Imaging and Processing Ⅲ)

Abstract

:
The detection and restoration of subsurface defects are essential for ensuring the structural reliability of airport runways. Subsurface inspections can be performed with the aid of a robot equipped with a Ground Penetrating Radar (GPR). However, interpreting GPR data is extremely difficult, as GPR data usually contains severe clutter interference. In addition, many different types of subsurface defects present similar features in B-scan images, making them difficult to distinguish. Consequently, this makes later maintenance work harder as different subsurface defects require different restoration measures. Thus, to automate the inspection process and improve defect identification accuracy, a novel deep learning algorithm, MV-GPRNet, is proposed. Instead of traditionally using GPR B-scan images only, MV-GPRNet utilizes multi-view GPR data to robustly detect regions with defects despite significant interference. It originally fuses the 3D feature map in C-scan data and the 2D feature map in Top-scan data for defect classification and localization. With our runway inspection robot, a large number of real runway data sets from three international airports have been used to extensively test our method. Experimental results indicate that the proposed MV-GPRNet outperforms state-of-the-art (SOTA) approaches. In particular, MV-GPRNet achieves F1 measurements for voids, cracks, subsidences, and pipes at 91%, 69%, 90%, and 100%, respectively.

1. Introduction

As an area for aircraft taking off and landing, the airport runway is the most fundamental infrastructure to guarantee the safety of aircraft operation [1]. Due to the environmental factors and repetitive landing, the runway structure inevitably deteriorates over time which results in many various subsurface defects, such as voids, subsidences, and cracks. With the continuous development of subsurface defects, it changes the stress state of the airport runway pavement, which seriously threatens the safety of aircraft operations. To ensure aircraft safety and reduce maintenance cost [2], airport runway defect inspection at their early stage is necessary.
GPR is one of the important tools in nondestructive testing (NDT) [3]. It has been widely used in runway inspection tasks to minimize the risk of disrupting aircraft normal operations [4]. While the GPR moves along a linear trajectory, it produces two-dimensional data referred to as B-scan. A Top-scan image is formed by imaging the data points at the same horizontal depth in multiple B-scans. Due to the interference of hardware modules, the inhomogeneity of underground media and the interaction of echoes, GPR data usually contain a high level of clutter interference. In addition, various subsurface defects produce similar patterns in B-scan images, such as voids, cracks, and subsidences. In the subsequent maintenance work, different defects acquire different restoration measures. Thus, to provide more reliable forecasts, fine-grained detection [5] of subsurface defects is required. The interpretation of GPR data mostly still relies on experienced human experts. Manual interpreting, however, is subjective, time-consuming, and cost-prohibitive, so it is not appropriate for large amounts of GPR data. Therefore, it is necessary to develop automatic subsurface object detection methods for GPR data analysis. In recent years, several studies have been conducted on the automatic detection of subsurface defects with GPR via traditional machine learning techniques [6,7], and deep learning methods [8,9]. However, all of the proposed algorithms for detection are mostly based on B-scan images only and primarily designed to detect subsurface regular targets with hyperbolic properties [10,11]. There are few studies on complicated fine-grained defects detection which is hard to realize based on B-scan data only.
Our novel airport runway inspection robot is shown in Figure 1, which performs automatic data collection along a pre-defined route. With the data obtained, we propose a hybrid convolutional neural network (CNN) architecture, named MV-GPRNet, which combines multi-view GPR data to predict the location and classification of the subsurface defects automatically. The MV-GPRNet consists of three main modules: The B-scan-based 3D CNN module, the Top-scan-based 2D CNN module, and the multi-view fusion module. By applying the 3D CNN module to C-scan data composed of multiple adjacent B-scan data, the characteristic relationship between radar channels can be obtained. To better supplement the texture features of subsurface defects, we extract the ROI-focused (Region of Interest) features from Top-scan data. The different representations of GPR data cannot be easily merged by adding or concatenating their feature maps. In the multi-view fusion module, a fusion strategy is designed to enable the interaction of 3D features and 2D region-wise features. We have evaluated our approach on a GPR data set that is collected from three international airports. Comparative results show that our proposed MV-GPRNet significantly outperforms four recent object detection methods. For voids, cracks, subsidences, and pipes, our method achieves F1-measures of 91%, 69%, 90%, and 100%, respectively. Furthermore, we have applied the proposed algorithm to raw GPR data collected from the entire airport runway. The experimental results demonstrated that our algorithm can satisfy the requirements of field applications.
The main contributions of this research are as follows.
(1) The MV-GPRNet is proposed to detect the subsurface defect which is hard to distinguish in GPR B-scan data. To the best of our knowledge, this is the first hybrid deep learning algorithm that employs multiple view GPR data via combing 3D CNNs and 2D CNN for airport runway inspection.
(2) The designed MV-GPRNet was successfully applied to detect various fine-grained defects pre-buried in the artificial airport runway. The experimental results validated that the proposed method could provide reliable results on real data.
(3) The MV-GPRNet has been deployed on the robot platform and applied in practice. The proposed method was tested on GPR data set collected from three airports in comparison to four existing methods. These comparative results demonstrate the superior performance of our proposed method.

2. Related Work

Our work relates to subsurface transportation infrastructure inspection and GPR data analysis. A Falling Weight Deflectometer (FWD) is considered to be a commonly-used solution to monitor the health situation of the airport runway. However, it can only measure pre-defined sampling positions and cannot quantitatively detect minor defects. Other NDT technologies, such as impact-echo (IE) [12], electrical resistivity (ER) [13], and GPR, are often used in the applications of subsurface transportation infrastructure inspection. Among these technologies, GPR is currently the best choice for airport runway inspection since it can achieve fast and full coverage detection with high resolution and good coverage in range and depth.
Researchers have invested much effort in GPR-based subsurface inspection. A GPR can not provide 3D shape information, but a convoluted reflection image with cluttered signals, making it difficult to recognize subsurface defects automatically [14]. It is possible to detect subsurface objects using spatially informative GPR data, but interpreting GPR data automatically remains a challenge.
Standard signal processing methods have been used in GPR data interpretation for subsurface object detection [15]. The template-matching algorithm has been used in [16] to locate and detect pipe signatures. Zhang et al. [17] use the short-time Fourier transform to process the GPR signals of cavern filling with different properties. Lu et al. [18] propose a multi-frequency and multi-attribute GPR data fusion algorithm based on wavelet transform. Szymczyk [19] develops 3D S-transform to detect sinkholes in geological structures. However, these approaches are susceptible to clutter interference, resulting in unreliable results, especially for field GPR data.
Deep learning methods have been widely applied in GPR data analysis in recent years. To address the difficulties of mapping the GPR B-scan data to intricate permittivity maps of subsurface structures, Liu et al. [8] propose a DNN architecture called GPRInvNet. Xu et al. [20] propose an adaptive 1D convolution neural network algorithm for concrete pavement detection. Hou et al. [21] develop an end-to-end framework to simultaneously detect and segment object signatures in GPR scans. Kang et al. [22] propose an underground cavity detection network (UcNet) based on CNN incorporated with phase analysis of super-resolution (SR) GPR images. Feng [23] proposes MigrationNet, a learning-based method for locating and visualizing subsurface objects. Ling et al. [24] propose a 3D object detection framework for urban transportation scenes based on the fusion of Lidar remote sensing technology and optical image sensing technology. To extract the hyperbolic features from GPR B-scan images, Faster R-CNN [25,26], a classic 2D object detection algorithm, is employed. It is important to note that most existing methods employ only 2D B-scan images without considering the correlated information from multiple B-scans for 3D defect detection. Despite the fast decay of the GPR’s electromagnetic waves, the majority of the usable signals produced by a defect will be recorded in multiple adjacent GPR traces [8]. Making full use of the data collected from adjacent GPR channels as opposed to each channel individually would therefore be more practical.
Research on 3D object detection has obtained great progress in the field of autonomous driving, aiming to obtain geometric data in 3D space, such as target location, size, and posture. Wu [27] proposes using a powerful deep model called MotionNet to conduct motion prediction and perception from 3D point clouds. A 3D object detector called KDA3D [28] achieves robust detection by utilizing key-point densification and multi-attention guidance. To turn a 3D point cloud into a 2D image, PIXOR [29] employs different point cloud projection techniques. MV3D [30] and AVOD [31] methods combine laser point cloud and visual information to achieve multi-modal three-dimensional target detection. Li [32] proposes a novel one-stage and keypoints-based framework for monocular 3D object detection using only RGB images. However, it is not an optimal choice to directly utilize the existing 3D CNN methods designed for autonomous driving. First, the characteristics of 3D GPR data are quite different from the 3D point clouds which are generally used as input in these methods. Second, most subsurface defects are irregular, so it is difficult to use the position dimension regression method to reestablish the geometric dimensions of the object in 3D space. Thus, the 3D object detection method specially designed for GPR data interpretation is needed.
The use of Top-scan images is also beneficial for subsurface object detection, as it overcomes the limitations associated with the sole use of B-scan images. For example, as shown in Figure 2, subsidence, crack and void exhibit similar reflection patterns in B-scan images. Thus, it is challenging to identify them using simply B-scan images. However, they can be distinguished by comparing their Top-scan images. Recently, some studies have been conducted using different views of GPR data simultaneously to detect various types of subsurface objects [33]. Kim et al. [34] propose a deep-learning-based underground object classification technique that uses triplanar GPR images. Kang et al. [35] present CNN-based autonomous underground cavity detection using 3D GPR data. However, the performances of these methods was focused mainly on detecting objects that show hyperbolic patterns in B-scan images, such as pipes, cavities, and rebars. Although these proposed techniques have been validated with field GPR data, they are all implemented based on the 2D network AlexNet and can only achieve object classification but not localization. Hence, there is still no published research using GPR B-scan and Top-scan images simultaneously for subsurface object classification and localization via combing 3D CNNs and 2D CNNs.
For several years, our group has been using a robotic platform to inspect both surface and subsurface infrastructure. [36,37,38]. For GPR-based subsurface inspection, we have made progress in subsurface pipeline mapping [37] and B-scan-based subsurface defect detection [38]. Specifically, the proposed GPR-RCNN algorithm [38] aims to detect subsurface defects from GPR data containing a lot of clutter, but it is not effective for fine-grained defects identification. By contrast, the MV-GPRNet enables to detect the defects with indistinguishable features via combining different views, providing more reliable forecasts for subsequent runway maintenance.

3. Problem Formulation

We perform the subsurface data collection with an airport runway inspection robot equipped with a multi-channel GPR, as shown in Figure 3, the GPR antenna array is fixed perpendicular to the robot’s travel direction, and the antennae are equally spaced along the array. The transmitter radiates the electromagnetic wave into the ground while performing inspection tasks. In turn, the receiver gathers the signal reflected off underground objects or ground layer interfaces and generates a one-dimensional waveform called GPR A-scan [39]. With the robot moving, each GPR antenna generates a set of A-scans at regular spatial intervals. This forms a two-dimensional data known as the B-scan which represents a vertical slice of the ground. When the GPR is moved over a regular grid on the ground to gather several parallel B-scans, a three-dimensional data set known as a C-scan can be recorded. The recording across all antennae at a certain depth forms another 2D data representation, named Top-scan, which represents a horizontal slice of a plane at a specific depth. Coloring amplitude of the received signal in the grayscale, a set of 2D B-scan images and Top-scan images can be produced and visualized. Here, only one vertical B-scan image and two horizontal Top-scan images are selectively shown among them as shown in Figure 4.
A collection of B-scan images combined with Top-scan images from multiple depths serve as inputs. To describe them, we define the following notations.
  • { W } , the right handed 3D world coordinate system, with y-axis pointing to the robot’s forward direction in motion, and z-axis pointing to the downward (see Figure 3).
  • C = { B i i = 1 , , n b } , a GPR C-scan consisting of all B-scan images, with B i denoting the i-th B-scan image.
  • T = { T j j = 1 , , n t } , the set of all Top-scan images, with T j representing the j-th Top-scan image.
  • Ω = { b m m = 1 , , n d } , the set of bounding box of the subsurface defect detected from C, with b m = [ x , y , z , l , w , h ] T . [ x , y , z ] T denoting the left top corner coordinate, l, w and h being the length, width and height, respectively.
Our subsurface defect detection problem is defined as follows.
Definition 1.
Given GPR C-scan C and Top-scan set T, obtain Ω.

4. Methodology

To process GPR data, we propose a deep CNN-based algorithm referred to as the MV-GPRNet. The overall framework is presented in Figure 5. This network consists of three modules: (a) B-scan based 3D CNN module: a 3D CNN is designed to extract the 3D feature map from C-scan data. (b) Top-scan based 2D CNN module: It generates ROI-focused (Region of Interest) features via projecting 2D proposals to the 2D feature map extracted from Top-scan data. (c) Multi-view fusion network: a fusion network is used to combine multi-view features via region alignment, and to jointly predict object classification and 3D bounding box regression.

4.1. B-Scan Based 3D CNN Module

We design a 3D network to generate a 3D feature map. Considering 3D defects usually occur across multiple scans, C-scan data which are composed of adjacent multiple B-scans are input to this module. Given a C-scan C, we first subdivide C into equally spaced voxels v with spatial resolution of v x × v y × v z , where v x , v y , and v z indicate the length, height and width of one voxel, respectively. Generally, we set v x to be the total number of GPR array channels, since it is preferred to avoid the data alignment error across multiple scans by keeping the data inside one voxel only from one scanning. This module utilizes a series of 3D convolutions to gradually transform the divided C-scan voxel into feature volumes of 1, 2, 4, and 8 downsampled sizes using a sequence of 3D convolutions. A set of voxel-wise feature vectors could be represented by such feature volumes.
To strengthen the use of the voxel-wise feature, we introduce a three-layer dense network in the 3D feature extraction process, each of which implements a non-linear transformation H l . DenseNet [40] increases the number of channels while maintaining the value of the feature maps [41]. The input for the l t h dense layer is the voxel-wise feature maps of all x 0 , , x l 1 previous dense layers:
x l = H l ( [ x 0 , x 1 , , x l 1 ] )
where [ x 0 , x 1 , , x l 1 ] refers to the concatenation of the voxel-wise feature maps generated in layers 0 , , l 1 .
Then, we design a radar-channel attention module to enhance the feature analysis of effective domains from certain radar channels. Inspired by the channel attention module in CBAM [42], both average-pooled and m max-pooled features are used simultaneously. The radar-channel attention module takes the last dense layer feature map x l as input. Feature tensor X avg is acquired by the average pooling operation, and X max is obtained with the max pooling operation. Based on an element-wise addition strategy, the radar-channel attention map Atte rc , whose size is set to C × 1 × 1 , is generated after a dynamic activation function, which constrains the weights in the range of 0 , 1 . Both descriptors are then forwarded to a shared network. The shared network consists of a fully connected layer F c and one hidden layer. For reducing parameter overhead, the hidden activation size is set to C / r × 1 × 1 , where r represents the reduction ratio. This module generates radar-channel attention feature cube by pixel-wise product between x l and Atte rc , where function is defined as A rc X l . The process can be formulated as follows:
A rc X l = σ F c ( A v g P o o l ( x l ) ) + F c ( M a x P o o l ( x l ) ) x l = σ F c ( X avg ) + F c ( X max ) x l = Atte rc x l
where σ · is a sigmoid activate function, ⊗ denotes element-wise multiplication.
Given C-scan as input, two sub-modules, dense layer and radar-channel attention, focus on ‘where’ and ‘what’, respectively. Due to this, it is possible to arrange two sub-modules in either a parallel or sequential manner. It is found that both the submodules play a significant role in the extraction of 3D feature maps, with a sequential arrangement achieving better results than a parallel arrangement. We will discuss experimental results in discussion section.
With the 3D feature map from C-scan obtained, we designed a 2D network to capture the feature in Top-scan which serves as the supplement for the 3D feature.

4.2. Top-Scan Based 2D CNN Module

The 3D CNNs utilized in the previous module focused on extracting correlation features between adjacent B-scans. To supplement some more effective information for distinguishing subsurface defects, the 2D CNNs utilized in this module are used to extract the spatial features in Top-scan images. We notice that some defects have strong signatures in the Top-scan but weak features in the B-scan. Figure 6 shows an example defect with a strong feature in the Top-scan image but a weak feature in the B-scan image. Thus, we process the multi-view data independently and then fuse their feature maps. To utilize the strong signature accurately, we first obtain the proposals via extracting the features from Top-scan. Then, we propose a strategy to obtain the ROI-focused feature, which is more useful for the fusion module than ROI-pooling operation.
We first design a network to generate 2D object proposals, inspired by the Region Proposal Network (RPN) of the SOTA 2D object detector. Since the continuity of the adjacent Top-scan is not as obvious as the adjacent B-scan, we design a 2D CNN to extract the panel-wise feature map from Top-scan. As same as the 3D feature map extraction network, we add the dense layer in the process of the 2D feature extraction process. Given a top view image, the network generates 2D box proposals from a set of 2D prior boxes. Each 2D box is parameterized by [ x , y , w , h ] T , which [ x , y ] T and [ w , h ] T denote the center and size of the 2D box, respectively. We adopt smooth L 1 loss for 2D proposal regression, as
L t r = 1 2 t i t ^ i 2 , if t i t ^ i   <   1 ( t i t ^ i 1 2 ) , otherwise
where t i = [ x , y , w , h ] T is a vector representing a labeled 2D bounding box, t ^ i indicates the predicted bounding box.
With the 2D proposal boxes obtained, we project these boxes to the corresponding position of the 2D feature, and set the feature value of the non-suggested region position in the feature map to zero. Thus, the ROI-focused feature map is generated in which only the value of the corresponding proposal area is reserved as shown in Figure 7. Therefore, the ROI-focused feature map is easy to be integrated into C-scan feature map according to the depth of the Top-scan without calculating the spatial position corresponding to the C-scan according to the ROI position.
With the 2D and 3D feature maps obtained, the two kinds of features are sent to the following fusion network as input.

4.3. Multi-View Fusion Module

We designed a region-based fusion strategy to effectively combine features from multiple views. We use element-wise add operation for fusing the ROI-focused feature maps and the 3D feature maps obtained in the previous modules. Specifically, the fusion method is to add these ROI-focused feature maps at different depths according to the actual depth and the corresponding position in the 3D voxel-wise feature as shown in Figure 8. The fused feature map can focus on the diseased areas which are not distinguished in the C-scan but have strong features in the Top-scan. In addition, we observe that subsurface defect is usually reflected in the top view but there is no fixed shape. Thus, fusing the ROI-focused feature from Top-scan and the voxel-wise feature from C-scan will cover up those interference features caused by clutter extracted in C-scan, which is helpful to distinguish the subsurface defects accurately.
As the subsurface defects are naturally separated without overlapping each other, we employ a bottom-up manner to generate 3D proposals. We summarize the proposed 3D proposal generation method in Algorithm 1. The pixels inside 3D boxes are considered foreground pixels. Thus, we send the fused feature to the deconvolution sub-network to segment the foreground pixels in each voxel, and then the 3D bounding box is generated from these foreground pixels as the initial 3D proposal. To fulfill the dense prediction, the 3D deconvolution network is designed to continue up-sampling the fused feature. Ground-truth segmentation masks are naturally generated by 3D ground-truth boxes. Considering the number of foreground pixels is generally much smaller than that of the background pixels for a large-scale airport runway, we use the focal loss [43] L f to handle the class imbalance problem as follows,
L f ( p t ) = α t ( 1 p t ) γ log p t ,
with
p t = 1 p , for background pixel p , otherwise
where α denotes the weighting factor, and γ is the focusing parameter.
Finally, the 3D proposal box is projected to the corresponding position in the fused feature map, and the “ROI pooling” operation is performed to obtain a fixed-size feature map. These fixed-size feature maps are used to regress the 3D bounding box. We combine cross-entropy loss and Huber loss for classification and regression, as
L s r = α L c + β L h ,
with
L c = i n c y i log y ^ i L h = 1 2 b i b ^ i 2 , if b i b ^ i     δ δ ( b i b ^ i 1 2 δ ) , otherwise
where n c indicates the total number of classes, for each point belonging to class i, y i and y ^ i represents the label value and predicted value, b i is a vector representing a labeled 3D bounding box, b ^ i indicates the predicted bounding box, δ is a parameter, and α , β is a balancing weighting factor.
Algorithm 1: 3D Proposal Generation
Remotesensing 14 04472 i001

5. Experiments

We have conducted three sets of experiments, based on simulation data, artificial and real airport runway data, respectively. Based on a representative dataset that our inspection robot collected from three civil aviation airports, we assess the performance of the proposed MV-GPRNet and compare it with the SOTA approaches quantitatively. To validate the performance of MV-GPRNet more intuitively, the simulation experiments were designed to obtain the synthetic data of subsurface defects that are pre-set. The ground truth of defects in real airport runway GPR data is labeled by human experts, whose judgments are inevitably subjective. Considering this, we established an artificial airport runway with various kinds of prior-designed defects pre-buried. To verify the superiority of the proposed method in defect detection, a further comparative study was performed on the artificial model data where manufactured defects can be observed directly. Moreover, the selection of dielectric constant in this paper is based on calibration in the artificial model and back calculation to determine the approximate depth of the disease. Specifically, according to the thickness of each layer of material, the depth of the pre-buried defect, and the travel time for receiving the electromagnetic wave, the propagation speed of the electromagnetic wave can be inversely calculated, thereby obtaining the dielectric constant. Through such multiple calibrations, we take the average value of the dielectric constant as the standard for subsequent target depth calculations.

5.1. Implementation

TensorFlow [44] framework is used for the implementation of our algorithm. The modules of MV-GPRNet are trained together. The entire network is trained for 60 epochs using a batch size of 1. Too high learning rates lead to unstable learning, while too low rates result in long training times. In the beginning, the learning rate is set to 0.01, but it is not fixed. As training times increase, it changes. Initially, the learning rate is relatively high, but as the training progresses, the learning rate continues to decrease until the model converges. For avoiding gradient explosions, we use gradient clipping algorithms. Gradient descent is accelerated with a momentum equal to 0.9 by using momentum optimization. To prevent loss of information and reduce calculations, we reformat each GPR B-scan into an image with a resolution of 448 × 448 pixels, which corresponded to a 7 m × 1.5 m concrete segment. For Equation (5), we set α = 1 , β = 1 , and δ = 1 . In learning box refinement, a ground-truth box is assigned to a box proposal if its IoU exceeds a specific threshold T r . Given that the number of candidate boxes obtained by segmentation is often small, we set the IoU threshold to T r = 0.3 .

5.2. Evalution Metrics

Precision (Prec), Recall (Rec), and F1-measure (F1), which are common metrics of object detection, are used to quantitatively evaluate the performance of different methods. The ratio of overlap between the candidate box and the original label box, known as Intersection over Union (IoU) [45], is the threshold to determine the three metrics. A true positive (TP) box is one with an IoU value exceeding the pre-set threshold T I o U , otherwise a false positive (FP) is considered. We set T I o U = 0.5 in our experiments. According to TP, FP, and false negative (FN), the three metrics can be computed as follows:
Precision = TP TP + FP Recall = TP TP + FN F 1 - measure = 2 × Precision × Recall Precision + Recall .

5.3. Experiments on Real Airport Runway

5.3.1. Airport Data Collection

The data set used for our experiments on real airport GPR data was collected using a robot-mounted GPR system. A Raptor™ GPR with isometric fourteen-channel is mounted on the robot equipped with 900 MHZ antennas. The GPR is configured for distance trigger mode. The coverage width of the multi-channel GPR is 1.2 m, and the channel spacing is 87.5 mm. When performing defect detection, at a guaranteed travel speed of 5 km per hour and within a depth range of 1m, the device can detect defects that are no less than 15 cm × 15 cm in the concrete pavement, base, and subbase. As illustrated in Figure 9, the robot conducts defect inspection by moving along a predefined trajectory on the airport runway to collect GPR data. The robot goes from a starting position and then follows a linear path through each scan. Once the current scan is completed, the robot continues on to the next scan until the whole surveyed region is thoroughly covered. We calibrate sensors using a mirror-assisted method [46]. By combining an onboard GPS-RTK receiver and an IMU, the robot performs self-localize. The robot simultaneously transfers image and GPR data using 4G/5G connections to the nearby data analysis center while scanning. The offline analysis will then be performed on the collected data.
We have collected real GPR data from three Chinese international airports. To obtain the B-scan and Top-scan data, we establish two datasets extracted from the GPR data by different views named AUD-B and AUD-T, respectively. The AUD-B dataset contains 5300 B-scan images, while 4526 Top-scan images are contained in the AUD-T dataset. There are 21,083 m 2 of inspection area at three airports, and 1.53 m of subsurface depth are surveyed. The horizontal sampling rate of the GPR is 390 A-scans per meter, and the vertical sampling rate is 1024 sample/A-scan. There are four types of subsurface defects/objects that have been labeled by two human experts individually, namely voids, subsidences, cracks, and pipes. In the training process, we extract features from different views of the same defect, that is, C-scan data composed of multiple parallel B-scan and top scan data from different depths at the same time for feature fusion. The number of different defect samples in B-scan and Top-scan datasets are listed in Table 1. With a set ratio of 7:3, each configuration of defects was randomly assigned to the training data and the testing data in this study.

5.3.2. Comparison with Object Detection Network

To evaluate the benefits of learning using the 3D data and multi-view data of GPR, we have conducted a comparison with four existing object detection networks, including:
  • Faster R-CNN [47]. Faster R-CNN is a two-stage object detection network utilizing RPN to extract candidate boxes.
  • YOLO v5 [48]. You only look once (YOLO) is a real-time one-stage object detection method.
  • PIXOR [29]. In the field of autonomous driving, PIXOR (ORiented 3D object detection from PIXel-wise neural network predictions) is a SOTA, real-time 3D object detection method.
  • GPR-RCNN [38]. A GPR B-scan-based deep learning method for subsurface defect detection.
A major problem that is common to all existing methods which apply CNNs to subsurface object detection in GPR data is that they rely on the detection of 2D B-scan profiles extracted from 3D GPR volumes. This may be sufficient to detect underground defects with a cylindrical shape. However, they cannot detect irregular targets such as subsurface defects, which have to be observed simultaneously from multiple slices through 3D convolution. To further verify the performance of the MV-GPRNet, an additional comparative experiment is conducted based on our previous work on a 3D convolutional network named as GPR-RCNN algorithm which is developed to detect the subsurface defect detection. These four methods can only use single view data as input, so we have implemented, trained, and tested these models on AUD-B dataset.
As shown in Table 2, for each defect class, MV-GPRNet achieves the highest Precision and F1-measure values. It is clear that our algorithm successfully outperforms its four counterparts. We believe the reason for the performance difference is that both Faster R-CNN and YOLO v5 only capture 2D features from B-scans. PIXOR also just extracts 2D features through a single view, and then estimates the 3D position of the object. Although GPR RCNN combines the 2D and 3D features, they are obtained from one single view. It indicates that adopting hybrid 2D and 3D CNN and extracting features from different views of GPR data is more suitable for airport runway subsurface defect detection problems. It can be noticed that the recall for cracks is relatively low, and we discovered that the missed cracks are all of the particular small shapes. On one hand, the spatial resolution based on the channel spacing is 87.5 mm, which means that cracks of less than 40 mm wide between the channels are difficult to be detected well. On the other hand, in B-scan images, the feature of the minor crack is not obvious and easily interfered with by clutter, which results in the performance of our proposed method being decreased.

5.4. Simulation Experiments

To validate the feasibility of the MV-GPRNet in known specific cases, we conduct a set of simulation experiments. Through gprMax [49], various synthetic GPR data of airport runway subsurface defects have been generated for different pre-designed cases. GprMax simulates a given environment using mathematical formulas and material physical properties. According to the different materials used in construction, the real airport runway structure is divided into the following layers: surface layer, base layer, under layer, and soil layer. The dielectric constants for each layer are determined in accordance with the underground structure of the real airport runway [50]. Different subsurface defects have different dielectric constants due to their different materials. In particular, the dielectric constant of voids and cracks is set to 1 since they contain mostly air. A 3D model is created for generating simulation data of complex underground scenarios. Table 3 lists the main parameters for the simulation. Simulating the GPR antenna, we set a ricker waveform source with an amplitude of 10 and a center frequency of 400 MHz.
As shown in Figure 10a–c, an airport runway structure model has been constructed with different types and sizes of subsurface defects. Specifically, we have designed three voids with different shapes, three pipes of the same length, one crack, and one subsidence, all of which are at different locations. As shown in Figure 10d–f, all buried objects are displayed in B-scan images as hyperbola. This proves that our experimental setup is effective, and the simulation data of different shapes of subsurface diseases are successfully collected. Then, we generate the C-scan data and Top-scan images by combining simultaneous B-scan images. For example, the Top-scan image of a subsidence defect is produced as shown in Figure 11. The simulated GPR data are fed into the trained MV-GPRNet model, and all test cases are successfully detected as expected.

5.5. Field Test on Artificial Airport Runway

The ground truth of defects in airport runways can only be labeled by human experts, who are inevitably biased or prone to error. To avoid the ambiguity of manual labeling, we have established an experimental model imitating the real airport runway structure with various defects embedded. The experimental model is approximately 8 m × 8 m × 1.5 m (length × width × height). During the construction of the experimental model, the same materials in each layer were used as in the real airport. The model is divided into four zones with approximate dimensions of 2 m × 2 m × 1.5 m for different experimental processes, as shown in Figure 12e. Pipes, cracks, subsidences and voids are deployed in different zones of the concrete model. As shown in Figure 12a, four plastic plates of different sizes and thicknesses were placed flat at a depth of approximately 0.5 m to simulate interlayer voids. A plastic board with dimensions of 200 × 500 × 10 mm and an angle of 45 degrees was used in the experiment to represent the cracks, as shown in Figure 12b. There was a steel pipe with a length of 2 m and a radius of 4.2 cm in the concrete model, as shown in Figure 12c. A small round pit was dug at about 0.7 m, which was filled with structural material of the upper layer to simulate the subsidence as shown in Figure 12d. It was important to maintain a large distance between each of the defects in order to avoid interference from two signals and to mitigate the impacts of any movement during concrete pouring. Experiments were conducted 30 days after the experimental model was formed.
It is more complicated to analyze collected GPR data than synthetic data due to the inhomogeneity of real structures and clutter under real environmental conditions. To prove the superiority of our proposed method in test sites with known and established features, we conducted further comparative experiments based on the artificial model. With Table 1 and Table 2, it can be seen that YOLO v5, GPR RCNN and MV-GPRNet have the highest detection rates of diseases, so we only compare these three algorithms. As shown in Figure 13, each artificially constructed underground defect has been collected by the GPR utilized in this study. These B-scan and Top-scan images containing pre-buried defects are fed into the trained comparative models for testing. As shown in Figure 14, we can easily observe that MV-GPRNet improves the accuracy of subsurface defect detection, especially for cracks and subsidences. It indicates that adopting 3D CNN and extracting features from different views of GPR data is more suitable for the detection of subsurface defects on airport runways. Furthermore, the testing results demonstrate that the transferability and robustness of MV-GPRNet are optimal.

6. Discussion

To prove the function of the submodules, we apply the Grad-CAM [51] to the dense layer sub-module and attention sub-module. Moreover, we visualize the feature map from the two sub-modules for qualitative analysis. Furthermore, ablation experiments are performed to assess the effectiveness of our design choice.

6.1. Qualitative Analysis

We visualize the trained models using the Grad-CAM and observe that the dense layer and attention sub-modules improve the network’s focus on target objects properly. A recent visualization method called Grad-CAM uses gradients to calculate the importance of spatial locations in convolutional layers. To understand the importance of each neuron for a particular decision, Grad-CAM uses gradient information flowing into the last dense layer of the CNN. Grad-CAM results clearly demonstrate attended regions since gradients are calculated according to a unique class. We examine how this network makes use of features by observing which regions it considers important for class prediction. We apply the Grad-CAM to the last dense layer and the last layer of the attention sub-module, respectively. In Figure 15, we can clearly see that the dense layer sub-module and the attention sub-module learn well to utilize information from a target object region and aggregate features from them.

6.2. Ablation Study

In this subsection, we conduct some extensive ablation experiments to validate the effectiveness of our design choice in the B-scan-based 3D CNN module and the necessity of merging the 2D features from Top-scan data and the 3D features from B-scan data. We compare several approaches to organizing and reserving the dense and attention sub-modules. As each module has different functions, the order may have an impact on total performance. For proving that adding the depth dimension feature performs better than only extracting features from B-scan using 3D CNN, we focus on two variants of our approach, the purely side view-based variant which only uses B-scan images as input, and the multi-view variant which combines B-scan and Top-scan images. As shown in Table 4, without anyone of the dense layer module or radar-channel attention module, the performance of the C-scan feature map which determines the quality of the 3D proposal decreases significantly. In addition, Table 5 shows the results of an ablation experiment on different sub-module arranging methods. We observe that generating an attention map sequentially infers a finer 3D feature map than doing so parallelly. Note that all the arranging methods outperform using only the attention or dense sub-module independently, indicating that using both sub-modules is crucial, while the best arranging approach further improves performance.

7. Conclusions and Future Work

We proposed an MV-GPRNet model for automatic subsurface defect detection for the robotic inspection of airport runways. Our MV-GPRNet originally fused the multi-view feature maps extracted from 3D C-scan data and 2D Top-scan data, which was able to take great advantage of the GPR information. Therefore, it was determined that the MV-GPRNet had the potential to improve fine-grained defect identification accuracy. The MV-GPRNet was first validated by using synthetic GPR data and artificial runway data. Then, the NV-GPRNet was successfully applied to detect the pre-buried irregular defects with similar features. Furthermore, a large number of real runway data sets from three international airports have been used to extensively test our method. The comparative results demonstrated that the MV-GPRNet can effectively detect the airport runway subsurface defects and had outperformed the SOTA techniques. Specifically, our method achieved F1-measures at 91%, 69%, 90%, and 100% for a void, crack, subsidence and pipe, respectively.
However, for minor defects, due to their insignificance feature in the real GPR data, they are easily neglected. In the future, we will consider adding the prior knowledge in the CNN model to further improve the performance of defect detection.

Author Contributions

Methology, N.L. and H.L.; Project administration, H.L.; Supervision & Validation, R.W. and H.W.; Investigation & Data collection, Z.G.; Writing—original draft preparation, N.L.; Writing—review and editing, H.L. and D.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by National Key Research and Development Project of China under Grant 2019YFB1310400.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Acknowledgments

We thank the Shanghai Guimu Robot Co., Ltd., Shanghai, China. for their support.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, Y.; Dai, Y.; Liu, Y. Design and Implementation of Airport Runway Robot Based on Artificial Intelligence. In Proceedings of the 2021 IEEE 5th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC), Chongqing, China, 12–14 March 2021; pp. 2636–2640. [Google Scholar]
  2. D’Amico, F.; Gagliardi, V.; Ciampoli, L.B.; Tosti, F. Integration of InSAR and GPR techniques for monitoring transition areas in railway bridges. NDT E Int. 2020, 115, 102291. [Google Scholar] [CrossRef]
  3. Iftimie, N.; Savin, A.; Steigmann, R.; Dobrescu, G.S. Underground Pipeline Identification into a Non-Destructive Case Study Based on Ground-Penetrating Radar Imaging. Remote Sens. 2021, 13, 3494. [Google Scholar] [CrossRef]
  4. Vyas, V.; Singh, A.P.; Srivastava, A. A decision making framework for condition evaluation of airfield pavements using non-destructive testing. In Airfield and Highway Pavements 2019: Innovation and Sustainability in Highway and Airfield Pavement Technology; American Society of Civil Engineers: Reston, VA, USA, 2019; pp. 343–353. [Google Scholar]
  5. Zhang, L.; Fan, Y.; Yan, R.; Shao, Y.; Wang, G.; Wu, J. Fine-Grained Tidal Flat Waterbody Extraction Method (FYOLOv3) for High-Resolution Remote Sensing Images. Remote Sens. 2021, 13, 2594. [Google Scholar] [CrossRef]
  6. Kaur, P.; Dana, K.J.; Romero, F.A.; Gucunski, N. Automated GPR rebar analysis for robotic bridge deck evaluation. IEEE Trans. Cybern. 2015, 46, 2265–2276. [Google Scholar] [CrossRef]
  7. Peng, M.; Wang, D.; Liu, L.; Shi, Z.; Shen, J.; Ma, F. Recent Advances in the GPR Detection of Grouting Defects behind Shield Tunnel Segments. Remote Sens. 2021, 13, 4596. [Google Scholar] [CrossRef]
  8. Liu, B.; Ren, Y.; Liu, H.; Xu, H.; Wang, Z.; Cohn, A.G.; Jiang, P. GPRInvNet: Deep Learning-Based Ground-Penetrating Radar Data Inversion for Tunnel Linings. IEEE Trans. Geosci. Remote Sens. 2021, 59, 8305–8325. [Google Scholar] [CrossRef]
  9. Qiu, Z.; Zhao, Z.; Chen, S.; Zeng, J.; Huang, Y.; Xiang, B. Application of an Improved YOLOv5 Algorithm in Real-Time Detection of Foreign Objects by Ground Penetrating Radar. Remote Sens. 2022, 14, 1895. [Google Scholar] [CrossRef]
  10. Lei, W.; Luo, J.; Hou, F.; Xu, L.; Wang, R.; Jiang, X. Underground cylindrical objects detection and diameter identification in GPR B-scans via the CNN-LSTM framework. Electronics 2020, 9, 1804. [Google Scholar] [CrossRef]
  11. Jaufer, R.M.; Ihamouten, A.; Goyat, Y.; Todkar, S.S.; Guilbert, D.; Assaf, A.; Dérobert, X. A Preliminary Numerical Study to Compare the Physical Method and Machine Learning Methods Applied to GPR Data for Underground Utility Network Characterization. Remote Sens. 2022, 14, 1047. [Google Scholar] [CrossRef]
  12. Pospisil, K.; Manychova, M.; Stryk, J.; Korenska, M.; Matula, R.; Svoboda, V. Diagnostics of Reinforcement Conditions in Concrete Structures by GPR, Impact-Echo Method and Metal Magnetic Memory Method. Remote Sens. 2021, 13, 952. [Google Scholar] [CrossRef]
  13. Bai, D.; Lu, G.; Zhu, Z.; Zhu, X.; Tao, C.; Fang, J. Using Electrical Resistivity Tomography to Monitor the Evolution of Landslides’ Safety Factors under Rainfall: A Feasibility Study Based on Numerical Simulation. Remote Sens. 2022, 14, 3592. [Google Scholar] [CrossRef]
  14. Kang, M.S.; An, Y.K. Frequency–wavenumber analysis of deep learning-based super resolution 3D GPR images. Remote Sens. 2020, 12, 3056. [Google Scholar] [CrossRef]
  15. Jin, Y.; Duan, Y. 2d wavelet decomposition and fk migration for identifying fractured rock areas using ground penetrating radar. Remote Sens. 2021, 13, 2280. [Google Scholar] [CrossRef]
  16. Sagnard, F.; Tarel, J.P. Template-matching based detection of hyperbolas in ground-penetrating radargrams for buried utilities. J. Geophys. Eng. 2016, 13, 491–504. [Google Scholar] [CrossRef]
  17. Zhang, S.; He, W.; Cao, F.; Hong, L. Time-Frequency Analysis of GPR Simulation Signals for Tunnel Cavern Fillings Based on Short-Time Fourier Transform. In Earth and Space 2021; ASCE: Washington, DC, USA, 2021; pp. 572–581. [Google Scholar]
  18. Lu, G.; Zhao, W.; Forte, E.; Tian, G.; Li, Y.; Pipan, M. Multi-frequency and multi-attribute GPR data fusion based on 2-D wavelet transform. Measurement 2020, 166, 108243. [Google Scholar] [CrossRef]
  19. Szymczyk, P.; Szymczyk, M. Non-destructive building investigation through analysis of GPR signal by S-transform. Autom. Constr. 2015, 55, 35–46. [Google Scholar] [CrossRef]
  20. Xu, J.; Zhang, J.; Sun, W. Recognition of the typical distress in concrete pavement based on GPR and 1D-CNN. Remote Sens. 2021, 13, 2375. [Google Scholar] [CrossRef]
  21. Hou, F.; Lei, W.; Li, S.; Xi, J.; Xu, M.; Luo, J. Improved Mask R-CNN with distance guided intersection over union for GPR signature detection and segmentation. Autom. Constr. 2021, 121, 103414. [Google Scholar] [CrossRef]
  22. Kang, M.S.; Kim, N.; Im, S.B.; Lee, J.J.; An, Y.K. 3D GPR image-based UcNet for enhancing underground cavity detectability. Remote Sens. 2019, 11, 2545. [Google Scholar] [CrossRef]
  23. Feng, J.; Yang, L.; Wang, H.; Tian, Y.; Xiao, J. Subsurface Pipes Detection Using DNN-based Back Projection on GPR Data. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2021; pp. 266–275. [Google Scholar]
  24. Bai, L.; Li, Y.; Cen, M.; Hu, F. 3D instance segmentation and object detection framework based on the fusion of LIDAR remote sensing and optical image sensing. Remote Sens. 2021, 13, 3288. [Google Scholar] [CrossRef]
  25. Pham, M.T.; Lefèvre, S. Buried object detection from B-scan ground penetrating radar data using Faster-RCNN. In Proceedings of the IGARSS 2018-2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 6804–6807. [Google Scholar]
  26. Lei, W.; Hou, F.; Xi, J.; Tan, Q.; Xu, M.; Jiang, X.; Liu, G.; Gu, Q. Automatic hyperbola detection and fitting in GPR B-scan image. Autom. Constr. 2019, 106, 102839. [Google Scholar] [CrossRef]
  27. Wu, P.; Chen, S.; Metaxas, D.N. MotionNet: Joint Perception and Motion Prediction for Autonomous Driving Based on Bird’s Eye View Maps. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11385–11395. [Google Scholar]
  28. Wang, J.; Zhu, M.; Wang, B.; Sun, D.; Wei, H.; Liu, C.; Nie, H. Kda3d: Key-point densification and multi-attention guidance for 3d object detection. Remote Sens. 2020, 12, 1895. [Google Scholar] [CrossRef]
  29. Yang, B.; Luo, W.; Urtasun, R. Pixor: Real-time 3d object detection from point clouds. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7652–7660. [Google Scholar]
  30. Chen, X.; Ma, H.; Wan, J.; Li, B.; Xia, T. Multi-view 3d object detection network for autonomous driving. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1907–1915. [Google Scholar]
  31. Ku, J.; Mozifian, M.; Lee, J.; Harakeh, A.; Waslander, S.L. Joint 3d proposal generation and object detection from view aggregation. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 1–8. [Google Scholar]
  32. Li, P.; Huaici, Z. Monocular 3D Detection with Geometric Constraint Embedding and Semi-supervised Training. IEEE Robot. Autom. Lett. 2021, 6, 5565–5572. [Google Scholar] [CrossRef]
  33. Ling, J.; Qian, R.; Shang, K.; Guo, L.; Zhao, Y.; Liu, D. Research on the Dynamic Monitoring Technology of Road Subgrades with Time-Lapse Full-Coverage 3D Ground Penetrating Radar (GPR). Remote Sens. 2022, 14, 1593. [Google Scholar] [CrossRef]
  34. Kim, N.; Kim, S.; An, Y.K.; Lee, J.J. Triplanar imaging of 3-D GPR data for deep-learning-based underground object detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 4446–4456. [Google Scholar] [CrossRef]
  35. Kang, M.S.; Kim, N.; An, Y.K.; Lee, J.J. Deep learning-based autonomous underground cavity detection using 3D GPR. Struct. Health Monit. 2020, 19, 173–185. [Google Scholar] [CrossRef]
  36. Li, H.; Song, D.; Liu, Y.; Li, B. Automatic Pavement Crack Detection by Multi-Scale Image Fusion. IEEE Trans. Intell. Transp. Syst. 2019, 20, 2025–2036. [Google Scholar] [CrossRef]
  37. Li, H.; Chou, C.; Fan, L.; Li, B.; Wang, D.; Song, D. Toward automatic subsurface pipeline mapping by fusing a ground-penetrating radar and a camera. IEEE Trans. Autom. Sci. Eng. 2019, 17, 722–734. [Google Scholar] [CrossRef]
  38. Li, H.; Li, N.; Wu, R.; Wang, H.; Gui, Z.; Song, D. GPR-RCNN: An Algorithm of Subsurface Defect Detection for Airport Runway based on GPR. IEEE Robot. Autom. Lett. 2021, 6, 3001–3008. [Google Scholar] [CrossRef]
  39. Kim, N.; Kim, S.; An, Y.K.; Lee, J.J. A novel 3D GPR image arrangement for deep learning-based underground object classification. Int. J. Pavement Eng. 2021, 22, 740–751. [Google Scholar] [CrossRef]
  40. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  41. Wang, W.; Dou, S.; Jiang, Z.; Sun, L. A fast dense spectral–spatial convolution network framework for hyperspectral images classification. Remote Sens. 2018, 10, 1068. [Google Scholar] [CrossRef]
  42. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  43. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  44. Martín, A.; Paul, B.; Jianmin, C. Tensorflow: A system for large-scale machine learning. In Proceedings of the 12th USENIX Symposium on Operating Systems Design and Implementation (OSDI 16), Savannah, GA, USA, 2–4 November 2016; pp. 265–283. [Google Scholar]
  45. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized intersection over union: A metric and a loss for bounding box regression. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 658–666. [Google Scholar]
  46. Chou, C.; Li, H.; Song, D. Encoder-Camera-Ground Penetrating Radar Sensor Fusion: Bimodal Calibration and Subsurface Mapping. IEEE Trans. Robot. 2020, 37, 67–81. [Google Scholar] [CrossRef]
  47. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  48. Jocher, G.; Stoken, A. ultralytics/yolov5: V3.0. 2020. Available online: https://github.com/ultralytics/yolov5 (accessed on 13 August 2020).
  49. Giannopoulos, A. Modelling ground penetrating radar by GprMax. Constr. Build. Mater. 2005, 19, 755–762. [Google Scholar] [CrossRef]
  50. Cao, Q.; Al-Qadi, I.L. Effect of Moisture Content on Calculated Dielectric Properties of Asphalt Concrete Pavements from Ground-Penetrating Radar Measurements. Remote Sens. 2021, 14, 34. [Google Scholar] [CrossRef]
  51. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
Figure 1. An illustration of our subsurface defect detection problem: (a) Our inspection robot equipped with a multi-channel GPR, and generates a set of B-scans and Top-scans GPR data; (b) Detected 3D bounding boxes of subsurface defects. The different color of the box indicates different categories.
Figure 1. An illustration of our subsurface defect detection problem: (a) Our inspection robot equipped with a multi-channel GPR, and generates a set of B-scans and Top-scans GPR data; (b) Detected 3D bounding boxes of subsurface defects. The different color of the box indicates different categories.
Remotesensing 14 04472 g001
Figure 2. Typical GPR patterns of void, crack, subsidence, and pipe: B-scan images of a (a) void; (b) crack; (c) subsidence; and (d) pipe. Top-scan images of a (e) void; (f) crack; (g) subsidence; and (h) pipe.
Figure 2. Typical GPR patterns of void, crack, subsidence, and pipe: B-scan images of a (a) void; (b) crack; (c) subsidence; and (d) pipe. Top-scan images of a (e) void; (f) crack; (g) subsidence; and (h) pipe.
Remotesensing 14 04472 g002
Figure 3. GPR antenna and 3D GPR data arrangement.
Figure 3. GPR antenna and 3D GPR data arrangement.
Remotesensing 14 04472 g003
Figure 4. An example of B−scan and Top−scan images.
Figure 4. An example of B−scan and Top−scan images.
Remotesensing 14 04472 g004
Figure 5. Architecture of the proposed MV-GPRNet. The framework consists of three components: (a) B-scan-based 3D CNN module (green block): a 3D CNN is designed to extract the 3D feature map from C-scan data. (b) Top-scan-based 2D CNN module (yellow block): It first generates 2D object proposals from the top view map and projects them to the 2D feature map. (c) Multi-view fusion network (purple block): A fusion network is used to combine multi-view features via region alignment. The fused features are used to generate 3D proposals and jointly predict object class and perform 3D box regression after the ROI-pooling layer.
Figure 5. Architecture of the proposed MV-GPRNet. The framework consists of three components: (a) B-scan-based 3D CNN module (green block): a 3D CNN is designed to extract the 3D feature map from C-scan data. (b) Top-scan-based 2D CNN module (yellow block): It first generates 2D object proposals from the top view map and projects them to the 2D feature map. (c) Multi-view fusion network (purple block): A fusion network is used to combine multi-view features via region alignment. The fused features are used to generate 3D proposals and jointly predict object class and perform 3D box regression after the ROI-pooling layer.
Remotesensing 14 04472 g005
Figure 6. Multi-view scans of a sample defect with a strong feature in the Top-scan and a weak feature in the B-scan: (a) B-scan image; (b) Top-scan image. Horizon lines in (a,b) indicate the depth.
Figure 6. Multi-view scans of a sample defect with a strong feature in the Top-scan and a weak feature in the B-scan: (a) B-scan image; (b) Top-scan image. Horizon lines in (a,b) indicate the depth.
Remotesensing 14 04472 g006
Figure 7. An illustration of ROI-focused feature extraction.
Figure 7. An illustration of ROI-focused feature extraction.
Remotesensing 14 04472 g007
Figure 8. An illustration of the region-based fusion process.
Figure 8. An illustration of the region-based fusion process.
Remotesensing 14 04472 g008
Figure 9. Our runway inspection robotic system and GPR data collection process.
Figure 9. Our runway inspection robotic system and GPR data collection process.
Remotesensing 14 04472 g009
Figure 10. Simulation results of the airport runway subsurface model. (ac) Different view of the simulated airport runway model with subsurface defects. (df) Simulated GPR B-scan images.
Figure 10. Simulation results of the airport runway subsurface model. (ac) Different view of the simulated airport runway model with subsurface defects. (df) Simulated GPR B-scan images.
Remotesensing 14 04472 g010
Figure 11. Generated C-scan data and Top-scan image based on simulated B-scan images.
Figure 11. Generated C-scan data and Top-scan image based on simulated B-scan images.
Remotesensing 14 04472 g011
Figure 12. Schematic of testing model and defects: (a) Four plastic plates of different thicknesses laid flat simulating the voids; (b) A plastic plate was placed at an angle to simulate cracks; (c) A steel pipe located at depth of 0.5 m; (d) A small round pit filled with structural material of the upper layer simulating the subsidence; (e) Testing model.
Figure 12. Schematic of testing model and defects: (a) Four plastic plates of different thicknesses laid flat simulating the voids; (b) A plastic plate was placed at an angle to simulate cracks; (c) A steel pipe located at depth of 0.5 m; (d) A small round pit filled with structural material of the upper layer simulating the subsidence; (e) Testing model.
Remotesensing 14 04472 g012
Figure 13. GPR B-scan and Top-scan images from the field test.
Figure 13. GPR B-scan and Top-scan images from the field test.
Remotesensing 14 04472 g013
Figure 14. Representative comparison results for complex subsurface defects on B-scan image of artificial model.
Figure 14. Representative comparison results for complex subsurface defects on B-scan image of artificial model.
Remotesensing 14 04472 g014
Figure 15. Grad-CAM visualization results: (a) Original B-scan image; (b) is Grad-CAM visualization for dense layer; (c) is Grad-CAM visualization for Radar-channel attention submodule. Red regions correspond to high score for class. Figure best viewed in color. For simplicity, we only add a coordinate axis to the first graph in the upper left corner, and the rest are the same.
Figure 15. Grad-CAM visualization results: (a) Original B-scan image; (b) is Grad-CAM visualization for dense layer; (c) is Grad-CAM visualization for Radar-channel attention submodule. Red regions correspond to high score for class. Figure best viewed in color. For simplicity, we only add a coordinate axis to the first graph in the upper left corner, and the rest are the same.
Remotesensing 14 04472 g015
Table 1. Defect Number In Different Dataset.
Table 1. Defect Number In Different Dataset.
ClassAUD-BAUD-T
Void16,0584547
Crack19553
Subsidence55452
Pipe59201547
Table 2. Defect detection results on AUD dataset.
Table 2. Defect detection results on AUD dataset.
MethodVoidCrackSubsidencePipe
PrecRecF1PrecRecF1PrecRecF1PrecRecF1
Faster R-CNN32%74%44%43%83%57%22%66%44%66%81%73%
YOLO v540%91%55%8%72%14%23%74%35%63%98%77%
PIXOR82%29%43%7%100%13%9%69%16%25%100%40%
GPR-RCNN76%95%85%25%66%36%77%86%81%87%100%93%
MV-GPRNet92%89%91%73%66%69%94%87%90%100%100%100%
Table 3. Parameters of the modeling methods in Figure 10.
Table 3. Parameters of the modeling methods in Figure 10.
Model Model Size (Length, Height, Width)
6 m1.7 m4 m
Spatial Step
0.010.010.01
Model ConstructionLayersShape (Radius; Depth)HeightDielectric Constant
Surface layerbox (6 m × 4 m; 0.4 m)1.5 m6.4
Base layerbox (6 m × 4 m; 0.35 m)1.1 m5.5
Under layerbox (6 m × 4 m; 0.25 m)0.75 m5
Soil layerbox (6 m × 4 m; 0.5 m)0.5 m3
Subsurface Defect
In 3D Model
ClassShape (Radius; Thickness)HeightDielectric Constant
Voidcylinder (0.15 m; 0.5 m)1.26 m1
Voidbox (1 m × 0.3 m; 0.3 m)1.2 m1
Voidtriangle (0.36 m × 0.43 m × 0.25 m; 0.85 m)1.1 m1
Crackbox (0.02 m × 0.55 m; 0.3 m)1.2 m1
Subsidencetriangle (1.07 m × 1 m × 0.4 m; 0.3 m)1 m2
Pipecylinder (0.01 m; 1 m)1.2 m8
Table 4. Ablation experiments on retention of dense layer module and attention module.
Table 4. Ablation experiments on retention of dense layer module and attention module.
Dense Layer
Module
Radar-Channel
Attention Module
VoidCrackSubsidencePipe
PrecRecF1PrecRecF1PrecRecF1PrecRecF1
92%86%89%45%61%52%78%90%84%91%95%93%
93%87%90%49%57%53%84%94%89%97%93%95%
92%85%89%25%37%30%52%64%57%94%90%92%
92%89%91%73%66%69%94%87%90%100%100%100%
Table 5. Ablation experiments on combining methods of dense layer module and attention module.
Table 5. Ablation experiments on combining methods of dense layer module and attention module.
Combining MethodsVoidCrackSubsidencePipe
PrecRecF1PrecRecF1PrecRecF1PrecRecF1
Dense + Attention92%89%91%73%66%69%94%87%90%100%100%100%
Attention + Dense91%88%90%57%47%52%92%81%86%97%93%95%
Dense and Attention in parallel90%88%89%62%51%56%63%91%74%92%93%92%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, N.; Wu, R.; Li, H.; Wang, H.; Gui, Z.; Song, D. MV-GPRNet: Multi-View Subsurface Defect Detection Network for Airport Runway Inspection Based on GPR. Remote Sens. 2022, 14, 4472. https://doi.org/10.3390/rs14184472

AMA Style

Li N, Wu R, Li H, Wang H, Gui Z, Song D. MV-GPRNet: Multi-View Subsurface Defect Detection Network for Airport Runway Inspection Based on GPR. Remote Sensing. 2022; 14(18):4472. https://doi.org/10.3390/rs14184472

Chicago/Turabian Style

Li, Nansha, Renbiao Wu, Haifeng Li, Huaichao Wang, Zhongcheng Gui, and Dezhen Song. 2022. "MV-GPRNet: Multi-View Subsurface Defect Detection Network for Airport Runway Inspection Based on GPR" Remote Sensing 14, no. 18: 4472. https://doi.org/10.3390/rs14184472

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop