Next Article in Journal
A Novel Decomposition Model for Visual Rail Surface Inspection
Previous Article in Journal
A Huffman-Based Joint Compression and Encryption Scheme for Secure Data Storage Using Physical Unclonable Functions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

GPR B-Scan Image Denoising via Multi-Scale Convolutional Autoencoder with Data Augmentation

1
School of Computer Science and Engineering, Central South University, Changsha 410075, China
2
China Research Institute of Radio Wave Propagation, Qingdao 266103, China
*
Author to whom correspondence should be addressed.
Electronics 2021, 10(11), 1269; https://doi.org/10.3390/electronics10111269
Submission received: 13 April 2021 / Revised: 19 May 2021 / Accepted: 21 May 2021 / Published: 26 May 2021
(This article belongs to the Section Microwave and Wireless Communications)

Abstract

:
Ground-penetrating radar (GPR), as a non-invasive instrument, has been widely used in civil engineering. In GPR B-scan images, there may exist random noise due to the influence of the environment and equipment hardware, which complicates the interpretability of the useful information. Many methods have been proposed to eliminate or suppress the random noise. However, the existing methods have an unsatisfactory denoising effect when the image is severely contaminated by random noise. This paper proposes a multi-scale convolutional autoencoder (MCAE) to denoise GPR data. At the same time, to solve the problem of training dataset insufficiency, we designed the data augmentation strategy, Wasserstein generative adversarial network (WGAN), to increase the training dataset of MCAE. Experimental results conducted on both simulated, generated, and field datasets demonstrated that the proposed scheme has promising performance for image denoising. In terms of three indexes: the peak signal-to-noise ratio (PSNR), the time cost, and the structural similarity index (SSIM), the proposed scheme can achieve better performance of random noise suppression compared with the state-of-the-art competing methods (e.g., CAE, BM3D, WNNM).

1. Introduction

Ground-penetrating radar (GPR), as an effective tool of underground non-destructive detection, is widely used to study near-surface geophysical structures and detect buried targets [1,2,3,4,5]. It transmits high-frequency electromagnetic waves and receives reflections to infer the composition of the underground medium and the dielectric properties, spatial location, and structural dimensions of the detected targets [6,7,8]. The GPR antenna scans in a certain direction and transmits electromagnetic waves at different positions. Each position will receive a one-dimensional signal. Many one-dimensional signals are combined into a two-dimensional signal, which is called GPR B-scan image. However, due to the inhomogeneity of the underground medium and the complexity of the embedded targets, the GPR signal is easily damaged by random noise, and it often appears in the form of non-stationary signals and spikes in the GPR data. Therefore, there exists random noise in the GPR B-scan images that complicates the interpretability of the useful information and seriously affects the detection performance of GPR [9]. It is necessary to develop an advanced algorithm to suppress random noise in GPR B-scan images.
In the field of GPR, image denoising methods can be divided into four groups: those based on spatial filtering, those based on transform domain, those based on subspace, and those based on deep learning (DL) [10,11,12]. Lee et al. [13] proposed a denoising filter, Lee filter, which is based on a linear noise model and a minimum mean square error model to obtain enhanced pixel points by calculating the neighborhood of a pixel. Frost [14] and Kuan [15] filters are improved from the Lee filter. The improved filters can not only better suppress noise but also better preserve image texture information. However, the performance of these spatial filters is greatly affected by the size of the filter window. A smaller window cannot effectively suppress noise, while a larger window inevitably leads to the loss of image texture details. Kumlu et al. [16] employed the non-local mean (NLM) denoising algorithm to denoise GPR images. NLM uses sub-block similarity to filter noisy images, and according to the similarity between the current noisy image block and adjacent blocks, calculates the weight. Although the output of the NLM algorithm is ideal for removing low-level noise, the performance deteriorates sharply when the noise increases.
Compared with the methods based on spatial filtering, the methods based on transform domain are more effective for the separation between signal and noise. In References [17,18], wavelet transform and multi-wavelet transform were proposed to remove the random noise of GPR images. Although it has been proven that wavelet-based denoising methods have better efficiency than classical filters, the limitation of applying wavelet transform is that the basis of wavelet transform is usually fixed and cannot fully represent the image. In order to overcome the problem of non-sparseness and lack of direction selectivity of coefficients in high-dimensional wavelet transform, Wang et al. [19] proposed Shearlet transform to remove clutter from GPR images. However, the translation robustness of Shearlet transform is poor, and the edge pseudo-Gibbs distortion phenomenon is obvious. Yahya et al. [20] proposed a block-matching and 3D filtering (BM3D) image denoising algorithm based on adaptive filtering that can achieve optimal noise reduction performance and preserve the high spatial frequency detail. However, BM3D has too large an amount of calculation to achieve real-time processing.
The GPR image denoising methods based on subspace include singular value decomposition (SVD) [21,22], principal component analysis (PCA) [11,23], and independent component analysis (ICA) [24]. These methods use various constraints in the cost function to decompose the matrix. Liu et al. [25] used the SVD method to decompose the signal and then used the PCA method to reconstruct the signal to achieve the PCA–SVD hybrid method for noise reduction. This method effectively removes the random noise of GPR images. In References [26,27], the weighted nuclear norm minimization (WNNM) algorithm was used to denoise the image. The algorithm is improved on the nuclear norm minimization (NNM) algorithm and uses the non-local similarity of images to construct a low-rank matrix to achieve noise cancellation. Compared with NNM, WNNM performs shrinking according to the size of the singular value itself, which can better retain effective information. However, the WNNM algorithm refers to a non-convex minimization problem, and its optimization is difficult.
In recent years, with the rapid development of DL technology, a large number of research methods using DL for image denoising have emerged. The DL-based technology handles the complex relationship between high-quality images and low-quality images by training a DL model so that denoised images can be obtained in a short time. In Reference [28], a convolutional neural network (CNN) with residual learning and batch normalization was applied for synthetic aperture radar (SAR) image denoising. In Reference [29], an autoencoder (AE) method was used to remove noise and proved the feasibility for noise reduction in the spectral domain of hyperspectral images. Choi et al. [30] combined the convolutional structure with the autoencoder and used the convolutional autoencoder (CAE) to remove image noise, which achieved a good denoising effect for random noise. However, the performance of CAE will drop sharply when the image in a high-level noise. In terms of GPR research, image denoising algorithms based on DL are few and immature.
In this paper, we propose a multi-scale convolutional autoencoder (MCAE) algorithm for GPR image denoising. In order to solve the problem of insufficient training dataset, we designed a data augmentation strategy—Wasserstein generative adversarial network (WGAN)—to increase the training dataset of MCAE, and a better denoising effect was achieved. The flow chart is shown in Figure 1. First, the obtained 1800 noise-free GPR images were used to train the discriminator and generator of the WGAN, and then the trained WGAN generated 3200 noise-free GPR images. Second, the 5000 noise-free GPR images had Gaussian white noise added in order to form a noisy GPR dataset. Finally, the noisy dataset and its corresponding noise-free data label were used to train MCAE. After the training was completed, the noisy GPR image was put into the trained MCAE and reconstructed to a denoised GPR image. The generated data, simulated data, and field data examples demonstrated that the proposed method can achieve better performance in random noise suppression when compared with the existing methods.
The contributions of this paper are summarized as follows:
(1)
We developed a scheme for GPR image denoising based on DL techniques.
(2)
We proposed a MCAE algorithm that improved the CAE model by introducing a multi-scale convolution kernel to extract higher-level features of the image.
(3)
We designed a WGAN algorithm to augment the GPR training dataset of MCAE.

2. Data Augmentation via the WGAN Algorithm

The MCAE for GPR image denoising requires a large GPR training dataset. If the training dataset is too small, it will cause problems such as unstable network training, difficulty in decreasing loss function, or over-fitting of the network [31,32]. At present, there are almost no public GPR datasets. In the field measurement process, collecting and storing large, heterogeneous radar data is quite time-consuming and expensive. In the simulation process, gprMax3.0 software [33] uses the finite difference time domain (FDTD) method to simulate GPR data, which is also relatively time-consuming. At present, generative adversarial networks (GANs) are widely used in data augmentation, such as medical [34], construction [35], climate science [36], and earth observation [37,38,39], but they are rarely used in GPR. Compared with the other types of augmentation, such as rotation, translation, and cropping, the GAN-based augmentation provides a powerful tool to generate more diverse artificial data by learning the data distribution of the training set [40,41]. In order to solve the problem of training dataset insufficiency, we designed a WGAN network to generate GPR images for expanding the training dataset of MCAE.
As depicted in Figure 2, WGAN consists of a generator (G) and a discriminator (D). The generator is composed of five deconvolutional layers. The size of kernels is 4 × 4 in all layers. The number of kernels in each layer is 512, 256, 128, 64, and 1, respectively. The stride in each layer is 1, 4, 4, 4, and 1, respectively. The discriminator is composed of three convolutional layers and one fully connected layer. The size of kernels in each convolutional layer is 4 × 4, 4 × 4, and 3 × 3, respectively. The number of kernels in each convolutional layer is 64, 128, and 256, respectively, and the stride is 4 in all layers. In the fully connected (FC) layer, the number of nodes is set as 1.
In the generator, a random variable z is input to the generator and a fake GPR image is generated through a series of deconvolutional layers. In the discriminator, the images generated by the generator and real images are put into the discriminator. Then, the discriminator outputs a value, which indicates the probability that the input image is the real image and guides the generator update. The purpose of the generator is to make the generated sample the same as the real sample, and the purpose of the discriminator is to distinguish whether the input image is the real image or the fake image generated by the generator. After WGAN training is completed, the data generated by the generator is close to the distribution of real GPR data, which can expand the training dataset of MCAE.
The generator and discriminator of WGAN have their own training goals. The generator learns the distribution of real data, and the discriminator estimates the probability that the sample comes from the real data instead of the data generated by the generator. Wasserstein loss [42] is used as the optimization target in order to minimize the difference between the generated data and the real data. At the same time, in order to improve the stability of network training, we introduced a gradient penalty term [43]. Therefore, the training goal of the discriminator can be expressed as
max θ L ( G , D ) = E x f p g [ D θ ( x f ) ] E x r p r [ D θ ( x r ) ] + λ E x ^ p x ^ [ ( x ^ D θ ( x ^ ) 2 1 ) 2 ]
where E x f p g [ D θ ( x f ) ] E x r p r [ D θ ( x r ) ] is Wasserstein loss and E x ^ p x ^ [ ( x ^ D ( x ^ ) 2 1 ) 2 ] is the gradient penalty. D θ ( x r ) represents the output of the real sample x r through the discriminator. θ is the parameter set of the discriminator, and D θ ( x f ) represents the output of the generated sample x f through the discriminator. λ is the penalty coefficient. x ^ comes from the linear difference between x r and x f . The equation is as follows:
x ^ = t x r + ( 1 t ) x f , t [ 0 , 1 ]
The goal of the discriminator D is to maximize the above-mentioned error L ( G , D ) and force the Wasserstein distance between the distribution p g of the generator and the real distribution p r to be as small as possible. The training goal of WGAN’s generator is as follows:
min ϕ L ( G , D ) = E z p z [ D θ ( G ϕ ( z ) ]
where z is a random variable sampled from the prior distribution p z . G ϕ ( z ) represents the generated sample, which is the output of the random variable z through the generator G ϕ , D θ ( G ϕ ( z ) ) represents the output of the generated sample G ϕ ( z ) through the discriminator, and ϕ is the parameter set of the generator. The goal of the generator is to minimize the above error L ( G , D ) so that the Wasserstein distance between the generator’s distribution p g and the real distribution p r is as small as possible. Algorithm 1 is the process we follow for training the WGAN.
Algorithm 1 WGAN with Gradient Penalty
Default values of λ = 5, n c r i t i c = 5, α = 0.0001, β 1 = 0.5, β 2 = 0.5
1: Require: The gradient penalty coefficient λ, the number of discriminator iterations per generator n c r i t i c , the batch size m, Adam hyper-parameter α , β 1 , β 2 .
2: Require: initial discriminator parameters θ 0 , initial generator parameters ϕ 0 .
3: while loss is still decreasing do
   4: for j = 1, …, n c r i t i c do
     5: for i = 1, …, m do
       6: Sample GPR B-scan x r ~ p r , latent variable z ~ p z
       7:  x f G 0 ( z )
       8:  x ^ t x r + ( 1 t ) x f
       9:  L ( i ) D θ ( x f ) D θ ( x r ) + λ ( x ^ D ( x ^ ) 2 1 ) 2
     10: end for
     11: Update the discriminator by gradient ascent:
     12:  θ A d a m ( θ 1 m i = 1 m L ( i ) , θ , α , β 1 , β 2 )
   13: end for
   14: Sample a batch of latent variable { z ( i ) } i = 1 m ~ p z
   15: Update the generator by gradient ascent:
   16:  ϕ A d a m ( ϕ 1 m i = 1 m D θ ( G ϕ ( z ) ) , ϕ , α , β 1 , β 2 )
17: end while

3. Image Denoising via the MCAE Algorithm

The convolutional auto-encoder is a DL algorithm that uses convolutional neural networks to learn image features. It can be used for image denoising and generally consists of an encoder and a decoder. The encoder is used to encode and compress the information of the noisy image, and the decoder is used to reconstruct and output the denoised image. A multi-scale convolution operation is introduced to learn multi-scale features of GPR images in MCAE. The proposed MCAE consists of an encoder and a decoder. The encoder is composed of three multi-scale convolutional blocks (ConvBlock). Each multi-scale convolutional block includes three parallel convolutional layers and one feature map fusion layer. The decoder consists of three multi-scale deconvolutional blocks (DeconvBlock) and a 3 × 3 convolutional layer. The deconvolutional block includes three parallel deconvolutional layers and a feature map fusion layer. Figure 3 shows the diagram of the MCAE model.
In the encoder, the input image is first processed by the first multi-scale convolutional block. The three parallel convolutional layers of the first convolutional block extract features of the input image using eight kernels. A total of 24 kernels in the first convolutional block and the size of the kernels in each convolutional layer are 1 × 1, 3 × 3, and 5 × 5 respectively. In addition, the dimension of the input image is reduced by half due to the convolution stride is 2. After feature map confusion processing, 24 feature maps are obtained and put into the second multi-scale convolutional block. The processing flow of the second and the third multi-scale convolutional block is similar to that of the first multi-scale convolutional block. The difference is that the three parallel convolutional layers of the second convolutional block use 16 kernels to extract the feature of the input maps, and the three parallel convolution layers of the third convolutional block use 32 kernels to extract the feature of the input maps. Therefore, there are a total of 48 kernels in the second convolutional block, and a total of 96 kernels in the third convolutional block. The input image is processed by the three convolutional blocks of the encoder to output encoded feature maps, and the encoded feature map are input to the decoder.
In the decoder, the encoded feature maps are first processed by the first multi-scale deconvolutional block. The three parallel deconvolutional layers of the first deconvolutional block extract features of the encoded feature maps using 32 kernels. A total of 96 kernels in the first deconvolutional block and the size of the kernels in each deconvolutional layer are 1 × 1, 3 × 3, and 5 × 5, respectively. In addition, the dimension of the encoded feature maps is increased twofold due to the deconvolution stride being 2. After feature map confusion processing, 96 feature maps are obtained and put into the second multi-scale deconvolutional block. The processing flow of the second and the third multi-scale deconvolutional blocks is similar to that of the first multi-scale deconvolutional block. The difference is that the three parallel deconvolutional layers of the second deconvolutional block use 16 kernels to extract the feature of the input maps, and the three parallel deconvolutional layers of the third deconvolutional block use 8 kernels to extract the feature of the input maps. Therefore, there are a total of 48 kernels in the second deconvolutional block, and a total of 24 kernels in the third deconvolutional block. After the third deconvolution block processing, the feature maps are put into a convolutional layer that uses a 3 × 3 kernel to process the input feature maps, and finally output the reconstructed image.
In the convolutional layer, the processing steps include convolutional operation, batch normalization (BN) and rectified linear unit (ReLu) activation. First, N convolutional kernels with the same size k × k are convolved with the input feature map to generate N output feature maps. The convolutional operation [44] between the k × k kernel and the input feature map can be expressed as follows:
h m , n k × k = c = 1 C ( i = 1 k j = 1 k w i , j c x m i + k , n j + k c ) + b
where h m , n k × k indicates the value of the output feature map at position (m,n), w i , j c represents the value of the c-th channel of the convolutional kernel at position (i,j), x m i + k , n j + k c denotes the value of the c-th channel of the input feature map at position (mi + k, nj + k), and b is the bias. After that, in order to speed up the training process, BN is performed on the feature maps obtained by convolution. The BN [45] can be expressed as
h ^ m , n c = γ h m , n c μ c ( σ c ) 2 + ε + β
with
μ c = 1 S s = 1 S ( 1 M × N m = 1 M n = 1 N h m , n c , s )
σ c = 1 S s = 1 S ( 1 M × N m = 1 M n = 1 N ( h m , n c , s μ c ) 2 )
where h m , n c represents the value of the c-th channel of the feature map at position (m,n), and h ^ m , n c denotes the corresponding normalized result. S is the batch size; γ and β are the scale and shift, which are set to 1 and 0 in this paper, respectively; and ε is a constant to ensure the numerical stability. After BN, the ReLu activation [46] is executed. ReLu activation can improve the non-linear expression ability of the model and overcome the problem of gradient disappearance, which is a common operation in neural network. Its processing expression is as follows:
y ^ m , n = max ( h ^ m , n , 0 )
In the deconvolution layer, the processing steps are similar to those in the convolutional layer, including deconvolution operation, batch normalization, and ReLu activation. The deconvolution operation is also called transposed convolution [47], and its processing formula is the same as that of the convolution operation.
A loss function that is able to evaluate the difference between the output image denoised by the MCAE model and the noise-free image is set to mean squared error (MSE). MSE is one of the most generally used loss functions for evaluating the accuracy of the DL model. The MCAE model is optimized by minimizing the pixel-by-pixel difference using Equation (9).
L ( w , b ) = 1 M × N m = 1 M n = 1 N y ^ m , n ( w , b ) y m , n 2
where y m , n represents the pixel value of the m-th row and n-th column in the noise-free GPR image. y ^ m , n indicates the pixel value of the m-th row and n-th column in the denoised GPR image. Algorithm 2 is the process we follow for training the MCAE.
Algorithm 2 MCAE
α = 0.000005
1: Input: Train data T = { ( x 1 , y 1 ) , , ( x n , y n ) } ( x i : d a t a , y i : l a b e l )
2: Output: MCAE parameters ω
3: Require: Initialize MCAE parameters ω 0
4: Require: The batch size S, Adam hyper-parameter α
5: for epochs do
  6: for batch do
    7: Sample { ( x 1 , y 1 ) } i = 1 S a batch from T
    8: L ( i ) M C A E ω ( x ) y 2
    9: ω A d a m ( ω 1 S i = 1 S L ( i ) , ω , α )
10: end for
11: Return MCAE

4. Experiment and Analysis

A series of simulated and field data were used to evaluate the proposed method. In our experiment, the gprMax3.0 [33] tool based on the principle of FDTD was used to carry out GPR detection simulations and obtain 1400 simulated noise-free GPR images of different rebars in the concrete. At the same time, the SIR 4000 GPR equipment manufactured by GSSI (Nashua, NH, USA) was used to obtain 400 GPR field images of different rebars in the concrete. The mean preprocessing method [4] was executed to eliminate the air-ground reflectance strip in GPR images. Its basic principle is that the value of each pixel in the GPR image subtracts the mean value of the row where the pixel is located. In addition, in order to verify the performance of the proposed method, we compared the proposed MCAE algorithm with several state-of-the-art image denoising methods. All the programs were executed on an Intel (Santa Clara, CA, USA) Core i7-9700k server clocked at 3.6 GHz with 32 GB RAM and an NVIDIA (Santa Clara, CA, USA) GeForce RTX 2070 super GPU with 8 GB.

4.1. Data Augmentation Results by WGAN

We employed the WGAN to learn hyperbola features from simulated GPR images and field GPR images. The 1400 simulated GPR images generated by the gprMax tool and the 400 field GPR images obtained by SIR 4000 were used as the training dataset of the WGAN. The dimensions of the input and output images were set to 256 × 256 pixels. In addition, in order to verify the advantages of WGAN, other GANs such as deep convolutional GAN (DCGAN) and auxiliary classifier GAN (ACGAN) were used for comparison.
A visual inspection of the WGAN training process is displayed in Figure 4a. The generator of WGAN learnt the hyperbolic characteristics of GPR images from the training dataset and updated the network parameters under the guidance of the discriminator to generate realistic GPR images. The generated image started from an image with random noise. After 10 epochs, there was a concentration of noise in the generated image. After 50 epochs, some prominent features appeared in the generated image, but it lacked details. When it reached 250 epochs, some subtle features similar to the reflection of rebar were observed, but there were cases where the tail of the hyperbola was missing. After 500 epochs, the generated image revealed more details and was very similar to the real GPR image. Figure 4b,c shows that when the epoch was 1000, the hyperbola of GPR image generated by DCGAN and ACGAN was still blurred, while the image generated by WGAN was clear and realistic. Therefore, the performance of WGAN was better than that of DCGAN and ACGAN. With the help of well-trained WGAN, we were able to generate 3200 generated images to expand the MCAE training dataset.

4.2. Denoising Results by MCAE

After data augmentation, the dataset contained 400 field GPR images, 1400 FDTD-simulated images, and 3200 WGAN-generated images. A total of 5000 noise-free GPR images were used as the label. Then, these 5000 images were added with noise to form the noisy dataset of MCAE, for which PSNR was 11.0 dB. Among them, 4000 (80%) noisy GPR images were used as the training dataset to train the MCAE model, and the remaining 1000 (20%) were used as the test dataset to test the MCAE model performance.
In the training process of the MCAE model, for each noisy GPR image, the data format was first converted, and the size of the converted image was 256 × 256 × 1. The input noisy image was encoded and compressed into a low-dimensional feature map by the MCAE encoder, and the denoised image was output through the decoder. After that, the mean square error between the image output by the decoder and the noise-free GPR image was calculated, and the weight parameters of MCAE were continuously optimized through the reverse gradient propagation algorithm to reduce the mean square error. In the experiment, the batch size was set to 100, and the learning rate was set to 0.000005. After 400 epochs, the mean square error L stabilized, and the model was well trained. As shown in Figure 5, after implementing the data augmentation method, we found that the model converged faster, and the stable loss value was lower.
In the verification stage, the noisy image was put into the trained MCAE model, and the denoised GPR image was reconstructed and output through encoding and decoding. The peak signal-to-noise ratio (PSNR) of the denoised image was calculated. The calculation expression of PSNR is shown as follows:
PSNR = 10 log 10 ( 2 a 1 ) 2 1 M × N m = 1 M n = 1 N y ^ m , n y m , n 2
where y m , n represents the value of the m-th row and n-th column of the noise-free GPR image; y ^ m , n represents the value of the m-th row and n-th column of the GPR image after denoising; and M and N are the height and width of the image, respectively. a is the bit per pixel and is usually set to 8, which means that the number of pixel gray levels is 256. The larger the PSNR value, the smaller the distortion.
The experimental results are shown in Figure 6. In the simulated image, the noisy GPR image with PSNR of 11.23 dB was denoised by MCAE, and its PSNR can be increased to 30.18 dB. In the field image, the noisy GPR image with PSNR of 11.25 dB was denoised by MCAE, and its PSNR can be increased to 32.41 dB. In the generated image, the noisy GPR image with PSNR of 11.28 dB was denoised by MCAE, and its PSNR can be increased to 31.82 dB.

4.3. Comparison of Experimental Results

The proposed MCAE algorithm was compared with several state-of-the-art image denoising methods, including BM3D, WNNM, NLM, and CAE. The classic Wavelet denoising method was also compared. PSNR is an important performance metric in terms of analyzing the performance of the different denoising methods; in addition, the structural similarity index (SSIM) and time cost were also introduced to analyze the different denoising methods. The value range of SSIM was between 0 and 1. The larger the SSIM value, the smaller the image distortion.
We evaluated the competing methods on 30 test images consisting of 10 simulated images obtained by the gprMax tool, 10 field images obtained by SIR 4000, and 10 generated images obtained by WGAN. Figure 7 shows the image scenes. Zero mean additive white Gaussian noise with different standard deviations (σ = 10, 30, 50, and 70) were added to these test images to generate the noisy images.
Table 1 shows the PSNR results of different denoising methods. The result for each image and on each noisy level is highlighted in bold, which denotes the highest PSNR. The proposed MCAE achieved the highest PSNR in almost every case. When the noise level was relatively low (σ = 10), on the simulated image, the proposed method achieved the highest PSNR of 31.61 dB. On the field and generated images, the BM3D method achieved the highest PSNRs of 38.92 dB and 36.59 dB, respectively. However, as the noise level continued to increase (σ = 30, 50, and 70), the performance of the proposed method could be maintained, and it achieved the highest PSNR on all images. This was followed by CAE and WNNM methods, which could also maintain a relatively high PSNR. However, when the noise level σ exceeded 50, the PSNR values achieved by both BM3D and Wavelet were relatively low, and the NLM method was not even effective. These indicate that the proposed MCAE method is more robust to noise intensity than other methods.
Table 2 shows the SSIM results of different denoising algorithms. In each noise level, the highest SSIM result of each image is bolded. In all four noise levels, the proposed method achieved the highest SSIM. Under a low noise level (σ = 10), BM3D and WNNM achieved the same SSIM of 0.95 as the proposed method on the field image, which had good denoising performance. However, as the noise level increased, the SSIM of the BM3D, Wavelet, and NLM methods dropped sharply, being lower than 0.5 when the σ was 70. The CAE method and the WNNM method can also maintain a high SSIM, but they were still lower than the proposed method.
Table 3 shows the time cost of different denoising algorithms for processing each image. It can be seen that BM3D consumed the most time, with an average time consumption of 116.28 s on the three images, followed by WNNM, with an average time consumption of 40.31 s. Both the proposed method and the CAE method had good performance. Their average consumption times were only 0.241 s and 0.231 s, respectively, which could realize real-time processing.
In Figure 8 and Figure 9, we show a comparison of the visual quality of denoised images processed by different methods. Figure 8 shows that when the noise level was relatively low (σ = 30), all denoising methods obtained good image visual quality and the proposed method was able to reconstruct more image details from noisy observations. Figure 9 shows that when the image was at a strong noise level (σ = 70), the proposed method could still reconstruct the details of the image and maintain a clear edge structure. However, the image visual quality of other methods was greatly reduced. Compared with the MCAE method, the method such as NLM, BM3D, and Wavelet still retained a large amount of noise. Although CAE and WNNM could suppress random noise effectively, the reconstructed image lost the hyperbolic tail features and edge information. In summary, MCAE had a powerful denoising ability, producing a visually pleasing denoising output while having higher PSNR.

5. Conclusions

In this paper, we propose a scheme for GPR image denoising based on DL techniques, which consisted of WGAN and MACE. The MCAE algorithm improved the CAE model by introducing a multi-scale convolution kernel to extract higher-level features of the image. The WGAN was designed for increasing the GPR images to solve the problem of insufficient MCAE training dataset. The MCAE achieved a better denoising effect with the data augmentation. The experimental results showed that MCAE could not only bring significant PSNR improvements over state-of-the-art methods such as CAE, WNNM, and BM3D, but also could reconstruct better local structures of the image and less visual artifacts. However, the factors that affect GPR image quality included not only random noise but also clutter caused by multi-target interference. In future work, we will continue to study the methods based on DL technology to remove the GPR image clutter.

Author Contributions

Data curation, J.L., C.W., and Q.R.; funding acquisition, W.L.; methodology, J.L.; software, J.L., L.X., and S.Z.; supervision, F.H.; validation, J.L.; visualization, S.Z., S.L., and Y.W.; writing—original draft, J.L.; writing—review and editing, W.L. and F.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Fundamental Research Funds for the Central Universities of Central South University (grant number 2021zzts0732) and project-“Research on GPR focusing imaging algorithm of targets embedded in layer mediums” (grant number [2019]267).

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to restriction of privacy.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, C.S.; Jeng, Y. GPR investigation of the near-surface geology in a geothermal river valley using contemporary data decomposition techniques with forward simulation modeling. Geothermics 2016, 64, 439–454. [Google Scholar] [CrossRef]
  2. Neal, A. Ground-penetrating radar and its use in sedimentology: Principles, problems and progress. Earth Sci. Rev. 2004, 66, 261–330. [Google Scholar] [CrossRef]
  3. Dong, Z.; Ye, S.; Gao, Y.; Fang, G.; Zhang, X.; Xue, Z.; Zhang, T. Rapid Detection Methods for Asphalt Pavement Thicknesses and Defects by a Vehicle-Mounted Ground Penetrating Radar (GPR) System. Sensors 2016, 16, 2067. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Lei, W.; Luo, J.; Hou, F.; Xu, L.; Jiang, X. Underground Cylindrical Objects Detection and Diameter Identification in GPR B-Scans via the CNN-LSTM Framework. Electronics 2020, 9, 1804. [Google Scholar] [CrossRef]
  5. Faize, A.; Alsharahi, G.; Hamdaoui, M. Radar GPR Application to Explore and Study Archaeological Sites: Case Study. Int. J. Adv. Comput. Sci. Appl. 2020, 11, 179–182. [Google Scholar]
  6. Jol, H. Ground Penetrating Radar: Theory and Applications; Elsevier Science: Amsterdam, The Netherlands, 2009. [Google Scholar]
  7. Ciampoli, L.B.; Tosti, F.; Economou, N.; Benedetto, F. Signal Processing of GPR Data for Road Surveys. Geosciences 2019, 9, 96. [Google Scholar] [CrossRef] [Green Version]
  8. Bugarinovi, E.; Pajewski, L.; Ristic, A.; Vrtunski, M.; Borisov, M. On the Introduction of Canny Operator in an Advanced Imaging Algorithm for Real-Time Detection of Hyperbolas in Ground-Penetrating Radar Data. Electronics 2020, 9, 541. [Google Scholar] [CrossRef] [Green Version]
  9. Oskooi, B.; Parnow, S.; Smirnov, M.; Varfinezhad, R.; Yari, M. Attenuation of random noise in GPR data by image processing. Arab. J. Geosci. 2018, 11, 677. [Google Scholar] [CrossRef]
  10. Kumlu, D.; Erer, I.; Kaplan, N.H. Low complexity clutter removal in GPR images via lattice filters. Digit. Signal Process. 2020, 101, 102724. [Google Scholar] [CrossRef]
  11. Zhu, J.; Xue, W.; Rong, X.; Yu, Y. A clutter suppression method based on improved principal component selection rule for ground penetrating radar. Prog. Electromagn. Res. M 2017, 53, 29–39. [Google Scholar] [CrossRef] [Green Version]
  12. Mursal, A.; Ibrahim, H. Median Filtering Using First-Order and Second-Order Neighborhood Pixels to Reduce Fixed Value Impulse Noise from Grayscale Digital Images. Electronics 2020, 9, 2034. [Google Scholar] [CrossRef]
  13. Lee, J.S. Speckle filtering of synthetic aperture radar images: A review. Remote Sens. Rev. 1994, 8, 313–340. [Google Scholar] [CrossRef]
  14. Frost, V.S.; Stiles, J.A.; Shanmugan, K.S.; Holtzman, J.C. A Model for Radar Images and Its Application to Adaptive Digital Filtering of Multiplicative Noise. IEEE Trans. Pattern Anal. Mach. Intell. 1982, 4, 157–166. [Google Scholar] [CrossRef]
  15. Kuan, D.; Sawchuk, A.; Strand, T.; Chavel, P. Adaptive restoration of images with speckle. IEEE Trans. Acoust. Speech Signal Process. 1987, 35, 373–383. [Google Scholar] [CrossRef]
  16. Kumlu, D.; Erer, I. Clutter removal techniques in ground penetrating radar by using non-local means approach. J. Fac. Eng. Archit. Gazi Univ. 2020, 35, 1269–1284. [Google Scholar] [CrossRef] [Green Version]
  17. Guangyin, L.U.; Luo, S.; Zhu, Z.; Zhao, Y.; Feiyan, X.I. Application of random noise suppression method based on two-dimensional wavelet transform in GPR data. Comput. Tech. Geophys. Geochem. Explor. 2019, 41, 234–240. [Google Scholar]
  18. Zou, H.L.; Sui, Y.L.; Xu, J.Y.; Ning, S.N. Study on methods of GPR image de-noising based on multi-wavelets transform. Acta Simulata Syst. Sin. 2005, 4, 855–858. [Google Scholar]
  19. Wang, X.; Liu, S. Noise suppressing and direct wave arrivals removal in GPR data based on Shearlet transform. Signal Process. 2017, 132, 227–242. [Google Scholar] [CrossRef]
  20. Yahya, A.A.; Tan, J.Q.; Su, B.Y.; Hu, M.; Wang, Y.B.; Liu, K.; Hadi, A.N. BM3D image denoising algorithm based on an adaptive filtering. Multimed. Tools Appl. 2020, 79, 20391–20427. [Google Scholar] [CrossRef]
  21. Xue, W.; Luo, Y.; Yang, Y.; Huang, Y. Noise Suppression for GPR Data Based on SVD of Window-Length-Optimized Hankel Matrix. Sensors 2019, 19, 3807. [Google Scholar] [CrossRef] [Green Version]
  22. Riaz, M.M.; Ghafoor, A. Information Theoretic Criterion Based Clutter Reduction for Ground Penetrating Radar. Prog. Electromagn. Res. B 2012, 45, 147–164. [Google Scholar] [CrossRef] [Green Version]
  23. Chen, G.; Fu, L.; Chen, K.; Boateng, C.D.; Ge, S. Adaptive Ground Clutter Reduction in Ground-Penetrating Radar Data Based on Principal Component Analysis. IEEE Trans. Geosci. Remote Sens. 2019, 57, 3271–3282. [Google Scholar] [CrossRef]
  24. Abujarad, F.; Omar, A. Comparison of independent component analysis (ICA) algorithms for GPR detection of non-metallic land mines. In Proceedings of the SPIE International Society for Optical Engineering, San Diego, CA, USA, 15–17 August 2006. [Google Scholar]
  25. Liu, H.B.; Wang, X.; Zheng, M. A clutter suppression method of ground penetrating radar for detecting shallow surface target. In Proceedings of the International Radar Conference, Hangzhou, China, 14–16 October 2015. [Google Scholar]
  26. Wu, J.; Lee, X. An Improved WNNM Algorithm for Image Denoising. J. Phys. Conf. Ser. 2019, 1237, 022037. [Google Scholar] [CrossRef]
  27. Gu, S.; Lei, Z.; Zuo, W.; Feng, X. Weighted Nuclear Norm Minimization with Application to Image Denoising. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  28. Wang, P.; Zhang, H.; Patel, V.M. SAR Image Despeckling Using a Convolutional Neural Network. IEEE Signal Process. Lett. 2017, 24, 1763–1767. [Google Scholar] [CrossRef] [Green Version]
  29. Zhang, C.; Zhou, L.; Zhao, Y.; Zhu, S.; He, Y. Noise reduction in the spectral domain of hyperspectral images using denoising autoencoder methods. Chemom. Intell. Lab. Syst. 2020, 203, 104063. [Google Scholar] [CrossRef]
  30. Choi, S.H.; Choi, H.J.; Min, C.H.; Chung, Y.H.; Ahn, J.J. Development of de-noised image reconstruction technique using Convolutional AutoEncoder for fast monitoring of fuel assemblies. Nucl. Eng. Technol. 2020, 53, 888–893. [Google Scholar] [CrossRef]
  31. Hawkins, D.M. The problem of overfitting. ChemInform 2004, 35, 1. [Google Scholar] [CrossRef]
  32. Zhang, C.; Vinyals, O.; Munos, R.; Bengio, S. A Study on Overfitting in Deep Reinforcement Learning. arXiv 2018, arXiv:1804.06893. Available online: https://arxiv.org/abs/1804.06893 (accessed on 15 March 2021).
  33. Warren, C.; Giannopoulos, A.; Giannakis, I. gprMax: Open source software to simulate electromagnetic wave propagation for Ground Penetrating Radar. Comput. Phys. Commun. 2016, 209, 163–170. [Google Scholar] [CrossRef] [Green Version]
  34. Frid-Adar, M.; Diamant, I.; Klang, E.; Amitai, M.; Goldberger, J.; Greenspan, H. GAN-based Synthetic Medical Image Augmentation for increased CNN Performance in Liver Lesion Classification. Neurocomputing 2018, 321, 321–331. [Google Scholar] [CrossRef] [Green Version]
  35. Bang, S.; Baek, F.; Park, S.; Kim, W.; Kim, H. Image augmentation to improve construction resource detection using generative adversarial networks, cut-and-paste, and image transformation techniques. Autom. Constr. 2020, 115, 103198. [Google Scholar] [CrossRef]
  36. Klemmer, K.; Saha, S.; Kahl, M.; Xu, T.; Zhu, X.X. Generative modeling of spatio-temporal weather patterns with extreme event conditioning. arXiv 2021, arXiv:2104.12469. Available online: https://arxiv.org/abs/2104.12469 (accessed on 10 April 2021).
  37. Yu, Y.; Li, X.; Liu, F. Attention GANs: Unsupervised Deep Feature Learning for Aerial Scene Classification. IEEE Trans. Geosci. Remote Sens. 2019, 58, 519–531. [Google Scholar] [CrossRef]
  38. Roy, S.; Sangineto, E.; Sebe, N.; Demir, B. Semantic-Fusion Gans for Semi-Supervised Satellite Image Classification. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2016. [Google Scholar]
  39. Saha, S.; Bovolo, F.; Bruzzone, L. Unsupervised Multiple-Change Detection in Vhr Multisensor Images via Deep-Learning Based Adaptation. In Proceedings of the 2019 IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 28 July–2 August 2019; pp. 5033–5036. [Google Scholar]
  40. Lashgari, E.; Liang, D.; Maoz, U. Data Augmentation for Deep-Learning-Based Electroencephalography. J. Neurosci. Methods 2020, 346, 108885. [Google Scholar] [CrossRef]
  41. Wang, X.; Wang, K.; Lian, S. A survey on face data augmentation for the training of deep neural networks. Neural Comput. Appl. 2020, 32, 15503–15531. [Google Scholar] [CrossRef] [Green Version]
  42. Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein GAN. arXiv 2017, arXiv:1701.07875. Available online: https://arxiv.org/abs/1701.07875 (accessed on 20 March 2021).
  43. Gulrajani, I.; Ahmed, F.; Arjovsky, M.; Dumoulin, V.; Courville, A. Improved Training of Wasserstein GANs. arXiv 2017, arXiv:1704.00028. Available online: https://arxiv.org/abs/1704.00028 (accessed on 20 March 2021).
  44. Jin, L.; Zhang, W.; Ma, G.; Song, E. Learning deep CNNs for impulse noise removal in images. J. Vis. Commun. Image Represent. 2019, 62, 193–205. [Google Scholar] [CrossRef]
  45. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. JMLR 2015, 37, 448–456. [Google Scholar]
  46. Wang, D.P.; Teng, G.; Computer, S.O.; University, J.N. Optimal Design of ReLU Activation Function in Convolutional Neural Networks. Inf. Commun. 2018, 1, 42–43. [Google Scholar]
  47. Noh, H.; Hong, S.; Han, B. Learning Deconvolution Network for Semantic Segmentation. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015. [Google Scholar]
Figure 1. Illustration of GPR B-scan image denoising by the proposed scheme.
Figure 1. Illustration of GPR B-scan image denoising by the proposed scheme.
Electronics 10 01269 g001
Figure 2. WGAN structure for generating GPR images.
Figure 2. WGAN structure for generating GPR images.
Electronics 10 01269 g002
Figure 3. Structure of the multi-scale convolutional autoencoder (MCAE) for GPR image denoising.
Figure 3. Structure of the multi-scale convolutional autoencoder (MCAE) for GPR image denoising.
Electronics 10 01269 g003
Figure 4. Generated GPR images using different GAN methods with respect to different training epochs: (a) WGAN; (b) DCGAN; (c) ACGAN.
Figure 4. Generated GPR images using different GAN methods with respect to different training epochs: (a) WGAN; (b) DCGAN; (c) ACGAN.
Electronics 10 01269 g004
Figure 5. Training loss of the MCAE with/without data augmentation.
Figure 5. Training loss of the MCAE with/without data augmentation.
Electronics 10 01269 g005
Figure 6. The denoised results of MCAE algorithm. (a) Simulated noisy image (PSNR: 11.23 dB). (b) Field noisy image (PSNR: 11.25 dB). (c) Generated noisy image (PSNR: 11.28 dB). (d) Simulated image denoised by MCAE (PSNR: 30.18 dB). (e) Field image denoised by MCAE (PSNR: 32.41 dB). (f) Generated image denoised by MCAE (PSNR: 31.82 dB).
Figure 6. The denoised results of MCAE algorithm. (a) Simulated noisy image (PSNR: 11.23 dB). (b) Field noisy image (PSNR: 11.25 dB). (c) Generated noisy image (PSNR: 11.28 dB). (d) Simulated image denoised by MCAE (PSNR: 30.18 dB). (e) Field image denoised by MCAE (PSNR: 32.41 dB). (f) Generated image denoised by MCAE (PSNR: 31.82 dB).
Electronics 10 01269 g006
Figure 7. The 30 test images.
Figure 7. The 30 test images.
Electronics 10 01269 g007
Figure 8. Denoising results on images by different methods (noise level σ = 30).
Figure 8. Denoising results on images by different methods (noise level σ = 30).
Electronics 10 01269 g008
Figure 9. Denoising results on images by different methods (noise level σ = 70).
Figure 9. Denoising results on images by different methods (noise level σ = 70).
Electronics 10 01269 g009
Table 1. Average PSNR (dB) for different standard deviations with various denoising methods.
Table 1. Average PSNR (dB) for different standard deviations with various denoising methods.
-Noisy ImageProposedCAENLMWNNMBM3DWavelet
σ = 10Simulated images31.6130.4630.7728.8930.9529.92
Field images38.1235.6234.3636.9838.9238.06
Generated images36.3532.6832.1232.5336.5935.74
σ = 30Simulated images30.6128.9229.1328.7529.9628.66
Field images35.8232.5631.5032.6331.8530.23
Generated images34.4232.4630.8330.6331.3629.83
σ = 50Simulated images30.3927.6416.1227.9225.9525.61
Field images33.3230.8915.4429.1427.2426.15
Generated images32.6629.7216.6029.8326.8226.22
σ = 70Simulated images30.1827.1812.2827.8122.5623.67
Field images32.4130.3415.9827.0924.5123.69
Generated images31.8228.2312.4328.6721.8523.74
Table 2. Average SSIM for different standard deviations with various denoising methods.
Table 2. Average SSIM for different standard deviations with various denoising methods.
-Noisy ImageProposedCAENLMWNNMBM3DWavelet
σ = 10Simulated images0.910.860.850.890.910.87
Field images0.950.930.930.950.950.93
Generated images0.930.880.870.900.920.90
σ = 30Simulated images0.870.850.780.840.710.68
Field images0.940.920.840.930.760.74
Generated images0.900.870.800.860.720.69
σ = 50Simulated images0.890.820.200.810.470.51
Field images0.930.900.210.900.570.56
Generated images0.860.830.190.840.460.52
σ = 70Simulated images0.850.790.110.820.420.43
Field images0.920.890.120.880.520.44
Generated images0.850.810.090.830.370.41
Table 3. Average time cost (s) for different denoising methods.
Table 3. Average time cost (s) for different denoising methods.
Noisy ImageProposedCAENLMWNNMBM3DWavelet
Simulated images0.2420.2273.86540.56120.710.751
Field images0.2400.2314.06239.67114.370.762
Generated images0.2410.2363.73540.72113.750.663
AVE.0.2410.2313.88740.31116.280.725
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Luo, J.; Lei, W.; Hou, F.; Wang, C.; Ren, Q.; Zhang, S.; Luo, S.; Wang, Y.; Xu, L. GPR B-Scan Image Denoising via Multi-Scale Convolutional Autoencoder with Data Augmentation. Electronics 2021, 10, 1269. https://doi.org/10.3390/electronics10111269

AMA Style

Luo J, Lei W, Hou F, Wang C, Ren Q, Zhang S, Luo S, Wang Y, Xu L. GPR B-Scan Image Denoising via Multi-Scale Convolutional Autoencoder with Data Augmentation. Electronics. 2021; 10(11):1269. https://doi.org/10.3390/electronics10111269

Chicago/Turabian Style

Luo, Jiabin, Wentai Lei, Feifei Hou, Chenghao Wang, Qiang Ren, Shuo Zhang, Shiguang Luo, Yiwei Wang, and Long Xu. 2021. "GPR B-Scan Image Denoising via Multi-Scale Convolutional Autoencoder with Data Augmentation" Electronics 10, no. 11: 1269. https://doi.org/10.3390/electronics10111269

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop