Next Article in Journal
Can Shared Control Improve Overtaking Performance? Combining Human and Automation Strengths for a Safer Maneuver
Next Article in Special Issue
Irregular Scene Text Detection Based on a Graph Convolutional Network
Previous Article in Journal
Remaining Useful Life Prediction Based on Adaptive SHRINKAGE Processing and Temporal Convolutional Network
Previous Article in Special Issue
An Improved Frequency Domain Guided Thermal Imager Strips Removal Algorithm Based on LRSID
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Super-Resolution Reconstruction Method of Pavement Crack Images Based on an Improved Generative Adversarial Network

School of Information Engineering, Chang’an University, Xi’an 710064, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(23), 9092; https://doi.org/10.3390/s22239092
Submission received: 13 October 2022 / Revised: 12 November 2022 / Accepted: 21 November 2022 / Published: 23 November 2022

Abstract

:
A super-resolution reconstruction approach based on an improved generative adversarial network is presented to overcome the huge disparities in image quality due to variable equipment and illumination conditions in the image-collecting stage of intelligent pavement detection. The nonlinear network of the generator is first improved, and the Residual Dense Block (RDB) is created to serve as Batch Normalization (BN). The Attention Module is then formed by combining the RDB, Gated Recurrent Unit (GRU), and Conv Layer. Finally, a loss function based on the L1 norm is utilized to replace the original loss function. The experimental findings demonstrate that the self-built pavement crack dataset’s Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity (SSIM) of the reconstructed images reach 29.21 dB and 0.854, respectively. The results improved compared to the Set5, Set14, and BSD100 datasets. Additionally, by employing Faster-RCNN and a Fully Convolutional Network (FCN), the effects of image reconstruction on detection and segmentation are confirmed. The findings indicate that the segmentation results’ F1 is enhanced by 0.012 to 0.737 and the detection results’ confidence is increased by 0.031 to 0.9102 when compared to state-of-the-art methods. It has a significant engineering application value and can successfully increase pavement crack-detecting accuracy.

1. Introduction

Many external variables, such as movement speed, equipment performance, relative light intensity, etc., interfere with pavement crack image collecting, resulting in poor image quality and drastically reduced detection and segmentation accuracy. Image super-resolution reconstruction [1] is proposed to improve the original image’s resolution and enhance the image’s content features, thereby resolving the faults produced by low-resolution images. This approach efficiently reduces hardware limits, enhances identification accuracy, and serves as a valuable reference for long-term pavement maintenance [2].
Previously, the network topology that resulted in good model performance was continually deeper. Despite the fact that images with a higher PSNR are recreated, texture details are not realistic or natural, and model degradation is a possibility. Excessive focus on the reconstructed image’s PSNR value while neglecting the visual effects’ performance results in extremely smooth texture detail.
Due to their remarkable generating capabilities, generative adversarial networks (GANs) theory and applications have rapidly expanded in recent years [3,4,5], and multiple variations [6,7,8,9] have been proposed and are frequently employed in diverse tasks of image processing. GANs are also employed in a number of areas [10,11,12,13,14] as important methods of virtual generation [15] and data set augmentation [16]. Christine et al. [17] proposed a DCGAN-based synthetic data method for improving the recognition of traffic signs. In this method, CNN models with different backbone networks and feature extractors are analyzed and discussed, and Resnet50 and Densenet are introduced into DNGAN to evaluate the generation effect. After mixing the synthetic image with the original image, the detection network is trained again, which improves the accuracy of traffic sign recognition performance and reduces the detection speed accordingly. Zhang et al. [18] proposed a permanent magnet motor fault diagnosis method based on DCGAN-RCCNN, which used DCGAN to generate virtual fault data, and used RCCNN to feature extract and classify stator current data, and finally realized the expansion of training samples. This method effectively solves the problems of insufficient fault data and difficult fault diagnosis of permanent magnet motors.
The Super-Resolution Using Generative Adversarial Networks (SRGAN) [19] was presented for image super-resolution reconstruction, and the results were impressive. SRGAN separates the network into generative and discriminant networks, allowing for increased image resolution and restoration of image detail characteristics without sacrificing PSNR.
In order to improve the recognition ability of low-resolution number plates, the problem of poor boundary and contrast is solved. Kabiraj et al. [20] used an enhanced super-resolution generative adversarial network to upscale from low resolution to high resolution to recover useful information that would improve the accuracy of number plate detection and recognition. They used the residual module to improve the native dense blocks of the generative adversarial network and combined the OCR model to segment the number plates to improve the detection performance of number plate information. Rashid et al. [21] applied this method to enhance the spatial resolution of two-dimensional MR images. By modifying ESRGAN’s network structure, it can train 2D Magnetic Resonance images (MRI). Xu et al. [22] proposed a new network architecture based on the principle of SRGAN: Enhanced High-Resolution Semantic Network (EHRS-Net). This method maintains and enhances the details of the feature map through the convolution process to improve the overall crack detection accuracy. By evaluating two pavement crack datasets, it is proved that the proposed pavement crack detection method is superior. Pei et al. [11] expanded the crack training set based on DCGAN, used Faster-RCNN to train the data set before and after the expansion, and found that the detection accuracy improved.
Even though SRGAN can generate sharper images compared to other networks, its network structure is complex. If the depth of the model is raised blindly for special tasks, the network parameters will drastically increase, and the performance improvement will be restricted. The analysis found that SRGAN has the following problems: (1) SRGAN is mainly composed of residual blocks of the same structure, consisting of two convolutional layers and two BN layers. As discussed in [23], the computational cost soars when the number of network layers increases. (2) SRGAN inherits the characteristics of GAN and pays more attention to the authenticity of the generated images while ignoring the differences in the details of the generated images. It leads to the problem that visual perception and practical application cannot be better satisfied. (3) The performance of pavement crack detection is highly dependent on the image quality sent into the network. The direct use of the original SRGAN, although the image’s resolution has been improved, there is still room for improvement in the reconstruction of crack characteristics.
This research is presented to solve the above problems and combine the engineering needs of pavement detection. The creativity of the research mainly lies in the super-resolution model based on the GAN and the attention mechanism. The main technical contributions are as follows:
  • Improved the non-linear network in the generator to create an attention recurrent network that removes the Batch Normalization (BN) and is fused by the Residual Dense Block (RDB), the Gated Recurrent Unit (GRU), and Attention. The improved generator pays more attention to crack features, and the reconstructed crack images are more detailed. This research aims to solve the shortcomings of high PSNR value but imperfect detail reconstruction in the original network reconstruction process.
  • Replace the original GAN’s loss function with an L1 norm-based loss function. Using this loss function enhances the generator’s learning ability while also enhancing the image reconstruction quality.
  • For the unique field of pavement crack image super-resolution applications, not only the PSNR and SSIM metrics are adopted, but also Faster-RCNN (the detection model) and FCN (segmentation model) are utilized to evaluate the detection and segmentation effects of reconstructed high-resolution images. These detailed evaluation methods can more comprehensively verify the methods’ effectiveness and provide reference ideas for other research in this field.

2. Methodology

This research incorporates the attention mechanism into the SRGAN model to address the model’s weaknesses. In terms of parameter control, a recursive structure is utilized to share network parameters so that the parameter scale does not expand dramatically as the network depth grows. In the performance of image details, feature extraction is performed on low-resolution images, and the attention mechanism is employed to distinguish between low-frequency information and high-frequency information.

2.1. Attention Generative Adversarial Network Structure

A generator and a discriminator are often used in GAN. The attention mechanism and recurrent network are introduced into the original network, and the components of GAN are redesigned. The upgraded component allows the super-resolution reconstruction model to combine the attention mechanism’s visual performance advantage with the recurrent network-sharing parameters’ feature. Figure 1 depicts the design model’s general structure.
The generator’s major components are shallow feature extraction, nonlinear mapping network, and up-sampling. The discriminator receives the super-resolution reconstructed image as input, which consists mostly of feature extraction and complete connection. It is used to determine whether the generator’s output and the real label are similar and whether the result is True or False.

2.2. Generator Design

The shallow feature extraction network, nonlinear mapping network, and up-sampling network are the primary components of the generator. As illustrated in Figure 2, the nonlinear mapping network is an attention recurrent network.
I L R represents the low-resolution image, and I S R represents the result after network reconstruction. The low-resolution image input by the network extracts shallow feature information through the Shallow Feature Extraction Network, mainly composed of two convolutional layers (Conv). The first Conv layer extracts the feature F 1 from the I L R , and the second performs further shallow feature extraction on F 1 , and the output is F 0 , as shown in Equations (1) and (2), respectively.
F 1 = H S F E 1 I L R
F 0 = H S F E 1 F 1
H S F E 1 ( · ) and H S F E 2 ( · ) serve as the convolution operations of the shallow feature extraction network. After the original input L R is subjected to shallow feature extraction, its output F 0 is sent to the attention recurrent network as a new input. The structure of the attention recurrent network is shown in Figure 3. The main purpose of the attention recurrent network is to extract the texture details that need to be recovered from the input image. The extraction results are enhanced by subsequent nonlinear mapping and used to generate high-resolution images. Therefore, the effect of reconstructing the image depends on the quality of texture detail extraction by the attention recurrent network.
(1) RDB: To replace the original structure, the redesigned dense residual block is shown in Figure 3. The BN layer creates pseudo-textures in the output images when the statistics of the training and test datasets change significantly, reducing the generalization ability. The BN layer is deleted to stabilize the network’s training, minimize computational complexity, and reduce computational overhead. RDB achieves feature fusion and dimensionality reduction by merging residual blocks with skip connections [23], resulting in dense blocks. To boost network capacity, the dense residual block not only preserves the feedforward information, but also fully extracts the local feature layer information.
(2) GRU: In Attention Recurrent Networks, GRU are introduced. As illustrated in Equations (3)–(6), the most significant structures are gate structures, namely Reset Gate rt and Update Gate zt.
r t = σ W r h t 1 , x t
z t = σ W z h t 1 , x t
h ^ t = tan h W r t     h t 1 , x t
h t = 1 z t     h t 1 + z t     h ^ t
In the hidden layer, h t 1 is the state at the previous time, h t is the state at the current time, and h ^ t is the updated state at the current time. W z produces the weight corresponding to the update state, W r represents the weight in the reset state, and * represents the convolution operation. To produce 2D attention images, the GRU’s output features are fed to successive convolutional layers. The resulting attention image is concatenated with the input image at each time step during training and utilized as the input to the next layer of the recurrent network. As shown in Equation (7), it is assumed that the output of the recurrent attention network after n layers is F n .
F n = H A T T , n F n 1 = H A T T , n H A T T , n 1 H A T T , 1 F 0
H A T T represents the recurrent attention network function, and n represents the number of layers of recursion. Feature mapping is performed after the texture detail features are extracted through the attention recurrent network. The designed feature mapping network has 8 Conv + ReLUs structures, and skip connections are added to improve the stability of network training, as shown in the right half of Figure 3. The feature map network is shown in Equation (8):
F N M N = H N M N F n
FNMN is the output result of using the nonlinear network function H N M N . The up-sampling network consists of convolutional layers before generating the high-resolution image to obtain the super-resolution image output scaled by a factor of 4 in Equation (9):
I S R = H U P S F N M N
H U P S is the up-sampling function, and I S R is the super-resolution image output.
(3) Attention: The attention feature of each layer of a recurrent network is a matrix with values between 0 and 1. The stronger the associated attention, the higher the element values in each matrix. The first four Attention (A1A4) in the Attention recursive network are selected for the output of the attention feature map, and they are colored using pseudocolor, as shown in Figure 4, where   A n are the generated visual attention maps. As recursion increases, attention feature maps highlight texture details and edges.
The generated visual attention map output by the first attention module A1 is basically blue. When recursive to Attention (A4), A4 corresponding to the input (the crack target area) is yellow, while the background is still blue. Therefore, the feature maps output by the four attention modules is weak in blue and strong in yellow.

2.3. Discriminator Design

After redesigning the generator, the discriminator needs to be improved to match the new generator in Figure 5. GAN uses a discriminator to judge the real and generated images, which are then fed back to the generator. The visual characteristics of the high-resolution-produced image to be assessed are extracted using seven layers of convolutional layers, then dimensionally flattened. Lastly, the results are discriminated using the fully connected layer (FC) and the Sigmoid function.
In the first training, the performance of the original generator was considerably inferior to that of the discriminator, causing the model to crash. To avoid this from happening, the loss function must be enhanced. The original loss function is replaced with a new loss function based on the L1 norm, which includes the discriminator’s reconstruction error. The loss function improves the learning ability of the generator while improving the image quality.
The image pixel-level error is considered to obey the Gaussian distribution, and the loss function based on the L1 norm is defined in Equation (10).
L I = I H R G I L R
I H R indicates the high-resolution images, and I L R indicates for low-resolution images. The optimization rules of the generator and discriminator loss functions that need to be iterated are expressed in Equations (11)–(15):
L D = L D r k t L D f , f o r θ D L G = L G z x , f o r θ G k t + 1 = k t + λ k γ L D r L G
L D f = L D x ; θ D x
L D f = L D y ; θ D y = L D G z ; θ G G z ; θ G ; θ D
y = G x ; θ D
γ = E L G z E L x
x performs a high-resolution image, z represents a low-resolution image, y is a super-resolution image, L D f represents the loss of a high-resolution image by the discriminator, and L D f indicates a loss of a low-resolution image. λ k represents the increment of k , k t means the result of the t -th iteration of k , and the change of the value of k can be used to improve the learning ability of the generator. The γ is the ratio of the expected value of the super-resolution image error to the expected value of the high-resolution image. The value of this parameter can improve the quality of the generated image.

3. Experimental Evaluation and Results Analysis

3.1. Training Parameter Setting

The dataset, the number of network layers, and the convolution kernel size are all elements that affect network performance. In this experiment, the training set includes 800 images from the ImageNet public dataset and 300 images from self-built pavement crack datasets in Figure 6. The self-built dataset mainly contains three types of pavement cracks: horizontal, longitudinal, and reticular. A Garmin VIRB professional sports area scan camera is used for sampling, and a 100W LED supplementary light source is used to avoid the interference of road shadows. The calibration shooting area is 1.5 m wide and 1 m long, and the resolution is 2048 × 1024. The camera is placed 1 m from the ground to shoot vertically. To avoid shooting shadows caused by direct sunlight on sunny days, try to choose cloudy days and set the exposure time to 300 us. The training set’s images are downsampled at a predetermined ratio of four times (4).
SRCNN [24] with different numbers of convolution kernels and varied convolution sizes and DRCN [25] with varying convolution sizes are evaluated independently using the PSNR measure to assess the training conditions.
Set the number of convolution kernels in the first layer to n 1 and the second layer to n 2 , and conduct comparative experiments in three cases: (i)   n 1 = 32, n 2 = 16; (ii)   n 1 = 64, n 2 = 32, (iii) n 1 = 128, n 2 = 64. The size of the convolution kernel’s first layer of the convolution kernel is f 1 = 9, and the last layer is f 2 = 5. The results are shown in Table 1. When the first layer and the second layer have 64 and 32 convolution kernels, respectively, the PSNR value is the highest. The network produces high-resolution images with the greatest possible performance and optimal performance.
Based on the above set, expand f 1   = 11, f 2 = 7, and compare other parameters unchanged. The result is 25.91 dB, which is only 0.08 dB higher than the previous result. However, after expanding the convolution kernel’s size, the network’s training parameters increase greatly.
Set the size of the recursive layer convolution kernels to 3 × 3 , 5 × 5 , and 9 × 9 for DRCN. Table 2 shows that when the kernel size is 3 × 3 , the network can get the best performance of 27.29 dB.
SRGAN’s generator is composed of five residual blocks with the same structure. In each residual block, the size of the kernel in the convolution layer is set to 3 × 3 , 5 × 5 , and 9 × 9 , respectively. Table 3 shows that when the kernel is 3 × 3 , the PSNR value is greater, resulting in 29.24 dB.
Based on the pre-training experimental findings mentioned above, the improved method sets all convolutional layers in the attention feature extraction module of the generator to 64, the size of the convolution kernel is 3 × 3 , and the up-sampling network utilizes a 3 × 3 size convolution kernel. The number of attention feature extraction modules are set to 4 (4 × magnification). The Adam optimizer’s learning rate for network training is set to 1 × 10 4 .

3.2. Experimental Evaluation

3.2.1. Evaluation Index

The evaluation index is to achieve an evaluation result consistent with human visual perception as much as possible. The major method is a full-reference evaluation based on the direct difference between all pixels in the image. Therefore, the first step is to get the data for all the pixels, and the second step is to diff all the pixels, find the differences between the images, and sort them to evaluate the quality of the images. Mean Square Error (MSE), PSNR, and SSIM are the most frequently utilized full-reference assessment techniques. PSNR and SSIM are frequently used as assessment indicators in image super-resolution reconstruction.
(1) PSNR.
The PSNR is the ratio of a signal’s maximum possible power to the destructive noise power that impacts its accuracy. Equation (16) defines PSNR as follows:
P S N R = 10 l n L 2 M S E
The meaning of L refers to the maximum grayscale an image can represent, and different images have different L values. If an eight-bit binary system represents the pixel information of an image, then L = 255 .
(2) SSIM.
SSIM is different from the direct solution method of MSE and PSNR. SSIM introduces the specific structure in the image into the evaluation method, strips off the influence factors of different illumination and contrast between images, and evaluates the difference in image quality through specific structural differences between images. SSIM is defined in Equation (17):
S S I M = 2 μ x μ y + C 1 2 σ x y + C 2 μ 2 x + μ 2 y + C 1 σ 2 x + σ 2 y + C 2
x and y represent the image to be evaluated and the reference image, respectively. μ and σ , respectively, refer to the mean and variance between image pixels. C 1 and C 2 are constants, and they are set to C 1 = k 1 L 2 , C 2 = k 2 L 2 , where k 1 = 0.01 , k 2 = 0.03 , L = 255 . The maximum value of SSIM is 1, which signifies that the image to be assessed is precisely the same as the original image. When the estimated value of SSIM approaches 1, it indicates that the similarity between the images is rising, while when it approaches 0, it indicates that the images are becoming increasingly dissimilar.

3.2.2. Performance Comparison and Analysis

Bicubic [26], SRCNN [24], DRCN [25], SRGAN [27], and the improved method are compared. Firstly, the convergence speed and final accuracy of different networks are compared to demonstrate the performance advantages of the innovative model. The PSNR values of various epoch outcomes during the training process of various approaches are compared in Figure 7. The improved approach provides the quickest convergence time and the highest final PSNR value. It also demonstrates that the new approach has a better level of training stability.
Secondly, the parameter complexities of various networks are compared. Figure 8 provides a comparison of network parameters and method performance. Under the same test set, the improved method’s parameters are about 1.1 × 10 6 . Compared with the SRGAN method, the parameters of the improved method are reduced by about 0.65 × 10 6 . However, the PSNR value of the image reconstructed by the improved method is 2.01 dB higher. Compared with the SRCNN and DRCN methods, although the improved method has more network parameters, the number of network layers has increased significantly, which also improves the final PSNR value compared to the shallow network.
Then, the experiment adopts a fixed four-times image magnification, and tests are carried out on Set5, Set14, BSDS100 and self-built pavement crack test data, respectively. The final performance results of different methods are shown in Table 4. On the self-built fracture data, PSNR and SSIM reach 29.21 and 0.854, respectively. In addition, the improved method has the best performance under different datasets.
Real-world photos of pavement cracks are used to evaluate the improved approach. The enhanced method’s PSNR value is 27.47 dB, and the SSIM value is 0.865, which are 0.54 and 0.018 higher than SRGAN, respectively, as shown in Figure 9. The enhanced approach restores the texture features of the pavement better, and the pavement cracks are also sharper and clearer.

3.3. Evaluation Based on Faster-RCNN and FCN

The novel method’s main goal is to provide technical support for pavement crack identification. Therefore, the detection and segmentation effect of pavement image super-resolution reconstruction is the key to evaluating the improved method, which directly proves its reliability and stability in the application field.
The F1 index is utilized as a comprehensive evaluation indicator to evaluate the accuracy of the crack detection and segmentation results in this case. To calculate the F1 value, precision and recall must be obtained ahead of time. In Equations (18)–(20), they must employ four types of pixel data for segmentation results: true positive (TP), true negative (TN), false positive (FP), and false negative (FN).
p r e c i s i o n = T P T P + F P
r e c a l l = T p T p + F N
F 1 = 2 × p r e c i s i o n     r e c a l l p r e c i s i o n + r e c a l l

3.3.1. Pavement Crack Detection Evaluation

The basic structure of the pavement crack detection method depends on the Faster R-CNN widely used in the field of image detection [28]. The model is shown in Figure 10, and is mainly composed of four parts: Conv layers, Region Proposal Networks (RPN), Region of Interest (RoI) Pooling, and Classifier.
This network inputs the image into the Conv layers. The size of the input image is uniformly adjusted to M × N . Through the VGG16-based network for feature extraction. The network comprises 13 convolutional layers, 13 ReLU layers and four pooling layers. After feature extraction, the feature map is output. Feature maps will be input into the RPN network. After convolution by a 3 × 3 convolution kernel, it is divided into two branches. The upper branch classifies the candidate frame through Softmax to obtain the detection target and background. The lower branch is used to calculate the offset of the candidate frame to adjust the accurate position of the candidate frame (Proposal). Immediately after that, the Feature map and proposal output from the previous step will be input into RoI Pooling simultaneously. The main function of RoI Pooling is to adjust to a unified feature map size. Finally, the Classifier judges the feature map output by RoI Pooling and adjusts the position and size of the detection frame again to achieve a more accurate result.
Table 5 shows the precision, recall and F1 of crack images of different super-resolution methods when Confidence_threshold is 0.5 and 0.75.
The results show that when Confidence_threshold = 0.5, although the recall of the dataset after the improved method is similar to the rest of the methods, the precision reaches 91.02%, and the F1 value is as high as 0.9027. The super-resolution images obtained by the improved method maintain a high recall rate when passing through the detection network, the model detection accuracy is higher, and the comprehensive performance is better. To further improve the detection requirements, the Confidence_threshold value is increased to 0.75. Even though the precision and recall of the super-resolution image of the method in this paper are reduced by 10.29% and 16.8%, respectively, compared with the Confidence_threshold = 0.5. However, the F1 reaching 0.7868 is higher than the rest of the methods. It is verified that the improved method has more stable properties and engineering application value under higher requirements.
Three images were randomly sampled containing horizontal, longitudinal, and reticular cracks. They are processed by different super-resolution methods and then fed into the Faster-RCNN for crack detection. The detection effect and confidence are shown in Table 6.

3.3.2. Pavement Crack Segmentation Evaluation

To assess the reconstruction effect of super-resolution crack images, the most representative FCN was chosen, which has been widely utilized in image segmentation. This end-to-end network can automatically segment the cracks from the input pavement crack image, eliminating useless information and noise interference. The FCN must first adjust the size of the input image, use the convolutional layer to extract features from the image to generate a feature map, use the FC to identify pavement cracks, and finally restore the cracks to the original input image after the deconvolution layer size. The network is mainly composed of five convolutional layers, two fully connected layers and four deconvolutional layers. The specific structure is shown in Figure 11. The pavement crack image after super-resolution reconstruction is fed into the FCN network, and the effectiveness of the designed super-resolution model is evaluated by comparing crack segmentation accuracy.
For super-resolution, the self-built pavement image data is reconstructed using several methods, and the results are sent into the FCN network for segmentation. Table 7 shows the relevant F1 and SSIM indications. The improved method’s F1 value is 0.737, up 0.012 from SRGAN. The SSMI value increased by 0.0043 to 0.9945, compared to SRGAN’s 0.0.9902.
The crack segmentation effect is seen in Figure 12. After going through the FCN segmentation network, the enhanced method’s reconstructed image has a more complete and continuous result (red area). The SSIM value of the segmented image and label is 0.9945, implying that the enhanced method’s crack image is the most similar to the original label in the segmentation result.

4. Conclusions and Future Work

This research proposes a GAN-based super-resolution reconstruction method for pavement cracks. The reconstructed crack image by the improved method is compared with other methods through an objective evaluation index, crack detection and segmentation network. The main conclusions are as follows:
(1) The research adopts RDB, GRU and Attention to upgrade the nonlinear network of the original generator and remove the BN layer. The feature learning ability of the network for the crack foreground is improved, and the network learning parameters aer reduced further. In the training process of the network, the loss function based on the L1 norm is introduced to speed up the convergence of the network and make the network pay more attention to the loss of detailed features of cracks, avoiding the defect of only focusing on the improvement of PSNR and SSIM indicators but insufficient detail reconstruction.
(2) Comparing with Bicubic, the SRCNN, DRCN and SRGAN methods on Set5, Set14, BSD100 and self-built crack datasets. Compared with the previous best results, the improved method improves the PSNR of the above dataset by 2.38, 1.91, 1.38 and 0.68, respectively. SSIM increased by 0.062, 0.043, 0.036 and 0.027, respectively. The reconstructed PSNR and SSIM of the self-built fracture dataset are 29.21 and 0.854, respectively.
(3) For the actual engineering scene, the image reconstructed by the self-built crack dataset with the improved method is fed into Faster-RCNN and FCN for detection and segmentation. The comprehensive detection accuracy is 91.02%, and the F1 reaches 0.9027. The F1 and SSIM of the segmentation effect are 0.737 and 0.9945, respectively.
Currently, the training model requires that the image magnification be chosen in advance, which is inflexible in practice. As a result, future research will focus on developing a super-resolution reconstruction model applicable to any magnification. At the same time, pavement noise (shadows, uneven illumination, etc.) can also indirectly impact pavement crack detection and segmentation performance. Preserving pavement crack features and eliminating noise interference as much as possible during the super-resolution processing of images will be the focus of our next research work.

Author Contributions

Conceptualization, Z.S., B.Y. and M.D.; data curation, M.D. and L.P.; formal analysis, Z.S., B.Y. and W.L.; methodology, Z.S., B.Y. and M.D.; project administration, B.Y. and M.D.; resources, B.Y., X.H. and W.L.; funding acquisition, Z.S.; investigation, B.Y., L.P. and X.H.; validation, B.Y., W.L., X.H. and Z.S.; writing—original draft preparation, B.Y., Z.S. and L.P.; writing—review and editing, B.Y., Z.S., L.P., W.L., M.D. and X.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research is funded by the National Key Research and Development Program, “Comprehensive Transportation and Intelligent Transportation”(grant number: 2018YFB1600202); National Natural Science Foundation of China (grant number: 51978071); Key R&D Projects in Shaanxi Province (2022JBGS3-08); The Fundamental Research Funds for the Central Universities, CHD (grant number: 300102242901).

Data Availability Statement

Some or all data, models, or code that support the findings of this study are available from the corresponding author upon reasonable request. The images from all sample sets used in this paper can be obtained from https://github.com/juhuyan/CrackDataset_DL_HY accessed on 10 October 2022 and http://www.cs.toronto.edu/~kriz/cifar.html. The modes and code used in this paper can be obtained from the corresponding authors after publication.

Acknowledgments

The authors would like to express sincere appreciation to the students in the Electronic Information Lab for taking photos so we could build good data sets and make the data so readily available. Meanwhile, the authors thank the anonymous reviewers for providing valuable comments.

Conflicts of Interest

No competing interest were disclosed.

References

  1. Chen, Y.; Liu, L.; Phonevilay, V.; Gu, K.; Xia, R.; Xie, J.; Zhang, Q.; Yang, K. Image super-resolution reconstruction based on feature map attention mechanism. Appl. Intell. 2021, 51, 4367–4380. [Google Scholar] [CrossRef]
  2. Ju, H.; Li, W.; Tighe, S.; Xu, Z.; Zhai, J. CrackU-net: A novel deep convolutional neural network for pixelwise pavement crack detection. Struct. Control Health Monit. 2020, 27, e2551. [Google Scholar]
  3. Cai, Z.; Xiong, Z.; Xu, H.; Wang, P.; Li, W.; Pan, Y. Generative adversarial networks: A survey toward private and secure applications. ACM Comput. Surv. (CSUR) 2021, 54, 1–38. [Google Scholar] [CrossRef]
  4. Alankrita, A.; Mittal, M.; Battineni, G. Generative adversarial network: An overview of theory and applications. Int. J. Inf. Manag. Data Insights 2021, 1, 100004. [Google Scholar]
  5. Kammoun, A.; Slama, R.; Tabia, H.; Ouni, T.; Abid, M. Generative Adversarial Networks for face generation: A survey. ACM Comput. Surv. (CSUR) 2022. [Google Scholar] [CrossRef]
  6. Lu, Y.; Chen, D.; Olaniyi, E.; Huang, Y. Generative adversarial networks (GANs) for image augmentation in agriculture: A systematic review. Comput. Electron. Agric. 2022, 200, 107208. [Google Scholar] [CrossRef]
  7. Singh, N.K.; Raza, K. Medical image generation using generative adversarial networks: A review. Health Inform. A Comput. Perspect. Healthc. 2021, 932, 77–96. [Google Scholar]
  8. AlAmir, M.; AlGhamdi, M. The Role of generative adversarial network in medical image analysis: An in-depth survey. ACM Comput. Surv. (CSUR) 2022. [Google Scholar] [CrossRef]
  9. Liu, M.-Y.; Huang, X.; Yu, J.; Wang, T.-C.; Mallya, A. Generative adversarial networks for image and video synthesis: Algorithms and applications. Proc. IEEE 2021, 109, 839–862. [Google Scholar] [CrossRef]
  10. Abdallah, M.S.; Cho, Y.-I. Virtual Hairstyle Service Using GANs & Segmentation Mask (Hairstyle Transfer System). Electronics 2022, 11, 3299. [Google Scholar]
  11. Pei, L.; Sun, Z.; Xiao, L.; Li, W.; Sun, J.; Zhang, H. Virtual generation of pavement crack images based on improved deep convolutional generative adversarial network. Eng. Appl. Artif. Intell. 2021, 104, 104376. [Google Scholar] [CrossRef]
  12. Nie, D.; Shen, D. Adversarial Confidence Learning for Medical Image Segmentation and Synthesis. Int. J. Comput. Vis. 2020, 128, 2494–2513. [Google Scholar] [CrossRef] [PubMed]
  13. He, Y.-L.; Li, X.-Y.; Ma, J.-H.; Lu, S.; Zhu, Q.-X. A novel virtual sample generation method based on a modified conditional Wasserstein GAN to address the small sample size problem in soft sensing. J. Process Control. 2022, 113, 18–28. [Google Scholar] [CrossRef]
  14. Toshpulatov, M.; Lee, W.; Lee, S. Generative adversarial networks and their application to 3D face generation: A survey. Image Vis. Comput. 2021, 108, 104119. [Google Scholar] [CrossRef]
  15. Han, Z.; Huang, H. Gan based three-stage-training algorithm for multi-view facial expression recognition. Neural Process. Lett. 2021, 53, 4189–4205. [Google Scholar] [CrossRef]
  16. Wang, X.; Liu, H.; Li, L.; Zhang, Y. Dual adversarial learning-based virtual sample generation method for data expansion of soft senors. Measurement 2022, 198, 111409. [Google Scholar] [CrossRef]
  17. Dewi, C.; Chen, R.-C.; Liu, Y.-T.; Tai, S.-K. Synthetic Data generation using DCGAN for improved traffic sign recognition. Neural Comput. Appl. 2021, 34, 21465–21480. [Google Scholar] [CrossRef]
  18. Zhang, D.; Ning, Z.; Yang, B.; Wang, T.; Ma, Y. Fault diagnosis of permanent magnet motor based on DCGAN-RCCNN. Energy Rep. 2022, 8, 616–626. [Google Scholar] [CrossRef]
  19. Zhang, H.; Zhu, T.; Chen, X.; Zhu, L.; Jin, D.; Fei, P. Super-resolution generative adversarial network (SRGAN) enabled on-chip contact microscopy. J. Phys. D Appl. Phys. 2021, 54, 394005. [Google Scholar] [CrossRef]
  20. Kabiraj, A.; Pal, D.; Ganguly, D.; Chatterjee, K.; Roy, S. Number plate recognition from enhanced super-resolution using generative adversarial network. Multimed. Tools Appl. 2022, 1–17. [Google Scholar] [CrossRef]
  21. Rashid, S.I.; Shakibapour, E.; Ebrahimi, M. Single MR Image Super-Resolution using Generative Adversarial Network. arXiv 2022, arXiv:2207.08036. [Google Scholar]
  22. Xu, Z.; Sun, Z.; Huyan, J.; Li, W.; Wang, F. Pixel-level pavement crack detection using enhanced high-resolution semantic network. Int. J. Pavement Eng. 2021, 1–15. [Google Scholar] [CrossRef]
  23. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; IEEE: New York, NY, USA, 2016. [Google Scholar]
  24. Mohan, A.; Dwivedi, R.; Kumar, B. Image Restoration of Landslide Photographs Using SRCNN. Recent Trends Electron. Commun. 2022, 777, 1249–1259. [Google Scholar]
  25. Park, S.; Gach, H.M.; Kim, S.; Lee, S.J.; Motai, Y. Autoencoder-inspired convolutional network-based super-resolution method in MRI. IEEE J. Transl. Eng. Health Med. 2021, 9, 1–13. [Google Scholar] [CrossRef] [PubMed]
  26. Yuan, S.; Abe, M.; Taguchi, A.; Kawamata, M. High accuracy bicubic interpolation using image local features. IEICE Trans. Fundam. Electron. Commun. Comput. Sci. 2007, 90, 1611–1615. [Google Scholar] [CrossRef]
  27. Ai, J.; Fan, G.; Mao, Y.; Jin, J.; Xing, M.; Yan, H. An Improved SRGAN Based Ambiguity Suppression Algorithm for SAR Ship Target Contrast Enhancement. IEEE Geosci. Remote Sens. Lett. 2021, 19, 8017505. [Google Scholar] [CrossRef]
  28. Sun, Z.; Pei, L.; Li, W.; Hao, X.; Chen, Y. Pavement encapsulation crack detection method based on improved Faster R-CNN. J. South China Univ. Technol. (Nat. Sci. Ed.) 2020, 48, 84–93. [Google Scholar]
Figure 1. The structure of the attention generative adversarial network.
Figure 1. The structure of the attention generative adversarial network.
Sensors 22 09092 g001
Figure 2. Generator structure with attention recursive network.
Figure 2. Generator structure with attention recursive network.
Sensors 22 09092 g002
Figure 3. Attention recursive network structure design.
Figure 3. Attention recursive network structure design.
Sensors 22 09092 g003
Figure 4. Attention map at different stages.
Figure 4. Attention map at different stages.
Sensors 22 09092 g004
Figure 5. Discriminator structure.
Figure 5. Discriminator structure.
Sensors 22 09092 g005
Figure 6. Training dataset.
Figure 6. Training dataset.
Sensors 22 09092 g006
Figure 7. Comparison of different super-resolution methods.
Figure 7. Comparison of different super-resolution methods.
Sensors 22 09092 g007
Figure 8. Comparison of network parameters and PSNR.
Figure 8. Comparison of network parameters and PSNR.
Sensors 22 09092 g008
Figure 9. Reconstruction results comparison.
Figure 9. Reconstruction results comparison.
Sensors 22 09092 g009
Figure 10. Pavement crack detection network structure.
Figure 10. Pavement crack detection network structure.
Sensors 22 09092 g010
Figure 11. Pavement crack segmentation network structure.
Figure 11. Pavement crack segmentation network structure.
Sensors 22 09092 g011
Figure 12. Pavement cracks segmentation comparison.
Figure 12. Pavement cracks segmentation comparison.
Sensors 22 09092 g012
Table 1. Comparison of PSNR values of SRCNN reconstruction. Results with different numbers of convolution kernels.
Table 1. Comparison of PSNR values of SRCNN reconstruction. Results with different numbers of convolution kernels.
n1n2PSNR (dB)
321625.17
643225.83
1286425.20
Table 2. PSNR values of DRCN with different kernels.
Table 2. PSNR values of DRCN with different kernels.
Convolution Kernel SizePSNR (dB)
3 × 327.29
5 × 526.86
9 × 927.05
Table 3. Comparison of PSNR values of SRGAN reconstruction results with different convolution kernel sizes.
Table 3. Comparison of PSNR values of SRGAN reconstruction results with different convolution kernel sizes.
Convolution Kernel SizePSNR (dB)
3 × 329.24
5 × 528.59
9 × 928.94
Table 4. PSNR and SSIM comparison on different data sets.
Table 4. PSNR and SSIM comparison on different data sets.
MethodEvaluation IndexDataset
Set5Set14BSD100Self-Built Crack Dataset
Bicubic [26]PSNR (dB)24.7222.5624.9921.68
SSIM0.6950.6310.7040.605
SRCNN [24]PSNR (dB)25.8323.1625.4922.63
SSIM0.8110.7260.7850.687
DRCN [25]PSNR (dB)27.2927.0226.6626.27
SSIM0.8320.8100.7940.759
SRGAN [27]PSNR (dB)29.2427.4628.8128.53
SSIM0.8410.8220.8360.827
Improved methodPSNR (dB)31.6229.3730.1929.21
SSIM0.9030.8650.8720.854
Table 5. Self-built dataset pavement crack detection matrix.
Table 5. Self-built dataset pavement crack detection matrix.
MethodConfidence_ThresholdPrecision (%)F1Recall (%)
Bicubic0.576.740.768276.91
SRCNN0.582.970.858288.84
DRCN0.584.680.864388.26
SRGAN0.587.920.883888.84
Improved method0.591.020.902789.53
Bicubic0.7563.770.653166.92
SRCNN0.7574.030.766879.53
DRCN0.7577.750.689061.87
SRGAN0.7579.200.769974.90
Improved method0.7580.730.786876.73
Table 6. Comparison of crack detection results and confidence.
Table 6. Comparison of crack detection results and confidence.
MethodHorizontalLongitudinalReticular
Benchmark (HR) Sensors 22 09092 i001 Sensors 22 09092 i002 Sensors 22 09092 i003
Confidence: 0.981Confidence: 0.964Confidence: 0.972
Bicubic Sensors 22 09092 i004 Sensors 22 09092 i005 Sensors 22 09092 i006
Confidence: 0.783Confidence: 0.892,0.905Confidence: 0.865
SRCNN Sensors 22 09092 i007 Sensors 22 09092 i008 Sensors 22 09092 i009
Confidence: 0.819Confidence: 0.954Confidence: 0.916
DRCN Sensors 22 09092 i010 Sensors 22 09092 i011 Sensors 22 09092 i012
Confidence: 0.825Confidence: 0.934Confidence: 0.962,0.985
SRGAN Sensors 22 09092 i013 Sensors 22 09092 i014 Sensors 22 09092 i015
Confidence: 0.889Confidence: 0.979Confidence: 0.971
Improved Method Sensors 22 09092 i016 Sensors 22 09092 i017 Sensors 22 09092 i018
Confidence: 0.947Confidence: 0.991Confidence: 0.976
Table 7. Comparison of crack segmentation F1 value and SSIM value of different methods.
Table 7. Comparison of crack segmentation F1 value and SSIM value of different methods.
MethodF1SSIM
Bicubic0.7030.9823
SRCNN0.7110.9835
DRCN0.7230.9840
SRGAN0.7250.9902
Improved method0.7370.9945
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yuan, B.; Sun, Z.; Pei, L.; Li, W.; Ding, M.; Hao, X. Super-Resolution Reconstruction Method of Pavement Crack Images Based on an Improved Generative Adversarial Network. Sensors 2022, 22, 9092. https://doi.org/10.3390/s22239092

AMA Style

Yuan B, Sun Z, Pei L, Li W, Ding M, Hao X. Super-Resolution Reconstruction Method of Pavement Crack Images Based on an Improved Generative Adversarial Network. Sensors. 2022; 22(23):9092. https://doi.org/10.3390/s22239092

Chicago/Turabian Style

Yuan, Bo, Zhaoyun Sun, Lili Pei, Wei Li, Minghang Ding, and Xueli Hao. 2022. "Super-Resolution Reconstruction Method of Pavement Crack Images Based on an Improved Generative Adversarial Network" Sensors 22, no. 23: 9092. https://doi.org/10.3390/s22239092

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop