Next Article in Journal
Preliminary Study on the Behaviour of Fibre-Reinforced Polymer Piles in Sandy Soils
Next Article in Special Issue
Load versus Strain Relationships of Single and Continuous Span Full-Scale Pre-cast Prestressed Concrete Girders for Monorail Systems
Previous Article in Journal
On the Seismic Behavior of Masonry Infilled Frame Structures
Previous Article in Special Issue
Research on Seismic Performance of Frame Structure with Beam Staircases
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Image-Based Crack Detection Method for FPSO Module Support

1
School of Ocean Science and Technology, Dalian University of Technology, Dalian 124221, China
2
Faculty of Infrastructure Engineering, Dalian University of Technology, Dalian 116024, China
*
Author to whom correspondence should be addressed.
Buildings 2022, 12(8), 1147; https://doi.org/10.3390/buildings12081147
Submission received: 7 July 2022 / Revised: 23 July 2022 / Accepted: 30 July 2022 / Published: 1 August 2022
(This article belongs to the Special Issue Structural Health Monitoring)

Abstract

:
Floating Production Storage and Offloading (FPSO) is essential offshore equipment for developing offshore oil and gas. Due to the complex sea conditions, FPSOs will be subjected to long-term alternate loads under some circumstances. Thus, it is inevitable that small cracks occur in the upper part of the module pier. Those cracks may influence the structure’s safety evaluation. Therefore, this paper proposes a method for the FPSO module to support crack identification based on the PSPNet model. The main idea is to introduce an attention mechanism into the model with Mobilenetv2 as the backbone of the PSPNet, which can fuse multiple feature maps and increase context information. The detail feature loss caused by multiple convolutions and compressions in the original model was solved by applying the proposed method. Moreover, the attention mechanism is introduced to enhance the extraction of adequate information and suppress invalid information. The mPA value and MIoU value of the improved model increased by 2.4% and 1.8%, respectively, through verification on FPSO datasets.

1. Introduction

A crack is a common form of structural damage that will lead to a decrease in the overall strength of the structure, with stress concentrating at both ends of the crack. A welding crack is a common defect in welding parts specifically. Under the joint action of welding stress and other brittle factors, the bonding force of metal atoms in the local area of the welded joint is destroyed and a new interface is formed by the gap. Generally, small cracks will not cause safety hazards in a short period. Yet, if the equipment is exposed to cracks in harsh conditions for a long time, cracks may expand and cause structural damage. This paper focuses on Floating Production Storage and Offloading (FPSO) cracks, which are specifically welding cracks. Those occur in the welding process and have a certain incubation period. With the discovery of large numbers of deep-sea, associated, and marginal oil and gas fields, and as we move away from fixed oil and gas exploration platforms, FPSO cracks are becoming increasingly prominent thanks to the low investment, short construction periods, and reusable materials applied to harvest the new fields. As such, FPSO cracks are coming to play an increasingly important role in the field of offshore engineering. When working in the offshore alternating load environment, the fatigue life of the upper pier module directly affects the service life of FPSO [1].
According to the results of a recent 5-year special comprehensive inspection of the upper pier module of FPSO, 44 of the 84 support piers of a ship were found to have cracks in the weld joints. Cracks were found in 71 out of 336 welds connected to the main deck, accounting for 21.1% of the total number of welds.
When FPSOs are subjected to wave action, the main load is created by the vertical shear force, torsional moment, and vertical bending moment. The wave causes buoyancy to change along the length of the ship, resulting in a total longitudinal moment, in which case, there will be mid-arch and mid-sag bends along the length of the ship. When the wavelength is equal to or close to the length, the total longitudinal moment of the wave is most significant, and the total longitudinal strength of the ship hull of FPSO is determined by the sagging state. According to the statistics on marine accidents in ship and ocean engineering, structural fatigue damage is the main cause of marine accidents. Seventy percent of the damage to large ships with captains over 200 m is fatigue damage. With fatigue damage to the hull structure, the focus is on the mutation area and stress concentration of the connection part of the component, namely, the fatigue of the connection node and high-stress area of the structure. FPSO is subjected to alternating stresses of the wave load for its whole life. In cases of strong wind and waves, cracks have been produced at the weld joint between the upper module support pier and the main deck. In the case of arch bends, the deck may be torn, resulting in an oil tank leakage accident. Figure 1 shows a crack distribution diagram for the upper module support pier.
FPSO is an important part of the offshore oil industry. Figure 2 shows a diagram of a crack in the upper module. If manual inspection is to be used to detect cracks, this usually requires the cooperation of engineers and technicians, project management personnel, ocean engineering experts, safety experts, and other relevant personnel. When human eyes detect cracks, it is easy to introduce subjective factors, and there is a residual phenomenon, which is inefficient. Since neural networks have undergone many innovations in recent years, a crack-recognition method based on deep learning has come to be widely used, providing an efficient method for structural health detection.

2. Related Work

In early research, computer vision was applied to automatically detect cracks through image processing. Image processing [2] mainly used edge detection and segmentation technology. Cheng et al. [3] proposed a threshold segmentation algorithm based on reducing the sample space. Song et al. [4] proposed a method to detect surface cracks based on the adaptive Canny algorithm and iterative threshold segmentation algorithm.
With the continuous development of computer science, machine learning methods are gradually coming to be applied to image processing. In 2012, Hinton’s team participated in the ImageNet image-recognition competition and won the championship by establishing AlexNet [5], proving the potential of deep learning and its Convolutional Neural Network (CNN). Since then, CNN has entered the industrial world. Farooq et al. [6] carried out experiments with different material properties and loading conditions and then simulated a healthy structure and two damaged structures with one and two small cracks, respectively. For binary classification of structures with or without damage, the prediction accuracies of neural networks and support vector machines were 93.2% and 96.66%, respectively. The average prediction accuracies of neural networks and support vector machines were 83.5% and 90.05%, respectively, for multi-class classification problems. In 2014, Ross et al. [7] improved the CNN algorithm, used it to generate candidate regions for the image, used the deep convolution network to extract features for each region, and applied the SVM algorithm to classify and judge the features. Finally, the R-CNN algorithm was proposed, which greatly improved the efficiency of target detection. The algorithm [8] only focuses on important information and ignores the rest. In 2015, the Visual Geometry Group of Oxford University [9] proposed the VGG algorithm, which improved the depth and accuracy of neural networks by using 3 × 3 convolution kernels. In the same year, Redmon proposed [10] the YOLO algorithm; it is different from the R-CNN algorithm-CNN algorithm, which has high accuracy but a slow speed, while the YOLO algorithm sacrifices accuracy for the detection speed. Recently, the fifth-generation YOLO algorithm has achieved a balance between speed and accuracy. The ResNet proposed by He [11] makes the existence of ultra-high-depth neural networks possible. It is based on the VGG network, and the algorithm adds residual cells for modification through a short circuit mechanism, as well as adding a short circuit connection between every two layers to enable residual learning. At present, most mature solutions in various fields of artificial intelligence are based on the VGG, YOLO, and R-CNN algorithms. In 2017, through further innovation of the ResNet algorithm, the DenseNet algorithm [12] was born. Then came the lightweight neural network based on mobile devices. Transformer [13] also came into being. It abandoned the traditional CNN and RNN; in place of those, the entire network structure is composed of attention mechanisms. These new algorithms, updated with the latest developments of the times, have opened up a new world in the field of deep learning.
Sometimes object detection does not apply to certain objects in a scene, meaning another method of target recognition is required, that is, semantic segmentation. Chorowski et al. [14] introduced the FCN full-convolution network to image crack detection and proposed a crack FCN model that effectively reduced the error of crack detection against complex backgrounds. In the research by Hamishebahar et al. [15], the Att-Unet model was proposed. The model is improved based on the attention mechanism of a full-convolution neural network, and it can realize end-to-end pixel-level crack segmentation. Zheng et al. [16] proposed an algorithm based on SegNet and separable convolution of bottleneck depth with residuals, which can be used for high-precision light bridge concrete crack detection. Kang and Cha [17] proposed a new network called STRNet for real-time segmentation of pixel-level cracks in complex scenes. Li et al. [18] proposed a crack-detection method consisting of two stages, which can support effective detection based on large-scale collected images. The first stage is crack recognition, in which crack images are recognized by finetuning the VGG16 model. In the second stage, UNet++ is used as a model [19] for pixel-level crack semantic segmentation of images. Fan et al. [20] proposed a new depth residual convolution neural network to establish a high-performance pavement crack-detection system.
In the field of structural health detection, the CNN model has also been fully applied. Chen et al. [21] proposed high-dimensional phase space reconstruction with a CNN model, converting the high-dimensional reconstructed attractors to images and feeding them into a CNN model. Pathak [22] proposed using CNN to monitor ridge health, and encoding methods are promising for future studies of CNN-based seismic damage evaluation. Zhao et al. [23] used a conventional and metaheuristic-tuned artificial neural network to predict the compressive strength of manufactured-sand concrete.

3. Methods

3.1. Baseline PSPNet Model

The PSPNet model is based on the FCN model and adopts a residual network to extract features. This model puts forward a Pyramid Scene Parsing module, which has been greatly improved for image segmentation. The network used in this paper is a modified feature-extraction network based on the PSPNet model. The original ResNet was composed of conv1_x, conv2_x, conv3_x, conv4_x, and conv5_x, among which conv1_1, conv2_1, conv3_1, conv4_1, and conv5_1 were downsampled once, respectively, giving a total of five instances of downsampling. The output feature map is 1/32 of the input size. In PSPNet, the ResNet module applies a dilated convolution, where conv1_x, conv2_x, and conv3_x remain unchanged. The stride of the first 1 × 1 convolution in conv4_1 is changed from 2 to 1, and the 3 × 3 convolution in conv4_x adopts dilated convolution with sampling rates of 2. The stride of the first 1 × 1 convolution in conv5_1 is changed from 2 to 1, and the 3 × 3 convolution in conv5_x adopted an expansion convolution with a sampling rate of 4. In this way, the downsampling times of the ResNet module when applying the expansion convolution are reduced by two to three times, and the output characteristic graph is one-eighth the input size; at the same time, the sensibility field of the feature map is enlarged by expanding the convolution.

3.2. Comparison of MobileNet and ResNet Structures

The main ideas of MobileNetv2 are deeply separable convolution, a linear bottleneck layer, and inverted residual [24]. Deeply separable convolution can greatly reduce the parameters of the model. Feature extraction and combination of the ordinary convolution layer are completed and outputted in one, while deeply separable convolution first uses a 3 × 3 convolution kernel with a thickness of 1, then uses a 1 × 1 convolution kernel (pointwise convolution) to adjust the numbers of channels, to separate feature extraction from feature combination. An inverse residual structure is the projection convolution of modules connected with residuals to raise the dimension, followed by depthwise convolution, and finally, using projection convolution to reduce the dimension. In MobileNetv1, the parameters of the depth convolution kernel are mostly 0, which means that its convolution kernel does not play a role in feature extraction. Due to the greater numbers of input and output channels in depthwise convolution, more information can be extracted. Figure 3 shows a diagram of MobileNetv2; as can be seen, feature maps of different scales can be outputted from MobileNetv2.
MobileNetv2 is a lightweight network. It retains the residual module, but compared to the traditional residual network, the number of parameters is greatly reduced. Figure 4 presents a comparison of the MobileNet and ResNet structures.
The ResNet module uses standard convolution to extract features, which reduces the dimension, convolves, and then increases the dimension. The MobileNet module has a depthwise convolution function, so the process of extracting features is to increase the dimension, convolve, and then reduce the dimension. Visually, ResNet has an hourglass microstructure, while MobileNetv2 has a rotating-type microstructure. Accordingly, the structure of MobileNetv2 is called a residual block.
The important module in PSPNet is the Pyramid Scene Parsing module. The pyramid pooling module combines four features with different scales: 1 × 1, 2 × 2, 3 × 3, and 6 × 6. Features of different depths are pooled at different scales according to the input, and the feature dimension is reduced to one-quarter of the original through the 1 × 1 convolution layer. Finally, these pyramid features are directly upsampled to the same size as the input features, then combined with the input features to obtain the final output features. The process of feature merging involves fusing the detail feature (shallow feature) and global feature (deep feature) of the target. Figure 5 shows a structure diagram of PSPNet. Figure 6 shows a PSPNet structure diagram with Mobile as the backbone. The difference compared to the original PSPNet is that MobileNetv2 is used as the backbone, and the feature maps at six different scales generated by MobileNetv2 are used as the input.

3.3. Improved PSPNet Model

The main idea of an attention mechanism is to enhance useful information and suppress useless or minor information. This paper introduces an attention mechanism to the Mobilev2-SSD model, applying the SENet model to introduce attention between channels. Figure 7 shows the SE block structure.
In this paper, on the premise of converting the PSPNet backbone network into MobileNetv2, six feature maps of different scales are extracted, which can aggregate the information of different feature maps and ensure that the network can extract richer semantic information. Since the original network undergoes various convolution and compression operations, the original feature information is lost, and so this method is beneficial for the model to extract fracture texture features and edge information. The five different feature maps were first passed through the attention mechanism network SENet to increase the weight of important information; accordingly, the weight of unimportant information was suppressed before entering the Pyramid Scene Parsing module. Figure 8 shows the architecture of the proposed model.

4. Data Preparation

4.1. Datasets Build

An FPSO dataset was collected from cracks of the upper module support pier. Figure 9 shows part of the FPSO dataset. As a total of 336 welds connecting the module support pier to the main deck were detected manually, where a total of 71 welds were reported to have cracks. The detection of these cracks was time-consuming and inefficient, so the collected photos were arranged, and 272 photos were selected as the dataset.
Labelme is an open-source annotation tool mainly for semantic segmentation class datasets. Labelme can annotate image data in various forms. Labelme stores annotation information in JSON files. In this work, we used the software to create a semantic segmentation class annotation of the FPSO dataset. The crack data annotation is shown in Figure 10.

4.2. Data Enhancement

To avoid issues associated with insufficient data and enhance the generalization capability of the model, methods of data enhancement to increase the image data were adopted. The specific methods include flipping, cropping, scaling, Gaussian noise, data loss, brightness adjustment, Gaussian blur, etc. The original image can usefully be flipped horizontally or vertically. It may be cropped in the following ways: random crop, target area crop, and split crop. The original image may also be rotated. Zoom enables the researcher to reduce or enlarge the original image. Gaussian noise adds pixel noise with a normal distribution to the image. Data loss removes image information of a predefined size in a randomly located area. Brightness adjustment tweaks the lighting of the original image. Gaussian blurring uses Gaussian kernel smoothing for each pixel of the image to obtain information about each nearby pixel.
In this step of our work, each image in the dataset was randomly selected for data enhancement by four of the above methods. One or more of the above methods was randomly selected to expand the data volume. Figure 11 presents details of the data enhancement effects.

5. Training

5.1. Loss Function

The loss function used in this paper consists of two parts: one is the classification error generated by classification using Cross-Entropy Loss [25], and the other is the segmentation error generated by pixel segmentation using Dice Loss [26]. Cross-Entropy Loss is used when the network applies the SoftMax function to classify pixels. In the case of binary classification, the model only needs to predict two cases, and the predicted probabilities are P and 1-P. The expression is as follows:
L = 1 N i L i = 1 N i y i log p i + 1 y i log 1 p i
where y i represents the label of sample I, with positive class 1 and negative class 0. p is the probability that sample I is positive. In the case of multiple categories, the dichotomies are extended as follows:
L = 1 N i L i = 1 N i c = 1 M y i c log p i c
where M refers to the number of categories, and y i c is the indicator variable (0 or 1). If the category is the same as sample I, it is 1, and if the category is different, it is 0, which refers to the predicted probability of sample I belonging to class C.
The similarity measure coefficient (dice) is used to calculate the similarity of the two samples, where the range is [0, 1]. The calculation formula of the similarity measure function is as follows:
dice = 2 | X Y | | X | + | Y |
where X is the predicted result and Y is the real result. The larger the result, the higher the coincidence degree between the predicted and real results. Therefore, the closer the dice coefficient is to 1, the better. If used as an error, the loss value should be as small as possible, so the loss of semantic segmentation should be set as Dice Loss = 1 − dice.

5.2. Metrics

The confusion matrix is another comprehensive measure of computer vision models, which includes the TP, FN, FP, and TN metrics. The confusion matrix is the basis for all other types of metrics. These metrics are obtained by mathematical operations on the data in the confusion matrix. The details of the confusion matrix are shown in Table 1.
The recall metric, shown in the formula below, consists of TP and FN and represents the proportion of positive samples that are both positive and predicted to be positive.
R e c a l l = T P T P + F N
Precision, shown in the formula below, is an indicator composed of TP and FP. Generally, a specified threshold is set to determine the confidence level of image classification to determine the TP and FP of image classification; this indicator represents the proportion of positive samples correctly predicted.
P r e c i s i o n = T P T P + F P
The mean pixel accuracy and mean intersection over union are usually used as evaluation indexes in image segmentation. The higher the values of these indexes, the better the performance of the model.
Mean pixel accuracy: calculate the proportion of correctly classified pixels in each class, then calculate the average of this proportion in all classes:
m P A = 1 m + 1 i = 0 m p i i j = 0 m p i j
Mean intersection over union: calculate the ratio of the intersection and union of two sets, which are real and predicted in semantic segmentation problems. This ratio can be deformed to be the sum of true positives, true negatives, false negatives, and false negatives (unions). Calculate the IoU for each class, then take an average.
M IoU = 1 m + 1 i = 0 m p i i j = 0 m p i j + j = 0 m p j i p i i

5.3. Experimental Conditions

The division ratio of the training, validation, and test sets in this paper was 8:1:1. To prove the generalization of the model, the subsets consisting of training and validation sets, and the subset consisting of the test set, were enhanced separately. Specifically, the training set and verification set was composed of 1210 enhanced images produced from 242 of the original images. The training set then contained 1060 images and the verification set contained 150 images. The test set, meanwhile, was composed of 150 enhanced images produced from 30 of the original images.
Using the adaptive moment estimation (Adam) optimization algorithm, we found the learning rate was 0.001, β1 was 0.9, and β2 was 0.999 [27]. We set the batch size to 16. The model was trained for 350 epochs in total, and the training time was about 6 h. The experiment took about 0.33 s to process each image. Table 2 shows experimental conditions in this paper.

6. Experimental Results and Analysis

To verify the effectiveness of the proposed model, different algorithms were used for training. The prediction results of different models on the same test set are shown in Table 3 and Table 4.
MobileNetv2 as the backbone showed an excellent performance, where a basic attention mechanism was introduced to enhance the performance of the model. This is because the background of the dataset in this paper was similar to the target object, and it was easy to recognize tiny cracks in the background during semantic segmentation. After the attention mechanism was introduced before feature extraction, the ability to identify cracks and extract contextual information was enhanced, as this is effective at recognizing tiny target objects. Table 3 and Figure 12 show the four performance indicators—recall, precision, PA, and IoU—for assessment of the background and cracks with different models. It can be seen that the crack-detection performance of the improved model was improved by 3.5%, 7.78%, 3.5%, and 8.44%, respectively, compared to PSPNet, and 3.22%, 3.91%, 3.22%, and 4.43% versus Mobilev2-PSPNet.
To further verify the effectiveness of the model, we compared it to the classic semantic segmentation FCN model and the most popular semantic segmentation model at present. It can be seen that the model proposed in this paper for FPSO cracks is practical. Table 4 compares the performance levels of PSPNet, Mobilev2-PSPNet, and the improved model. It can be seen that the MioU, PA, and accuracy of the improved model were 4.63%, 2.12%, and 0.87% higher than those of the PSPNet model, respectively. Compared to the Mobilev2-PSPNet model, it was 2.4%, 1.81%, and 0.73% better. Table 5 shows the number of parameters for different models. It can be seen from the table that using MobileNetv2 as the backbone greatly reduces the parameters of the model, and the model parameters after introducing the attention mechanism increase only slightly. As such, the performance metrics of the model are improved without a significant increase in model parameters, demonstrating the significance of our improved model.
As can be seen from the chart, we used MobileNetv2 as the model backbone, and the model performance was effectively promoted. Since the original network involves multiple convolutions and compressions, it is difficult to perform effective feature extraction. Moreover, cracks and the background can be easily confused due to the low resolution and the effects of light intensity, meaning the ResNet model may not be as good as MobileNetv2 at identifying small cracks. To solve this problem, the proposed model extracts different feature maps from MobileNetv2 and fuses them together after activating the attention mechanism. Experiments show that the method is effective. All the performance indexes are improved to some extent, which indicates that the model can improve the crack identification ability through different receptive fields after fusing different feature maps.
In Figure 13, Figure 14 and Figure 15, (a) represents the loss curve of the PSPNet model, and (b) represents its MioU curve. As can be seen from the training curve, the loss gradually converges to a stable value, while MioU gradually rises to a stable state. This demonstrates that the above three models are of practical significance. Figure 16 shows the results for crack identification by different means of semantic segmentation.

7. Conclusions

For FPSO, real-time detection of cracks is an important guarantee of structural safety, and it is of great significance to know how to identify cracks using computer vision instead of manual detection and identification. In that context, the existing PSPNet model was improved in this paper. In a case where the original model performed excellently, the main trunk was changed to MobileNet and attention mechanism SENet was introduced. Training was carried out on the FPSO dataset we made, and the performance of the original model was improved. The important evaluation index values of mPA and MIoU semantic segmentation models were improved obviously, and an improved model for FPSO crack identification was obtained. The improved model has engineering significance for FPSO crack identification. The proposed method of multi-feature graph fusion enhances the feature extraction ability of the model for fine cracks and effectively integrates contextual information. In our research presented here, the MIoU and mPA of the improved model reached 83.38% and 94.49%, respectively, indicating the effectiveness of the model we propose.

Author Contributions

Conceptualization, Z.J.; methodology, X.S.; software, X.S and G.M.; validation, Z.J., X.S. and C.Q.; formal analysis, T.D.; investigation, L.R.; resources, G.M.; data curation, X.S.; writing—original draft preparation, Z.J.; writing—review and editing, C.Q.; visualization, G.M.; supervision, T.D.; project administration, L.R.; funding acquisition, Z.J. and C.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Key Research and Development Program of China grant number 2019YFB1504303, the National Natural Science Foundation of China grant number 52078100, Anhui international joint research center of data diagnosis and smart maintenance on bridge structures grant number 2022AHGHYB03 and the Fundamental Research Funds for the Central Universities grant number DUT22JC19.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No data were used to support this study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yuan, G. Research on Fatigue Characteristics and Antifatigue Design of FPSO Topside Module; Tianjin University: Tianjin, China, 2018. [Google Scholar] [CrossRef]
  2. Ye, X.-W.; Jin, T.; Chen, P.-Y. Structural crack detection using deep learning–based fully convolutional networks. Adv. Struct. Eng. 2019, 22, 3412–3419. [Google Scholar] [CrossRef]
  3. Cheng, H.D.; Shi, X.J.; Glazier, C. Real-time image thresholding based on sample space reduction and interpolation approach. J.Comput. Civ. Eng. 2003, 17, 264–272. [Google Scholar] [CrossRef]
  4. Qiang, S.; Guoying, L.; Jingqi, M.; Hongmei, Z. An Edge-Detection Method Based on Adaptive Canny Algorithm and Iterative Segmentation Threshold. In Proceedings of the 2nd International Conference on Control Science and Systems Engineering (ICCSSE), Singapore, 27–29 July 2016. [Google Scholar]
  5. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. In Proceedings of the 25th International Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA; 2012; Volume 1, pp. 1097–1105. [Google Scholar]
  6. Farooq, M.; Zheng, H.; Nagabhushana, A.; Roy, S.; Burkett, S.; Barkey, M.; Kotru, S.; Sazonov, E. Damage Detection and Identification in Smart Structures using SVM and ANN. In Proceedings of the Conference on Smart Sensor Phenomena, Technology, Networks, and Systems Integration, San Diego, CA, USA, 12–14 March 2012. [Google Scholar]
  7. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  8. Cui, X.; Wang, Q.; Dai, J.; Xue, Y.; Duan, Y. Intelligent crack detection based on attention mechanism in convolution neural network. Adv. Struct. Eng. 2021, 24, 1859–1868. [Google Scholar] [CrossRef]
  9. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015 Conference Track Proceedings (Vienna: ICLR), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  10. Joseph, R.; Santosh, D.; Ross, G.; Ali, F. Learning Deep Features for Discriminative Localization. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 December 2016; pp. 779–788. [Google Scholar]
  11. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; IEEE: Las Vegas, NV, USA, 2016. [Google Scholar] [CrossRef] [Green Version]
  12. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR, Honolulu, HI, USA, 22–25 July 2017; IEEE: Honolulu, HI, USA, 2017. [Google Scholar] [CrossRef] [Green Version]
  13. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N. Attention Is All You Need. arXiv 2017. [Google Scholar] [CrossRef]
  14. Chorowski, J.; Bahdanau, D.; Serdyuk, D. Attention-based models for speech recognition. In Proceedings of the 28th International Conference on Neural Information Processing Systems, Montreal, QB, Canada, 7–12 December 2015; Volume 1, pp. 577–585. [Google Scholar]
  15. Hamishebahar, Y.; Guan, H.; So, S.; Jo, J. A Comprehensive Review of Deep Learning-Based Crack Detection Approaches. Appl. Sci. 2022, 12, 1374. [Google Scholar] [CrossRef]
  16. Zheng, X.; Zhang, S.; Li, X.; Li, G.; Li, X. Lightweight Bridge Crack Detection Method Based on SegNet and Bottleneck Depth-Separable Convolution With Residuals. IEEE Access 2021, 9, 161649–161668. [Google Scholar] [CrossRef]
  17. Kang, D.H.; Cha, Y.-J. Efficient Attention-Based Deep Encoder and Decoder for Automatic Crack Segmentation. Struct. Health Monit. 2021. [Google Scholar] [CrossRef]
  18. Li, Y.; Zhao, W.; Zhang, X.; Zhou, Q. A Two-Stage Crack Detection Method for Concrete Bridges Using Convolutional Neural Networks. IEICE Trans. Inf. Syst. 2018, 101, 3249–3252. [Google Scholar] [CrossRef] [Green Version]
  19. Yang, Q.N.; Ji, X.D. Automatic Pixel-Level Crack Detection for Civil Infrastructure Using Unet plus and Deep Transfer Learning. IEEE Sens. J. 2021, 21, 19165–19175. [Google Scholar] [CrossRef]
  20. Fan, Z.; Lin, H.; Li, C.; Su, J.; Bruno, S.; Loprencipe, G. Use of Parallel ResNet for High-Performance Pavement Crack Detection and Measurement. Sustainability 2022, 14, 1825. [Google Scholar] [CrossRef]
  21. Chen, Y.-L.; Chiang, Y.; Chiu, P.-H.; Huang, I.-C.; Xiao, Y.-B.; Chang, S.-W.; Huang, C.-W. High-Dimensional Phase Space Reconstruction with a Convolutional Neural Network for Structural Health Monitoring. Sensors 2021, 21, 3514. [Google Scholar] [CrossRef] [PubMed]
  22. Pathak, N. Bridge Health Monitoring Using CNN. In Proceedings of the 2020 International Conference on Convergence to Digital World—Quo Vadis (ICCDW), Mumbai, India, 18–20 February 2020; pp. 1–4. [Google Scholar] [CrossRef]
  23. Yinghao, Z.; Zeyu, W. Subset Simulation with Adaptable Intermediate Failure Probability for Robust Reliability Analysis: An Unsupervised Learning-Based Approach. Struct. Multidiscip. Optim. 2022, 65, 1–22. [Google Scholar]
  24. Zou, Y.; Zhao, L.; Qin, S.; Pan, M.; Li, Z. Ship target detection and identification based on SSD_MobilenetV2. In Proceedings of the 2020 IEEE 5th Information Technology and Mechatronics Engineering Conference, Chongqing, China, 12–14 June 2020. [Google Scholar]
  25. Shore, J.; Johnson, R. Axiomatic derivation of the principle of maximum entropy and the principle of minimum cross-entropy. Inf. Theory IEEE Trans. 1980, 26, 26–37. [Google Scholar] [CrossRef] [Green Version]
  26. Wei, H.; Peng, P. The segmentation of retinal layer and fluid in SD-OCT images using mutex dice loss based fully convolutional networks. IEEE Access 2020, 8, 60929–60939. [Google Scholar] [CrossRef]
  27. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. 2014. Available online: https://arxiv.org/abs/1412.6980 (accessed on 12 May 2022).
Figure 1. Crack distribution diagram of upper module support pier. (A–D represents the line number).
Figure 1. Crack distribution diagram of upper module support pier. (A–D represents the line number).
Buildings 12 01147 g001
Figure 2. Diagram of the crack in the upper module.
Figure 2. Diagram of the crack in the upper module.
Buildings 12 01147 g002
Figure 3. Structure diagram of MobileNetv2.
Figure 3. Structure diagram of MobileNetv2.
Buildings 12 01147 g003
Figure 4. Comparison of MobileNet and ResNet structures: (a) ResNet; (b) MobileNetv2.
Figure 4. Comparison of MobileNet and ResNet structures: (a) ResNet; (b) MobileNetv2.
Buildings 12 01147 g004
Figure 5. Structure diagram of PSPNet.
Figure 5. Structure diagram of PSPNet.
Buildings 12 01147 g005
Figure 6. PSPNet structure diagram with Mobile as the backbone.
Figure 6. PSPNet structure diagram with Mobile as the backbone.
Buildings 12 01147 g006
Figure 7. SE block structure.
Figure 7. SE block structure.
Buildings 12 01147 g007
Figure 8. PSPNet architecture diagram with MobileNetv2 as backbone.
Figure 8. PSPNet architecture diagram with MobileNetv2 as backbone.
Buildings 12 01147 g008
Figure 9. FPSO datasets.
Figure 9. FPSO datasets.
Buildings 12 01147 g009
Figure 10. The crack data annotation.
Figure 10. The crack data annotation.
Buildings 12 01147 g010
Figure 11. Image enhancement example of two images: (a) Random Image Enhancement Combination 1; (b) Random Image Enhancement Combination 2.
Figure 11. Image enhancement example of two images: (a) Random Image Enhancement Combination 1; (b) Random Image Enhancement Combination 2.
Buildings 12 01147 g011aBuildings 12 01147 g011b
Figure 12. Result of training.
Figure 12. Result of training.
Buildings 12 01147 g012
Figure 13. Training curve and MIoU curve of PSPNet model.
Figure 13. Training curve and MIoU curve of PSPNet model.
Buildings 12 01147 g013
Figure 14. Training curve and MIoU curve of MobileNetv2-PSPNet model.
Figure 14. Training curve and MIoU curve of MobileNetv2-PSPNet model.
Buildings 12 01147 g014
Figure 15. Training curve and MIoU curve of the improved model.
Figure 15. Training curve and MIoU curve of the improved model.
Buildings 12 01147 g015
Figure 16. Results of crack identification by different semantic segmentation: (a) original image; (b) PSPNet; (c) MobileNetv2-PSPNet; (d) improved model.
Figure 16. Results of crack identification by different semantic segmentation: (a) original image; (b) PSPNet; (c) MobileNetv2-PSPNet; (d) improved model.
Buildings 12 01147 g016
Table 1. Confusion matrix.
Table 1. Confusion matrix.
CategoryPredicted
PositiveNegative
Real PositiveTPFN
NegativeFPTN
Table 2. Experimental conditions.
Table 2. Experimental conditions.
Memory32 GB
GPUGeForce RTX 2080 Ti
OSUbuntu 18.04
Python3.6.8
CUDA10.1
Pytorch1.7.1
Table 3. Results of different models for training.
Table 3. Results of different models for training.
Model RecallPrecisionPAIOU
PSPNetBackground97.8599.0397.8597.47
crack86.8966.2886.8960.03
MobileNet-PSPNetBackground98.1998.8398.1997.91
crack87.1770.1587.1764.04
Improved modelBackground98.5899.5298.5898.29
crack90.3974.0690.3968.47
Table 4. Performance comparison of the different models.
Table 4. Performance comparison of the different models.
ModelMioUmPAAccuracy
FCN62.6478.5189.75
SegNet71.9483.6793.32
DeepLabv379.1191.0897.96
Unet75.2988.2186.14
PSPNet78.7592.3797.87
MobileNet-PSPNet80.9892.6898.01
Improved model83.3894.4998.74
Table 5. Number of parameters for different models.
Table 5. Number of parameters for different models.
ModelNo. of Parameters
PSPNet46,706,626
MobileNet-PSPNet2,375,874
Improved model2,388,058
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Su, X.; Jia, Z.; Ma, G.; Qu, C.; Dai, T.; Ren, L. Image-Based Crack Detection Method for FPSO Module Support. Buildings 2022, 12, 1147. https://doi.org/10.3390/buildings12081147

AMA Style

Su X, Jia Z, Ma G, Qu C, Dai T, Ren L. Image-Based Crack Detection Method for FPSO Module Support. Buildings. 2022; 12(8):1147. https://doi.org/10.3390/buildings12081147

Chicago/Turabian Style

Su, Xin, Ziguang Jia, Guangda Ma, Chunxu Qu, Tongtong Dai, and Liang Ren. 2022. "Image-Based Crack Detection Method for FPSO Module Support" Buildings 12, no. 8: 1147. https://doi.org/10.3390/buildings12081147

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop