Next Article in Journal
New Coach, New Risks? Injury Trends in Elite Football After Coaching Changes During an Eight-Season Period
Previous Article in Journal
Toward More Efficient Large-Scale Green Hydrogen Systems via Waste Heat Recovery and ORC
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Crack Defect Detection Method for Plunge Pool Corridor Based on the Improved U-Net Network

1
China Yangtze Power Co., Ltd., Yichang 443002, China
2
Sichuan Energy Internet Research Institute, Tsinghua University, Chengdu 610000, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(10), 5225; https://doi.org/10.3390/app15105225
Submission received: 19 March 2025 / Revised: 29 April 2025 / Accepted: 2 May 2025 / Published: 8 May 2025

Abstract

:
Crack defects pose a significant threat to the safe operation of the foundational structures of hydropower stations. Therefore, crack detection in a plunge pool corridor is crucial for the safe operation and maintenance of hydropower hubs. Addressing challenges such as missing crack datasets, complex and diverse crack shapes, and difficulties in feature extraction, this paper proposes an improved U-Net model. The model integrates the CDIAM attention module within the U-Net architecture and is optimized through the incorporation of the MIDConv convolutional block. The enhanced U-Net model demonstrates outstanding performance on a self-constructed plunge pool corridor crack dataset, achieving an MIoU, F1-score, MPA, and accuracy of 84.44%, 93.05%, 89.75%, and 99.46%, respectively. Compared to the original U-Net model, these metrics show improvements of 5.34%, 4.90%, 3.39%, and 0.36%, respectively. The improved U-Net model exhibits higher detection accuracy in corridor crack segmentation tasks, effectively extracting feature information from irregular shapes, and provides a superior solution for crack defect detection in a plunge pool corridor, thereby enhancing the safety and efficient maintenance of hydropower stations.

1. Introduction

The corridors of hydropower stations are critical components of hydraulic hubs, directly affecting the safety and economic benefits of the hydropower station. The appearance of cracks threatens structural integrity and can lead to water leakage and equipment damage. Therefore, regular crack detection is crucial as it enables the timely identification and resolution of potential risks, ensuring the health and stability of the corridors and thereby safeguarding the safe operation and efficient maintenance of the hydropower station. Conventional manual inspection not only involves substantial consumption of human and material resources but also suffers from extremely low detection efficiency, being unable to meet the requirement of rapid detection [1]. Alongside the rapid advancement of computer science, the accurate detection of concrete cracks via computer vision has emerged as a primary research goal.
Yu et al. [2] employed the Sobel operator to detect crack edges in images, identifying genuine cracks through region filling, labeling, and adjacent region connection operations. Li et al. [3] utilized a wavelet-based algorithm to eliminate noise from images and integrated the C-V model with a Canny edge detector for crack segmentation. Through comparison with the original Canny edge detector, the proposed method outperforms the original Canny edge detector in performance. Likewise, Li et al. [4] employed the Canny edge detector to extract cracks from images and optimized crack shapes with a region-based active contour model. Subsequently, support vector machines (SVM) and a greedy search strategy were applied to eliminate noise induced by crack surface contamination and uneven lighting, ultimately yielding pixel-level crack detection results with distinct shapes. Despite extensive research on crack detection using conventional digital image processing techniques, their limitations are evident. Conventional image processing techniques depend on manual feature extraction and segmentation, restricting their applicability to specific scenarios.
In recent years, convolutional neural networks (CNN) [5] have been widely applied in various tasks, with segmentation techniques based on fully convolutional networks demonstrating exceptional performance in infrastructure inspection, such as concrete crack detection, providing strong technical support for crack recognition and monitoring. The U-Net model [6], originally developed for the medical field, along with its improved variants, has been widely applied in concrete infrastructure inspection, including residential building crack detection [7,8,9,10,11], bridge surface crack detection [12,13,14], and highway surface crack detection [15,16,17,18,19,20,21]. For instance, Ren et al. [22] integrated dilated convolution and spatial pyramid pooling modules into the deep fully convolutional neural network to enable efficient multi-scale feature extraction, aggregation, and resolution reconstruction, demonstrating high precision and generalization capability in crack detection. Aiming at complex scenarios in tunnel images, including water stains, scratches, and structural joints, QING et al. [23] developed an objective and fast recognition algorithm for tunnel cracks via computer vision-based semantic segmentation. Wooram et al. [24] developed the SDDNet network, comprising standard convolution, a DenSep module, an improved atrous spatial pyramid pooling (ASPP) module, and a decoder module. Experimental results indicate that SDDNet is highly effective for segmented crack segmentation. Liu et al. [25] opted to learn and aggregate multi-scale and multi-level features from low to high convolutional layers during training, followed by guided filtering and conditional random fields (CRFs) for model optimization, resulting in the DeepCrack deep neural network with an F1-score of 86.5%. The CrackU-Net proposed by Ju et al. [26] enhances the depth of the U-Net model and surpasses the original U-Net in pavement crack detection. To tackle the issue of crack discontinuities in segmentation outcomes, Cao et al. [27] introduced the HC-Unet++ network segmentation model for road crack detection. SUN et al. [28] incorporated a multi-scale attention module into the decoder of DeepLabv3+, which effectively reduces the instances of missed crack detections.
  • Semantic segmentation approaches are capable of achieving detailed identification of specified objects, showcasing notable advantages in concrete crack detection. Yet most current mainstream crack detection networks are tailored for scenarios like building walls, road surfaces, or bridges. Plunge pool corridors, long-term exposed to high-speed water flow erosion, thermal stress, and wet–dry cycles, exhibit concrete cracks with morphological features, spatial distributions, and data distributions that differ substantially from target objects in conventional public datasets. This makes it challenging for generic models to accurately differentiate targets from backgrounds. Additionally, the nonlinear morphology and multi-scale characteristics of cracks in plunge pool corridors render traditional fixed-size convolution kernels ineffective in capturing edge details. While techniques like deformable convolutions attempt to dynamically adjust the receptive field, existing approaches lack the exploitation of crack morphological priors, making them susceptible to feature deviation in complex backgrounds and resulting in segmentation issues such as discontinuities and missed detections.
  • To address the aforementioned research gaps, this study develops a specialized dataset for crack segmentation in plunge pool corridors via on-site data collection and presents an improved U-Net model incorporating morphology-aware dynamic convolution and cross-dimensional interactive attention. The contributions of this work are as follows:
    (1)
    Development of a scenario-specific dataset: Using an inspection robot outfitted with a multi-angle vision system, we acquired crack images of plunge pool corridors with complex backgrounds and constructed a high-quality dataset comprising 4075 images via data augmentation techniques, addressing the lack of data for plunge pool corridor scenarios.
    (2)
    Proposing morphology-integrated irregular dynamic convolution (MIDConv): By integrating the linear morphological features of cracks and constraining the offset accumulation of deformable convolutions, the receptive field of the convolution kernel is dynamically expanded along the crack orientation, effectively enhancing the ability to extract features of irregular edges.
    (3)
    Development of cross-dimensional interactive attention module (CDIAM): By partitioning channels into groups and introducing a global information calibration mechanism, semantic interaction between channel and spatial dimensions is established, enhancing the model’s capacity to model pixel-level relationships of cracks in complex backgrounds and markedly reducing misdetections and omissions.
    (4)
    Substantial enhancement in detection performance: On the self-constructed dataset, the improved model achieves an MIoU, F1-Score, MPA, and accuracy of 84.44%, 93.05%, 89.75%, and 99.46%, reflecting improvements of 5.34%, 4.90%, 3.39%, and 0.36% compared to the original U-Net, effectively resolving the problem of inadequate detection accuracy of traditional methods in complex scenarios. Additionally, to validate the generalization capability of the proposed approach, experiments were performed on two public datasets, both yielding promising results.

2. Methods

2.1. Crack Segmentation Model for Plunge Pool Corridors Based on the Improved U-Net

U-Net, proposed by O. Ronneberger in 2015, is a general-purpose semantic segmentation model with a typical symmetric encoder–decoder network architecture, which has shown outstanding performance in the field of semantic segmentation for medical images. The U-Net model performs feature extraction of target objects through four down-sampling operations and then restores the spatial resolution of the image through four up-sampling operations and skip connections, ultimately achieving the semantic segmentation task. The original U-Net model is shown in Figure 1.
The crack images in plunge pool corridors have complex backgrounds and diverse target pixel shapes. The original U-Net model uses traditional convolution, where the fixed kernel shape and size result in poor performance when processing different receptive field sizes. Additionally, using the same convolutional kernel for the entire image limits the feature extraction capability in different regions, which affects the segmentation results. The up-sampling and down-sampling structure makes the network insensitive to the precise location of the target, leading to positional deviations. To address these issues, this paper proposes the integration of morphological irregular dynamic convolution (MIDConv) to capture the morphological features of cracks and extract edge feature information. Furthermore, a cross-dimensional interaction attention module (CDIAM) is incorporated to enhance the model’s ability to extract spatial semantic information of cracks. The improved model is shown in Figure 2.

2.2. Morphological Irregular Dynamic Convolution Method

The morphological irregular dynamic convolution (MIDConv) method is a feature extraction approach, based on deformable convolution [29], aimed at extracting features according to the morphological characteristics of cracks. Deformable convolution enhances feature extraction by introducing offsets to the original convolution operation, allowing the convolutional kernel to dynamically adjust according to the input feature map, thereby improving the model’s ability to handle irregular shapes and structures. Inspired by deformable convolution, this paper integrates the morphological features of cracks into the convolution process, introducing deformable offsets to enhance the model’s ability to extract crack features. However, if the convolution kernel can freely deform and shift, the feature extraction area may gradually deviate from the target, especially in images of cracks with complex shapes. Therefore, it is necessary to restrict the introduced offsets to ensure that the feature extraction process does not deviate from the target region. The process is shown in Figure 3.
In MIDConv, the convolution kernel is straightened along the X-axis and Y-axis. Taking a convolution kernel of size 3 × 3 as an example, the coordinates of the convolution kernel along the X-axis are K i ± b = x i ± b , y i ± b , b = 0 , 1 , 2 , 3 , 4 , where b represents the horizontal distance from each convolution grid to the center grid. The selection of each grid position in the convolution kernel is progressively accumulated, starting from the central grid position of the kernel. The position of the kernel center grid is determined by the previous grid position. Compared to the central grid K i , the grid position K i + 1 is offset by a value of σ σ 1 , 1 . The offset values in different convolution kernel positions need to be continually accumulated to ensure that the convolution kernel conforms to a linear morphological structure. Therefore, the coordinates of each grid position in the convolution kernel along the X-axis can be represented as follows:
K i ± b = x i + b , y i + b = x i + b , y i + i i + b σ y x i b , y i b = x i b , y i i b i σ y
From the above equation, the coordinates of each grid position in the convolution kernel along the Y-axis can be represented as follows:
K j ± b = x j + b , y j + b = x j + b + j j + b σ x , y j x j b , y j b = x j b j b j σ x , y j
Since the offset σ σ 1 , 1 is usually not an integer, the coordinates of the convolution kernel K can be obtained through bilinear interpolation and are represented as follows:
K = K B I K , K K
where K represents the fractional part of the coordinate offset along the X-axis and Y-axis, K denotes the integer spatial coordinates of the corresponding grid point in the convolution kernel, and B I represents the bilinear interpolation operation.
As shown in Figure 4, after processing the input features using the morphological irregular dynamic convolution method, the receptive field of the features is expanded, allowing for more detailed information about the crack features to be obtained, thereby enhancing the model’s ability to extract crack features.

2.3. Cross-Dimensional Interaction Attention Module

The attention module [30] primarily enhances the model’s detection performance of target features through spatial information, channel information of the feature data, and their combination. Currently, the widely acknowledged attention modules that effectively enhance feature detection capabilities mainly consist of CBAM [31], SE [32], ECA [33], and CA [34]. In the crack segmentation task, the channel or spatial attention mechanism has been demonstrated to be highly effective in extracting crack features. Nevertheless, when handling cross-dimensional information, the approach of channel dimensionality reduction may result in information loss during the extraction of higher-dimensional features. To address this issue, this paper, based on the existing research of the attention mechanism, puts forward a cross-dimensional interaction attention module (CDIAM). This attention module remolds part of the channel information into the batch-processing dimension, divides the channel dimension into multiple sub-feature groups, and guarantees the even distribution of spatial semantic features. Meanwhile, this module utilizes global information to calibrate the weights of the channel dimension, integrates the features output from parallel branches, captures pixel-level relationships, and enhances the model’s crack feature extraction capacity and crack recognition precision. The architecture of the cross-dimensional interaction attention module is shown in Figure 5.
To enable the model to better differentiate between foreground and background pixels, the input feature map is subjected to cross-dimensional division, splitting it into N sub-feature groups. Suppose the input feature map is X C × H × W . Following cross-dimensional division, the feature map can be denoted as X = X 0 , X 1 , X 2 X N , of which X i ϵ C / / N × H × W , typically N C . Meanwhile, it is hypothesized that the learned attention weight descriptor will be employed to enhance the feature representation of the regions of interest within each sub-feature. This implies that the model can concentrate its attention on specific areas, thus enhancing the quality and significance of the feature representation in those areas. By this means, the network can capture the crucial information in the input data more efficiently, thereby enhancing the performance and expressivity of the task.
The attention module consists of three parallel branches, which are the H-direction feature map-processing branch (the first branch), the W-direction feature map-processing branch (the second branch), and the input sub-feature-group-processing branch (the third branch). After the features of average pooling and convolution in the first and second branches are concatenated, weight redistribution is carried out with the input sub-feature group, thus obtaining the output features. Subsequently, the output features are interactively integrated across dimensions with the feature map of the third branch that has undergone 3 × 3 convolution processing, and ultimately, the output feature map is acquired.
To realize diverse cross-channel feature interactions between the first and second branches, the product operation is utilized to aggregate the channel features of these two branches. Meanwhile, to capture local cross-channel interactions and broaden the feature space, the third branch conducts processing on the input features with a 3 × 3 convolution. By means of the above operations, the cross-dimensional interaction attention module is capable of not only encoding the inter-channel information to adjust the weights of various channel features but also of precisely retaining the spatial structure information within the channels. The output features in the H and W directions after going through the first and second branches are denoted, respectively, as follows:
Z c H H = 1 W 0 i W x c H , i
Z c W W = 1 H 0 j H x c j , W
Cross-dimensional interaction can well establish a relationship between the channel-position features and spatial-position features of the input feature map. To better integrate these two types of features to obtain a feature map with more abundant semantic information, cross-dimensional interaction is performed between the features processed by the first and second branches and the features of the third branch after a regular 3 × 3 convolution. Global average pooling is employed to encode the global spatial information in the output of the 1 × 1 branch. Prior to the joint activation mechanism of channel features, the output of the smallest branch is directly transformed into the corresponding dimensional form, namely, 1 1 × C / / N × 3 C / / N × HW . Two-dimensional global average pooling is expressed as the formula below:
Z c = 1 H × W J H I W x c i , j
Two-dimensional global average pooling is capable of establishing the reliance between global and dimensional information. Employing a two-dimensional Gaussian map to approximate the linear transformation of the global average pool’s output by Softmax renders the calculation more efficient. The multiplication of the parallel-processed output and the matrix dot-product operation results in the first spatial attention and integrates multi-scale spatial information. Likewise, global average pooling is employed to encode the global information of the 3 × 3 branch, while the 1 × 1 branch changes the dimensional form to 3 1 × C / / N × 1 C / / N × HW . We generate the second spatial attention map to preserve the accurate spatial location information. The output feature map represents the combination of the two spatial-attention weight values. Following Sigmoid processing, the global context is emphasized.

3. Dataset

3.1. Data Collection

The corridor crack data in this study were collected by an inspection robot fitted with a visible-light camera and a fill light. This inspection robot is equipped with a mechanical arm system, allowing the acquisition device to be mounted on the mechanical arm to achieve omnidirectional rotation and multi-position control at the transverse central position of the standard corridor. Regarding hardware parameters, the visible-light camera features a pixel resolution of 2 megapixels, with a horizontal field of view of 65° and a vertical field of view of 63°. At the software control level, taking full account of on-site conditions and practical requirements, in the horizontal direction, upon reaching each data acquisition point, the inspection robot extends the mechanical arm to the front and positions it at the transverse central position of the standard corridor (e.g., point P1 or P2 in Figure 6a below). Additionally, an overlapping image area of approximately 5° is designed to ensure image overlap between acquisition points. In the cross-sectional direction, the mechanical arm is set at a height of approximately 1 m from the ground. To ensure full coverage of the corridor side walls and vault, a total of five images are designed to be captured, with an overlap of approximately 5° employed to achieve image overlap in the cross-sectional direction.

3.2. Data Augmentation

Given the limitations of manual inspection during flood-discharge operations, the structural damage is often concealed, making it difficult for regular manual inspections to promptly and accurately identify corridor defects and operating states. Additionally, the scarcity of effective defect data in the early stage makes it challenging to support the design of concrete-defect identification methods. This paper adopts the data augmentation algorithm. The collected data were processed with random flipping and random color adjustment, the effect is shown in Figure 7 and Figure 8, including horizontal flipping, transposed flipping, vertical flipping, brightness enhancement, and brightness reduction, resulting in a plunge pool corridor crack dataset consisting of 4075 images in total.

4. Experiment

4.1. Experimental Environment

All experiments in this paper were conducted under the Windows operating system, using PyTorch 1.11.0 on a personal server equipped with an Intel i5-13600kf processor, 32 GB of RAM, and an NVIDIA GeForce RTX 4070 GPU (12GB). In the experiments, the SGD optimizer was employed to update the weight parameters of the plunge pool corridor crack detection model. The DiceLoss [35] was used as the loss function. The number of training epochs was 200, the batchsize was 8, the initial learning rate was 0.007, and the learning rate decayed by 0.0001 every 10 epochs. Throughout the model training process, the loss change curve and pixel accuracy change curve are depicted in the Figure 9.

4.2. Evaluation Metrics

Evaluation metrics refer to the use of various data indicators to determine the detection performance of the model. The metrics adopted in this paper include mean intersection over union (MIoU), F1-score, mean pixel accuracy (MPA), and accuracy.
In crack detection, the MIoU metric can help evaluate the performance of the model in image segmentation, especially in capturing the degree of overlap between the detected crack area and the actual label. Its calculation formula is as follows:
M I o U = 1 N I = 1 N T P i T P i + F P i + F N i
Among them, N denotes the total number of categories; T P i denotes the true positives of the i-th category, that is, the number of pixels that the model correctly predicts as this category; F P i denotes the false positives of the i-th category, namely, the number of pixels that the model wrongly predicts as this category; and F N i denotes the false negatives of the i-th category, which means the number of pixels that the model fails to predict as this category.
The F1-score represents the balance coefficient between the precision and recall of the model, providing a comprehensive evaluation of the overall performance of the trained model.
P r e c i s i o n = T P i T P i + F P i
R e c a l l = T P i T P i + F N i
F 1 s c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
As an intuitive and straightforward evaluation metric, mean pixel accuracy (MPA) can comprehensively consider the classification accuracy of an image segmentation model at the pixel level. It provides an intuitive understanding of the overall performance of the model and is applicable to the evaluation and comparison of various image segmentation tasks.
M P A = i = 1 N TP i i = 1 N ( TP i + FP i )

4.3. Comparative Experiment

This paper verifies the superiority of the improved model by comparing three segmentation models: U-Net, DeepLabV3+ [36], U2Net [37], SegFormer [38], and CrackSegNet [39]. The figure below shows the detection results of U-Net, DeepLabV3+, U2Net, and the improved model on the plunge pool corridor crack dataset. The visual comparison is shown in Figure 10, The red box represents a relatively large difference, the comparison of evaluation metrics are shown in Table 1.
In Figure 10, the U-Net and DeepLabV3+ models show little overall difference in performance within this dataset. Nevertheless, DeepLabV3+ uses spatial pyramid pooling, allowing it to achieve a larger receptive field. As a result, it performs better in crack-feature extraction, with fewer occurrences of crack breakpoints, undetected cracks, and misidentifications. By introducing the residual module, U2Net can acquire more diverse crack features; SegFormer integrates the transformer with a lightweight multi-layer perceptron (MLP) decoder, allowing the encoder to produce high-resolution fine-grained features and low-resolution coarse-grained features; and the CrackSegNet decoder incorporates a spatial attention module to improve the feature representation of cracks. Yet there are still instances of crack breakpoints, undetected cracks, and misidentifications. In comparison to the previous models, the improved U-Net network model in this paper can capture crack features more effectively, and there is a significant reduction in the number of crack breakpoints, misidentifications, and undetected cracks.
By calculating the four metrics of MIOU, MPA, F1-score, and accuracy for the test dataset, the statistical results are presented in the table below. The U-Net and DeepLabV3+ models exhibit similar performances, and U2Net shows an improvement compared to the previous two. Compared with the pre-improvement U-Net model, the improved model in this paper has an increase of 5.34% in MIOU, 4.90% in MPA, 3.39% in F1-score, and 0.36% in accuracy. This shows that the improved model has improved in terms of both the extraction of crack features in different channel dimensions and the handling of the features of irregular shapes.

4.4. Ablative Experiment

To more effectively validate the performance of the morphology-integrated irregular dynamic convolution (MIDConv) and the cross-dimensional interactive attention module (CDIAM) in corridor crack detection, an ablation experiment was devised to examine the effectiveness of these two improvements in extracting crack features, as show in Figure 11, the red box represents a relatively large difference. The comparison of evaluation metrics are shown in Table 2.
Figure 11 indicates that by introducing irregular dynamic convolution to expand the fixed-shaped and sized convolution kernels in the original U-Net network architecture, the receptive field of the improved network is correspondingly enlarged. During crack feature extraction, U-Net-MIDConv acquires more abundant feature information, allowing the segmentation model to identify more crack features in the prediction map. It can be easily observed that while the crack segmentation results of U-Net-MIDConv are more effective than the original U-Net network and visually more comprehensive, U-Net-MIDConv still exhibits slight omissions in some details. In this case, by incorporating the cross-dimensional interactive attention module, pixel-level relationships between cracks are reconstructed using global channel information, minimizing information loss during high-dimensional feature extraction. Additionally, experimental results from adding the cross-dimensional interactive attention module to the original U-Net network alone validate its ability to handle details at a fine-grained level.
After computing the four metrics of MIOU, MPA, F1-score, and accuracy for the test dataset, the results are presented in the table. The results indicate that for the original U-Net network, independently adopting the morphology-integrated irregular dynamic convolution method (MIDConv) or adding the cross-dimensional interactive attention module (CDIAM) both improve the model’s detection performance, demonstrating that while MIDConv or CDIAM alone are effective in corridor crack feature extraction, they also have shortcomings. By combining the two, the model can both extract features according to the morphological features of the cracks and also, remarkably, extract crack features across different dimensions.
Additionally, to further validate the generalization ability of the algorithm proposed in this paper, the crack dataset used in the literature DeepCrack and the public crack dataset Crack500 were trained and tested using the method in this paper. The indicators of the test set are shown in Table 3. Through comparative analysis of metrics results, it is evident that the improved algorithm proposed herein demonstrates strong generalization capability under the same environmental configuration.

5. Discussion

5.1. Validity and Scenario Adaptability of Data Acquisition Scheme

The multi-angle data acquisition scheme proposed herein achieves full-area coverage of the side walls and vaults in plunge pool corridors via the mechanical arm system integrated with the inspection robot. The 5° overlapping region design in the horizontal direction and stratified shooting strategy with five images in the cross-sectional direction effectively address the issue of detection blind spots arising from traditional single-view acquisition, ensuring the integrity and diversity of crack samples. In comparison to manual handheld device acquisition, this scheme notably enhances the standardization and efficiency of data collection while eliminating the influence of human perspective bias on subsequent model training. Furthermore, data augmentation techniques, including random flipping and color adjustment, expand the original dataset from limited samples to 4075 images, effectively mitigating the typical small-sample learning issue in water conservancy engineering and enhancing the model’s robustness to complex backgrounds.

5.2. The Synergy of Improved Modules

The ablation experiments and visual analyses in Section 4.4 demonstrate that MIDConv and CDIAM achieve effective synergy via the complementarity of their feature extraction mechanisms. MIDConv adjusts convolution kernel offsets to enable dynamic expansion of the network’s receptive field, improving the capability to capture features of irregularly shaped cracks. CDIAM employs channel grouping and global information calibration to enhance pixel-level semantic relationships in complex backgrounds and suppress interference from similar textures. In their combination, the dynamic spatial features provided by MIDConv at the encoder stage and the cross-dimensional semantic enhancement realized by CDIAM at the decoder stage form hierarchical complementarity, jointly resolving the issues of geometric feature extraction for irregular morphologies and semantic ambiguity in complex backgrounds. Quantitative results indicate that the joint use of the two modules raises MIoU to 84.44%, representing increases of 2.08% and 1.66% over single modules, respectively, which validates the positive impact of the two modules’ synergy on improving crack detection accuracy.

5.3. Comparison with Existing Methods and Domain Migration Potential

Comparisons with existing semantic segmentation approaches reveal that the improved model in this study exhibits notable advantages in crack detection under complex backgrounds via scenario-tailored design. In contrast to mainstream open-source models, its core distinction lies in the redesign of the feature extraction framework to address the irregular morphologies and cross-dimensional semantic requirements of plunge pool corridor cracks. Regarding cross-domain recognition capabilities, the model achieves promising results on public building wall crack and pavement crack datasets, demonstrating that its core mechanisms possess cross-scenario applicability.

6. Conclusions

In the domain of hydropower station infrastructure safety, traditional manual inspection and crack detection methods based on conventional digital image processing are plagued by issues such as low efficiency and dependence on manual feature design. Current deep learning models encounter challenges in plunge pool corridor crack detection, including a lack of datasets, complex crack morphologies, and inadequate cross-dimensional semantic information extraction capabilities. To tackle these challenges, this study employs an inspection robot to acquire full-area crack data in plunge pool corridors and establishes a plunge pool corridor crack dataset comprising 4075 images through data augmentation algorithms applied to the original data. Through investigating the effects of integrating morphology-integrated irregular dynamic convolution (MIDConv) and the cross-dimensional interactive attention module (CDIAM) on model feature extraction, a pixel-level crack segmentation model based on an improved U-Net network tailored for this corridor dataset is proposed. The performance of the improved U-Net model is thoroughly validated via quantitative metrics and visual analysis.
(1)
Quantitative assessments reveal that the improved model achieves an MIoU, F1-score, MPA, and accuracy of 84.44%, 93.05%, 89.75%, and 99.46%, respectively. When compared with mainstream semantic segmentation models like DeepLabV3+ and U2Net, it exhibits superior fine-grained segmentation capability for overlapping areas between crack regions and backgrounds. The notable enhancement in F1-score signifies that the model has achieved breakthroughs in balancing precision and recall, effectively minimizing the missed detections of slender cracks (e.g., missing bifurcation structures) and misjudgments of complex backgrounds (e.g., stains, uneven lighting areas) of traditional models.
(2)
Visual comparisons further disclose that the improved model, through the morphology-integrated irregular dynamic convolution (MIDConv), binds the convolution kernel’s offsets in the X/Y axis directions with the linear morphological features of cracks, allowing the receptive field to dynamically expand along crack orientations and thus fully capture continuous edges. Simultaneously, the cross-dimensional interactive attention module (CDIAM) significantly enhances the capability to model pixel-level semantic relationships by partitioning channels into sub-feature groups and calibrating cross-dimensional weights via global average pooling, effectively suppressing background interference and retaining crack details.
(3)
Independently introducing the MIDConv or CDIAM module elevates the MIoU to 82.78% and 82.36%, respectively, whereas their combination further boosts the metric to 84.44%, demonstrating significant complementarity between dynamic feature extraction and cross-dimensional semantic enhancement. In generalization capability tests, the improved model validates its accuracy–efficiency balance advantage—achieved via structured feature group partitioning and convolution kernel offset constraints—on public datasets.
For future work, plans involve expanding multimodal datasets, optimizing model architecture, exploring cross-scenario transfer learning, and developing an integrated detection–repair system to further improve model performance and practical application value.

Author Contributions

Conceptualization, X.Z. and Y.S.; writing—original draft preparation, X.Z. and Y.S.; writing—review and editing, Y.S.; supervision, Y.L.; data curation, C.H., Z.L. and F.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Three Gorges Jinsha River Chuanyun Hydropower Development Co., Ltd. Yongshan Xiluodu Power Plant, No. Z412302022; the National Natural Science Foundation of China, No. U21A20157; and the Sichuan Science and Technology Program No. 2025ZHCG0016.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used in this article are not easily available, and due to third-party, commercial, or privacy restrictions, complete datasets cannot be provided for the time being.

Conflicts of Interest

Author Chunyao Hou, Zhihui Liu and Fan Xia was employed by the company China Yangtze Power Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest. The authors declare that this study received funding from Three Gorges Jinsha River Chuanyun Hydropower Development Co., Ltd. Yongshan Xiluodu Power Plant No. Z412302022. The funder was not involved in the study design, collection, analysis, interpretation of data, the writing of this article, or the decision to submit it for publication.

References

  1. Tian, W.; Shen, H.; Li, X.; Shi, L. Research on corridor surface crack detection technology based on image processing. Electron. Des. Eng. 2020, 5, 148–151. (In Chinese) [Google Scholar] [CrossRef]
  2. Yu, S.N.; Jang, J.H.; Han, C.S. Auto inspection system using a mobile robot for detecting concrete cracks in a tunnel. Autom. Constr. 2007, 16, 255–261. [Google Scholar] [CrossRef]
  3. Li, G.; He, S.; Ju, Y.; Du, K. Long-distance precision inspection method for bridge cracks with image processing. Autom. Constr. 2014, 41, 83–95. [Google Scholar] [CrossRef]
  4. Li, G.; Zhao, X.; Du, K.; Ru, F.; Zhang, Y. Recognition and evaluation of bridge cracks with modified active contour model and greedy search-based support vector machine. Autom. Constr. 2017, 78, 51–61. [Google Scholar] [CrossRef]
  5. Kim, Y. Convolutional Neural Networks for Sentence Classification. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1746–1751. [Google Scholar]
  6. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings Part III 18. Springer International Publishing: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  7. Fan, H.; Liu, X. Concrete Crack Detection Based on ST-UNet and Target Features. Comput. Syst. Appl. 2024, 33, 77–84. (In Chinese) [Google Scholar]
  8. Pan, Y.; Zhou, S.; Yang, D. Research on Concrete Crack Segmentation Based on Improved Unet Model. J. East China Jiaotong Univ. 2024, 41, 11–19. (In Chinese) [Google Scholar]
  9. Meng, Q.; Li, M.; Wan, D.; Hu, L.; Wu, H.; Qi, X. Real-time segmentation algorithm of concrete cracks based on M-Unet. J. Civ. Environ. Eng. 2024, 46, 215–222. (In Chinese) [Google Scholar]
  10. Zhang, H.; Ma, L.; Yuan, Z.; Liu, H. Enhanced concrete crack detection and proactive safety warning based on I-ST-UNet model. Autom. Constr. 2024, 166, 105612. [Google Scholar] [CrossRef]
  11. Liu, Z.; Cao, Y.; Wang, Y.; Wang, W. Computer vision-based concrete crack detection using U-net fully convolutional networks. Autom. Constr. 2019, 104, 129–139. [Google Scholar] [CrossRef]
  12. Liang, D.; Li, Y.; Zhang, S. Identification of cracks in concrete bridges through fusing improved ResNet-14 and RS-Unet models. J. Beijing Jiaotong Univ. 2023, 47, 10–18. (In Chinese) [Google Scholar]
  13. Zhou, T.; Zhou, Z.; Wu, Y.; Wen, X.; Pu, Q. Self-detection method of concrete bridge cracks based on improved U-Net network. Sichuan Cem. 2023, 47, 10–18. (In Chinese) [Google Scholar]
  14. Ma, Y.; Sun, W.; He, Y.; Wang, L. Surface crack identification method of concrete bridge based on DC-Unet. J. Chang’Univ. (Nat. Sci. Ed.) 2024, 44, 66–75. (In Chinese) [Google Scholar] [CrossRef]
  15. Shi, M.; Gao, J. Research on Pavement Crack Detection Based on Improved U-Net Algorithm. Autom. Instrum. 2022, 37, 52–55+67. (In Chinese) [Google Scholar] [CrossRef]
  16. Hui, B.; Li, Y. A Detection Method for Pavement Cracks Based on an Improved U-Shaped Network. J. Transp. Inf. Saf. 2023, 41, 105–114. (In Chinese) [Google Scholar] [CrossRef]
  17. Zhao, Z.; Hao, Z.; He, P. Combining attention mechanism with GhostUNet method for pavement crack detection. Electron. Meas. Technol. 2023, 46, 164–171. (In Chinese) [Google Scholar] [CrossRef]
  18. Zhang, M.; Xu, J.; Liu, X.; Zhang, Y.; Zhang, C. Improved U-Net pavement crack detection method. Comput. Eng. Appl. 2024, 60, 306–313. (In Chinese) [Google Scholar]
  19. Sun, Y.; Bi, F.; Gao, Y.; Chen, L.; Feng, S. A multi-attention UNet for semantic segmentation in remote sensing images. Symmetry 2022, 14, 906. [Google Scholar] [CrossRef]
  20. Chen, S.; Feng, Z.; Xiao, G.; Chen, X.; Gao, C.; Zhao, M.; Yu, H. Pavement Crack Detection Based on the Improved Swin-Unet Model. Buildings 2024, 14, 1442. [Google Scholar] [CrossRef]
  21. Al-Huda, Z.; Peng, B.; Riyadh, N.A.A.; Mugahed, A.A.; Rabea, A.; Omar, A.; Zhai, D. Asymmetric dual-decoder-U-Net for pavement crack semantic segmentation. Autom. Constr. 2023, 156, 105138. [Google Scholar] [CrossRef]
  22. Ren, Y.; Huang, J.; Hong, Z. Image-based concrete crack detection in tunnels using deep fully convolutional networks. Constr. Build. Mater. 2020, 234, 117367. [Google Scholar] [CrossRef]
  23. Qing, S.; Yingqi, W.; Xueshi, X.; Lu, Y.; Min, Y.; Hongming, C. Real-Time Tunnel Crack Analysis System via Deep Learning. IEEE Access 2019, 7, 64186–64197. [Google Scholar] [CrossRef]
  24. Choi, W.; Cha, Y.-J. SDDNet: Real-Time Crack Segmentation. IEEE Trans. Ind. Electron. 2020, 67, 8016–8025. [Google Scholar] [CrossRef]
  25. Yahui, L.; Jian, Y.; Xiaohu, L.; Renping, X.; Li, L. DeepCrack: A deep hierarchical feature learning architecture for crack segmentation. Neurocomputing 2019, 338, 139–153. [Google Scholar]
  26. Ju, H.; Li, W.; Tighe, S.; Xu, Z.; Zhai, J. CrackU-net: A novel deep convolutional neural network for pixelwise pavement crack detection. Struct. Control Health Monit. 2020, 27, e2551. [Google Scholar]
  27. Cao, H.; Gao, Y.; Cai, W.; Xu, Z.; Li, L. Segmentation Detection Method for Complex Road Cracks Collected by UAV Based on HC-Unet++. Drones 2023, 7, 189. [Google Scholar] [CrossRef]
  28. Sun, X.; Xie, Y.; Jiang, L.; Cao, Y.; Liu, B. DMA-Net: DeepLab With Multi-Scale Attention for Pavement Crack Segmentation. IEEE Trans. Intell. Transp. Syst. 2022, 23, 18392–18403. [Google Scholar] [CrossRef]
  29. Dai, J.; Qi, H.; Xiong, Y.; Li, Y.; Zhang, G.; Hu, H.; Wei, Y. Deformable Convolutional Networks. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 764–773. [Google Scholar] [CrossRef]
  30. Vaswani, A. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems—NIPS’17, Red Hook, NY, USA, 4–9 December 2017; pp. 6000–6010. [Google Scholar]
  31. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  32. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  33. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient channel attention for deep convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11534–11542. [Google Scholar]
  34. Hou, Q.; Zhou, D.; Feng, J. Coordinate attention for efficient mobile network design. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13713–13722. [Google Scholar]
  35. Li, X.; Sun, X.; Meng, Y.; Liang, J.; Wu, F.; Li, J. Dice Loss for Data-imbalanced NLP Tasks. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 2020. [Google Scholar] [CrossRef]
  36. Chen, L.C.; Zhu, Y.; George, P.; Florian, S.; Hartwig, A. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. In European Conference on Computer Vision; Springer: Cham, Switzerland; Glasgow, KY, USA, 2018. [Google Scholar] [CrossRef]
  37. Qin, X.; Zhang, Z.; Huang, C.; Dehghan, M.; Zaiane, O.R.; Jagersand, M. U2-Net: Going deeper with nested U-structure for salient object detection. Pattern Recognit. 2020, 106, 107404. [Google Scholar] [CrossRef]
  38. Xie, E.; Wang, W.; Yu, Z.; Anandkumar, A.; Alvarez, J.M.; Luo, P. SegFormer: Simple and efficient design for semantic segmentation with transformers. Adv. Neural Inf. Process. Syst. 2021, 34, 12077–12090. [Google Scholar]
  39. Ran, R.; Xu, X.; Qiu, S.; Cui, X.; Wu, F. Crack-SegNet: Surface Crack Detection in Complex Background Using Encoder-Decoder Architecture. In Proceedings of the 2021 4th International Conference on Sensors, Signal and Image Processing (SSIP’21). Association for Computing Machinery, New York, NY, USA, 15–17 October 2021; pp. 15–22. [Google Scholar] [CrossRef]
Figure 1. U-Net model.
Figure 1. U-Net model.
Applsci 15 05225 g001
Figure 2. Improved U-Net model.
Figure 2. Improved U-Net model.
Applsci 15 05225 g002
Figure 3. Morphological irregular dynamic convolution method.
Figure 3. Morphological irregular dynamic convolution method.
Applsci 15 05225 g003
Figure 4. Comparison of FOV between traditional convolution and morphological irregular dynamic convolution methods: (a) FOV of traditional convolution; (b) FOV of MIDConv.
Figure 4. Comparison of FOV between traditional convolution and morphological irregular dynamic convolution methods: (a) FOV of traditional convolution; (b) FOV of MIDConv.
Applsci 15 05225 g004
Figure 5. Cross-dimensional interaction attention module structure.
Figure 5. Cross-dimensional interaction attention module structure.
Applsci 15 05225 g005
Figure 6. Data collection scheme for cracks in plunge pool corridor: (a) horizontal image acquisition scheme; (b) cross-sectional image acquisition scheme.
Figure 6. Data collection scheme for cracks in plunge pool corridor: (a) horizontal image acquisition scheme; (b) cross-sectional image acquisition scheme.
Applsci 15 05225 g006
Figure 7. Results of random flipping of the dataset.
Figure 7. Results of random flipping of the dataset.
Applsci 15 05225 g007
Figure 8. Results of random color adjustment of the dataset.
Figure 8. Results of random color adjustment of the dataset.
Applsci 15 05225 g008
Figure 9. Loss change and pixel accuracy change curves: (a) loss change; (b) pixel accuracy change.
Figure 9. Loss change and pixel accuracy change curves: (a) loss change; (b) pixel accuracy change.
Applsci 15 05225 g009
Figure 10. Comparison of visualization results of different algorithms.
Figure 10. Comparison of visualization results of different algorithms.
Applsci 15 05225 g010
Figure 11. Comparison of crack detection results in ablation experiments.
Figure 11. Comparison of crack detection results in ablation experiments.
Applsci 15 05225 g011
Table 1. Comparative experiments of different detection models.
Table 1. Comparative experiments of different detection models.
ModelMIOU (%)MPA (%)F1-Score (%)Accuracy (%)
U-Net79.187.1584.3699.1
DeepLabV3+80.588.0484.7999.2
U2Net82.1189.4585.5299.11
SegFormer82.6489.69.85.7899.32
CrackSegNet81.3788.5384.9499.23
Ours84.4492.0587.7599.46
Table 2. Comparative experiments of different model structures.
Table 2. Comparative experiments of different model structures.
ModelMIOU (%)MPA (%)F1-Score (%)Accuracy (%)
U-Net79.187.1584.3699.1
U-Net+CDIAM82.3689.5485.6899.27
U2Net+MIDConv82.7891.5687.0299.30
Ours84.4492.0587.7599.46
Table 3. Comparison table of model generalization capability indicators.
Table 3. Comparison table of model generalization capability indicators.
DatasetMIOU (%)MPA (%)F1-Score (%)Accuracy (%)
DeepCrack83.6488.8386.3499.37
Crack50084.3789.2087.3699.41
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hou, C.; Liu, Z.; Xia, F.; Zhou, X.; Shui, Y.; Li, Y. Crack Defect Detection Method for Plunge Pool Corridor Based on the Improved U-Net Network. Appl. Sci. 2025, 15, 5225. https://doi.org/10.3390/app15105225

AMA Style

Hou C, Liu Z, Xia F, Zhou X, Shui Y, Li Y. Crack Defect Detection Method for Plunge Pool Corridor Based on the Improved U-Net Network. Applied Sciences. 2025; 15(10):5225. https://doi.org/10.3390/app15105225

Chicago/Turabian Style

Hou, Chunyao, Zhihui Liu, Fan Xia, Xun Zhou, Yuhang Shui, and Yonglong Li. 2025. "Crack Defect Detection Method for Plunge Pool Corridor Based on the Improved U-Net Network" Applied Sciences 15, no. 10: 5225. https://doi.org/10.3390/app15105225

APA Style

Hou, C., Liu, Z., Xia, F., Zhou, X., Shui, Y., & Li, Y. (2025). Crack Defect Detection Method for Plunge Pool Corridor Based on the Improved U-Net Network. Applied Sciences, 15(10), 5225. https://doi.org/10.3390/app15105225

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop