Next Article in Journal
Clinical Features and Rehabilitation Needs in Patients with Neuroinvasive West Nile Virus Infection: A Retrospective Analysis in an Area of High Incidence
Previous Article in Journal
Improved Binary Grey Wolf Optimization Approaches for Feature Selection Optimization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Feature Fusion Model Using Transfer Learning and Bidirectional Attention Mechanism for Plant Pipeline Leak Detection

1
Department of AI and Big Data Engineering, Daegu Catholic University, 13-13, Hayang-ro, Hayang-eup, Gyeongsan-si 38430, Republic of Korea
2
Smart Structural Safety & Prognosis Research Division, Korea Atomic Energy Research Institute (KAERI), 62, Gwahak-ro, Yuseong-gu, Daejeon 34142, Republic of Korea
3
Department of Computer Education, Korea National University of Education, Cheongju-si 28173, Republic of Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(2), 490; https://doi.org/10.3390/app15020490
Submission received: 26 October 2024 / Revised: 7 December 2024 / Accepted: 10 December 2024 / Published: 7 January 2025

Abstract

:
The aging of power plant pipelines has led to significant leaks worldwide, causing environmental damage, human safety risks, and economic losses. Rapid leak detection is critical for mitigating these issues, but challenges such as varying leak characteristics, ambient noise, and limited real-world data complicate their accurate detection and model development. To address these issues, we propose a leak detection model that integrates stepwise transfer learning and an attention mechanism. The proposed model utilizes a two-stage deep learning process. In Stage 1, one-dimensional convolutional neural networks (1D CNNs) are pre-trained to extract root mean square (RMS) and frequency-domain features from acoustic signals. In Stage 2, the classifier layers of the pre-trained models are removed, and the extracted features are fused and processed using a bidirectional long short-term memory (LSTM) network. An attention mechanism is incorporated within the LSTM to prioritize critical features, enhancing the ability of the model to distinguish leak signals from noise. The model achieved an accuracy of 99.99%, significantly outperforming the traditional methods considered in this study. By effectively addressing noise interference and data scarcity, this robust approach demonstrates its potential to enhance safety, reduce risks, and improve cost efficiency in industrial pipelines and critical infrastructure.

1. Introduction

The aging of power plant piping, both domestically and internationally, has increasingly resulted in severe leaks, causing substantial harm to human safety, the environment, and economic stability [1,2,3]. Consequently, the prompt identification of leaks in aging piping systems is important for ensuring safety and economic efficiency [4,5,6]. Historically, various sensors, including acoustic, vibration, pressure, and ultrasonic types, have been explored for early leak detection [7,8,9]. However, accurately detecting leaks remains challenging due to the varying physical characteristics of leaks, which are influenced by factors such as pressure, material properties, and environmental conditions. Ambient noise, in particular, significantly complicates the differentiation of leak signals, leading to reduced detection rates [10,11].
With the advent of large-scale data, advancements in deep learning (DL) and machine learning (ML) technologies have progressed rapidly, enabling solutions to a range of complex problems [12,13,14]. These technologies have become integral in industrial applications, excelling in modeling complex patterns and delivering accurate predictions [15,16,17]. Artificial intelligence (AI), in particular, has emerged as a transformative tool for enhancing productivity, reducing costs, and improving safety across various sectors, owing to its advanced learning and inference capabilities [18,19,20]. This has raised significant expectations for the role of AI in detecting minor leaks in aging piping systems, fostering active research in this area [21,22,23]. For instance, Mysorewala et al. achieved 93% accuracy in detecting vibrations caused by leaks using low-power accelerometers and a support vector machine (SVM) model [24]. Similarly, Ahmad et al. reported detection performances of 98.4% and 96.67% using convolutional autoencoders and convolutional neural network (CNN) models, respectively [25]. These studies demonstrate the potential of AI to advance fine leak detection in aging piping systems.
Despite these advancements, existing methods face significant limitations. Traditional signal processing techniques often fail to accurately distinguish leak signals from noise, as noise can distort the defining characteristics of leak signals. Additionally, achieving high accuracy and performance in leak detection requires extensive data, yet collecting leak data in real-world environments is both challenging and hazardous. Even under controlled conditions that simulate real-world scenarios, the data collected are often insufficient due to the limitations of the experimental setup. Therefore, there is an urgent need for research focused on differentiating leak signals from noise with high accuracy and reliability, even when working with small datasets.
To address these challenges, this paper proposes a feature fusion DL model based on stepwise transfer learning (TL) [26,27,28] that utilizes an attention mechanism. Time-domain root mean square (RMS) and frequency-domain frequency patterns are generated from one-dimensional (1D) time-series acoustic signals acquired using multiple acoustic sensors. These patterns are then used to train the feature fusion model based on TL. The proposed model comprises a two-stage process (Stages 1 and 2). In Stage 1, key features are extracted from RMS and frequency pattern signals using 1D CNN models for each domain after preprocessing the source domain data. In Stage 2 of the TL process, the classifiers of the 1D CNN models trained in Stage 1 for each domain are removed, and the output features of these models are combined. The preprocessed RMS and frequency pattern signals from the target domain data are then input into a model enhanced with bidirectional long short-term memory (LSTM) [29].
The bidirectional LSTM captures both forward and backward temporal dependencies, enriching the understanding of the model of sequence data. Unlike unidirectional LSTMs, which process input sequences in a single direction from past to future, bidirectional LSTMs consider the entire context of the data, thereby improving the detection of subtle patterns and the prediction accuracy. Numerous studies have demonstrated that bidirectional LSTMs often outperform unidirectional LSTMs in tasks requiring temporal feature extraction due to their ability to capture more comprehensive temporal dependencies [30,31,32]. By processing sequence information in both directions simultaneously, the bidirectional LSTM enables the model to utilize richer contextual information. Additionally, a dot-product-based attention mechanism [33] applied to the LSTM in each direction helps the model focus on important patterns in the input from both directions, effectively extracting fused sequence features in the time and frequency domains. Subsequently, the outputs of these attention layers are concatenated into a single layer. Finally, a classifier is added to this combined layer to extract domain-specific features and make the final prediction. This approach allows the model to achieve more accurate predictions by leveraging attended features extracted from both directions of the input.
This paper is organized as follows: In Section 1, the research background is discussed, emphasizing the significance of addressing leak issues in aging pipes. Section 2 outlines the process of feature extraction in the RMS and frequency domains, along with the data preprocessing techniques applied. Section 3 introduces the proposed feature fusion leak detection model, which integrates TL and bidirectional attention mechanisms. Section 4 describes the data collection setup and assesses the performance of the model. Finally, Section 5 provides a conclusion and suggests potential avenues for future research.

2. Data Preprocessing for Feature Extraction

This section describes the data preprocessing methods used to differentiate between leak and normal states based on 1D time-series acoustic signals obtained from several microphone sensors. As outlined in [34] and illustrated in Figure 1, 1D features representing RMS patterns and frequency patterns are extracted. At this stage, two signal processing techniques, designated as ( P a t h α ,   P a t h β ), are applied to extract the RMS pattern features. The detailed nomenclature and description of the data preprocessing methods are provided in Table 1.
First, in P a t h α , the input data x are sampled at intervals of T d without any filtering, and the sampled data are represented as x n = x n T d . A moving average window is then applied to the sampled input data x = x n to calculate their RMS level. Through this process, y α = [ y n ] n = 1,2 , , T is generated, where y n is calculated as shown in Equation (1):
y n = 1 C j = 1 L n 1 x n + j 2 ,   n = 1,2 , , T ,  
where C denotes a coefficient used to normalize the data and L n represents the window size for the RMS level calculation of the received data.
In P a t h β , the preprocessed signal y * is first generated by passing the input data through a band-pass filter (BPF) block, as defined in Equation (2):
y * = B P F x
where the B P F operates within the 20 kHz to 40 kHz frequency range, capturing ultrasonic leak signals primarily caused by turbulence at the leak site. This frequency range effectively reduces low-frequency mechanical noise, focusing on critical leak signal features to improve detection accuracy. Additional parameters, such as the amplitude weighting function and window size for RMS level calculation, are carefully selected to optimize signal processing. These parameters highlight key signal features, minimize irrelevant noise, and prepare the signal y * for subsequent processing stages. The preprocessed signal y * then passes through the “amplitude weighting block”, where a weighting function W ( f ) is applied in the frequency domain. This function emphasizes desired frequency components, suppresses external noise, and amplifies signals with concentrated spectral presence. The processed signal y * * is calculated as shown in Equation (3):
y * * = I F T W f · F T y * ,
where F T · denotes the 1D Fourier transform (FT) and I F T · represents the 1D inverse FT.
The weighting function W ( f ) is defined as shown in Equation (4):
W f = k ,   f f L , f H 0 ,   o t h e r w i s e ,
where f L and f H are the beginning and ending frequencies for the windowing operation, and k represents the amplification ratio of the signal within this range. In the RMS block of P a t h β , the RMS level for y * * is calculated based on Equation (1), similar to P a t h α , to generate y β . Finally, the y α and y β feature vectors from both paths are combined to produce the final RMS pattern feature y r m s = [ y α , y β ] .
Frequency pattern features are then derived from the received time-series data by analyzing the frequency response magnitude of the leak signal. Similar to P a t h β of the RMS pattern features in Figure 1, the frequency pattern feature is obtained by converting the preprocessed signal y * , after passing through the BPF block, into frequency-domain data using the 1D FT, as shown in Equation (5):
y f r e q = F T y * ,
This process results in the final frequency pattern feature y f r e q .
Figure 2 provides a visual representation of the pattern features extracted from the RMS and frequency components of leak and noise signals. Figure 2a illustrates the RMS pattern features, with the x -axis divided into two sections. The left side corresponds to RMS levels computed directly without filtering in P a t h α , while the right side represents RMS levels calculated after filtering and amplitude weighting in P a t h β . For the leak feature shown in Figure 2a, the RMS pattern exhibits an increasing trend in P a t h β , whereas for the noise feature, external noise is effectively filtered, leading to a decreasing trend in the RMS pattern size. Figure 2b displays the frequency pattern.
The 1D RMS and frequency pattern features are further integrated with domain-specific features obtained from multiple sensors to train the proposed model. Figure 3 demonstrates the sequential alignment of these pattern features along the channel direction, using data gathered from four sensors to form two-dimensional (2D) tensor features. This method combines features from various sensors, structuring the resulting 2D tensor as input for the proposed model, as discussed in Section 3.

3. Proposed Leak Detection Model

Using the data preprocessing methods for feature extraction described in Section 2, the 1D RMS and frequency-domain data were converted and utilized for training the proposed model for fine leak detection. The proposed model comprises two stages, as shown in Figure 4. TL is carried out by transferring the knowledge acquired in Stage 1 in (source domain) to Stage 2 (target domain). This approach utilizes the pre-trained model from Stage 1 to mitigate overfitting issues caused by the limited data available in the target domain.
In Stage 1, a 1D CNN structure is utilized to extract local feature maps from the preprocessed 1D RMS and frequency data, focusing on identifying fine leak or noise signals. The 1D convolution operation is mathematically defined in Equation (6):
x d , n l = f l m R n l x d , m l 1 × w d , m , n l + b d , n l , d T , F , n = 1,2 ,   , N d l ,   l = 1,2 ,   ,   L d ,
where d denotes the domain, with T denoting the time domain and F the frequency domain. F . x d , n l denotes the 1D feature map n of domain d at layer l , with dimensions 1 × W d l . x d , m l 1 denotes the 1D feature map m of domain d at layer l 1 , whereas w ( d , m , n ) l denotes the kernel function with a window size of 1 × C d l , mapping from feature map m of domain d at layer l 1 to feature map n at layer l . The bias term is denoted by b d , n l , and f l ( · ) denotes the activation function. N d l denotes the number of output feature maps at layer l , L d denotes the total number of convolution layers in domain d , and R n l denotes the list of 1D feature maps at layer l 1 connected to feature map n .
To pre-extract features for leak and noise signals, a 1D convolution layer with a kernel size of 1 × 3   and eight filters is applied, as summarized in Table 2. This convolution operation is followed by batch normalization (BN) and MaxPooling1D, which compress the extracted features. This process corresponds to the 1D CNN block (1) within the 1D CNN module, forming the initial feature extraction phase for fine leak detection.
Next, a second 1D convolution layer with the same kernel size and 16 filters is applied, followed by BN and global max-pooling to extract the main representative features. This corresponds to the 1D CNN block (2) within the same module. The 1D convolution employs the rectified linear unit (ReLU) as its activation function. To enable independent learning for each domain, the final output feature map x d of domain d is used as the input for a binary classifier. The first layer of this classifier contains eight neurons activated by a ReLU function, while the output layer includes a single neuron with a sigmoid activation function.
In Stage 2, the pre-trained “1D CNN Module” for each domain, independently learned and stored in Stage 1, is loaded, as shown in Figure 4. Feature fusion is performed by removing the classifiers from each model and combining the output feature maps x T and x F , representing the time and frequency domains, respectively, into a single higher-dimensional layer Z . This fusion process is mathematically defined in Equation (7) as follows:
Z = f c o n c a t ( [ x T ; x F ] ) ,
where f c o n c a t denotes the concatenation function, which merges the two feature maps. Through this feature fusion process, temporal- and frequency-domain information are integrated, resulting in richer features that capture more comprehensive insights.
A bidirectional LSTM network is then applied to the fused feature data. This network consists of a forward output h t f with 128 units and a backward output h t b with 128 units, which collectively consider sequence information in both forward and backward directions. BN is applied to the bidirectional LSTM outputs to stabilize the learning process. By leveraging the bidirectional LSTM, the model processes sequence data more effectively, utilizing context from both input directions.
To emphasize important information at each time step, a dot-product-based attention mechanism is applied to the outputs of each LSTM layer. The attention score is calculated using Equation (8) as follows:
S c o r e h t i , h s i = ( h t i ) T h s i ,   i { f , b } ,
where i indicates the LSTM direction f for forward and b for backward, h t i denotes the hidden state at time step t , h s i denotes the hidden state at all time steps s , and h t i T represents the transpose of h t i . A softmax operation is applied to the attention scores to normalize them into probabilities, which are then used to compute the attention weights. This process is defined in Equation (9) as follows:
α t , s i = e x p S c o r e h t i , h s i s e x p S c o r e h t i ,   h   s i ,   i { f , b } ,
where α t , s i represents the attention weight between the hidden state h t at the current time step and the hidden state h s at a different time step. The softmax operation in Equation (4) converts each attention score into a probability value, enabling the calculation of weights for each time step. Additionally, s denotes the summation index that includes all possible time steps, ensuring the denominator normalizes the total similarity scores into a probability distribution across the sequence.
The context vector c t i is then computed as the weighted sum of the attention weight α t , s i and the hidden state h s i across each time step s , as shown in the following Equation (10):
c t i = s α t , s i h s i ,   i f , b .
Using Equation (5), the hidden state at each time step is weighted by its corresponding attention weight to form the context vector c t i . This allows the model to summarize and extract the most important information from the entire input sequence effectively.
Finally, the attention vector α t i for each direction is calculated in Equation (11) by combining the context vector c t i with the current hidden state h t i , providing domain-specific trainable attention weights.
α t i = tanh W a i c t i ; h t i ,   i f , b ,
where W a i denotes the learnable weight matrix for direction i and c t i ; h t i denotes the concatenation of c t i and h t i . The hyperbolic tangent function (tanh) introduces non-linear transformation, generating the final attention vector and refining the focus of the model on domain-specific features.
Attention vectors α t f and α t b , obtained from the forward and backward directions of the bidirectional LSTM, are concatenated to form a single vector y α . This process is mathematically defined in Equation (12) as follows:
y α = f c o n c a t α t f ; α t b .
Subsequently, a fully connected (FC) layer comprising 64 units is added, with y α serving as the input. This layer employs a ReLU activation function to introduce non-linearity, enhancing the ability of the model to capture complex patterns in the fused features. The final output layer consists of a single neuron with a sigmoid activation function, designed to perform binary classification, distinguishing between leak and noise signals. The architecture of Stage 2 in the proposed feature fusion model, incorporating TL and the described layers, is summarized in Table 3. Additionally, the nomenclature for the proposed leak detection model is summarized in Table 4, which defines the key variables and their corresponding descriptions for clarity and reference.

4. Experimental Results

4.1. Data Collection Environment and Experimental Setup for Leak Detection

This paper investigated the model transfer across different testbed environments. In Stage 1, the model was pre-trained using leak signals and mechanical noise data collected from a source domain within a controlled testbed environment, as shown in Figure 5 [34]. The dataset was obtained using four microphone acoustic sensors, each operating at a sampling frequency of 100 kHz, with a signal measurement duration of 1.5 s per recording. The commercial ICP microphone (Model HT378A06) from PCB Piezotronics, located in Depew, NY, USA, was utilized as the sensor for this study [22]. Leak signals were recorded at 100 distinct locations along the pipeline under 25 unique combinations of leak pressure and leak size. The leak pressure was varied from 1 bar to 5 bar in increments of 1 bar, while the leak diameter ranged from 0.5 mm to 2.5 mm, increasing in intervals of 0.5 mm.
In the target domain, data were collected from 10 leak locations using the same four microphone sensors as employed in the source domain, as shown in Figure 6. The sampling frequency and measurement duration were set to 100 kHz and 5 s, respectively. According to micro-leak detection standards [35], the leak pressure was varied between 1 and 2 bar, and the leak diameter was fixed at 0.5 mm. Both the source and target domains used identical microphone acoustic sensors and applied the same signal preprocessing techniques described in this study. However, the source domain provided a more diverse data collection environment, incorporating additional measurement locations and leakage conditions. In contrast, the target domain focused on conditions specifically tailored to detect low-level leaks.
The data collected from the source domain were preprocessed using the methods detailed in Section 2 and converted into 65,985 RMS and frequency features. To assess the performance of the proposed model in detecting leaks, the time- and frequency-domain data were divided into training and test sets using the hold-out method, with a split ratio of 6:4. The source domain included 33,000 leak data points and 32,985 background noise data points, while the target domain comprised 33,000 combined leak and noisy data points. The training procedure involved two stages. In Stage 1, two separate models were trained independently, while in Stage 2, the proposed feature fusion model was developed. For both stages, a batch size of 128 was employed, with training conducted for 30 epochs. Binary cross-entropy was used as the loss function, and the learning rate was configured to 1 × 10 3 .

4.2. Analysis of Results for the Proposed Pipe Leak Detection

Figure 7 illustrates the classification results of the models with TL and without TL (non-TL) using confusion matrices, which display the correct and incorrect classifications of leak and normal states. The non-TL model was trained end-to-end in Stage 2 without leveraging Stage 1 pre-training, as shown in Figure 4. The confusion matrix on the left demonstrates the performance of the non-TL model, correctly classifying 12,940 noise instances and 13,450 leak instances, while misclassifying one noise instance and nine leak instances. In contrast, the matrix on the right shows the results of TL model, with 12,939 noise instances and 13,458 leak instances correctly classified, and only two noise instances and one leak instance incorrectly classified. These results confirm that the proposed TL model achieves superior classification accuracy for leak and normal states by leveraging pre-trained fused features compared to the non-TL model.
The performance of the leak detection model was further evaluated based on precision, recall, and F1-score metrics. Precision, defined in Equation (13), quantifies the proportion of true leaks among all samples identified as leaks. Recall, defined in Equation (14), measures the proportion of actual leaks correctly detected by the model. The F1-score, shown in Equation (15), combines precision and recall into a single metric by calculating their harmonic mean.
P r e c i s i o n = T P T P + F P ,
R e c a l l = T P T P + F N ,
F 1 s c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l ,
where true positive (TP) refers to cases where the model correctly identifies a sample as a leak when it truly is one. A false positive (FP) occurs when the model incorrectly classifies a noise sample as a leak, while false negative (FN) signifies instances where the model mistakenly classifies a leak sample as noise.
Table 5 summarizes the micro-leak detection accuracies of the proposed model compared to various existing ML and DL approaches. Each model was independently trained on 1D features derived from the time and frequency domains, with their outputs combined using an ensemble approach. The multi-layer perceptron (MLP) [36,37] ensemble employed an MLP model to learn RMS and frequency features, achieving an ensemble accuracy of 80.18%. The SVM-ensemble, which employed an SVM model, achieved a higher ensemble accuracy of 85.36%. The LSTM-ensemble, combining results learned through an LSTM network, achieved an accuracy of 74.47%. The 1D CNN-based model introduced in Stage 1 learned RMS and frequency features with accuracies of 98.42% and 98.97%, respectively, culminating in an ensemble accuracy of 99.32% (1D CNN-ensemble). In contrast, the TL-based model introduced in Stage 2 of the proposed framework exhibited superior performance, achieving an exceptionally high accuracy of 99.99%.

4.3. Comprehensive Performance Analysis of the Proposed Method

Figure 8 illustrates the performance comparison among the FC-TL model, the non-TL model, and the proposed model as the dataset size decreases progressively from 100% to 1%. The FC-TL model employs transfer learning by pre-training individual models for different domains during Stage 1, as shown in Figure 4. In this stage, domain-specific features are learned independently for each domain. After training, the classifiers of the domain-specific models are removed, and feature fusion is performed to combine the extracted features from all domains. Following this step, a new classifier is added to utilize the fused features. However, the FC-TL model lacks the bidirectional LSTM and attention mechanism, which are crucial to Stage 2 of the proposed architecture, as depicted in Figure 4. While the FC-TL model effectively uses pre-trained features and benefits from transfer learning and feature fusion, its performance diminishes as the dataset size decreases. Specifically, with approximately 1% of the dataset, the FC-TL model’s accuracy drops to 83.45%. While it benefits from TL and feature fusion, the absence of mechanisms to process temporal dependencies or selectively emphasize important features limits its capability, exposing its limitations in achieving robust performance under data-scarce scenarios.
The non-TL model is trained end-to-end in Stage 2 without utilizing Stage 1 pre-training, as shown in Figure 4. Unlike the FC-TL model, it does not utilize transfer learning or pre-trained features from other domains, relying solely on domain-specific features extracted during the training process. When the dataset size was large, the non-TL model performed comparably to the FC-TL model, maintaining stable accuracy. However, as the dataset size decreased, particularly below 30%, its performance declined rapidly. This significant drop demonstrates the heavy reliance of the model on a large volume of labeled data for effective learning and its limited ability to generalize in data-scarce conditions.
In contrast, the proposed model outperformed both the FC-TL and non-TL models, demonstrating exceptional robustness across all dataset sizes. Even with approximately 1% of the data, the proposed model achieved an impressive accuracy of 92.21%, far surpassing the other models. This superior performance is attributed to its efficient utilization of pre-trained features and its ability to combine diverse data through feature fusion. Furthermore, the integration of a bidirectional LSTM enables the model to process temporal information in both forward and backward directions, while the attention mechanism effectively focuses on critical features. These advanced components allow the model to differentiate leak signals from noise with high precision. These advanced components have been experimentally verified, demonstrating that the proposed model can effectively generalize and maintain consistent performance, even with minimal data.

5. Conclusions

In this study, a feature fusion leak detection model was proposed, combining TL and a bidirectional attention mechanism to identify micro-leaks in aging pipelines. The model utilized TL and feature fusion based on pre-trained information, incorporating an attention mechanism within the bidirectional LSTM to focus on critical independent features. This approach effectively addressed challenges such as data scarcity and ambient machine noise by enabling seamless transfer from the source domain to the target domain across distinct environments. A detection accuracy of 99.99% was achieved under controlled experimental conditions using a dataset comprising leak signals and ambient noise collected from multiple acoustic sensors. The ability of the model to suppress irrelevant noise and enhance key signal features through optimized signal preprocessing and feature fusion significantly contributed to this high level of accuracy. However, this exceptional performance requires further validation in real-world environments, where factors such as pipeline material variability, sensor placement, and diverse noise sources could impact the robustness of the model. Despite these potential limitations, the proposed method demonstrated strong potential for practical applications in industrial pipelines and critical infrastructure, where accurate and timely leak detection is crucial for safety and operational efficiency. Future research directions include deploying the model in embedded systems for real-time monitoring, validating its performance in diverse industrial settings, and exploring its scalability for multi-sensor environments. These efforts aim to develop a robust and adaptable early leak detection technology capable of operating effectively under real-world conditions, thereby enhancing industrial safety and efficiency.

Author Contributions

Conceptualization, Y.H. and J.-H.B.; methodology, Y.H. and J.-H.B.; software, Y.H.; validation, Y.H., Y.C., and J.L.; investigation, Y.H.; resources, Y.C.; data curation, Y.C.; draft writing and preparation, Y.H.; manuscript review and editing, J.L. and J.-H.B.; visualization, Y.H.; supervision, J.-H.B.; project administration, Y.H. and J.-H.B.; funding acquisition, Y.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation of Korea through the Government (Ministry of Science and Information and Communications Technology) in 2022 under Grant RS-2022-00165225 and Grant RS-2022-00144000.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets presented in this article are not readily available because the data are part of an ongoing study and security reasons. Requests to access the datasets should be directed to the Korea Atomic Energy Research Institute.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Hussain, M.; Zhang, T.; Dwight, R.; Jamil, I. Energy pipeline degradation condition assessment using predictive analytics—Challenges, issues, and future directions. J. Pipeline Sci. Eng. 2024, 4, 100178. [Google Scholar] [CrossRef]
  2. Di-Sarno, L.; Majidian, A. Risk assessment of a typical petrochemical plant with ageing effects subjected to seismic sequences. Eng. Struct. 2024, 310, 118110. [Google Scholar] [CrossRef]
  3. Nnoka, M.; Jack, T.A.; Szpunar, J. Effects of different microstructural parameters on the corrosion and cracking resistance of pipeline steels: A review. Eng. Fail. Anal. 2024, 159, 108065. [Google Scholar] [CrossRef]
  4. Nati, M.; Nastasiu, D.; Stanescu, D.; Digulescu, A.; Ioana, C. Pipe leaks detection and localization using non-intrusive acoustic monitoring. In Proceedings of the 2024 15th International Conference on Communications (COMM), Bucharest, Romania, 3–4 October 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1–5. [Google Scholar] [CrossRef]
  5. Yan, W.; Liu, W.; Zhang, Q.; Bi, H.; Jiang, C.; Liu, H.; Wang, T.; Dong, T.; Ye, X. Multi-source multi-modal feature fusion for small leak detection in gas pipelines. IEEE Sens. J. 2023, 24, 1857–1865. [Google Scholar] [CrossRef]
  6. Obike, A.; Uwakwe, K.; Abraham, E.; Ikeuba, A.; Emori, W. Review of the losses and devastation caused by corrosion in the Nigeria oil industry for over 30 years. Int. J. Corros. Scale Inhib. 2020, 9, 74–91. [Google Scholar] [CrossRef]
  7. Wang, T.; Wang, X.; Wang, B.; Fan, W. Gas leak location method based on annular ultrasonic sensor array. In Proceedings of the 2018 IEEE International Instrumentation and Measurement Technology Conference (I2MTC), Houston, TX, USA, 14–17 May 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–6. [Google Scholar] [CrossRef]
  8. Schenck, A.; Daems, W.; Steckel, J. AirleakSlam: Detection of pressurized air leaks using passive ultrasonic sensors. In Proceedings of the 2019 IEEE SENSORS, Montreal, QC, Canada, 27–30 October 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–4. [Google Scholar] [CrossRef]
  9. Exaudi, K.; Passarella, R.; Rendyansyah; Zulfahmi, R. Using pressure sensors towards pipeline leakage detection. In Proceedings of the 2018 International Conference on Electrical Engineering and Computer Science (ICECOS), Las Vegas, NV, USA, 2–4 October 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 89–92. [Google Scholar] [CrossRef]
  10. Jiao, J.; Zhang, J.; Ren, Y.; Li, G.; Wu, B.; He, C. Sparse representation of acoustic emission signals and its application in pipeline leak location. Measurement 2023, 216, 112899. [Google Scholar] [CrossRef]
  11. Yeo, D.; Bae, J.H.; Lee, J.C. Unsupervised learning-based pipe leak detection using deep auto-encoder. J. Korea Soc. Comput. Inf. 2019, 24, 21–27. [Google Scholar] [CrossRef]
  12. Jordan, M.I.; Mitchell, T.M. Machine learning: Trends, perspectives, and prospects. Science 2015, 349, 255–260. [Google Scholar] [CrossRef]
  13. Rane, N.; Mallick, S.; Kaya, O.; Rane, J. Techniques and optimization algorithms in machine learning: A review. In Applied Machine Learning and Deep Learning: Architectures and Techniques; CRC Press: Boca Raton, FL, USA, 2024; pp. 39–58. [Google Scholar] [CrossRef]
  14. Selmy, H.A.; Mohamed, H.K.; Medhat, W. Big data analytics deep learning techniques and applications: A survey. Inf. Syst. 2024, 120, 102318. [Google Scholar] [CrossRef]
  15. Gao, J.; Heng, F.; Yuan, Y.; Liu, Y. A novel machine learning method for multiaxial fatigue life prediction: Improved adaptive neuro-fuzzy inference system. Int. J. Fatigue 2024, 178, 108007. [Google Scholar] [CrossRef]
  16. Meghraoui, K.; Sebari, I.; Pilz, J.; Ait El Kadi, K.; Bensiali, S. Applied deep learning-based crop yield prediction: A systematic analysis of current developments and potential challenges. Technologies 2024, 12, 43. [Google Scholar] [CrossRef]
  17. Chen, J.; Chen, G.; Li, J.; Du, R.; Qi, Y.; Li, C.; Wang, N. Efficient seismic data denoising via deep learning with improved MCA-SCUNet. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5903614. [Google Scholar] [CrossRef]
  18. Plathottam, S.J.; Rzonca, A.; Lakhnori, R.; Iloeje, C.O. A review of artificial intelligence applications in manufacturing operations. J. Adv. Manuf. Process. 2023, 5, e10159. [Google Scholar] [CrossRef]
  19. Waltersmann, L.; Kiemel, S.; Stuhlsatz, J.; Sauer, A.; Miehe, R. Artificial intelligence applications for increasing resource efficiency in manufacturing companies—A comprehensive review. Sustainability 2021, 13, 6689. [Google Scholar] [CrossRef]
  20. Arinze, C.A.; Izionworu; Onuegbu, V.; Isong, D.; Daudu, C.D.; Adefemi, A. Integrating artificial intelligence into engineering processes for improved efficiency and safety in oil and gas operations. Open Access Res. J. Eng. Technol. 2024, 6, 039–051. [Google Scholar] [CrossRef]
  21. Rahimi, M.; Alghassi, A.; Ahsan, M.; Haider, J. Deep learning model for industrial leakage detection using acoustic emission signal. Informatics 2020, 7, 49. [Google Scholar] [CrossRef]
  22. Kwon, S.; Jeon, S.; Park, T.J.; Bae, J.H. Automatic weight redistribution ensemble model based on transfer learning to use in leak detection for the power industry. Sensors 2024, 24, 4999. [Google Scholar] [CrossRef]
  23. Hu, X.; Zhang, H.; Ma, D.; Wang, R. A tnGAN-based leak detection method for pipeline network considering incomplete sensor data. IEEE Trans. Instrum. Meas. 2020, 70, 3510610. [Google Scholar] [CrossRef]
  24. Mysorewala, M.F.; Cheded, L.; Ali, I.M.; Alshebeili, S.; Alotaibi, A.B.; Al-Fakheri, F.M.; Haque, M. Leak detection using flow-induced vibrations in pressurized wall-mounted water pipelines. IEEE Access 2020, 8, 188673–188687. [Google Scholar] [CrossRef]
  25. Ahmad, S.; Ahmad, Z.; Kim, C.H.; Kim, J.M. A method for pipeline leak detection based on acoustic imaging and deep learning. Sensors 2022, 22, 1562. [Google Scholar] [CrossRef] [PubMed]
  26. Pan, S.J.; Yang, Q. A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 2009, 22, 1345–1359. [Google Scholar] [CrossRef]
  27. Chen, X.; Yang, R.; Xue, Y.; Huang, M.; Ferrero, R.; Wang, Z. Deep transfer learning for bearing fault diagnosis: A systematic review since 2016. IEEE Trans. Instrum. Meas. 2023, 72, 3508221. [Google Scholar] [CrossRef]
  28. Guo, L.; Lei, Y.; Xing, S.; Yan, T.; Li, N. Deep convolutional transfer learning network: A new method for intelligent fault diagnosis of machines with unlabeled data. IEEE Trans. Ind. Electron. 2018, 66, 7316–7325. [Google Scholar] [CrossRef]
  29. Sak, H.; Senior, A.; Beaufays, F. Long short-term memory recurrent neural network architectures for large scale acoustic modeling. In Proceedings of the Interspeech 2014, Singapore, 14–18 September 2014; pp. 338–342. [Google Scholar] [CrossRef]
  30. Alawneh, L.; Mohsen, B.; Al-Zinati, M.; Shatnawi, A.; Al-Ayyoub, M. A comparison of unidirectional and bidirectional LSTM networks for human activity recognition. In Proceedings of the 2020 IEEE International Conference on Pervasive Computing and Communications Workshops (PerCom Workshops), Austin, TX, USA, 23–27 March 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–6. [Google Scholar] [CrossRef]
  31. Atef, S.; Eltawil, A.B. Assessment of stacked unidirectional and bidirectional long short-term memory networks for electricity load forecasting. Electr. Power Syst. Res. 2020, 187, 106489. [Google Scholar] [CrossRef]
  32. Siami-Namini, S.; Tavakoli, N.; Namin, A.S. The performance of LSTM and BiLSTM in forecasting time series. In Proceedings of the 2019 IEEE International Conference on Big Data (Big Data), Los Angeles, CA, USA, 9–12 December 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 3285–3292. [Google Scholar] [CrossRef]
  33. Luong, M.T. Effective approaches to attention-based neural machine translation. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing (EMNLP), Lisbon, Portugal, 17–21 September 2015; pp. 1412–1421. [Google Scholar] [CrossRef]
  34. Bae, J.H.; Yeo, D.Y.; Yoon, D.B.; Oh, S.W.; Kim, G.J.; Kim, N.S.; Pyo, C.S. Deep-learning-based pipe leak detection using image-based leak features. In Proceedings of the 2018 IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 2361–2365. [Google Scholar]
  35. Choi, Y.R.; Lee, J.C.; Cho, J.W. A technology of micro-leak detection. In Proceedings of the Korean Society of Computer Information Conference; Korean Society of Computer Information: Seoul, Republic of Korea, 2021; pp. 685–687. [Google Scholar]
  36. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  37. Hornik, K.; Stinchcombe, M.; White, H. Multilayer feedforward networks are universal approximators. Neural Netw. 1989, 2, 359–366. [Google Scholar] [CrossRef]
Figure 1. Structure for RMS and frequency pattern extraction.
Figure 1. Structure for RMS and frequency pattern extraction.
Applsci 15 00490 g001
Figure 2. Features of RMS and frequency patterns for leak and noise signals: (a) RMS pattern feature, (b) frequency pattern features, where the spectrum size and pattern of the leak signal, generated through BPF filtering, are clearly distinguishable from those of the noise signal.
Figure 2. Features of RMS and frequency patterns for leak and noise signals: (a) RMS pattern feature, (b) frequency pattern features, where the spectrum size and pattern of the leak signal, generated through BPF filtering, are clearly distinguishable from those of the noise signal.
Applsci 15 00490 g002
Figure 3. Fusion of 1D pattern features using four sensors for model training.
Figure 3. Fusion of 1D pattern features using four sensors for model training.
Applsci 15 00490 g003
Figure 4. Architecture of a feature fusion DL model based on stepwise TL using an attention mechanism.
Figure 4. Architecture of a feature fusion DL model based on stepwise TL using an attention mechanism.
Applsci 15 00490 g004
Figure 5. Source domain data collection testbed environment using plant piping specimens.
Figure 5. Source domain data collection testbed environment using plant piping specimens.
Applsci 15 00490 g005
Figure 6. Target domain data collection testbed environment using plant piping specimens.
Figure 6. Target domain data collection testbed environment using plant piping specimens.
Applsci 15 00490 g006
Figure 7. Comparison of confusion matrices for model performance with and without TL: (a) results of the proposed model without TL; (b) results of the proposed model with TL.
Figure 7. Comparison of confusion matrices for model performance with and without TL: (a) results of the proposed model without TL; (b) results of the proposed model with TL.
Applsci 15 00490 g007
Figure 8. Performance comparison among various models according to training data size.
Figure 8. Performance comparison among various models according to training data size.
Applsci 15 00490 g008
Table 1. Nomenclature for data preprocessing.
Table 1. Nomenclature for data preprocessing.
Parameter
Notation
Parameter Definition
x n Sampled input data at interval T d
T d Sampling interval for input data
y n RMS level for the n -th sample in P a t h α
C Normalization coefficient for RMS calculation
L n Window size for RMS calculation
y α RMS pattern feature from P a t h α
y * Preprocessed signal after passing through the B P F in P a t h β
W f Weighting function applied in the frequency domain
f L Lower frequency bound for the frequency windowing operation
f H Upper frequency bound for the frequency windowing operation
k Amplification ratio applied within the frequency range f L ,   f H
y * * Signal after applying the amplitude weighting and I F T
F T · 1D Fourier transform operation
I F T · 1D Inverse Fourier transform operation
y β RMS pattern feature from P a t h β
y r m s Final RMS pattern feature after combining y α and y β
y f r e q Frequency pattern feature derived from the frequency response magnitude
Table 2. Detailed structure of single 1D CNN model for both time and frequency domains.
Table 2. Detailed structure of single 1D CNN model for both time and frequency domains.
StageOutput DimensionLayer
Input 40 × 4 Input Layer
1D CNN block (1) 38 × 8 [ 1 × 3 , 8 ]
1D CNN block (2) 17 × 16 [ 1 × 3 , 16 ]
Classifier 1 Global Max Pooling
FC: [ 8 1 ]
sigmoid
# params745
Table 3. Detailed structure of Stage 2 in the proposed leak detection model.
Table 3. Detailed structure of Stage 2 in the proposed leak detection model.
StageOutput DimensionLayer
Input 40 × 4 Input Layer
1D CNN Module (RMS) 17 × 16 [ 1 × 3 ,   16 ]
1D CNN Module (Freq)
Feature Fusion 34 × 16 -
LSTM Forward 34 × 128 LSTM (128 units)
LSTM Backward
Attention Forward128-
Attention Backward
Concatenate256-
Classifier 1 FC: [ 64 1 ]
sigmoid
# params265,521
Table 4. Nomenclature for the proposed leak detection model.
Table 4. Nomenclature for the proposed leak detection model.
Parameter
Notation
Parameter Definition
x t Input data at time step t
x d , n l 1D feature map n of domain d at layer l
w ( d , m , n ) l Kernel function connecting feature maps m and n
b d , n l Bias term for feature map n
f l · Activation function at layer l
N d l Number of output feature maps at layer l
L d Total number of convolution layers in domain d
R n l List of feature maps at layer l 1 connected to feature map n
Z The resulting map obtained by fusing the features derived from the time domain x T and the frequency domain x F
h t f , h s b Forward and backward outputs of the bidirectional LSTM
α t , s i Attention weight for time steps t and s in direction i
c t i Context vector summarizing important sequence information
W a i Weight matrix for calculating attention vectors in direction i
y α Attention vector concatenated from forward and backward directions
f c o n c a t · Feature fusion function that combines the features from the time and frequency domains
Table 5. Comparative analysis of accuracy in detecting low-level pipe leak.
Table 5. Comparative analysis of accuracy in detecting low-level pipe leak.
MethodsAccuracy (%)F1-Score
MLP-Ensemble80.1875.81
SVM-Ensemble85.3683.95
LSTM-Ensemble74.4766.61
1D CNN-RMS (Only Stage 1)98.4298.31
1D CNN-Freq (Only Stage 1)98.9798.92
1D CNN-Ensemble99.3299.33
Proposed Methods99.9999.99
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Han, Y.; Choi, Y.; Lee, J.; Bae, J.-H. Feature Fusion Model Using Transfer Learning and Bidirectional Attention Mechanism for Plant Pipeline Leak Detection. Appl. Sci. 2025, 15, 490. https://doi.org/10.3390/app15020490

AMA Style

Han Y, Choi Y, Lee J, Bae J-H. Feature Fusion Model Using Transfer Learning and Bidirectional Attention Mechanism for Plant Pipeline Leak Detection. Applied Sciences. 2025; 15(2):490. https://doi.org/10.3390/app15020490

Chicago/Turabian Style

Han, Yujin, Yourak Choi, Jonghyuk Lee, and Ji-Hoon Bae. 2025. "Feature Fusion Model Using Transfer Learning and Bidirectional Attention Mechanism for Plant Pipeline Leak Detection" Applied Sciences 15, no. 2: 490. https://doi.org/10.3390/app15020490

APA Style

Han, Y., Choi, Y., Lee, J., & Bae, J.-H. (2025). Feature Fusion Model Using Transfer Learning and Bidirectional Attention Mechanism for Plant Pipeline Leak Detection. Applied Sciences, 15(2), 490. https://doi.org/10.3390/app15020490

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop