Next Article in Journal
TableBorderNet: A Table Border Extraction Network Considering Topological Regularity
Previous Article in Journal
Fault Identification Model Using Convolutional Neural Networks with Transformer Architecture
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Long-Time Series Forecast Method for Wind Turbine Blade Strain with Incremental Bi-LSTM Learning

School of Mechanical Engineering, Xinjiang University, Urumqi 830047, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(13), 3898; https://doi.org/10.3390/s25133898
Submission received: 23 April 2025 / Revised: 27 May 2025 / Accepted: 21 June 2025 / Published: 23 June 2025

Abstract

This article presents a novel incremental forecast method to address the challenges in long-time strain status prediction for a wind turbine blade (WTB) under wind loading. Taking strain as the key indicator of structural health, a mathematical model is established to characterize the long-time series forecast forecasting process. Based on the Bi-directional Long Short-Term Memory (Bi-LSTM) framework, the proposed method incorporates incremental learning via an error-supervised feedback mechanism, enabling the dynamic self-updating of the model parameters. The experience replay and elastic weight consolidation are integrated to further enhance the prediction accuracy. Ultimately, the experimental results demonstrate that the proposed incremental forecast method achieves a 24% and 4.6% improvement in accuracy over the Bi-LSTM and Transformer, respectively. This research not only provides an effective solution for long-time prediction of WTB health but also offers a novel technical framework and theoretical foundation for long-time series forecasting.

1. Introduction

Wind energy, as the most popular renewable energy source, is gradually replacing fossil fuels [1]. Projections indicate an anticipated increase of 1210 GW in installed capacity by 2030, potentially supplying over 36% of global electricity demand [2,3]. Therefore, the research on the technologies related to wind turbine has attracted much attention [4]. As the most critical component of the wind turbine, the wind turbine blade (WTB) has undergone substantial development, with current designs reaching 90–150 m and exceeding 30 t [5]. Such a huge size and weight create significant health concerns. Furthermore, continuous exposure to harsh environmental conditions subjects WTBs to alternating loads and surface erosion, increasing vulnerability to multiple failure modes including surface cracking, abrasion, icing, and catastrophic breakage. According to the research by Wen [6], failure of the WTB represents the primary cause of downtime, accounting for 34%. In addition, the maintenance cost of WTBs accounts for about 20% of the total equipment maintenance costs [7,8]. Therefore, it is important to accurately forecast the health status of the WTB as early as possible in order to make reasonable maintenance decisions in advance.
At first, the structural response analysis and fatigue damage mechanisms are employed to forecast the health status of the WTB. For example, Meng [9] analyzes the fatigue life of a WTB based on anisotropic beam model and stress-based fatigue method. Caous [10] uses ply scale damage model to evaluate WTB fatigue life. Li [11] proposes a method for a combined high and low cycle fatigue life prediction model based on Miner’s rule. How-ever, the accuracy of these forecast methods is compromised by environmental variability.
Subsequently, the method based on data analysis have received lots of attention. The commonly employed carrier for WTB health information is the SCADA data. For example, Lu [12] proposes a small-sample forecast method via a combination of multiple neural networks based on SCADA data. Antoine [13] evaluates the WTB damage under multiple conditions based on SCADA data. But the information reflected via SCADA is too vague for large-scale WTBs because it is low-frequency and indirect. Strain is the physical quantity that most directly reflects the status of a structure, and the techniques based on strain detection have been widely employed to monitor the health status of WTBs. For example, the research studies [14,15,16] evaluated the performance of WTBs via monitoring strain changes. However, the time course of WTB strain signals tends to exhibit nonlinear, periodic and unstable characteristics during the degradation process of WTBs. Therefore, how to find out the change rule of data characteristics through the highly complex strain information is the most difficult problem in the prediction of the health status of WTBs.
In order to address the above problem, the machine learning and intelligent algorithms are gradually employed to forecast the health status of WTBs. For example, Deng [17] forecasts the deformation of WTBs via combining of machine learning and mechanism analysis. Liu [18] employs deep learning to forecast the stiffness of the WTB. Choe [19] pro-poses a sequence-based modeling for structural damage detection of the WTB using Long Short-Term Memory and Gated Recurrent Unit neural networks. Although these methods can accurately forecast the short-term health status of a WTB, they perform poorly in the long-term. Unfortunately, the demand for forecasting the future health status of WTBs as early as possible poses a serious challenge to long-time series forecast method.
This study proposes a long-time series forecast method in order to improve the prediction accuracy of strain. Compared to other networks currently, the Bi-LSTM network is the most mature in solving time series problems and it receives the widest application because of its stable performance. Therefore, the Bi-LSTM network is chosen as the basic network for the long-time series forecast method proposed in this article, and the incremental learning is adopted to improve the long-time forecasting performance.

2. Theoretical Background

2.1. Problem Statement

The health degradation of a WTB is a gradual process over time. Then, the forecast of the health status of a WTB via strain data is a time series forecast problem. The idea of using neural networks to solve the time series prediction problem is originally proposed by Weigend in International Journal of Neural Systems [20]. In this idea, the input to the model is defined as a historical vector in a sliding time window, and the output is defined as a future value. A time series with length L and dimension n can be recorded as historical data X = ( x 1 , x 2 , , x L ) R L × n , and a movable retrospect window with length e is denoted as X i , e = ( x i e + 1 , x i e + 2 , , x i ) . It is the goal of time series forecast to estimate the series in the future period after time i based on the latent regularity mined via the retrospect window from the historical data before time i. The length of this future period is known as the forecast window and is denoted as H. Then the mathematical expression of the time series forecast is
f o r e c a s t v a l u e s = f ω ( i n p u t v a l u e s ) ( x i + 1 , x i + 2 , x i + H ) = f ω ( X i , e ) ( x i + 1 , x i + 2 , x i + H ) = f ω ( x i e + 1 , x i e + 2 , x i )
where ω is the parameters of the forecast model. It is required to minimize the error between the forecast value and the true value throughout the forecasting process.
In the practical applications, the data arrive sequentially as continuous streams, as shown in Figure 1. The online time series forecasting task necessitates incremental learning new knowledge from a continuous input data stream, and the training and prediction of the model occur alternated. Since the current health status exhibits temporal dependencies on previous states, comprehensive historical data must be considered when predicting future conditions. However, historical data are constantly increasing, and the huge amount of data will be accompanied by high computing cost and computational time. Consequently, another tricky issue in forecasting the health of WTBs is how to reduce memory occupation while ensuring high accuracy.

2.2. Bi-LSTM Networks

The LSTM network, originally proposed by Hochreiter in Neural Computation [21], has unique capabilities which make it very suitable for time series forecasting problems. The LSTM network allows the entire network to have longer feature memory via designing a unique gate structure in every cell. In addition, the unique gate structure enables the LSTM network to discard the invalid components of the transmission data from previous neuronal and retain useful information effectively. Thus, the LSTM network shows good stability in time series prediction. In order to improve the forecast accuracy, the Bi-LSTM network is adopted as the fundamental framework of the model to forecast the health status of WTBs.
The LSTM network is essentially an improved RNN. As shown in Figure 2, the cellular structure inside the LSTM network mainly consists of the conveyor belt, forget gate, input gate and output gate [22].
The method of updating the cellular status of the LSTM network can be expressed as follows in Equation (2):
F t = f s i g ( W f [ h t 1 , x t ] + b f ) I t = f s i g ( W i [ h t 1 , x t ] + b i ) C ˜ t = f tanh ( W c [ h t 1 , x t ] + b c ) C t = F t C t 1 + I t C ˜ t O t = f s i g ( W o [ h t 1 , x t ] + b o ) h t = O t f tanh ( C t )
where Ft, It, C ˜ t , Ct, Ot, ht and xt are, respectively, forget gate vector, input gate vector, new value vector, status vector, output gate vector, output vector and input vector. W is the weight coefficients matrix, b is the bias coefficient vector, and f is the activation function. It is worth noting that all parameter matrices in the model are independent of each other and need to be trained via back propagation.
In order to further improve the memory ability, the Bi-LSTM architecture, initially introduced by Graves in Neural Networks [23], incorporates an additional reverse direction propagation path to the classical LSTM framework, as shown in Figure 3.
The forward and backward propagation paths operate independently with distinct model parameters. In the forward path, the output h t F gradually forgets the earliest inputs at the left. Conversely, the output h t B of the backward propagation path gradually forgets the earliest inputs at the right. The forward propagation and back propagation for updating the cellular status of the Bi-LSTM network can be expressed as Equation (3) and Equation (4), respectively.
F t = f s i g ( W f [ h t 1 , x t ] + b f ) I t = f s i g ( W i [ h t 1 , x t ] + b i ) C ˜ t = f tanh ( W c [ h t 1 , x t ] + b c ) C t = F t C t 1 + I t C ˜ t O t = f s i g ( W o [ h t 1 , x t ] + b o ) h t = O t f tanh ( C t )
F t = f s i g ( W f [ h t 1 , x t ] + b f ) I t = f s i g ( W i [ h t 1 , x t ] + b i ) C ˜ t = f tanh ( W c [ h t 1 , x t ] + b c ) C t = F t C t 1 + I t C ˜ t O t = f s i g ( W o [ h t 1 , x t ] + b o ) h t = O t f tanh ( C t )
The output is obtained via combining h t F and h t B , as shown in Equation (5). Thus, the problem of forgetting the earliest information is solved. It is noteworthy that the combination of h t F and h t B is achieved via vector concatenation. The dimension of the new vector H t is the sum of the dimensions of h t F and h t B .
h t F = L S T M F ( h , x , c ) h t B = L S T M B ( h , x , c ) H t = [ h t F , h t B ]

2.3. Incremental Learning

In the engineering of WTB health status prediction, the target often occupies a long-time span in the future. Forecasting the future strain as early and accurately as possible is of great engineering importance for the predictive operation and maintenance of WTBs. However, it is still limited in a short and fixed time span to accurately forecast the future data via Bi-LSTM network. Therefore, this research introduces an incremental learning strategy to construct a forecast model that can self-update dynamically in order to enhance the long-time forecasting performance based on the Bi-LSTM network.
Incremental learning [24,25], also referred to as continuous learning, is a learning strategy designed for sequentially arriving data. Consequently, the forecast model can gradually adapt to the changes of the external information environment. Within neural network systems, incremental learning enables continuous new knowledge learned from a data stream while preserving memory learned previously. As the new data become available over time, the pre-trained forecast model learns the new knowledge in real time under this strategy. In the process of incremental learning, the prior knowledge is conducive to improving the learning efficiency for new knowledge, while the new data are conducive to improving the forecasting performance in the new information environment. Therefore, incremental learning has the following advantages compared to traditional learning strategies: (1) Improve prediction accuracy via continuously and dynamically optimizing model parameters based on new data. (2) Avoid retraining the model from scratch using a large amount of historical data, and instead utilize a small amount of new data to reduce and memory occupation. Figure 4 illustrates the difference between incremental and traditional learning strategies.
The primary objective of incremental learning is to update the model parameters via gradually accepting new data samples without retraining the entire model. Particularly in the prediction of WTB strain, incremental learning permits the forecast model to automatically adapt the model parameters according to accumulating strain data.

3. Proposed Incremental Health Status Forecast Method

3.1. Overall Architecture of Method

In order to address the two critical challenges mentioned in Section 2.1, this research presents an incremental forecast method based on Bi-LSTM networks and incremental learning theory. The complete workflow of this incremental forecast method is shown in Figure 5.
The WTB strain data are continuously acquired via sensors, forming a data stream that serves as input to the incremental Bi-LSTM model. This model is the computational core in the health status forecast method. It is worth noting that the incremental Bi-LSTM model at this point is not in a fully initialized state, but has been trained previously via historical data. After the strain data stream of the current period passes through the model, the model generates the WTB strain information for the future period and output the strain forecast value. When the strain forecast value reaches the warning threshold, it is identified that fatigue failure will occur at that moment. This warning threshold is commonly determined via fatigue testing of the WTB and historical experience. Usually, a WTB is considered to fail when its stiffness degrades by more than 10% [26]. Therefore, the warning threshold of strain can be set to 110% of the initial strain at the same wind speed.
Over time, the measured value at future moments corresponding to the forecast value will be measured via the strain sensors installed on the WTB. The measured value can be employed to check the accuracy of the forecast value. The process is a biased estimation problem based on small sample, so this research adopts the root mean square error (RMSE) to evaluate the deviation between the forecast value and the actual value. The formula for RMSE is as Equation (6),
RMSE = 1 n i = 1 n ( y i y ^ i ) 2
where y ^ i is the ith forecast value, y i is the ith actual value. If the RMSE is less than the threshold, the forecast value is considered accurate and credible, and the new measured data are subsequently stored in a data buffer. According to engineering experience, 10% of the mean of the training dataset is selected as the error threshold in this article. It is important to emphasize that the threshold for deviation judgments is not the same one as the warning threshold for critical state of health.
A normal-off gate is designed to regulate the data flow between the data buffer and the incremental Bi-LSTM model. As the new data arrive continually, errors continue to accumulate and the deviation between forecast value and the actual value will become larger. When the RMSE is greater than the threshold, the forecast value is considered to be unreliable. Then, the channel between the data buffer and the incremental Bi-LSTM model is activated. The new data stored in the data buffer are fed into the incremental Bi-LSTM model to incrementally train the historical model. The design of the normal-off gate structure with error judgment and the data buffer realizes the adaptive updating of the incremental Bi-LSTM model. The updated forecast model can continuously output accurate and credible forecast values. The pseudocode for the forecast process is shown in Figure 6.
The incremental Bi-LSTM model, as the key component of the WTB health status forecast method, is mainly composed of two modules: the experience replay module and the parameter regularization module. The experience replay module generates an incremental sample set via feature relevance computation for subsequent incremental training. The parameter regularization module adopts the elastic weight consolidation (EWC) method to regulate the parameter updates of the historical model during incremental training. The updated forecast model has the historical priori knowledge while adapting to the new information environment and can be employed for forecasting subsequent strain value. The experience replay module and the parameter regularization module are described in detail in the following sections, respectively.

3.2. Incremental Learning Based on Experience Replay

With the degradation of the WTB health status, the characteristics of new strain data progressively diverge from historical patterns. Then, historical knowledge will be covered when the model learns new knowledge via the new data. This is the key challenge to incremental learning, namely catastrophic forgetting. The lack of historical knowledge can degrade the forecasting performance of the model. In order to mitigate catastrophic for-getting, this research integrates the experience replay [27] technology into the incremental learning. As shown in Figure 7, the core idea of the experience replay technology is to add historical samples to the new samples, forming an incremental sample set. In addition to the new sample information, the incremental sample set also carries the characteristics of the historical task. In this way, the historical samples participate in the incremental training of the model, enabling to effectively prevent forgetting previous knowledge. However, incorporating the complete historical dataset would incur prohibitive computational cost and increase risks of overfitting or convergence difficulties. Consequently, selectively incorporating representative historical samples into the incremental sample set presents a practical approach.
A new problem arises as how to store more important samples in the limited memory space. Typically, classical experience replay technology uses random sampling with uniform probability to select the samples from a large number of historical samples. However, the method ignores the differential contributions of individual historical samples to the model performance. In this research, the K-Means clustering algorithm is employed to select representative samples from a large number of historical samples because of its computational efficiency and good clustering results. The K-Means clustering algorithm firstly divides the historical samples into several feature clusters based on the distance between the feature vectors. The distance of each sample point in the dataset from each cluster center is calculated according to Equation (7), and each sample point is assigned to the cluster center with the closest distance;
d i s t ( x , c i ) = j = 1 d ( x j c i j ) 2
where x is sample point, ci is the cluster center, d is the dimensions of data, xj and cij are the values on j, respectively. Then, the new clustering center is recalculated based on each feature cluster as Equation (8):
c i = 1 S i x S i x
where Si is the set of sample points for the ith cluster.
Subsequently, some samples closest to the center of each feature cluster are selected as representative samples. These representative samples carry historical knowledge and reflect the overall characteristics of the historical sample. The representative samples are combined with new samples to form the incremental sample set.
The cross entropy is usually employed as the loss function for training historical models on incremental sample sets, providing the optimization objective for updating of forecast model parameters. However, the imbalance in the number of new and representative samples induces significant gradient instability, which can cause drastic changes in the model parameters. In order to address this issue, the losses of the representative and new samples are calculated separately. This separation eliminates direct comparison between historical and new samples. The final loss function is defined as the sum of the cross entropies of the representative and new samples, as shown in Equation (9);
L m o d e l ( θ ) = L h i s ( θ ) + L n e w ( θ )
where θ is the model parameters of the current status, and L m o d e l , L h i s , and L n e w are the final loss of the model, the loss of the representative sample and the loss of the new sample, respectively.

3.3. Parameters Regularization via Elastic Weight Consolidation

Although the representative samples mentioned above are effective in expressing the overall characteristics of historical samples, some samples located at the boundaries of feature clusters are equally crucial to historical knowledge. It may also affect the fore-casting accuracy to ignore these boundary samples. Unfortunately, the identification of boundary samples is difficult and computationally expensive. Therefore, in addition to optimizing the selection of historical sample, other technical methods need to be incorporated to further improve the learning performance of the model.
The fundamental mechanism by which training samples influence model performance lies in the differential informational contribution of sample features to parameter updates. Certain parameters exhibit high sensitivity to variations in the sample. Conversely, other parameters have less information from the historical samples and the knowledge already learned cannot be covered when updating these parameters. Thus, the problem of forgetting knowledge can be mitigated via imposing regularization on the highly sensitive parameters. Based on this idea, the EWC is incorporated into the training process on new samples to compensate for the neglect of boundary samples, as shown in Figure 6.
The EWC [28] employs the Fisher information matrix to impose regularization on the updating of model parameters based on varying importance, because the Fisher information matrix is a good measure of the sensitivity of the model parameters. Specifically, the EWC redefines the loss of a new sample as the sum of the cross entropy of the new sample and the total deviation of all the parameters of the model, as shown in Equation (10);
L n e w ( θ ) = L n e w * ( θ ) + λ 2 i Ω i θ i * θ i 2
where L n e w * ( θ ) is the cross entropy of the new sample, which is numerically equal to that in Equation (9). θ i and θ i * are the model parameters before and after the update, respectively. Ω i represents the importance of the model parameters. λ is the regulating factor.
The second term in Equation (10) is the regularization term demonstrating that the parameter with high weight should be close before and after updating in order to minimize the overall loss. The knowledge learned from the historical samples in the model is retained via the regularization.
The regularization term can be obtained based on Bayesian theory. The strain data of the WTB is noted as D = D h i s , D n e w , where Dhis and Dnew are historical and new data, respectively. In the Bayesian perspective, the goal of training the model parameters changes from minimizing the loss function to maximizing the posterior probability of the model parameters log p θ | D . Then, the parameters of the model should obey the posterior probability distribution:
log p θ | D = log p D h i s , D n e w | θ + log p θ log p D h i s , D n e w = log p D h i s | θ + log p D n e w | θ + log p θ log p D h i s log p D n e w = log p D n e w | θ + log p θ | D h i s log p D n e w
In Equation (11), the first log probability log p D n e w | θ can be viewed as the negative of the loss of the new sample, i.e., L n e w * ( θ ) in Equation (10). The third term log p D n e w is a constant independent of the parameters and can be disregarded. The second term log p θ | D h i s is written into functional form as Equation (12);
p θ | D h i s = 1 Z f ( θ )
where Z is normalization factor. The Taylor expansion of its likelihood function is then performed as Equation (13);
ln f ( θ ) = ln f ( θ * ) + J ln f ( θ * ) ( θ θ * ) + 1 2 ( θ θ * ) T H ln f ( θ * ) ( θ θ * ) + R n ( θ )
where θ i * is the parameter trained previously, H ln f is the Hessian matrix. The first order terms and terms above the third order are approximated as 0 because the loss is small in the model trained. Thus, only the second order and zero order terms are retained as Equation (14).
ln f ( θ ) ln f ( θ * ) 1 2 ( θ θ * ) T F ln f ( θ * ) ( θ θ * ) F ln f = H ln f ( θ * )
Since the Fisher information matrix is the negative of the expectation of the Hessian matrix of the log-likelihood function, the Fisher information matrix is used instead of the Hessian matrix for convenience of calculation. Taking the index of Equation (14) yields Equation (15).
f ( θ ) 1 Z f ( θ * ) e 1 2 ( θ θ * ) T F ln f ( θ * ) ( θ θ * )
Therefore, the log p θ | D h i s can be regarded as a Gaussian distribution obeying μ = θ i , 1 σ 2 = F ln f ( θ ) . Then, the maximization of log p θ | D h i s is equivalent to maximization of 1 2 ( θ θ * ) T F ln f ( θ * ) ( θ θ * ) , and Equation (10) can ultimately be ex-pressed as
L n e w ( θ ) = L n e w * ( θ ) + λ 2 i F i θ i * θ i 2 F i = L n e w * ( θ i * ) θ i * 2
It should be emphasized that health status prediction of the WTB is an online time series forecasting task. A new training task is initiated upon activation of the normal-off gate described in Figure 5. As the number of history tasks continues to increase, the EWC calculates a regularization term for all history tasks. Then, the regularization term grows linearly with the data flow, resulting in a large computational cost. Since each new task is trained via a regularization term to the previous history task, the regularization term of the previous task can contain the constraints of all previous history tasks. Then, the weighted sum of the Fisher information matrix of all historical tasks is performed, and only the parameters of the previous one task need to be regularized. Therefore, the online form of Equation (16) can be expressed as
L n e w ( θ ) = L n e w * ( θ ) + 1 2 i t < n e w λ t F t , i θ i * θ n e w 1 , i 2
Equation (17) can retain the information characteristics of historical samples as much as possible, so as to achieve more accurate online prediction for the health status of the WTB.

4. Experiments and Data Analysis

4.1. Load Scheme Design

In order to validate the proposed long-time series forecast method, the WTB strain measurement experiment is designed in this research. The strain data of the WTB under simulated wind load in experimental conditions is collected, and is used as the time series dataset to validate the proposed method.
This research collects the wind speed data spanning 168 h in a week from a wind farm in Darbancheng, Xinjiang, in order to guide the load scheme design. Each wind speed value is sampled as the average wind speed value during the current 1 min time period. Invalid data points in the data are eliminated, such as exceeding the maximum operating wind speed. The final wind speed data time series contains 6500 wind speed sampling points, as shown in Figure 8.
Ideally, wind speed data should be set as a load spectrum and then loaded on the WTB via a wind tunnel test. However, the large-scale wind tunnel experimental conditions are often demanding and costly. Consequently, researchers often design concentrated loads instead of wind loads in the WTB tests, as demonstrated in other peer studies [29]. The experiments in this research require precise control of the loading history based on the wind speed data. Because it has the advantage of being able to respond accurately and quickly to irregular fluctuations in load, the forced displacement loading method is finally selected to simulate the wind loads after comprehensive consideration.
Furthermore, given that each wind speed sampling point represents the average value in one minute, the corresponding load should not be an instantaneous load, but a sustained load over a period of time. Therefore, in order to simulate the actual wind load, the loading of each wind speed sampling point should continue for a period of time be-fore transitioning to the subsequent wind speed sampling point. Notably, even if the wind speed is constant in actual conditions, the WTB does not keep a static deformation, but vibrate repeatedly within the range of the maximum deformation corresponding to the wind speed (only the flap direction is considered in this research). Accordingly, the WTB is designed to repeatedly cycle within a range of deformation corresponding to each wind speed sampling point. The duration is defined as the flap-period of a single wind speed sampling point. The repeated cycles during each flap-period are defined as flap-cycles. Then, it proceeds to the next flap-period of wind speed sampling points until the loading history of all wind speed sampling points is completed.

4.2. Data Acquisition

After designing the loading scheme, this research adopts a servo motor as the controllable power output, a crank-slider mechanism as the load transfer mechanism, and a microcontroller to control load fitting the wind speed data. The overall schematic of the experimental loading equipment is shown in Figure 9.
In order to ensure smoother operation of the whole equipment, a one-stage planetary reducer is installed between the servo motor and the crank to increase the input torque. The crank, as the original moving part of the load transmission mechanism, converts the initial circular motion into a linear reciprocating motion via the crank slider mechanism. The slider is fixed on the WTB via the clamp to realize the forced displacement loading on the WTB. The displacement x of the loading mechanism can be determined by Equation (18);
x = r cos α + l 2 r 2 sin 2 α
where α is crank rotation angle, α = ω t . ω and t are the crank angular velocity and time, respectively, r and l are the length of crank and connecting rods, respectively.
A 2.4 m scaled WTB is used for strain measurement experiment. The unidirectional deformation value of the WTB corresponding to the maximum wind speed data is 7.21 cm, which is calculated via the fluid–structure coupling simulation of rotating impellers in finite element software. Accordingly, the crank length is set to 7.21 cm, then the maximum displacement stroke is 14.42 cm. The crank angulars corresponding to the maximum and minimum limit positions are 0° and 180°, respectively. The displacements corresponding to different wind speed are fitted based on the deformation values of the WTB obtained from the simulation, then the crank angular corresponding to a certain of wind speed is calculated via Equation (18). During the experiment, the initial position of the crank was set at 90° and the flap-period corresponding to each wind speed sampling point is set to 3 s. The “90° to maximum angle to 90° to minimum angle to 90°” is a flap-cycle lasting 1 s, and the crank undergoes a total of 3 cycles in each flap-period.
According to the wind speed course shown in Figure 7 and the above loading scheme, a control program for motor shaft rotation angle is written and embedded into the microcontroller to control the servo motor to complete the entire loading course. Notably, the crank angle is the motor output shaft value after deceleration via the reducer.
The servo motor used in the experiment is a Hongsen AC servo motor, and the model is JXE-751K30-BJ-B (made in China), with a power of 750 W and a rated output torque of 2.4 Nm. The reducer is a First single-stage planetary reducer, and the model is WLF080 (made in China), with a gear reduction ratio of 10, maximum output torque of 40 Nm, and maximum input speed of 6000 rpm. The crank-slider mechanism is designed and machined in-house. The WTB is fixed on the experimental fixture with flanges in the form of end restraints. The experimental site for strain status detection of the WTB under simulated wind load is shown in Figure 10.
The entire length of the WTB is defined as R. Three strain sensors, which are employed to measure and record the WTB strain data, are installed at 0.4R, 0.5R, and 0.6R along the WTB span direction with epoxy resin adhesive, and numbered as strain sensor 1, strain sensor 2, and strain sensor 3, respectively. These positions have received attention in many studies [30,31]. The sensors are resistance strain gauges and Table 1 describes the strain sensors models and characteristics. The strain data were sampled at a frequency of 1000 Hz.
In addition, the host computer used in the experiment is a personal laptop (PC) with Windows 7 64-bit operating system and a CPU of Intel Core i5-3230M. The data acquisition model is DT9857 (made by Measurement Computing Corporation in USA) and the program accompanying Quick DAQ (DAQami 3.2) is used as software. Model calculations are done under MATLAB R2023b software.

4.3. Signal Preprocessing and Analysis

The preprocessing flow of signal is shown in Figure 11. The data acquired via the sensors are a one-dimensional time series of voltage values. Firstly, the direct current drift in original signal is removed via mean subtraction in order to correct the baseline.
The signal contains unavoidable noise, which mainly originate from the current interference inherent within the system, such as switching power supply noise, grid harmonics, and inductive coupling. The frequency of the grid harmonics is 50 Hz, which is the lowest of the above noises. The frequency of the WTB strain is below 10 Hz; therefore, a low-pass filter can be used for noise reduction. The 1st order Butterworth filter is adopted because it preserves the amplitude accuracy better and avoids phase delays. Typically, the cutoff frequency is greater than twice the target frequency and less than the noise frequency. Four frequencies, 25, 30, 35, and 40 Hz, are compared, and the best 35 Hz is selected as the cutoff frequency.
After the noise reduction process, every voltage value is converted to strain value according to Equation (19);
ε = 4 × U × 10 6 K × 5.02 × 1 × 405.321
where U is voltage value and K is sensitivity coefficient of sensors.
The strain data is a row vector with dimensions 1 by 19,380,000 and it is shown in Figure 12 (taking strain sensor 1 as an example). The absolute value of the strain in the figure reflects the magnitude of strain, and the sign reflects the direction of the load on the WTB, with tensile strain as positive and compressive strain as negative. According to the alternating changes in the peak and valley values of the strain, it is evident that the strain gauges are repeatedly subjected to tensile and compressive loads as a result of the continuous vibration of the WTB in the flap direction under the linear reciprocating load. The strains in the figure exhibit approximately symmetrical distribution about the baseline, indicating that the flap deformation of the WTB always remains symmetry in experiments. During the initial 20 min, the baseline gradually increased from 0 to about 1.7 × 10−5. This phenomenon is not the zero-point drift caused by the measurement system, but the real change of the WTB strain state. The WTB enters the rapid degradation period at the beginning of vibration. The tensile stiffness decreases rapidly and causes the tensile strain to become larger and the compressive strain to become smaller. After about 20 min, the WTB enters the period of stabilization, during which the stiffness remains stable, and the tensile and compressive strains also remain stable. In other words, the change of baseline reflects the fatigue process of the WTB.
Analysis of the strain data in the locally magnified graph around the 28th minute reveals a consistent pattern where every three continuous peaks (flap-cycles) constitute one complete flap-period. The phenomenon aligns with the load design above, wherein the flap-period corresponding to a certain wind speed consists of three flap-cycles.
In order to observe more clearly the process of strain values following the load, the strain data are subjected to envelope analysis and the extracted peak envelope is shown in Figure 13. The horizontal axis of the graph is the time course, the blue line represents the wind speed data, and the red line represents the strain data. It can be clearly found that the strain data show a strong positive correlation with wind speed variations. The maximum peak of wind speed occurs at approximately the 28th and 250th minutes, and accordingly, the strain data similarly reach the maximum values at the 28th and 250th minutes. This consistent temporal alignment is evident both in the global trend analysis and localized short-term observations, confirming that strain fluctuations faithfully track wind speed changes throughout the entire monitoring period. As the wind speed declines to its trough, the resulting load and deformation on the WTB diminish, causing the strain value to decrease correspondingly to a local minimal value. Conversely, when the wind speed rises to its peak, the increased load and deformation strive on the strain value to a local maximum.
In summary, the strain data collected in this experiment are highly fitted to the wind speed. It can be assumed that the collected strain data capture the characteristics of the simulated wind speed. It shows that it is feasible and effective to use this data as a dataset to test the proposed long-time series forecast method under wind load.

5. Results and Discussions

5.1. Forecast Results Analysis

The first 60% of the strain dataset in time series is employed as the training set and the last 40% serves as the test set. During training the model via the training set, every 100 consecutive data points are taken as a set of inputs, and the 10 consecutive data points following immediately are taken as the corresponding prediction targets. In other words, the length of retrospect window is 100 and the length of forecast window is 10. The Table 2 shows the prediction accuracy for different lengths of retrospect window. The model performs best at the length of 100 and decreases significantly when the length is greater than 200.
Crucially, preserving temporal order is fundamental for time series forecasting. Both input and target vectors must maintain strict chronological alignment of their constituent elements to ensure the integrity of temporal dependencies in the data.
Consistent with the training, the lengths of the retrospect and forecast windows are 100 and 10, respectively, during the testing. In the conventional testing method, each input sequence operates independently, with no temporal dependencies between successive test samples. The evaluation remains unaffected even if the original chronological order of test input sequences is randomized. In practical applications, however, only the elements contained in the first set of inputs are known, while all subsequent inputs represent future unknown states.
Consequently, this research adopts an alternative testing method that more accurately reflects the actual engineering problems, departing from conventional approaches. The testing process iteratively incorporates the forecast values into new test inputs to advance the forecast window stepwise. In this way, the propagation error is cumulative and will have some influence on the subsequent prediction. It is more in line with future unknown problems in actual engineering. Therefore, in the testing phase of this research, only the first set of actual values in the test set are employed as inputs, while the other actual values are only employed as the corresponding labels to evaluate the forecast accuracy.
It is assumed that the strain in the first 60% has already occurred while the strain in the last 40% has not yet occurred, which is unknown future data. The incremental forecast model is trained with the first 60% of strain data to learn the changing regularity of strain. Then, the model forecasts the last 40% of the strain data. The forecast results are shown in Figure 14. The vertical axis of the figure is the strain value and the horizontal axis is the index of the data series point. The results of three different datasets, 0.4R, 0.5R, and 0.6R, are distributed from top to bottom. The black lines are the actual strain values measured from sensors and the red lines are the forecast strain values computed from model. The first 60% is the training phase and the last 40% is the testing phase. It can be found that the three datasets exhibit consistent temporal patterns despite minor variations in strain values. After training, the forecast values of the model fit well with the actual values. Some relatively significant errors occur only at some peaks, valleys where strain values change abruptly.
In order to better analyze the forecasting performance of the model, the root mean square error (RMSE) between the forecast and actual values during the forecast phase is tracked to quantify the forecast accuracy, as shown in Figure 15. The RMSE counts the current moment as well as all previous errors, and is able to reflect the effect on forecast values from accumulation of errors over the time. The RMSE in figure remains at a small value overall, indicating that the model has a high accuracy in prediction. The high RMSE in the short period of time at the beginning is due to the fact that the small amount of data only reflects limited information as a statistical indicator. As the number of forecast values increases, the RMSE decreases rapidly to lower values and is maintained over time. There are two significant bumps which indicate a larger deviation in the current forecast values due to the effect of the errors from the previous forecast values. Larger slopes indicate that the error accumulates rapidly and has a great impact on the subsequent forecast values. The subsequent decrease after the bump is due to the fact that the model detects an anomaly in the error and initiates the incremental learning to dynamic update the parameters. The forecast accuracy of the updated model returns to a higher level, and the RMSE gradually decreases. It is clear that the incremental learning mechanism of the model is able to dynamically update the parameters according to the arrival of new data, and thus maintain a high forecast accuracy over time.

5.2. Comparative Analysis

Comparative analysis is an essential part of evaluating the proposed forecast method. According to the operational mechanisms, recurrent neural networks (RNN) and self-attention are the two most dominant classifications in current forecast model research. One of the classical representatives of RNN is Bi-LSTM, while the classical representative of Self-attention is Transformer, whose performance has a significant improvement over RNN. Therefore, the above two models are selected as the reference representatives for RNN and Self-attention, respectively, to compare and discuss the performance with the incremental forecast method proposed in this research.
Take the strain data on 0.4R as an example; Figure 16 shows the test results of the different methods. In each of these figures, the forecast results of Bi-LSTM, Transformer and incremental forecast methods are in order from the top to the bottom, with the black lines being the actual strain values measured from sensors and the red lines being the forecast strain values computed from model.
It can be seen that the forecast values of the three methods can keep the same change regularity with the actual values. However, the three methods show significantly different performance in the final stage. Since the accumulation of errors in the early stage has a serious impact on the prediction accuracy, it can be clearly observed that both the Bi-LSTM and Transformer show a significant deviation between the forecast and actual values in the final stage. Conversely, the incremental forecast method can effectively eliminate the effect of error accumulation via the error-supervised feedback mechanism. As a result, the incremental forecast method exhibits superior prediction accuracy in the final stage.
In addition, the Bi-LSTM shows a significant lag near the peak mutation. This is due to the inevitable gradient vanishing, which leads to the forgetting of some historical features and short-term memory struggling to respond to the data mutations in time. This phenomenon is significantly improved because the model structure of the Transformer and incremental forecast method can capture historical features better than Bi-LSTM. Moreover, the incremental forecast method outperforms the Transformer.
In this research, the coefficient of determination R2, which is widely used to evaluate regression models, is adopted to further quantify the forecast performance of the three methods. When its value is closer to 1, it indicates that the data fit better. Whereas, its value tends to 0 when the model does not capture the characteristics of changes in the data and performs poorly. The formula for R2 is as follows:
R 2 = i = 1 n ( y i y ^ i ) 2 i = 1 n ( y i y ¯ ) 2
where y ^ i is the ith forecast value, y i is the ith actual value and y ¯ i is the mean value of actual values.
The R2 of the three methods on each of the three datasets is shown in Figure 17. It is clearly seen that the incremental forecast outperforms the other two methods on all three datasets, with R2 values being 0.929 (mean of 0.4R, 0.5R and 0.6R). Additionally, the R2 values of Bi-LSTM and Transformer are 0.749 and 0.889, respectively. According to the above, the forecast accuracy is improved by 24% and 4.6% compared with Bi-LSTM before improvement and Transformer. Although the performance of the Transformer lightly decreases compared to the incremental forecast, it has a significant improvement compared to the Bi-LSTM. The above fully illustrates that although the Bi-LSTM performs the worst among the three methods, the incremental forecast method based on the Bi-LSTM is able to substantially improve the long-time forecast, even surpassing the Transformer. The reason for this phenomenon is the unique self-updating structural designed in the proposed incremental forecast method.
The inference time and memory usage of the above three methods are shown in the Table 3. The performance of the incremental forecast method is obviously optimal.

5.3. Ablation Analysis

In order to analyze the preprocessing sensitivity on prediction accuracy, the preprocessed and original data are input into the incremental forecast model under the same model settings, respectively. The R2 of the forecast results are shown in the Table 4. Obviously, the preprocessing improves the prediction accuracy by 2.9%.
The K-means clustering during experience replay and the EWC regularization during training are two very important designs for the proposed incremental forecast architecture. In order to validate the contribution of both for the proposed method, an ablation experiment is designed to analyze the effects of K-means clustering and EWC regularization on the performance of proposed method, respectively.
The ablation experiment is designed with four schemes: Scheme 1 is a complete incremental forecast method. Scheme 2 only retains the EWC regularization, and the K-means clustering is instead of the conventional uniform probability of random sampling. Scheme 3, opposite of Scheme 2, retains only the K-means clustering and removes the EWC regularization during training. Scheme 4 further eliminates the K-means clustering on the basis of Scheme 3, i.e., only retains the initial incremental learning process. The three datasets are entered into the model for each of the four schemes described above, with all settings remaining the same as before.
The R2 of the forecast results for each group of schemes is calculated to evaluate the performance, as shown in Figure 18. Scheme 4 substantially improves the R2 score by almost 10% compared to that without incremental forecast, Bi-LSTM in the above section, demonstrating the importance of the incremental learning structure. However, Scheme 4 performs the worst in ablation comparisons. The fitting effects of Scheme 2 and 3 are relatively close to each other, indicating that the EWC regularization and the K-means clustering, contribute equally to the improvement of forecast accuracy, with the former being slightly higher than the latter. The combined effect of the two will further improve the forecast performance to the level of Scheme 1. All three datasets exhibit the same phenomena.
The error statistics for all forecast results are shown in Table 5. The mean absolute percentage error, MAPE, reflects the overall relative difference between the forecast and actual values. The mean absolute error, MAE, provides the most intuitive view of the average deviation of the forecast results. The root mean square error, RMSE, reflects fluctuations in error and is sensitive to error outliers. The data in Table 3 illustrate again that the forecast accuracy and stability of Scheme 4, Scheme 3, Scheme 2, and Scheme 1 are sequentially increasing, and it validates the effectiveness of the proposed incremental forecast method.

5.4. Discussion of the Experiment Limitations

It must be emphasized that the loading methods and the test setup in this research are simplifications. However, the actual operating state of the WTB is very complex. For example, the aeroelastic effect makes load variable even if the wind speed is constant, and the strain will exhibit periodic characteristics under the rotating centrifugal force. In addition, the vibration contains flap-wise, edge-wise and twisting directions. But it is difficult to restore all the states under laboratory conditions. Therefore, this research simplifies the loading methods and only considers the vibration in the flap-wise direction.

6. Conclusions

This article conducts a comprehensive study on methods to address the challenge of accurately predicting the long-time health status of a WTB. The research reveals that the accumulation of errors is the primary reason for the suboptimal performance of current popular methods in long-term time series forecasting. Consequently, an incremental forecast method endowed with adaptive updating capabilities is proposed to mitigate the phenomenon of error accumulation, thereby enhancing the accuracy of long-time series forecasting. The method adopts the idea of incremental learning to improve the Bi-LSTM model via the error-supervised feedback structure designed in model. Then, the model is capable of continuously updating the parameters as new data become available, thereby maintaining a high level of accuracy in long-time forecast.
Furthermore, an experimental setup for strain detection on WTBs is constructed in this research. The strain data collected from this experimental platform are utilized to test the accuracy of the proposed incremental forecast method. The results show that the proposed method can maintain high accuracy in accomplishing the long-time series forecast task, with a fitting accuracy of 0.929 to the actual values and an average error below 4.9%. The forecast accuracy is improved by 24% compared with that of Bi-LSTM before improvement, and by 4.6% compared with that of Transformer.
The experimental results demonstrate that the incremental forecast method pro-posed in this research can effectively forecast the long-term strain of WTBs, showing the great potential value for engineering applications. Meanwhile, it provides a novel technical idea and theoretical guidance for the long-time series forecast problem. On this basis, the methodologies of setting thresholds more rationally and evaluating the health status of WTBs based on predicted strain values are very promising for future research. In the meantime, the improvement of the experimental setup to take into account more actual operating conditions is equally exciting.

Author Contributions

Conceptualization: B.W.; data curation, B.W. and H.W.; formal analysis, B.W.; funding acquisition, W.S.; investigation, B.W.; methodology, B.W.; project administration, B.W.; resources, W.S.; software, B.W. and H.W.; supervision, W.S.; validation, B.W.; visualization, B.W.; writing—original draft, B.W.; writing—review and editing, B.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Science and Technology Department of Xinjiang Uygur Autonomous Region, China, Science and Technology Plan Project—Key R & D Special Project (202107120033), and Xinjiang University, Outstanding Doctor Innovation Project (XJU2022BS090).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

References

  1. Dai, J.; Li, M.; Chen, H.; He, T.; Zhang, F. Progress and challenges on blade load research of large-scale wind turbines. Renew. Energy 2022, 196, 482–496. [Google Scholar] [CrossRef]
  2. Miao, H.; Xu, H.; Huang, G.; Yang, K. Evaluation and future projections of wind energy resources over the Northern Hemisphere in CMIP5 and CMIP6 models. Renew. Energy 2023, 211, 809–821. [Google Scholar] [CrossRef]
  3. Song, M.; Moaveni, B.; Ebrahimian, H.; Hines, E.; Bajric, A. Joint parameter-input estimation for digital twinning of the Block Island wind turbine using output-only measurements. Mech. Syst. Signal Process. 2023, 198, 110425. [Google Scholar] [CrossRef]
  4. Sun, R.; Guan, L.; Diao, N. rlaNet: A Residual Convolution Nested Long-Short-Term Memory Model with an Attention Mechanism for Wind Turbine Fault Diagnosis. Mathematics 2024, 12, 3460. [Google Scholar] [CrossRef]
  5. Jaramillo, F.; Gutiérrez, J.; Orchard, M.; Guarini, M.; Astroza, R. A Bayesian approach for fatigue damage diagnosis and prognosis of wind turbine blades. Mech. Syst. Signal Process. 2022, 174, 109067. [Google Scholar] [CrossRef]
  6. Wen, B.; Li, Z.; Jiang, Z.; Tian, X.; Dong, X.; Peng, Z. Blade loading performance of a floating wind turbine in wave basin model tests. Ocean Eng. 2020, 199, 107061. [Google Scholar] [CrossRef]
  7. Chen, X.; Yan, R.; Liu, Y. Wind turbine condition monitoring and fault diagnosis in China. IEEE Instrum. Meas. Mag. 2016, 19, 22–28. [Google Scholar] [CrossRef]
  8. Artigao, E.; Martín-Martínez, S.; Honrubia-Escribano, A.; Gómez-Lázaro, E. Wind turbine reliability: A comprehensive review towards effective condition monitoring development. Appl. Energy 2018, 228, 1569–1583. [Google Scholar] [CrossRef]
  9. Meng, H.; Lien, F.; Glinka, G.; et al. Study on fatigue life of bend-twist coupling wind turbine blade based on anisotropic beam model and stress-based fatigue analysis method. Compos. Struct. 2018, 208, 678–701. [Google Scholar] [CrossRef]
  10. Caous, D.; Bois, C.; Wahl, J.C.; Palin-Luc, T.; Valette, J. Toward composite wind turbine blade fatigue life assessment using ply scale damage mod-el. Procedia Eng. 2018, 213, 173–182. [Google Scholar] [CrossRef]
  11. Li, M.; Gao, J.; Zhou, J. A Combined High and Low Cycle Fatigue Life Prediction Model for Wind Turbine Blades. Appl. Sci. 2025, 15, 1173. [Google Scholar] [CrossRef]
  12. Lu, Q.; Ye, W.; Yin, L. ResDenIncepNet-CBAM with principal component analysis for wind turbine blade cracking fault prediction with only short time scale SCADA data. Measurement 2023, 212, 112696. [Google Scholar] [CrossRef]
  13. Chrétien, A.; Tahan, A.; Pelletier, F. Wind Turbine Blade Damage Evaluation under Multiple Operating Conditions and Based on 10-Min SCADA Data. Energies 2024, 17, 1202. [Google Scholar] [CrossRef]
  14. Qin, Z.; Qiang, S.; Zhang, M.; Rong, X.; Liao, C.; Wang, J.; Xu, J. Design and structural responses of a 38-meter sectional wind turbine blade under extreme static loads. Compos. Struct. 2022, 290, 115487. [Google Scholar] [CrossRef]
  15. Sierra-Pérez, J.; Torres-Arredondo, M.; Güemes, A. Damage and nonlinearities detection in wind turbine blades based on strain field pattern recognition. FBGs, OBR and strain gauges comparison. Compos. Struct. 2016, 135, 156–166. [Google Scholar] [CrossRef]
  16. Wang, B.; Sun, W.; Wang, H.; Wan, Y.; Xu, T. Location determination of impact on the wind turbine blade surface based on the FBG and the Time Difference. Sensors 2021, 21, 232. [Google Scholar] [CrossRef]
  17. Deng, W.; Liu, L.; Dai, Y.; Wu, H.; Yuan, Z. A prediction method for blade deformations of large-scale FVAWTs using dynamics theory and machine learning techniques. Energy 2024, 304, 132211. [Google Scholar] [CrossRef]
  18. Liu, H.; Zhang, Z.; Jia, H.; Li, Q.; Liu, Y.; Leng, J. A novel method to predict the stiffness evolution of in-service wind turbine blades based on deep learning models. Compos. Struct. 2020, 252, 112702. [Google Scholar] [CrossRef]
  19. Do-Eun, C.; Hyoung-Chul, K.; Moo-Hyun, K. Sequence-based modeling of deep learning with LSTM and GRU networks for structural damage detection of floating offshore wind turbine blades. Renew. Energy 2021, 174, 218–235. [Google Scholar]
  20. Weigend, A.S.; Huberman, B.A.; Rumelhart, D.E. Predicting the Future: A Connectionist Approach. Int. J. Neural Syst. 1990, 1, 193–209. [Google Scholar] [CrossRef]
  21. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  22. Yang, H.; Yang, S.; Meng, D.; Hu, C.; Wu, C.; Yang, B.; Nie, P.; Si, Y.; Su, X. Optimization of Analog Circuit Parameters Using Bidirectional Long Short-Term Memory Coupled with an Enhanced Whale Optimization Algorithm. Mathematics 2025, 13, 121. [Google Scholar] [CrossRef]
  23. Graves, A.; Schmidhuber, J. Framewise phoneme classification with bidirectional LSTM and other neural network architectures. Neural Netw. 2005, 18, 602–610. [Google Scholar] [CrossRef] [PubMed]
  24. Peng, Y.; Zhang, Y.; Zhang, L.; Yao, L.; Guo, X. Prediction of corrosion fatigue crack growth rate in aluminum alloys based on incremental learning strategy. Int. J. Fatigue 2024, 187, 14. [Google Scholar] [CrossRef]
  25. Alaeiyan, H.; Mosavi, M.; Ayatollahi, A. Improving the performance of GPS/INS integration during GPS outage with incremental regularized LSTM learning. Alex. Eng. J. 2024, 105, 137–155. [Google Scholar] [CrossRef]
  26. Shi, K.; Zhao, X.; Xu, J. Research on fatigue test of large horizontal axis wind turbine blade. Acta Energiae Sin. 2011, 32, 1264–1268. (In Chinese) [Google Scholar]
  27. Buzzega, P.; Boschini, M.; Porrello, A.; Abati, D.; Calderara, S. Dark experience for general continual learning: A strong, simple baseline. Adv. Neural Inf. Process. Syst. 2020, 33, 15920–15930. [Google Scholar]
  28. Kirkpatrick, J.; Pascanu, R.; Rabinowitz, N.; Veness, J.; Desjardins, G.; Rusu, A.A.; Milan, K.; Quan, J.; Ramalho, T.; Grabska-Barwinska, A.; et al. Overcoming catastrophic forgetting in neural networks. Natl. Acad. Sci. 2017, 114, 3521–3526. [Google Scholar] [CrossRef]
  29. Zhang, L.; Guo, Y.; Yu, L.; Wei, X.; Liu, W.; Huang, X. Structural collapse characteristics of a 48.8 m wind turbine blade under ultimate bending loading. Eng. Fail. Anal. 2019, 106, 104150. [Google Scholar] [CrossRef]
  30. García, D.; Tcherniak, D. An experimental study on the data-driven structural health monitoring of large wind turbine blades using a single accelerometer and actuator. Mech. Syst. Signal Process. 2019, 127, 102–119. [Google Scholar] [CrossRef]
  31. Wang, B.; Sun, W.; Wang, H.; Xu, T.; Zou, Y. Research on rapid calculation method of wind turbine blade strain for digital twin. Renew. Energy 2024, 221, 119783. [Google Scholar] [CrossRef]
Figure 1. Online time series forecast task.
Figure 1. Online time series forecast task.
Sensors 25 03898 g001
Figure 2. Cell structure of LSTM model.
Figure 2. Cell structure of LSTM model.
Sensors 25 03898 g002
Figure 3. Structure of Bi-LSTM networks.
Figure 3. Structure of Bi-LSTM networks.
Sensors 25 03898 g003
Figure 4. The difference between incremental learning and traditional learning.
Figure 4. The difference between incremental learning and traditional learning.
Sensors 25 03898 g004
Figure 5. The overall process of incremental health status forecasting.
Figure 5. The overall process of incremental health status forecasting.
Sensors 25 03898 g005
Figure 6. The pseudocode for the workflow.
Figure 6. The pseudocode for the workflow.
Sensors 25 03898 g006
Figure 7. Principles of experience replay technology.
Figure 7. Principles of experience replay technology.
Sensors 25 03898 g007
Figure 8. Wind speed data.
Figure 8. Wind speed data.
Sensors 25 03898 g008
Figure 9. The experimental loading equipment.
Figure 9. The experimental loading equipment.
Sensors 25 03898 g009
Figure 10. The random load WTB strain measurement experiments.
Figure 10. The random load WTB strain measurement experiments.
Sensors 25 03898 g010
Figure 11. The signal preprocessing flow.
Figure 11. The signal preprocessing flow.
Sensors 25 03898 g011
Figure 12. The original strain data.
Figure 12. The original strain data.
Sensors 25 03898 g012
Figure 13. The strain envelope and wind speed time series course.
Figure 13. The strain envelope and wind speed time series course.
Sensors 25 03898 g013
Figure 14. Forecast results of the three datasets.
Figure 14. Forecast results of the three datasets.
Sensors 25 03898 g014
Figure 15. Error dynamics during the forecast.
Figure 15. Error dynamics during the forecast.
Sensors 25 03898 g015
Figure 16. Comparison of different methods for forecast on 0.4R strain data.
Figure 16. Comparison of different methods for forecast on 0.4R strain data.
Sensors 25 03898 g016
Figure 17. Comparison of different method for forecast on R2.
Figure 17. Comparison of different method for forecast on R2.
Sensors 25 03898 g017
Figure 18. Comparison of different scheme for forecast on R2.
Figure 18. Comparison of different scheme for forecast on R2.
Sensors 25 03898 g018
Table 1. The characteristics of the strain sensor.
Table 1. The characteristics of the strain sensor.
Wire Grid Size
(mm)
Substrate Size
(mm)
Resistance
(Ω)
Strain LimitSensitivity
Coefficient
2.0 × 1.03.6 × 3.1350 ± 0.12%2.0 ± 1%
Table 2. The sensitivity analysis of window length.
Table 2. The sensitivity analysis of window length.
Window Length50100200300
R20.8970.9290.9020.838
Table 3. The inference time and memory usage.
Table 3. The inference time and memory usage.
Bi-LSTMTransformerIncremental Forecast
Inference time4 ms15 ms4 ms
Memory usage3.2 G14.9 G5.2 G
Table 4. The R2 of result of the preprocessed and original data.
Table 4. The R2 of result of the preprocessed and original data.
0.4R0.5R0.5RMean
Preprocessed0.9270.9290.9310.929
Original0.8960.9010.9030.900
Table 5. The error statistics of different ablation scheme on each dataset.
Table 5. The error statistics of different ablation scheme on each dataset.
IndicatorsDatasetScheme 1Scheme 2Scheme 3Scheme 4
MAPE0.4R5.36%6.08%6.38%7.10%
0.5R4.56%4.76%5.34%6.12%
0.6R4.77%5.32%5.96%6.61%
MAE0.4R2.15 × 10−52.55 × 10−52.58 × 10−52.89 × 10−5
0.5R1.77 × 10−51.87 × 10−52.11 × 10−52.45 × 10−5
0.6R1.86 × 10−52.17 × 10−52.39 × 10−52.67 × 10−5
RMSE0.4R3.13 × 10−53.56 × 10−53.77 × 10−54.18 × 10−5
0.5R2.57 × 10−52.72 × 10−53.05 × 10−53.57 × 10−5
0.6R2.73 × 10−53.13 × 10−53.38 × 10−53.83 × 10−5
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, B.; Sun, W.; Wang, H. A Long-Time Series Forecast Method for Wind Turbine Blade Strain with Incremental Bi-LSTM Learning. Sensors 2025, 25, 3898. https://doi.org/10.3390/s25133898

AMA Style

Wang B, Sun W, Wang H. A Long-Time Series Forecast Method for Wind Turbine Blade Strain with Incremental Bi-LSTM Learning. Sensors. 2025; 25(13):3898. https://doi.org/10.3390/s25133898

Chicago/Turabian Style

Wang, Bingkai, Wenlei Sun, and Hongwei Wang. 2025. "A Long-Time Series Forecast Method for Wind Turbine Blade Strain with Incremental Bi-LSTM Learning" Sensors 25, no. 13: 3898. https://doi.org/10.3390/s25133898

APA Style

Wang, B., Sun, W., & Wang, H. (2025). A Long-Time Series Forecast Method for Wind Turbine Blade Strain with Incremental Bi-LSTM Learning. Sensors, 25(13), 3898. https://doi.org/10.3390/s25133898

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop