Next Article in Journal
Floating-Transport Characteristics of Integrated Mono-Column Composite Bucket Foundation for Offshore Wind Turbines
Next Article in Special Issue
Exploring Dissipation Terms in the SPH Momentum Equation for Wave Breaking on a Vertical Pile
Previous Article in Journal
A Study on the Hydrodynamic Response Characteristics of Vessel-Shaped Cages Based on the Smoothed Particle Hydrodynamics Method
Previous Article in Special Issue
A Numerical Method on Large Roll Motion in Beam Seas Under Intact and Damaged Conditions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning-Based Prediction of Pitch Response for Floating Offshore Wind Turbines

1
Ocean College, Zhejiang University, Zhoushan 316021, China
2
Zhejiang Zhongnan Green Construction Technology Group Co., Ltd., Hangzhou 310051, China
*
Author to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2024, 12(12), 2198; https://doi.org/10.3390/jmse12122198
Submission received: 15 October 2024 / Revised: 15 November 2024 / Accepted: 26 November 2024 / Published: 1 December 2024

Abstract

:
Accurate dynamic response prediction is a challenging and crucial aspect for the fatigue or ultimate analysis of floating offshore wind turbines (FOWTs), which are increasingly recognized for their potential to harness wind energy in deep-water environments. However, traditional numerical modeling approaches like the finite element method are time-consuming, making them inefficient for generating the extensive datasets required. This paper presents an efficient deep learning-based approach, referred to as the CNN-GRU model, considering multiple external environments. This model integrates convolutional neural networks (CNNs) and gated recurrent units (GRUs), effectively extracting the coupling relationships among various input features and capturing the temporal dependencies to enhance predictive accuracy. The proposed model is applied to two distinct types of FOWTs under three sea states, and the results demonstrate its satisfactory accuracy, with an average correlation coefficient (CC) of 0.9962 and an average coefficient of determination (R²) of 0.9864. The high accuracy across all cases proves the model’s robustness and reliability. Furthermore, the model’s optimal configurations, including memory lengths, sample sizes, and optimizer, are identified through parametric studies. Moreover, the Shapley additive explanations (SHAP) interpretation is utilized to reveal the most significant features influencing structural responses. In addition, a comparative analysis with two other ensemble models, namely random forest and gradient boosting, is conducted. The proposed approach achieves superior accuracy, with computational time approximately half that of the other two models, thereby highlighting its efficiency and effectiveness. The comprehensive framework, which encompasses feature selection, data processing, deep learning model construction, and interpretation, demonstrates significant potential for addressing a broad range of engineering problems through deep learning methodologies.

1. Introduction

With the excessive exploitation of fossil fuels, global environmental pollution and energy crisis issues are becoming increasingly severe. Wind power, as a type of clean and renewable energy, not only aligns with the objectives of the Paris Agreement [1] but also plays a pivotal role in the global energy transition toward sustainable and low-carbon energy systems. Offshore wind energy generation, particularly in deep waters, presents a promising and scalable solution due to the presence of strong and consistent wind resources, as well as the capability of transporting large engineering facilities while minimizing noise and visual pollution for the public [2]. Floating offshore wind turbines (FOWTs) are especially suited for deep-water installations where seabed-fixed turbines are not feasible. Thus, the dynamic analysis of FOWT responses is essential for optimizing structural design and ensuring reliability.
However, accurate and efficient response analysis for FOWTs is challenging due to the need to account for the coupled effects of aerodynamics, structural dynamics, hydrodynamics, servo control, and mooring systems. Significant efforts have been made by researchers to address this issue. Open-source fatigue, aerodynamics, structures, and turbulence modeling (OpenFAST), developed for the dynamic analysis of wind turbines, has gained widespread application due to its open-source nature. It originated from the fatigue, aerodynamics, structures, and turbulence modeling (FAST) program, where Jonkman [3,4] developed a hydrodynamic module to enable coupled aero-hydro-servo-elastic analysis of FOWTs. Nevertheless, OpenFAST primarily relies on frequency-domain analysis and overlooks the dynamic effects of moorings, which can lead to inaccuracies in highly nonlinear wave conditions or scenarios involving significant structural motion. To overcome these limitations, other simulation tools have been integrated to enhance hydrodynamic modeling capabilities [5,6,7,8,9,10,11]. For instance, advanced quantitative wave analysis (AQWA) software has been integrated with FAST to simulate the floating foundation and mooring systems of FOWTs, leveraging AQWA’s ability to handle the nonlinear dynamics of substructures [11]. While these combined approaches are capable of producing accurate results, the underlying finite element analysis (FEA) and computational fluid dynamics (CFD) are inherently slow due to their implicit nature. This limitation becomes particularly pronounced in fatigue and ultimate response analyses, which require numerous simulations, leading to prohibitive computational times. Hence, there is an urgent need for more efficient methods while maintaining the same level of accuracy for the dynamic analysis of FOWTs.
In recent years, machine learning (ML) has gained prominence as a promising approach to tackle the highly nonlinear and complex coupled relationships inherent in engineering problems. Among the various ML techniques, artificial neural networks (ANNs) have found extensive application in offshore engineering. An efficient hybrid procedure that combines ANN with finite element methods (FEM) has been proposed, using surge, sway, and heave motions of floating production storage and offloading (FPSO) units as inputs to accurately predict the top tension of mooring lines [12]. Another study examined the dynamic response of a buoy supporting riser (BSR) and successfully employed ANN to estimate its tension under varying sea states, demonstrating strong performance with a small relative error compared to traditional FEM analyses [13]. Further investigations have shown the utility of ANN in predicting tensions and bending moments for fatigue analysis of steel catenary risers (SCRs), indicating the versatility of ANN in different wave environments [14,15,16]. A similar methodology has been utilized to conduct comprehensive fatigue analyses of mooring lines by predicting their top tensions based on surge, sway, heave, roll, pitch, and yaw motions. Notably, this approach achieved a deviation of only 1.6% in predicted fatigue life when compared to FEM results [17]. Building on these advancements, a new ANN-based procedure was proposed, enabling fresh predictions without the need for an additional dynamic analysis [18]. Further enhancements to the model have incorporated the ability to account for the directionality of environmental loads [19,20]. In addition, three novel schemes have been introduced to improve the generalization of the ANN model [21]. As an important branch of ML, ensemble models such as random forest and gradient boosting have also gained significant traction in various machine learning applications due to their strong generalization capabilities and computational efficiency. Random forest constructs multiple independent decision trees, aggregating their predictions to reduce variance and improve model robustness. In contrast, gradient boosting builds trees sequentially, with each new tree aiming to correct the errors of the previous ones by minimizing residual errors, often leading to superior predictive accuracy. These two methodologies have been effectively applied in forecasting wind power and modeling the structural response across diverse environmental conditions [22,23,24]. However, the main limitation of traditional ML methods lies in their ability to handle a limited number of input features. The performance of these methods may degrade when faced with highly nonlinear and complex systems like FOWTs, where multiple interacting variables influence the system’s response with unknown contributions beforehand.
Deep learning (DL), a rapidly emerging branch of machine learning driven by advancements in computational power, has revolutionized predictive modeling, particularly for complex and high-dimensional problems. DL excels at extracting features and capturing intricate patterns across multiple layers. This enables it to model the highly nonlinear and coupling relationships between numerous variables effectively, which is suited for applications like time-series forecasting, computer vision, and natural language processing [25]. As a novel technology, the application of DL in predicting the dynamic response of FOWTs remains relatively limited. For instance, a stacked sequential model comprising five layers, with a total of 500 neurons, has been employed to predict the mooring’s tension of a FOWT model [26]. Similarly, long short-term memory (LSTM) networks have been tested for predicting the dynamic behavior of mooring systems [27,28,29]. Additionally, a multilayer perceptron (MLP) model featuring three layers of 768 neurons has demonstrated its accurate identification of the tower top acceleration and tower root force [30]. A gated recurrent unit (GRU) model was subsequently applied and found to outperform both backpropagation neural networks (BPNN) and LSTM networks in predicting flapwise moments and platform pitching [31]. Recently, the hybrid convolutional neural network-GRU (CNN-GRU) model has gained attention for its ability to predict dynamic responses in FOWTs. However, existing approaches exhibit notable limitations. One study is restricted to shutdown conditions, utilizing only wave loads as the sole input for response prediction, which is not directly applicable to FOWTs during operational conditions [32]. Another approach incorporates additional features into the deep learning model, but it excludes past response values, relying solely on current features [33]. However, the inclusion of hydrodynamic forces on the platform introduces computational complexities, as the model requires additional time to compute these forces for each new prediction.
This paper presents a novel CNN-GRU deep learning model specifically designed to predict the pitch response of FOWTs. In contrast to previous studies, our model not only selects multiple relevant input features but also incorporates wave elevation and historical response values. This innovative design improves the efficiency of the model, enabling faster predictions once the training phase is complete. Meanwhile, the CNN component effectively extracts the coupling relationships among various features, while the GRU component captures the temporal dependencies between multiple input features and output responses. Dynamic analyses are performed by integrating the FAST and AQWA programs to generate accurate baseline data for the model’s predictions. In this study, two FOWTs with distinct floating foundations, namely OC4 and Umaine, are examined and compared using the same deep learning model. Several configurations, including different memory lengths, sample sizes, and optimization algorithms, are tested to identify the optimal solution. Finally, the optimal model is interpreted using SHAP, providing insights into the contribution of each feature to the predictions.
The case studies demonstrate that the proposed CNN-GRU approach achieves accurate and efficient predictions for the pitch response of FOWTs, exhibiting strong correlation and minimal discrepancies. Additionally, the optimal configuration is identified, indicating its suitability for different FOWTs. Furthermore, the most relevant features are selected through SHAP interpretation, which will guide researchers in selecting fewer features for simplicity or optimization of models in future studies.

2. Methodology

This section outlines the framework of the deep learning model employed for the response prediction of FOWTs. It begins with a review of CNNs and GRUs to explain their underlying mechanisms. Following this, the proposed hybrid method is presented in detail, showcasing its integration and application.

2.1. Review of Convolutional Neural Networks

Convolutional neural networks [34] are a specialized class of neural networks most commonly used for analyzing grid-structured data such as time series and images. Time-series data can be regarded as a 1D grid of samples taken at regular time intervals while image data can be thought of as a 2D grid of pixels. Convolutional neural networks can effectively preserve the grid structure of the input, enabling the network to capture spatial or temporal dependencies in the data.
A typical CNN consists of three stages of data processing: the convolution stage, detector stage, and pooling stage, which transform the input data into a refined set of output features.

2.1.1. Convolution Stage

The first stage applies an affine transformation to the input data. In this process, a set of learnable filters, known as kernels, is convolved over the input, extracting feature maps that highlight important characteristics such as edges, textures, or other local features. For two-dimensional input data, the equation for convolution operation is as follows:
S ( i , j ) = ( K I ) ( i , j ) = m n I ( i m , j n ) K ( m , n )
where I is the input data, K is the filter (or kernel), and S is the output referred to as the feature map.
In traditional signal processing, the convolution operation involves flipping the kernel relative to the input; as m increases, the index into the kernel decreases but the index into the input increases. However, in deep learning frameworks, this flipping is typically omitted, and the kernel is directly applied to the input during convolution by using a related function referred to as the cross-correlation:
S ( i , j ) = ( I K ) ( i , j ) = m n I ( i + m , j + n ) K ( m , n )
During the convolution stage, each output is computed as a weighted sum of the input values within the filter’s receptive field, as illustrated in Figure 1. This process leverages three key principles to enhance the efficiency and performance of deep learning systems: sparse interactions, parameter sharing, and equivariant representations. Sparse interactions mean that each filter only interacts with a small local region of the input, reducing computational cost and capturing local features more effectively. Parameter sharing ensures that the same filter is applied across different regions of the input, leading to fewer parameters and improved generalization. Equivariant representations mean that if the input shifts, the output shifts accordingly, allowing CNNs to detect features regardless of their location in the input. These properties make convolution highly effective for tasks such as image and time-series analysis.

2.1.2. Detector Stage

After the convolution operation, the detector stage introduces nonlinearity through the application of an activation function, typically a rectified linear unit (ReLU) or other nonlinear functions such as sigmoid or tanh. The ReLU function is defined as f(x) = max(0, x), which means that all negative values are replaced by zero while leaving positive values unchanged. Compared with sigmoid or tanh, ReLU is more computationally efficient as it only outputs zero for negative values. Meanwhile, it does not saturate for positive values, helping mitigate the vanishing gradient issue during backpropagation in the training of deep networks. However, if many neurons only receive negative values during training, they will “die” and stop updating their weights, leading to non-learning neurons. To fix the “dead neuron” problem, an improved function called Leaky ReLU is proposed [35]. The function is defined by the following equations:
x               if   x 0   α x           if   x < 0
where α is a small constant, typically set as 0.01, ensuring the negative values are not completely ignored.

2.1.3. Pooling Stage

Generally, the last stage of CNNs is the pooling stage. This stage performs a down-sampling operation to reduce the spatial dimensions (height and width) of the input features, using methods such as selecting the maximum value from a defined region, known as max pooling, or computing the average of the values in that region, referred to as average pooling. Max pooling can capture the most prominent features by selecting the maximum value while average pooling provides a smoother representation by using the average value but may sometimes dilute significant information. The pooling stage reduces the number of parameters, and consequently decreases the computational complexity. Meantime, the risk of overfitting is alleviated by concentrating on the most important features. As a result, CNNs become more robust as they are invariant to small translations in the input data. On the other hand, dropout is also aimed at preventing overfitting. This method randomly “drops out” or deactivates a fraction of neurons (typically 20–50%) during the training process, forcing the model to learn more generalized representations of the input.

2.2. Review of Gated Recurrent Units

Gated recurrent units (GRUs), proposed by Cho et al. [36], are a powerful variant of recurrent neural networks (RNNs) to deal with sequential data. GRUs introduce an innovative gating mechanism to regulate the sequential data by deciding what information to discard and what to pass on to future time steps. Due to the gating mechanism, GRUs overcome the limitations of traditional RNNs in recognizing long-term dependencies in the sequential data and mitigating the vanishing gradient problem. Compared to the more complex architecture of RNNs, known as long short-term memory (LSTM) networks, GRUs have fewer parameters, making them more computationally efficient and helping to reduce the risk of overfitting to a certain extent.
The gating mechanism of GRUs is governed by two main gates: the update gate and the reset gate, as illustrated in Figure 2. The update gate determines the amount of information from the previous hidden state to be carried over to the current hidden state. The update gate, zt, is defined as follows:
z t = σ ( W z [ h t 1 , x t ] )
where ht−1 is the previous hidden state and xt is the current input. Wz is the weight matrix, which is learned during the training process of the neural network. σ represents the sigmoid function, which increases monotonically within the range of (0,1) with an “S” shape.
The reset gate controls how much of the past information to omit with a similar equation:
r t = σ ( W r [ h t 1 , x t ] )
where Wr is the weight matrix that needs to be learned during the training. In the formulation, when rt is close to 0, the previous hidden state is disregarded, forcing the network to focus on the current state; on the other hand, when rt is close to 1, the previous hidden state is fully considered.
Next, the candidate activation h ~ t is computed as:
h ˜ t = ϕ ( W h [ r t h t 1 , x t ] )
where Wh is the weight matrix to be learned and represents element-wise multiplication. ϕ is the hyperbolic tangent function within the range of (−1,1).
The actual activation ht is calculated as a weighted sum of the previous hidden state and the candidate activation:
h t = ( 1 z t ) h t 1 + z t h ˜ t
Owing to the capability of capturing temporal dependencies in sequential data, GRUs are successfully employed in time-series prediction, natural language processing, and various other fields.

2.3. CNN-GRU Hybrid Deep Learning Model

Considering the complexity of FOWTs’ response prediction in time series, this paper proposes a CNN-GRU model that integrates CNNs and GRUs to leverage the strengths of the above two deep learning techniques. The detailed architecture is depicted in Figure 3. CNNs are employed to extract the coupling relationships between different inputs while GRUs are designed to handle temporal dependencies between various inputs and the structural response.
Nine variables in time series are selected as the input of the deep learning model, which are rearranged in a matrix of 3 × 3. The CNN part includes three 2-dimensional convolutional layers, each with a kernel size of 2 × 2 and a stride of 1, ensuring that the receptive field of the final layer covers the entire area of the input matrix. The padding is configured as “same” to guarantee that edge information is fully preserved. The filter numbers in the three layers are 8, 16, and 32, respectively. The progressively increasing numbers of filters enable the model to learn more complex features as the network goes deeper. In addition, each CNN layer incorporates a Leaky ReLU activation function and a max pooling operation.
In the GRU part, 64 hidden neurons are set in the GRU layer to provide adequate pattern recognition capability while maintaining manageable complexity.
Typically, a fully connected network is utilized as the final layer in various deep learning models, serving a crucial function to utilize the features extracted in past layers for accurate prediction. In this model, the final fully connected layer consists of 128 neurons and is responsible for outputting the predicted structural response.
In the training process, the loss function is a key component that measures how well the model’s predictions match the actual target values. Based on this evaluation, the optimizer adjusts the model’s parameters (i.e., weight matrices) to minimize the difference between the predictions and the true values. For classification tasks, cross-entropy loss is commonly used, whereas mean squared error (MSE) is typically chosen for regression problems for its sensitivity to outlier data and smooth gradient during backpropagation.
In addition to MSE, four additional metrics are included for model evaluation: correlation coefficient (CC), mean absolute percentage error (MAPE), symmetric mean absolute percentage error (SMAPE), and coefficient of determination (denoted as R2), as presented in Equations (8)–(11), respectively. Here, the correlation coefficient refers to Pearson’s correlation coefficient, which is the most widely used measure of linear correlation. It ranges from −1 to +1, where values closer to +1 or −1 indicate a stronger linear relationship, and values near 0 indicate little to no linear relationship. It provides a clearer measure of the strength of the linear relationship between variables, making it particularly useful for evaluating model performance in machine learning. MAPE measures the model’s accuracy by calculating the percentage difference between prediction and actual values. SMAPE can provide a symmetric view of overestimate and underestimate as it divides by the average of actual and predicted values. The coefficient of determination, R2, is widely used to assess how well the model captures the relationship between predictors and the actual response. A value close to 1 indicates a strong fit, meaning the model explains most of the variance in the response variable. Conversely, a value close to 0 suggests that the model’s predictions are no better than simply using the mean value of the response variable. Additionally, a negative R2 indicates that the model performs worse than a simple mean model, often pointing to issues such as model misspecification or overfitting.
CC = ( y i y ¯ i ) ( y ^ i y ^ ¯ ) ( y i y ¯ ) 2 ( y ^ i y ^ ¯ ) 2
MAPE = 1 n i = 1 n y i y ^ i y i × 100
SMAPE = 1 n i = 1 n y i y ^ i ( y i + y ^ i ) / 2 × 100
R 2 = 1 ( y i y ^ i ) ( y i y ¯ ) 2
where yi is the true value, y ^ i is the predicted value, and y ¯ and y ^ ¯ are the mean values of the true and predicted values, respectively.

3. Simulation and Data Preprocessing

In this section, a traditional coupled analysis is conducted to establish a baseline for our simulations. To ensure the accuracy of the results, two software programs are integrated for the simulation process. Subsequently, the results are analyzed to identify the most relevant features, which are then preprocessed for the training of the proposed deep learning model.

3.1. Coupled Analysis of FOWTs

In dynamic analysis of floating wind turbines, OpenFAST has gained widespread application due to its open-source nature. It includes five major analytical modules: AeroDyn (aerodynamic analysis module), ElastoDyn (structural elasticity analysis module), ServoDyn (servo analysis module), HydroDyn (hydrodynamic analysis module), and MAP++ (mooring analysis module). These modules work together to enable coupled analysis of wind turbines. However, HydroDyn is primarily based on frequency-domain analysis, which may become inaccurate under highly nonlinear wave conditions or when the structure undergoes significant changes in submerged surface area due to motion. Meanwhile, MAP++ is based on the quasi-static approach to account for the static stiffness of mooring lines, but it cannot capture the dynamic effects.
To obtain more accurate simulated data as the input of the CNN-GRU model, the study employs commercial software AQWA to calculate the dynamic response of the substructure and couples it with the upper structure results obtained from FAST [11]. Two FOWTs with different floating foundations, i.e., OC4 [37] and Umaine [38] (see Figure 4), are tested with the proposed model. The weight of the ballast of Umaine is adjusted to keep the same draft of 20 m with OC4 under the same NREL 5 MW wind turbine and mooring system. The detailed parameters are listed in Table 1, Table 2 and Table 3.
In parametric studies, one common sea state around the east coast of USA is selected where the average wind speed at the hub height of 90 m is 11.4 m/s, accompanied by a wave with a significant height of 1.836 m and a spectral peak period of 7.441 s [39]. Multiple 1-h simulations (in steady state) under the same sea state but with different seeds are generated as the training data for the deep learning model.

3.2. Data Preparation for CNN-GRU Model

To identify the most relevant features for the pitch response of FOWTs, the Pearson’s correlation study is conducted, as illustrated in Figure 5. Although Spearman’s or Kendall’s correlation coefficient are particularly useful for detecting nonlinear or monotonic relationships due to their rank-based nature, they may overlook the actual numerical values and distributions of the data. This limitation can reduce their effectiveness when dealing with features and targets that exhibit smooth or continuous variations. Pearson’s correlation coefficient is commonly used as an initial screening tool in feature selection to identify those features that exhibit a strong linear relationship with the target. Subsequently, the deep learning model can learn more complex, nonlinear relationships through its internal transformations. While Pearson’s correlation emphasizes linear relationships, it provides a suitable starting point for the model, minimizing unnecessary computational burden and allowing for more efficient processing of the data.
Based on the results, the following parameters are selected as inputs for the deep learning model: TTDspFA (tower fore-aft displacement), RotThrust (rotor thrust), RotTorq (rotor torque), YawBrFxp (fore-aft shear at the top of the tower), YawBrMxp (side-to-side bending at the top of the tower), TwrBsFxt (fore-aft shear at the base of the tower), and TwrBsMyt (fore-aft bending at the base of the tower), with the minimum correlation coefficient with pitch response being 0.67, indicating a strong correlation, which is significantly higher than the correlations with other listed parameters. Given that correlation analysis only captures linear relationships, wave elevation (donated as WvEle) is also included as a feature since it induces nonlinear forces on the floating foundation of FOWTs. Additionally, based on the principle in structural dynamics that the initial state of motion influences subsequent motion, the previous time history of pitch (denoted as PtfmPitch) is incorporated into the model training and prediction. The autocorrelation of the pitch response is presented in Figure 6. It is observed that for small lag values (ranging from 1 s to 5 s), the autocorrelation coefficient gradually decreases, but remains positive and above 0.5, indicating a strong short-term correlation within the time series. As the lag increases to around 10 s, the autocorrelation coefficient becomes negative, suggesting the presence of some periodicity or inverse relationship. For multiple lags, up to 45 s, the autocorrelation coefficient still exceeds the significance level (with the blue area typically representing the 95% confidence interval), which indicates the inherent dependence of the pitch response on its past values. Based on this, in this CNN-GRU model, the past pitch response with a memory length of 40 s is used as the initial test value, serving as one of the inputs for the model training and prediction. Other features are also treated using the same memory length to prevent potential biases that may arise from differing time dependencies among features.
The nine features selected above are normalized using min-max scaling, as defined in Equation (12). This normalization method is widely employed during the preprocessing stage of deep learning models. It scales the input features to a uniform range of (0, 1), thereby mitigating bias in weight updates that may occur due to features with varying magnitudes. This practice also enhances the convergence speed of the model by creating a smoother loss surface, facilitating more efficient navigation during the optimization process. After that, the input features are reorganized into a 3 × 3 matrix for the CNN component of the proposed model, enabling it to capture local coupling relationships.
x = x min ( x ) max ( x ) min ( x )
where x is the original data, and x′ is the normalized data.

4. Results and Discussion

The preprocessed data mentioned above are fed into the CNN-GRU model to optimize the weight matrices. During the training process, the data are split into two parts: 80% is allocated for training and 20% for validation. The validation data are utilized to gauge the model’s performance on unseen data and to mitigate the risk of overfitting. Figure 7a,b depicts the changes in training and validation loss as the number of epochs increases. The model with the lowest validation loss is saved during the training process for subsequent predictions. In the case of OC4 and Umaine, the training loss and validation loss of the saved model are 2.6657 × 10−4, 2.6098 × 10−4 and 4.5337 × 10−4, 2.1228 × 10−4, respectively. The validation losses in both cases are quite low, even lower than the training loss, indicating that the models effectively capture generalizable patterns and perform well on new data. One particular simulation for each case is supplied into the models to detect the accuracy of the pitch prediction. Figure 8 shows that the predictions for both cases closely align with the trend of the real values, with only minimal discrepancies.
To investigate the effect of different model settings on the two types of floating wind turbines, this study explores different memory lengths, training data sizes, and optimizers. Then the trained models are tested on a consistent set of 10 simulations for comparison. The prediction accuracy is measured using five metrics: correlation coefficient (CC), mean squared error (MSE), mean absolute percentage error (MAPE), symmetric mean absolute percentage error (SMAPE), and coefficient of determination, R2.

4.1. Different Memory Length

Different memory lengths, namely 50 s, 40 s, 30 s, 20 s, and 10 s, are set for the training of the CNN-GRU model, with a uniform sample size of 40 to predict the pitch response of the OC4 and Umaine cases, as listed in Table 4. From the values of the five metrics, the deep learning models with different memory lengths for both FOWTs demonstrate accurate predictions, characterized by strong correlations and very small differences, as indicated by the MSE. In both cases, as the memory length increases from 10 s to 40 s, the MSE, MAPE, and SMAPE decrease while R2 increases. However, when the memory length is extended from 40 s to 50 s, MSE, MAPE, and SMAPE start to rise, accompanied by a decrease in R². This trend may be attributed to an increase in parameters leading to overfitting. On the other hand, as the memory length varies, the correlation coefficient remains relatively stable, fluctuating around 0.99. Thus, the optimal memory length is determined to be 40 s for both cases.

4.2. Different Sample Size

While keeping the memory length at 40 s, the sample size of the model is adjusted from 10 to 40. The results are presented in Table 5. In both cases, MSE, MAPE, and SMAPE essentially decrease as CC and R2 increase with the sample size, which aligns with our expectations. With more samples, the model can better learn the relationship between inputs and outputs. However, taking into account both accuracy and training speed, a sample size of 40 is selected as the final choice.

4.3. Different Optimizer

After setting the memory length to 40 s and the sample size to 40, three different optimizers, i.e., stochastic gradient descent (SGD), adaptive moment estimation (Adam), and Nesterov-accelerated adaptive moment estimation (Nadam), are applied to the CNN-GRU model to evaluate their prediction performance. The results are provided in Table 6. Across both cases, the proposed model with the Nadam optimizer is the most accurate, followed by those with Adam, while the SGD optimizer consistently produces the least accurate results. This result demonstrates the superiority of Nadam, which benefits from both adaptive learning rates and Nesterov momentum. The adaptive learning rate allows the optimizer to adjust based on the gradient at each step, improving convergence efficiency. Simultaneously, the anticipatory nature of Nesterov momentum enables the optimizer to “look ahead” in the gradient’s future direction, helping to mitigate issues like overshooting and oscillation. While Adam also incorporates adaptive learning rates, it lacks the anticipatory updates provided by Nesterov momentum. Conversely, SGD employs Nesterov momentum but uses a fixed learning rate, limiting its flexibility. These combined advantages of Nadam allow the model to converge to an optimal solution more efficiently, which is particularly important for addressing the complexity of this problem.

4.4. Feature Contribution Evaluation

In traditional machine learning models, the contribution of features is often more straightforward to assess. For example, in a linear regression model, a large absolute weight value indicates a strong influence on the output and a positive weight value represents a positive impact on the output. However, it is challenging to evaluate the impact of features in a deep learning model due to its complex architecture with multiple layers and nonlinear interactions between features, acting as a black box.
In this study, an interpretability method known as SHAP (SHapley Additive exPlanations) is utilized to “open” the black box and provide insights into the contribution of each feature to the prediction. The method is based on the principles from cooperative game theory, specifically Shapley values. The core idea is to assess the marginal contribution of a feature to the prediction by measuring how much the prediction changes when the feature is included versus when it is not included.
The input features of the CNN-GRU models are evaluated using SHAP. Figure 9a,b illustrates the feature interaction heatmaps for OC4 and Umaine, respectively. The SHAP values for the nine features are computed for each sample, taking into account their individual values, which are represented by distinct colors. Additionally, the mean SHAP value for each feature is calculated to compare their contributions to the output, as shown in Figure 10. Although the contributions of features differ between the two cases, the past pitch response of the floating platform, fore-aft bending at the base of the tower, fore-aft shear at the base of the tower, and wave elevation emerge as the most significant features influencing the present pitch response. In contrast, tower fore-aft displacement, fore-aft shear at the top of the tower, rotor torque, side-to-side bending at the top of the tower, and rotor thrust have a comparatively lesser impact.
The identified significant features are critical because they directly relate to the structural dynamics and environmental conditions impacting the platform pitch response. Conversely, features like tower fore-aft displacement and rotor torque show lesser contributions, suggesting they may have more indirect or less critical roles in this pitch prediction. This distinction in feature importance underscores the necessity of focusing on relevant factors for accurate modeling and prediction, particularly in complex systems like floating wind turbines, where environmental interactions are paramount. These insights can guide future research and the optimization of predicted models in similar applications.

4.5. Robustness and Comparative Evaluation

To assess the robustness and reliability of the optimized CNN-GRU model, two additional representative sea states from the east coast of the USA [39] (denoted as sea state 1 and sea state 3) were considered, alongside the sea state used in the aforementioned parametric studies (denoted as sea state 2), as detailed in Table 7. The simulated data of the FOWTs under the two sea states are used in the same manner for the training and prediction of the proposed model, ensuring consistency across all cases. As shown in Table 8, the proposed model exhibits outstanding performance for both types of FOWTs across all sea states, with an average correlation coefficient (CC) of 0.9962, an average coefficient of determination (R²) of 0.9864, and consistently low values of MSE, MAPE, and SMAPE.
The efficiency of the proposed model is further evaluated by comparing its performance with two ensemble models, namely random forest (RF) and gradient boosting (GB), under the same sea state 2 conditions. The RF and GB model both use 100 trees (estimators), a minimum sample split of 2, and a minimum sample leaf of 1, with the former having no maximum tree depth and the latter having a maximum tree depth of 3 and a learning rate of 0.1. As shown in Table 9, the CNN-GRU model outperforms both RF and GB in terms of higher CC and lower values of MSE, MAPE, and SMAPE. Notably, the CNN-GRU model also achieves a significantly higher coefficient of determination (R²), demonstrating its superior suitability for accurately predicting the dynamic response of FOWTs. From Figure 11, the CNN-GRU model displays superior performance at both the peak and valley points. Furthermore, the computational efficiency of the CNN-GRU model is significantly superior, with its training time approximately half that of the RF and GB models. The computational system is equipped with an AMD Ryzen 5 5600X 6-core CPU, an NVIDIA RTX 2080 Super GPU with 8 GB of VRAM, and 16 GB of RAM. This can be attributed to the inherent parallelism of deep learning models, which are highly optimized for parallel computation, particularly when utilizing GPUs, thereby accelerating the training process. In contrast, both RF and GB rely on sequential decision tree construction, a process that becomes increasingly time-consuming as the number of trees grows, thus contributing to the longer computational time for these models.
However, the CNN-GRU model still has certain limitations. These include its reliance on complex convolutional and recurrent layers, which require high-performance GPUs to perform efficiently. Additionally, in contrast to ensemble models like RF and GB, the internal mechanisms of the CNN-GRU model are more intricate and lack straightforward interpretability. As a result, understanding the importance of individual input variables often requires supplementary methods, such as SHAP, to assess feature contributions.

5. Conclusions

This paper presents a novel deep learning-based approach, referred to as the CNN-GRU model, for predicting the response of floating wind turbines in the time domain. This model harnesses the strengths of both convolutional neural networks (CNNs) and gated recurrent units (GRUs). The CNN component is specifically designed to extract the coupling relationships among various features, while the GRU component excels at capturing the temporal dependencies between input features and output responses. This approach is successfully applied to two different types of floating wind turbines, i.e., OC4 and Umaine, achieving high accuracy with minimal discrepancies and strong correlation.
To identify the optimal model for this case, multiple configurations regarding memory lengths, sample sizes, and optimizers were explored. The accuracies of different settings were compared using five metrics: correlation coefficient (CC), mean squared error (MSE), mean absolute percentage error (MAPE), symmetric mean absolute percentage error (SMAPE), and coefficient of determination (R2). The optimal memory lengths for both OC4 and Umaine were found to be 40 s; shorter memory lengths slightly reduced the accuracy, while longer memory lengths tended to cause overfitting, which also negatively impacted accuracy. Additionally, accuracy increased with larger sample sizes; however, the sample size was determined to be 40 to balance training speed and efficiency. Furthermore, three optimizers, i.e., SGD, Adam, and Nadam, were tested. The Nadam optimizer demonstrated the best performance, benefiting from both adaptive learning rates and the anticipatory nature of Nesterov momentum.
The optimal model was interpreted using SHAP to provide insights into the contribution of each feature to the prediction. The results indicate that the past pitch response of the floating platform, fore-aft bending at the base of the tower, fore-aft shear at the base of the tower, and wave elevation are the most significant features influencing the present pitch response. These features are directly related to structural dynamics and environmental conditions, in contrast to tower fore-aft displacement, fore-aft shear at the top of the tower, rotor torque, side-to-side bending at the top of the tower, and rotor thrust. These insights can guide researchers in selecting fewer features for simplicity or the optimization of models in future studies.
To evaluate the robustness and reliability of the proposed model, three distinct sea states were considered, where the model consistently demonstrated accurate predictions for both types of FOWTs. Furthermore, in a comparative analysis with random forest (RF) and gradient boosting (GB) models, the CNN-GRU model exhibited superior performance, highlighting its superior suitability for accurately predicting the dynamic response of FOWTs.
The proposed method has the advantage of rapid prediction once training is complete, which can be attributed to its explicit formulation. In contrast, traditional finite element methods are substantially slower due to their implicit formulation. Therefore, the proposed method is well-suited for rapidly generating numerous samples required for fatigue analysis or ultimate response analysis. Furthermore, the framework—encompassing feature selection, data processing, CNN-GRU model construction and optimization, and SHAP interpretation—can be promoted for solving a wide range of engineering problems with deep learning models.

Author Contributions

R.C.: Conceptualization, Methodology, Software, Investigation, Writing—original draft preparation. K.Z.: Writing—review and editing, Supervision. M.L.: Conceptualization, Methodology, Writing—review and editing, Supervision. Y.A.: Writing—review and editing. L.G.: Writing—review and editing, Supervision. All authors have read and agreed to the published version of the manuscript.

Funding

This study was funded by the National Postdoctoral Fund for Overseas Scholars.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to their relevance to ongoing and future research.

Acknowledgments

The authors would like to acknowledge the assistance of AI tools in providing language editing support during the preparation of this manuscript.

Conflicts of Interest

The authors Ruifeng Chen, Ke Zhang and Lixiang Guo are employed by Zhejiang Zhongnan Green Construction Technology Group Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be perceived as a potential conflict of interest.

References

  1. United Nations. Paris Agreement. Report of the Conference of the Parties to the United Nations Framework Convention on Climate Change. In Proceedings of the Parties on its Twenty-First Session, Paris, France, 30 November–13 December 2015; Volume 4. [Google Scholar]
  2. Wang, C.M.; Utsunomiya, T.; Wee, S.C.; Choo, Y.S. Research on Floating Wind Turbines: A Literature Survey. IES J. Part A Civ. Struct. Eng. 2010, 3, 267–277. [Google Scholar] [CrossRef]
  3. Jonkman, J.M. Dynamics Modeling and Loads Analysis of an Offshore Floating Wind Turbine. Ph.D. Thesis, University of Colorado at Boulder, Boulder, CO, USA, 2007. [Google Scholar]
  4. Jonkman, J.M. Dynamics of Offshore Floating Wind Turbines—Model Development and Verification. Wind. Energy 2009, 12, 459–492. [Google Scholar] [CrossRef]
  5. Cermelli, C.; Roddier, D.; Aubault, A. WindFloat: A Floating Foundation for Offshore Wind Turbines—Part II: Hydrodynamics Analysis. In Proceedings of the International Conference on Offshore Mechanics and Arctic Engineering, Honolulu, HI, USA, 31 May–5 June 2009; Volume 43444, pp. 135–143. [Google Scholar]
  6. Kvittem, M.I.; Bachynski, E.E.; Moan, T. Effects of Hydrodynamic Modelling in Fully-Coupled Simulations of a Semi-Submersible Wind Turbine. Energy Procedia 2012, 24, 351–362. [Google Scholar] [CrossRef]
  7. Shim, S. Coupled Dynamic Analysis of Floating Offshore Wind Farms. Doctoral Dissertation, Texas A & M University, College Station, TX, USA, 2010. [Google Scholar]
  8. Bae, Y.H.; Kim, M.H.; Shin, Y.S. Rotor-Floater-Mooring Coupled Dynamic Analysis of Mini TLP-Type Offshore Floating Wind Turbines. In Proceedings of the ASME 2010 29th International Conference on Ocean, Offshore and Arctic Engineering, Shanghai, China, 6–11 June 2010; American Society of Mechanical Engineers Digital Collection. pp. 491–498. [Google Scholar]
  9. Bae, Y.H.; Kim, M.H. Turbine Floater-Tether Coupled Dynamic Analysis Including Second-Order Sum-Frequency Wave Loads for a TLP-Type FOWT (Floating Offshore Wind Turbine). In Proceedings of the ASME 2013 32nd International Conference on Ocean, Offshore and Arctic Engineering, Nantes, France, 9–14 June 2013. American Society of Mechanical Engineers Digital Collection. [Google Scholar]
  10. Bae, Y.H.; Kim, M.H. Aero-Elastic-Control-Floater-Mooring Coupled Dynamic Analysis of Floating Offshore Wind Turbine in Maximum Operation and Survival Conditions. J. Offshore Mech. Arct. Eng. 2014, 136, 020902. [Google Scholar] [CrossRef]
  11. Yang, Y.; Bashir, M.; Michailides, C.; Li, C.; Wang, J. Development and Application of an Aero-Hydro-Servo-Elastic Coupling Framework for Analysis of Floating Offshore Wind Turbines. Renew. Energy 2020, 161, 606–625. [Google Scholar] [CrossRef]
  12. Guarize, R.; Matos, N.A.F.; Sagrilo, L.V.S.; Lima, E.C.P. Neural Networks in the Dynamic Response Analysis of Slender Marine Structures. Appl. Ocean. Res. 2007, 29, 191–198. [Google Scholar] [CrossRef]
  13. de Aguiar, C.S.; de Lacerda, T.A.G.; Sagrilo, L.V.; Siqueira, W.B. Comparison Between Finite Element Model and an Artificial Neural Networks Procedure for Riser Analysis. In Proceedings of the ASME 2015 34th International Conference on Ocean, Offshore and Arctic Engineering, St. John’s, NL, Canada, 31 May–5 June 2015. American Society of Mechanical Engineers Digital Collection. [Google Scholar]
  14. Chaves, V.; Sagrilo, L.V.; da Silva, V.R.M.; Vignoles, M.A. Artificial Neural Networks Applied to Flexible Pipes Fatigue Calculations. In Proceedings of the ASME 2015 34th International Conference on Ocean, Offshore and Arctic Engineering, St. John’s, NL, Canada, 31 May–5 June 2015; American Society of Mechanical Engineers: New York, NY, USA, 2015; p. V05BT04A022. [Google Scholar]
  15. Cortina, J.P.; de Sousa, F.J.; Sagrilo, L.V. Neural Networks Applied to the Wave-Induced Fatigue Analysis of Steel Risers. Math. Probl. Eng. 2018, 2018, 2719682. [Google Scholar] [CrossRef]
  16. Chaves, V.; Sagrilo, L.V.; Ribeiro Machado da Silva, V. Optimization of Flexible Pipes Dynamic Analysis Using Artificial Neural Networks. In Proceedings of the ASME 2016 35th International Conference on Ocean, Offshore and Arctic Engineering, Busan, Republic of Korea, 19–24 June 2016. American Society of Mechanical Engineers Digital Collection. [Google Scholar]
  17. Christiansen, N.H.; Torbergsen Voie, P.E.; Høgsberg, J.; Sødahl, N. Efficient Mooring Line Fatigue Analysis Using a Hybrid Method Time Domain Simulation Scheme. In Proceedings of the ASME 2013 32nd International Conference on Ocean, Offshore and Arctic Engineering, Nantes, France, 9–14 June 2013. American Society of Mechanical Engineers Digital Collection. [Google Scholar]
  18. Chen, R.; Low, Y.M. Reducing Uncertainty in Time Domain Fatigue Analysis of Offshore Structures Using Control Variates. Mech. Syst. Signal Process. 2021, 149, 107192. [Google Scholar] [CrossRef]
  19. Chen, R.; Low, Y.M. Efficient Long-Term Fatigue Analysis of Deepwater Risers in the Time Domain Including Wave Directionality. Mar. Struct. 2021, 78, 103002. [Google Scholar] [CrossRef]
  20. Cheng, A.; Low, Y.M. A New Metamodel for Predicting the Nonlinear Time-Domain Response of Offshore Structures Subjected to Stochastic Wave Current and Wind Loads. Comput. Struct. 2024, 297, 107340. [Google Scholar] [CrossRef]
  21. Cheng, A.; Low, Y.M. Improved Generalization of NARX Neural Networks for Enhanced Metamodeling of Nonlinear Dynamic Systems Under Stochastic Excitations. Mech. Syst. Signal Process. 2023, 200, 110543. [Google Scholar] [CrossRef]
  22. Shi, K.; Qiao, Y.; Zhao, W.; Wang, Q.; Liu, M.; Lu, Z. An Improved Random Forest Model of Short-Term Wind-Power Forecasting to Enhance Accuracy, Efficiency, and Robustness. Wind Energy 2018, 21, 1383–1394. [Google Scholar] [CrossRef]
  23. Lahouar, A.; Slama, J.B.H. Hour-Ahead Wind Power Forecast Based on Random Forests. Renew. Energy 2017, 109, 529–541. [Google Scholar] [CrossRef]
  24. Dong, X.; Miao, Z.; Li, Y.; Zhou, H.; Li, W. One Data-Driven Vibration Acceleration Prediction Method for Offshore Wind Turbine Structures Based on Extreme Gradient Boosting. Ocean. Eng. 2024, 307, 118176. [Google Scholar] [CrossRef]
  25. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2017. [Google Scholar]
  26. Lin, Z.; Liu, X. Assessment of Wind Turbine Aero-Hydro-Servo-Elastic Modelling on the Effects of Mooring Line Tension via Deep Learning. Energies 2020, 13, 2264. [Google Scholar] [CrossRef]
  27. Qiao, D.; Li, P.; Ma, G.; Qi, X.; Yan, J.; Ning, D.; Li, B. Realtime Prediction of Dynamic Mooring Lines Responses with LSTM Neural Network Model. Ocean. Eng. 2021, 219, 108368. [Google Scholar] [CrossRef]
  28. Wang, Z.; Qiao, D.; Yan, J.; Tang, G.; Li, B.; Ning, D. A New Approach to Predict Dynamic Mooring Tension Using LSTM Neural Network Based on Responses of Floating Structure. Ocean. Eng. 2022, 249, 110905. [Google Scholar] [CrossRef]
  29. Liu, J.; Li, B. A Deep Learning Model for Predicting Mechanical Behaviors of Dynamic Power Cable of Offshore Floating Wind Turbine. Mar. Struct. 2025, 99, 103705. [Google Scholar] [CrossRef]
  30. Wang, Z.; Qiao, D.; Tang, G.; Wang, B.; Yan, J.; Ou, J. An Identification Method of Floating Wind Turbine Tower Responses Using Deep Learning Technology in the Monitoring System. Ocean. Eng. 2022, 261, 112105. [Google Scholar] [CrossRef]
  31. Zhang, Y.; Yang, X.; Liu, S. Data-Driven Predictive Control for Floating Offshore Wind Turbines Based on Deep Learning and Multi-Objective Optimization. Ocean. Eng. 2022, 266, 112820. [Google Scholar] [CrossRef]
  32. He, G.; Xue, J.; Zhao, C.; Cui, T.; Liu, C. Deep Learning Based Short-Term Motion Prediction of Floating Wind Turbine Under Shutdown Condition. Appl. Ocean. Res. 2024, 151, 104147. [Google Scholar] [CrossRef]
  33. Barooni, M.; Velioglu Sogut, D. Forecasting Pitch Response of Floating Offshore Wind Turbines with a Deep Learning Model. Clean Technol. 2024, 6, 418–431. [Google Scholar] [CrossRef]
  34. LeCun, Y. Generalization and Network Design Strategies; Connections in Perspective; University of Toronto: Toronto, ON, Canada, 1989. [Google Scholar]
  35. Maas, A.L.; Hannun, A.Y.; Ng, A.Y. Rectifier Nonlinearities Improve Neural Network Acoustic Models. In Proceedings of the International Conference on Machine Learning (ICML), Atlanta, GA, USA, 16–21 June 2013; Volume 30, p. 3. [Google Scholar]
  36. Cho, K. On the Properties of Neural Machine Translation: Encoder-Decoder Approaches. arXiv 2014, arXiv:1409.1259. [Google Scholar]
  37. Robertson, A.; Jonkman, J.; Masciola, M.; Song, H.; Goupee, A.; Coulling, A.; Luan, C. Definition of the Semisubmersible Floating System for Phase II of OC4; National Renewable Energy Laboratory: Golden, CO, USA, 2014. [Google Scholar]
  38. Allen, C.; Viscelli, A.; Dagher, H.; Goupee, A.; Gaertner, E.; Abbas, N.; Matthew, H.; Garrett, B. Definition of the UMaine VolturnUS-S Reference Platform Developed for the IEA Wind 15-Megawatt Offshore Reference Wind Turbine; No. NREL/TP-5000-76773; National Renewable Energy Lab. (NREL): Golden, CO, USA; Univ. of Maine: Orono, ME, USA, 2020. [Google Scholar]
  39. Stewart, G.M.; Robertson, A.; Jonkman, J.; Lackner, M.A. The Creation of a Comprehensive Metocean Data Set for Offshore Wind Turbine Simulations. Wind Energy 2016, 19, 1151–1159. [Google Scholar] [CrossRef]
Figure 1. An example of 2-D convolution.
Figure 1. An example of 2-D convolution.
Jmse 12 02198 g001
Figure 2. Gated recurrent unit (GRU).
Figure 2. Gated recurrent unit (GRU).
Jmse 12 02198 g002
Figure 3. Proposed CNN-GRU model.
Figure 3. Proposed CNN-GRU model.
Jmse 12 02198 g003
Figure 4. Two types of floating foundations for a floating offshore wind turbine.
Figure 4. Two types of floating foundations for a floating offshore wind turbine.
Jmse 12 02198 g004
Figure 5. Correlation between different features.
Figure 5. Correlation between different features.
Jmse 12 02198 g005
Figure 6. Autocorrelation of pitch response.
Figure 6. Autocorrelation of pitch response.
Jmse 12 02198 g006
Figure 7. Training and validation loss for OC4 and Umaine.
Figure 7. Training and validation loss for OC4 and Umaine.
Jmse 12 02198 g007
Figure 8. Comparison of real values and model predictions for OC4 and Umaine.
Figure 8. Comparison of real values and model predictions for OC4 and Umaine.
Jmse 12 02198 g008
Figure 9. Feature interaction heatmap for OC4 and Umaine.
Figure 9. Feature interaction heatmap for OC4 and Umaine.
Jmse 12 02198 g009
Figure 10. Mean SHAP value for OC4 and Umaine.
Figure 10. Mean SHAP value for OC4 and Umaine.
Jmse 12 02198 g010
Figure 11. Performance comparison of CNN-GRU, RF, and GB models for OC4 and Umaine.
Figure 11. Performance comparison of CNN-GRU, RF, and GB models for OC4 and Umaine.
Jmse 12 02198 g011
Table 1. Properties of wind turbine.
Table 1. Properties of wind turbine.
Elevation of the tower base above still water level10 m
Tower length77.6 m
Hub height above still water level90 m
Hub mass56,780 kg
Nacelle mass240,000 kg
Tower mass249,718 kg
Rated power of the wind turbine5 MW
Cut-in, rated, cut-out wind speed3 m/s, 11.4 m/s, 25 m/s
Table 2. Properties of mooring system.
Table 2. Properties of mooring system.
Number of mooring lines3
Unstretched length of mooring lines835.35 m
Depth of anchor below still water level200 m
Fairlead depth below still water level14 m
Diameter of mooring lines0.0766 m
Equivalent density of mooring lines113.35 kg/m
Equivalent tensile stiffness of mooring lines753.6 MN
Hydrodynamic drag coefficient of mooring lines1.1
Hydrodynamic added mass coefficient of mooring lines1.0
Table 3. Properties of two floating foundations.
Table 3. Properties of two floating foundations.
Type of Floating FoundationOC4Umaine
Mass of floating foundation (including ballast)1.3473 × 107 kg1.9935 × 107 kg
Distance from center of gravity to still water level13.46 m14.79 m
Roll moment of inertia about
the center of gravity
6.827 × 109 kg.m21.530 × 1010 kg.m2
Pitch moment of inertia about the center of gravity6.827 × 109 kg.m21.530 × 1010 kg.m2
Yaw moment of inertia about
the center of gravity
1.226 × 109 kg.m22.924 × 1010 kg.m2
Table 4. The effect of different memory lengths on prediction performance.
Table 4. The effect of different memory lengths on prediction performance.
TypeMemory LengthCCMSEMAPESMAPER2
OC4500.99625.8150 × 10−51.13721.14500.9836
400.99561.3647 × 10−50.45900.46040.9872
300.99513.1327 × 10−50.69060.69500.9724
200.99575.0763 × 10−50.95950.96680.9683
100.99637.0648 × 10−51.24851.23870.9492
Umaine500.98812.7902 × 10−50.62440.62770.9659
400.99211.1912 × 10−50.41940.41940.9811
300.97464.8175 × 10−50.96030.96190.9616
200.98652.2693 × 10−50.60100.60280.9531
100.97615.4823 × 10−50.91560.90930.9341
Table 5. The effect of different sample sizes on prediction performance.
Table 5. The effect of different sample sizes on prediction performance.
TypeSample SizeCCMSEMAPESMAPER2
OC4400.99561.3647 × 10−50.45900.46030.9872
300.99477.7729 × 10−51.16891.16160.9869
200.99424.6157 × 10−50.83220.83830.8897
100.98264.7637 × 10−50.88770.88890.8358
Umaine400.99351.4156 × 10−50.49500.49380.9811
300.98923.4255 × 10−50.81020.81030.9696
200.98673.8938 × 10−50.86120.85670.9359
100.98394.3954 × 10−50.83450.83620.8408
Table 6. The effect of different optimizers on prediction performance.
Table 6. The effect of different optimizers on prediction performance.
TypeOptimizerCCMSEMAPESMAPER2
OC4SGD0.97136.5691 × 10−43.40623.33020.9665
Adam0.99453.7636 × 10−43.00823.05970.9745
Nadam0.99561.3647 × 10−50.45900.46030.9872
UmaineSGD0.93012.8640 × 10−42.34750.74480.9249
Adam0.99273.6731 × 10−50.74910.74480.9695
Nadam0.99351.4156 × 10−50.49500.49380.9811
Table 7. Characteristics of different sea states.
Table 7. Characteristics of different sea states.
Type Average Wind Speed at the Hub Height (m/s)Significant Wave Height (m)Spectral Peak Period (s)
Sea state 18.01.3168.006
Sea state 211.41.8367.441
Sea state 316.02.5987.643
Table 8. CNN-GRU model’s performance in different sea states.
Table 8. CNN-GRU model’s performance in different sea states.
Type of Sea StateType of FOWTCCMSEMAPESMAPER2
Sea state 1OC40.99742.3727 × 10−50.61970.62220.9816
Umaine0.99731.2009 × 10−50.42400.42280.9911
Sea state 2OC40.99561.3647 × 10−50.45900.46030.9872
Umaine0.99351.4156 × 10−50.49500.49380.9811
Sea state 3OC40.99701.5628 × 10−50.59460.59320.9857
Umaine0.99621.3206 × 10−50.46180.46110.9915
Table 9. The effect of different models on prediction performance.
Table 9. The effect of different models on prediction performance.
TypeModelTimeCCMSEMAPESMAPER2
OC4CNN-GRU410 s0.99561.3647 × 10−50.45900.46030.9872
Random forest1010 s0.96666.4487 × 10−50.95550.95180.9277
Gradient boosting732 s0.97535.1615 × 10−50.83270.82950.9421
UmaineCNN-GRU407 s0.99351.4156 × 10−50.49500.49380.9811
Random forest1040 s0.84462.1472 × 10−41.87381.86870.7066
Gradient boosting744 s0.88541.6190 × 10−41.60871.60430.7787
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, R.; Zhang, K.; Luo, M.; An, Y.; Guo, L. Deep Learning-Based Prediction of Pitch Response for Floating Offshore Wind Turbines. J. Mar. Sci. Eng. 2024, 12, 2198. https://doi.org/10.3390/jmse12122198

AMA Style

Chen R, Zhang K, Luo M, An Y, Guo L. Deep Learning-Based Prediction of Pitch Response for Floating Offshore Wind Turbines. Journal of Marine Science and Engineering. 2024; 12(12):2198. https://doi.org/10.3390/jmse12122198

Chicago/Turabian Style

Chen, Ruifeng, Ke Zhang, Min Luo, Ye An, and Lixiang Guo. 2024. "Deep Learning-Based Prediction of Pitch Response for Floating Offshore Wind Turbines" Journal of Marine Science and Engineering 12, no. 12: 2198. https://doi.org/10.3390/jmse12122198

APA Style

Chen, R., Zhang, K., Luo, M., An, Y., & Guo, L. (2024). Deep Learning-Based Prediction of Pitch Response for Floating Offshore Wind Turbines. Journal of Marine Science and Engineering, 12(12), 2198. https://doi.org/10.3390/jmse12122198

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop