Next Article in Journal
Performance of Sewer Concretes with Calcium Sulpho-Aluminate Cement and Portland Cement Blends: Field and Laboratory Studies
Next Article in Special Issue
Accelerating Multi-Objective Optimization of Composite Structures Using Multi-Fidelity Surrogate Models and Curriculum Learning
Previous Article in Journal
Effect of Hydrothermal Coatings of Magnesium AZ31 Alloy on Osteogenic Differentiation of hMSCs: From Gene to Protein Analysis
Previous Article in Special Issue
Prediction and Experimental Study of Low-Frequency Acoustic and Vibration Responses for an Aircraft Instrument Compartment Based on the Virtual Material Method
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dynamic Load Identification of Thin-Walled Cabin Based on CNN-LSTM-SA Neural Network

1
National Key Laboratory of Aerospace Liquid Propulsion, Xi’an Aerospace Propulsion Institute, Xi’an 710100, China
2
School of Aerospace Engineering, Xi’an Jiaotong University, Xi’an 710049, China
3
State Key Laboratory for Strength and Vibration of Mechanical Structures, School of Mechanical Engineering, Xi’an Jiaotong University, Xi’an 710049, China
*
Author to whom correspondence should be addressed.
Materials 2025, 18(6), 1255; https://doi.org/10.3390/ma18061255
Submission received: 17 January 2025 / Revised: 24 February 2025 / Accepted: 28 February 2025 / Published: 12 March 2025
(This article belongs to the Special Issue Numerical Methods and Modeling Applied for Composite Structures)

Abstract

:
Spacecraft are subjected to various external loads during flight, and these loads have a direct impact on the structural safety and functional stability of the spacecraft. Obtaining external load information can provide reliable support for spacecraft health detection and fault warning, so accurate load identification is very important for spacecraft. Compared with the traditional time-domain load identification method, the neural network-based time-domain load identification method can avoid the establishment of the inverse model and realize the response-load time-sequence mapping, which has a broad application prospect. In this paper, a CNN-LSTM-SA neural network-based load identification method is proposed for load acquisition of a thin-walled spacecraft model. Simulation results show that the method has higher identification accuracy and robustness (RMSE and MAE of 8.47 and 10.83, respectively, at a 20% noise level) in the load identification task compared to other network structures. The experimental results show that the coefficients of determination (R2) of the proposed neural network load recognition model for time-domain identification tasks of sinusoidal and random loads are 0.98 and 0.93, respectively, indicating excellent fitting performance. This study provides a reliable new method for load identification in thin-walled spacecraft cabin structures.

1. Introduction

In the field of aerospace, the acquisition of external load information is directly related to the stability and reliability of spacecraft. However, due to the harsh external environment and complex dynamic loads during spacecraft flight, it is difficult to measure external loads directly with sensors, so it is necessary to use load identification techniques to estimate external loads from the response inside the spacecraft [1]. The concept of load identification originated in the 1970s. At that time, load identification techniques were used to determine actual aircraft loads in order to improve aircraft performance [2]. It has since been widely applied in various engineering fields due to its ability to improve product reliability and durability. Dynamic load identification is divided into two categories: time-domain load identification methods and frequency-domain load identification methods. Time-domain load identification technology does not need to consider the frequency characteristics of the structure and is suitable for dynamic problems with significant nonlinearity and uncertainty. At the same time, the time-domain method can realize the on-line monitoring of structural loads, so the research on time-domain load identification methods has received more and more attention in recent years [3].
Time-domain load identification usually relies on the Green’s kernel function deconvolution method (GKFM) [4], which is mainly for linear systems. Meanwhile, in order to solve the matrix ill-posedness problem, the regularization method is needed in the solution process. However, the selection of the regularization parameters is a more complicated task [5]. With the development of deep learning, neural networks are gradually being applied to the task of load identification. Compared with the traditional time-domain load identification methods, the neural network-based method has the ability of automatic feature extraction and powerful nonlinear modeling capability, as well as better robustness [6]. Luis et al. [7] used a neural network to achieve the identification of horizontal and vertical forces on automobile wheel hubs and discussed the potential of the method for online monitoring of loads applied to automobiles. Tang et al. [8] proposed a method combining stochastic response power spectral density and deep convolutional neural network (CNN) to accurately recognize vehicle load information. Zhang et al. [9]. proposed a CNN network based on a transform domain approach, using decomposed signals from wavelet transforms of multiple vibration signals as input; combined with a CNN network to achieve load identification in the time domain. Zhou et al. [10] performed impact load identification for nonlinear structures using a deep recurrent neural network consisting of two long short-term memory network (LSTM) layers and one bi-directional long short-term memory network (BiLSTM) layer; the network trained on numerous dynamic responses and impact loads demonstrated the ability to identify complex impact loads, even when the impact location was unknown. Yang et al. [11] proposed a neural network model with a bi-directional LSTM layer, an LSTM layer, and two fully connected layers to identify typical dynamic loads (sinusoidal, impact, and random loads) for simply supported beams. Furthermore, Yang et al. [12] developed a depth-expanded convolutional neural network, which directly constructs a transfer model between the structural acceleration response and excitation for data-driven dynamic load identification. Wang et al. [13] proposed a deep regression adaptive network based on model migration that learns to improve the accuracy and efficiency of neural networks for load identification.
Due to the different characteristics of different neural networks, recent research in time-series data prediction has focused on combining multiple networks, thereby leveraging their strengths and building more accurate and efficient models. Lu et al. [14] proposed a CNN-BiLSTM-AM (attention mechanism) neural network for predicting stock data, which confirms the significant advantages of this method over other neural network methods. Huy et al. [15] used a CNN-LSTM model combined with an attention mechanism to predict short-term power loads, addressing issues with input–output relationships, mitigating information loss from long input time-series data and improving prediction accuracy. Marios [16] examined dynamic structural loads of gated cyclic units, using an LSTM and CNN trained on small datasets, and compared the results with a physically based residual Kalman filter (RKF). Zhang et al. [17] proposed a method to predict wave height based on a CNN-LSTM neural network, and the results show that the method can effectively improve the prediction accuracy as well as robustness. Hu et al. [18] proposed a fall detection method based on a CNN-LSTM neural network and compared the CNN network and LSTM network alone, and the results showed that the network can significantly improve the accuracy of detection. Divya et al. [19] proposed a new hybrid deep learning method SSA (Singular Spectrum Analysis)-CNN-LSTM, which demonstrated the superior performance of this network for solar power generation prediction over a long time period in the future. Numerous studies have shown that higher prediction accuracy as well as robustness can usually be achieved by employing hybrid neural networks in time-series tasks. Whereas most of the related studies are currently focused on time-series prediction tasks, there are fewer related studies on time-series mapping tasks similar to load identification.
In order to further improve the identification accuracy and applicability of neural networks in load identification tasks, and to promote the engineering application of neural network load identification models, this paper proposes a load identification model based on a CNN-LSTM-SA (self-attention mechanism) hybrid neural network. The CNN layer performs feature extraction on the time-series data, the LSTM layer captures the long-term dependencies in the time series, and the SA layer establishes the global dependencies of the time series and assigns different weights. In this paper, the initial time-series data are segmented and used as inputs to the neural network for network training, the load identification accuracy of this neural network is compared to other network models, and its noise immunity is discussed. Finally, experimental studies are conducted on the identification of sinusoidal and random load time-domain information using the method to prove its value for engineering applications.

2. CNN-LSTM-SA Neural Network Load Identification Model Building

2.1. Load–Response Relationship in the Time Domain

For a linear system, the response induced by the load in the time domain satisfies the principle of linear superposition, and the time-domain response data R ( t ) and the time-domain load F ( t ) at the response measurement point satisfy the following Equation [20]:
R ( t ) = 0 t F ( t ) h ( t τ ) d τ
where R ( t ) denotes the dynamic response of the system measurement points, e.g., strain, stress, displacement, velocity, acceleration, etc. h ( t ) is the response point–excitation point impulse response function. By discretizing the continuous time domain into a number of discrete time periods ∆t, Equation (1) can be expressed as follows:
R 1 R 2 R n = h 1 0 0 h 2 h 1 0 h n h n 1 h 1 F 0 F 1 F n 1 Δ t
That is to say, for time-series data with time sampling length, the relationship between response and loading satisfies Equation (3). From Equation (3), it can be seen that the response r ( t ) of the structure at the moment t is not only affected by the current moment f ( t ) but also by all previous moments f ( 0 ) ~ f ( t 1 ) . In other words, the response at any moment t contains all the information about the loads between the moments 0 ~ t . For the inverse of the above process, the load f ( t ) at the moment t can be written as Equation (4).
r ( 1 ) r ( 2 ) r ( k ) = h ( 1 ) Δ t h ( 2 ) Δ t h ( 1 ) Δ t h ( k ) Δ t h ( k 1 ) Δ t h ( 1 ) Δ t f ( 0 ) f ( 1 ) f ( k 1 )
f ( t ) = G r ( t ) , r ( t + Δ t ) , , r ( t + n Δ t )

2.2. CNN-LSTM-SA Network Architecture

The CNN model has excellent feature extraction ability and is widely used in various classification problems. The LSTM model has a special gate structure and weight sharing mechanism, which can avoid losing the long-term features of time-series data and is widely used in time-series analysis. The SA model has the importance of adding the past feature state of the time-series data to the output results and is widely used in adjusting the prediction results from an LSTM model. In this paper, according to the characteristics of the CNN, LSTM, and SA models, a load identification model based on a CNN-LSTM-SA architecture is established, and the structure of the model is shown in Figure 1. The main structure of the model consists of an input layer, CNN layer (one-dimensional convolutional layer and maximum pooling layer), LSTM layer, SA layer, fully connected layer, and output layer.
CNN is a feed-forward neural network with good performance in digital image processing, image denoising [21], and classification problems [22]. Its main structure includes a convolutional layer, pooling layer, fully connected layer, and output layer, as shown in Figure 2. The convolutional layer is the core of the CNN, which performs convolutional operations on input data by convolutional kernels of specific sizes. Each convolutional layer contains multiple convolutional kernels, which can extract high-dimensional features from the data and establish the relationship between input and output units through sparse connections. Its unique weight sharing feature allows all output units to share a common set of parameters to connect with the input units, which greatly reduces the training parameters required for the model and saves training time. The pooling layer is used to downsize the data, reducing redundant information while retaining key information and reducing the number of parameters required by the network. The calculational procedure of the convolutional layer is shown in Equation (5):
y t = f x t k t + b t
where x t is the input, k t is the weight of the convolution kernel, b t is the bias, f is the activation function, and y t is the output.
The LSTM neural network is a special type of recurrent neural network which is especially used for processing and predicting temporal data [23]. Compared with a recurrent neural network (RNN), a LSTM solves the problems of gradient vanishing and gradient explosion through a special network structure, which is especially suitable for tasks involving long-term dependencies. The memory cell of an LSTM is shown in Figure 3, which controls the flow of data through a “gating mechanism” to selectively remember or forget information to preserve the long-term dependencies of the temporal data. The memory cell of the LSTM consists of three gates (forget gate, input gate, and output gate), each consisting of a sigmoid activation function and a point multiplication operation with values between 0 and 1, which are used to control the flow of data. In Figure 3, C t 1 is the state of the cell at the previous moment, h t 1 is the final output value of the LSTM cell at the previous moment, f t is the output value of the forget gate, i t is the input of the current input gate, C t is the temporary state of the current cell, o t is the output value of the output gate, C t is the state of the cell at the current moment, and h t is the final output value of the current cell. The computational procedure of an LSTM is as follows:
(1)
The forget gate generates an output value f t between 0 and 1 by reading the final output h t 1 of the previous moment and the input x t of the current moment and using Equation (6), where 1 represents the complete retention of information and 0 represents the complete discarding of information.
f t = σ W f h t 1 , x t + b f
where W f is the weight matrix of the forget gate, b f is the bias term, and σ is the Sigmoid activation function;
(2)
The input gate generates the temporary state of the cell at the current moment C t by reading the final output of the previous moment h t 1 and the input of the current moment x t and then updates the cell state in conjunction with the output of the forget gate to obtain the new cell state C t which taking the values from 0 to 1.
i t = σ W i h t 1 , x t + b i
C t = tanh W c h t 1 , x t + b c
C t = f t * C t 1 + i t *   C t
where i t is the output value of the input gate at the current moment, which determines the extent to which the current input x t affects the update of the unit state C t . W c and b c are the weight matrix and bias terms for computing the temporary unit state C t . tanh is the hyperbolic tangent activation function and * denotes the matrix element-by-element multiplication;
(3)
The output gate extracts and outputs key information from the current unit state. It also reads the final output value h t 1 of the cell at the previous moment and the input value x t of the cell at the current moment and calculates o t through Equation (10).
o t = σ W o h t 1 , x t + b o
where o t takes values from 0 to 1, W o is the weight of the output gate, and b o is the bias of the output gate;
(4)
Finally, the final output of the unit at the current moment is calculated from the output value of the output gate o t and the current state of the unit C t by Equation (11).
h t = o t * tanh C t
SA (self-attention mechanism) is a technique that can dynamically capture the dependencies between different positions in a sequence and is widely used in natural language processing, computer vision, and time-series prediction. Its core idea is to decide how to summarize information by calculating the correlation weights of each element in a sequence with other elements. This mechanism allows the model to model the backward and forward relationships of the input sequences on a global scale, effectively capturing long-term dependencies. The main computational procedure for SA is as follows:
The input sequence X is mapped to the Query (Q), Key (K), and Value (V) for representing the characteristics of the current element, the importance of other elements, and the information to be summarized, respectively, through three sets of learnable linear transformations as in Equation (12).
Q = X W Q , K = X W K , V = X W V
where W Q , W K , and W V are trainable weight matrices.
The similarity score is obtained by computing the dot product of the Query and Key as in Equation (13), which is usually divided by d k in order to balance the numerical stability, where d k is the Key dimension:
S i m i l a r i t y i , j = Q i K j T d k
Use the softmax function to normalize the scores to a probability distribution:
α i , j = softmax S i m i l a r i t y i , j = exp S i m i l a r i t y i , j k = 1 n exp S i m i l a r i t y i , k
A new representation Z is obtained by weighted summation of V using the attention weights:
Z i = j = 1 n α i , j V j

2.3. Constructing the CNN-LSTM-SA Load Identification Model

Since the load identification problem is a time-series mapping problem rather than a direct time-series prediction, therefore, it is necessary to construct the obtained response data and load data into a time series that the network can handle. The method used here is to construct the time-series mapping dataset by cutting the time-series data X ( t ) with a sampling length of n into m segments with a length of k short time-series data X 1 ( t ) , X 2 ( t ) X m ( t ) , as shown in Equation (16). The time series constructed by this method can fully utilize the information of the original time domain data, ensure that the length of each time sample is consistent, and can also perform load identification with an arbitrary time-series length. At the same time, the data were normalized to the interval [−1,1] to prevent too large a gap in the order of magnitude of the data from causing the model to converge slowly or fail to converge.
X 1 ( t ) X 2 ( t ) X m ( t ) = X 1 ( 1 ) X 1 ( k ) X 2 ( 1 ) X 2 ( k ) X m ( 1 ) X m ( k ) = x ( 1 ) x ( 2 ) x ( k ) x ( 2 ) x ( 3 ) x ( k + 1 ) x ( m ) x ( m + 1 ) x ( n )
The structural diagram of the constructed CNN-LSTM-SA network is shown in Figure 4, and the main steps of the model applied to the load identification work are as follows:
(1)
Segment the load and response data according to the method described above and perform data normalization as well as division of the training and test sets;
(2)
After initialization of the network, the response data first pass through a one-dimensional convolutional CNN network for convolutional operation and average pooling operation to extract the high-dimensional features of the data;
(3)
In order to make the highly dimensional data adapt to the LSTM neural network, the data output from the CNN network need to be flattened. The flattening process consists of pulling the feature maps of each channel into one-dimensional vectors in order and then connecting the vectors of all channels;
(4)
The flattened data go sequentially through the LSTM neural network, the SA network, and the fully connected network, and finally the predicted load is generated;
(5)
Judge whether to terminate network training based on the error between the network’s predicted load and the actual load;
(6)
The trained network performs load identification on the test set to verify the recognition effect.
The hyperparameters of this network are as follows:
(1)
CNN layer: 1D convolutional kernel size, number of convolutional kernels, number of convolutional layers, and pooling size;
(2)
LSTM layer: number of LSTM units, number of LSTM layers, and dropout rate;
(3)
SA layer: number of attention heads, attention head dimension, and dropout rate;
(4)
Hyperparameters of the network: optimizer type, learning rate, learning rate decay, batch size, max epochs, etc.
The hyperparameters of the network structure described above can be adjusted according to the complexity of the task and whether it is overfitted or not. The network hyperparameters can be determined by grid search, Bayesian optimization, and empirically. Due to the limitation of the arithmetic power, the key hyperparameters of the neural network are determined in this study by small batch grid search.
It is worth noting that the optimal value of the segmentation sequence length is affected by a variety of factors, which will not be investigated in detail here, and the length of the segmentation time sequence set in the subsequent simulation and experimental studies in this paper is 200.

3. Numerical Simulation Study

Thin-walled cabins are efficient structural designs widely used in aerospace, industrial production, and scientific research, and have attracted considerable attention due to their light weight, high strength, and excellent material utilization. This paper presents both numerical simulations and experiments using a thin-walled cabin as the research object, applying the proposed method for load identification.
The simulation model is shown in Figure 5. The bottom face of the model is constrained with a fixed constraint and a simple harmonic excitation is applied at a point on the bulkhead wall, and the response is measured at the bulkhead bracket. Both the wall thickness of the bulkhead model and the bracket thickness inside the bulkhead are 5 mm. The model is made of aluminum alloy with a density of 2770 kg/m3, Young’s modulus of 71 GPa, and Poisson’s ratio of 0.33. The simulation was performed by transient analysis. The transient simulation module of ANSYS(2024R2) Mechanical software is used to calculate the simulation results. The simulation was performed using transient analysis with a time step of 0.001 s and a damping ratio of 0.03. Considering the calculation time and the accuracy of the simulation results, the mesh size is set to 2 mm.
Fifty sets of transients were analyzed for sinusoidal excitations of different frequencies and amplitudes lasting 1 s. The acceleration response at the measurement points was recorded. The sinusoidal excitation frequency ranges from 5 to 7 Hz, and a simulation calculation frequency is selected every 0.2 Hz (except 7 Hz), and transient analyses of sinusoidal excitations with amplitudes of 80 N, 110 N, 140 N, 170 N, and 200 N are performed at each frequency, respectively. The applied load time-domain data and the acquired acceleration time-domain data were divided into training (80%) and test (20%) sets for network training and load identification. To evaluate the effectiveness of the model, LSTM, CNN-LSTM, and CNN-LSTM-SA networks were selected for load identification, and their performance in load identification and ability to handle noisy data were compared.
This section uses test set error metrics to evaluate the effectiveness of the neural network load identification. The root mean square error (RMSE) and mean absolute error (MAE) are chosen as the error metrics, as defined by the following formulas:
RMSE = 1 n i = 1 n y ^ i y i 2
MAE = 1 n i = 1 n ( y ^ i y i ) × 100 %
where y ^ i is the neural network predictive value load, y i is the true load of the test set, and n is the number of samples in the test set; the smaller the value of RMSE and MAE, the higher the prediction accuracy of the model.
The three neural networks are trained separately, and the load identification results of each network under 0% noise data are shown in Table 1. It can be seen that the load identification accuracy of CNN-LSTM-SA has a significant advantage over the other two neural networks. Figure 6 shows the recognition effect and absolute error of the three neural networks for load identification under the data without the influence of noise. From the figure, it can be seen that the CNN-LSTM-SA network can not only recognize the peak of sinusoidal excitation well, but also the absolute error is smaller than the other two networks in the whole sampling time, indicating that the CNN-LSTM-SA load identification method proposed in this paper has obvious advantages.
In practical engineering problems, there is often noise interference during signal data acquisition using sensing devices, so it is necessary to verify the effectiveness of the CNN-LSTM-SA network for load recognition tasks in the presence of noisy signals. Researchers have confirmed in previous studies that CNN networks have filtering capabilities [23]; in this paper, we use the CNN-LSTM neural network as a reference to evaluate the adaptability of the CNN-LSTM-SA for load recognition tasks with noisy signals. White noise signals of 2%, 5%, 10%, and 20% are added to the simulated collected dataset, and the load recognition task is performed using the above two neural networks. Here, RMSE and MAE are still used as the indexes to evaluate the model load recognition effect, and the specific results are shown in Table 2. It can be seen that the CNN-LSTM-SA neural network has higher recognition accuracy and better noise immunity than the CNN-LSTM neural network in the load recognition work in the case of containing noise. Figure 7 shows the effect and absolute error of load recognition of both networks under different noise levels.
Based on the results of the analysis of the graphical data, it can be clearly seen that with the gradual increase in the noise level, the performance of both neural networks in the load recognition task shows a significant decrease, which is manifested by the trend of increasing their absolute error and root mean square error (RMSE). However, it is worth noting that under the influence of noise interference, the CNN-LSTM-SA neural network is still able to maintain the accuracy of load recognition well, with a relatively small increase in its error and a more stable recognition performance. In contrast, the performance of the CNN-LSTM neural network decreases more significantly, and the recognition accuracy is much lower than that of the CNN-LSTM-SA network. This indicates that the CNN-LSTM-SA network has obvious advantages in robustness and accuracy in the noisy environment.

4. Experimental Study

4.1. Test Object and Test System

The bottom of the test object is a fixed constraint, as shown in Figure 8, which is used to simulate the thin-walled segment of the spacecraft; the material is aluminum alloy, and the wall thickness of the model is 5 mm. Excitation is applied by the shaker to simulate the external load of the thin-walled cabin, and the acceleration sensor is used to measure the structural response. The excitation position of the shaker and the mounting position of the acceleration sensor are shown in Figure 9a, and the test system consists of a personal computer, a power amplifier, and a data acquisition system, as shown in Figure 9b. In order to reflect the superiority of the load recognition method proposed in this paper, two load recognition methods, an LSTM neural network and a CNN-LSTM neural network, are still used for comparison, and the evaluation indexes are the RMSE, MAE, and coefficient of determination R2, where R2 is used to evaluate the fitting effect of the regression model, and the closer the value is to 1, the better the model fitting effect.

4.2. Time-Domain Identification of Sinusoidal Load

The experimental study of sinusoidal load identification is first carried out, and the sampling frequency of the time-series data is 6400 Hz. In this paper, 30 sets of sinusoidal excitation information with different frequencies and amplitudes and the corresponding structural acceleration responses are collected. A total of 25 sets were used to train the neural network load recognition model, and 5 test sets were used to evaluate the load recognition effectiveness of the model. The sinusoidal excitation frequency ranges from 50 to 100 Hz, and an experimental frequency (including 50 Hz) is selected every 10 Hz; data acquisition is performed at each frequency with amplitudes of 1 N, 1.5 N, 2 N, 2.5 N, and 3 N, respectively.
The sinusoidal time-domain load recognition effect of the model is shown in Figure 10a–c. From the figure, it can be observed that all three neural networks exhibit high accuracy for sinusoidal load identification. Specifically, Figure 10d shows the absolute errors of the three networks, where the CNN-LSTM-SA neural network has significantly lower errors than the other two networks in the time domain of the test set. This indicates that this network outperforms the other two models in the sinusoidal load identification task. Therefore, the CNN-LSTM-SA neural network used in this study has more superior accuracy and robustness in the sinusoidal load identification task. The model evaluation indices are shown in Table 3; compared with the other two neural networks, the neural network load identification method used in this paper has a smaller RMSE and MAE. This indicates that the method has excellent load recognition performance. Meanwhile, the R2 value of the proposed network in this paper is closest to 1, indicating that the method used in this paper has the best fitting effect among the three neural networks.

4.3. Time-Domain Identification of Random Load

Since the structure is often subjected to not only stable sinusoidal loads but also random loads under real working conditions, this paper also applies the proposed method to the time-domain load identification of random vibration tests to verify the generalizability of the method. Random signals are commonly described by power spectral density (PSD). In the experimental part of this study, 30 sets of time-domain excitation force and acceleration response information are collected under the same PSD with a sampling frequency of 12,800 Hz. A total of 25 of these sets are used as a training set to train the neural network load identification model, and 5 sets are used as a test set to evaluate the load identification effect of the model. The evaluation metrics used here are the same as those used for sinusoidal excitation identification. The PSD for the experimental setup is shown in Figure 11.
The random time-domain load recognition effect of the model is shown in Figure 12a–c, from which it can be seen that the three kinds of neural networks for random time-domain load have this good recognition effect. But the CNN-LSTM-SA network still has this certain advantage, which can be proved from the absolute error diagram; the absolute error diagram of the three kinds of neural networks is shown in Figure 12d. The model evaluation indices are shown in Table 4. It can be seen that, compared with the other two neural networks, the RMSE and MAE of the neural network load identification method used in this paper are smaller, which indicates that this method has an excellent load identification effect. At the same time, the R2 value of this network is closest to 1, which indicates that the method used in this paper has the best fitting effect among the three neural networks, and it also indicates that the neural network model used in this paper has the best result for recognizing the time-domain information of random loads. In summary, the neural network model used in this paper has good applicability for the time-domain information recognition of random loads.
It can be seen from the above study that under the experimental data, the load identification effect of the neural network decreases compared to the simulation data. This is mainly due to the fact that the experimental data are not only disturbed by white noise but also by abnormal time-series data and other uncertainties. Nevertheless, the neural network model proposed in this paper still shows good identification ability, which fully proves its adaptability and effectiveness in complex environments. On the other hand, the number of training parameters for the model increases significantly due to the mixing of multiple network structures. In addition, under high-frequency excitation, a higher sampling frequency is required to ensure sufficient accuracy, which also leads to an increase in the size of the dataset. The combination of these factors increases the cost of network training.

5. Conclusions

This paper proposes a time-domain load identification method based on a CNN-LSTM-SA neural network. The method takes segmented time-series data as input and combines the advantages of the three networks to achieve higher recognition accuracy and robustness. The main conclusions are the following:
(1)
Simulation results show that for sinusoidal load identification, the CNN-LSTM-SA network has obvious advantages in terms of recognition accuracy and noise immunity. The RMSE and MAE are 0.47 and 0.53 under 0% noise and 8.8 and 8.5 under 20% noise, respectively;
(2)
The experimental results show that the CNN-LSTM-SA network achieves high identification accuracies in both sinusoidal and random load identification tasks (RMSE of 0.08 and 0.83; R2 of 0.98 and 0.93, respectively);
(3)
The CNN-LSTM-SA-based load identification method provides researchers with a tool with higher accuracy and noise immunity, as well as a reliable method for structural health monitoring and optimal design.

Author Contributions

Conceptualization, Y.Z. and J.W.; methodology, S.S. and J.W.; software, J.W. and Y.Z.; validation, S.S. and Y.Z.; formal analysis, J.W.; investigation, J.W.; resources, S.S. and C.L.; data curation, S.S.; writing—original draft preparation, Y.Z.; writing—review and editing, J.W. and C.L.; visualization, J.W.; supervision, J.W.; project administration, J.W.; funding acquisition, J.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data provided in this study can be provided at the request of the corresponding author. The research data presented in the paper consist of simulation and experimental results. We have provided the necessary simulation settings and experimental setups within this paper to enable scholars to replicate our results. Therefore, it is not necessary for us to specifically upload the data. Additionally, we have not uploaded the data from this paper to any public datasets. However, if there is a need, we are willing to provide the data upon request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Liu, R.X.; Dobriban, E.; Hou, Z.C.; Qian, K. Dynamic Load Identification for Mechanical Systems: A Review. Arch. Comput. Methods Eng. 2022, 29, 831–863. [Google Scholar] [CrossRef]
  2. Bartlett, F.D.; Flannelly, W.G. Model Verification of Force Determination for Measuring Vibratory Loads. J. Am. Helicopter Soc. 1979, 24, 10–18. [Google Scholar] [CrossRef]
  3. Li, K.; Liu, J.; Wen, J.; Lu, C. Time domain identification method for random dynamic loads and its application on reconstruction of road excitations. Int. J. Appl. Mech. 2020, 12, 2050087. [Google Scholar] [CrossRef]
  4. Kazemi, M.; Hematiyan, M.; Ghavami, K. An efficient method for dynamic load identification based on structural response. In Proceedings of the International Conference on Engineering Optimization, Rio de Janeiro, Brazil, 1–5 June 2008; pp. 1–6. [Google Scholar]
  5. Liu, J.; Meng, X.; Zhang, D.; Jiang, C.; Han, X. An efficient method to reduce ill-posedness for structural dynamic load identification. Mech. Syst. Signal Process. 2017, 95, 273–285. [Google Scholar] [CrossRef]
  6. Roseiro, L.; Alcobia, C.; Ferreira, P.; Baïri, A.; Laraqi, N.; Alilat, N. Identification of the forces in the suspension system of a race car using artificial neural networks. In Computational Intelligence and Decision Making: Trends and Applications; Springer: Berlin, Germany, 2012; pp. 469–477. [Google Scholar]
  7. Tang, Q.; Xin, J.; Jiang, Y.; Zhou, J.; Li, S.; Chen, Z. Novel identification technique of moving loads using the random response power spectral density and deep transfer learning. Measurement 2022, 195, 111120. [Google Scholar] [CrossRef]
  8. Zhang, X.; He, W.; Cui, Q.; Bai, T.; Li, B.; Li, J.; Li, X. WavLoadNet: Dynamic Load Identification for Aeronautical Structures Based on Convolution Neural Network and Wavelet Transform. Appl. Sci. 2024, 14, 1928. [Google Scholar] [CrossRef]
  9. Zhou, J.; Dong, L.; Guan, W.; Yan, J. Impact load identification of nonlinear structures using deep Recurrent Neural Network. Mech. Syst. Signal Process. 2019, 133, 106292. [Google Scholar] [CrossRef]
  10. Yang, H.; Jiang, J.; Chen, G.; Mohamed, M.S.; Lu, F. A recurrent neural network-based method for dynamic load identification of beam structures. Materials 2021, 14, 7846. [Google Scholar] [CrossRef] [PubMed]
  11. Yang, H.; Jiang, J.; Chen, G.; Zhao, J. Dynamic load identification based on deep convolution neural network. Mech. Syst. Signal Process. 2023, 185, 109757. [Google Scholar] [CrossRef]
  12. Wang, C.; Chen, D.; Chen, J.; Lai, X.; He, T. Deep regression adaptation networks with model-based transfer learning for dynamic load identification in the frequency domain. Eng. Appl. Artif. Intell. 2021, 102, 104244. [Google Scholar] [CrossRef]
  13. Lu, W.; Li, J.; Wang, J.; Qin, L. A CNN-BiLSTM-AM method for stock price prediction. Neural Comput. Appl. 2021, 33, 4741–4753. [Google Scholar] [CrossRef]
  14. Huy, T.H.B.; Vo, D.N.; Nguyen, K.P.; Huynh, V.Q.; Huynh, M.Q.; Truong, K.H. Short-term load forecasting in power system using CNN-LSTM neural network. In Proceedings of the 2023 Asia Meeting on Environment and Electrical Engineering (EEE-AM), Hanoi, Vietnam, 15 November 2023; pp. 1–6. [Google Scholar]
  15. Impraimakis, M. Deep recurrent-convolutional neural network learning and physics Kalman filtering comparison in dynamic load identification. Struct. Health Monit. 2024, 1, 31. [Google Scholar] [CrossRef]
  16. Zhang, J.; Luo, F.; Quan, X.; Wang, Y.; Shi, J.; Shen, C.; Zhang, C. Improving wave height prediction accuracy with deep learning. Ocean Model. 2024, 188, 102312. [Google Scholar] [CrossRef]
  17. Hu, X.; Yu, S.; Zheng, J.; Fang, Z.; Zhao, Z.; Qu, X. A hybrid CNN-LSTM model for involuntary fall detection using wrist-worn sensors. Adv. Eng. Inform. 2025, 65, 103178. [Google Scholar] [CrossRef]
  18. Venkateswaran, D.; Cho, Y. Efficient solar power generation forecasting for greenhouses: A hybrid deep learning approach. Alex. Eng. J. 2024, 91, 222–236. [Google Scholar] [CrossRef]
  19. Liu, J.; Meng, X.; Jiang, C.; Han, X.; Zhang, D. Time-domain Galerkin method for dynamic load identification. Int. J. Numer. Methods Eng. 2016, 105, 620–640. [Google Scholar] [CrossRef]
  20. Ilesanmi, A.E.; Ilesanmi, T.O. Methods for image denoising using convolutional neural network: A review. Complex Intell. Syst. 2021, 7, 2179–2198. [Google Scholar] [CrossRef]
  21. Wen, L.; Gao, L.; Li, X.; Zeng, B. Convolutional Neural Network with Automatic Learning Rate Scheduler for Fault Classification. IEEE Trans. Instrum. Meas. 2021, 70, 3509912. [Google Scholar] [CrossRef]
  22. Sang, S.; Li, L. A Novel Variant of LSTM Stock Prediction Method Incorporating Attention Mechanism. Mathematics 2024, 12, 945. [Google Scholar] [CrossRef]
  23. Uchida, K.; Tanaka, M.; Okutomi, M. Coupled convolution layer for convolutional neural network. Neural Netw. 2018, 105, 197–205. [Google Scholar] [CrossRef] [PubMed]
Figure 1. CNN-LSTM-SA load identification model.
Figure 1. CNN-LSTM-SA load identification model.
Materials 18 01255 g001
Figure 2. CNN network model (The dashed squares are convolution kernels and the circles indicate the flattened data).
Figure 2. CNN network model (The dashed squares are convolution kernels and the circles indicate the flattened data).
Materials 18 01255 g002
Figure 3. LSTM network memory cell.
Figure 3. LSTM network memory cell.
Materials 18 01255 g003
Figure 4. CNN-LSTM-SA network structure diagram.
Figure 4. CNN-LSTM-SA network structure diagram.
Materials 18 01255 g004
Figure 5. (a) Simulation object model; (b) simple harmonic excitation application position and response pickup position.
Figure 5. (a) Simulation object model; (b) simple harmonic excitation application position and response pickup position.
Materials 18 01255 g005
Figure 6. (a) Effectiveness of three neural network load recognition (0% noise); (b) absolute error of three neural network load recognition (0% noise).
Figure 6. (a) Effectiveness of three neural network load recognition (0% noise); (b) absolute error of three neural network load recognition (0% noise).
Materials 18 01255 g006
Figure 7. Effectiveness and absolute error (with noise) of two neural network’s load recognition. (a) effectiveness of two neural network’s load recognition (2% noise); (b) absolute error of two neural network’s load recognition (2% noise); (c) effectiveness of two neural network’s load recognition (5% noise); (d) absolute error of two neural network’s load recognition (5% noise); (e) effectiveness of two neural network’s load recognition (10% noise); (f) absolute error of two neural network’s load recognition (10% noise); (g) effectiveness of two neural network’s load recognition (20% noise); (h) absolute error of two neural network’s load recognition (20% noise).
Figure 7. Effectiveness and absolute error (with noise) of two neural network’s load recognition. (a) effectiveness of two neural network’s load recognition (2% noise); (b) absolute error of two neural network’s load recognition (2% noise); (c) effectiveness of two neural network’s load recognition (5% noise); (d) absolute error of two neural network’s load recognition (5% noise); (e) effectiveness of two neural network’s load recognition (10% noise); (f) absolute error of two neural network’s load recognition (10% noise); (g) effectiveness of two neural network’s load recognition (20% noise); (h) absolute error of two neural network’s load recognition (20% noise).
Materials 18 01255 g007aMaterials 18 01255 g007b
Figure 8. Test object.
Figure 8. Test object.
Materials 18 01255 g008
Figure 9. (a) Data acquisition location; (b) test system.
Figure 9. (a) Data acquisition location; (b) test system.
Materials 18 01255 g009
Figure 10. (a) LSTM sinusoidal load recognition effect; (b) CNN-LSTM sinusoidal load recognition effect; (c) CNN-LSTM-SA sinusoidal load recognition effect; (d) absolute errors in sinusoidal load recognition by three neural networks.
Figure 10. (a) LSTM sinusoidal load recognition effect; (b) CNN-LSTM sinusoidal load recognition effect; (c) CNN-LSTM-SA sinusoidal load recognition effect; (d) absolute errors in sinusoidal load recognition by three neural networks.
Materials 18 01255 g010
Figure 11. PSD for random vibration experiments.
Figure 11. PSD for random vibration experiments.
Materials 18 01255 g011
Figure 12. (a) LSTM random load recognition effect; (b) CNN-LSTM random load recognition effect; (c) CNN-LSTM-SA random load recognition effect; (d) absolute errors in random load recognition by three neural networks.
Figure 12. (a) LSTM random load recognition effect; (b) CNN-LSTM random load recognition effect; (c) CNN-LSTM-SA random load recognition effect; (d) absolute errors in random load recognition by three neural networks.
Materials 18 01255 g012
Table 1. Neural network load recognition effect (0% noise).
Table 1. Neural network load recognition effect (0% noise).
RMSEMAE
LSTM1.100.61
CNN-LSTM0.740.55
CNN-LSTM-SA0.470.53
Table 2. Neural network load recognition effect (with noise).
Table 2. Neural network load recognition effect (with noise).
Noise Level2% Noise5% Noise10% Noise20% Noise
RMSEMAERMSEMAERMSEMAERMSEMAE
CNN-LSTM1.621.273.553.186.435.1013.8912.73
CNN-LSTM-SA1.291.132.582.864.764.458.4710.83
Table 3. Effect of time-domain identification of sinusoidal loads.
Table 3. Effect of time-domain identification of sinusoidal loads.
RMSEMAER2
LSTM0.120.140.97
CNN-LSTM0.090.110.97
CNN-LSTM-SA0.080.100.98
Table 4. Random load time-domain recognition effect.
Table 4. Random load time-domain recognition effect.
RMSEMAER2
LSTM2.073.420.88
CNN-LSTM1.392.570.90
CNN-LSTM-SA0.831.690.93
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, J.; Song, S.; Liu, C.; Zhao, Y. Dynamic Load Identification of Thin-Walled Cabin Based on CNN-LSTM-SA Neural Network. Materials 2025, 18, 1255. https://doi.org/10.3390/ma18061255

AMA Style

Wang J, Song S, Liu C, Zhao Y. Dynamic Load Identification of Thin-Walled Cabin Based on CNN-LSTM-SA Neural Network. Materials. 2025; 18(6):1255. https://doi.org/10.3390/ma18061255

Chicago/Turabian Style

Wang, Jun, Shaowei Song, Chang Liu, and Yali Zhao. 2025. "Dynamic Load Identification of Thin-Walled Cabin Based on CNN-LSTM-SA Neural Network" Materials 18, no. 6: 1255. https://doi.org/10.3390/ma18061255

APA Style

Wang, J., Song, S., Liu, C., & Zhao, Y. (2025). Dynamic Load Identification of Thin-Walled Cabin Based on CNN-LSTM-SA Neural Network. Materials, 18(6), 1255. https://doi.org/10.3390/ma18061255

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop