Next Article in Journal
A Multi-Objective Optimization Design Method for High-Aspect-Ratio Wing Structures Based on Mind Evolution Algorithm Backpropagation Surrogate Model
Previous Article in Journal
Hybrid Deep Learning Model for Fault Diagnosis in Centrifugal Pumps: A Comparative Study of VGG16, ResNet50, and Wavelet Coherence Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Data Augmentation and Deep Learning Methods for Pressure Prediction in Ignition Process of Solid Rocket Motors

1
College of Aerospace Engineering, Shenyang Aerospace University, Shenyang 110136, China
2
Key Laboratory of Education Ministry for Modern Design and Rotor-Bearing System, Xi’an Jiaotong University, Xi’an 710049, China
*
Author to whom correspondence should be addressed.
Machines 2024, 12(12), 906; https://doi.org/10.3390/machines12120906
Submission received: 13 November 2024 / Revised: 2 December 2024 / Accepted: 8 December 2024 / Published: 10 December 2024
(This article belongs to the Section Electrical Machines and Drives)

Abstract

:
During the ignition process of a solid rocket motor, the pressure changes dramatically and the ignition process is very complex as it includes multiple reactions. Successful completion of the ignition process is essential for the proper operation of solid rocket motors. However, the measurement of pressure becomes extremely challenging due to several issues such as the enormity and high cost of conducting tests on solid rocket motors. Therefore, it needs to be investigated using numerical calculations and other methods. Currently, the fundamental theories concerning the ignition process have not been fully developed. In addition, numerical simulations require significant simplifications. To address these issues, this study proposes a solid rocket motor pressure prediction method based on bidirectional long short-term memory (CBiLSTM) combined with adaptive Gaussian noise (AGN). The method utilizes experimental pressure data and simulated pressure data as inputs for co-training to predict pressure data under new operating conditions. By comparison, the AGN-CBiLSTM method has a higher prediction accuracy with a percentage error of 3.27% between the predicted and actual data. This method provides an effective way to evaluate the performance of solid rocket motors and has a wide range of applications in the aerospace field.

1. Introduction

Solid rocket motors are frequently employed in domains such as missiles. The propellant within solid rocket motors is directly stored in the combustion chamber and can be preserved for an extended period, typically ranging from 5 to 20 years. Solid rocket motors possess a simplistic structure, are facile to utilize, and entail reduced maintenance requirements [1]. In the future, solid rocket motors will continue to play an important role in missile, aerospace, and other fields, making greater contributions to military and scientific and technological development.
The ignition process is the initial step of generating power for a solid rocket motor, and it is the key for the motor to work normally. Although the ignition process generally has an extremely short duration, the pressure undergoes significant alterations during this course, completing the entire pressure-building process of the motor. Consequently, it exerts a substantial influence on the operational process of the solid rocket motor. Figure 1 illustrates the pressure-time variation curve during the whole working process of the theoretical combustion chamber. In Figure 1, t0 is the ignition process time of the motor, t1 is the normal working time of the motor, t2 is the shutdown process time, P0 is the initial pressure of the combustion chamber, and Pm is the maximum ignition pressure.
The pressure variation in the ignition of the same type of solid rocket motor under diverse conditions is also dissimilar. As illustrated in Figure 2, the solid rocket motor pressure varies under the storage temperature conditions of high temperature (+50 °C), low temperature (−40 °C), and room temperature (+20 °C).
The pressure variation in the ignition process can be acquired through ground tests, and pressure sensors are utilized to collect data during this procedure [2,3]. Since the internal working environment of solid rocket motors is extremely harsh, it is difficult to collect abundant data in practice. And relying only on empirical or semi-empirical design methods cannot satisfy the requirements of modern advanced solid rocket motor design. Therefore, numerical simulation has become one of the effective tools for solid rocket motor design.
At present, research on numerical simulation in the ignition process has gradually matured. Researchers have proposed several numerical models, such as the fluid–structure interaction model [4,5] and the two-dimensional/three-dimensional turbulence model [6,7]. With the development of computer technology, fluid simulation software is gradually being applied to the simulation of the solid rocket motor ignition process. In recent years, numerous new methods have been proposed. For instance, Francois et al. [8] proposed coupling the computational fluid dynamics (CFD) solver of the combustion chamber flow field with the one-dimensional solver of unsteady propellant combustion and proposed two solution methods. Li et al. [9] numerically studied the second pulse ignition transients in a double-pulse solid rocket motor with an elastic barrier pulse separation device, and proposed a governing equation for unsteady compressible flow, heat conduction, and structural dynamics.
Artificial intelligence methods have received more attention because of their powerful inference and generalization abilities for complex engineering systems [10,11,12]. Data-driven machine learning methods such as the Support Vector Machine (SVM) [13] and Bayesian Network (BN) [14] are widely used. However, machine learning methods face challenges such as poor learning ability and slow learning speed when dealing with large amounts of data. Deep learning can better address these problems. Methods such as Convolutional Neural Network (CNN) [15,16], Recurrent Neural Network (RNN) [17], Deep Belief Network (DBN) [18], and Long Short-Term Memory (LSTM) [19,20] have achieved good results in speech recognition [21], image recognition [22], and natural language processing [23]. Artificial intelligence methods have been widely applied in industrial fields [24,25], such as mechanical fault diagnosis [26,27], prediction of instrument residual life [28], and evaluation of instrument health status [29,30]. With the continuous development and improvement of artificial intelligence technology, its application prospects in various fields will become even broader.
The application of deep learning methods to solve solid rocket motor problems has attracted significant attention. Surina et al. [31] proposed an image-based deep learning tool to verify the boundary layer combustion mechanism of a rocket motor with mixed fuels by predicting the fuel regression rate. Lee et al. [32,33] proposed an optimization method for solid propellant grain design based on neural networks. Six neural network-based methods were tested and compared to optimize solid propellant grain design. Liu et al. [34] proposed a Deep CNN model for defect scale prediction of solid rocket motors in the case of chamber splitting and delamination. Zhang et al. [35] proposed a data-driven method using deep learning to predict the thrust of a solid rocket motor.
Due to the difficulty of obtaining data samples for solid rocket motors, the adoption of data augmentation methods to enhance the richness of data samples is essential. Data augmentation is a technical measure that generates new data samples through various transformations and expansions of the original data, thereby augmenting the size and diversity of the dataset [36]. Data augmentation is a regularization strategy that implements functional regularization by adding noise to the intermediate activations of the network or imposing constraints on the functional form [37].
In this study, a pressure prediction method for the solid rocket motor based on data enhancement and a deep learning model is proposed by combining experimental data and simulation data. Firstly, the one-dimensional data are enhanced by the AGN method, and the sample size of the data is expanded. Then, the CBiLSTM neural network model is established to learn the pressure data under different working conditions and predict the pressure under new working conditions. This process is concluded in Figure 3.

2. Methodology

This section primarily focuses on the solid rocket motor pressure prediction method of AGN-CBiLSTM. AGN is incorporated into the data augmentation method. The network models comprise a one-dimensional convolutional layer, a BiLSTM layer, and Dropout. The pressure prediction process primarily consists of three stages: data preparation, network model training, and network model testing. This section will introduce the AGN method, the CBiLSTM model structure, and the network training process respectively.

2.1. Data Augmentation Method

Through data augmentation technology [38], various operating environments and working conditions can be simulated to generate data samples that are closer to reality. This can not only improve the training effect of the deep learning model but also enhance the model generalization ability so that it can better adapt to various complex operating environments. Therefore, in the field of solid rocket motors, it is very necessary and meaningful to use data augmentation methods to increase the richness of data samples.
AGN is a type of Gaussian noise that adaptively adjusts as the maximum difference in sequence data changes [39]. This noise is applied directly to the original sequence data, assuming that the input sequence is x i n p u t = [ x 1 , x 2 , , x n ] , The augmentation sequence is x A G N , α g a u s s is the Adaptive Gaussian Noise coefficient.
x A G N = x i n p u t + α g a u s s G ( x m a x x m i n )
G = 1 2 π e x 2 2
Different pressure data can be obtained given different sizes of α g a u s s . Applying the AGN method with the same α g a u s s multiple times to the same sequence of data can also generate different pressure data, thus improving the diversity of data to a certain extent. Figure 4 illustrates the change of a set of pressure data after adding Gaussian noise. The number of data augmentation is 1, the α g a u s s size is 0.1.
Data augmentation relies on specific assumptions for Gaussian noise coefficients. However, this approach based on specific assumptions may lead to poor performance when generalized to other application scenarios or different datasets. The characteristics, structure, and distribution of the dataset need to be considered in practical applications. Blindly applying the data augmentation strategy based on Gaussian noise coefficient assumptions will destroy the original characteristics of the data, making the output violate the expected results.
The selection of the number of data augmentations is equally important. Although too much expansion contains more feature information, it will increase the computational complexity, increase the computational cost, and aggravate the training cycle of the neural network. However, if the number of expansions is too small, it is difficult for the model to fully learn the laws and patterns embedded in the data, which is a dilemma.
In this study, a reasonable noise factor and expansion number are selected based on the characteristics of solid rocket motor pressure data.

2.2. Deep Learning Model Structure

2.2.1. One-Dimensional Convolution Structure

One-dimensional convolution is a convolution calculation method used in CNNs, which is primarily used to process one-dimensional sequence data, such as text and time series data [40]. Convolution works by sliding along the direction of the input sequence data. During each sliding process, the convolution kernel performs convolution operations on only a small portion of the sequence data, called the local connection. The design of this local connection allows the convolutional neural network to effectively extract local features from the input data. The pressure time series data are taken as the input data to obtain features by convolution. By shifting the convolution kernel, a local convolution operation can be performed on the input data to obtain local features. These local features are combined to form a feature map that contains the characteristics of the pressure changes of the solid rocket motors during the ignition process. The operation process of one-dimensional convolution is illustrated in Figure 5.

2.2.2. Long Short-Term Memory Structure

LSTM is an improvement over RNN, designed to solve the problem of gradient disappearance or gradient explosion in RNN [41,42]. The pressure data of the ignition process are long series data, which are suitable for LSTM prediction. Unlike the structure of RNN, LSTM has a memory unit to store past information. In addition, LSTM introduces three gate control units: the input gate, the forgetting gate, and the output gate. The input gate determines how much of the current input information can enter the memory cell. The forget gate controls how much of the previous moment’s information is retained in the memory cell and can selectively discard past information that is no longer important. The output gate is responsible for determining which information in the memory cell will be output for prediction at the current moment and its impact on the subsequent network state. Through the coordinated work of these three gates, LSTM can handle the long-term dependencies in long sequence data more intelligently, effectively alleviating the dilemma that traditional RNNs cannot effectively learn long-term information due to the vanishing or exploding gradients when dealing with long sequences.
The sigmoid function is used to control the importance of different elements to the current state, the degree of forgetting historical information, and the hidden state output corresponding to the current state. The basic unit of the LSTM network is concluded in Figure 6.
BiLSTM is a further improvement based on LSTM and is composed of a forward LSTM and a backward LSTM [43]. As concluded in Figure 7, pressure data are first input into the forward LSTM input layer to obtain a set of forward output pressure data change information. This change information is input into the backward LSTM layer to obtain the reverse LSTM output, and the forward and reverse outputs are spliced to obtain the final pressure change information. BiLSTM combines forward LSTM and reverse LSTM to obtain a more complete picture of pressure data changes, enabling more accurate prediction of pressure data.

2.2.3. Convolutional Bidirectional Long Short-Term Structure

The output dimension of the first layer of the convolutional layer used in this article is 8, and the output dimension of the second layer is 16. The size of the convolution kernel of both convolutional layers is 3 × 1, the stride of the convolution is 1, and the convolution formula is as follows:
y 1 = K i · x i ( i = 1,2 , 3 )
After convolution calculation, the pressure characteristic sequence Y 1 = [ y 1 , y 2 , , y n ]   is obtained. The pressure feature sequence Y 1 is placed in the second layer of convolution for calculation to obtain the pressure feature sequence Y 2 . Y 2 contains all the change characteristics of the pressure data. To effectively obtain more dimensional information, nonlinear mapping must be introduced in the modeling process. This nonlinear mapping is called an activation function. This article uses LeakyRelu as the activation function. The calculation formula of the LeakyRelu function is as follows:
L e a k y R e L u ( x ) = x , x > 0 a x , x 0
Input the convolved pressure feature sequence into BiLSTM. The input dimension of the BiLSTM layer is 16, the hidden layer dimension is 100, and the number of network layers (Num Layers) is 1. Output the change information ht of pressure data:
h t = L S T M ( x t , h t 1 ) h t = L S T M ( x t , h t 1 ) h t = W f d h t + W b d h t + b t
Among them, h t is the forward hidden layer state, h t is the backward hidden layer state, W f d and W b d represent the weight parameters of the forward hidden layer and backward hidden layer, respectively, and b t is the bias parameter of the hidden layer.
Enter the calculated result h t into the fully connected layer. The linear layer plays a key role in neural networks. It fully connects all neurons in the previous layer with all neurons in the current layer and realizes the combination and transformation of features by learning weight parameters. Output predicted pressure data after full connection layer integration.
The structure of the CBiLSTM model established in this study is shown in Figure 8. The model contains two convolution layers and activation function layers, one BiLSTM layer and one fully connected layer.
CBiLSTM is designed to combine the advantages of CNN and BiLSTM to better process sequence data. Traditional LSTM networks may encounter the problem of gradient vanishing or gradient explosion when dealing with long sequences, while BiLSTM can improve the model’s ability to capture contextual information by considering both forward and backward information of sequences. However, BiLSTM may encounter computational efficiency problems when processing high-dimensional data. In contrast, CNNs perform well in processing images or high-dimensional features with feature extraction and dimensionality reduction capabilities. Therefore, by combining CNN and BiLSTM, CBiLSTM aims to improve the ability of the model to process sequential data while maintaining computational efficiency. This combination allows the CBiLSTM model to have higher accuracy and robustness in processing complex sequence data.
The model designed in this study includes two convolutional layers because convolutional layers are designed to extract useful features from the input data and learn more complex and abstract feature representations by stacking multiple convolutional layers. These high-level feature representations help subsequent BiLSTM layers to better understand and process sequence data. By stacking multiple convolutional layers, the CBiLSTM model learns richer feature representations, which helps the model improve its performance on complex tasks. In addition, additional convolutional layers increase the depth of the model, which enhances the model’s learning and generalization capabilities.
The modification of CBiLSTM is to combine the advantages of CNN and BiLSTM to improve the model performance, while the introduction of 2 convolutional layers is to improve the feature extraction ability as well as the learning ability of the model. These changes make CBiLSTM more efficient and accurate in processing sequence data.

2.3. Model Training Process

Neural network prediction is mainly divided into three stages: data preparation stage, model training stage, and model testing stage. As shown in Figure 9, the data preparation stage involves processing the ignition test data and simulation data of the solid rocket motor into one-dimensional data, dividing the data into a training set and a test set, and performing data augmentation processing on the training set data. The model training phase requires establishing a network model and determining model parameters such as convolution kernel size, stride, and number of LSTM hidden layers. Backpropagation is used to optimize and update the weight parameters, and the Adam optimizer is used. The learning rate is adjusted during training. Initially, a larger learning rate is used for fast convergence, and then a smaller learning rate is used to continue training until the training round is complete. In the model testing phase, the input portion of the test set is fed into the model, the pressure prediction is performed, and the prediction results are compared with the actual results to analyze the errors.

3. Results and Comparisons

In this section, the AGN-CBiLSTM method is used to predict the pressure data during the solid rocket motor ignition process. The sources of pressure data are solid rocket motor ground experiments and fluid simulations. These two types of data are processed into one-dimensional data and normalized, and finally the training set and test set are divided to complete the data preparation process. The superiority of the AGN-CBiLSTM method is highlighted by comparing it with mainstream deep learning methods.

3.1. Data Preparation

3.1.1. Solid Rocket Motor Turbulence Simulation Model

The main structure of a solid rocket motor consists of a propellant, a combustion chamber, an igniter, a long tail nozzle, and a nozzle plug cap. After receiving the ignition command, the igniter generates a high-temperature flame to ignite the propellant. After the propellant is ignited, it begins to burn vigorously, producing high-temperature, high-pressure gas. The long-tailed nozzle has a special design shape that can efficiently convert the thermal energy of the gas into kinetic energy, so that the gas is ejected at high speed, generating a huge reaction force. The nozzle cap plays the role of sealing and protecting the internal structure of the nozzle before the motor is started, and after the motor is ignited, it will be broken under the strong pressure of the gas, and the gas can be sprayed out of the nozzle.
Due to the problems of long computation time using 3D simulation, some necessary simplifications of the model structure of the solid rocket motor are made in this study. Considering the axisymmetric distribution of the solid rocket motor structure, only the upper part is simulated. Therefore, a 1/2 axisymmetric two-dimensional model is used to simulate the transient ignition process. The entire flow field region is discretized into a mesh, and each region is treated accordingly based on the mesh density. The meshes of the wall and inlet regions should be relatively dense. Since the simulation time is relatively short, the opening of the plug cover is not considered, and the nozzle expansion part is truncated during the simulation to simplify the calculation. The mesh structure of the solid motor is shown in Figure 10.
The two main parts in Figure 11 are the combustion chamber and the long tail nozzle. The igniter is placed inside the combustion chamber during the simulation, and three locations are selected to be zoomed in to show the mesh boundaries more clearly.
The S-A model is a one-equation turbulence model with the turbulent viscosity transport equation as the core. The model equation is:
t ρ v 1 + x i ρ v 1 u i = C b 1 ρ S 1 v 1 C w 1 ρ f w v 1 d 2 + 1 σ v ~ x j μ + ρ v 1 v 1 x j + C b 2 ρ v 1 x j 2
f v 1 = χ 3 χ 3 + C v 1 3 , χ v 1 v , f w = g 1 + C w 3 6 g 6 + C w 3 6 1 6 , g = r + C w 2 r 6 r , r v 1 S 1 κ 2 d 2
C b 1 = 0.1355 , C b 2 = 0.622 , C w 2 = 0.3 , C w 3 = 2 , C v 1 = 7.1 , κ = 0.41
In this study, an axial igniter is used with black powder as the propellant. During ignition, the igniter injects high-temperature gas along the axis of the combustion chamber to ignite the propellant surface.
The source of the simulation pressure data is the normal temperature ignition simulation under different igniter propellant weights. Different igniter propellant weights will affect the ignition mass flow rate of the igniter, thereby affecting the pressure change process. The ignition mass flow rate equation used in this article is:
m ˙ = a t 2 + b t + c m ˙ t = 0 = 0 m ˙ t = t i g = 0
0 t i g m ˙ d t = M i g
Mig is the mass of igniter propellant weight, tig is the time when ignition ends, and m ˙ is the mass flow rate of ignition propellant.
In this study, a polynomial fitting method is used to simulate the flame propagation process of the igniter, and the propellant weight of the igniter is set from 12 g to 54 g. A third-order polynomial equation is used to fit the mass flow rate of the igniter, and the ignition duration is set to 100 ms, and the mass flow rate of the igniter reaches its maximum value at 30 ms. This is then set as the mass flow rate input condition using the User Define Function (UDF) functionality.
The initial pressure of the model, the axial and radial velocities of the gas, and the temperature of the combustion chamber should also be considered in the calculation of this model. The inlet to the igniter is a conventional mass flow inlet and the injection duration must be set. The initial pressure of the model is set to 0.101325 MPa, the axial and radial velocities are set to 0, and the temperature is set to 300 K. The mass flow inlet is used at the inlet, and the injection duration is set to 0.01 s with a time step of 0.00001 s.

3.1.2. Fluid Simulation Pressure Data

As concluded in Figure 11, this paper conducted a total of 8 sets of simulations. The selected igniter propellant weights changed from 12 g to 54 g, and the length of each set of data was 900 data points. It can be seen from Figure 11 that when the same ignition device is used, the greater the igniter propellant weight, the faster the pressure changes.
Figure 11. Simulation of pressure data using different ignition propellant qualities.
Figure 11. Simulation of pressure data using different ignition propellant qualities.
Machines 12 00906 g011

3.1.3. Solid Rocket Motor Ground Experiment Pressure Data

As concluded in Figure 12, the data used in this article come from the same model of solid rocket motor ground test experiment. There are 10 sets of test data, and the length of each set of data is 900 data points. Each different-colored curve represents ignition ground experiments conducted under different solid rocket motor ignition initial conditions. The yellow lines represent all ground experiments conducted under high temperature conditions, the green lines represent ground experiments conducted under normal temperature conditions, and the blue lines represent ground experiments conducted under low temperature conditions.

3.1.4. Data Set Partition Processing

Under the conditions of high temperature and normal temperature, 8 sets of pressure experiment data were selected. When the igniter propellant weight is 12 g to 54 g, 8 sets of simulation data are selected as the input data of the model. Two sets of pressure test curves with the igniter propellant weight of 12 g at low temperatures were used as labels. The AGN method was applied to improve the data. The noise coefficient was set to 0.002, and each data set was enhanced 20 times. After data enhancement, there are 20 data sets for each working condition, with a total of 320 data sets. During the training, 1 set of pressure data was randomly selected under each working condition, and a total of 16 sets of data were used as input to the model. Then, 1 set was randomly selected from 2 sets of pressure data under low temperature and 12 g experimental conditions as a label. The ignition pressure data after data augmentation are shown in Figure 13.

3.1.5. Data Normalization Processing

To facilitate subsequent model learning, the min-max normalization method is used to normalize the input data. Min-max normalization is a linear transformation method aimed at scaling the original data into a specified range, typically [0, 1]. Through this method, the numerical values among different features become comparable, while reducing the sensitivity of the model to feature scales. Consequently, it enhances the accuracy and stability of the algorithm. The mathematical expression for Min-max normalization is:
x * = x x m i n x m a x x m i n
x m a x is the maximum value in the data, and x m i n is the minimum value in the data.

3.2. Different Network Prediction Results

To verify the superiority of the CBiLSTM method, this study compares it with four different neural network models, namely, Convolutional Neural Network (CNN), Long Short-Term Memory (LSTM), Bidirectional Long Short-Term Memory (BiLSTM), and Gated Recurrent Unit (GRU). After processing the ignition pressure data using the data augmentation method under the same conditions, different neural network models are used for comparison. The α g a u s s used for experimental and simulation data is 0.002 and the number of data augmentations is 20. This verifies that the AGN-CBiLSTM model is more advantageous under the same training conditions.
In all comparison methods in this article, the input layer and output layer of the model are the same as in the CBiLSTM method, and the optimizer used for updating the model parameters is Adam. The maximum training epochs are kept the same for all models. In the variable learning rate method, the first 25 epochs of training are set to 0.0001 to quickly optimize the network, and then a learning rate of 0.00001 is used for accurate convergence. Table 1 illustrates the hyperparameters adopted by all the neural networks.
To better illustrate the change in the model loss function during the training process, we plot the loss function values of the first 20 epochs for comparison, as shown in Figure 14.
The loss values of CBiLSTM decrease faster for the same epochs, indicating that the CBiLSTM model exhibits higher learning efficiency at the early stage of training. CBiLSTM can converge to lower loss values faster than comparison models (e.g., traditional BiLSTM or other variants), which usually means that its parameter optimization process is more efficient, and the model is able to capture key features in the data faster.
The predicted pressure results of 50 training courses were averaged. The prediction results of the four models are concluded in Figure 15.
It can be seen from Figure 15 that the predicted pressure data are basically consistent with the pressure data in the test data set. There is some fluctuation in the forecast pressure data at data points 200–400. Figure 16 illustrates a zoomed-in view for sequence points 200–400, and the predicted pressure of the AGN-CBiLSTM model is much closer to the real data, indicating that its prediction is more satisfactory.
Each neural network was trained 50 times and t-distributed Stochastic Neighbor Embedding (t-SNE) images of the prediction results and the test set stress data were plotted. t-SNE is a technique commonly used for downscaling and visualization and is particularly suited for exploring complex structures and distributions in high-dimensional data. In Figure 17, the distribution of the predictions of each model mapped to a two-dimensional space after 50 training sessions is shown. Through the t-SNE image, the clustering effect and distribution characteristics of the prediction results of different models can be visualized.
For the CBiLSTM model, the t-SNE images of its prediction results show a clear clustering structure, with the predicted values closely surrounding the cluster centers of the real test set stress data, with only two points deviating from the larger ones. This indicates that the prediction results of the CBiLSTM model are not only accurate but also have a good generalization ability to capture the intrinsic features of the test set data well.
It should be noted that although t-SNE images can provide intuitive visualization, they may distort the high-dimensional structure of the original data to a certain extent due to the nonlinear transformations in their dimensionality reduction process. Therefore, when interpreting t-SNE images, we need to combine the actual prediction performance of the model with other evaluation indicators to make a comprehensive judgment.
To test the prediction effect of the model, the Mean Absolute Error (MAE), Root Mean Square Error (RMSE), and Mean Absolute Percent Error (MAPE) are used as the evaluation functions of the model.
M A E = 1 m i = 1 n y i y ^ i
R M S E = 1 m i = 1 n y i y ^ i 2
M A P E = 1 m i = 1 n y i y ^ i × 100 %
Figure 18 and Table 2 show the experimental results of different network models for pressure prediction. In these results, we can clearly see the differences in the performance of each model in terms of prediction accuracy. The GRU method achieves the maximum values of MAE, RMSE, and MAPE in the experiment. This indicates that the accuracy of the GRU model in predicting pressure data is relatively low, and there is a large deviation between its prediction results and the actual values.
In contrast, the CBiLSTM method exhibits minimum values in all four error metrics. This fully proves that the CBiLSTM model has excellent performance in predicting pressure data, and its prediction results are not only accurate but also stable. The CBiLSTM model can fully exploit the temporal and spatial features of the data to make more accurate predictions.
The errors of BiLSTM, LSTM, and CNN, on the other hand, are intermediate between GRU and CBiLSTM. This indicates that these three models have some accuracy in predicting the stress data, but their performance could be improved relative to the CBiLSTM model.
It is worth noting that among the two models based on long and short-term memory networks, BiLSTM and LSTM, the error of BiLSTM is slightly smaller than that of LSTM. This suggests that the bi-directional structure can capture the temporal characteristics of the data more efficiently, thus improving the prediction accuracy of the model. The bi-directional structure allows the model to consider both past and future information when processing data at the current time step and therefore provides a more comprehensive understanding of the dynamics of the data.
In summary, the four errors of the CBiLSTM method are the smallest, while the error of the GRU method is the largest. The CBiLSTM method has the advantage of high precision in predicting solid rocket motor pressure data.
By comparing the performance of these four networks, the CBiLSTM method has significant advantages in accuracy and other aspects. In contrast, the four networks of CNN, LSTM, BiLSTM, and GRU have certain limitations when processing sequence data and cannot capture complex patterns and dependencies in sequence data as effectively as the CBiLSTM method. Therefore, the CBiLSTM method is a more excellent neural network model with broad application prospects and potential value.
In this way, after inputting the pressure data of similar solid rocket motors under different operating conditions, it is possible to predict the pressure data under new operating conditions. This helps to better understand the pressure dynamics inside the engine and thus optimize the structural design of the engine. If higher pressure peaks are predicted under certain operating conditions, then the combustion chamber walls need to be designed with higher-strength materials or increased wall thickness to ensure engine safety and reliability.
In the future, the failure of the model can be analyzed by introducing data from extreme working conditions (e.g., ultra-high temperatures, strong corrosive environments) and using simulation software to simulate the extreme and abnormal conditions, generating a large amount of synthetic data to supplement the lack of actual test data. Inputting the extreme working condition data to verify whether the neural network can maintain stability for extreme conditions. Develop methods to quantify the uncertainty of model predictions for extreme conditions.

4. Conclusions

An AGN-CBiLSTM solid rocket motor pressure prediction method has been proposed to solve the problem of different pressure changes during the ignition process of solid rocket motors under different operating conditions. Through the processing of experimental and simulation data, experimental pressure data and fluid simulation pressure data are simultaneously used as input data for collaborative training. A model suitable for solid rocket motor data was established with the purpose of predicting the pressure data under new working conditions by inputting the pressure data of the same type of solid rocket motor under different working conditions. By fully training the model, the method achieves good results on the test set. The AGN-CBiLSTM method has MAE = 0.1431, RMSE = 0.1779, MAPE = 3.27%. The prediction results concluded a small error and are close to the actual data.
Compared with other commonly used deep learning methods, this method has the accuracy and effectiveness of prediction. The two error indicators of the AGN-CBiLSTM method are smaller than the errors of the other three deep learning methods. The results indicate that this method can effectively reflect the changes in actual pressure.
Although the currently proposed AGN-CBiLSTM method has achieved favorable results in experiments, further optimization of the model structure is still required to reduce the prediction error and reflect the pressure changes during the ignition process of solid rocket motors more accurately. Although it has been pointed out in the research that the AGN method is helpful for expanding data samples, deep learning models still rely heavily on a large amount of training data. In the future, more efficient data augmentation strategies or methods to reduce data dependence need to be explored to ensure that the model can maintain a high level of prediction accuracy with limited data. Currently, this method has shown good performance on the test set, and it can be attempted to be widely applied in the design practices of solid rocket motors under different types and working conditions, to better provide theoretical support and practical guidance for the design of new solid rocket motors.

Author Contributions

Conceptualization, X.L.; methodology, H.Y.; software, P.Y.; validation, Y.C.; formal analysis, Y.C. and B.L.; investigation, P.Y.; data curation, B.L.; writing—original draft preparation, P.Y. and Y.C.; writing—review and editing, H.Y. and X.L.; visualization, B.L.; supervision, H.Y.; project administration, X.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Dataset available on request from the authors.

Acknowledgments

We would like to thank the National University of Defense Technology Aircraft. Conceptual Design team for technical guidance and suggestions.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ren, P.; Wang, H.; Zhou, G.; Li, J.; Cai, Q.; Yu, J.; Yuan, Y. Solid Rocket Motor Propellant Grain Burnback Simulation Based on Fast Minimum Distance Function Calculation and Improved Marching Tetrahedron Method. Chin. J. Aeronaut. 2021, 34, 208–224. [Google Scholar] [CrossRef]
  2. Migliorino, M.T.; Aiello, M.; Berti, M.; Rotondi, M.; D’Alessandro, S.; Bianchi, D.; Jahjah, M.; Pizzarelli, M. Student Firing Tests and Launches with Commercial and Self-Made Solid Rocket Motors. Acta Astronaut. 2022, 197, 23–34. [Google Scholar] [CrossRef]
  3. Gao, Y.; Liu, Y.; Dong, Z.; Ma, D.; Yang, B.; Qiu, C. Preliminary Experimental Study on Combustion Characteristics in a Solid Rocket Motor Nozzle Based on the TDLAS System. Energy 2023, 268, 126741. [Google Scholar] [CrossRef]
  4. Liu, X.; Xu, F.; Cheng, L.; Pan, W.; Jiao, W. Stress Characteristics Analysis of Vertical Bi-Directional Flow Channel Axial Pump Blades Based on Fluid–Structure Coupling. Machines 2022, 10, 368. [Google Scholar] [CrossRef]
  5. Guo, C.; Zhang, H.; Qian, Z.; Liu, M. Smoothed-Interface SPH Model for Multiphase Fluid-Structure Interaction. J. Comput. Phys. 2024, 518, 113336. [Google Scholar] [CrossRef]
  6. Tian, H.; Duan, Y.; Zhu, H. Three-Dimensional Numerical Analysis on Combustion Performance and Flow of Hybrid Rocket Motor with Multi-Segmented Grain. Chin. J. Aeronaut. 2020, 33, 1181–1191. [Google Scholar] [CrossRef]
  7. Xiao, Q.; Wang, J.; Jiang, B.; Ding, Y.; Yang, X. Study on Nonlinear Correlation in Modal Coefficients of the Bionic Airfoil. Machines 2023, 11, 88. [Google Scholar] [CrossRef]
  8. Francois, L.; Dupays, J.; Massot, M. A New Simulation Strategy for Solid Rocket Motor Ignition: Coupling a CFD Code with a One-Dimensional Boundary Flame Model, Verification against a Fully Resolved Approach. In Proceedings of the AIAA Propulsion and Energy 2021 Forum, Online, 9–11 August 2021; American Institute of Aeronautics and Astronautics: Reston, VA, USA, 2021. [Google Scholar]
  9. Li, Y.; Chen, X.; Cheng, H.; Zhao, Z. Fluid–Structure Coupled Simulation of Ignition Transient in a Dual Pulse Motor Using Overset Grid Method. Acta Astronaut. 2021, 183, 211–226. [Google Scholar] [CrossRef]
  10. Song, L.; Jin, Y.; Lin, T.; Zhao, S.; Wei, Z.; Wang, H. Remaining Useful Life Prediction Method Based on the Spatiotemporal Graph and GCN Nested Parallel Route Model. IEEE Trans. Instrum. Meas. 2024, 73, 3511912. [Google Scholar] [CrossRef]
  11. Chakraborty, S.; Guha, S.; Mishra, D.; Pal, S.K. Federated Learning for Weld Quality Prediction. J. Dyn. Monit. Diagn. 2024. [Google Scholar] [CrossRef]
  12. Wang, X.; Zhou, D.; Zhuang, X.; Ge, J.; Xiang, J. SGG-DGCN: Wind Turbine Anomaly Identification by Using Deep Graph Convolutional Networks with Similarity Graph Generation Strategy. J. Dyn. Monit. Diagn. 2024. [Google Scholar] [CrossRef]
  13. Khelif, R.; Chebel-Morello, B.; Malinowski, S.; Laajili, E.; Fnaiech, F.; Zerhouni, N. Direct Remaining Useful Life Estimation Based on Support Vector Regression. IEEE Trans. Ind. Electron. 2017, 64, 2276–2285. [Google Scholar] [CrossRef]
  14. Melani, A.H.D.A.; Michalski, M.A.D.C.; Da Silva, R.F.; De Souza, G.F.M. A Framework to Automate Fault Detection and Diagnosis Based on Moving Window Principal Component Analysis and Bayesian Network. Reliab. Eng. Syst. Saf. 2021, 215, 107837. [Google Scholar] [CrossRef]
  15. Zhu, Y.; Liu, X. A Lightweight CNN for Wind Turbine Blade Defect Detection Based on Spectrograms. Machines 2023, 11, 99. [Google Scholar] [CrossRef]
  16. Ma, Y.; Du, Z.; Xu, Q.; Qi, J. Flow Field Reconstruction of Compressor Blade Cascade Based on Deep Learning Methods. Aerosp. Sci. Technol. 2024, 155, 109637. [Google Scholar] [CrossRef]
  17. Dudukcu, H.V.; Taskiran, M.; Taskiran, Z.G.C.; Yildirim, T. Temporal Convolutional Networks with RNN Approach for Chaotic Time Series Prediction. Appl. Soft Comput. 2023, 133, 109945. [Google Scholar] [CrossRef]
  18. Sarangi, S.; Dash, P.K.; Bisoi, R. Probabilistic Prediction of Wind Speed Using an Integrated Deep Belief Network Optimized by a Hybrid Multi-Objective Particle Swarm Algorithm. Eng. Appl. Artif. Intell. 2023, 126, 107034. [Google Scholar] [CrossRef]
  19. Sontheimer, M.; Singh, A.-K.; Verma, P.; Chou, S.-Y.; Kuo, Y.-L. LSTM for Modeling of Cylinder Pressure in HCCI Engines at Different Intake Temperatures via Time-Series Prediction. Machines 2023, 11, 924. [Google Scholar] [CrossRef]
  20. Zhou, R.; Zhang, Z.; Zhang, H.; Cai, S.; Zhang, W.; Fan, A.; Xiao, Z.; Li, L. Reliable Monitoring and Prediction Method for Transmission Lines Based on FBG and LSTM. Adv. Eng. Inform. 2024, 62, 102603. [Google Scholar] [CrossRef]
  21. Hema, C.; Garcia Marquez, F.P. Emotional Speech Recognition Using CNN and Deep Learning Techniques. Appl. Acoust. 2023, 211, 109492. [Google Scholar] [CrossRef]
  22. Hossain, S.; Umer, S.; Rout, R.K.; Tanveer, M. Fine-Grained Image Analysis for Facial Expression Recognition Using Deep Convolutional Neural Networks with Bilinear Pooling. Appl. Soft Comput. 2023, 134, 109997. [Google Scholar] [CrossRef]
  23. Chou, J.-S.; Chong, P.-L.; Liu, C.-Y. Deep Learning-Based Chatbot by Natural Language Processing for Supportive Risk Management in River Dredging Projects. Eng. Appl. Artif. Intell. 2024, 131, 107744. [Google Scholar] [CrossRef]
  24. Zhang, W.; Hao, H.; Zhang, Y. State of Charge Prediction of Lithium-Ion Batteries for Electric Aircraft with Swin Transformer. IEEE/CAA J. Autom. Sin. 2024, 1–3. [Google Scholar] [CrossRef]
  25. Balshaw, R.; Heyns, P.S.; Wilke, D.N.; Schmidt, S. Temporally-Preserving Latent Variable Models: Offline and Online Training for Reconstruction and Interpretation of Fault Data for Gearbox Condition Monitoring. J. Dyn. Monit. Diagn. 2024, 3, 156–177. [Google Scholar] [CrossRef]
  26. Li, X.; Zhang, C.; Li, X.; Zhang, W. Federated Transfer Learning in Fault Diagnosis under Data Privacy with Target Self-Adaptation. J. Manuf. Syst. 2023, 68, 523–535. [Google Scholar] [CrossRef]
  27. Li, X.; Yu, S.; Lei, Y.; Li, N.; Yang, B. Dynamic Vision-Based Machinery Fault Diagnosis with Cross-Modality Feature Alignment. IEEE/CAA J. Autom. Sin. 2024, 11, 2068–2081. [Google Scholar] [CrossRef]
  28. Li, X.; Zhang, W.; Li, X.; Hao, H. Partial Domain Adaptation in Remaining Useful Life Prediction With Incomplete Target Data. IEEE/ASME Trans. Mechatron. 2024, 29, 1903–1913. [Google Scholar] [CrossRef]
  29. Yang, S.; Ling, L.; Li, X.; Han, J.; Tong, S. Industrial Battery State-of-Health Estimation with Incomplete Limited Data Towards Second-Life Applications. J. Dyn. Monit. Diagn. 2024. [Google Scholar] [CrossRef]
  30. Qiang, X.; Liu, W.; Lyu, Z.; Ruan, H.; Li, X. A Data-Fusion-Model Method for State of Health Estimation of Li-Ion Battery Packs Based on Partial Charging Curve. Green Energy Intell. Transp. 2024, 3, 100169. [Google Scholar] [CrossRef]
  31. Surina, G.; Georgalis, G.; Aphale, S.S.; Patra, A.; DesJardin, P.E. Measurement of Hybrid Rocket Solid Fuel Regression Rate for a Slab Burner Using Deep Learning. Acta Astronaut. 2022, 190, 160–175. [Google Scholar] [CrossRef]
  32. Lee, H.S.; Ko, S.C.; Kwon, S.W.; Lee, J.S. Optimization of a Solid Rocket Motor Design through Neural Network-Assisted Performance Prediction. IEEE Trans. Aerosp. Electron. Syst. 2023, 59, 8769–8781. [Google Scholar] [CrossRef]
  33. Lee, H.S.; Kwon, S.W.; Lee, J.S. Neural Networks for the Burn Back Performance of Solid Propellant Grains. Aerosp. Sci. Technol. 2023, 137, 108283. [Google Scholar] [CrossRef]
  34. Liu, D.; Sun, L.; Miller, T.C. Defect Diagnosis in Solid Rocket Motors Using Sensors and Deep Learning Networks. AIAA J. 2021, 59, 276–281. [Google Scholar] [CrossRef]
  35. Zhang, W.; Xu, M.; Yang, H.; Wang, X.; Zheng, S.; Li, X. Data-Driven Deep Learning Approach for Thrust Prediction of Solid Rocket Motors. Measurement 2024, 225, 114051. [Google Scholar] [CrossRef]
  36. Liu, Y.; Wang, X.; Zhang, Z.; Deng, F. Deep Learning Based Data Augmentation for Large-Scale Mineral Image Recognition and Classification. Miner. Eng. 2023, 204, 108411. [Google Scholar] [CrossRef]
  37. Shorten, C.; Khoshgoftaar, T.M.; Furht, B. Text Data Augmentation for Deep Learning. J. Big Data 2021, 8, 101. [Google Scholar] [CrossRef]
  38. Qiu, X.; Wang, S.; Chen, K. A Conditional Generative Adversarial Network-Based Synthetic Data Augmentation Technique for Battery State-of-Charge Estimation. Appl. Soft Comput. 2023, 142, 110281. [Google Scholar] [CrossRef]
  39. Jiang, T.; Chen, J.; Wang, Y. A Novel Fractional Nonlinear State Estimation Algorithm in Non-Gaussian Noise Environment. Measurement 2024, 227, 114261. [Google Scholar] [CrossRef]
  40. Parineh, H.; Sarvi, M.; Bagloee, S.A. Detecting Emergency Vehicles With 1D-CNN Using Fourier Processed Audio Signals. Measurement 2023, 223, 113784. [Google Scholar] [CrossRef]
  41. Wang, J.; Zhang, W.; Yang, H.; Yeh, C.-C.M.; Wang, L. Visual Analytics for RNN-Based Deep Reinforcement Learning. IEEE Trans. Vis. Comput. Graph. 2022, 28, 4141–4155. [Google Scholar] [CrossRef]
  42. Mansouri, M.; Dhibi, K.; Hajji, M.; Bouzara, K.; Nounou, H.; Nounou, M. Interval-Valued Reduced RNN for Fault Detection and Diagnosis for Wind Energy Conversion Systems. IEEE Sens. J. 2022, 22, 13581–13588. [Google Scholar] [CrossRef]
  43. Siami-Namini, S.; Tavakoli, N.; Namin, A.S. The Performance of LSTM and BiLSTM in Forecasting Time Series. In Proceedings of the 2019 IEEE International Conference on Big Data (Big Data), Los Angeles, CA, USA, 9–12 December 2019; pp. 3285–3292. [Google Scholar]
Figure 1. Solid rocket motor working process theoretical pressure curve.
Figure 1. Solid rocket motor working process theoretical pressure curve.
Machines 12 00906 g001
Figure 2. Ignition pressure data under different temperature conditions.
Figure 2. Ignition pressure data under different temperature conditions.
Machines 12 00906 g002
Figure 3. The solid rocket motor pressure prediction process based on AGN-CBiLSTM architecture.
Figure 3. The solid rocket motor pressure prediction process based on AGN-CBiLSTM architecture.
Machines 12 00906 g003
Figure 4. Comparison between the AGN method and the initial pressure data.
Figure 4. Comparison between the AGN method and the initial pressure data.
Machines 12 00906 g004
Figure 5. One-dimensional convolution calculation principle.
Figure 5. One-dimensional convolution calculation principle.
Machines 12 00906 g005
Figure 6. Schematic illustration of the operation principle of the LSTM unit.
Figure 6. Schematic illustration of the operation principle of the LSTM unit.
Machines 12 00906 g006
Figure 7. Schematic diagram of BiLSTM operation principle.
Figure 7. Schematic diagram of BiLSTM operation principle.
Machines 12 00906 g007
Figure 8. Schematic diagram of the structural framework of the CBiLSTM model.
Figure 8. Schematic diagram of the structural framework of the CBiLSTM model.
Machines 12 00906 g008
Figure 9. Solid rocket motor pressure prediction workflow based on AGN-CBiLSTM network.
Figure 9. Solid rocket motor pressure prediction workflow based on AGN-CBiLSTM network.
Machines 12 00906 g009
Figure 10. Mesh Structure of 2D Simulation Model of Solid Rocket Motor.
Figure 10. Mesh Structure of 2D Simulation Model of Solid Rocket Motor.
Machines 12 00906 g010
Figure 12. Solid rocket motor ignition ground test experiment pressure data.
Figure 12. Solid rocket motor ignition ground test experiment pressure data.
Machines 12 00906 g012
Figure 13. The ignition pressure data processed by the AGN method.
Figure 13. The ignition pressure data processed by the AGN method.
Machines 12 00906 g013
Figure 14. Loss functions for different neural network models.
Figure 14. Loss functions for different neural network models.
Machines 12 00906 g014
Figure 15. Predicted pressure and true pressure comparison.
Figure 15. Predicted pressure and true pressure comparison.
Machines 12 00906 g015
Figure 16. Enlarged view of predicted pressure.
Figure 16. Enlarged view of predicted pressure.
Machines 12 00906 g016
Figure 17. t-SNE plots of different network prediction results and test samples. (a) CBiLSTM model predicted value distribution; (b) CNN model predicted value distribution; (c) GRU model predicted value distribution; (d) LSTM model predicted value distribution; (e) BiLSTM model predicted value distribution.
Figure 17. t-SNE plots of different network prediction results and test samples. (a) CBiLSTM model predicted value distribution; (b) CNN model predicted value distribution; (c) GRU model predicted value distribution; (d) LSTM model predicted value distribution; (e) BiLSTM model predicted value distribution.
Machines 12 00906 g017
Figure 18. Predicted pressure and true pressure comparison.
Figure 18. Predicted pressure and true pressure comparison.
Machines 12 00906 g018
Table 1. Parameter Settings for neural network.
Table 1. Parameter Settings for neural network.
Parameter NameValuesParameter NameMethods
Learning rate11 × 10−4Learning rate21 × 10−5
Activation functionLeakyReLUoptimizerAdam
Batch size10Loss functionMSE
Maximum epochs50Dropout0.1
Table 2. Error of predicted pressure data.
Table 2. Error of predicted pressure data.
ErrorAGN-CBiLSTMCNNGRULSTMBiLSTM
MAE0.14310.2458 0.2772 0.2390 0.2233
RMSE0.17790.2989 0.3210 0.2803 0.2672
MAPE3.27%5.46% 6.90% 5.88% 4.43%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, H.; Yu, P.; Cui, Y.; Lou, B.; Li, X. Data Augmentation and Deep Learning Methods for Pressure Prediction in Ignition Process of Solid Rocket Motors. Machines 2024, 12, 906. https://doi.org/10.3390/machines12120906

AMA Style

Yang H, Yu P, Cui Y, Lou B, Li X. Data Augmentation and Deep Learning Methods for Pressure Prediction in Ignition Process of Solid Rocket Motors. Machines. 2024; 12(12):906. https://doi.org/10.3390/machines12120906

Chicago/Turabian Style

Yang, Huixin, Pengcheng Yu, Yan Cui, Bixuan Lou, and Xiang Li. 2024. "Data Augmentation and Deep Learning Methods for Pressure Prediction in Ignition Process of Solid Rocket Motors" Machines 12, no. 12: 906. https://doi.org/10.3390/machines12120906

APA Style

Yang, H., Yu, P., Cui, Y., Lou, B., & Li, X. (2024). Data Augmentation and Deep Learning Methods for Pressure Prediction in Ignition Process of Solid Rocket Motors. Machines, 12(12), 906. https://doi.org/10.3390/machines12120906

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop