Next Article in Journal
Machine Learning Quantitative Structure–Property Relationships as a Function of Ionic Liquid Cations for the Gas-Ionic Liquid Partition Coefficient of Hydrocarbons
Previous Article in Journal
Physiological and Transcriptome Analysis on Diploid and Polyploid Populus ussuriensis Kom. under Salt Stress
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Prediction of Time-Series Transcriptomic Gene Expression Based on Long Short-Term Memory with Empirical Mode Decomposition

1
State Key Laboratory of Bioelectronics, School of Biological Science & Medical Engineering, Southeast University, Nanjing 210096, China
2
School of Medicine, Southeast University, Nanjing 210097, China
*
Author to whom correspondence should be addressed.
Int. J. Mol. Sci. 2022, 23(14), 7532; https://doi.org/10.3390/ijms23147532
Submission received: 14 June 2022 / Revised: 3 July 2022 / Accepted: 4 July 2022 / Published: 7 July 2022
(This article belongs to the Section Molecular Informatics)

Abstract

:
RNA degradation can significantly affect the results of gene expression profiling, with subsequent analysis failing to faithfully represent the initial gene expression level. It is urgent to have an artificial intelligence approach to better utilize the limited data to obtain meaningful and reliable analysis results in the case of data with missing destination time. In this study, we propose a method based on the signal decomposition technique and deep learning, named Multi-LSTM. It is divided into two main modules: One decomposes the collected gene expression data by an empirical mode decomposition (EMD) algorithm to obtain a series of sub-modules with different frequencies to improve data stability and reduce modeling complexity. The other is based on long short-term memory (LSTM) as the core predictor, aiming to deeply explore the temporal nonlinear relationships embedded in the sub-modules. Finally, the prediction results of sub-modules are reconstructed to obtain the final prediction results of time-series transcriptomic gene expression. The results show that EMD can efficiently reduce the nonlinearity of the original data, which provides reliable theoretical support to reduce the complexity and improve the robustness of LSTM models. Overall, the decomposition-combination prediction framework can effectively predict gene expression levels at unknown time points.

1. Introduction

The fate of RNA transcripts in dead tissues and the decay of isolated RNA are not subject to strict regulation, unlike in vivo normal cells [1]. Whether most transcripts decay at a similar rate remains unknown at present [2]. Understanding RNA degradation patterns is critical for studies of samples collected in the field or samples collected in clinical settings, as these tissue samples are often not immediately stored under conditions that can prevent RNA degradation [3]. For example, obtaining donated samples in the clinic may require several hours of transport for RNA extraction and sequencing analysis. It is crucial to investigate the decay pattern of RNA degradation over time, because neglecting the initial RNA quality changes in some clinical samples may even lead to incorrect judgments of disease. In addition, some RNA extraction methods need to take several hours. Currently, samples collected during certain low-quality fieldwork remain the only means of solving specific problems [4]. In these cases, the extracted RNA is usually partially degraded and may not faithfully represent the initial gene expression level [5]. Moreover, ignoring the initial RNA quality changes in clinical samples may lead to misjudgment of the disease. Analyzing the effect of RNA degradation on gene expression is essential to obtain meaningful and reliable gene expression data and to ensure reproducible results. Therefore, an artificial intelligence method is needed to predict the decay pattern of RNA degradation over time. If the irreversible change in gene expression level of RNA that occurred during this time can be understood clearly, it will be possible to predict the true gene expression level of RNA when degradation has not occurred by the time of degradation and the existing gene expression level.
Time-series-based gene expression data have become one of the most fundamental methods for studying biological processes. Most biological processes are dynamic, and using time series data can characterize the function of specific genes, discover linkages and regulatory relationships between genes, and find their clinical applications [6,7,8]. Time series data can not only capture transient changes in gene expression, but also demonstrate the order of events that occur during biological processes and allow the study of the temporal patterns of gene expression [9]. These transcriptome time series experiments require observational sampling of biological systems, the preparation of transcriptome libraries, and sequencing at each time point. The current time series experiments have a high overhead. Their sampling time points are generally limited and they are only used with some specific study subjects. Artificial intelligence computational methods are particularly important for humans, which can only perform limited experiments directly. Therefore, there is an urgent need for tools that can study limited time-series-based gene expression data or frameworks and methods that can use time-series-based gene expression data for analytical modeling [10]. So far, this field has been developed and studied by lots of scientists. For example, Lakizadeh et al. [11] proposed the use of time-series-based gene expression data to complement protein–protein interaction (PPI) networks to detect protein complexes. Wise et al. [12] developed SMARTS, a method that combines static and time-series-based data from multiple individuals to reconstruct condition-specific unsupervised feedback networks. Qian et al. [13] used deep neural networks for time-series classification tasks. However, there is no method that combines long short-term memory (LSTM) [14] deep learning models for gene expression prediction of time series.
It is the first prediction framework constructed by deep learning methods for time series data of isolated RNA from tissue samples. The time-series-based prediction is able to extract causality because the information on gene expression flows over time in one direction. Overall, we developed a new LSTM-based deep learning model named Multi-LSTM to predict gene expression at target time points in transcriptome time series data. (Figure 1). This model is capable of predicting the gene expression at initial time points from the gene expression at subsequent time points. It can also predict the gene expression at future times from previous time points. This study has important implications for those studies that lack initial time samples. Moreover, applying the LSTM deep learning model to the analysis of transcriptome time series data can obtain the missing RNA-seq gene expression data, which can facilitate the functional prediction of transcriptome gene expression profiles and the discovery of disease-related genes. With the increase in the number of time-series-based biological experimental data, it will be possible to make full use of the time-series information based on this study to gain a deeper understanding of the gene functions and molecular mechanisms in biological processes.

2. Results and Discussions

2.1. Multi-Scale Analysis Based on EMD

Our data are the expressions of transcripts corresponding to each gene obtained by extracting RNA from mouse brain tissue after transcriptional library building and sequencing. Usually, the normalized Fragments Per Kilobase of exon model per Million mapped fragments (FPKM) value is a standard value to judge the high or low expression of that gene [15]. In this study, the initial gene expression values were normalized as input values and predicted values. Two main time points were predicted in this study. One was the time point when the samples were placed for 0 h when no degradation has occurred. As the target time for prediction, the gene expression data at that time point are most desirable to be known. As a complement, another time point was chosen that was the time point of the sample place for a period of time (6 h). A list of genes associated with RNA degradation in existing studies was found through NCBI. The sampling points represent the expression values of genes corresponding to 12 time points for the 60 genes.
The raw gene expression data are shown in Figure 2a, and the data to be predicted are highlighted with blue dots. Through Figure 2a, it can be seen that the values of the expression of the genes in these time series show a periodic oscillation. In general, these data present a nonlinear, nonsmooth time–frequency signal. Specifically, gene expression data related to the apoptosis and degradation of RNA samples at different time points were utilized, and the gene names and expressions are provided (Supplementary Materials Table S1). These prediction points are the gene expression data at 0 h when the RNA has not yet undergone degradation. It can be seen that these data exhibit complex nonlinearity and nonstationarity, which makes the prediction of gene expression at specific locations extremely difficult. How to solve the nonlinearity problem of gene expression data and make the deep learning neural network operate on a more regular time is the key to improve the prediction accuracy and model robustness. In this regard, the empirical mode decomposition (EMD) [16] signal decomposition technique was utilized, which aimed to reduce the nonlinearity of the original gene expression data and decompose them into a series of stable components with different frequencies. This study is based on each component and is combined with a deep learning predictor, which is expected to improve the accuracy of model analysis.
The decomposition results are shown in Figure 2b. After the EMD component, the original gene expression signal is divided into nine different frequency components of intrinsic mode functions (IMFs) [17]. As a new data analysis method, EMD has obvious advantages in dealing with nonsmooth and nonlinear data. EMD is especially suitable for handling terahertz time-domain signals. Compared with the simple harmonic function, IMF does not have a constant amplitude and frequency in the simple harmonic function, but has an amplitude and frequency that vary as a function of time. The system can achieve remarkably less reconstruction errors at extremely low oversampling rates. The different components contain different gene expression information. The correlation between each component and the original data is quantified by Spearman’s [18] rank-order relationship, and the correlation heat map is shown in Figure 2c. It is obvious that different components contain different gene expression information, and aggregating all components will obtain the total gene expression information. This indicates the high reliability of the multi-scale analysis method proposed in this study. Meanwhile, compared with the original data in Figure 2a, each component shows the trend of the recent harmonic function and has a strong change pattern. IMF can not only be amplitude-modulated but also can modulate frequency. The variable instantaneous frequency and instantaneous amplitude can enhance the efficiency of signal decomposition. At the same time, the EMD method can better preserve the nonsmooth and nonlinear characteristics of the original signal. The eigenmode function obtained by EMD also has the characteristic of intra-wave modulation, which can cover the information of the same component expressed by different Fourier frequencies into one IMF [19]. The EMD method is not only intuitive, simple, complete, approximately orthogonal, and modulated in the IMF components, but also has good adaptive properties. The completeness means that the IMF components obtained by the decomposition can fully recover the original signal. The adaptiveness of the EMD method is mainly manifested in the adaptiveness of the generated basis functions, the adaptive frequency resolution of the IMF, and the adaptive filtering characteristics. The adaptive basis function means that the decomposition quantities obtained by EMD for signal decomposition are an adaptive generalized basis from the perspective of signal analysis. A series of IMFs with variable frequency and variable amplitude are obtained by the fully adaptive signal in the decomposition process. Finally, the trends of the data can be effectively captured using LSTM, which provides an accurate prediction of unknown gene expression.

2.2. Construction of Core Predictor Model Using LSTM

First, the LSTM model was built for each IMF component to obtain the corresponding prediction results. The prediction results of the nine components are shown in Figure 3a–i. Then, we performed the predictions of the gene expression according to these nine modules separately. Specifically, the white dots represent the true values and the blue dots represent the predicted values. The line of blue dots (predicted values) in the results will be very close to the line of white dots (true values). From Figure 3a–i, it can be seen that the initial signal is decomposed into nine IMFs by EMD, and the nonsmooth signal becomes increasingly smooth. It can also be seen that the LSTM can accurately capture the change trend of each component after decomposition, and the prediction accuracy is satisfactory. Figure 4a–i show the comparison of the predicted results of each component with the original results. As a result, the prediction values of each component are closer to the original data. There is no difference between the prediction results of each component and the original data, and all the p-values > 0.05.
In order to quantify the prediction accuracy of each component more directly, the root-mean-square error (RMSE), mean absolute error (MAE), and coefficient of determination (R2) were selected as the prediction performance evaluation indexes. Among them, RMSE can be used to assess the degree of inconsistency between the predicted and true values of the model. Theoretically, the smaller its value is, the better the performance of the model. In addition, the MAE can be used to complement the evaluation of the performance of this model. Meanwhile, the induced effects of the model can be evaluated by the determinants statistic R2. The R2 is an excellent statistical indicator of the closeness between the predicted and true values. The results of R2 of each component are shown in Figure 5a–i. In general, the performance of this model can be fully evaluated by the above three items. The results of the evaluation of each component are shown in Table 1. These results illustrate the effectiveness of the multi-scale analysis based on EMD in various aspects.

2.3. Target Time Point Prediction Based on LSTM

The final gene expression prediction results can be obtained by recombining the prediction results of each component. We made predictions for the two time points of 0 h (Figure 6a,b) and 6 h (Figure 6c,d) in the original data, respectively. Both predictions were excellent either by predicting the initial unknown time point from the known later time point or by predicting the later unknown time point from the known earlier time point. In addition, we calculated the determinant statistics R2 between the predicted results and the true values. It can be seen that the fit between the predicted and true values is satisfactory, and more than 90% of the data points are located in the elliptical confidence band of the fitted curve (Figure 6b,d). The forward time prediction results have an RMSE of 0.2558, MAE of 0.2105, and R2 of 0.4771. The overall prediction error is low and there is a high agreement between the prediction results and the true values. The RMSE value for the reverse time prediction is 0.4002, the MAE value is 0.3169, and the R2 is 0.6145. It also shows low error and excellent agreement between the prediction results and the true values. The prediction results in both directions indicate that this prediction model can obtain well the gene expression results at an unknown destination time. In summary, it can be seen that our proposed Multi-LSTM model can dig deeper into the change pattern of gene expression. Moreover, the prediction results can excellently reflect the expression level of genes at an unknown time.
The model combines signal processing technology and deep learning with biomedicine, and can be widely applied to gene function prediction, disease-related gene discovery, and transcriptomic analysis. In the present study, only the expression data of RNA degradation-related genes were utilized, and the data volume can be expanded in the future to apply Multi-LSTM to a wider range of fields. This model can also be used for time-related disease prediction, for example, degenerative brain diseases and Parkinson’s disease (PD). To use this model for disease prediction in Parkinson’s disease, the genes related to RNA degradation can simply be replaced with genes related to Parkinson’s disease. Of course, this model is not limited to the brain, but can also be used to predict retinal degenerative diseases, such as age-related macular degeneration. In conclusion, this model can be very widely used in disease prediction.

3. Materials and Methods

3.1. Data Acquisition and Pre-Processing

The data used in this study were obtained from brain tissue samples of normal mice. Both tissues and blood cells should be preserved by proper methods as soon as possible after leaving the optimal cellular life state; otherwise, they will all undergo degradation. This study did not involve biological experiments, and mainly explored the mechanism of RNA degradation in mouse brain tissue. RNA from mouse brain tissue was stored at room temperature (RT) prior to extraction for 3 samples at 0 h time points (labeled 0.1 h, 0.2 h, and 0.3 h), 3 samples at 2 h time points (labeled 2.1 h, 2.2 h, and 2.3 h), 3 samples at 4 h time points (labeled 4.1 h, 4.2 h, and 4.3 h), and 3 samples at 6 h time points (labeled 6.1 h, 6.2 h, and 6.3 h). The data obtained from the final sequencing of the samples were used for the analysis of this study. The 12 time points are 12 consecutive points in time. We used the next 11 time points as the training set and the gene expression data from the very first time point as the prediction set. Then, we found the list of genes associated with RNA degradation in existing studies from NCBI. In total, there are 60 genes corresponding to the expression values from time point 1 to time point 12, respectively. The data used in this study were partly obtained from previously published articles [20]. Time-specific clean data of the build sequencing results have been uploaded to the NCBI database. The alignment of genes from time-series transcriptome sequencing results was performed by Hisat2 [21]. The conversion between data was performed using Samtools [22]. The number of reads mapped to each gene for each sample was calculated using FeatureCounts. A catalog of RNA degradation-related genes was downloaded from NCBI, and the corresponding expressions of the genes associated with each sample were extracted. Data pre-processing includes data cleaning, data integration, data transformation, and the reduction in experimental data before model construction. These processes can improve the training speed of subsequent models and the accuracy and credibility of experimental results. In this study, the raw gene expression data were preprocessed with censoring and normalization. The RNA-seq data were normalized by a heat map using ggplots in the R package.

3.2. EMD-Based Decomposition Signal Processing

EMD has been recognized as an effective time–frequency analysis method for dealing with nonlinear, nonstationary signals since it was proposed in 1998 [23]. This method adaptively decomposes the signal into the sum of several IMFs based on the characteristics of the input signal itself, without any prior knowledge. The method has gradually developed its unique advantages in nonstationary signal processing over the years and has important theoretical research value and broad application prospects. EMD is considered as a major breakthrough from the traditional time–frequency analysis methods such as Fourier analysis and wavelet transform, which are based on linear and smooth assumptions. The EMD method is no longer limited by Heisenberg’s inaccuracy principle and can obtain a high-frequency resolution. The method is based on the decomposition of the signal itself, which does not need to define the basis function in advance, and does not need to adopt the prior knowledge of the signal. Theoretically, each IMF must satisfy the following two conditions. First, on the whole signal, the difference between the number of extreme value points and the number of over-zero points is not greater than 1. Secondly, at any point, the mean value of the upper and lower envelopes is 0. Usually, the actual signals are complex signals and do not satisfy the above conditions. Therefore, Huang et al. [23] made the following assumptions. First, any signal is composed of several eigenmode functions. Second, each eigenmode function can be linear or nonlinear. The number of local zeros and extreme points of each eigenmode function is the same, while the upper and lower envelopes are on the time axis of local symmetry. Third, at any time, a signal can contain a number of eigenmodes. If the modal functions overlap each other, the composite signal is formed.
The variable instantaneous frequency and instantaneous amplitude can enhance the efficiency of signal decomposition. At the same time, the EMD method can better preserve the nonsmooth and nonlinear characteristics of the original signal. The eigenmode function obtained by EMD also has the characteristic of intra-wave modulation, which can cover the information of the same component expressed by different Fourier frequencies into one IMF [19]. The resolution of the IMF is adaptive in the sense that the eigenmode functions obtained by EMD have different characteristic time scales. The instantaneous frequency resolution f i of the i th IMF can be expressed as
f i = f i   m a x N
f i   m a x denotes the highest frequency contained in the i th IMF and N is the number of signal samples. From Equation (1), it can be seen that the frequency resolution of each IMF is different. The frequency resolution of the components containing low-frequency components is high, and the frequency resolution of the components containing high-frequency components is low. The frequency resolution corresponding to each IMF is adaptively obtained. At the same time, there is no relationship with the practical resolution. In this regard, it is completely different from the wavelet analysis in the time and frequency resolution. It is not subject to the constraints of the principle of inaccuracy.
The purpose of the EMD algorithm is to decompose the original signal into some series of IMFs, and then obtain the time–frequency relationship of the signal by the Hilbert transform. Taking the terahertz time-domain spectral signal as an example, the basic calculation process of the EMD algorithm is as follows.
Step 1: The terahertz time-domain spectral signal x ( t ) is calculated for all the maxima and minima points. All the curves from the upper envelope u ( t ) are made by connecting all the maxima points with the cubic spline function. The resulting curves from the lower envelope v ( t ) are made by connecting all the minima points with the cubic spline function.
Step 2: The mean value of the upper and lower envelopes is calculated by the formula:
m ( t ) = u ( t ) + v ( t ) 2
The components of the EMD algorithm are defined as:
h ( t ) = x ( t ) m ( t )
Step 3: Determine whether the component h ( t ) satisfies the definition of IMF. If h ( t ) satisfies the definition of IMF, then h ( t ) is the first IMF component filtered out, and it is noted as c 1 ( t ) . If h ( t ) does not satisfy the definition of IMF, h ( t ) is taken as the original data and the above two steps are repeated until the first IMF component is calculated, which is also labeled as c 1 ( t ) .
Step 4: The first IMF component is raised from the original signal to obtain the residual signal, which is calculated as follows.
r 1 ( t ) = h ( t ) c 1 ( t )
Repeat the above three steps with r 1 ( t ) as the new signal to be analyzed, so that the second IMF component c 2 ( t ) is obtained. Remove the second IMF component from r 1 ( t ) to obtain the new residual term.
r 2 ( t ) = r 1 ( t ) c 2 ( t )
The above steps are repeated continuously until the stopping criterion of the EMD algorithm is satisfied. The stopping criterion of the EMD algorithm uses the Cauchy convergence criterion, a test that requires the normalized squared difference between two adjacent extraction operations to be sufficiently small, as defined by:
S D k = t = 0 T |   h k 1   ( t )     h k   ( t ) | 2 t = 0 T   h k 1   ( t ) 2
In the formula: T denotes the signal length, and the decomposition ends when S D k is less than the set threshold value.
The root-mean-square error (RMSE), also known as root-mean-square deviation, is a commonly used measure of the difference between the values. Its formula is:
R M S E = ( i = 1 n ( y i y ^ i ) 2 ) ( n 1 )
The mean absolute error (MAE) is the average of the absolute values of the deviations of all individual observations from the arithmetic mean. The mean absolute error avoids the problem of errors canceling each other out and thus accurately reflects the magnitude of the actual prediction error. Its formula is:
M A E = i = 1 n | y i y ^ i | n
The coefficient of determination (R2) is a statistical measure of how close the regression prediction is to the true data point. The formula for R 2 is:
R 2 = i = 1 n ( y ^ i y ¯ ) 2 i = 1 n ( y i y ¯ ) 2
In the equations of RMSE, MAE, and R2, n is the number of samples and y ^ i is the expression value of the target gene predicted by the i th sample. y i is the true expression value of the target gene in the ith sample and y ¯ is the sample mean.

3.3. Construction of Core Predictors Using LSTM

LSTM is a special recurrent neural network (RNN) proposed to solve the deficiency of RNN in the long-range dependency problem [24]. LSTM achieves the selection of new information addition and the control of information accumulation rate by adding a gating mechanism. The addition of gating control makes up for the shortcomings of the RNN network structure. Compared with the classical RNN network in which there is only a single tanh cyclic body, LSTM adds three gate structures and one memory unit. The flow chart is shown in Figure 1b. Let W be the gate weight and b be the bias. The gate structure can be described as:
g ( x ) = σ   ( W x   + b )
The three gates in the LSTM gate structure are the input gate, the output gate, and the forget gate. The input gate mainly reflects the amount of information stored in the cell state c t at the current moment of the input sequence x t at this time. The output state of the forget gate mainly determines the amount of information from the cell state c t to the output value of the output gate at this moment. Each gate state update of the LSTM is calculated as:
i t = σ   ( W i x t + U i h t 1 + b i )
f t = σ   ( W f x t + U f h t 1 + b f )
o t = σ   ( W o x t + U o h t 1 + b o )
Memory unit update:
c ¯   t = tan h   ( W c x t + U c h t 1 + b c )
Status update:
c t = f t · c t + i t · c ¯   t  
h t = o t   · tan h ( c t )
The formula σ is the sigmoid activation function. W i , W f , W o , b i , b f , and b o are the weight matrices and biases of the input gate, forget gate, and output gate, respectively. W c and b c are the weight matrices and biases of the memory cell after updating, respectively. U i , U f , U o , and U c are the state quantity weight matrices. h t 1 is the previous moment state quantity.

4. Conclusions

In this study, we propose a method based on signal decomposition techniques and deep learning, called Multi-LSTM. It is a prediction framework to explore the expression changes of RNA degradation-related genes in the initial hours of the sample, which is important for studying the effect of RNA degradation on gene expression levels. The first module decomposes the collected gene expression data by the empirical mode decomposition algorithm to obtain a series of sub-modules with different frequencies in order to improve data stability and reduce modeling complexity. The second module uses an LSTM neural network as the core predictor, aiming to deeply explore the temporal nonlinear relationships embedded in the sub-modules. Finally, the prediction results of sub-modules are reconstructed to obtain the final transcriptome time series gene expression level prediction results. The results show that EMD can efficiently reduce the nonlinearity of the original data, which provides a reliable theoretical support to reduce the complexity and improve the robustness of LSTM models. Meanwhile, the decomposition-combination prediction framework can effectively predict gene expression levels at unknown time points by combining the robust temporal nonlinearity analysis capability of LSTM. The combination of LSTM and the effective EMD decomposition algorithm not only improves the accuracy of prediction results, but also has low prediction errors. The results show that both forward and reverse time series can be predicted accurately. The combination of deep learning neural networks with biomedicine will lead to great breakthroughs in gene function prediction, disease-related gene discovery, and transcriptome time series analysis.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/ijms23147532/s1.

Author Contributions

Y.Z., investigation, formal analysis, writing—original draft, data analysis; E.J., investigation, methodology; H.S., investigation, methodology; Z.L. (Zhiyu Liu), investigation, methodology; Y.S., data analysis; M.P., investigation, methodology; J.T., supervision, resources; Q.G., methodology, conceptualization, writing—review and editing; Z.L. (Zuhong Lu), supervision. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Natural Science Foundation of China (61801108) and Natural Science Foundation of Jiangsu Province (BK20201148, BK20211166).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data of this study are freely available at https://www.ncbi.nlm.nih.gov/bioproject/PRJNA826142 (accessed on 12 April 2022). The source code used to support the findings of this study is freely available to the academic community at https://github.com/habbyzy/Multi-LSTM (accessed on 23 April 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ioannidis, J.P.A. Is Molecular Profiling Ready for Use in Clinical Decision Making? Oncologist 2007, 12, 301–311. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Gallego, R.I.; Pai, A.A.; Tung, J.; Gilad, Y. RNA-seq: Impact of RNA Degradation On Transcript Quantification. BMC Biol. 2014, 12, 42. [Google Scholar] [CrossRef] [Green Version]
  3. Fouda, M.A.; Elzefzafy, N.A.; Taha, I.I.; Mohemmed, O.M.; Wahab, A.A.A.; Farahat, I.G. Effect of Seasonal Variation in Ambient Temperature On RNA Quality of Breast Cancer Tissue in a Remote Biobank Setting. Exp. Mol. Pathol. 2020, 112, 104334. [Google Scholar] [CrossRef] [PubMed]
  4. Opitz, L.; Salinas-Riester, G.; Grade, M.; Jung, K.; Jo, P.; Emons, G.; Ghadimi, B.M.; Beißbarth, T.; Gaedcke, J. Impact of RNA Degradation On Gene Expression Profiling. BMC Med. Genom. 2010, 3, 36. [Google Scholar] [CrossRef] [Green Version]
  5. Shen, Y.; Li, R.; Tian, F.; Chen, Z.; Lu, N.; Bai, Y.; Ge, Q.; Lu, Z. Impact of RNA Integrity and Blood Sample Storage Conditions On the Gene Expression Analysis. Oncol. Targets Ther. 2018, 11, 3573–3581. [Google Scholar] [CrossRef] [Green Version]
  6. Jin, C.; Cukier, R.I. Machine Learning Can be Used to Distinguish Protein Families and Generate New Proteins Belonging to those Families. J. Chem. Phys. 2019, 151, 175102. [Google Scholar] [CrossRef]
  7. El-Attar, N.E.; Moustafa, B.M.; Awad, W.A. Deep Learning Model to Detect Diabetes Mellitus Based on DNA Sequence. Intell. Autom. Soft Comput. 2022, 31, 325–338. [Google Scholar] [CrossRef]
  8. Liang, J.; Huang, X.; Li, W.; Hu, Y. Identification and External Validation of the Hub Genes Associated with Cardiorenal Syndrome through Time-Series and Network Analyses. Aging 2022, 14, 1351–1373. [Google Scholar] [CrossRef]
  9. Zhou, J.; Theesfeld, C.L.; Yao, K.; Chen, K.M.; Wong, A.K.; Troyanskaya, O.G. Deep Learning Sequence-Based Ab Initio Prediction of Variant Effects On Expression and Disease Risk. Nat. Genet. 2018, 50, 1171–1179. [Google Scholar] [CrossRef]
  10. Karim, F.; Majumdar, S.; Darabi, H.; Chen, S. LSTM Fully Convolutional Networks for Time Series Classification. IEEE Access. 2018, 6, 1662–1669. [Google Scholar] [CrossRef]
  11. Lakizadeh, A.; Jalili, S.; Marashi, S. PCD-GED: Protein Complex Detection Considering PPI Dynamics Based On Time Series Gene Expression Data. J. Theor. Biol. 2015, 378, 31–38. [Google Scholar] [CrossRef] [PubMed]
  12. Wise, A.; Bar-Joseph, Z. SMARTS: Reconstructing Disease Response Networks From Multiple Individuals Using Time Series Gene Expression Data. Bioinformatics 2015, 31, 1250–1257. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Qian, B.; Xiao, Y.; Zheng, Z.; Zhou, M.; Zhuang, W.; Li, S.; Ma, Q. Dynamic Multi-Scale Convolutional Neural Network for Time Series Classification. IEEE Access. 2020, 8, 109732–109746. [Google Scholar] [CrossRef]
  14. Cheng, X.; Wang, J.; Li, Q.; Liu, T. BiLSTM-5mC: A Bidirectional Long Short-Term Memory-Based Approach for Predicting 5-Methylcytosine Sites in Genome-Wide DNA Promoters. Molecules 2021, 26, 7414. [Google Scholar] [CrossRef] [PubMed]
  15. Trapnell, C.; Williams, B.A.; Pertea, G.; Mortazavi, A.; Kwan, G.; Van Baren, M.J.; Salzberg, S.L.; Wold, B.J.; Pachter, L. Transcript assembly and quantification by RNA-seq reveals unannotated transcripts and isoform switching during cell differentiation. Nat. Biotechnol. 2010, 28, 511–515. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Zhang, Q.; Zheng, X.Y. Walsh Transform and Empirical Mode Decomposition Applied to Reconstruction of Velocity and Displacement from Seismic Acceleration Measurement. Appl. Sci. 2020, 10, 3509. [Google Scholar] [CrossRef]
  17. Yang, C.; Ling, B.W.; Kuang, W.; Gu, J. Decimations of Intrinsic Mode Functions Via Semi-Infinite Programming Based Optimal Adaptive Nonuniform Filter Bank Design Approach. Signal Process. 2019, 159, 53–71. [Google Scholar] [CrossRef]
  18. Zhou, Y.; Li, S. BP Neural Network Modeling with Sensitivity Analysis On Monotonicity Based Spearman Coefficient. Chemometr. Intell. Lab. 2020, 200, 103977. [Google Scholar] [CrossRef]
  19. Waindim, M.; Agostini, L.; Larchêveque, L.; Adler, M.; Gaitonde, D.V. Dynamics of separation bubble dilation and collapse in shock wave/turbulent boundary layer interactions. Shock Waves 2020, 30, 63–75. [Google Scholar] [CrossRef]
  20. Jia, E.; Zhou, Y.; Shi, H.; Pan, M.; Zhao, X.; Ge, Q. Effects of Brain Tissue Section Processing and Storage Time On Gene Expression. Anal. Chim. Acta. 2021, 1142, 38–47. [Google Scholar] [CrossRef]
  21. Kim, D.; Langmead, B.; Salzberg, S.L. HISAT: A Fast Spliced Aligner with Low Memory Requirements. Nat. Methods 2015, 12, 357–360. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Li, H.; Handsaker, B.; Wysoker, A.; Fennell, T.; Ruan, J.; Homer, N.; Marth, G.; Abecasis, G.; Durbin, R. The Sequence Alignment/Map Format and SAMtools. Bioinformatics 2009, 25, 2078–2079. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Huang, N.E.; Shen, Z.; Long, S.R.; Wu, M.C.; Shih, H.H.; Zheng, Q.; Yen, N.-C.; Tung, C.C.; Liu, H.H. The empirical mode decomposition and the Hilbert spectrum for nonlinear and non-stationary time series analysis. Proc. R. Soc. Lond. A 1998, 454, 903–995. [Google Scholar] [CrossRef]
  24. Xie, W.; Luo, J.; Pan, C.; Liu, Y. SG-LSTM-FRAME: A Computational Frame Using Sequence and Geometrical Information Via LSTM to Predict miRNA–Gene Associations. Brief Bioinform. 2021, 22, 2032–2042. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Multi-LSTM flowchart. (a): Flowchart of data processing. Step 1: Data mining. RNA degradation-related genes and the expression corresponding to each time are extracted and normalized. Step 2: Multi-scale analysis. The original signal is decomposed into IMF1-IMF9 by the EMD algorithm, and then the time–frequency relationship of the signal is obtained by the Hilbert transform. Step 3: Deep learning combined prediction. The core predictor is constructed by the LSTM and the prediction results of each subseries are integrated. (b): LSTM flow chart. The three gates in the LSTM gate structure are the input gate, the output gate, and the forget gate. The input gate mainly reflects the amount of information stored in the cell state c t at the current moment of the input x t . The output state of the forget gate mainly determines the amount of information from the cell state c t to the output value of the output gate at this moment. The output gates of the gate control often use the sigmoid function as the activation function, while the activation functions of the input gates and memory cells usually use tanh.
Figure 1. Multi-LSTM flowchart. (a): Flowchart of data processing. Step 1: Data mining. RNA degradation-related genes and the expression corresponding to each time are extracted and normalized. Step 2: Multi-scale analysis. The original signal is decomposed into IMF1-IMF9 by the EMD algorithm, and then the time–frequency relationship of the signal is obtained by the Hilbert transform. Step 3: Deep learning combined prediction. The core predictor is constructed by the LSTM and the prediction results of each subseries are integrated. (b): LSTM flow chart. The three gates in the LSTM gate structure are the input gate, the output gate, and the forget gate. The input gate mainly reflects the amount of information stored in the cell state c t at the current moment of the input x t . The output state of the forget gate mainly determines the amount of information from the cell state c t to the output value of the output gate at this moment. The output gates of the gate control often use the sigmoid function as the activation function, while the activation functions of the input gates and memory cells usually use tanh.
Ijms 23 07532 g001
Figure 2. EMD-based decomposition of the original data. (a) Expression of the genes to be analyzed. The data to be predicted are highlighted with blue dots. (b) The results obtained from the original data based on EMD decomposition for each component of IMF1-IMF9. (c) Spearman correlation of each component of IMF1-IMF9 after EMD decomposition with the original data.
Figure 2. EMD-based decomposition of the original data. (a) Expression of the genes to be analyzed. The data to be predicted are highlighted with blue dots. (b) The results obtained from the original data based on EMD decomposition for each component of IMF1-IMF9. (c) Spearman correlation of each component of IMF1-IMF9 after EMD decomposition with the original data.
Ijms 23 07532 g002
Figure 3. The prediction results of each component using LSTM. (ai) correspond to the results of components IMF1-IMF9, respectively. The expression values of a total of 60 degradation-related genes were decomposed into 9 IMFs by EMD, and the prediction of gene expression was performed according to these 9 modules separately. Specifically, the white dots represent the true values and the blue dots represent the predicted values. From (ai), it can be seen that the nonsmooth signal is becoming smoother, and the line of blue dots (predicted values) is also approaching the line of white dots (true values).
Figure 3. The prediction results of each component using LSTM. (ai) correspond to the results of components IMF1-IMF9, respectively. The expression values of a total of 60 degradation-related genes were decomposed into 9 IMFs by EMD, and the prediction of gene expression was performed according to these 9 modules separately. Specifically, the white dots represent the true values and the blue dots represent the predicted values. From (ai), it can be seen that the nonsmooth signal is becoming smoother, and the line of blue dots (predicted values) is also approaching the line of white dots (true values).
Ijms 23 07532 g003
Figure 4. Proximity between the predicted values and the original data. (ai) correspond to the results of components IMF1-IMF9, respectively. The results show that there is no difference between the predicted and true values of each component, and all p-values > 0.05.
Figure 4. Proximity between the predicted values and the original data. (ai) correspond to the results of components IMF1-IMF9, respectively. The results show that there is no difference between the predicted and true values of each component, and all p-values > 0.05.
Ijms 23 07532 g004
Figure 5. R2 between predicted and true values of each component. (ai) correspond to the results of components IMF1-IMF9, respectively.
Figure 5. R2 between predicted and true values of each component. (ai) correspond to the results of components IMF1-IMF9, respectively.
Ijms 23 07532 g005
Figure 6. Overall prediction results. (a,b): RMSE value of 0.2558, MAE value of 0.2105, and R2 of 0.4771 at 0 h. The overall prediction error is low and there is a high agreement between the prediction results and the true values. (c,d): RMSE value of 0.4002, MAE value of 0.3169, and R2 of 0.6145 for the reverse time prediction of 6 h. It also shows low error and excellent agreement between the prediction results and the true values.
Figure 6. Overall prediction results. (a,b): RMSE value of 0.2558, MAE value of 0.2105, and R2 of 0.4771 at 0 h. The overall prediction error is low and there is a high agreement between the prediction results and the true values. (c,d): RMSE value of 0.4002, MAE value of 0.3169, and R2 of 0.6145 for the reverse time prediction of 6 h. It also shows low error and excellent agreement between the prediction results and the true values.
Ijms 23 07532 g006
Table 1. RMSE, MAE, and R2 values for each IMF component.
Table 1. RMSE, MAE, and R2 values for each IMF component.
IMF1IMF2IMF3IMF4IMF5IMF6IMF7IMF8IMF9
RMSE0.13320.08710.09590.08060.08170.06880.02020.00810.0172
MAE0.10590.06610.08120.06470.06390.04920.01600.00600.0151
R20.90160.83320.86190.87650.94420.95500.96170.97770.9914
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhou, Y.; Jia, E.; Shi, H.; Liu, Z.; Sheng, Y.; Pan, M.; Tu, J.; Ge, Q.; Lu, Z. Prediction of Time-Series Transcriptomic Gene Expression Based on Long Short-Term Memory with Empirical Mode Decomposition. Int. J. Mol. Sci. 2022, 23, 7532. https://doi.org/10.3390/ijms23147532

AMA Style

Zhou Y, Jia E, Shi H, Liu Z, Sheng Y, Pan M, Tu J, Ge Q, Lu Z. Prediction of Time-Series Transcriptomic Gene Expression Based on Long Short-Term Memory with Empirical Mode Decomposition. International Journal of Molecular Sciences. 2022; 23(14):7532. https://doi.org/10.3390/ijms23147532

Chicago/Turabian Style

Zhou, Ying, Erteng Jia, Huajuan Shi, Zhiyu Liu, Yuqi Sheng, Min Pan, Jing Tu, Qinyu Ge, and Zuhong Lu. 2022. "Prediction of Time-Series Transcriptomic Gene Expression Based on Long Short-Term Memory with Empirical Mode Decomposition" International Journal of Molecular Sciences 23, no. 14: 7532. https://doi.org/10.3390/ijms23147532

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop