Next Article in Journal
Research on Braking Efficiency of Master-Slave Electro-Hydraulic Hybrid Electric Vehicle
Previous Article in Journal
Investigation of Near-Field Source Localization Using Uniform Rectangular Array
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Effective Deep Learning-Based Architecture for Prediction of N7-Methylguanosine Sites in Health Systems

1
Department of Computer Science, Abdul Wali Khan University Mardan, Mardan 23200, Pakistan
2
Department of Electronics and Information Engineering, Jeonbuk National University, Jeonju 54896, Korea
3
Advanced Electronics and Information Research Center, Chonbuk National University, Jeonju 54896, Korea
*
Authors to whom correspondence should be addressed.
Electronics 2022, 11(12), 1917; https://doi.org/10.3390/electronics11121917
Submission received: 22 April 2022 / Revised: 12 June 2022 / Accepted: 14 June 2022 / Published: 20 June 2022
(This article belongs to the Section Computer Science & Engineering)

Abstract

:
N7-methylguanosine (m7G) is one of the most important epigenetic modifications found in rRNA, mRNA, and tRNA, and performs a promising role in gene expression regulation. Owing to its significance, well-equipped traditional laboratory-based techniques have been performed for the identification of N7-methylguanosine (m7G). Consequently, these approaches were found to be time-consuming and cost-ineffective. To move on from these traditional approaches to predict N7-methylguanosine sites with high precision, the concept of artificial intelligence has been adopted. In this study, an intelligent computational model called N7-methylguanosine-Long short-term memory (m7G-LSTM) is introduced for the prediction of N7-methylguanosine sites. One-hot encoding and word2vec feature schemes are used to express the biological sequences while the LSTM and CNN algorithms have been employed for classification. The proposed “m7G-LSTM” model obtained an accuracy value of 95.95%, a specificity value of 95.94%, a sensitivity value of 95.97%, and Matthew’s correlation coefficient (MCC) value of 0.919. The proposed predictive m7G-LSTM model has significantly achieved better outcomes than previous models in terms of all evaluation parameters. The proposed m7G-LSTM computational system aims to support the drug industry and help researchers in the fields of bioinformatics to enhance innovation for the prediction of the behavior of N7-methylguanosine sites.

1. Introduction

To date, about 150 various types of RNA alteration/modification have been recognized. These changes in RNA perform important functions in regulating the expression of genes at different levels. For example, it was previously confirmed that changes in RNA can affect RNA transport, processing, mRNA translation, and stability [1,2]. The most abundant RNA modifications are the N7-methylguanosine (m7G), which happens in the 5’ cap position of mRNA molecules and transfer RNA (tRNA) loop eukaryotic S ribosomal RNA (rRNA); these modifications are preserved amid the three different kingdoms. This modification plays a precarious role in the regulation of RNA function, post-transcription modifications, and metabolism [3]. Within the context of aging, Zago et al. have discussed the emerging importance of microRNAs as biomarkers for Parkinson’s disease [4]. Unfortunately, data on the functional mechanisms are very limited. Recently, it has been revealed that m7G sites can be efficiently identified by modern sequencing techniques [5,6,7]. Using deep sequencing technology, Marchand et al. investigated AlkAniline-Seq for identifying m7G in RNA at single-nucleotide resolution in yeast, human, and bacterial mitochondrial and cytoplasmic rRNAs and tRNAs [5]. Furthermore, differently modifying internal m7G sites to certain basic sites, Zhang et al. effectively established a MeRIP-seq approach to predict m7G sites at single-base resolution. The single-base resolution approach used in m7G-seq data revealed the profile of m7G in human mRNA and tRNA, to enhance the knowledge of m7G distribution in human cells [6].
However, the transcriptome-wide dissemination and vibrant regulation of m7G within internal mRNA areas are still unknown. According to Zhao et al., the internal mRNA m7G methyltransferase METTL1, and not WDR4, is a key responder to post-ischemic insults, resulting in a global reduction in m7G methylation inside mRNA [8]. In addition, Liu et al. introduced m7GPredictor for predicting internal m7G modification sites using sequence properties [9]. In this model, the authors used various numerical descriptor methods and a random forest was used for the selection of optimal feature sets. Likewise, Bi et al., developed a computational model for the identification of m7G sites [10]. In this model, they have used different types of sequence encoding schemes in combination with the XGBoost algorithm. Further, Shoombuatong et al. proposed a new predictor known as THRONE for discrimination of human RNA N7-methylguanosine sites [11]. The THRONE was designed in three steps using an ensemble learning predictor. Likewise, the m7G-DPP web predictor was introduced by Zou and Yin by using physicochemical properties of RNA for the prediction of m7G sites [12]. Here, Pearson correlation coefficient, dynamic time warping, and distance correlation were utilized for extracting numerical features. Next, the LASSO algorithm was employed to select highly discriminative features [12]. Likewise, Zhang et al. introduced a predictor, namely BERT-m7G, by utilizing in staking ensemble approach for the identification of RNA m7G sites [13]. In this model, a BERT-based multilingual model was utilized to represent the information RNA sequences. Similarly, to specifically detect the internal mRNA m7G mutation, Malbec et al. developed the m7G individual nucleotide-resolution cross-linking and immunoprecipitation with sequencing (miCLIP-seq) approach [7]. Finally, this group of researchers determined that m7G modifications are enriched in AG-rich contexts, which are highly preserved in different mouse tissues and human cell lines. However, the advanced sequencing techniques revealed significant findings in this area, although these methods are still costly for transcriptome-wide detection. In this context, computational analysis m7G site predictors have been introduced, namely m7GFinder [14], iRNA-m7G [15], and m7G-IFL [16]. In these predictors, Yang et al. introduced a computational m7GFinder tool that can predict m7G sites in H. sapiens RNA using a sequence-based approach. The optimal feature subset was determined using mRMR, F-score, and Relief; and a support vector machine (SVM) was used as a classifier. Similarly, in sequential, the iRNA-m7G model was performed by Chen et al. for the identification of m7-methylguanosine sites by fusing multiple feature spaces. In this model, sequential- and structural-based features were integrated in order to form a hybrid space. Three types of features were combined using the feature fusion method, including secondary structure components, pseudo-nucleotide composition, and nucleotide property and frequency, to extract important RNA sequence features. Experiments have shown that the feature fusion technique outperforms the use of a single type of feature in detecting m7G sites [15]. Similarly, Ning et al. presented a predictor for the identification of m7G, namely m7G-DLSTM based on an LSTM model and natural language processing (NLP), nucleotide chemical property, and binary code feature extraction methods [17]. Most recently, an m7G-IFL computational model for identifying m7G sites was developed by Dai et al. [18]. This model uses an RNA sequence-encoding iterative feature representation approach to discover probabilistic distribution information from various sequential models and improve feature representation skills in a supervised iterative manner. The m7G-IFL predictor used various feature extraction techniques such as ring-function-hydrogen properties (RFH), physical-chemical-properties (PCP), and binary k-mer frequency (BKF). Then, extreme gradient boosting (XGBoost) was applied as a classifier. Furthermore, it was discovered that the proposed iterative feature method can improve feature representation capability during the iterative phase through feature analysis [16].
Furthermore, enhanced efficiency of existing computational models is still needed in the detection process. Thus, there is a dire need for the development of novel computational methods for the accurate, fast, and precise detection of m7G modification. In our recent study, we tried to focus on deep learning-based prediction methodologies to develop an accurate computational system called “m7G-LSTM” to predict N7-methylguanosine sites, which could directly determine m7G sites based on sequence information. The proposed m7G-LSTM system contains two stages i.e., distributed attributes representation and long short-term memory (LSTM) model. In the attributes representation stage, the NLP-based approach word2vec is applied to fragment the RNA instance into words (3-mers). Likewise, in the second stage, the N7-methylguanosine site is identified by using the LSTM model. The proposed prediction model m7G-LSTM has shown better performance and obtained promising outcomes.

2. Methods and Materials

2.1. Benchmark Dataset

Here, we select and download a benchmark dataset from Chen et al. [15,16] to train the proposed computational system. The benchmark dataset consists of 741 positive sequences, which are m7G sites, and 741 negative sequences, which are non-m7G sites; both have the same length (41 nucleotides). The benchmark dataset is mathematically expressed in Equation (1).
S = S + S
The benchmark dataset S consists of m7G sites and non-m7G sites, S+ with positive m7G sites sequences, and S with negative non-m7G sites sequences. To examine the performance of the proposed model, cross-validation can be used. The dataset is split into three sections: 20% for testing, 10% for validation, and 70% for training.

2.2. Encoding Scheme

The one-hot encoding approach is a simple but useful feature extraction technique, frequently used in deep learning, but shows effective performance in bioinformatics [19] and computer science [20]. It is employed to illustrate the nucleotide acid composition along the RNA/DNA sequence. In previous studies [21,22,23,24,25], one-hot encoding was employed. In this encoding technique ‘A’, ‘C’, ‘G’, and ‘U’ are represented by binary vectors of (1, 0, 0, 0), (0, 1, 0, 0), (0, 0, 1, 0), and (0, 0, 0, 1), respectively. As a result, an n nucleotide RNA/DNA sequence is encoded as a 4×n dimensional binary vector, which is used as input to the CNN and LSTM models in this study. The vector has a length of n = 41 nucleotides. Figure 1 demonstrates the graphical representation of the one-hot encoding scheme.

2.3. Distributed Attributes Formulation

The distributed attributes formulation scheme reduces the classification error of the computational model by obtaining noiseless data. Genetic data are usually expressed as biological sequences; hence, it may be thought of as a language through which information moves between cells. Natural language processing (NLP) has been used for a variety of biological problems in this area, such as EP2vec [26], alternative splicing site [27], G2Vec [28], and iN6-Methyl (5-step) [29]. We approach this sequence analysis problem from a new angle, manipulated by NLP. Indeed, there are several effective deep learning applications in the NLP, i.e., word2vec, which embeds words into a vector space. The paragraph vector is built on word2vec, and it embeds whole phrases into vectors that encode their semantic content. In this regard, treating the sequence of RNA/DNA as a sentence rather than an image is more natural since DNA sequences are just one-dimensional data, whereas images are frequently two-dimensional data. Consequently, we consider a DNA/RNA sequence to be a sentence made up of k-mers (or words) [26]. Here, an NLP-based method, i.e., word2vec, is applied to obtain decipherable demonstrations for RNA sequences. For discontinuity, the RNA sequences are first fragmented into multiple words represented by overlapping k-mers. Here, the value of k = 3 indicates a 3-mer. Commonly, genomes are collected from the Genbank databank by using the following link: http://hgdownload.soe.ucsc.edu. The genome is split into distant 21 chromosomes (C1, C2, C3, C4, C5, …, C20, and C21). Additionally, the chromosome is fragmented with sentences of 100 nt residues. Lastly, the words are created by cutting each sentence into overlapping 3-mers. The word2vec model is trained using the continuous bag-of-words (CBOW) technique. The current word w(t) is predicted using the context words around it in a predetermined frame in the CBOW technique. Table 1 shows the training parameters of the word2vec model. Finally, each 3-mer word is expressed by a 100-dimensional vector, and each sequence of length L is represented by an array of shapes (L − 2) × 100.

2.4. Convolutional Neural Networks (CNN)

A CNN is a deep learning algorithm frequently used in image processing, natural language processing, and bioinformatics studies [30,31,32,33,34,35,36,37]. In image data, CNN works with two-dimensional; however, CNN can also be employed with three-dimensional and one-dimensional data. In this regard, the 1D (1-dimensional) CNN model in the field of bioinformatics is effectively applied [24,38,39,40,41]. A CNN comprises one input layer, multiple hidden layers such as pooling layers, ReLU (activation function) layers, convolutional layers, normalization layers, fully connected layers, and an output layer. In this study, various optimal hyper-parameters, such as the size of the masks [3, 5, 7, 9, 11, 13, and 15], the number of masks [4, 6, 8, 10, 12, 14, and 16], and convolution layers [1, 2, and 3] are used for training CNN model. The dropout probability range was [0.2, 0.25, 0.3, 0.35, 0.4, 0.5, 0.6, 0.7, and 0.75]. The selections of hyper-parameters are performed on the best success rates in terms of all performance metrics to discriminate N7-methylguanosine sites. The convolution layer, ReLU layer, and sigmoid function are mathematical as follows:
Conv 1 D ( R ) j f = ReLU ( s = 0 S 1 n = 0 N 1 W s n f R j + s , n )
In Equation (2), R, f, and j stand for the input, filter index, and output index position, respectively. N shows the number of input channels, and S denotes the size of the window.
Dense layer with dropout: The scalar output score of the dense layer is transformed from the feature vector z.
f = w d + 1 + k = 1 d w k z k
f = w d + 1 + k = 1 d m k w k z k
In Equations (3) and (4), wd+1 represents the term of additive bias and the previous layer zk weight is wk. The rectified linear function is denoted by ReLU and mathematically stated in Equation (5).
ReLU ( z ) = max ( 0 , z )
As its output is scaled to the [0, 1], the sigmoid function is responsible for predicting whether a given sequence is an m7G site or not. Equation (6) expresses the sigmoid function mathematically.
Sigmoid ( z ) = 1 1 + e z

2.5. Long Short-Term Memory Layer (LSTM)

The recurrent neural network (RNN) is a type of deep learning that can learn only from sequential data such as time-series data and textual data [42]. However, it has the issue of gradient vanishing, and thus the parameters are not updated during the backpropagation [36,43,44,45,46]. Therefore, LSTM is a type of RNN that may store information regarding long-distance data dependence and added gating function by addressing the issue of RNN gradient [30,47,48]. LSTM gating mechanisms enable the network to effectively decide to keep it remember or ignore it. Furthermore, speech recognition and language translation also have great contributions [49,50,51]. Figure 2 illustrates our proposed model, which is composed of an input layer, two LSTM layers, and followed by a dense layer.
The first LSTM layer has an output channel of size 32, which is fed into the second LSTM layer, where the second layer has 64 output channels. Moreover, the dropout rate of 35% is applied to the input connection within the LSTM layers. The outcome of the second LSTM layer is flattened and passes to the dense layer. The dense layer is a fully-connected layer with x output channels, and it is followed by a sigmoid activation function. Finally, the sigmoid function generates the outcomes.
The proposed model has been trained as follows. Let x i be a vector demonstrating the input RNA sequence (Equation (7)). The LSTM computes z i for x i (Equation (9)). Sigmoid (Equation (10)) changes z i to a vector of values between 0 and 1. The loss is the binary cross-entropy of the prediction (Equation (10)). It is used for updating the hidden neurons at the hidden layer utilizing the Adam optimization algorithm, with 0.0005 being set as the learning rate.
x i = R N A   sequence   where   x i     { A ,   C ,   G ,   U }
y i = { 0   if   x i = non - m 7 G   sites 1   if   x i = m 7 G   sites
z i = l s t m ( x i )
s i g m o i d ( z i ) = 1 1 + e y

2.6. Evaluation Parameters

In the literature [15,52,53,54,55,56,57,58,59,60,61,62,63], the following four equations were employed to measure the prediction performance of the computational method: specificity (sp), sensitivity (sen), accuracy (acc), MCC (Matthew’s correlation coefficient), and auROC. In the below equations ‘FP’ is a false positive, ‘TP’ is a true positive, ‘TN’ is a true negative, and ‘FN’ is a false negative.
{ S p = T N TN + FP   × 100 S e n = T P TP + FN × 100 A c c = T N + T P FN + TP + TN + FP   × 100 M C C = TP × TN FP × FN ( TP + FP ) ( TP + FN ) ( TN + FP ) ( TN + FN )   × 100
Accuracy: assesses the precision of a computational algorithm for distinguishing m7G sites and non-m7G sites. Sensitivity and specificity are the true positive (TP) and true negative (TN) rates of a test. MCC reveals the correlation between target classes in the case of the imbalanced dataset; here, the ratio of both classes is the same. The area under the ROC curve (auROC) is another measurement metric that shows the predicted outcomes of the model. The auROC indicates the quality of the model. In the above equation, FN and FP denote false negative and false positive, respectively.

3. Results and Discussion

An intelligent computational method, namely m7G-LSTM, is designed based on a natural language processing approach, i.e., word2vec, in combination with the deep learning algorithm LSTM. The efficiency is reported on the basis of various measuring metrics, which are mentioned above. The proposed m7G-LSTM model has an accuracy of 95.95%, specificity of 95.97%, the sensitivity of 95.94%, MCC of 0.919, and auROC of 0.980 on the LSTM model, whereas the CNN model achieved 94.94% accuracy, 93.28% specificity, 96.62% sensitivity, 0.899 MCC, and 0.979 auROC. Table 2 shows the detailed projected outcomes of the proposed predictor on LSTM and CNN.
Figure 3 and Figure 4 show the auROC as well as a graphical illustration of the confusion matrix.
We compared our proposed m7G-LSTM model to the state-of-the-art models, such as m7GFinder, m7G-IFL, and iRNA-m7G, to calculate its predictive performance. In this study, we develop two various deep learning-based approaches i.e., LSTM and CNN approach to present the proposed m7G-LSTM model. As a result, we compared the highest-performing model of the m7G-LSTM to the most recent m7G-IFL model. To achieve a valid comparison, we execute and evaluate the proposed model on the same benchmark dataset as existing models. Table 3 summarizes the prediction performance. Our computational model, as can be shown, outperforms the other three models, with an accuracy of 95.95%, the sensitivity of 95.94%, specificity of 95.97%, and MCC of 0.919, respectively. Our proposed m7G-LSTM computational model outperformed the existing latest m7G-IFL predictor by 3.54% in specificity, 3.37% in sensitivity, 3.45% in accuracy, and 0.069 in MCC. We observe that our predictor outperforms other models, with an AUROC of 0.980. Our m7G-LSTM model improves upon state-of-the-art prediction models for predicting m7G site modification, based on the results.
The graphical depiction of the execution outcomes is demonstrated in Figure 5 in which the m7G-LSTM method achieved remarkable results compared to the current prediction models. This shows the significance of our proposed model.

4. Conclusions

The proposed m7G-LSTM method is a reliable and novel deep learning-based prediction model for m7G sites. The proposed model utilized the distributed feature representations that are exploited by the LSTM model. Two-feature encoding schemes were used, i.e., word2vec and one-hot encoding. The input of RNA sequence is divided into 3-mers or words in feature representation, and each word is mapped to its corresponding feature representation using the NLP method, i.e., word2vec. The one-hot encoding converts categorical data to binary data that can be processed by computational models efficiently. Then, the LSTM and CNN models were applied to identify N7-methylguanosine sites, but the LSTM model produced better performance than the CNN model. In terms of all performance measures for discriminating N7-methylguanosine sites, the m7G-LSTM model highly outperforms state-of-the-art models, according to the prediction results. The predicted outcome demonstrates that the proposed m7G-LSTM computational system is efficient and reliable and that it might be useful in drug-related applications and academics.

Author Contributions

Conceptualization, M.T.; methodology, M.T.; software, M.H. and R.K.; validation, M.T., M.H., R.K. and K.T.C.; resources, K.T.C.; writing—original draft, M.T. and R.K.; writing—review & editing, M.T., M.H. and K.T.C.; visualization, M.T. and R.K.; supervision, M.H. and K.T.C.; project administration, K.T.C.; funding acquisition, K.T.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korean government (MSIT) (No. 2020R1A2C2005612).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Roost, C.; Lynch, S.R.; Batista, P.J.; Qu, K.; Chang, H.Y.; Kool, E.T. Structure and thermodynamics of N6-methyladenosine in RNA: A spring-loaded base modification. J. Am. Chem. Soc. 2015, 137, 2107–2115. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Wang, X.; Lu, Z.; Gomez, A.; Hon, G.C.; Yue, Y.; Han, D.; Fu, Y.; Parisien, M.; Dai, Q.; Jia, G. N 6-methyladenosine-dependent regulation of messenger RNA stability. Nature 2014, 505, 117–120. [Google Scholar] [CrossRef] [PubMed]
  3. Cowling, V.H. Regulation of mRNA cap methylation. Biochem. J. 2010, 425, 295–302. [Google Scholar] [CrossRef] [Green Version]
  4. Zago, E.; Molin, A.D.; Dimitri, G.M.; Xumerle, L.; Pirazzini, C.; Bacalini, M.G.; Maturo, M.G.; Azevedo, T.; Spasov, S.; Gómez-Garre, P. Early downregulation of hsa-miR-144-3p in serum from drug-naïve Parkinson’s disease patients. Sci. Rep. 2022, 12, 1330. [Google Scholar] [CrossRef]
  5. Marchand, V.; Ayadi, L.; Ernst, F.G.; Hertler, J.; Bourguignon-Igel, V.; Galvanin, A.; Kotter, A.; Helm, M.; Lafontaine, D.L.; Motorin, Y. AlkAniline-Seq: Profiling of m7G and m3C RNA Modifications at Single Nucleotide Resolution. Angew. Chem. Int. Ed. 2018, 57, 16785–16790. [Google Scholar] [CrossRef]
  6. Zhang, L.-S.; Liu, C.; Ma, H.; Dai, Q.; Sun, H.-L.; Luo, G.; Zhang, Z.; Zhang, L.; Hu, L.; Dong, X. Transcriptome-wide mapping of internal N7-methylguanosine methylome in mammalian mRNA. Mol. Cell 2019, 74, 1304–1316.e1308. [Google Scholar] [CrossRef]
  7. Malbec, L.; Zhang, T.; Chen, Y.-S.; Zhang, Y.; Sun, B.-F.; Shi, B.-Y.; Zhao, Y.-L.; Yang, Y.; Yang, Y.-G. Dynamic methylome of internal mRNA N 7-methylguanosine and its regulatory role in translation. Cell Res. 2019, 29, 927–941. [Google Scholar] [CrossRef]
  8. Zhao, Y.; Kong, L.; Pei, Z.; Li, F.; Li, C.; Sun, X.; Shi, B.; Ge, J. m7G methyltransferase METTL1 promotes post-ischemic angiogenesis via promoting VEGFA mRNA translation. Front. Cell Dev. Biol. 2021, 9, 1376. [Google Scholar] [CrossRef]
  9. Liu, X.; Liu, Z.; Mao, X.; Li, Q. m7GPredictor: An improved machine learning-based model for predicting internal m7G modifications using sequence properties. Anal. Biochem. 2020, 609, 113905. [Google Scholar] [CrossRef]
  10. Bi, Y.; Xiang, D.; Ge, Z.; Li, F.; Jia, C.; Song, J. An interpretable prediction model for identifying N7-methylguanosine sites based on XGBoost and SHAP. Mol. Ther.-Nucleic Acids 2020, 22, 362–372. [Google Scholar] [CrossRef]
  11. Shoombuatong, W.; Basith, S.; Pitti, T.; Lee, G.; Manavalan, B. THRONE: A New Approach for Accurate Prediction of Human RNA N7-Methylguanosine Sites. J. Mol. Biol. 2022, 434, 167549. [Google Scholar] [CrossRef] [PubMed]
  12. Zou, H.; Yin, Z. m7G-DPP: Identifying N7-methylguanosine sites based on dinucleotide physicochemical properties of RNA. Biophys. Chem. 2021, 279, 106697. [Google Scholar] [CrossRef] [PubMed]
  13. Zhang, L.; Qin, X.; Liu, M.; Liu, G.; Ren, Y. BERT-m7G: A Transformer Architecture Based on BERT and Stacking Ensemble to Identify RNA N7-Methylguanosine Sites from Sequence Information. Comput. Math. Methods Med. 2021, 2021, 7764764. [Google Scholar] [CrossRef] [PubMed]
  14. Yang, Y.-H.; Ma, C.; Wang, J.-S.; Yang, H.; Ding, H.; Han, S.-G.; Li, Y.-W. Prediction of N7-methylguanosine sites in human RNA based on optimal sequence features. Genomics 2020, 112, 4342–4347. [Google Scholar] [CrossRef] [PubMed]
  15. Chen, W.; Feng, P.; Song, X.; Lv, H.; Lin, H. iRNA-m7G: Identifying N7-methylguanosine sites by fusing multiple features. Mol. Ther.-Nucleic Acids 2019, 18, 269–274. [Google Scholar] [CrossRef] [Green Version]
  16. Dai, C.; Feng, P.; Cui, L.; Su, R.; Chen, W.; Wei, L. Iterative feature representation algorithm to improve the predictive performance of N7-methylguanosine sites. Brief. Bioinform. 2020, 22, bbaa278. [Google Scholar] [CrossRef]
  17. Ning, Q.; Sheng, M. m7G-DLSTM: Intergrating directional Double-LSTM and fully connected network for RNA N7-methlguanosine sites prediction in human. Chemom. Intell. Lab. Syst. 2021, 217, 104398. [Google Scholar] [CrossRef]
  18. Wei, L.; Su, R.; Luan, S.; Liao, Z.; Manavalan, B.; Zou, Q.; Shi, X. Iterative feature representations improve N4-methylcytosine site prediction. Bioinformatics 2019, 35, 4930–4937. [Google Scholar] [CrossRef]
  19. Li, F.; Li, C.; Wang, M.; Webb, G.I.; Zhang, Y.; Whisstock, J.C.; Song, J. GlycoMine: A machine learning-based approach for predicting N-, C-and O-linked glycosylation in the human proteome. Bioinformatics 2015, 31, 1411–1419. [Google Scholar] [CrossRef]
  20. Uriarte-Arcia, A.V.; López-Yáñez, I.; Yáñez-Márquez, C. One-hot vector hybrid associative classifier for medical data classification. PLoS ONE 2014, 9, e95715. [Google Scholar] [CrossRef] [Green Version]
  21. Khanal, J.; Nazari, I.; Tayara, H.; Chong, K.T. 4mCCNN: Identification of N4-methylcytosine sites in prokaryotes using convolutional neural network. IEEE Access 2019, 7, 145455–145461. [Google Scholar] [CrossRef]
  22. Lv, Z.; Ding, H.; Wang, L.; Zou, Q. A Convolutional Neural Network Using Dinucleotide One-hot Encoder for identifying DNA N6-Methyladenine Sites in the Rice Genome. Neurocomputing 2021, 422, 214–221. [Google Scholar] [CrossRef]
  23. Zeng, F.; Fang, G.; Yao, L. A deep neural network for identifying DNA N4-methylcytosine sites. Front. Genet. 2020, 11, 209. [Google Scholar] [CrossRef] [Green Version]
  24. Tahir, M.; Tayara, H.; Hayat, M.; Chong, K.T. kDeepBind: Prediction of RNA-Proteins binding sites using convolution neural network and k-gram features. Chemom. Intell. Lab. Syst. 2021, 208, 104217. [Google Scholar] [CrossRef]
  25. Shao, Y.; Chou, K.-C. pLoc_Deep-mVirus: A CNN Model for Predicting Subcellular Localization of Virus Proteins by Deep Learning. Nat. Sci. 2020, 12, 388–399. [Google Scholar] [CrossRef]
  26. Zeng, W.; Wu, M.; Jiang, R. Prediction of enhancer-promoter interactions via natural language processing. BMC Genom. 2018, 19, 13–22. [Google Scholar] [CrossRef] [Green Version]
  27. Oubounyt, M.; Louadi, Z.; Tayara, H.; Chong, K.T. Deep learning models based on distributed feature representations for alternative splicing prediction. IEEE Access 2018, 6, 58826–58834. [Google Scholar] [CrossRef]
  28. Choi, J.; Oh, I.; Seo, S.; Ahn, J. G2Vec: Distributed gene representations for identification of cancer prognostic genes. Sci. Rep. 2018, 8, 13729. [Google Scholar] [CrossRef]
  29. Nazari, I.; Tahir, M.; Tayara, H.; Chong, K.T. iN6-Methyl (5-step): Identifying RNA N6-methyladenosine sites using deep learning mode via Chou’s 5-step rules and Chou’s general PseKNC. Chemom. Intell. Lab. Syst. 2019, 193, 103811. [Google Scholar] [CrossRef]
  30. Tan, K.K.; Le, N.Q.K.; Yeh, H.-Y.; Chua, M.C.H. Ensemble of deep recurrent neural networks for identifying enhancers via dinucleotide physicochemical properties. Cells 2019, 8, 767. [Google Scholar] [CrossRef] [Green Version]
  31. Tayara, H.; Soo, K.G.; Chong, K.T. Vehicle detection and counting in high-resolution aerial images using convolutional regression neural network. IEEE Access 2017, 6, 2220–2230. [Google Scholar] [CrossRef]
  32. Bai, S.; Kolter, J.Z.; Koltun, V. An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv 2018, arXiv:1803.01271. [Google Scholar]
  33. Tahir, M.; Tayara, H.; Chong, K.T. iRNA-PseKNC (2methyl): Identify RNA 2′-O-methylation sites by convolution neural network and Chou’s pseudo components. J. Theor. Biol. 2019, 465, 1–6. [Google Scholar] [CrossRef] [PubMed]
  34. Tahir, M.; Tayara, H.; Chong, K.T. iPseU-CNN: Identifying RNA pseudouridine sites using convolutional neural networks. Mol. Ther.-Nucleic Acids 2019, 16, 463–470. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Tayara, H.; Tahir, M.; Chong, K.T. Identification of prokaryotic promoters and their strength by integrating heterogeneous features. Genomics 2020, 112, 1396–1403. [Google Scholar] [CrossRef]
  36. Saini, V.K.; Kumar, R.; Mathur, A.; Saxena, A. Short term forecasting based on hourly wind speed data using deep learning algorithms. In Proceedings of the 2020 3rd International Conference on Emerging Technologies in Computer Engineering: Machine Learning and Internet of Things (ICETCE), Jaipur, India, 7–8 February 2020; pp. 1–6. [Google Scholar]
  37. Sharma, A.K.; Saxena, A.; Soni, B.P.; Gupta, V. Voltage stability assessment using artificial neural network. In Proceedings of the 2018 IEEMA Engineer Infinite Conference (eTechNxT), New Delhi, India, 13–14 March 2018; pp. 1–5. [Google Scholar]
  38. Tahir, M.; Hayat, M.; Chong, K.T. A convolution neural network-based computational model to identify the occurrence sites of various RNA modifications by fusing varied features. Chemom. Intell. Lab. Syst. 2021, 211, 104233. [Google Scholar] [CrossRef]
  39. Tahir, M.; Hayat, M.; Khan, S.; Chong, K.T. Prediction of piwi-interacting RNAs and their functions via Convolutional Neural Network. IEEE Access 2021, 9, 54233–54240. [Google Scholar] [CrossRef]
  40. Tahir, M.; Hayat, M.; Ullah, I.; Chong, K.T. A deep learning-based computational approach for discrimination of dna n6-methyladenosine sites by fusing heterogeneous features. Chemom. Intell. Lab. Syst. 2020, 206, 104151. [Google Scholar] [CrossRef]
  41. Tahir, M.; Hayat, M.; Chong, K.T. Prediction of n6-methyladenosine sites using convolution neural network model based on distributed feature representations. Neural Netw. 2020, 129, 385–391. [Google Scholar] [CrossRef]
  42. Wang, L.; Zhong, X.; Wang, S.; Zhang, H.; Liu, Y. A novel end-to-end method to predict RNA secondary structure profile based on bidirectional LSTM and residual neural network. BMC Bioinform. 2021, 22, 169. [Google Scholar] [CrossRef]
  43. Zhang, J.; Chen, Q.; Liu, B. DeepDRBP-2L: A new genome annotation predictor for identifying DNA binding proteins and RNA binding proteins using convolutional neural network and long short-term memory. IEEE/ACM Trans. Comput. Biol. Bioinform. 2019, 18, 1451–1463. [Google Scholar] [CrossRef] [PubMed]
  44. Zhang, Y.; Qiao, S.; Ji, S.; Li, Y. DeepSite: Bidirectional LSTM and CNN models for predicting DNA–protein binding. Int. J. Mach. Learn. Cybern. 2020, 11, 841–851. [Google Scholar] [CrossRef]
  45. Dutta, A.; Dalmia, A.; Athul, R.; Singh, K.K.; Anand, A. Using the Chou’s 5-steps rule to predict splice junctions with interpretable bidirectional long short-term memory networks. Comput. Biol. Med. 2020, 116, 103558. [Google Scholar] [CrossRef] [PubMed]
  46. Niu, K.; Luo, X.; Zhang, S.; Teng, Z.; Zhang, T.; Zhao, Y. iEnhancer-EBLSTM: Identifying Enhancers and Strengths by Ensembles of Bidirectional Long Short-Term Memory. Front. Genet. 2021, 12, 385. [Google Scholar] [CrossRef] [PubMed]
  47. di Gangi, M.; Bosco, G.L.; Rizzo, R. Deep learning architectures for prediction of nucleosome positioning from sequences data. BMC Bioinform. 2018, 19, 127–135. [Google Scholar] [CrossRef] [Green Version]
  48. Saxena, A. Grey forecasting models based on internal optimization for Novel Corona virus (COVID-19). Appl. Soft Comput. 2021, 111, 107735. [Google Scholar] [CrossRef]
  49. van Houdt, G.; Mosquera, C.; Nápoles, G. A review on the long short-term memory model. Artif. Intell. Rev. 2020, 53, 5929–5955. [Google Scholar] [CrossRef]
  50. Shewalkar, A. Performance evaluation of deep neural networks applied to speech recognition: RNN, LSTM and GRU. J. Artif. Intell. Soft Comput. Res. 2019, 9, 235–245. [Google Scholar] [CrossRef] [Green Version]
  51. Zazo, R.; Lozano-Diez, A.; Gonzalez-Dominguez, J.; Toledano, D.T.; Gonzalez-Rodriguez, J. Language identification in short utterances using long short-term memory (LSTM) recurrent neural networks. PLoS ONE 2016, 11, e0146917. [Google Scholar] [CrossRef] [Green Version]
  52. Chen, W.; Feng, P.; Yang, H.; Ding, H.; Lin, H.; Chou, K.-C. iRNA-AI: Identifying the adenosine to inosine editing sites in RNA sequences. Oncotarget 2017, 8, 4208. [Google Scholar] [CrossRef] [Green Version]
  53. Chen, W.; Yang, H.; Feng, P.; Ding, H.; Lin, H. iDNA4mC: Identifying DNA N4-methylcytosine sites based on nucleotide chemical properties. Bioinformatics 2017, 33, 3518–3523. [Google Scholar] [CrossRef] [Green Version]
  54. Chen, W.; Feng, P.-M.; Lin, H.; Chou, K.-C. iRSpot-PseDNC: Identify recombination spots with pseudo dinucleotide composition. Nucleic Acids Res. 2013, 41, e68. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  55. Liu, B.; Li, K.; Huang, D.-S.; Chou, K.-C. iEnhancer-EL: Identifying enhancers and their strength with ensemble learning approach. Bioinformatics 2018, 34, 3835–3842. [Google Scholar] [CrossRef]
  56. Feng, P.; Yang, H.; Ding, H.; Lin, H.; Chen, W.; Chou, K.-C. iDNA6mA-PseKNC: Identifying DNA N6-methyladenosine sites by incorporating nucleotide physicochemical properties into PseKNC. Genomics 2019, 111, 96–102. [Google Scholar] [CrossRef] [PubMed]
  57. Liu, B.; Yang, F.; Huang, D.-S.; Chou, K.-C. iPromoter-2L: A two-layer predictor for identifying promoters and their types by multi-window-based PseKNC. Bioinformatics 2017, 34, 33–40. [Google Scholar] [CrossRef] [PubMed]
  58. Sun, J.; Wang, X.; Xiong, N.; Shao, J. Learning sparse representation with variational auto-encoder for anomaly detection. IEEE Access 2018, 6, 33353–33361. [Google Scholar] [CrossRef]
  59. Yi, B.; Shen, X.; Liu, H.; Zhang, Z.; Zhang, W.; Liu, S.; Xiong, N. Deep matrix factorization with implicit feedback embedding for recommendation system. IEEE Trans. Ind. Inform. 2019, 15, 4591–4601. [Google Scholar] [CrossRef]
  60. Qu, Y.; Xiong, N. RFH: A resilient, fault-tolerant and high-efficient replication algorithm for distributed cloud storage. In Proceedings of the 2012 41st International Conference on Parallel Processing, Pittsburgh, PA, USA, 10–13 September 2012; pp. 520–529. [Google Scholar]
  61. Lin, B.; Zhu, F.; Zhang, J.; Chen, J.; Chen, X.; Xiong, N.N.; Mauri, J.L. A time-driven data placement strategy for a scientific workflow combining edge computing and cloud computing. IEEE Trans. Ind. Inform. 2019, 15, 4254–4265. [Google Scholar] [CrossRef] [Green Version]
  62. Li, H.; Liu, J.; Liu, R.W.; Xiong, N.; Wu, K.; Kim, T.-H. A dimensionality reduction-based multi-step clustering method for robust vessel trajectory analysis. Sensors 2017, 17, 1792. [Google Scholar] [CrossRef] [Green Version]
  63. Fang, W.; Yao, X.; Zhao, X.; Yin, J.; Xiong, N. A stochastic control approach to maximize profit on service provisioning for mobile cloudlet platforms. IEEE Trans. Syst. Man Cybern. Syst. 2016, 48, 522–534. [Google Scholar] [CrossRef]
Figure 1. Description of the one-hot encoding scheme.
Figure 1. Description of the one-hot encoding scheme.
Electronics 11 01917 g001
Figure 2. The proposed m7G-LSTM computational model Schema.
Figure 2. The proposed m7G-LSTM computational model Schema.
Electronics 11 01917 g002
Figure 3. The auROC curve of the m7G-LSTM model.
Figure 3. The auROC curve of the m7G-LSTM model.
Electronics 11 01917 g003
Figure 4. Confusion matrix of the proposed m7G-LSTM model.
Figure 4. Confusion matrix of the proposed m7G-LSTM model.
Electronics 11 01917 g004
Figure 5. Comparison of proposed m7G-LSTM computational model with existing methods.
Figure 5. Comparison of proposed m7G-LSTM computational model with existing methods.
Electronics 11 01917 g005
Table 1. Training parameters of word2vec.
Table 1. Training parameters of word2vec.
List of ParametersWord2vec Model
Training MethodCBOW
CorpusHuman Genome
Context words3-mer
Vector size100
Window size5
Minimum Count5
Negative Sampling5
Epochs20
Table 2. Performance of CNN and LSTM models using cross-validation.
Table 2. Performance of CNN and LSTM models using cross-validation.
ModelsSensitivitySpecificityAccuracyMCC
LSTM95.9495.9795.950.919
CNN96.6293.2894.940.899
Table 3. Model comparison between the proposed m7G-LSTM and current models.
Table 3. Model comparison between the proposed m7G-LSTM and current models.
ModelsAccuracySensitivitySpecificityMCC
m7G-LSTM95.9595.9495.970.919
m7G-IFL [18]92.592.492.60.850
m7GFinder [14]89.990.889.10.799
iRNA-m7G [15]89.989.190.70.798
m7GPredictor [9]85.7083.288.20.715
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tahir, M.; Hayat, M.; Khan, R.; Chong, K.T. An Effective Deep Learning-Based Architecture for Prediction of N7-Methylguanosine Sites in Health Systems. Electronics 2022, 11, 1917. https://doi.org/10.3390/electronics11121917

AMA Style

Tahir M, Hayat M, Khan R, Chong KT. An Effective Deep Learning-Based Architecture for Prediction of N7-Methylguanosine Sites in Health Systems. Electronics. 2022; 11(12):1917. https://doi.org/10.3390/electronics11121917

Chicago/Turabian Style

Tahir, Muhammad, Maqsood Hayat, Rahim Khan, and Kil To Chong. 2022. "An Effective Deep Learning-Based Architecture for Prediction of N7-Methylguanosine Sites in Health Systems" Electronics 11, no. 12: 1917. https://doi.org/10.3390/electronics11121917

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop