Abstract
Micro RNAs (miRNA) are a type of non-coding RNA involved in gene regulation and can be associated with diseases such as cancer, cardiovascular, and neurological diseases. As such, identifying the entire genome of miRNA can be of great relevance. Since experimental methods for novel precursor miRNA (pre-miRNA) detection are complex and expensive, computational detection using Machine Learning (ML) could be useful. Existing ML methods are often complex black boxes that do not create an interpretable structural description of pre-miRNA. In this paper, we propose a novel framework that makes use of generative modeling through Variational Auto-Encoders to uncover the generative factors of pre-miRNA. After training the VAE, the pre-miRNA description is developed using a decision tree on the lower dimensional latent space. Applying the framework to miRNA classification, we obtain a high reconstruction and classification performance while also developing an accurate miRNA description.
1. Introduction
In living organisms, DNA encodes all the information used by the organism to develop and survive. To make use of the information encoded in the DNA, it needs to be translated to RNA. One type of RNA is messenger RNA (mRNA), which is used to create proteins. Sometimes, depending on the cell type, it is desirable that a certain mRNA is not expressed, which is where micro RNA (miRNA) comes into play by silencing and inhibiting the expression of certain mRNAs. MiRNAs can silence mRNAs by binding to them, following which the mRNA is degraded. As such, the definition of what constitutes an miRNA is functional since it needs to carry out this task. Depending on the shape of the miRNA, it can be used to silence (multiple) different mRNAs. Since over- and under-expression of miRNA can lead to various diseases, it is important to understand the genome of miRNA better.
It is estimated that there are around 2300 types of human miRNA [1] while only around 700 are documented [2]. Finding new miRNAs often requires complex lab conditions [3,4] where potential new miRNA needs to be differentiated from other RNA. To detect novel miRNAs, the focus is typically on identifying precursor miRNAs (pre-miRNAs), which are intermediate forms processed before becoming mature miRNAs. Initially, primary miRNA is transcribed from DNA and processed by enzymes into pre-miRNA [5]. Pre-miRNA is then further processed into mature miRNA within the cell. Folded in a hairpin structure, pre-miRNA contains more nucleotides and therefore possesses more distinguishing features (see Figure 1) compared to other types of RNA. As a result, detecting novel pre-miRNAs is generally easier than detecting mature miRNAs. While some features, like the presence of a terminal loop and a large fraction of base pairs (C-G/A-U) in the stem, are prevalent in pre-miRNA, they are not enough to differentiate it from other RNA molecules and hence are not sufficient to form a structural description of pre-miRNA.
Figure 1.
Artificial pre-miRNA strand [6] labeled with different properties that can be present in (non) pre-miRNA.
A number of data-driven methods have been developed to identify novel pre-miRNA. Some of the pre-miRNA detection methods are based on traditional ML methods, which use engineered features [7] such as Random Forests (RFs) [8] and Support Vector Machines (SVMs) [3,8,9,10,11], where only limited interpretability is possible through feature importance. In addition, several Deep Learning methods have been proposed, which rely on Convolutional Neural Networks (CNNs) [12,13,14,15,16,17], Recurrent Neural Networks (RNNs) [14,18,19], and transformers [20,21]. While these methods are able to achieve state-of-the-art accuracy on the miRNA classification task, interpretation is only possible through methods that use activation maps or techniques such as concept whitening [22]. As such, these black-box models do not directly allow us to form a description of pre-miRNA, which would, in turn, allow us to better understand these molecules and their roles in the cell.
In recent years, generative models such as Variational Auto-Encoders (VAEs) and Generative Adversarial Networks (GANs) have been successfully applied to a wide variety of tasks such as protein [23,24], drug [25,26,27], and DNA design [28,29]. Typically, VAEs and GANs generate data by mapping a low-dimensional latent space to high-dimensional data. The low-dimensional latent space in these models is usually entangled, meaning that one latent variable can encode multiple aspects of the generative process. To tackle this problem, different disentanglement techniques have been proposed [30,31,32], where the loss function or model architecture forces the model to learn a latent representation in which each variable only encodes one aspect of the generative process. However, it is not inherently clear which latent variable corresponds to which property of the data and how these properties could be used to describe the class of a datapoint.
In this work, we develop a method for developing pre-miRNA descriptions, using the lower dimensional latent space of a latent variable model (Figure 2). Our contributions are two-fold: (i) We propose a novel framework based on VAEs and Decision Trees (DTs), which can be used to develop interpretable descriptions of the different object classes y being modeled in terms of features f; (ii) We apply this framework to the domain of precursor microRNA where we demonstrate that it obtains biologically sound structural descriptions.
Figure 2.
Overview of the proposed description generation framework. The Variational Autoencoder (VAE) is trained to encode the input into a latent space, which is then decoded. A Decision Tree (DT) is trained on the latent space to split based on features, using an SVM for each feature to assess which split maximizes information gain for the classification of pre-miRNA versus non-pre-miRNA. The resulting tree provides interpretable descriptions of pre-miRNA.
2. Methods
2.1. Data
We use the modmiRBase dataset [12], which contains 49602 RNAs from various organisms and is balanced in terms of class. The pre-miRNA sequences in the dataset were sourced from mirbase.org [33] and mirgenedb.org [34]. The non-pre-miRNA sequences were obtained from various datasets, which included existing non-pre-miRNAs with characteristics similar to pre-miRNAs [3,35,36], as well as shuffled or suboptimally folded pre-miRNAs [7]. The RNA sequences are folded using the RNAFold algorithm [37], which predicts the secondary structure by identifying stable configurations where complementary nucleotides bond. Following this, the RNA sequences are encoded as images with dimensions of 100 by 25 pixels [12]. In this encoding, each nucleotide is represented as a colored bar—adenine (A) in blue, cytosine (C) in yellow, guanine (G) in green, and uracil (U) in red. Gaps (represented in black) are added for nucleotides lacking complementary pairs. The length of each bar is inversely related to the bond strength between nucleotides, with strong bonds occurring between base pairs and weaker bonds between other pairs or gaps. Additionally, the length of each bar increases with multiple consecutive weak bonds, while gaps are consistently represented with a length of 2 pixels. This encoding captures the physical structure of the RNA, with the hairpin formations resembling actual hairpins in the images. Additionally, any bulges present in the RNA structure are represented as protruding bars, accurately reflecting the spatial arrangement of the nucleotides. An example of an image encoding can be found in Figure 3.
Figure 3.
Example of RNA image encoding.
Since the shape plays an important role in the process of silencing mRNAs, we also encode the bond strength between nucleotides. This is performed in vector m of size 100, where a strong bond is assigned a value of 1, while a weak bond (or no bond) is assigned a value of 0.
For each datapoint, we also calculated multiple features, using the algorithms from [22]. Several of the features can be found in Figure 1, such as the presence of a terminal loop, the size of the terminal loop, and the length of the stem. These features are not used directly in our VAE but are used in the decision tree algorithm instead.
2.2. Model
To model the (non) pre-miRNA, we made use of the Domain Invariant Variational Autoencoder (DIVA) framework [32], with a latent space for bond strength (), class () and remaining variance (). Each latent space has a size of 64. For each latent space, a separate encoder based on ResNet [38] was used, with an overview of the architecture in Figure 4a. Furthermore, and have conditional priors, whereas has a standard normal prior. The prior for consists of a 1D convolutional network (Figure 4b), followed by a fully connected layer, while the prior of consists of two fully connected layers. In addition, auxiliary classifiers were used to further enhance the disentanglement. The auxiliary classifier for consists of a fully connected layer followed by 1D deconvolutional layers (Figure 4c), while the auxiliary classifier for y consists of two fully connected layers. Since the shape of the miRNA plays an important role in the mRNA silencing process, we made use of an additional auxiliary classifier for y on . Finally, to organize the latent space better, Inverse Autoregressive Flow (IAF) [39] was applied to all 3 latent spaces. Each IAF consists of 8 blocks of 2 Masked Autoencoder for Distribution Estimation (MADE) [40] layers, which have a context size of 32 (obtained from the encoder), as well as a hidden size of 1080.
Figure 4.
Architecture of different model components. The first number following the layer name indicates the kernel size, and the second number indicates the amount of filters. Unless mentioned otherwise, each convolutional layer uses a stride (s) of 1. Convolutional layers in the encoder use same padding (p), while other layers use valid padding unless mentioned otherwise. (a) Encoder architecture. Sampling from the final layer is followed by IAF to obtain the latent space. (b) Prior () architecture. (c) Classifier for m architecture. (d) Decoder for bar height/color. Both decoders are preceded by the same fully connected layer and reshape operation. The height decoder uses 2D transposed convolutions, while the decoder for color first sums over the height dimension and then uses 1D transposed convolutions. For the final layer, the color decoder gives the probability for each color (top and bottom row, 5 colors in each bar). Outputs of the two decoders are combined to obtain .
Since our data are highly structured, both the encoders and decoders were designed with the data complexity in mind. In the input representation, we one-hot encoded the 5 possible different colors to ensure that distances between different colors are always the same, which results in an image with 5 channels. Furthermore, using a regular deconvolutional image decoder could yield invalid reconstructions, as respecting the structure of the data would not be possible. To limit the amount of invalid images, we separately decode the height and color of each bar. After both height and color distributions are decoded, we combine the two distributions to obtain a distribution over both height and color, which is used to calculate the reconstruction loss. The decoder of the bar height is a 2D deconvolutional decoder. It outputs the probability for the different possible bar heights after softmax activation. To obtain the actual bars, the probability distribution is multiplied by the different bar heights, similar to the approach in StampNet [41]. The decoder for the bar color is a 1D deconvolutional decoder and outputs the color probabilities of the bars in the bottom and top rows. Finally, the heights and colors are multiplied to obtain the reconstruction. When sampling from the decoder, a discrete value should be sampled from both the height and color distributions. The architecture of both decoders can be found in Figure 4d. In both the encoders and decoders, the number of filters was chosen based on Cordero et al. [12].
An overview of the full model architecture can be found in Figure 5. In the model, each layer is followed by batch normalization and uses the ELU activation function (unless mentioned otherwise). For training the model, the loss functions in Equation (2) were used. In Equation (1), the first term represents the reconstruction loss, while the remaining terms represent the KL divergence between the encoded prior () and the (learned) prior (). The penalty terms for each KL divergence are represented by . In Equation (2), each additional term represents the loss of the different auxiliary classifiers (), with being the penalty term for each classifier. The dataset was split into a training set containing 34,721 samples and a test set containing 14,881 samples, with both datasets being balanced. For training, we set each parameter to 0.5. Both and were set to 12, while was set to 1. The hyperparameters for and were selected to ensure the model learns a well-organized latent space while still generating high-quality samples. The model was trained using the Adam optimizer [42] with a batch size of 64 and a learning rate of 0.0005 until convergence.
Figure 5.
Full DIVA model architecture. represents the concatenation of the different latent spaces.
To verify that the model architecture is optimal, we performed several ablation studies where we compared the aforementioned model with a standard(/) VAE using a decoder consisting of fully connected layers (VAE), a -VAE with IAF applied on the latent space (-IAF-VAE), as well as with a deconvolutional decoder and IAF (DC--IAF-VAE). For all models, was set to 0.5. Code for all models and experiments can be found on https://www.github.com/marko-petkovic/mirna (accessed on 22 September 2024).
2.3. Decision Tree
To develop a description of pre-miRNA, we create a decision tree algorithm on top of the latent space, as this latent space contains both information regarding the shape and class of RNA. Since this latent space is organized, various features of the RNA should be linearly separable within this space. As such, a linear SVM can be trained to partition the space based on a feature, such as the length of the RNA.
Compared to a normal decision tree algorithm, such as CART [43] or C4.5 [44], our algorithm makes splits based on the learned latent representation rather than the features (f) themselves. To create a split using our decision tree algorithm, for each feature, we train an SVM to partition the latent space (), with the feature being the target variable. In case a feature is continuous, we make multiple binary features out of it, where the feature is divided according to different thresholds. Alternatively, one could train a regression model rather than a classifier and decide the threshold for a binary feature afterward. For each split, we assess whether the performance of the SVM is above a threshold, and only features classified with an accuracy above the threshold are considered. Then, to decide which split to make, we asses which split yields the highest information gain based on the class of the RNA. The SVM used to make the split is stored as an attribute of the decision tree node. After training a full decision tree, we can assess which descriptions of pre-miRNA the model developed by following the paths down the tree that lead to a pre-miRNA classification. The full algorithm for making splits for the DT can be found in Algorithm 1.
| Algorithm 1 MakeSplit() |
|
To predict whether a new datapoint is a pre-miRNA, it should first be encoded by the encoder for . Then, the decision tree should be followed, where it is checked on which side of the split by the SVM the point lies.
After training the model, we trained a decision tree with , , and using the training set and features f from [22]. Each of the continuous features was partitioned into multiple binary features.
3. Results
3.1. Performance
In Table 1, the reconstruction performance of our model and the different ablations can be found. The reconstruction performance was calculated based on three different metrics: the Mean Absolute Error (MAE) of the entire image, the nucleotide accuracy, and the MAE of the length of the nucleotides. In Equations (3)–(5), the evaluation metrics are defined. In Equation (3), represents the true pixel value, while is the predicted pixel value for each of the P pixels. In Equation (4), is the true color label, and is the predicted nucleotide for the b-th bar, with B being the total number of bars. The indicator function evaluates to 1 if the colors match and 0 otherwise. Finally, in Equation (5), and represent the true and predicted lengths of the b-th bar, respectively, for all B bars. As can be seen, the DC-IAF-DIVA model achieved the highest performance in terms of all three metrics.
Table 1.
Reconstruction statistics for different model types. Bold statistics represent the lowest obtained errors.
To verify that the model has learned an organized latent space, we have visualized the latent spaces using t-SNE [45] dimensionality reduction. As can be seen in Figure 6, the disentanglement of the DIVA model indeed appears to be successful, as there seems to be a clear separation between classes in and , while this is not present in . Taking a further look at the distribution of the fraction of base pairs in the stem, terminal loop length, and the stem length over the latent space (Figure 7), we see that the latent space is well organized. Therefore, it can be used by our decision tree algorithm to develop a pre-miRNA description.
Figure 6.
Latent spaces of DIVA model after dimensionality reduction using t-SNE. Yellow dots represent pre-miRNA.
Figure 7.
latent space colored according to the fraction base pairs in stem, terminal loop length, and stem length. Lighter colors indicate higher values for each variable.
3.2. Conditional Generation
In addition to inspecting the latent space embeddings, we also qualitatively verify the learned representation of . Here, we first encode a selected RNA image to obtain , , and . Following this, the bond strength m of the sample is modified and passed through the conditional prior to obtain . Afterward, we reconstruct the RNA based on , , and . Finally, we verify whether the newly generated RNA image indeed follows the provided bond strength.
In Figure 8, we can find the results of the conditional generation experiment. In the top left, the original image can be found, with the bond strength m displayed below the picture. In each column, a different sample from the same conditional prior is reconstructed, while for each row a unique m was used. Overall, we see that there are a few (minor) errors in reconstruction, but in general, the generated images follow the provided bond strength m. As such, we can conclude that the learned latent space is well organized both in terms of features and bond strength.
Figure 8.
Generated RNA images, with sampled from the prior of various m. In each row, a different m was used. In the second row, the same m as in the original image was used.
3.3. Decision Tree
The obtained description from the decision tree can be found in Figure 9. Overall, the model achieved an accuracy of 0.912, a sensitivity of 0.923, and a specificity of 0.955. The concept whitening method achieved an accuracy of 0.928 on the same dataset while offering less interpretation compared to the decision tree. An overview of the performance comparison between the two methods can be found in Table 2.
Figure 9.
Learned pre-miRNA description.
Table 2.
Model Performance on pre-miRNA Classification.
In addition to the high classification performance, the decision tree learned biologically sound classification rules. The first split in the decision tree is made based on the amount of base pairs in the stem, where samples with a high amount of base pairs are considered to be pre-miRNA, which is also supported by literature [6]. Similarly, short RNA sequences are also considered unlikely to be pre-miRNA [6], which can be seen in the second split of the tree. The final feature used by the tree to discriminate pre-miRNA from other RNA is the absence of large asymmetric bulges, which is also a known property of pre-miRNA [46]. In general, given the supplied features, we see that the decision tree is able to generate a biologically valid description of pre-miRNA.
4. Discussion
In this work, we have proposed a new method for latent space interpretability in VAEs and applied it to pre-miRNA generation. The trained VAE for pre-miRNA generation has shown excellent performance, validating its suitability for the application of our interpretability method. Our analysis demonstrated that the latent space descriptions developed through the method align with existing biological knowledge of pre-miRNAs, suggesting both the VAE model and our interpretability approach perform robustly in capturing meaningful structural features.
Despite these promising results, there are limitations to our approach. The model is focused solely on pre-miRNA structures and does not account for interactions with target genes, which could provide further insights into miRNA functionality. Additionally, the interpretability method relies on existing biological knowledge for validation, meaning its utility may be limited in less well-characterized regions of the latent space or for novel pre-miRNA variants. Expanding the model to account for more diverse and less-studied pre-miRNA types could pose challenges in interpretability and generalization.
A potential direction for future work could be to explore the use of geometric GNNs instead of CNN-based models for learning pre-miRNA structures. Since miRNA is primarily distinguished from other RNA by its function, which is closely linked to its shape, GNNs may provide a more effective representation of the spatial and structural properties of miRNAs. By using a geometric representation of the RNA molecule, GNNs can more accurately capture and process its shape, potentially improving interpretability and generalization across more complex datasets compared to image-based representations. Additionally, target interaction data, such as miRNA-mRNA interactions, could be integrated to account for functional relationships that influence miRNA activity. Recent machine learning methods that predict miRNA-target interactions [47,48,49] could be leveraged to enhance the model’s understanding of these regulatory networks. Incorporating such interaction data, either as additional input features or for refining the latent space, could significantly improve the model’s ability to generalize across diverse datasets and capture the biological relevance of pre-miRNAs more effectively.
Author Contributions
Conceptualization, M.P. and V.M.; methodology, M.P. and V.M.; software, M.P.; validation, M.P. and V.M.; formal analysis, M.P.; investigation, M.P.; resources, V.M.; writing—original draft preparation, M.P.; writing—review and editing, M.P. and V.M.; visualization, M.P.; supervision, V.M.; project administration, V.M. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Data Availability Statement
The data as well as the code for the experiments and results can be found on https://github.com/marko-petkovic/mirna (accessed on 22 September 2024).
Conflicts of Interest
The authors declare no conflicts of interest.
Abbreviations
The following abbreviations are used in this manuscript:
| VAE | Variational Auto-Encoder |
| RNA | Ribonucleic Acid |
| KL | Kullback–Leibler |
| DIVA | Domain Independent Variational Auto-encoder |
| DT | Decision Tree |
References
- Alles, J.; Fehlmann, T.; Fischer, U.; Backes, C.; Galata, V.; Minet, M.; Hart, M.; Abu-Halima, M.; Grässer, F.A.; Lenhof, H.P.; et al. An estimate of the total number of true human miRNAs. Nucleic Acids Res. 2019, 47, 3353–3364. [Google Scholar] [CrossRef] [PubMed]
- Kozomara, A.; Birgaoanu, M.; Griffiths-Jones, S. miRBase: From microRNA sequences to function. Nucleic Acids Res. 2019, 47, D155–D162. [Google Scholar] [CrossRef]
- Ng, K.L.S.; Mishra, S.K. De novo SVM classification of precursor microRNAs from genomic pseudo hairpins using global and intrinsic folding measures. Bioinformatics 2007, 23, 1321–1330. [Google Scholar] [CrossRef]
- Saçar, M.D.; Allmer, J. Machine learning methods for microRNA gene prediction. In miRNomics: microRNA Biology and Computational Analysis; Springer: Berlin/Heidelberg, Germany, 2014; pp. 177–187. [Google Scholar]
- Erson-Bensan, A.E. Introduction to microRNAs in biological systems. In miRNomics: microRNA Biology and Computational Analysis; Springer: Berlin/Heidelberg, Germany, 2014; pp. 1–14. [Google Scholar]
- Allmer, J. Computational and bioinformatics methods for microRNA gene prediction. In miRNomics: microRNA Biology and Computational Analysis; Springer: Berlin/Heidelberg, Germany, 2014; pp. 157–175. [Google Scholar]
- Saçar Demirci, M.D.; Baumbach, J.; Allmer, J. On the performance of pre-microRNA detection algorithms. Nat. Commun. 2017, 8, 330. [Google Scholar] [CrossRef]
- Jiang, P.; Wu, H.; Wang, W.; Ma, W.; Sun, X.; Lu, Z. MiPred: Classification of real and pseudo microRNA precursors using random forest prediction model with combined features. Nucleic Acids Res. 2007, 35, W339–W344. [Google Scholar] [CrossRef]
- Batuwita, R.; Palade, V. microPred: Effective classification of pre-miRNAs for human miRNA gene prediction. Bioinformatics 2009, 25, 989–995. [Google Scholar] [CrossRef]
- Ding, J.; Zhou, S.; Guan, J. MiRenSVM: Towards better prediction of microRNA precursors using an ensemble SVM classifier with multi-loop features. BMC Bioinform. 2010, 11, S11. [Google Scholar] [CrossRef]
- Xue, C.; Li, F.; He, T.; Liu, G.P.; Li, Y.; Zhang, X. Classification of real and pseudo microRNA precursors using local structure-sequence features and support vector machine. BMC Bioinform. 2005, 6, 310. [Google Scholar] [CrossRef]
- Cordero, J.; Menkovski, V.; Allmer, J. Detection of pre-microRNA with Convolutional Neural Networks. BioRxiv 2019. [Google Scholar] [CrossRef]
- Do, B.T.; Golkov, V.; Gürel, G.E.; Cremers, D. Precursor microRNA identification using deep convolutional neural networks. BioRxiv 2018. [Google Scholar] [CrossRef]
- Tasdelen, A.; Sen, B. A hybrid CNN-LSTM model for pre-miRNA classification. Sci. Rep. 2021, 11, 14125. [Google Scholar] [CrossRef] [PubMed]
- Zheng, X.; Xu, S.; Zhang, Y.; Huang, X. Nucleotide-level convolutional neural networks for pre-miRNA classification. Sci. Rep. 2019, 9, 628. [Google Scholar] [CrossRef] [PubMed]
- Yones, C.; Raad, J.; Bugnon, L.A.; Milone, D.H.; Stegmayer, G. High precision in microRNA prediction: A novel genome-wide approach with convolutional deep residual networks. Comput. Biol. Med. 2021, 134, 104448. [Google Scholar] [CrossRef] [PubMed]
- Bugnon, L.A.; Raad, J.; Merino, G.A.; Yones, C.; Ariel, F.; Milone, D.H.; Stegmayer, G. Deep Learning for the discovery of new pre-miRNAs: Helping the fight against COVID-19. Mach. Learn. Appl. 2021, 6, 100150. [Google Scholar] [CrossRef] [PubMed]
- Park, S.; Min, S.; Choi, H.S.; Yoon, S. Deep recurrent neural network-based identification of precursor micrornas. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
- Bell, J.; Hendrix, D.A. Predicting Drosha and Dicer Cleavage Sites with DeepMirCut. Front. Mol. Biosci. 2022, 8, 799056. [Google Scholar] [CrossRef]
- Raad, J.; Bugnon, L.A.; Milone, D.H.; Stegmayer, G. miRe2e: A full end-to-end deep model based on transformers for prediction of pre-miRNAs. Bioinformatics 2022, 38, 1191–1197. [Google Scholar] [CrossRef]
- Gupta, S.; Shankar, R. miWords: Transformer-based composite deep learning for highly accurate discovery of pre-miRNA regions across plant genomes. Briefings Bioinform. 2023, 24, bbad088. [Google Scholar] [CrossRef]
- Van den Brandt, I. Towards Concept-Based Interpretability of Pre-miRNA Detection Using Convolutional Neural Networks. Master’s Thesis, Eindhoven University of Technology, Eindhoven, The Netherlands, 2021. [Google Scholar]
- Ingraham, J.; Garg, V.; Barzilay, R.; Jaakkola, T. Generative models for graph-based protein design. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; Volume 32. [Google Scholar]
- Strokach, A.; Kim, P.M. Deep generative modeling for protein design. Curr. Opin. Struct. Biol. 2022, 72, 226–236. [Google Scholar] [CrossRef]
- Cheng, Y.; Gong, Y.; Liu, Y.; Song, B.; Zou, Q. Molecular design in drug discovery: A comprehensive review of deep generative models. Briefings Bioinform. 2021, 22, bbab344. [Google Scholar] [CrossRef]
- Grisoni, F.; Huisman, B.J.; Button, A.L.; Moret, M.; Atz, K.; Merk, D.; Schneider, G. Combining generative artificial intelligence and on-chip synthesis for de novo drug design. Sci. Adv. 2021, 7, eabg3338. [Google Scholar] [CrossRef]
- Tong, X.; Liu, X.; Tan, X.; Li, X.; Jiang, J.; Xiong, Z.; Xu, T.; Jiang, H.; Qiao, N.; Zheng, M. Generative models for de novo drug design. J. Med. Chem. 2021, 64, 14011–14027. [Google Scholar] [CrossRef] [PubMed]
- Killoran, N.; Lee, L.J.; Delong, A.; Duvenaud, D.; Frey, B.J. Generating and designing DNA with deep generative models. arXiv 2017, arXiv:1712.06148. [Google Scholar]
- Gupta, A.; Zou, J. Feedback GAN for DNA optimizes protein functions. Nat. Mach. Intell. 2019, 1, 105–111. [Google Scholar] [CrossRef]
- Chen, R.T.; Li, X.; Grosse, R.B.; Duvenaud, D.K. Isolating sources of disentanglement in variational autoencoders. In Proceedings of the Advances in Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018; Volume 31. [Google Scholar]
- Higgins, I.; Matthey, L.; Pal, A.; Burgess, C.; Glorot, X.; Botvinick, M.; Mohamed, S.; Lerchner, A. beta-vae: Learning basic visual concepts with a constrained variational framework. In Proceedings of the International Conference on Learning Representations, San Juan, Puerto Rico, 2–4 May 2016. [Google Scholar]
- Ilse, M.; Tomczak, J.M.; Louizos, C.; Welling, M. Diva: Domain invariant variational autoencoders. In Proceedings of the Medical Imaging with Deep Learning–PMLR, Montreal, QC, Canada, 6–8 July 2020; pp. 322–348. [Google Scholar]
- Griffiths-Jones, S. The microRNA registry. Nucleic Acids Res. 2004, 32, D109–D111. [Google Scholar] [CrossRef] [PubMed]
- Fromm, B.; Domanska, D.; Høye, E.; Ovchinnikov, V.; Kang, W.; Aparicio-Puerta, E.; Johansen, M.; Flatmark, K.; Mathelier, A.; Hovig, E.; et al. MirGeneDB 2.0: The metazoan microRNA complement. Nucleic Acids Res. 2019, 48, D132–D141. [Google Scholar] [CrossRef]
- Gudyś, A.; Szcześniak, M.W.; Sikora, M.; Makałowska, I. HuntMi: An efficient and taxon-specific approach in pre-miRNA identification. BMC Bioinform. 2013, 14, 83. [Google Scholar] [CrossRef]
- Wei, L.; Liao, M.; Gao, Y.; Ji, R.; He, Z.; Zou, Q. Improved and promising identification of human microRNAs by incorporating a high-quality negative set. IEEE/ACM Trans. Comput. Biol. Bioinform. 2013, 11, 192–201. [Google Scholar] [CrossRef]
- Hofacker, I.L.; Fontana, W.; Stadler, P.F.; Bonhoeffer, L.S.; Tacker, M.; Schuster, P. Fast folding and comparison of RNA secondary structures. Monatshefte Chem./Chem. Mon. 1994, 125, 167–188. [Google Scholar] [CrossRef]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Kingma, D.P.; Salimans, T.; Jozefowicz, R.; Chen, X.; Sutskever, I.; Welling, M. Improved variational inference with inverse autoregressive flow. In Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; Volume 29. [Google Scholar]
- Germain, M.; Gregor, K.; Murray, I.; Larochelle, H. Made: Masked autoencoder for distribution estimation. In Proceedings of the International Conference on Machine Learning—PMLR, Lille, France, 7–9 July 2015; pp. 881–889. [Google Scholar]
- Visser, J.; Corbetta, A.; Menkovski, V.; Toschi, F. StampNet: Unsupervised multi-class object discovery. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 2951–2955. [Google Scholar]
- Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
- Breiman, L.; Friedman, J.H.; Olshen, R.A.; Stone, C.J. Classification and Regression Trees; Routledge: London, UK, 2017. [Google Scholar]
- Quinlan, J.R. C4. 5: Programs for Machine Learning; Elsevier: Amsterdam, The Netherlands, 2014. [Google Scholar]
- Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 11. [Google Scholar]
- Kang, W.; Friedländer, M.R. Computational prediction of miRNA genes from small RNA sequencing data. Front. Bioeng. Biotechnol. 2015, 3, 7. [Google Scholar] [CrossRef] [PubMed]
- Wen, M.; Cong, P.; Zhang, Z.; Lu, H.; Li, T. DeepMirTar: A deep-learning approach for predicting human miRNA targets. Bioinformatics 2018, 34, 3781–3787. [Google Scholar] [CrossRef]
- Yu, X.; Jiang, L.; Jin, S.; Zeng, X.; Liu, X. preMLI: A pre-trained method to uncover microRNA–lncRNA potential interactions. Briefings Bioinform. 2022, 23, bbab470. [Google Scholar] [CrossRef] [PubMed]
- Cheng, S.; Guo, M.; Wang, C.; Liu, X.; Liu, Y.; Wu, X. MiRTDL: A deep learning approach for miRNA target prediction. IEEE/ACM Trans. Comput. Biol. Bioinform. 2015, 13, 1161–1169. [Google Scholar] [CrossRef] [PubMed]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).