Next Article in Journal
Targeting Phospholipase D Pharmacologically Prevents Phagocytic Function Loss of Retinal Pigment Epithelium Cells Exposed to High Glucose Levels
Next Article in Special Issue
Regulatory Small RNAs for a Sustained Eco-Agriculture
Previous Article in Journal
CTH/H2S Regulates LPS-Induced Inflammation through IL-8 Signaling in MAC-T Cells
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

PINC: A Tool for Non-Coding RNA Identification in Plants Based on an Automated Machine Learning Framework

1
Anhui Province Key Laboratory of Smart Agricultural Technology and Equipment, Anhui Agricultural University, Hefei 230036, China
2
College of Information and Computer Science, Anhui Agricultural University, Hefei 230036, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Int. J. Mol. Sci. 2022, 23(19), 11825; https://doi.org/10.3390/ijms231911825
Submission received: 23 August 2022 / Revised: 26 September 2022 / Accepted: 30 September 2022 / Published: 5 October 2022
(This article belongs to the Special Issue Plant Non-coding RNAs in the Era of Biological Big Data 2.0)

Abstract

:
There is evidence that non-coding RNAs play significant roles in the regulation of nutrient homeostasis, development, and stress responses in plants. Accurate identification of ncRNAs is the first step in determining their function. While a number of machine learning tools have been developed for ncRNA identification, no dedicated tool has been developed for ncRNA identification in plants. Here, an automated machine learning tool, PINC is presented to identify ncRNAs in plants using RNA sequences. First, we extracted 91 features from the sequence. Second, we combined the F-test and variance threshold for feature selection to find 10 features. The AutoGluon framework was used to train models for robust identification of non-coding RNAs from datasets constructed for four plant species. Last, these processes were combined into a tool, called PINC, for the identification of plant ncRNAs, which was validated on nine independent test sets, and the accuracy of PINC ranged from 92.74% to 96.42%. As compared with CPC2, CPAT, CPPred, and CNIT, PINC outperformed the other tools in at least five of the eight evaluation indicators. PINC is expected to contribute to identifying and annotating novel ncRNAs in plants.

1. Introduction

RNA is the template that codes for the proteins required to create cellular functions. RNA is structurally similar to DNA, but its function and chemical composition are fundamentally different. At a higher level, RNA is divided into two main groups: coding RNA that accounts for approximately 2% of all RNAs, and non-coding RNA (ncRNA) that accounts for the majority (>90%) of RNAs [1]. Non-coding RNA refers to all RNAs that are transcribed from DNA but do not code for proteins. Additionally, ncRNA can be categorized into two groups according to the size of the sequence: long non-coding RNAs (lncRNAs) with sequences >200 nucleotides and small non-coding RNAs (sncRNAs) with sequences shorter than 200 nucleotides [2]. In previous research, ncRNAs have frequently been referred to as “useless genes” or transcriptional “noise” [3,4]. In contrast, a growing number of experiments have demonstrated that ncRNAs play important biological roles in a variety of biological processes, including gene regulation/expression, gene silencing, RNA modification and processing, as well as multiple important roles in life activities [5,6,7]. Numerous plant-specific biological processes, including the regulation of plant nutrient homeostasis, development, and stress responses, have been linked to ncRNAs [8,9,10]. MiRNAs and trans-acting siRNAs, for instance, contribute to leaf senescence in Arabidopsis; miR164 and its target ORE1 control leaf senescence in Arabidopsis, and as miR164 expression declines, ORE1 expression eventually increases [11]. In addition, overexpression of miR398b has been shown to decrease the transcript levels of genes encoding superoxide dismutase (CSD1, CSD2, SODX, and CCSD), which resulted in the production of reactive oxygen species (ROS) and increased rice resistance to Magnaporthe oryzae [12,13]. In recent years, to facilitate subsequent analyses and research of transcripts, ncRNA identification has been one of the tasks that needs to be addressed. Numerous bioinformatics methods and experiments have been developed for ncRNA identification and to evaluate their functions [14,15]. Genomic SELEX, microarray analysis, and chemical RNA-Seq are the most commonly used experimental techniques [16]; however, they are costly and time-consuming. Therefore, bioinformatics may be a more effective means of addressing the biological aspects of the problem.
Kong et al. developed the Coding Potential Calculator (CPC) in 2007 [17]. The CPC selected a number of biologically significant features, including ORF quality, coverage, and integrity. These features were incorporated into a support vector machine for coding potential identity, but its performance was dependent on sequence comparisons. CPC was revised in 2017 with the release of CPC2 [18]. CPC2 is faster and more accurate than CPC, and, as an input to the SVM model, it uses ORF size and integrity, a Fickett score, and the isoelectric point extracted from the original RNA sequence. CPC2 is a relatively neutral tool, which makes it somewhat more applicable to transcriptomes of non-model organisms. CPAT, developed by Wang et al. in 2013, is a logistic-regression-model-based ncRNA identification tool that classifies ncRNAs and cRNAs based on features such as ORF size and coverage, Fickett score, and hexamer score [19]. CNCI was proposed by Liang et al. in 2013, and while it is also based on the same SVM classifier as CPC2, it uses different features, categorizing ncRNA and cRNA based on ANT features [20]. CNIT is an updated version of CNCI that was released in 2019. CNIT employs the more robust integrated machine model XGBoost for classification [21]. Tong et al. introduced CPPred in 2019 [22] as an SVM-based tool. This tool distinguishes between ncRNAs and coding RNAs using the same ORF features as CPC2, as well as the isoelectric point, stability index, gravity three peptide, hexamer score, CTD, and Fickett score features. A number of tools have been published that can distinguish between ncRNA and coding RNA; however, the tools have some limitations, for example, their application is mainly limited to vertebrates and mammals. In addition, these tools rarely consider using plants for model training. Most tools only use the model plant Arabidopsis, and rarely involve other non-model plants. Moreover, since ncRNAs of animals are mainly transcribed by polymerase II, while ncRNAs of plants are mainly transcribed by RNA polymerase II, IV, and V [23], and ncRNAs are characterized by low-level expression and cross-species conservation [24], these tools for ncRNA identification in animals cannot guarantee the reliability in plants. Therefore, it is necessary to construct a powerful tool for ncRNA identification in plants.
Automatic machine learning (AutoML) is the process of applying machine learning to real-world problems in an automated manner. Since 2013, frameworks have been developed that have been based on the AutoML concept. AutoWEKA was the first AutoML framework to emerge [25]; it automatically selected models and hyperparameters. Additionally, H2O [26] and TPOT [27] were created. H2O is a JAVA-based framework that supports multiple types of grid searches to identify the optimal parameters following the generation of an integrated model. At its core, TPOT is a tree-based process optimization tool based on a genetic algorithm. Today, more and more frameworks, such as AutoGluon [28] and AutoKeras [29], have been developed based on the concept of AutoML. These frameworks have also been applied to Alzheimer’s disease diagnosis [30], biomedical big data [31], and additional bioinformatics fields [32].
In this experiment, we developed PINC, an AutoML-based instrument for the identification of ncRNAs and cRNAs in plants. The AutoML framework does not require a great deal of effort and time to optimize the model; it simply accepts the processed data as an input, tunes and sets the framework’s parameters, and then outputs the model automatically. Our experimental results include a number of significant contributions: (1) By combining the F-test and variance threshold, 10 out of 91 features were identified as being able to strongly distinguish between ncRNAs and coding RNA in plants. (2) Using the AutoML framework, a neutral model for non-coding RNA identification was obtained. (3) We combined the two previous points and developed a tool called PINC for ncRNA identification. After comparing PINC with the CPC2, CPAT, CNIT, and CPPred identification tools on nine independent test sets to validate the performance of PINC, we discovered that PINC performed exceptionally well on these independent test sets. This suggests that PINC is a reliable method for ncRNA identification in plants. In addition, users can upload their data for identification, which facilitates the study of plants that have received less attention.

2. Results

2.1. Training Setup

Once the features were selected, the models were tuned to find the best parameters, and the results were validated using a five-fold cross-validation procedure. A benchmark dataset of 4000 randomly selected data from each class was constructed for training and validation. Meanwhile, to ensure the validity of the experiments, we repeated the above experiments 100 times. As shown in Figure 1A, the highest accuracy of the 100 experiments was 95.32% and the lowest was 94.52%, mostly distributed between 94.6% and 94.9%, with very small fluctuations. For further proof, we averaged the accuracy of every fifth experiment, as shown in Figure 1B, and the curve fluctuates even less. This result shows that the randomly selected data is representative of the entire data set. Therefore, we took 4000 randomly selected data from each class as our baseline dataset.

2.2. Performance Comparison of the Feature Selection Methods

In this research, 91 features were filtered using four feature selection methods: F-test, variance threshold, RF, and variance threshold combined with F-test (VT-F). These feature selection methods were compared in order to assess their usefulness. These feature selection methods use learning curves to continuously reduce the number of available features and to select the most appropriate features. The maximum validation set accuracy was 94.77 percent when the first 31 features were chosen using F-test filtering, and it was 94.29 percent when the first 25 features were chosen using variance threshold filtering. For VT-F, features below the mean were first filtered out using a variance threshold, and then the remaining features were filtered using the F-test, with a maximum accuracy of 95.25 percent when the first 10 features were selected. The evaluation of the three previously described feature selection methods was based on the AutoGluon model. For RF, the range of features was narrowed down based on the importance of the features, with the highest accuracy of 93.27 percent when the first 21 features were selected, and the 21 features were then fed into AutoGluon with an accuracy of 94.72 percent. In addition to accuracy, we compared SE, F1, MCC, and SPC performance metrics. Table 1 demonstrates that the method combining the F-test and variance threshold for feature selection outperformed the other methods, and the 10 features it selected were GC content, score, cdsStop, cdsSize, and T, C, GT, GC, ACG, and TAT frequencies. The experiments analyzed the distribution of ncRNAs and coding RNAs on the dataset for these 10 features, and based on Figure 2, it can be seen that they play a significant role in the identification of discriminatory power. In addition, we conducted a correlation analysis between the ten features selected for the classification task. Figure 3 showed that, GC content had a weak correlation with the other features. Score, cdsStop, and cdsSize showed a stronger correlation with the other features. T, C, GT, GC, ACG, and TAT frequencies had the strongest correlation with the other features.

2.3. Comparison of Models

Regarding the validation set, in this study, we compared the five-fold cross-validation results of four AutoML frameworks, AutoGluon, TPOT, H2O, and AutoKeras, to those of three conventional machine learning models, i.e., random forest, SVM, and Naive Bayes (Table 2). It is evident that, in general, conventional machine learning models are less effective than AutoML; three of the four automated machine learning frameworks produced more effective models than the random forest, the best performing conventional machine learning model. AutoGluon achieved the best results for five of the eight evaluation metrics within the AutoML framework: ACC, F1, MCC, NPV, and SE. H2O achieved the best results for AUC, while Autokeras achieved the best results for PPV and SPC. It is evident that the AutoGluon framework is more effective than the other frameworks, possibly because AutoGluon employs per-variable embedding, which improves quality via gradient flow, whereas the other frameworks merely apply the standard feed-forward architecture to hot-coded data. The Autokeras effect, which is based on NAS that combines multiple search strategies such as random search, grid search, etc., is only marginally weaker than the AutoGluon effect. The goal of NAS is to reduce human intervention and to allow the algorithm to design the neural network automatically, which consists of three key components: the search space, the search strategy, and the evaluation strategy. However, this process is typically very time-consuming. H2O had the highest AUC score, but its overall performance was comparable to that of conventional machine learning models and TPOT. H2O is a distributed machine learning platform based on the Java programming language, unlike other AutoML frameworks. TPOT was the least effective AutoML and the only framework with overall lower results than conventional machine learning models. This is likely due to the genetic algorithm employed by TPOT, which tends to converge on a locally optimal solution prematurely. Consequently, the comparison demonstrates that the models created by the AutoGluon framework are superior to those created by the other four automatic machine learning frameworks and the three conventional machine learning models.

2.4. Comparison Tools against Plant Datasets

To evaluate the accuracy of PINC in ncRNA and coding RNA identification, we compared it to CPC2, CPAT, CNIT, and CPPred. We compared the identification accuracy for nine plant species from four databases, GreeNC, CANTATA, RNAcentral, and Phytozome, using five different tools. It is evident from the results shown in Figure 4 that our tool has the highest degree of precision for all nine plants. The large fluctuation of CPPred indicates that it has poor generalization performance, whereas the other three tools have some stability. However, it can be seen that the identification accuracy of PINC is greater than that of the other three tools, indicating that our tool performs the best among the different plant species. To compare the performances of these five tools further, we used eight metrics: sensitivity (SE), specificity (SPC), accuracy (ACC), F1-score, PPV, NPV, MCC, and AUC to evaluate and compare the five tools for these nine independent test sets (Table 3). We plotted the ROC curve (Figure 5); it can be seen that the ROC curve for PINC differs from the other tools. A true positive rate is rapidly achieved (1.0) at the cost of a relatively high false positive rate. Therefore, we have also plotted PR curves (Figure 6) to further illustrate the performance of PINC. The results showed that the PR curve of PINC did not fluctuate markedly and had a decreasing trend when the threshold was greater than 0.8. Meanwhile, PR curves illustrated that Precision and Recall values of five plants (Cicer arietinum, Manihot esculenta, Nymphaea colorata, Sorghum bicolor, and Zea mays) were higher than the other tools at the same threshold. All those results showed that PINC had the superior performance for distinguishing ncRNAs from coding RNAs. Solanum tuberosum outperformed the other tools in seven of the eight evaluation metrics and at least five of the remaining eight test sets, namely, SE ACC, F1, NPV, and MCC. The high Se score indicates that the probability of missing is small; therefore, PINC is the best choice for ncRNA identification. For the specificity SPC score, only one dataset was higher than the other tools, with four datasets performing best on CNIT and two datasets performing best on CPC2 and CPAT, respectively. However, the difference between the SPC of PINC and the SPC of the other tools was not large, and all tools had high performances above 86.99%. Among the five tools, PINC was the most effective for ncRNA identification in the nine plants. This indicates that our tool has a strong generalization to plants, which is crucial for non-model plants.

3. Discussion

In the field of bioinformatics, automated machine learning methods are now beginning to be implemented. In our experiments, we compared four automatic machine learning frameworks that are good matches for the more recently introduced frameworks and the older frameworks. For all the automatic machine learning frameworks, we used the same preprocessing methods to process the data as a raw input, then, we adjust the parameters of each framework in order to find the most suitable parameters, and finally we output the model. In general, we consider automatic machine learning frameworks to be black boxes and do not examine frame-specific methods for automatically optimizing parameters and integrating the model for direct output. Automated machine learning frameworks automatically optimize models, thereby reducing the time and effort devoted by researchers and, to a certain extent, allowing non-experts in machine learning to solve bioinformatics problems.
Utilizing high-quality features is one way to improve performance in machine learning. It is necessary to find features that are suitable for ncRNA identification in the study because providing or discovering good features is one of the most important tasks in machine learning. We extracted k-mer frequency features, coding sequence features, and other features during our experiments. Despite the fact that traditional k-mer features have been used in a variety of studies, such as gene identification [33], subcellular localization [34], and sequence analysis, it has been demonstrated that the k-mer frequency is highly effective at detecting ncRNAs [35]. Many tools have also used features related to coding sequences and some other features [36]. Ninety-one extracted features were filtered using our feature selection method; the filtered features successfully identified ncRNAs and it was the most precise tool, to date, for ncRNA identification in all plant species.
For ncRNA identification, there are additional factors to consider, such as the trade-off between sensitivity and specificity. At present, the number of ncRNAs is small as compared with the number of coding RNAs identified. To prevent ncRNAs from being missed, high sensitivity is important. Currently, CPAT, CNCI, CPPred, and CPC2 are less sensitive and focus more on identifying coding RNAs, but this requires an additional step to screen for non-coding RNAs. In contrast, the high sensitivity of PINC reduces the necessity for additional filtering processes. Moreover, PINC demonstrated a higher rate of accuracy than any other tool among the nine plants evaluated. Although some tools for non-coding RNA identification have reached over 85 percent accuracy, increased accuracy is not meaningless, as large amounts of data have become available due to advances in sequencing technology, and it is possible that for every one percent increase, hundreds of additional correct RNAs can be identified. Here, PINC achieves a high degree of ncRNA identification precision. This may be because the model in PINC adopts the stacking strategy, while other tools use single models such as SVM, logistic regression, and xgboost. For a long time, the performance of combining the predicted results of multiple models has been better than that of a single model, and the variance has been significantly reduced [37]. In the experiment, we selected the default basic model in the AutoGluon framework. Here, the basic model is trained separately, and then the prediction of the basic model is used as a feature to train the stacked model. Stacked models can improve the shortcomings of a single-model prediction and can take advantage of their interactions to improve the prediction ability [38]. In addition, it can be seen from the feature level distribution map described earlier that these features also have strong discrimination ability.
In addition, we plan to continue research in two areas: first, deep learning, which can automatically extract features, reduce the time required to extract features, and can improve the accuracy of cross-species recognition. In contrast, we should consider machine learning techniques to gain a deeper understanding of these RNA types and to investigate their biological significance. In addition, for plants, only a handful of ncRNA functions have been identified; once these functions are identified, new mechanisms can be explored and new features can be added to PINC to improve our tool further.

4. Materials and Methods

Figure 7 depicts the tool’s overall workflow, which consists of three steps: (1) dataset construction, (2) feature extraction and selection, and (3) model construction.
To create the dataset, RNA sequences were obtained from the GreeNC, CANTATA, RNAcentral, and Phytozome databases. Secondly, feature selection methods were used to extract and filter features. Finally, machine learning models were compared to determine the most effective model for ncRNA identification.

4.1. Dataset Construction

To construct the experimental dataset, we considered two factors. On the one hand, the diversity of plants and the abundance of annotation data were taken into consideration. On the other hand, considering the balance of the data, we chose four plants as our training and validation datasets (Table 4), which included two model plants, i.e., Arabidopsis thaliana and Oryza sativa, in addition to two non-model plants, i.e., Glycine max and Vitis vinifera. We used ncRNAs as the positive sample data and coding RNAs as the negative sample data in the dataset. Negative samples were obtained from Phytozome.v13 [39]. Positive samples were obtained from three public databases, including GreeNC [40], CANTATA [41], and RNAcentral [42]. For all data, first, we used cd-hit-est-2D in the CD-hit tool [43] to eliminate redundant sequences between the test and training sets at a threshold of 80% [22,35,44,45]. Second, in order to balance the datasets, random selections of 4000 data were made for each plant, of which 2000 were positive samples and 2000 were negative samples. The positive sample data consisted of 1800 lncRNAs and 200 sncRNAs, and the negative sample data consisted of 2000 mRNAs (Table 5) [18,46]. Thus, the baseline dataset consisted of a total of 16,000 protein sequences from four plants. Meanwhile, we analyzed the length distribution of the positive and negative datasets, as shown in Figure 8. The median length of the coding RNAs data was 1029 and the data were mostly concentrated in the range of 0–2000. The ncRNA data had a median length of 321 and the data were mostly concentrated in the range of 0–1000. Finally, we proportionally divided the dataset into 70% training data and 30% validation data. Additionally, nine independent test sets were created for nine plants. (Table 6): Cicer arietinum, Gossypium darwinii, Lactuca sativa, Manihot esculenta, Musa acuminata, Nymphaea colorata, Solanum tuberosum, Sorghum bicolor, and Zea mays. To eliminate redundant sequences, the data for these nine independent test sets were taken from the four databases mentioned above and filtered at a threshold of 80%.

4.2. Feature Extraction and Selection

This experiment initially extracted 91 features (Table 7). The 86 features of k-mer frequency, sequence length, and GC content were obtained using the Python script program (https://github.com/midisec/PINC, accessed on 22 August 2022); the five features of Score and CDS were obtained using the UCSC Genome txCdsPredict program in the browser (http://hgdown-load.soe.ucsc.edu/admin/jksrc.zip, accessed on 11 November 2014) [47]. These features can be classified into three categories: k-mer frequency features, CDS-related features, and other features. The k-mer frequency describes all possible frequencies for the presence of k nucleotides in a sequence, based on methods that have initially been implemented in whole genome shotgun assemblers. When k = 1, each nucleotide can contain a maximum of four A, C, G, or T. When k equals 2, the calculation involves the dinucleotide frequency (i.e., AA, AT, AG, AC, …, TT) and a total of 4 2 = 16 species. When k = 3, the calculated three-nucleotide frequencies (i.e., AAA, AAT, AAG, AAC, …, TTT) are computed for a total of 4 3 = 64 species. By combining 1–3-mer frequencies for a total of 84 features, k-mer frequencies can capture rich statistical information about negative profiles in plant genomes, according to some research [48]. CDS is the result of encoded proteins that are interchangeable with ORF in some ways, but differ slightly [49]. The features Score, cdsStarts, cdsStop, cdsSize, and cdsPercent comprise the second major category of features. Score is the predicted protein score; if it is >800, there is a 90% chance that it is a protein, and if it is >1000, it is virtually certain that it is a protein. cdsStop is the end of the coding region in the transcript, cdsSize is cdsStop minus cdsStart, and cdsPercent is the ratio of cdsSize to the total sequence length. Other features include sequence length and GC content, which are widely used for ncRNA identification. Sequence length indicates the total length of the sequence. GC content is the ratio of guanine and cytosine to the other four DNA bases.
There may be redundant features among the 91 features listed above; therefore, we employed feature selection to filter them. For the feature selection method, redundant features were filtered out using a combination of variance threshold filtering and the F-test. Variance threshold filtering is used to filter features based on their own variance. The smaller a feature’s variance, the less significant its variation, and these insignificant features are eliminated. F-test is a method to determine the relationship between each feature and label. The GC content, Score, cdsStop, cdsSize, and T, C, GT, GC, ACG, and TAT frequencies were among the 91 features identified by this combined feature selection method. Finally, these 10 features were used as the model input.

4.3. Model Construction

Machine learning (ML) is currently utilized in a variety of fields to solve numerous difficult problems. Nevertheless, model construction for machine learning requires human intervention. Manual intervention is required in the feature extraction, model selection, and parameter adjustment processes, which require professionals to optimize and can waste a significant amount of time and resources if errors occur. To reduce these repetitive development costs, the concept of automating the entire machine learning process, automatic machine learning, has been conceived (AutoML). The definition of AutoML is that it is a combination of automation and ML [50]. From an automation standpoint, AutoML can be viewed as the design of a framework to automate the entire machine learning process, allowing models to automatically learn the correct parameters and configurations without manual intervention. From the standpoint of machine learning, AutoML is a system that is highly capable of learning and generalizing given data and tasks. Recent research on AutoML has focused on the neural network architecture search (NAS) method, which employs a search strategy to test and evaluate a large number of architectures in a search space, and then selects the one that best meets the objectives of a given problem by maximizing the adaptation function. However, the NAS faces two obstacles to the method: first, the amount of computation is excessive, resulting in increased resource consumption. Second, instability may vary each time and the search structure is altered, resulting in varying precision. In our experiments, we compared four automatic machine learning frameworks, AutoGluon, H2O, TPOT, and Autokeras, with three conventional machine learning models, SVM, RF, and Naive Bayes. We determined that AutoGluon was the superior framework, and therefore it was used as the classifier. AutoGluon contains 26 base models including random forest, XGBoost, and a neural network, and in our experiments, we used all the base models for training the model [51]. AutoGluon is an open-source machine learning training framework for tabular data. It is a framework that attempts to avoid a hyperparametric search as much as possible, training multiple models concurrently and weighting them using a multi-layer stacking strategy to obtain the final output.

4.4. Performance Evaluation

Several widely used performance metrics were evaluated in the experiments, including sensitivity (SE), specificity (SPC), accuracy (ACC), F1-score, positive predictive value (PPV), negative predictive value (NPV), and the Matthews correlation coefficient (MCC). To evaluate the performance of the classifier numerically and visually, the area under the curve (AUC) and ROC curves were also used. These definitions are as follows:
S e n s i t i v i t y ( S E ) = T P T P + F N S p e c i f i c i t y ( S P C ) = T N T N + F P A c c u r a c y ( A C C ) = T P + T N T P + F N + F P + T N F 1 = 2 × T P 2 × T P + F P + F N P P V = T P T P + F P N P V = T N T N + F N M C C = T P × T N F P × F N ( T P + F N ) × ( T P + F P ) × ( T N + F P ) × ( T N + F N )
TP represents true positives, the number of correctly identified positive samples, while FN, TN, and FP represent false negatives, true negatives, and false positives, the number of incorrectly identified positive samples, correctly identified negative samples, and incorrectly identified negative samples, respectively.

5. Conclusions

Various tools have been developed to distinguish between ncRNAs and coding RNAs, the majority of which have used scientific computational methods to differentiate sequences and to accelerate the annotation of various human genes. In addition to nucleotides with high discriminatory power in 1–3-mer, we also extracted other features such as the sequence’s definition, composition, and function. Moreover, we combined F-test and variance threshold filtering and found that the combined method was superior to the individual methods of F-test and variance threshold filtering. A number of automated machine learning and traditional machine learning frameworks were also used for modeling, in which the validation set was carefully evaluated and analyzed, including the use of cross-validation on the validation set available, with AutoGluon performing the best. Then, we compiled these into a tool called PINC and compared it to nine other tools on nine test sets, demonstrating that PINC performed better than other tools on all of these species. For user convenience, a user-friendly web (http://www.pncrna.com/, accessed on 22 August 2022) has been developed, where the output can be obtained simply by entering a FASTA sequence or file. Overall, PINC has excellent predictive properties, permits cross-species plant identification, and is a practical and user-friendly tool.

Author Contributions

Conceptualization, X.Z. (Xiaodan Zhang) and X.J.; methodology, X.Z. (Xiaohu Zhou); software, M.W.; validation, J.X., X.Z. (Xiaohu Zhou) and M.W.; resources, X.Z. (Xiaohu Zhou); data curation, X.Z. (Xiaodan Zhang); writing—original draft preparation, X.Z. (Xiaodan Zhang); writing—review and editing, X.Z. (Xiaodan Zhang); visualization, X.Z. (Xiaohu Zhou) and M.W.; supervision, S.L.; project administration, M.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Nature Science Research Project of Education Department in Anhui Province grant number KJ2020A0108.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available at https://github.com/midisec/PINC, accessed on 22 August 2022.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Eddy, S.R. Non-coding RNA genes and the modern RNA world. Nat. Rev. Genet. 2001, 2, 919–929. [Google Scholar] [CrossRef]
  2. Wang, K.C.; Chang, H.Y. Molecular mechanisms of long noncoding RNAs. Mol. Cell 2011, 43, 904–914. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Palazzo, A.F.; Lee, E.S. Non-coding RNA: What is functional and what is junk? Front. Genet. 2015, 6, 2. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Ponting, C.P.; Oliver, P.L.; Reik, W. Evolution and functions of long noncoding RNAs. Cell 2009, 136, 629–641. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Hannon, G.J. RNA interference. Nature 2002, 418, 244–251. [Google Scholar] [CrossRef]
  6. Jamalkandi, S.A.; Masoudi-Nejad, A. Reconstruction of Arabidopsis thaliana fully integrated small RNA pathway. Funct. Integr. Genom. 2009, 9, 419–432. [Google Scholar] [CrossRef]
  7. Morris, K.V.; Mattick, J.S. The rise of regulatory RNA. Nat. Rev. Genet. 2014, 15, 423–437. [Google Scholar] [CrossRef] [Green Version]
  8. Bazin, J.; Bailey-Serres, J. Emerging roles of long non-coding RNA in root developmental plasticity and regulation of phosphate homeostasis. Front. Plant Sci. 2015, 6, 400. [Google Scholar] [CrossRef] [Green Version]
  9. Wang, D.; Qu, Z.; Yang, L.; Zhang, Q.; Liu, Z.H.; Do, T.; Adelson, D.L.; Wang, Z.Y.; Searle, I.; Zhu, J.K. Transposable elements (TE s) contribute to stress-related long intergenic noncoding RNA s in plants. Plant J. 2017, 90, 133–146. [Google Scholar] [CrossRef] [Green Version]
  10. Yu, Y.; Zhang, Y.; Chen, X.; Chen, Y. Plant Noncoding RNAs: Hidden Players in Development and Stress Responses. In Annual Review of Cell and Developmental Biology; Lehmann, R., Ed.; Annual Reviews: Palo Alto, CA, USA, 2019; Volume 35, pp. 407–431. [Google Scholar]
  11. Woo, H.R.; Kim, H.J.; Nam, H.G.; Lim, P.O. Plant leaf senescence and death—Regulation by multiple layers of control and implications for aging in general. J. Cell Sci. 2013, 126, 4823–4833. [Google Scholar] [CrossRef]
  12. Li, Y.; Cao, X.-L.; Zhu, Y.; Yang, X.-M.; Zhang, K.-N.; Xiao, Z.-Y.; Wang, H.; Zhao, J.-H.; Zhang, L.-L.; Li, G.-B.; et al. Osa-miR398b boosts H2O2 production and rice blast disease-resistance via multiple superoxide dismutases. New Phytol. 2019, 222, 1507–1522. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Li, Y.; Lu, Y.-G.; Shi, Y.; Wu, L.; Xu, Y.-J.; Huang, F.; Guo, X.-Y.; Zhang, Y.; Fan, J.; Zhao, J.-Q.; et al. Multiple Rice MicroRNAs Are Involved in Immunity against the Blast Fungus Magnaporthe oryzae. Plant Physiol. 2014, 164, 1077–1092. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Cheng, L.; Hu, Y.; Sun, J.; Zhou, M.; Jiang, Q. DincRNA: A comprehensive web-based bioinformatics toolkit for exploring disease associations and ncRNA function. Bioinformatics 2018, 34, 1953–1956. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Noviello, T.M.R.; Ceccarelli, F.; Ceccarelli, M.; Cerulo, L. Deep learning predicts short non-coding RNA functions from only raw sequence data. PLoS Comput. Biol. 2020, 16, 1008451. [Google Scholar] [CrossRef]
  16. Huttenhofer, A.; Vogel, J. Experimental approaches to identify non-coding RNAs. Nucleic Acids Res. 2006, 34, 635–646. [Google Scholar] [CrossRef] [Green Version]
  17. Kong, L.; Zhang, Y.; Ye, Z.-Q.; Liu, X.-Q.; Zhao, S.-Q.; Wei, L.; Gao, G. CPC: Assess the protein-coding potential of transcripts using sequence features and support vector machine. Nucleic Acids Res. 2007, 35 (Suppl. S2), W345–W349. [Google Scholar] [CrossRef]
  18. Kang, Y.-J.; Yang, D.-C.; Kong, L.; Hou, M.; Meng, Y.-Q.; Wei, L.; Gao, G. CPC2: A fast and accurate coding potential calculator based on sequence intrinsic features. Nucleic Acids Res. 2017, 45, W12–W16. [Google Scholar] [CrossRef] [Green Version]
  19. Wang, L.; Park, H.J.; Dasari, S.; Wang, S.; Kocher, J.-P.; Li, W. CPAT: Coding-Potential Assessment Tool using an alignment-free logistic regression model. Nucleic Acids Res. 2013, 41, e74. [Google Scholar] [CrossRef]
  20. Sun, L.; Luo, H.; Bu, D.; Zhao, G.; Yu, K.; Zhang, C.; Liu, Y.; Chen, R.; Zhao, Y. Utilizing sequence intrinsic composition to classify protein-coding and long non-coding transcripts. Nucleic Acids Res. 2013, 41, e166. [Google Scholar] [CrossRef]
  21. Guo, J.-C.; Fang, S.-S.; Wu, Y.; Zhang, J.-H.; Chen, Y.; Liu, J.; Wu, B.; Wu, J.-R.; Li, E.-M.; Xu, L.-Y. CNIT: A fast and accurate web tool for identifying protein-coding and long non-coding transcripts based on intrinsic sequence composition. Nucleic Acids Res. 2019, 47, W516–W522. [Google Scholar] [CrossRef]
  22. Tong, X.; Liu, S. CPPred: Coding potential prediction based on the global description of RNA sequence. Nucleic Acids Res. 2019, 47, e43. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Zhang, H.; He, X.; Zhu, J.-K. RNA-directed DNA methylation in plants: Where to start? RNA Biol. 2013, 10, 1593–1596. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Schneider, H.W.; Raiol, T.; Brigido, M.M.; Walter, M.E.M.; Stadler, P.F. A support vector machine based method to distinguish long non-coding RNAs from protein coding transcripts. BMC Genom. 2017, 18, 804. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Thornton, C.; Hutter, F.; Hoos, H.H.; Leyton-Brown, K. Auto-WEKA: Combined selection and hyperparameter optimization of classification algorithms. In Proceedings of the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Chicago, IL, USA, 11–14 August 2013; pp. 847–855. [Google Scholar]
  26. LeDell, E.; Poirier, S. H2o automl: Scalable automatic machine learning. In Proceedings of the 7th ICML Workshop on Automated Machine Learning (2020), Vienna, Austria, 17–18 July 2020. [Google Scholar]
  27. Olson, R.S.; Moore, J.H. TPOT: A tree-based pipeline optimization tool for automating machine learning. In Workshop on Automatic Machine Learning; Springer: Berlin/Heidelberg, Germany, 2016; pp. 66–74. [Google Scholar]
  28. Erickson, N.; Mueller, J.; Shirkov, A.; Zhang, H.; Larroy, P.; Li, M.; Smola, A. Autogluon-tabular: Robust and accurate automl for structured data. arXiv 2020, arXiv:2003.06505. [Google Scholar]
  29. Jin, H.; Song, Q.; Hu, X. Auto-keras: An efficient neural architecture search system. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; Association for Computing Machinery: New York, NY, USA; pp. 1946–1956. [Google Scholar]
  30. Karaglani, M.; Gourlia, K.; Tsamardinos, I.; Chatzaki, E. Accurate blood-based diagnostic biosignatures for Alzheimer’s disease via automated machine learning. J. Clin. Med. 2020, 9, 3016. [Google Scholar] [CrossRef]
  31. Le, T.T.; Fu, W.; Moore, J.H. Scaling tree-based automated machine learning to biomedical big data with a feature set selector. Bioinformatics 2020, 36, 250–256. [Google Scholar] [CrossRef] [Green Version]
  32. Xiang, J.S.; Kaplan, M.; Dykstra, P.; Hinks, M.; McKeague, M.; Smolke, C.D. Massively parallel RNA device engineering in mammalian cells with RNA-Seq. Nat. Commun. 2019, 10, 4327. [Google Scholar] [CrossRef] [Green Version]
  33. Liu, Y.; Guo, J.; Hu, G.; Zhu, H. Gene Prediction in metagenomic fragments based on the SVM algorith. In BMC Bioinformatics; Springer: Berlin/Heidelberg, Germany, 2013; pp. 1–12. [Google Scholar]
  34. Zeng, M.; Wu, Y.; Lu, C.; Zhang, F.; Wu, F.-X.; Li, M. DeepLncLoc: A deep learning framework for long non-coding RNA subcellular localization prediction based on subsequence embedding. Brief. Bioinform. 2022, 23, bbab360. [Google Scholar] [CrossRef]
  35. Li, A.; Zhang, J.; Zhou, Z. PLEK: A tool for predicting long non-coding RNAs and messenger RNAs based on an improved k-mer scheme. BMC Bioinform. 2014, 15, 311. [Google Scholar] [CrossRef] [Green Version]
  36. Negri, T.d.C.; Alves, W.A.L.; Bugatti, P.H.; Saito, P.T.M.; Domingues, D.S.; Paschoal, A.R. Pattern recognition analysis on long noncoding RNAs: A tool for prediction in plants. Brief. Bioinform. 2019, 20, 682–689. [Google Scholar] [CrossRef]
  37. Dietterich, T.G. Ensemble methods in machine learning. In International Workshop on Multiple Classifier Systems; Springer: Berlin/Heidelberg, Germany, 2000; pp. 1–15. [Google Scholar]
  38. Van der Laan, M.J.; Polley, E.C.; Hubbard, A.E. Super learner. Stat. Appl. Genet. Mol. Biol. 2007, 6, 1309. [Google Scholar] [CrossRef] [PubMed]
  39. Goodstein, D.M.; Shu, S.; Howson, R.; Neupane, R.; Hayes, R.D.; Fazo, J.; Mitros, T.; Dirks, W.; Hellsten, U.; Putnam, N. Phytozome: A comparative platform for green plant genomics. Nucleic Acids Res. 2012, 40, D1178–D1186. [Google Scholar] [CrossRef] [PubMed]
  40. Paytuví Gallart, A.; Hermoso Pulido, A.; de Anzar Martínez Lagrán, I.; Sanseverino, W.; Aiese Cigliano, R. GREENC: A Wiki-based database of plant lncRNAs. Nucleic Acids Res. 2016, 44, D1161–D1166. [Google Scholar] [CrossRef] [PubMed]
  41. Szczesniak, M.W.; Bryzghalov, O.; Ciomborowska-Basheer, J.; Makalowska, I. CANTATAdb 2.0: Expanding the Collection of Plant Long Noncoding RNAs. In Methods in Molecular Biology (Clifton, N.J.); Humana Press: New York, NY, USA, 2019; pp. 415–429. [Google Scholar]
  42. Petrov, A.I.; Kay, S.J.E.; Kalvari, I.; Howe, K.L.; Gray, K.A.; Bruford, E.A.; Kersey, P.J.; Cochrane, G.; Finn, R.D.; Bateman, A.; et al. RNAcentral: A comprehensive database of non-coding RNA sequences. Nucleic Acids Res. 2017, 45, D128–D134. [Google Scholar] [PubMed]
  43. Li, W.; Godzik, A. Cd-hit: A fast program for clustering and comparing large sets of protein or nucleotide sequences. Bioinformatics 2006, 22, 1658–1659. [Google Scholar] [CrossRef] [Green Version]
  44. Lertampaiporn, S.; Thammarongtham, C.; Nukoolkit, C.; Kaewkamnerdpong, B.; Ruengjitchatchawalya, M. Identification of non-coding RNAs with a new composite feature in the Hybrid Random Forest Ensemble algorithm. Nucleic Acids Res. 2014, 42, e93. [Google Scholar] [CrossRef]
  45. Sun, L.; Liu, H.; Zhang, L.; Meng, J. lncRScan-SVM: A tool for predicting long non-coding RNAs using support vector machine. PLoS ONE 2015, 10, e0139654. [Google Scholar] [CrossRef]
  46. Ito, E.A.; Katahira, I.; Vicente, F.F.d.R.; Pereira, L.F.P.; Lopes, F.M. BASiNET—BiologicAl Sequences NETwork: A case study on coding and non-coding RNAs identification. Nucleic Acids Res. 2018, 46, e96. [Google Scholar] [CrossRef]
  47. Haeussler, M.; Zweig, A.S.; Tyner, C.; Speir, M.L.; Rosenbloom, K.R.; Raney, B.J.; Lee, C.M.; Lee, B.T.; Hinrichs, A.S.; Gonzalez, J.N. The UCSC genome browser database: 2019 update. Nucleic Acids Res. 2019, 47, D853–D858. [Google Scholar] [CrossRef] [Green Version]
  48. Kurtz, S.; Narechania, A.; Stein, J.C.; Ware, D. A new method to compute K-mer frequencies and its application to annotate large repetitive plant genomes. BMC Genom. 2008, 9, 517. [Google Scholar] [CrossRef] [Green Version]
  49. Park, S.-G.; Kim, K.-B. ORF Miner: A web-based ORF search tool. Genom. Inform. 2009, 7, 217–219. [Google Scholar] [CrossRef]
  50. Yao, Q.; Wang, M.; Chen, Y.; Dai, W.; Li, Y.-F.; Tu, W.-W.; Yang, Q.; Yu, Y. Taking human out of learning applications: A survey on automated machine learning. arXiv 2018, arXiv:1810.13306. [Google Scholar]
  51. Xiao, S.; Tian, H.; Tao, P. PASSer2. 0: Accurate Prediction of Protein Allosteric Sites Through Automated Machine Learning. Front. Mol. Biosci. 2022, 9, 879251. [Google Scholar] [CrossRef] [PubMed]
Figure 1. (A) Graph showing the accuracy of 100 experiments; (B) graph showing the average accuracy of every 5th experiment out of 100 experiments.
Figure 1. (A) Graph showing the accuracy of 100 experiments; (B) graph showing the average accuracy of every 5th experiment out of 100 experiments.
Ijms 23 11825 g001
Figure 2. Differential distribution of ten features in coding RNAs and ncRNAs.
Figure 2. Differential distribution of ten features in coding RNAs and ncRNAs.
Ijms 23 11825 g002
Figure 3. Correlation analysis chart of 10 features selected for the classification task.
Figure 3. Correlation analysis chart of 10 features selected for the classification task.
Ijms 23 11825 g003
Figure 4. Comparing the identification accuracy of nine independent test sets across five tools.
Figure 4. Comparing the identification accuracy of nine independent test sets across five tools.
Ijms 23 11825 g004
Figure 5. ROC curves for 5 tools on 9 plants.
Figure 5. ROC curves for 5 tools on 9 plants.
Ijms 23 11825 g005
Figure 6. The PR curves obtained by PINC and four existing tools.
Figure 6. The PR curves obtained by PINC and four existing tools.
Ijms 23 11825 g006
Figure 7. Overall workflow: (A) dataset construction: a dataset was constructed using four species of plants together for training and validation, and nine independent test sets were constructed for testing; (B) feature extraction: features were extracted from the original sequence species and redundant features were filtered out using feature selection methods; (C) model construction: a stacking strategy was used to integrate multiple models.
Figure 7. Overall workflow: (A) dataset construction: a dataset was constructed using four species of plants together for training and validation, and nine independent test sets were constructed for testing; (B) feature extraction: features were extracted from the original sequence species and redundant features were filtered out using feature selection methods; (C) model construction: a stacking strategy was used to integrate multiple models.
Ijms 23 11825 g007
Figure 8. Distribution of positive and negative sample lengths in the benchmark dataset.
Figure 8. Distribution of positive and negative sample lengths in the benchmark dataset.
Ijms 23 11825 g008
Table 1. Comparing the performance of different feature selection methods.
Table 1. Comparing the performance of different feature selection methods.
SESPCACCMCCF1
F-test99.4990.1894.7789.9594.93
VT99.4989.2494.2989.0894.49
VT-F99.7990.795.2790.9195.49
RF90.3596.5293.2786.7693.39
RF-AutoGluon99.4590.1494.7289.8794.89
Table 2. Performance comparisons among five automated machine learning frameworks and three conventional machine learning models.
Table 2. Performance comparisons among five automated machine learning frameworks and three conventional machine learning models.
ModelACC (%)F1 (%)AUC (%)MCC (%)NPV (%)PPV (%)SE (%)SPC (%)
AutoGluon95.2595.4995.2590.9199.7691.5599.7990.70
Naive Bayes86.9787.9686.9374.8379.2294.6582.1693.63
SVM53.1413.6753.3716.9399.330.0791.6551.51
RFC92.1092.2692.0984.2690.4593.7390.8693.47
H2O92.9893.3896.8086.6086.9898.9888.3898.84
TPOT86.1486.1986.1572.2986.1886.1086.2886.00
Autokeras93.7094.0694.5787.9588.1099.2589.3999.15
Table 3. Nine plants’ performance indicators were compared using five tools.
Table 3. Nine plants’ performance indicators were compared using five tools.
SpeciesToolSE (%)SPC (%)ACC (%)F1 (%)PPV (%)NPV (%)MCC (%)AUC (%)
Cicer arietinumPINC98.7692.4696.4297.2095.7097.7292.3495.61
CPC276.0192.9184.4583.0491.5079.4269.9296.50
CPAT89.2788.7589.0189.0588.8489.1878.0296.26
CNIT65.6594.6780.1076.8192.5473.2362.9994.36
CPPred71.2487.7079.4677.6585.3275.2459.7589.72
Gossypium darwiniiPINC98.6186.8492.7493.1688.2998.4186.0892.72
CPC285.2590.6287.9487.6090.0986.0075.9995.23
CPAT95.0384.0289.5390.0785.6194.4279.5493.65
CNIT63.2390.7376.9873.3187.2171.1656.1390.67
CPPred80.7886.5983.8982.3684.0083.8067.5991.66
Lactuca sativaPINC98.8087.0092.9993.4788.6898.6086.5492.90
CPC270.1490.9680.5678.3088.5875.3062.4993.56
CPAT87.2482.5084.8485.0382.9486.9069.7992.00
CNIT51.9592.2472.0365.0887.0965.5948.2689.79
CPPred64.3984.0674.2371.4280.1670.2449.4284.12
Manihot esculentaPINC99.8287.1294.4595.4091.3699.7288.9993.47
CPC287.8285.1486.4886.6685.5387.4872.9992.15
CPAT93.5581.7387.6488.3383.6692.6875.8191.13
CNIT62.7386.1874.4671.0681.9469.8250.3391.30
CPPred87.580.0884.7985.1983.0086.7869.6888.97
Musa acuminataPINC99.2287.2293.2393.6188.6199.1187.0893.22
CPC290.1288.9989.6290.691.0987.8379.0294.71
CPAT71.6988.8280.2578.4186.5475.7961.4291.84
CNIT65.0888.2876.6673.6384.7771.654.8590.14
CPPred76.4486.181.2780.3384.6478.4862.8489.24
Nymphaea colorataPINC98.8291.9495.4495.6692.6998.7091.0895.38
CPC282.6993.8588.2887.5793.0684.4577.0397.08
CPAT82.991.3987.1486.5790.5984.2474.5695.10
CNIT55.2192.3873.7967.8187.8867.3451.2792.24
CPPred84.8986.5985.7485.6286.3685.1471.4991.56
Solanum tuberosumPINC99.7384.5392.9994.0689.0099.6086.4192.13
CPC267.2386.9977.1174.6083.7972.6355.3190.61
CPAT86.6978.6182.6583.3180.1885.5465.5189.47
CNIT58.7688.4973.6269.0283.6368.2049.4988.12
CPPred60.6481.7571.2067.8076.8767.5043.3881.24
Sorghum bicolorPINC99.987.6995.7996.9294.1199.7990.6993.79
CPC294.3887.3290.8591.1688.1693.9581.9196.42
CPAT86.6588.7187.6887.5588.4686.9375.3895.71
CNIT75.0485.3480.1979.1083.6377.4060.7192.89
CPPred91.8185.0488.4288.8085.9891.2177.0294.28
Zea maysPINC99.7190.3896.3097.1694.7499.4592.1295.04
CPC290.8190.8890.8590.8490.8790.8281.7096.63
CPAT76.5291.3783.9682.6489.8279.6368.6795.07
CNIT65.2490.1077.6974.4986.7872.2457.1592.50
CPPred84.8387.9286.3886.1687.5485.2972.8093.05
Table 4. Training set data for the model.
Table 4. Training set data for the model.
SpeciesNoncoding Coding
TotalUsedTotalUsed
Arabidopsis thaliana45,910200027,4162000
Glycine max8599200071,3582000
Oryza sativa11,338200042,1892000
Vitis vinifera4301200055,5642000
Total70,1488000196,5278000
Table 5. Detailed description of the training set data.
Table 5. Detailed description of the training set data.
Size
Non-coding RNAsLong ncRNAs1800
Small ncRNAs200
Coding RNAsmRNAs2000
Overall 4000
Table 6. Plant dataset for testing.
Table 6. Plant dataset for testing.
SpeciesCodingNoncodingTotal
Cicer arietinum209920994198
Gossypium darwinii5622562211,244
Lactuca sativa468246829364
Manihot esculenta280828085616
Musa acuminata205920634122
Nymphaea colorata170817083416
Solanum tuberosum8282828216,564
Sorghum bicolor8657865717,314
Zea mays7406740614,812
Table 7. All features considered in this paper.
Table 7. All features considered in this paper.
FeaturesDescriptionSource
k-mer frequency1–3 k-mer = 84PINC
1 nt = 4 features; 2 nt = 16 features
3 nt = 64 features
ScoreValues >800 are likely to be a protein, >1000 must be proteintxCdsPredict
cdsStartsNT position of CDS starts from the transcript and is based on zerotxCdsPredict
cdsStopnt position for the CDS endtxCdsPredict
cdsSizescdsStop-cdsStarttxCdsPredict
cdsPercent(cdsStop + cdsStart)/total nt sequence sizetxCdsPredict
Sequence lengthTotal nucleotide length of the sequencePINC
GC content C + G A + C + G + T PINC
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, X.; Zhou, X.; Wan, M.; Xuan, J.; Jin, X.; Li, S. PINC: A Tool for Non-Coding RNA Identification in Plants Based on an Automated Machine Learning Framework. Int. J. Mol. Sci. 2022, 23, 11825. https://doi.org/10.3390/ijms231911825

AMA Style

Zhang X, Zhou X, Wan M, Xuan J, Jin X, Li S. PINC: A Tool for Non-Coding RNA Identification in Plants Based on an Automated Machine Learning Framework. International Journal of Molecular Sciences. 2022; 23(19):11825. https://doi.org/10.3390/ijms231911825

Chicago/Turabian Style

Zhang, Xiaodan, Xiaohu Zhou, Midi Wan, Jinxiang Xuan, Xiu Jin, and Shaowen Li. 2022. "PINC: A Tool for Non-Coding RNA Identification in Plants Based on an Automated Machine Learning Framework" International Journal of Molecular Sciences 23, no. 19: 11825. https://doi.org/10.3390/ijms231911825

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop