Next Article in Journal
Updating Our Knowledge on Fish Immunology
Previous Article in Journal
Comparative Analysis of Water Extracts from Roselle (Hibiscus sabdariffa L.) Plants and Callus Cells: Constituents, Effects on Human Skin Cells, and Transcriptome Profiles
Previous Article in Special Issue
In Silico Exploration of Metabolically Active Peptides as Potential Therapeutic Agents against Amyotrophic Lateral Sclerosis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using the Random Forest for Identifying Key Physicochemical Properties of Amino Acids to Discriminate Anticancer and Non-Anticancer Peptides

1
College of Biomedical Engineering, Sichuan University, Chengdu 610065, China
2
College of Life Science, Sichuan University, Chengdu 610065, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Int. J. Mol. Sci. 2023, 24(13), 10854; https://doi.org/10.3390/ijms241310854
Submission received: 8 May 2023 / Revised: 17 June 2023 / Accepted: 26 June 2023 / Published: 29 June 2023
(This article belongs to the Special Issue Computer-Aided Screening and Action Mechanism of Bioactive Peptides)

Abstract

:
Anticancer peptides (ACPs) represent a promising new therapeutic approach in cancer treatment. They can target cancer cells without affecting healthy tissues or altering normal physiological functions. Machine learning algorithms have increasingly been utilized for predicting peptide sequences with potential ACP effects. This study analyzed four benchmark datasets based on a well-established random forest (RF) algorithm. The peptide sequences were converted into 566 physicochemical features extracted from the amino acid index (AAindex) library, which were then subjected to feature selection using four methods: light gradient-boosting machine (LGBM), analysis of variance (ANOVA), chi-squared test (Chi2), and mutual information (MI). Presenting and merging the identified features using Venn diagrams, 19 key amino acid physicochemical properties were identified that can be used to predict the likelihood of a peptide sequence functioning as an ACP. The results were quantified by performance evaluation metrics to determine the accuracy of predictions. This study aims to enhance the efficiency of designing peptide sequences for cancer treatment.

1. Introduction

Identifying novel anticancer compounds has long been a major focus of medical research. Conventional cancer treatments, such as surgery, radiotherapy, and chemotherapy, often adversely affect healthy cells and tissues and can lead to the development of treatment resistance. Therefore, it is essential to identify additional effective therapeutic options. Anticancer peptides (ACPs) are short peptide sequences that typically contain fewer than 50 amino acids. They exert their anticancer effects via a variety of mechanisms, for example, by inhibiting the proliferation and migration of cancer cells, inducing apoptosis, changing the pH in the internal and external cellular environment, or by damaging the cell membrane of cancer cells without affecting healthy tissues [1]. Compared to conventional therapeutic compounds, ACPs offer several advantages, such as high specificity, low intrinsic toxicity, high tissue permeability, and the convenience of sequence modification [2,3,4,5]. These features make ACPs a promising new therapeutic option in the management of cancer.
Historically, the identification of ACPs involved conventional laboratory experiments that were time-consuming and costly. However, with the accumulation of considerable ACP sequence data and the establishment of experimentally validated ACP databases, such as CancerPPD [2], DADP [6], CAMP [7], and APD [8], the rapid identification of novel ACP sequences using machine learning algorithms is becoming increasingly feasible. For example, in 2018, Wei et al. [9] used amino acid binary profile, amino acid type group, composition–transition–distribution, and twenty-one-bit features to represent peptides and then adopted a support vector machine learning method named ACPred_FL. In 2020, Rao et al. [10] developed another model named ACPred_Fuse, which applied 114 type features to represent peptides. In the same year, Agrawal et al. [11] used amino acid composition, dipeptide composition, terminus composition, and binary profile to develop an extra-tree-based model named AntiCP-2.0. With the rise in deep learning and protein representation learning, many new anticancer peptide recognition methods [11,12,13,14,15,16] (e.g., iACP-DRLF and TriNet) continue to emerge, and the performances of the methods are becoming increasingly better.
Nonetheless, current machine-learning-based ACP prediction models are limited. Analyzing standard datasets with alternative sequence feature extraction methods often results in greatly variable outcomes, and the causes of this divergence are currently unclear. In addition, studies to date have paid little attention to the key physicochemical features that differentiate ACPs from peptides with no ACP activity, resulting in an insufficient understanding of what features determine whether an amino acid sequence can act as an ACP. To overcome these shortcomings, we used the AAindex database [17,18,19,20] as the feature database of peptide sequences. This database consists of two subdatabases, AAindex1 and AAindex2. AAindex1 is a database representing the different physical, chemical, and biological properties of amino acids, currently containing 566 amino acid features. AAindex2 is an amino acid mutation matrix database that represents the similarity between amino acids, currently containing 94 matrices. The AAindex database is primarily used in protein-related research fields and as a machine learning database in protein prediction applications. However, the information contained within this database has not been previously utilized for extracting the key features of ACP sequences. Thus, no information is available on the physicochemical properties of key amino acids that affect the function of ACPs from a holistic perspective.
To address the problems of current ACP prediction models, we explored the physicochemical characteristics of amino acids predicting ACP activity. We combined multiple feature selection techniques and a random forest (RF) model to construct a single-feature model based on the physicochemical characteristics of ACP sequences [21,22,23]. The resulting model performed well in predicting ACPs in the calculation-based analysis of the physicochemical properties of key amino acids within the peptide sequences. This approach allowed us to identify 19 key amino acid properties that were useful in detecting ACP sequences in various benchmark datasets. A user-friendly webserver (https://www.aibiochem.net/servers/RFaaindexACP, accessed on 18 June 2023) is provided.

2. Results and Discussion

2.1. Analysis of Feature Selection Methods

Various feature selection methods can result in marked differences in the ranking of key features. Therefore, we compared the performance of the models using selected features with the baseline of non-selected 566-dimensional feature vector results.

2.1.1. Model Performance Analysis before and after Feature Selection

After establishing an RF model based on 566-dimensional features, the four studied datasets, ACPred-Fuse, ACPred-FL, ACP20Alt, and ACP20main, were each processed using four different feature selection methods: ANOVA, Chi2, LGBM, and MI. Table 1 displays the optimized feature space dimensions for different datasets. For instance, for the ACPred_Fuse dataset, the best feature numbers were 73, 77, 28, and 40 after ANOVA, Chi2, LGBM, and MI were applied. A histogram was created for each index, as well as the baseline metrics without feature selection, and the corresponding ACC, MCC, Sn, Sp, and AUC values were compared, as shown in Figure 1. Irrespective of the dataset being analyzed, the best performing feature selection was invariably achieved using LGBM. Therefore, feature selection and optimization were carried out using this approach.

2.1.2. Feature Selection Results for Different Datasets Using the Same Feature Selection Method

After completing the feature selection steps using the four datasets, the key features obtained were compared longitudinally to identify overlaps. As shown in Table 1, the four distinct feature selection strategies resulted in a marked variation in the number of features being selected.
Through data records, common features present in three or more datasets after each feature selection strategy were selected and analyzed using Venn diagrams, by intersecting the obtained features. These Venn diagrams, drawn using the Venny 2.1 online tool [24], are shown in Figure 2 (the tool is available at https://bioinfogp.cnb.csic.es/tools/venny/index.html, accessed on 1 January 2021).
In order to obtain universally applicable features and eliminate the influence of different feature selection methods, features obtained from the four datasets were intersected and merged. Since the number of features was relatively small, we only retained merged features common to at least three datasets. The number of common features obtained from the four feature selection methods were 27 features from ANOVA, 57 features from Chi2, 12 features from LGBM, and 25 features from MI. These were combined, obtaining a total of 105 features for the next round of feature selection.
The RF analysis results, using the 105 selected features, showed that the ACC, MCC, and AUC metrics for each dataset were almost optimal, with these metrics with three datasets being better than what could be achieved using the original, unselected data. However, when analyzing the ACP20Alt dataset, the performance of the selected features was slightly reduced, although the difference compared to the full 566-dimensional data was very small. Therefore, the selected 105 features appeared to be sufficient to identify ACPs in the analyzed datasets.

2.2. Model Analysis Based on 105 Selected Features

2.2.1. Model Performance

In order to reduce the number of features further and make the selection results more representative, a second round of feature selection was conducted, starting with the 105 features described above. This round of feature selection was performed using LGBM on each of the four datasets, respectively. The results of this analysis are shown in Table 2. As shown in Table 2, the model based on the ACPred-Fuse dataset has the best independent test accuracy, while the model based on ACP20main has the smallest. Moreover, it can be concluded from Table 2 that the feature space dimensions of the models constructed on the basis of different datasets are different to obtain the best independent test accuracy. This means that these features are not sufficiently representative of all datasets.

2.2.2. First Optimization Based on LGBM Feature Selection

To obtain a common feature set represented by the four datasets, feature importance analysis was performed on the features obtained from the four datasets. Among the identified features, those with importance values greater than 0.01 were selected from the intersecting areas of the data. The intersection of the four dataset feature spaces is shown in a Venn diagram in Figure 3. Only 19 features satisfied these requirements.

2.2.3. Second Optimization of the 19 Features by LGBM

To explore how well the 19 selected features represented the full data, we created RF models using all four datasets. The cross-validation and independent testing results are shown in Table 3. We presented the results as histograms for the three feature selections and compared the three most important metrics: ACC, MCC, and AUC. The result of these comparisons is summarized in Figure 4.
When the performance of the final 19 features was compared to the original 566-dimensional information or the previously selected 105 features, the results were very similar, indicating that the reduced feature set contained possibly all the necessary information to determine the characteristic properties of ACPs. During cross-validation, the performance metrics analyzing the ACPred-Fuse dataset showed some reduction, but the difference compared to the performance of the previously selected 105 features was small, indicating that the 19 features still captured sufficient information. In contrast, the metrics when testing the ACPred-FL dataset generally improved and were more representative. Comparing the performance of the full 566-dimensional data, the 105 and 19 selected features using peptide data from the ACP20Alt and ACP20main datasets only showed negligible differences, supporting the notion that the selected features were representative. The 19 physicochemical properties of amino acids from the AAindex database that were sufficient to predict ACP characteristics in peptides are shown in Table 4.
Further, it shows pairwise correlations for all 19 physicochemical properties (see Supplementary Figures S7–S10). The diagonal lines of Figures S7–S10 show that there was a large overlap in the numerical distribution of all 19 features across the positive and negative samples. For a certain feature correlated to 18 other features, a two pairs graph showed that it was not enough to distinguish the anticancer and non-anti-cancer peptides well. These results meant that relying on the feature alone or a combination of two features is not enough to identify a peptide with anticancer activity from the peptide sequence, i.e., the primary structure. A fine numerical analysis of all 19 features must be relied upon to obtain better results. Figure S5 and the support material file show the random forest binary classification trees and forests. From Figure S5, it can be seen that through the fine division of each feature value, many binary classification judgments are formed, and finally, the peptides with anticancer activity can be concluded. Here, these 19 features shown in Table 4 detected most of the critical features in all four datasets; it is likely that they can be used to distinguish ACPs from peptides with no ACP activity in general.

2.3. Statistical Comparison with Previously Reported Prediction Methods

We then compared the predictive performance of the RF-based prediction algorithm analyzing the initial model with 566-dimensional data versus those with 105- and 19- dimensional data and the previously reported ACP prediction methodologies. Obviously, the lack of statistically significant differences between the performance of the models would indicate that the 19 features captured sufficient information to distinguish ACPs from non-ACPs. The performance of the predictive algorithms previously reported in the literature is summarized in Table 5, while the comparison between the averaged performance of these previously reported approaches and our three models with 566-, 105-, and 19-dimensional features is shown in Table 6.
As shown in Table 6, the performance metrics of our predictive algorithm are generally lower than those for previously reported approaches. This is because of several reasons. The first is that the previously reported algorithms used two or more amino acid sequences in their feature representation; a multi-feature representation model will usually be better than a single-feature one. Here, in this study, we used a single feature (i.e., amino acid index), so the performance of the model will be slightly worse. Second, the focus of the previous algorithms was not the same as ours. The previous algorithms performed feature engineering optimization for a specific dataset. As a result, the optimized feature space of them usually only showed a better performance for that dataset, and the performance was worse for different datasets. That is, the generalization performance of the model was not good. Instead, we studied four standard datasets at the same time, looking for common feature representations that can be applied to different datasets, so as to build a more generalized model. Third, we also noticed a difference in performance compared to previous algorithms. For this reason, we performed a statistical significance test (see Table 7 and Table 8). The study showed that there was no statistically significant difference between the metrics of our results and the mean of the results optimized for specific datasets reported in the literature. This means that the 19 amino acid physicochemical properties we used can be applied to different datasets and obtain a performance without statistically significant difference from the literature algorithms.

3. Materials and Methods

We constructed an amino acid sequence feature extraction tool based on the AAindex database, converting peptide amino acid sequences into 566-dimensional feature vectors, where each dimension represents a physicochemical property of an amino acid [25,26,27]. First, the ACP datasets were divided into training and test datasets, and we analyzed them by a random forest (RF) model based on the full 566-dimensional features of the AAindex database to select the most informative features. Next, we used light gradient-boosting machine (LGBM), analysis of variance (ANOVA), chi-squared test (Chi2), and mutual information (MI) analysis for feature selection. By examining the performance of the RF model based on the best top n features under the four methods, we initially selected 105 features. RF modeling was performed again based on these 105 features on the same benchmark datasets, and the best features were selected based on hyperparameter optimization and feature importance analysis. As a result, we identified the best performing top 19 features. Finally, the RF model was trained based on the 19 features using all four datasets. The best performing models were compared with previously reported ACP prediction algorithms described in the literature, and the statistical significance of the differences in prediction performance indices was calculated. The overall flowchart of the conducted work is summarized in Figure 5.

3.1. Benchmark Datasets

Throughout the work presented here, we used the ACP20Alt [28], ACP20main [28], ACPred-FL [9], and ACPred-Fuse [10] datasets. Of these, the positive samples of ACPred-FL and ACPred-Fuse were primarily published by Chen et al. [29], and Tyagi et al. [2] in their CancerPPD datasets. Peptides with no ACP activity were represented by antimicrobial peptides (AMPs) and by a collection of peptides that had no anticancer effect during experimental testing. We randomly selected 250 ACP and 250 non-ACP sequences from the ACPred-FL dataset to act as the training dataset, and 82 ACPs and 92 non-ACPs as the test dataset. The 250 positive samples in the ACPred-Fuse training dataset were selected from the work of Wei et al. [9] Half of the 250 negative samples were also derived from Wei et al., while the other half was collected from the AMP dataset. The test dataset contained all the remaining ACPred-Fuse data (82 ACPs and 2628 non-ACPs) as positive and negative samples. The ACP20main and ACP20Alt databases were compiled by Lv et al. [28]. ACP20main contains 861 experimentally verified ACPs and 861 peptides with no documented ACP activity. Peptides from this dataset were divided into two subsets for 5-fold cross-validation and independent testing. Finally, the ACP20Alt database contains 970 ACPs and 970 non-ACPs. These were also subdivided into a training set and an independent test subset. The main difference between the ACP20main and ACP20Alt databases is that the negative samples in the former are represented by AMPs, while the negative samples in ACP20Alt are randomly selected peptides, assumed to have no antitumor activity. The main details of the used benchmark datasets are summarized in Table 9. The shared sequences numbers are shown in Figure S6 in the supporting information.

3.2. Feature Extraction

To investigate the physicochemical properties of key amino acids, we first had to extract the information contained in the ACP sequences and convert raw sequences into interpretable features [30,31]. The data for this feature extraction were obtained from the AAindex databases. The AAindex1 database, available at https://www.genome.jp/aaindex/ [32] (accessed on 1 January 2020), describes the physicochemical properties of amino acids using 566 indices.
The physicochemical properties of a peptide sequence with a length of L were extracted from the amino acid index database as numerical values for each amino acid [33,34]. It means that each amino acid in the peptide sequence is represented by a 566-dimensional vector, and then a peptide sequence of L amino acids will be transformed into an L × 566 matrix. Next, an average pooling of the matrix is conducted, resulting in a 566-dimensional vector representing the peptide sequence. This approach transformed each peptide sequence into a 566-dimensional feature vector, where each dimension represents a particular physicochemical property of amino acids [21]. The source code for peptides from the AAindex feature extraction can be downloaded from https://github.com/zhibinlv/RFaaindexACP (accessed on 19 June 2023).

3.3. Feature Selection Methods

Feature selection is the process of filtering the most relevant features by machine learning [35]. The main purpose of feature selection is to identify irrelevant or redundant features, reducing the runtime of machine learning algorithms and obtaining more accurate results. Four feature selection methods were used in our research: light gradient-boosting machine (LGBM), analysis of variance (ANOVA), chi-squared test (Chi2), and mutual information (MI). The LGBM feature selection code can be found at https://github.com/zhibinlv/iACP-DRLF/tree/main/feature_selection (accessed on 1 January 2020). For Chi2, ANOVA, and MI, it can be found using the scikit-learn toolkits at https://scikit-learn.org/stable/index.html (accessed on 1 January 2020).

3.3.1. LGBM

LGBM adopts the histogram algorithm [36], in which continuous features are turned into k discrete values in order to construct a histogram with a width of k. The algorithm is trained to count the value of each discrete value in the histogram. Based on these discrete values, the optimal feature segmentation points can be determined, and the exact number of key features can be obtained.

3.3.2. ANOVA

ANOVA analyzes the relationship between independent variables and dependent variables by studying whether the variance of multiple samples is equal to the overall mean value [37,38]. It can perform feature extraction before the data enter classifier training, thus reducing data dimensionality. Taking binary classification as an example, using ANOVA for feature selection divides the value of a certain feature into two groups, a positive and a negative group. The greater the difference between these groups in ANOVA, the greater their impact on the sample.

3.3.3. Chi-Squared Test

A chi-squared test can be used to determine whether two variables are correlated and to calculate the extent of this correlation [39], by performing a test between the feature and the real label. Assuming that the number of independent variables is A, and the number of dependent variables is B, the following equation χ can be constructed:
  χ 2 = ( A B ) 2 B  
The result indicates the degree of dependence between the independent variables and dependent variables.

3.3.4. MI

Mutual information [40,41] can also be used to test the correlation between two variables, which can be defined as follows:
I ( X ,   Y ) = y Y   x X   p ( x ,   y ) log ( p ( x ,   y ) p ( x ) p ( y ) )
If x and y are random variables that are independent of each other, then
  p x , y = p x p y  
Thus, the result of I(X, Y) is 0. Therefore, a larger I(X, Y) indicates a greater correlation between the two variables, allowing feature filtering.

3.4. Random Forest Algorithm

In this study, we used a random forest (RF) [42,43,44,45] machine learning algorithm to analyze the importance of features. By determining the contribution of each feature to each tree in the models, the importance of key features can be ranked, identifying key amino acid physicochemical properties that determine the likelihood of a given peptide having ACP activity.
During this process, the RF model selects several samples from the sample set to construct a training dataset with replacement. It then uses the obtained training dataset to generate a decision tree and randomly selects multiple non-repetitive features at each node, using these features to divide the sample set [46]. After the optimal division features are found, the process is repeated until all the decision trees in a random forest are generated [47]. Finally, the model trained by the above steps is used to predict the sample set, and the prediction result is determined by the number of classifications.

3.5. Feature Importance Analysis

Common evaluation metrics for calculating feature importance include the Gini index [48] and the out-of-bag (OOB) error. The Gini index is calculated according to the following equation:
G i n i p = k = 1 k p k 1 p k = 1 k = 1 k p k 2
where k indicates the number of features divided into k categories, and p k represents the importance of category k .
The OOB dataset refers to the data that are not chosen in the sampling process. For a tree in the RF model, the error e 1 is obtained by the out-of-bag data sample, and then, error e 2 is derived by randomly permuting the i-th column of the out-of-bag data matrix. This way, the importance of feature i can be represented by e 1 e 2 :
I m p i = e 1 e 2 n  
where n represents the number of decision trees in the model.

3.6. Evaluation of Model Performance

To evaluate the performance of each model, the sensitivity (Sn), specificity (Sp), accuracy (ACC), Matthews correlation coefficient (MCC), and area under receiver-operating characteristic curve (AUC) [49,50] were calculated as follows:
  S n = T P T P + F N × 100 %  
  S p = T N T N + F P × 100 %  
  A C C = T P + T N T P + F P + F N + T N × 100 %
  M C C = T P × T N F P × F N T P + F P × T P + F N × T N + F P × T N + F N  
TP, TN, FP, and FN represent the number of true positive, true negative, false positive, and false negative samples, respectively. In addition, we also analyzed the ROC curve and AUC area, where a larger AUC value indicates a stronger predictive performance.

3.7. Websever

After the model was developed and optimized, we developed a simple and easy-to-use website for interested readers to use. The reader only needs to enter the peptide sequence in FASTA format and after a few minutes can find out whether those peptides have anticancer activity. The website can be found at https://www.aibiochem.net/servers/RFaaindexACP (accessed on 18 June 2023). A simple screenshot of the application is shown in Figures S1–S4.

4. Conclusions

The work presented in this paper proposed to identify the key physicochemical properties of ACPs based on existing machine learning algorithms, using existing ACP datasets available in the literature and amino acid features collated in the AAindex database. We analyzed the influence of four feature selection methods, ANOVA, Chi2, LGBM, and MI, on identifying key features. The comparison of these methods revealed that LGBM was the best approach for selecting features that led to the creation of the best fitting RF model and resulted in the best performance indices. Ultimately, this work identified 19 key amino acid features, which compared favorably with machine learning models reported previously in the literature. Furthermore, statistical tests revealed that the 19 key identified features provided as much information as predictions based on a much larger feature dimensionality or alternative machine learning algorithms, enhancing the credibility of our approach. Based on these 19 key features, we can develop new machine learning models with better effects or further refine existing models. In addition, based on these key properties, investigators will be able to design prospective ACPs with an improved probability of therapeutic effectiveness, thereby increasing the speed of transitioning peptides into clinical ACP research.

Supplementary Materials

The following supporting information can be downloaded at https://www.mdpi.com/article/10.3390/ijms241310854/s1.

Author Contributions

Conceptualization, Z.L.; methodology, Z.L.; software, Z.L.; validation, Y.D., S.M. and Z.L.; formal analysis, Y.D., S.M. and Z.L.; investigation, J.L. and B.Z.; data curation, Y.D. and Z.L.; writing—original draft preparation, Y.D. and S.M.; writing—review and editing, Y.D., S.M., J.L., B.Z. and Z.L.; visualization, Y.D. and S.M.; supervision, Z.L.; funding acquisition, Z.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (No. 62001090), the Fundamental Research Funds for the Central Universities of Sichuan University (No. YJ2021104), and the 2023 Foundation Cultivation Research—Basic Research Cultivation Special Funding (No. 20826041G4189).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used to support the findings of this study can be made available by the corresponding author upon request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, L.; Yu, L.; Gao, L. Potent antibiotic design via guided search from antibacterial activity evaluations. Bioinformatics 2023, 39, btad059. [Google Scholar] [CrossRef]
  2. Tyagi, A.; Tuknait, A.; Anand, P.; Gupta, S.; Sharma, M.; Mathur, D.; Joshi, A.; Singh, S.; Gautam, A.; Raghava, G.P.S. Cancerppd: A database of anticancer peptides and proteins. Nucleic Acids Res. 2015, 43, D837–D843. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Yang, S.; Wang, Y.; Chen, Y.; Dai, Q. Masqc: Next generation sequencing assists third generation sequencing for quality control in n6-methyladenine DNA identification. Front. Genet. 2020, 11, 269. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Cao, C.; Mak, L.; Jin, G.; Gordon, P.; Ye, K.; Long, Q. Presm: Personalized reference editor for somatic mutation discovery in cancer genomics. Bioinformatics 2019, 35, 1445–1452. [Google Scholar] [CrossRef] [PubMed]
  5. Wang, H.; Guo, F.; Du, M.; Wang, G.; Cao, C. A novel method for drug-target interaction prediction based on graph transformers model. BMC Bioinform. 2022, 23, 459. [Google Scholar] [CrossRef]
  6. Novkovic, M.; Simunic, J.; Bojovic, V.; Tossi, A.; Juretic, D. Dadp: The database of anuran defense peptides. Bioinformatics 2012, 28, 1406–1407. [Google Scholar] [CrossRef] [Green Version]
  7. Thomas, S.; Karnik, S.; Barai, R.S.; Jayaraman, V.K.; Idicula-Thomas, S. Camp: A useful resource for research on antimicrobial peptides. Nucleic Acids Res. 2010, 38, D774–D780. [Google Scholar] [CrossRef] [Green Version]
  8. Wang, G.; Li, X.; Wang, Z. Apd2: The updated antimicrobial peptide database and its application in peptide design. Nucleic Acids Res. 2009, 37, D933–D937. [Google Scholar] [CrossRef] [Green Version]
  9. Wei, L.; Zhou, C.; Chen, H.; Song, J.; Su, R. Acpred-fl: A sequence-based predictor using effective feature representation to improve the prediction of anti-cancer peptides. Bioinformatics 2018, 34, 4007–4016. [Google Scholar] [CrossRef] [Green Version]
  10. Rao, B.; Zhou, C.; Zhang, G.; Su, R.; Wei, L. Acpred-fuse: Fusing multi-view information improves the prediction of anticancer peptides. Brief. Bioinform. 2020, 21, 1846–1855. [Google Scholar] [CrossRef]
  11. Agrawal, P.; Bhagat, D.; Mahalwal, M.; Sharma, N.; Raghava, G.P.S. Anticp 2.0: An updated model for predicting anticancer peptides. Brief. Bioinform. 2020, 22, bbaa153. [Google Scholar] [CrossRef]
  12. Yao, L.; Li, W.; Zhang, Y.; Deng, J.; Pang, Y.; Huang, Y.; Chung, C.R.; Yu, J.; Chiang, Y.C.; Lee, T.Y. Accelerating the discovery of anticancer peptides through deep forest architecture with deep graphical representation. Int. J. Mol. Sci. 2023, 24, 4328. [Google Scholar] [CrossRef]
  13. Jiang, J.; Lin, X.; Jiang, Y.; Jiang, L.; Lv, Z. Identify bitter peptides by using deep representation learning features. Int. J. Mol. Sci. 2022, 23, 7877. [Google Scholar] [CrossRef]
  14. Su, D.; Xiong, Y.; Wei, H.; Wang, S.; Ke, J.; Liang, P.; Zhang, H.; Yu, Y.; Zuo, Y.; Yang, L. Integrated analysis of ovarian cancer patients from prospective transcription factor activity reveals subtypes of prognostic significance. Heliyon 2023, 9, e16147. [Google Scholar] [CrossRef]
  15. Jiang, L.; Jiang, J.; Wang, X.; Zhang, Y.; Zheng, B.; Liu, S.; Zhang, Y.; Liu, C.; Wan, Y.; Xiang, D.; et al. Iup-bert: Identification of umami peptides based on bert features. Foods 2022, 11, 3742. [Google Scholar] [CrossRef]
  16. Zhou, W.; Liu, Y.; Li, Y.; Kong, S.; Wang, W.; Ding, B.; Han, J.; Mou, C.; Gao, X.; Liu, J. Trinet: A tri-fusion neural network for the prediction of anticancer and antimicrobial peptides. Patterns 2023, 4, 100702. [Google Scholar] [CrossRef]
  17. Nakai, K.; Kidera, A.; Kanehisa, M. Cluster analysis of amino acid indices for prediction of protein structure prediction and function. Protein Eng. 1988, 2, 93–100. [Google Scholar] [CrossRef]
  18. Tomii, K.; Kanehisa, M. Analysis of amino acid indices and mutation matrices for sequence comparison and structure prediction of proteins. Protein Eng. 1996, 9, 27–36. [Google Scholar] [CrossRef]
  19. Kawashima, S.; Pokarowski, P.; Pokarowska, M.; Kolinski, A.; Katayama, T.; Kanehisa, M. Aaindex: Amino acid index database, progress report 2008. Nucleic Acids Res. 2008, 36, D202–D205. [Google Scholar] [CrossRef] [Green Version]
  20. Ao, C.; Jiao, S.; Wang, Y.; Yu, L.; Zou, Q. Biological sequence classification: A review on data and general methods. Research 2022, 2022, 0011. [Google Scholar] [CrossRef]
  21. Dai, Q.; Ma, S.; Hai, Y.; Yao, Y.; Liu, X. A segmentation based model for subcellular location prediction of apoptosis protein. Chemom. Intell. Lab. Syst. 2016, 158, 146–154. [Google Scholar] [CrossRef]
  22. Tao, J.; Liu, X.; Yang, S.; Bao, C.; He, P.; Dai, Q. An efficient genomic signature ranking method for genomic island prediction from a single genome. J. Theor. Biol. 2019, 467, 142–149. [Google Scholar] [CrossRef] [PubMed]
  23. Cui, F.; Li, S.; Zhang, Z.; Sui, M.; Cao, C.; Hesham, A.E.-L.; Zou, Q. Deepmc-inabp: Deep learning for multiclass identification and classification of nucleic acid-binding proteins. Comput. Struct. Biotechnol. J. 2022, 20, 2020–2028. [Google Scholar] [CrossRef]
  24. Venny. An Interactive Tool for Comparing Lists with Venn’s Diagrams. Available online: https://bioinfogp.cnb.csic.es/tools/venny/index.html (accessed on 1 January 2020).
  25. Yang, Z.; Yi, W.; Tao, J.; Liu, X.; Zhang, M.Q.; Chen, G.; Dai, Q. Hpvmd-c: A disease-based mutation database of human papillomavirus in China. Database-J. Biol. Databases Curation 2022, 2022, baac018. [Google Scholar] [CrossRef] [PubMed]
  26. Cui, F.; Zhang, Z.; Cao, C.; Zou, Q.; Chen, D.; Su, X. Protein-DNA/rna interactions: Machine intelligence tools and approaches in the era of artificial intelligence and big data. Proteomics 2022, 22, e2100197. [Google Scholar] [CrossRef]
  27. Zhang, Z.; Cui, F.; Zhou, M.; Wu, S.; Zou, Q.; Gao, B. Single-cell rna sequencing analysis identifies key genes in brain metas-tasis from lung adenocarcinoma. Curr. Gene Ther. 2021, 21, 338–348. [Google Scholar] [CrossRef]
  28. Lv, Z.; Cui, F.; Zou, Q.; Zhang, L.; Xu, L. Anticancer peptides prediction with deep representation learning features. Brief. Bioinform. 2021, 22, bbab008. [Google Scholar] [CrossRef]
  29. Chen, W.; Ding, H.; Feng, P.; Lin, H.; Chou, K.-C. Iacp: A sequence-based tool for identifying anticancer peptides. Oncotarget 2016, 7, 16895–16909. [Google Scholar] [CrossRef] [Green Version]
  30. Saxena, D.; Sharma, A.; Siddiqui, M.H.; Kumar, R. Development of machine learning based blood-brain barrier permeability prediction models using physicochemical properties, maccs and substructure fingerprints. Curr. Bioinform. 2021, 16, 855–864. [Google Scholar] [CrossRef]
  31. Dao, F.-Y.; Liu, M.-L.; Su, W.; Lv, H.; Zhang, Z.-Y.; Lin, H.; Liu, L. Acrpred: A hybrid optimization with enumerated machine learning algorithm to predict anti-crispr proteins. Int. J. Biol. Macromol. 2023, 228, 706–714. [Google Scholar] [CrossRef]
  32. Aaindex: Amino Acid Index Database. Available online: https://www.genome.jp/aaindex/ (accessed on 13 February 2017).
  33. Onesime, M.; Yang, Z.; Dai, Q. Genomic island prediction via chi-square test and random forest algorithm. Comput. Math. Methods Med. 2021, 2021, 9969751. [Google Scholar] [CrossRef]
  34. Pan, G.; Jiang, L.; Tang, J.; Guo, F. A novel computational method for detecting DNA methylation sites with DNA sequence information and physicochemical properties. Int. J. Mol. Sci. 2018, 19, 511. [Google Scholar] [CrossRef] [Green Version]
  35. Dao, F.-Y.; Lv, H.; Zhang, Z.-Y.; Lin, H. Bdselect: A package for k-mer selection based on the binomial distribution. Curr. Bioinform. 2022, 17, 238–244. [Google Scholar]
  36. Sharma, A.; Singh, B. Ae-lgbm: Sequence-based novel approach to detect interacting protein pairs via ensemble of autoencoder and lightgbm. Comput. Biol. Med. 2020, 125, 103964. [Google Scholar] [CrossRef]
  37. Liao, X.; Gu, X.; Peng, D. Identification of plasmodium secreted proteins based on monodikgap and distance-based top-n-gram methods. Curr. Bioinform. 2022, 17, 804–813. [Google Scholar] [CrossRef]
  38. Panthakkan, A.; Anzar, S.M.; Jamal, S.; Mansoor, W. Concatenated xception-resnet50-a novel hybrid approach for accurate skin cancer prediction. Comput. Biol. Med. 2022, 150, 106170. [Google Scholar] [CrossRef]
  39. Yan, C.; Wu, B.; Ma, J.; Zhang, G.; Luo, J.; Wang, J.; Luo, H. A novel hybrid filter/wrapper feature selection approach based on improved fruit fly optimization algorithm and chi-square test for high dimensional microarray data. Curr. Bioinform. 2021, 16, 63–79. [Google Scholar] [CrossRef]
  40. Carballido, J.A. Microarray analysis workflow based on a genetic algorithm to discover potential hub genes. Curr. Bioinform. 2022, 17, 787–792. [Google Scholar] [CrossRef]
  41. Ding, Y.; Tang, J.; Guo, F. Predicting protein-protein interactions via multivariate mutual information of protein sequences. BMC Bioinform. 2016, 17, 398. [Google Scholar] [CrossRef] [Green Version]
  42. Ao, C.; Zou, Q.; Yu, L. Nmrf: Identification of multispecies rna 2′-o-methylation modification sites from rna sequences. Brief. Bioinform. 2022, 23, bbab480. [Google Scholar] [CrossRef]
  43. Zhou, L.; Wang, H. A combined feature screening approach of random forest and filter-based methods for ultra-high dimensional data. Curr. Bioinform. 2022, 17, 344–357. [Google Scholar]
  44. Zhang, H.; Chi, M.; Su, D.; Xiong, Y.; Wei, H.; Yu, Y.; Zuo, Y.; Yang, L. A random forest-based metabolic risk model to assess the prognosis and metabolism-related drug targets in ovarian cancer. Comput. Biol. Med. 2023, 153, 106432. [Google Scholar] [CrossRef] [PubMed]
  45. Zhang, Y.-F.; Wang, Y.-H.; Gu, Z.-F.; Pan, X.-R.; Li, J.; Ding, H.; Zhang, Y.; Deng, K.-J. Bitter-rf: A random forest machine model for recognizing bitter peptides. Front. Med. 2023, 10, 1052923. [Google Scholar] [CrossRef] [PubMed]
  46. Kong, R.; Xu, X.; Liu, X.; He, P.; Zhang, M.Q.; Dai, Q. 2sigfinder: The combined use of small-scale and large-scale statistical testing for genomic island detection from a single genome. BMC Bioinform. 2020, 21, 159. [Google Scholar] [CrossRef]
  47. Dai, Q.; Bao, C.; Hai, Y.; Ma, S.; Zhou, T.; Wang, C.; Wang, Y.; Huo, W.; Liu, X.; Yao, Y.; et al. Mtgipick allows robust identification of genomic islands from a single genome. Brief. Bioinform. 2018, 19, 361–373. [Google Scholar] [CrossRef]
  48. Yang, H.; Luo, Y.; Ren, X.; Wu, M.; He, X.; Peng, B.; Deng, K.; Yan, D.; Tang, H.; Lin, H. Risk prediction of diabetes: Big data mining with fusion of multifarious physical examination indicators. Inf. Fusion 2021, 75, 140–149. [Google Scholar] [CrossRef]
  49. Pan, J.; Lin, H.; Dong, Y.; Wang, Y.; Ji, Y. Mamf-gcn: Multi-scale adaptive multi-channel fusion deep graph convolutional network for predicting mental disorder. Comput. Biol. Med. 2022, 148, 105823. [Google Scholar] [CrossRef]
  50. Yang, Y.; Gao, D.; Xie, X.; Qin, J.; Li, J.; Lin, H.; Yan, D.; Deng, K. Deepidc: A prediction framework of injectable drug combination based on heterogeneous information and deep learning. Clin. Pharmacokinet. 2022, 61, 1749–1759. [Google Scholar] [CrossRef]
Figure 1. Comparison of performance metrics selected after the first feature selection.
Figure 1. Comparison of performance metrics selected after the first feature selection.
Ijms 24 10854 g001
Figure 2. The intersection data derived from the four datasets.
Figure 2. The intersection data derived from the four datasets.
Ijms 24 10854 g002
Figure 3. Intersection of features with a feature importance value greater than 0.01.
Figure 3. Intersection of features with a feature importance value greater than 0.01.
Ijms 24 10854 g003
Figure 4. Comparison of performance metrics based on feature dimensionality. (ad) is for metrics of models based on different datasets, respectively.
Figure 4. Comparison of performance metrics based on feature dimensionality. (ad) is for metrics of models based on different datasets, respectively.
Ijms 24 10854 g004
Figure 5. The overall flowchart of experimental design.
Figure 5. The overall flowchart of experimental design.
Ijms 24 10854 g005
Table 1. The number of features selected from the four datasets using the indicated feature selection methods.
Table 1. The number of features selected from the four datasets using the indicated feature selection methods.
DatasetsANOVAChi2LGBMMI
ACPred-Fuse73772840
ACPred-FL45467592
ACP20Alt949850100
ACP20main76616425
Table 2. The metrics of selected features used to create the 105-feature matrix.
Table 2. The metrics of selected features used to create the 105-feature matrix.
DatasetsNumber of Features5-Fold Cross-ValidationIndependent Test
ACC (/%)MCCSn (/%)Sp (/%)AUCACC (/%)MCCSn (/%)Sp (/%)AUC
ACPred-Fuse5880.20.61288.472.00.85892.90.45793.970.70.897
ACPred-FL3380.00.60183.276.80.86376.20.52881.770.70.856
ACP20Alt6189.70.79489.290.20.96090.90.81989.192.70.960
ACP20main8074.20.48967.281.20.80069.30.38667.371.30.792
Table 3. Performance metrics using the final 19 selected features.
Table 3. Performance metrics using the final 19 selected features.
DatasetsNumber of Features5-Fold Cross-ValidationIndependent Test
ACC (/%)MCCSn (/%)Sp (/%)AUCACC (/%)MCCSn (/%)Sp (/%)AUC
ACPPred-Fuse1978.60.57684.872.40.85493.00.44794.168.30.900
ACPred-FL1980.60.61686.075.20.86079.90.60185.474.40.846
ACP20Alt1989.30.78689.389.30.95889.40.79185.093.80.952
ACP20main1971.90.44066.777.00.79770.50.41068.472.50.789
Table 4. List of the selected 19 key features and the corresponding physicochemical properties.
Table 4. List of the selected 19 key features and the corresponding physicochemical properties.
FeaturePhysicochemical Properties
QIAN880113Weights for alpha-helix at the window position of 6
QIAN880126Weights for beta-sheet at the window position of 6
BROC820102Retention coefficient in HFBA
RACS820101Average relative fractional occurrence in A0(i)
SNEP660103Principal component III
OOBM850104Optimized average non-bonded energy per atom
SNEP660104Principal component IV
FINA910103Helix termination parameter at position j-2, j-1,j
OOBM850101Optimized beta-structure-coil equilibrium constant
RICJ880110Relative preference value at C5
RICJ880112Relative preference value at C3
CHAM830102A parameter defined from the residuals obtained from the best correlation of the Chou–Fasman parameter of beta-sheet
ZASB820101Dependence of partition coefficient on ionic strength
KLEP840101Net charge
FINA910101Helix initiation parameter at position i-1
MEEJ800101Retention coefficient in HPLC, pH7.4
WOLS870103Principal property value z3
AURR980112Normalized positional residue frequency at helix termini C4
KARP850103Flexibility parameter for two rigid neighbors
Table 5. Comparison of previously reported machine learning algorithms from the literature.
Table 5. Comparison of previously reported machine learning algorithms from the literature.
5-Fold Cross-ValidationIndependent Test
ACC (/%)MCCSn (/%)Sp (/%)AUCACC (/%)MCCSn (/%)Sp (/%)AUC
ACPPred-Fuse82.40.65277.287.60.88289.00.32072.089.50.868
ACPred-FL89.00.78384.893.20.94087.80.75884.191.50.960
ACP20Alt93.00.86189.796.30.97892.50.85288.696.4-
ACP20main76.70.53672.281.10.84173.40.46876.070.8-
Table 6. Comparison of the averaged literature results with the performance of RF based on three feature selections.
Table 6. Comparison of the averaged literature results with the performance of RF based on three feature selections.
5-Fold Cross-ValidationIndependent Test
ACC (/%)MCCSn (/%)Sp (/%)AUCACC (/%)MCCSn (/%)Sp (/%)AUC
566D82.60.65485.879.30.89185.80.57886.578.90.904
105D80.70.61681.679.70.87682.10.54181.778.00.874
19D80.10.60481.778.50.86783.20.56283.277.20.872
Literature average85.30.70881.089.60.91085.70.599580.287.1-
Table 7. Results of statistical tests based on 5-fold cross-validation with a significance level α = 0.05.
Table 7. Results of statistical tests based on 5-fold cross-validation with a significance level α = 0.05.
Comparison of ResultsACCMCCSnSpAUC
105D vs. 566D0.5380.5380.4660.6120.606
19D vs. 566D0.4350.4370.4730.3390.375
19D vs. 105D0.1820.1690.7950.1290.229
566D vs. Literature results<0.0010.449<0.001<0.0010.548
105D vs. Literature results<0.0010.043 <0.0010.115
19D vs. Literature results<0.0010.018<0.001<0.0010.047
Table 8. Results of statistical tests based on an independent test with a significance level α = 0.05.
Table 8. Results of statistical tests based on an independent test with a significance level α = 0.05.
Comparison of ResultsACCMCCSnSpAUC
105D vs. 566D0.4170.4640.40.602-
19D vs. 566D0.5890.7860.5720.433-
19D vs. 105D0.4330.350.4160.528-
566D vs. Literature results0.9830.7920.3170.298-
105D vs. Literature results0.2790.4240.8180.186-
19D vs. Literature results0.3830.5750.6770.175-
Table 9. The number of ACPs and non-ACPs in the four datasets derived from the literature.
Table 9. The number of ACPs and non-ACPs in the four datasets derived from the literature.
DatasetsTraining DatasetTesting DatasetSequence
Similarity
Sequence
Length (L)
Average
ACPNon-ACPACPNon-ACPLength
ACPred-Fuse125125821871<80%7 ≤ L ≤ 20725
ACPred-FL2502508282<90%11 ≤ L ≤ 20727
ACP20Alt775775193193<80%2 ≤ L ≤ 5024
ACP20main688688171171<80%3 ≤ L ≤ 5024
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Deng, Y.; Ma, S.; Li, J.; Zheng, B.; Lv, Z. Using the Random Forest for Identifying Key Physicochemical Properties of Amino Acids to Discriminate Anticancer and Non-Anticancer Peptides. Int. J. Mol. Sci. 2023, 24, 10854. https://doi.org/10.3390/ijms241310854

AMA Style

Deng Y, Ma S, Li J, Zheng B, Lv Z. Using the Random Forest for Identifying Key Physicochemical Properties of Amino Acids to Discriminate Anticancer and Non-Anticancer Peptides. International Journal of Molecular Sciences. 2023; 24(13):10854. https://doi.org/10.3390/ijms241310854

Chicago/Turabian Style

Deng, Yiting, Shuhan Ma, Jiayu Li, Bowen Zheng, and Zhibin Lv. 2023. "Using the Random Forest for Identifying Key Physicochemical Properties of Amino Acids to Discriminate Anticancer and Non-Anticancer Peptides" International Journal of Molecular Sciences 24, no. 13: 10854. https://doi.org/10.3390/ijms241310854

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop