Next Article in Journal
SPEEK and SPPO Blended Membranes for Proton Exchange Membrane Fuel Cells
Next Article in Special Issue
Insight into the Structure, Functions, and Dynamics of the Leptospira Outer Membrane Proteins with the Pathogenicity
Previous Article in Journal
Influence of the Sludge Retention Time on Membrane Fouling in an Anaerobic Membrane Bioreactor (AnMBR) Treating Lipid-Rich Dairy Wastewater
Previous Article in Special Issue
Structures and Dynamics of Dengue Virus Nonstructural Membrane Proteins
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

iAcety–SmRF: Identification of Acetylation Protein by Using Statistical Moments and Random Forest

1
Faculty of Computing and Information Technology, King Abdulaziz University, Jeddah 21911, Saudi Arabia
2
Department of Computer Science, Abdul Wali Khan University Mardan, Mardan 23200, Pakistan
*
Author to whom correspondence should be addressed.
Membranes 2022, 12(3), 265; https://doi.org/10.3390/membranes12030265
Submission received: 10 December 2021 / Revised: 25 January 2022 / Accepted: 1 February 2022 / Published: 25 February 2022
(This article belongs to the Special Issue The Structure, Dynamics and Function of Membrane Proteins)

Abstract

:
Acetylation is the most important post-translation modification (PTM) in eukaryotes; it has manifold effects on the level of protein that transform an acetyl group from an acetyl coenzyme to a specific site on a polypeptide chain. Acetylation sites play many important roles, including regulating membrane protein functions and strongly affecting the membrane interaction of proteins and membrane remodeling. Because of these properties, its correct identification is essential to understand its mechanism in biological systems. As such, some traditional methods, such as mass spectrometry and site-directed mutagenesis, are used, but they are tedious and time-consuming. To overcome such limitations, many computer models are being developed to correctly identify their sequences from non-acetyl sequences, but they have poor efficiency in terms of accuracy, sensitivity, and specificity. This work proposes an efficient and accurate computational model for predicting Acetylation using machine learning approaches. The proposed model achieved an accuracy of 100 percent with the 10-fold cross-validation test based on the Random Forest classifier, along with a feature extraction approach using statistical moments. The model is also validated by the jackknife, self-consistency, and independent test, which achieved an accuracy of 100, 100, and 97, respectively, results far better as compared to the already existing models available in the literature.

1. Introduction

Proteins are the basic and key part of the human body that perform many kinds of major functions in and outside a cell. The proteins are translated or synthesized from messenger RNA which is first codified into ribosomes and makes a chain of amino acid or polypeptide. After the translation process, certain amino acids can experience chemical changes at the protein’s C-termini or N-termini or in amino acid side chains, known as post-translation modifications (PTLM or PTM). The PTM can modify or may introduce the new functional group to the protein, such as in Acetylation, an example of Acetyl-lysine is shown in the Figure 1) [1]. It plays a key role in making protein products [2,3,4]. Each protein in the proteome may be altered either before or after it is translated. The charge state, hydrophobicity, conformation, and stability of a protein are all affected by various changes, which, in turn, influence its function. Protein modification has a variety of functions in different organs: (1) It ensure the fast and complex response of cells to regulate intra-cellular communication, division, and growth of cells (2) also pivotal for various physiological and pathological mechanism.
Protein acetylation can be achieved using a variety of methods, this adds the acetyl functional group into a chemical compound, which make another ester, the acetate.
Another form of lysine residue is usually acetylated [1,2,3,4]. The active substance, acetic anhydride, is commonly used to react with free hydroxyl groups as an acetylating agent. It is used in, for example, aspirin, heroin, and THC-O-acetate synthesis. Thousands of acetylated mammalian proteins have been identified [1], in addition to protein analysis. The acetylation takes place, for example, via a co-translation and a post-translational adaptation of proteins, histone, and p53, in addition to tubulins. Among these proteins, there is a high representation of chromatin proteins and metabolic enzymes, suggesting that acetylation, in addition to digestion, has an extremely important influence on the appearance of the genetic material. Among the microbes, 90 percent of the proteins that are surrounded by the central metabolism of Salmonella Centrica are acetylated [1,2].
Further, acetylation sites also play an important role in regulating membrane protein functions of multiple families, as documented in Reference [5]. It is supported by examples that acetylation is significantly enhanced in membrane-binding regions, where it is often located directly in critical membrane-binding pockets ideally positioned to modulate membrane interactions. Moreover, it was found that acetylation and acetylation mimetics strongly affected the membrane interaction of proteins, resulting in decreased membrane affinity and, in the case of amphiphysin and EHD2, altered membrane remodeling [6]. In cells, mimicking even a single acetylation event within the membrane interaction region reduced the binding affinity to membranes, resulting in cytoplasmic dispersal. In another report, acetylation affected the effects on membrane interaction, as well as membrane remodeling [7]. Similarly, the ability to control the membrane-binding activity of C2 domains via acetylation could allow the cell to further regulate Ca-dependent transmembrane transport and signaling events. It has also been documented that acetylation is present on the membrane-binding surface of the phosphatase domain of K163/K164, as it appears to be that Alanine mutations reduce membrane binding [8]. In addition, two reports on proteins with PH domains indicate that acetylation has opposite effects on membrane localization in cells (either increased or decreased) [9,10]. Much has been learned about the acetyltransferases and deacetylases that regulate protein-DNA-protein-protein interactions [2,5,11]. Some of these enzymes may also be involved in controlling protein membrane interactions. Consistent with this idea, localization of acetyltransferase in the cytoplasm and deacetylases in cell membranes have been observed [12,13,14].
Acetylation also plays a prominent role in numerous important cellular processes, such as stability and localization of protein [4,5]. In addition, modification of S/T/Y sites by acetylation, glycosylation, sulfation, and nitration has been reported [5,6]. Moreover, it also plays a role in the modulation of gene expression by histone alteration, as well as is a very significant function in controlling cellular metabolism and protein folding [15,16,17,18].
From the above discussion it revealed that Acetylation is an important post-translation modification, and it is necessary to correctly identify them; however, it remains a major challenge to understand the functions and regulations of the molecular acetylation mechanism. Many traditional approaches are in use for their identification, including high-throughput mass spectrometry (MS) [19,20]. However, since the acetylation mechanism is complicated, rapid, and reversible, such methods remain time-consuming, expensive, and laborious [21,22].
To overcome the existing problems in its identification, many researchers have developed a computational model for fast and inexpensive prediction of PTM sites [15,16,23,24], such as the ubiquitination [17,18,24], the phosphorylation [15,16,19,25,26,27], sumoylation [28,29,30], and the acetylation [31,32,33,34], etc. The important step for the PTM prediction model to correctly transform the biological sequences into their equivalent numerical form, for this purpose, many feature extraction methods are developed which are documented like the amino acids composition (AAC), the dipeptides composition (DPC), and Pseudo Amino Acid Composition (PseAAC) [35]. For such feature extractions many methods are discussed in Reference [20].
Reference [1] proposed a novel measurement procedure iAcet-PseFDA, a classification model for acetylation proteins by extracting features come from sequence conservation information using a gray structure model and KNN scoring based on functional domain annotation databases including GO [36] and subcellular localization for acetylation protein recognition. The authors achieved 77.10 percent accuracy using 5-fold cross-validation on three datasets, with a significant amount of attribute analysis and the discovery algorithm for relief functionality.
Reference [37] proposed a method ProAcePred to predict prokaryote-specific lysine acetylation sites, using SVM, 10-fold cross-validation, and the elastic net mathematical approach for optimizing the dimensionality of feature vectors, which greatly increased prediction accuracy and yielded promising results.
Wuyun et al. [38], developed a model, KA-predictor, to predict species-specific lysine acetylation sites using the classifier SVM. They achieved highly competitive for the majority of species as compared with other methods.
Hou [39] suggested a predictor for lysine acetylation prediction called LAceP, based on logistic regression classifiers and various biological characteristics. Using Random Forest classifiers, Li [40] developed SSPKA, a tool for species-specific lysine acetylation prediction.
From the above discussion, it has been observed that many predictors have been developed for the identification of acetylation sites; however, the maximum prediction accuracy established in all previous models was 77.10%, which is very poor for a correct identification of the acetylation sites.
Therefore, in the present study, we use statistical moments as feature extractions and Random Forest and PNN as a classifier. Further, the model evaluation is done by 10-fold cross-validation, self-consistency, independent, and by jackknife testing. We obtained dominant results as compared to the existing models that were developed earlier.
Therefore, to improve the predictor model, we use statistical moments as feature extractions and Random Forest and PNN as classifiers. The model evaluation is carried out through 10-fold cross-validation, self-consistency, independently, and the jackknife tests. We achieved dominant results as compared to the existing models developed earlier.

2. Materials and Methods

In this review, we follow the 5-step procedure mentioned in References [16,41] to establish a predictive predictor for biological sequences. It consists of the following steps: (1) choosing or generating an appropriate benchmark dataset to be used for training and testing; (2) transforming a biological sequence into its equivalent mathematical form, which returns the basic correlation with the biological sequence, and transforming a biological sequence into its equivalent mathematical form, which returns the required correlation with the biological sequence; (3) developing or using existing classifier for the required predictor; (4) using cross-validation experiments to determine the accuracy of the suggested indicator; and, finally, (5) developing an influential website/GitHub resource for public use for the benefit of future research and development. All the above steps are presented in Figure 2. Further, details of the above can be found in the following subsections.

3. Benchmark Dataset

We begin with collection of a valid benchmark dataset for training and testing, which is the first step in the 5-step rule [1]. The dataset is collected from the well-known data repository, the UNIPROT http://www.uniprot.org, retrieved dated 8 December 2021. It contains 2900 protein samples, of which 725 were positive denoted by Sposi, and 2175 were negative denoted by Snegt. Further, for the effectiveness of the proposed predictor, the set of negative data samples is equally divided in three sets, S ¯ 1 , S ¯ 2 , and S ¯ 3 as in Reference [1] such that,
S ¯ i S ¯ j = ϕ ,             ( i j ; ,   i ,   j = 1 ,   2 ,   3 ) ,
where the sign ∩ represents the intersection of sets. Furthermore, we individually combine these three negative datasets S ¯ 1 , S ¯ 2 and S ¯ 3 with S_posi and created three new datasets with the same number of positive and negative samples as expressed in Equation (1) below:
( S posi S ¯ 1 ) ,   ( S posi S ¯ 2 ) ,   ( S posi S ¯ 3 ) .
The symbol denotes the union of two sets. It is important to note that the dataset under discussion was used for prediction by [1], whereas the positive dataset was collected based on the given three steps: (1) The possible proteins to be acetylated are identified by a single fixed keyword, i.e., {N acetylcysteine, N acetylserine, N acetylglutamate, N acetylglycine, N acetylcholine, N acetylthreonine, N acetylcholine, N acetylmethionine, N acetylmethionylc, N acetyltyrosine, O-acetylserine, or N6-acetyllysine O-acetyltheronine O}; (2) here, protein collection was validated using some assertion technique; and (3) the 30, or additional, amino acids in proteins, along with the redundant proteins, were removed as discussed in [1].
Whereas the negative data was generated in a similar way to the positive data, except those proteins are not a member being searched by the above keywords? As a result, it produced a large number of negative samples, moreover, random selection is made from those that were balanced in size to positive samples.

4. Feature Extraction

The existing traditional classifiers, such as SVM, KNN, ANN, and many others, are not as powerful in classifying the biological data and making the required prediction. Therefore, a medium is needed to convert biological data into the necessary numerical form to make it suitable for traditional classifiers. For this purpose, many models are developed to extract the required characteristics from biological data, e.g., PseAAC, AAC, Pse-in-One, Pse-Analysis, and many more [27,28,29]. In feature extraction, the emphasis is on preserving the critical properties of the protein, its location, and functions. The statistical moment [42] is used to derive features in this study, which is discussed in detail below.

4.1. Statistical Moments

In statistics and probability distributions, some form remains beneficial when performing analysis of a particular sequence. The study of such configuration of data collection in pattern matching is known as moments [25]. There are useful moments when there are various pattern recognition problems related to feature development that do not depend on the pattern or sequence parameters provided [27,29,30,31,43]. Particular moments are used to calculate data size, data alignment, and data eccentricity. In this study, we extract the necessary features of acetylation proteins using Hahn, raw, and central moments. The raw moment is used to estimate the probability distribution by using mean, variance, and asymmetry, these moments are neither location invariant nor scale invariant [32]. Similarly, the same procedure is used in case of the central moment, but the calculation is based on the data centroid. This moment is a scale variant and location invariant. The Hahn moments, on the other hand, are dependent on Hahn polynomials, it is neither scale invariant nor location variant [33,34,44]. These moments are very important for extracting obscure features from protein sequences, as they contain complex orderly details about biotic sequences. In the proposed work, a linearly planned structural of a protein sequence is used, as given in Equation (2).
P = R 1 R 2 R 3 R L ,
where R 1 is the 1st amino acid, represented in proteins P, the last amino acid is R L , and total length is ‘L’. Transforming the information of the protein linear structure as given in Equation (2) into 2D matrix representation of dimension k as computed by the following equation.
k = [√P]
where “P” represents the protein sequence length, and k represents the dimension of the obtained 2D square matrix.
Hence the Equation (4), represents the matrix denoted by N is constructed by using the order obtained from Equation (3) that is k × k
N = [ T 1 1 T 1 2 T 1 j T 1 k T 2 1 T 2 2 T 2 j T 2 k T i 1 T i 2 T i j T i k     T k 1 T k 2 T k j T k k ]
The raw moment R (a, b) is computed by the values of N’, which is a continuous 2D function of order (a + b), as shown in Equation (5):
R ab = p · q · p a q b N ( p , q ) .
Up to order 3, the equation calculates the raw moments. This raw moments are measured using the data’s roots as a starting point [45,46,47,48]. R00, R01, R10, R11, R02, R20, R21, R30, and R03 are the raw moment’s characteristics, weighed up to order 3rd.
The centroid is a point from where all points are equivalently dispersed in all directions with a weighted average [45,48,49,50]. The following equation, which uses the centroid, calculates the special characteristics of central moments up to order 3 (6).
C ab = p · q · ( p p ¯ ) a ( q q ¯ ) b N ( p , q ) .
The unique features are calculated up to 3rd order as: C00, C01, C10, C11, C02, C20, C30, and C31. Further, the centroids are calculated, as given by Equations (7) and (8), as p ¯ and q ¯ .
p ¯ = R 10 R 00 ,
q ¯ = R 01 R 00 .
The Hahn moments must be converted from 1D notation to a square matrix before they can be calculated. Discrete Hahn’s moments, also known as 2D moments, necessitate square matrix input data in a 2D structure [51]. Since these moments are orthogonal possess inverse properties, therefore, the construction of original data can be constructed using the inverse discrete Hahn moment. The aforementioned remains observed, and the positional and compositional features are somehow preserved in the measured moments [25,32,33,34,44,52] Two-dimensional input data in the form N’ is used to calculate the orthogonal Hahn moments, as seen in Equation (9).
h m x , y   ( r , M ) = ( M + y 1 ) m ( M 1 ) m   s = 0 m 1 s × ( m ) s   ( r ) s   ( 2 M + x + y m 1 ) s ( M + y 1 ) s ( M 1 ) s · 1 s ! ,
where ‘p’ and ‘q’ (p > −1, q > −1) controlling the shape of polynomials by using the adjustable parameters. The Pochhammer symbol is defined by Equation (10), as follows:
( ϼ ) s = ϼ   ( ϼ + 1 )   ( ϼ + s 1 ) .
The equation is further simplified by the Gamma operator:
( ϼ ) s = Γ   ( ϼ + s ) ϼ ( ϼ ) .
The raw values of Hahn moments are usually scaled using a square norm and weighting formula, as seen in Equation (12):
h m x ~ , y ( r , M ) = h m x , y   ( r , M ) ϼ ( r ) s m 2 ,   m   = 0 , 1 , , M 1 .
Meanwhile, in Equation (13),
ϼ ( r ) = Γ   ( p + r + q ) Γ   ( q + r + 1 ) ( p + q + r + 1 ) M   ( p + q + 2 r + 1 ) m   ! ( M r 1 ) ! .
Hahn moments are computed for the discrete 2D data up to the 3rd order through the following, Equation (14):
H pq   j = 0 M 1 · i = 0 M 1 · N i , j   h p x ~ , y ( j , M )   h q x ~ , y ( i , M )   ,   p ,   q   = 0 , 1 , , M 1 .
The special features based on the Hahn moments are represented by H00, H01, H10, H11, H02, H20, H12, H21, H30, and H03. For each protein sequence up to the third order, we produced 10 central, 10 raw, and 10 Hahn moments and added them to the miscellaneous Super Feature Vector at random (SFV).

4.2. Position Relative Incident Matrix (PRIM)

The amino acids’ order and location in a protein sequence have crucial importance for the recognition of protein characteristics [47,50,53]. In any protein sequence, the relative position of an amino acid remains an essential pattern for understanding its physical properties. The Position Relative Incident Matrix (PRIM) uses a square matrix of order 20 to depict the relative location of amino acids in protein sequences, which is expressed by Equation (15):
N PRIM = [ O 1 1 O 1 2 O 1 j O 1 20 O 2 1 O 2 2 O 2 j O 2 20 O i 1 O i 2 O i j O i 20     O k 1 O k 2 O k j O k 20 ]
O i j represents the position of the jth amino acid for the first occurrence of the ith amino acid in the chain.
The score is of biological evolutional process accomplished by amino-acid of type ‘J’. The matrix, N PRIM has 400 coefficients based on the relative position of amino acids occurrence.
Ten central moments, 10 raw moments, and 10 Hahn moments are calculated using the 2D N PRIM and 30 additional special features randomly applied to the miscellaneous SFV.

4.3. Reverse Position Relative Incident Matrix (R-PRIM)

There are several instances of cell biology where biochemical sequences are homologous in origin. This normally occurs where a single ancestor is involved in the evolution process, and several sequences are derived from it. In such situations, using these homologous sequences has a significant impact on the classifier’s output. For the purpose of obtaining correct results, successful and efficient sequence similarity searching is carried out. In machine learning, efficiency and accuracy are urgently needed for the preciseness of feature extraction algorithms through which the most relevant features are extracted from biological data [43,47,50,53].
The methods used in R-PRIM and PRIM computations are the same, but R-PRIM is only useful for reverse protein sequence ordering. The R-PRIM computations revealed hidden trends in the data and removed ambiguities between homologous sequences. R-PRIM was created as a 20 × 20 matrix containing 400 hundred coefficients, as seen in Equation (16):
N R - PRIM = [ B 1 1 B 1 2 B 1 j B 1 20 B 2 1 B 2 2 B 2 j B 2 20 B i 1 B i 2 B i j B i 20     B k 1 B k 2 B k j B k 20 ]
The NR-PRIM 2D matrix is used to measure 10 raw, 10 central, and 10 Hahn moments up to 3rd order, as well as more than 30 special features that are randomly applied to the SFV range.

4.4. Frequency Distribution Vector (FDV)

A frequency distribution vector (FDV) can be generated by using the distribution rate of each amino acid in a protein chain, as expressed in Equation (17).
µ = { a i , a 2 , a 3 , a 4 , a 20 } .
Here a i is the occurrence of frequency of ith (1≤ i ≤ 20) amino acid in each protein chain. Twenty more special functions have been randomly added to the SFV’s miscellany.

4.5. AAPIV (Accumulative Absolute Position Incidence Vector)

The AAPIV was used to retrieve relevant amino acid positional information, which retrieves and stores amino acid positional information for 20 native amino acids in a protein sequence [50,53]. This creates 20 critical features associated with each amino acid in a sequence, as expressed by Equation (18). These 20 new features were thrown into the SFV at random.
AAPIV = { μ 1 , μ 2 , μ 3 , , μ 20 } ,
where Φ i is expressed by Equation (19).
μ i = x = 1 n R x .
The μ i comes from the protein sequence Rx, which has a cumulative amino acid count of ‘n’, which can be determined using Equation (19).

4.6. R-AAPIV (Reverse Accumulative Absolute Position Incidence Vector)

R-AAPIV uses reverse sequence ordering to extract and store positional information of amino acids with respect to 20 native amino acids in a protein sequence, which is in reverse order relative to AAPIV [50,53].
This creates 20 critical features associated with each amino, as expressed by Equation (20).
R AAPIV = { Φ i , Φ 2 , Φ 3 , ,   Φ 20 } ,
where Φ i is expressed by Equation (21).
Φ i = x = 1 n   Reverse   ( R ) x .
where R 1 , R 2 , R 3 , R n are the ordinal locations at which the residue of protein sequence occurs in the reverse sequence? The values of an arbitrary element of Φ i are given by Equation (21).

5. Machine Learning Classifiers

The Random Forest and Probabilistic Neural Network are used as a training model to predict acetylation and non-acetylation sites in this research. The following sections go into these classification algorithms in greater depth.

5.1. Probabilistic Neural Network

In this paper, we used a Probabilistic Neural Network (PNN) as a classifier. The Probabilistic Neural Network (PNN) is a powerful algorithm that is mostly employed for classification problems. PNN was first introduced by Specht in 1990 [54]. It is a feed-forward neural network that works on the principle of Kernel Fisher Discriminant Analysis. PNN uses the probability density function and generates better prediction results compared to other neural network algorithms [55,56].
The PNN operates on four layers, i.e., (1) input layer, (2) hidden layer, (3) pattern layer/summation layer, and (4) output layer [57]. The input layer accepts the training samples as input; further, the input data is forwarded to the hidden layer to operate the computational functions. In the hidden layer, an individual node is a computational unit that has some weighted input connections. The third layer received the probability results along with its given classes [58]. The output layer takes the decision and assigns the respective class label to the unknown sample. The schematic view of the PNN is shown in Figure 3.

5.2. Random Forest

In this paper, we used Random Forest (RF) as a classifier [59]. Random Forest (RF) is an ensemble learning method for classification and regression problems that is widely used [60,61,62]. RF, as its name implies, contains a large number of individual decision trees that operate as an ensemble. Each individual tree in the Random Forest splits the dataset into training and testing subsets. The training subset is used for training the model, and the testing subset is employed for testing the prediction performance of the trained model classification. During the classification, the class label is assigned to the testing sample by receiving the majority votes of all trees. The variation or bias of a single tree has little effect on the overall prediction accuracy because of the majority voting principle. RF also implements the concept of the weight model by providing a weight value that is low when a particular tree consumes a high error rate. Random Forest builds multiple decision trees and merges them together to get a more accurate and stable prediction. Random Forest has nearly the same hyper parameters as a decision tree or a bagging classifier. Random Forest adds additional randomness to the model while growing the trees [59]. The working of the Random Forest algorithm is illustrated by the following procedure.
The first step starts with the selection of random samples from a given dataset; in the second step, this algorithm will construct a decision tree for every sample, and, in the last step, voting will be performed for every predicted result [63].
We tested the proposed model performance using the different numbers of trees and found the best results when the number of trees is 500. The RF offers several advantages, including its optimal accuracy, it works efficiently with large datasets, and its detection of outliers. Many researchers used RF for solving several biological problems and achieve better performance. The working mechanism of the RF is depicted in Figure 4.

6. Performance Evaluation Parameters and Testing Methods

6.1. Performance Evaluation Parameters

The performance of the proposed model can be measured by effective evaluation metrics. We use the subsequent four metrics to measure the forecast quality: (1) Overall Accuracy (ACC), (2) Sensitivity (Sn), (3) Specificity (Sp), and (4) Mathews Correlation Coefficient (MCC). We computed these parameters using a binary confusion matrix. These metrics remain the most common metrics used to measure efficiency of the proposed model. We computed these parameters using the following equations.
{ ACC = TP + TN TP + TN + FP + FN S n = TP TP + FN S p = TN TN + FP MCC = ( TP × TN ) ( FP × FN ) ( TP + FP ) ( TP + FN ) ( TN + FP ) ( TN + FN ) , Recall = TP TP + FN Precision = TP TP + FP F . measure = 2 × Recall × Precision Recall + Precision
where TP is True Positive, FN is False Negative, TN is True Negative, and FP is False Positive, respectively.
As per the given confusion matrix as shown in the Table 1, we subsequently calculate the following:
TP: Production prognosis, such as True Positive (TP), where we found that acetylation subject stays properly categorized, as well as classified, then the subjects have acetylation proteins.
TN: Production forecast, such as True Negative (TN), where we found that an non-acetylation protein remains properly classified, and then the subject remains non-acetylation protein.
FP: Production prognosis, such as false positive (FP), where we found that non- acetylation protein remains inaccurately classified as containing acetylation proteins known as “type 1 error”.
FN: Forecast of production, such as false negative (FN), where we found that acetylation proteins remain inaccurately classified and that the subject has non- acetylation proteins, this is the “type 2 error”.
ROC and AUC: The optimistic receiver curves evaluate the predictability of the machine learning classifiers at various threshold settings. The ROC exam remains a graphical demonstration that relates the “true positive rate” to the “false positive rate” in the grouping results of the machine learning algorithm. AUC describes a classifier of the ROC. The higher value of the AUC more than 0.5, suggest discrimination, whereas the value of 0.5 doesn’t suggest any discrimination in true positive and true negative of classifier, more the AUC value more the efficiency in the performance of a classifier.

6.2. Testing Methods

Several cross-validation techniques have been used to examine the statistical forecaster’s results in the literature. The jackknife test, independent dataset test, and k-fold cross-validation test are three experiments that are commonly used by various researchers.
When testing a forecaster designed for its efficiency, we use the following cross-validation methods in this paper to estimate the expected accuracy of the forecaster, self-consistency, independent, K-fold cross-validation, and jackknife testing for the assessment of the proposed model.
The following sub-sections contain the details.

6.3. K-Fold (10-Fold) Cross-Validation Test

The K-fold cross-validation (KFCV) test is a technique to estimate predictive models by partitioning the original dataset A into disjoint k-folds {A1, A2, A3, Ai…, Ak}, where it uses A-Ai folds for training, and the remaining ith fold Ai for testing, where i = 1, 2, 3, …k as shown in the Figure 5. The method iterates the process for each i, and calculate the performance that is the accuracy, sensitivity, precession, recall, F-Measure, and MCC. Further, for the overall result, the average is taken of all the iterations performed for each fold. This technique has many benefits, such as the fact that it validates the model based on multiple datasets to reduce the bias and reach to a stable evaluation that how the model performs. This technique is much more powerful compared to other cross-validation techniques. In literature, the K-fold method is quite popular for k = 10 and 5. In this research, a 10-fold cross-validation is used: the overall result obtained is 100% with random forest classifier as presented in Table 2, and through ROC curves shown in Figure 6.
The results obtained from three dataset using 10-folds Cross validation with the Random forest classifier, and achieved the best result with ACC 100%, MCC, Sn, Precision, and F-measure all are 1.
We also used a 10-fold cross-validation test to evaluate the PNN-based model on the three datasets, and obtained the ACC 66.83, MCC 0.36, Sn 0.72, Precision 0.65, and F-measure 0.72 as presented in the Table 3 and by the ROC curve in the Figure 7.

6.4. Jackknife Test

Several cross-validation tests are extensively useful to estimate the performance of the statistical predictors. Amongst these, the jackknife test is considered to be the supreme in being consistent and reliable. Consequently, the jackknife test is comprehensively applied by researchers to estimate the performance of the predictor model. In this test, if the dataset has N records in the dataset, then it trains the model for N − 1 records and tests the model for the remaining one record, which is why it is also called leave-one-out cross-validation. Further, this process is repeated N-times, and the label of each record is predicted. Finally, we accumulate all the results to make the overall prediction based on accuracy, sensitivity, precession, recall, F-measure, and MCC.
In present research work, the jackknife test is used to measure the performance of models by using the classifiers Random Forest and PNN, and we achieved the result of 100% through RF but got 66.87 through PNN, as presented in Table 4 and Table 5, and by the ROC curves in Figure 8 and Figure 9.
In this evaluation process, three datasets were used which give the overall accuracy, sensitivity, specificity, precession, MCC, Recall, and F-Measure, given in detail in Table 4.

6.5. Self-Consistency Test

Self-consistency test is a technique referred to as the ultimate test for the validation of efficiency and efficacy of the prediction model, this method uses the same data for both training and testing A representation of these proposed parameters, by conducting the self-consistency testing, the results for the acetylation protein prediction based on the Random Forest classifier as presented in the Table 6 and by the ROC curve is shown in the Figure 10.
Similarly, the evaluation results based on the PNN for the three datasets, S1, S2, and S3, based on RF and PNN as presented in the Table 7 and by the ROC curve as shown in the Figure 11.

6.6. Independent Test

An independent test is a cross-validation method that objectively finds out the predictor’s performance metrics of the planned model, which obtains values from a confusion matrix to evaluate the performance of the model. In this method, the dataset is divided into two parts, training, and the testing part. In the proposed work the data is split in two parts that is 70% of the data for the training and the remaining 30% for the testing as shown in Figure 12. The method used to train and test the models based on the classifiers, the Random Forest and the PNN, and we obtained the result of 98% through RF, and 50.8 through PNN. The area under the curve (AUC), obtained by Random Forest and PNN, is 98% and 50.8%, respectively. The remaining detailed results, based on the two classifiers, is presented in Table 8 and Table 9 and by the ROC curve in Figure 13 and Figure 14, respectively.

7. Results and Discussions

The most critical thing is to compare the proposed novel model to other state-of-the-art models in order to assess its prediction accuracy. When compared to well-known current classifiers, the RF and the PNN. In this work, the model with RF achieved significantly higher accuracy and efficiency in predicting acetylation from non-acetylation, as seen in Table 10 and by Figure 15 and Figure 16.

Comparison to Existing Models

To show the efficiency of our proposed model, iAcety–SmRF, which achieved the highest accuracy of 100% with Sensitivity, Precision, and MCC, is 1, as shown in Section 4.2 (B) (i), using a 10-fold cross-validation. The proposed model was compared to several models, including the latest iAcet-PseFDA by Reference [1]; they used the same dataset and developed a method for predicting acetylation proteins by extracting features from sequence conservation information using a gray frame model and an ANN score based on information from the annotation of the functional domains and subcellular localization. For model validation, they used a 5-fold cross-validation for all three datasets and achieved an average accuracy of 77.10%. The All-Mean JK model by Nakai and Horton [36] achieved an accuracy of 74.64%. Hunter constructed a predictive model InterPro, with accuracy of 68.25% reference [64]. Table 11 lists all comparative analyses of the proposed study, based on our two models iAcety-SmRF and iAcety-SmPNN, in which iAcety-SmRF achieved the superior accuracy as compared to all existing models.

8. Conclusions

A computational model for predicting Acetylation sites from non-Acetylation sites is developed in this paper. The model contained Statistical Moments used for extraction of features into the equivalent numerical form of the original biological data. Further, Random Forest and Probabilistic Neural Network were applied for its classification to predict acetylation from non-acetylation. Furthermore, independent testing, 10-fold cross-validation, self-consistency test, and jackknife testing are used to evaluate accuracy, with results of 97%, 100%, and 100%, respectively, based on the Random Forest. Further, the model was compared with the already existing relevant models available in the literature, which revealed the remarkable performance of our work. Finally, the final step is to develop an influential website/GitHub resource for public use for the benefit of future research and development which can be accessed by the following link: https://github.com/shaistarahmanmcs/My-Website-identifying-Proteins-Acetylation-.git (accessed on 8 December 2021).

Author Contributions

Conceptualization, S.M. and S.R.; methodology, O.B.; software, S.R.; validation, S.A.K. and R.A.; formal analysis, investigation, resources, data curation, S.A.K.; writing—original draft preparation, S.R.; writing—review and editing, S.M. and O.B.; supervision, S.A.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset and classification code is available on the GitHub: https://github.com/shaistarahmanmcs/My-Website-identifying-Proteins-Acetylation-.git (accessed on 8 December 2021).

Acknowledgments

The authors extend their appreciation to the Deputyship for Research & Innovation, Ministry of Education in Saudi Arabia for funding this research work through the project number IFPHI-359-830-2020 and King Abdulaziz University, DSR, Jeddah, Saudi Arabia.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Qiu, W.-R.; Xu, A.; Xu, Z.-C.; Zhang, C.-H.; Xiao, X. Identifying Acetylation Protein by Fusing Its PseAAC and Functional Domain Annotation. Front. Bioeng. Biotechnol. 2019, 7, 311. [Google Scholar] [CrossRef] [Green Version]
  2. Chunaram, C.; Chanchal, K.; Florian, G. Lysine Acetylation Targets Protein Complexes and Co-Regulates Major Cellular Functions. Science 2009, 325, 834–840. [Google Scholar] [CrossRef] [Green Version]
  3. Drazic, A.; Myklebust, L.M.; Ree, R.; Arnesen, T. The world of protein acetylation. Biochim. Biophys. Acta—Proteins Proteom. 2016, 1864, 1372–1401. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Zhang, K.; Tian, S.; Fan, E. Protein lysine acetylation analysis: Current MS-based proteomic technologies. Analyst 2013, 138, 1628. [Google Scholar] [CrossRef] [PubMed]
  5. Choudhary, C.; Weinert, B.T.; Nishida, Y.; Verdin, E.; Mann, M. The growing landscape of lysine acetylation links metabolism and cell signalling. Nat. Rev. Mol. Cell Biol. 2014, 15, 536–550. [Google Scholar] [CrossRef] [PubMed]
  6. Yang, Y.; Rao, R.; Shen, J.; Tang, Y.; Fiskus, W.; Nechtman, J.; Atadja, P.; Bhalla, K. Role of Acetylation and Extracellular Location of Heat Shock Protein 90α in Tumor Cell Invasion. Cancer Res. 2008, 68, 4833–4842. [Google Scholar] [CrossRef] [Green Version]
  7. Bozelli, J.C.; Kamski-Hennekam, E.; Melacini, G.; Epand, R.M. α-Synuclein and neuronal membranes: Conformational flexibilities in health and disease. Chem. Phys. Lipids 2021, 235, 105034. [Google Scholar] [CrossRef]
  8. Okada, A.K.; Teranishi, K.; Ambroso, M.R.; Isas, J.M.; Vazquez-Sarandeses, E.; Lee, J.Y.; Melo, A.A.; Pandey, P.; Merken, D.; Berndt, L.; et al. Lysine acetylation regulates the interaction between proteins and membranes. Nat. Commun. 2021, 12, 6466. [Google Scholar] [CrossRef]
  9. Sundaresan, N.R.; Pillai, V.B.; Wolfgeher, D.; Samant, S.; Vasudevan, P.; Parekh, V.; Raghuraman, H.; Cunningham, J.M.; Gupta, M.; Gupta, M.P. The Deacetylase SIRT1 Promotes Membrane Localization and Activation of Akt and PDK1 During Tumorigenesis and Cardiac Hypertrophy. Sci. Signal. 2011, 4, ra46. [Google Scholar] [CrossRef] [Green Version]
  10. Fischer, A.; Mühlhäuser, W.W.D.; Warscheid, B.; Radziwill, G. Membrane localization of acetylated CNK1 mediates a positive feedback on RAF/ERK signaling. Sci. Adv. 2017, 3, e1700475. [Google Scholar] [CrossRef] [Green Version]
  11. Gräff, J.; Tsai, L.-H. Histone acetylation: Molecular mnemonics on the chromatin. Nat. Rev. Neurosci. 2013, 14, 97–111. [Google Scholar] [CrossRef]
  12. Sadoul, K.; Wang, J.; Diagouraga, B.; Khochbin, S. The Tale of Protein Lysine Acetylation in the Cytoplasm. J. Biomed. Biotechnol. 2011, 2011, 970382. [Google Scholar] [CrossRef]
  13. Longworth, M.S.; Laimins, L.A. Histone deacetylase 3 localizes to the plasma membrane and is a substrate of Src. Oncogene 2006, 25, 4495–4500. [Google Scholar] [CrossRef] [Green Version]
  14. Budayeva, H.G.; Cristea, I.M. Human Sirtuin 2 Localization, Transient Interactions, and Impact on the Proteome Point to Its Role in Intracellular Trafficking. Mol. Cell. Proteom. 2016, 15, 3107–3125. [Google Scholar] [CrossRef] [Green Version]
  15. Zhang, Z.-H.; Wang, Z.-H.; Zhang, Z.-R.; Wang, Y.-X. A novel method for apoptosis protein subcellular localization prediction combining encoding based on grouped weight and support vector machine. FEBS Lett. 2006, 580, 6169–6174. [Google Scholar] [CrossRef] [Green Version]
  16. Shi, S.-P.; Qiu, J.-D.; Sun, X.-Y.; Suo, S.-B.; Huang, S.-Y.; Liang, R.-P. A method to distinguish between lysine acetylation and lysine methylation from protein sequences. J. Theor. Biol. 2012, 310, 223–230. [Google Scholar] [CrossRef]
  17. Jiao, Y.-S.; Du, P.-F. Predicting protein sub mitochondrial locations by incorporating the positional-specific physicochemical properties into Chou’s general pseudo-amino acid compositions. J. Theor. Biol. 2017, 416, 81–87. [Google Scholar] [CrossRef]
  18. Chou, K.-C. Some remarks on protein attribute prediction and pseudo amino acid composition. J. Theor. Biol. 2011, 273, 236–247. [Google Scholar] [CrossRef]
  19. Liu, B.; Wu, H.; Chou, K.-C. Pse-in-One 2.0: An Improved Package of Web Servers for Generating Various Modes of Pseudo Components of DNA, RNA, and Protein Sequences. Nat. Sci. 2017, 9, 67–91. [Google Scholar] [CrossRef] [Green Version]
  20. Liu, B.; Liu, F.; Wang, X.; Chen, J.; Fang, L.; Chou, K.-C. Pse-in-One: A web server for generating various modes of pseudo components of DNA, RNA, and protein sequences. Nucleic Acids Res. 2015, 43, W65–W71. [Google Scholar] [CrossRef] [Green Version]
  21. Chou, K.-C. Impacts of Bioinformatics to Medicinal Chemistry. Med. Chem. 2015, 11, 218–234. [Google Scholar] [CrossRef]
  22. Chou, K.-C. Prediction of protein cellular attributes using pseudo-amino acid composition. Proteins Struct. Funct. Genet. 2001, 43, 246–255. [Google Scholar] [CrossRef]
  23. Kaur, H.; Raghava, G.P.S. A neural network method for prediction of -turn types in proteins using evolutionary information. Bioinformatics 2004, 20, 2751–2758. [Google Scholar] [CrossRef]
  24. Chen, Z.; Chen, Y.-Z.; Wang, X.-F.; Wang, C.; Yan, R.-X.; Zhang, Z. Prediction of Ubiquitination Sites by Using the Composition of k-Spaced Amino Acid Pairs. PLoS ONE 2011, 6, e22930. [Google Scholar] [CrossRef] [Green Version]
  25. Papademetriou, R.C. Reconstructing with moments. Proc. Int. Conf. Pattern. Recognit. 1992, 3, 476–480. [Google Scholar]
  26. Butt, A.H.; Khan, S.A.; Jamil, H.; Rasool, N.; Khan, Y.D. A Prediction Model for Membrane Proteins Using Moments Based Features. Biomed. Res. Int. 2016, 2016, 8370132. [Google Scholar] [CrossRef] [Green Version]
  27. Butt, A.H.; Rasool, N.; Khan, Y.D. A Treatise to Computational Approaches Towards Prediction of Membrane Protein and Its Subtypes. J. Membr. Biol. 2017, 250, 55–76. [Google Scholar] [CrossRef]
  28. Han, Z.-J.; Feng, Y.-H.; Gu, B.-H.; Li, Y.-M.; Chen, H. The post-translational modification, SUMOylation, and cancer (Review). Int. J. Oncol. 2018, 52, 1081–1094. [Google Scholar] [CrossRef] [Green Version]
  29. Butt, A.H.; Rasool, N.; Khan, Y.D. Predicting membrane proteins and their types by extracting various sequence features into Chou’s general PseAAC. Mol. Biol. Rep. 2018, 45, 2295–2306. [Google Scholar] [CrossRef]
  30. Butt, A.H.; Rasool, N.; Khan, Y.D. Prediction of antioxidant proteins by incorporating statistical moments based features into Chou’s PseAAC. J. Theor. Biol. 2019, 473, 1–8. [Google Scholar] [CrossRef]
  31. Butt, A.H.; Khan, Y.D. Prediction of S-Sulfenylation Sites Using Statistical Moments Based Features via CHOU’S 5-Step Rule. Int. J. Pept. Res. Ther. 2020, 26, 1291–1301. [Google Scholar] [CrossRef]
  32. Khan, Y.D.; Khan, S.A.; Ahmad, F.; Islam, S. Iris Recognition Using Image Moments and k-Means Algorithm. Sci. World J. 2014, 2014, 723595. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Zhu, H.; Shu, H.; Zhou, J.; Luo, L.; Coatrieux, J.L. Image analysis by discrete orthogonal dual Hahn moments. Pattern Recognit. Lett. 2007, 28, 1688–1704. [Google Scholar] [CrossRef] [Green Version]
  34. Yap, P.-T.; Paramesran, R.; Ong, S.-H. Image Analysis Using Hahn Moments. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 2057–2062. [Google Scholar] [CrossRef]
  35. Kumar, R.; Panwar, B.; Chauhan, J.S.; Raghava, G.P. Analysis and prediction of cancerlectins using evolutionary and domain information. BMC Res. Notes 2011, 4, 237. [Google Scholar] [CrossRef] [Green Version]
  36. Harris, M.A.; Clark, J.; Ireland, A.; Lomax, J.; Ashburner, M.; Foulger, R.; Eilbeck, K.; Lewis, S.; Marshall, B.; Mungall, C. The Gene Ontology (GO) database and informatics resource. Nucleic Acids Res. 2004, 32, D258–D261. [Google Scholar]
  37. Chen, G.; Cao, M.; Luo, K.; Wang, L.; Wen, P.; Shi, S. ProAcePred: Prokaryote lysine acetylation sites prediction based on elastic net feature optimization. Bioinformatics 2018, 34, 3999–4006. [Google Scholar] [CrossRef]
  38. Wuyun, Q.; Zheng, W.; Zhang, Y.; Ruan, J.; Hu, G. Improved species-specific lysine acetylation site prediction based on a large variety of features set. PLoS ONE 2016, 11, e0155370. [Google Scholar] [CrossRef] [Green Version]
  39. Hou, T.; Zheng, G.; Zhang, P.; Jia, J.; Li, J.; Xie, L.; Wei, C.; Li, Y. LAceP: Lysine Acetylation Site Prediction Using Logistic Regression Classifiers. PLoS ONE 2014, 9, e89575. [Google Scholar] [CrossRef] [Green Version]
  40. Li, T.; Du, Y.; Wang, L.; Huang, L.; Li, W.; Lu, M.; Zhang, X.; Zhu, W.G. Characterization and Prediction of Lysine (K)-Acetyl-Transferase Specific Acetylation Sites. Mol. Cell. Proteom. 2012, 11, M111.011080. [Google Scholar] [CrossRef] [Green Version]
  41. Nawaz, S.; Fatima, K.; Ashraf, A. Prediction of Allergen and Non-Allergen Proteins Sequence via Chou’s 5-Step Rule. VFAST Trans. Softw. Eng. 2021, 9. [Google Scholar] [CrossRef]
  42. Ashraf, A.; Ashraf, R.A.R. A Technique for Prediction Cytokines based On Statistical Moments and a Random Forest Classifier. VFAST Trans. Softw. Eng. 2021, 9. [Google Scholar] [CrossRef]
  43. Albugami, N. Prediction of Saudi Arabia SARS-COV 2 diversifications in protein strain against China strain. VAWKUM Trans. Comp. Sci. 2020, 8. [Google Scholar] [CrossRef]
  44. Goh, H.-A.; Chong, C.-W.; Besar, R.; Abas, F.S.; Sim, K.-S. Translation and scale invariants of HAHN moments. Int. J. Image Graph. 2009, 9, 271–285. [Google Scholar] [CrossRef]
  45. Khan, Y.D.; Rasool, N.; Hussain, W.; Khan, S.A.; Chou, K.-C. IPhosYPseAAC: Identify phosphotyrosine sites by incorporating sequence statistical moments into PseAAC. Mol. Biol. Rep. 2018, 45, 2501–2509. [Google Scholar] [CrossRef] [PubMed]
  46. Yang, C.-H.; Lin, Y.-D.; Chuang, L.-Y. TRNAfeature: An algorithm for tRNA features to identify tRNA genes in DNA sequences. J. Theor. Biol. 2016, 404, 251–261. [Google Scholar] [CrossRef]
  47. Akmal, M.A.; Rasool, N.; Khan, Y.D. Prediction of N-linked glycosylation sites using position relative features and statistical moments. PLoS ONE 2017, 12, e0181966. [Google Scholar] [CrossRef] [Green Version]
  48. Khan, Y.D.; Jamil, M.; Hussain, W.; Rasool, N.; Khan, S.A.; Chou, K.-C. PSSbond-PseAAC: Prediction of disulfide bonding sites by integration of PseAAC and statistical moments. J. Theor. Biol. 2019, 463, 47–55. [Google Scholar] [CrossRef]
  49. Khan, Y.D.; Batool, A.; Rasool, N.; Khan, S.A.; Chou, K.-C. Prediction of nitrosocysteine sites using position and composition variant features. Lett. Org. Chem. 2019, 16, 283–293. [Google Scholar] [CrossRef]
  50. Hussain, W.; Khan, Y.D.; Rasool, N.; Khan, S.A.; Chou, K.-C. SPrenylC–PseAAC: A sequence-based model developed via Chou’s 5-steps rule and general PseAAC for identifying S-prenylation sites in proteins. J. Theor. Biol. 2019, 468, 1–11. [Google Scholar] [CrossRef]
  51. Reiss, T.H. Features invariant to linear transformations in 2D and 3D. Proc. Int. Conf. Pattern Recognit. 1992, 3, 493–496. [Google Scholar]
  52. Pawlak, M.; Liao, X. On image analysis by orthogonal moments. Proc. Int. Conf. Pattern Recognit. 1992, 3, 549–552. [Google Scholar]
  53. Awais, M.; Hussain, W.; Khan, Y.D.; Rasool, N.; Khan, S.A.; Chou, K.-C. IPhosH-PseAAC: Identify phosphohistidine sites in proteins by blending statistical moments and position relative features according to the Chou’s 5-step rule and general pseudo amino acid composition. IEEE/ACM Trans. Comput. Biol. Bioinf. 2019. to be published. [Google Scholar] [CrossRef] [PubMed]
  54. Specht, D.F. Probabilistic neural networks. Neural Netw. 1990, 3, 109–118. [Google Scholar] [CrossRef]
  55. Khan, Z.U.; Hayat, M.; Khan, M.A. Discrimination of acidic and alkaline enzyme using pseudo amino acid composition in conjunction with probabilistic neural network model. J. Theor. Biol. 2014, 365, 197–203. [Google Scholar] [CrossRef]
  56. Paliwal, M.; Kumar, U.A. Neural networks and statistical techniques: A review of applications. Expert Syst. Appl. 2009, 36, 2–17. [Google Scholar] [CrossRef]
  57. Huang, Y.; Li, Y. Application of probabilistic neural networks to the class prediction of leukemia and embryonal tumor of central nervous system. Neural Process. Lett. 2004, 19, 211–226. [Google Scholar] [CrossRef]
  58. Hayat, M.; Khan, A. Prediction of membrane protein types by using dipeptide and pseudo amino acid composition based composite features. IET Commun. 2012, 6, 3257–3264. [Google Scholar] [CrossRef]
  59. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  60. Dai, Q.; Ma, S.; Hai, Y.; Yao, Y.; Liu, X. A segmentation based model for subcellular location prediction of apoptosis protein. Chemometr. Intell. Lab. Syst. 2016, 158, 146–154. [Google Scholar] [CrossRef]
  61. Kabir, M.; Hayat, M. iRSpot-GAEnsC: Identifing recombination spots via ensemble classifier and extending the concept of Chou’s PseAAC to formulate DNA samples. Mol. Genet Genom. 2016, 291, 285–296. [Google Scholar] [CrossRef]
  62. Farman, A.; Maqsood, H. Classification of membrane protein types using Voting Feature Interval in combination with Chou׳s Pseudo Amino Acid Composition. J. Theor. Biol. 2015, 384, 78–83. [Google Scholar] [CrossRef]
  63. Ashraf, A.; Muhammad, S.R.; Muhammad, S.A. Identifying Key Genes of Liver Cancer by Using Random Forest Classification. VFAST Trans. Softw. Eng. 2021. [Google Scholar]
  64. Hunter, S.; Jones, P.; Mitchell, A.; Apweiler, R.; Attwood, T.K.; Bateman, A.; Bernard, T.; Binns, D.; Bork, P.; Burge, S.; et al. InterPro in 2011: New developments in the family and domain prediction database. Nucleic Acids Res. 2012, 40, 4725. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Acetylation protein.
Figure 1. Acetylation protein.
Membranes 12 00265 g001
Figure 2. Flowchart of the proposed predictor.
Figure 2. Flowchart of the proposed predictor.
Membranes 12 00265 g002
Figure 3. Schematic view of PNN.
Figure 3. Schematic view of PNN.
Membranes 12 00265 g003
Figure 4. The working mechanism of the Random Forest classifier.
Figure 4. The working mechanism of the Random Forest classifier.
Membranes 12 00265 g004
Figure 5. K-fold cross-validation (KFCV).
Figure 5. K-fold cross-validation (KFCV).
Membranes 12 00265 g005
Figure 6. 10-fold Random Forest ROC curve.
Figure 6. 10-fold Random Forest ROC curve.
Membranes 12 00265 g006
Figure 7. Ten-fold Probabilistic Neural Network ROC curve.
Figure 7. Ten-fold Probabilistic Neural Network ROC curve.
Membranes 12 00265 g007
Figure 8. Jackknife Random Forest ROC curve.
Figure 8. Jackknife Random Forest ROC curve.
Membranes 12 00265 g008
Figure 9. Jackknife PNN ROC curve.
Figure 9. Jackknife PNN ROC curve.
Membranes 12 00265 g009
Figure 10. Self- consistency Random Forest ROC curve.
Figure 10. Self- consistency Random Forest ROC curve.
Membranes 12 00265 g010
Figure 11. Self- consistency Probabilistic Neural Network ROC curve.
Figure 11. Self- consistency Probabilistic Neural Network ROC curve.
Membranes 12 00265 g011
Figure 12. Independent test.
Figure 12. Independent test.
Membranes 12 00265 g012
Figure 13. Independent Random Forest ROC curve.
Figure 13. Independent Random Forest ROC curve.
Membranes 12 00265 g013
Figure 14. Independent Probabilistic Neural Network ROC curve.
Figure 14. Independent Probabilistic Neural Network ROC curve.
Membranes 12 00265 g014
Figure 15. ROC curves through Random Forest.
Figure 15. ROC curves through Random Forest.
Membranes 12 00265 g015
Figure 16. ROC curves through Probabilistic Neural Network.
Figure 16. ROC curves through Probabilistic Neural Network.
Membranes 12 00265 g016
Table 1. Confusion matrix.
Table 1. Confusion matrix.
Status Person Predicted Patient (1)Predicted Healthy Person (0)
Actual patient (1)TPFN
Actual healthy person (0)FPTN
Table 2. Result of 10-fold cross-validation based on the Random Forest classifier.
Table 2. Result of 10-fold cross-validation based on the Random Forest classifier.
10-Fold Cross-Validation Random Forest
Dataset 1Dataset 2Dataset 3
K-FoldsACCSnSpPrecMCCRecallF.mACCSnSpPrecMCCRecallF.mACCSnSpPrec MCCRecallF.m
1100111111100111111100111111
2100111111100111111100111111
3100111111100111111100111111
4100111111100111111100111111
5100111111100111111100111111
6100111111100111111100111111
7100111111100111111100111111
8100111111100111111100111111
9100111111100111111100111111
10100111111100111111100111111
result100111111100111111100111111
Table 3. 10-fold cross-validation Result for Probabilistic neural network.
Table 3. 10-fold cross-validation Result for Probabilistic neural network.
10-Fold Cross-Validation Probabilistic Neural Network
Dataset 1Dataset 2Dataset 3
K-Folds ACCSnSpPrecMCCRecallF.mACCSnSpPrecMCCRecallF.mACCSnSpPrecMCCRecallF.m
Final Score66.830.720.600.650.360.950.72600.260.930.810.210.260.4057.170.920.220.540.360.920.72
Table 4. Jackknife test score based on Random Forest.
Table 4. Jackknife test score based on Random Forest.
PredictonJackknife Random Forest
Dataset 1Dataset 2Dataset 3
FoldACCSnSpPreMCCRecallF.mACCSnSpPrecMCCRecallF.mACCSnSpPrec MCCRecallF.m
Result1000.550.50.550.050.0030.0199.860.550.50.550.050.549310.5599.860.550.50.550.050.549310.5
Table 5. Jackknife score based on the Probabilistic Neural Network.
Table 5. Jackknife score based on the Probabilistic Neural Network.
PredictionJackknife Probabilistic Neural Network
Dataset 1Dataset 2Dataset 3
JackknifeACCSnSpPrecMCCRecallF.mACCSnSpPrecMCCRecallF.mACCSnSpPrecMCCRecallF.m
Final Score66.870.550.50.50.60.410.4259.77o.50.50.50.60.130.2057.410.550.40.50.60.500.48
Table 6. Results of self-consistency based on Random Forest.
Table 6. Results of self-consistency based on Random Forest.
Self-Consistency Random Forest
Dataset 1Dataset 2Dataset 3
ACCSnSpPrecMCCRecallF.mACCSnSpPrecMCCRecallF.mACCSnSpPrecMCCRecallF.m
10010.9910.990.99711001110.990.9971100111111
Table 7. Self-consistency test result for probabilistic neural network.
Table 7. Self-consistency test result for probabilistic neural network.
Self-Consistency Probabilistic Neural Network
Dataset 1Dataset 2Dataset 3
ACCSnSpPrecMCCRecallF.mACSnSpPrecMCCRecallF.mACCSnSpPrecMCCRecallF.m
66.830.720.600.640.360.720.68600.260.930.800.200.260.3957.170.920.220.540.360.921.84
Table 8. Results of independent test based on the Random Forest.
Table 8. Results of independent test based on the Random Forest.
Independent Test Results Random Forest
Dataset 1Dataset 2Dataset 3
Training DatasetTraining DatasetTraining Dataset
ACCSnSpPrecMCCRecallF.mACCSn SpPrecMCCRecallF.mACCSnSpPrecMCCRecallF.m
9711110.96919611110.9719611110.951
Testing Dataset Testing DatasetTesting Dataset
ACCSnSpPrecMCCRecallF.mACCSn SpPrecMCCRecallF.mACCSnSpPrecMCCRecallF.m
9811110.9619511110.9319711110.961
Table 9. Result of independent test based on Probabilistic Neural Network.
Table 9. Result of independent test based on Probabilistic Neural Network.
Independent Test Result Probabilistic Neural Network
Dataset 1Dataset 2Dataset 3
Independent Training Dataset Confusion MatrixIndependent Training Dataset Confusion MatrixIndependent Training Dataset Confusion Matrix
ACC SnPrecMCCRecallF.mACCSnSpPrecMCCRecallF.mACCSnSpPrecMCCRecallF.m
50.80.110.60.1150.2052.60.70.30.530.050.750.6251.3310.110.50.970.98
Independent Testing Dataset Confusion MatrixIndependent Testing Dataset Confusion MatrixIndependent Testing Dataset Confusion Matrix
ACCSnPrecMCCRecallF.mACCSnSpPrecMCCRecallF.mACCSnSpPrecMCCRecallF.m
50.8110.60.920.9654.00.20.90.540.060.192.8651.0310.910.60.070.13
Table 10. Performance of proposed model based on RF and PNN.
Table 10. Performance of proposed model based on RF and PNN.
PredictionComparative Analysis of RF and PNN
Dataset 1Dataset 2Dataset 3
ClassifierACCSnSpPrecMCCRecallF.mACCSnSpPrec MCCRecallF.mACCSnSpPrecMCCRecallF.m
RF10011111111111111111111
PNN66.830.720.60.650.360.950.72600.260.930.810.210.260.4057.170.920.220.540.360.920.72
Table 11. Comparative analysis of the proposed acetylation model with the existing models.
Table 11. Comparative analysis of the proposed acetylation model with the existing models.
Prediction ModelsACC%MCC%Sn%Sp%Prec%F.m%
All-Mean JK 74.64% 0.4980 81.38%67.91%71.78%76.24%
iAcet-PseFDA77.55% 0.5883 96.41%71.26%52.79%68.23%
InterPro68.25% 0.365871.40% 65.10% 67.17% 69.22%
iAcety–SmRF1001.01.01.01.01.0
iAcety–SmPNN66.830.360.720. 600.650.72
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Malebary, S.; Rahman, S.; Barukab, O.; Ash’ari, R.; Khan, S.A. iAcety–SmRF: Identification of Acetylation Protein by Using Statistical Moments and Random Forest. Membranes 2022, 12, 265. https://doi.org/10.3390/membranes12030265

AMA Style

Malebary S, Rahman S, Barukab O, Ash’ari R, Khan SA. iAcety–SmRF: Identification of Acetylation Protein by Using Statistical Moments and Random Forest. Membranes. 2022; 12(3):265. https://doi.org/10.3390/membranes12030265

Chicago/Turabian Style

Malebary, Sharaf, Shaista Rahman, Omar Barukab, Rehab Ash’ari, and Sher Afzal Khan. 2022. "iAcety–SmRF: Identification of Acetylation Protein by Using Statistical Moments and Random Forest" Membranes 12, no. 3: 265. https://doi.org/10.3390/membranes12030265

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop