Next Article in Journal
Recent Achievements in the Development of Biomaterials Improved with Platelet Concentrates for Soft and Hard Tissue Engineering Applications
Next Article in Special Issue
Comprehensive Research on Druggable Proteins: From PSSM to Pre-Trained Language Models
Previous Article in Journal
Longitudinal NMR-Based Metabolomics Study Reveals How Hospitalized COVID-19 Patients Recover: Evidence of Dyslipidemia and Energy Metabolism Dysregulation
Previous Article in Special Issue
Enhancer Recognition: A Transformer Encoder-Based Method with WGAN-GP for Data Augmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Attenphos: General Phosphorylation Site Prediction Model Based on Attention Mechanism

Qingdao Institute of Software, College of Computer Science and Technology, China University of Petroleum, Qingdao 266555, China
*
Author to whom correspondence should be addressed.
Int. J. Mol. Sci. 2024, 25(3), 1526; https://doi.org/10.3390/ijms25031526
Submission received: 3 January 2024 / Revised: 18 January 2024 / Accepted: 23 January 2024 / Published: 26 January 2024
(This article belongs to the Special Issue Deep Learning in Bioinformatics and Biological Data Analysis)

Abstract

:
Phosphorylation site prediction has important application value in the field of bioinformatics. It can act as an important reference and help with protein function research, protein structure research, and drug discovery. So, it is of great significance to propose scientific and effective calculation methods to accurately predict phosphorylation sites. In this study, we propose a new method, Attenphos, based on the self-attention mechanism for predicting general phosphorylation sites in proteins. The method not only captures the long-range dependence information of proteins but also better represents the correlation between amino acids through feature vector encoding transformation. Attenphos takes advantage of the one-dimensional convolutional layer to reduce the number of model parameters, improve model efficiency and prediction accuracy, and enhance model generalization. Comparisons between our method and existing state-of-the-art prediction tools were made using balanced datasets from human proteins and unbalanced datasets from mouse proteins. We performed prediction comparisons using independent test sets. The results showed that Attenphos demonstrated the best overall performance in the prediction of Serine (S), Threonine (T), and Tyrosine (Y) sites on both balanced and unbalanced datasets. Compared to current state-of-the-art methods, Attenphos has significantly higher prediction accuracy. This proves the potential of Attenphos in accelerating the identification and functional analysis of protein phosphorylation sites and provides new tools and ideas for biological research and drug discovery.

1. Introduction

Protein is an important living organism component, regulating life activities, controlling growth and energy metabolism, etc. Post-translational modifications (PTMs) are important cellular control mechanisms that increase the functional diversity of the proteome through the covalent addition of functional groups or proteins, the hydrolytic cleavage of regulatory subunits, or the degradation of whole proteins [1]. Protein phosphorylation is the most studied PTM, and the most studied phosphorylation to date is on serine (S), threonine (T), and tyrosine (Y). Phosphorylation site prediction can help researchers understand protein function and interactions, particularly in the areas of signal transduction and metabolic pathways [2]. For example, it can help researchers to study the relationship between proteins and biological processes such as cell cycle, apoptosis, and cell proliferation. And phosphorylation site prediction can provide important information for protein structure studies, as phosphorylation can change the conformation and charge of proteins and can help researchers predict protein secondary structures, tertiary structures, and interactions [3]. In addition, some studies have shown that more than 30% of eukaryotic proteins can be phosphorylated, and most of these proteins are closely associated with the development of human diseases, such as cancer [4]. Therefore, the study of the principles, types, and mechanisms of protein phosphorylation is important for understanding the pathogenesis of human diseases. Research on protein phosphorylation site prediction can also help to identify new drug targets, thus providing new ideas and directions for disease treatment.
Early researchers performed many bioassay identification methods to identify phosphorylation sites, such as low-throughput 32P labeling [5] and high-throughput mass spectrometry [6], but such methods are quite time-consuming and laborious. With the rapid development of natural language processing, artificial intelligence, and other disciplines, and the integration of various new technologies with bioinformatics, the difficulties of traditional experimental methods have been effectively solved. Dozens of different methods, including machine learning, statistics, feature extraction, and other techniques [7,8,9], have been used to predict and identify phosphorylation sites. For example, Musite, proposed in 2010, used multiple features and algorithms to improve prediction accuracy by using local amino acid sequence frequencies, k-nearest neighbor features, and protein disorder scores. This was followed by MusiteDeep [10], the first deep learning framework for predicting general and kinase-specific phosphorylation sites, which was proposed by the team in 2017, using a convolutional neural network with a novel two-dimensional attention mechanism for prediction. In 2019, Fenglin Luo et al. proposed DeepPhos [11], which utilizes densely connected convolutional neuron network blocks to capture multiple representations of sequences. Final phosphorylation prediction was performed through intra-block crosstalk layers and inter-block crosstalk layers. These deep learning architectures for phosphorylation site prediction are far superior to traditional experimental methods. However, there is still room for improvement in prediction accuracy and prediction efficiency. It is essential to develop a prediction tool that performs well at S, T, and Y sites.
These deep learning architectures for phosphorylation site prediction are far superior to traditional experimental methods. However, there are more negative samples than positive samples for almost all PTM types, including protein phosphorylation [10], which makes the performance of these methods limited. By training with unbalanced datasets, the model can have better generalization ability in real situations. In addition, there is still a lot of room for improvement in the generalization ability of these models, which are not yet able to achieve a high level of prediction performance on S, T, and Y sites with small sample sizes at the same time. Therefore, there is still room for further improvement in terms of prediction accuracy and efficiency. Therefore, it has become crucial to develop a prediction tool that can perform well on both balanced and unbalanced datasets with higher accuracy and better generalization ability.
In response to the above challenges, this study proposes a new deep learning framework to accurately predict general phosphorylation sites using protein sequence information with good robustness and generalization ability. Unlike previous deep learning prediction methods, Attenphos consists of self-attention and CNN blocks that use three structurally identical modules with different window sizes to capture significant sequence representations of protein phosphorylation sites. The feature vector output of the attention module is converted using the Encoding conversion module, increasing the representation capability of the neural network. The transformed coded feature matrix has more structural information and learns features more fully than if the output features were processed directly. The final three outputs are further integrated to make the final prediction. To evaluate the performance of Attenphos, we collected a large number of validated phosphorylation examples for model training and evaluation, and we conducted experimental comparisons with some other prediction tools and deep learning methods. The experimental results show that our proposed method outperforms some existing prediction tools on both balanced and unbalanced datasets. Attenphos achieves the best overall prediction performance on T, S, and Y sites with small sample sizes.

2. Results

2.1. Comparison Experiments on Balanced Datasets

In this section, we provide a discussion of the comparative performance of Attenphos for general phosphorylation site prediction using a balanced dataset from human protein phosphorylation. We compared Attenphos with several existing general phosphorylation site prediction tools, including GPS 2.1 [12], NetPhos 3.0 [13], PPRED [14], Musite [15], PhosphoSVM [16], SKIPHOS [17], Deepphos [11], and Transphos [18]. We performed 10-fold cross-validation to train the model using the training datasets of the S, T, and Y sites. We tested it using an independent PPA test set. As shown in Table 1, the results indicate that our method obtains the best overall performance in S and T bit prediction. And on the Y dataset with a smaller sample size, our method still outperforms existing state-of-the-art deep learning architectures to obtain the best overall performance.
Specifically, at the S site, our model obtains the highest Mcc value of 0.443 and the highest Auc value of 0.788, while other metrics such as Sn, Sp, and F1 values are also in the lead. In the case of the Mcc metric, we achieve an improvement of 11% compared to the state-of-the-art prediction tool Transphos. At the T site, our method shows the best performance, with Attenphos obtaining the highest Mcc and Auc values compared to other methods. Compared to GPS 2.1, NetPhos 3.0, PPRED, Musite, PhosphoSVM, SKIPHOS, Deepphos, and Transphos, our architecture improves the Auc values by 2.1%, 5.1%, 11.3%, 6.9%, 2.6%, 4.8%, 1.7%, and 1.9%, respectively, and obtains the highest Mcc value (0.254). At the Y site, our method obtains the Auc value of 0.625, second only to the Auc value of SKIPHOS (0.634), and the highest Mcc value (0.201). Compared to state-of-the-art deep learning prediction tools such as Deepphos and Transphos, Attenphos improves a lot on Auc metrics, especially on Mcc metrics. To summarize, our model Attenphos performs well in phosphorylation prediction at the S, T, and Y sites, showing the best overall results, with Attenphos being able to accurately predict phosphorylation sites compared to existing state-of-the-art tools.

2.2. Comparison Experiments on Unbalanced Datasets

As shown in Table 2, we conducted performance comparison experiments on a dataset with positive and negative sample unbalance distribution, which is a phosphorylation dataset from mouse proteins. In order to assess the generalization capability of our model on unbalanced datasets with limited samples, we specifically selected the advanced deep learning models Transphos and Deepphos for a focused comparison. Both Transphos and Deepphos perform well on the balanced dataset. We trained and validated individual models, including Deepphos, Transphos, and Attenphos. Then, we tested them using independent test sets. Considering the unbalance in the number of positive and negative samples, we used evaluation metrics of SN, SP, MCC, AUC, and F1 scores for comparison. The results are shown in the table above. At the T site, Attenphos achieves the highest SP, AUC, MCC, SN, and F1 values of 87.79%, 0.775, 0.416, 50.91%, and 0.566, respectively. Compared with Transphos, its AUC value is increased by 0.9% and its MCC value by 2.2%. At the Y site, Attenphos achieves the highest SP, SN, AUC, F1, and MCC values. Attenphos obtains the highest MCC and F1 values in S-site phosphorylation prediction. Compared to Transphos, Attenphos shows an increase of 0.5% in the AUC value and 0.3% in the F1 score value. In addition, the MCC, SN, and SP values also perform better. Our research findings unveil the superiority of Attenpphos over Deepphos and Transphos, as confirmed by the ROC curve analysis. The ROC curves of Attenpphos consistently exhibit a larger area under the curve (AUC), indicating the better classification performance of our model compared to the other two models. The ROC graph is shown in Figure 1 below. This indicates that our model can still show excellent performance in the presence of positive and negative sample unbalance.
In real-world protein phosphorylation problems, where the distributions of positive and negative samples are often unbalanced, our model demonstrates significant performance improvement in phosphorylation prediction at the S, T, and Y sites, achieving higher overall performance. This suggests that Attenphos is not only applicable to datasets with balanced distributions but also to datasets with unbalanced distributions that are close to the real situation.

2.3. Ablation Experiments

We investigated the contribution of the Encoding conversion block to general phosphorylation site prediction. Figure 2 shows that removing the Encoding conversion block in the Attenphos architecture decreases the overall performance of general phosphorylation site prediction. For example, at the S site, removing the Encoding conversion causes the Sp, F1, Auc, and Mcc values to be reduced at the S site. At the T site, the removal of Encoding conversion reduces the values of Sn, F1, Auc, and Mcc. At the Y site, the removal of Encoding conversion reduces the values of Sn, F1, Auc, and Mcc. Therefore, the use of the Encoding conversion module allows the model to better capture the relational features prior to the amino acids, focusing on key information and thus improving the overall predictive performance of the model.
In this section, we performed ablation experiments on the effect of the attention mechanism in the Attenphos method. The experiment results are shown in Table 3 below. As shown in the table, the performance of Attenphos-noAttention undergoes a large decrease compared to Attenphos at the S, T, and Y sites. Although the F1 score improves by 1% and 0.6% at the S and T sites, other metrics such as Mcc and Auc show a large decrease. At the Y site, Attenphos-noAttention decreases by 6.1% in AUC, 7.2% in mcc, and 5.1% in F1 score. Self-attention mechanisms can focus on the relationship between different amino acids. The self-attention mechanism module enables our model to better and more efficiently capture the long-range dependencies of protein sequences and extract global feature information, thus improving the prediction accuracy.

2.4. Feature Visualization

To more visually demonstrate the effectiveness of the method in this paper for the classification of phosphorylated versus non-phosphorylated sites, the features extracted by Attenphos were visualized. We chose a commonly used dimensionality reduction visualization method, t-SNE (t-Distributed Stochastic Neighbor Embedding) [19], which can be used to map high-dimensional data into a low-dimensional space and preserve the relative distance relationship between the data. It can reduce the dimensionality of the high-dimensional features and visualize the results. Figure 3 shows the visualization of the original features encoded by embedding and the features extracted by Attenphos. It is clear that the features encoded by embedding are visualized in a cluttered manner. However, the extracted features of Attenphos can easily separate phosphorylated and unphosphorylated samples, indicating that our method can predict the generalized phosphorylated site with high accuracy.

3. Discussion

We propose a new deep learning method named Attenphos. It can predict potential general phosphorylation sites. And experiments comparing our model with existing methods show that our method has better overall performance compared to other existing methods. The protein sequences are centered on the prediction sites, intercepting short sequences with different window sizes. Short sequences are encoded and input into three modules with the same structure, in which the output features of the self-attention block are transformed and encoded. Then, when convolution is used, effective features can be better extracted. Finally, the three parts are combined into a combined feature to fully improve the generalization ability of the enhanced model and the prediction of the model accuracy. And self-attention can effectively solve the problem of long-distance dependence on protein phosphorylation.
Although our method has achieved better performance, there are still some shortcomings for improvement. (1) In future work, we can improve the framework using some deep learning modules that work well for classification, such as BiGRU [20], which can be explored in the next work. (2) We could attempt to include information other than the protein sequence. For example, protein structure information can provide information about the environment of amino acid residues, such as residue accessibility, local secondary structure, inter-residue distance, etc. PPI [21] information can provide association information in protein interaction networks, such as the tendency of proteins in the same signaling pathway to share phosphorylation sites. So, we could try to combine protein sequence, structure, and PPI information and use deep learning models to perform prediction, which can be further explored in future work. (3) Due to the small kinase-specific phosphorylation dataset available and the small number of samples, our model does not have good prediction performance on small samples, and more advanced model architecture is needed for kinase phosphorylation prediction, which will be the goal of our next work.
In summary, our deep learning method Attenphos shows superior performance in protein general phosphorylation site prediction with high accuracy and stability. The method is based on an attention mechanism which works by filtering out the most predictive features and optimal model parameters for prediction. Our method can provide a useful tool not only in basic research but also in biomedical and drug discovery applications, helping to accelerate the identification and study of phosphorylation sites. This study provides a strong reference for future bioinformatics and proteomics studies and offers new ideas for understanding the mechanisms of protein phosphorylation regulation.

4. Material and Methods

4.1. Datasets

This study is based on the prediction of general phosphorylation sites. We used the phosphorylation site training dataset provided by Phospho.ELM [22], a database containing experimentally validated, manually curated phosphorylation sites. By integrating information from literature reports and experimental data, the Phospho.ELM database provides extensive annotations of phosphorylation sites and related biological information. We used human protein phosphorylation data and mouse protein phosphorylation data from this database.
For mouse phosphorylation data, as shown in Table 4, our study used the mouse protein phosphorylation dataset from the PELM database and removed redundant data using the CD-HIT tool [23] with a redundancy threshold set at 40%. Since the number of positive and negative samples for phosphorylation is not balanced in the real world, we chose negative samples in this study at a ratio of 1:2 for positive and negative samples. The negative samples were randomly selected on the same protein sequence, excluding identical amino acids outside the labeled phosphorylation sites. Firstly, the dataset was randomly divided into an 80% training set and 20% independent test set, and secondly, 10% of the training set was used as a validation set during training to prevent overfitting.
For phosphorylation data from human proteins, we used the same training dataset, validation set, and independent test set as in Transphos [18]. The number of positive and negative samples in the dataset was balanced. Among them, the training dataset was obtained from the human phosphorylation data in the Phospho.ELM database, as shown in Table 5. Meanwhile, the independent test set was obtained from the PhosPhAt (PPA) [24,25,26] database, as shown in Table 6. In the above way, we constructed the phosphorylation site datasets for training and evaluation to support our research work.
In order to better extract the features near the phosphorylation site, we separately cut the original sequence into short sequences of different lengths centered on the prediction site for experiments. We explored many window combinations by conducting experiments and finally chose the optimal input window combination of 21, 33, and 51. We chose sequence fragments with left and right lengths of 10, 16, and 25, centered on the prediction site. If the left and right lengths of the predicted site were insufficient, * was filled in to ensure that each subsequence had the same length, and the negative samples were manipulated as above. We digitally encoded the short protein sequences. Then, the encoded sequence digital representation was encoded using embedding. When encoding with one-hot, it cannot represent the relationship between adjacent amino acids very well, meaning that the efficiency of training is not very high. The use of embedding can optimize this problem very well. The feature vectors after embedding encoding were 21 × 16, 33 × 16, and 51 × 16.

4.2. Model Architecture

The deep learning architecture of Attenphos is shown in Figure 4. We first processed the protein sequence length and extracted short sequences with window sizes of 21, 33, and 51. Experiments have shown that the length combinations of 21, 33, and 51 are the most appropriate. Then, we encoded short sequences by embedding and input the encoded feature vectors to three structurally identical self-attention blocks for feature extraction. Using self-attention could fully obtain the long-range dependence information of protein sequences. Then, the output features were separately transformed and encoded using the Encoding conversion module, converting each feature of length 16 into 4 × 4 features to better learn the relationship features between amino acids. Then, the transformed feature vectors were input into a one-dimensional convolutional block. The resulting three-part output vectors were combined into an integrated feature. Then, it was input to a flatten layer, transformed into a one-dimensional tensor. And then, it passed through several layers of fully connected layers. Finally, the output results were obtained by softmax.

4.2.1. Self-Attention

For neural networks, the input vectors are generally of different sizes. There is a certain relationship between different vectors, which cannot be reflected by the embedding vectors. But the actual training cannot give full play to the relationship among these inputs, resulting in very poor model training results.
The self-attention mechanism is introduced in this model, the specific structure of which is shown in Figure 1a, and the self-attention mechanism is one of the attention mechanisms. The problem of poor model performance due to the inability to establish a correlation between the relevant inputs above can be solved by this. This allows our model to notice correlations between different parts of the whole input and to learn the importance of the features, providing more weight to the more important features to represent the sequence.
In the self-attention mechanism, the input sequence is first subjected to feature extraction. It usually uses a linear transformation to map the input sequence into the Query, Key, and Value vector spaces. Specifically, for the input sequence X = {x1, x2, …, xn}, Query, Key, and Value can be generated in the following ways:
Q = X W Q
K = X W k
V = X W V
where X denotes the input sequence; W Q , W k , and W V is the learnable weight matrix; and Q, K, and V denote the generated Query, Key, and Value, respectively.
s i j = q i k j
α i j = exp ( s i j / d k ) j = 1 n exp s i j d k
where q i denotes the i-th element of the Query vector, k j denotes the j-th element of the Key vector, d k denotes the dimension of the Key vector, and α i j denotes the attention weight between position i and position j.
o i = j = 1 n α i j v j
where v j denotes the j-th element of the Value vector and o i denotes the i-th element of the Output vector.
In our model, we used a self-attention layer with 16 input units and output features of the win (window length) ×16, which allowed the correlation between the features of each amino acid to be better expressed.

4.2.2. Encoding Conversion

We transformed the feature vector after the output of self-attention by changing the features of length 16 into 4 × 4 features, as shown in Figure 4b and Figure 5a,b. This is because turning the 16-length features into 4 × 4 features increased the representation capability of the neural network. The 4 × 4 feature matrix had more structural information. This approach can better take advantage of the one-dimensional convolutional layer when processing sequential data to improve the model efficiency and the generalization capability of the model, and it also can reduce the number of model parameters. For example, if the original data are considered as image data, a higher dimensional feature matrix retains more detailed information in the image and helps the network to better learn the features in the image. This transformation encoding the features in a way that allows for better learning of amino acid sequences. The transformed feature vectors were fed into a one-dimensional convolutional layer with the stride set to 4 and the convolutional kernel set to 8. This configuration allowed the relationships between amino acids to be fully learned and more detailed information to be retained. We verified through experimental results that this method helped in the extraction of model features and improved the accuracy of the model prediction.

4.2.3. Model Training

Attenphos is a deep learning framework designed to predict sites in protein sequences where phosphorylation is likely to occur. The framework uses a self-attention mechanism to learn the interactions between amino acid residues to extract effective feature vectors for phosphorylation site prediction. To prevent overfitting, Attenphos uses dropout layers in the convolutional and fully connected layers. The model uses the categorical cross-entropy loss function, which is minimized by the Adam algorithm, and L2 regularization, which adjusts the model parameters to better fit the data in training. The specific hyper-parameter choices are shown in Table 7.
L o s s c = 1 N j = 1 N y i ln P ( y i = 1 |   x j ) + ( 1 y i ) ln P ( y i = 0 |   x j )
N is the number of samples, x j refers to the jth input localized sequence, and y j refers to the corresponding phosphorylated state label of the jth input sequence.

4.2.4. Metrics

Several commonly used measurements were used to assess the performance of Attenphos, including the area under the ROC curve (AUC), sensitivity (Sn), specificity (Sp), F1 score, and the Matthew correlation coefficient (Mcc). The calculation formulas are shown below:
S n = T P T P + F N
S p = T N T N + F P
F 1 = 2 × P r e × S n P r e + S n
M c c = T P × T N F P × F N T P + F N × T P + F P × T N + F N × T N + F P
TP (true positive): a prediction of 1 and a true value of 1, i.e., true positive; FP (false positive): a prediction of 1 and a true value of 0, i.e., false positive; TN (true negative): a prediction of 0 and a true value of 0, i.e., true negative; FN (false negative): a prediction of 0 and a true value of 1, i.e., false negative.

5. Conclusions

In this study, we propose a new framework, Attenphos, based on deep learning to accurately predict general phosphorylation sites using protein sequence information. Attenphos has good robustness and generalization ability. Unlike previous deep learning prediction methods, Attenphos consists of self-attention and CNN modules that use three modules with the same structure and different window sizes to capture important sequence representations of protein phosphorylation sites. On the balanced dataset, Attenphos achieved AUC values of 0.788, 0.691, and 0.625 at the S, T, and Y sites, respectively. On the unbalanced dataset for mouse species, AUC values of 0.831, 0.775, and 0.658 and MCC values of 0.511, 0.416, and 0.262 were achieved at the S, T, and Y sites, respectively. It was demonstrated that Attenphos achieved better results on both balanced and unbalanced datasets. The experiment proves that our model outperforms other existing methods and deep learning models in all aspects. Also, we experimentally proved the importance of the Encoding conversion block. Adding the Encoding conversion block can significantly improve the performance of the model. We chose a commonly used downscaling visualization method, t-SNE, and the features extracted by our method, Attenphos, can easily separate phosphorylated sites from unphosphorylated sites compared to the features encoded by embedding.

Author Contributions

Conceptualization, X.W.; methodology, Q.Y.; validation, P.Q. and L.Q.; supervision, T.S.; investigation, Q.Y.; writing—original draft preparation, Q.Y. and X.W.; visualization, P.Q. All authors have read and agreed to the published version of the manuscript.

Funding

We thank the anonymous reviewers for their valuable feedback and insightful comments. This work was funded by the National Natural Science Foundation of China (No. 61972416), Natural Science Foundation of Shandong Province (No. ZR2022LZH009) and National Key R&D Program of China (No. 2021YFA1000103-3).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The Attenphos project’s source code and datasets are publicly available at https://github.com/uuy99/Attenphos, accessed on 18 January 2024.

Acknowledgments

We thank our partners who provided all the help during the research process and the team for their great support.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Mann, M.; Jensen, O.N. Proteomic analysis of post-translational modifications. Nat. Biotechnol. 2003, 21, 255–261. [Google Scholar] [CrossRef] [PubMed]
  2. Cohen, P.T. Protein phosphatase 1–targeted in many directions. J. Cell Sci. 2002, 115, 241–256. [Google Scholar] [CrossRef] [PubMed]
  3. Groban, E.S.; Narayanan, A.; Jacobson, M.P. Conformational changes in protein loops and helices induced by post-translational phosphorylation. PLoS Comput. Biol. 2006, 2, e32. [Google Scholar] [CrossRef] [PubMed]
  4. Li, F.; Li, C.; Marquez-Lago, T.T.; Leier, A.; Akutsu, T.; Purcell, A.W.; Smith, A.I.; Lithgow, T.; Daly, R.J.; Song, J. Quokka: A comprehensive tool for rapid and accurate prediction of kinase family-specific phosphorylation sites in the human proteome. Bioinformatics 2018, 34, 4223–4231. [Google Scholar] [CrossRef]
  5. Aponte, A.M.; Phillips, D.; Harris, R.A.; Blinova, K.; French, S.; Johnson, D.T.; Balaban, R.S. 32P labeling of protein phosphorylation and metabolite association in the mitochondria matrix. Methods Enzymol. 2009, 457, 63–80. [Google Scholar] [PubMed]
  6. Beausoleil, S.A.; Villén, J.; Gerber, S.A.; Rush, J.; Gygi, S.P. A probability-based approach for high-throughput protein phosphorylation analysis and site localization. Nat. Biotechnol. 2006, 24, 1285–1292. [Google Scholar] [CrossRef] [PubMed]
  7. Trost, B.; Maleki, F.; Kusalik, A.; Napper, S. DAPPLE 2: A tool for the homology-based prediction of post-translational modification sites. J. Proteome Res. 2016, 15, 2760–2767. [Google Scholar] [CrossRef]
  8. Qin, G.-M.; Li, R.-Y.; Zhao, X.-M. PhosD: Inferring kinase–substrate interactions based on protein domains. Bioinformatics 2017, 33, 1197–1204. [Google Scholar] [CrossRef]
  9. Huang, S.-Y.; Shi, S.-P.; Qiu, J.-D.; Liu, M.-C. Using support vector machines to identify protein phosphorylation sites in viruses. J. Mol. Graph. Model. 2015, 56, 84–90. [Google Scholar] [CrossRef]
  10. Wang, D.; Zeng, S.; Xu, C.; Qiu, W.; Liang, Y.; Joshi, T.; Xu, D. MusiteDeep: A deep-learning framework for general and kinase-specific phosphorylation site prediction. Bioinformatics 2017, 33, 3909–3916. [Google Scholar] [CrossRef]
  11. Luo, F.; Wang, M.; Liu, Y.; Zhao, X.-M.; Li, A. DeepPhos: Prediction of protein phosphorylation sites with deep learning. Bioinformatics 2019, 35, 2766–2773. [Google Scholar] [CrossRef] [PubMed]
  12. Xue, Y.; Ren, J.; Gao, X.; Jin, C.; Wen, L.; Yao, X. GPS 2.0, a tool to predict kinase-specific phosphorylation sites in hierarchy. Mol. Cell. Proteom. 2008, 7, 1598–1608. [Google Scholar] [CrossRef] [PubMed]
  13. Blom, N.; Gammeltoft, S.; Brunak, S. Sequence and structure-based prediction of eukaryotic protein phosphorylation sites. J. Mol. Biol. 1999, 294, 1351–1362. [Google Scholar] [CrossRef] [PubMed]
  14. Basu, S.; Plewczynski, D. AMS 3.0: Prediction of post-translational modifications. BMC Bioinform. 2010, 11, 210. [Google Scholar] [CrossRef] [PubMed]
  15. Gao, J.; Thelen, J.J.; Dunker, A.K.; Xu, D. Musite, a tool for global prediction of general and kinase-specific phosphorylation sites. Mol. Cell. Proteom. 2010, 9, 2586–2600. [Google Scholar] [CrossRef] [PubMed]
  16. Dou, Y.; Yao, B.; Zhang, C. PhosphoSVM: Prediction of phosphorylation sites by integrating various protein sequence attributes with a support vector machine. Amino Acids 2014, 46, 1459–1469. [Google Scholar] [CrossRef] [PubMed]
  17. Dang, T.H.; Trac, Q.T.; Phan, H.K.; Nguyen, M.C.; Thi, Q.T.P. SKIPHOS: Non-kinase specific phosphorylation site prediction with random forests and amino acid skip-gram embeddings. BioRxiv 2019, 793794. [Google Scholar] [CrossRef]
  18. Wang, X.; Zhang, Z.; Zhang, C.; Meng, X.; Shi, X.; Qu, P. Transphos: A deep-learning model for general phosphorylation site prediction based on transformer-encoder architecture. Int. J. Mol. Sci. 2022, 23, 4263. [Google Scholar] [CrossRef]
  19. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  20. Lin, X.; Quan, Z.; Wang, Z.-J.; Huang, H.; Zeng, X. A novel molecular representation with BiGRU neural networks for learning atom. Brief. Bioinform. 2020, 21, 2099–2111. [Google Scholar] [CrossRef]
  21. Jones, S.; Thornton, J.M. Principles of protein-protein interactions. Proc. Natl. Acad. Sci. USA 1996, 93, 13–20. [Google Scholar] [CrossRef]
  22. Dinkel, H.; Chica, C.; Via, A.; Gould, C.M.; Jensen, L.J.; Gibson, T.J.; Diella, F. Phospho.ELM: A database of phosphorylation sites—Update 2011. Nucleic Acids Res. 2010, 39 (Suppl. S1), D261–D267. [Google Scholar] [CrossRef]
  23. Fu, L.; Niu, B.; Zhu, Z.; Wu, S.; Li, W. CD-HIT: Accelerated for clustering the next-generation sequencing data. Bioinformatics 2012, 28, 3150–3152. [Google Scholar] [CrossRef]
  24. Durek, P.; Schmidt, R.; Heazlewood, J.L.; Jones, A.; MacLean, D.; Nagel, A.; Kersten, B.; Schulze, W.X. PhosPhAt: The Arabidopsis thaliana phosphorylation site database. An update. Nucleic Acids Res. 2010, 38 (Suppl. S1), D828–D834. [Google Scholar] [CrossRef]
  25. Heazlewood, J.L.; Durek, P.; Hummel, J.; Selbig, J.; Weckwerth, W.; Walther, D.; Schulze, W.X. PhosPhAt: A database of phosphorylation sites in Arabidopsis thaliana and a plant-specific phosphorylation site predictor. Nucleic Acids Res. 2007, 36 (Suppl. S1), D1015–D1021. [Google Scholar] [CrossRef]
  26. Zulawski, M.; Braginets, R.; Schulze, W.X. PhosPhAt goes kinases—Searchable protein kinase target information in the plant phosphorylation site database PhosPhAt. Nucleic Acids Res. 2012, 41, D1176–D1184. [Google Scholar] [CrossRef] [PubMed]
Figure 1. ROC curve plots of Attenphos, Transphos, and Deepphos on the mouse dataset ((ac) indicate the S, T, and Y sites).
Figure 1. ROC curve plots of Attenphos, Transphos, and Deepphos on the mouse dataset ((ac) indicate the S, T, and Y sites).
Ijms 25 01526 g001
Figure 2. Comparison of ablation experiment results. (a), (b), and (c) indicate the comparison of experimental results at the S site, T site, and Y site, respectively.
Figure 2. Comparison of ablation experiment results. (a), (b), and (c) indicate the comparison of experimental results at the S site, T site, and Y site, respectively.
Ijms 25 01526 g002
Figure 3. t-SNE visualization of Attenphos extracted features and embedding encoded features. (a), (b), and (c) represent the visualization plots after feature extraction using Attenphos at the S, T, and Y sites, respectively, and (d), (e), and (f) represent the feature visualization plots after encoding using embedding at the S, T, and Y sites, respectively.
Figure 3. t-SNE visualization of Attenphos extracted features and embedding encoded features. (a), (b), and (c) represent the visualization plots after feature extraction using Attenphos at the S, T, and Y sites, respectively, and (d), (e), and (f) represent the feature visualization plots after encoding using embedding at the S, T, and Y sites, respectively.
Ijms 25 01526 g003
Figure 4. Attenphos framework. (a) represents the self-attention mechanism, and (b) represents encoding conversion.
Figure 4. Attenphos framework. (a) represents the self-attention mechanism, and (b) represents encoding conversion.
Ijms 25 01526 g004
Figure 5. (a) Conversion example. (b) Encoding conversion module converted feature representation.
Figure 5. (a) Conversion example. (b) Encoding conversion module converted feature representation.
Ijms 25 01526 g005
Table 1. Model comparison on balanced dataset. (a), (b), and (c) represent the results of model comparison for S site, T site, and Y site, respectively.
Table 1. Model comparison on balanced dataset. (a), (b), and (c) represent the results of model comparison for S site, T site, and Y site, respectively.
(a)
MethodsSn(%)Sp(%)AUCMccF1
GPS 2.122.2095.260.6700.1350.350
NetPhos 3.028.5587.230.6430.0810.404
PPRED21.3294.000.6760.1070.335
Musite28.6095.210.7260.1820.429
PhosphoSVM34.0195.900.7760.2370.492
SKIPHOS46.2068.600.6910.2650.520
Deepphos66.4375.890.7750.4250.697
transphos67.1675.890.7870.4320.702
Attenphos62.1581.360.7880.4430.688
(b)
MethodsSn(%)Sp(%)AUCMccF1
GPS 2.122.2095.260.6700.1350.350
NetPhos 3.028.5587.230.6400.0810.404
PPRED26.4383.510.5780.0520.370
Musite15.5695.360.6220.0980.259
PhosphoSVM21.7993.410.6650.1150.339
SKIPHOS65.8058.600.6430.1970.635
Deepphos46.0276.040.6740.2310.542
transphos47.3276.220.6720.2460.553
Attenphos48.9375.680.6910.2540.565
(c)
MethodsSn(%)Sp(%)AUCMccF1
GPS 2.147.9360.830.5520.0430.512
NetPhos 3.063.9146.100.5540.0480.587
PPRED42.0165.080.5390.0640.475
Musite28.8581.710.5870.0640.392
PhosphoSVM28.5584.390.5950.0840.396
SKIPHOS65.8058.600.6340.1970.635
Deepphos49.9366.370.6210.1650.545
transphos38.5272.300.6010.1150.463
Attenphos59.5660.590.6250.2010.599
Data in bold indicates that the model performs best for that evaluation metric.
Table 2. Performance comparison on unbalanced mouse dataset. (a), (b), and (c) denote the comparison of the prediction results of Attenphos with the state-of-the-art deep learning prediction tools at the S site, T site, and Y site, respectively.
Table 2. Performance comparison on unbalanced mouse dataset. (a), (b), and (c) denote the comparison of the prediction results of Attenphos with the state-of-the-art deep learning prediction tools at the S site, T site, and Y site, respectively.
(a)
MethodsSP(%)SN(%)AUCMCCF1
Attenphos84.0166.980.8310.5110.672
Transphos85.4165.080.8260.5120.669
Deepphos84.6461.270.8210.4690.637
(b)
MethodsSP(%)SN(%)AUCMCCF1
Attenphos87.7950.910.7750.4160.566
Transphos86.2650.910.7660.3930.554
Deepphos89.3130.910.770.2480.395
(c)
MethodsSP(%)SN(%)AUCMCCF1
Attenphos75.6850.000.6580.2620.524
Transphos72.0050.000.6420.2210.506
Deepphos67.5738.640.5840.1260.400
Data in bold indicates that the model performs best for that evaluation metric. For each column, deeper colors mean that the model performs better on that metric.
Table 3. Ablation experiments of attention mechanisms in the Attenphos method. (a), (b), and (c) indicate the S, T, and Y sites.
Table 3. Ablation experiments of attention mechanisms in the Attenphos method. (a), (b), and (c) indicate the S, T, and Y sites.
(a)
MethodsSn(%)Sp(%)AUCMccF1
Attenphos62.1581.360.7880.4430.688
Attenphos-noAttention67.5973.790.7070.4150.698
(b)
MethodsSn(%)Sp(%)AUCMccF1
Attenphos48.9375.680.6910.2540.565
Attenphos-noAttention52.6568.250.6040.2110.571
(c)
MethodsSn(%)Sp(%)AUCMccF1
Attenphos59.5660.590.6250.2010.599
Attenphos-noAttention52.7460.150.5640.1290.548
Deepphos67.5738.640.5840.1260.400
Data in bold indicates that the model performs best for that evaluation metric.
Table 4. PELM. Mouse protein phosphorylation dataset.
Table 4. PELM. Mouse protein phosphorylation dataset.
SitePositiveNegative
S15893178
T310620
Y201402
Table 5. PELM. Human protein phosphorylation training set.
Table 5. PELM. Human protein phosphorylation training set.
SitePositiveNegative
S20,96820,968
T56855685
Y21632163
Table 6. PPA test dataset.
Table 6. PPA test dataset.
SitePositiveNegative
S54375437
T16861686
Y676676
Table 7. Model parameter values.
Table 7. Model parameter values.
Parameters SettingsParameters Settings
Epochs100
Learning rate0.0001
Dropout0.3
EMBEDDING DIM16
DENSE256, 84, 2
Batch size32
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Song, T.; Yang, Q.; Qu, P.; Qiao, L.; Wang, X. Attenphos: General Phosphorylation Site Prediction Model Based on Attention Mechanism. Int. J. Mol. Sci. 2024, 25, 1526. https://doi.org/10.3390/ijms25031526

AMA Style

Song T, Yang Q, Qu P, Qiao L, Wang X. Attenphos: General Phosphorylation Site Prediction Model Based on Attention Mechanism. International Journal of Molecular Sciences. 2024; 25(3):1526. https://doi.org/10.3390/ijms25031526

Chicago/Turabian Style

Song, Tao, Qing Yang, Peng Qu, Lian Qiao, and Xun Wang. 2024. "Attenphos: General Phosphorylation Site Prediction Model Based on Attention Mechanism" International Journal of Molecular Sciences 25, no. 3: 1526. https://doi.org/10.3390/ijms25031526

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop