Next Article in Journal
Cationic Polymers Remarkably Boost Haloalkane Dehalogenase Activity in Organic Solvent Solutions and the Molecular Implications
Previous Article in Journal
High-Efficiency Mono-Cyclopentadienyl Titanium and Rare-Earth Metal Catalysts for the Production of Syndiotactic Polystyrene
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Prediction of Protein Ion–Ligand Binding Sites with ELECTRA

Department of Electrical Engineering and Computer Science, Bond Life Sciences Center, University of Missouri, Columbia, MO 65211, USA
*
Author to whom correspondence should be addressed.
Molecules 2023, 28(19), 6793; https://doi.org/10.3390/molecules28196793
Submission received: 25 August 2023 / Revised: 15 September 2023 / Accepted: 19 September 2023 / Published: 25 September 2023

Abstract

:
Interactions between proteins and ions are essential for various biological functions like structural stability, metabolism, and signal transport. Given that more than half of all proteins bind to ions, it is becoming crucial to identify ion-binding sites. The accurate identification of protein–ion binding sites helps us to understand proteins’ biological functions and plays a significant role in drug discovery. While several computational approaches have been proposed, this remains a challenging problem due to the small size and high versatility of metals and acid radicals. In this study, we propose IonPred, a sequence-based approach that employs ELECTRA (Efficiently Learning an Encoder that Classifies Token Replacements Accurately) to predict ion-binding sites using only raw protein sequences. We successfully fine-tuned our pretrained model to predict the binding sites for nine metal ions (Zn2+, Cu2+, Fe2+, Fe3+, Ca2+, Mg2+, Mn2+, Na+, and K+) and four acid radical ion ligands (CO32−, SO42−, PO43−, NO2). IonPred surpassed six current state-of-the-art tools by over 44.65% and 28.46%, respectively, in the F1 score and MCC when compared on an independent test dataset. Our method is more computationally efficient than existing tools, producing prediction results for a hundred sequences for a specific ion in under ten minutes.

Graphical Abstract

1. Introduction

Many biological processes are facilitated by the interactions between proteins and ligand ions [1]. These interactions are necessary for the proteins to carry out their functions properly [2,3]. More than fifty percent of proteins, when observed, interact with metal ions (cations) and acid radicals to stabilize their structure, and regulate their biological functions [4,5]. Fe3+ binding to hemoglobin is critical for transporting oxygen through the blood [6]. Ca2+ intracellular signaling triggers T-cell activation, and the development of B-cell response to antigens, differentiation, and development [7,8]. Zn2+ maintains the stability of the protein’s tertiary structure and is also essential for over 300 enzyme activities [9]—a lack or excess of it may cause central nervous system diseases [10]. The interaction of proteins with phosphate ions (PO43−) can result in phosphorylation, which switches enzymes on and off, thereby altering their function and activity [11]. Sulfate ions (SO42−) play a variety of structural roles, as well as binding to a variety of cytokines, growth factors, cell-surface receptors, adhesion molecules, enzymes, and fibrillary glycoproteins, to carry out various essential biological functions [12]. From these examples, we see that ions play significant roles in a wide range of cellular processes. Hence, the accurate identification of the protein–ion binding sites is important for understanding the mechanism of protein function and new drug discovery.
To understand the mechanism of protein–ion interactions, biological experiments, such as Nuclear Magnetic Resonance (NMR) spectroscopy [13] and fluorescence [14] methods, are carried out to measure the structure information of protein–ligand complexes and target ligand-binding proteins and their corresponding binding sites. As this is a very tedious and time-consuming process, computational methods to identify protein–ion binding sites are essential. The various computational methods proposed for predicting protein–ion binding sites can be grouped into sequence-based [15,16] and structure-based methods [17,18]. IonCom [19] proposed a new ligand-specific approach to predict the binding sites of nine metal ions (Zn2+, Cu2+, Fe2+, Fe3+, Ca2+, Mg2+, Mn2+, Na+ and K+) and four acid radical ion ligands (CO2−, PO3−, NO2−) using a sequence-based ab initio model that was first trained on sequence profiles, then extended using a modified AdaBoost algorithm to balance binding and non-binding residue samples. Sobolev and Edelman predicted the binding sites of protein chains and transition-metal ions by implementing the ‘CHED’ algorithm, obtaining a specificity of 96%; when predicting 349 whole proteins, 95% specificity was obtained [20]. Lu et al. used the fragment transformation method to predict metal ions’ (Ca2+, Mg2+, Cu2+, Fe2+, Mn2+, Zn2+) ligand binding sites, and obtained an overall accuracy of 94.6% and a sensitivity of 60.5% [21]. Ref. [22] identified four metal ions in the BioLip [23] database by implementing both sequence-based and template-based methods and obtained a Matthew’s correlation coefficient (MCC) greater than 0.5. Cao et al. used the SVM algorithm to identify ten metal ion-binding sites based on amino acid sequences, which obtained a good result using five-fold cross-validation [24]. Greenside et al. used an interpretable confidence-rated boosting algorithm to predict protein–ligand interactions with high accuracy from ligand chemical substructures and protein 1D sequence motifs, which produced decent results [25].
A major drawback of some of the existing computational tools is that they involve complex 3D computations, the threading of protein sequences to potential structural templates, and integrating multiple data types from both sequences and structures that are computationally intensive and time-consuming. In addition, several sequence-based tools have limited predictive performances (i.e., low precision) since they do not include tertiary structure information.
In this work, we propose IonPred, a Deep Learning framework based on ELECTRA [26] for predicting ion binding in proteins. The model adopts a sequence-based method for predicting the binding sites of nine metal ions and four acidic radicals. It takes raw protein sequences from all the protein chains, with at least one binding site as input for the model. IonPred is based on the Transformer architecture, which adopts a two-stage pretraining and fine-tuning process. In the initial pretraining phase, it employs the replacement token detection technique to learn contextual representations within the protein sequences from unlabeled protein sequence fragments. In contrast, in the fine-tuning phase, the model is trained with labeled sequence fragments to perform various binary classification tasks for various types of ion-binding sites.

2. Results

To benchmark the performance of our method, we compared its predictive performance with existing tools, and selected the Zinc dataset as a case study to understand how its performance is affected by different model configurations.

2.1. Comparison with Other Tools

We compared IonPred with six state-of-the-art tools. Three of these are sequence-based tools (i.e., TargetS [27], ZinCaps [28], and LMetalSite [29]), while the other three are structure-based (i.e., MIB [30], IonCom, and DELIA [31]) predictors. ZinCaps only supports the prediction of Zn2+, while DELIA does not. We also compared the performance of our tool for predicting the binding sites of non-metal ions with IonCom. For the metals, as reported in Table 1, alkali metals (Na+ and K+) are the hardest to differentiate according to their low-performance scores, followed by the alkali earth metals (Ca2+, Mg2+). This could probably be due to the wide variability of ion binding in these ion categories, even among the homologous proteins, and subtle differences in their binding affinities across various amino acid residues.
Except for Zn2+, where LMetalSite surpassed IonPred by 1.8% and 26.67% in the F1 and MCC, respectively, IonPred significantly outperformed all the sequence- and structure-based tools in most of the ion categories. The performance of LMetalSite is comparable to that of IonPred as both tools are based on a similar architecture (i.e., pretrained language models), and it shows that the sequence representation and contextual information learned from directly fine-tuning pretrained language models is both more insightful and more robust than when it is just used for feature extraction.
For the non-metal ion category, as seen in Table 2, IonPred outperforms IonCom in all metrics for all the acid radical by 50–117% for recall, 8.03–38.07% for precision, 44.65–67.03% in the F1 score, and 28.46–67.13% in the MCC.
We also plotted the ROC curves for the metal ions to further illustrate the superior performance of our method. As seen in Figure 1, except for Zn2+, the ROC curves for IonPred are all located at the upper portion of the plots to show more coverage and a higher AUC score. This indicates that IonPred has a greater capability to distinguish between positive and negative classes.
IonPred is far more computationally efficient than other tools, as it takes about 5 min to generate prediction results for 50–100 protein sequences of various lengths. It takes about 8 min to predict the same number of sequences with ZinCaps. It takes about 3 min to obtain the prediction results on just one protein sequence with TargetS, whereas it takes several hours to obtain one prediction result on just one protein sequence using IonCom and MIB.

2.2. Ablation Tests

To understand the efficiency of IonPred with different configurations, we used the Zn2+ dataset as a case study. This is because of its abundance in nature and the availability of quality datasets available for this ion. We evaluated the effect of the number of pretraining steps and the discriminator size on model performance. We pretrained and fine-tuned several ELECTRA models with various configurations for generator discriminator sizes, namely ELECTRA-0.25G-100K (the generator is 25% the size of the discriminator, with 100,000 training steps), ELECTRA-0.25G-200K (the generator is 25% the size of the discriminator, with 200,000 training steps), IonPred-0.25G-1M (the generator is 25% the size of the discriminator, with 1 million pretraining steps), ELECTRA-0.5G-200K (the generator is 50% the size of the discriminator, with 100,000 training steps) and ELECTRA-no-pretraining. We report their performance on the test dataset for Zinc in Table 3.
From the results, we see that of all the three models created with 200 K training steps, ELECTRA-0.25G-200K had the highest performance. This indicates that a generator size of 25% gives an optimal performance. Then, for all the configurations with a generator size of 0.25, we see that IonPred-0.25G-1M provided a better and overall superior performance. This indicates that a higher number of training steps gives a better performance. While ELECTRA-0.25G-100K has the same generator size as IonPred-0.25G-1M, it reports lower metric scores due to a lack of convergence of the model during pretraining. The model ELECTRA-no-pretraining, which was created without pretraining, reported the lowest performance for both AUC and AUPR.

2.3. Running Some Test Examples

We demonstrated the capability of our tool to identify ion-binding residues by running predictions on two known proteins that bind to Fe3+ and Mg2+, respectively. These proteins were obtained from RSCB. Table 4 contains the metal binding residues obtained from the protein database for each of the proteins, the position of the residue in the protein sequence, and the probability score from IonPred.
We see from the results that IonPred clearly identified three out of the five ion-binding residues in 3GKR_A for Fe3+ and accurately identified all six binding residues in 3DHG_D for Mg2+. Our model generally does a better job distinguishing residues that bind to Fe3+ than those that bind to Mg2+.

2.4. Tool

The pretrained ELECTRA model for ion-binding site prediction is provided as an open-source command-line tool, available at https://github.com/clemEssien/IonPred, accessed on 12 September 2023. It takes a Fasta file containing one or more protein sequences. Its instructions for use have been properly documented and the test datasets used are made available in the code repository. The output of the tool is a text file that contains the probability scores for each candidate site of the specified ion. The residues whose probability scores are higher than 0.5 are considered binding sites. IonPred was trained on a GPU, and it requires a GPU to run the prediction. The development environment requirements are Python 3, TensorFlow-GPU 1.15. CUDA 10, NumPy, Pandas, Scikit-learn, and SciPy. The default batch size for running predictions is 128.

3. Discussion

In this work, we presented IonPred, a pretrained ELECTRA model for predicting some of the most frequently seen ion-binding sites that have significant impact on protein structures and functions. Our method used raw sequence-based prediction because many proteins have no known structures or reliably predicted structures. The model was pretrained on a large corpus of unlabeled protein fragments in an unsupervised method and fine-tuned on a smaller quantity of non-redundant semi-manually curated labeled datasets. The model provided better predictive performance on alkali and alkali earth metal ions, which are typically difficult to predict. This is because the self-attention mechanism is adept at understanding the structural contexts of amino acid residues within protein sequences. This mechanism excels at assimilating conserved protein information by inherently focusing on neighboring residues and utilizes the transformer architecture to discern long-range dependencies.
However, there’s room for improvement for both metal categories. The attention mechanism of IonPred learns from the imbalanced dataset and provides improvement in the recall. We compared the different ELECTRA configurations of the training steps and generator sizes before we settled on the best configuration. IonPred significantly outperformed existing sequence and structure-based tools in all ion categories except Zinc, where LMetalSite slightly outperformed it. Here, we saw that directly fine-tuning the pretrained model on each specific binding site gave a better performance than just using it for feature extraction, as was demonstrated in LMetalSite.
The performance of the fine-tuning stage is mainly dependent on the availability of large high-quality labeled datasets. For ion-binding sites that have limited labels, its performance would not be as good. For future work, meta-learning could be explored as this could speed up the adaptation of binding sites with very limited labels. Also, the use of large protein information like ESM [32] or Sequence profile, and predicted structures from alpha fold [33] could also be incorporated to improve context-dependent biological properties learned by the discriminator with the purpose of significantly improving the recall.

4. Materials and Methods

4.1. Data and Data Processing

This study developed a new pretraining dataset by first downloading all protein chains from RCSB Protein Data Bank [34] using Biopython [35]. A total of 521,419 chains with their corresponding protein sequences were obtained. We excluded RNA and DNA components and protein chains that had less than 50 amino acid residues. We then made a series of API calls to the RCSB graph-based API [36], passing each protein chain ID and the keyword ‘UNIPROT’ as the parameters. The API response contained a lot of information, but we were only interested in the annotations. We obtained a total of 27,626 unique ligand-binding sites. While we identified various categories of binding sites, such as anions, cations, organic compounds, etc., we only focused on anions and cations for this study. Then, we used the sliding-window technique to extract fragments of a length of 25 (i.e., 12 amino acid residues on each side of the candidate binding residue). We chose fragment lengths of 25, because from the literature, fragment lengths of 7–25 have been tested, and it has been demonstrated by different methods that optimal fragments vary between 17 and 25. We used positive fragments for pretraining because through experimentation, we determined that pretraining with positive fragments made it easier to learn features related to ion-binding fragments more effectively. This process is illustrated in Figure 2.

4.2. Candidate Residue Selection

Almost all the amino acid residues are potential binding sites to varying degrees. A few of them participate more frequently in ion binding than others. Some of these residues are regular candidates for specific ions. To determine which candidate binding residues should be the focus when applying the sliding window for fragment extraction (which was used to extract positive fragments, as illustrated in Figure 2), we used a binary heatmap to plot the distribution of each amino acid residue with respect to the number of ion-binding sites. Through this process, we plotted two sets of heatmaps for the thirteen ions from the IonCom dataset (Figure 3a,b) and from BioLip database (Figure 3c,d). The x-axis represents the twenty amino acids, while the y-axis represents the ion ligands.
The plot is a frequency distribution of amino acid residues in relation to the number of ion-binding sites. From the figure, we determined that a residue was a binding site if its frequency was greater than or equal to the mean of the total residues in each row for a particular binding site. For the metal ions in Figure 3a, every amino acid residue was a candidate residue, but we observed the highest representation of candidate residues at Aspartate (D), Glutamate (E), and Histidine (H), followed by Leucine (L) and Cysteine (C). For the acid radical ions, we observed a high frequency of candidate residues at Histidine (H), Arginine (R), Glycine (G), Threonine (T), Lysine (K), and Serine (S), and using the sliding window technique, we extracted protein fragments of a length of 25 (i.e., 12 amino acid residues to the left and right of each candidate residue of interest), as implemented in [30], around the following amino acid residues: CYS (C), ASP (D), GLU (E), GLY (G), HIS (H), THR(T), LYS (K), ARG (R) and SER (S) at the center. If the amino acid residue at the center was an ion-binding site, the whole fragment was considered a positive sample; otherwise, it was regarded as a negative sample. We excluded negative fragments that contained a binding residue.
After eliminating duplicate fragments and excluding the negative fragments, we obtained 283,526 positive fragments. The complete process for obtaining the protein sequences used, annotation, and input fragment generation for pretraining is summarized in Figure 2. For the second stage, where we needed labeled data for finetuning, we obtained the labeled data of nine metal ions and four acid radicals from IonCom. The CD-HIT [37] tool was used to split the fine-tuning dataset into training, test and validation sets using a 40% similarity threshold to avoid over-fitting. The distribution of proteins used for fine-tuning is displayed in Table 5.
We extracted fragments from the labeled fine-tuning dataset. To label the fragments, we used the candidate binding residues, determined from the distribution in Figure 3. Fragments extracted around a binding residue were labeled as positive examples, while fragments extracted around a candidate binding site that were not binding residues were labeled as negative training examples. The statistics of the training, test, and validation fragments are summarized in Table 6.

4.3. Problem Definition

The ion-binding-site prediction in this study was formulated as a binary classification problem. For example, given a protein sequence for which the binding sites are unknown, we selected a particular ion (i.e., Zn2+, Cu2+, Fe2+, Fe3+, etc.) for which we wanted to determine the binding sites. Then, the aim would be to ascertain if the candidate binding residues (from Figure 3) for the selected ion(s) were binding site(s) or not. This would output probabilities for each candidate residue. A probability of 0.5 and above was considered a positive prediction (i.e., an ion-binding site), while a probability less than 0.5 was regarded as a negative prediction (i.e., not a binding site).

4.4. Deep Learning Model

The architecture of the proposed IonPred, as shown in Figure 4, was based on the ELECTRA (i.e., “Efficiently Learning an Encoder that Classifies Token Replacements Accurately”) learning model. This architecture comprised two neural networks, a generator, and a discriminator. These networks basically mapped a sequence of input tokens x = x 1 , , x n into a sequence of contextualized vector representations h x = h 1 , , h n .
So, for any given position t, where xt is a masked amino acid residue [MASK], the generator used a SoftMax layer to produce the probability of generating a particular token xt.
p G x t x = exp e x t T h G x t / x exp   ( e x T h G x t )
In the equation above, e denotes the embeddings for the amino acid residues. The generator was trained using masked language modeling (MLM). For a given input, x = [ x 1 , , x n ] , MLM selected a random set of positions ranging from 1 to n to mask out. This produced the vector m = [ m 1 , , m k ] . The residues in these positions were replaced with a [MASK] token, which was represented as x m a s k e d = R E P L A C E   ( x , m ,   [ M A S K ] ) . The generator learned to predict the original amino acid residues. The discriminator predicted whether the amino acid residue was originally from the input data or if it was a replacement from the generator distribution using a sigmoid output layer, as shown in the equation below:
D x , t = sigmoid w T h D x t
The masked-out residues were replaced by samples from the generator. This sample is represented as x c o r r u p t . The discriminator was trained to predict which residues in x c o r r u p t matched the original input   x . The model inputs were described as shown below:
m i   ~   u n i f { 1 , n }   f o r   i = 1   t o   k                           x m a s k e d = R E P L A C E ( x , m ,   M A S K )
x ^ i ~ p G x i x m a s k e d   f o r   i   m x c o r r u p t = R E P L A C E ( x , m ,   x ^ )
And the loss functions used for the generator and discriminator are shown in Equations (5) and (6) below:
L M L M   x , θ G = E     i m n l o g p G x i x m a s k e d  
L D i s c   x , θ D = E   t = 1 n 1 x t c o r r u p t = x t log D ( x t c o r r u p t , t ) 1 ( x t c o r r u p t   x t )   l o g ( 1 D ( x t c o r r u p t , t ) )  
The minimized combined loss for both the generator and discriminator was given as
m i n θ G , θ D   x X L M L M   x , θ G + λ L D i s c x , θ D

4.5. Pretraining

As shown in Figure 5, the pretraining consisted of the generator and discriminator, which are essentially two transformer models. Here, the Generator corrupted a percentage of the tokens (i.e., amino acid residues) from the input fragments, and the discriminator was trained to detect the replaced tokens. This enabled the model to learn context-dependent biological properties of protein sequence fragments from a large-scale task-independent and unlabeled protein dataset. The patterns learned during this stage were then embedded into a smaller task-specific and labeled dataset in the downstream tasks, i.e., binary classification prediction for various protein–ion binding sites. This significantly reduced the amount of labeled data needed since the pretrained model had already learned the underlying patterns related to classification. We selected the ELECTRA-small model, which comprised 12 layers, 256 hidden layers, and 128-dimension embedding.
This model was chosen due to the relatively small size of our pretraining corpus and the fact that a larger-size model would have been computationally expensive to train, which may not have led to any significant improvement. The vocabulary size used was 25, which included all 20 amino acid residues, the ‘-’ character to pad positions at the protein terminus, [MASK] as the masking character, [CLS] to mark the start of a fragment, [SEP] to mark the end of a fragment, and [UNK] for out-of-vocabulary words, i.e., unknown amino acid residues. We masked 15% of each input fragment in the embedding layer, which was then encoded into the token embeddings matrix, having a dimension of [27 × 128]. Both the token and position embeddings were summed and presented as input tokens, i.e., x = [x1, …, x27], into the generator. The generator used was 25% of the size of the discriminator [32], with 12 layers and a hidden size of 64. The generator trained using maximum likelihood to predict the original masked-out amino acid residues based on the contextual information from neighboring amino acid residues in the protein fragment. The model was trained over 1 million training steps, using a batch size of 128 and a learning rate of 0.0001.

4.6. Fine-Tuning

After pretraining, the generator was discarded, and the discriminator was then fine-tuned using labeled data for various specific classification tasks. For this, a fully connected layer was built over the pretrained ELECTRA model and the entire network was fine-tuned with 12 layers of the discriminator. This was performed to ensure the error was backpropagated throughout the whole architecture and that the weights of the discriminator were updated based on the fragments in the fine-tuned dataset. We fine-tuned separate models for each ligand–ion binding site using labeled fragments generated from the protein sequence, as described in Table 1. The candidate binding residues used for the metals were C, H, E, and D, while the ones used for acidic radicals were G, H, K, R, and S. The training, testing and dev fragments were split by a ratio of 80%, 10%, and 10%, respectively. We added a fully connected layer at the end of the pretrained ELECTRA model and fine-tuned the entire network consisting of 12 layers of the discriminator, so that the error was backpropagated across the entire architecture and the discriminator weights were updated using the labeled data, as shown in Figure 6. Similar hyperparameters used in the pretraining were implemented at this stage, except for the learning rate and the number of training steps, which were set at 0.00001 and 200 epochs, respectively. Fine-tuning runs much quicker than pretraining.

4.7. Model Assessment

We evaluated IonPred using the following metrics: Recall, Precision, F1 score, Matthew’s correlation coefficient (MCC), and the Receiver operating characteristic (ROC) curve, which are defined below:
R e c a l l = T P T P + F N × 100
P r e c i s i o n = T P T P + F P × 100
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l × 100
M C C = T P × T N F P × F N T P + F P × T P + F N × T N + F P × T N + F N
The ROC curve is a graphical representation used in binary classification to assess the performance of a model across all possible classification thresholds. It is used to understand the trade-off between the true positive rate (TPR) and false positive rate (1-specificity) at different threshold settings.
T P R = T P T P + F N
F P R = F P F P + T N × 100
where TP represents the number of binding residues correctly predicted as binding residues, TN is the number of non-binding residues that are correctly predicted as non-binding residues, FP is the number of non-binding residues that are incorrectly predicted as binding residues, and FN represents the number of binding residues incorrectly predicted as non-binding residues.
We also reported the AUC score and AUPR score. These results are reported in Table 1 and Table 2.

Author Contributions

Conceptualization, D.X., D.W. and C.E.; methodology, C.E. and L.J.; software, C.E.; validation, C.E., L.J. and D.X.; formal analysis, C.E.; data curation, C.E.; writing—original draft preparation, C.E.; writing—review and editing, D.X.; supervision, D.X.; funding acquisition, D.X. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the US National Institutes of Health, grant R35-GM126985. In addition, this work used the high-performance computing infrastructure provided by Research Computing Support Services at the University of Missouri, as well as the Pacific Northwest National Laboratory (PNNL). We would like to thank Negin Manshour for technical assistance.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All the source codes and data used for this project are available at https://github.com/clemEssien/IonPred (accessed on 12 September 2023).

Conflicts of Interest

The authors declare no conflict of interest.

Sample Availability

Not applicable.

References

  1. Alberts, B.; Johnson, A.; Lewis, J.; Raff, M.; Roberts, K.; Walter, P. Molecular biology of the cell. Scand. J. Rheumatol. 2003, 32, 125. [Google Scholar]
  2. Gao, M.; Skolnick, J. The distribution of ligand-binding pockets around protein-protein interfaces suggests a general mechanism for pocket formation. Proc. Natl. Acad. Sci. USA 2012, 109, 3784–3789. [Google Scholar] [CrossRef] [PubMed]
  3. Gao, M.; Skolnick, J. A comprehensive survey of small-molecule binding pockets in proteins. PLoS Comput. Biol. 2013, 9, e1003302. [Google Scholar] [CrossRef] [PubMed]
  4. Tainer, J.A.; Roberts, V.A.; Getzoff, E.D. Metal-binding sites in proteins. Curr. Opin. Biotechnol. 1991, 2, 582–591. [Google Scholar] [CrossRef]
  5. Thomson, A.J.; Gray, H.B. Bio-inorganic chemistry. Curr. Opin. Chem. Biol. 1998, 2, 155–158. [Google Scholar] [CrossRef] [PubMed]
  6. Hsia, C.C.W. Respiratory function of hemoglobin. N. Engl. J. Med. 1998, 338, 239–248. [Google Scholar] [CrossRef] [PubMed]
  7. Fracchia, K.M.; Pai, C.; Walsh, C.M. Modulation of t cell metabolism and function through calcium signaling. Front. Immunol. 2013, 4, 324. [Google Scholar] [CrossRef] [PubMed]
  8. Baba, Y.; Kurosaki, T. Role of calcium signaling in B cell activation and biology. In B Cell Receptor Signaling; Springer: Cham, Switzerland, 2015; pp. 143–174. [Google Scholar]
  9. McCall, K.A.; Huang, C.-c.; Fierke, C.A. Function and mechanism of zinc metalloenzymes. J. Nutr. 2000, 130, 1437S–1446S. [Google Scholar] [CrossRef]
  10. Gower-Winter, S.D.; Levenson, C.W. Zinc in the central nervous system: From molecules to behavior. BioFactors 2012, 38, 186–193. [Google Scholar] [CrossRef]
  11. Wang, J.P.; Chuang, L.; Loziuk, P.L.; Chen, H.; Lin, Y.C.; Shi, R.; Qu, G.Z.; Muddiman, D.C.; Sederoff, R.R.; Chiang, V.L. Phosphorylation is an on/off switch for 5-hydroxyconiferaldehyde o-methyl-transferase activity in poplar monolignol biosynthesis. Proc. Natl. Acad. Sci. USA 2015, 112, 8481–8486. [Google Scholar] [CrossRef]
  12. Zhang, B.; Chi, L. Chondroitin sulfate/dermatan sulfate-protein interactions and their biological functions in human diseases: Implications and analytical tools. Front. Cell Dev. Biol. 2021, 9, 693563. [Google Scholar] [CrossRef] [PubMed]
  13. Sletten, E. The binding of transition metal ions to DNA oligonucleotides studied by nuclear magnetic resonance spectroscopy. In Cytotoxic, Mutagenic and Carcinogenic Potential of Heavy Metals Related to Human Environment; Springer: Dordrecht, The Netherlands, 1997; pp. 493–509. [Google Scholar]
  14. Yonezawa, M.; Doi, N.; Higashinakagawa, T.; Yanagawa, H. DNA display of biologically active proteins for in vitro protein selection. J. Biochem. 2004, 135, 285–288. [Google Scholar] [CrossRef] [PubMed]
  15. Chen, P.; Huang, J.Z.; Gao, X. Ligandrfs: Random forest ensemble to identify ligand-binding residues from sequence information alone. BMC Bioinform. 2014, 15, S4. [Google Scholar] [CrossRef] [PubMed]
  16. Chen, P.; Hu, S.; Zhang, J.; Gao, X.; Li, J.; Xia, J.; Wang, B. A sequence-based dynamic ensemble learning system for protein ligand-binding site prediction. IEEE/ACM Trans. Comput. Biol. Bioinform. 2015, 13, 901–912. [Google Scholar] [CrossRef] [PubMed]
  17. Roy, A.; Yang, J.; Zhang, Y. Cofactor: An accurate comparative algorithm for structure-based protein function annotation. Nucleic Acids Res. 2012, 40, W471–W477. [Google Scholar] [CrossRef] [PubMed]
  18. Yang, J.; Roy, A.; Zhang, Y. Protein–ligand binding site recognition using complementary binding-specific substructure comparison and sequence profile alignment. Bioinformatics 2013, 29, 2588–2595. [Google Scholar] [CrossRef] [PubMed]
  19. Hu, X.; Dong, Q.; Yang, J.; Zhang, Y. Recognizing metal and acid radical ion-binding sites by integrating ab initio modeling with templatebased transferals. Bioinformatics 2016, 32, 3260–3269. [Google Scholar] [CrossRef]
  20. Sobolev, V.; Edelman, M. Web tools for predicting metal binding sites in proteins. Isr. J. Chem. 2013, 53, 166–172. [Google Scholar] [CrossRef]
  21. Lu, C.H.; Lin, Y.F.; Lin, J.J.; Yu, C.S. Prediction of metal ion–binding sites in proteins using the fragment transformation method. PLoS ONE 2012, 7, e39252. [Google Scholar] [CrossRef]
  22. Hu, X.; Wang, K.; Dong, Q. Protein ligand-specific binding residue predictions by an ensemble classifier. BMC Bioinform. 2016, 17, 470. [Google Scholar] [CrossRef]
  23. Yang, J.; Roy, A.; Zhang, Y. Biolip: A semi-manually curated database for biologically relevant ligand–protein interactions. Nucleic Acids Res. 2012, 41, D1096–D1103. [Google Scholar] [CrossRef] [PubMed]
  24. Cao, X.; Hu, X.; Zhang, X.; Gao, S.; Ding, C.; Feng, Y.; Bao, W. Identification of metal ion binding sites based on amino acid sequences. PLoS ONE 2017, 12, e0183756. [Google Scholar] [CrossRef] [PubMed]
  25. Greenside, P.; Hillenmeyer, M.; Kundaje, A. Prediction of protein-ligand interactions from paired protein sequence motifs and ligand sub-structures. In Pacific Symposium on Biocomputing 2018: Proceedings of the Pacific Symposium; World Scientific: Singapore, 2018; pp. 20–31. [Google Scholar]
  26. Clark, K.; Luong, M.T.; Le, Q.V.; Manning, C.D. ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators. arXiv 2020, arXiv:2003.10555. [Google Scholar]
  27. Yu, D.J.; Hu, J.; Yang, J.; Shen, H.B.; Tang, J.; Yang, J.Y. Designing template-free predictor for targeting protein-ligand binding sites with classifier ensemble and spatial clustering. IEEE/ACM Trans. Comput. Biol. Bioinform. 2013, 10, 994–1008. [Google Scholar]
  28. Essien, C.; Wang, D.; Xu, D. Capsule network for predicting zinc binding sites in metalloproteins. In Proceedings of the 2019 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), San Diego, CA, USA, 18–21 November 2019; pp. 2337–2341. [Google Scholar]
  29. Yuan, Q.; Chen, S.; Wang, W. Prediction of ligand binding residues in protein sequences using machine learning. In Proceedings of the 2019 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), San Diego, CA, USA, 18–21 November 2019; pp. 2298–2304. [Google Scholar]
  30. Lin, Y.-F.; Cheng, C.-W.; Shih, C.-S.; Hwang, J.-K.; Yu, C.-S.; Lu, C.-H. Mib: Metal ion-binding site prediction and docking server. J. Chem. Inf. Model. 2016, 56, 2287–2291. [Google Scholar] [CrossRef] [PubMed]
  31. Xia, C.-Q.; Pan, X.; Shen, H.-B. Protein–ligand binding residue prediction enhancement through hybrid deep heterogeneous learning of sequence and structure data. Bioinformatics 2020, 36, 3018–3027. [Google Scholar] [CrossRef] [PubMed]
  32. Lin, Z.; Akin, H.; Rao, R.; Hie, B.; Zhu, Z.; Lu, W.; Smetanin, N.; Costa, A.d.S.; Fazel-Zarandi, M.; Sercu, T.; et al. Language models of protein sequences at the scale of evolution enable accurate structure prediction. bioRxiv 2022. [Google Scholar] [CrossRef]
  33. Jumper, J.; Evans, R.; Pritzel, A.; Green, T.; Figurnov, M.; Ronneberger, O.; Tunyasuvunakool, K.; Bates, R.; Žídek, A.; Potapenko, A.; et al. Highly accurate protein structure prediction with Alphafold. Nature 2021, 596, 583–589. [Google Scholar] [CrossRef]
  34. Berman, H.M.; Westbrook, J.; Feng, Z.; Gilliland, G.; Bhat, T.N.; Weissig, H.; Shindyalov, I.N.; Bourne, P.E. The protein data bank. Nucleic Acids Res. 2000, 28, 235–242. [Google Scholar] [CrossRef]
  35. Cock, P.J.A.; Antao, T.; Chang, J.T.; Chapman, B.A.; Cox, C.J.; Dalke, A.; Friedberg, I.; Hamelryck, T.; Kauff, F.; Wilczynski, B.; et al. Biopython: Freely available python tools for computational molecular biology and bioinformatics. Bioinformatics 2009, 25, 1422–1423. [Google Scholar] [CrossRef]
  36. Segura, J.; Rose, Y.; Westbrook, J.; Burley, S.K.; Duarte, J.M. Rcsb protein data bank 1d tools and services. Bioinformatics 2020, 36, 5526–5527. [Google Scholar] [CrossRef] [PubMed]
  37. Li, W.; Jaroszewski, L.; Godzik, A. Clustering of highly homologous sequences to reduce the size of large protein databases. Bioinformatics 2001, 17, 282–283. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Comparison of ROC curves of IonPred with existing tools for the metal-ion predictions.
Figure 1. Comparison of ROC curves of IonPred with existing tools for the metal-ion predictions.
Molecules 28 06793 g001
Figure 2. Data preprocessing for generating positive protein fragments used for pretraining.
Figure 2. Data preprocessing for generating positive protein fragments used for pretraining.
Molecules 28 06793 g002
Figure 3. Frequency distribution of amino acid residues with respect to ion ligands derived from the IonCom dataset. (a) Metal ions (IonCom). (b) Acid radicals (IonCom). (c) Metal ions (BioLip). (d) Acid radicals (BioLip).
Figure 3. Frequency distribution of amino acid residues with respect to ion ligands derived from the IonCom dataset. (a) Metal ions (IonCom). (b) Acid radicals (IonCom). (c) Metal ions (BioLip). (d) Acid radicals (BioLip).
Molecules 28 06793 g003
Figure 4. Electra architecture, which illustrates token corruption and replacement by the generator and discriminator.
Figure 4. Electra architecture, which illustrates token corruption and replacement by the generator and discriminator.
Molecules 28 06793 g004
Figure 5. Pretraining process taking in raw protein fragments as input.
Figure 5. Pretraining process taking in raw protein fragments as input.
Molecules 28 06793 g005
Figure 6. Fine-tuning with labeled dataset to generate probabilities for binary classification.
Figure 6. Fine-tuning with labeled dataset to generate probabilities for binary classification.
Molecules 28 06793 g006
Table 1. Performance comparison of IonPred with other tools on metal-ion test sets.
Table 1. Performance comparison of IonPred with other tools on metal-ion test sets.
IonMethodRecPreF1MCCAUCAUPR
MIB0.7390.2200.3390.3890.9220.388
TargetS0.4500.7500.5630.5780.8680.594
Zn2+ZinCaps0.7530.7800.7660.6010.9150.768
IonCom0.7790.1370.2330.3170.8920.671
LMetalSite0.7220.8590.7850.7600.9710.801
IonPred0.7900.8400.8140.6000.9580.780
MIB0.3410.0820.1320.1390.7640.105
TargetS0.1190.4870.1910.2440.7750.165
Ca2+DELIA0.1720.6300.2700.3300.7820.251
IonCom0.2970.2470.2700.2580.6970.166
LMetalSite0.4130.7200.5250.5400.9040.490
IonPred0.4670.7590.5780.6150.9230.520
MIB0.2460.0430.0730.0820.6730.053
TargetS0.1180.4910.1900.2370.7150.148
Mg2+IonCom0.2400.2500.2450.2370.6850.184
DELIA0.1290.0650.0860.2870.7400.198
LMetalSite0.2450.7280.3670.4190.8660.316
IonPred0.4000.7800.5290.4700.8890.450
MIB0.4620.0960.1590.1930.8550.168
TargetS0.2710.4960.3500.3620.8620.322
Mn2+DELIA0.5020.6650.5720.5740.8870.489
IonCom0.5110.2450.3310.3440.8310.304
LMetalSite0.6130.7190.6620.6610.9630.625
IonPred0.6200.7000.6580.6700.9700.670
MIB0.5860.6200.6030.5730.9090.354
Fe2+TargetS0.3450.2540.2930.2450.7600.299
IonPred0.7490.7280.7380.7230.9370.771
IonCom0.6100.4980.5480.5790.9090.567
Fe3+MIB0.4740.3990.4330.3830.8130.438
IonPred0.7430.6120.6710.6520.9280.724
IonCom0.5960.3980.4770.5920.8900.399
Cu2+MIB0.4660.2800.3500.3580.8700.419
IonPred0.7890.6340.7030.6200.9390.677
IonCom0.2100.1780.1930.1600.7230.156
K+TargetS0.3890.4110.4000.3410.8760.336
IonPred0.4980.6720.5720.5240.9120.478
Na+IonCom0.4510.2920.3550.2180.7090.233
IonPred0.5230.7310.6100.5950.9040.487
Rec refers to Recall, Pre refers to Precision, MCC refers to Matthew’s correlation coefficient, F1 refers to F1 score, AUC refers to Area under the curve, and AUPR refers to Area under precision recall curve. Bold font indicates metric with best performance.
Table 2. Performance comparison of IonPred with other tools on non-metal-ion test sets.
Table 2. Performance comparison of IonPred with other tools on non-metal-ion test sets.
RadicalsMethodRecPreF1MCCAUCAUPR
CO32−IonCom0.6100.4980.5480.5790.9090.567
IonPred0.7430.6120.6710.6520.9280.724
NO2IonCom0.5960.3980.4770.5920.8900.399
IonPred0.7890.6340.7030.6200.9390.677
SO43−IonCom0.2100.1780.1930.1600.7230.156
IonPred0.3890.4110.4000.3410.8760.336
PO43−IonCom0.4510.2920.3550.2180.7090.233
IonPred0.5230.7310.6100.5950.9040.487
Rec refers to Recall, Pre refers to Precision, MCC refers to Matthew’s correlation coefficient, F1 refers to F1 score, AUC refers to Area under the curve, and AUPR refers to Area under precision recall curve. Bold font indicates metric with best performance.
Table 3. Performance evaluation of several ELECTRA model configurations on Zinc dataset.
Table 3. Performance evaluation of several ELECTRA model configurations on Zinc dataset.
ConfigurationAUCAUPR
ELECTRA-0.25G-100K0.9160.698
ELECTRA-0.25G-200K0.9510.756
IonPred-0.25G-1M0.9580.780
ELECTRA-0.5G-200K0.9260.739
ELECTRA-1.0G-200K0.9040.676
ELECTRA-no-pretraining0.8570.519
Bold font indicates metric with best performance.
Table 4. Sample predictions of known proteins that bind to Fe3+ and Mg2+.
Table 4. Sample predictions of known proteins that bind to Fe3+ and Mg2+.
ProteinResidueResidue PositionPredicted Probability
3GKR_A
(Fe3+)
D650.558
D670.551
D1510.516
E3090.410
E3110.499
3DHG_D
(Mg2+)
E1040.891
E1340.912
H1370.896
E1970.903
E2310.920
H2340.899
Table 5. Statistics of the residue distribution of each ion dataset used for fine-tuning.
Table 5. Statistics of the residue distribution of each ion dataset used for fine-tuning.
CategoryIonNprotRposRneg
Metal ionsCa2+1791360119,192
Cu2+11053538,488
Fe2+227111573,813
Fe3+10343934,113
K+5353618,776
Mg2+10339176,382
Mn2+3791778148,618
Na+7848927,408
Zn2+14269793,952
Acid radicalsCO32−6231622,766
NO222988144
PO43−303212599,729
SO42−3392168112,279
Nprot represents the number of protein chains, while Rpos and Rneg represent the number of binding residues and the number of non-binding residues, respectively.
Table 6. Statistics of the training, test, and validation fragments used for fine-tuning.
Table 6. Statistics of the training, test, and validation fragments used for fine-tuning.
CategoryIonTrainingTestValidation
Metal ionsCa2+849,087108,85795,919
Cu2+23,97740743070
Fe2+51,39865896345
Fe3+106,11413,60413,100
K+26,86458486010
Mg2+594,19376,17973,357
Mn2+195,49924,06524,672
Na+46,07074505493
Zn2+712,169104,85691,922
Acid radicalsCO32−11,46519191417
NO2905713051180
PO43−114,23423,83613,240
SO42−76,13412,93711,534
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Essien, C.; Jiang, L.; Wang, D.; Xu, D. Prediction of Protein Ion–Ligand Binding Sites with ELECTRA. Molecules 2023, 28, 6793. https://doi.org/10.3390/molecules28196793

AMA Style

Essien C, Jiang L, Wang D, Xu D. Prediction of Protein Ion–Ligand Binding Sites with ELECTRA. Molecules. 2023; 28(19):6793. https://doi.org/10.3390/molecules28196793

Chicago/Turabian Style

Essien, Clement, Lei Jiang, Duolin Wang, and Dong Xu. 2023. "Prediction of Protein Ion–Ligand Binding Sites with ELECTRA" Molecules 28, no. 19: 6793. https://doi.org/10.3390/molecules28196793

Article Metrics

Back to TopTop