Next Article in Journal
Effects of Calcium-Oxide-Modified Biochar on the Anaerobic Digestion of Vacuum Blackwater
Next Article in Special Issue
Two Methods for Superposing the Structures of Like-Molecule Assemblies: Application to Peptide and Protein Oligomers and Aggregates
Previous Article in Journal
Recovery of Polysaccharides from Red Grape Marc and White Grape Pomace by Degradation of Cell Walls by Enzymes with Different Activities
Previous Article in Special Issue
Bluues_cplx: Electrostatics at Protein–Protein and Protein–Ligand Interfaces
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Deep Learning Approaches for the Prediction of Protein Functional Sites

Computational Systems Biology Group, National Center for Biotechnology (CNB-CSIC), 28049 Madrid, Spain
*
Author to whom correspondence should be addressed.
Molecules 2025, 30(2), 214; https://doi.org/10.3390/molecules30020214
Submission received: 3 November 2024 / Revised: 20 December 2024 / Accepted: 1 January 2025 / Published: 7 January 2025

Abstract

:
Knowing which residues of a protein are important for its function is of paramount importance for understanding the molecular basis of this function and devising ways of modifying it for medical or biotechnological applications. Due to the difficulty in detecting these residues experimentally, prediction methods are essential to cope with the sequence deluge that is filling databases with uncharacterized protein sequences. Deep learning approaches are especially well suited for this task due to the large amounts of protein sequences for training them, the trivial codification of this sequence data to feed into these systems, and the intrinsic sequential nature of the data that makes them suitable for language models. As a consequence, deep learning-based approaches are being applied to the prediction of different types of functional sites and regions in proteins. This review aims to give an overview of the current landscape of methodologies so that interested users can have an idea of which kind of approaches are available for their proteins of interest. We also try to give an idea of how these systems work, as well as explain their limitations and high dependence on the training set so that users are aware of the quality of expected results.

1. Introduction

The relative ease in obtaining the amino acid sequences of proteins contrasts with the difficulty in finding functional information for them. Thanks to modern sequencing technologies [1], today, it is possible to obtain the sequences of an organism’s repertory of proteins (proteome) by sequencing its genome. It is also possible to obtain the sequences of millions of protein and protein fragments from a metagenomics or environmental sample with a complex mixture of uncharacterized organisms [2]. Nevertheless, associating functional information to these sequences, either globally (i.e., infer the function of a given protein) or locally (which residues/regions of a protein are important for different aspects of its function), is not so straightforward. As a consequence, most proteins for which we know the amino acid sequence are not functionally characterized. Even in the well-curated UniProt database [3], less than 10% of the proteins have some kind of associated functional information [4], a figure that is much lower than that of other less annotated protein sequence databases. The functional characterization at the residue level is one order of magnitude lower (~1%).
The experimental characterization of a protein’s functional sites is slow and expensive, as it basically involves mutating every residue of the protein and assaying the effect on its function. Such experimental approaches cannot cope with the increasing deluge of newly sequenced proteins. Nevertheless, knowing which residues in a protein are related to its function is very important, not only to obtain insight into it from a molecular point of view but also to devise ways of modifying it for our benefit [5]. Consequently, much effort was put into developing methodologies for predicting a protein’s functional regions from the wealth of available sequence information [4,5,6].
Most of these approaches are sequence-based. They take as input multiple sequence alignments of homologous proteins and look for position variation patterns related to functionality, such as full conservation, subfamily-dependent conservation, or co-evolving positions [6]. While these sequence-based methods have existed for a long time, they are still highly used as they take advantage of the abundant sequence information, are quite reliable, and complement other more sophisticated methodologies.
There is also a plethora of methodologies that use structural information for detecting functional regions, normally in combination with the sequence-based approaches commented on earlier. These approaches have been recently boosted by the development of deep learning-based methodologies. In fact, deep learning-based AlphaFold [7] or RosettaFold [8] are successful in predicting accurate three-dimensional (3D) structures, which are comparable to experimental structures. These approaches can generate high-quality models for virtually every sequenced protein. Notably, the developers of AlphaFold were awarded the 2024 Nobel Prize in Chemistry.
As in many other areas in bioinformatics, in the area of functional site prediction, there was a progression in the methodologies from statistics to “classic” machine learning and now to deep learning. Deep learning is a class of machine learning approaches (systems that learn from examples to relate inputs and outputs) that use large neural networks with complex architectures and many internal layers. As in human activities in general, and in research in particular, deep learning is revolutionizing inference tasks for which large enough datasets are available for training [9,10]. These methodologies are not only assisting the prediction of functional sites through the generation of accurate 3D structural models, commented on in the previous paragraph, but there are deep learning-based methods specifically trained for this task as well.
In this review, we aim to provide an overview of the main deep learning-based approaches for predicting protein functional regions. We use a general definition of “functional site”: any residue whose modification affects any functional aspect of the protein without affecting the structure. Hence, this definition includes active sites, binding sites, post-translational modifications, non-functional variants related to diseases, etc., although methodologies are generally focused on a particular type of functional site. This review does not try to be comprehensive but covers only some representative examples of the many approaches used in this area to give readers an idea of the main tendencies of the field.
This review does not cover the prediction of global protein function (i.e., molecular activity or cellular role of the whole protein) and only focuses on the prediction of functional residues. The prediction of global protein function using deep learning approaches is also a hot topic and has been covered in other excellent reviews (e.g., [11]). The target audience of this review is composed by life scientists, potential users of these methodologies. Hence, the technical details of these complex methods are kept to a minimum, although a primer on deep learning, references, and tables with further information are included.

2. A Primer on Deep Learning

Deep learning is a very complex area involving many technical aspects. It is impossible to give an overview of the field at any depth here. Nevertheless, a basic description of the main concepts can help users of these methodologies to understand what they have in their hands, as well as their capabilities and limitations. For a deeper introduction to the subject, the reader is referred, for example, to [9].
Machine learning methods are systems designed to carry out inference tasks without explicit instructions or a procedural algorithm. They do that by learning from solved examples of that task. A set of examples is shown to the system (“training”), and it gets ready for processing a new case not seen in its training set (“prediction”). Neural networks are a class of machine learning systems designed to relate numerical inputs and outputs (i.e., predict the output for a given input) when that relationship is complex and cannot be expressed by an explicit mathematical function. Their architecture and functioning try to resemble what we know about the information processing by the human brain. The inputs and outputs are coded as vectors of real numbers, whose components are depicted as “neurons” in the standard graphical representation of these systems (Figure 1). Apart from the input and output layers of artificial neurons, these systems have one or more internal layers, termed “hidden layers”. The neurons of a given layer are connected to those of the next layer, and these connections have associated numerical values representing their strengths (weights). The values of the input layer neurons are determined by the input vector. The value of a given neuron in the forthcoming layers is given, in the simplest case, by the sum of the values of the neurons connected to it in the preceding layers multiplied by the corresponding weights (Figure 1). In most cases, more complex “activation functions” are applied to calculate these values, eventually involving thresholds, again trying to replicate how the activation of biological neurons works. So, with this procedure, starting with a given input vector, a final output vector is generated. During the training process, a set of known input–output pairs of vectors is shown to the network, and the weights of the connections (which start with random values) are adjusted to minimize empirical risk: the difference between the predicted output and the actual target vector. After this training process, the weights of the connections should have values optimized to relate inputs and outputs, at least those in the training set. With these trained weights, the network is ready to “predict” (generate an output vector for an unseen input). Note that in a neural network without hidden layers, the values of the output neurons can be calculated from the inputs with a simple linear function. The objective of the architecture of hidden layers is to approximate complex non-linear functions with many parameters (weights of the connections) relating to inputs and outputs.
Although described in a very simplified way here, the training process is complex and a critical part of the development of a predictor. The training set has to be well balanced in terms of examples and try to uniformly cover the spaces of possible inputs and outputs the network is going to face later at the prediction stage. If the training set is biased to a particular class of examples, the network will be “too adapted” to that class and will not generalize well for others (overfitting). Besides being well balanced, the training set has to be as large as possible, especially in deep learning systems, which require a lot of training examples in order to optimize the many parameters they have
“Deep” neural networks differ from these “classic” neural networks in their size (they have many more internal layers and hence neurons) as well as their architecture, which deviates from that succinctly described above and is more complex and adapted to specific tasks. Table 1 contains descriptions of some deep learning architectures and approaches widely used for making predictions on biological data. For example, the well-known ChatGPT system for generating human-like text in response to a prompt is a deep neural network with a “transformer” architecture (Table 1) and hundreds of billions of connections. The input of ChatGPT is a sequence of “words”, and the output is a ranked list of possible words that would fit as a continuation of that sequence, sorted by a probability. Starting with the user prompt, iterative applications of the network to the growing chain of words generate the text we see. ChatGPT was trained with a large corpus of texts to optimize these hundred billion parameters. This particular class of deep learning systems trained to predict the next item in a sequence are called language models, and they are very successful learning features of biological sequences (“biological language model”) due to their intrinsic sequential character [12].
It has been shown that, in a neural network already trained for some objective, the vectors represented by some internal layers, even if not having an evident “translation” into the “objects” originally coded in the input layer, contain modified representations of them that can be useful for other tasks. These internal vectors are called embeddings [12] (Figure 1). They can be used as they are, for example, for clustering input objects based on this new vectorial representation; the clustering generated is better than the equivalent cluster of the original input vectors concerning the problem the neural network is solving. However, they can also be used as input for another neural network trained for a different task, which is termed “transfer learning” (Figure 1). In transfer learning, during the training phase, the first network (previously trained for another task) is only used for re-coding the inputs; its weights are not changed, and only the weights of the second network are tuned for its new task.

3. Deep Learning Approaches for Predicting Protein Functional Sites

There is an enormous variety of machine learning and deep learning approaches for predicting protein functional sites from sequence and/or structural information. Figure 2 schematizes an example of a possible simple deep neural network setup for this task. The input layer of the network codes for different structural and/or sequence features of a particular protein residue. The sequence features usually come from a multiple sequence alignment, as these are rich sources of functional information [13] and represent, for example, the evolutionary pattern of that residue (conservation level, amino-acid variations, etc.). Eventually, it can also include that information for a sequence window around the residue of interest or even the whole alignment to include long-range information (on co-varying positions, for example). At first, most methods tried to use sequence information alone as that is abundant. But now that it is possible to generate high-quality structural models for virtually any protein thanks to AlphaFold and similar systems, most methods also include structural information in one way or another. The structural information coded in the input layer for the residue of interest can include, for example, its amino acid type, those of other residues close to it in the 3D structure, solvent accessibility, secondary structure, distances to other residues, etc. The output layer codes for whether that residue is functional or not, for example, with a single neuron coding “1” for functional and “0” for non-functional. The neural network is trained with (many) examples of functional and non-functional residues (from the same and different proteins) according to the definition of “functional” we are interested in (e.g., active sites, binding sites, post-translational modifications, all, etc.). As commented earlier, during this training process, the network optimizes its internal parameters to maximize the matching between the numerical codification of the features of a residue (input) and its functional character (output). After this training, the network is ready to make a prediction for a new residue.
In the following, we describe some representative examples of deep learning-based approaches for predicting functional and active protein sites. These are also listed in Table 2, together with links to the corresponding source code or web server. Table 1 contains short descriptions of the main deep learning architectures and approaches they use.
The most obvious impact of deep learning on protein functional studies is through the high-quality structural models generated by AlphaFold [7] and similar methods, which shine light on many protein structural and functional aspects as they can model the (virtually) whole protein universe [14,15,16]. In many cases, a 3D protein structure alone gives clues about possible functional sites (e.g., pockets, tunnels, certain secondary structure motifs associated with function, known 3D arrangements of residues in active sites, etc.). Consequently, AlphaFold is predicting functional sites indirectly. For example, AlphaFold has been shown to be a good predictor of disordered regions [17], and these protein segments are usually protein binding sites. Improved versions of the program can also predict the structure of multimers (hence pointing to protein binding sites) and protein–ligand complexes (ligand binding sites and active sites).
Transmembrane helices can also be inferred from deep learning-generated structural models (previous paragraph), for example, looking for α-helices with a specific length and hydrophobicity pattern. But there are also deep learning approaches specifically designed for this task, such as DeepTMpred [18]. There are also predictors for the other type of membrane-spanning structures, the beta-barrels, e.g., BetAwareDeep [19], as well as predictors able to deal with both types of motifs at the same time [20].
There are many works focused on the generation of general-purpose embeddings (see above and Figure 1) for protein local sites. These embeddings are then used for various downstream tasks (“transfer learning”), including the prediction of different kinds of functional features. For example, COLLAPSE [21] uses as input structural features (atoms close in 3D to the residue of interest) as well as evolutionary data (multiple sequence alignments) to generate embeddings coding for a residue and its structural neighborhood. These embeddings are then used to predict protein interaction sites, sites whose mutation changes the stability of the protein, and PROSITE [22] functional motifs. COLLAPSE embeddings can also be used outside prediction scenarios to compare sets of residues (e.g., similarity between the active sites of two proteins).
Some deep learning methods aimed at predicting a protein’s global function can also take those predictions to a local or even residue level thanks to their internal architecture, hence effectively predicting functional sites. For example, the method developed by Jang et al. uses sequence information alone as input for predicting protein function, and the contribution of each residue to that function can be inferred, hence generating also predictions at the residue level [23]. DeepFRI is another predictor of protein function that uses sequences and structures as input [24]. It is able to “zoom in” the representation from the protein to the region level, allowing the detection of sites related to specific functions. PARSE [25] uses the embeddings of local structural features generated with COLLAPSE (commented above) in a transfer learning framework for predicting enzyme functions and can also take these predictions to the residue level, effectively detecting the associated catalytic sites.
ScanNet uses structural information as input and, representing atoms and amino acids, including 3D neighborhood information, effectively predicts protein binding sites in these structures [26]. The program is available through a web server. Another method based on structural information uses a very general representation (raw atom distributions) to predict functional sites represented by PROSITE motifs or catalytic sites [27].
A particular type of functional residues, the post-translational modifications (PTMs), are targeted by an important number of deep learning predictors. For example, Zhu et al. developed a predictor of different PTMs (phosphorylation, acetylation, and ubiquitination) that uses as input the sequence, structure, and dynamical features [28]. The dynamical information is a distinctive characteristic of this predictor, and interestingly, it is shown to contribute largely to the predictive performance of some PTMs. Other examples of PTM predictors are DeepNphos [29], for predicting N-phosphorylation sites, and NetGPI [30], for predicting glycosylphosphatidylinositol anchoring sites.
Another specific type of functional site, the ligand binding site, has been targeted by many works due to its importance for drug binding. For example, [31] predict drug binding sites with a deep learning approach that uses sequences as input, while the predictor developed by [32] uses structural information and was trained to predict ligand binding sites in general, not only those for drugs.
There are also many predictors dealing with protein local features related to protein targeting and sorting [33], usually using sequence information as input, as these features are quite evident at that level. For example, SignalP 6.0 predicts all types of signal peptides using protein language models [34], and TargetP predicts transit peptides, determining the sorting of the protein to different subcellular compartments [35]. Cellular localization predictors usually report the protein region responsible for the targeting. Within these, we can distinguish “two-class” predictors, trained to discriminate whether a protein is going to a particular compartment or not (e.g., [36]), and “multi-class” predictors, able to select among a number of compartments in a single prediction shot (e.g., [37]).
A plethora of deep learning methods have been developed for predicting the effect of amino acid changes (mutations) on protein stability and interactions, mainly with the aim of interpreting disease-associated variants and locating interface hot spots [38]. These can also be considered functional site predictors as these residues are important for different aspects of the protein’s function. For example, DDMut-PPI [39] predicts the effect of mutations on protein interfaces and can be used through a web server. ProMEP [40] uses sequence and structural information to predict mutation effects. The method developed by Brandes et al. uses a protein language model to predict variant effects, and it is fast enough to scan the genome for millions of possible variants [41]. Cagiada et al. trained a predictor on a large dataset of known variant effects and showed how it could effectively predict functional sites of different natures (active sites, allosteric sites, binding sites, etc.) [42]. Some methods are tailored to the prediction of the effect of mutations on a specific type of function, such as EnzyACT, which predicts the effect of mutations on enzyme activity and hence can be used as a predictor of active sites [43].
Table 2. Representative deep learning methods for predicting protein functional sites. All URLs were accessed on 6 December 2024.
Table 2. Representative deep learning methods for predicting protein functional sites. All URLs were accessed on 6 December 2024.
NamePrediction GoalURL(s) and/or ReferenceDeep Learning Architecture/Approach
AlphaFoldProtein 3D structurehttps://alphafoldserver.com/Deep neural network with a diffusion generative model
[7]
DeepTMpredTransmembrane heliceshttps://github.com/ISYSLAB-HUST/DeepTMpredCombination of different DL models
[18]
BetAware-DeepTransmembrane beta barrelshttps://busca.biocomp.unibo.it/betaware2Deep recurrent neural network
[19]
TMbedTransmembrane helices and barrelshttps://github.com/BernhoferM/TMbedEmbbedings of protein language models
[20]
PhiGnetProtein function and associated functional siteshttps://doi.org/10.5281/zenodo.12496869Graph neural networks
[23]
DeepFriProtein function and associated functional siteshttps://beta.deepfri.flatironinstitute.org/Convolutional neural network and pre-trained embedding with protein language models
[24]
PARSEEnzyme functions and associated catalytic siteshttps://github.com/awfderry/PARSECOLLAPSE embeddings [21] and traditional statistics
[25]
ScanNetProtein binding siteshttp://bioinfo3d.cs.tau.ac.il/ScanNet/Geometric deep learning
[26]
--PROSITE motifs [22] and catalytic siteshttps://simtk.org/projects/fscnn3D Convolutional neural network
[27]
cDL-PAU, cDL- FuncPhosDifferent post-translational modificationshttps://github.com/ComputeSuda/PTM_MLCombination of different neural networks
[28]
DeepNphosN-phosphorylation siteshttps://github.com/ChangXulinmessi/DeepNPhosConvolutional neural network
[29]
NetGPIGlycosylphosphatidylinositol anchoring siteshttps://services.healthtech.dtu.dk/service.php?NetGPIRecurrent neural network
https://github.com/mhgislason/netgpi-1.1
[30]
DeepConv-DTIDrug targets and drug-binding residueshttps://github.com/GIST-CSBL/DeepConv-DTIConvolutional neural network
[31]
DeepDrug3DLigand binding siteshttps://github.com/pulimeng/DeepDrug3DConvolutional neural network
[32]
SignalPAll five types of signal peptideshttps://services.healthtech.dtu.dk/service.php?SignalP-6.0Protein language models
https://github.com/fteufel/signalp-6.0
[34]
TargetPTransit peptideshttp://www.cbs.dtu.dk/services/TargetP-2.0/Recurrent neural network
https://github.com/JJAlmagro/TargetP-2.0/
[35]
SCLpred-EMSSorting signal to endomembrane/secretory pathwayhttp://distilldeep.ucd.ie/SCLpred2/Convolutional neural networks
[36]
DeepLocSorting signals for different subcellular compartmentshttps://services.healthtech.dtu.dk/service.php?DeepLoc-2.0Protein language models
[37]
DDMut-PPIEffect of mutations on protein interactionshttps://biosig.lab.uq.edu.au/ddmut_ppiGraph convolutional neural network
[39]
ProMEPEffect of mutationshttps://github.com/wenjiegroup/ProMEPProtein language models
[40]
ESM1bDisease variant effecthttps://github.com/ntranoslab/esm-variantsProtein language model
[41]

4. Example

To illustrate the kind of predictions these methods generate and the input they take, we applied two of these approaches to the well-studied RasH human protein. This is a GTPase that acts in signal transduction, interacting with other proteins in a functional cycle driven by GTP binding and hydrolysis. We used the DeepFRI method [24] through its web server (Table 2) to predict the function(s) of this protein as well as the associated functional sites, in this case, using the sequence as input (UniProt accession: P01112). The server correctly predicts the molecular functions (Gene Ontology terms) known for this protein, “GTP binding” and “GTPase activity”, as well as their ancestors in the GO hierarchy (e.g., “nucleotide binding”, “small molecule binding”, etc.). Figure 3 shows the representation generated by the server of the predicted functional sites associated with two of these GO terms related to GTP binding. These residues clearly map around the known binding site for this nucleotide and comprise the well-described phosphate-binding loop (“P-loop”, at residues 10–17). To complement this prediction, we used another server, ScanNet [26] (Table 2), to predict protein binding sites for this protein using its 3D structure as input (PDB ID: 4q21). The predicted sites (Figure 3) include the “switch-I” and “switch-II” regions of RasH (residues 26–38 and 61–73, respectively), known to be involved in the interactions with effectors [44].

5. Discussion

The prediction of protein functional sites/regions has always been an important area in bioinformatics due to the importance of detecting such sites in protein sequences and structures. Many methodologies exist, and they are routinely applied to proteins of biomedical and biotechnological interest, usually in combination with experimental approaches aimed at confirming the predictions or implementing the proposed mutations for tuning or changing the protein’s function [5].
Among other techniques, classic machine learning methods such as support vector machines or neural networks were applied to this problem [45]. With the advent of deep learning, the field exploded; as for some functional aspects, the large amounts of available data were perfectly suited for these data-hungry systems. This is because the performance of deep learning can be impressive, but these systems require large amounts of data for training. As a rule of thumb, these methods require training a number of examples ten times the number of tunable parameters (i.e., number of connections) [46]. This precludes the usage of these systems for functional aspects for which not enough examples are available for training [33]. This is in part alleviated by “transfer learning” approaches, where a neural network pre-trained for a different problem (for which enough data is available) can generate embeddings that might work better than the original inputs in a simpler neural network trained for a second learning task for which fewer examples are available.
Besides the large amounts of protein sequences available for training and generating embeddings, their intrinsic sequential nature (forgive the redundancy) makes them very suitable for a successful class of deep learning approaches, such as language models. Indeed, revising the temporal trends of the landscape of deep learning methodologies for predicting protein functional sites, a tendency towards protein language models to the detriment of other architectures is evident [12,33]. Finally, a third factor that contributed to the adoption of deep learning (and machine learning in general) in this field is the quite trivial codification of protein sequences into vectors for feeding these systems. For other biological entities, such as chemical compounds, the codification is not so trivial, and some assumptions have to be made, which move away a little from the input vectors from the entities they try to represent.
Another problem with the datasets of functional residues used for training/testing is that they might be biased and/or not represent a protein’s whole repertory of functional sites. For example, the experimentally known binding sites for a particular protein depend on its known complexes, as well as the particular conformational states it presents on them. In this context, many “negatives” (residues considered non-functional) might not be such.
It can be seen that most of these problems will be alleviated when more experimental information (both in size and diversity) becomes available for training these methods.
Although the predictive performance of these methods is excellent, an important drawback compared with other approaches is their intrinsic “black box” nature. This means that, while they predict functional sites with high reliability, in general, it is not possible to obtain either the knowledge on how they made these decisions or the explicit rules to make the predictions out of the network itself. In other words, they do not provide scientific knowledge on functional sites. This is a general problem for neural networks. For example, we can say that AlphaFold is able to “fold” proteins in silico (a long-sought goal in bioinformatics), but we learned nothing about the folding process itself.
As more methods are presented in the literature, independent initiatives for fairly benchmarking these systems using common datasets of functional features are required. This is particularly important for machine learning systems as they are prone to overfitting and lack of generalization, and consequently, the performance values calculated by the authors for a particular training/testing set might not hold for other independent datasets. In this sense, initiatives such as CAFA (Critical Assessment of Function Annotation) [47] are crucial. Establishing standards and good practices for training and evaluating deep learning approaches is also important [48,49]. In any case, as these approaches are based on different concepts and predict different types of functional residues, in general, it is advisable to run more than one method in order to combine the results, as illustrated with the RasH example.
Deep learning is also changing the field of protein design, with methods able to design new sequences compatible with a given fold, function, or able to bind a given molecule, as well as the other ways around [50]. A remarkable milestone in the field was the work of David Baker and his group on a deep learning-based methodology aimed at designing a protein scaffold able to “host” a desired functional site (arrangement of residues in 3D) given as input [51]. Although this can be considered just the opposite of what this review is about (i.e., knowing an active site predicts a protein sequence/structure for it), we include it here for its enormous importance for understanding protein functional sites and how to engineer them for our benefit.
Contrary to protein sequence and structure, which are univocal and well-defined concepts, “protein function” is less defined and difficult to quantify, depends on our available experimental toolkit, and, in many cases, is something we impose on proteins. This problem with the definition of protein function itself permeates to the definition/annotation of “functional site”, and probably our datasets of functional sites are biased or incomplete due to that. A more univocal definition of protein function and functional site, grounded on a molecular basis, as well as associated experimental approaches for determining them, will for sure improve the datasets of experimental functional sites and, consequently, the deep learning methods trained on them.

Author Contributions

Conceptualization, F.P.; investigation, F.P. and B.P.; resources, F.P. and B.P.; writing—original draft preparation, F.P.; writing—review and editing, F.P. and B.P.; visualization, F.P. and B.P.; supervision, F.P. and B.P.; project administration, F.P.; funding acquisition, F.P. All authors have read and agreed to the published version of the manuscript.

Funding

Spanish Ministry of Economy and Competitiveness with European Regional Development Fund (grant number PID2022-140047OB-C22 to F.P.).

Data Availability Statement

No datasets associated to this review article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hu, T.; Chitnis, N.; Monos, D.; Dinh, A. Next-Generation Sequencing Technologies: An Overview. Hum. Immunol. 2021, 82, 801–811. [Google Scholar] [CrossRef] [PubMed]
  2. Shokralla, S.; Spall, J.L.; Gibson, J.F.; Hajibabaei, M. Next-Generation Sequencing Technologies for Environmental DNA Research. Mol. Ecol. 2012, 21, 1794–1805. [Google Scholar] [CrossRef] [PubMed]
  3. The UniProt Consortium. UniProt: The Universal Protein Knowledgebase in 2021. Nucleic Acids Res. 2021, 49, D480–D489. [Google Scholar] [CrossRef] [PubMed]
  4. Rauer, C.; Sen, N.; Waman, V.P.; Abbasian, M.; Orengo, C.A. Computational Approaches to Predict Protein Functional Families and Functional Sites. Curr. Opin. Struct. Biol. 2021, 70, 108–122. [Google Scholar] [CrossRef]
  5. Chagoyen, M.; Garcia-Martin, J.A.; Pazos, F. Practical Analysis of Specificity-Determining Residues in Protein Families. Brief. Bioinform. 2016, 17, 255–261. [Google Scholar] [CrossRef]
  6. Pazos, F. Computational Prediction of Protein Functional Sites-Applications in Biotechnology and Biomedicine. Adv. Protein. Chem. Struct Biol. 2022, 130, 39–57. [Google Scholar] [CrossRef]
  7. Jumper, J.; Evans, R.; Pritzel, A.; Green, T.; Figurnov, M.; Ronneberger, O.; Tunyasuvunakool, K.; Bates, R.; Žídek, A.; Potapenko, A.; et al. Highly Accurate Protein Structure Prediction with AlphaFold. Nature 2021, 596, 583–589. [Google Scholar] [CrossRef]
  8. Baek, M.; DiMaio, F.; Anishchenko, I.; Dauparas, J.; Ovchinnikov, S.; Lee, G.R.; Wang, J.; Cong, Q.; Kinch, L.N.; Schaeffer, R.D.; et al. Accurate Prediction of Protein Structures and Interactions Using a Three-Track Neural Network. Science 2021, 373, 871–876. [Google Scholar] [CrossRef]
  9. Baldi, P. Deep Learning in Science; Cambridge University Press: Cambridge, UK, 2021; ISBN 978-1-108-84535-9. [Google Scholar]
  10. Zou, J.; Huss, M.; Abid, A.; Mohammadi, P.; Torkamani, A.; Telenti, A. A Primer on Deep Learning in Genomics. Nat. Genet. 2019, 51, 12–18. [Google Scholar] [CrossRef]
  11. Boadu, F.; Lee, A.; Cheng, J. Deep Learning Methods for Protein Function Prediction. PROTEOMICS 2024, 12, 2300471. [Google Scholar] [CrossRef]
  12. Simon, E.; Swanson, K.; Zou, J. Language Models for Biological Research: A Primer. Nat. Methods 2024, 21, 1422–1429. [Google Scholar] [CrossRef] [PubMed]
  13. Pazos, F.; Sanchez-Pulido, L. Protein Superfamilies. In eLS; John Wiley & Sons, Ltd: Chichester, UK, 2014. [Google Scholar] [CrossRef]
  14. Bordin, N.; Lau, A.M.; Orengo, C. Large-Scale Clustering of AlphaFold2 3D Models Shines Light on the Structure and Function of Proteins. Mol. Cell 2023, 83, 3950–3952. [Google Scholar] [CrossRef] [PubMed]
  15. Porta-Pardo, E.; Ruiz-Serra, V.; Valentini, S.; Valencia, A. The Structural Coverage of the Human Proteome before and after AlphaFold. PLOS Comput. Biol. 2022, 18, e1009818. [Google Scholar] [CrossRef] [PubMed]
  16. Yang, Z.; Zeng, X.; Zhao, Y.; Chen, R. AlphaFold2 and Its Applications in the Fields of Biology and Medicine. Signal Transduct. Target. Ther. 2023, 8, 115. [Google Scholar] [CrossRef]
  17. Wilson, C.J.; Choy, W.-Y.; Karttunen, M. AlphaFold2: A Role for Disordered Protein/Region Prediction? Int. J. Mol. Sci. 2022, 23, 4591. [Google Scholar] [CrossRef]
  18. Wang, L.; Zhong, H.; Xue, Z.; Wang, Y. Improving the Topology Prediction of α-Helical Transmembrane Proteins with Deep Transfer Learning. Comput. Struct. Biotechnol. J. 2022, 20, 1993–2000. [Google Scholar] [CrossRef]
  19. Madeo, G.; Savojardo, C.; Martelli, P.L.; Casadio, R. BetAware-Deep: An Accurate Web Server for Discrimination and Topology Prediction of Prokaryotic Transmembrane β-Barrel Proteins. J. Mol. Biol. 2021, 433, 166729. [Google Scholar] [CrossRef]
  20. Bernhofer, M.; Rost, B. TMbed: Transmembrane Proteins Predicted through Language Model Embeddings. BMC Bioinform. 2022, 23, 326. [Google Scholar] [CrossRef]
  21. Derry, A.; Altman, R.B. COLLAPSE: A Representation Learning Framework for Identification and Characterization of Protein Structural Sites. Protein. Sci. 2023, 32, e4541. [Google Scholar] [CrossRef]
  22. Sigrist, C.J.A.; de Castro, E.; Cerutti, L.; Cuche, B.A.; Hulo, N.; Bridge, A.; Bougueleret, L.; Xenarios, I. New and Continuing Developments at PROSITE. Nucleic Acids Res. 2013, 41, D344–D347. [Google Scholar] [CrossRef]
  23. Jang, Y.J.; Qin, Q.-Q.; Huang, S.-Y.; Peter, A.T.J.; Ding, X.-M.; Kornmann, B. Accurate Prediction of Protein Function Using Statistics-Informed Graph Networks. Nat. Commun. 2024, 15, 6601. [Google Scholar] [CrossRef] [PubMed]
  24. Gligorijević, V.; Renfrew, P.D.; Kosciolek, T.; Leman, J.K.; Berenberg, D.; Vatanen, T.; Chandler, C.; Taylor, B.C.; Fisk, I.M.; Vlamakis, H.; et al. Structure-Based Protein Function Prediction Using Graph Convolutional Networks. Nat. Commun. 2021, 12, 3168. [Google Scholar] [CrossRef] [PubMed]
  25. Derry, A.; Altman, R.B. Explainable Protein Function Annotation Using Local Structure Embeddings. bioRxiv 2023. [Google Scholar] [CrossRef]
  26. Tubiana, J.; Schneidman-Duhovny, D.; Wolfson, H.J. ScanNet: An Interpretable Geometric Deep Learning Model for Structure-Based Protein Binding Site Prediction. Nat. Methods 2022, 19, 730–739. [Google Scholar] [CrossRef]
  27. Torng, W.; Altman, R.B. High Precision Protein Functional Site Detection Using 3D Convolutional Neural Networks. Bioinformatics 2019, 35, 1503–1512. [Google Scholar] [CrossRef]
  28. Zhu, F.; Yang, S.; Meng, F.; Zheng, Y.; Ku, X.; Luo, C.; Hu, G.; Liang, Z. Leveraging Protein Dynamics to Identify Functional Phosphorylation Sites Using Deep Learning Models. J. Chem. Inf. Model. 2022, 62, 3331–3345. [Google Scholar] [CrossRef]
  29. Chang, X.; Zhu, Y.; Chen, Y.; Li, L. DeepNphos: A Deep-Learning Architecture for Prediction of N-Phosphorylation Sites. Comput. Biol. Med. 2024, 170, 108079. [Google Scholar] [CrossRef]
  30. Gíslason, M.H.; Nielsen, H.; Almagro Armenteros, J.J.; Johansen, A.R. Prediction of GPI-Anchored Proteins with Pointer Neural Networks. Curr. Res. Biotechnol. 2021, 3, 6–13. [Google Scholar] [CrossRef]
  31. Lee, I.; Keum, J.; Nam, H. DeepConv-DTI: Prediction of Drug-Target Interactions via Deep Learning with Convolution on Protein Sequences. PLOS Comput. Biol. 2019, 15, e1007129. [Google Scholar] [CrossRef]
  32. Pu, L.; Govindaraj, R.G.; Lemoine, J.M.; Wu, H.-C.; Brylinski, M. DeepDrug3D: Classification of Ligand-Binding Pockets in Proteins with a Convolutional Neural Network. PLOS Comput. Biol. 2019, 15, e1006718. [Google Scholar] [CrossRef]
  33. Savojardo, C.; Martelli, P.L.; Casadio, R. Finding Functional Motifs in Protein Sequences with Deep Learning and Natural Language Models. Curr. Opin. Struct. Biol. 2023, 81, 102641. [Google Scholar] [CrossRef] [PubMed]
  34. Teufel, F.; Almagro Armenteros, J.J.; Johansen, A.R.; Gíslason, M.H.; Pihl, S.I.; Tsirigos, K.D.; Winther, O.; Brunak, S.; von Heijne, G.; Nielsen, H. SignalP 6.0 Predicts All Five Types of Signal Peptides Using Protein Language Models. Nat. Biotechnol. 2022, 40, 1023–1025. [Google Scholar] [CrossRef] [PubMed]
  35. Armenteros, J.J.A.; Salvatore, M.; Emanuelsson, O.; Winther, O.; von Heijne, G.; Elofsson, A.; Nielsen, H. Detecting Sequence Signals in Targeting Peptides Using Deep Learning. Life Sci. Alliance 2019, 2, e201900429. [Google Scholar] [CrossRef] [PubMed]
  36. Kaleel, M.; Zheng, Y.; Chen, J.; Feng, X.; Simpson, J.C.; Pollastri, G.; Mooney, C. SCLpred-EMS: Subcellular Localization Prediction of Endomembrane System and Secretory Pathway Proteins by Deep N-to-1 Convolutional Neural Networks. Bioinformatics 2020, 36, 3343–3349. [Google Scholar] [CrossRef] [PubMed]
  37. Thumuluri, V.; Almagro Armenteros, J.J.; Johansen, A.R.; Nielsen, H.; Winther, O. DeepLoc 2.0: Multi-Label Subcellular Localization Prediction Using Protein Language Models. Nucleic Acids Res. 2022, 50, W228–W234. [Google Scholar] [CrossRef]
  38. Diaz, D.J.; Kulikova, A.V.; Ellington, A.D.; Wilke, C.O. Using Machine Learning to Predict the Effects and Consequences of Mutations in Proteins. Curr. Opin. Struct. Biol. 2023, 78, 102518. [Google Scholar] [CrossRef]
  39. Zhou, Y.; Myung, Y.; Rodrigues, C.H.M.; Ascher, D.B. DDMut-PPI: Predicting Effects of Mutations on Protein-Protein Interactions Using Graph-Based Deep Learning. Nucleic Acids Res. 2024, 52, W207–W214. [Google Scholar] [CrossRef]
  40. Cheng, P.; Mao, C.; Tang, J.; Yang, S.; Cheng, Y.; Wang, W.; Gu, Q.; Han, W.; Chen, H.; Li, S.; et al. Zero-Shot Prediction of Mutation Effects with Multimodal Deep Representation Learning Guides Protein Engineering. Cell Res 2024, 34, 630–647. [Google Scholar] [CrossRef]
  41. Brandes, N.; Goldman, G.; Wang, C.H.; Ye, C.J.; Ntranos, V. Genome-Wide Prediction of Disease Variant Effects with a Deep Protein Language Model. Nat. Genet. 2023, 55, 1512–1522. [Google Scholar] [CrossRef]
  42. Cagiada, M.; Bottaro, S.; Lindemose, S.; Schenstrøm, S.M.; Stein, A.; Hartmann-Petersen, R.; Lindorff-Larsen, K. Discovering Functionally Important Sites in Proteins. Nat. Commun. 2023, 14, 4175. [Google Scholar] [CrossRef]
  43. Li, G.; Zhang, N.; Dai, X.; Fan, L. EnzyACT: A Novel Deep Learning Method to Predict the Impacts of Single and Multiple Mutations on Enzyme Activity. J. Chem. Inf. Model. 2024, 64, 5912–5921. [Google Scholar] [CrossRef] [PubMed]
  44. Harrison, R.A.; Lu, J.; Carrasco, M.; Hunter, J.; Manandhar, A.; Gondi, S.; Westover, K.D.; Engen, J.R. Structural Dynamics in Ras and Related Proteins upon Nucleotide Switching. J. Mol. Biol. 2016, 428, 4723–4735. [Google Scholar] [CrossRef] [PubMed]
  45. Baldi, P.; Brunak, S. Bioinformatics, Second Edition: The Machine Learning Approach; MIT Press: Cambridge, MA, USA, 2001. [Google Scholar]
  46. Alwosheel, A.; van Cranenburgh, S.; Chorus, C.G. Is Your Dataset Big Enough? Sample Size Requirements When Using Artificial Neural Networks for Discrete Choice Analysis. J. Choice Model. 2018, 28, 167–182. [Google Scholar] [CrossRef]
  47. Zhou, N.; Jiang, Y.; Bergquist, T.R.; Lee, A.J.; Kacsoh, B.Z.; Crocker, A.W.; Lewis, K.A.; Georghiou, G.; Nguyen, H.N.; Hamid, M.N.; et al. The CAFA Challenge Reports Improved Protein Function Prediction and New Functional Annotations for Hundreds of Genes through Experimental Screens. Genome Biol. 2019, 20, 244. [Google Scholar] [CrossRef]
  48. Jones, D.T. Setting the Standards for Machine Learning in Biology. Nat. Rev. Mol. Cell Biol. 2019, 20, 659–660. [Google Scholar] [CrossRef]
  49. Walsh, I.; Fishman, D.; Garcia-Gasulla, D.; Titma, T.; Pollastri, G.; Harrow, J.; Psomopoulos, F.E.; Tosatto, S.C.E. DOME: Recommendations for Supervised Machine Learning Validation in Biology. Nat. Methods 2021, 18, 1122–1127. [Google Scholar] [CrossRef]
  50. Khakzad, H.; Igashov, I.; Schneuing, A.; Goverde, C.; Bronstein, M.; Correia, B. A New Age in Protein Design Empowered by Deep Learning. Cell Syst. 2023, 14, 925–939. [Google Scholar] [CrossRef]
  51. Wang, J.; Lisanza, S.; Juergens, D.; Tischer, D.; Watson, J.L.; Castro, K.M.; Ragotte, R.; Saragovi, A.; Milles, L.F.; Baek, M.; et al. Scaffolding Protein Functional Sites Using Deep Learning. Science 2022, 377, 387. [Google Scholar] [CrossRef]
Figure 1. Neural networks. Top: representation of a simple neural network with one hidden layer. The input is encoded as a two-component vector. These input values are “propagated” through the network, taking into account the weights of the connections (blue) to finally generate the output vector (a single value in this case). During the training phase, the weights (blue) are tuned to maximize the matching between the generated outputs and the real ones of the training set. Bottom: schematic representation of a larger neural network with many hidden layers and internal neurons. One of these layers (light blue box) can be taken as an alternative coding for the input vectors (embedding) and used, among other things, as input for another neural network (right) trained for a different task (“transfer learning”).
Figure 1. Neural networks. Top: representation of a simple neural network with one hidden layer. The input is encoded as a two-component vector. These input values are “propagated” through the network, taking into account the weights of the connections (blue) to finally generate the output vector (a single value in this case). During the training phase, the weights (blue) are tuned to maximize the matching between the generated outputs and the real ones of the training set. Bottom: schematic representation of a larger neural network with many hidden layers and internal neurons. One of these layers (light blue box) can be taken as an alternative coding for the input vectors (embedding) and used, among other things, as input for another neural network (right) trained for a different task (“transfer learning”).
Molecules 30 00214 g001
Figure 2. Example of a deep neural network setup for predicting functional sites. The input layer of the network (green) codes a set of sequence and/or structural features of a particular protein residue, such as amino acid type, evolutionary pattern in a multiple sequence alignment, neighbor residues in 3D, solvent accessibility, etc. The output layer is, in this case, a single neuron (blue) whose value represents whether the residue is functional (e.g., “1.0”) or not (“0.0”).
Figure 2. Example of a deep neural network setup for predicting functional sites. The input layer of the network (green) codes a set of sequence and/or structural features of a particular protein residue, such as amino acid type, evolutionary pattern in a multiple sequence alignment, neighbor residues in 3D, solvent accessibility, etc. The output layer is, in this case, a single neuron (blue) whose value represents whether the residue is functional (e.g., “1.0”) or not (“0.0”).
Molecules 30 00214 g002
Figure 3. Example of a prediction of functional sites and protein binding sites from sequence and structure information using two deep learning approaches. The top panel shows the functional sites associated with two Gene Ontology molecular function terms (“nucleotide binding” and “GTP binding”) predicted by DeepFRI [24] for the human RasH protein, using its sequence as the only input. The color scale represents the prediction score, from blue (low) to red (high). Only the N-terminal part of the protein, where the top predictions are, is shown. The three residues with the highest scores (G10, G12, and V14) are also highlighted in yellow in the 3D structure of this protein (lower panel), where the bound nucleotide (GTP) is shown in stick representation. The lower panel shows the prediction of protein binding sites generated with ScanNet [26] for this same protein using structural information as input, in a color scale going from bluish colors (low) to reddish (high). This 3D representation was generated by the Molstar molecular viewer (https://molstar.org/) integrated into the ScanNet server.
Figure 3. Example of a prediction of functional sites and protein binding sites from sequence and structure information using two deep learning approaches. The top panel shows the functional sites associated with two Gene Ontology molecular function terms (“nucleotide binding” and “GTP binding”) predicted by DeepFRI [24] for the human RasH protein, using its sequence as the only input. The color scale represents the prediction score, from blue (low) to red (high). Only the N-terminal part of the protein, where the top predictions are, is shown. The three residues with the highest scores (G10, G12, and V14) are also highlighted in yellow in the 3D structure of this protein (lower panel), where the bound nucleotide (GTP) is shown in stick representation. The lower panel shows the prediction of protein binding sites generated with ScanNet [26] for this same protein using structural information as input, in a color scale going from bluish colors (low) to reddish (high). This 3D representation was generated by the Molstar molecular viewer (https://molstar.org/) integrated into the ScanNet server.
Molecules 30 00214 g003
Table 1. Deep learning architectures and approaches used by the described methods.
Table 1. Deep learning architectures and approaches used by the described methods.
TypeDescription
TransformerArchitecture specifically designed for processing data with a sequential character, mainly text. It can dynamically weigh the significance of different components of the input via a mechanism called “attention”.
Language modelDeep learning approach that, using transformers or other architectures, is trained to predict the next item in a sequence.
Protein language modelLanguage model that handles protein amino-acid sequences
Generative modelApproach designed to generate new data “similar” to those it was trained on. For example, generating “feasible” new protein sequences.
Recurrent neural networksNeural network architecture in which the inputs and outputs are of the same nature and codified in the same way. Consequently, the output of a given iteration can be used as input for the next. They are used for processing sequential or temporal data, for example.
Graph neural networkNeural network whose architecture (neurons and connection) is not general but that of a graph representing the phenomenon the network is intended to model.
Convolutional neural networkNeural network architecture in which matrix operations are applied on some of its layers besides the typical forward propagation operations. Especially suited for handling spatial data, either 2D (e.g., images) or 3D (3D convolutional NN, e.g., three-dimensional structures)
Geometric deep neural networkDeep neural network architecture especially suited for handling generic geometrical data.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Pitarch, B.; Pazos, F. Deep Learning Approaches for the Prediction of Protein Functional Sites. Molecules 2025, 30, 214. https://doi.org/10.3390/molecules30020214

AMA Style

Pitarch B, Pazos F. Deep Learning Approaches for the Prediction of Protein Functional Sites. Molecules. 2025; 30(2):214. https://doi.org/10.3390/molecules30020214

Chicago/Turabian Style

Pitarch, Borja, and Florencio Pazos. 2025. "Deep Learning Approaches for the Prediction of Protein Functional Sites" Molecules 30, no. 2: 214. https://doi.org/10.3390/molecules30020214

APA Style

Pitarch, B., & Pazos, F. (2025). Deep Learning Approaches for the Prediction of Protein Functional Sites. Molecules, 30(2), 214. https://doi.org/10.3390/molecules30020214

Article Metrics

Back to TopTop