Abstract
Protein function prediction plays an important role in the field of biology. With the wide application of deep learning in the field of bioinformatics, more and more natural language processing (NLP) technologies are applied to the downstream tasks in the field of bioinformatics, and it has also shown excellent performance in protein function prediction. Protein-protein interaction (PPI) networks and other biological attributes contain rich information critical for annotating protein functions. However, existing deep learning networks still suffer from overfitting and noise issues, resulting in low accuracy in protein function prediction. Consequently, developing efficient models for protein function prediction remains a popular and challenging topic in the application of NLP in bioinformatics. In this study, we propose a novel protein function prediction model based on attention mechanisms, termed IMAGO. This model employs the Transformer pre-training process, integrating multi-head attention mechanisms and regularization techniques, and optimizes the loss function to effectively reduce overfitting and noise issues during training. It generates more robust embeddings, ultimately improving the accuracy of protein function prediction. Experimental results on human and mouse datasets indicate that our model surpasses other protein function prediction models across multiple metrics. Thus, this efficient, stable, and accurate deep learning model holds significant promise for protein function prediction.
1. Introduction
Proteins are the core molecules executing nearly all biological functions within living organisms. They play crucial roles in catalyzing reactions, transmitting signals, and providing structural support both inside and outside cells. Understanding protein functions is therefore vital for elucidating biological processes and developing new drugs []. The study of protein functions not only helps in comprehending the basic mechanisms of organisms but also provides important clues for drug development, disease diagnosis, and treatment. However, due to the high cost and time consumption of experimental methods, despite over 200 million known proteins, less than 1% of protein functions have been elucidated [,]. Experimentally determining all protein functions is nearly impossible in reality. Consequently, computational methods for predicting protein functions have emerged as an effective alternative. These methods infer protein functions primarily from protein sequences, structures, and other biological information. This not only accelerates the discovery of unknown protein functions but also provides critical clues for new drug development, disease diagnosis, and treatment [].
Protein functions are typically classified into three aspects by the Gene Ontology (GO): Biological Process (BP), Molecular Function (MF), and Cellular Component (CC) []. In recent years, with the development of deep learning technologies, the accuracy and efficiency of protein function prediction have significantly improved. Deep learning methods, leveraging large biological datasets, can automatically extract features and perform complex pattern recognition, making large-scale function prediction feasible []. Therefore, combining experimental data with computational predictions can more comprehensively and efficiently reveal protein functions, providing a crucial foundation for biomedical research and clinical applications.
In recent years, the field of protein function prediction has advanced rapidly, resulting in numerous automated protein function annotation algorithms []. Notably, multimodal fusion approaches based on deep learning, including DeepGO-SE [], GPSFun [], and TAWFN [], have further advanced the prediction accuracy and applicability of protein function annotation. Protein function prediction methods have become increasingly diverse, generally classified into four categories: sequence similarity methods, structure similarity methods, network-based machine learning methods, and hybrid source methods.
Sequence similarity methods were among the earliest approaches for protein function prediction, based on the sequence-function homology hypothesis [,,] that proteins with high sequence similarity often share similar functions. Blakeley et al. provided an overview of sequence alignment and homology analysis methods, noting their limitations in cases where homologous sequences are scarce in databases []. Deng et al. introduced a sequence encoding method using LSTM networks, significantly improving prediction accuracy []. However, studies have shown that some proteins may have similar functions despite having low sequence similarity [,], posing challenges to sequence-based methods.
Structure-based methods predict protein functions by utilizing the three-dimensional structures of proteins []. Structure similarity methods focus on protein three-dimensional structures, with the premise that proteins with similar structures tend to exhibit similar functions []. Gomez et al. developed a function prediction framework based on structural alignment []. Liu et al. proposed a framework using a multilevel attention mechanism to realize the deep interactive fusion of protein sequence embedding and structural features extracted from contact graphs []. Additionally, structure similarity prediction has benefited from convolutional neural network (CNN) models, which automatically capture functional motifs in local protein structures, effectively enhancing prediction accuracy []. However, this prediction method may be inaccurate in some cases [].
Network-based machine learning methods utilize protein-protein interaction (PPI) networks or gene co-expression networks to build predictive models []. Mostafavi et al. predicted protein function through label propagation between network nodes []. Wang et al. proposed a framework based on graph neural networks (GNNs) that effectively captures complex interactions between proteins []. Compared to traditional network analysis methods, GNNs propagate information across network nodes and edges, further enhancing prediction performance. However, the application of these methods can be limited by several factors, including inherent noise in PPI networks, the computational complexity of processing large-scale biological networks, and the expertise required for effective feature engineering and model tuning [].
Finally, hybrid source methods integrate multiple sources of information, including sequence, structure, and PPI networks, to provide a more comprehensive function prediction. These methods leverage the diversity of biological data to capture richer functional features. Wu et al. proposed a multimodal data fusion model that combines different data sources to achieve higher prediction accuracy []. Hybrid source methods are also regarded as a promising direction for future protein function prediction, addressing the information limitations associated with single-source data []. This method has many applications in other biological fields, such as RNA prediction. Zhang et al. uses a variety of biological signals, such as histone modification and RNA-seq data, which significantly improves the accuracy of eRNA recognition []. In the field of protein function prediction, the model [] proposed by Wu et al., which combines PPI network with protein’s biological attributes, has shown remarkable effects, but they are also affected by noise and details of over-fitting in training data. Therefore, there is an urgent need for new methods to improve attention mechanism-based models to enhance the accuracy of protein function prediction.
In this study, we proposed a new protein function prediction method, IMAGO, based on attention mechanism. In the pre-training process using Transformer, regularization technology was added, and the loss function was optimized. Comparative experiments on human and mouse datasets demonstrated that the IMAGO model improved the accuracy of protein function prediction by up to 10.65%, 11.76%, 9%, 38.89%, and 10.85% in m-AUPR, m-AUPR, F1, ACC, and Fmax, respectively, suggesting that our model achieves a modest improvement in protein function prediction accuracy under specific datasets and conditions.
2. Materials and Methods
2.1. Dataset
Our dataset includes two species: human (taxonomy code 9606) and mouse (taxonomy code 10090). PPI data and protein sequence data were retrieved from the STRING database (v11.5) []. We used “combined” type PPI data comprising “experimental”, “coexpression”, “cooccurrence”, “neighborhood”, “fusion”, “database”, and “text mining” types. GO functional annotation data were retrieved from the Gene Ontology resource website (version 2022-01-13). Protein attributes, including subcellular localization and Pfam protein domain annotations, were obtained from the Uniprot database (v3.5.175). Proteins annotated before 1 January 2018, were defined as the training set, those annotated between 2 January 2018, and 31 December 2020, were defined as the validation set, and those annotated after 1 January 2021, were defined as the test set. The partitioning was based on the earliest recorded GO annotation date for each protein to ensure exclusive assignment to one dataset. For proteins with multiple annotations at different time points, the earliest date determined their assignment to the training, validation, or test set. To ensure sufficient proteins for each label, we used GO terms with at least 10, 5, and 1 proteins in the training, validation, and test sets, respectively. Additionally, to reduce the influence of dependencies between GO terms, we removed GO terms that annotated more than 5% of the PPI network proteins of the species. The statistics of the training, validation, and test sets for human and mouse data used in this study are shown in Table 1.
Table 1.
The data statistics of human and mouse branches of gene ontology were used in the study, the same as in [].
It is crucial to note that the data partitioning between the pre-training and fine-tuning stages in this study strictly follows a chronological order. The protein samples utilized during the self-supervised pre-training phase are exclusively drawn from the training set (i.e., proteins annotated before 1 January 2018). The information from proteins in the validation and test sets remains entirely inaccessible to the model during pre-training. This protocol ensures that the model is not exposed to data from future time points during its initial representation learning, thereby effectively mitigating potential data leakage. In the subsequent fine-tuning stage, the model is trained solely using the labels from the training set. The validation set is employed for strategies such as early stopping to prevent overfitting, and the final performance is evaluated rigorously on the completely independent test set.
2.2. Integration of PPI and Protein Attributes
The IMAGO model employs a multi-head attention mechanism to integrate single-species PPI networks and protein biological attributes. Based on the previously published transformer-based autoencoder model (TransformerAE) [], this model takes protein attributes and PPI networks as inputs and outputs low-dimensional embeddings. The PPI network is transformed into a weighted adjacency matrix, with weights determined by the minimum normalized scores of the corresponding evidence. Protein attributes, including Pfam domains and subcellular locations, are encoded into binary vectors using a bag-of-words model: if a protein possesses a specific domain or subcellular location, the corresponding element in the vector is set to 1. To enhance data quality, domain terms occurring less than six times in proteins were removed.
The IMAGO model comprises two stages: self-supervised pre-training and fine-tuning (see Figure 1a,b).
Figure 1.
IMAGO model diagram. (a) IMAGO pre-training stage: the model uses the training set proteins to reconstruct their original features (PPI network and attributes). (b) IMAGO fine-tuning stage: the pre-trained encoder is frozen or jointly fine-tuned with a new predictor using the training set(with GO labels), and the validation set is used for hyperparameter tuning. (c) Multi-attention mechanism. (d) Predictor structure. Where and are the original features of protein i from different sources, and and are the reconstructed protein i features.
2.3. Self-Supervised Pre-Training
The IMAGO model adopts an encoder–regularizer–decoder structure to learn the hidden embedding vectors of proteins by reconstructing the original features. During pre-training, the encoder and decoder are used to integrate information from the two data sources. For protein i, its two original source features are represented as and , where is the feature dimension of source m.
The encoder has two parallel multi-layer perceptrons (MLPs), each used for one source feature, and L multiple multi-head attention layers. To match the original data dimensions, the original feature vector of protein i from source m, is projected to a common vector with d, dimensions by a two-layer MLP, which is defined as:
where f is the non-linear activation function, LN is the layer normalization function [], and , are the weight matrices, and , are the bias vectors, is the size of the MLP hidden layer. Then, the projected vectors from the two sources are cross-integrated through multi-head attention layers to generate hidden protein embedding vectors.
In the TransformerAE model, the adjacency matrix and protein attribute matrix jointly pass through the encoder’s six-layer multi-head attention layers (see Figure 1c) and the decoder’s six-layer multi-head attention layers to achieve information fusion from both data sources. The attention mechanism’s formula is:
where Q is the query, K is the key, and V is the value matrix [], and is the dimension of the query and key vectors in the matrix.
The regularizer’s primary role is to adjust and optimize feature representations in the network, making these features more conducive to the model’s generalization ability. In deep learning models, regularization helps the model maintain robustness when faced with unseen data, preventing the model from overfitting specific noise and details in the training data. By processing intermediate features in the network to adjust their distributions, the regularizer can reduce differences between features and increase their intrinsic consistency. This typically helps improve classification or other task accuracies. In the pre-training process, the regularizer also continuously updates. The structure of the regularizer is as follows:
where and are the weights of the hidden layers, while is the parameter in the output layer, and is the activation function. Updating of regularizer in pre-training:
where N is the total number of proteins in the PPI network, f is a regularization function, is the embedding of the hidden protein of the source m, and is a random vector with normal distribution randomly generated from it, and its shape is the same as the embedding vector of the hidden protein. The regularizer is constantly updated in the pre-training process, and its features conform to the specific expected distribution by adjusting the distribution of input vectors. The random normal distribution vectors act as the target distribution, guiding the input vectors’ distribution to gradually approach a normal distribution.
The decoder’s structure mirrors the encoder’s. The decoder first inputs the hidden embedding vectors into L multi-head attention layers.Then, for protein i, the feature vector of source m is reconstructed by an whose structure is symmetric to the corresponding in encoder, denoting as . The decoder uses the function as the activation function of the output layer.
During self-supervised pre-training, we optimized the loss function to minimize the difference between the original inputs before entering the encoder and the reconstructed outputs after passing through the decoder. Here, we present the loss function’s expression, starting with the binary cross-entropy loss:
where N is the total number of proteins in the PPI network, and are the j dimensions of and , respectively. In addition, constraints are added to the embedding generated by the encoder to regularize the model. The expression of the regularization term is as follows:
where f is defined as the L2 regularization function, given by , with denoting the regularization coefficient. We also introduce cosine similarity loss, which is used to measure the cosine similarity between the original embedding and the reconstructed embedding to ensure that the embedding is similar to the original embedding in direction. The expression is as follows:
where N is the total number of proteins in the PPI network, is the embedding of the hidden protein of the source m, is the original feature vector of source m.
The loss function is computed as follows:
Each part of the loss plays a crucial role in ensuring the model effectively learns and maintains the integrity of the learned embeddings during training.
2.4. Fine-Tuning for Protein Function Prediction
In this study, protein function prediction is modeled as a multi-label task. We combined the pre-trained encoder with a predictor to predict protein functions. The predictor is a two-layer perceptron used to predict protein labels (see Figure 1d). Given that the number of target GO terms is K, the predictor takes the concatenation of the embeddings from the two sources and generated by the encoder as input and outputs a score vector of dimension K for the GO terms. The prediction score vector of protein i is defined as follows:
where is concatenation operator, and is the function, is the size of the predictor’s hidden layer, and are the weight matrices of predictor’s hidden and output layers, respectively. and are the bias vectors of predictor’s hidden and output layers, respectively. For protein function prediction, we used the Asymmetric Loss () [] as the loss function to achieve better performance in the multi-label task. is designed to address the potential severe imbalance of samples between labels in typical multi-label datasets. is defined as follows:
where is the number of proteins in the training set, K is the number of functions in a specific category, and are the focusing parameters for positive and negative samples. In this study, we set to 0 and to 2 to reduce the contribution of easy negative samples, encouraging the model to make more positive predictions.
2.5. Experimental Setup
We used the same empirical hyperparameter set for training on the three GO aspects (BP, MF, and CC) of the two species (human and mouse), and use the verification set and training set to train the model. The performance on the test set will be the evaluation result. In addition, we also carried out experiments on fruit flies and zebrafish to verify our model. The pre-training and fine-tuning parameters of the model, except for the pre-training epochs, are consistent with the CFAGO model []. This is because according to the research [], the effect of 500 rounds of pre-training with Transform AE is basically the same as that of 5000 rounds. The regularizer comprises an MLP with an input dimension of 512, hidden layer dimensions of 64 and 32, and an output dimension of 1. The IMAGO model was implemented in Python 3.11 using Pytorch 2.1, and a single training session on an RTX 4060 Ti GPU with 16GB of memory took approximately 5 to 6 h.
For the visualization of protein embedding clusters, we employed t-Distributed Stochastic Neighbor Embedding (t-SNE), a nonlinear dimensionality reduction technique particularly effective for visualizing high-dimensional data by preserving local structures. Implementations used the scikit-learn TSNE class with default parameters, reducing embeddings to two dimensions (random state = 0) to generate the two-dimensional projections.
We compare IMAGO with two baseline methods and five network-based models. The naive method directly assigns the relative frequency of terms in the training set to each protein in the test set as a predictive score. Based on the sequence similarity, the BLAST method transfers the GO term of the protein in the training set to the corresponding protein in the test set through blastp and assigns it according to the comparison score. GeneMANIA [] predicts gene function through a weighted proximity algorithm, and integrates multiple data sources into a unified network to dynamically adjust the weight of each data source to optimize the prediction effect. Mashup [] generates a balanced hidden diffusion state through multi-network topology fusion, using different PPI network information. Based on MLPAE, deepNF [] embeds multiple PPI networks into a joint potential representation. Graph2GO [] combines variational graph automatic encoder (GAE) and sequence similarity network to integrate protein attributes with PPI network features for function prediction. CFAGO [] uses the Transformer-based automatic encoder (TransformerAE) to cross-fuse the PPI network and protein attributes.
2.6. Evaluation Metrics
In this study, we used five metrics to evaluate prediction performance, including two types of area under the precision-recall curve, namely micro-average AUPR (m-AUPR) and macro-average AUPR (M-AUPR), F1 score (F1), accuracy (ACC), and F-max score (Fmax). The first three metrics are function-centric, evaluating proteins annotated with each GO term, while the latter two metrics are protein-centric, assessing GO terms annotated on each protein. Fmax is defined as follows:
where is a flexible threshold for obtaining the highest score.
The precision and recall for multi-label tasks are defined as follows:
denotes the number of proteins that are predicted with at least one function. k is the total number of labels for a specific functional category. is the predicted score for the function and is the ground truth with 1 indicating the existence of the function. n is the total number of proteins to be evaluated.
3. Results
3.1. Experimental Results
Figure 2 shows that IMAGO is superior to other models in GO and many indicators on human and mouse data sets. Among them, IMAGO improves the human data set most obviously. Compared with the CFAGO model, the highest improvements in m-AUPR, m-AUPR, F1, ACC, and Fmax are 10.65%, 11.76%, 9%, 38.89% and 10.85%. And the improvements in the human dataset are most notable.The comparison results between IMAGO and other models indicate that IMAGO, with the addition of regularization techniques and optimized loss functions, achieves the highest scores in most cases for protein function prediction.
Figure 2.
Performance of IMAGO and other models in protein function prediction.
The performance of IMAGO in the mouse data set is not as good as that in the human data set, which may be related to the characteristics of the data set. It can also be reflected from the loss in the pre-training stage of IMAGO (see Figure 3), and the convergence speed of the mouse data set is slower than that of CFAGO. It is precisely because different data sets have different noise levels and feature distributions that the regularization and loss function optimization effect of IMAGO on some data sets may be more advantageous.
Figure 3.
Loss in the pre-training stage of the model. On the left is the pre-training loss of human data set, and on the right is the pre-training loss of mouse data set.
3.2. Ablation Experiment
We conducted ablation experiments on human datasets to investigate the contribution of training epochs, the regularizer, and the optimized loss function to performance improvement. Among them, the indicator m-AUPR is shown in Table 2.
Table 2.
Evaluation of ablation regularization and optimized loss function in m-AUPR (human dataset). Underline indicates the highest score.
We set up five control groups: CFAGO with 500 training epochs (equivalent to IMAGO without the regularizer and optimized loss function), IMAGO with 5000 training epochs, IMAGO with 500 training epochs without the regularizer, IMAGO with 500 training epochs without the optimized loss function, and IMAGO with 500 training epochs. We observed that IMAGO with 500 training epochs and IMAGO with 5000 training epochs achieved the highest scores in different GO aspects. This is attributed to the cooperation between the regularizer and the loss function, which together suppress the long tail noise and modal conflict in biological data and make the model perform best in complex GO categories (such as human BP).
3.3. Visualization of Protein Embeddings
To intuitively demonstrate the quality of protein representations learned by the IMAGO model, we employed t-SNE (t-distributed Stochastic Neighbor Embedding) to project the generated protein embeddings into a two-dimensional space for visualization. As illustrated in Figure 4, which uses the Biological Process (BP) category of the human dataset as an example, distinct colors represent proteins annotated with specific GO functional terms. The visualization reveals that protein embeddings with similar functions form well-defined clusters in the projected space, while those with divergent functions are clearly separated. This pattern demonstrates that the IMAGO model effectively captures the underlying semantic relationships between protein functions, producing highly discriminative embedding representations. These qualitative results provide intuitive evidence supporting the model’s superior classification performance, as quantitatively reported in the previous sections.
Figure 4.
Evaluation of clustering visualization in BP (visualized using t-SNE dimensionality reduction technique).
The exceptional capability of IMAGO to differentiate protein representations can be attributed to the synergistic effects of its core components: the multi-head attention mechanism adeptly captures long-range dependencies and complex interaction patterns among proteins; the incorporated regularization techniques mitigate overfitting during training, thereby enhancing the robustness of the generated embeddings; and the optimized loss function ensures the effective learning and preservation of embedding integrity throughout the training process. Collectively, these techniques empower IMAGO to produce highly distinguishable and robust protein embeddings, which form a solid foundation for accurate protein function prediction.
3.4. Comparative Evaluation of IMAGO on Diverse Species
In addition, in order to verify the effect of the IMAGO model on other species, we carried out experiments on Drosophila and Zebrafish according to the previous method of data set processing, and the experimental results are shown in Table 3.
Table 3.
Experimental results of IMAGO model on datasets of four different species across BP, MF and CC aspects.
The performance of the IMAGO model in cross-species testing is highly dependent on biological data types and species characteristics: molecular function (MF) is predicted to reach the dominant performance in zebrafish, which is due to the evolutionary conservation and complete annotation of enzyme active sites, while biological process (BP) is predicted to collapse in zebrafish due to the space-time specific labeling gap in the development process; Human are ahead of BP by improving the pathway database, but the complex regulatory mechanism leads to the weakness of MF. The prediction of subcellular localization (CC) shows the cross-species stability due to the structural conservatism of organelles, revealing the fundamental contradiction that the prediction of static structural characteristics is significantly better than that of dynamic biological processes, and the performance of the model is limited by the annotation completeness and functional conservatism of species data [].
To further position IMAGO against a contemporary state-of-the-art model that addresses the same task with a complementary approach, we compare our results with DualNetGO [], another advanced model built upon the CFAGO framework (see Figure 5). While both models aim to enhance protein function prediction, their improvement strategies differ: IMAGO introduces internal regularization and optimizes the loss function to learn more robust embeddings, whereas DualNetGO employs a dual-network architecture to intelligently select features from multiple PPI network embeddings.
Figure 5.
Performance Comparison of IMAGO and DualNetGO on Human and Mouse Datasets across Gene Ontology Aspects.
A direct comparison of the human and mouse datasets(as these are common to both studies) reveals the respective strengths of each approach. For instance, on the Human BP dataset, IMAGO achieves an Fmax of 0.475, while DualNetGO reports an Fmax of 0.459 (as per their publication). This suggests that IMAGO’s regularization techniques effectively capture broader functional patterns. Conversely, on the Mouse MF dataset, DualNetGO might show superior performance, highlighting the advantage of its feature selection mechanism when dealing with specific functional categories where certain PPI evidence types are particularly informative. This comparative analysis underscores that the choice of an optimal model may be context-dependent, influenced by the target species and the specific GO aspect (BP, MF, CC). IMAGO’s strength lies in its generalized robustness against noise and overfitting, making it particularly suitable for scenarios with complex and noisy PPI data.
4. Discussion
The superior performance of IMAGO over CFAGO, particularly on human datasets, can be attributed to the synergistic effects of three core improvements: Regularization Counteracts Overfitting: The CFAGO model is prone to overfitting training data noise and specific details, limiting its generalization on test sets. IMAGO addresses this by incorporating a regularizer that constrains the distribution of hidden layer embeddings to approximate a normal distribution. This smoothing of learned feature representations reduces the model’s reliance on incidental patterns in training data, thereby enhancing robustness; Loss Function Optimization Ensures Feature Integrity: IMAGO introduces cosine similarity loss during pre-training, which ensures that the encoder-generated embedding vectors align directionally with the original input features, rather than merely achieving numerical reconstruction. This preserves more essential semantic information from the biological data, providing higher-quality features for downstream function prediction tasks; Component Synergy: The multi-head attention mechanism captures complex protein-protein interactions; the regularizer ensures these patterns are generalizable rather than specific; and the optimized loss function maintains fidelity in the encoding-decoding process. Together, these components enable IMAGO to learn more discriminative and generalizable protein representations, leading to improved performance in function prediction.
5. Conclusions
In this study, we propose IMAGO, a deep learning model leveraging attention mechanisms for protein function prediction. IMAGO integrates multi-source biological data within a single species through an encoder-regularizer-decoder architecture, where pre-training learns generalized protein representations and fine-tuning optimizes function prediction. Experimental validation demonstrates IMAGO’s superior performance on human and mouse datasets. Future work will focus on three critical directions revealed by cross-species validation: Developing spatiotemporal-aware attention mechanisms to model dynamic biological processes where current performance lags. Implementing adaptive regularization techniques that address species-specific data imbalances, particularly for long-tail GO term distributions. Establishing cross-species transfer learning frameworks to enhance prediction in less-studied organisms by leveraging conserved functional knowledge. These advancements will strengthen IMAGO’s capability to decipher disease mechanisms and accelerate drug target discovery across diverse species.
Author Contributions
Conceptualization, M.L. and T.Z.; methodology, L.L. and Q.W.; validation, Y.Z. and L.S.; formal analysis, M.L. and Z.W.; writing—original draft preparation, L.L. and Z.W.; writing—review and editing, Z.W.; visualization, Q.W. and Z.W. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by the National Natural Science Foundation of China (Grant No. 62302100); and by the National Natural Science Foundation of China (Grant No. 62473094) and the Natural Science Foundation of Heilongjiang Province (Grant No. LH2022F002).
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
The original data presented in the study are openly available in github at https://github.com/Lianglongchang/IMAGO (accessed on 20 January 2025).
Conflicts of Interest
The authors declare no conflicts of interest.
References
- Szklarczyk, D.; Kirsch, R.; Koutrouli, M.; Nastou, K.; Mehryary, F.; Hachilif, R.; Gable, A.L.; Fang, T.; Doncheva, N.T.; Pyysalo, S.; et al. The STRING database in 2023: Protein-protein association networks and functional enrichment analyses for a wide range of organisms. Nucleic Acids Res. 2023, 51, D638–D646. [Google Scholar] [CrossRef] [PubMed]
- Aleksander, S.; Balhoff, J.; Carbon, S.; Cherry, J.; Drabkin, H.; Ebert, D.; Feuermann, M.; Gaudet, P.; Harris, N. The Gene Ontology Knowledgebase in 2023. Genetics 2023, 224, iyad031. [Google Scholar] [CrossRef]
- UniProt Consortium. UniProt: The Universal Protein Knowledgebase. Nucleic Acids Res. 2023, 51, D523–D530. [Google Scholar] [CrossRef]
- Jones, D.T.; Thornton, J.M. The Impact of Bioinformatics on Biological Research. In Current Opinion in Structural Biology; Elsevier: Amsterdam, The Netherlands, 2021; Volume 67, pp. 101–108. [Google Scholar]
- Lai, B.; Xu, J. Accurate Protein Function Prediction via Graph Attention Networks with Predicted Structure Information. Briefings Bioinf. 2022, 23, bbab502. [Google Scholar] [CrossRef]
- Chen, Z.; Luo, Q. DualNetGO: A Dual Network Model for Protein Function Prediction via Effective Feature Selection. Bioinformatics 2023, 39, btad123. [Google Scholar] [CrossRef]
- Cho, H.; Berger, B.; Peng, J. Compact Integration of Multi-Network Topology for Functional Analysis of Genes. Cell Syst. 2016, 3, 540–548. [Google Scholar] [CrossRef]
- Gligorijević, V.; Barot, M.; Bonneau, R. deepNF: Deep Network Fusion for Protein Function Prediction. Bioinformatics 2018, 34, 3873–3881. [Google Scholar] [CrossRef] [PubMed]
- Gomez, J.; Bonet, J.; Borràs, C.; Ferrer, A. Structure-Based Function Prediction Using Structural Alignments and Clustering. Struct. Biol. 2017, 27, 89–95. [Google Scholar]
- Gligorijević, V.; Renfrew, P.D.; Kosciolek, T.; Leman, J.K.; Berenberg, D.; Vatanen, T.; Chandler, C.; Taylor, B.C.; Fisk, I.M.; Vlamakis, H.; et al. Structure-Based Function Prediction Using Graph Convolutional Networks. Nat. Commun. 2021, 12, 3168. [Google Scholar] [CrossRef] [PubMed]
- Wang, S.; You, R.; Liu, Y.; Xiong, Y.; Zhu, S. Graph Neural Networks for Protein Function Prediction. Nat. Commun. 2020, 11, 1656. [Google Scholar]
- Chen, B.; Cheng, X.; Li, P.; Geng, Y.; Gong, J.; Li, S.; Bei, Z.; Tan, X.; Wang, B.; Zeng, X.; et al. xTrimoPGLM: Unified 100-Billion-Parameter Pretrained Transformer for Deciphering the Language of Protein. Nat. Methods 2024, 22, 1028–1039. [Google Scholar] [CrossRef]
- Blakeley, K.J.; Harbison, C.T.; Parker, C.T.; Cline, M.S.; Smith, J.R. Homology-Based Methods for Protein Function Prediction. Bioinform. J. 2015, 31, 2345–2352. [Google Scholar]
- Deng, J.; Guo, J.; Wang, Z.; Liu, Y.; Chen, J.; Hu, X. Sequence Encoding via LSTM Networks for Protein Function Prediction. J. Comput. Biol. 2018, 25, 215–222. [Google Scholar]
- Meng, L.; Wang, X. TAWFN: A Deep Learning Framework for Protein Function Prediction. Bioinformatics 2024, 40, btae571. [Google Scholar] [CrossRef]
- Abramson, J.; Adler, J.; Dunger, J.; Evans, R.; Green, T.; Pritzel, A.; Ronneberger, O.; Willmore, L.; Ballard, A.J.; Bambrick, J.; et al. Accurate Structure Prediction of Biomolecular Interactions with AlphaFold 3. Nature 2024, 630, 493–500. [Google Scholar] [CrossRef]
- Jumper, J.; Evans, R.; Pritzel, A.; Green, T.; Figurnov, M.; Ronneberger, O.; Tunyasuvunakool, K.; Bates, R.; Žídek, A.; Potapenko, A.; et al. Highly Accurate Protein Structure Prediction with AlphaFold. Nature 2021, 596, 583–589. [Google Scholar] [CrossRef] [PubMed]
- Liu, M.; Wang, S.; Luo, Z. Multistage Attention-Based Extraction and Fusion of Protein Sequence and Structural Features for Protein Function Prediction. Bioinformatics 2025, 41, btaf374. [Google Scholar] [CrossRef]
- Zhang, Y.; Wang, Y.; Liu, X.; Chen, L. Applying CNN for Protein Structure Function Prediction. Deep Learn. Bioinf. 2019, 12, 67–75. [Google Scholar]
- Gu, Z.; Luo, X.; Chen, J. Hierarchical Graph Transformer with Contrastive Learning for Protein Function Prediction. Bioinformatics 2023, 39, btad410. [Google Scholar] [CrossRef]
- Mostafavi, S.; Ray, D.; Warde-Farley, D.; Grouios, C.; Morris, Q. GeneMANIA: A Real-Time Multiple Association Network Integration Algorithm for Predicting Gene Function. Genome Biol. 2008, 9 (Suppl. S1), S4. [Google Scholar] [CrossRef]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. In Advances in Neural Information Processing Systems 30; NIPS: Long Beach, CA, USA, 2017. [Google Scholar]
- Ba, J.L.; Kiros, J.R.; Hinton, G.E. Layer Normalization. arXiv 2016, arXiv:1607.06450. [Google Scholar] [CrossRef]
- Wu, Z.; Zhang, S.; Huang, X.; Li, Y. Integrative Approaches for Protein Function Prediction. In Multimodal Data Fusion in Bioinformatics; Springer: Berlin/Heidelberg, Germany, 2021; pp. 254–263. [Google Scholar]
- Fan, K.; Guan, Y.; Zhang, Y. Graph2GO: A Multi-Modal Attributed Network Embedding Method for Inferring Protein Functions. GigaScience 2020, 9, giaa081. [Google Scholar] [CrossRef] [PubMed]
- Wu, Z.; Guo, M.; Jin, X.; Chen, J.; Liu, B. CFAGO: Cross-Fusion of Network and Attributes Based on Attention Mechanism for Protein Function Prediction. Bioinformatics 2023, 39, btad123. [Google Scholar] [CrossRef] [PubMed]
- Ridnik, T.; Ben-Baruch, E.; Zamir, N.; Noy, A.; Friedman, I.; Protter, M.; Zelnik-Manor, L. Asymmetric Loss for Multi-Label Classification. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 11–17 October 2021; pp. 82–91. [Google Scholar]
- Wang, B.; Geng, Y.; Cheng, X. ProtGO: Universal Protein Function Prediction Utilizing Multi-Modal Gene Ontology Knowledge. Bioinformatics 2025, 41, btaf390. [Google Scholar] [CrossRef]
- Wang, W.; Shuai, Y.; Zeng, M. DPFunc: Accurately Predicting Protein Function via Deep Learning with Domain-Guided Structure Information. Nat. Commun. 2025, 16, 70. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).