Abstract
Emergent Large Language Models (LLMs) show impressive capabilities in performing a wide range of tasks. These models can be harnessed for biophysical use as well. The main challenge in this endeavor lies in transforming 3D chemical data into 1D language-like data. We developed a method to transform molecular data into language-like data and tokenize it for LLM use in a biophysical context. We then trained a model and validated it with a known protein–ligand complex. Using the pre-trained result, the model can assess the chemical properties of targets, detect shared binding properties and structures, and reveal related drugs. The model and the synthetic language to describe binding interactions uncovered novel protein–protein networks influenced by ligands, indicating functionally related yet previously unreported interactions.
1. Introduction
The investigation into forecasting binding interactions between ligands and proteins has been instrumental in revealing novel therapeutic target proteins and identifying potential pharmaceutical agents []. To address these complexities inherent to chemical binding interactions, the use of artificial intelligence (AI) combined with three-dimensional (3D) virtual screening (VS) has become widespread and shows promising potential as a viable approach []. Biophysical data depict spatially and temporally complex systems with intricate interactions, thus requiring specialized approaches, data structures, and computational methods such as deep learning for processing and analysis.
A central challenge in the computational elucidation of protein–ligand interactions lies in developing computationally tractable representations for molecular structures. Advanced and recent AI models, such as AlphaFold3 [] and Boltz-2 [], exemplify a new paradigm by training directly on 3D raw atomic coordinates. In contrast, models based on the AlphaFold2 architecture, like OpenFold [], are trained primarily on 1D protein sequences and 2D multiple-sequence alignments (MSAs) to infer structural information. This architectural distinction is critical: OpenFold focuses on structural prediction for proteins by determining residue-specific reference frames and torsion angles. This protein-centric representation is well-suited for predicting protein–protein interactions but is fundamentally ill-equipped to model the diverse chemical space of small-molecule ligands. Conversely, AlphaFold3 and Boltz-2 overcome this limitation. AlphaFold3 employs a diffusion module to generatively predict the 3D atomic coordinates for the entire biomolecular complex simultaneously. Boltz-2 utilizes a similar denoising module to generate the 3D structure but further extends this capability by incorporating a dedicated affinity module, which explicitly calculates binding strength to provide a quantitative measure of interaction potency. While these models have achieved formidable success, significant performance gaps remain in capturing accurate physical interaction behaviors [,]. This limitation may be a direct consequence of their training paradigm: the models are primarily supervised on static, end-state structures rather than the binding interactions themselves, potentially because the dynamic, biophysical process of binding is not yet well-represented in a computationally trainable format.
A few techniques have been developed to address this challenge, including transforming SMILES (the simplified molecular input line entry system) [] format into 2D or 3D data structures []. More recently, AI language models are used to better dissect these string-based data []. Large Language Models (LLMs), an emerging deep-learning method, show impressive capabilities in performing a multitude of tasks of different natures [,,], promising to overcome this challenge. Many of these tasks focus on common language uses such as summarization, editing, and translation []. These models can also be harnessed to perform less conventional tasks, including analyzing molecular structures and chemical features, especially in the context of protein–ligand interactions, protein–protein interactions, biophysical network analysis, protein design, and drug discovery [,,].
A critical step in applying LLMs to biophysical data is transforming complex, 3D spatial data into language-like, 1D string-based data that can be further processed by LLMs. The challenge in this transformation lies in preserving information beyond atom identity, chemical bonds, and electrochemical charges. Deep learning models such as 3D-convolutional neural networks (3D-CNNs) and spatial graph convolutional neural networks (SG-CNNs) [] can use graph networks to represent compounds; however, this representation does not necessarily encompass information related to electrochemical interactions or medium-dependent phenomena [,,].
We introduce a novel approach for integrating 3D binding interactions data into LLMs, addressing a central challenge in transforming biological data into a “machine-readable” format, especially by LLMs. Our approach defines a new synthetic, “machine-readable” language that can represent molecular, 3D spatial data in a fashion that encapsulates biochemical features. We trained a Generative Pre-trained Transformer (GPT) model, a model originally developed by OpenAI [], to analyze and predict text-converted 3D binding simulation data. The model, named “3D binding mode GPT” (3bmGPT), demonstrates that our method can effectively represent biochemical data in a format amenable to analysis by LLMs, both by reproducing previously validated results and by uncovering new features to known biophysical interactions.
2. Materials and Methods
2.1. Text Conversion for 3D Binding Prediction
We developed a synthetic language based on the Dictionary of Secondary Structure in Proteins (DSSP) algorithm [] to represent binding interactions data. The interaction of binding residues in the binding pocket has a unique fragment composed of three atoms, and one specific atom of each fragment was defined as a donor or acceptor (Figure 1a). From these atoms, a bond forms between an atom in a ligand and a target protein. This bond formation constitutes a “binding interaction word.” Subsequently, each discrete interaction, whether it is a simple contact or one of several interactions originating from a single atom, is individually represented as a word. The complete aggregation of all such words then constitutes the comprehensive binding interaction, forming a “binding interaction sentence” (Figure 1b). Because each word depicts a single protein residue and ligand state at a specific time point (i.e., a specific ligand in a specific spatial orientation), changes in this state, such as shifts in the spatial orientation that occur over time, will yield different words and new binding interaction sentences.
Figure 1.
Overview of the text conversion, dataset generation pipeline, and model application framework. (a) Detailed format of text conversion from binding mode and sample case. The upper part (Text Format of Binding Mode) shows the structure of a word representing a single binding interaction. The lower part (Text Conversion Case) depicts the conversion of the binding part in a ligand and the binding amino acid in a receptor. The table from the text conversion case illustrates how binding properties change by given notation, resulting in a binding interaction word. (b) Illustration of dataset generation for binding interaction sentences. The first step simulates binding interactions between a small molecule (blue) and amino acids (yellow). The second step visualizes their conversion into binding interaction words (ligand in purple, receptor in yellow). The final step illustrates their transformation into sentences, generating approximately 60 million sentences from multiple simulations. Ellipses denote continuous words and sentences. (c) Flowchart detailing the model, analysis steps, and input/output data. Square boxes indicate input/output data, either initially introduced or generated by a process. Round-edged square boxes represent a process, such as a programmed model or analysis step, that takes input data and generates output data. Lines with arrows illustrate the flow of data and the sequence of operations.
To encapsulate these biochemical dynamics, we collected data from the 3D interaction binding dataset, CrossDocked2020 [], which has 18,450 protein–ligand complexes with 22.6 million interaction poses specified by Pocketome [], as input for our method. We then simulated data using a CNN affinity algorithm [] to find more interaction poses, generating 120 additional poses per complex to account for binding flexibility, such as side chain variations. The interaction of binding residues in the pocket had a unique fragment composed of three atoms, and a specific atom of each fragment was defined as a donor and an acceptor. The ligand also contained donors and acceptors according to the same principle. As a rule, the donor of a residue interacts with the acceptor of a ligand, and the acceptor of a residue interacts with the donor of a ligand. Therefore, if the combination of the donor–acceptor of a residue–ligand pair was defined as a word, and all other ligand fragments interacting with a specific residue were defined as sentences. These sentences could be created with slightly different characteristics, either when new ligands were used as input or when new sentences were created by each motion of an existing ligand. This conversion from 3D information to 1D, text format, generated approximately 60 million trainable paragraph data by the language AI model.
We trained an LLM on the text-converted data to comprehensively analyze the nature of binding interactions and to evaluate the potential of this approach for studying protein–ligand interactions. A schematic overview of the model’s architecture and its applications is illustrated in Figure 1c, which delineates the input/output data formats and the analysis processes for different objectives. The detailed methodology is described in the following sub-sections, and the corresponding results and their interpretation are discussed in Section 3 (Results).
2.2. Validation of Text Conversion Using Tight-Binding Designed Data
To simplify molecular dynamics and ensure accurate predictions, we used 463 simulated binding datasets with tight-binding ligands for 16 targets: BACE, CDK2, CDK8, CMET, EG5, HIF2A, JNK1, MCL1, P38, PFKFB3, PTP1B, SHP2, SYK, THROMBIN, TNKS2, and TYK2 []. Ligands were extracted from the Protein Data Bank (PDB) [] data to SMILES sequences; then, similarities were computed using the Tanimoto coefficient (TC) []. TCs were calculated using the interaction fingerprint method []. Firstly, SMILES sequences were parsed to generate fingerprint format by the rcdk package version 3.8.1 []. The distance between fingerprints was calculated by the fingerprint package version 3.5.7. A heatmap was drawn by the NMF package version 0.27 []. Shared words were defined as those containing the same elements, including adjacent elements, and the percentage of shared words was calculated as the number of shared words divided by the total query word set. Statistical significance was assessed using a t-test in R (version 4.2.2).
2.3. Data Preparation with Tokenization
As a data filtration step, we excluded binding sentences with five or fewer words. This threshold was based on the observation that insignificant or non-specific interactions typically generate a correspondingly low number of binding words. Then, the protein–ligand dataset was partitioned into training, validation, and test sets in a 7:2:1 ratio. Each paragraph was a list of words represented in strings. We adopted two tokenization methods, and each method included special tokens [PAD], [SEP], and [UNK] in the vocabulary. First, we used word-based tokenization by treating each word as a token. All unique tokens in the entire dataset, plus the special tokens, constituted the vocabulary, which had a size of 143,018. Second, we trained a Byte-Pair Encoding (BPE) tokenizer on the dataset text, where all words within a paragraph were concatenated by a space character and paragraphs were separated by a newline character. The total vocabulary size was set to be 50,257. One potential benefit of BPE tokenization, which needs to be further examined, is the ability to keep the vocabulary size contained even when additional new words are added to the protein–ligand dataset. After the BPE tokenization, space and newline characters were taken out of the training tokens to avoid the excessive appearance of these two characters. For both tokenization schemes, tokens from the same paragraph constituted one sample. For all models trained, the maximum sequence length was set to 1071, equivalent to the number of tokens in the longest sample. Shorter samples were padded with the [PAD] token to the maximum sequence length. Training was performed with one epoch of the training tokens. To verify the reproducibility of our results, the whole process was performed an additional time.
2.4. Training Data for GPT and BERT LLMs
We used the protein–ligand token data to train GPT-2 []. Two GPT-2 backbone sizes were trained in PyTorch (version 2.2.1). GPT-2 small had a hidden size of 768, 12 hidden layers, and 12 attention heads and was trained with a batch size of 112, AdamW optimizer, a 6e-04 peak learning rate, 1600 warm-up steps, and a linear learning rate scheduler. GPT-2 medium had a hidden size of 1024, 24 hidden layers, and 16 attention heads and was trained with a batch size of 100, AdamW optimizer, and a 2e-04 peak learning rate. Then, we ran the studies on the Cerebras Wafer Scale Cluster with weight streaming technology [] and software release 1.8.0. Cerebras uses the Wafer Scale Engine (WSE), which has 850,000 cores with 40 GB of on-chip memory. Training times are shown in Table 1.
Table 1.
Training times by WSE.
The BERT (Bidirectional Encoder Representations from Transformers) model [] was trained on the same dataset using PyTorch (version 2.2.1). Similarly to the GPT model, it was configured with a hidden size of 768, 12 hidden layers, and 12 attention heads. Training was performed using three NVIDIA Tesla V100 GPUs with multithreaded processing.
2.5. Deriving Binding Words and Logits Using the Trained LLM
The trained model was archived as a set of files comprising model weights, configuration, and tokenizer data. As the model was trained on binding sentences described in Section 2.1, the input data must also be binding sentences generated from query PDB files. These query PDBs must contain a binding complex with a target protein and ligand to facilitate the analysis of further bindings or related binding properties. The PDBs were preprocessed for trimming and standardization, then subjected to binding detection and binding text generation. The resulting text can be directly applied to the loaded model to generate subsequent binding sentences or to calculate logits that quantify binding properties. To interpret the results, users can directly examine the generated binding sentences or compare the logit values against a reference dataset (e.g., the provided 10k dataset).
2.6. Validation of LLMs Using Practical Binding Data
To validate word generation and compare LLM performance, we used binding data from the drug Rebastinib and its target. ABL1. This target was chosen because its key bindings have been reported, and its binding is not highly unique, as it interacts with multiple proteins in the same kinase family []. This allowed us to assess various predictive capabilities of LLMs. From its PDB data, 30 binding interaction words were generated, and the first 10 words were input into the pre-trained GPT and BERT models to predict the next binding interaction word using the same parameters: temperature 0.7 and repetition penalty 2. This process was repeated 1000 times for each model, and the total results were analyzed for performance comparison.
Generated words were categorized into four distinct groups: (1) a word was labeled “query adjacent” if it exactly matched the next word in the input sequence (e.g., the 11th word in the Rebastinib case); (2) it was labeled “exists in original” if the word was not the correct adjacent word but was present elsewhere in the original words (e.g., the model predicts the 15th word or 3rd word instead of the 11th); (3) it was considered a “variation of existing words” if it shared binding elements with an original word but was not an exact match (e.g., the generated word is “1_6_C:N_C_C-E_O:C_N_CA,” while the 16th word is “1_4_C:N_C_C-E_O:C_N_CA”); and (4) it was classified as “completely new” if it did not fall into any of these categories. Prediction accuracy was assessed using logit values to calculate AUC scores, with separate measurements: word (1) was considered as the only true positive, and words (2) and (3) were combined as true positives. The first measurement indicated how accurately the model predicted the next word, while the second reflected its ability to generate related bindings.
We also generated words from binding simulation data between EGFR and ligand AEE788, a dual-targeting inhibitor for both EGFR and ERBB2 []. A total of 23 binding interaction words were generated from the simulation, and the first 7 words were input into the GPT model to generate 12 additional words with default parameters. These 12 words were then searched in the previously generated word dataset to identify shared words with other targets, and protein sequence similarities were calculated using BLAST [] (version 2.16).
2.7. Cluster Analysis
To compare prediction results for target proteins, we randomly selected 10,000 sentences and applied them to GPT model, which was pre-trained using PyTorch (version 2.2.1) based on the modified GPT-2 medium configuration (See Section 2.4). The average logit values for each token were calculated, producing a 10,000 × 50,257 matrix. Because the matrix was of high order, we applied feature selection and dimensionality reduction techniques using the Seurat package version 4.2.1 [] in R version 4.2.2. High standard deviation token vectors were selected to reduce the number of parameters to 737, followed by principal component analysis (PCA) for further dimension reduction. Utilizing 100 principal components, UMAP analysis with Louvain clustering was conducted to group target proteins with similar characteristics; then, we identified token vectors with distinct values for each cluster, ordered by the false discovery rate (FDR) value. We repeated our analysis 3 times (hereby noted as “trials”), all yielding consistent results, thus ascertaining the analysis reproducibility (Supplementary Figure S4).
2.8. Analysis for Biological Function
Each sentence’s data was annotated based on its target protein, with each data point representing a distinct target. Data points associated with the same target protein were color-coded on the UMAP plot to visualize the distribution of these target proteins. Subsequently, we examined the proportions of clusters corresponding to each target protein. To discern the biological functions, we converted the names of target proteins into gene names using UniProt ID mapping []. Genes related to tyrosine kinase (TK) and serine/threonine kinase (STK) were retrieved from Gene Ontology [] for further analysis of kinase-related binding interactions. Cluster characteristics were determined by applying gene members within each cluster to GO enrichment analysis for biological processes []. Function terms that appeared in more than 10 clusters were filtered out. From the remaining terms, we selected the function term with the lowest FDR value within a cluster to represent the biological function of that cluster. Additionally, we computed the average vector values for each cluster and used hierarchical clustering to compare these values. This hierarchical ordering allowed us to group the clusters into six categories, revealing common biological function terms identified through GO analysis that described the larger cluster’s characteristics. Furthermore, we identified genes related to insulin, sex hormones, and vitamin receptors from Gene Ontology to investigate the distribution of their binding interaction sentences.
2.9. Analysis from Applied Interaction Data for Biological Function and Ligand Similarity
First, we used the model to generate binding interaction words and logit values from the PDB-formatted data. These newly generated logit values were then combined with the 10,000 pre-analyzed reference data to generate new UMAP coordinates. Subsequently, we marked the location of the newly introduced data on the UMAP plot and generated a radar plot depicting the normalized distances between the new data’s location and the centroids of each cluster, annotated by biological function. To quantify structural similarity, Tanimoto coefficients (TCs) were calculated between the ligand extracted from the input PDB complex and each of the 10,000 ligands derived from the reference datasets, yielding 10,000 corresponding similarity scores. To visualize the TC trend across the UMAP space, the coordinates were first partitioned into 100 segments (bins). The average TC within each bin was then calculated and rendered as a line plot, revealing the relationship between UMAP proximity and structural similarity. Finally, the 20 nearest interaction data points to the input data were selected based on UMAP coordinates for network analysis. For the network analysis, connections were drawn exclusively for the 10 pairs with the highest TC values among all combinations. The ligands from the 20 nearest interactions were queried against the drugbank database [], which identified several known drugs with potential for repurposing. Subsequently, the structures of these known drugs were compared with that of the input ligand to identify shared molecular features.
2.10. Benchmarking Analysis
The target identification performance of 3bmGPT was benchmarked against ProBiS [] and Foldseek [] using a test set of 23 randomly selected PDBs. For ProBis and Foldseek, predictions were generated via their web servers. For 3bmGPT, PDBs were converted to text to generate logit values. The target was then assigned based on the most common protein among the 100 nearest neighbors, as determined by Euclidean distance in the logit space. To further assess the model’s capacity to discriminate between structurally similar targets, we used a set of 9 kinases from the KLIFS [] dataset. The 3bmGPT logit distance, ProBiS E-value, and Foldseek Z-score were calculated for this set and used to determine the AUC, allowing for a direct comparison of their discriminative power.
3. Results
3.1. Synthetic Language to Represent 3D Protein–Ligand Binding Data and Evaluation
Validation of the text conversion methodology was performed using a curated benchmark set of 463 datasets []. These datasets were intentionally selected because they represent static and high-affinity interactions, providing a clear test of our model’s ability to capture fundamental binding properties. Although the ligands within this set were structurally distinct—reflecting different molecular design approaches for the same target—the textual representations generated by our model showed remarkable convergence. As a result, the shared binding words were distinctive enough to allow for accurate clustering of these datasets based on their target proteins (Figure 2a). Moreover, ligand similarity showed a moderate linear correlation with the percentage of shared words (R2: 0.6937), and all structurally distant ligand groups exhibited statistically significant differences (Figure 2a). To further explore the potential of binding interaction words, we applied this binding data to LLMs to gain deeper insight into binding dynamics.
Figure 2.
Validation of text conversion and large language model performance. (a) Heatmap of 463 tightly designed bindings across 16 target proteins. Rows represent individual binding samples, ordered by hierarchical clustering and annotated with colors corresponding to each target protein. Columns are ordered by hierarchical clustering of textualized binding interactions. Distinct clusters are separated by lines. (b) Relationship between ligand similarity and percentage of shared binding interaction words. Left: The dot plot shows the correlation between ligand similarity (x-axis, Tanimoto coefficient) and the percentage of shared binding interaction words (y-axis). A linear regression line is shown in red, with R2 and p-value indicated. Right: Box plots display the distribution of shared word percentages across quartile groups; 1st to 4th represent each quartile range. Statistical significance between groups was measured by a t-test: *** p ≤ 0.001. (c) Example of key binding capture (Rebastinib). The red dashed lines indicate bonds between the ligand and protein atoms. (d) Comparison of GPT and BERT based on generated word proportions and area under the curve (AUC). Left: Bar plots show the proportion of words generated by each model. Right: AUC plots for adjacent and related word prediction, with GPT in red and BERT in blue.
To test text-converted bindings and select a suitable LLM, we first converted binding data between the drug Rebastinib and its target protein ABL1, which involves four key bindings []. These key bindings were accurately detected and represented as binding interaction words, regardless of binding type (Figure 2c). We then applied partial words from the Rebastinib–ABL1 bindings to two LLMs, GPT and BERT, both trained on our 60 million binding interaction sentences, to assess prediction performance. With identical parameters, BERT and GPT generated similar proportions of adjacent words (20.2% and 19.2%, respectively), but BERT generated more unrelated new words (Figure 2d). When considering true predictions as adjacent words, both models showed similar performance based on AUC values. However, when true predictions were defined as words related to query words, GPT outperformed BERT with an AUC of 0.8486 versus 0.7073 (Figure 2d). Based on this result, we focused on developing the GPT model.
We then used the epidermal growth factor receptor (EGFR), a well-studied target protein renowned for its substantial implications in tumor growth and metastasis [], to evaluate the model’s application in word generation. From its simulation with ligand AEE788, we generated 12 words, and four of these words corresponded to interaction words observed in complexes of other protein targets: ABL1, VGFR2, EPHA2, and CDK2 (Supplementary Figure S1a). Notably, all these targets, except for CDK2, are well-known targets for tyrosine kinases (TKs), the same family to which EGFR also belongs [,]. CDK2, a well-established target for serine/threonine kinase (STK) [], was also identified as sharing residues with EGFR. Both TK and STK are known to possess similar structural traits []. Within our trained dataset, the target proteins associated with TK and STK exhibited no significant differences in sequences, whereas their distinction from other proteins demonstrated statistical significance (Supplementary Figure S1b). Thus, the GPT model aptly captured this structural resemblance, pointing to key binding interactions that can provide valuable insights into the relationship between EGFR and the other related target proteins.
3.2. The Model Preserves the Characteristics of Coexisting Binding Nature
In natural language processing (NLP) tasks, results can be classified to identify aspects like the emotional tone of a text. This is often achieved not by focusing on the single most likely next token, but by treating the entire logit distribution as a feature vector that represents the input’s overall properties. In our model, protein–ligand interactions were predicted based on binding interaction words, utilizing the logit values corresponding to each tokenized vector generated by the NLP model. This approach enabled the application of NLP clustering techniques to identify trends in binding natures. We clustered 10,000 randomly selected binding representations into 23 clusters (Figure 3a) and observed that binding complexes tended to aggregate based on their respective target proteins (Supplementary Figure S2). The 23 unsupervised clusters also grouped sentences targeting the same proteins into coherent clusters. For example, cluster 2 and cluster 5 contained approximately 88% and 85% of the sentences generated from mouse ACES and horse ADH1E, respectively (Supplementary Table S1). Similarly, sentences generated from proteins encoded by analogous genes were closely located, suggesting a higher likelihood of clustering together.
Figure 3.
Target protein characteristics unveiled via cluster analysis of token vectors. (a) UMAP analysis with 23 clusters. (b) Token vector values heatmap. Dark-blue color indicates a higher value of vector values, and light-green color indicates a lower value. (c) Dot plot for proportion and number of sentences in cluster 7, and density plot on UMAP coordinates for sentences targeting DNA polymerase. (d) Plotting sentences for serine/threonine kinase (STK) and tyrosine kinase (TK), and the proportion of sentences for the kinases. STK in red, TK in green.
Of note, binding interaction sentences generated from the DNA polymerase protein family demonstrated that shared structural or binding attributes transcended factors such as species and potential data biases in shaping probabilities. Within cluster 7, over 75% of sentences related to DNA polymerase genes—including DNA polymerase beta from humans (denoted as DPOLB HUMAN), DNA polymerase 1 from Thermus aquaticus (denoted as DPO1 THEAQ), and DNA polymerase IV from Sulfolobus tengchongensis (denoted as DPO4 SULSO)—were congregated (Figure 3c), and their corresponding vectors (881st, 1022nd, 1649th, 1964th, and 1611th) values were significantly higher compared to other clusters (Figure 3b). This suggested that these specific vectors could be pivotal in representing distinct binding properties unique to DNA polymerase, setting them apart from other target-binding complexes. Instead of originating from the same genes, the target proteins involved in the kinases, TK and STK, held particular significance due to their substantial implication in cancer therapy [] and the similarities in their binding interactions. Intriguingly, the involved sentences appeared to intermingle predominantly within clusters 0 and 3 (Figure 3d), suggesting that the binding interactions indeed possessed striking similarities, accurately captured by the 3bmGPT. However, while TK and STK were not isolated into distinct clusters, the separation of clusters 0 and 3 indicated the potential existence of finer-grained characteristics that could serve to distinguish between them (Figure 3b). This pattern of distinct vector values was similarly evident across the remaining clusters associated with each target protein, potentially serving to uncover cofactors and distinguishing elements essential for the binding interactions of diverse target proteins.
3.3. The 3bmGPT Model Clustered Binding Interactions from Associated Molecular Functions by Distinct Features
Diverse biological functions have been reported to exhibit a robust correlation with distinct types of proteins []. Particularly, ligand binding sites within protein complexes have a substantial impact on their functions []. Given these factors, our cluster results also revealed distinct biological and molecular functions within each trained dataset. Employing Gene Ontology annotations [], we identified common molecular functions within each cluster group, resulting in six functionally similar categories, including nuclear receptor, protein kinase, protease, phospholipase, catalysis, and transferase functions (Figure 4a). Importantly, the generated functional categories were limited by the genes found in the ligand–receptor binding data, which seldom encompassed genes associated with tissue construction and bodily mobility. As a result, the categories exclusively revolved around chemical activities. Nuclear receptors are extensively investigated transcription factors owing to their noteworthy connection with metabolic disorders. Proteins falling under this category can be triggered by the binding of hormones, including sex hormones and insulin, as well as bile acids that play a role in the digestion and absorption of fat-soluble vitamins [].
Figure 4.
Functional group for clusters and multifunctional target proteins. (a) Hierarchically ordered clusters with biological functions. Cluster numbers are the same as in Figure 1a. (b) Percentage of binding interaction sentences related to specific functions within the nuclear receptor clusters: vitamin (yellow), sex hormone (red), and insulin (blue). Cluster order is hierarchical. (c) Target proteins across functional groups ordered by the nearest functional group in hierarchical clustering.
Based on our findings, the cluster designated as nuclear receptor displayed six distinct subgroups, each containing a collection of genes associated with abundant target proteins (Figure 4a). The gene that appeared most frequently across various subgroups was the albumin gene (ALB), accounting for 108 out of 2159 sentences within the nuclear receptor group. Unlike albumin, sentences originating from different sub-functions of the nuclear receptor were observed to coalesce into distinct subgroups. In the trained dataset, the gene for the vitamin D receptor (VDR) was the sole target associated with fat-soluble vitamins, and its corresponding sentences predominantly clustered in cluster 18. Similarly, genes linked to sex hormones, such as androgen (AR), estrogen (ESR), and progestogen (PGR), exhibited a concentration of sentences in cluster 6 (Figure 4b). Importantly, sentences directed towards the peroxisome proliferator-activated receptor (PPAR), strongly linked to adipogenesis, were prominently located in cluster 8. Sentences originating from retinoid X receptor alpha (RXRA) binding as a heterodimer of PPAR [] were situated in cluster 8 as well (Supplementary Figure S3), suggesting this cluster contains insulin-related binding information (Figure 4b). These results were reproducible using independent, randomly selected data (Supplementary Figure S4).
The model captured characteristics from other functional groups as well. Among the five clusters within the protease group, cluster 19, an independent cluster, emerged with 72.86% of its sentences relating to the human immunodeficiency virus type 1. Of note, HIV protease binding interactions are critical to AIDS antiviral therapy []. Interestingly, this cluster featured sentences generated from the renin hormone (REN), following those related to HIV. Given that renin is a dual target for HIV [], its placement within the cluster further demonstrated the model’s capability to capture biological features.
These observations demonstrate that sentences generated for closely related target proteins are closely located within clusters, suggesting that the sentences generated for multifunctional proteins [] would be distributed across multiple clusters. We therefore examined sentence distribution across clusters to test this hypothesis. In total, 16 target proteins demonstrated multifunctionality order (Figure 4c), exhibiting instances of singular functionality in cases like RXRA and kinesin-like protein KIF11, as well as spanning across different functional groups, such as PPARD and nitric oxide synthase 3 (NOS3). The presence of PPAR in the protein kinase group stands in line with prior studies []. Similar findings included fibroblast growth factor receptor 1 (FGFR1), a TK family member commonly reported to resist tyrosine kinase inhibitors (TKIs) []. In our findings, 68.42% of binding interaction sentences from FGFR1 were positioned within the protein kinase group, 26.32% within the nuclear receptor group, and the remaining 5.26% in other groups. Similarly, the renin hormone (RENI) exhibited multifunctional patterns between protein kinases and proteases, and coactivator-associated arginine methyltransferase 1 (CARM1) showed interactions between proteases and phospholipase. However, the catalysis and transferase groups did not showcase distinct multifunctional target proteins.
3.4. The Method Highlighted Key Structural Elements Crucial for Drug Discovery
To evaluate the 3bmGPT model’s capability in analyzing new data, we introduced new interaction data from the Protein Data Bank (PDB) [] into the model. We then conducted an in-depth analysis of its binding characteristics, focusing particularly on its ligand component. We utilized the PDB data structure with the ID 5EDQ, representing EGFR kinase, along with its associated ligand (ID 5N3) [] (Figure 5a). We deliberately excluded the annotations for the PDB, such as information about the target protein and ligand, to assess the model’s performance with unknown data. To elucidate the input’s underlying characteristics, the data were contextualized against the pre-annotated 10,000-sample dataset, which contains biological clusters, ligand information, and known drug associations, to elucidate the input’s latent characteristics. The biological property of the target protein (EGFR) from the input data was accurately determined as a protein kinase through mapping onto a previously annotated UMAP and assessing the prominence of its functional group in a radar plot (Figure 5b). While the data point positioned itself within the protein kinase cluster rather than any other clusters, the radar plot’s depiction of relative weights for each group further indicated that the binding interaction from the input data was closer to the protease cluster than the nuclear receptor cluster. These results are valuable for assessing multifunctionality, especially when the input interaction cannot be unequivocally categorized into a single function.
Figure 5.
Results of applying 3bmGPT onto new EGFR interaction data. (a) Input interaction data and the 3bmGPT analysis process of the input data. (b) Biological function analysis. UMAP shows 6 annotated biological functions with approximate areas. A red diamond indicates the location of newly introduced data. The radar chart shows the weights of biological functions calculated for the input data. (c) UMAP with binned Tanimoto coefficient (TC) values between the ligand of the input data and the other ligands. Each square with a scaled color indicates the average values of TC values from the included data. Blue color indicates a lower value of TC, while red color indicates higher values. The red circle pinpoints the location of the input data. The bottom line plot shows the TC value pattern according to horizontal UMAP coordinates (UMAP_1). The right line plot shows the TC value pattern according to vertical UMAP coordinates (UMAP_2). The red dashed lines indicate the corresponding position of the input data point. (d) Network for ligand similarity with the top 20 closely located data and common structures for connected ligands. In the network plot, circles indicate ligands labeled by PDB IDs. The red node indicates input data, and the line shows similar molecular structures between connected ligands. Common molecular substructures for two connected ligands to the input ligand are illustrated on the right.
Moreover, we computed the structural similarity between the ligand from the input PDB and the ligands within the reference datasets. A positive correlation was observed between this structural similarity and proximity in our model’s UMAP space (Figure 5c), suggesting our AI model successfully captures key ligand features. More importantly, the analysis revealed that the interaction-induced features provide a distinct and more holistic grouping compared to clustering based on ligand structure alone. For instance, the protein kinase group, which includes the input EGFR data, was positioned in proximity to the protease and nuclear receptor groups in our interaction space. In contrast, a purely ligand structure-based result showed the kinases closer to the phospholipase group. This distinction is a key finding: it demonstrates that our model’s numerical representation is not solely dependent on ligand structure. Instead, it successfully captures comprehensive binding interaction, underscoring that ligand structure alone is an insufficient descriptor to fully characterize these complex events.
As we discovered the algorithm’s potential to analyze ligands in terms of binding, we attempted to compare the ligands from other closely located binding data with the input PDB’s ligand (ID 5N3). For this, we selected 20 ligand data points closest to the input data. Among these 20 ligands, two (with IDs 0S9 and AQU) displayed molecular structures similar to the input ligand 5N3 (Figure 5d) and shared structural motifs. Notably, two small molecules in the drug database [], namely merestinib (ID L1X) and brigatinib (ID 6GY), were found in close proximity to the input data. These two drugs are significantly associated with cancer treatment; for instance, merestinib inhibits neurotrophic receptor kinase (NTRK) [], and brigatinib inhibits anaplastic lymphoma kinase []. Although the molecular structures of the two drugs differed from that of 5N3, their inhibition targets were closely related to tyrosine kinase, the same category as the target protein EGFR of the input data.
3.5. Benchmarking Highlights the Model’s Capacity for Sensitive Target Detection
We benchmarked our method against ProBiS [] and Foldseek [] to evaluate its performance in target protein identification. On a dataset of unlabeled binding data, all three methods successfully identified the correct targets with over 90% accuracy (Figure 6a). While overall detection was comparable, our model’s ability to sensitively distinguish between similar targets, quantified by AUC, was notably superior (Figure 6b). Unlike the other tools, which excel at single-target analysis, our model’s numerical approach to binding characteristics makes it particularly effective for quantitative and comparative assessments.
Figure 6.
Performance benchmark for target and similar protein identification. (a) Bar plot comparing the accuracy of the three tools. The horizontal axis represents the performance accuracy, and the vertical axis lists each of the compared tools. (b) Receiver operating characteristic (ROC) curves and corresponding AUC values for the discriminative power in similarity assessment.
4. Discussion
Transforming 3D chemical data into 1D language-like data has various applications, such as calculating binding affinity [] and designing novel molecules []. New data enriches the model’s “vocabulary” and strengthens its link to pre-trained datasets, improving residue identification and uncovering novel molecular relationships. In our results, key residues from EGFR-TK binding modes indicated possible essential residues for the kinase (Supplementary Figure S1). Adding new bindings to training data can refine the understanding of TK binding characteristics. Clustering quantified binding characteristics as vector values revealed biochemical characteristics consistent with experimental studies. Notably, STK and TK, known for similar ligand binding and biological functions [,], showed converging binding traits within the same cluster (Figure 3d), highlighting the LLM’s ability to capture binding attributes numerically. Further molecular insights from a protease cluster (Figure 4a) confirmed prior findings, such as darunavir and aliskiren’s potential dual therapeutic roles []. These insights from closely related proteins offer promise for identifying key residues and drug repurposing. The benchmark highlighted the model’s enhanced sensitivity for comparative target evaluation (Figure 6). Thus, protein associations, defined by binding characteristics, are quantifiable via the vector matrix, enabling numerical assessments of interaction and multifunctionality.
Research on ligand–protein complexes suggests the existence of distinct binding pockets for small molecules []. AI-driven analyses have since been used to detect and compare macromolecular indentations for pharmaceutical applications []. Our study used clustered characteristics to identify ligand interactions in DNA polymerases from different species (Figure 3c). We also identified a nearby cluster of prenyltransferases, grouping them as transferases (Figure 4a), suggesting shared binding pockets or dual-targeting ligands. Notably, farnesyltransferase, a type of prenyltransferase, is linked to DNA synthesis [,]. However, its relationship with DNA polymerases remains underexplored. Since DNA polymerase is a key target in drug development [], leading to agents like actinomycin D [], mithramycin A [], and zelpolib [], understanding its connection to transferases could improve drug discovery.
The use of measurable coordinates facilitated comparisons to identify biological functions proximate to the target protein, thereby unveiling opportunities to uncover potential multifunctionality. This exploration may provide insights into important scaffolds [] and off-target binding []. Drawing from this observation, we demonstrated the potential of this tool in pharmaceutical research (Figure 5d). Specifically, 3bmGPT efficiently positioned an input interaction in close proximity to pre-trained interactions, even with known drugs. This capability enables the tool to effectively search for related drugs, and if drug targets are incorporated into the model, it can be further applied to drug repurposing. As the model was trained with additional drugs and their targets, the representation of simplified binding units as words will be easily archived with minimized data size, connoting progress in the AI model for more accurate output generation.
Apparent limitations inherent to 3bmGPT are as follows: (1) a restricted number of trained target proteins; (2) the presence of uninformative vocabulary vectors stemming from tokenized units; (3) an insufficient number of comparative AI models; and (4) a lack of experimental validation, both for the training dataset and the model’s subsequent predictions. Regarding the first limitation, the training dataset, consisting of a modest 2907 proteins, is relatively limited when compared to the vast number of existing proteins, including their various isoforms. Consequently, substantial room for enhancement exists through the introduction of additional data, thereby enabling the model to scrutinize more intricate protein characteristics. The second limitation pertains to the presence of distinct vectors that fall short in capturing binding interactions comprehensively, as the text-converted binding interactions remain fragmented, aligning with challenges encountered by other LLMs []. To enable the reconstruction of tokenized binding modes, further refinement is imperative, potentially involving the development of attention maps to establish connections among the vectors. The third limitation applies to all AI developments as we strive to improve model fit. In this study, only GPT and BERT were compared, but other models could enhance performance. Additionally, variations in parameters, epochs, and tuning could yield different results. Lastly, the model was trained primarily on structural binding data, which lacks quantitative experimental metrics such as binding affinities derived from bioassays. This presents a key limitation, as the correlation between the model’s outputs and actual in vitro binding strength remains unconfirmed, leaving its practical application potential in question. Although 3bmGPT’s results show a favorable alignment with established findings, its capacity to discover novel ligands still necessitates rigorous prospective validation under laboratory conditions. Consequently, further research is essential to enhance and validate the model’s performance. Promising future direction includes (1) fine-tuning the model using feedback from a true/false binding classification task and (2) employing knowledge distillation to create smaller, specialized models tailored to the unique binding characteristics of specific protein families.
5. Conclusions
In this study, we developed a robust method and implemented it in a novel language model, 3bmGPT. Our results demonstrate that 3bmGPT can effectively predict key residues for proteins with similar biological functions (Figure 3) and successfully convert 3D binding modes into text to classify molecular characteristics (Figure 4). Furthermore, the model identified meaningful connections between biological functions via biochemical interactions (Figure 5). Its superior performance in sensitive target detection was validated through benchmarking (Figure 6). To facilitate further research and ensure reproducibility, the pre-trained model and all associated codes for training and analysis are openly available.
Supplementary Materials
The following supporting information can be downloaded at https://www.mdpi.com/article/10.3390/molecules30234516/s1, Figure S1: Binding interaction words generation for EGFR by 3bmGPT; Figure S2: Plotting target proteins on UMAP coordinates. Red dots represent interaction data points for each target protein; Figure S3: Density plot on UMAP coordinates for sentences from PPARG, PPARD, and RXRA; Figure S4: Reproduced result of density plot on UMAP coordinates; Table S1: Majority proportion of sentences per target protein in matched cluster.
Author Contributions
Conceptualization, T.L., H.J., A.J., J.M., J.H.H. and J.J.; Data curation, H.J.; Formal analysis, T.L.; Funding acquisition, J.J.; Investigation, T.L. and H.J.; Methodology, T.L., J.J., A.J., H.J., J.M. and B.C.Z.; Project administration, J.J. and B.C.Z.; Resources, H.J.; Software, A.J., T.L. and H.J.; Supervision, J.J.; Validation, T.L.; Visualization, T.L.; Writing—original draft, T.L.; Writing—review and editing, T.L. and J.J. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Data Availability Statement
The raw data utilized in this study were sourced from the Crossdocked2020 dataset [], which is publicly available at https://github.com/gnina/models (accessed on 1 April 2023). The pre-trained model and associated analysis data generated during this study are available on Zenodo (https://doi.org/10.5281/zenodo.17053576 (accessed on 4 September 2025)). The open-source code used for training and analysis can be accessed at https://github.com/LeeTS1001/3bmGPT (accessed on 1 September 2025).
Acknowledgments
We would like to thank the Cerebras team for providing access to their specialized hardware tailored for AI and deep-learning applications.
Conflicts of Interest
T.L., H.J., A.J., J.M., J.H.H., and J.J. have a financial interest in Syntekabio Incorporated, a biotechnology company focused on drug development using AI algorithms. J.J. also has a financial interest in Syntekabio USA Incorporated, a business subsidiary of Syntekabio Incorporated. B.C.Z. has a financial interest in Cerebras Systems, an AI company building computer systems for AI applications.
References
- Hughes, J.P.; Rees, S.S.; Kalindjian, S.B.; Philpott, K.L. Principles of early drug discovery. Br. J. Pharmacol. 2011, 162, 1239–1249. [Google Scholar] [CrossRef]
- Dhakal, A.; McKay, C.; Tanner, J.J.; Cheng, J. Artificial intelligence in the prediction of protein-ligand interactions: Recent advances and future directions. Brief. Bioinform. 2022, 23, bbab476. [Google Scholar] [CrossRef]
- Abramson, J.; Adler, J.; Dunger, J.; Evans, R.; Green, T.; Pritzel, A.; Ronneberger, O.; Willmore, L.; Ballard, A.J.; Bambrick, J.; et al. Accurate structure prediction of biomolecular interactions with AlphaFold 3. Nature 2024, 630, 493–500. [Google Scholar] [CrossRef]
- Passaro, S.; Corso, G.; Wohlwend, J.; Reveiz, M.; Thaler, S.; Somnath, V.R.; Getz, N.; Portnoi, T.; Roy, J.; Stark, H.; et al. Boltz-2: Towards Accurate and Efficient Binding Affinity Prediction. bioRxiv 2025. Available online: http://www.ncbi.nlm.nih.gov/pubmed/40667369 (accessed on 26 October 2025). [CrossRef] [PubMed]
- Ahdritz, G.; Bouatta, N.; Floristean, C.; Kadyan, S.; Xia, Q.; Gerecke, W.; O’donnell, T.J.; Berenberg, D.; Fisk, I.; Zanichelli, N.; et al. OpenFold: Retraining AlphaFold2 yields new insights into its learning mechanisms and capacity for generalization. Nat. Methods 2024, 21, 1514–1524. [Google Scholar] [CrossRef] [PubMed]
- Masters, M.R.; Mahmoud, A.H.; Lill, M.A. Investigating whether deep learning models for co-folding learn the physics of protein-ligand interactions. Nat. Commun. 2025, 16, 8854. [Google Scholar] [CrossRef]
- Ma, Y.; Chen, X. Deep-Learning vs Physics-Based Docking Tools for Future Coronavirus Pandemics. J. Chem. Inf. Model. 2025, 65, 12006–12013. [Google Scholar] [CrossRef] [PubMed]
- Weininger, D. SMILES, a Chemical Language and Information System 1: Introduction to Methodology and Encoding Rules. J. Chem. Inf. Comput. Sci. 1988, 28, 31–36. [Google Scholar] [CrossRef]
- Lee, K.; Kim, D. In-silico molecular binding prediction for human drug targets using deep neural multi-task learning. Genes 2019, 10, 906. [Google Scholar] [CrossRef]
- Flam-Shepherd, D.; Zhu, K.; Aspuru-Guzik, A. Language models can learn complex molecular distributions. Nat. Commun. 2022, 13, 3293. [Google Scholar] [CrossRef]
- Radford, A.; Wu, J.; Child, R.; Luan, D.; Amodei, D.; Sutskever, I. Language Models are Unsupervised Multitask Learners. Available online: https://github.com/codelucas/newspaper (accessed on 1 September 2023).
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. arXiv 2017. Available online: http://arxiv.org/abs/1706.03762 (accessed on 1 September 2023).
- Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv 2018. Available online: http://arxiv.org/abs/1810.04805 (accessed on 30 August 2023).
- Kaddour, J.; Harris, J.; Mozes, M.; Bradley, H.; Raileanu, R.; McHardy, R. Challenges and Applications of Large Language Models. arXiv 2023. Available online: http://arxiv.org/abs/2307.10169 (accessed on 1 June 2024).
- Merk, D.; Friedrich, L.; Grisoni, F.; Schneider, G. De Novo Design of Bioactive Small Molecules by Artificial Intelligence. Mol. Inform. 2018, 37, 1700153. [Google Scholar] [CrossRef]
- Bagal, V.; Aggarwal, R.; Vinod, P.K.; Priyakumar, U.D. MolGPT: Molecular Generation Using a Transformer-Decoder Model. J. Chem. Inf. Model. 2021, 62, 2064–2076. [Google Scholar] [CrossRef] [PubMed]
- Ferruz, N.; Schmidt, S.; Höcker, B. ProtGPT2 is a deep unsupervised language model for protein design. Nat. Commun. 2022, 13, 4348. [Google Scholar] [CrossRef] [PubMed]
- Jones, D.; Kim, H.; Zhang, X.; Zemla, A.; Stevenson, G.; Bennett, W.F.D.; Kirshner, D.; Wong, S.E.; Lightstone, F.C.; Allen, J.E. Improved Protein-Ligand Binding Affinity Prediction with Structure-Based Deep Fusion Inference. J. Chem. Inf. Model. 2021, 61, 1583–1592. [Google Scholar] [CrossRef] [PubMed]
- AEwing, T.J.; Makino, S.; Geoffrey Skillman, A.; Kuntz, I.D. DOCK 4.0: Search strategies for automated molecular docking of flexible molecule databases. J. Comput. Aided Mol. Des. 2001, 15, 411–428. [Google Scholar] [CrossRef]
- Velec, H.F.G.; Gohlke, H.; Klebe, G. DrugScoreCSD-knowledge-based scoring function derived from small molecule crystal data with superior recognition rate of near-native ligand poses and better affinity prediction. J. Med. Chem. 2005, 48, 6296–6303. [Google Scholar] [CrossRef]
- Wang, R.; Lai, L.; Wang, S. Further development and validation of empirical scoring functions for structure-based binding affinity prediction. J. Comput. Aided Mol. Des. 2002, 16, 11–26. [Google Scholar] [CrossRef]
- Kabsch, W.; Sander, C. Dictionary of protein secondary structure: Pattern recognition of hydrogen-bonded and geometrical features. Biopolymers 1983, 22, 2577–2637. [Google Scholar] [CrossRef]
- Francoeur, P.G.; Masuda, T.; Sunseri, J.; Jia, A.; Iovanisci, R.B.; Snyder, I.; Koes, D.R. Three-dimensional convolutional neural networks and a crossdocked data set for structure-based drug design. J. Chem. Inf. Model. 2020, 60, 4200–4215. [Google Scholar] [CrossRef]
- Kufareva, I.; Ilatovskiy, A.V.; Abagyan, R. Pocketome: An encyclopedia of small-molecule binding sites in 4D. Nucleic Acids Res. 2012, 40, D535–D540. [Google Scholar] [CrossRef]
- McNutt, A.T.; Francoeur, P.; Aggarwal, R.; Masuda, T.; Meli, R.; Ragoza, M.; Sunseri, J.; Koes, D.R. GNINA 1.0: Molecular docking with deep learning. J. Cheminform. 2021, 13, 43. [Google Scholar] [CrossRef]
- Wang, L.; Wu, Y.; Deng, Y.; Kim, B.; Pierce, L.; Krilov, G.; Lupyan, D.; Robinson, S.; Dahlgren, M.K.; Greenwood, J.; et al. Accurate and reliable prediction of relative ligand binding potency in prospective drug discovery by way of a modern free-energy calculation protocol and force field. J. Am. Chem. Soc. 2015, 137, 2695–2703. [Google Scholar] [CrossRef] [PubMed]
- Berman, H.M.; Westbrook, J.; Feng, Z.; Gilliland, G.; Bhat, T.N.; Weissig, H.; Shindyalov, I.N.; Bourne, P.E. The Protein Data Bank. Nucleic Acids Res. 2000, 28, 235–242. [Google Scholar] [CrossRef] [PubMed]
- Rácz, A.; Bajusz, D.; Héberger, K. Life beyond the Tanimoto coefficient: Similarity measures for interaction fingerprints. J. Cheminform. 2018, 10, 48. [Google Scholar] [CrossRef] [PubMed]
- Guha, R. Journal of Statistical Software Chemical Informatics Functionality in R. 2007. Available online: http://www.jstatsoft.org/ (accessed on 24 November 2023).
- Gaujoux, R.; Seoighe, C. A flexible R package for nonnegative matrix factorization. BMC Bioinform. 2010, 11, 367. [Google Scholar] [CrossRef]
- Lauterbach, G. The Path to Successful Wafer-Scale Integration: The Cerebras Story. IEEE Micro 2021, 41, 52–57. [Google Scholar] [CrossRef]
- Chan, W.W.; Wise, S.C.; Kaufman, M.D.; Ahn, Y.M.; Ensinger, C.L.; Haack, T.; Hood, M.M.; Jones, J.; Lord, J.W.; Lu, W.P.; et al. Conformational Control Inhibition of the BCR-ABL1 Tyrosine Kinase, Including the Gatekeeper T315I Mutant, by the Switch-Control Inhibitor DCC-2036. Cancer Cell 2011, 19, 556–568. [Google Scholar] [CrossRef]
- Traxler, P.; Allegrini, P.R.; Brandt, R.; Brueggen, J.; Cozens, R.; Fabbro, D.; Grosios, K.; Lane, H.A.; McSheehy, P.; Mestan, J. AEE788: A Dual Family Epidermal Growth Factor Receptor/ErbB2 and Vascular Endothelial Growth Factor Receptor Tyrosine Kinase Inhibitor with Antitumor and Antiangiogenic Activity. Cancer Res. 2004, 64, 4931–4941. [Google Scholar] [CrossRef]
- Altschul, S.F.; Gish, W.; Miller, W.; Myers, E.W.; Lipman, D.J. Basic local alignment search tool. J. Mol. Biol. 1990, 215, 403–410. [Google Scholar] [CrossRef]
- Hao, Y.; Hao, S.; Andersen-Nissen, E.; Mauck, W.M.; Zheng, S.; Butler, A.; Lee, M.J.; Wilk, A.J.; Darby, C.; Zager, M.; et al. Integrated analysis of multimodal single-cell data. Cell 2021, 184, 3573–3587.e29. [Google Scholar] [CrossRef]
- Bateman, A.; Martin, M.J.; Orchard, S.; Magrane, M.; Ahmad, S.; Alpi, E.; Bowler-Barnett, E.H.; Britto, R.; Bye-A-Jee, H.; Cukura, A.; et al. UniProt: The Universal Protein Knowledgebase in 2023. Nucleic Acids Res. 2023, 51, D523–D531. [Google Scholar]
- Aleksander, S.A.; Balhoff, J.; Carbon, S.; Cherry, J.M.; Drabkin, H.J.; Ebert, D.; Feuermann, M.; Gaudet, P.; Harris, N.L.; Hill, D.P. The Gene Ontology knowledgebase in 2023. Baryshnikova A, editor. Genetics 2023, 224, iyad031. [Google Scholar]
- Thomas, P.D.; Ebert, D.; Muruganujan, A.; Mushayahama, T.; Albou, L.P.; Mi, H. PANTHER: Making genome-scale phylogenetics accessible to all. Protein Sci. 2022, 31, 8–22. [Google Scholar] [CrossRef]
- Wishart, D.S.; Feunang, Y.D.; Guo, A.C.; Lo, E.J.; Marcu, A.; Grant, J.R.; Sajed, T.; Johnson, D.; Li, C.; Sayeeda, Z.; et al. DrugBank 5.0: A major update to the DrugBank database for 2018. Nucleic Acids Res. 2018, 46, D1074–D1082. [Google Scholar] [CrossRef] [PubMed]
- Konc, J.; Janežič, D. ProBiS-ligands: A web server for prediction of ligands by examination of protein binding sites. Nucleic Acids Res. 2014, 42, W215–W220. [Google Scholar] [CrossRef] [PubMed]
- van Kempen, M.; Kim, S.S.; Tumescheit, C.; Mirdita, M.; Lee, J.; Gilchrist, C.L.M.; Söding, J.; Steinegger, M. Fast and accurate protein structure search with Foldseek. Nat. Biotechnol. 2024, 42, 243–246. [Google Scholar] [CrossRef] [PubMed]
- Kanev, G.K.; de Graaf, C.; Westerman, B.A.; de Esch, I.J.P.; Kooistra, A.J. KLIFS: An overhaul after the first 5 years of supporting kinase research. Nucleic Acids Res. 2021, 49, D562–D569. [Google Scholar] [CrossRef]
- Uribe, M.L.; Marrocco, I.; Yarden, Y. Cancers EGFR in Cancer: Signaling Mechanisms, Drugs, and Acquired Resistance. Cancers 2021, 13, 2748. [Google Scholar] [CrossRef]
- Paul, M.D.; Grubb, H.N.; Hristova, K. Quantifying the strength of heterointeractions among receptor tyrosine kinases from different subfamilies: Implications for cell signaling. J. Biol. Chem. 2020, 295, 9917–9933. [Google Scholar] [CrossRef]
- Du, Z.; Lovly, C.M. Mechanisms of receptor tyrosine kinase activation in cancer. Mol. Cancer 2018, 17, 58. [Google Scholar] [CrossRef] [PubMed]
- Saturno, G.; Pesenti, M.; Cavazzoli, C.; Rossi, A.; Giusti, A.M.; Gierke, B.; Pawlak, M.; Venturi, M. Expression of serine/threonine protein-kinases and related factors in normal monkey and human retinas: The mechanistic understanding of a CDK2 inhibitor induced retinal toxicity. Toxicol. Pathol. 2007, 35, 972–983. [Google Scholar] [CrossRef] [PubMed]
- Taylor, S.S.; Radzio-Andzelm, E.; Hunter, T. How do protein kinases discriminate between serine/threonine and tyrosine? Structural insights from the insulin receptor protein-tyrosine kinase. FASEB J. 1995, 9, 1255–1266. [Google Scholar] [CrossRef] [PubMed]
- Bhullar, K.S.; Lagarón, N.O.; McGowan, E.M.; Parmar, I.; Jha, A.; Hubbard, B.P.; Rupasinghe, H.P.V. Kinase-targeted cancer therapies: Progress, challenges and future directions. Mol. Cancer 2018, 17, 48. [Google Scholar] [CrossRef]
- Bergendahl, L.T.; Marsh, J.A. Functional determinants of protein assembly into homomeric complexes. Sci. Rep. 2017, 7, 4932. [Google Scholar] [CrossRef]
- Abrusán, G.; Marsh, J.A. Ligand Binding Site Structure Influences the Evolution of Protein Complex Function and Topology. Cell Rep. 2018, 22, 3265–3276. [Google Scholar] [CrossRef]
- Tobin, J.F.; Freedman, L.P. Nuclear receptors as drug targets in metabolic diseases: New approaches to therapy. Trends Endocrinol. Metab. 2006, 17, 284–290. [Google Scholar] [CrossRef]
- Chandra, V.; Huang, P.; Hamuro, Y.; Raghuram, S.; Wang, Y.; Burris, T.P.; Rastinejad, F. Structure of the intact PPAR-γ-RXR-α nuclear receptor complex on DNA. Nature 2008, 456, 350–356. [Google Scholar] [CrossRef]
- Weber, I.T.; Agniswamy, J. HIV-1 Protease: Structural perspectives on drug resistance. Viruses 2009, 1, 1110–1136. [Google Scholar] [CrossRef]
- Tzoupis, H.; Leonis, G.; Megariotis, G.; Supuran, C.T.; Mavromoustakos, T.; Papadopoulos, M.G. Dual inhibitors for aspartic proteases HIV-1 PR and renin: Advancements in AIDS-hypertension-diabetes linkage via Molecular dynamics, inhibition assays, and binding free energy calculations. J. Med. Chem. 2012, 55, 5784–5796. [Google Scholar] [CrossRef]
- Espinosa-Cantú, A.; Cruz-Bonilla, E.; Noda-Garcia, L.; DeLuna, A. Multiple Forms of Multifunctional Proteins in Health and Disease. Front. Cell Dev. Biol. 2020, 8, 451. [Google Scholar] [CrossRef]
- Lazennec, G.; Canaple, L.; Saugy, D.; Wahli, W. Activation of Peroxisome Proliferator-Activated Receptors (PPARs) by Their Ligands and Protein Kinase A Activators. Mol. Endocrinol. 2000, 14, 1962–1975. Available online: https://academic.oup.com/mend/article/14/12/1962/2751066 (accessed on 28 August 2023). [CrossRef] [PubMed]
- Yue, S.; Li, Y.; Chen, X.; Wang, J.; Li, M.; Chen, Y.; Wu, D. FGFR-TKI resistance in cancer: Current status and perspectives. J. Hematol. Oncol. 2021, 14, 23. [Google Scholar] [CrossRef] [PubMed]
- Hanan, E.J.; Baumgardner, M.; Bryan, M.C.; Chen, Y.; Eigenbrot, C.; Fan, P.; Gu, X.-H.; La, H.; Malek, S.; Purkey, H.E.; et al. 4-Aminoindazolyl-dihydrofuro[3, 4-d]pyrimidines as non-covalent inhibitors of mutant epidermal growth factor receptor tyrosine kinase. Bioorganic Med. Chem. Lett. 2016, 26, 534–539. [Google Scholar] [CrossRef]
- Konicek, B.W.; Capen, A.R.; Credille, K.M.; Ebert, P.J.; Falcon, B.L.; Heady, G.L.; Patel, B.K.R.; Peek, V.L.; Stephens, J.R.; Stewart, J.A. Merestinib (LY2801653) inhibits neurotrophic receptor kinase (NTRK) and suppresses growth of NTRK fusion bearing tumors. Oncotarget 2018, 9, 13796–13806. [Google Scholar] [CrossRef]
- Huang, W.S.; Liu, S.; Zou, D.; Thomas, M.; Wang, Y.; Zhou, T.; Romero, J.; Kohlmann, A.; Li, F.; Qi, J.; et al. Discovery of Brigatinib (AP26113), a Phosphine Oxide-Containing, Potent, Orally Active Inhibitor of Anaplastic Lymphoma Kinase. J. Med. Chem. 2016, 59, 4948–4964. [Google Scholar] [CrossRef]
- Shi, L.; Potts, M.; Kennelly, P.J. The serine, threonine, and/or tyrosine-specific protein kinases and protein phosphatases of prokaryotic organisms: A family portrait. FEMS Microbiol. Rev. 1998, 22, 229–253. [Google Scholar] [CrossRef]
- Laskowski, R.A.; Luscombe, N.M.; Swindells, M.B.; Thornton, J.M.; Thornton, J.M. Protein clefts in molecular recognition and function. Protein Sci. 1996, 5, 2438–2452. [Google Scholar]
- Govindaraj, R.G.; Brylinski, M. Comparative assessment of strategies to identify similar ligand-binding pockets in proteins. BMC Bioinform. 2018, 19, 91. [Google Scholar] [CrossRef]
- Shydlovskyi, S.; Zienert, A.Y.; Ince, S.; Dovengerds, C.; Hohendahl, A.; Dargazanli, J.M.; Blum, A.; Günther, S.D.; Kladt, N.; Stürzl, M.; et al. Nucleotide-dependent farnesyl switch orchestrates polymerization and membrane binding of human guanylate-binding protein. Proc. Natl. Acad. Sci. USA 2017, 114, E5559–E5568. [Google Scholar] [CrossRef]
- Law, B.K.; Nørgaard, P.; Gnudi, L.; Kahn, B.B.; Poulson, H.S.; Moses, H.L. Inhibition of DNA synthesis by a farnesyltransferase inhibitor involves inhibition of the p70(s6k) pathway. J. Biol. Chem. 1999, 274, 4743–4748. [Google Scholar] [CrossRef] [PubMed]
- Berdis, A.J. DNA polymerases as therapeutic targets. Biochemistry 2008, 47, 8253–8260. [Google Scholar] [CrossRef] [PubMed]
- Liu, X.F.; Xiang, L.; Zhou, Q.; Carralot, J.P.; Prunotto, M.; Niederfellner, G.; Pastan, I. Actinomycin D enhances killing of cancer cells by immunotoxin RG7787 through activation of the extrinsic pathway of apoptosis. Proc. Natl. Acad. Sci. USA 2016, 113, 10666–10671. [Google Scholar] [CrossRef] [PubMed]
- Quarni, W.; Dutta, R.; Green, R.; Katiri, S.; Patel, B.; Mohapatra, S.S.; Mohapatra, S. Mithramycin A Inhibits Colorectal Cancer Growth by Targeting Cancer Stem Cells. Sci. Rep. 2019, 9, 15202. [Google Scholar] [CrossRef]
- Mishra, B.; Zhang, S.; Zhao, H.; Darzynkiewicz, Z.; Lee, E.Y.C.; Lee, M.Y.W.T.; Zhang, Z. Discovery of a novel DNA polymerase inhibitor and characterization of its antiproliferative properties. Cancer Biol. Ther. 2019, 20, 474–486. [Google Scholar] [CrossRef]
- Zheng, S.; Lei, Z.; Ai, H.; Chen, H.; Deng, D.; Yang, Y. Deep scaffold hopping with multimodal transformer neural networks. J. Cheminform. 2021, 13, 87. [Google Scholar] [CrossRef]
- Xie, L.; Xie, L.; Bourne, P.E. Structure-based systems biology for analyzing off-target binding. Curr. Opin. Struct. Biol. 2011, 21, 189–199. [Google Scholar] [CrossRef]
- Chang, Y.; Wang, X.; Wang, J.; Wu, Y.; Yang, L.; Zhu, K.; Chen, H.; Yi, X.; Wang, C.; Wang, Y. A Survey on Evaluation of Large Language Models. arXiv 2023. Available online: http://arxiv.org/abs/2307.03109 (accessed on 4 September 2023). [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).