Next Article in Journal
Topical Probiotics as a Novel Approach in the Treatment of Chronic Dermatoses Associated with Skin Dysbiosis: A Narrative Review
Previous Article in Journal
Palmitoylation Code and Endosomal Sorting Regulate ABHD17A Plasma Membrane Targeting and Activity
Previous Article in Special Issue
Computational Discovery of Selective Carbonic Anhydrase IX (CA IX) Inhibitors via Pharmacophore Modeling and Molecular Simulations for Cancer Therapy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Local Functional Structure Features to Improve Drug–Target Interaction Prediction

1
Center for Artificial Intelligence Driven Drug Discovery, Faculty of Applied Sciences, Macao Polytechnic University, Macao 999708, China
2
School of Physics and Technology, Nanjing Normal University, Nanjing 210023, China
3
State Key Laboratory of Food Nutrition and Safety, College of Food Science and Engineering, Tianjin University of Science and Technology, Tianjin 300457, China
*
Authors to whom correspondence should be addressed.
Int. J. Mol. Sci. 2025, 26(20), 10194; https://doi.org/10.3390/ijms262010194
Submission received: 29 August 2025 / Revised: 14 October 2025 / Accepted: 17 October 2025 / Published: 20 October 2025

Abstract

Molecular simulation is central to modern drug discovery but is often limited by high computational cost and the complexity of molecular interactions. Deep-learning drug–target interaction (DTI) prediction can accelerate screening; however, many models underuse the local functional structure features—binding motifs, reactive groups, and residue-level fragments—that drive recognition. We present LoF-DTI, a framework that explicitly represents and couples such local features. Drugs are converted from SMILES into molecular graphs and targets from sequences into feature representations. On the drug side, a Jumping Knowledge (JK) enhanced Graph Isomorphism Network (GIN) extracts atom- and neighborhood-level patterns; on the target side, residual CNN blocks with progressively enlarged receptive fields, augmented by N-mer substructural statistics, capture multi-scale local motifs. A Gated Cross-Attention (GCA) module then performs atom-to-residue interaction learning, highlighting decisive local pairs and providing token-level interpretability through attention scores. By prioritizing locality during both encoding and interaction, LoF-DTI delivers competitive results across multiple benchmarks and improves early retrieval relevant to virtual screening. Case analyses show that the model recovers known functional binding sites, suggesting strong potential to provide mechanism-aware guidance for molecular simulation and to streamline the drug design pipeline.

1. Introduction

Since its inception in the 1950s with foundational techniques like Monte Carlo sampling and molecular dynamics, molecular simulation technology (MST) has become an indispensable tool across scientific disciplines, including pharmacology, phytomedicine, and molecular physics [1,2,3,4]. In the realm of health promotion, MST is particularly valuable for exploring the interaction mechanisms between bioactive compounds and therapeutic protein targets, as well as interactions between bioactive compounds and their environments [5]. By providing atomic-level insights, it significantly reduces the labor and resource costs associated with traditional experimentation. However, the utility of MST for large-scale applications, such as screening vast compound libraries, remains hampered by substantial computational burdens, creating a critical bottleneck in the discovery pipeline [6].
In response to these computational demands, the rapidly evolving field of machine learning (ML) offers powerful approximation methods for complex atomic systems, effectively bypassing the need to solve computationally intractable equations [7]. First, traditional structure-based approaches, most notably molecular docking with physics/empirical/knowledge-based scoring functions explicitly model the 3D binding pose and approximate binding energies when reliable protein structures are available; these methods provide physical interpretability but depend on structural availability/quality and can be sensitive to receptor flexibility and scoring errors [8]. Second, data-driven learning approaches infer interactions from observed patterns. Among them, classic kernel-based baselines offer data efficiency when informative similarity matrices are available, albeit with reliance on handcrafted similarities and potential cold-start limitations [9]. While deep learning (DL)-based DTI techniques leverage known interaction data and molecular features to learn rich representations and rapidly estimate binding affinity between compounds and targets [10]. By providing an efficient means for preliminary screening and prioritization, these DTI models guide subsequent in-depth molecular simulation studies and experimental validation [11].
Recent advancements in DL have led to a suite of sophisticated DTI models that learn from heterogeneous data sources like chemical structures and protein sequences [12]. MolTrans [13] uses knowledge-guided substructure mining and an enhanced Transformer to model interactions. MCANet [14] adopts a cross-attention mechanism to enhance drug–target feature associations. DrugBAN [15] introduces a bilinear attention network with domain-adaptive learning to improve generalization. DefuseDTI [16] employs a dual-branch encoder and multi-view attention to extract fine-grained multimodal features. DMHGNN [17] builds a dual-view heterogeneous GNN to learn drug–protein representations via topological and semantic graphs. MHAN-DTA [18] proposes a multiscale hybrid attention network to enhance feature extraction capabilities, thereby improving the performance of affinity prediction. Multi-ITI [19] integrates a pre-trained biological feature learning module with a heterogeneous graph neural network equipped with dynamic graph attention, effectively capturing drug–target network relationships while mitigating the impact of noise.
However, despite their impressive performance, a critical limitation persists: most existing models do not explicitly enhance the representation of local functional structures of compounds or targets. They often focus on global representations or abstract interaction patterns, thereby overlooking the fine-grained, functionally relevant substructures, such as reactive groups, binding motifs, or local sequence/graph neighborhoods—that ultimately govern molecular recognition and binding [20]. The importance of focusing on local features is gaining attention. For example, Liu et al. [21] employed recurrent neural networks to capture intrinsic features and dependencies within sequences, and introduced incremental learning to improve the adaptability and accuracy of miRNA prediction. Gao et al. [22] incorporated pathway information and contrastive learning [23] into spatial transcriptomics studies, taking into account the local functional correlations between spatial points to better uncover underlying biological characteristics.
To address this gap and better support MST-driven discovery, we propose LoF-DTI, a deep learning model explicitly architected to strengthen local functional structure features. The model systematically decomposes inputs on both sides: drug SMILES are converted to molecular graphs and encoded by a Graph Isomorphism Network (GIN) with Jumping Knowledge to capture hierarchical, atom- and neighborhood-level patterns in drugs [24], while protein sequences are processed by a CNN-based residual [25] module. We further augment locality with N-mer substructural semantic features, which emphasize motif-scale signals in both molecules and sequences. The resulting drug and target representations are fused by a Gated Cross-Attention (GCA) module that uses multi-head attention to highlight interaction cues between key atoms and residues, and a gating mechanism to adaptively balance raw local evidence with global context. This design improves the identification of pharmacophores and yields high-confidence, interpretable structure–activity associations to guide downstream molecular simulation and experimental validation. In contrast to approaches that rely primarily on global similarity or coarse feature aggregation, our method explicitly prioritizes local functional structures during both representation learning and cross-modal interaction, enabling precise alignment of atom–residue signals. This design not only improves the identification of key pharmacophores but also provides high-confidence, interpretable structure-activity associations to guide subsequent molecular simulation experiments.

2. Results and Discussion

2.1. Performance Evaluation

To benchmark the predictive power of LoF-DTI, we conducted a comparative analysis against six state-of-the art baseline models—DeepConv-DTI [26], GraphDTA [27], MolTrans [13], DrugBAN [15], NFSA-DTI [28], and IHDFN-DTI [29] across four benchmark datasets. The experimental results are presented in Table 1, where the bolded values indicate the best performance and underlined values denote the second-best results.
As shown in Table 1, LoF-DTI consistently achieves the best overall performance across all four benchmark datasets, demonstrating its superior ability to model complex drug–target interactions. Across the BindingDB, BioSNAP, DAVIS, and Human datasets, LoF-DTI achieves the highest scores in most evaluation metrics, including AUROC, AUPRC, and Accuracy. Among them, the low AUPRC on the DAVIS dataset is primarily driven by evaluation and distributional factors that are specific to DAVIS, a low positive prevalence combined with a thresholding strategy that makes AUPRC highly sensitive. Notably, while IHDFN-DTI shows competitive results on some metrics, LoF-DTI achieves its state-of-the-art performance with a model that is four times smaller in parameter size. This underscores the exceptional computational efficiency of our proposed architecture. Collectively, these findings establish LoF-DTI as a robust, generalizable, and highly efficient framework that strikes an optimal balance between predictive accuracy and model complexity.

2.2. Ablation Studies

To dissect the contribution of each architectural component to the overall performance of LoF-DTI, we conducted a comprehensive series of ablation studies. The study systematically evaluated the impact of removing key components: the Jumping Knowledge (JK) mechanism, the residual connections (RS) in the protein encoder, and the entire Gated Cross-Attention (GCA) module. We also replaced the Graph Isomorphism Network (GIN) with a standard Graph Convolutional Network (GCN) to assess its specific contribution.
Figure 1 presents the results of ablation studies on different modules. In the figure, “_JK” denotes the removal of the JK mechanism from the full model; “_RS” indicates the removal of residual connections, replacing the structure with a basic CNN; “_GCA” represents the absence of the GCA module, where drug and target features are directly concatenated instead; and “GCN” refers to replacing GIN with a standard GCN [30]. During each ablation experiment, all other components remain consistent with the original LoF-DTI configuration. All experimental results presented in Figure 1 were obtained using consistent hyperparameter configurations.
Every tested component contributes positively to the model’s performance. The most substantial performance drop was observed upon removal of the GCA module, highlighting its critical role in learning effective drug–target interaction representations. Compared to simple concatenation or averaging strategies, GCA effectively highlights functional interaction regions between drugs and targets, thereby greatly enhancing performance in both classification and regression tasks. Furthermore, the DrugGIN and ProteinRS modules provide strong unimodal representations that serve as high-quality inputs to the GCA module, demonstrating the synergy between the specialized encoders and the interaction module.
Given the importance of the GCA module, we further investigated the impact of different gating strategies on the Human dataset. Figure 2 shows the AUROC and AUPRC values on the validation set over the first 60 training epochs.
The results demonstrate that the dynamic gating mechanism consistently outperforms other gating strategies, maintaining higher AUROC and AUPRC scores throughout the training process. It also exhibits better stability and faster convergence. This suggests that incorporating learnable dynamic weights during feature fusion enables the model to adaptively adjust the importance of drug and target information for each specific sample. Such adaptability helps the model capture critical interaction signals more accurately. Compared to static or averaged gating approaches, the dynamic mechanism more effectively distinguishes between redundant and informative features, leading to improved overall prediction performance. These findings further confirm the pivotal role of the GCA module in enhancing the model’s discriminative capability.
Further ablation studies, shown in Figure 3, explored the sensitivity of LoF-DTI to hyperparameters and the impact of N-mer features. While the number of attention heads had a modest effect on performance, the inclusion of 3-mer substructural features provided a consistent and significant boost across all metrics, particularly in sensitivity and accuracy. Notably, dynamically learned projections for these features were essential, as static random mappings offered minimal benefit, confirming the value of adaptive feature engineering in our model.

2.3. Case Studies

2.3.1. DTI Prediction as a Guide for Molecular Simulation

Traditional techniques such as molecular simulation are often limited in their ability to efficiently screen potential compounds in advance. In contrast, our model leverages large-scale learning of relevant knowledge to effectively identify candidate compounds that are likely to interact with specific drug molecules or protein targets, thereby avoiding repetitive and aimless experimental efforts [31].
We conducted a case study involving the drug Sertraline and the complex 3QMN. Sertraline is a selective serotonin reuptake inhibitor, primarily used for the treatment of depression, anxiety disorders, obsessive–compulsive disorder, and related mood disorders [32]. The complex 3QMN corresponds to the crystal structure of 4′-phosphopantetheinyl transferase AcpS from Vibrio cholerae O1 biovar eltor [33].
In Table 2, the model produces affinity scores above 0.9 for all drug–target and target–compound pairs, indicating predicted matches. Specifically, the interaction between Sertraline and its corresponding target is validated by DrugBank [34], while the interaction between the target 3QMN and its associated compound is verified using the PDB [35]. Given that our model is capable of learning from large-scale data, it can serve as an efficient and accurate initial guide for MS, such as predicting potential drug–target binding modes or screening high-confidence candidate molecules. This significantly reduces the computational resources and time costs required by traditional simulation techniques. In comparison to conventional approaches, our method offers data-driven prior knowledge, enabling dual improvements in both simulation accuracy and efficiency [36].

2.3.2. Interpretable Prediction of Functional Structures

Beyond predicting interaction likelihood, a central goal of LoF-DTI is to provide interpretable, atom-level insights to directly guide molecular simulations. To more effectively support traditional methods such as MS, LoF-DTI provides targeted guidance at the molecular level. We extracted the “Att map” from the GCA module, which captures the correlation between target residues and individual atoms within the drug molecule. After removing virtual nodes, we traversed the attention matrix and selected the top 20% of high-weight indices. These indices were then mapped back onto the molecular graph of the drug to identify the most critical atomic regions involved in the specific drug–target interaction. This process provides precise structural insights that can guide downstream tasks in molecular simulation, such as docking and conformational sampling.
As shown in Figure 4, we conducted a case study on the8DL1 complex (BoGH13ASus-E523Q from Bacteroides ovatus bound to maltoheptaose) to validate the interpretability and practical applicability of our model predictions.
The upper and lower ends of the complex are composed of Maltopentaose, while the central ligand is Maltoheptaose. Although the structures may appear complex, they are essentially oligosaccharide-based organic compounds. Oligosaccharides, consisting of 3 to 10 monosaccharide units linked by glycosidic bonds, have a relatively simple elemental composition, allowing focused analysis of functional moieties with minimal interference from rare elements [42]. For the prediction involving Maltopentaose, we input only the protein sequences from the upper and lower domains into the model. In contrast, for Maltoheptaose, the complete protein sequence was used as input.
According to relevant biochemical experiments and literature [43], the classical catalytic pocket is composed of a catalytic triad—D477, E523 (mutation site E523Q), and D581—surrounded by aromatic residues (Y365, F478, Y440, F442, F525) that precisely position the oligosaccharide substrate at the −1 to +2 subsites through π–π interactions and hydrogen bonds. Additionally, an α-1,6-branched glucose at the −2 position (−2′) is specifically recognized and stabilized via a hydrogen bond network involving R641, the backbone of W363, and the side chain of N366. Two newly identified sites include: (1) an aromatic platform on the N-terminal CBM98 module (residues 44–163), formed by W92 and W98, which interacts with the O2/O3 atoms of the oligosaccharide through hydrogen bonding with polar residues; and (2) a surface site on the catalytic domain, located at W555 and Y592, which binds to a pentasaccharide unit via direct hydrogen bonding and water-mediated bridges.
LoF-DTI successfully predicted all these critical sites. The extracted attention weights were significantly enriched around the catalytic pocket and the newly identified sites, indicating that the model not only captures known functional residues but also learns potential structure–function association patterns from complex 3D structures. These results demonstrate the potential of LoF-DTI to assist in experimental structural-function annotation and the identification of novel active sites, providing strong support for a deeper understanding of drug–target mechanisms of action.

3. Materials and Methods

3.1. Evaluation Metrics and Implementation

The performance of LoF-DTI was evaluated using five metrics: Area Under the Receiver Operating Characteristic Curve (AUROC), Area Under the Precision–Recall Curve (AUPRC), accuracy, sensitivity, and specificity. Among them, AUROC (Area Under the Receiver Operating Characteristic Curve) and AUPRC (Area Under the Precision-Recall Curve) serve as the primary indicators of classification effectiveness. AUROC reflects the trade-off between true positive and false positive rates across varying thresholds, while AUPRC highlights the balance between precision and recall. Additionally, accuracy, sensitivity, and specificity are reported at the threshold corresponding to the optimal F1 score. The definitions of metrics are as follows:
S e n s i t i v i t y = F P F P + T N ,
S p e c i f i c i t y = T P T P + F N ,
P r e c i s i o n = T P T P + F P ,
A c c u r a c y = T P + T N T P + F P + T N + F N ,
here, T P ,   T N ,   F P , F N denote true positives (The model predicts positive and the pair is labeled positive in the dataset/by the thresholding rule), true negatives (The model predicts negative and the pair is labeled negative), false positives (The model predicts positive but the pair is labeled negative, either measured inactive or below the activity threshold), and false negatives (The model predicts negative but the pair is labeled positive), respectively. In general, higher values across these metrics indicate better model performance. This work focuses on prediction correctness rather than structural geometric accuracy; therefore, we do not employ interface-level structural metrics such as I-RMSD [44] or I-INF [45]. These metrics require ground-truth three-dimensional complex structures, which are not available under the current data conditions.
Our experimental setup was deployed on a system equipped with two RTX 4070 units, made by NVIDIA in the USA, each with specifications of 8 GB of graphics memory, 12 virtual CPUs, and 32 GB of RAM. The execution environment and versions of tools adopted by LoF-DTI include: PyTorch 2.3.2+CUDA 11.8, Python 3.9.23, DGL 0.9.1post1, Numpy 1.23.5, Pandas 1.5.3, and RDKit 2022.09.5. Key hyperparameter settings are summarized in Table 3.

3.2. Datasets

We evaluated the performance of LoF-DTI and six state-of-the-art baseline models on four benchmark datasets: BindingDB [46], BioSNAP [47], Human [48], and DAVIS [49]. Detailed statistics for each dataset are provided in Table 4.
Each dataset was randomly split into training, validation, and test sets with a fixed ratio of 7:1:2. The test sets consist exclusively of unseen drug–target pairs. For each dataset, we conducted five independent experiments. The model achieving the best AUROC on the validation set was selected and subsequently evaluated on the test set to report the final performance.

3.3. Method

LoF-DTI comprises two main modules, as shown in Figure 5. First, DrugGIN and ProteinRS encode the drug and the target, respectively, extracting features that capture fine-grained, functionally relevant substructures (e.g., molecular motifs and active-site residues). Second, the GCA module further abstracts and fuses these features, strengthening the representation, interaction modeling, and alignment of local functional structures between small molecules and their protein targets. This design not only improves DTI predictive accuracy but also enhances structural-level interpretability: the GCA module’s attention mechanism computes atom-to-residue relevance for every drug–protein pair, enabling us to attribute binding interactions to specific regions of the compound and of the protein.

3.3.1. Structure-Enhanced Drug Feature Encoder

To effectively capture the local structural characteristics of drug molecules, we propose a structure-enhanced feature encoder based on the GIN. This encoder first translates each atom’s chemical properties into a numerical format that a neural network can process, and then uses the GIN architecture to learn from the molecules’ graph structure.
Step 1: From SMILES String to Raw Atomic Features
The process begins with the drug’s SMILES string, a standard text-based representation of a molecule. Using the RDKit library, we parse this string to create a computational molecular graph, where atoms are nodes and bonds are edges. For each atom (node) in this graph, we extract a set of 8 key chemical attributes: its atom type (e.g., Carbon, Oxygen), atomic degree (number of bonds), number of implicit hydrogens, formal charge, number of radical electrons, hybridization state (e.g., sp2, sp3), total number of attached hydrogens, and an indicator for whether it is part of an aromatic ring. These attributes are then converted into a single numerical vector for each atom. For instance, categorical features like atom type are one-hot encoded, while others are represented as integers. By concatenating all these numerical values, we create a 74-dimensional feature vector for every atom in the molecule. For a drug molecule containing N atoms, we stack these individual vectors to form an input feature matrix, which we denote as X. The dimensions of this matrix are N × 74.
Step 2: Learning Initial Atom Embeddings via Linear Projection
This raw feature matrix X is sparse and high-dimensional. It represents chemical properties in a way that is human-readable but not optimized for a neural network. For example, the raw numbers do not inherently capture the chemical similarity between different elements. Therefore, our first task is to transform this raw data into a more compact, dense, and meaningful representation—a process known as creating an “embedding.”
We achieve this using a simple linear transformation followed by a non-linear activation function ( R e L u ), as shown in Equation 5. This is a standard input layer in a neural network.
H ( 0 ) = R e L u ( X W 0 ) ,
X is the input feature matrix described above, with dimensions (N, 74), where N is the number of atoms. W 0 is a learnable weight matrix with dimensions. The values in this matrix are not fixed but are automatically adjusted during model training. The network learns the optimal transformation to convert the 74 raw features into meaningful features. The matrix multiplication X W 0 projects each atom’s 74-dimensional raw feature vector into a new dimensional vector. ReLU is the Rectified Linear Unit activation function, a standard component that introduces non-linearity, allowing the model to learn more complex patterns. H ( 0 ) is the resulting output matrix, with dimensions. Each row of H ( 0 ) is now a dense vector, or “embedding,” that represents an atom in a way that is optimized for the subsequent neural network layers. This matrix serves as the initial node representation for the drug molecule. For processing molecules in batches, all molecular graphs are padded with zero-valued dummy nodes to a fixed maximum node count.
Step 3: Aggregating Neighborhood Information with GIN and Jumping Knowledge
With these initial atom embeddings H ( 0 ) , we then use a stack of Graph Isomorphism Network (GIN) layers to extract structure-aware representations. Each GIN layer updates an atom’s feature vector by aggregating information from its immediate neighbors
h v ( k ) = ML P ( k ) ( 1 + ε ( k ) ) h v ( k 1 ) + u N ( v ) h u ( k 1 ) ,
where h v ( k )   is the feature vector of node v at layer k , N ( v ) denotes the neighbors of node v , MLP is a multi-layer perceptron, and ε is a learnable parameter. After each GIN layer, batch normalization and ReLU activation are applied to stabilize training.
To alleviate the limitations of deep GNNs in capturing hierarchical features and to enhance gradient flow, we incorporate J K aggregation. Specifically, the outputs of all GIN layers are concatenated and projected back to a unified feature space:
h J K = W J K C o n c a t h ( 1 ) , h ( 2 ) , , h ( K ) ,
W J K denotes the weight matrix for the linear transformation applied after J K aggregation. This allows the model to flexibly combine information from different receptive field sizes and network depths, enriching the encoded representation. These node-level embeddings are then passed to downstream modules for further fusion or prediction tasks.

3.3.2. Structure-Enhanced Protein Feature Encoder

To extract informative and structure-aware representations from protein sequences, we design a convolutional encoder named ProteinRS, which leverages residual convolutional blocks to enhance the hierarchical modeling of local sequential patterns. This encoder not only captures local motifs in amino acid sequences but also preserves gradient flow and alleviates degradation through residual connections. This structure-aware design makes it particularly effective in modeling the biochemical and functional relevance of amino acid sequences in tasks such as drug–target interaction prediction.
Each protein sequence is treated as a string over a 25-character alphabet representing amino acids. Let D p denote the dimensionality of the target (protein) latent space. We initialize a learnable embedding matrix X V R 25 × D p covering all 25 amino-acid tokens. By looking up X V , each target is encoded into an initial feature matrix. Analogous to the drug side, we then apply a simple linear projection to obtain the dense target input features X p :
X p = σ ( W 0 X d ) , X p R Θ p × D p
Each residual block is composed of a sequence of 1D convolutional layers with batch normalization, activation, and dropout, the residual block computes:
Y = σ F ( X ) + G ( X ) ,
where F ( ) denotes the stacked convolutional transformation, G ( ) denotes a dimension-matching shortcut connection. After the residual blocks, the feature maps are transposed and reshaped to fit downstream processing requirements. This residual design ensures both feature richness and gradient stability during training.
To effectively extract local functional structure information from protein sequences, we adopt a 3-mer-based statistical encoding scheme combined with a random projection-based dimensionality reduction method. This approach preserves local sequence patterns while projecting high-dimensional representations into a more compact embedding space, enhancing efficiency and maintaining functional relevance.
Given a protein sequence composed of characters from an alphabet of 25 amino acid symbols, we first enumerate all possible combinations of 3 continuous characters. The total number of possible 3-mer patterns is D = 25 3 = 15,625 . Each input sequence is then scanned with a sliding window of size 3 to extract all its 3-mers. We define a count vector x D for each sequence, where each element x i corresponds to the frequency of the i -th 3-mer in the sequence:
x i = n i j = 1 D n j ,
where n i is the number of times the i -th 3-mer appears in the sequence. This normalization ensures that i = 1 D x i = 1 , making the representation invariant to sequence length.
To mitigate the curse of dimensionality and reduce computation cost while retaining discriminative power, we employ a random projection matrix. Each element of R is sampled from a Gaussian distribution:
R i j ~ N 0 , 1 d .
This operation projects the original high-dimensional k-mer frequency vector into a lower-dimensional space while approximately preserving pairwise distances and inner products. Unlike fixed handcrafted descriptors or learned embeddings that may overlook local motifs, our method explicitly models short-range interactions via 3-mers, aligning well with the notion of local functional domains in proteins.

3.3.3. Gated Cross-Attention Module

To enable fine-grained and structure-aware interactions between drug and protein representations, we propose a GCA module. This module performs multi-head attention from one modality (query) to another (key-value), followed by a learnable fusion mechanism that selectively combines the attended features with the original query. This design allows for dynamic and adaptive integration of cross-modal information while preserving the semantics of the original features.
Set a query Q = X p , a key K = X d , and a value V = X d , these representations are then reshaped for multi-head attention [50], where hidden dimension is D H .
Q = W Q Q , K = W K K , V = W V V ,
here, W Q R L q × D H , W K R L k × D H , W V R L v × D H . For each h e a d i , the attention weights are computed as:
A t t e n t i o n ( Q , K , V ) = S o f t max Q K d h / h V .
After applying attention across all heads, the output is concatenated and linearly transformed:
A = L i n e a r ( C o n c a t ( h e a d 1   , , h e a d h   ) ) .
To integrate the attended features A with the original query Q , we introduce gated fusion strategy. This strategy enables token-level, context-aware control over how much attention-driven interaction should influence the final representation, allowing the model to adaptively balance between self-information and cross-modal context:
H = α A + ( 1 α ) P r o j ( Q ) , α ( 0 , 1 )
here, P r o j ( · ) refers to the linear mapping function used to transform feature dimensions. α can be manually fine-tuned according to the task type, or it can be learned adaptively by the model. The formulation for dynamic adaptive learning is as follows:
α = S i g m o i d ( F c N e t ( [ A ; P r o j ( Q ) ] ) ) .
The GCA Module is a core component in our framework for drug–target interaction modeling. This design enables the model to learn both shared and complementary patterns from cross-modal pairs, significantly improving downstream prediction performance.

4. Conclusions

In this study, we introduced Local Functional Structure-aware Drug–Target Interaction, a deep-learning framework that strengthens both the modeling and the alignment of local functional structure features for drug–target interaction prediction. On the drug side, a Jumping-Knowledge-enhanced Graph Isomorphism Network extracts atom-level and neighborhood-level patterns from molecular graphs. On the target side, a convolutional neural network-based residual encoder captures multi-scale sequence motifs, further reinforced with N-mer substructural semantics to emphasize locality. A gated cross-attention module then performs bidirectional, multi-head cross-attention between atom and residue tokens; for each side it forms a cross-attended summary from the other side and uses a learnable gate to modulate the contribution of this cross-attended signal relative to the original local embedding, after which the gated summaries are pooled to obtain a fused representation with token-level attribution. Across four benchmark datasets, the method consistently surpasses six comparative approaches, achieving the best overall performance while maintaining a compact model size. Case studies show that it accurately identifies known catalytic residues and suggests previously unreported functional sites, offering mechanism-aware hypotheses with high confidence. Overall, prioritizing local functional structure features improves both accuracy and interpretability and provides precise guidance for downstream molecular simulation and experimental design, helping to streamline the drug discovery pipeline.

Author Contributions

Conceptualization, B.F. and H.D.; methodology, B.F.; software, B.F.; validation, X.W. and H.D.; formal analysis, B.F.; investigation, K.L.; resources, K.L.; data curation, H.D.; writing—original draft preparation, B.F.; writing—review and editing, K.L. and X.W.; visualization, H.H.Y.T.; supervision, H.H.Y.T.; project administration, K.L. and X.W.; funding acquisition, K.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the fund from Macao Polytechnic University (RP/FCA-14/2023), The Science and Technology Development Funds (FDCT) of Macao (0033/2023/RIB2), and Macau Science and Technology Development Fund and the Department of Science and Technology of Guangdong Province (FDCT-GDST, 0009/2024/AGJ) with the submission approval code of fca.f920.d578.8.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The complete code of LoF-DTI and all associated datasets are available at: https://github.com/fbbgood/LoFDTI.git (accessed on 1 September 2025). The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding authors.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ciccotti, G.; Dellago, C.; Ferrario, M.; Hernández, E.R.; Tuckerman, M.E. Molecular Simulations: Past, Present, and Future (a Topical Issue in EPJB). Eur. Phys. J. B 2022, 95, 3, Correction to Eur. Phys. J. B 2022, 95, 17. [Google Scholar] [CrossRef]
  2. Liu, Y.; Luo, J.; Meenu, M.; Xu, B. Anti-Obesity Mechanisms Elucidation of Essential Oil Components from Artemisiae Argyi Folium (Aiye) by the Integration of GC-MS, Network Pharmacology, and Molecular Docking. Int. J. Food Prop. 2025, 28, 2452445. [Google Scholar] [CrossRef]
  3. Yu, Y.; Xu, S.; He, R.; Liang, G. Application of molecular simulation methods in food science: Status and prospects. J. Agric. Food Chem. 2023, 71, 2684–2703. [Google Scholar] [CrossRef]
  4. Son, A.; Kim, W.; Park, J.; Lee, W.; Lee, Y.; Choi, S.; Kim, H. Utilizing Molecular Dynamics Simulations, Machine Learning, Cryo-EM, and NMR Spectroscopy to Predict and Validate Protein Dynamics. Int. J. Mol. Sci. 2024, 25, 9725. [Google Scholar] [CrossRef] [PubMed]
  5. Jin, Z.; Wei, Z. Molecular Simulation for Food Protein–Ligand Interactions: A Comprehensive Review on Principles, Current Applications, and Emerging Trends. Compr. Rev. Food Sci. Food Saf. 2024, 23, e13280. [Google Scholar] [CrossRef] [PubMed]
  6. Müller, K.-R.; Mika, S.; Tsuda, K.; Schölkopf, K. An Introduction to Kernel-Based Learning Algorithms. In Handbook of Neural Network Signal Processing, 1st ed.; CRC Press: Boca Raton, FL, USA, 2002; ISBN 978-1-315-22041-3. [Google Scholar]
  7. Prašnikar, E.; Ljubič, M.; Perdih, A.; Borišek, J. Machine Learning Heralding a New Development Phase in Molecular Dynamics Simulations. Artif. Intell. Rev. 2024, 57, 102. [Google Scholar] [CrossRef]
  8. Meng, X.-Y.; Zhang, H.-X.; Mezei, M.; Cui, M. Molecular Docking: A Powerful Approach for Structure-Based Drug Discovery. Curr. Comput. Aided Drug Des. 2011, 7, 146–157. [Google Scholar] [CrossRef]
  9. Pahikkala, T.; Airola, A.; Pietilä, S.; Shakyawar, S.; Szwajda, A.; Tang, J.; Aittokallio, T. Toward More Realistic Drug-Target Interaction Predictions. Brief. Bioinform. 2015, 16, 325–337. [Google Scholar] [CrossRef]
  10. Noé, F.; Tkatchenko, A.; Müller, K.-R.; Clementi, C. Machine Learning for Molecular Simulation. Annu. Rev. Phys. Chem. 2020, 71, 361–390. [Google Scholar] [CrossRef]
  11. Zhou, J.; Huang, M. Navigating the Landscape of Enzyme Design: From Molecular Simulations to Machine Learning. Chem. Soc. Rev. 2024, 53, 8202–8239. [Google Scholar] [CrossRef]
  12. Askr, H.; Elgeldawi, E.; Aboul Ella, H.; Elshaier, Y.A.M.M.; Gomaa, M.M.; Hassanien, A.E. Deep Learning in Drug Discovery: An Integrative Review and Future Challenges. Artif. Intell. Rev. 2023, 56, 5975–6037. [Google Scholar] [CrossRef]
  13. Huang, K.; Xiao, C.; Glass, L.M.; Sun, J. MolTrans: Molecular Interaction Transformer for Drug–Target Interaction Prediction. Bioinformatics 2021, 37, 830–836. [Google Scholar] [CrossRef] [PubMed]
  14. Bian, J.; Zhang, X.; Zhang, X.; Xu, D.; Wang, G. MCANet: Shared-Weight-Based MultiheadCrossAttention Network for Drug–Target Interaction Prediction. Brief. Bioinform. 2023, 24, bbad082. [Google Scholar] [CrossRef]
  15. Bai, P.; Miljković, F.; John, B.; Lu, H. Interpretable Bilinear Attention Network with Domain Adaptation Improves Drug–Target Prediction. Nat. Mach. Intell. 2023, 5, 126–136. [Google Scholar] [CrossRef]
  16. Feng, B.-M.; Zhang, Y.-Y.; Niu, N.-W.-J.; Zheng, H.-Y.; Wang, J.-L.; Feng, W.-F. DeFuseDTI: Interpretable Drug Target Interaction Prediction Model with Dual-Branch Encoder and Multiview Fusion. Future Gener. Comput. Syst. 2024, 161, 239–247. [Google Scholar] [CrossRef]
  17. Ning, Q.; Wang, Y.; Zhao, Y.; Sun, J.; Jiang, L.; Wang, K.; Yin, M. DMHGNN: Double Multi-View Heterogeneous Graph Neural Network Framework for Drug-Target Interaction Prediction. Artif. Intell. Med. 2025, 159, 103023. [Google Scholar] [CrossRef]
  18. Li, J.; Bi, X.; Ma, W.; Jiang, H.; Liu, S.; Lu, Y.; Wei, Z.; Zhang, S. MHAN-DTA: A Multiscale Hybrid Attention Network for Drug-Target Affinity Prediction. IEEE J. Biomed. Health Inform. 2024, 1–12. [Google Scholar] [CrossRef]
  19. Liang, X.; Lai, G.; Yu, J.; Lin, T.; Wang, C.; Wang, W. Herbal Ingredient-Target Interaction Prediction via Multi-Modal Learning. Inf. Sci. 2025, 711, 122115. [Google Scholar] [CrossRef]
  20. Feng, B.-M.; Zhang, Y.-Y.; Zhou, X.-C.; Wang, J.-L.; Feng, Y.-F. MolLoG: A Molecular Level Interpretability Model Bridging Local to Global for Predicting Drug Target Interactions. J. Chem. Inf. Model. 2024, 64, 4348–4358. [Google Scholar]
  21. Liu, Y.; Qiu, L.; Li, A.; Fei, R.; Li, J.; Wu, F.-X. Prediction of miRNA Family Based on Class-Incremental Learning. In Proceedings of the 2024 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Lisbon, Portugal, 3–6 December 2024; pp. 197–200. [Google Scholar]
  22. Gao, Y.; Li, F.; Meng, F.; Ge, D.; Ren, Q.; Shang, J. Spatial Domains Identification Based on Multi-View Contrastive Learning in Spatial Transcriptomics. In Proceedings of the 2024 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Lisbon, Portugal, 3–6 December 2024; pp. 521–527. [Google Scholar]
  23. Hu, H.; Wang, X.; Zhang, Y.; Chen, Q.; Guan, Q. A Comprehensive Survey on Contrastive Learning. Neurocomputing 2024, 610, 128645. [Google Scholar] [CrossRef]
  24. Xu, K.; Hu, W.; Leskovec, J.; Jegelka, S. How Powerful Are Graph Neural Networks? arXiv 2019. [Google Scholar] [CrossRef]
  25. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. arXiv 2015. [Google Scholar] [CrossRef]
  26. Lee, I.; Keum, J.; Nam, H. DeepConv-DTI: Prediction of Drug-Target Interactions via Deep Learning with Convolution on Protein Sequences. PLoS Comput. Biol. 2019, 15, e1007129. [Google Scholar] [CrossRef]
  27. Nguyen, T.; Le, H.; Quinn, T.P.; Nguyen, T.; Le, T.D.; Venkatesh, S. GraphDTA: Predicting Drug–Target Binding Affinity with Graph Neural Networks. Bioinformatics 2021, 37, 1140–1147. [Google Scholar] [CrossRef] [PubMed]
  28. Liu, F.; Xu, H.; Cui, P.; Li, S.; Wang, H.; Wu, Z. NFSA-DTI: A Novel Drug–Target Interaction Prediction Model Using Neural Fingerprint and Self-Attention Mechanism. Int. J. Mol. Sci. 2024, 25, 11818. [Google Scholar] [CrossRef]
  29. Zhang, Y.; Wang, Q.; Zhang, C.; Feng, B.; Shang, J.; Zhang, L. IHDFN-DTI: Interpretable Hybrid Deep Feature Fusion Network for Drug–Target Interaction Prediction. Interdiscip. Sci. Comput. Life Sci. 2025, 1–15. [Google Scholar] [CrossRef]
  30. Kipf, T.N.; Welling, M. Semi-Supervised Classification with Graph Convolutional Networks. arXiv 2017. [Google Scholar] [CrossRef]
  31. Randeni, N.; Luo, J.; Wu, Y.; Xu, B. Elucidating the Anti-Diabetic Mechanisms of Mushroom Chaga (Inonotus obliquus) by Integrating LC-MS, Network Pharmacology, Molecular Docking, and Bioinformatics. Int. J. Mol. Sci. 2025, 26, 5202. [Google Scholar] [CrossRef]
  32. Davis, L.L.; Behl, S.; Lee, D.; Zeng, H.; Skubiak, T.; Weaver, S.; Hefting, N.; Larsen, K.G.; Hobart, M. Brexpiprazole and Sertraline Combination Treatment in Posttraumatic Stress Disorder: A Phase 3 Randomized Clinical Trial. JAMA Psychiatry 2025, 82, 218–227. [Google Scholar] [CrossRef]
  33. Halavaty, A.S.; Kim, Y.; Minasov, G.; Shuvalova, L.; Dubrovska, I.; Winsor, J.; Zhou, M.; Onopriyenko, O.; Skarina, T.; Papazisi, L.; et al. Structural Characterization and Comparison of Three Acyl-Carrier-Protein Synthases from Pathogenic Bacteria. Acta Crystallogr. Sect. D Biol. Crystallogr. 2012, 68, 1359–1370. [Google Scholar] [CrossRef] [PubMed]
  34. Knox, C.; Wilson, M.; Klinger, C.M.; Franklin, M.; Oler, E.; Wilson, A.; Pon, A.; Cox, J.; Chin, N.E.; Strawbridge, S.A.; et al. DrugBank 6.0: The DrugBank Knowledgebase for 2024. Nucleic Acids Res. 2024, 52, D1265–D1275. [Google Scholar] [CrossRef]
  35. Burley, S.K.; Berman, H.M.; Kleywegt, G.J.; Markley, J.L.; Nakamura, H.; Velankar, S. Protein Data Bank (PDB): The Single Global Macromolecular Structure Archive. In Protein Crystallography: Methods and Protocols; Wlodawer, A., Dauter, Z., Jaskolski, M., Eds.; Springer: New York, NY, USA, 2017; pp. 627–641. ISBN 978-1-4939-7000-1. [Google Scholar]
  36. Li, J.; Zhai, X.; Liu, J.; Lam, C.K.; Meng, W.; Wang, Y.; Li, S.; Wang, Y.; Li, K. Integrated Causal Inference Modeling Uncovers Novel Causal Factors and Potential Therapeutic Targets of Qingjin Yiqi Granules for Chronic Fatigue Syndrome. Acupunct. Herb. Med. 2024, 4, 122. [Google Scholar] [CrossRef]
  37. Zhou, Y.; Zhang, Y.; Zhao, D.; Yu, X.; Shen, X.; Zhou, Y.; Wang, S.; Qiu, Y.; Chen, Y.; Zhu, F. TTD: Therapeutic Target Database Describing Target Druggability Information. Nucleic Acids Res. 2024, 52, D1465–D1477. [Google Scholar] [CrossRef] [PubMed]
  38. Ghanizadeh, A. Sertraline-Associated Hair Loss. J. Drugs Dermatol. 2008, 7, 693–694. [Google Scholar]
  39. Fishback, J.A.; Robson, M.J.; Xu, Y.-T.; Matsumoto, R.R. Sigma Receptors: Potential Targets for a New Class of Antidepressant Drug. Pharmacol. Ther. 2010, 127, 271–282. [Google Scholar] [CrossRef] [PubMed]
  40. Sørensen, L.; Andersen, J.; Thomsen, M.; Hansen, S.M.R.; Zhao, X.; Sandelin, A.; Strømgaard, K.; Kristensen, A.S. Interaction of Antidepressants with the Serotonin and Norepinephrine Transporters: Mutational Studies of the S1 Substrate Binding Pocket. J. Biol. Chem. 2012, 287, 43694–43707. [Google Scholar] [CrossRef]
  41. Daws, L.C. Unfaithful Neurotransmitter Transporters: Focus on Serotonin Uptake and Implications for Antidepressant Efficacy. Pharmacol. Ther. 2009, 121, 89–99. [Google Scholar] [CrossRef] [PubMed]
  42. Li, B.; Cui, J.; Xu, T.; Xu, Y.; Long, M.; Li, J.; Liu, M.; Yang, T.; Du, Y.; Xu, Q. Advances in the Preparation, Characterization, and Biological Functions of Chitosan Oligosaccharide Derivatives: A Review. Carbohydr. Polym. 2024, 332, 121914. [Google Scholar] [CrossRef] [PubMed]
  43. Brown, H.A.; DeVeaux, A.L.; Juliano, B.R.; Photenhauer, A.L.; Boulinguiez, M.; Bornschein, R.E.; Wawrzak, Z.; Ruotolo, B.T.; Terrapon, N.; Koropatkin, N.M. BoGH13ASus from Bacteroides Ovatus Represents a Novel α-Amylase Used for Bacteroides Starch Breakdown in the Human Gut. Cell. Mol. Life Sci. 2023, 80, 232. [Google Scholar] [CrossRef]
  44. Nithin, C.; Kmiecik, S.; Błaszczyk, R.; Nowicka, J.; Tuszyńska, I. Comparative Analysis of RNA 3D Structure Prediction Methods: Towards Enhanced Modeling of RNA–Ligand Interactions. Nucleic Acids Res. 2024, 52, 7465–7486. [Google Scholar] [CrossRef]
  45. Ludwiczak, O.; Antczak, M.; Szachniuk, M. Assessing Interface Accuracy in Macromolecular Complexes. PLoS ONE 2025, 20, e0319917. [Google Scholar] [CrossRef] [PubMed]
  46. Bai, P.; Miljković, F.; Ge, Y.; Greene, N.; John, B.; Lu, H. Hierarchical Clustering Split for Low-Bias Evaluation of Drug-Target Interaction Prediction. In Proceedings of the 2021 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Houston, TX, USA, 9–12 December 2021; pp. 641–644. [Google Scholar]
  47. Zitnik, M.; Agrawal, M.; Leskovec, J. Modeling Polypharmacy Side Effects with Graph Convolutional Networks. Bioinformatics 2018, 34, i457–i466. [Google Scholar] [CrossRef] [PubMed]
  48. Liu, H.; Sun, J.; Guan, J.; Zheng, J.; Zhou, S. Improving Compound–Protein Interaction Prediction by Building up Highly Credible Negative Samples. Bioinformatics 2015, 31, i221–i229. [Google Scholar] [CrossRef]
  49. Davis, M.I.; Hunt, J.P.; Herrgard, S.; Ciceri, P.; Wodicka, L.M.; Pallares, G.; Hocker, M.; Treiber, D.K.; Zarrinkar, P.P. Comprehensive Analysis of Kinase Inhibitor Selectivity. Nat. Biotechnol. 2011, 29, 1046–1051. [Google Scholar] [CrossRef]
  50. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
Figure 1. Ablation study on different modules.
Figure 1. Ablation study on different modules.
Ijms 26 10194 g001
Figure 2. Impact of gating strategies on model performance.
Figure 2. Impact of gating strategies on model performance.
Ijms 26 10194 g002
Figure 3. Impact of N-mer features and the number of attention heads in the GCA module on model performance.
Figure 3. Impact of N-mer features and the number of attention heads in the GCA module on model performance.
Ijms 26 10194 g003
Figure 4. LoF-DTI Predicted Key Atom Visualizations and Validation Images. (A) Overview of the 8DL1 complex, with Maltoheptaose in the binding domain shown as colored spheres. (BD) Validated binding site visualizations. (EG) Key functional structures predicted by LoF-DTI, highlighted in red.
Figure 4. LoF-DTI Predicted Key Atom Visualizations and Validation Images. (A) Overview of the 8DL1 complex, with Maltoheptaose in the binding domain shown as colored spheres. (BD) Validated binding site visualizations. (EG) Key functional structures predicted by LoF-DTI, highlighted in red.
Ijms 26 10194 g004
Figure 5. The Architecture of the LoF-DTI Model. Drug SMILES strings are converted into molecular graphs, while protein sequences are mapped into feature matrices. The DrugGIN module employs a multi-layer GIN combined with a Jumping Knowledge mechanism to effectively extract local structural features from drug molecular graphs. The ProteinRS module leverages a residual convolutional network to extract protein representations that preserve local domain-level information from the amino acid sequence, while concurrently incorporating N-mer-level substructural semantic features to capture scale patterns. Subsequently, the GCA module facilitates cross-modal feature interaction between drugs and targets, highlighting the coupling relationships of key functional regions. The fused features f are finally passed through a MLP classifier to predict potential drug–target interactions.
Figure 5. The Architecture of the LoF-DTI Model. Drug SMILES strings are converted into molecular graphs, while protein sequences are mapped into feature matrices. The DrugGIN module employs a multi-layer GIN combined with a Jumping Knowledge mechanism to effectively extract local structural features from drug molecular graphs. The ProteinRS module leverages a residual convolutional network to extract protein representations that preserve local domain-level information from the amino acid sequence, while concurrently incorporating N-mer-level substructural semantic features to capture scale patterns. Subsequently, the GCA module facilitates cross-modal feature interaction between drugs and targets, highlighting the coupling relationships of key functional regions. The fused features f are finally passed through a MLP classifier to predict potential drug–target interactions.
Ijms 26 10194 g005
Table 1. Comparative Performance of Different Models.
Table 1. Comparative Performance of Different Models.
MethodAUROCAUPRCAccuracySensitivitySpecificity
BindingDB
DeepConv-DTI0.944 ± 0.0040.925 ± 0.0050.882 ± 0.0070.873 ± 0.0180.884 ± 0.009
GraphDTA 0.950 ± 0.0030.934 ± 0.0020.888 ± 0.0050.882 ± 0.0120.887 ± 0.008
MolTrans 0.952 ± 0.0020.933 ± 0.0040.887 ± 0.0060.884 ± 0.0190.883 ± 0.011
DrugBAN 0.956 ± 0.0030.943 ± 0.0030.897 ± 0.0030.890 ± 0.0150.896 ± 0.008
NFSA-DTI 0.951 ± 0.0030.933 ± 0.0040.883 ± 0.0040.892 ± 0.0080.908 ± 0.012
IHDFN-DTI0.955 ± 0.0020.939 ± 0.0030.893 ± 0.0030.884 ± 0.0120.912 ± 0.009
LoF-DTI0.963 ± 0.0050.947 ± 0.0050.902 ± 0.0020.896 ± 0.0150.918 ± 0.007
BioSNAP
DeepConv-DTI 0.886 ± 0.0060.890 ± 0.0060.805 ± 0.0090.760 ± 0.0290.851 ± 0.011
GraphDTA 0.887 ± 0.0080.890 ± 0.0070.800 ± 0.0070.745 ± 0.0320.854 ± 0.025
MolTrans 0.890 ± 0.0060.891 ± 0.0050.804 ± 0.0030.755 ± 0.0210.846 ± 0.022
DrugBAN 0.903 ± 0.0050.900 ± 0.0040.836 ± 0.0090.825 ± 0.0180.849 ± 0.013
NFSA-DTI 0.897 ± 0.0040.895 ± 0.0080.832 ± 0.0100.807 ± 0.0150.844 ± 0.011
IHDFN-DTI 0.903 ± 0.0050.908 ± 0.0060.835 ± 0.0070.815 ± 0.0220.862 ± 0.008
LoF-DTI0.905 ± 0.0030.904 ± 0.0020.841 ± 0.0050.812 ± 0.0200.872 ± 0.014
DAVIS
DeepConvDTI0.884 ± 0.0080.299 ± 0.0390.774 ± 0.0120.754 ± 0.0400.876 ± 0.013
DeepDTA 0.880 ± 0.0070.301 ± 0.0440.773 ± 0.0100.765 ± 0.0450.880 ± 0.024
MolTrans0.892 ± 0.0040.371 ± 0.0310.779 ± 0.0170.781 ± 0.0230:878 ± 0.012
DrugBAN0.892 ± 0.0050.333 ± 0.0390.770 ± 0.0150.751 ± 0.0240.869 ± 0.011
NFSA-DTI0.884 ± 0.0080.329 ± 0.0280.774 ± 0.0120.754 ± 0.0300.866 ± 0.013
IHDFN-DTI0.876 ± 0.0050.348 ± 0.0320.778 ± 0.0100.778 ± 0.0130.874 ± 0.007
LoF-DTI0.894 ± 0.0050.354 ± 0.0230.782 ± 0.0150.782 ± 0.0150.882 ± 0.005
Human
DeepConvDTI0.975 ± 0.0020.969 ± 0.0030.941 ± 0.0020.915 ± 0.0080.934 ± 0.015
DeepDTA0.975 ± 0.0020.969 ± 0.0030.941 ± 0.0020.915 ± 0.0080.934 ± 0.015
MolTrans0.973 ± 0.0030.968 ± 0.0030.943 ± 0.0030.918 ± 0.0070.936 ± 0.013
DrugBAN0.981 ± 0.0040.974 ± 0.0060.938 ± 0.0050.927 ± 0.0110.938 ± 0.018
NFSA-DTI0.980 ± 0.0020.966 ± 0.0050.943 ± 0.0050.930 ± 0.0070.947 ± 0.014
IHDFN-DTI0.983 ± 0.0040.980 ± 0.0030.945 ± 0.0020.938 ± 0.0090.953 ± 0.006
LoF-DTI0.985 ± 0.0040.977 ± 0.0020.948 ± 0.0070.944 ± 0.0120.953 ± 0.008
Table 2. Predicted Interactions for the Drug “Sertraline” and Protein “3QMN”.
Table 2. Predicted Interactions for the Drug “Sertraline” and Protein “3QMN”.
DrugTargetPronteinCompound
Ijms 26 10194 i001P31645 [37]Ijms 26 10194 i002COA
Q01950 [38]A3P
Q99720 [39]MRD
P23975 [40]MPD
SertralinP08684 [41]3QMNACT
Table 3. Key Hyperparameter Settings.
Table 3. Key Hyperparameter Settings.
HyperparameterSetting
OptimizerAdam
Learning rate1 × 10−5
MAX_Epoch100
BATCH_SIZE64
Number of residual blocks2
GIN layers4
CNN kernel size[3, 6, 9]
Heads of attention4
Attention pooling size3
Table 4. Datasets statistic.
Table 4. Datasets statistic.
Dataset#Drugs#Proteins#Interactions
BindingDB14,643262349,200
BioSNAP4510218127,465
Human272620016728
DAVIS7238211,885
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Feng, B.; Du, H.; Tong, H.H.Y.; Wang, X.; Li, K. Enhancing Local Functional Structure Features to Improve Drug–Target Interaction Prediction. Int. J. Mol. Sci. 2025, 26, 10194. https://doi.org/10.3390/ijms262010194

AMA Style

Feng B, Du H, Tong HHY, Wang X, Li K. Enhancing Local Functional Structure Features to Improve Drug–Target Interaction Prediction. International Journal of Molecular Sciences. 2025; 26(20):10194. https://doi.org/10.3390/ijms262010194

Chicago/Turabian Style

Feng, Baoming, Haofan Du, Henry H. Y. Tong, Xu Wang, and Kefeng Li. 2025. "Enhancing Local Functional Structure Features to Improve Drug–Target Interaction Prediction" International Journal of Molecular Sciences 26, no. 20: 10194. https://doi.org/10.3390/ijms262010194

APA Style

Feng, B., Du, H., Tong, H. H. Y., Wang, X., & Li, K. (2025). Enhancing Local Functional Structure Features to Improve Drug–Target Interaction Prediction. International Journal of Molecular Sciences, 26(20), 10194. https://doi.org/10.3390/ijms262010194

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop