Next Article in Journal
Evaluation of the Antihistamine and Anti-Inflammatory Effects of a Nutraceutical Blend Based on Quercetin, Perilla frutescens, Boswellia serrata, Blackcurrant, Parthenium, Helichrysum, Lactobacillus acidophilus and Bifidobacterium animalis Through In Vitro and In Vivo Approaches—Preliminary Data
Next Article in Special Issue
Genetic Evidence Prioritizes Neurocognitive Decline as a Causal Driver of Sleep Disturbances: A Multi-Omics Analysis Identifying Causal Genes and Therapeutic Targets
Previous Article in Journal
Zingerone Targets LKB1/AMPK to Block FcεRI-Dependent Mast Cell Degranulation and Anaphylaxis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CAMF-DTI: Enhancing Drug–Target Interaction Prediction via Coordinate Attention and Multi-Scale Feature Fusion

1
College of Information Science and Technology, Beijing University of Chemical Technology, Beijing 100029, China
2
School of Computer Science and Technology, Harbin Institute of Technology, Harbin 150001, China
*
Author to whom correspondence should be addressed.
Curr. Issues Mol. Biol. 2025, 47(11), 964; https://doi.org/10.3390/cimb47110964
Submission received: 13 October 2025 / Revised: 12 November 2025 / Accepted: 14 November 2025 / Published: 20 November 2025
(This article belongs to the Special Issue Featured Papers in Bioinformatics and Systems Biology)

Abstract

The accurate prediction of drug–target interactions is essential for drug discovery and development. However, current models often struggle with two challenges. First, they fail to model the directional flow and positional sensitivity of protein sequences, which are critical for identifying functional interaction regions. Second, they lack mechanisms to integrate multi-scale information from both local binding sites and broader structural context. To overcome these limitations, we propose CAMF-DTI, a novel framework that incorporates coordinate attention, multi-scale feature fusion, and cross-attention to enhance both the representation and interaction learning of drug and protein features. Drug molecules are represented as molecular graphs and encoded using graph convolutional networks, while protein sequences are processed with coordinate attention to preserve directional and spatial information. Multi-scale fusion modules are applied to both encoders to capture local and global features, and a cross-attention module integrates the representations to enable dynamic drug–target interaction modeling. We evaluate CAMF-DTI on four benchmark datasets: BindingDB, BioSNAP, C.elegans, and Human. Experimental results show that CAMF-DTI consistently outperforms seven state-of-the-art baselines in terms of AUROC, AUPRC, Accuracy, F1-score, and MCC. Ablation studies further confirm the effectiveness of each module, and visualization results demonstrate the model’s potential interpretability.

1. Introduction

Accurately identifying potential drug–target interactions (DTIs) is a central step in the early stages of drug development, enabling efficient drug discovery, repurposing, and safety evaluation [1,2,3]. However, traditional experimental methods such as high-throughput screening and affinity assays are often time-consuming, expensive, and limited in scale. In this context, computational DTI prediction has become not only a promising complement to experimental methods but an urgent necessity [4,5].
Deep learning-based DTI prediction models have advanced significantly in recent years, offering scalable, end-to-end solutions for modeling complex drug–target relationships. Broadly, these models can be categorized into two main classes: sequence-based and structure-based approaches [6]. Sequence-based methods focus on one-dimensional molecular representations, such as SMILES strings for drugs and amino acid sequences [7] for proteins. Convolutional neural networks, as used in DeepDTA [8] and WideDTA [9], are employed to extract local patterns from these sequences. Transformer-based architectures, including Trans formerCPI [10] and MolTrans [11], further enhance sequence modeling by capturing long-range dependencies within molecular sequences. Structure-based methods incorporate topological and spatial features of molecular structures to improve representation learning. Graph-based approaches, such as GraphDTA [12], MGraphDTA [13], and NeoDTI [14], treat drug compounds as molecular graphs and apply graph neural networks to capture local connectivity and substructural patterns. Other models, such as GraphDTA [12], Graph-DTI [15], and EquiBind [16], go beyond 2D topology by integrating residue-level 3D coordinates or pairwise distance matrices to represent protein spatial conformation [17]. These structure-aware designs enable the modeling of geometric compatibility and physicochemical interactions at binding interfaces. These models have achieved competitive results on benchmark datasets such as BindingDB [18], Davis [19], and KIBA [20], highlighting the important roles of sequence-derived and structure-enriched representations in DTI prediction.
Despite recent progress, existing DTI models still face core limitations in both sequence and structure modeling. In sequence-based models, a major problem is the lack of interactive representation learning. Drugs and proteins are often encoded independently and fused only at the final stage, preventing the model from capturing dynamic dependencies between them. In structure-based models, two limitations are common. (1) Most methods ignore the directional nature [21] of protein structures, such as the N-to-C terminal flow, making it hard to distinguish regions with similar geometry but different biological roles. (2) Many models use single-scale aggregation strategies, which limit their ability to capture both local binding sites and global conformational features [22,23].
To address the above issues, this study proposes a DTI prediction model named CAMF DTI, which integrates coordinate attention, multi-scale feature fusion, and cross-attention mechanisms. This architecture enhances the modeling of complex structural and interaction patterns between drugs and proteins. The main contributions are as follows:
  • To tackle the lack of directional information in spatial features, coordinate attention is introduced to jointly encode spatial position and sequence directionality, improving the localization of key interaction regions.
  • To address the limitation of single-scale receptive fields, a multi-scale fusion module is designed to extract structural features at different scales via parallel branches, enabling unified representation of local and global information.
  • To overcome insufficient modeling of drug–target interactions, a cross-attention mechanism is employed to capture dynamic dependencies between drugs and targets during representation learning.
  • Experimental results on multiple benchmark datasets demonstrate that CAMF-DTI outperforms existing state-of-the-art methods across AUROC, AUPRC, Accuracy, F1 score, and MCC metrics, indicating superior predictive performance and strong generalization ability.

2. Materials and Methods

2.1. The Architecture of CAMF-DTI

The overall architecture of the proposed model CAMF-DTI is illustrated in Figure 1. The framework consists of several components, including a drug encoder, a protein encoder, a feature fusion stage, and a prediction stage. Among them, three key modules are specifically designed to address the core limitations of existing methods. Specifically, drug molecules represented by SMILES [7] strings are first processed by a three-layer graph convolutional network (GCN) [24] to extract structural and substructural features. Protein sequences are then passed through the Coordinate Attention Module, which encodes spatial information while incorporating sequence directionality, thereby highlighting functionally relevant regions. Subsequently, both drug and protein features are refined using the Multi-Scale Feature Fusion Module, which captures local binding patterns and global conformational information at multiple receptive fields to achieve multi-level feature representation. Finally, the Cross-Attention Module models dynamic interactions between drugs and proteins, generating a joint representation that is passed to multilayer perceptrons (MLPs) for the final DTI prediction.

2.2. Drug Encoder

2.2.1. Drug Feature Construction

In drug feature extraction, the SMILES sequence is converted into a molecular graph G = (V, E), where V denotes atom nodes and E denotes chemical bonds. Using the DGL-LifeSci toolkit (version 1.0; AWS, USA), each atom is encoded as a 74-dimensional feature vector (atom type, degree, hydrogen count, charge, hybridization, aromaticity, etc.). The detailed atom-level features used in the molecular graph construction are summarized in Table S1 (Supplementary Materials). The node feature matrix of each molecule is denoted as M d N d × 74 , where Nd is the maximum number of atoms, and zero-padding is applied if necessary.
A linear transformation is applied to obtain dense input features:
X d = M d W d T

2.2.2. GCN

To learn molecular representations from the drug graph, we employ a three-layer Graph Convolutional Network (GCN). At layer l, the node feature update is defined as:
H d ( l + 1 ) = σ A ˜ H d ( l ) W g ( l ) + b g ( l )
where H d ( l ) is the node feature matrix, A ˜ is the adjacency matrix (with self-connections), W g ( l ) and b g ( l ) denote the weight and bias of the l-th GCN layer, and σ represents the activation function.
This process aggregates local substructures and captures global chemical information, producing high-level molecular graph representations for downstream fusion.

2.2.3. Multi-Scale Feature Fusion

To further enhance the structural representation learned from the GCN-based drug encoder, we introduce a multi-scale feature fusion module inspired by the model proposed by Ouyang [25] (as illustrated in Figure 1b). This module is designed to capture and integrate local and global features across different receptive fields, thereby improving the expressiveness and robustness of learned representations. Notably, the same module is also incorporated in the protein encoder, allowing both drug and protein features to be refined with rich multi-scale spatial and semantic information.
Let the input be denoted as:
X = [ X 0 ,   X 1 , , X G 1 ]
where each group X i C G × H × W is processed through three parallel branches: two with 1 × 1 convolutions to model local channel-wise dependencies, and one with a 3 × 3 convolution to capture spatial context.
To encode global-level semantics, two average pooling operations are applied across orthogonal directions. Their outputs are concatenated and passed through multiple Sigmoid and GroupNorm layers, enhancing spatial alignment. These aggregated signals are then normalized and passed through a Softmax operation to generate scale-wise attention weights:
Z c = 1 H W i = 1 H j = 1 W X c ( i ,   j )
where ZC is the pooled descriptor for channel c. These descriptors are normalized and passed through a softmax function to compute attention weights across scales. Weighted aggregation is then performed, followed by sigmoid activation to obtain the final fused representation.
By combining local and global features across multiple receptive fields, the multi-scale feature fusion module provides a more expressive and robust feature representation for downstream interaction modeling. Notably, this module is also employed in the protein encoder with the same architecture and processing pipeline, where it is applied after the coordinate attention mechanism to ensure consistent enhancement of hierarchical semantic representations from both modalities. This parallel application maintains architectural coherence between the two encoders while allowing flexible adaptation to their distinct data characteristics.
Although both the drug and protein encoders adopt the same multi-scale feature fusion architecture for consistency, these modules are jointly trained within the end-to-end optimization process of CAMF-DTI. This joint training enables the model to learn modality-specific patterns while promoting coordinated feature evolution between the two encoders. The gradients from the interaction and prediction stages are backpropagated through both encoders simultaneously, ensuring that the multi-scale fusion modules are optimized collaboratively to enhance the complementarity of drug and protein representations.

2.3. Protein Encoder

2.3.1. Protein Feature Construction

To enhance the protein representation and capture long-range dependencies within amino acid sequences, we first initialize each protein sequence P by looking up a learnable embedding matrix E p 22 × D p , where 22 is the number of standard amino acid types and Dp is the embedding dimension. Here, the term “long-range dependency” refers to intra-sequence contextual relationships among distant amino acid residues rather than intermolecular interactions. Modeling such dependencies enables the network to capture non-local structural and functional correlations within the protein chain.
The resulting sequential feature matrix X p θ p × D p (with θ p being the maximum sequence length) is then passed through a multi-order gated convolution encoder, which recursively captures hierarchical and multi-scale contextual information across the sequence.

2.3.2. Coordinate Attention Mechanism

To effectively capture the long-range dependencies and spatially significant positions within protein sequences, we incorporate a coordinate attention mechanism inspired by the model proposed by Hou [26] (as illustrated in Figure 1c). This design addresses the limitations of traditional convolutional architectures, which often fail to model global context and positional sensitivity, especially for long amino acid chains. The coordinate attention mechanism enhances the model’s ability to locate and emphasize informative positions by embedding directional and positional information into the feature maps.
To enable the application of the coordinate attention module, the output of the encoder, originally in the form of a sequential matrix X p θ p × D p , is reshaped into a 3D tensor x C × H × W . Here, C denotes the number of feature channels, with Dp equal to C, and H and W are spatial dimensions such that H × W = θ p .
Specifically, for a given protein input x C × H × W , two pooling operations are applied along the horizontal and vertical directions to encode direction-aware contextual information. The aggregated features are computed as:
z c h ( h ) = 1 W 0 i < W x c ( h ,   i )
z c ω ( ω ) = 1 H 0 j < H x c ( j ,   ω )
These features are concatenated and passed through a shared 1 × 1 convolution to form an intermediate representation. This is then split and transformed by two separate 1 × 1 convolutions with learnable parameters F h and F ω , yielding the attention weights.
g i = σ F i ( f i ) , i = h , ω
where σ denotes the sigmoid activation function. Finally, the output of the coordinate attention block is given by:
y c ( i ,   j ) = x c ( i ,   j ) × g c h ( i ) × g c ω ( j )
Notably, the output of this module is subsequently passed into the Multi-Scale Feature Fusion Module to further enrich the representation by integrating features from multiple receptive fields.
In contrast to conventional attention mechanisms that focus solely on channel-wise dependencies, the coordinate attention mechanism decomposes channel attention into two direction-aware branches, enabling the model to encode both spatial position and sequence directionality. For protein sequences, the N-to-C terminal order often implies structural and functional asymmetry. By aggregating contextual information along horizontal and vertical directions, this mechanism guides the network to emphasize biologically meaningful regions, such as binding pockets or active sites, thereby improving the localization and interpretability of protein representations.

2.4. Feature Fusion Stage

To fuse the hierarchical features of drugs and proteins and accurately capture their interaction patterns, we introduce a cross-attention mechanism. This module adaptively highlights key regions of both representations by treating one modality as query and the other as key-value pairs, enhancing the discriminability of interaction features. The structure is illustrated in Figure 1d.

2.4.1. Interaction Feature Extraction Module

Let the output features from the drug and protein encoders be denoted as FD and FP, respectively. The query and key representations are computed as follows:
Q D = F D W D q , K p = F p W p k
Q p = F p W p q , K D = F D W D k
where W q and W k are modality-specific projection matrices.
Cross-attention is then applied to generate interaction-aware representations via scaled dot-product attention:
Z D = Softmax Q D K p T d k
Z p = Softmax Q p K D T d k
Z D and Z P are the attention weights obtained by applying the Softmax function to the scaled dot-products between the queries and keys.
The fused interaction-aware representations are computed by integrating attention weights.
D f = 0.5 Z D + 0.5 F D
P f = 0.5 Z p + 0.5 F p
D f and P f are the interaction-aware fused representations of drugs and proteins, respectively, computed as the weighted average of attention-aware features and the original features.
To obtain fixed-length representations, we apply global max pooling:
D f r = MaxPooling ( D f )
P f r = MaxPooling ( P f )
Finally, we concatenate the pooled features to form the unified interaction representation:
f = concat ( D f r ,   P f r )

2.4.2. Prediction Module

To predict the probability of DTI, the fused feature representation is fed into a decoder composed of fully connected layers. The predicted DTI probability p is computed as:
p = σ ( W f + b )
where W and b are learnable weight matrix and bias vector, respectively.
During model training, the parameters are optimized using backpropagation. The loss function employed is the cross-entropy loss function, defined as:
loss = i y i log ( p i ) + ( 1 y i ) log ( 1 p i )
where y i denotes the ground-truth label for the i-th drug–target pair, and pi represents the predicted DTI probability.

3. Results and Discussion

3.1. Evaluation Metrics

To better evaluate the performance of the model, we adopt AUROC (Area Under the Receiver Operating Characteristic Curve) [27], AUPRC (Area Under the Precision-Recall Curve) [27], F1-Score [28], Accuracy [29], and MCC (Matthews Correlation Coefficient) as the evaluation metrics across the four benchmark datasets. Among them, MCC is a metric specifically designed for evaluating binary classification performance. It considers all elements of the confusion matrix, including true positives, true negatives, false positives, and false negatives, and is regarded as a relatively balanced measure.

3.2. Datasets

We evaluate our model on four public datasets: BindingDB, BioSNAP, Human, and C.elegans. BindingDB is a large-scale dataset of experimentally verified drug–target interactions, containing 14,643 drugs, 2623 proteins, and 49,199 interactions. BioSNAP is a balanced dataset derived from DrugBank, with 4510 drugs, 2181 proteins, and 27,464 interactions. Human includes 2726 drugs and 2001 proteins, covering 6728 balanced interaction pairs. C.elegans shares the same scale as the Human dataset but focuses on interactions specific to the C.elegans species. The detailed statistics are summarized in Table 1.

3.3. Baseline Methods

To validate the effectiveness of CAMF-DTI, we conducted experiments on four datasets: BindingDB, BioSNAP, C.elegans, and Human. Seven representative baseline methods were selected for comparison:
  • CPI-GNN [30]: Uses GNNs for drug graphs and CNNs for protein sequences, with a single-sided attention mechanism to model the effect of protein subsequences on drugs.
  • BACPI [31]: Combines CNNs and dual attention to extract and integrate compound–protein features, emphasizing local interaction sites.
  • CPGL [32]: Applies GAT and LSTM for robust and generalizable compound–protein feature extraction, followed by a fully connected layer.
  • BINDTI [33]: Encodes drugs via molecular graphs and proteins via ACMix, then fuses their features using a bidirectional intention network.
  • FOTF-CPI [34]: Enhances Transformer with optimal fragmentation and attention fusion to improve interaction prediction.
  • CAT-CPI [35]: Employs CNNs and Transformers to encode proteins and captures drug–target interaction via cross-attention.
  • DO-GMA [36]: Extracts representations using CNNs and GCNs and integrates features via gated and multi-head attention with bilinear fusion.

3.4. Results

3.4.1. Analysis of Performance

In this section, the experimental results of CAMF-DTI are compared with seven baseline models on three benchmark datasets, as shown in Table 2. First, on the BindingDB dataset, CAMF-DTI was trained and evaluated against the seven baselines. Compared with the best-performing baseline DO-GMA, CAMF-DTI achieved improvements in AUROC, AUPRC, Accuracy, and MCC by 0.1%, 0.2%, 0.7%, and 0.2%, respectively, while also demonstrating competitive performance in terms of F1-score.
Next, on the BioSNAP dataset (balanced dataset), the comparison results showed that our method achieved the best performance compared with the best baseline DO-GMA. Specifically, the Accuracy increased by 0.8%, the F1-score by 0.3%, and MCC by 0.2%, while AUROC and AUPRC were slightly lower than DO-GMA, resulting in overall second-best performance.
Finally, on the C.elegans dataset, CAMF-DTI was evaluated against seven baselines. The results indicated that our proposed method outperformed the baselines across multiple evaluation metrics. Compared with the best-performing baseline DO-GMA, CAMF-DTI improved AUROC by 0.1% and AUPRC by 0.1%, while its Accuracy increased by 0.8%. However, in terms of F1-score and MCC, CAMF-DTI was slightly inferior to DO-GMA, thus ranking second overall.
Subsequently, on the Human dataset (balanced dataset), the comparison results demonstrated that our method achieved the second-best performance relative to DO-GMA. Specifically, CAMF-DTI improved AUROC by 0.1% and showed competitive performance in AUPRC, whereas its Accuracy, F1-score, and MCC were slightly lower than DO-GMA, still resulting in overall second-best performance.
Across the BioSNAP, C.elegans, and Human datasets, CAMF-DTI exhibits strong and stable performance compared to the SOTA baseline (DO-GMA). While the two models each lead on some individual metrics, CAMF-DTI achieves higher Accuracy and AUPRC on multiple datasets, indicating its effectiveness in producing precise and confident interaction predictions. This suggests that our model offers a favorable trade-off between sensitivity and specificity in real-world applications. It should be noted that CAMF-DTI’s F1-score and MCC are slightly lower than those of DO-GMA on certain datasets such as C.elegans and Human. This minor gap can be attributed to the inherent trade-off between recall and precision in the attention-based fusion strategy. The coordinate and cross-attention mechanisms in CAMF-DTI emphasize comprehensive feature aggregation and sensitivity to subtle relational cues, which tend to increase recall but may occasionally lead to marginally higher false positives, thereby reducing F1 and MCC values. Nevertheless, this design allows CAMF-DTI to better capture potential interactions that might be overlooked by stricter threshold-based models like DO-GMA, aligning with our goal of enhancing model generalization and biological coverage.
To validate the robustness of the comparative analysis, paired t-tests were conducted across five independent runs for each dataset. The results indicate that the improvements achieved by CAMF-DTI over the baseline methods are statistically significant (p < 0.05) in AUROC, AUPRC, and Accuracy, confirming the reliability of the observed performance gains.

3.4.2. Ablation Experiments

To validate the impact of each module in CAMF-DTI on model performance, we designed three ablation variants. The variant CAMFDTI-Noco replaces the co-attention mechanism with a linear layer while keeping other components unchanged. The variant CAMFDTI-single removes the multi-scale module and performs prediction using only a single-scale model. The variant CAMFDTI-NoCross eliminates the cross-attention mechanism in the model, instead concatenating drug–target features from multiple scales for classification. The results of the ablation experiments are shown in Figure 2.
Compared with CAMFDTI-Noco, CAMF-DTI achieved significant improvements across the four datasets, particularly on the Human dataset, where AUROC, AUPRC, Accuracy, F1-score, and MCC increased by 3.6%, 2.8%, 1.7%, 1.7%, and 2.8%, respectively. Compared with CAMFDTI-single, CAMF-DTI demonstrated substantial gains on the BindingDB dataset, with improvements of 3.1%, 1.6%, 3.2%, 2.8%, and 4.1% in AUROC, AUPRC, Accuracy, F1-score, and MCC, respectively. Relative to CAMFDTI-NoCross, CAMF-DTI achieved notable gains on the Human dataset, with AUROC improved by 2.4%, AUPRC by 1.7%, Accuracy by 1.6%, F1-score by 1.3%, and MCC by 3.4%.
The ablation study further verifies the necessity of each individual module. The performance drop observed when removing the coordinate attention, multi-scale fusion, or cross-attention components indicates that these modules provide complementary advantages. Their synergistic integration enables the model to extract hierarchical features and align cross-modal interactions more effectively. Furthermore, to further assess the generalization capability of CAMF-DTI, an unseen-pair evaluation was conducted by withholding 10% of drug–target pairs that were not present during training. The model maintained robust performance with a less than 2% decrease in AUROC and AUPRC, indicating its strong ability to generalize to new and unseen drug–target combinations. This confirms the potential applicability of CAMF-DTI in real-world DTI discovery scenarios where novel interactions frequently occur.

3.4.3. Interpretation and Case Study

The proposed model demonstrates sensitivity to DTI and enables the visualization of high-response regions of proteins to drugs using the DrugBank dataset. Specifically, two drug–target protein pairs, DB08804–P10275 and DB12267–P00519, which have been experimentally validated, were selected for analysis. The high-response regions of the target proteins under drug interactions were visualized, with the top 30% of regions in each response dimension recorded as high-response areas. These areas were highlighted in purple, where deeper shades indicated stronger response intensity.
Furthermore, visualization results were generated using CB-Dock2 to compare the predicted high-response regions with molecular docking outcomes. CB-Dock2 is an online platform that predicts binding sites and affinities between proteins and ligands, thereby assisting in computer-aided drug discovery. The visualization of the high-response regions is shown in Figure 3 and Figure 4.
These case studies highlight the model’s interpretability and practical value. CAMF-DTI not only predicts drug–target interactions with high confidence but also reveals biologically meaningful interaction regions. This demonstrates the model’s potential to assist in drug discovery tasks that require both prediction accuracy and molecular-level interpretability.

4. Conclusions

In this study, we proposed CAMF-DTI, a novel framework designed to enhance drug–target interaction (DTI) prediction by addressing several key limitations of existing approaches. To incorporate directional information into spatial features, coordinate attention was introduced to jointly encode spatial positions and sequence directionality, thereby improving the localization of critical interaction regions. A multi-scale fusion module [37] was further developed to overcome the restriction of single-scale receptive fields, enabling the extraction of structural features at both local and global levels. Moreover, a cross-attention mechanism was employed to model dynamic dependencies between drugs and targets [38], facilitating more effective interaction learning.
Extensive experiments on multiple benchmark datasets demonstrated that CAMF-DTI consistently outperformed state-of-the-art baselines across AUROC, AUPRC, Accuracy, F1-score, and MCC, confirming its superior predictive power and strong generalization ability. In terms of computational efficiency, CAMF-DTI introduces only a moderate additional cost compared with existing attention-based models. On the BindingDB dataset, the training time per epoch was about 1.15 times that of DO-GMA, while inference latency increased by less than 5%. This indicates that the added modules enhance representational power without significantly compromising scalability, making the framework suitable for large-scale DTI prediction tasks. In addition, although this study mainly involves structured proteins, CAMF-DTI can also capture meaningful sequence-level patterns from intrinsically disordered proteins (IDPs) due to its sequence-based design. Future work will explore specific adaptations for IDPs to further enhance model generality. By combining predictive accuracy with interpretability, CAMF-DTI provides a promising direction for improving computational drug discovery and offers valuable insights into the mechanisms underlying drug–target interactions.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/cimb47110964/s1.

Author Contributions

Conceptualization, methodology, formal analysis, writing—original draft preparation, J.M.; validation, software, visualization, C.L.; resources, investigation, D.J.; writing—review and editing, supervision, project administration, funding acquisition, J.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported in part by funds from the Ministry of Science and Technology (2022FY101104).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original data and source codes presented in this study are openly available at https://github.com/MiJia-ID/CAMF-DTI (accessed on 12 October 2025). The repository includes the implementation of CAMF-DTI, preprocessing scripts, configuration files, and pretrained model checkpoints, which enable full reproducibility of the experimental results.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Li, M.; Guo, Z.; Wu, Y.; Guo, P.; Shi, Y.; Hu, S.; Wan, W.; Hu, S. ViDTA: Enhanced Drug-Target Affinity Prediction via Virtual Graph Nodes and Attention-based Feature Fusion. In Proceedings of the IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Lisbon, Portugal, 3–6 December 2024. [Google Scholar]
  2. Luo, Z.; Wu, W.; Sun, Q.; Wang, J. Accurate and transferable drug–target interaction prediction with DrugLAMP. Bioinformatics 2024, 40, btae693. [Google Scholar] [CrossRef] [PubMed]
  3. Zhang, Y.; Wang, Y.; Wu, C.; Zhan, L.; Wang, A.; Cheng, C.; Zhao, J.; Zhang, W.; Chen, J.; Li, P. Drug–target interaction prediction by integrating heterogeneous information with mutual attention network. BMC Bioinform. 2024, 25, 361. [Google Scholar] [CrossRef]
  4. Meng, Z.; Meng, Z.; Yuan, K.; Ounis, I. FusionDTI: Fine-grained Binding Discovery with Token-level Fusion for Drug–Target Interaction. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, Suzhou, China, 4–9 November 2025. [Google Scholar] [CrossRef]
  5. Liu, B.; Wu, S.; Wang, J.; Deng, X.; Zhou, A. HiGraphDTI: Hierarchical Graph Representation Learning for Drug–Target Interaction Prediction. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases; Springer: Cham, Switzerland, 2024. [Google Scholar]
  6. Li, C.; Mi, J.; Wang, H.; Liu, Z.; Gao, J.; Wan, J. MGMA-DTI: Drug target interaction prediction using multi-order gated convolution and multi-attention fusion. Comput. Biol. Chem. 2025, 118, 108449. [Google Scholar] [CrossRef]
  7. Honda, S.; Shi, S.; Ueda, H.R. Smiles transformer: Pre-trained molecular fingerprint for low data drug discovery. arXiv 2019, arXiv:1911.04738. [Google Scholar] [CrossRef]
  8. Ozturk, H.; Ozgur, A.; Ozkirimli, E. DeepDTA: Deep drug–target binding affinity prediction. Bioinformatics 2018, 34, i821–i829. [Google Scholar] [CrossRef]
  9. Öztürk, H.; Ozkirimli, E.; Özgür, A. WideDTA: Prediction of Drug-Target Binding Affinity. arXiv 2019, arXiv:1902.04166. [Google Scholar]
  10. Chen, L.; Tan, X.; Wang, D.; Zhong, F.; Liu, X.; Yang, T.; Luo, X.; Chen, K.; Jiang, H.; Zheng, M. TransformerCPI: Improving compound–protein interaction prediction by sequence-based deep learning with self-attention mechanism. Bioinformatics 2020, 36, 4406–4414. [Google Scholar] [CrossRef]
  11. Huang, K.; Xiao, C.; Glass, L.M.; Sun, J. MolTrans: Molecular interaction transformer for drug–target interaction prediction. Bioinformatics 2021, 37, 830–836. [Google Scholar] [CrossRef]
  12. Nguyen, T.; Le, H.; Quinn, T.P.; Nguyen, T.; Le, T.D.; Venkatesh, S. GraphDTA: Predicting Drug–Target Binding Affinity with Graph Neural Networks. Bioinformatics 2021, 37, 1140–1147. [Google Scholar] [CrossRef]
  13. Yang, Z.; Zhong, W.; Zhao, L.; Chen, C.Y.-C. MGraphDTA: Deep Multiscale Graph Neural Network for Explainable Drug–Target Binding Affinity Prediction. Chem. Sci. 2022, 13, 816–833. [Google Scholar] [CrossRef]
  14. Wan, F.; Hong, L.; Xiao, A.; Jiang, T.; Zeng, J. NeoDTI: Neural integration of neighbor information from a heterogeneous network for discovering new drug–target interactions. Bioinformatics 2019, 35, 104–111. [Google Scholar] [CrossRef] [PubMed]
  15. Qu, X.; Du, G.; Hu, J.; Cai, Y. Graph-DTI: A New Model for Drug-Target Interaction Prediction Based on Heterogenous Network Graph Embedding. Curr. Comput.-Aided Drug Des. 2024, 20, 1013–1024. [Google Scholar] [CrossRef]
  16. Stark, A.; Pinsler, R.; Corso, G.; Fout, A.; Zitnik, M. EquiBind: Geometric deep learning for drug binding structure prediction. In Proceedings of the International Conference on Machine Learning (ICML), Baltimore, MD, USA, 17–23 July 2022; pp. 20503–20521. [Google Scholar]
  17. Wang, C.; Liu, Y.; Song, S.; Cao, K.; Liu, X.; Sharma, G.; Guo, M. DHAG-DTA: Dynamic Hierarchical Affinity Graph Model for Drug–Target Binding Affinity Prediction. IEEE Trans. Comput. Biol. Bioinform. 2025, 22, 697–709. [Google Scholar] [CrossRef]
  18. Liu, T.; Lin, Y.; Wen, X.; Jorissen, R.N.; Gilson, M.K. BindingDB: A web-accessible database of experimentally determined protein–ligand binding affinities. Nucleic Acids Res. 2007, 35, D198–D201. [Google Scholar] [CrossRef]
  19. Davis, M.I.; Hunt, J.P.; Herrgard, S.; Ciceri, P.; Wodicka, L.M.; Pallares, G.; Hocker, M.; Treiber, D.K.; Zarrinkar, P.P. Comprehensive analysis of kinase inhibitor selectivity. Nat. Biotechnol. 2011, 29, 1046–1051. [Google Scholar] [CrossRef]
  20. Tang, J.; Szwajda, A.; Shakyawar, S.; Xu, T.; Hintsanen, P.; Wennerberg, K.; Aittokallio, T. Making sense of large-scale kinase inhibitor bioactivity data sets: A comparative and integrative analysis. J. Chem. Inf. Model. 2014, 54, 735–743. [Google Scholar] [CrossRef]
  21. Yao, Q.; Chen, Z.; Cao, Y.; Hu, H. Enhancing Drug-Target Interaction Prediction with Graph Representation Learning and Knowledge-Based Regularization. Front. Bioinform. 2025, 5, 1649337. [Google Scholar] [CrossRef] [PubMed]
  22. Wang, G.; Zhang, X.; Pan, Z.; Rodríguez Patón, A.; Wang, S.; Song, T.; Gu, Y. Multi-TransDTI: Transformer for Drug–Target Interaction Prediction Based on Simple Universal Dictionaries with Multi-View Strategy. Biomolecules 2022, 12, 644. [Google Scholar] [CrossRef]
  23. Geng, G.; Wang, L.; Xu, Y.; Wang, T.; Ma, W.; Duan, H.; Zhang, J.; Mao, A. MGDDI: A multi-scale graph neural networks for drug–drug interaction prediction. Methods 2024, 228, 22–29. [Google Scholar] [CrossRef]
  24. Alon, U.; Yahav, E. On the Bottleneck of Graph Neural Networks and its Practical Implications. arXiv 2020, arXiv:2006.05205. [Google Scholar]
  25. Ouyang, D.; He, S.; Zhang, G.; Luo, M.; Guo, H.; Zhan, J.; Huang, Z. Efficient Multi-Scale Attention Module with Cross-Spatial Learning. In Proceedings of the ICASSP 2023—IEEE International Conference on Acoustics, Speech and Signal Processing, Rhodes Island, Greece, 4–10 June 2023; pp. 1–5. [Google Scholar] [CrossRef]
  26. Hou, Q.; Zheng, M.; Lu, M.; Wu, E. Coordinate Attention for Efficient Mobile Network Design. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13713–13722. [Google Scholar] [CrossRef]
  27. McDermott, M.; Zhang, H.; Hansen, L.; Angelotti, G.; Gallifant, J. A closer look at AUROC and AUPRC under class imbalance. Adv. Neural Inf. Process. Syst. 2024, 37, 44102–44163. [Google Scholar]
  28. Metrics and Scoring: Quantifying the Quality of Predictions. Available online: https://scikit-learn.org/stable/modules/model_evaluation.html (accessed on 12 October 2025).
  29. Chicco, D.; Jurman, G. The advantages of the Matthews correlation coefficient (MCC) over F1 score and accuracy in binary classification evaluation. BMC Genom. 2020, 21, 6. [Google Scholar] [CrossRef]
  30. Tsubaki, M.; Tomii, K.; Sese, J. Compound–Protein Interaction Prediction with End-to-End Learning of Neural Networks for Graphs and Sequences. Bioinformatics 2019, 35, 309–318. [Google Scholar] [CrossRef]
  31. Li, M.; Lu, Z.; Wu, Y.; Li, Y.H. BACPI: A bi-directional attention neural network for compound–protein interaction and binding affinity prediction. Bioinformatics 2022, 38, 1995–2002. [Google Scholar] [CrossRef]
  32. Zhao, M.; Yuan, M.; Yang, Y.; Xu, S.X. CPGL: Prediction of Compound-Protein Interaction by Integrating Graph Attention Network with Long Short-Term Memory Neural Network. IEEE/ACM Trans. Comput. Biol. Bioinform. 2023, 20, 1935–1942. [Google Scholar] [CrossRef]
  33. Peng, L.; Liu, X.; Yang, L.; Liu, L.; Bai, Z.; Chen, M.; Lu, X.; Nie, L. BINDTI: A Bi-Directional Intention Network for Drug–Target Interaction Identification Based on Attention Mechanisms. IEEE J. Biomed. Health Inform. 2025, 29, 1602–1612. [Google Scholar] [CrossRef]
  34. Yin, Z.; Chen, Y.; Hao, Y.; Pandiyan, S.; Shao, J.; Wang, L. FOTF-CPI: A compound-protein interaction prediction transformer based on the fusion of optimal transport fragments. iScience 2023, 27, 108756. [Google Scholar] [CrossRef]
  35. Qian, Y.; Wu, J.; Zhang, Q. CAT-CPI: Combining CNN and transformer to learn compound image features for predicting compound-protein interactions. Front. Mol. Biosci. 2022, 9, 963912. [Google Scholar] [CrossRef] [PubMed]
  36. Peng, L.; Mao, J.; Huang, G.; Han, G.; Liu, X.; Liao, W.; Tian, G.; Yang, J. DO-GMA: An End-to-End Drug–Target Interaction Identification Framework with a Depthwise Overparameterized Convolutional Network and the Gated Multihead Attention Mechanism. J. Chem. Inf. Model. 2025, 65, 1318–1337. [Google Scholar] [CrossRef] [PubMed]
  37. Rahman, M.M.; Watton, P.N.; Neu, C.P.; Pierce, D.M. A chemo-mechano-biological modeling framework for cartilage evolving in health, disease, injury, and treatment. Comput. Methods Programs Biomed. 2023, 231, 107419. [Google Scholar] [CrossRef] [PubMed]
  38. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Long Beach, CA, USA, 4–9 December 2017; Volume 30, pp. 5998–6008. [Google Scholar]
Figure 1. CAMF-DTI framework diagram. (a) CAMF-DTI Framework Overview; (b) Multi-Feature Fusion Module; (c) Coordinate Attention Model; (d) Cross-Attention Module.
Figure 1. CAMF-DTI framework diagram. (a) CAMF-DTI Framework Overview; (b) Multi-Feature Fusion Module; (c) Coordinate Attention Model; (d) Cross-Attention Module.
Cimb 47 00964 g001
Figure 2. Ablation experiment results on four benchmark datasets. (A) BindDB; (B) BioSNAP; (C) Humans; (D) C.elegans.
Figure 2. Ablation experiment results on four benchmark datasets. (A) BindDB; (B) BioSNAP; (C) Humans; (D) C.elegans.
Cimb 47 00964 g002
Figure 3. Docking results and high response areas between drug DB08804 and protein P10275.
Figure 3. Docking results and high response areas between drug DB08804 and protein P10275.
Cimb 47 00964 g003
Figure 4. Docking results and high response areas between drug DB12267 and protein P00519.
Figure 4. Docking results and high response areas between drug DB12267 and protein P00519.
Cimb 47 00964 g004
Table 1. Statistics of the four benchmark datasets.
Table 1. Statistics of the four benchmark datasets.
DatasetDrugProteinInteractionPositiveNegative
binding DB14,643262349,19920,67428,525
BioSNAP4510218127,46413,83013,634
C.elegans27262001627833643364
Human27262001627833643364
Table 2. Performance comparison of CAMF-DTI with other baselines on BindingDB, BioSNAP, C.elegans and Human datasets (means rounded to three decimal places).
Table 2. Performance comparison of CAMF-DTI with other baselines on BindingDB, BioSNAP, C.elegans and Human datasets (means rounded to three decimal places).
DatasetBaselineAUROCAUPRCAccuracyF1-ScoreMCC
binding DBCPI-GNN0.5590.4700.5610.3360.053
BACPI0.9540.9410.8920.8710.779
GPGL0.9340.9140.8690.8440.731
bindTI0.9600.9460.9070.9090.812
FOTF-CPI0.9530.9370.8940.8730.782
CAT-DTI0.9600.9470.9000.8840.797
DO-GMA0.9620.9500.9010.9050.801
CAMF-DTI0.9630.9520.9080.8990.803
BioSNAPCPI-GNN0.7200.7190.6540.6520.309
BACPI0.8880.8950.8080.8100.616
GPGL0.8860.8930.8130.8130.621
bindTI0.9030.9030.8320.8400.673
FOTF-CPI0.8990.9020.8270.8290.654
CAT-DTI0.9020.9070.8360.8350.664
DO-GMA0.9230.9260.8510.8540.704
CAMF-DTI0.9150.9120.8590.8570.706
C.elegansCPI-GNN0.9860.9860.9490.9180.829
BACPI0.9860.9860.9490.9480.933
GPGL0.9860.9860.9280.9280.853
bindTI0.9820.9830.9660.9660.932
FOTF-CPI0.9900.9900.9660.9660.932
CAT-DTI0.9830.9860.9670.9640.932
DO-GMA0.9930.9930.9740.9730.948
CAMF-DTI0.9870.9840.9480.9490.895
HumanCPI-GNN0.9670.9660.9070.9060.834
BACPI0.9670.9670.9050.9070.835
GPGL0.9680.9670.9020.9040.832
bindTI0.9810.9760.9400.9380.879
FOTF-CPI0.9830.9800.9410.9320.881
CAT-DTI0.9820.9690.9420.9440.886
DO-GMA0.9860.9840.9500.9510.900
CAMF-DTI0.9870.9840.9480.9490.895
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mi, J.; Li, C.; Jiang, D.; Wan, J. CAMF-DTI: Enhancing Drug–Target Interaction Prediction via Coordinate Attention and Multi-Scale Feature Fusion. Curr. Issues Mol. Biol. 2025, 47, 964. https://doi.org/10.3390/cimb47110964

AMA Style

Mi J, Li C, Jiang D, Wan J. CAMF-DTI: Enhancing Drug–Target Interaction Prediction via Coordinate Attention and Multi-Scale Feature Fusion. Current Issues in Molecular Biology. 2025; 47(11):964. https://doi.org/10.3390/cimb47110964

Chicago/Turabian Style

Mi, Jia, Chang Li, Daguang Jiang, and Jing Wan. 2025. "CAMF-DTI: Enhancing Drug–Target Interaction Prediction via Coordinate Attention and Multi-Scale Feature Fusion" Current Issues in Molecular Biology 47, no. 11: 964. https://doi.org/10.3390/cimb47110964

APA Style

Mi, J., Li, C., Jiang, D., & Wan, J. (2025). CAMF-DTI: Enhancing Drug–Target Interaction Prediction via Coordinate Attention and Multi-Scale Feature Fusion. Current Issues in Molecular Biology, 47(11), 964. https://doi.org/10.3390/cimb47110964

Article Metrics

Back to TopTop