Next Article in Journal
Endothelin and the Cardiovascular System: The Long Journey and Where We Are Going
Next Article in Special Issue
A Novel Ensemble Learning-Based Computational Method to Predict Protein-Protein Interactions from Protein Primary Sequences
Previous Article in Journal
Cancer-Associated Microbiota: From Mechanisms of Disease Causation to Microbiota-Centric Anti-Cancer Approaches
Previous Article in Special Issue
RoFDT: Identification of Drug–Target Interactions from Protein Sequence and Drug Molecular Structure Using Rotation Forest
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

BioChemDDI: Predicting Drug–Drug Interactions by Fusing Biochemical and Structural Information through a Self-Attention Mechanism

1
School of Information Engineering, Xijing University, Xi’an 710123, China
2
College of Grassland and Environment Sciences, Xinjiang Agricultural University, Urumqi 830052, China
3
School of Computer Science, Northwestern Polytechnical University, Xi’an 710129, China
*
Authors to whom correspondence should be addressed.
Biology 2022, 11(5), 758; https://doi.org/10.3390/biology11050758
Submission received: 12 April 2022 / Revised: 12 May 2022 / Accepted: 13 May 2022 / Published: 16 May 2022
(This article belongs to the Special Issue Intelligent Computing in Biology and Medicine)

Abstract

:

Simple Summary

Throughout history, combining drugs has been a common method in the fight against complex diseases. However, potential drug–drug interactions could give rise to unknown toxicity issues, which requires the urgent proposal of efficient methods to identify potential interactions.We use computer technology and machine learning techniques to propose a novel computational framework to calculate scores of drug–drug interaction probability for simplifying the screening process. Additionally, we built an online prescreening tool for biological researchers to further verify possible interactions in the fields of biomedicine and pharmacology. Overall, our study can provide new insights and approaches for rapidly identifying potential drug–drug interactions.

Abstract

During the development of drug and clinical applications, due to the co-administration of different drugs that have a high risk of interfering with each other’s mechanisms of action, correctly identifying potential drug–drug interactions (DDIs) is important to avoid a reduction in drug therapeutic activities and serious injuries to the organism. Therefore, to explore potential DDIs, we develop a computational method of integrating multi-level information. Firstly, the information of chemical sequence is fully captured by the Natural Language Processing (NLP) algorithm, and multiple biological function similarity information is fused by Similarity Network Fusion (SNF). Secondly, we extract deep network structure information through Hierarchical Representation Learning for Networks (HARP). Then, a highly representative comprehensive feature descriptor is constructed through the self-attention module that efficiently integrates biochemical and network features. Finally, a deep neural network (DNN) is employed to generate the prediction results. Contrasted with the previous supervision model, BioChemDDI innovatively introduced graph collapse for extracting a network structure and utilized the biochemical information during the pre-training process. The prediction results of the benchmark dataset indicate that BioChemDDI outperforms other existing models. Moreover, the case studies related to three cancer diseases, including breast cancer, hepatocellular carcinoma and malignancies, were analyzed using BioChemDDI. As a result, 24, 18 and 20 out of the top 30 predicted cancer-related drugs were confirmed by the databases. These experimental results demonstrate that BioChemDDI is a useful model to predict DDIs and can provide reliable candidates for biological experiments. The web server of BioChemDDI predictor is freely available to conduct further studies.

1. Introduction

Drugs play a crucial role in curing diseases and enhancing quality of life [1]. During drug development, drug–drug interactions (DDIs) are a critical consideration and the drug targeting of the selected protein should be bioavailable (e.g., favorable absorption and metabolism) [2]. However, potential DDIs may lead to a strong rise or drop in the plasma concentration of the drug or metabolite, and even generate toxic compounds [3]. From a clinical perspective, a combination of drugs is used for the treatment of complex diseases, but unexpected DDIs may induce adverse reactions, which can give rise to drug withdrawal and even to the death of the patient [4,5]. Thus, the early identification of potential DDIs is very critical for drug development and medical safety.
Alterations in drug pharmacokinetics and drug pharmacodynamics can be caused by DDIs [6]. Pharmacokinetic DDIs crop up when the perpetrator drug disrupts the absorption, distribution, metabolism and elimination (ADME) of the victim drug, and also when the perpetrator drug interacts with the protein of the victim drug or other protein within the same signaling pathway [4]. To screen and analyze unknown DDIs, biological techniques are mainly used, which are regarded as the ultimate way to judge and validate the DDIs, containing metabolism-based and transporter-based DDIs, such as testing whether the drug is the inhibitor/inducer or substrate of CYP enzymes [7], and testing whether the drug is the inhibitor/inducer or substrate of a P-gp transporter [3]. Then, based on the in vitro parameters, a cumbersome dynamic model (e.g., PBPK) and expensive in vivo experiment should be constructed and analyzed for the final validation.
Traditional biological technologies face the challenges of high cost, limited participant number, low efficiency and large number of pairwise drugs waiting for identification [2]. Additionally, the constantly increasing demand for drug therapy makes the identification of potential DDIs before clinical medications are administered [8] more and more urgent. Consequently, exploiting large-scale computational prediction methods as a decision aids for a large number of DDI candidates prescreening to provide a direction or prioritize of the in vitro–in vivo experiments, and improve efficiency for DDI research and development (R&D). Computational methods have gained concern from the academy and the industry, due to their promise to discover drug–drug interactions on the large scale [9,10]. The performance of computers has been greatly improved as the precondition to consider computational methods to realize DDI prediction. Furthermore, researchers constructed many reliable bioinformatic databases on drugs through many biomedical experiments, such as DrugBank [11], ChEBI [12], PubChem [13] and KEGG [14]. Recently, several computational methods have been proposed to reduce the cost of predicting potential DDIs. These methods can be roughly divided into four categories: similarity-based, network-based, matrix-factorization-based and semanticity-based.
Similarity-based methods are one of the relatively main approaches, assuming that, if drugs have similar function structures, they are more likely to have a similar interaction structure. Gottlieb et al. put forward a model named INDI, extracting feature vectors by calculating seven drug similarities and predicting interactions of the drugs by logistic regression [15]. Cheng et al. merged many drug similarities to express drug–drug pairs and exploited five classifiers to build predicting models [16]. Ferdousi et al. provided a method to construct embedding vectors of drugs, using four biological elements, including carriers, transporters, enzymes and targets (CTET), to predict potential DDIs through a Russell–Rao similarity [17]. Rohani et al. predicted DDIs based on fusing similarity matrices and tested the performance on three different scales of drug similarity datasets [18]. Due to the structure information of the interaction network and chemical sequence information not being considered, the final prediction effect was not very good.
Network-based methods infer novel drug–drug edges by the topological structure of the network and biological network that involves biomedical entities, or learning the high-order drug similarity and the propagating similarity. Zhang et al. predicted DDIs using an integrative label propagation method with high-order similarity transitivity on the multi-scale similarity-based network. It also can rank multiple drug information sources [19]. Park et al. exploited the random walk with restarting on the protein–protein network method to the analog propagation of signals to predict drug–drug interactions [20]. Deepika et al. proposed a meta-learning framework, extracting representation on four types of feature networks through Node2vec and containing chemical feature networks, biological feature networks, phenotypic feature networks and disease feature networks, then integrating the results of four classifiers to predict unknown DDIs [21]. Liu et al. drew an approach DANN-DDI that contains five types of drug networks and learned the features through the graph embedding algorithm SDNE. An attention neural network is designed to learn concatenating representation and a deep neural network is used to generate prediction results [22]. Although these methods have shown a good prediction performance, most of them preserve higher-order structure features with difficulty and stick to the local optimum with ease. The attention mechanism is not used to process multiple pieces of information, which leads to a limited performance.
Matrix-factorization-based methods turn the DDI adjacency matrix into several decomposed matrices and then re-establish adjacency by the decompositions. The model, named TMFUF, was developed by Shi et al. to identify DDIs, which uses drug additional information to rebuild the interaction matrix through triple matrix factorization [23]. Zhang et al. proposed an ensemble model that is based on sparse feature learning for predicting drug–drug interactions [24]. Zhang et al. designed a method for DDIs prediction that uses eight types of background information based on the matrix factorization of a manifold regularization [25]. Yu et al. proposed DDINMF applying semi-nonnegative matrix factorization to conclude the enhanced and degressive prediction of pairwise drugs [26]. Shi et al. introduced a BRSNMF model, an optimization of the DDINMF model, technically utilizing drug-binding protein as the feature to predict DDIs of new drugs [27]. These approaches exploit the biological information of many supplements to ensure generalization, but the important information on the chemical sequence is not fully considered.
Semanticity-based methods generally abstract the information from the semanticity of sentences about drugs through text-mining, and then the candidates of drug interactions are detected and classified. Chowdhury et al. applied a framework that can extract information on multi-phase relations, exploiting the scope of negation cues and semantic roles to reduce the skewness of the data, and used SVM to calculate the possibility of DDIs [28]. Zhu et al. exploited the BioBERT method to pre-train word vectors of drug descriptions and extracted the semantic representation of sentences by the BiGRU, integrating drug entity information by entity-aware attention, obtaining prediction results by the MLP [29]. However, these methods are heavily dependent on the clinical evidence in the post-market, which means that there is no capability of providing potential DDI alerts before clinical medications are administered.
Although the aforementioned methods have their own advantages and play a crucial role in computational method development for drug–drug interaction prediction, there are still some limitations. (i) Due to several existing relevant computational methods only focusing on single information of drugs, they still cannot satisfy the demand for prediction accuracy in reality applications. (ii) Furthermore, most of them rely on artificially designed molecular representation, limited by the knowledge of domain experts. (iii) Ignoring deep network structure information and integrating drug features without the attention machine could lead to limited prediction performance.
In in silico research, data availability and accessibility are crucial factors in determining the accuracy and precision of calculational methods [30]. In this paper, in order to address the existing deficiencies, we propose a novel framework for fusing drug chemical sequence, drug biological function similarity and deep network topology structure with an attention machine (BioChemDDI) to predict potential DDIs. In particular, we first obtain the drug biochemical features regarding chemical sequence information and biological function information through a word-embedding algorithm and the similarity matrix fusing method, respectively. Notably, the chemical sequence feature of each drug is firstly represented as a matrix, whose dimension can be reduced by a Convolutional Neural Network (CNN). Then, the rich structural information of the interaction network is extracted by efficiently graph embedding with graph collapse. Thirdly, we introduce the attention mechanism to fuse multiple features and enhance the feature of each drug node. Finally, the interaction scores are generated by the fully connected layer. The proposed model is trained end-to-end, and each feature vector can be further screened for the characteristic factors through the hidden layers. Additionally, in this work, four datasets are used to verify the robustness of our model and are compared with state-of-the-art methods to demonstrate the high efficiency of BioChemDDI. Then, the results of the five-fold cross-validation and the case studies further investigate whether our model is suitable for interaction studies between drugs. More meaningfully, case studies related to three cancer diseases can give insight to reveal unknown drug–drug interactions for inferring combinations of the drug in treating complex diseases. The BioChemDDI model can provide accurate predictions for the potential DDIs and is anticipated to serve as the prioritization tool for the development of drug and clinical applications, which can be used as pre-screening tools for potential DDIs. The computational platform web server is accessible at: http://120.77.11.78/BioChemDDI/ (accessed on 11 April 2022).

2. Materials and Methods

2.1. Dataset Description

In this study, to demonstrate the robustness of BioChemDDI, we used four datasets called DS1, DS2, D-DS3 and E-DS3. The DS1 was processed by Ren et al. [31], and its biochemical information was downloaded from DrugBank [11], containing one chemical sequence information and four types of biological function information, which are expressed by drug-related receptor-types of carriers, enzymes, targets and transporters. It also includes 1940 drug nodes, and the 219,247 known pairwise DDIs have been proved [32]. Zhang et al. [33] collected DS2 from the databases of DrugBank [11], SIDER [34], KEGG [35], PubChem [13] and OFFSIDES [36]. Additionally, the D-DS3 and the E-DS3 were processed by Shi et al. [27], which, respectively, contain 1562 drugs with 55,278 depressive interactions and 125,298 enhancive interactions and 4 types of biological functions, the tsame as the DS1. Table 1 illustrates his information in more detail.
Noteworthily, due to the use of testing datasets from different sources, high unbalanced data regarding biological function information is contained, which can be addressed by the biological function fusing module. Known drug–drug pairs were considered positive samples and the negative samples can be presented by other drug–drug pairs in the network. To avoid bias from the imbalance of positive and negative samples, the negative samples were randomly chosen to be equivalent to the positive samples. We used a 5-fold cross-validation and split train–test on our datasets to proceed with a meaningful contrast. After disrupting the samples, 70% of them were seen as training datasets. We treated 20% of them as test datasets, and regarded the other 10% as validation datasets. Inspired by Feng et al. [37], different scales of datasets were used to validate the superior performance of BioChemDDI. Because DS2 is often utilized to evaluate the ability of prediction, we also used DS2 as the primary comparison experimental dataset to save time and costs. DS1, D-DS3 and E-DS3 were selected to evaluate the stability of BioChemDDI.

2.2. Overview of the Methods

In this paper, we propose a learning framework, BioChemDDI, to efficiently predict potential DDIs. As Figure 1 illustrates, BioChemDDI consists of four modules comprising a drug chemical sequence information learning module, drug–drug network structural information learning module, biological function information learning module and attentional multi-feature integration module. At the beginning, the chemical sequence was indicated by learning their corresponding SMILES through the NLP method of CBOW. Then, we represented each type of biological function as the adjacency matrix and calculated the drug similarity matrix under different function types. After that, all of the similarity matrixes were fused efficiently by SNF. Before Hierarchical Representation Learning for Networks (HARP) [38] was used to extract drug network structural information, we constructed the DDI network from the positive samples in the train set. Furthermore, the attentional integration mechanism integrates biochemical features and network structural features to finally predict potential DDIs. Additionally, we used the binary cross-entropy loss and error backpropagation (BP) feedforward network to implement end-to-end training. A detailed description is given below, according to the order of the function module in the flowchart.

2.3. Drug Chemical Sequence Information

With deep learning technology springing up, the methods of word embedding are utilized to raise downstream task performance. In order to make full use of the chemical sequence of drugs, researchers regard SMILES as sentences and treat each symbol of a chemical element or atom as a word. Chemical sequence learning consists in transforming each symbol into a digital vector, thereby each drug sequence can be represented as a matrix. Based on the “natural biological language”, chemical sequence information is learned by word2vec [39], which is a method for learning distributed vector representation of words through a large corpus. CBOW is a word2vec model, which calculates the probability of a certain center word according to the context of the word. It contains an input layer, a hidden layer and an output layer. The embedding vectors of each context word are calculated by the weight matrix ω . Additionally, the arithmetic average of embedding vectors h of the center word is calculated by the hidden layer, as follows:
h = 1 c ω ( x 1 + x 2 + + x c )
where x C is the one-hot vector representation of C-th context words, which are the V dimension. The h is the average of the embedding vectors of context words. By the weight matrix ω of the output layer, the co-occurrence probabilities are calculated, and each word is the center word to calculate the embedding vector. Moreover, the probability of the actual central word occurring is maximized by the function:
E = log P ( ω O | ω I 1 , ω I 2 , , ω I C ) = log j = 1 V exp ( v ω j × h ) v ω o × h
where v ω j means the j-th row of the weight matrix ω . In this paper, SMILES was seen as the drug structure corpus to make use of the word embedding method. The Python package of gensim was used to obtain the drug feature matrix. Then, in addition to the parameters of “size” being set to 64 and “min_count” being set to 1, the parameter values were default. So, as shown in Figure 2, each symbol is shown as a 64-dimension vector, thereby obtaining the drug feature matrix.

2.4. Network Structural Information

The network structural information is a crucial character for potential interaction analysis and prediction. Under the entire drug network, each node can be represented by the structural relationship of nodes. In this subsection, we focus on the task of extracting the embedding vector-containing node structural information. The node structure is more similar, and the embedding vectors of the nodes are closer to each other. To fully utilize the network relation of DDIs and to globally express the straight or potential information flow between nodes, we formulated graph embedding representation by a method named HARP, proposed by Chen et al. [38]. Recently, for large-scale networks, some existing learning algorithms of network representation demand complex computational complexity and many of them are local embedding methods. Relationships of long-distance global network may be ignored by focusing on local embedding configurations, and the learning representations are incapable of revealing vital global patterns of distribution [38]. HARP-LINE, improved from LINE, is suitable for embedding a large global information network into a low-dimensional space of vectors and can catch the deep structure of the interaction network. HARP-LINE consists of three parts: graph coarsening, graph embedding and representation refining, as Figure 3 shows. For a given graph G = ( V , E ) , finding a mapping function Φ : V R | V | × d , d | V | is the first important task of graph representation.
In order to gain a better global embedding, HARP firstly performs a graph coarsening operation, and then a small subgraph G S = ( V S , E S ) is generated, in which | V S | | V | and | E S | | E | . In this way, because of the smaller subgraph containing more coarse-grained information, embedding learning on G S can gain a wider global structure, although the learning method is based on a partial structure. Therefore, as the original graph structure collapses constantly until the number of nodes reaches a certain threshold, LINE [40] is exploited to learn representations in the part of graph embedding. LINE defines the first-order proximity and second-order proximity to indicate, respectively, the nodes’ direct relationship and the relationship of nodes that are not direct, which have a common neighbor node. If the nodes share the same neighbor, they are proximal to each other. The co-occurrence probability of vertexes, v j and v i , keeping a similar context can be defined as:
p 2 ( v j | v i ) = exp ( u j × u i ) k = 1 | V | exp ( u k × u i )
where u i and u j are, respectively, the vector representation of the vertex of v i and its context v j , and | V | indicates the quantity of context vertex. To assure proximity, the following objective function should be minimized:
m i n i m i z e O 2 = i V λ i d ( p 2 ¯ ( | v i ) , p 2 ( | v j ) )
The function of d ( , ) is used to calculate the distance of distribution, λ i , viewed as the importance of v i ; the empirical distribution probability p 2 ¯ ( | v i ) is designed as:
p 2 ¯ ( v j | v i ) = w i j d e g r e e i
where w i j is the weight between v j and v i , and d e g r e e i stands for the degree of the vertex v i . KL divergence is utilized to measure the similarity of distributions. For the sake of simplicity, λ i was set in this paper as the degree of v i , i.e., λ i = d e g r e e i and certain constants were left out. The loss function can be reduced as follows:
m i n i m i z e O 2 = ( i , j ) V w i j log ( p 2 ( v j | v i ) )
After obtaining the subgraph nodes embedding vectors, using them to initialize the node of the upper-level subgraph can finely collect global information of the coarse graph. Finally, improved embedding vectors can be obtained after optimizing coarse graph embedding by upward iteration. We used the default parameters of the HARP Python package. As all graph embedding methods have a common drawback, that is, that extremely few nodes of the test set do not exist in the train set, these nodes cannot be embedded by graph embedding methods. The biochemical features can increase the information content of the feature to some extent and reduce the impact of the missing graph structural information on the model performance. On the one hand, in reality, many new drugs may have no interaction with known drugs, which can be regarded as no embedded nodes in the network. On the other hand, a few drugs missing some information can also indicate the generalization of our model. In this work, the no embedded nodes were given priority to use the embedding characteristics of their nearest neighbor as their embedding characteristics; otherwise, their embedding was set to a zero vector as their embedding characteristic.

2.5. Biological Function Information

The biological function of drugs is another important feature that can be represented by drug-related receptors. All of the biological function information can be fused by the multiple calculating similarity matrixes. In this work, M types of receptor bound to drugs are used to establish the adjacency matrixes with different dimension spaces. In each matrix, the values 1 or 0 indicate the existence or the inexistence of the drug-receptor relation, respectively, and the dimension is determined by the number of receptors. For example, there are 1660 receptors of the target type, so a drug can be embedded as a 1660-dimensional initial vector. Given a drug D 1 and a drug D 2 , their initial embedding vectors are d 1 and d 2 , and the similarity between them is based on the Euclidean distance and satisfies the formula as follows:
S ( D 1 , D 2 ) = i = 1 n ( d 1 i d 2 i ) 2
where d 1 i and d 2 i are, respectively, the element representation of the vectors d 1 and d 2 , and n means the number of dimensions. Therefore, the weak similarity matrices among these initial vectors are calculated in each function space type. Then, it is particularly important to integrate M weak similarity matrices into a strong similarity matrix. Similarity Network Fusion (SNF), proposed by Wang et al. [41], is a competent approach to fusing massive biological context similarities [18]. SNF is a non-linear similarity fusion method, combining multi-similarity matrices into a single integrated similarity matrix, which carries all appropriate representation information. Eventually, SNF is applied to fuse M similarity N × N matrix into one N × N matrix, where N is the number of drugs. Figure 4 reveals the details of this process.

2.6. Attentional Multi-Feature Integration and Prediction

To integrate the chemical sequence feature, the network structural distribution feature and the biological function feature for interaction prediction, the attentional multi-feature integration module is proposed. The attention [42] module can fuse multi-level features by using attention scores to reflect adaptive weights, namely, fusing each feature according to a score of feature significance level. Given the initial representation vectors θ and the attention enhance vectors θ ˜ can be obtained as the following function:
θ ˜ = A t t e n t i o n ( Q , K , V ) = s o f t m a x ( Q K d k ) V
where Q , K and V stand on behalf of the product of input vector and three matrices. The d k represents a normalization coefficient. Then, after calculating by a hidden layer, the classification result can be captured.
For the further screening of the efficient characteristic factors, before the three features were input into the attention module, a series of dimension-reducing processes were carried out. Firstly, we used CNN to reduce the feature matrix dimension of the chemical sequence, which contains 64 kernels. Inspired by NIN [43], we used a two-dimensional global average pooling, instead of the flattening operation. Then, the network structural feature vector was put into a fully connected layer containing 64 neurons, and the biological function features passed through a hidden layer of 300 neurons and a hidden layer with 64 neurons. Ultimately, after the elements’ weight of feature vectors were redistributed over attention, all of the processed features were put into the attention module. During the end-to-end training process, the Adam optimizer was used to adjust the learning rate automatically from 1 × 10−3. To avoid overfitting, we adopted the dropout rate of 0.3 on the hidden layers.

3. Experimental Results and Discussion

3.1. Evaluation Criteria

In this paper, classifying each pair of drugs as interaction or non-interaction, we utilized metrics that are frequently used in classification to assess the effectiveness and robustness of our model through distinct perspectives, including five parameters: Accuracy (Acc), Sensitivity (Sen), Precision (Prec), F-measure (F1) and Matthews’s Correlation Coefficient (MCC). These evaluation parameters are represented as follows:
A c c = T P + T N T N + T P + F N + F P  
P r e c = T P T P + F P  
S e n = T P T P + F N  
F 1 = 2 × P r e c × S e n P r e c + S e n  
M C C = T P × T N F P × F N ( T P + F P ) × ( T N + F N ) × ( T N + F P ) × ( T P + F N )  
where true positive, false negative, true negative and false positive are, respectively, are represented by TP, FN, TN and FP.
Meanwhile, according to these parameters, we evaluated the performance of the proposed model by constructing an AUC, the area under the receiver operating characteristic (ROC) curves, and AUPR, which is the area under the precision-recall (PR) curve. AUPR represents an appropriate criterion if the count of negative and positive samples is not the same, just as suggested by [44]. A larger AUC shows a better predictor. We also exploited a five-fold cross-validation and the mean value to assess the performance of methods.

3.2. Assessment of Prediction Ability

In this experiment, to demonstrate the superior performance and robustness of BioChemDDI, we employed it on the four datasets. The jack-knife and q-fold cross-validation (CV) tests [45] are often used to examine whether the predictor is effective or not. We adopted the five-fold cross-validation to enhance the persuasion of the results, which are displayed in Table 2. The average values of AUC and AUPR can reach 0.9711 and 0.9701 or more in all four datasets. Specifically, on DS1, reaching the highest score of 0.9927 and 0.9921. In addition, the average scores of other evaluation criteria are also high; thus, the MCC of the four datasets gained the scores of 92.32, 81.48, 91.08 and 92.05. It can be concluded that the data scale produces an impact on model performance and results, but the worse performance, among the four datasets, also reached 90.72%, 89.54%, 92.21%, 90.86% and 81.48% on the evaluation criteria.
It is obvious that the proposed model has an excellent capability to identify positive and negative samples and predict novel DDIs. Furthermore, the higher AUC and AUPR show that our framework has a superior predictive performance, and the stability and the robustness of the proposed framework can be indicated by the lower variance of the results. As all of the evaluation results show, the performance can improve as the number of data increase. Four ROC curves and four PR curves were drawn as shown in Figure 5 and Figure 6. Finally, it can be concluded that the framework is effective and robust in drug–drug interaction prediction.

3.3. Ablation Experiments

A set of ablation experiments were conducted to validate the contribution levels of the biochemical and network structural information on DS1. The strategy of the training and the evaluation was similar to the procession of model training in Section 3.1. For convenience, we set the experimental number for each ablation experiment and Table 3 shows the results. For a fair comparison of each contribution level of feature, the attention machine should not be used, that is, comparing between (a), (b), (c) and (d). Without chemical sequence level, the values of AUC and ACPR declined by 0.0028 and 0.0031, respectively. Without network structure information, the performance decreased by 0.0135 and 0.0157 in terms of AUC and AUPR, respectively. Without biological function information, the AUC and AUPR were 0.0015 and 0.0019 lower than BioChemDDI, respectively. The experimental results show that the network structural information performs the most significant contribution. One of the possible reasons for this is that the representation pattern between known correlations can be viewed as direct features. Additionally, on the system constituted by abundant direct features, these features are more important than the other correlation features. Due to the important role of the graph structural feature in the proposed model, we discuss the related impact factors in Section 3.4 and Section 3.5.
Apart from this, we performed ablation studies to investigate the contributions of attentional integration at different features. As Table 3 illustrates, respectively comparing (a) with (A), (b) with (B), (c) with (C) and (d) with (D), the performance of the prediction dropped by 0.0073 and 0.0087, 0.0077 and 0.0095, 0.0024 and 0.0031, 0.0057 and 0.0066, respectively, in terms of AUC and ACPR, without attention redistribution. The ablation experiments show that the model can reach optimum performance only when all the information and attentional integration are used.

3.4. Influence of Graph Embedding Methods

The graph embedding method of HARP-LINE based on graph collapse was used in our experiment to extract information on the network structure. In order to test the impact of the graph embedding method on the performance of the model, we tested four embedding methods containing LINE [40], Deepwalk [46], HARP-Deepwalk and Laplace Eigenmaps [47] on the four datasets. All the parameters of each embedding method were default, and the embedding dimension was set to 64, the same as in the case of HARP-LINE. To reduce the influence of biochemical features on the comparison experiment, the attention module is not available. From Figure 7, we can see that, although there are a few fluctuations in the evaluation criteria, from the overall trend, HARP-LINE obtained a better performance. That is attributed to HARP-LINE being able to obtain deep network structural information by focusing on long-distance embedding configurations.

3.5. Influence of the Embedding Dimension of the Graph Structural Feature

Embedding dimension has a crucial influence on the learning ability of graph structural information, in which a large value could cause inefficiency for training, while a small value could limit the learning ability. When generating the feature vectors of the network structure, we set the embedding dimension to 64 to obtain the rich topology information. To test the influence of the graph embedding dimension on the model performance, a comparative experiment was implemented. In this section, we analyzed the performance in different embedding dimensions containing 16, 32, 64 and 128. As Figure 8 shows, when the dimension is set to 64, there is a significant promotion of all the evaluation indicators on DS1 and D-DS3. Additionally, the competitive results were also obtained on the other datasets. Thus, we set the dimension of graph embedding as 64.

3.6. Comparison with Various Classifier Models

In the BioChemDDI model, we identified the potential DDIs through the feedforward neural network with all feature information. To verify the influence of the attentional neural network on the model performance, we compared the effects of various classifiers. Specifically, we kept the feature descriptors unchanged andused Gaussian NB (GNB), Decision Tree (DT) and Logistic Regression (LR) to implement the prediction task. For intuitive comparison, we displayed the experimental results in the form of a histogram in Figure 9. It can be easily seen that BioChemDDI shows a clear predominance on all evaluation criteria. These results demonstrate that the performance of our method can be enhanced with an attentional neural network.

3.7. Comparison with Other Methods

In order to comprehensively evaluate the performance of our proposed model, seven computational methods were chosen as baseline to make the comparison. As mentioned above, we utilized DS2, usually used to test model performance, as the comparison dataset. We compared models containing NDD [18], ISCMF [48], DPDDI [37], GCN-BMP [49], AttentionDDI [50], BioDKG-DDI [31] and an ensemble model [33]. These methods use the dataset of DS2 and five-fold cross-validation. We selected all of the reported evaluation criteria of each comparison method in this paper. The comparison results of the evaluation measure are shown in Table 4, which shows that BioChemDDI reaches the best performance with the outstanding improvement of 0.0043~0.2521, 0.0083~0.1856, 0.0174, 0.0043~0.0721, 0.0299~0.1631 regarding Sen, F1, MCC, AUC and AUPR. The possible reason is that the BioChemDDI model can fully learn the network structural information and extract more effective information through the attention mechanism integrating biochemical information. Nevertheless, in terms of Acc and Prec, our model was lower than the Ensemble Model and ISCMF. On the one hand, the Ensemble Model has an ensemble classifier, which obtains accurate classification results by classifier voting. On the other hand, ISCMF obtained a higher value of Prec, but a lower value of F1, because it only ensures the ability of predicting positive samples. On the whole, although BioChemDDI is a powerful method, there is still room for further improvement.

3.8. Case Studies: Cyclophosphamide, Regorafenib and Allopurinol

To further estimate the ability of BioChemDDI to predict novel DDIs, we conducted case studies on the drugs cyclophosphamide, regorafenib and allopurinol, which are associated with three cancers: breast cancer, hepatocellular carcinoma and malignancies. In the experiment, all 219,247 known drug–drug pairs, among the 1940 drugs, were used to train the prediction model, and all the remaining pairs related to these three test drugs were used as the test set to predict the existence of DDIs; in other words, the test pairs were not included in the train set. The positive samples in the test set were unvisualizable to our model and we predicted the unvisualizable DDIs, which can be seen as unknown DDI, through learning visualizable samples. Finally, the prediction of DDIs can be validated by these positive samples not known beforehand. In particular, we used BioChemDDI to identify all unknown interactions among the three drugs and ranked them according to the prediction score and then searched the DrugBank to verify our results. The top 30 ranked prediction results for each pair of cancer-related drugs are shown in Table 5, Table 6 and Table 7.
The results of the experiment show that 24, 18 and 20 interactions were confirmed by DrugBank with the top five predictions confirmed, except cyclophosphamide-amoxicillin. Unconfirmed DDIs should be further verified through the wet lab. Especially, the drug amoxicillin may potentially interact with Cyclophosphamide with high confidence. To further illustrate the potential interaction mechanism between cyclophosphamide and amoxicillin, we consulted their relevant pharmacokinetic information in DrugBank. We found that cyclophosphamide is the substrate and the inducer of the Cytochrome P450 2C8 enzyme, and meanwhile, amoxicillin is the inhibitor of the Cytochrome P450 2C8 enzyme, which means amoxicillin may impact the metabolism process of cyclophosphamide. To clearly observe the results of the identification of potential DDIs for these three cancer-related drugs, we visualized the newly discovered and known interactions of the top 30 in Figure 10. In addition, the full prediction scores in each case study are reported in Supplementary Materials Table S1.

4. Conclusions

During the development of drugs and their clinical application, combining drugs to treat complex diseases may induce adverse reactions, which makes the identification of potential DDIs before clinical medications are administered more and more urgent. In this paper, we studied how to exploit network topology structure and biochemical information to predict potential DDIs. Additionally, we proposed a novel model, named BioChemDDI, to adaptively learn drug features from the chemical sequence information, network structural information and biological function information, and fused them to enhance features through the attention module. After sufficient supervised learning, the model accurately predicted potential DDIs. On the one hand, the case studies of three cancer-related drugs indicate the good prediction ability of our model. On the other hand, our model could be seen as a pre-screening tool for potential DDIs. In this way, the workload of exploring the unknown complex interactions of drugs can be reduced. In the future, to improve our framework, choosing negative samples with a more reasonable way to reduce the noise brought by unbalancing the original dataset and transferring our framework to predict interactions between unknown drugs will be considered.

5. Patents

The computational platform web server was built and is accessible at: http://120.77.11.78/BioChemDDI/ (accessed on 11 April 2022).

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/biology11050758/s1: Table S1: The full prediction scores in each case study.

Author Contributions

Conceptualization, Z.-H.R.; methodology, Z.-H.R.; software, validation, resources and data curation, Z.-H.R.; writing—original draft preparation, L.-P.L. and C.-Q.Y.; writing—review and editing, Y.-J.G., L.-X.G. and J.P.; visualization, C.-Q.Y.; supervision, L.-P.L.; project administration, Z.-H.Y.; funding acquisition, Z.-H.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Science and Technology Innovation 2030—New Generation Artificial Intelligence Major Project (No. 2018AAA0100103); the National Natural Science Foundation of China, under grant numbers 62002297, 61722212 and 62072378; and the Neural Science Foundation of Shanxi Province, under grant number 2022JQ-700.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

BioChemDDI is publicly available as an online predictor at: http://120.77.11.78/BioChemDDI/ (accessed on 11 April 2022). The original data of the drugs are available at: https://go.drugbank.com/ (accessed on 3 June 2021).

Acknowledgments

This work was mainly supported by the Science and Technology Innovation 2030—New Generation Artificial Intelligence Major Project (No. 2018AAA0100103).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Huang, L.; Luo, H.; Li, S.; Wu, F.-X.; Wang, J. Drug-drug similarity measure and its applications. Brief. Bioinform. 2021, 22, bbaa265. [Google Scholar] [CrossRef] [PubMed]
  2. Whitebread, S.; Hamon, J.; Bojanic, D.; Urban, L. Keynote review: In vitro safety pharmacology profiling: An essential tool for successful drug development. Drug Discov. Today 2005, 10, 1421–1433. [Google Scholar] [CrossRef]
  3. Huang, S.M.; Temple, R.; Throckmorton, D.; Lesko, L. Drug interaction studies: Study design, data analysis, and implications for dosing and labeling. Clin. Pharmacol. Ther. 2007, 81, 298–304. [Google Scholar] [CrossRef] [PubMed]
  4. Percha, B.; Altman, R.B. Informatics confronts drug-drug interactions. Trends Pharmacol. Sci. 2013, 34, 178–184. [Google Scholar] [CrossRef] [Green Version]
  5. Prueksaritanont, T.; Chu, X.; Gibson, C.; Cui, D.; Yee, K.L.; Ballard, J.; Cabalu, T.; Hochman, J. Drug-drug interaction studies: Regulatory guidance and an industry perspective. AAPS J. 2013, 15, 629–645. [Google Scholar] [CrossRef] [Green Version]
  6. Kusuhara, H. How far should we go? Perspective of drug-drug interaction studies in drug development. Drug Metab. Pharmacokinet. 2014, 29, 227–228. [Google Scholar] [CrossRef]
  7. Veith, H.; Southall, N.; Huang, R.; James, T.; Fayne, D.; Artemenko, N.; Shen, M.; Inglese, J.; Austin, C.P.; Lloyd, D.G. Comprehensive characterization of cytochrome P450 isozyme selectivity across chemical libraries. Nat. Biotechnol. 2009, 27, 1050–1055. [Google Scholar] [CrossRef] [Green Version]
  8. Zhao, X.-M.; Iskar, M.; Zeller, G.; Kuhn, M.; Van Noort, V.; Bork, P. Prediction of drug combinations by integrating molecular and pharmacological data. PLoS Comput. Biol. 2011, 7, e1002323. [Google Scholar] [CrossRef]
  9. Wiśniowska, B.; Polak, S. The role of interaction model in simulation of drug interactions and QT prolongation. Curr. Pharmacol. Rep. 2016, 2, 339–344. [Google Scholar] [CrossRef] [Green Version]
  10. Zhou, D.; Bui, K.; Sostek, M.; Al-Huniti, N. Simulation and prediction of the drug-drug interaction potential of naloxegol by physiologically based pharmacokinetic modeling. CPT Pharmacomet. Syst. Pharmacol. 2016, 5, 250–257. [Google Scholar] [CrossRef] [Green Version]
  11. Wishart, D.S.; Feunang, Y.D.; Guo, A.C.; Lo, E.J.; Marcu, A.; Grant, J.R.; Sajed, T.; Johnson, D.; Li, C.; Sayeeda, Z. DrugBank 5.0: A major update to the DrugBank database for 2018. Nucleic Acids Res. 2018, 46, D1074–D1082. [Google Scholar] [CrossRef] [PubMed]
  12. Hastings, J.; Owen, G.; Dekker, A.; Ennis, M.; Kale, N.; Muthukrishnan, V.; Turner, S.; Swainston, N.; Mendes, P.; Steinbeck, C. ChEBI in 2016: Improved services and an expanding collection of metabolites. Nucleic Acids Res. 2016, 44, D1214–D1219. [Google Scholar] [CrossRef] [PubMed]
  13. Kim, S.; Chen, J.; Cheng, T.; Gindulyte, A.; He, J.; He, S.; Li, Q.; Shoemaker, B.A.; Thiessen, P.A.; Yu, B. PubChem 2019 update: Improved access to chemical data. Nucleic Acids Res. 2019, 47, D1102–D1109. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Kanehisa, M.; Sato, Y.; Kawashima, M.; Furumichi, M.; Tanabe, M. KEGG as a reference resource for gene and protein annotation. Nucleic Acids Res. 2016, 44, D457–D462. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Gottlieb, A.; Stein, G.Y.; Oron, Y.; Ruppin, E.; Sharan, R. INDI: A computational framework for inferring drug interactions and their associated recommendations. Mol. Syst. Biol. 2012, 8, 592. [Google Scholar] [CrossRef] [PubMed]
  16. Cheng, F.; Zhao, Z. Machine learning-based prediction of drug–drug interactions by integrating drug phenotypic, therapeutic, chemical, and genomic properties. J. Am. Med. Inform. Assoc. 2014, 21, e278–e286. [Google Scholar] [CrossRef] [Green Version]
  17. Ferdousi, R.; Safdari, R.; Omidi, Y. Computational prediction of drug-drug interactions based on drugs functional similarities. J. Biomed. Inform. 2017, 70, 54–64. [Google Scholar] [CrossRef]
  18. Rohani, N.; Eslahchi, C. Drug-drug interaction predicting by neural network using integrated similarity. Sci. Rep. 2019, 9, 13645. [Google Scholar] [CrossRef] [Green Version]
  19. Zhang, P.; Wang, F.; Hu, J.; Sorrentino, R. Label propagation prediction of drug-drug interactions based on clinical side effects. Sci. Rep. 2015, 5, 12339. [Google Scholar] [CrossRef]
  20. Park, K.; Kim, D.; Ha, S.; Lee, D. Predicting pharmacodynamic drug-drug interactions through signaling propagation interference on protein-protein interaction networks. PLoS ONE 2015, 10, e0140816. [Google Scholar] [CrossRef] [Green Version]
  21. Deepika, S.; Geetha, T. A meta-learning framework using representation learning to predict drug-drug interaction. J. Biomed. Inf. 2018, 84, 136–147. [Google Scholar] [CrossRef] [PubMed]
  22. Liu, S.; Zhang, Y.; Cui, Y.; Qiu, Y.; Deng, Y.; Zhang, W.; Zhang, Z. Enhancing Drug-Drug Interaction Prediction Using Deep Attention Neural Networks. IEEE/ACM Trans. Comput. Biol. Bioinform. 2021. [Google Scholar] [CrossRef] [PubMed]
  23. Shi, J.-Y.; Huang, H.; Li, J.-X.; Lei, P.; Zhang, Y.-N.; Dong, K.; Yiu, S.-M. TMFUF: A triple matrix factorization-based unified framework for predicting comprehensive drug-drug interactions of new drugs. BMC Bioinform. 2018, 19, 411. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Zhang, W.; Jing, K.; Huang, F.; Chen, Y.; Li, B.; Li, J.; Gong, J. SFLLN: A sparse feature learning ensemble method with linear neighborhood regularization for predicting drug-drug interactions. Inf. Sci. 2019, 497, 189–201. [Google Scholar] [CrossRef]
  25. Zhang, W.; Chen, Y.; Li, D.; Yue, X. Manifold regularized matrix factorization for drug-drug interaction prediction. J. Biomed. Inform. 2018, 88, 90–97. [Google Scholar] [CrossRef]
  26. Yu, H.; Mao, K.-T.; Shi, J.-Y.; Huang, H.; Chen, Z.; Dong, K.; Yiu, S.-M. Predicting and understanding comprehensive drug-drug interactions via semi-nonnegative matrix factorization. BMC Syst. Biol. 2018, 12, 14. [Google Scholar] [CrossRef] [Green Version]
  27. Shi, J.-Y.; Mao, K.-T.; Yu, H.; Yiu, S.-M. Detecting drug communities and predicting comprehensive drug–drug interactions via balance regularized semi-nonnegative matrix factorization. J. Cheminform. 2019, 11, 1–16. [Google Scholar] [CrossRef] [Green Version]
  28. Chowdhury, M.F.M.; Lavelli, A. FBK-irst: A multi-phase kernel based approach for drug-drug interaction detection and classification that exploits linguistic information. In Second Joint Conference on Lexical and Computational Semantics (*SEM), Proceedings of the Seventh International Workshop on Semantic Evaluation (SemEval 2013), Atlanta, GA, USA, 14–15 June 2013; Association for Computational Linguistics: Stroudsburg, PA, USA, 2013; Volume 2, pp. 351–355. [Google Scholar]
  29. Zhu, Y.; Li, L.; Lu, H.; Zhou, A.; Qin, X. Extracting drug-drug interactions from texts with BioBERT and multiple entity-aware attentions. J. Biomed. Inform. 2020, 106, 103451. [Google Scholar] [CrossRef]
  30. Cases, M.; Briggs, K.; Steger-Hartmann, T.; Pognan, F.; Marc, P.; Kleinöder, T.; Schwab, C.H.; Pastor, M.; Wichard, J.; Sanz, F. The eTOX data-sharing project to advance in silico drug-induced toxicity prediction. Int. J. Mol. Sci. 2014, 15, 21136–21154. [Google Scholar] [CrossRef] [Green Version]
  31. Ren, Z.-H.; Yu, C.-Q.; Li, L.-P.; You, Z.-H.; Guan, Y.-J.; Wang, X.-F.; Pan, J. BioDKG–DDI: Predicting drug–drug interactions based on drug knowledge graph fusing biochemical information. Brief. Funct. Genom. 2022. [Google Scholar] [CrossRef]
  32. Yue, X.; Wang, Z.; Huang, J.; Parthasarathy, S.; Moosavinasab, S.; Huang, Y.; Lin, S.M.; Zhang, W.; Zhang, P.; Sun, H. Graph embedding on biomedical networks: Methods, applications and evaluations. Bioinformatics 2020, 36, 1241–1251. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Zhang, W.; Chen, Y.; Liu, F.; Luo, F.; Tian, G.; Li, X. Predicting potential drug-drug interactions by integrating chemical, biological, phenotypic and network data. BMC Bioinform. 2017, 18, 18. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Kuhn, M.; Letunic, I.; Jensen, L.J.; Bork, P. The SIDER database of drugs and side effects. Nucleic Acids Res. 2016, 44, D1075–D1079. [Google Scholar] [CrossRef] [PubMed]
  35. Kanehisa, M.; Furumichi, M.; Tanabe, M.; Sato, Y.; Morishima, K. KEGG: New perspectives on genomes, pathways, diseases and drugs. Nucleic Acids Res. 2017, 45, D353–D361. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Tatonetti, N.P.; Patrick, P.Y.; Daneshjou, R.; Altman, R.B. Data-driven prediction of drug effects and interactions. Sci. Transl. Med. 2012, 4, 125ra31. [Google Scholar] [CrossRef] [Green Version]
  37. Feng, Y.-H.; Zhang, S.-W.; Shi, J.-Y. DPDDI: A deep predictor for drug-drug interactions. BMC Bioinform. 2020, 21, 419. [Google Scholar] [CrossRef]
  38. Chen, H.; Perozzi, B.; Hu, Y.; Skiena, S. Harp: Hierarchical representation learning for networks. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  39. Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.S.; Dean, J. Distributed representations of words and phrases and their compositionality. In Advances in Neural Information Processing Systems 30, Proceedings of the 26th International Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 5–10 December 2013; Curran Associates Inc.: Red Hook, NY, USA, 2017; pp. 3111–3119. [Google Scholar]
  40. Tang, J.; Qu, M.; Wang, M.; Zhang, M.; Yan, J.; Mei, Q. Line: Large-scale information network embedding. In Proceedings of the 24th International Conference on World Wide Web, Florence, Italy, 18–22 May 2015; pp. 1067–1077. [Google Scholar]
  41. Wang, B.; Mezlini, A.M.; Demir, F.; Fiume, M.; Tu, Z.; Brudno, M.; Haibe-Kains, B.; Goldenberg, A. Similarity network fusion for aggregating data types on a genomic scale. Nat. Methods 2014, 11, 333–337. [Google Scholar] [CrossRef]
  42. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Advances in Neural Information Processing Systems 30, Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Curran Associates Inc.: Red Hook, NY, USA, 2017; pp. 5998–6008. [Google Scholar]
  43. Lin, M.; Chen, Q.; Yan, S. Network in network. arXiv 2013, arXiv:1312.4400. [Google Scholar]
  44. Jiao, Y.; Du, P. Performance measures in evaluating machine learning based bioinformatics predictors for classifications. Quant. Biol. 2016, 4, 320–330. [Google Scholar] [CrossRef] [Green Version]
  45. Chou, K.-C.; Zhang, C.-T. Prediction of protein structural classes. Crit. Rev. Biochem. Mol. Biol. 1995, 30, 275–349. [Google Scholar] [CrossRef]
  46. Perozzi, B.; Al-Rfou, R.; Skiena, S. Deepwalk: Online learning of social representations. In Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining, New York, NY, USA, 24–27 August 2014; pp. 701–710. [Google Scholar]
  47. Belkin, M.; Niyogi, P. Laplacian eigenmaps for dimensionality reduction and data representation. Neural Comput. 2003, 15, 1373–1396. [Google Scholar] [CrossRef] [Green Version]
  48. Rohani, N.; Eslahchi, C.; Katanforoush, A. Iscmf: Integrated similarity-constrained matrix factorization for drug–drug interaction prediction. Netw. Modeling Anal. Health Inform. Bioinform. 2020, 9, 11. [Google Scholar] [CrossRef]
  49. Chen, X.; Liu, X.; Wu, J. GCN-BMP: Investigating graph representation learning for DDI prediction task. Methods 2020, 179, 47–54. [Google Scholar] [CrossRef] [PubMed]
  50. Schwarz, K.; Allam, A.; Perez Gonzalez, N.A.; Krauthammer, M. AttentionDDI: Siamese attention-based deep learning method for drug-drug interaction predictions. BMC Bioinform. 2021, 22, 412. [Google Scholar] [CrossRef]
Figure 1. The pipeline of BioChemDDI. After inputting data, the sub-model of chemical sequence information obtains an embedding matrix of each drug that can be reduced as a 1-dimension vector by the CNN. The network structural information is learned by the method of HARP, and the feature vector can be further screened for the characteristic factors through the hidden layers. Multiple biological function matrices are used to generate multiple drug similarity matrices, which can be fused by SNF and further screened for characteristic factors by the hidden layers. The module of feature integration with attention contains three hidden layers that finish the final DDI prediction.
Figure 1. The pipeline of BioChemDDI. After inputting data, the sub-model of chemical sequence information obtains an embedding matrix of each drug that can be reduced as a 1-dimension vector by the CNN. The network structural information is learned by the method of HARP, and the feature vector can be further screened for the characteristic factors through the hidden layers. Multiple biological function matrices are used to generate multiple drug similarity matrices, which can be fused by SNF and further screened for characteristic factors by the hidden layers. The module of feature integration with attention contains three hidden layers that finish the final DDI prediction.
Biology 11 00758 g001
Figure 2. The guideline for extracting features of chemical sequence. The SMILES of all drugs were obtained from DrugBank. Additionally, vector representation, the matrix of 60 × 64, of all 60 biological symbols is learned by word2vec. Then, the SMILES of each drug corresponds to its own symbolic vector representation. Finally, N drugs can be indicated as N matrixes. To ensure the identical scale of the drug matrix, the strategies were adopted as follows: a. If the sequence length of the drug is shorter than 64, the vector of zero is attached to the drug matrix; b. If the sequence length of the drug is longer than 64, it is reduced to 64.
Figure 2. The guideline for extracting features of chemical sequence. The SMILES of all drugs were obtained from DrugBank. Additionally, vector representation, the matrix of 60 × 64, of all 60 biological symbols is learned by word2vec. Then, the SMILES of each drug corresponds to its own symbolic vector representation. Finally, N drugs can be indicated as N matrixes. To ensure the identical scale of the drug matrix, the strategies were adopted as follows: a. If the sequence length of the drug is shorter than 64, the vector of zero is attached to the drug matrix; b. If the sequence length of the drug is longer than 64, it is reduced to 64.
Biology 11 00758 g002
Figure 3. The process of the network representation learning module. There are three parts to HARP: graph coarsening, graph embedding and representation refining. After generating the drug network from the training of the positive samples, the original network is folded continuously to gain many coarse subnetworks. For example, in the level-1 coarsening network, according to the folded rule, node 1 and node 2 are seen as a whole sharing the same embedding vector, when extracting embedding vectors of nodes in the level-1 coarsening network. The final embedding vectors of the lower-level subnetwork are exploited as the initial vector of its superior level subnetwork. Iterating upward constantly, ending at the original graph, and finally obtaining all of the node embedding vectors, which have rich structural information. The LINE method is utilized to obtain embedding vectors for the nodes of each subnetwork.
Figure 3. The process of the network representation learning module. There are three parts to HARP: graph coarsening, graph embedding and representation refining. After generating the drug network from the training of the positive samples, the original network is folded continuously to gain many coarse subnetworks. For example, in the level-1 coarsening network, according to the folded rule, node 1 and node 2 are seen as a whole sharing the same embedding vector, when extracting embedding vectors of nodes in the level-1 coarsening network. The final embedding vectors of the lower-level subnetwork are exploited as the initial vector of its superior level subnetwork. Iterating upward constantly, ending at the original graph, and finally obtaining all of the node embedding vectors, which have rich structural information. The LINE method is utilized to obtain embedding vectors for the nodes of each subnetwork.
Biology 11 00758 g003
Figure 4. Illustration of the extracting process of the fusion drug similarity matrix. At first, M types of biological function information are used to construct adjacency matrixes, whose row indicates the drug’s initial vector and the number of receptors decides the dimension. Thus, each row is regarded as a high-dimensional space representation of drugs. Then, the drug similarity matrix based on different kinds of space can be obtained by calculating the Euclidean distance on each pairwise embedding vector of drugs. Finally, it exploits the SNF method to produce a comprehensive representation of the biological function information by fusing similarity matrixes of all types of space.
Figure 4. Illustration of the extracting process of the fusion drug similarity matrix. At first, M types of biological function information are used to construct adjacency matrixes, whose row indicates the drug’s initial vector and the number of receptors decides the dimension. Thus, each row is regarded as a high-dimensional space representation of drugs. Then, the drug similarity matrix based on different kinds of space can be obtained by calculating the Euclidean distance on each pairwise embedding vector of drugs. Finally, it exploits the SNF method to produce a comprehensive representation of the biological function information by fusing similarity matrixes of all types of space.
Biology 11 00758 g004
Figure 5. ROC curves of the five-fold cross-validation on the four datasets by the proposed method, BioChemDDI.
Figure 5. ROC curves of the five-fold cross-validation on the four datasets by the proposed method, BioChemDDI.
Biology 11 00758 g005
Figure 6. PR curves of the five-fold cross-validation on the four datasets by the proposed method, BioChemDDI.
Figure 6. PR curves of the five-fold cross-validation on the four datasets by the proposed method, BioChemDDI.
Biology 11 00758 g006
Figure 7. The performance of different graph embedding methods on each dataset.
Figure 7. The performance of different graph embedding methods on each dataset.
Biology 11 00758 g007
Figure 8. The performance of different dimensions of the graph embedding vector on each dataset.
Figure 8. The performance of different dimensions of the graph embedding vector on each dataset.
Biology 11 00758 g008
Figure 9. The performance of different classifiers on each dataset.
Figure 9. The performance of different classifiers on each dataset.
Biology 11 00758 g009
Figure 10. Network visualization of the newly discovered and known DDIs of the top 30 in each case study. The three colors of the node represent the three case studies. The red lines represent the interactions with no support and the green lines represent confirmed interactions. The thickness of the line between any two nodes represents the prediction score of the interaction.
Figure 10. Network visualization of the newly discovered and known DDIs of the top 30 in each case study. The three colors of the node represent the three case studies. The red lines represent the interactions with no support and the green lines represent confirmed interactions. The thickness of the line between any two nodes represents the prediction score of the interaction.
Biology 11 00758 g010
Table 1. The feature description and interaction details of the four datasets.
Table 1. The feature description and interaction details of the four datasets.
Datasets NameNumber of DrugsNumber of PairsNumber of InteractionsBiological Function Information
DS1 [31,32] 19403,763,600219,247Carrier, Target, Enzyme, Transporter
DS2 [33]548300,30497,168Pathway, Target, Enzyme, Indication, Offside, Side effect, Transporter
D-DS3 [27]15622,439,84455,278Target, Carrier, Enzyme, Transporter, Molecular fingerprint
E-DS3 [27]15622,439,844125,298Carrier, Target, Enzyme, Transporter, Molecular fingerprint.
Table 2. Five-fold cross-validation results through BioChemDDI.
Table 2. Five-fold cross-validation results through BioChemDDI.
DatasetsFoldAccPrecSenF1MCCAUCAUPR
DS100.95830.94360.97480.95890.91700.99150.9907
10.96190.94690.97870.96260.92440.99300.9925
20.96160.94580.97930.96220.92370.99300.9925
30.96100.94490.97920.96170.92270.99270.9920
40.96340.94690.98170.96400.92730.99340.9929
Average0.9612 ± 0.00190.9456 ± 0.00140.9787 ± 0.00250.9618 ± 0.00190.9232 ± 0.00380.9926 ± 0.00070.9921 ± 0.0009
DS200.91010.89910.92370.91130.82040.97080.9694
10.90630.89640.9190.90750.81300.97060.9695
20.90510.89130.92270.90670.81060.97070.9701
30.90730.89310.92530.90890.81510.97130.9702
40.90730.89730.91980.90840.81480.97210.9713
Average0.9072 ± 0.00180.8954 ± 0.00320.9221 ± 0.00270.9086 ± 0.00170.8148 ± 0.00360.9711 ± 0.00060.9701 ± 0.0008
D-DS300.95450.93700.97450.95540.90970.98480.9794
10.95490.93850.9760.95570.91050.98650.9822
20.95390.95150.95650.95400.90780.98720.9843
30.95590.94080.97300.95660.91230.98650.9826
40.95670.94320.97190.95730.91380.98630.9815
Average0.9552 ± 0.00110.9422 ± 0.00570.9699 ± 0.00760.9558 ± 0.00130.9108 ± 0.00230.9863 ± 0.00090.9820 ± 0.0018
E-DS300.95510.93600.97710.95610.91110.98970.9879
10.9660.95110.97730.96410.92750.99280.9915
20.95860.94690.97180.95920.91760.99120.9901
30.96150.94750.97710.96210.92350.99180.9904
40.96130.94930.97450.96180.92280.99170.9897
Average0.9600 ± 0.00330.9462 ± 0.00590.9756 ± 0.00240.9607 ± 0.00310.9205 ± 0.00630.9914 ± 0.00110.9900 ± 0.0013
Table 3. Results of the ablation on DS1.
Table 3. Results of the ablation on DS1.
Experimental NumberChemical SequenceNetwork StructureBiological FunctionAUCAUPR
Without the attention mechanism(a)FTT0.98420.9824
(b)TFT0.97350.9698
(c)TTF0.98550.9836
(d)TTT0.98700.9855
With the attention mechanism(A)FTT0.99150.9911
(B)TFT0.98120.9793
(C)TTF0.98790.9867
(D)TTT0.99270.9921
F means False; T means True and with attention. The bold values indicate the highest values of AUC and AUPR.
Table 4. Comparison of our proposed method with seven computational methods.
Table 4. Comparison of our proposed method with seven computational methods.
MethodAccPrecSenF1MCCAUCAUPR
NDD [18](-)0.83300.83600.8350(-)0.95400.9220
ISCMF [48]0.85100.98800.85100.8850(-)0.89900.8640
DPDDI [37]0.94000.75400.81000.8400(-)0.95600.9070
GCN-BMP [49](-)(-)(-)0.8500(-)0.96660.9402
AttentionDDI [50](-)(-)(-)(-)(-)0.95400.9240
BioDKG-DDI [31]0.89840.88350.91780.90030.79740.9668(-)
Ensemble Model [33]0.95500.78500.67000.7230(-)0.95700.8070
Proposed Method0.90720.89540.92210.90860.81480.97110.9701
The values in bold indicate the highest values in each column. The symbol (-) indicates the evaluation criteria are not reported in the original articles.
Table 5. The top 30 interactions of Cyclophosphamide.
Table 5. The top 30 interactions of Cyclophosphamide.
RankDrug NameDrugBank IDEvidenceRankDrug NameDrugBank IDEvidence
1AltretamineDB00488Confirmed16FludarabineDB01073Confirmed
2AmoxicillinDB01060N.A.17NaproxenDB00788Confirmed
3IfosfamideDB01181Confirmed18MercaptopurineDB01033Confirmed
4IxazomibDB09570Confirmed19PropylthiouracilDB00550Confirmed
5IndomethacinDB00328Confirmed20SulindacDB00605N.A.
6Aminolevulinic acidDB00855N.A.21DoxorubicinDB00997Confirmed
7Magnesium salicylateDB01397N.A.22FoscarnetDB00529N.A.
8CytarabineDB00987Confirmed23SuraminDB04786N.A.
9PirfenidoneDB04951Confirmed24AntipyrineDB01435Confirmed
10MelphalanDB01042Confirmed25MethotrexateDB00563Confirmed
11EstriolDB04573Confirmed26EtoricoxibDB01628Confirmed
12PentostatinDB00552Confirmed27LetrozoleDB01006Confirmed
13DaunorubicinDB00694Confirmed28FlutamideDB00499Confirmed
14RaltitrexedDB00293Confirmed29ProcarbazineDB01168Confirmed
15CarmustineDB00262Confirmed30TrabectedinDB05109Confirmed
Table 6. The top 30 interactions of Regorafenib.
Table 6. The top 30 interactions of Regorafenib.
RankDrug NameDrugBank IDEvidenceRankDrug NameDrugBank IDEvidence
1IndomethacinDB00328Confirmed16DoxorubicinDB00997Confirmed
2PitavastatinDB08860Confirmed17GrazoprevirDB11575Confirmed
3LamivudineDB00709Confirmed18RomidepsinDB06176Confirmed
4DaunorubicinDB00694Confirmed19AmodiaquineDB00613N.A.
5PrucaloprideDB06480Confirmed20CilostazolDB01166Confirmed
6DocetaxelDB01248Confirmed21DebrisoquineDB04840N.A.
7EstriolDB04573Confirmed22ThiotepaDB04572N.A.
8Taurocholic acidDB04348N.A.23CisplatinDB00515N.A.
9UlipristalDB08867N.A.24IbuprofenDB01050Confirmed
10OmbitasvirDB09296Confirmed25TheophyllineDB00277N.A.
11EzetimibeDB00973Confirmed26WarfarinDB00682Confirmed
12ZidovudineDB00495Confirmed27IndacaterolDB05039Confirmed
13SparfloxacinDB01208N.A.28PaclitaxelDB01229N.A.
14AlitretinoinDB00523N.A.29AminophyllineDB01223N.A.
15VismodegibDB08828Confirmed30DyphyllineDB00651N.A.
Table 7. The top 30 interactions of Allopurinol.
Table 7. The top 30 interactions of Allopurinol.
RankDrug NameDrugBank IDEvidenceRankDrug NameDrugBank IDEvidence
1PropylthiouracilDB00550Confirmed16FlucloxacillinDB00301N.A.
2AmoxicillinDB01060Confirmed17AntipyrineDB01435Confirmed
3IbrutinibDB09053Confirmed18DicloxacillinDB00485N.A.
4CarboplatinDB00958Confirmed19TrabectedinDB05109Confirmed
5TretinoinDB00755Confirmed20PodofiloxDB01179N.A.
6AmodiaquineDB00613N.A.21TioguanineDB00352Confirmed
7CilostazolDB01166Confirmed22LomefloxacinDB00978N.A.
8AzathioprineDB00993Confirmed23Aluminum hydroxideDB06723Confirmed
9MevastatinDB06693N.A.24TheophyllineDB00277Confirmed
10MethimazoleDB00763Confirmed25Magnesium hydroxideDB09104Confirmed
11IdarubicinDB01177Confirmed26OxaliplatinDB00526Confirmed
12StavudineDB00649N.A.27DyphyllineDB00651Confirmed
13MercaptopurineDB01033Confirmed28VerteporfinDB00460N.A.
14AminophyllineDB01223Confirmed29Magnesium oxideDB01377Confirmed
15TemafloxacinDB01405N.A.30CarbimazoleDB00389N.A.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ren, Z.-H.; Yu, C.-Q.; Li, L.-P.; You, Z.-H.; Pan, J.; Guan, Y.-J.; Guo, L.-X. BioChemDDI: Predicting Drug–Drug Interactions by Fusing Biochemical and Structural Information through a Self-Attention Mechanism. Biology 2022, 11, 758. https://doi.org/10.3390/biology11050758

AMA Style

Ren Z-H, Yu C-Q, Li L-P, You Z-H, Pan J, Guan Y-J, Guo L-X. BioChemDDI: Predicting Drug–Drug Interactions by Fusing Biochemical and Structural Information through a Self-Attention Mechanism. Biology. 2022; 11(5):758. https://doi.org/10.3390/biology11050758

Chicago/Turabian Style

Ren, Zhong-Hao, Chang-Qing Yu, Li-Ping Li, Zhu-Hong You, Jie Pan, Yong-Jian Guan, and Lu-Xiang Guo. 2022. "BioChemDDI: Predicting Drug–Drug Interactions by Fusing Biochemical and Structural Information through a Self-Attention Mechanism" Biology 11, no. 5: 758. https://doi.org/10.3390/biology11050758

APA Style

Ren, Z. -H., Yu, C. -Q., Li, L. -P., You, Z. -H., Pan, J., Guan, Y. -J., & Guo, L. -X. (2022). BioChemDDI: Predicting Drug–Drug Interactions by Fusing Biochemical and Structural Information through a Self-Attention Mechanism. Biology, 11(5), 758. https://doi.org/10.3390/biology11050758

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop