Next Article in Journal
Exploring the Synergistic Effect of Bergamot Essential Oil with Spironolactone Loaded Nano-Phytosomes for Treatment of Acne Vulgaris: In Vitro Optimization, In Silico Studies, and Clinical Evaluation
Next Article in Special Issue
Identification of a Family of Glycoside Derivatives Biologically Active against Acinetobacter baumannii and Other MDR Bacteria Using a QSPR Model
Previous Article in Journal
Terminalia arjuna, a Cardioprotective Herbal Medicine–Relevancy in the Modern Era of Pharmaceuticals and Green Nanomedicine—A Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SIRT2i_Predictor: A Machine Learning-Based Tool to Facilitate the Discovery of Novel SIRT2 Inhibitors

by
Nemanja Djokovic
1,*,
Minna Rahnasto-Rilla
2,
Nikolaos Lougiakis
3,
Maija Lahtela-Kakkonen
2 and
Katarina Nikolic
1,*
1
Department of Pharmaceutical Chemistry, Faculty of Pharmacy, University of Belgrade, Vojvode Stepe 450, 11221 Belgrade, Serbia
2
School of Pharmacy, University of Eastern Finland, P.O. Box 1627, 70210 Kuopio, Finland
3
Laboratory of Medicinal Chemistry, Section of Pharmaceutical Chemistry, Department of Pharmacy, School of Health Sciences, National and Kapodistrian University of Athens, Panepistimiopolis-Zografou, 15771 Athens, Greece
*
Authors to whom correspondence should be addressed.
Pharmaceuticals 2023, 16(1), 127; https://doi.org/10.3390/ph16010127
Submission received: 13 December 2022 / Revised: 10 January 2023 / Accepted: 11 January 2023 / Published: 14 January 2023

Abstract

:
A growing body of preclinical evidence recognized selective sirtuin 2 (SIRT2) inhibitors as novel therapeutics for treatment of age-related diseases. However, none of the SIRT2 inhibitors have reached clinical trials yet. Transformative potential of machine learning (ML) in early stages of drug discovery has been witnessed by widespread adoption of these techniques in recent years. Despite great potential, there is a lack of robust and large-scale ML models for discovery of novel SIRT2 inhibitors. In order to support virtual screening (VS), lead optimization, or facilitate the selection of SIRT2 inhibitors for experimental evaluation, a machine-learning-based tool titled SIRT2i_Predictor was developed. The tool was built on a panel of high-quality ML regression and classification-based models for prediction of inhibitor potency and SIRT1-3 isoform selectivity. State-of-the-art ML algorithms were used to train the models on a large and diverse dataset containing 1797 compounds. Benchmarking against structure-based VS protocol indicated comparable coverage of chemical space with great gain in speed. The tool was applied to screen the in-house database of compounds, corroborating the utility in the prioritization of compounds for costly in vitro screening campaigns. The easy-to-use web-based interface makes SIRT2i_Predictor a convenient tool for the wider community. The SIRT2i_Predictor’s source code is made available online.

Graphical Abstract

1. Introduction

SIRT2 is a NAD+-dependent protein deacetylase involved in the regulation of many important biological functions, including the maintenance of genome stability, metabolism, aging, tumorigenesis, and cell-cycle regulation [1,2,3,4,5]. Studies on cellular and animal models of disease revealed the promising potential of SIRT2 inhibition in the treatment of age-related diseases, including neurodegenerative disease and carcinoma [6,7]. Preclinical evidence generated during the last decade has resulted in growing interest in the development of small-molecule SIRT2 inhibitors, particularly as novel anticancer therapeutics [7]. The inhibition of SIRT2 was shown to be an important factor in the treatment of various aspects of tumor development and progression, including inhibition of proliferation, invasion, angiogenesis, and metastatic potential [8,9,10]. In addition to cancer development and progression, SIRT2 was proven to be involved in conferring re-sistance to cancer treatment. Great potential for synergistic combinations of SIRT2 inhibitors with clinically approved drugs was revealed just recently by examining the role of SIRT2 inhibitors in overcoming drug resistance to dasatinib, doxorubicin, or paclitaxel in treatment of melanoma or specific subtypes of breast cancer cells [11,12,13]. Furthermore, the recent study examined selective SIRT2 inhibitors as an augmentation to tumor immunotherapy due to their ability to activate tumor-infiltrating lymphocytes. This approach opened exciting opportunities for future usage of selective SIRT2 inhibitors in overcoming poor clinical response to the TIL (tumor-infiltrating lymphocyte) or CAR-T (chimeric antigen receptor–T cell) immunotherapies [14]. In spite of two decades of vigorous research efforts around the world and many discovered SIRT2 inhibitors, none of the described compounds have entered clinical trials, which signifies the need for novel advances in the field [15]. The most common limitations of known inhibitors includes poor selectivity, potency, or physicochemical properties [10,15,16].
The catalytic core of sirtuins consists of a larger Rossmann-fold domain and a smaller zinc-binding domain connected with several flexible loops (Figure 1A). All sirtuins share the same catalytic mechanism, which involves the formation of a positively charged O-alkylimidate intermediate between NAD+ and acetyl–lysine substrate. After several steps, this intermediate hydrolyzes to produce deacetylated polypeptide and 2′-O-acetyl-ADP-ribose (Figure 1B) [17]. Most known inhibitors interfere with this catalytic mechanism by binding to the catalytic site of sirtuins positioned in the cleft between two domains (Figure 1C). Due to the conserved structure of the catalytic site of sirtuins, achieving the selectivity of small-molecule SIRT2 inhibitors turned out to be one of the greatest challenges in the development of this group of compounds (Figure 1C) [15,18]. Recently described pharmacological advantages of selective SIRT2 inhibition over non-selective inhibition of other isoforms of the sirtuin family, particularly sirtuin 1 (SIRT1) and sirtuin 3 (SIRT3), positioned selectivity as one of the most important objectives in development of novel SIRT2 inhibitors [19]. Furthermore, a recent study indicated that the complex conformational behavior of SIRT2 in interaction with inhibitors represents one of the major obstacles in the discovery of novel inhibitors through structure-based computer-aided drug-design (CADD) approaches [20]. However, years of searching for novel SIRT2 inhibitors resulted in large and diverse datasets that could greatly benefit ligand-based CADD approaches relying on specific machine-learning techniques.
Discovery of novel drugs under the scope of the precision medicine (NIH) initiative heavily relies on integration of large datasets into the drug-discovery pipelines through cheminformatics approaches [21]. The big-data-driven era in modern drug discovery recognized artificial intelligence (AI) as one of the most important tools that could drastically reduce the time and cost of drug discovery in preclinical phases [22,23]. Escalated development and the usage of machine-learning (ML) tools in modern drug discovery was mainly allowed by the availability of large datasets and the democratization of AI. Public databases of pharmacologically active compounds with an ever-increasing number of records on biological activities allowed for a more comprehensive approach to modern drug discovery by the utilization of ML in the modeling of structure–activity relationships [21,22,23]. Quantitative structure–activity relationship (QSAR) modeling is a well-known computational technique for establishing classification-based or regression-based relationships between structural properties of compounds and biological activities [24]. The QSAR technique represents one of the most successful strategies to avoid inactive compounds or to eliminate side effects in pre-clinical drug development [21,22,25]. Retrospective analysis indicated that updating QSAR models as more data become available generally leads to improvements in the accuracy and usefulness of predictions. QSAR models trained on the larger and more diverse datasets are more likely to have a wider applicability domain and to exert a larger coverage of the chemical space. Considering the general improvements in quality as well as broadening the applicability domains of QSAR models trained on larger datasets, global or large-scale QSAR models (i.e., models trained on large data sets of higher compound diversity) are becoming more and more popular [25,26,27]. Currently, there is a lack of large-scale and robust QSAR models for prediction of SIRT2 inhibitor potency and selectivity. Development of such models could aid virtual-screening studies, lead-optimization studies, and repurposing studies, or the integration of cheminformatics with omics data under the more complex precision-medicine pipelines.
With preclinical proof of pharmacological potential of selective SIRT2 inhibitors to treat various modalities of cancer, or to synergize with existing therapies, including immunotherapies, selective SIRT2 inhibitors could prove to be a valuable asset to the existing palette of drugs in the emerging era of personalized medicine. In order to provide open-source computational tools to facilitate the development of SIRT2 inhibitors, in this work we aimed to develop a framework for fast screening and evaluation of novel compounds on SIRT2 inhibitory potency and selectivity. The defined framework, named SIRT2i_Predictor, was built on set of high-quality large-scale classification and regression QSAR models implementing publicly available datasets on selectivity and potency of SIRT2 inhibitors. By creating an appealing and easy-to-use web-based interface, SIRT2i_Predictor was made available to the broader community.

2. Results and Discussion

2.1. Datasets for Modelling

Available data on the structures and activities of SIRT2 inhibitors were collected from ChEMBL data and literature (see Section 3), resulting in a total of 1797 unique records. Considering the biological-activity measurements (e.g., some of the compounds were tested on additional activities on isoforms SIRT1 or SIRT3, whereas some others were not), the initial pool of data was distributed into four datasets (Datasets 1–4) (Figure 2 and Figure 3 and Table 1). Dataset 1 was intended for building a regression QSAR model, whereas Datasets 2–4 were intended for building different types of classification models. Distributions of activity values (pIC50 in Dataset 1) across different activity classes (Datasets 2–4) are depicted in Figure 2, whereas the main characteristics of Datasets 1–4 are summarized in Table 1.
Dataset 1 was built solely on records of compounds with reported SIRT2 inhibitory activity expressed as pIC50 values. This dataset contained 1002 compounds with a range of pIC50 values from 4 to 7.96. With 1797 entries, Dataset 2 was the largest among the datasets and encompassed all compounds, including activities expressed as both pIC50 and Inh%. Considering the criteria presented in Section 3, compounds within Dataset 2 were assigned to the two classes—SIRT2 active and SIRT2 inactive—resulting in almost one-third of compounds assigned as inactive (Figure 2).
Dataset 3 was composed of the compounds that reported inhibitory activities on both SIRT1 and SIRT2 (expressed either as pIC50 or Inh%), whereas Dataset 4 was composed of the compounds that reported SIRT2 and SIRT3 inhibitory data, expressed as pIC50 or Inh%. Therefore, Datasets 3 and 4 were composed of three classes of compounds: SIRT1(3)/SIRT2 inactive compounds, SIRT2 selective compounds, and SIRT1(3)/SIRT2 nonselective compounds (Figure 2 and Table 1). It should be noted that the bioactivity values used in this study were heterogeneous in origin, determined with different experimental approaches (fluorimetric assay, luminescence assay, electrophoretic mobility shift, scintillation counting) and conditions (time of incubation, acetyl–lysine substrates of different Km values, etc.). In order to make a clearer distinction between classes and eliminate potential noise coming from different experimental conditions, during the creation of Datasets 2–4 compounds with activities lying near fuzzy borders of different classes were omitted. This small group of compounds, referred to in the manuscript as “twilight zone” compounds, had an activity range of IC50 = 50–90 µM (for Inh% criteria see Section 3). The recent growth of interest in large-scale QSAR models utilizing datasets from ChEMBL that are heterogeneous in origin has resulted in many studies with similar data-acquisition and -processing strategies as those presented in this work [28,29,30,31,32,33].
Descriptive analysis across different datasets indicated that most of the available compounds obeyed Lipinski’s rule of 5, with several outliers in each dataset (Figure 2) [34]. Compounds from the datasets were pre-processed, and molecular descriptors and fingerprints were calculated using approaches explained in the Section 3. Prior to modeling, all datasets were split into training and test sets using stratified random sampling (70% training set and 30% test set) to ensure the sampling from the same activity distributions. Principal component analysis of different datasets indicated that the splitting strategy was able to maintain equal coverage of the chemical space with training- and test-set compounds (Figure 2). Considering the slight imbalance inside classification datasets (Datasets 2–4), the SMOTE algorithm was used to oversample the minority classes by synthesizing new minority instances prior to training the classification models.

2.2. Model Development and Validation

In this study, different regression, binary, and multiclass classification machine-learning models were developed using the combination of five machine-learning algorithms (random forest (RF), support-vector machines (SVM–support-vector classification (SVC) and support-vector regression (SVR)), k-nearest neighbors (KNN), extreme gradient boosting (XGBoost), and deep neural networks (DNN)), as well as four descriptors/fingerprints (Mordred descriptors, ECFP4, ECFP6, and MACCS key fingerprints) (see Section 3). Due to the their easy calculation and the very positive results obtained through cheminformatics studies over the years, recent literature recognized ECFP and MACCS as the most popular ones among commonly used fingerprints [21]. The abovementioned reasons were the discriminatory criteria for the selection of ECFP and MACCS fingerprints in our study, as well. The general workflow of the study is presented in Figure 3. Different regression or classification models for each dataset were trained thorough a process of hyperparameter tuning using Bayesian optimization with five-fold cross-validation. The list of values of hyperparameters used for Bayesian optimization is presented in the Supplementary Materials (Table S1). For DNN models, more comprehensive optimization of hyperparameters related to the network structure (number of hidden layers, units, and dropouts) was performed using the Keras Tuner with the Bayesian search method. The list of optimized hyperparameters for each combination of modeling approach and descriptor/fingerprint is provided in the Supplementary Materials (Table S2). The quality of each model trained with optimized hyperparameters was primarily evaluated through an inspection of internal and cross-validation parameters. Afterwards, predictive performance of each model on the “unseen” data was accessed through external validation with the test set. The top-performing models were selected using the consensus approach after evaluating the predictive performance of each model with additional tests. Additional evaluation approaches were chosen depending on the type and purpose of the model. Specificities of each of additional evaluation test are addressed below. The final selected models were considered for constructing the framework of SIRT2i_Predictor (Figure 3).

2.2.1. Regression Models

The combinations of five machine-learning methods with four features (MACCS, ECFP4, and ECFP6 fingerprints, as well as Mordred descriptors [35]) were explored for the development of global regression-based QSAR models using Dataset 1. After the feature-selection procedure (see Section 3), 52 Mordred descriptors were selected for the final QSAR modeling (Supplementary Materials, Table S3). Fingerprints were used without further reduction of the number of beats. After training the models through the Bayesian hyperparameter optimization procedure with five-fold cross-validation (CV), the quality of each model was accessed initially using internal validation parameters: the coefficient of determination for the training set (R2) and cross-validated correlation coefficient (Q2), root mean square error (RMSE) of fitting the training set (RMSEint), and the RMSE of cross-validation (RMSECV). Thresholds for R2 and Q2 were set according to the criteria proposed by Golbraikh and Tropsha (R2 > 0.6, Q2 > 0.5) [36]. Parameters of internal validations, presented in the Supplementary Materials (Table S4), indicated the internal predictive power, stability, and robustness of each model. Y-scrambling was performed as a part of the internal validation procedure by generating 100 models on randomly shuffled data. The Y-scrambling procedure indicated that the models were not obtained by chance and were highly reliable (Figure 4).
External validation was performed in order to examine the predictive power of the models outside the training data. In alignment with the Organization for Economic Co-operation and Development (OECD) principles, the external validation of the QSAR models could be performed by measuring the goodness-of-fit by coefficient of determination ( R e x t 2 , should be > 0.6) and root mean square error (RMSEext parameter, should be as low as possible) [37,38]. All models performed almost equally well according to these two criteria, with slight predominance in the quality of the models build with ECFP4 and ECFP6 fingerprints (Table 2). However, as discussed by many authors before, relying solely on the simplistic R e x t 2 in some cases could lead to an overoptimistic estimation of the model’s external predictive performance due to the dependence of R e x t 2 on the range of the response values of the test set and their distribution pattern around the training-/test-set mean [39,40,41,42]. Therefore, additional evaluation of the QSAR models’ external predictive power was performed using a set of additional parameters developed accounting for precision (deviation of observations from the fitting line), accuracy (deviation of the regression line from the slope 1 line passing through the origin of Yobserved vs. Ypredicted curve), and by ensuring that no bias was introduced based on the response scale [40,42]. As additional criteria, the r ¯ m 2 metrics by Roy et al. ( r ¯ m 2 > 0.5 and Δ r m 2 < 0.2) [39,43], the Q F n 2 metric, and the CCC with thresholds proposed by Chirico and Gramatica ( Q F 1 2 , Q F 2 2 , and Q F 3 2   > 0.7, and CCC > 0.85) [40,44] were used (Table 2). Furthermore, the criteria proposed by Golbraikh and Tropsha (( R 2 R 0 2 )/ R 2 < 0.1 or ( R 2 R 0 2 )/ R 2 < 0.1, 0.85 ≤ k (or k’) ≤ 1.15 and R 2 R 0 2 < 0.3) were assessed as well (Supplementary Materials, Table S4) [36]. The created models satisfied almost all of the proposed additional criteria, except Δ r m 2 and CCC, which failed for some of the models (see Table 2). After discarding the models according to the Δ r m 2 and CCC criteria, two the most promising models were selected: the XGBoost:ECFP4 model (Figure 4) and the KNN:ECFP6 model (Table 2 and Supplementary Materials, Table S4).
Following the OECD principles, chemical space boundaries where the model achieved reliable predictions, also known as the applicability domain (AD) of the QSAR model, should be defined as part of external validation and considered when the model is applied for predictions of unknown compounds [37,38,45]. One of the most widely used methods for estimating the boundaries of AD in regression QSAR models is the leverage method [45]. Leverage values are considered proportional to the distances of each compound from the centroid of the training set in the molecular-features space. Leverage values of selected QSAR models are presented through Williams plots (leverage versus standardized residuals) where compounds with leverages and/or residuals above threshold values could be easily detected (Figure 4). Several compounds from the test set were detected to be outside of the applicability domains for the two of the most promising models (XGBoost:ECFP4 and KNN:ECFP6) (Table 2 and Figure 4). The KNN:ECFP6 model was more limited in respect to the coverage of the chemical space, with 39 compounds of the test set being out of the defined AD boundaries compared to the 24 compounds of the XGBoost:ECFP4 model. Additionally, excluding compounds from outside the AD borders resulted in significant improvements in the XGBoost:ECFP4 model statistics compared to those of the KNN:ECFP6 model (Table 2). These results may indicate lower predictive power and lower coverage of chemical space of the KNN:ECFP6 model inside AD borders. Additionally, the KNN:ECFP6 model showed lower robustness through internal validation (Supplementary Materials, Table S4, cross-validation parameters). Since the global QSAR models are aimed at having broader coverage of chemical space, robustness, and optimal external predictive power, the XGBoost:ECFP4 model was selected for further work. However, it is important to note that regression models were trained on Dataset 1, where most of the compounds were active according to the class assignments (914 active compounds, whereas only 88 compounds were in the “twilight zone” (IC50 = 50–90 µM) or “inactive” (IC50 > 90 µM)) (see Section 3). This largely limits the usage of regression models only for the pIC50 predictions of the active compounds or compounds predicted to be active by classification models, which is further discussed in Section 2.3. Furthermore, the heterogeneity of the data introduced by the numerous sources that populate the ChEMBL database could contribute to the prediction error of the regression models [28]. Therefore, classification-based models that circumvent the abovementioned issue could be expected to have better performance in the identification of active compounds.

2.2.2. Binary Classification Models

According to the general protocol of the study (Figure 3), the combination of five different ML algorithms and four molecular features was explored through Bayes hyperparameter optimization with five-fold CV to develop binary classification models using Dataset 2. The main aim of this part of the study was to train models for the classification of SIRT2 inhibitors and inactive compounds. The rules for the assignment of compounds to the SIRT2 active/inactive class of inhibitors were addressed in the Section 3. Mordred descriptors used for training the binary models were, prior to model training, submitted to the feature-selection protocol (see Section 3 and Supplementary Materials, Table S3). 233 Mordred descriptors were selected for final modeling. Fingerprints were used without further reduction of the number of beats.
Parameters of internal validation indicated the internal predictive power, stability, and robustness of each model (Supplementary Materials, Table S5). The external predictive power of the trained binary models was evaluated using the test set. The following parameters were used for monitoring performances on the external set of different models: balanced accuracy (BA), Matthews correlation coefficient (MCC), area under the receiver operating characteristics curve (ROC_AUC), precision, recall, and F1-score (Table 3). Almost all of the modeling algorithms displayed equally good external predictive power on the external set, with mild predominance of the RF, SVC, and DNN models build using descriptors.
Aiming to recover active molecules from large databases enriched with inactive compounds, virtual screening (VS) is one of the possible real-life applications of this type of ML model. VS models that are able to encompass a larger portion of chemical space are considered more useful since the main objective of screening studies is to find chemically novel and diverse active compounds. Trained on the largest dataset—Dataset 2—the binary model could be expected to have the largest chemical space coverage and to be more useful in VS purposes compared to the models trained on other datasets. In order to additionally evaluate the applicability of selected binary models in the VS, real-life application was simulated by generating almost 20,000 virtual decoy molecules and assigning them to the inactive class. Decoys were created by enforcing 2D topological dissimilarity with known active molecules while retaining similar physical properties. The decoy dataset was merged with an external set, creating the imbalanced database with ratio active:inactive = 1:40. The models were further tested on their ability to recover the active molecules.
In these settings, statistical parameters of the models were recalculated with the addition of early enrichment metrics (Table 4). Early enrichment metrics represent one of the most important parameters for the evaluation of early recognition of VS models since only the top-ranked compounds are usually considered for experimental evaluation. Early recognition is the direct reflection of the models’ ability to rank active molecules very early in an ordered list. Herein, we used ROC EF 0.5%, 1%, 2%, and 5%, which quantified the area covered by the curve at 0.5%, 1%, 2%, and 5% of the screened false positives, respectively [20,46]. With the dataset containing a significantly larger number of chemically diverse inactive compounds, the RF:ECFP4 binary model stood out as the model with the greatest predictive power (Table 4 and Figure 5). In the heavily disbalanced decoy set, the RF:ECFP4 binary model displayed better sensitivity, specificity, precision, and robustness but also better early recognition. In 0.5% of false positives, the RF:ECFP4 binary model was able to find over 70% of true active molecules (Table 4). It is worth noting that the most of the true inactive compounds from Dataset 2 were chemically similar to the active compounds, whereas the decoy dataset was enriched in topologically dissimilar compounds. Since the parameters calculated on the decoy dataset represented the more reliable estimation of the model’s predictive performance in VS settings, the RF:ECFP4 model was selected for further work.
The applicability domain of the selected models was defined according to the indeterminate-zone approach [47,48,49]. Predictions that predicted probabilities falling into indeterminate zones (in-zone predictions) are considered unconfident and vice versa. For binary models, this zone was set to 0.5 ± 0.1 of the prediction probability for corresponding classes. Considering the applicability-domain corrections, the performance of the selected RF:ECFP4 model significantly improved (Table 4 and Figure 5).

2.2.3. Multiclass Classification Models

Being assigned to the same class of sirtuins—Class I—SIRT1 and 3 are the closest homologues to SIRT2, which explains the difficulties in achieving selectivity of SIRT2 inhibitors [50]. With recent links established between the safety of SIRT2 inhibitors and selectivity profiles, achieving the selectivity of novel compounds appears to be a rising trend in the drug-discovery efforts of SIRT2 inhibitors [19]. A notable amount of available structure–activity data on ChEMBL contains molecules with parallel inputs on activity towards different sirtuin isoforms, the most abundant being SIRT1 and 3. However, the size of the subsets of SIRT2 inhibitors with parallel records on SIRT1 or SIRT3 inhibitory activity were still significantly smaller and more imbalanced than the global SIRT2 dataset (Figure 2), which could hamper the predictive power and applicability of models created using these datasets. The main goal of this part of the study was to create and validate models for prediction of the selectivity of the potential inhibitors.
In alignment with the general protocol (Figure 3), all previously mentioned ML algorithms and molecule features were used to build and validate selectivity models. Two types of models were built: a SIRT1/2 selectivity model and a SIRT2/3 selectivity model. Due to the limited number of compounds with concomitant SIRT1/2/3 records accompanied by significant class imbalances, a joint SIRT1/2/3 model was not built. A total of 270 Mordred descriptors was selected for SIRT1/2 selectivity models, whereas 316 Mordred descriptors were selected for SIRT2/3 modeling (Supplementary Materials, Table S3). Fingerprints were used without further reduction of the number of beats. Selectivity models were built as multi-class prediction models, where different classes represented selective SIRT2, non-selective SIRT1/2 or SIRT2/3 and SIRT1/2, or SIRT2/3 inactive compounds (Figure 2). Similar to the VS binary models, external and internal predictive performance of selectivity models was evaluated using the same set of statistical parameters (Figure 6 and the Supplementary Materials, Tables S6–S10). Internal validation parameters indicated good internal predictive performance of the trained model (Supplementary Materials, Table S6). Interestingly, for the selectivity models, models built with descriptors performed significantly better on external validation with the test set compared to the models built with topological fingerprints (Supplementary Materials, Tables S7 and S9), with a predominance of RF, DNN, and SVC ML approaches. It is more likely that physicochemical properties encoded in selected descriptors for each modeling approach played greater importance in selectivity than structural features of the molecules. It is interesting to note that DNN models achieved the greatest performance across all types of molecular features for the SIRT2/3 selectivity model, which indicates the ability of the deep learning approach to more efficiently learn from the limited number of training data (Dataset 4–the smallest dataset) (Figure 6A).
Practical application of the selectivity models could involve making predictions across the large number of inactive compounds. Although models have been trained to recognize inactive compounds, limited chemical space coverage of true inactive compounds of Datasets 3 and 4 may limit the applicability of these models. In order to emulate behavior in real-life application of these models, when facing the large number of inactive compounds, models have been additionally evaluated on decoy datasets (active: inactive = 1:40) similar to binary models. When enriching the inactive class in a decoy dataset, statistical parameters showed a slight shift in the model’s predictive power. Surprisingly, the decoy-set analysis revealed the advantage of ECFP4 molecule representations in the case of the SIRT1/2 models (Figure 6B and Supplementary Materials, Tables S7 and S8). After evaluation of the decoy dataset, the RF:ECFP4 SIRT1/2 model stood out as significantly better compared to the other models. A similar situation was seen in the SIRT2/3 models, where the RF:ECFP4 SIRT2/3 model displayed increased predictive performance on the decoy set. However, in the case of the SIRT2/3 models, the situation was less clear and the DNN:descriptors SIRT2/3 model showed comparable performance (Figure 6B and Supplementary Materials, Tables S9 and S10). It should be noted that the SIRT2/3 models generally performed poorly on the highly imbalanced decoy datasets created by maximization of 2D topological dissimilarity of decoy compounds. This limits usage of SIRT2/3 models only on compounds topologically similar to the already-known active compounds. Generally poor performance of SIRT2/3 models on the decoy dataset could be attributed to the limited size and diversity of the dataset used for training.
In order to further explore the applicability of the selected models, applicability domains were defined. Similar to the binary models, the indeterminate-zone approach was used to define AD for the selectivity models. Since our models have three classes as outcomes, the AD zone where predictions are considered confident was defined as >0.5 for the probability of the predicted class. Considering only the data points within AD, the greatest improvement in predictive statistics was observed in the case of the SIRT1/2 model (Figure 6B and Supplementary Materials, Table S8). On the other hand, two promising SIRT2/3 models with similar performances on the decoy dataset (RF:ECFP4 SIRT2/3 and DNN:descriptors SIRT2/3) displayed just slight improvements after AD corrections (Figure 6B and Supplementary Materials, Table S10). However, the DNN model had much better coverage, including almost 19,000 compounds (compared to 9000 for RF) within the AD borders, so it was retained for further work.
In summary, the SIRT1/2 model showed excellent predictive power, whereas the SIRT2/3 models demonstrated lower quality on topologically dissimilar compounds. The most probable explanation for this result may be in the size differences of the utilized datasets. Considering the abovementioned limitations in the dataset’s size and diversity as well as the quality of the models, selectivity models SIRT1/2 and SIRT2/3 are the most appropriate to be applied as tools for additional selectivity analysis of the virtual screening results where the activity of the compounds is already predicted with the more accurate binary models. Conflicting predictions (e.g., when binary models judge a compound as active, whereas selectivity models judge a compound as inactive) inside the applicability domain of the models should be addressed with caution, and in that context selectivity models (specifically, the SIRT1/2 model) could be used as additional confirmation of the compound activity.

2.3. SIRT2i_Predictor’s Framework for Discovery of Novel Inhibitors

In order to increase the availability and ensure the best practice in the application of the created models, a framework for prediction of the activity/selectivity of novel compounds was defined and encoded into the Python-based application named SIRT2i_Predictor. The workflow of SIRT2i_Predictor is presented in Figure 7 and roughly consists of (1) a module selector, (2) a SMILES preprocessor, (3) predictors, and (4) analyzers. With the aim of making SIRT2i_Predictor easily accessible to the wider community, the appealing web-based graphical user interface (GUI) was created as well (Figure 8).
The module selector allows the user to choose between the two modules of the framework: VS and SMILES-Analyzer modules. In the VS module, SIRT2i_Predictor uses CSV files as an input, which contain compounds in SMILES format with or without compound IDs (up to 200 MB in size). SMILES are prepared for predictions automatically by a SMILES pre-processor. The predictor of the VS module relies solely on the binary-classification RF:ECFP4 model. Compared to the binary model, selectivity models, especially the SIRT2/3 model, demonstrated limited utility on the decoy set, which could be attributed to the limited size and the diversity of the training sets (see Section 2.2.3). On the other hand, regression models, trained mostly on the active compounds, may not be the best choice for VS purposes. Due to the limited presence of inactive compounds in the training set (as discussed above in Section 2.2.1), the regression model could represent a valuable analysis tool for detailed analysis of compounds predicted to be active by the binary model. Therefore, the binary RF:ECFP4 classification model, due to having the largest and the most diverse dataset, as well as superior performance in the real-life application, was selected as the primary virtual-screening model. As an output, the VS module of SIRT2i_Predictors creates and writes the table of screened molecules (CSV output file), together with the predicted probabilities and class assignments (“Yes” class if the molecule is predicted to be an inhibitor, “No” class if not) (Figure 8). This utility could represent a great asset in the prioritization of compounds and cost reductions in large-scale in vitro screening campaigns.
The other module, SMILES-Analyzer, is intended for more detailed analysis of the VS results. However, it can be used independently of the VS module for analysis of desirable compounds. The SMILES-Analyzer module requires list of SMILES as an input, which are pasted by the user into the input box (Figure 8). The predictor of this module encompass all four ML models (binary RF:ECFP4 model, regression XGBoost:ECFP4 model, selectivity RF:ECFP4 SIRT1/2, and DNN:descriptors SIRT2/3 models), whereas different analyzers provide the user with detailed reports on the predicted potencies and selectivity of predicted inhibitors (Figure 7). The main objective of the SMILES-Analyzer module is to assist with the analysis and selection of VS results for further experimental evaluation.
Similar to the VS module, the SMILES-Analyzer module could also process a large number of compounds. However, due to the different requirements of each model on the inputs, the preprocessing step of this module requires more computational time, which could be problematic for screening very large databases. In addition to the text and numerical summaries of predictions for compounds made by all four ML models, the SMILES-Analyzer module allows for the detailed inspection of each compound independently through the graphical interpretations of the predictions (Figure 8). Graphical interpretation of the predictions includes radar-chart summary of predictions made by all four models, a histogram of predicted probabilities for estimation of the confidences (applicability domain) of classification-model predictions, a leverage plot for estimation of the confidence (applicability domain) of pIC50 predictions, and the maps of the per-atom contributions (positive and negative) for predicted SIRT2 activity (with regression and binary models) (Figure 8). Atomic-contribution maps are built according to the concept of similarity maps, where the contribution of each atom is calculated with respect to the differences in predicted probabilities obtained when the bits in the fingerprint corresponding to the atom are removed [51]. Additionally, GUI provides the predictions of the most similar compound in the ChEMBL database using Tanimoto similarity and ECFP4 fingerprints. Tanimoto analysis is provided in order to judge the confidence of SIRT2/3 selectivity predictions since the DNN:descriptors SIRT2/3 model displayed limited applicability on the diverse decoy set (see Section 2.2.3). The SMILES-Analyzer module, similar to the VS module, also outputs a tabular report (CSV file) for predictions on all input molecules.

2.4. Benchmarking SIRT2i_Predictor against the Structure-Based VS Approach

The novel structure-based virtual-screening (SBVS) approach relying on alternative conformational states of SIRT2 discovered through computationally intensive simulations of the binding-pocket dynamics was recently published by our group [20]. The utilization of alternative binding-pocket conformational states, besides showing significant improvements in validation metrics compared to the single-structure approach, resulted in the expansion of the chemical space coverage of virtual hits. In order to test SIRT2i_Predictor’s ability to expand the chemical space of virtual hits, we repeated the prospective SBVS campaign from the aforementioned paper encompassing around 200,000 compounds from the SPECS database [52]. Self-organizing maps representing chemical space coverage were constructed and compared between different approaches. SIRT2i_Predictor binary models showed a surprising expansion of chemical space of virtual hits compared to the known chemical space of SIRT2 inhibitors (Figure 9A). Compared to the SBVS approach, the expansion of chemical space was just slightly lower, which places SIRT2i_Predictor as a comparable tool in search of novel chemical scaffolds of inhibitors. To further explore the reach of SIRT2i_Predictor in unexplored portions of chemical space, one of the top-ranked compounds occupying a portion of chemical space not covered either by SBVS or by ChEMBL compounds was singled out and analyzed using the SMILES-Analyzer module. This portion of chemical space was occupied by a thiohydantoin derivative, which was predicted to be a non-selective SIRT2 inhibitor with all probabilities being inside applicability domains (Figure 9B). Structure–activity analysis of SIRT2i_Predictor based on similarity maps revealed the thiohydantoin scaffold as crucial for predicted activity (Figure 9C). To our knowledge, none of the thiohydantoin derivatives had been reported as sirtuin inhibitors, which further corroborated SIRT2i_Predictor’s capacity to provide novel scaffolds of inhibitors and demonstrated its utility in creating structure–activity hypotheses. As expected, the VS module of SIRT2i_Predictor was able to screen 200,000 compounds in a matter of minutes, which was a significant improvement compared to the SBVS campaign, requiring hours.
In our SBVS campaign, nine hit molecules extracted from previously unexplored portions of chemical space were shown to have potency on SIRT2. However, the IC50 values of two lead compounds as well as the Inh% of five compounds were in the “twilight zone” of the binary models (IC50 = 50–90 µM; Inh%@200 µM = 40–80%), whereas only two compounds were true inactive compounds. In order to explore the performance of SIRT2i_Predictor on this group of “twilight zone” compounds from distant portions of the chemical space of known inhibitors, nine hit compounds were analyzed using the VS module of the framework. The results of the predictions were in agreement with the experimental observations (Supplementary Materials, Table S11), where SIRT2i_Predictor predicted that none of the compounds would have IC50 < 50 μM, which was the criterion for classifying a compound as active in binary models. Most of the compounds were assigned as being outside the applicability domain (5 of 9), which corresponds to the “twilight zone” and the fact that the compounds were selected from distant portions of chemical space in the SBVS study. The rest of the compounds were classified as inactive (4 of 9).
Further analysis of the utility of SIRT2i_Predictor was performed by analyzing our small in-house database of compounds (Supplementary Materials, Table S12) [53,54]. Unfortunately, predicted probabilities for binary models for all in-house compounds were outside the defined applicability domain or predicted to be inactive. In order to further justify SIRT2i_Predictor as a valuable tool in processing the VS data, we utilized our previously published SBVS model, and four compounds predicted to be active according to the SBVS model were tested in vitro. Although the compounds showed some level of SIRT2 inhibitory activity (Supplementary Materials, Table S12), all compounds were shown to be poor inhibitors (three were in the “twilight zone” and one was inactive according to the criteria for classification models). These results were in agreement with SIRT2i_Predictor’s estimations, which predicted that the compounds would be either inactive or outside the applicability domain. As a standard inhibitor, we used compound EX-527 (IC50 (SIRT2) = 20 µM) [55]. This compound was not included in the training sets for the binary model of SIRT2i_Predictor. SIRT2i_Predictor clearly distinguished EX-527 from the in-house compounds, whereas the predictions for in-house compounds were in agreement with the experimental results.
In summary, SIRT2i_Predictor demonstrated utility in discarding the compounds of lower potency while retaining comparable coverage of chemical space as some of more computationally demanding SBVS approaches. The presented benchmarking results place the SIRT2i_Predictor as a complementary tool to the SBVS approaches, which can be utilized as standalone virtual-screening tool, but also as an additional fast and convenient filter after virtual or in vitro screening studies to further prioritize the most promising compounds for biological evaluations.

3. Materials and Methods

3.1. Dataset Preparation

The initial database of structures and biological activities was prepared by collecting all records from the ChEMBL database (release 30) with the reported SIRT2 inhibitory activity (expressed as half-maximal inhibitory concentration (IC50) values or percent of inhibition (Inh%)) [56]. Additional compounds were acquired from patent US20160376238A1 [57]. The compounds were extracted from the patent document using ChemDataExtractor (v 1.3.0) software [58]. The raw data were initially divided into the four datasets (Datasets 1–4) according to the intended purpose (see below). After collecting the raw data, Datasets 1–4 were manually curated by removing records with activities reported for anything but inhibition of deacetylation reaction (e.g., different defatty-acylation activities, etc.). Curated datasets were further pre-processed by canonizing the SMILES, removing the duplicates, stripping the salts, and unchanging the molecules. All pre-processing steps were performed using the RDKit (v 2021.03.4) [59]. Before removal, duplicated records with multiple activities were manually inspected, and only records with activity closest to the average within the group were retained. If the same molecule contained IC50 and %Inh records, the record with IC50 values was retained.
For the regression models, the IC50 values were converted into pIC50 (pIC50 = −log10 (IC50)), whereas for the classification models the IC50 values and Inh% were encoded into different classes. In this study, compounds were assigned to the “SIRT2 active” class if IC50 ≤ 50 µM, or Inh% ≥ 80% (if assayed at 200 µM), or Inh% ≥ 70% (if assayed at 100 µM), or Inh% ≥ 60% (if assayed between 50 and 100 µM), or Inh% ≥ 50% (if assayed bellow 50 µM). Compounds were assigned to the SIRT2 “inactive” class if IC50 ≥ 90 µM or Inh% ≤ 40% (if assayed above 100 µM). For multiclass models, the same criteria were applied with the records on SIRT1 and SIRT3 activities. In subsequent modeling steps, Datasets 1–4 were further divided into the training (internal) set (70%) and test (external) set (30%) using the stratified train–test split algorithm of the scikit-learn library (v 1.1.1) [60]. Class imbalance before training the classification models was accessed using the SMOTE (Synthetic Minority Oversampling Technique) algorithm from the imbalanced-learn library (v 0.9.1) [61].

3.2. Calculation of Molecular Features and Feature Selection

After preparation of the dataset, all molecules were encoded using 166 bit-long MACCS key fingerprints, 1024 bit-long extended-connectivity fingerprints (ECFP4 and ECFP6) as implemented in RDKit, or 1613 two-dimensional descriptors calculated by Mordred (v 1.2.0) [35]. The number of descriptors was further reduced for each dataset independently. Firstly, descriptors with zero or “NaN” values were removed. After standardization, low-variance descriptors were removed using the cut-off value 0.1. Correlations across all pairs of descriptors was calculated using Pearson correlation coefficient. Any two descriptors with correlation values higher than 0.9 were regarded as redundant and only one was retained. The final selection of descriptors for the modeling was performed according to recursive feature elimination with cross-validation (CV), as implemented in Python library scikit-learn (v 1.1.1) [60]. A decision tree classifier was used as an estimator, with 10-fold cross-validation. All descriptors selection steps were performed only for training-set compounds.

3.3. Model Building and Evaluation

In this study, five ML algorithms were utilized for model building: random forest (RF), support-vector machines (SVM: support-vector classification (SVC) and support-vector regression (SVR)), k-nearest neighbors (KNN), extreme gradient boosting (XGBoost), and deep neural networks (DNN). Repressors and classifiers for RF, SVM, XGBoost, and KNN models were built using the scikit-learn library with the addition of the XGBoost Python library (v 1.5.1). DNN models were generated using TensorFlow (v 2.9.1) [62]. For RF, SVC (SVR), KNN, and XGBoost hyperparameter tuning was performed with the Bayesian optimizations using the five-fold CV as implemented in the scikit-optimized library (v 0.8.1). For DNN models, hyperparameter optimizations, together with comprehensive optimization of the neural network architectures, was performed with custom Keras Tuner (v 1.1.1) [63] scripts and using the Bayesian optimization with five-fold CV. A detailed list of all hyperparameters optimized, together with objectives and other specificities of ML approaches, are presented in Supplementary Materials, Table S1.
Briefly, three types of ML models were trained on training sets of corresponding datasets: regression-based (Dataset 1), binary classification-based (Dataset 2), and multiclass classification-based models (Datasets 3 and 4). Internal validation of the regression models was performed using the coefficient of determination for the training set (R2) and the cross-validated correlation coefficient (Q2), the root mean square error (RMSE) of fitting the training set (RMSEint), and the RMSE of cross-validation (RMSECV) [64]. Y-scrambling was performed as a part of an internal validation procedure to check the robustness and reliability of top-performing models. Y-scrambling was performed by generating 100 models using randomly shuffled data with the same hyperparameters used for training the initial model. Evaluation of external predictive power of regression-based QSAR models was performed using the set of different validation metrics: determination coefficient of the test set ( R e x t 2 ) (Equation (1)), RMSE of the external set ( R M S E e x t ) (Equation (2)) [64], Q F n 2 metrics ( Q F 1 2 , Q F 2 2 , Q F 3 2 ) (Equations (3)–(5)) [37,65,66], r m 2 metrics ( r m 2 , r ¯ m 2 , Δ r m 2 ) (Equations (6) and (7)) [39,43], and the concordance correlation coefficient (CCC) (Equation (8)) [40]. Formulas for the calculation of internal validation metrics are presented in the Supplementary Materials, Supplementary Note S1.
R e x t 2 = 1 i = 1 n E X T y i y ^ i 2 i = 1 n E X T y i y ¯ 2
R M S E e x t = n = 1 n E X T y i y ^ i 2 n E X T
Q F 1 2 = 1 i = 1 n E X T y i y i ^ 2 i = 1 n E X T ( y i y ¯ T R ) 2
Q F 2 2 = 1 i = 1 n E X T y i y i ^ 2 i = 1 n E X T ( y i y ¯ E X T ) 2
Q F 3 2 = 1 i = 1 n E X T y i y i ^ 2 / n E X T i = 1 n T R ( y i y ¯ T R ) 2 / n T R
r m 2 = r 2 1 r 2 r 0 2
Δ r m 2 = r m 2 r m 2
C C C = 2 i = 1 n E X T y i y ¯ y ^ i y ^ ¯ i = 1 n E X T y i y ¯ 2 + i = 1 n E X T y ^ i y ^ ¯ 2 + n E X T y ¯ i y ^ ¯ 2
In Equations (1)–(8), TR represents the training set, EXT represents the test set (external set), y i represents the experimental data values, y ^ i represents the predicted data values, y ¯ i represents the average of the experimental data values, and y ^ ¯ represents the average of the predicted data values. r 0 2   and r 2 are the determination coefficients of the regression function calculated using the experimental and predicted data of the external set, forcing the regression through the origin ( r 0 2 ) or not ( r 2 ). r m 2 was calculated using the experimental values on the ordinate axis, whereas r m 2 was calculated using the same values on the abscissa. r ¯ m 2 is the average of r m 2 and r m 2 .
h i = x i T X T X 1 x i
h * = 3 m + 1 p
The applicability domain of the created regression-based models was performed according to the leverage method [45]. The leverage values (hi) were computed according to Equation (9), where X is the matrix formed with rows corresponding to the most important descriptors/bits of molecules from the training set and xi is the descriptor/bit vector for a query molecule. Typically, the threshold, h*, is computed with Equation (10), where m is the number of features and p is the number of molecules in the training set. The feature importance was calculated using the permutation importance approach from scikit-learn with 30 repeats.
The classification models were evaluated using the following metrics: balanced accuracy, recall, precision, F1-score, Matthews correlation coefficient (MCC), and area under the ROC (receiver operating characteristics) curve (ROC_AUC). All metrics were derived from confusion matrices created from the number of true-positive (TP), true-negative (TN), false-positive (FP), and false-negative (FN) predictions. Sensitivity (true-positive rate, or recall) (Sensitivity = TP/(TP + FN)) and specificity (true-negative rate) (Specificity = TN/(TN + FP)) reflect the ability of the model to correctly classify a sample as positive (sensitivity), or negative (specificity) considering all positive data points or all negative data points, respectively. Balanced accuracy represents the average of the sensitivity and specificity, which prevents inflated performance estimates in imbalanced datasets (BA  =  (Sensitivity + Specificity)/2). Precision reflects the ability of the classifier to correctly label all positive samples as positive (Precision = TP/(TP + FP)), whereas the F1-score represents the harmonic mean between precision and recall (sensitivity), which summarizes the precision and robustness of the classifier (F1 = 2 × (Precision × Recall)/(Precision + Recall)). MCC could be seen intuitively as the summary of all categories in the confusion matrix (Supplementary Materials, Supplementary Note S1). This balanced metric can be used even if the classes have different sizes. Values of MCC above 0 indicate that the classifier performed well in all four confusion-matrix categories. The ROC curve is created by plotting the fraction of true-positive rates vs. the fraction of true-negative rates at various thresholds. Additionally, the ROC_AUC value reflects the probability of the classifier to rank the randomly chosen positive example higher than the randomly chosen negative example. The values of ROC_AUC, precision, recall and F1 for multiclass models were calculated as macro-averages using a one-vs-rest approach. Additional evaluation of external performance of the classification models was performed by creating a decoy dataset with a DUD-E server, which was further concatenated to the external set [67]. For interpretation of atomic contributions and graphical interpretation of the structure–activity relationships, similarity maps were built according to the RDKit implementation of the approach proposed by Riniker et al. [51]. Chemical space projections of virtual hits from virtual screening of the SPECS database [52] were calculated using self-organizing maps, as described in previous work [20].
In vitro enzymatic evaluation of potency for the in-house library of compounds was performed using the fluorometric assay, as described elsewhere [68]. The percent of inhibition for all compounds was evaluated in triplicate at 200 µM. Details on the performed assay are provided in the Supplementary Materials, Supplementary Note S2.

4. Conclusions

An increasing number of preclinical evidence demonstrates the potential of SIRT2 inhibitors as novel therapeutics for the treatment of a number of age-related disorders. Despite the growing interest in the development of small-molecule SIRT2 inhibitors in the past decade, none of the SIRT2 inhibitors have entered clinical trials. Currently, there is a lack of large-scale and robust structure–activity relationship models for the prediction of SIRT2 inhibitor potency and selectivity, which could greatly reduce the time and cost of developing novel inhibitors. Inspired to facilitate the discovery of novel SIRT2 inhibitors, we collected all of the currently available structure–activity information and developed a set of high-quality machine-learning models for predictions of novel inhibitor potency and selectivity. After extensive validation of the external predictive power of the models, four models were singled out: the binary RF:ECFP4 model and the regression XGBoost:ECFP4 model for the prediction of inhibitor potencies, as well as the RF:ECFP4 SIRT1/2 and DNN:descriptors SIRT2/3 models for the prediction of inhibitor selectivity. To provide the best practice in the application of the created models, a framework for the prediction of the activity/selectivity of novel compounds was defined and encoded into the Python-based application named SIRT2i_Predictor. SIRT2i_Predictor was equipped with an appealing easy-to-use web-based graphical user interface, which was aimed at enabling usage of the framework for the wider community. With automatic processing of input format (SMILES) and a demonstrated ability to rapidly and efficiently evaluate large databases of compounds on SIRT2 inhibitory potency and SIRT1–3 selectivity, SIRT2i_Predictor’s main utility is to support virtual-screening campaigns and prioritization of compounds for costly in vitro studies. Visualization functionalities, which allow for inspection of parts of molecules that contribute to the activity, make SIRT2i_Predictor a valuable resource for lead-optimization campaigns as well. Our benchmarking study indicated SIRT2i_Predictor’s complementarity to the recently published SBVS approach. A set of codes generated for database curation, model trainings, and GUI could be generalized on a number of pharmacologically relevant targets as part of the development of wider in silico platforms, which is one of our future directions.
Intending to aid future virtual-screening studies, lead-optimization studies, repurposing studies, or the integration of cheminformatics with omics data under the more complex precision-medicine pipelines, we made SIRT2i_Predictor’s GUI code with trained ML models freely available at https://github.com/echonemanja/SIRT2i_Predictor, accessed on 13 December 2022.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/ph16010127/s1, Table S1: List of hyperparameters and their values and objectives considered for optimization through Bayesian search; Table S2: Optimized hyperparameters for final models included in SIRT2i_Predictor; Table S3: Mordred descriptors selected after feature-selection procedure; Table S4: Results of internal validation (training set (int) and cross-validation (CV)) and external (ext) validation for regression models; Table S5: Results of internal validation (training set (int) and cross-validation (CV)) for binary classification models; Table S6: Results of internal validation (training set (int) and cross-validation (CV)) for multiclass classification models; Table S7: External (test set) validation parameters of the multiclass SIRT1/2 models; Table S8: The predictive performance parameters of the multiclass SIRT1/2 models on the decoy dataset; Table S9: External (test set) validation parameters of the multiclass SIRT2/3 models; Table S10: The predictive performance parameters of the multiclass SIRT2/3 models on the decoy dataset; Table S11: Utility SIRT2i_Predictor for the processing of the previously published structure-based virtual-screening results; Table S12: Results of in vitro evaluation of the in-house compounds; Note S1: Internal-validation parameters; Note S2: Supplementary methods.

Author Contributions

Conceptualization, N.D. and K.N.; methodology, N.D., M.R.-R., N.L. and M.L.-K.; software, N.D.; validation, N.D., M.L.-K. and K.N.; formal analysis, N.D., M.R.-R. and N.L.; investigation, N.D. and M.R.-R.; resources, N.L., M.L.-K. and K.N.; data curation, N.D.; writing—original draft preparation, N.D.; writing—review and editing, N.D., M.R.-R., N.L., M.L.-K. and K.N.; supervision, M.L.-K. and K.N.; funding acquisition, K.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Ministry of Education, Science and Technological Development, Republic of Serbia, through a grant agreement with the University of Belgrade—Faculty of Pharmacy, No: 451-03-68/2022-14/200161.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article and Supplementary Materials. The code for the developed software, together with the machine-learning models generated through the study, are freely available at https://github.com/echonemanja/SIRT2i_Predictor, accessed on 13 December 2022.

Acknowledgments

The authors kindly acknowledge COST-Action CM1406 “Epigenetic Chemical Biology” (EpiChemBio) for supporting this research project. The authors kindly acknowledge Panagiotis Marakos, Nicole Pouli, and Emmanuel Mikros from the Department of Pharmaceutical Chemistry, Faculty of Pharmacy, University of Athens, for providing the molecules for in vitro assays.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Finkel, T.; Deng, C.-X.; Mostoslavsky, R. Recent Progress in the Biology and Physiology of Sirtuins. Nature 2009, 460, 587–591. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Haigis, M.C.; Sinclair, D.A. Mammalian Sirtuins: Biological Insights and Disease Relevance. Annu. Rev. Pathol. Mech. Dis. 2010, 5, 253–295. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Saunders, L.R.; Verdin, E. Sirtuins: Critical Regulators at the Crossroads between Cancer and Aging. Oncogene 2007, 26, 5489–5504. [Google Scholar] [CrossRef] [Green Version]
  4. Wang, Y.; Yang, J.; Hong, T.; Chen, X.; Cui, L. SIRT2: Controversy and Multiple Roles in Disease and Physiology. Ageing Res. Rev. 2019, 55, 100961. [Google Scholar] [CrossRef]
  5. Zhang, H.; Dammer, E.B.; Duong, D.M.; Danelia, D.; Seyfried, N.T.; Yu, D.S. Quantitative Proteomic Analysis of the Lysine Acetylome Reveals Diverse SIRT2 Substrates. Sci. Rep. 2022, 12, 3822. [Google Scholar] [CrossRef]
  6. de Oliveira, R.M.; Sarkander, J.; Kazantsev, A.; Outeiro, T. SIRT2 as a Therapeutic Target for Age-Related Disorders. Front. Pharmacol. 2012, 3, 82. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Hong, J.Y.; Lin, H. Sirtuin Modulators in Cellular and Animal Models of Human Diseases. Front. Pharmacol. 2021, 12, 735044. [Google Scholar] [CrossRef] [PubMed]
  8. Zhang, L.; Kim, S.; Ren, X. The Clinical Significance of SIRT2 in Malignancies: A Tumor Suppressor or an Oncogene? Front. Oncol. 2020, 10, 1721. [Google Scholar] [CrossRef] [PubMed]
  9. Jing, H.; Hu, J.; He, B.; Negron Abril, Y.L.; Stupinski, J.; Weiser, K.; Carbonaro, M.; Chiang, Y.-L.; Southard, T.; Giannakakou, P.; et al. A SIRT2-Selective Inhibitor Promotes c-Myc Oncoprotein Degradation and Exhibits Broad Anticancer Activity. Cancer Cell 2016, 29, 297–310. [Google Scholar] [CrossRef] [Green Version]
  10. Nielsen, A.L.; Rajabi, N.; Kudo, N.; Lundø, K.; Moreno-Yruela, C.; Bæk, M.; Fontenas, M.; Lucidi, A.; Madsen, A.S.; Yoshida, M.; et al. Mechanism-Based Inhibitors of SIRT2: Structure–Activity Relationship, X-Ray Structures, Target Engagement, Regulation of α-Tubulin Acetylation and Inhibition of Breast Cancer Cell Migration. RSC Chem. Biol. 2021, 2, 612–626. [Google Scholar] [CrossRef]
  11. Wawruszak, A.; Luszczki, J.; Czerwonka, A.; Okon, E.; Stepulak, A. Assessment of Pharmacological Interactions between SIRT2 Inhibitor AGK2 and Paclitaxel in Different Molecular Subtypes of Breast Cancer Cells. Cells 2022, 11, 1211. [Google Scholar] [CrossRef] [PubMed]
  12. Karwaciak, I.; Sałkowska, A.; Karaś, K.; Sobalska-Kwapis, M.; Walczak-Drzewiecka, A.; Pułaski, Ł.; Strapagiel, D.; Dastych, J.; Ratajewski, M. SIRT2 Contributes to the Resistance of Melanoma Cells to the Multikinase Inhibitor Dasatinib. Cancers 2019, 11, 673. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Cheng, W.-L.; Chen, K.-Y.; Lee, K.-Y.; Feng, P.-H.; Wu, S.-M. Nicotinic-NAChR Signaling Mediates Drug Resistance in Lung Cancer. J. Cancer 2020, 11, 1125–1140. [Google Scholar] [CrossRef] [Green Version]
  14. Hamaidi, I.; Zhang, L.; Kim, N.; Wang, M.-H.; Iclozan, C.; Fang, B.; Liu, M.; Koomen, J.M.; Berglund, A.E.; Yoder, S.J.; et al. Sirt2 Inhibition Enhances Metabolic Fitness and Effector Functions of Tumor-Reactive T Cells. Cell Metab. 2020, 32, 420–436.e12. [Google Scholar] [CrossRef]
  15. Ružić, D.; Đoković, N.; Nikolić, K.; Vujić, Z. Medicinal Chemistry of Histone Deacetylase Inhibitors. Arh. Farm. 2021, 71, 73–100. [Google Scholar] [CrossRef]
  16. Yang, W.; Chen, W.; Su, H.; Li, R.; Song, C.; Wang, Z.; Yang, L. Recent Advances in the Development of Histone Deacylase SIRT2 Inhibitors. RSC Adv. 2020, 10, 37382–37390. [Google Scholar] [CrossRef]
  17. Sauve, A.A.; Youn, D.Y. Sirtuins: NAD+-Dependent Deacetylase Mechanism and Regulation. Curr. Opin. Chem. Biol. 2012, 16, 535–543. [Google Scholar] [CrossRef] [PubMed]
  18. Wang, Y.; He, J.; Liao, M.; Hu, M.; Li, W.; Ouyang, H.; Wang, X.; Ye, T.; Zhang, Y.; Ouyang, L. An Overview of Sirtuins as Potential Therapeutic Target: Structure, Function and Modulators. Eur. J. Med. Chem. 2019, 161, 48–77. [Google Scholar] [CrossRef]
  19. Hong, J.Y.; Fernandez, I.; Anmangandla, A.; Lu, X.; Bai, J.J.; Lin, H. Pharmacological Advantage of SIRT2-Selective versus Pan-SIRT1-3 Inhibitors. ACS Chem. Biol. 2021, 16, 1266–1275. [Google Scholar] [CrossRef]
  20. Djokovic, N.; Ruzic, D.; Rahnasto-Rilla, M.; Srdic-Rajic, T.; Lahtela-Kakkonen, M.; Nikolic, K. Expanding the Accessible Chemical Space of SIRT2 Inhibitors through Exploration of Binding Pocket Dynamics. J. Chem. Inf. Model. 2022, 62, 2571–2585. [Google Scholar] [CrossRef]
  21. Carracedo-Reboredo, P.; Liñares-Blanco, J.; Rodríguez-Fernández, N.; Cedrón, F.; Novoa, F.J.; Carballal, A.; Maojo, V.; Pazos, A.; Fernandez-Lozano, C. A Review on Machine Learning Approaches and Trends in Drug Discovery. Comput. Struct. Biotechnol. J. 2021, 19, 4538–4558. [Google Scholar] [CrossRef]
  22. Qian, T.; Zhu, S.; Hoshida, Y. Use of Big Data in Drug Development for Precision Medicine: An Update. Expert Rev. Precis. Med. Drug Dev. 2019, 4, 189–200. [Google Scholar] [CrossRef] [PubMed]
  23. Zhu, H. Big Data and Artificial Intelligence Modeling for Drug Discovery. Annu. Rev. Pharmacol. Toxicol. 2020, 60, 573–589. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Roy, K.; Kar, S.; Das, R.N. A Primer on QSAR/QSPR Modeling; SpringerBriefs in Molecular Science; Springer International Publishing: Cham, Switzerland, 2015; ISBN 978-3-319-17280-4. [Google Scholar]
  25. Cherkasov, A.; Muratov, E.N.; Fourches, D.; Varnek, A.; Baskin, I.I.; Cronin, M.; Dearden, J.; Gramatica, P.; Martin, Y.C.; Todeschini, R.; et al. QSAR Modeling: Where Have You Been? Where Are You Going To? J. Med. Chem. 2014, 57, 4977–5010. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Dixon, S.L.; Duan, J.; Smith, E.; Von Bargen, C.D.; Sherman, W.; Repasky, M.P. AutoQSAR: An Automated Machine Learning Tool for Best-Practice Quantitative Structure-Activity Relationship Modeling. Future Med. Chem. 2016, 8, 1825–1839. [Google Scholar] [CrossRef] [PubMed]
  27. Gramatica, P. Principles of QSAR Modeling: Comments and Suggestions from Personal Experience. IJQSPR 2020, 5, 61–97. [Google Scholar] [CrossRef]
  28. Bosc, N.; Atkinson, F.; Felix, E.; Gaulton, A.; Hersey, A.; Leach, A.R. Large Scale Comparison of QSAR and Conformal Prediction Methods and Their Applications in Drug Discovery. J. Cheminform. 2019, 11, 4. [Google Scholar] [CrossRef] [Green Version]
  29. Suvannang, N.; Preeyanon, L.; Ahmad Malik, A.; Schaduangrat, N.; Shoombuatong, W.; Worachartcheewan, A.; Tantimongcolwat, T.; Nantasenamat, C. Probing the Origin of Estrogen Receptor Alpha Inhibition via Large-Scale QSAR Study. RSC Adv. 2018, 8, 11344–11356. [Google Scholar] [CrossRef] [Green Version]
  30. Zakharov, A.V.; Zhao, T.; Nguyen, D.-T.; Peryea, T.; Sheils, T.; Yasgar, A.; Huang, R.; Southall, N.; Simeonov, A. Novel Consensus Architecture to Improve Performance of Large-Scale Multitask Deep Learning QSAR Models. J. Chem. Inf. Model. 2019, 59, 4613–4624. [Google Scholar] [CrossRef]
  31. Li, S.; Ding, Y.; Chen, M.; Chen, Y.; Kirchmair, J.; Zhu, Z.; Wu, S.; Xia, J. HDAC3i-Finder: A Machine Learning-Based Computational Tool to Screen for HDAC3 Inhibitors. Mol. Inform. 2021, 40, e2000105. [Google Scholar] [CrossRef]
  32. Li, R.; Tian, Y.; Yang, Z.; Ji, Y.; Ding, J.; Yan, A. Classification Models and SAR Analysis on HDAC1 Inhibitors Using Machine Learning Methods. Mol. Divers. 2022. [Google Scholar] [CrossRef]
  33. Machado, L.A.; Krempser, E.; Guimarães, A.C.R. A Machine Learning-Based Virtual Screening for Natural Compounds Capable of Inhibiting the HIV-1 Integrase. Front. Drug Discov. 2022, 2, 954911. [Google Scholar] [CrossRef]
  34. Lipinski, C.A. Drug-like Properties and the Causes of Poor Solubility and Poor Permeability. J. Pharmacol. Toxicol. Methods 2000, 44, 235–249. [Google Scholar] [CrossRef] [PubMed]
  35. Moriwaki, H.; Tian, Y.-S.; Kawashita, N.; Takagi, T. Mordred: A Molecular Descriptor Calculator. J. Cheminform. 2018, 10, 4. [Google Scholar] [CrossRef] [Green Version]
  36. Tropsha, A.; Golbraikh, A. Predictive QSAR Modeling Workflow, Model Applicability Domains, and Virtual Screening. Curr. Pharm. Des. 2007, 13, 3494–3504. [Google Scholar] [CrossRef]
  37. Validation of (Q)SAR Models—OECD. Available online: https://www.oecd.org/chemicalsafety/risk-assessment/validationofqsarmodels.htm (accessed on 20 August 2022).
  38. Czub, N.; Pacławski, A.; Szlęk, J.; Mendyk, A. Do AutoML-Based QSAR Models Fulfill OECD Principles for Regulatory Assessment? A 5-HT1A Receptor Case. Pharmaceutics 2022, 14, 1415. [Google Scholar] [CrossRef] [PubMed]
  39. Ojha, P.K.; Mitra, I.; Das, R.N.; Roy, K. Further Exploring Rm2 Metrics for Validation of QSPR Models. Chemom. Intell. Lab. Syst. 2011, 107, 194–205. [Google Scholar] [CrossRef]
  40. Chirico, N.; Gramatica, P. Real External Predictivity of QSAR Models: How to Evaluate It? Comparison of Different Validation Criteria and Proposal of Using the Concordance Correlation Coefficient. J. Chem. Inf. Model. 2011, 51, 2320–2335. [Google Scholar] [CrossRef]
  41. Roy, K.; Das, R.N.; Ambure, P.; Aher, R.B. Be Aware of Error Measures. Further Studies on Validation of Predictive QSAR Models. Chemom. Intell. Lab. Syst. 2016, 152, 18–33. [Google Scholar] [CrossRef]
  42. Consonni, V.; Todeschini, R.; Ballabio, D.; Grisoni, F. On the Misleading Use of Q2F3 for QSAR Model Comparison. Mol. Inform. 2019, 38, e1800029. [Google Scholar] [CrossRef]
  43. Roy, K.; Chakraborty, P.; Mitra, I.; Ojha, P.K.; Kar, S.; Das, R.N. Some Case Studies on Application of “Rm2” Metrics for Judging Quality of Quantitative Structure–Activity Relationship Predictions: Emphasis on Scaling of Response Data. J. Comput. Chem. 2013, 34, 1071–1082. [Google Scholar] [CrossRef] [PubMed]
  44. Chirico, N.; Gramatica, P. Real External Predictivity of QSAR Models. Part 2. New Intercomparable Thresholds for Different Validation Criteria and the Need for Scatter Plot Inspection. J. Chem. Inf. Model. 2012, 52, 2044–2058. [Google Scholar] [CrossRef] [PubMed]
  45. Sahigara, F.; Mansouri, K.; Ballabio, D.; Mauri, A.; Consonni, V.; Todeschini, R. Comparison of Different Approaches to Define the Applicability Domain of QSAR Models. Molecules 2012, 17, 4791–4810. [Google Scholar] [CrossRef] [Green Version]
  46. Truchon, J.-F.; Bayly, C.I. Evaluating Virtual Screening Methods:  Good and Bad Metrics for the “Early Recognition” Problem. J. Chem. Inf. Model. 2007, 47, 488–508. [Google Scholar] [CrossRef]
  47. Kuhn, M.; Johnson, K. Applied Predictive Modeling; Springer: New York, NY, USA, 2013; ISBN 978-1-4614-6848-6. [Google Scholar]
  48. Li, X.; Kleinstreuer, N.C.; Fourches, D. Hierarchical Quantitative Structure-Activity Relationship Modeling Approach for Integrating Binary, Multiclass, and Regression Models of Acute Oral Systemic Toxicity. Chem. Res. Toxicol. 2020, 33, 353–366. [Google Scholar] [CrossRef]
  49. Klingspohn, W.; Mathea, M.; ter Laak, A.; Heinrich, N.; Baumann, K. Efficiency of Different Measures for Defining the Applicability Domain of Classification Models. J. Cheminform. 2017, 9, 44. [Google Scholar] [CrossRef]
  50. Costantini, S.; Sharma, A.; Raucci, R.; Costantini, M.; Autiero, I.; Colonna, G. Genealogy of an Ancient Protein Family: The Sirtuins, a Family of Disordered Members. BMC Evol. Biol. 2013, 13, 60. [Google Scholar] [CrossRef]
  51. Riniker, S.; Landrum, G.A. Similarity Maps—A Visualization Strategy for Molecular Fingerprints and Machine-Learning Methods. J. Cheminform. 2013, 5, 43. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  52. Specs. Compound Management Services and Research Compounds for the Life Science Industry. Available online: https://www.specs.net/ (accessed on 8 January 2019).
  53. Lougiakis, N.; Gavriil, E.-S.; Kairis, M.; Sioupouli, G.; Lambrinidis, G.; Benaki, D.; Krypotou, E.; Mikros, E.; Marakos, P.; Pouli, N.; et al. Design and Synthesis of Purine Analogues as Highly Specific Ligands for FcyB, a Ubiquitous Fungal Nucleobase Transporter. Bioorgan. Med. Chem. 2016, 24, 5941–5952. [Google Scholar] [CrossRef]
  54. Sklepari, M.; Lougiakis, N.; Papastathopoulos, A.; Pouli, N.; Marakos, P.; Myrianthopoulos, V.; Robert, T.; Bach, S.; Mikros, E.; Ruchaud, S. Synthesis, Docking Study and Kinase Inhibitory Activity of a Number of New Substituted Pyrazolo [3,4-c]Pyridines. Chem. Pharm. Bull. 2017, 65, 66–81. [Google Scholar] [CrossRef]
  55. Blum, C.A.; Ellis, J.L.; Loh, C.; Ng, P.Y.; Perni, R.B.; Stein, R.L. SIRT1 Modulation as a Novel Approach to the Treatment of Diseases of Aging. J. Med. Chem. 2011, 54, 417–432. [Google Scholar] [CrossRef] [PubMed]
  56. Mendez, D.; Gaulton, A.; Bento, A.P.; Chambers, J.; De Veij, M.; Félix, E.; Magariños, M.P.; Mosquera, J.F.; Mutowo, P.; Nowotka, M.; et al. ChEMBL: Towards Direct Deposition of Bioassay Data. Nucleic Acids Res. 2019, 47, D930–D940. [Google Scholar] [CrossRef] [PubMed]
  57. Chen, L.; Ai, T.; More, S. Therapeutic Compounds 2016. French Patent WO2016140978A1, 9 September 2016. [Google Scholar]
  58. Swain, M.C.; Cole, J.M. ChemDataExtractor: A Toolkit for Automated Extraction of Chemical Information from the Scientific Literature. J. Chem. Inf. Model. 2016, 56, 1894–1904. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  59. Landrum, G. RDKit. Available online: http://rdkit.org (accessed on 15 April 2022).
  60. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-Learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  61. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic Minority Over-Sampling Technique. JAIR 2002, 16, 321–357. [Google Scholar] [CrossRef]
  62. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems. arXiv 2016, arXiv:1603.04467. [Google Scholar]
  63. O’Malley, T.; Bursztein, E.; Long, J.; Chollet, F.; Jin, H.; Invernizzi, L. Others Keras Tuner. 2019. Available online: https://keras.io/keras_tuner/ (accessed on 30 April 2022).
  64. Gramatica, P. On the Development and Validation of QSAR Models. Methods Mol. Biol. 2013, 930, 499–526. [Google Scholar] [CrossRef]
  65. Consonni, V.; Ballabio, D.; Todeschini, R. Comments on the Definition of the Q2 Parameter for QSAR Validation. J. Chem. Inf. Model. 2009, 49, 1669–1678. [Google Scholar] [CrossRef]
  66. Schüürmann, G.; Ebert, R.-U.; Chen, J.; Wang, B.; Kühne, R. External Validation and Prediction Employing the Predictive Squared Correlation Coefficient Test Set Activity Mean vs Training Set Activity Mean. J. Chem. Inf. Model. 2008, 48, 2140–2145. [Google Scholar] [CrossRef]
  67. Mysinger, M.M.; Carchia, M.; Irwin, J.J.; Shoichet, B.K. Directory of Useful Decoys, Enhanced (DUD-E): Better Ligands and Decoys for Better Benchmarking. J. Med. Chem. 2012, 55, 6582–6594. [Google Scholar] [CrossRef]
  68. Heger, V.; Tyni, J.; Hunyadi, A.; Horáková, L.; Lahtela-Kakkonen, M.; Rahnasto-Rilla, M. Quercetin Based Derivatives as Sirtuin Inhibitors. Biomed. Pharmacother. 2019, 111, 1326–1333. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Summary of the sirtuin structures and catalytic mechanism. (A) Two domains of sirtuins exemplified on the structure of SIRT3 (PDB ID: 4FVT). NAD+ and substrate are presented in green sticks; (B) overview of the mechanism of sirtuin-mediated deacetylation; (C) problem of achieving sirtuin inhibitor selectivity exemplified through aligned structures of SIRT1 (yellow) (PDB ID: 4I5I), SIRT2 (pink) (PDB ID: 5D7P), and SIRT3 (gray) (PDB ID: 4BV3) (some parts omitted for clarity). Structurally related inhibitors (gray, pink, or yellow sticks) share the same binding mode across all isoforms. NAD+ and ADP–ribose are presented in gray, pink, or yellow lines.
Figure 1. Summary of the sirtuin structures and catalytic mechanism. (A) Two domains of sirtuins exemplified on the structure of SIRT3 (PDB ID: 4FVT). NAD+ and substrate are presented in green sticks; (B) overview of the mechanism of sirtuin-mediated deacetylation; (C) problem of achieving sirtuin inhibitor selectivity exemplified through aligned structures of SIRT1 (yellow) (PDB ID: 4I5I), SIRT2 (pink) (PDB ID: 5D7P), and SIRT3 (gray) (PDB ID: 4BV3) (some parts omitted for clarity). Structurally related inhibitors (gray, pink, or yellow sticks) share the same binding mode across all isoforms. NAD+ and ADP–ribose are presented in gray, pink, or yellow lines.
Pharmaceuticals 16 00127 g001
Figure 2. Descriptive statistics of the datasets used in the study. (A) PCA analysis of the chemical space of the datasets. PCA plots were calculated in accordance with the descriptors/fingerprints of the final ML models. (B) Distribution of data within each dataset. (C) Decomposition of datasets onto the Lipinski’s rule of 5.
Figure 2. Descriptive statistics of the datasets used in the study. (A) PCA analysis of the chemical space of the datasets. PCA plots were calculated in accordance with the descriptors/fingerprints of the final ML models. (B) Distribution of data within each dataset. (C) Decomposition of datasets onto the Lipinski’s rule of 5.
Pharmaceuticals 16 00127 g002
Figure 3. Overall design of the protocol used for generation and validation of the ML models.
Figure 3. Overall design of the protocol used for generation and validation of the ML models.
Pharmaceuticals 16 00127 g003
Figure 4. Top promising regression-based XGBoost:ECFP4 model. (A) Plot of experimental vs. predicted pIC50 values; (B) results of Y-scrambling; (C) applicability domain of the model. Dashed line indicates the leverage-threshold value (h*).
Figure 4. Top promising regression-based XGBoost:ECFP4 model. (A) Plot of experimental vs. predicted pIC50 values; (B) results of Y-scrambling; (C) applicability domain of the model. Dashed line indicates the leverage-threshold value (h*).
Pharmaceuticals 16 00127 g004
Figure 5. ROC curves obtained for the top-performing RF:ECFP4 binary model. (A) Results of external validation; (B) results of external validation on the decoy dataset; (C) results of external validation on the decoy dataset after applicability-domain corrections.
Figure 5. ROC curves obtained for the top-performing RF:ECFP4 binary model. (A) Results of external validation; (B) results of external validation on the decoy dataset; (C) results of external validation on the decoy dataset after applicability-domain corrections.
Pharmaceuticals 16 00127 g005
Figure 6. Summary of the predictive performances for multiclass selectivity models presenting the top-performing model across each feature type. (A) Validation parameters obtained for the external (test) set; (B) validation parameters obtained on the decoy set. Values of precision, recall and F1 are calculated as macro averages.
Figure 6. Summary of the predictive performances for multiclass selectivity models presenting the top-performing model across each feature type. (A) Validation parameters obtained for the external (test) set; (B) validation parameters obtained on the decoy set. Values of precision, recall and F1 are calculated as macro averages.
Pharmaceuticals 16 00127 g006
Figure 7. SIRT2i_Predictor framework.
Figure 7. SIRT2i_Predictor framework.
Pharmaceuticals 16 00127 g007
Figure 8. Preview of the main functionalities of SIRT2i_Predictor.
Figure 8. Preview of the main functionalities of SIRT2i_Predictor.
Pharmaceuticals 16 00127 g008
Figure 9. Comparison of SIRT2i_Predictor to the multi-structure SBVS approach. (A) Comparisons of the chemical space covered by ChEMBL SIRT2 inhibitors (left), SBVS virtual hits (middle), and SIRT2i_Predictor virtual hits (right); (B) analysis of probabilities obtained for virtual hits extracted from a unique portion of chemical space covered by SIRT2i_Predictor (star sign in (A)); (C) structure–activity relationships calculated using similarity maps for a binary model (upper plot) and a regression model (lower plot) for virtual hits extracted from a unique portion of chemical space (star sign in (A)). A red glow indicates a region that has a positive influence on the activity, whereas a blue glow indicates a negative influence.
Figure 9. Comparison of SIRT2i_Predictor to the multi-structure SBVS approach. (A) Comparisons of the chemical space covered by ChEMBL SIRT2 inhibitors (left), SBVS virtual hits (middle), and SIRT2i_Predictor virtual hits (right); (B) analysis of probabilities obtained for virtual hits extracted from a unique portion of chemical space covered by SIRT2i_Predictor (star sign in (A)); (C) structure–activity relationships calculated using similarity maps for a binary model (upper plot) and a regression model (lower plot) for virtual hits extracted from a unique portion of chemical space (star sign in (A)). A red glow indicates a region that has a positive influence on the activity, whereas a blue glow indicates a negative influence.
Pharmaceuticals 16 00127 g009
Table 1. Description of the datasets.
Table 1. Description of the datasets.
Dataset 1Dataset 2Dataset 3Dataset 4
No. of compounds10021797984612
Expressed activitypIC50pIC50 and Inh%Inh%Inh%
Activity towardsSIRT2SIRT2SIRT1, SIRT2SIRT2, SIRT3
Encoded activitypIC50Active, inactiveSelective, nonselective, inactiveSelective, nonselective, inactive
Table 2. External validation parameters of regression QSAR models.
Table 2. External validation parameters of regression QSAR models.
ML AlgorithmMolecular Feature R e x t 2 RMSEext r ¯ m 2   Δ r m 2 Q F 1 2 Q F 2 2   Q F 3 2 CCC
RFDescriptors0.70.550.520.270.70.70.70.81
ECFP40.750.50.60.230.750.750.750.85 a
MACCS0.710.530.550.260.710.710.710.82
ECFP60.770.480.620.210.770.770.760.86 a
SVRDescriptors0.620.610.440.310.620.620.620.77
ECFP40.740.510.630.13 *0.740.740.730.84
MACCS0.680.570.550.210.680.680.680.81
ECFP60.740.510.630.18 *0.740.740.740.86 a
XGBoostDescriptors0.670.580.530.250.680.680.680.82
ECFP40.75
(0.79) b
0.5
(0.46) b
0.64
(0.7) b
0.17
(0.17) *,b
0.74
(0.75) b
0.74
(0.75) b
0.74
(0.74) b
0.86
(0.86) a,b
MACCS0.710.530.580.240.70.70.70.82
ECFP60.730.520.620.20.730.730.730.87 a
KNNDescriptors0.680.560.560.230.680.680.680.86 a
ECFP40.740.510.640.13 *0.740.740.740.87 a
MACCS0.60.630.470.16 *0.60.60.60.79
ECFP60.76
(0.77) b
0.49
(0.48) b
0.66
(0.68) b
0.12
(0.11) *,b
0.76
(0.76) b
0.76
(0.76) b
0.76
(0.76) b
0.87
(0.87) a,b
DNNDescriptors0.660.580.570.03 *0.660.660.660.81
ECFP40.740.510.630.18 *0.730.730.730.84
MACCS0.680.560.560.16 *0.680.680.670.80
ECFP60.730.520.630.17 *0.730.730.730.81
Criteria>0.6 >0.5<0.2>0.7>0.7>0.7>0.85
* Marks the values within the threshold Δ r m 2 < 0.2. a Marks the values above the proposed threshold CCC > 0.85. b Number in brackets represent values after applicability domain corrections.
Table 3. External validation parameters of the binary classification models.
Table 3. External validation parameters of the binary classification models.
ML AlgorithmMolecular FeatureBAMCCROC_AUCPrecision aRecall aF1 a
RFDescriptors0.880.740.940.860.880.87
ECFP40.840.660.920.820.840.83
MACCS0.820.620.910.80.820.81
ECFP60.850.680.920.830.850.84
SVRDescriptors0.880.740.950.870.880.87
ECFP40.810.630.90.820.810.82
MACCS0.80.590.870.790.80.79
ECFP60.790.620.90.830.790.81
XGBoostDescriptors0.860.720.940.850.860.85
ECFP40.810.620.910.80.810.81
MACCS0.80.60.90.80.80.8
ECFP60.810.620.910.810.810.81
KNNDescriptors0.790.560.880.770.790.77
ECFP40.820.620.90.80.820.81
MACCS0.820.620.880.80.820.81
ECFP60.840.650.910.810.840.82
DNNDescriptors0.890.750.940.850.860.86
ECFP40.830.650.910.80.810.8
MACCS0.80.580.890.80.80.8
ECFP60.820.640.90.790.820.8
a Average values across classes are presented.
Table 4. The predictive performance parameters of the binary models on the decoy dataset.
Table 4. The predictive performance parameters of the binary models on the decoy dataset.
ML AlgorithmMolecular FeatureBAMCCROC_AUCPrecision aRecall aF1 aEF05%EF1%EF2%EF5%
RFDescriptors0.680.090.870.510.680.350.630.670.680.73
ECFP40.81 (0.9) b0.19 (0.52) b0.87 (0.89) b0.53 (0.67) b0.81 (0.9) b0.49 (0.73) b0.74 (0.74) b0.74 (0.74) b0.76 (0.76) b0.77 (0.8) b
MACCS0.660.080.820.510.660.350.550.560.590.62
ECFP60.750.140.870.520.750.430.720.740.760.78
SVRDescriptors0.690.10.890.510.690.360.430.560.620.71
ECFP40.46−0.060.80.480.460.050.750.750.750.76
MACCS0.620.060.830.510.620.320.390.610.680.74
ECFP60.47−0.070.80.470.470.030.760.760.770.77
XGBoostDescriptors0.710.110.850.510.710.390.410.440.480.54
ECFP40.740.130.870.520.740.420.350.390.430.52
MACCS0.640.070.730.510.640.32000.020.2
ECFP60.710.110.850.510.710.390.370.380.440.5
KNNDescriptors0.660.080.760.510.660.370.090.230.260.29
ECFP40.720.120.80.520.720.410000
MACCS0.640.070.750.510.640.330000
ECFP60.720.110.80.520.720.410000
DNNDescriptors0.720.120.80.510.710.380000
ECFP40.730.130.840.520.730.430.10.250.320.41
MACCS0.690.10.790.510.620.290.040.080.170.23
ECFP60.670.090.810.510.670.380.170.250.340.43
a Average values across classes are presented. b Number in brackets represents value after applicability domain corrections.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Djokovic, N.; Rahnasto-Rilla, M.; Lougiakis, N.; Lahtela-Kakkonen, M.; Nikolic, K. SIRT2i_Predictor: A Machine Learning-Based Tool to Facilitate the Discovery of Novel SIRT2 Inhibitors. Pharmaceuticals 2023, 16, 127. https://doi.org/10.3390/ph16010127

AMA Style

Djokovic N, Rahnasto-Rilla M, Lougiakis N, Lahtela-Kakkonen M, Nikolic K. SIRT2i_Predictor: A Machine Learning-Based Tool to Facilitate the Discovery of Novel SIRT2 Inhibitors. Pharmaceuticals. 2023; 16(1):127. https://doi.org/10.3390/ph16010127

Chicago/Turabian Style

Djokovic, Nemanja, Minna Rahnasto-Rilla, Nikolaos Lougiakis, Maija Lahtela-Kakkonen, and Katarina Nikolic. 2023. "SIRT2i_Predictor: A Machine Learning-Based Tool to Facilitate the Discovery of Novel SIRT2 Inhibitors" Pharmaceuticals 16, no. 1: 127. https://doi.org/10.3390/ph16010127

APA Style

Djokovic, N., Rahnasto-Rilla, M., Lougiakis, N., Lahtela-Kakkonen, M., & Nikolic, K. (2023). SIRT2i_Predictor: A Machine Learning-Based Tool to Facilitate the Discovery of Novel SIRT2 Inhibitors. Pharmaceuticals, 16(1), 127. https://doi.org/10.3390/ph16010127

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop