Next Article in Journal
Exploring The Relative Astringency of Tea Catechins and Distinct Astringent Sensation of Catechins and Flavonol Glycosides via an In Vitro Assay Composed of Artificial Oil Bodies
Previous Article in Journal
Structure–Activity Relationship Development Efforts towards Peripherally Selective Analogs of the Cannabinoid Receptor Partial Agonist BAY 59-3074
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Development of GBRT Model as a Novel and Robust Mathematical Model to Predict and Optimize the Solubility of Decitabine as an Anti-Cancer Drug

by
Walid Kamal Abdelbasset
1,2,*,
Shereen H. Elsayed
3,
Sameer Alshehri
4,
Bader Huwaimel
5,
Ahmed Alobaida
6,
Amal M. Alsubaiyel
7,
Abdulsalam A. Alqahtani
8,
Mohamed A. El Hamd
9,10,*,
Kumar Venkatesan
11,
Kareem M. AboRas
12 and
Mohammed A. S. Abourehab
13,14
1
Department of Health and Rehabilitation Sciences, College of Applied Medical Sciences, Prince Sattam bin Abdulaziz University, P.O. Box 173, Al-Kharj 11942, Saudi Arabia
2
Department of Physical Therapy, Kasr Al-Aini Hospital, Cairo University, Giza 12613, Egypt
3
Department of Rehabilitation Sciences, College of Health and Rehabilitation Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
4
Department of Pharmaceutics and Industrial Pharmacy, College of Pharmacy, Taif University, P.O. Box 11099, Taif 21944, Saudi Arabia
5
Department of Pharmaceutical Chemistry, College of Pharmacy, University of Hail, Hail 81442, Saudi Arabia
6
Department of Pharmaceutics, College of Pharmacy, University of Hail, Hail 81442, Saudi Arabia
7
Department of Pharmaceutics, College of Pharmacy, Qassim University, Buraidah 52571, Saudi Arabia
8
Department of Pharmaceutics, College of Pharmacy, Najran University, Najran 11001, Saudi Arabia
9
Department of Pharmaceutical Sciences, College of Pharmacy, Shaqra University, Al Dwadmi 11961, Saudi Arabia
10
Department of Pharmaceutical Analytical Chemistry, Faculty of Pharmacy, South Valley University, Qena 83523, Egypt
11
Department of Pharmaceutical Chemistry, College of Pharmacy, King Khalid University, Abha 62529, Saudi Arabia
12
Department of Electrical Power and Machines, Faculty of Engineering, Alexandria University, Alexandria 21928, Egypt
13
Department of Pharmaceutics, College of Pharmacy, Umm Al-Qura University, Makkah 21955, Saudi Arabia
14
Department of Pharmaceutics and Industrial Pharmacy, Faculty of Pharmacy, Minia University, Minia 61519, Egypt
*
Authors to whom correspondence should be addressed.
Molecules 2022, 27(17), 5676; https://doi.org/10.3390/molecules27175676
Submission received: 5 July 2022 / Revised: 15 August 2022 / Accepted: 17 August 2022 / Published: 2 September 2022

Abstract

:
The efficient production of solid-dosage oral formulations using eco-friendly supercritical solvents is known as a breakthrough technology towards developing cost-effective therapeutic drugs. Drug solubility is a significant parameter which must be measured before designing the process. Decitabine belongs to the antimetabolite class of chemotherapy agents applied for the treatment of patients with myelodysplastic syndrome (MDS). In recent years, the prediction of drug solubility by applying mathematical models through artificial intelligence (AI) has become known as an interesting topic due to the high cost of experimental investigations. The purpose of this study is to develop various machine-learning-based models to estimate the optimum solubility of the anti-cancer drug decitabine, to evaluate the effects of pressure and temperature on it. To make models on a small dataset in this research, we used three ensemble methods, Random Forest (RFR), Extra Tree (ETR), and Gradient Boosted Regression Trees (GBRT). Different configurations were tested, and optimal hyper-parameters were found. Then, the final models were assessed using standard metrics. RFR, ETR, and GBRT had R2 scores of 0.925, 0.999, and 0.999, respectively. Furthermore, the MAPE metric error rates were 1.423 × 10−1 7.573 × 10−2, and 7.119 × 10−2, respectively. According to these facts, GBRT was considered as the primary model in this paper. Using this method, the optimal amounts are calculated as: P = 380.88 bar, T = 333.01 K, Y = 0.001073.

1. Introduction

Recent studies in the area of clinical pharmacology have necessitated the invention of novel, promising, and environmentally friendly tools to increase the performance of therapeutic drugs [1,2]. In order to achieve this purpose, numerous efforts have been made to develop disparate approaches to reduce the application of potentially detrimental/deleterious organic solvents.
Decitabine (currently sold under the brand name DACOGEN®) is an intravenously administered chemotherapeutic drug, which acts as a nucleoside metabolic inhibitor [3,4,5]. Despite the emergence of some adverse events such as neutropenia, thrombocytopenia, and embryo–fetal toxicity in patients, the drug’s great efficiency in ameliorating MDS has encouraged researchers to use it extensively [6,7,8]. Having long been considered, the most important duty of research and development (R&D) centers of pharmaceutical companies has been to concentrate on the development of supercritical fluid technology (SCFT) [9].
The existence of noteworthy advantages such as negligible processing times, the manufacturing of no organic co-solvent, and its great capability of extracting bioactive molecules has encouraged the researchers to use SCFT for drug discovery from natural sources [10,11,12]. In recent years, CO2 supercritical fluid (CO2SCF) has received more attention within SCFT as an efficient solvent, due to significant privileges such as chemical inactivity, availability, cost-effectiveness, low critical temperature/pressure, and its approval as a food-grade solvent [13].
In recent years, artificial intelligence (AI) has found its place as a versatile tool, offering a high potential for applications in different industries such as separation, extraction, and nanotechnology, as well as drug development, including the identification, validation, and designation of novel drugs [14,15]. The various advantages of AI technology, such as robustness and time-effectiveness, have provided an appropriate chance to overcome the incompetence and discrepancies which may take place during conventional drug optimization and development techniques [16,17]. Machine learning (ML) is a predictive mathematical approach based on AI, which has paved the way to estimate the solubility of drugs in CO2SCF. To increase the generalization and performance of a single model, an ensemble of models is used in ML. Because of the blending of diverse predictions, ensembles generate effective predictive algorithms with enormous generalization capability [18]. Some ML techniques, such as decision tree and linear regression, are inherently unstable, which means that changing the training dataset results in a significantly different estimator. Unstable estimators have a low bias and high variability. Ensemble approaches have been proposed to reduce generalization error, that is, to reduce variance, bias, or both. In these approaches, the training dataset is modified, and an ensemble of different base estimators is created. These estimators are then combined to create a single estimator [19]. This section provides a quick overview of three main ensemble algorithms: bagging, gradient boosting, and Extra Trees.
In this study, a few basic models were first studied. Given that the decision tree gave significantly better outcomes, but these findings were not general enough to be presented as a powerful model backend, it was decided to employ models that reinforce it. Bagging and boosting are known as the most efficient advanced methods with decision trees. Bagging (bootstrap aggregating), created by Breiman et al. [20], can be considered as both a principal approach and a straight ensemble approach, illustrating brilliant efficiency as long as it reduces variance and avoids overfitting. The bootstrap technique, which replicates training datasets and creates training data subsets, contributes to the diversity of the bagging algorithm. Each subset is utilized to fit a various basic estimator, and the ultimate prediction outcomes are compiled by applying a majority vote procedure.
The other ensemble technique which is introduced from the work of Freund and Schapire is boosting [21]. In contrast to bagging, that provided a variety of basic learners by gradually reweighting the training data. Each sample weakly estimated by the previous estimator is given a higher weight in the next training step. As a result, training samples weakly estimated by predecessors are more likely to occur in the following bootstrap sample, and bias can be removed effectively. The final boosting algorithm model integrates all the underlying base estimators, which are weighted using their prediction performance.
There is a recently developed decision tree model called Extremely Random Tree (ExtraTree) that is an improved version of the traditional top-down decision tree (DT) model. There is an ensemble of DTs that has been trained in a recursive manner. The final model is built using a massive DT that is trained in a recursive manner. In each case, the tree must be expand using the entire set of data, and the proper cut point for each split can be calculated through the amount of data gained from each split [22,23].
The three algorithms selected for this study are:
  • Random Forest (Bagging of Regression Trees);
  • Extra Trees (Bagging of Regression Trees);
  • Gradient Boosting (Boosting of Regression Trees).

2. Dataset

Solubility models were created using a dataset with 32 input vectors, similar to [24]. The dataset is illustrated in Table 1. The distribution of features and output is shown in Figure 1. The diagonal subplots (when the x-axis and y-axis are identical) of this figure also show the kernel density estimate (KDE) plot. KDE plots visualize the distribution of observations in a dataset, like histograms. With KDEs, one or more dimensions of probability density curves are used to represent the data.

3. Methodology

3.1. Random Forest Regression (RFR)

This regression method is an ML procedure which estimates the targeted output by combining the results of several DT learners [25,26]. When Random Forest receives an (x) input data point, it includes the amounts of the various input features probed for a given training area, which creates K regression trees and the averages of their results. The RF regression predictor, after such K-trees {T(x)}1K have been trained, is [27]:
f ^ r f K x = 1 K k = 1 K T x
For bypass tree correlation, Random Forest makes the trees grow using different training subsets generated across a routine called bagging. This process is a subset creation technique that involves resampling the original dataset randomly with replacement, in order to generate the next training subset {h (x, Θk), k = 1, 2, …, K}, where each Θk is the same distributed independent random vector. Accordingly, some data can be applied many times during training, while some data points may never be used. By creating a tree using RF, the best split point will be created through a set of input light [28,29,30].
Moreover, the data which are not used in training step in the k-th tree model in the bagging method, they have been utilized in an out-of-bag subset (oob). k-th tree model be able to utilize the oob items to calculate accuracy [29]. Moreover, the non-selected items in the training step of the k-th tree along the bagging routine are considered in the out-of-bag subset (oob). On the other hand, the k-th tree must be able to utilize these oob items in order to evaluate the efficiency. Increasing the quantity of trees results in a reduction in error, which is illustrated by the fact that the Random Forest does not have an overfitting issue. The relative importance of the features is likewise determined using RF. In order to select the best features in multi-source investigations, it is critical to find the relationship between each item and predicted procedure, and this feature can assist with that understanding [30,31].

3.2. Extra Tree Regression (ETR)

Geurts et al. [32] introduced the extremely randomized tree (ExtraTree), which is an improved version of traditional top-down decision tree (DT) models. The ExtraTree is an ensemble of DTs that has been trained in a recursive manner, and the final model was built employing a massive DT. Each develops the tree utilizing the whole dataset, and the suitable cut point for each split can be decided through achieved information [22,23].
This model is very close to the Random Forest and the Extra Tree model’s primary innovations in that (i) the nodes are separated randomly when applying cut points, and (ii) the whole training dataset was used for developing the decision tree instead of subset generation using the bootstrap in Random Forest method [23,33].

3.3. Gradient Boosting Regression Trees (GBRT)

To improve prediction accuracy, boosting uses a series of base estimator compare to a single predictor to get an average between them. Base estimators/models (such as decision trees) are coordinated to clear bias in a stage-wise process. In order to modify the loss function, a new learner is introduced at each phase. Using training data, the first learner decreases the loss function to the lowest amount [34,35,36]. The following estimators make use of the previous estimators’ residuals. The following Algorithm 1 demonstrates the gradient boosting procedure [35,36,37]:
Algorithm 1
Initialize F 0 x = argmin p i 1 N L y i , P
For   m   1 ,   2 ,   ,   M :
   1. Compute the negative gradient
y ¯ i = L y i , F x i F x i
   2. Create a model
a m =   argmin a , β i = 1 N y ¯ β h ( x i , a m 2
   3. Select a gradient descent step size as
p k = argmin p i = 1 N L y i , F m 1 x i + ph x i , a
   4. Modify the estimation of F(x)
F m x = F m 1 x + p k h x , a m
Output: the aggregated regression function F m x
Here, x demonstrates the feature vector and y demonstrates the corresponding class label. {xi, yi}Ni = 1,as training data and the aim is to calculate F*(x), be able to design x to y graph, a specific loss function L(y, F(x)) could be reduced to the lowest amount.

4. Results

After tweaking the hyper-parameters of models by testing various combinations of them, we employed MAPE and R2 [38] to verify the accuracy and generality of the models.
The performance success of estimating findings is frequently measured using R-squared, which is without a doubt the most often utilized criterion. Using this metric, you can see how closely the predicted results match up with the observed data [39].
R 2 = 1 e i o i 2 o i   o ¯ i 2
MAPE is also one of the most used evaluation metrics. MAPE illustrates error size, which is between 0 and 1 [40].
M A P E = 1 n i = 1 n o i e i o i
ei and oi are predicted and actual (observed) values. o ¯ i is the average of the actual data. n indicates the size of the dataset. Comparisons of the actual and estimated values using RFR, ETR, and GBRT ML-based mathematical models are depicted in Figure 2, Figure 3 and Figure 4. In these figures, points are estimated amount (black is used as training and the other one, red can be used as the test) and the line indicates the actual values. The results imply that the GBRT method is the most general and precise model. Consideration of the R2 and RMSE amounts through Table 2 confirms the greater accuracy of the GBRT model with the best generality.
The reason why GBRT is superior to the other two models is because the dataset is small. Therefore, every data point can have a profound impact on the final model. In order to improve air performance, it has been shown that the boosting method, by which the points that are incorrectly predicted are corrected by weighting, is more effective than the conventional method.
Figure 5 illustrates the three-dimensional result based on the GBRT predictive model to simultaneously evaluate the effect of those two parameters, pressure and temperature as input parameters on the solubility of the anti-cancer drug decitabine as the only output. Additionally, Figure 6 and Figure 7 schematically demonstrate two-dimensional variations in pressure and temperature versus decitabine solubility. For all evaluated isotherms, an increase in pressure considerably improves the density of CO2SCF due to an increase in the compaction of molecules. The increment in density resulted in an enhancement in the efficiency of the solvent and, therefore, the solubility value of decitabine in CO2SCF increases. Despite the correlation between pressure and drug solubility, the influence of temperature is not straightforward, and a reverse alteration is seen after the 200-bar pressure point. It is worth pointing out that the solvent density and the sublimation pressure are considered as two competing parameters which entirely affect the effect of temperature on the solubility of the drug. By increasing the temperature, the density of the solvent significantly reduces, since greater molecular energy eventuates in the free movement of solvent molecules. Moreover, enhancement in the temperature of the system can improve the sublimation pressure, which presents a positive influence on drug solubility in a supercritical system. Considering the description, the net effect of the abovementioned competing factors can determine the positive or negative role of temperature on drug solubility. Analysis of the graphs illustrate a threshold pressure, the temperature increment positively encourages drug solubility owing to the significant role of sublimation pressure in comparison to density. Further, in pressure below the cross-over pressure, an increase in the temperature results in a substantial decrement in the decitabine solubility because of the decline in the solvent density. According to Table 3, 380.88 bar and 333.01 K, can be mentioned as the optimum values at highest decitabine solubility.

5. Conclusions

In this study, various predictive models were developed using AI approaches to estimate the optimum solubility value of the anti-cancer drug decitabine inside carbon dioxide supercritical fluid (CO2SCF). We used three ensemble methods to build models on a small dataset: Random Forest (RFR), Extra Tree (ETR), and Gradient Boosted Regression Trees (GBRT). Various configurations were tested, and optimal hyper-parameters were discovered. The final models were then evaluated using industry-standard metrics. RFR, ETR, and GBRT all had R2 scores of 0.925, 0.999, and 0.999, respectively. Furthermore, the MAPE metric error rates were 1.423 × 10−2, 7.573 × 10−2, and 7.119 × 10−2, respectively. GBRT was selected as the primary method for this study through these facts and other visual considerations. The optimal values were calculated using this model as P = 380.88, T = 333.01, and Y = 0.001073.

Author Contributions

W.K.A.: conceptualization, writing, supervision, methodology, S.H.E.: writing, visualization, resources, analysis, S.A.: validation, writing, software, methodology, B.H.: editing, analysis, resources, investigation, A.A.: editing, methodology, investigation, resources, analysis, A.M.A.: supervision, writing, analysis, conceptualization, methodology, A.A.A.: software, analysis, methodology, editing, M.A.E.H.: writing, editing, analysis, visualization, resources, K.V.: analysis, editing, resources, formal analysis, K.M.A.: writing, editing, analysis, visualization, resources, M.A.S.A.: writing, editing, analysis, visualization, resources, investigation. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2022R99), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia. The author would like to thank the Deanship of Scientific Research at Shaqra University for supporting this work. The authors extend their appreciation to the Deanship of Scientific Research at King Khalid University for funding this work through Large Groups Project under grant number (RGP.2/50/43). The authors would like to thank the Deanship of Scientific Research at Umm Al-Qura University for supporting this work by Grant Code: 22UQU4290565DSR61.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All data are available within the published paper.

Acknowledgments

This research was funded by Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2022R99), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia. The author would like to thank the Deanship of Scientific Research at Shaqra University for supporting this work. The authors extend their appreciation to the Deanship of Scientific Research at King Khalid University for funding this work through Large Groups Project under grant number (RGP.2/50/43). The authors would like to thank the Deanship of Scientific Research at Umm Al-Qura University for supporting this work by Grant Code: 22UQU4290565DSR61.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, Q.; Gilbert, J.A.; Zhu, H.; Huang, S.-M.; Kunkoski, E.; Das, P.; Bergman, K.; Buschmann, M.; ElZarrad, M.K. Emerging clinical pharmacology topics in drug development and precision medicine. In Atkinson’s Principles of Clinical Pharmacology; Elsevier: Amsterdam, The Netherlands, 2022; pp. 691–708. [Google Scholar]
  2. Docherty, J.R.; Alsufyani, H.A. Pharmacology of drugs used as stimulants. J. Clin. Pharmacol. 2021, 61, S53–S69. [Google Scholar] [CrossRef] [PubMed]
  3. Gore, S.D.; Jones, C.; Kirkpatrick, P. Decitabine. Nat. Rev. Drug Discov. 2006, 5, 891–893. [Google Scholar] [CrossRef] [PubMed]
  4. Jabbour, E.; Issa, J.P.; Garcia-Manero, G.; Kantarjian, H. Evolution of decitabine development: Accomplishments, ongoing investigations, and future strategies. Cancer Interdiscip. Int. J. Am. Cancer Soc. 2008, 112, 2341–2351. [Google Scholar] [CrossRef]
  5. Saba, H.I. Decitabine in the treatment of myelodysplastic syndromes. Ther. Clin. Risk Manag. 2007, 3, 807. [Google Scholar] [PubMed]
  6. Xiao, J.; Liu, P.; Wang, Y.; Zhu, Y.; Zeng, Q.; Hu, X.; Ren, Z.; Wang, Y. A Novel Cognition of Decitabine: Insights into Immunomodulation and Antiviral Effects. Molecules 2022, 27, 1973. [Google Scholar] [CrossRef]
  7. Senapati, J.; Shoukier, M.; Garcia-Manero, G.; Wang, X.; Patel, K.; Kadia, T.; Ravandi, F.; Pemmaraju, N.; Ohanian, M.; Daver, N. Activity of decitabine as maintenance therapy in core binding factor acute myeloid leukemia. Am. J. Hematol. 2022, 97, 574–582. [Google Scholar] [CrossRef]
  8. Issa, J.-P. Decitabine. Curr. Opin. Oncol. 2003, 15, 446–451. [Google Scholar] [CrossRef]
  9. Bhusnure, O.; Gholve, S.; Giram, P.; Borsure, V.; Jadhav, P.; Satpute, V.; Sangshetti, J. Importance of supercritical fluid extraction techniques in pharmaceutical industry: A Review. Indo Am. J. Pharm. Res. 2015, 5, 3785–3801. [Google Scholar]
  10. Khaw, K.-Y.; Parat, M.-O.; Shaw, P.N.; Falconer, J.R. Solvent supercritical fluid technologies to extract bioactive compounds from natural sources: A review. Molecules 2017, 22, 1186. [Google Scholar] [CrossRef]
  11. Tran, P.; Park, J.-S. Application of supercritical fluid technology for solid dispersion to enhance solubility and bioavailability of poorly water-soluble drugs. Int. J. Pharm. 2021, 610, 121247. [Google Scholar] [CrossRef]
  12. Zhuang, W.; Hachem, K.; Bokov, D.; Javed Ansari, M.; Taghvaie Nakhjiri, A. Ionic liquids in pharmaceutical industry: A systematic review on applications and future perspectives. J. Mol. Liq. 2022, 349, 118145. [Google Scholar] [CrossRef]
  13. Zhang, Q.-W.; Lin, L.-G.; Ye, W.-C. Techniques for extraction and isolation of natural products: A comprehensive review. Chin. Med. 2018, 13, 20. [Google Scholar] [CrossRef] [PubMed]
  14. Huang, Z.; Juarez, J.M.; Li, X. Data Mining for Biomedicine and Healthcare; Hindawi: London, UK, 2017. [Google Scholar]
  15. Zhang, Y.; Zhang, G.; Shang, Q. Computer-aided clinical trial recruitment based on domain-specific language translation: A case study of retinopathy of prematurity. J. Healthc. Eng. 2017, 2017, 7862672. [Google Scholar] [CrossRef] [PubMed]
  16. Seddon, G.; Lounnas, V.; McGuire, R.; van den Bergh, T.; Bywater, R.P.; Oliveira, L.; Vriend, G. Drug design for ever, from hype to hope. J. Comput.-Aided Mol. Des. 2012, 26, 137–150. [Google Scholar] [CrossRef] [PubMed]
  17. Mak, K.-K.; Pichika, M.R. Artificial intelligence in drug development: Present status and future prospects. Drug Discov. Today 2019, 24, 773–780. [Google Scholar] [CrossRef]
  18. Dietterich, T.G. Ensemble methods in machine learning. In Proceedings of the International Workshop on Multiple Classifier Systems, Cagliari, Italy, 21–23 June 2000. [Google Scholar]
  19. Izenman, A.J. Modern Multivariate Statistical Techniques—Regression, Classification, and Manifold Learning; Springer: New York, NY, USA, 2008; Volume 10. [Google Scholar]
  20. Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef]
  21. Freund, Y.; Schapire, R.E. Experiments with a new boosting algorithm. In Proceedings of the 13th International Conference on Machine Learning (ICML’96), Bari, Italy, 3–6 July 1996; pp. 148–156. [Google Scholar]
  22. Kocev, D.; Ceci, M. Ensembles of extremely randomized trees for multi-target regression. In Proceedings of the International Conference on Discovery Science, Banff, AB, Canada, 4–6 October 2015. [Google Scholar]
  23. Wehenkel, L.; Ernst, D.; Geurts, P. Ensembles of extremely randomized trees and some generic applications. In Proceedings of the Workshop “Robust Methods for Power System State Estimation and Load Forecasting”; Paris, France, 29-–30 May 2006.
  24. Pishnamazi, M.; Zabihi, S.; Jamshidian, S.; Borousan, F.; Hezave, A.Z.; Marjani, A.; Shirazian, S. Experimental and thermodynamic modeling decitabine anti cancer drug solubility in supercritical carbon dioxide. Sci. Rep. 2021, 11, 1075. [Google Scholar] [CrossRef]
  25. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  26. Rodriguez-Galiano, V.F.; Ghimire, B.; Rogan, J.; Chica-Olmo, M.; Rigol-Sanchez, J.P. An assessment of the effectiveness of a random forest classifier for land-cover classification. ISPRS J. Photogramm. Remote Sens. 2012, 67, 93–104. [Google Scholar] [CrossRef]
  27. Nguyen Duc, M.; Ho Sy, A.; Nguyen Ngoc, T.; Hoang Thi, T.L. An Artificial Intelligence Approach Based on Multi-layer Perceptron Neural Network and Random Forest for Predicting Maximum Dry Density and Optimum Moisture Content of Soil Material in Quang Ninh Province, Vietnam. In CIGOS 2021, Emerging Technologies and Applications for Green Infrastructure, Proceedings of the 6th International Conference on Geotechnics, Civil Engineering and Structures, Ha Long, Vietnam, 28–29 October 2021; Springer: Berlin/Heidelberg, Germany, 2022; pp. 1745–1754. [Google Scholar]
  28. Li, Y.; Zou, C.; Berecibar, M.; Nanini-Maury, E.; Chan, J.C.-W.; Van den Bossche, P.; Van Mierlo, J.; Omar, N. Random forest regression for online capacity estimation of lithium-ion batteries. Appl. Energy 2018, 232, 197–210. [Google Scholar] [CrossRef]
  29. Peters, J.; De Baets, B.; Verhoest, N.E.; Samson, R.; Degroeve, S.; De Becker, P.; Huybrechts, W. Random forests as a tool for ecohydrological distribution modelling. Ecol. Model. 2007, 207, 304–318. [Google Scholar] [CrossRef]
  30. Verikas, A.; Gelzinis, A.; Bacauskiene, M. Mining data with random forests: A survey and results of new tests. Pattern Recognit. 2011, 44, 330–349. [Google Scholar] [CrossRef]
  31. Gislason, P.O.; Benediktsson, J.A.; Sveinsson, J.R. Random forests for land cover classification. Pattern Recognit. Lett. 2006, 27, 294–300. [Google Scholar] [CrossRef]
  32. Geurts, P.; Ernst, D.; Wehenkel, L. Extremely randomized trees. Mach. Learn. 2006, 63, 3–42. [Google Scholar] [CrossRef]
  33. Ahmad, M.W.; Mourshed, M.; Rezgui, Y. Tree-based ensemble methods for predicting PV power generation and their comparison with support vector regression. Energy 2018, 164, 465–474. [Google Scholar] [CrossRef]
  34. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  35. Mason, L.; Baxter, J.; Bartlett, P.; Frean, M. Boosting algorithms as gradient descent. In Advances in Neural Information Processing Systems 12; MIT Press: Cambridge, MA, USA, 1999; Volume 12. [Google Scholar]
  36. Truong, V.-H.; Vu, Q.-V.; Thai, H.-T.; Ha, M.-H. A robust method for safety evaluation of steel trusses using Gradient Tree Boosting algorithm. Adv. Eng. Softw. 2020, 147, 102825. [Google Scholar] [CrossRef]
  37. Xu, Q.; Xiong, Y.; Dai, H.; Kumari, K.M.; Xu, Q.; Ou, H.-Y.; Wei, D.-Q. PDC-SGB: Prediction of effective drug combinations using a stochastic gradient boosting algorithm. J. Theor. Biol. 2017, 417, 1–7. [Google Scholar] [CrossRef]
  38. Willmott, C.J. Some comments on the evaluation of model performance. Bull. Am. Meteorol. Soc. 1982, 63, 1309–1313. [Google Scholar] [CrossRef]
  39. Gouda, S.G.; Hussein, Z.; Luo, S.; Yuan, Q. Model selection for accurate daily global solar radiation prediction in China. J. Clean. Prod. 2019, 221, 132–144. [Google Scholar] [CrossRef]
  40. Kim, S.; Kim, H. A new metric of absolute percentage error for intermittent demand forecasts. Int. J. Forecast. 2016, 32, 669–679. [Google Scholar] [CrossRef]
Figure 1. Data distribution, P (pressure), T (temperature), and Y (solubility).
Figure 1. Data distribution, P (pressure), T (temperature), and Y (solubility).
Molecules 27 05676 g001
Figure 2. RFR Model: test and train data predictions.
Figure 2. RFR Model: test and train data predictions.
Molecules 27 05676 g002
Figure 3. ETR Model: test and train data predictions.
Figure 3. ETR Model: test and train data predictions.
Molecules 27 05676 g003
Figure 4. GBRT Model: test and train data predictions.
Figure 4. GBRT Model: test and train data predictions.
Molecules 27 05676 g004
Figure 5. Three-dimensional illustration of inputs/outputs (GBRT Model).
Figure 5. Three-dimensional illustration of inputs/outputs (GBRT Model).
Molecules 27 05676 g005
Figure 6. Tendency of P.
Figure 6. Tendency of P.
Molecules 27 05676 g006
Figure 7. Tendency of T.
Figure 7. Tendency of T.
Molecules 27 05676 g007
Table 1. The whole dataset.
Table 1. The whole dataset.
NoX1 = P (Bar)X2 = T (K)Y = Solubility (Mole Fraction)
11203085.04 × 10−5
21203184.51 × 10−5
31203283.69 × 10−5
41203382.84 × 10−5
51603088.23 × 10−5
61603189.37 × 10−5
71603289.11 × 10−5
81603387.79 × 10−5
92003081.18 × 10−4
102003181.55 × 10−4
112003281.77 × 10−4
122003382.05 × 10−4
132403081.37 × 10−4
142403181.87 × 10−4
152403282.82 × 10−4
162403383.71 × 10−4
172803081.76 × 10−4
182803182.40 × 10−4
192803283.42 × 10−4
202803384.90 × 10−4
213203081.97 × 10−4
223203182.69 × 10−4
233203284.27 × 10−4
243203387.15 × 10−4
253603082.18 × 10−4
263603183.40 × 10−4
273603285.60 × 10−4
283603388.74 × 10−4
294003082.83 × 10−4
304003185.06 × 10−4
314003287.88 × 10−4
324003381.07 × 10−3
Table 2. Outputs.
Table 2. Outputs.
ModelsR2 ScoreMAPE
RFR0.9251.423 × 10−1
ETR0.9997.573 × 10−2
GBRT0.9997.119 × 10−2
Table 3. Optimal values (GBRT Model).
Table 3. Optimal values (GBRT Model).
P (Bar)T (K)Y
380.88333.010.001073
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Abdelbasset, W.K.; Elsayed, S.H.; Alshehri, S.; Huwaimel, B.; Alobaida, A.; Alsubaiyel, A.M.; Alqahtani, A.A.; El Hamd, M.A.; Venkatesan, K.; AboRas, K.M.; et al. Development of GBRT Model as a Novel and Robust Mathematical Model to Predict and Optimize the Solubility of Decitabine as an Anti-Cancer Drug. Molecules 2022, 27, 5676. https://doi.org/10.3390/molecules27175676

AMA Style

Abdelbasset WK, Elsayed SH, Alshehri S, Huwaimel B, Alobaida A, Alsubaiyel AM, Alqahtani AA, El Hamd MA, Venkatesan K, AboRas KM, et al. Development of GBRT Model as a Novel and Robust Mathematical Model to Predict and Optimize the Solubility of Decitabine as an Anti-Cancer Drug. Molecules. 2022; 27(17):5676. https://doi.org/10.3390/molecules27175676

Chicago/Turabian Style

Abdelbasset, Walid Kamal, Shereen H. Elsayed, Sameer Alshehri, Bader Huwaimel, Ahmed Alobaida, Amal M. Alsubaiyel, Abdulsalam A. Alqahtani, Mohamed A. El Hamd, Kumar Venkatesan, Kareem M. AboRas, and et al. 2022. "Development of GBRT Model as a Novel and Robust Mathematical Model to Predict and Optimize the Solubility of Decitabine as an Anti-Cancer Drug" Molecules 27, no. 17: 5676. https://doi.org/10.3390/molecules27175676

Article Metrics

Back to TopTop