Next Article in Journal
Proposal for a System Model for Offline Seismic Event Detection in Colombia
Next Article in Special Issue
Knowledge-Graph-Based Drug Repositioning against COVID-19 by Graph Convolutional Network with Attention Mechanism
Previous Article in Journal
Before and after the Quarantine: An Approximate Study on the Psychological Impact of COVID-19 on the Italian Population during the Lockdown Period
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

IgA Nephropathy Prediction in Children with Machine Learning Algorithms

1
School of Mathematics and Statistics, Henan University of Science and Technology, Luoyang 471000, China
2
School of Information Engineering, Zhengzhou University, Zhengzhou 450000, China
3
School of Computer and Information Engineering, Luoyang Institute of Science and Technology, Luoyang 471000, China
*
Author to whom correspondence should be addressed.
Future Internet 2020, 12(12), 230; https://doi.org/10.3390/fi12120230
Submission received: 15 November 2020 / Revised: 11 December 2020 / Accepted: 14 December 2020 / Published: 17 December 2020
(This article belongs to the Special Issue Curative Power of Medical Data 2020)

Abstract

:
Immunoglobulin A nephropathy (IgAN) is the most common primary glomerular disease all over the world and it is a major cause of renal failure. IgAN prediction in children with machine learning algorithms has been rarely studied. We retrospectively analyzed the electronic medical records from the Nanjing Eastern War Zone Hospital, chose eXtreme Gradient Boosting (XGBoost), random forest (RF), CatBoost, support vector machines (SVM), k-nearest neighbor (KNN), and extreme learning machine (ELM) models in order to predict the probability that the patient would not reach or reach end-stage renal disease (ESRD) within five years, used the chi-square test to select the most relevant 16 features as the input of the model, and designed a decision-making system (DMS) of IgAN prediction in children that is based on XGBoost and Django framework. The receiver operating characteristic (ROC) curve was used in order to evaluate the performance of the models and XGBoost had the best performance by comparison. The AUC value, accuracy, precision, recall, and f1-score of XGBoost were 85.11%, 78.60%, 75.96%, 76.70%, and 76.33%, respectively. The XGBoost model is useful for physicians and pediatric patients in providing predictions regarding IgAN. As an advantage, a DMS can be designed based on the XGBoost model to assist a physician to effectively treat IgAN in children for preventing deterioration.

1. Introduction

Immunoglobulin A nephropathy (IgAN) is the most common primary glomerular disease all over the world and it is a major cause of end-stage renal disease (ESRD). Additionally, there are similarities and differences of IgAN in adults and children. Clinical studies showed that the survival rate of IgAN patients over 18 years old was 70–80% within 10 years. Among them, 20–30% patients will progress to ESRD and need renal replacement therapy [1]. Wyatt et al. [2] found that the five-year survival rate of children with IgAN was 94–98%, and the 20-year survival rate was 70–89%. Although IgAN in children has relatively lower incidence and less progress in childhood as compared to adults, there is a continuous hazard of progression as children grow older [3]. The prognosis of children with IgAN varies from remission to progression to ESRD. Early detection and effective intervention are important in improving the outcome of IgAN, and the renal biopsy is still the cornerstone of the correct diagnosis of IgAN. However, most patients have entered ESRD at the time of diagnosis, since it is difficult for patients to accept invasive renal biopsy. It is difficult to determine a fixed treatment plan for IgAN in children because it is more troublesome to gather up the course of IgAN in children than in adults. Therefore, it is necessary to develop new predictive models for the progression of IgAN in children in order to guide the selection of cases to be treated [4]. Machine learning has been successfully applied in many fields, providing a shield for diagnosis of IgAN for non-invasive diagnose in asymptomatic cases.
Traditional medical treatment methods rely entirely on doctors’ diagnosis and the treatment of patients. In this way, it is difficult to distinguish between diseases with similar symptoms and discover the hidden diseases, leading to misdiagnosis, which may delay the patient’s treatment or endanger the patient’s life. With the explosive growth of electronic medical records, machine learning in medicine has attracted the attention of many scholars [5,6,7]. Ali et al. proposed intelligent smart healthcare monitoring systems that are based on machine learning approaches to extract useful features from the collected healthcare data of patients, reduce the dimensionality of the data, and improve the classification precision [8,9]. ESRD has the characteristics of high disability, high mortality, and high medical expenses. A lot of researches were invested in the early detection and prediction of kidney disease by machine learning algorithms in order to avoid the occurrence of ESRD, such as the prediction of the kidney disease [10,11], the formulation of the best course of treatment [12], the evaluation of the risk stratification [13], and so on.
The goal of treating IgAN is still to delay the progression of IgAN, since the exact treatment plan for IgAN has not yet been established. Researchers mainly focus on the prognosis [14,15,16], risk stratification [17,18,19,20], and ESRD prediction of IgAN [21,22], whose patients are over 18 years old. The researches on IgAN in children focused on pathological characteristics and statistics [23,24,25], but there are few on the prediction of IgAN in children based on machine learning. Machine learning algorithms are an effective method for solving the high-dimensional problem of medical data, and it plays an important role in smart medical care.
In this paper, a novel method is proposed for predicting the probability of children patients with IgAN reaching ESRD in five years. The main contributions of this paper are:
  • A dataset about IgAN in children was created. And the chi-square test was used to extract the most useful features from the dataset.
  • EXtreme Gradient Boosting (XGBoost) was adopted in order to predict whether IgAN disease in children patients would reach ESRD or not within five years using a new dataset instead of the traditional clinical pathology. A decision-making system that was based on the XGBoost algorithm was designed with the Django framework.
  • Comparation of the performance of XGBoost with random forest (RF), CatBoost, support vector machines (SVM), k-nearest neighbor (KNN), and extreme learning machine (ELM) was conducted.

2. Materials and Methods

2.1. Dataset

The initial data include electronic medical records of 1167 patients aged 0–18 years old from 2003 to 2019, which came from the electronic medical records of the Nanjing Eastern War Zone Hospital. After the initial data cleaning and deletion of information with missing values, the dataset contained 1146 records. Each record contained 37 attributes, where the first 36 attributes are independent variables that correspond to the patient information, while the remaining attribute is the dependent variable of clinical interest. The 36 independent variables were collected according to five aspects, which are epidemiology, blood test, urine test, renal pathology, and treatment options. In detail, age, sex and hypertension are classified as epidemiology. Serum creatinine (Scr), cholesterol (CHOL), triglycerides (TG), albumin (ALB) complement C3, and glomerular filtration rate (eGFR) are blood test indicators. Urine tests include urine C3 (Ur_C3), α2-m, urine NAG enzyme (Ur_NAG), urine RBP (Ur_RBP), and uric acid (UA). ACEI_ARB, immunosuppression therapy, lipid lowering, and tonsillectomy are the treatment options. The remaining 19 attributes M, E, S, T, C, IgA, IgG, IgM, C3, C4, C1q, loop necrosis, focal segmental glomerular sclerosis (FSGS), glomerulosclerosis, arterial hyaline degeneration, crescent ratio, medullary interstitial fibrosis, thickening and stratification of elastic layer of interlobular artery, and vacuolar degeneration of arteriole smooth muscle cells are included in renal pathology. The dependent variable was whether the patient would reach the end-stage renal disease within five years, which was expressed by ESRD.
These independent variable attributes are divided into continuous independent variables and categorical independent variables according to whether they are continuous values or not, as shown in Table 1 and Table 2. Specifically, Table 1 shows the range, mean, and standard deviation for continuous independent variables and Table 2 displays the possible values, numeric value, as well as the number of records for categorical independent variables. Table 3 displays the possible values, numeric values, as well as the number of records for categorical dependent variables. The ESRD is distributed in 567 and 578 records, which correspond to the categories yes and no, respectively, which indicates that patients reach the final stage of IgAN or do not reach that within five years, respectively.

2.2. Feature Selection

We evaluated the importance and relevance of predictors with ESRD by the chi-square test for the purpose of identifying significant predictors of ESRD to be applied as inputs for the data mining methods. Feature analysis, as illustrated in Table 4, displays that the importance and relevance of all the predictors with ESRD. The p-value is called Pierce the correlation coefficient. A p-value of less than 0.05 means that there is significant difference. The score stands for the Chi-square statistics, which can be calculated according to Equation (1).
χ 2 = ( A T ) 2 / T
where A represents the actual value and T represents the theoretical value.
The higher the score, the more important the attribute. Sixteen features, which are shown in the first 16 rows of Table 4, were selected based on p-value that is equal to 0 and score greater than 10 for data dimension reduction in this paper.

2.3. Model

In this section, XGBoost has been introduced as the best performing algorithm. For processing high-dimensional data, dimensionality reduction, feature extraction, etc., it has a higher accuracy than traditional algorithms. XGBoost is an improved gradient boosting algorithm [26]. The innovation of XGBoost lies in the optimization of the objective function with the second-order Taylor expansion. It merges multiple weak classifiers in order to evolve into a strong classifier, and the base classifier is a classification and regression tree (CART).
The objective function of XGBoost consists of a loss function and a regularization term, which are defined, as follows:
O b j = i = 1 n l ( y i , y ^ i ) + k = 1 K Ω ( f k )
where f k is the function expression of the k-th tree model and y i and y ^ i are the true label and predicted value of the i-th sample x i , respectively. XGBoost is an additive model, so the predicted value is the sum of the predicted values of each tree, i.e., y ^ i = k = 1 K f k ( x i ) , f k F .
The sum of the complexity of K trees is used as a regularization term for preventing the model from over-fitting. Assuming that the tree model that is trained on the t-th iteration is f t , then:
y ^ i ( t ) = k = 1 t f k ( x i ) = y ^ i ( t 1 ) + f t ( x i )
Substitute Equation (3) into Equation (2) to obtain Equation (4).
O b j ( t ) = i = 1 t l ( y i , y ^ i ( t ) ) + i = 1 t Ω ( f i ) = i = 1 t l ( y i , y ^ i ( t 1 ) + f t ( x i ) ) + Ω ( f t )
Expand the loss function i = 1 t l ( y i , y ^ i ( t ) ) to second-order Taylor
O b j ( t ) i = 1 t [ l ( y i , y ^ i ( t 1 ) + f t ( x i ) ) ] = i = 1 t [ l ( y i , y ^ i ( t 1 ) + g i f t ( x i ) + 1 2 h i f t 2 ( x i ) ) ] + Ω ( f t )
where g i and h i are the first-order partial derivative and second-order partial derivative of the loss function l with regard to y ^ i ( t 1 ) , respectively, and Ω ( f t ) = γ T + 1 2 λ j = 1 T ω j 2 .
Define the leaf node I j = { i | q ( x i ) = j } , and the objective function is finally reduced to
O b j ( t ) = j T [ ( i I j g i ) ω j + 1 2 ( i I j h i ) ω j 2 ] + γ T
During the training process of the XGBoost model, when the t-th tree is established, the greedy strategy is adopted in order to split the tree nodes. Every time the tree node splits into two left and right leaf nodes, it will bring gain to the loss function, which is defined, as follows:
G a i n = O b j L + R ( O b j L + O b j R ) = 1 2 [ ( i I L g i ) 2 ( i I L h i ) + λ + ( i I R g i ) 2 ( i I R h i ) + λ ( i I g i ) 2 ( i I h i ) + λ ] λ
If G a i n > 0 , then the result of this split is added to the model construction.
XGBoost provides three calculation methods for feature importance. The first way is gain, which refers to the average gain of the feature when it is used in trees. The second way is weight, which is the number of times that a feature is used to split the data across all trees. The last way is cover, which relates to the average coverage of the feature when it is used in trees. In this study, the gain method was mainly used for calculating feature importance.

2.4. Performance Evaluation

ROC curve and area under curve (AUC) were used in order to evaluate the pros and cons of a binary classifier (binary classifier) in our paper. The abscissa of the curve is the false positive rate (FPR) and the ordinate is the true positive rate (TPR).
T P R = T P T P + F N
F P R = F P F P + T N
where TP stands for True Positive, TN represents True Negative, FP symbolizes False Positive, and FN means False Negative.
Different ( F P R , T P R ) points can be obtained by adjusting the threshold value that is predicted by the model, and these points can be connected into a curve, which is the ROC curve. After the curve is drawn, qualitatively analyze the model that if you want, you need to calculate the AUC area.
AUC refers to the area under the ROC curve [27]. Calculating the AUC value only needs to integrate F P R on the horizontal axis of ROC. In a real scene, the ROC curve is generally above the line y = x , so the value of AUC is generally between 0.5 and 1. The larger the value of AUC, the better the performance of the model.
In addition, compare the recall rate and f 1 score of all the classifiers.
r e c a l l = T P R = T P T P + F N
f 1 = 2 T P 2 T P + F P + F N
The training set and test set were randomly selected with a ratio of 3 to 1, during model training. Training and testing applied the 10-fold cross validation method to separate the dataset into several partitions or fold, calculated the average of accuracy from all folds. In addition, the above-mentioned performances, such as ROC, AUC, recall, and f1 scores were used to evaluate all techniques.

3. System Implementation

XGBoost, which showed the best classification performance, was used in order to implement an online decision-making system. The core framework of the system is the Django framework, which was used to apply the machine learning models and build the web tool, and it made use of two programming languages, including Python (version 3.7.0) and HTML (version HTML 5). Django is an open source python web framework. Programmers can easily and quickly create high-quality, easy-to-maintain, database-driven applications with the framework. In addition, in the Django framework, it also contains many powerful third-party plug-ins, which makes Django highly extensible.
In the current implementation of the system, an HTML communicates with the Python service and formats the information that is shown to the user. The training model of the system that is shown in Figure 1 can be used for a single prediction. Additionally, Figure 1 shows a screenshot of the initial web page. When users enter this page, fill in the data of the initial web page according to the feature description in Section 2.1. The system backend predicts whether the patient will reach ESRD or not in five years based on the data that were submitted by the initial web page. Moreover, the web-based decision-making assistance system will obtain the probability of a patient reaching and not reaching ESRD within five years, which may help doctors to alert to some borderline patients. Figure 2 shows the prediction outcome of the decision-making system (DMS) that is based on XGBoost.

4. Results

The best performing model was selected in order to predict whether children suffering from IgAN would reach the end-stage renal disease after five years among six kinds of machine learning algorithms of XGBoost, RF, CatBoost, KNN, SVM, and ELM.
Table 5 shows the accuracy, precision, recall, f1-score, and AUC values of XGBoost, RF, CatBoost, KNN, SVM, and ELM models. It can be concluded from the table that all of the performance indicators of XGBoost are the best, AUC, that accuracy, precision, recall, and f1-score are 85.11%, 78.60%, 75.96%, 76.70%, and 76.33%, respectively. Table 6 illustrates the importance scores of XGBoost on 16 variables that were selected from Table 4. The AUC indicators of RF and XGBoost are almost equal, but the accuracy, precision, recall, and f1-score of the RF model are all smaller than those of the XGBoost model.
Figure 3 depicts the ROC curve of six machine learning models. From the figure, it can be seen that the ROC curves of the XGBoost and RF models are close at the top; the ROC curve of the KNN model is at the bottom, which means that the XGBoost and RF models have the best performance and the KNN model has the worst performance according to the ROC curve. The ROC curves of the CatBoost, SVM, and ELM models are between that of the XGBoost and KNN models.

5. Discussion

In the paper, the use of machine learning algorithms for predicting IgAN in children was researched. At first, 16 features (serum creatinine, focal segmental glomerular sclerosis, crescent ratio, albumin, uric acid, glomerulosclerosis, cholesterol, urine NAG enzyme, eGFR, triglycerides, E, T, C, M, IgM, and C3) were chosen as the input of the classifiers by the chi-square test for dimension reduction. Subsequently, the XGBoost, RF, CatBoost, KNN, SVM, and ELM models were applied to predict IgAN in children. Finally, a decision-making system was build based on the best performing model and Django framework. The results that are shown in Table 5 and Figure 3 indicate that the XGBoost model can provide better performance when compared to other models for the medical application that was considered in this study. The AUC value, accuracy, precision, recall, and f1-score of XGBoost were 85.11%, 78.60%, 75.96%, 76.70%, and 76.33%, respectively. While the AUC value, accuracy, precision, recall and f1-score of RF (76.42%, 74.26%, 72.82%, 73.53%, 85.07%), CatBoost (76.42%, 73.79%, 73.79%, 73.79%, 84.54%), KNN (75.55%, 73.27%, 71.84%, 72.55%, 80.90%), SVM (76.42%, 73.33%, 74.76%, 74.04%, 82.72%), and ELM (75.98%, 72.64%, 74.76%, 73.68%, 81.74%) models are all lower than those of XGBoost. Here, we can highlight an advantage of XGBoost, because we not only need interpretable models to assist clinical decision-making, but also help clinicians to discover hidden factors that affect the disease. The XGBoost algorithm has a regularization term to prevent overfitting. Moreover, XGBoost can specify the default direction of the branch for missing values or specified values, which can greatly improve the efficiency of the algorithm. More importantly, the XGBoost model has high generalization performance and it can clearly output the important scores of each attribute, namely it is interpretable, which is required by clinical medicine.
Despite the potential of this research, there are several limitations. First, the collected dataset is relatively small, and it cannot fully cover kidney cases in children, which leads to inaccurate predictions of special kidney disease cases in children. Therefore, the dataset needs to be increased and feature processing needs to be more refined by applying data mining techniques for predicting IgAN in children. Second, the prediction system is suitable for children with IgAN in the age range of 0–18 years old, but not for adults with IgAN.
In the future, we will devote to improve the accuracy of the model, perfect the system, add database to the system, expand the training dataset, and complement the system with an error correction function.

Author Contributions

Conceptualization, P.Z. and N.S.; methodology, R.W. and N.S.; software, R.W. and N.S.; validation, R.W. and N.S.; formal analysis, P.Z. and R.W.; investigation, P.Z. and N.S.; resources, N.S.; data curation, R.W.; writing—original draft preparation, R.W.; writing—review and editing, P.Z. and R.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number 11401172, Key Scientific Research Project Plan of Colleges and Universities in Henan Province, grant number 20A520012, and the Pediatric Medical Innovation Team of Jiangsu Province, grant number CXTDA2017022.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Damico, G. The Commonest Glomerulonephritis in the World: IgA Nephropathy. QJM Int. J. Med. 1987, 64, 709–727. [Google Scholar]
  2. Wyatt, R.J.; Kritchevsky, S.B.; Woodford, S.Y.; Miller, P.M.; Roy, S., III; Holland, N.H.; Jackson, E.; Bishof, N.A. IgA nephropathy: Long-term prognosis for pediatric patients. J. Pediatr 1995, 127, 913–919. [Google Scholar] [CrossRef]
  3. Coppo, R.; Robert, T. IgA nephropathy in children and in adults: Two separate entities or the same disease? J. Nephrol. 2020, 33, 1219–1229. [Google Scholar] [CrossRef] [PubMed]
  4. Coppo, R. Treatment of IgA nephropathy in children: A land without KDIGO guidance. Pediatr. Nephrol. 2020, 2020, 1–6. [Google Scholar] [CrossRef]
  5. Jonsson, A. Deep Reinforcement Learning in Medicine. Kidney Dis. 2019, 5, 18–22. [Google Scholar] [CrossRef]
  6. Toback, F.G. Regeneration after acute tubular necrosis. Kidney Int. 1992, 41, 226–246. [Google Scholar] [CrossRef] [Green Version]
  7. Agar, J.W.; Webb, G.I. Application of machine learning to a renal biopsy database. Nephrol. Dialys. Transplant. 1992, 7, 472–478. [Google Scholar] [CrossRef] [Green Version]
  8. Ali, F.; El-Sappagh, S.; Islam, S.M.R.; Ali, A.; Attique, M.; Imran, M.; Kwak, K.-S. An intelligent healthcare monitoring framework using wearable sensors and social networking data. Futur. Gener. Comput. Syst. 2020, 114, 23–43. [Google Scholar] [CrossRef]
  9. Ali, F.; El-Sappagh, S.H.A.; Islam, S.R.; Kwak, D.; Ali, A.; Imran, M.; Kwak, K.-S. A smart healthcare monitoring system for heart disease prediction based on ensemble deep learning and feature fusion. Inf. Fusion 2020, 63, 208–222. [Google Scholar] [CrossRef]
  10. Leung, R.K.; Wang, Y.; Ma, R.C.; Luk, A.O.; Lam, V.; Ng, M.; So, W.Y.; Tsui, S.K.; Chan, J.C. Using a multi-staged strategy based on machine learning and mathematical modeling to predict genotype-phenotype risk patterns in diabetic kidney disease: A prospective case–control cohort analysis. BMC Nephrol. 2013, 14, 162. [Google Scholar] [CrossRef]
  11. Hamedan, F.; Orooji, A.; Sanadgol, H.; Sheikhtaheri, A. Clinical Decision Support System to Predict Chronic Kidney Disease: A Fuzzy Expert System Approach. Int. J. Med. Info. 2020, 138, 104134. [Google Scholar] [CrossRef] [PubMed]
  12. Gupta, R.; Sharma, A.; Singh, S.; Dinda, A.K. Rule-based decision support system in the biopsy diagnosis of glomerular diseases. J. Clin. Pathol. 2011, 64, 862. [Google Scholar] [CrossRef] [PubMed]
  13. Takayuki, K.; Masaki, O.; Akira, K.; Michiharu, K.; Kyoichi, H.; Jun, K.; Masaki, M.; Ryosuke, Y.; Atsushi, S. Risk Prediction of Diabetic Nephropathy via Interpretable Feature Extraction from EHR Using Convolutional Autoencoder. Stud. Health Technol. Info. 2018, 247, 106–110. [Google Scholar]
  14. Lemley, K.V.; Lafayette, R.A.; Derby, G.C.; Blouch, K.L.; Anderson, L.; Efron, B.; Myers, B.D. Prediction of early progression in recently diagnosed IgA nephropathy. Nephrol. Dialys. Transplant. 2007, 23, 213–222. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Noh, J.; Punithan, D.; Lee, H.; Lee, J.; Kim, Y.; Kim, D.; McKay, R.B. Machine Learning Models and Statistical Measures for Predicting the Progression of IgA Nephropathy. Int. J. Soft. Eng. Know. Eng. 2015, 25, 829–849. [Google Scholar] [CrossRef]
  16. Ducher, M.; Kalbacher, E.; François, C.; de Vilaine, J.F.; McGregor, B.; Fouque, D.; Fauvel, J.P. Comparison of a Bayesian Network with a Logistic Regression Model to Forecast IgA Nephropathy. BioMed Res. Int. 2013, 2013, 686150. [Google Scholar] [CrossRef] [Green Version]
  17. Barbour, S.J.; Reich, H.N. Risk Stratification of Patients with IgA Nephropathy. Am. J. Kidney Dis. 2012, 59, 865–873. [Google Scholar] [CrossRef] [Green Version]
  18. Barbour, S.J.; John, F. Predicting the future in immunoglobulin A nephropathy: A new international risk prediction tool. Nephrol. Dialys. Transplant. 2020, 35, 379–382. [Google Scholar] [CrossRef]
  19. Barbour, S.J.; Coppo, R.; Zhang, H.; Liu, Z.-H.; Suzuki, Y.; Matsuzaki, K.; Katafuchi, R.; Er, L.; Espino-Hernandez, G.; Kim, J.; et al. Evaluating a New International Risk-Prediction Tool in IgA Nephropathy. JAMA Intern. Med. 2019, 179, 942–952. [Google Scholar] [CrossRef]
  20. Chen, T.; Li, X.; Li, Y.; Xia, E.; Qin, Y.; Liang, S.; Xu, F.; Liang, D.; Zeng, C.; Liu, Z. Prediction and Risk Stratification of Kidney Outcomes in IgA Nephropathy. Am. J. Kidney Dis. 2019, 74, 300–309. [Google Scholar] [CrossRef]
  21. Liu, Y.; Zhang, Y.; Liu, D.; Tan, X.; Tang, X.; Zhang, F.; Xia, M.; Chen, G.; He, L.; Zhou, L.; et al. Prediction of ESRD in IgA Nephropathy Patients from an Asian Cohort: A Random Forest Model. Kidney Blood Press. Res. 2018, 43, 1852–1864. [Google Scholar] [CrossRef] [PubMed]
  22. Han, X.; Zheng, X.; Wang, Y.; Sun, X.; Xiao, Y.; Tang, Y.; Qin, W. Random forest can accurately predict the development of end-stage renal disease in immunoglobulin a nephropathy patients. Annal. Translat. Med. 2019, 7, 234. [Google Scholar] [CrossRef] [PubMed]
  23. Bertram, J.F.; Goldstein, S.L.; Pape, L.; Schaefer, F.; Shroff, R.C.; Warady, B.A. Kidney disease in children: Latest advances and remaining challenges. Nat. Rev. Nephrol. 2016, 12, 182–191. [Google Scholar] [CrossRef] [PubMed]
  24. Coppo, R.; Troyanov, S.; Camilla, R.; Hogg, R.J.; Cattran, D.C.; Terence Cook, H.; Feehally, J.; Roberts, I.S.D.; Amore, A.; Alpers, C.E.; et al. The Oxford IgA nephropathy clinicopathological classification is valid for children as well as adults. Kidney Int. 2010, 77, 921–927. [Google Scholar] [CrossRef] [PubMed]
  25. Shima, Y.; Nakanishi, K.; Sato, M.; Hama, T.; Mukaiyama, H.; Togawa, H.; Tanaka, R.; Nozu, K.; Sako, M.; Iijima, K.; et al. IgA nephropathy with presentation of nephrotic syndrome at onset in children. Pediatr. Nephrol. 2017, 32, 457–465. [Google Scholar] [CrossRef] [PubMed]
  26. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. arXiv 2016, arXiv:1603.02754. [Google Scholar] [CrossRef] [Green Version]
  27. Hanley, J.A.; Mcneil, B.J. The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology 1982, 143, 29–36. [Google Scholar] [CrossRef] [Green Version]
Figure 1. A screenshot of the initial web page of the implemented decision-making system (DMS).
Figure 1. A screenshot of the initial web page of the implemented decision-making system (DMS).
Futureinternet 12 00230 g001
Figure 2. Outcome of the eXtreme Gradient Boosting (XGBoost)-based DMS.
Figure 2. Outcome of the eXtreme Gradient Boosting (XGBoost)-based DMS.
Futureinternet 12 00230 g002
Figure 3. Receiver operating characteristic (ROC) curve of models.
Figure 3. Receiver operating characteristic (ROC) curve of models.
Futureinternet 12 00230 g003
Table 1. Continuous independent variables.
Table 1. Continuous independent variables.
VariableRangeMeanStandard Deviation
Age (years)0–18143.59
Scr (μmol/L)16.00–1154.5071.1267.86
CHOL (mmol/L)0.20–19.005.492.70
TG (mmol/L)0.17–22.001.621.25
UA (μmol/L)0.53–777.00347.8797.63
ALB (g/mL)7.60–65.2041.88120.48
Complement C3 (g/L)0.24–25.001.060.74
FSGS0.00–61.504.208.75
Spherically sclerotic0.00–84.203.759.21
Crescent ratio0.00–77.805.259.50
eGFR (mL·min−1·(1.72 m2)−1)4.90–141.75111.7721.11
Scr-Serum creatinine; CHOL-cholesterol; TG-triglycerides; UA-uric acid; ALB-albumin; FSGS-focal segmental glomerular sclerosis; eGFR-glomerular filtration rate.
Table 2. Categorical independent variables.
Table 2. Categorical independent variables.
VariablePossible ValuesNumeric ValueNumber of RecordsPercentage (%)
GenderM/F1/0776/36967.78/32.22
Ur_C3normal/abnormal0/1347/79830.31/69.69
α 2-mnormal/abnormal0/1214/93118.69/81.31
Ur_NAGnormal/abnormal0/1909/23679.39/20.61
Ur_RBPnormal/abnormal0/1164/98114.32/85.68
M 0/1/2/3283/562/267/3324.72/49.08/23.32/2.88
E 0/1/2756/386/366.03/33.71/0.26
S 0/1728/41763.58/36.42
T 0/1/2/3845/251/42/773.80/21.92/3.67/0.61
C 0/1/2648/444/5356.59/38.78/4.63
IgA 0/1/2/316/31/1082/161.40/2.70/94.50/1.40
IgG 0/0.5/1/2882/1/177/8577.03/0.09/15.46/7.42
IgM 0/0.5/1/2669/13/418/4558.43/1.14/36.50/3.93
C3 0/0.5/1/2207/9/173/75618.08/0.79/15.11/66.02
C4 0/11133/1298.95/1.05
C1q 0/1/21094/49/295.55/4.28/0.17
Hypertensionyes/no1/0322/82328.12/71.88
ACEI_ARB 0/1862/28375.28/24.72
Immunosuppressive therapy 0/1726/41963.41/36.59
Lipid lowering 0/1738/40764.45/35.55
Tonsillectomy 0/11124/2198.17/1.83
Loop necrosis 0/1/2/31039/91/7/890.74/7.95/0.61/0.70
Arterial hyaline degeneration 0/1/51015/129/188.65/11.26/0.09
Medullary interstitial fibrosis 0/1/2/31090/39/11/595.20/3.41/0.96/0.43
Thickening and stratification of elastic layer of interlobular artery 0/11108/3796.77/3.23
Vacuolar degeneration of arteriole smooth muscle cells 0/11085/6094.76/5.24
Ur_C3-urine C3; Ur_NAG-urine NAG enzyme; Ur_RBP-urine RBP.
Table 3. Categorical dependent variable.
Table 3. Categorical dependent variable.
VariablePossible ValuesNumeric ValueNumber of RecordsPercentage (%)
ESRDyes156749.52
no057850.48
Table 4. Chi-square analysis results of predictors.
Table 4. Chi-square analysis results of predictors.
VariableScorep-Value
Scr827.8850
FSGS716.5520
Crescent ratio658.5340
ALB621.3520
UA461.2230
Spherically sclerotic321.2430
CHOL279.5440
Ur_NAG119.2240
eGFR163.1700
TG56.1080
E27.4700
T23.4790
C18.6530
M18.4510
IgM16.9730
C313.1890
Gender0.2030.652
Ur_C30.0240.878
α 2-m7.2330.007
Ur_RBP5.8230.016
Complement C30.0030.953
S5.7600.016
IgA0.0250.875
IgG0.8160.366
C40.0010.973
C1q3.4440.063
Hypertension11.5920.001
ACEI_ARB0.2110.646
Immunosuppressive therapy0.2320.630
Lipid lowering9.7250.002
Tonsillectomy0.4880.485
Loop necrosis0.0240.877
Arterial hyaline degeneration0.0120.911
Medullary interstitial fibrosis5.6550.017
Thickening and stratification of elastic layer of interlobular artery6.3730.012
Vacuolar degeneration of arteriole smooth muscle cells0.4900.484
Table 5. Results for Immunoglobulin A nephropathy (IgAN) prediction.
Table 5. Results for Immunoglobulin A nephropathy (IgAN) prediction.
AlgorithmAccuracyPrecisionRecallF1_ScoreAUC
XGBoost0.78600.75960.76700.76330.8511
RF0.76420.74260.72820.73530.8507
CatBoost0.76420.73790.73790.73790.8454
KNN0.75550.73270.71840.72550.8090
SVM0.76420.73330.74760.74040.8272
ELM0.75980.72640.74760.73680.8174
Table 6. The corresponding variable importance score using XGBoost.
Table 6. The corresponding variable importance score using XGBoost.
VariablesImportance Score
Ur_NAG0.191591
ALB0.175141
CHOL0.104368
Crescent ratio0.085010
FSGS0.067488
Scr0.062304
TG0.050906
Spherically sclerotic0.034839
UA0.034522
M0.033158
IgM0.032431
T0.031299
C30.025893
C0.025279
eGFR0.024145
E0.021627
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, P.; Wang, R.; Shi, N. IgA Nephropathy Prediction in Children with Machine Learning Algorithms. Future Internet 2020, 12, 230. https://doi.org/10.3390/fi12120230

AMA Style

Zhang P, Wang R, Shi N. IgA Nephropathy Prediction in Children with Machine Learning Algorithms. Future Internet. 2020; 12(12):230. https://doi.org/10.3390/fi12120230

Chicago/Turabian Style

Zhang, Ping, Rongqin Wang, and Nianfeng Shi. 2020. "IgA Nephropathy Prediction in Children with Machine Learning Algorithms" Future Internet 12, no. 12: 230. https://doi.org/10.3390/fi12120230

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop