Previous Article in Journal
Notch Signalling Plays a Role in Patterning the Ventral Mesoderm During Early Embryogenesis in Drosophila melanogaster
Previous Article in Special Issue
Artificial Intelligence for Multiscale Spatial Analysis in Oncology: Current Applications and Future Implications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Artificial Intelligence in Nephrology—State of the Art on Theoretical Background, Molecular Applications, and Clinical Interpretation

by
Jakub Stojanowski
1,
Tomasz Gołębiowski
1 and
Kinga Musiał
2,*
1
Department of Nephrology and Transplantation Medicine, Wroclaw Medical University, Borowska 213, 50-556 Wroclaw, Poland
2
Department of Pediatric Nephrology, Wroclaw Medical University, Borowska 213, 50-556 Wroclaw, Poland
*
Author to whom correspondence should be addressed.
Int. J. Mol. Sci. 2026, 27(3), 1285; https://doi.org/10.3390/ijms27031285 (registering DOI)
Submission received: 9 December 2025 / Revised: 15 January 2026 / Accepted: 24 January 2026 / Published: 28 January 2026
(This article belongs to the Special Issue Machine Learning in Disease Diagnosis and Treatment: 2nd Edition)

Abstract

Artificial intelligence (AI) has transformed the clinical approach to analysis of large datasets, introducing the possibility of verifying long-term observations. AI tools ease the analysis of connections between multiple variable parameters and are particularly useful in the field of nephrology. These solutions enable the search for early diagnostic markers and predictors of renal function deterioration, both in acute and chronic conditions. Furthermore, AI techniques can be used as data mining tools, paving the way for future theories regarding the pathomechanisms of disease. Moreover, recently published papers focus on building models that facilitate decision-making, thus predicting renal involvement, its progression, and systemic complications. This review aims to demonstrate the multifunctionality of various AI methods from an omics perspective. To increase the power of argumentation, a mathematical background of each method is presented, followed by examples of molecular applications and anchorage in the nephrological clinical context. Our aim was to demonstrate the potential of AI tools in addressing diagnostic, prognostic, and therapeutic challenges, as well as to initiate the discussion on the pros and cons of future AI applications in nephrology.

1. Introduction

According to the founder of artificial intelligence, John McCarthy, AI relies on mimicking human decision-making [1]. From a clinical perspective, intelligence can mean predicting adverse outcomes in order to prevent them. In the era of big data, such results can only be achieved with advanced tools that enable in-depth analysis of a large number of variables, preferably assessed periodically during a long-term follow-up. International registries and multicenter studies, where the primary focus is on assessing risk factors associated with patient morbidity and mortality, serve as excellent examples. Clinical reasoning based on large-scale retrospective analysis approximates the most likely scenarios, identifies the most common endpoints, and provides solutions to prevent the disease, treat it effectively, or at least maintain in remission. The need for effective and early prediction of adverse events has dramatically increased during the COVID-19 pandemic.
There is a noticeable upward trend in interest in AI in medicine, as evidenced by bibliographic statistics. In the early 1990s, a single article per year was published on AI in medicine. Over the following years, the number of papers increased year after year, accelerating rapidly between 2017 and 2019, reaching several thousand papers annually [2]. This may be due to two leading factors: (1) advances in AI technology, and (2) the isolation caused by the COVID-19 pandemic and the pressure to conduct research under limited conditions and resources [3,4].
Global mortality risk management has fueled intensive research into the ability to effectively predict complications and improve decision-making. Nephrological aspects have also been of interest, including the prediction of acute kidney injury, morbidity and mortality in intensive care units, and thrombotic microangiopathies as the main nephrological complications of COVID-19 infection. Certain significant relationships also elude classical statistical analysis, and given this need, AI is prioritized for implementation.
Most diagnostic and therapeutic processes in nephrology utilize biochemical assays and are based on molecular aspects of the disease processes. Hence, there is significant interest in analyzing omics data, including proteomics and genomics.
Given the overwhelming abundance of data, rapid risk assessment may not directly translate to the needs of a specific patient. Experience and knowledge may not be sufficient in such cases, but with the help of AI, new diagnostic and therapeutic guidelines can be developed, based on a shift from macroscopic analysis to an individualized approach.
The potential for AI applications in nephrology is enormous, provided the method is tailored to specific requirements and endpoints. This review will discuss a wide range of AI tools, highlighting their importance in various clinical conditions and tailoring results to individual patient needs. It will also demonstrate, using examples from the literature, how to choose among many AI tools and which method suits best which disease.

2. General Classification

The most general classification of AI methods divides them into machine learning and deep learning. The first of these fields concerns rather simple methods, relatively less advanced than deep learning. The latter uses more powerful technical and algorithmic resources and provides a wider range of possible applications, although it is weak in certain areas which are reserved for machine learning.

2.1. Machine Learning

There are various machine learning techniques that are applicable to relevant topics. Some techniques require the supervision of a researcher, whose role is to determine the context of the data, i.e., assign them appropriate classes or labels. This type of AI that requires human intervention in the learning process is called supervised learning. In clinical practice, it may be useful to detect outlying structures and elements; then, the context is given by an algorithm. The mentioned technique is called unsupervised learning. However, when dynamic interaction of the algorithm with a changing environment is required, e.g., a changing patient’s condition, then reinforcement learning may be useful.

2.2. Unsupervised Learning

Unsupervised learning is based on finding patterns in data without given labels or target points. An example application is clustering and detecting deviations in a group of records or images, e.g., biopsy. Unsupervised learning methods allow for redefining existing diagnostic standards and reclassifying existing divisions by comparing them with existing classifications. Clustering, or cluster analysis, plays a special role in unsupervised machine learning. These techniques involve grouping elements based on similarity.
Consequently, they empower identification of parameters or constellations of parameters responsible for belonging to a specific cluster, also known as a category. This approach allows for reclassification of chronic kidney disease stages based on the results of analysis using the Self-Organizing Maps—an unsupervised ANN machine learning algorithm [5]. This approach has allowed the discovery of protein patterns, as well as sets of patterns characterizing CKD with an etiology confirmed by nephropathological examination.
A clustered dataset of prediabetic patients was analyzed using lasso regression. The stepwise application of several analysis techniques allowed for the stratification of patient groups based on selected peptidomes and the determination of the risk of developing diabetes and diabetic complications [6].
In other applications, unsupervised learning allows for the detection of important elements in images, e.g., examination of histopathological kidney preparations, and linking them with the patient’s laboratory parameters [7]. Modeling using unsupervised machine learning allows for stratification of the risk of transplant rejection and the risk of death in the observed period [8]. Patient clustering is possible with a clear distinction between patient group characteristics and graft survival in transplant patients [9].

2.3. Supervised Learning

Supervised learning requires input data that has been grouped into classes. This method typically involves labeling and categorizing data records. For example, a set of genes can be assigned to the degree of renal fibrosis, thus assessing the risk of kidney disease progression in a carrier of these genes. The use of several models in a parallel decision-making process allows to achieve the target point with an AUC of 0.923 based on the selected five genes: ARID4B, EOMES, KCNJ3 (alias GIRK-1), LIF, and STAT1 [10]. Looking back, hypotheses can be constructed based on the association of the endpoint with the designated genes, laying the foundation for further research to uncover the association of fibrosis with the expression of these genes.
Similarly, renal failure can be assessed based on the proteomic set detected in the patient’s urine using clustering [11]. Patients with CKD constitute a group with an interdisciplinary profile and complex outcomes. Medical history of CKD demonstrates a higher risk of cardiovascular complications, which can be predicted using a supervised machine learning model [12]. An example of molecular application is a database of patient records with parameters and the endpoint of achieving remission. In detail, the use of machine learning allows for the assessment of the risk of AKI in pediatric patients and adults staying in the ICU more effectively than traditional methods [13,14,15].

2.4. Reinforcement Learning

Reinforcement learning is like playing a game in which the player performs certain tasks in the game environment and receives feedback in the form of a reward and information about the game state. In a hospital setting, this mimics personalized medicine. While therapy is dynamically tailored to the patient’s condition, it is subject to subsequent adaptive changes. These changes involve not only drug dosages but also the introduction of new medications, the discontinuation of old ones, and the expansion of diagnostics. An example application is the personalized dosing of the sedatives (propofol and fentanyl) in response to variable patient characteristics [16]. Therapy modification, based on reinforcement learning, has been used to achieve optimal dry weight in patients undergoing hemodialysis [17]. Likewise, the above-mentioned AI method allowed the modulation of erythropoiesis-stimulating agent (ESA) dosing in relation to variable patient characteristics and outcomes [18,19].

3. Machine Learning Models

3.1. Random Trees

Random forest classifiers (RFCs) are derived from classification and regression trees (CARTs), which are models based on decision trees built using a splitting procedure. The result of such procedures is a decision tree, which, due to the algorithm and method of generation, is highly sensitive to variability in input parameters. Random forest classifiers allow for improved flexibility of the decision tree by utilizing multiple decision trees built on subsamples of the input dataset.
A specific example of decision trees are guideline-based algorithms. The essence of a decision tree is to apportion an input set using successive conditions, based on a minority relation, in the simplest possible way. The goal of a decision tree is to divide the set in such a way that one class is separated from another in the optimal number of steps. RFCs are insensitive to data that may be irrelevant. In practice, this means that the random forest’s construction method forces the ignoring of variables that do not provide crucial information for partitioning the dataset. An RFC classifier consists of independently generated decision trees. Therefore, RFCs are characterized by a certain degree of robustness to outliers.
Compared to other methods selected by Yuan et al., an RFC based on selected genes allows for the extraction of a dataset most helpful in analyzing the risk of renal fibrosis, which other machine learning techniques assess [10]. The role of the classifier can be played by data mining to find the most valuable input datasets. To this end, the small or large deterioration in prediction occurs after eliminating a variable. If such a variable has little impact on the classifier’s result, it is likely not significant in the prediction and can be eliminated.
The RFC has been used to assess potential drug interactions, achieving an accuracy of 96.51%. Competing models by Kha et al. achieved slightly better results with similar input data, but RFC offers several advantages, including lower resource requirements and faster evaluation [20].
RFCs, due to their simplicity, transparency, and short computational time, are used to discover features associated with mortality and risk factors for the development of diabetic kidney disease. An RFC-based model based on cystatin C, serum albumin, hemoglobin, 24-h total urinary protein, and eGFR predicted ESRD in patients with diabetic kidney disease with an accuracy of 82.65% and an AUC of 0.90 [21]. An RFC-based model based on diabetic retinopathy status, diabetes course, pulse pressure, hemoglobin level, serum creatinine level, albumin level, serum total cholesterol, and acute onset of severe proteinuria was able to distinguish the etiology of kidney disease between diabetic and nondiabetic kidney disease. In this case, adding variables slightly improved the predictive value [22].

3.2. Variants of Regression

Logistic regression is one of the basic techniques for analyzing dichotomous variables using odds analysis. The formal approach to logistic regression is a general linear model, where the link function is the logarithm of the odds of selecting a given category. Thus, logistic regression represents a very simple yet effective machine learning model. Machine learning analysis allows for the verification of known metabolic pathways, but it also generates a tool particularly useful in practical clinical applications. Knowing the developmental scenario of diabetic kidney disease and progression to subsequent stages based on logistic regression allows for the identification of a specific proteomic trace and the development of a tool for predicting patients at particularly high risk of disease progression [23].
Machine learning is widely used in data mining to identify key data. For this purpose, multiple models can be built, their performance assessed, and the best model identified. Decision-making parameters can be reviewed to draw clinical conclusions or suggestions for further research, including differentiation of nephropathies leading to chronic kidney disease [24].
LASSO regression (least absolute shrinkage and selection operator) is an extension of logistic regression in which regularization clarifies the leading problem and improves overall performance while reducing the risk of overfitting. Using RFC, Yuan et al. identified basement membrane gene markers associated with a leading mechanism of chronic kidney disease progression, namely renal parenchymal fibrosis. Based on parameters selected using three machine learning methods, a nomogram based on LASSO regression was constructed with satisfactory discriminative power, with an AUC of 0.923, enabling estimation of the risk of kidney disease progression.
A pre-clustered dataset of prediabetic patients, then analyzed with LASSO regression, allows for the stratification of patient groups based on selected peptidomes and the determination of the risk of developing diabetes and diabetic complications. Positive results for the CKD273 (CKD proteinogram classifier), HF2 (urine peptidogram to determine the risk of heart failure), and CAD238 (coronary artery disease proteomic classifier) tests were associated with increased insulin resistance. These parameters differentiated groups with a high risk of progression from groups with a very low and low risk of progression to diabetes and the development of diabetic complications [6].
The next evolution of logistic regression, and especially LASSO regression, is Elastic Net Regression with additional regularization; in short, a tool that improves predictions and reduces overfitting. Proteomic modeling and machine learning identify new risk management models and analyze the biological pathways of various proteins involved in the development of cardiovascular complications in patients with comorbidities and chronic kidney disease [12].
Regularized Cox Regression is a compilation of Elastic Net with the Cox hazard model. Combining selected techniques allows for the utilization of their advantages and mixing the partial results into a final solution to the overall problem. Massy et al. used clustering to determine CKD severity categories [11]. The number of peptides constituting the protein profile was narrowed using Regularized Cox Regression. Clustering allows for the identification of closely related data subpopulations and then the selection of data from which these populations can be successively differentiated. This technique is an extension of the Cox risk model, with a greater emphasis on avoiding overfitting [25].

3.3. eXtreme Gradient Boosting (XGBoost)

eXtreme Gradient Boosting (XGBoost) is a machine learning model containing decision trees based on the principle of boosting, i.e., enriching the existing model with a new additional model, provided it improves performance. The added models are decision trees that cover the errors of the existing model. XGBoost model is promising due to its good convergence to optimal performance and relatively low complexity. It has been used, among others, in diagnosing CKD, hyperkalemia in patients with CKD, or mortality in patients with acute kidney injury, with accuracy of 1.00, 0.933, 0.86, respectively [26,27,28]. XGBoost achieved the prediction of 50% reduction of estimated glomerular filtration rate (eGFR) and progression to end-stage kidney disease (ESKD) in patients with IgA nephropathy with C-statistic of 0.84 [29].

3.4. Support Vector Machine

Support vector machine (SVM) is a method that involves determining a hyperplane that optimally separates the elements of the set according to the assigned classes. This is a relatively simple method, but when used on a small amount of data in the form of a table, it may prove to be superior to RFC or XGBoost [30]. Proteomic analysis of potential factors with SVM associated with cellular aging may provide insight into the protein-mediated pathways responsible for the development of chronic kidney disease in patients without underlying disease compared to those with underlying disease but without established kidney disease. This sheds light on the independence of the renal disease process and provides a foundation for targeted medicine in nephrology [31].

3.5. K-Nearest Neighbors Classifier (kNN)

This technique relies on the similarity of an element to its nearest N neighbors, e.g., three. Based on the labels of the closest elements, the most frequently occurring one is selected and the similarity of a given element of the set to these labels is estimated.
The kNN classifier allows for group classification and clustering, in particular, effective differentiation of a group of patients with diabetic nephropathy and glomerulonephritis from a healthy control based on serum proteomics. Results had an accuracy of 96.3% for glomerulonephritris and 96.4% diabetic nephropathy [32]. Among other machine learning techniques, kNN also performed very well in kidney cell typing based on scRNA-seq and snRNA-seq data [33]. Models based on XGB, RFC, KNN, and SVM effectively identified blood cell types with a sufficiently high F1-score, varying depending on the model configurations, reaching values of 0.97–1.0. Identifying kidney cells proved more challenging. Here, F1-scores for nephron cell typing were significantly poorer, bordering on 0.6, and for proximal tubule cell identification, even values of 0.32–0.35. All the described models were based on input data consisting of scRNA and snRNA sequences obtained from kidney biopsies.
Compared to other machine learning techniques, kNN is suitable for data with a low number of dimensions, i.e., the number of features. Depending on the input data and selected variables, its performance is comparable to logistic regression, RFCs, or support vector machines [24,34]. It is clearly dominant in clustering due to the nature of the algorithm. However, an unusual application is data imputing, i.e., filling in missing data based on neighborhood similarity [35]. Data completeness is crucial for machine learning. Empty cells are interpreted as zero values or must be removed from columns and rows containing entire records, which reduces the size of the input data. Inputting the most likely values in empty cells improves model flexibility and performance [36].

4. Deep Learning and Multilayer Perceptron

Deep learning refers to AI techniques that are more advanced than machine learning and are based on convolutional neural networks. A convolutional neural network (CNN), like a multilayer perceptron, is a feedforward network, i.e., a one-way network in which information flows from input to output without returning to visited neurons or creating a reinforcement loop. Convolution means the optimization of neural networks with the use of mathematical operations based on the application of filters. Applying a filter ensures that subsequent layers gain some context about the input data. This application allows, for example, image recognition, finding details in the image, e.g., recognizing glomeruli in biopsies [37]. The anatomical analogy of a convolutional network is the visual cortex of the brain, where different neurons correspond to a specific visual field. Comprehensive CNN-based systems allow for the classification of biopsy specimens from rejected renal allografts in terms of the presence of pathology, and they also visually highlight pathological areas [38]. Deep learning allows for the differentiation of individual structures in histopathological images. Segmenting the histopathological image allows for automatic qualitative classification of the collected specimen into many groups, such as unilateral ureteral obstruction, ischemia-reperfusion injury, nephrotoxic serum nephritis, adenine-induced nephropathy, or Alport syndrome [39].
A niche, although essential, application of deep learning is in omics analysis. Its significance results from the ability to identify, e.g., the molecular profile of patients with advanced diabetic kidney disease, based on two lipid metabolites that have significant relationships with the level of glycated hemoglobin and glycemic profiles in patients with type 2 diabetes [40].

4.1. From Linear Regression to Artificial Neural Network

The main goal of statistical modeling is to describe a phenomenon using mathematical operations, with the smallest possible error between observed and calculated values. One of the simplest examples is linear regression, which assumes a linear relationship between the explained and explanatory variables. Assuming that the regression coefficients are betas (β) and the random error is epsilon (ε), then linear regression is a linear function that can be written as
y i = β 0 + β 1 x 1 + β 2 x 2 + + β k x k + ε i
To determine the above coefficients, the least squares method and its derivatives are used. The purpose of such calculations is to minimize differences in estimates from observed values. The main issue then becomes how to calculate the coefficients to minimize the loss function describing the discrepancy between observations and approximations.
A single artificial neuron is called a perceptron and can be described as a composite of two functions: the activation function and the sum of the products of inputs x1, x2, x3, …, xi and their corresponding weights marked by w1, w2, w3, …, wi plus an intercept called bias vector w0, which is the argument of the activation function.
o u t p u t = f w 0 + i = 1 k w i x i  
Sum of products of inputs and weights assigned to inputs:
i = 1 k w i x i = w 1 x 1 + w 2 x 2 + + w k x k
Activation functions can be various and may influence the behavior of the network differently, which is an empirical domain.
The simplest is rectifier or rectified linear unit (ReLU), which returns the positive part of the argument:
f x = m a x ( 0 , x )
The use of the sigmoid function reduces a single neuron to the level of logistic regression:
f x = 1 1 + e x
Yet another activation function is the hyperbolic tangent function:
f x = tanh x = e x e x e x + e x
By arranging many neurons into layers and connecting the outputs of others with the inputs of others in an organized and orderly way, the resulting neural network is obtained (Figure 1).
If such a network has only perceptrons, i.e., neurons described above, which are organized into layers in which the neurons are connected each with each other between the layers but not within them, we obtain a model of an ANN called a multilayer perceptron (MLP). A multilayer perceptron has an input layer, optional hidden layers, and an output layer. Each neuron in one layer is connected to each neuron in the other layer, creating a complete neural network.
The multilayer perceptron-based prediction model was able to effectively predict fibrosis in patients with chronic kidney disease (CKD) based on leading predictors SWE value, eGFR, age, UACR, RI, renal parenchyma thickness and hypertension, and several with less impact on prediction [41]. On the other hand, a model based on static and dynamic parameters allowed for ongoing monitoring of the risk of intradialytic hypotension in hemodialysis patients, which is, in a sense, an example of reinforcement learning [42]. The mentioned model based on static features such as age, diabetes status, hypertension and ultrafiltration, and dynamic parameters combined with the mentioned static ones, heart rate (HR) slope at 0, 30, and 60 min for each patient and depending on nervous tension. At the same time, the authors used extended dynamic variables derived from HR slope at appropriate moments of measurement, difference of slope (DoS) at 30–15 min and 60–30 min. The model designed in this way achieved an accuracy of 81.5%. In the model, diabetes, age more than 65 years old, and a negative history of hypertension created a cluster that brought patients closer to the endpoint, i.e., hypotension.
Attempts are being made to reclassify chronic kidney disease, based on the results of analysis using the Self-Organizing Maps unsupervised machine learning algorithm [2]. This approach has allowed the discovery of protein patterns, as well as sets of patterns, characterizing CKD with an etiology confirmed by nephropathological examination. A Self-Organizing Map is a type of neural network that contains a two-dimensional array of data in the input layer, which is connected to several neurons in the competitive layer. Case analysis allows to formulate hypotheses based on the aforementioned modeling approach.

4.2. Deep Learning and Convolutional Neural Networks

Deep learning is based on the use of many layers of a neural network. A convolutional neural network uses convolution operations to determine the context of the analyzed data. Deep learning is used in analyzing radiological and histopathological images, including detecting differences in unsupervised learning or detecting specific structures or pathologies in supervised learning [37,38,39,43]. For example, the AlexNet architecture allowed segmenting the glomeruli in the histopathological image [43,44]. In detail, convolutional networks consist of three types of layers arranged appropriately: (1) convolutional layers that enable the creation of features from single data, e.g., pixels in an image, (2) pooling layers that are responsible for combining features and deepen the neural network, and (3) dense layers, which are responsible for classification (Figure 2).
Classification of histopathological images of kidney specimens using a network with U-net architecture allows for the identification of structures visible in the image in an automatic and precise way [37,45,46,47]. Networks from the ResNet family (ResNet18, 50 and 101), where the numbers indicate the depth of the network, or the number of layers, allow the classification of kidney transplant pathology [38,48].
The summary of multiple machine learning applications in translation of proteomics to clinical practice is given below (Table 1).

5. Comparative Characteristics of Various Methods

All classification models gain power as the volume of analyzed data increases and the structure of the classifier is modified. The role of the model is to match the provided data, but redundant data may obscure the classification process. Each model tries to adapt as much as possible to the provided data to ensure optimal performance. The role of a physician is to avoid unnecessary data by deciding which variables are of clinical significance and should be included within the dataset. Consequently, balancing the number of variables should have a positive impact on the classifier’s performance.
RFC has some ability to reject irrelevant variables, although this feature has a natural limitation, because the model tries to fit the training data anyway.
There is no scaling or normalization required to prepare the data before use in the RFC. The smallest cell in the decision tree in the RFC is the node with the minority relationship stored. In the case of a neural network, data scaling is necessary for the network to function properly. On the other hand, artificial neural networks are more amenable to incremental learning, which involves introducing new data into the model. RFCs require virtually complete redesign to expand the training input data.
Both mentioned-above models allow the detection of linear and non-linear dependencies, with non-linear performance being a definite advantage.
The complexity of the neural network may both bring a profit and expose to harm. Better fitting to non-linear data is advantageous, whereas complexity and difficulty in interpretation are inconvenient. Unlike decision trees in an RFC, where the stages of decision-making can be traced, a neural network involves many calculations, the meaning of which is not clear without careful analysis. To maintain transparency of the processes taking place inside the model, it is worth focusing on models that are simpler than neural networks, e.g., RFCs.
Fernández-Delgado et al. compared various machine learning models on the publicly available UCI Repository training dataset [49]. The conclusion was reached that as the data complexity increases, the accuracy of models based on neural networks increases, which confirms the validity of using neural networks in large and complex datasets. RFCs on the UCI repository datasets showed the highest accuracy, although neural networks were not significantly worse. These findings further reinforce the need to adapt AI methods to the quality of available data.
Classifiers based on decision trees, including RFCs or XGBoost, are less time-consuming and resource-intensive. RFC is partially insensitive to irrelevant variables and requires neither scaling nor standardization. Decision trees perform better on small amounts of data compared to neural networks. A big advantage is the transparency of the model, in which the decision-making process in individual trees can be traced.
The RFC algorithm allows for the extraction of data relevant to analysis. Input data selection is performed during model development. The extracted data provides the basis for further research. Similarly, significant factors influencing acute rejection of transplanted organs can be assessed, identifying individuals at particular risk before the procedure. This approach also allows for the estimation of the possibility of delayed function. In summary, transplantology is not only about modeling but also about data selection. The random forest classifier consisting of donor BMI, recipient BMI, donor-recipient weight difference, and donor eGFR before collection, EPTS, KDRI, KDPI, recipient gender, and age was very effective in identifying the risk of delayed graft function with an AUC of 0.91 and an accuracy of 93.75% [50].
Similarly, Liu et al. successfully identified the most important predictors of progression from IgA nephropathy to end-stage renal disease. Using random forests, sensitive data was extracted and a predictive model was built based on it. The model based on clinical data and kidney biopsy results allowed for predicting end-stage renal disease in patients diagnosed with IgA nephropathy [51]. A similar analysis methodology based on the comparison of machine learning techniques was performed by Konieczny et al. to predict remission of IgA nephropathy [52].
ANN has a complex structure capable of discovering non-linear relationships, but requires a sufficiently large amount of data to learn. ANN requires data rescaling and the performance improves after data standardization. Compared to an RFC, a neural network such as a multilayer perceptron requires the adjustment of more parameters, e.g., the number of neurons in hidden layers.
More complex neural networks such as deep learning are used in analyzing complex structures, finding patterns, or detecting irregularities. Additionally, deep learning may use data augmentation to improve the quality of classification. Deep neural networks allow for the most advanced data analysis of the presented methods. However, the greatest technical resources and a sufficiently large amount of good-quality data are necessary.
Table 2 summarizes major applications of AI methods, characterizing their requirements and major features.

6. Limitations

6.1. Missing Data

Databases, especially those collected over a long period of time, may have missing data due to individual cells or blocks. There may be missing data in individual patient records, or in the measurements corresponding to column variables. Artificial intelligence allows to solve the problem of missing data with the technical improvement of the algorithm’s performance, but it also distorts real observations. Methods such as random forest are insensitive to irrelevant variables and can independently eliminate their involvement in modeling. However, care should be taken to ensure that the results obtained are primarily related to the clinical situation, and not merely correct.
Missing data can be filled with, among others, some version of random forests, as well as kNN imputer [35]. The kNN method involves deciding based on the local surroundings of the point. Empty cells may be interpreted as a zero value depending on the computer environment and software. While creating a database, one should clearly distinguish between an empty cell and a parameter with a value of zero.

6.2. Low Number of Patient Records

The main aspect that limits the use of machine learning with a small dataset is the difficulty in training a model that always converges mathematically to the optimal one. It is difficult to cross-validate such a model because the divided subsets are very small and, at the same time, single errors significantly worsen the final result. However, this difficulty does not exclude the potential of machine learning efficiency on small sets. In the literature, it is possible to encounter the use of a RFC in genomics, where the data is often small and contains many variable parameters [34,49,50]. Our own experience has shown the possibility of using RFC in the prediction of kidney dysfunction among children after hematopoietic stem cell transplantation and the need for dialysis in those with chronic kidney disease [54,55].
An additional problem, affecting small datasets, is the difficulty in validation. Whereas large datasets provide reasonable division into a training set, in small datasets, such division carries the risk of distorting the measured results. A sufficiently large training set ensures adequate representation of individual classes in cross-validation and reliability of the measured results. Small sets raise the risk of individual element prevalence and significant disturbance in the measured performance. In the case of small datasets, leave-one-out cross-validation (LOOCV) can be used [56]. LOOCV involves cross-validation with each element validating the remaining elements of the training set, so the number of obtained measurements corresponds to the number of elements in the set. For large sets of elements, 5 or 10, cross-validation is performed, which similarly creates 5 or 10 divisions into the training and validation sets [14].
The intuitive hypothesis that a larger dataset improves prediction and classification performance is confirmed by experiments [57]. Certain issues, however, reach the limit of further refinement after a certain number of samples is reached [58]. This means that collecting more additional data, which is labor-intensive and time-consuming, may not yield a sufficiently large improvement in model performance. A significant change in the amount of data does not significantly impact the final result. A small number of samples can lead to overfitting to the training data and poor performance on the testing data, resulting in poor model performance. Based on the literature, it appears that an appropriate dataset size range would be 10 to 100 times the number of features in the dataset [57,58,59,60,61]. Naturally, this is a rough estimate and, especially in proteomics, may not be applicable when analyzing many proteins and protein products.

6.3. Large Number of Variable Parameters

A large number of parameters with a relatively small number of patient records may lead to overfitting the model to the training data and, as a result, show poor performance on new data. The risk of overfitting also applies to the models that are too complex. More advanced models, such as neural networks, have a complexity that allows complete adaptation to the training data. However, such a model may not be able to operate effectively on new data that is outstandingly different from the training data.
While designing the database and generating predictive models, the appropriateness of selecting specific features should be considered. The clinical justification of chosen numbers is essential, because only the researcher is aware of the context of the study and the clinical environment.
RFC allows to measure the impact of variables on the decision-making process with the Gini importance index, quantifying the contribution of a given variable to classification or regression. In this way, the number of input variables can be reduced by discarding the least important ones. New parameters relevant to the clinical problem under study may be discovered during the development of the model [53,62,63].

6.4. Selection of the Method

As demonstrated by Konieczny et al., various methods may achieve different results with the same input data, owing to the algorithm that was chosen [52]. There are no unanimous recommendations regarding the choice of a specific method or the decisive advantage of one technique over another. Certain methods, as summarized in Table 2, offer advantages inherent to their design. These include complexity, adaptability, robustness to large datasets, flexibility, and insensitivity to missing or bad data. Method selection is largely a case study of the clinical problem and the question posed [64].

6.5. Dependent Variables and Augmented Data

Classical statistical analysis struggles with the dependence of variables. In the case of machine learning, augmented data can be introduced. Augmented data is used in deep learning and example operations involve scaling or rotating the analyzed images [65]. In the case of tabular data analysis, data expansion may consist of introducing variables that are derivatives of other variables. BMI as a derivative of weight and height could serve as an example. The introduction of augmented data with a valid clinical or real-world context should be considered, as it provides context assigned by the researcher. A simple example is a classifier that assigns points within a circle with a given radius. The introduction of derived variables—coordinate squares—significantly simplifies the task. It should be noted that the time and technical complexity of models increase with the number of variables, but imposing variables is an expression of the researcher’s control over the model [66].

7. Accuracy Assessment Methods in Machine Learning

Various machine learning techniques are characterized by differential performance in relation to the input data. The main requirement is the completeness of the data used to generate the model. Moreover, capabilities of a model usually increase with its complexity, although a simpler model may perform better under certain conditions. Simpler models show some insensitivity to missing data or outlier data. RFCs do not require either scaling or standardization of data, whereas deep networks use scaling, standardization, and augmentation of data to improve performance. Although greater computational complexity is the consequence, the significantly improved performance may compensate for it. Models based on decision trees are clearer in interpretation than neural networks, yet the neural networks provide more possibilities and potential applications. However, due to their complexity and extensive structure, the manual interpretation is difficult or practically impossible.
Below is a confusion matrix defined in a general sense:
M = T P F N F P T N
Correctly classified values are TP (true positive) and TN (true negative), respectively, and incorrectly classified values are FP (false positive) and FN (false negative), respectively. An ideal classifier is the one that correctly distinguishes true and false instances, and therefore, the FP and FN values under such conditions are zero.
Based on such a matrix, various measures of model behavior can be derived. However, it should be taken into account whether the dataset is balanced in terms of the presence of both positive and negative instances.
The Matthews correlation coefficient (MCC) is a binary classifier metric that is based on all fields of the confusion matrix and is particularly useful in unbalanced sets. Additionally, this metric is not oriented towards the positive class and remains symmetrical towards the negative class [67,68].
M C C = T P · T N F P · F N T P + F P · T P + F N · T N + F P · T N + F N
In our previous work, we have emphasized high MCC scores of our models [69]. MCC takes values in the range from −1.0 to 1.0
Accuracy is the most intuitive parameter determining the ratio of correct classifications to all instances.
A c c u r a c y = T P + T N A l l   t h e   i n s t a n c e s = T P + T N T P + F P + F N + T N
This parameter is sensitive to the balance of the set. For example, if the set contains 90 data of one class and 10 of the other class, the classifier assigning the dominant class to any sample will obtain TP 90, FP 10, FN 0, TN 0, which gives an accuracy of 0.9.
Precision, also called positive prediction value, is the probability that a positive result is true.
P r e c i s i o n = T P T P + F P
Recall, also called sensitivity, is the probability of detecting a positive instance.
R e c a l l = T P T P + F N
The receiver-operator curve (ROC) shows how the true positive rate (TPR) changes in relation to the false positive rate (FPR) and is a commonly used metric describing the discriminatory ability of a binary classifier. The area under the ROC (AUROC) determines how close the measured classifier is to the ideal classifier (AUROC = 1.0).
F1-score is the harmonic mean of precision and recall, a simplified measure relative to MCC, but not taking into account true negative classifications.
F 1 s c o r e = 2 p r e c i s i o n 1 + r e c a l l 1 = 2 · p r e c i s i o n · r e c a l l p r e c i s i o n + r e c a l l = 2 T P 2 T P + F P + F N

8. How to Choose an Appropriate AI Tool? Practical Summary

Research in the field of nephrology is often conducted on small groups of patients because it often involves disorders that are rare in the general population. The utilization of standard statistical techniques to analyze research outcomes on small sample sizes may result in the inability to achieve statistical significance. Additionally, missing data, outlier data, or a large number of variables may further complicate the interpretation of results.
In such conditions, AI techniques can be extremely helpful in prediction of development, risk factors, mortality in a variety of diseases, including glomerulonephritis, AKI, CKD, and indications for renal replacement therapy. The range of applications varies from imaging diagnostics, through surgery and omics, to genetic testing.
RFCs do not require either scaling or standardization of data, whereas deep networks use scaling, standardization, and augmentation of data to improve performance. Models based on decision trees are clearer in interpretation than neural networks, whereas the neural network provides more possibilities and potential applications. However, due to its complexity and extensive structure, manual interpretation is difficult.
Therefore, the appropriate selection of data for analysis is essential for machine learning. Missing data clouds the picture, so removing such records or columns and not including them in the analysis should be considered. However, such an attitude may become problematic in the case of rare diseases. Missing data imputation using the kNN method may be the best solution, as it can to some extent deduce what the missing cell may look like, based on the neighborhood similarity.
Regardless of the chosen technique, it is essential to assess the effectiveness of any model, based on the accuracy assessment, mainly the Matthews correlation coefficient and AUROC, as well as on the external validation.

9. AI-Driven Proteomic Diagnostics—Recent Nephrological Perspective

Recent years in medicine have been marked by extensive AI-driven research in the field of nephrology.
Among multiple trigger factors, the COVID-19 pandemic seemed the most stimulating one, paving the way for analysis with the use of artificial intelligence tools. Based on the known pathways of relationships between inflammatory markers, cytokines, and chemokines, one could specifically ask artificial intelligence what conclusions can be drawn using algorithms more complex than the classical analysis [70]. The virus was detected in kidney cells, and the pathogen itself, through metabolic and proteomic pathways, triggered a cytokine storm, damaging almost every organ and system in the body [71,72,73]. The Polish CRACoV-AKI prognostic model enabled the stratification of patients with SARS-CoV-2 infection, according to the risk of development of acute kidney injury [74]. Therefore, the SARS-CoV-2 pandemic has fueled medicine based on molecular diagnostics and the use of artificial intelligence tools, showing their superiority over classical statistics and encouraging the continuators. Some of the examples are listed below.
The introduction of newly discovered proteomic markers of early organ damage, such as the urinary exosomal WT-1 in diabetic nephropathy, is a promising new parameter that can detect patients before the clinical manifestation of organ damage, including increased UPCR, occurs [75,76].
Chronic kidney disease is a multisystem dysfunction, with cardiovascular complications leading to life-threatening conditions. Combining clinical parameters and patient/family history in a prognostic model with proteomic data offers the potential for tools to classify patients into risk groups for cardiovascular events, including sudden cardiac death. According to current knowledge, patients with chronic dialysis are at increased risk of cardiac events. Connective tissue growth factors (CTGF) and NT-proBNP play a significant diagnostic role, improving the ability to identify patients at risk for cardiovascular mortality (AUC 0.883) or sudden cardiac death (AUC 0.877) during a 2-year follow-up period. Although nonspecific, the CTGF and NT-proBNP biomarkers reflect the level of myocardial fibrosis and myocardial overload, respectively [77].
Idiopathic nephrotic syndrome is the most common primary glomerulopathy among children. Its beginning at a young age carries the burden of steroid-based and long-term immunosuppression. In a study by Yagin et al., predictive models were developed to identify steroid-dependent and steroid-resistant nephrotic syndrome in children based on a metabolic profile, taking into account many parameters, in particular serum glucose, creatinine, glycerate, carnitine, betaine [78]. The best model differentiated the two conditions with an accuracy of 0.87 and an AUC of 0.92. Indeed, in this case, artificial intelligence addressed a fundamental issue of nephrology in its entirety.
IgA nephropathy, the most common primary form of mesangial glomerulonephritis in adults, plays a significant role in the development of chronic kidney disease. There are promising biochemical marker studies that, when combined with artificial intelligence tools, enable identification of patients at risk of disease progression before glomerular structure is damaged [79].
Contrast-induced nephropathy constitutes a niche of nephrological conditions, yet the increasing demand for imaging diagnostics urges early identification of risk groups. Lee et al. and Gonzales et al. point to the potential for developing models predicting contrast-induced nephropathy, based on classic markers of renal function and multiple markers of early damage [80,81].
Kidney transplant patients are exposed to the adverse effects of the immunosuppressant tacrolimus, which, on one hand, protects the transplanted organ from rejection and, on the other, triggers nephrotoxicity, which becomes irreversible at some point. The potential targets for drugs, that can inhibit fibrosis induced by chronic tacrolimus administration, were identified. These include PRMT1, the presence of which triggers increased STAT3 levels and, consequently, increased expression of β6 subunits in the αVβ6 integrin, which is known to be elevated in chronic kidney disease [82].
The most common histological type of kidney cancer, accounting for 70% to 90% of all cases, is clear cell renal cell carcinoma (ccRCC), characterized by significant heterogeneity. Modeling based on upregulation and downregulation of genes selected through statistical analysis allows patient stratification for survival and therapeutic targeting [83].
The authors of all above-mentioned studies point out that the multitude of parameters allows for personalized medicine, tailored to every patient—those suffering from acute dysfunction, either developing chronic disease with comorbidities, or undergoing medical procedures. Therefore, the use of AI-driven analyses may serve for holistic purposes including diagnosis, treatment, and prophylaxis.

10. Back to the Future of Nephrology with New Markers

The future of artificial intelligence applications in medicine lies in network models, analyzing interactions between entities like proteins, genes, cells, or drugs, and leading to personalized management. Although the data on building such models in nephrology is scarce, the examples listed below, concerning other specialties, provide the area for future investigation and show the most effective tools.
Correctly registered molecular structures can be analyzed using artificial neural network models to discover the most effective catalytic reaction systems. The models then analyze the individual functional structures of the drug and target. Deep learning analysis of molecular models provides information that can be used to saturate molecular simulations with data on critical molecule–drug interactions. Such models achieve AUC above 0.9 [84]. Previous research highlights the importance of high-quality data provided for analysis [85].
The analysis of drug–target point interactions may enable the foundation of modern personalized medicine, focused on effective treatment with the most appropriate medication. The work of Mess et al. shows that models based on RFCs and XGBoost classifiers are effective. These models were tasked with predicting the target point for a selected drug in metastatic cancer (AUCs of 0.93 and 0.94, respectively), and in Parkinson’s disease (AUCs of 0.91 and 0.92, respectively) [86].
Artificial intelligence models are able to analyze information on the overexpression of numerous genes, representing the complexity of biochemical processes and numerous intercellular interactions involving proteins, cytokines, and chemokines, in the form of a digital, scalable model. Models examining intercellular interactions at the expression level of selected gene packages offer the potential to classify SARS-CoV-2 disease into severity categories. The deep learning network is able to identify gene expression patterns typical of immune diseases that correlate with the clinical severity of SARS-CoV-2 infection [87]. In short, the model established potential pathways of gene interactions based on measured gene expression. Machine learning models, despite their clearly simpler structure compared to deep learning, are not inferior in terms of predictive ability in the case of COVID-19. Classically determined biochemical parameters were of good predictive value, but extending the data with metabolomics significantly improved the precision of the prediction of SARS-CoV-2 infection severity. Not surprisingly, the best model was that of RFC. The kNN classifier proved to be of comparable quality, because they both achieved an AUC of 0.92 [88].
Deep learning identifies intercellular interactions at the molecular level that defy classical statistical analysis. Kim et al. were able to analyze over 36,000 genes in nearly 500,000 non-small cell lung cancer cells and over 700,000 colorectal cancer cells using this method [89]. On average, both models performed extremely well, achieving performance, as measured by accuracy and AUC, of 0.991 (AUC of 1.0) and 0.952 (AUC of 0.99), respectively [89].
Last, but not least, nonspecific markers, when combined with others, can provide significant prognostic potential. For example, the combination of interleukin 8- and 10-, TNFα, Brain-derived neurotrophic factor (BDNF), Nerve Growth Factor (NGF), and the chemokine CXCL10, arranged in a decision tree model, quite accurately differentiates treatment-resistant lower urinary tract dysfunction (LUTD) [90]. Therefore, it is clear that properly collected and processed data can yield promising results in AI-based research in the future.

11. Ethical Aspects of AI in Nephrology

Ethical considerations regarding AI in medicine focus primarily on responsibility for the results and outcomes of achieved diagnostic goals. In this regard, the degree of reliance on the chosen method and personal responsibility for the accuracy of the classification are crucial. The allocation of benefits from the achieved results remains an ethical issue, especially in the field of scientific work. Among the important technical issues, there are the reliability and flexibility of solutions [91,92].
The cited literature presents a developing picture of new tools that can support the care of patients with chronic kidney disease, end-stage renal failure, including dialysis patients and patients after kidney transplantation. It is important to critically assess the results achieved, so that they serve the patient’s well-being and have a positive impact on science.

12. Conclusions

Various machine learning techniques are characterized by differential performance in relation to the input data. The main requirement is the completeness of the data used to generate the model. Moreover, capabilities of a model usually increase with its complexity, although under certain conditions, a simpler model may perform better. Simpler models show some insensitivity to missing data or outlier data.
Further AI applications in nephrology are highly expected. In order to make the most of AI potential, basic rules of its efficient implementation, summarized in this review, have to be taken into account.

Author Contributions

Conceptualization, J.S.; methodology, J.S.; software, J.S.; validation, J.S.; formal analysis, J.S.; investigation, J.S.; resources, J.S.; data curation, J.S.; writing—original draft preparation, J.S.; writing—review and editing, T.G. and K.M.; visualization, J.S.; supervision, T.G. and K.M.; project administration, J.S. and K.M.; funding acquisition, T.G. and K.M. All authors have read and agreed to the published version of the manuscript.

Funding

The project is financed by the Foundation “Na Ratunek Dzieciom z Chorobą Nowotworową” (FNRD.C210.19.002).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data was created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
A1ATAlpha-1 antitrypsin
ADPKDAutosomal dominant polycystic kidneys disease
AFMAfamin
AIArtificial intelligence
AKIAcute kidney injury
AlexNetAlex Krizhevsky Network
ANNArtificial neural network
ANXA7Annexin A7
APODApolipoprotein D
ARID4BAT-rich interactive domain-containing protein 4B
AUCArea under curve
BDNFBrain-derived neurotrophic factor
BMIBody mass index
C9Complement component 9
CAD238 Coronary Artery Disease 238
CARTsClassification and regression trees
ccRCCClear cell renal carcinoma
CKAP4Cytoskeleton-associated protein 4
CKD273Chronic kidney disease 273
CNNConvolutional neural network
CPCeruloplasmin
CTGFConnective tissue growth factors
CXCL10C-X-C motif chemokine ligand 10
DGFDelayed graft function
DKDDiabetic kidney disease
DoSDifference of slope
eGFREstimated Gloimerular Filtration Rate
EOMESEomesodermin
EPTSEstimated post-transplant survival
ESAsErythropoiesis-stimulating agents
ESKDEnd-stage kidney disease
ETExtra trees
FNFalse negative
FPFalse positive
FPRFalse positive rate
GIRK-1G protein-activated inward rectifier potassium channel 1
GPX3Glutathione Peroxidase 3
HbHemoglobin
HF2Heart failure 2
HRHeart rate
ICUIntensive Care Unit
IGFBP2Insulin-like Growth Factor-Binding Protein 2
KDPIKidney Donor Profile Index
KDRIKidney Donor Risk Index
kNNK-nearest neighbors classifier
LASSOLeast Absolute Shrinkage and Selection Operator
LGBMLight Gradient-Boosting Machine
LIFLeukemia inhibitory factor
LOOCVLeave-one-out cross-validation
LUTDLower urinary tract dysfunction
MCCMatthews correlation coefficient
MLPMultilayer perceptron
NGFNerve Growth Factor
OPNOsteopontin
PLMNPlasminogen
PRMT1Protein Arginine Methyltransferase 1
PTX3Pentraxin-3
ResNetResidual Network
RFCRandom Forest Classifier
RFCsRandom Forest Classifiers
RIResistive index
ROCReceiver-operator curve
sAlbSerum albumin
SHBGSex hormone binding globulin
STAT1Signal transducer and activator of transcription 1
STAT3Signal transducer and activator of transcription 3
SVMSupport vector machine
SVM-RFESupport vector machine recursive feature elimination
SWEShear wave elastography
TFTranscription factor
TNTrue negative
TPTrue positive
TPRTrue positive rate
UACRUrinary albumin creatinine ratio
UCIUniversity of California Irvine
U-NetU-shaped architecture Network
VPS4AVacuolar protein sorting-associated protein 4A
XGBoosteXtreme Gradient Boosting

References

  1. McCarthy, J.; Minsky, M.L.; Rochester, N.; Shannon, C.E. A Proposal for the Dartmouth Summer Research Project on Artificial Intelligence. 1955. Available online: http://www-formal.stanford.edu/jmc/history/dartmouth/dartmouth.html (accessed on 12 January 2026).
  2. Xie, Y.; Zhai, Y.; Lu, G. Evolution of artificial intelligence in healthcare: A 30-year bibliometric study. Front. Med. 2025, 11, 1505692. [Google Scholar] [CrossRef]
  3. Statt, N. The AI boom is happening all over the world, and it’s accelerating quickly. The Verge, 12 December 2018. [Google Scholar]
  4. Ding, X.; Shang, B.; Xie, C.; Xin, J.; Yu, F. Artificial intelligence in the COVID-19 pandemic: Balancing benefits and ethical challenges in China’s response. Humanit. Soc. Sci. Commun. 2025, 12, 245. [Google Scholar] [CrossRef]
  5. Reznichenko, A.; Nair, V.; Eddy, S.; Fermin, D.; Tomilo, M.; Slidel, T.; Ju, W.; Henry, I.; Badal, S.S.; Wesley, J.D.; et al. Unbiased kidney-centric molecular categorization of chronic kidney disease as a step towards precision medicine. Kidney Int. 2024, 105, 1263–1278. [Google Scholar] [CrossRef]
  6. Schork, A.; Fritsche, A.; Schleicher, E.D.; Peter, A.; Heni, M.; Stefan, N.; von Schwartzenberg, R.J.; Guthoff, M.; Mischak, H.; Siwy, J.; et al. Differential risk assessment in persons at risk of type 2 diabetes using urinary peptidomics. Metabolism 2025, 167, 156174. [Google Scholar] [CrossRef] [PubMed]
  7. Sato, N.; Uchino, E.; Kojima, R.; Sakuragi, M.; Hiragi, S.; Minamiguchi, S.; Haga, H.; Yokoi, H.; Yanagita, M.; Okuno, Y. Evaluation of Kidney Histological Images Using Unsupervised Deep Learning. Kidney Int. Rep. 2021, 6, 2445–2454. [Google Scholar] [CrossRef]
  8. Thongprayoon, C.; Vaitla, P.; Jadlowiec, C.C.; Leeaphorn, N.; Mao, S.A.; Mao, M.A.; Pattharanitima, P.; Bruminhent, J.; Khoury, N.J.; Garovic, V.D.; et al. Use of Machine Learning Consensus Clustering to Identify Distinct Subtypes of Black Kidney Transplant Recipients and Associated Outcomes. JAMA Surg. 2022, 157, e221286. [Google Scholar] [CrossRef]
  9. Thongprayoon, C.; Vaitla, P.; Jadlowiec, C.C.; Mao, S.A.; Mao, M.A.; Acharya, P.C.; Leeaphorn, N.; Kaewput, W.; Pattharanitima, P.; Tangpanithandee, S.; et al. Differences between kidney retransplant recipients as identified by machine learning consensus clustering. Clin. Transplant. 2023, 37, e14943. [Google Scholar] [CrossRef]
  10. Yuan, Z.; Lv, G.; Liu, X.; Xiao, Y.; Tan, Y.; Zhu, Y. Machine learning selection of basement membrane-associated genes and development of a predictive model for kidney fibrosis. Sci. Rep. 2025, 15, 6567. [Google Scholar] [CrossRef]
  11. Massy, Z.A.; Lambert, O.; Metzger, M.; Sedki, M.; Chaubet, A.; Breuil, B.; Jaafar, A.; Tack, I.; Nguyen-Khoa, T.; Alves, M.; et al. Machine Learning-Based Urine Peptidome Analysis to Predict and Understand Mechanisms of Progression to Kidney Failure. Kidney Int. Rep. 2022, 8, 544–555. [Google Scholar] [CrossRef] [PubMed]
  12. Deo, R.; Dubin, R.F.; Ren, Y.; Wang, J.; Feldman, H.; Shou, H.; Coresh, J.; Grams, M.E.; Surapaneni, A.L.; Cohen, J.B.; et al. Proteomic Assessment of the Risk of Secondary Cardiovascular Events among Individuals with CKD. J. Am. Soc. Nephrol. 2025, 36, 231–241. [Google Scholar] [CrossRef]
  13. Dong, J.; Feng, T.; Thapa-Chhetry, B.; Cho, B.G.; Shum, T.; Inwald, D.P.; Newth, C.J.L.; Vaidya, V.U. Machine learning model for early prediction of acute kidney injury (AKI) in pediatric critical care. Crit. Care 2021, 25, 288. [Google Scholar] [CrossRef] [PubMed]
  14. Zhao, X.; Lu, Y.; Li, S.; Guo, F.; Xue, H.; Jiang, L.; Wang, Z.; Zhang, C.; Xie, W.; Zhu, F. Predicting renal function recovery and short-term reversibility among acute kidney injury patients in the ICU: Comparison of machine learning methods and conventional regression. Ren. Fail. 2022, 44, 1326–1337. [Google Scholar] [CrossRef]
  15. Zhang, Z.; Ho, K.M.; Hong, Y. Machine learning for the prediction of volume responsiveness in patients with oliguric acute kidney injury in critical care. Crit. Care 2019, 23, 112. [Google Scholar] [CrossRef]
  16. Eghbali, N.; Alhanai, T.; Ghassemi, M.M. Patient-Specific Sedation Management via Deep Reinforcement Learning. Front. Digit. Health 2021, 3, 608893. [Google Scholar] [CrossRef]
  17. Yang, Z.; Tian, Y.; Zhou, T.; Zhu, Y.; Zhang, P.; Chen, J.; Li, J. Optimization of Dry Weight Assessment in Hemodialysis Patients via Reinforcement Learning. IEEE J. Biomed. Health Inform. 2022, 26, 4880–4891. [Google Scholar] [CrossRef]
  18. Escandell-Montero, P.; Chermisi, M.; Martínez-Martínez, J.M.; Gómez-Sanchis, J.; Barbieri, C.; Soria-Olivas, E.; Mari, F.; Vila-Francés, J.; Stopper, A.; Gatti, E.; et al. Optimization of anemia treatment in hemodialysis patients via reinforcement learning. Artif. Intell. Med. 2014, 62, 47–60. [Google Scholar] [CrossRef] [PubMed]
  19. Gaweda, A.E.; Muezzinoglu, M.K.; Jacobs, A.A.; Aronoff, G.R.; Brier, M.E. Model predictive control with reinforcement learning for drug delivery in renal anemia management. In 2006 International Conference of the IEEE Engineering in Medicine and Biology Society; IEEE: Piscataway, NJ, USA, 2006; pp. 5177–5180. [Google Scholar]
  20. Kha, Q.-H.; Le, V.-H.; Hung, T.N.K.; Nguyen, N.T.K.; Le, N.Q.K. Development and Validation of an Explainable Machine Learning-Based Prediction Model for Drug–Food Interactions from Chemical Structures. Sensors 2023, 23, 3962. [Google Scholar] [CrossRef]
  21. Zou, Y.; Zhao, L.; Zhang, J.; Wang, Y.; Wu, Y.; Ren, H.; Wang, T.; Zhang, R.; Wang, J.; Zhao, Y.; et al. Development and internal validation of machine learning algorithms for end-stage renal disease risk prediction model of people with type 2 diabetes mellitus and diabetic kidney disease. Ren. Fail. 2022, 44, 562–570. [Google Scholar] [CrossRef]
  22. Zhang, W.; Liu, X.; Dong, Z.; Wang, Q.; Pei, Z.; Chen, Y.; Zheng, Y.; Wang, Y.; Chen, P.; Feng, Z.; et al. New Diagnostic Model for the Differentiation of Diabetic Nephropathy from Non-Diabetic Nephropathy in Chinese Patients. Front. Endocrinol. 2022, 13, 913021. [Google Scholar] [CrossRef]
  23. Fan, G.; Gong, T.; Lin, Y.; Wang, J.; Sun, L.; Wei, H.; Yang, X.; Liu, Z.; Li, X.; Zhao, L.; et al. Urine proteomics identifies biomarkers for diabetic kidney disease at different stages. Clin. Proteom. 2021, 18, 32. [Google Scholar] [CrossRef]
  24. Kononikhin, A.S.; Brzhozovskiy, A.G.; Bugrova, A.E.; Chebotareva, N.V.; Zakharova, N.V.; Semenov, S.; Vinogradov, A.; Indeykina, M.I.; Moiseev, S.; Larina, I.M.; et al. Targeted MRM Quantification of Urinary Proteins in Chronic Kidney Disease Caused by Glomerulopathies. Molecules 2023, 28, 3323. [Google Scholar] [CrossRef] [PubMed]
  25. Simon, N.; Friedman, J.H.; Hastie, T.; Tibshirani, R. Regularization Paths for Cox’s Proportional Hazards Model via Coordinate Descent. J. Stat. Softw. 2011, 39, 1–13. [Google Scholar] [CrossRef]
  26. Ogunleye, A.; Wang, Q.G. XGBoost Model for Chronic Kidney Disease Diagnosis. IEEE/ACM Trans. Comput. Biol. Bioinform. 2020, 17, 2131–2140. [Google Scholar] [CrossRef] [PubMed]
  27. Chang, H.H.; Chiang, J.H.; Tsai, C.C.; Chiu, P.F. Predicting hyperkalemia in patients with advanced chronic kidney disease using the XGBoost model. BMC Nephrol. 2023, 24, 169. [Google Scholar] [CrossRef]
  28. Liu, J.; Wu, J.; Liu, S.; Li, M.; Hu, K.; Li, K. Predicting mortality of patients with acute kidney injury in the ICU using XGBoost model. PLoS ONE 2021, 16, e0246306. [Google Scholar] [CrossRef]
  29. Chen, T.; Li, X.; Li, Y.; Xia, E.; Qin, Y.; Liang, S.; Xu, F.; Liang, D.; Zeng, C.; Liu, Z. Prediction and Risk Stratification of Kidney Outcomes in IgA Nephropathy. Am. J. Kidney Dis. Off. J. Natl. Kidney Found. 2019, 74, 300–309. [Google Scholar] [CrossRef]
  30. Lee, A.M.; Hu, J.; Xu, Y.; Abraham, A.G.; Xiao, R.; Coresh, J.; Rebholz, C.; Chen, J.; Rhee, E.P.; Feldman, H.I.; et al. Using Machine Learning to Identify Metabolomic Signatures of Pediatric Chronic Kidney Disease Etiology. J. Am. Soc. Nephrol. JASN 2022, 33, 375–386. [Google Scholar] [CrossRef]
  31. McCallion, S.; McLarnon, T.; Cooper, E.; English, A.R.; Watterson, S.; Chemaly, M.E.; McGeough, C.; Eakin, A.; Ahmed, T.; Gardiner, P.; et al. Senescence Biomarkers CKAP4 and PTX3 Stratify Severe Kidney Disease Patients. Cells 2024, 13, 1613. [Google Scholar] [CrossRef]
  32. Glazyrin, Y.E.; Veprintsev, D.V.; Ler, I.A.; Rossovskaya, M.L.; Varygina, S.A.; Glizer, S.L.; Zamay, T.N.; Petrova, M.M.; Minic, Z.; Berezovski, M.V.; et al. Proteomics-Based Machine Learning Approach as an Alternative to Conventional Biomarkers for Differential Diagnosis of Chronic Kidney Diseases. Int. J. Mol. Sci. 2020, 21, 4802. [Google Scholar] [CrossRef] [PubMed]
  33. Tisch, A.; Madapoosi, S.; Blough, S.; Rosa, J.; Eddy, S.; Mariani, L.; Naik, A.; Limonte, C.; McCown, P.; Menon, R.; et al. Identification of kidney cell types in scRNA-seq and snRNA-seq data using machine learning algorithms. Heliyon 2024, 10, e38567. [Google Scholar] [CrossRef]
  34. Yue, S.; Li, S.; Huang, X.; Liu, J.; Hou, X.; Zhao, Y.; Niu, D.; Wang, Y.; Tan, W.; Wu, J. Machine learning for the prediction of acute kidney injury in patients with sepsis. J. Transl. Med. 2022, 20, 215. [Google Scholar] [CrossRef]
  35. Emmanuel, T.; Maupong, T.; Mpoeleng, D.; Semong, T.; Mphago, B.; Tabona, O. A survey on missing data in machine learning. J. Big Data 2021, 8, 140. [Google Scholar] [CrossRef]
  36. Mahboob, T.; Ijaz, A.; Shahzad, A.; Kalsoom, M. Handling Missing Values in Chronic Kidney Disease Datasets Using KNN, K-Means and K-Medoids Algorithms. In Proceedings of the 2018 12th International Conference on Open Source Systems and Technologies (ICOSST), Lahore, Pakistan, 19–21 December 2018; pp. 76–81. [Google Scholar] [CrossRef]
  37. Hermsen, M.; de Bel, T.; den Boer, M.; Steenbergen, E.J.; Kers, J.; Florquin, S.; Roelofs, J.J.T.H.; Stegall, M.D.; Alexander, M.P.; Smith, B.H.; et al. Deep Learning-Based Histopathologic Assessment of Kidney Tissue. J. Am. Soc. Nephrol. JASN 2019, 30, 1968–1979. [Google Scholar] [CrossRef]
  38. Kers, J.; Bülow, R.D.; Klinkhammer, B.M.; Breimer, G.E.; Fontana, F.; Abiola, A.A.; Hofstraat, R.; Corthals, G.L.; Peters-Sengers, H.; Djudjaj, S.; et al. Deep learning-based classification of kidney transplant pathology: A retrospective, multicentre, proof-of-concept study. Lancet Digit. Health 2022, 4, e18–e26. [Google Scholar] [CrossRef] [PubMed]
  39. Bouteldja, N.; Klinkhammer, B.M.; Bülow, R.D.; Droste, P.; Otten, S.W.; Freifrau von Stillfried, S.; Moellmann, J.; Sheehan, S.M.; Korstanje, R.; Menzel, S.; et al. Deep Learning-Based Segmentation and Quantification in Experimental Kidney Histopathology. J. Am. Soc. Nephrol. JASN 2021, 32, 52–68. [Google Scholar] [CrossRef] [PubMed]
  40. Zhao, H.; Yuan, Y.; Chen, S.; Yao, Y.; Bi, C.; Liu, C.; Sun, G.; Su, H.; Li, X.; Li, X.; et al. Deep learning-based multi-omics study reveals the polymolecular phenotypic of diabetic kidney disease. Clin. Transl. Med. 2023, 13, e1301. [Google Scholar] [CrossRef] [PubMed]
  41. Chen, Z.; Ying, T.C.; Chen, J.; Wu, C.; Li, L.; Chen, H.; Xiao, T.; Huang, Y.; Chen, X.; Jiang, J.; et al. Using elastography-based multilayer perceptron model to evaluate renal fibrosis in chronic kidney disease. Ren. Fail. 2023, 45, 2202755. [Google Scholar] [CrossRef]
  42. Bae, T.W.; Kim, M.S.; Park, J.W.; Kwon, K.K.; Kim, K.H. Multilayer Perceptron-Based Real-Time Intradialytic Hypotension Prediction Using Patient Baseline Information and Heart-Rate Variation. Int. J. Environ. Res. Public Health 2022, 19, 10373. [Google Scholar] [CrossRef]
  43. Gallego, J.; Pedraza, A.; Lopez, S.; Steiner, G.; Gonzalez, L.; Laurinavicius, A.; Bueno, G. Glomerulus Classification and Detection Based on Convolutional Neural Networks. J. Imaging 2018, 4, 20. [Google Scholar] [CrossRef]
  44. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  45. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, Part III 18; Springer International Publishing: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar] [CrossRef]
  46. Zhao, Z.; Chen, H.; Li, J.; Wang, L. Boundary Attention U-Net for Kidney and Kidney Tumor Segmentation. In Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Glasgow, UK, 11–15 July 2022; pp. 1540–1543. [Google Scholar] [CrossRef]
  47. Rombolotti, M.; Sangalli, F.; Cerullo, D.; Remuzzi, A.; Lanzarone, E. Automatic cyst and kidney segmentation in autosomal dominant polycystic kidney disease: Comparison of U-Net based methods. Comput. Biol. Med. 2022, 146, 105431. [Google Scholar] [CrossRef] [PubMed]
  48. Uhm, K.H.; Jung, S.W.; Choi, M.H.; Shin, H.K.; Yoo, J.I.; Oh, S.W.; Kim, J.Y.; Kim, H.G.; Lee, Y.J.; Youn, S.Y.; et al. Deep learning for end-to-end kidney cancer diagnosis on multi-phase abdominal computed tomography. NPJ Precis. Oncol. 2021, 5, 54. [Google Scholar] [CrossRef]
  49. Fernández-Delgado, M.; Cernadas, E.; Barro, S. Do We Need Hundreds of Classifers to Solve Real World Classifcation Problems? J. Mach. Learn. Res. 2014, 15, 3133–3181. [Google Scholar]
  50. Konieczny, A.; Stojanowski, J.; Rydzyńska, K.; Kusztal, M.; Krajewska, M. Artificial Intelligence—A Tool for Risk Assessment of Delayed-Graft Function in Kidney Transplant. J. Clin. Med. 2021, 10, 5244. [Google Scholar] [CrossRef]
  51. Liu, Y.; Zhang, Y.; Liu, D.; Tan, X.; Tang, X.; Zhang, F.; Xia, M.; Chen, G.; He, L.; Zhou, L.; et al. Prediction of ESRD in IgA Nephropathy Patients from an Asian Cohort: A Random Forest Model. Kidney Blood Press. Res. 2018, 43, 1852–1864. [Google Scholar] [CrossRef]
  52. Konieczny, A.; Stojanowski, J.; Krajewska, M.; Kusztal, M. Machine Learning in Prediction of IgA Nephropathy Outcome: A Comparative Approach. J. Pers. Med. 2021, 11, 312. [Google Scholar] [CrossRef]
  53. Beghdadi, N.; Kitano, Y.; Golse, N.; Vibert, E.; Cunha, A.S.; Azoulay, D.; Cherqui, D.; Adam, R.; Allard, M.-A. Features Importance in Acute Kidney Injury After Liver Transplant: Which Predictors Are Relevant? Exp. Clin. Transplant. 2023, 21, 408–414. [Google Scholar] [CrossRef] [PubMed]
  54. Musiał, K.; Stojanowski, J.; Miśkiewicz-Bujna, J.; Kałwak, K.; Ussowicz, M. KIM-1, IL-18, and NGAL, in the Machine Learning Prediction of Kidney Injury among Children Undergoing Hematopoietic Stem Cell Transplantation—A Pilot Study. Int. J. Mol. Sci. 2023, 24, 15791. [Google Scholar] [CrossRef]
  55. Kawalec, A.; Stojanowski, J.; Mazurkiewicz, P.; Choma, A.; Gaik, M.; Pluta, M.; Szymański, M.; Bruciak, A.; Gołębiowski, T.; Musiał, K. Systemic Immune Inflammation Index as a Key Predictor of Dialysis in Pediatric Chronic Kidney Disease with the Use of Random Forest Classifier. J. Clin. Med. 2023, 12, 6911. [Google Scholar] [CrossRef]
  56. Acharjee, A.; Larkman, J.; Xu, Y.; Cardoso, V.R.; Gkoutos, G.V. A random forest based biomarker discovery and power analysis framework for diagnostics research. BMC Med. Genom. 2020, 13, 178. [Google Scholar] [CrossRef]
  57. Rajput, D.; Wang, W.J.; Chen, C.C. Evaluation of a decided sample size in machine learning applications. BMC Bioinform. 2023, 24, 48. [Google Scholar] [CrossRef]
  58. Bailly, A.; Blanc, C.; Francis, É.; Guillotin, T.; Jamal, F.; Wakim, B.; Roy, P. Effects of dataset size and interactions on the prediction performance of logistic regression and deep learning models. Comput. Methods Programs Biomed. 2022, 213, 106504. [Google Scholar] [CrossRef]
  59. Zantvoort, K.; Nacke, B.; Görlich, D.; Hornstein, S.; Jacobi, C.; Funk, B. Estimation of minimal data sets sizes for machine learning predictions in digital mental health interventions. NPJ Digit. Med. 2024, 7, 361. [Google Scholar] [CrossRef]
  60. Hammoudeh, Z.; Lowd, D. Training data influence analysis and estimation: A survey. Mach. Learn. 2024, 113, 2351–2403. [Google Scholar] [CrossRef]
  61. Gütter, J.; Kruspe, A.; Zhu, X.X.; Niebling, J. Impact of Training Set Size on the Ability of Deep Neural Networks to Deal with Omission Noise. Front. Remote Sens. 2022, 3, 932431. [Google Scholar] [CrossRef]
  62. Chen, X.; Ishwaran, H. Random forests for genomic data analysis. Genomics 2012, 99, 323–329. [Google Scholar] [CrossRef] [PubMed]
  63. Ghosh, D.; Cabrera, J. Enriched Random Forest for High Dimensional Genomic Data. IEEE/ACM Trans. Comput. Biol. Bioinform. 2022, 19, 2817–2828. [Google Scholar] [CrossRef] [PubMed]
  64. Ramezan, C.A.; Warner, T.A.; Maxwell, A.E.; Price, B.S. Effects of Training Set Size on Supervised Machine-Learning Land-Cover Classification of Large-Area High-Resolution Remotely Sensed Data. Remote Sens. 2021, 13, 368. [Google Scholar] [CrossRef]
  65. Hattab, G.; Arnold, M.; Strenger, L.; Allan, M.; Arsentjeva, D.; Gold, O.; Simpfendörfer, T.; Maier-Hein, L.; Speidel, S. Kidney edge detection in laparoscopic image data for computer-assisted surgery: Kidney edge detection. Int. J. Comput. Assist. Radiol. Surg. 2020, 15, 379–387. [Google Scholar] [CrossRef]
  66. Connor, C.W. Artificial Intelligence and Machine Learning in Anesthesiology. Anesthesiology 2019, 131, 1346–1359. [Google Scholar] [CrossRef] [PubMed]
  67. Chicco, D.; Jurman, G. The advantages of the Matthews correlation coefficient (MCC) over F1 score and accuracy in binary classification evaluation. BMC Genom. 2020, 21, 6. [Google Scholar] [CrossRef] [PubMed]
  68. Powers, D.M. Evaluation: From precision, recall and F-measure to ROC, informedness, markedness and correlation. arXiv 2020, arXiv:2010.16061. [Google Scholar] [CrossRef]
  69. Stojanowski, J.; Konieczny, A.; Lis, Ł.; Frosztęga, W.; Brzozowska, P.; Ciszewska, A.; Rydzyńska, K.; Sroka, M.; Krakowska, K.; Gołębiowski, T.; et al. The Artificial Neural Network as a Diagnostic Tool of the Risk of Clostridioides difficile Infection among Patients with Chronic Kidney Disease. J. Clin. Med. 2023, 12, 4751. [Google Scholar] [CrossRef] [PubMed]
  70. Eriksson, M.B.; Marks-Hultström, M.; Åberg, M.; Lipcsey, M.; Frithiof, R.; Larsson, A.O. Mapping Interactions Between Cytokines, Chemokines, Growth Factors, and Conventional Biomarkers in COVID-19 ICU-Patients. Int. J. Mol. Sci. 2025, 26, 11419. [Google Scholar] [CrossRef]
  71. Farkash, E.A.; Wilson, A.M.; Jentzen, J.M. Ultrastructural evidence for direct renal infection with SARS-CoV-2. J. Am. Soc. Nephrol. 2020, 31, 1683–1687, Erratum in J. Am. Soc. Nephrol. 2020, 31, 2494. [Google Scholar] [CrossRef]
  72. Su, H.; Yang, M.; Wan, C.; Yi, L.-X.; Tang, F.; Zhu, H.-Y.; Yi, F.; Yang, H.-C.; Fogo, A.B.; Nie, X. Renal histopathological analysis of 26 postmortem findings of patients with COVID-19 in China. Kidney Int. 2020, 98, 219–227. [Google Scholar] [CrossRef]
  73. Sur, S.; Khatun, M.; Steele, R.; Isbell, T.S.; Ray, R.; Ray, R.B. Exosomes from COVID-19 patients carry tenascin-C and fibrinogen-β in triggering inflammatory signals in cells of distant organ. Int. J. Mol. Sci. 2021, 22, 3184. [Google Scholar] [CrossRef]
  74. Krzanowska, K.; Batko, K.; Niezabitowska, K.; Woźnica, K.; Grodzicki, T.; Małecki, M.; Bociąga-Jasik, M.; Rajzer, M.; Sładek, K.; Wizner, B.; et al. Predicting acute kidney injury onset with a random forest algorithm using electronic medical records of COVID-19 patients: The CRACoV-AKI model. Pol. Arch. Intern. Med. 2024, 134, 16697. [Google Scholar] [CrossRef]
  75. Aljohani, M.A. Exosomes in Clinical Laboratory: From Biomarker Discovery to Diagnostic Implementation. Medicina 2025, 61, 1930. [Google Scholar] [CrossRef]
  76. Kalani, A.; Chaturvedi, S.; Chaturvedi, P. Wilms’ tumor 1 in urinary exosomes as a non-invasive biomarker for diabetic nephropathy. Clin. Chim. Acta 2026, 579, 120599. [Google Scholar] [CrossRef]
  77. Ko, W.-C.; Chen, C.-S.; Chang, Y.-P.; Wu, C.-S.; Yang, H.-C.; Chang, J.-F. CCN2/CTGF-Driven Myocardial Fibrosis and NT-proBNP Synergy as Predictors of Mortality in Maintenance Hemodialysis. Int. J. Mol. Sci. 2025, 26, 11350. [Google Scholar] [CrossRef]
  78. Yagin, F.H.; Inceoglu, F.; Colak, C.; Alkhalifa, A.K.; Alzakari, S.A.; Aghaei, M. Machine Learning-Integrated Explainable Artificial Intelligence Approach for Predicting Steroid Resistance in Pediatric Nephrotic Syndrome: A Metabolomic Biomarker Discovery Study. Pharmaceuticals 2025, 18, 1659. [Google Scholar] [CrossRef]
  79. Delrue, C.; Speeckaert, M.M. Transcriptomic Signatures in IgA Nephropathy: From Renal Tissue to Precision Risk Stratification. Int. J. Mol. Sci. 2025, 26, 10055. [Google Scholar] [CrossRef] [PubMed]
  80. Lee, P.-H.; Huang, S.M.; Tsai, Y.-C.; Wang, Y.-T.; Chew, F.Y. Biomarkers in Contrast-Induced Nephropathy: Advances in Early Detection, Risk Assessment, and Prevention Strategies. Int. J. Mol. Sci. 2025, 26, 2869. [Google Scholar] [CrossRef] [PubMed]
  81. González-Nicolás, M.Á.; González-Guerrero, C.; Goicoechea, M.; Boscá, L.; Valiño-Rivas, L.; Lázaro, A. Biomarkers in Contrast-Induced Acute Kidney Injury: Towards A New Perspective. Int. J. Mol. Sci. 2024, 25, 3438. [Google Scholar] [CrossRef]
  82. Nishida, S.; Ishima, T.; Iwami, D.; Nagai, R.; Aizawa, K. Trans-Omic Analysis Identifies the ‘PRMT1–STAT3–Integrin αVβ6 Axis’ as a Novel Therapeutic Target in Tacrolimus-Induced Chronic Nephrotoxicity. Int. J. Mol. Sci. 2025, 26, 10282. [Google Scholar] [CrossRef]
  83. Zhu, Y.; Yu, S.; Yang, D.; Yu, T.; Liu, Y.; Du, W. Integrated Multi-Omics Analysis Unveils Distinct Molecular Subtypes and a Robust Immune–Metabolic Prognostic Model in Clear Cell Renal Cell Carcinoma. Int. J. Mol. Sci. 2025, 26, 3125. [Google Scholar] [CrossRef] [PubMed]
  84. Feng, B.; Du, H.; Tong, H.H.Y.; Wang, X.; Li, K. Enhancing Local Functional Structure Features to Improve Drug–Target Interaction Prediction. Int. J. Mol. Sci. 2025, 26, 10194. [Google Scholar] [CrossRef]
  85. Manan, A.; Baek, E.; Ilyas, S.; Lee, D. Digital Alchemy: The Rise of Machine and Deep Learning in Small-Molecule Drug Discovery. Int. J. Mol. Sci. 2025, 26, 6807. [Google Scholar] [CrossRef]
  86. Messa, L.; Testa, C.; Carelli, S.; Rey, F.; Jacchetti, E.; Cereda, C.; Raimondi, M.T.; Ceri, S.; Pinoli, P. Non-Negative Matrix Tri-Factorization for Representation Learning in Multi-Omics Datasets with Applications to Drug Repurposing and Selection. Int. J. Mol. Sci. 2024, 25, 9576. [Google Scholar] [CrossRef]
  87. Park, H.; Miyano, S. DD-CC-II: Data Driven Cell–Cell Interaction Inference and Its Application to COVID-19. Int. J. Mol. Sci. 2025, 26, 10170. [Google Scholar] [CrossRef]
  88. Lepoittevin, M.; Remaury, Q.B.; Lévêque, N.; Thille, A.W.; Brunet, T.; Salaun, K.; Catroux, M.; Pellerin, L.; Hauet, T.; Thuillier, R. Advantages of Metabolomics-Based Multivariate Machine Learning to Predict Disease Severity: Example of COVID. Int. J. Mol. Sci. 2024, 25, 12199. [Google Scholar] [CrossRef]
  89. Kim, H.; Choi, E.; Shim, Y.; Kwon, J. PriorCCI: Interpretable Deep Learning Framework for Identifying Key Ligand–Receptor Interactions Between Specific Cell Types from Single-Cell Transcriptomes. Int. J. Mol. Sci. 2025, 26, 7110. [Google Scholar] [CrossRef] [PubMed]
  90. Jiang, Y.-H.; Jhang, J.-F.; Wang, J.-H.; Wu, Y.-H.; Kuo, H.-C. A Decision Tree Model Using Urine Inflammatory and Oxidative Stress Biomarkers for Predicting Lower Urinary Tract Dysfunction in Females. Int. J. Mol. Sci. 2024, 25, 12857. [Google Scholar] [CrossRef] [PubMed]
  91. Weightman, A.; Clayton, P.; Coghlan, S. Confronting the Ethical Issues with Artificial Intelligence Use in Nephrology. Kidney360 2025. Online ahead of print. [Google Scholar] [CrossRef]
  92. Llerena-Izquierdo, J.; Ayala-Carabajo, R. Ethics of the Use of Artificial Intelligence in Academia and Research: The Most Relevant Approaches, Challenges and Topics. Informatics 2025, 12, 111. [Google Scholar] [CrossRef]
Figure 1. Fully connected neural network with an input layer containing 6 neurons, two hidden layers containing 10 and 7 neurons, respectively, and an output neuron constituting the output layer. The neurons are represented by white circles with black edges. The red color shows the positive weight value, and the blue color shows the negative weight value of the connection between the appropriate neurons. The color intensity is proportional to weight. Each neuron in one layer has connections with every neuron in the next layer. The visualizations were made using online software from the following website: https://alexlenail.me/NN-SVG/index.html. Accessed on 5 January 2026.
Figure 1. Fully connected neural network with an input layer containing 6 neurons, two hidden layers containing 10 and 7 neurons, respectively, and an output neuron constituting the output layer. The neurons are represented by white circles with black edges. The red color shows the positive weight value, and the blue color shows the negative weight value of the connection between the appropriate neurons. The color intensity is proportional to weight. Each neuron in one layer has connections with every neuron in the next layer. The visualizations were made using online software from the following website: https://alexlenail.me/NN-SVG/index.html. Accessed on 5 January 2026.
Ijms 27 01285 g001
Figure 2. Visualization of a convolutional neural network on the example of the AlexNet architecture. The input image is processed in such a way that convolutional layers formulate the context and build features from the elements of the previous layer, and pooling layers combine these features, giving them a broader context. Ultimately, the set and arrangement of features allow for classifying the image, finding structures, segmenting or detecting deviations from the norm. The visualizations were made using online software from the following website: https://alexlenail.me/NN-SVG/index.html. Accessed on 5 January 2026.
Figure 2. Visualization of a convolutional neural network on the example of the AlexNet architecture. The input image is processed in such a way that convolutional layers formulate the context and build features from the elements of the previous layer, and pooling layers combine these features, giving them a broader context. Ultimately, the set and arrangement of features allow for classifying the image, finding structures, segmenting or detecting deviations from the norm. The visualizations were made using online software from the following website: https://alexlenail.me/NN-SVG/index.html. Accessed on 5 January 2026.
Ijms 27 01285 g002
Table 1. A compilation of various machine learning applications in proteomics and molecular sciences. ALB—albumin, AFM—afamin, ANXA7—annexin A7, APOD—Apolipoprotein D, C9—Complement component 9, SERPINA5—Protein C inhibitor, VPS4A—vacuolar protein sorting-associated protein 4A, CP—Ceruloplasmin, TF—transcription factor.
Table 1. A compilation of various machine learning applications in proteomics and molecular sciences. ALB—albumin, AFM—afamin, ANXA7—annexin A7, APOD—Apolipoprotein D, C9—Complement component 9, SERPINA5—Protein C inhibitor, VPS4A—vacuolar protein sorting-associated protein 4A, CP—Ceruloplasmin, TF—transcription factor.
AuthorsAI MethodInput VariablesTarget PointPerformance
Yuan et al. [10]LASSO logistic regression,
RFC,
SVM-RFE
Genes ARID4B, EOMES, KCNJ3,
LIF and STAT1
Kidney fibrosisAUC of 0.923
Kha et al. [20]RFC, XGBoost, ET, LGBM, MLP18 values derived from molecular computational methodsPossible drug–food constituent interactions (DFIs)Accuracy of 96.75% for XGBoost
Massy et al. [11]Clustering and
Regularized Cox Regression
Set of 90 urinary peptidesKidney failureAUC of 0.83
Reznichenko et al. [5]Self-Organizing Maps unsupervised ANN ML algorithmA set of upregulated and downregulated genes associated with faster progression of chronic kidney disease (CKD)CKD reclasification AUC of 0.825
McCallion et al. [31]SVMCKAP4, PTX3, IGFBP2, OPNSenescence in AKI and CKD vs. comorbiditiesAUC of 0.98 for CKAP4
Schork et al. [6]LASSO regressionAmong 112 peptides CKD273, HF2, and CAD238 were statistically significantClustering into groups at risk of developing diabetes and diabetes complicationsAUC of 0.868 with 95% CI 0.755–0.981
Deo et al. [12]Elastic Net RegressionSet of 16 proteinsSecondary cardiovascular eventsAUC of 0.77–0.80
Fan et al. [23]Logistic regressionSets of 2, 3 and 4 proteins: ALB + AFM, ANXA7 + APOD + C9,
SERPINA5 + VPS4A + CP + TF
DKD vs. uncomplicated diabetic patients and DKD3 vs. DKD4, progression to DKDAUC of 0.928,
AUC of 0.949,
AUC of 0.952, respectively
Kononikhin et al. [24]Logistic regression, k-NN, SVMSet of GPX3, PLMN, and A1AT or SHBGMild vs. severe glomerulopathiesAUC of 0.99
Glazyrin et al. [32]KNeighbors (kNN), logistic regression, support vector machine (SVM), and decision treeSet of biochemistry parameters, clustersVarious groups of chronic kidney disease of renal and postrenal or systemic etiology (prerenal)Accuracy of 96.3% for glomerulonephritris and 96.4% diabetic nephropaty
Table 2. General characteristics of machine learning methods applications, requirements, leading advantages and disadvantages.
Table 2. General characteristics of machine learning methods applications, requirements, leading advantages and disadvantages.
MethodApplicationRequirementsAdvantagesDisadvantages
Artificial intelligence/
Machine learning
Support
Vector
Machine
Classification and regression of tabular data
[27,30,34]
Complete dataSimplicityInsufficient in more complex tasks, possible low performance on large sets
Random
Forest
Classifier
Classification [34]
and regression of tabular data
Feature importation [50,53]
Complete data, partially insensitive to missing dataSimplicity, insensitivity to irrelevant variables, partial resistance to outliersMay be insufficient in more complex tasks
XGBoostClassification and regression of tabular data
[26,27,28,29,34]
Complete dataSimplicity
kNNData imputation [35]
regression,
classification [32,34]
Empty cells for imputationSimplicityPossible low performance on large sets with many variables.
Sensitive to missing and outlier data.
Except for data imputation, rather limited use
Artificial Neural NetworkMultilayer PerceptronClassification and regression of tabular data
[34,41,42]
Necessary data scaling, standardization, improved performanceCan learn non-linear relationships
Can learn up to date through additional data
Hyperparameter tuning required
Sensitive to scaling of input data
Deep learningConvolutional Neural NetworkClassification [38,43,48],
deviation detection [43],
segmentation [45,46,47,48]
Mainly imaging, histopathological, radiological data, etc.Complexity enabling deep image analysis.
Possibility to upload large amounts of data.
Application of a once trained network in various approaches
Complexity, risk of overfitting to too few input data due to the multitude of parameters inside the model.
Requirement of a sufficiently large number of class representatives
U-Net [45,46,47]
ResNet18 [38]
ResNet50 [38]
ResNet101
[38,48]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Stojanowski, J.; Gołębiowski, T.; Musiał, K. Artificial Intelligence in Nephrology—State of the Art on Theoretical Background, Molecular Applications, and Clinical Interpretation. Int. J. Mol. Sci. 2026, 27, 1285. https://doi.org/10.3390/ijms27031285

AMA Style

Stojanowski J, Gołębiowski T, Musiał K. Artificial Intelligence in Nephrology—State of the Art on Theoretical Background, Molecular Applications, and Clinical Interpretation. International Journal of Molecular Sciences. 2026; 27(3):1285. https://doi.org/10.3390/ijms27031285

Chicago/Turabian Style

Stojanowski, Jakub, Tomasz Gołębiowski, and Kinga Musiał. 2026. "Artificial Intelligence in Nephrology—State of the Art on Theoretical Background, Molecular Applications, and Clinical Interpretation" International Journal of Molecular Sciences 27, no. 3: 1285. https://doi.org/10.3390/ijms27031285

APA Style

Stojanowski, J., Gołębiowski, T., & Musiał, K. (2026). Artificial Intelligence in Nephrology—State of the Art on Theoretical Background, Molecular Applications, and Clinical Interpretation. International Journal of Molecular Sciences, 27(3), 1285. https://doi.org/10.3390/ijms27031285

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop