Next Article in Journal
Binary Particle Swarm Optimization with Manta Ray Foraging Learning Strategies for High-Dimensional Feature Selection
Previous Article in Journal
Energy-Efficient Fall-Detection System Using LoRa and Hybrid Algorithms
Previous Article in Special Issue
Harnessing Gut Microbiota for Biomimetic Innovations in Health and Biotechnology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Application of Metaheuristics for Optimizing Predictive Models in iHealth: A Case Study on Hypotension Prediction in Dialysis Patients

by
Felipe Cisternas-Caneo
1,
María Santamera-Lastras
2,3,
José Barrera-Garcia
1,4,
Broderick Crawford
1,*,
Ricardo Soto
1,
Cristóbal Brante-Aguilera
5,
Alberto Garcés-Jiménez
3,6,7,
Diego Rodriguez-Puyol
2,7,8 and
José Manuel Gómez-Pulido
3,6,7
1
Escuela de Ingeniería Informática, Pontificia Universidad Católica de Valparaíso, Avenida Brasil 2241, Valparaíso 2362807, Chile
2
Department of Medicine and Medical Specialties, Universidad de Alcalá, Alcalá de Henares, 28805 Madrid, Spain
3
Health Computing and Intelligent Systems Research Group (HCIS), Universidad de Alcalá, Alcalá de Henares, 28805 Madrid, Spain
4
Escuela de Negocios y Economía, Pontificia Universidad Católica de Valparaíso, Amunátegui 1838, Viña del Mar 2580129, Chile
5
Escuela de Ingeniería en Construcción y Transporte, Pontificia Universidad Católica de Valparaíso, Avenida Brasil 2147, Valparaíso 2362804, Chile
6
Department of Computer Science, Universidad de Alcalá, Alcalá de Henares, 28805 Madrid, Spain
7
Ramón y Cajal Institute for Health Research (IRYCIS), 28034 Madrid, Spain
8
Nephrology Department and Research Foundation, Hospital Universitario Principe de Asturias, 28805 Madrid, Spain
*
Author to whom correspondence should be addressed.
Biomimetics 2025, 10(5), 314; https://doi.org/10.3390/biomimetics10050314
Submission received: 24 March 2025 / Revised: 3 May 2025 / Accepted: 9 May 2025 / Published: 12 May 2025

Abstract

:
Intradialytic hypotension (IDH) is a critical complication in patients with chronic kidney disease undergoing dialysis, affecting both patient safety and treatment efficacy. This study examines the application of advanced machine learning techniques, combined with metaheuristic optimization methods, to improve predictive models for intradialytic hypotension (IDH) in hemodialysis patients. Given the critical nature of IDH, which can lead to significant complications during dialysis, the development of effective predictive tools is vital for improving patient safety and outcomes. Dialysis session data from 758 patients collected between January 2016 and October 2019 were analyzed. Particle Swarm Optimization, Grey Wolf Optimizer, Pendulum Search Algorithm, and Whale Optimization Algorithm were employed to reduce the feature space, removing approximately 45% of clinical and analytical variables while maintaining high recall for the minority class of patients experiencing hypotension. Among the evaluated models, the XGBoost classifier showed superior performance, achieving a macro F-score of 0.745 with a recall of 0.756 and a precision of 0.718. These results highlight the effectiveness of the combined approach for early identification of patients at risk for IDH, minimizing false negatives, and improving clinical decision-making in nephrology.

1. Introduction

Dialysis and transplantation have enabled many patients with advanced chronic kidney disease (CKD) to remain alive and have a good quality of life for many years. Although early renal engraftment is the desirable treatment for these patients, this is not always possible, partly because of the limited availability of organs and partly because of patient characteristics. The prevalence of CKD in its most advanced stages, and specifically those requiring RRT, has grown by almost 30% in Spain in the last decade. According to data from the O.N.T./S.E.N. Registry, the number of people on RRT hemodialysis, peritoneal dialysis, or transplantation has already reached 1363 per million population, totaling 64,600 people as of 2020, as reported in [1]. Consequently, in 2014, 26,533 patients in Spain needed dialysis techniques to stay alive, and of these, 23,519 patients required a technique known as hemodialysis (HD). If we compare these figures with the prevalence of other diseases, they are not very significant. However, the impact on individuals and families is brutal, and the economic cost is disproportionate, as 2.5% of all Spanish healthcare expenditure is devoted to treating patients with end-stage CKD [2].
Significant advances in hemodialysis techniques in recent decades have allowed us to improve the management of intradialytic hypotension (IDH) as a recurrent problem in patients receiving hemodialysis (HD) treatment. Nevertheless, this phenomenon is still too frequent (5–40% of sessions) [3,4] and is associated with high patient morbidity and mortality and multiple complications. There are no guidelines or systematic treatments to manage this complication, and different maneuvers are used to control it [5]. IDH has three essential components: a drop of more than 20 mmHg in systolic blood pressure (SBP) or more than 10 mmHg in mean arterial pressure (MAP), the presence of symptoms due to ischemia of different organs, and interventions by dialysis staff [6]. Identifying the occurrence of HDI during HD treatment is not a straightforward task due to the numerous variables or factors involved, including the type of dialyzer, dialyzer temperature, patient type, dialysis modality, and medical criteria, among others. HDI involves complex factors, including circulating volume, fluid removal rate, the rate of “filling” the intravascular space, the osmolar gradient between compartments, and impairment of compensatory mechanisms due to underlying medical conditions, diseases, or medications. In addition, some patients are at higher risk; studies suggest that women, the elderly with isolated systolic hypertension, diabetics, and those with documented autonomic neuropathy or heart disease are at higher risk of IDH.
Thus, this research hypothesizes that it is possible to detect an optimal combination of clinical and analytical parameters associated with the development of hypotension or heart failure. By measuring some combination of these parameters in a patient at the start of a dialysis session, the possible occurrence of hypotension during the session could be predicted, helping to reduce its incidence and the need for treatment by medical staff. Identifying the factors that most influence the occurrence of HDI and predicting its occurrence would provide ideal decision support for medical staff. Indeed, determining whether a patient has HDI during HD treatment is a complex task for traditional (statistical) models. However, machine learning (ML)-based models can discover and learn patterns in the data, and are therefore particularly well-suited to address this problem.
This research aims to extract and construct a subset of data consisting of several significantly relevant variables from an extensive database of hemodialysis sessions and blood tests. This subset will be capable of being processed by mass data processing tools to predict the occurrence of hypotension from patterns obtained from an optimal and reduced set of clinical and analytical parameters. Metaheuristic-based techniques will be applied to select an optimal combination of variables from the clinical and analytical blood records of the patient receiving HD treatment. The main contributions of this paper are as follows:
  • Use a metaheuristic approach to solve the feature selection problem in the context of hemodialysis and apply two objective functions for this problem.
  • Use four metaheuristic algorithms: Particle Swarm Optimization, Grey Wolf Optimizer, Pendulum Search Algorithm, and Whale Optimization Algorithm. In addition, use three classification algorithms, K-Nearest Neighbors, Random Forest, and XGBoost, to evaluate the efficiency of the metaheuristic algorithms.
  • Perform in-depth analysis using various performance metrics such as recall, F-score, precision, and number of selected features, graphs, and statistical tests.
As a result of the study and analysis, the XGBoost model demonstrated better performance in the analyzed metrics compared to the other ML models evaluated. It achieved values of around 77% for the recall in the minority class with around 45% of features eliminated in hypotensive patients, minimizing false positives. Consequently, these results suggest that the combination of the use of the metaheuristic PSO, WOA, PSA, GWO, and the ML model XGBoost can predict with high reliability and robustness the occurrence of IDH in patients and, in addition, offers medical staff facilities to support decision-making based on the information inferred by the model itself.
The remainder of this document is structured as follows: Section 2 provides the background and a review of related work. Section 3 introduces the dataset used in this research. In Section 4, we present the metaheuristics and their binarization process for the feature selection task. Section 5 describes the proposed methodology for predicting intradialytic hypotension. Section 6 details the experimental results and evaluates the performance of the different machine learning models. Finally, Section 7 discusses the conclusions drawn from the research and suggests directions for future work.

2. Background and Related Works

2.1. Soft Computing Techniques in Healthcare

Smart Health, or “Intelligent Health”, greatly interests the scientific and industrial community. Smart Health arises from two already known fronts, the Smart City and e-Health (also known as e-Health), as they aim to improve the health standards of citizens using technological paradigms [7]. Smart Health uses advanced technologies and digital solutions to improve medical care and health in general. The Smart Health dimension generates a considerable amount of information [8] coming from the constant monitoring of patients, who are essential to manage, process and analyze through the implementation of rigorous methods that will allow healthcare personnel to make a more informed decision regarding the care and/or treatment of the patient’s health [9]. For example, digitized clinical history allows medical personnel to keep records of patients’ diseases or pathologies to know the patient’s condition and/or evolution, thus providing effective clinical treatment and care for the patient. Likewise, the application of novel methods allows terminally ill patients, patients with genetic disorders, and/or renal patients to be treated and cared for at home, as the caregiver would make more informed decisions according to the analysis of the patients’ parameters while receiving home care. For example, this is applied to treating patients with in-home palliative care, home MRI, and home hemodialysis, among others [10,11]. Consequently, patients would experience healthier aging or more bearable hemodialysis since family members are present to support them. Considering the above, the use, processing, and analysis of data will contribute significantly to the creation of advanced services of great utility for the various actors involved in the dimensions of CS mentioned above [12]. In this sense, it is necessary to create advanced services through Soft Computing techniques that support intelligent decision-making. These offer a wide range of services with high added value for a better quality of life and welfare of society, as well as improved efficiency and environmental sustainability. On the other hand, the term Soft Computing was proposed by Lofti A. Zadeh, and he points out that Soft Computing is a set of methods or techniques based on imprecision, approximation, and uncertainty to solve complex real-world problems [13,14,15]. Therefore, these techniques can be learned from experience and applied to any problem domain. Because Soft Computing techniques make use of approximate calculations, they offer robust solutions to complex problems. Also, they have a lower computational cost than exact methods [8,15,16]. A characteristic of Soft Computing techniques is that they can be used in combination, offering the possibility of obtaining more reliable results thanks to the characteristics of each of the methods. Soft Computing techniques encompass various advanced algorithms, including fuzzy computing, evolutionary computing, neural networks, machine learning, and deep learning. Figure 1 shows a taxonomy of some of these techniques.
Each technique shown in Figure 1 is used to solve various real-world problems. For example, methods based on fuzzy computing (FC) are employed to solve vague or imprecise problems, i.e., it allows for the transformation of natural language adjectives to quantify the degree of membership of a given dataset [13,17]. Moreover, FC resembles the human ability to make decisions. This is observed in works [18,19] where they make use of fuzzy logic for patient monitoring. Similarly, Soft Computing employs evolutionary computation (EC), inspired by natural evolution. EC is applied to solving optimization problems, and a probabilistic approach is used to solve these problems [17]. Similarly, as in FC techniques, several studies have developed research in the dimensions of Smart Health by applying EC techniques [20,21,22,23,24,25]. On the other hand, machine learning (ML) techniques are widely used due to their potential and versatility in various computer science sectors. These algorithms learn directly from the information they provide, enabling them to extract valuable insights and knowledge from the data itself [26,27,28]. ML algorithms are generally used for classification and/or prediction tasks. As an example, in the study by [29], they used Random Forest and IoT in a healthcare monitoring system. Another Soft Computing technique that recently received much attention is models developed using artificial neural networks (ANNs). The ANN architecture typically consists of an input layer, one or more hidden layers, and an output layer. Examples of ANN architectures are MLP or multilayer perceptron, RBF or radial basis function, SOM networks, or self-organizing maps [17,30,31,32]. Within this framework, the robustness and reliability that ANNs provide today have led to an evolution in ANN architectures, known as deep learning (DL). DL architectures employ more complex processes that enable the capture of intrinsic and abstract relationships in data with greater depth. Some tasks that these DL architectures or models can perform are speech or text recognition, object detection, natural language processing, energy consumption forecasting, and weather forecasting, among others [33]. As a complement, in the Smart Health dimension, researchers have developed DL-based intelligent systems to predict epileptic seizures by analyzing electrocardiogram signals, as well as to predict heart abnormalities [34,35,36,37,38].

2.2. Optimization Techniques for IDH Prediction

In healthcare, innovative and efficient solutions for complex problems are crucial [39,40]. Metaheuristics, including genetic algorithms, Particle Swarm Optimization, and ant colony optimization, have found significant applications in this sector. These techniques optimize medical treatments, improve hospital scheduling, and facilitate the analysis of large medical datasets [41,42]. The ability of metaheuristics to find near-optimal solutions in a reasonable amount of time makes them ideal for addressing challenges in healthcare, where quick and accurate decisions can significantly impact patient outcomes.
A specific application of advanced computational techniques in healthcare is the prediction of complications during hemodialysis, such as intradialytic hypotension (IDH). Recent research has developed predictive models that utilize machine learning algorithms and optimization techniques to forecast IDH episodes accurately. These models enable healthcare professionals to anticipate and prevent IDH episodes, enhancing patient safety and well-being during hemodialysis treatments. By combining the power of machine learning with optimization techniques, significant progress has been made in managing and preventing health complications. Various studies have proposed different methodologies to predict and manage IDH episodes, aiming to improve patient outcomes and enhance clinical decision-making.
In [43], the authors present a hybrid model named BSCWJAYA_KELM that integrates serum biomarkers of nutrition with a novel optimization algorithm and machine learning model. By focusing on indicators such as serum uric acid, dialysis vintage, age, diastolic pressure, and albumin, this study aims to improve the accuracy of IDH predictions, ultimately enhancing patient management. In [44], the authors describe developing and validating two AI-based risk models to predict symptomatic IDH during hemodialysis sessions. Their work highlights the importance of personalized approaches in preventing IDH by identifying high-risk patients before the start of dialysis, thereby enhancing treatment efficacy and patient safety. In [45], the authors address the issue of IDH by proposing a predictive model that utilizes features from the photoplethysmography (PPG) signal. Their study emphasizes the potential of PPG in forecasting pre-IDH and IDH episodes, thereby aiding in the prevention of acute hypotension during hemodialysis sessions. In [46], the authors introduce a novel optimization algorithm, the Covariance Matrix-Driven Whale Optimizer with Orthogonal Structure-Assisted Extreme Learning Machine (bCOWOA-KELM), to enhance the prediction accuracy and reliability of IDH. This approach addresses the limitations of existing methods, providing a more solid tool for managing hemodialysis complications.
In another study [46], the authors explore ensemble machine learning techniques to predict various hemodialysis complications, including IDH. The objective is to develop high-performance models for early prediction, thus improving patient care and reducing the incidence of early dialysis session terminations. In [47], the authors discuss the development of an early alert system for predicting IDH before the initiation of hemodialysis. By employing artificial intelligence (AI) and machine learning techniques, their study aims to identify high-risk patients early, allowing for timely interventions and better management of IDH. In [48], the authors propose an optimized machine learning framework for predicting IDH, with a focus on indices related to chronic kidney disease–mineral and bone disorders (CKD-MBDs). Their model aims to predict IDH episodes effectively, thereby supporting clinical decision-making and reducing the risk of severe complications during hemodialysis.
These studies collectively contribute to IDH prediction and management knowledge, utilizing advanced machine learning algorithms and AI techniques to improve patient outcomes during hemodialysis.
In summary, metaheuristics, machine learning, and optimization techniques are pivotal in advancing healthcare solutions [49,50]. Their applications enhance operational efficiencies and have a direct impact on patient care and treatment outcomes.

3. Hypotension Dataset

3.1. Digitized Clinical Database of Hemodialysis Patients

Our research is based on two extensive databases containing the medical information of 758 patients undergoing hemodialysis sessions. These data were collected at the Hospital Príncipe de Asturias, Madrid, Spain, over almost four years, from 1 January 2016 to 30 October 2019. This dataset contains records relating to clinical variables of patients who received HD treatment between January 2016 and October 2019. The first database collected each patient’s clinical data for 98,015 hemodialysis sessions of about five hours’ duration, which are labeled as “Hour 0”, “Hour 1”, “Hour 2”, “Hour 3”, “Hour 4”, and “Hour 5”. This first data source contains the following patient and session identification data: patient identifier, gender, and age; whether the patient is hypertensive or diabetic; and the date of the session. The remaining 29 variables contain the sequence of the different clinical parameters measured at the corresponding time of the session. The number of variables associated with a dialysis session is 180. Additionally, a second database containing data from the same patients during the same period was available, with the variables and corresponding values of the information extracted from the hospital laboratory’s blood analyses. This second database contains the values of 141 variables. The total set of variables related to patient, dialysis session, and blood test data amounts to 221. The hospital’s ethics committee approved the collection of all these records, and the database was completely anonymized. From the two databases above, a new database was constructed incorporating the values of the blood test variables and linking them to the nearest hemodialysis session, either on the same date or immediately before. Figure 2 shows the unification process of both datasets. It should be noted that the different analytical variables were not obtained in the same tests and with the same periodicity.

3.2. Determination of IDH in the HD Session

The potential hypotension that may occur during a hemodialysis session is determined by medical judgment, calculated from the systolic blood pressure measurements taken at each of the five hours of the session. If any of the pressures measured at “Hour 2”, “Hour 3”, and “Hour 4” is lower by at least 20 mL than any of the pressures measured at “Hour 0” and “Hour 1”, then hypotension is present, quantified by the Hypotension Measurement (HYM) as the difference in systolic blood pressures; otherwise, there is no hypotension. Once the possible hypotension values for all sessions have been calculated, the HYM variable and the binary variable for the presence of IDH are incorporated into the database. Subsequently, a new database is generated which only takes into account the values of the clinical variables at the beginning of the HD session, the so-called “Hour 0”, given that the final objective of this work is to develop a predictive system capable of accurately anticipating the possible future appearance of hypotension during the dialysis session, as indicated by [51,52]. Figure 3 and Figure 4 show the hypotension warning process prior to Hour 0 and the mockup of the application visible to medical staff respectively.

3.3. Data Processing and Variables Considered in the Clinical Study

After obtaining the integrated database, the medical staff, utilizing their experience and clinical knowledge, identified a set of 80 variables (26 related to the patient and the dialysis session, and 54 to the analytical tests) to be considered for developing the predictive model. Subsequently, as is common in data science projects, pre-processing the data is essential before proceeding to the modeling phase. In this case, data cleaning tasks were performed after integrating and merging data from different sources to identify and correct potential errors. An analysis of the variables was conducted to identify potential quality issues, and cleaning tasks were performed to ensure a consistent dataset by correcting incorrect, atypical, or missing data ([53,54,55]). This analysis revealed that many dialysis sessions or analytical tests lacked quality data. As a result, the medical team retained the most significant variables and excluded patients and dialysis sessions with insufficient quality data. The final dataset comprised 68,574 samples, divided into two classes: the majority class (non-hypotensive during dialysis sessions) with 48,764 cases (71.11% of the dataset) and the minority class (hypotensive during dialysis sessions) with 19,810 cases (28.89%). These samples comprise 328 patients and 68,574 dialysis sessions, ensuring that both the cohort size and the number of sessions are representative of the population under study. After pre-processing, the dataset was expanded to 87 features, derived from the original 71 features through the one-hot encoding of 3 variables. Table 1 provides a summary regarding the IDH dataset, Instances refers to the total number of instances in the dataset. Labels categorizes the dataset into two classes NH, representing dialysis sessions where no hypotension occurred, and H, representing sessions where hypotension was observed. Features includes two stages of feature processing: Pre. for the number of features before pre-processing, and Post for the number of features after pre-processing.
Table 2 presents the classification results of three machine learning classifiers (XGBoost, Random Forest (RF), and K-Nearest Neighbors (KNN)) applied to the dataset without any optimization or feature selection processes. These results reflect the raw performance of the predictive models, with no techniques employed to reduce the number of features or address the class imbalance in the data. The table includes several key performance metrics: macro-averaged F-score (fs_macro), recall (r_macro), and precision (p_macro) for all classes, as well as these metrics for the minority class (fs_min, r_min, p_min) and the majority class (fs_may, r_may, p_may).
The macro-averaged metrics provide an overall evaluation across both classes. In contrast, the minority and majority class-specific metrics allow for a more detailed analysis of the classifier’s performance in handling the imbalanced dataset. As seen, XGBoost generally achieves the highest F-scores, recall, and precision, particularly in the majority class. In contrast, the performance of the minority class is notably lower across all classifiers, reflecting the challenge posed by class imbalance.

4. Optimization of Relevant Feature Selection for the Predictive Model

This section develops a method demonstrating the great capability and versatility of models based on Soft Computing techniques to adapt to the requirements of highly reliable clinical predictive systems. This allows for providing highly relevant and valuable information for medical decision-makers.
The feature selection (FS) problem is a multi-objective combinatorial optimization issue that stems from the need to eliminate irrelevant and redundant information from datasets used in machine learning training, as such information is detrimental to the learning tasks of prediction or classification algorithms.
In its mathematical definition, FS assumes a dataset O such as the original dataset, which contains a o number of f features, such that O = { f 1 , f 2 , f 3 , , f o } . The objective of the problem is to select the best subset of features B = { f 1 , f 2 , f 3 , , f b } with b < o so that the features belonging to the selected subset are the most representative of the set of original data.
Internally, FS has a binary domain; the solutions are represented by ones and zeros, where a one indicates that the feature is included in the data subset, while a zero indicates the opposite. Figure 5 shows the representation of the solutions for the feature selection problem. In this example, it is shown that the original dataset is composed of six features. Therefore, the solution is [ 1 , 1 , 1 , 1 , 1 , 1 ] . After applying the metaheuristics, we will obtain the best subset of features composed of the features f 1 , f 3 , and f 4 . Therefore, the solution is [ 1 , 0 , 1 , 1 , 0 , 0 ] .
The objective function to optimize within the feature selection problem is represented in various ways, with the weighted multi-objective function being the most commonly used, as noted by [56]. In this way, the objective function Z is calculated as follows:
m i n Z = α · m e t r i c + β · | R | | N |
where m e t r i c corresponds to a performance metric obtained from the machine learning technique, | R | is the number of selected features, and | N | is the total number of features, with α [ 0 , 1 ] and β = ( 1 α ) parameters that regulate the importance of the quality of the results and the size of the subset, respectively.

4.1. Two−Step Techniques

A review of the literature reveals a great diversity of implemented metaheuristics. The work carried out in [57] summarizes around 500 metaheuristics, highlighting that most of them are continuous population metaheuristics. On the other hand, there is significant interest in using continuous metaheuristics to solve binary combinatorial problems, particularly in relation to the feature selection problem, as highlighted by [56,58]. Figure 6 provides an overview of the two-step technique.

4.1.1. Transfer Function

A transfer function aims to transfer the values assigned to the decision variables from the R domain to the [0,1] domain. The first transfer function proposed in the literature within the context of metaheuristics was in 1997 with the presentation of the first binary metaheuristic, Binary Particle Swarm Optimization (BPSO), as noted by [59].
As research into binarization methods has deepened, [58] proposed new transfer functions in the literature, including the S-shaped and V-shaped transfer functions introduced by [60].
Table 3 and Figure 7 show the S-shaped transfer functions and V-shaped transfer functions found in the literature. The notation d i j observed in Table 3 corresponds to the continuous value of the j t h dimension of the i t h individual resulting after the perturbation performed by the continuous metaheuristic.

4.1.2. Binarization Rule

The second step is discretizing the value transferred from step 1 by applying a binarization rule. Various rules described in the scientific literature by [61,62] can be utilized for this binarization process. Table 4 shows the five binarization rules found in [61].
The notations X i j and X B e s t j observed in Table 4 correspond to the j-th dimension binary value of the i-th current individual and the j-th dimension binary value of the best solution, respectively.
Algorithm 1 shows the general process of the two-step technique, where the process is performed for each solution for each dimension.
Algorithm 1 Two-step technique.
Input: Continuous population X = { X 1 , X 2 , , X p o p }
Output: Binary population X = { X 1 , X 2 , , X p o p }
1:
for   i = 1   to   p o p   do                        ▹ Binarization Process
2:
     for  j = 1   to   d i m   do
3:
           Get T   ( d i j ) by applying transfer function
4:
           Get X n e w j by applying Binarization Rule
5:
      end for
6:
end for

4.2. Binary Particle Swarm Optimization

Particle Swarm Optimization is considered one of the first population metaheuristics presented in the literature by [63]. This metaheuristic simulates the collective behavior of swarms such as birds or fish. This algorithm is one of the most studied and referenced in the field of metaheuristics, demonstrating its versatility in different optimization problems, as highlighted by [57,58]. The following equation gives the equation of motion:
X i ( t + 1 ) = X i ( t ) + v i ( t + 1 )
where X i ( t ) is the current position of particle i, and X i ( t + 1 ) is the new position of particle i after updating the velocity v i ( t + 1 ) . The velocity v i ( t + 1 ) is calculated as follows:
v i ( t + 1 ) = w · v i ( t ) + c 1 · r 1 · ( p B e s t X i ( t ) ) + c 2 · r 2 ( g B e s t X i ( t ) )
where c 1 and c 2 are the acceleration coefficients that determine the influence of individual and collective knowledge on the motion of the particle, r 1 , and r 2 , are random numbers between [ 0 , 1 ] , p B e s t is the personal best position of particle i, g B e s t is the global best position found by the entire swarm up to that moment, X i ( t ) is the current position of particle i, v i ( t ) is the current velocity of particle i, and w is the inertia factor that controls the influence of the previous velocity on the new velocity.
The inertia factor (w) plays an important role in the metaheuristic since it allows for the balancing of exploration and exploitation. This inertia factor is calculated as follows:
w = w m a x w m a x w m i n m a x I t e r · t
where w m a x corresponds to the maximum inertia factor, w m i n corresponds to the minimum inertia factor, m a x I t e r corresponds to the maximum number of iterations, and t corresponds to the current iteration. This equation allows for a linear decrease from the maximum inertia factor ( w m a x ) to the minimum inertia factor ( w m i n ).
This metaheuristic was designed to solve continuous optimization problems. Thus, the creators of PSO,  [59], proposed the binary version of PSO, and Algorithm 2 shows its behavior.
Algorithm 2 Binary Particle Swarm Optimization.
 Input: The population X = { X 1 , X 2 , , X i }
 Output: The updated population X = { X 1 , X 2 , , X i } and X α
 1:
Initialize binary random population X
 2:
Evaluate the objective function of each individual and save the p b e s t of each solution
 3:
Identify the g B e s t in the population
 4:
Define c 1 , c 2 , w m a x , and w m i n
 5:
for   i t e r a t i o n   ( t )   do
 6:
      Update w using Equation (4)
 7:
      for  s o l u t i o n ( i )  do
 8:
            Update v i using Equation (3)
 9:
            Update X i using Equation (2)
10:
      end for
11:
      Binarization of population X using Algorithm 1
12:
      Evaluate the objective function of each individual and save the p b e s t of each solution
13:
      Update g B e s t
14:
end for
15:
Return ( g B e s t )

4.3. Binary Grey Wolf Optimizer

The Grey Wolf Optimizer was proposed by [64] and simulates the social hierarchy and hunting behavior of grey wolves in the wild. The wolf hierarchy is composed of the alpha wolf ( α ), who is the pack leader and is represented by the best solution in the population; the beta wolf ( β ), who can assume leadership in the absence of alpha and is represented by the second best solution in the population; the delta wolf ( δ ), who helps the alpha and beta wolves control the pack and is represented by the third best solution in the population; and the omega wolves ( ω ), who are the lowest-ranking wolves and are represented by the rest of the population. Stalking and surrounding behavior is modeled as follows:
X ( t + 1 ) = X 1 + X 2 + X 3 3
where X 1 , X 2 , and X 3 are the updated positions of the wolves based on the positions of the alpha wolf ( X α ), beta wolf ( X β ), and delta wolf ( X δ )
X 1 = X α ( A 1 · D α ) X 2 = X β ( A 2 · D β ) X 3 = X δ ( A 3 · D δ )
where D α , D β , and D δ corresponds to the distance between the position of the wolf X that we are updating and the wolves X α , X β , and X δ .
D α = | C 1 · X α X | D β = | C 2 · X β X | D δ = | C 3 · X δ X |
A and C are coefficients that play different roles in metaheuristic behavior. On the one hand, C interferes with the direction and intensity of the position adjustment relative to the best solutions, while A plays a vital role in the exploration–exploitation balance.
When | A | > 1 , the wolf moves away from the prey, allowing the exploration of new areas of the search space. On the other hand, when | A | < 1 , the wolf approaches the prey, promoting exploitation. The coefficient a decreases linearly from 2 to 0 during the optimization, meaning that the value of A also decreases, gradually shifting the focus of the algorithm from exploration to exploitation.
A = 2 a · r 1 a C = 2 · r 2 a = 2 2 · t m a x I t e r
This metaheuristic was also designed to solve continuous problems. Therefore, we must incorporate the two-step technique to solve binary combinatorial problems. Such an adaptation was proposed in [65], and Algorithm 3 shows its behavior.
Algorithm 3 Binary Grey Wolf Optimizer.
 Input: The population X = { X 1 , X 2 , , X i }
 Output: The updated population X = { X 1 , X 2 , , X i } and X α
 1:
Initialize binary random population X
 2:
Evaluate the objective function of each individual
 3:
Identify the X α , X β , and X δ in the population
 4:
for   i t e r a t i o n   ( t )   do
 5:
      for  s o l u t i o n ( i )  do
 6:
            Update a, A, and C using Equation (8)
 7:
            Update X 1 , X 2 , and X 3 using Equation (6)
 8:
            Update the position of X i t using Equation (5)
 9:
      end for
10:
    Binarization of population X using Algorithm 1
11:
    Evaluate the objective function of each individual
12:
    Update X α , X β , and X δ
13:
end for
14:
Return ( X α )

4.4. Binary Pendulum Search Algorithm

The Pendulum Search Algorithm is a new population-based metaheuristic created in 2022 by Nor Azlina Ab. It was inspired by [66] and the harmonic motion of the simple pendulum. The search agents are initialized randomly, and their position is updated using Equation (9).
X i , j t = X i , j t + p e n d i , j t · ( B e s t j X i , j t )
where X i , j t is the position of the i-th solution in the j-th dimension in t-th iteration, B e s t j it is the position of the best solution in the j-th dimension in the t-th iteration, and p e n d i , j t is a parameter which is calculated as follows:
p e n d i , j t = 2 · e ( t / t m a x ) · c o s ( 2 · π · r a n d )
where t is the current iteration, t m a x is the maximum number of iterations and r a n d is a uniform random number between [0,1].
This metaheuristic was also designed to solve continuous problems. Therefore, we must incorporate the two-step technique to solve binary combinatorial problems. Such an adaptation was proposed in [67], and Algorithm 4 shows its behavior.
Algorithm 4 Binary Pendulum Search Algorithm.
 Input: The population X = { X 1 , X 2 , , X i }
 Output: The updated population X = { X 1 , X 2 , , X i } and B e s t
 1:
Initialize binary random population X
 2:
Evaluate the objective function of each individual
 3:
Identify the best individual in the population ( B e s t )
 4:
for   i t e r a t i o n   ( t )   do
 5:
      for  s o l u t i o n ( i )  do
 6:
            for  d i m e n s i o n ( j )  do
 7:
                  Update p e n d i . j t by Equation (10)
 8:
                  Update the position of X i , j t using Equation (9)
 9:
            end for
10:
      end for
11:
      Binarization of population X using Algorithm 1
12:
      Evaluate the objective function of each individual
13:
      Update B e s t if there is a better solution
14:
end for
15:
return the updated population X where B e s t is the best result

4.5. Binary Whale Optimization Algorithm

The Whale Optimization Algorithm (WOA) is inspired by the hunting behavior of humpback whales, in particular how they search (Section 4.5.1), encircle (Section 4.5.2) and attack (Section 4.5.3) their prey with a technique known as “bubble nets”. This algorithm was devised by [68].

4.5.1. Searching for Prey

Searching for prey is related to the exploration phase of the search space. To ensure a good exploration, the value of A is used. When | A | > 1 , whales move away from a known solution, allowing a global search for space exploration. This behavior is modeled as follows:
X i t + 1 = X r a n d t A · D D = | C · X r a n d t X i t |
where A and C are coefficient vectors, t denotes the current iteration, and X r a n d t is a random whale in the population. The coefficient vectors A and C can be computed according to Equations (12) and (13):
A = 2 a · r a
C = 2 · r
where a decreases linearly from 2 to 0 over the course of the iterations (both in the exploration and exploitation phases) and r corresponds to a random vector of values between [ 0 , 1 ] .

4.5.2. Encircling the Prey

The first exploitation operator is related to the encircling prey. This occurs when a whale approaches the best solution found in the population so far. It is modeled using the following equations:
X i t + 1 = X * A · D D = | C · X * X i t |
where X * is the best whale obtained so far and X i t is the i-th whale in the t-th iteration. The vectors A and C are calculated as shown in Equations (12) and (13).

4.5.3. Spiral Movement

The second exploitation operator is related to spiral motion. This occurs when whales move towards the best known solution in a spiral-shaped path. The following equation models this:
X i t + 1 = D · e b l · cos ( 2 π l ) + X * D = | X * X i t |
where D is the distance of the i-th whale from the prey (the best solution obtained so far), and b is a constant to define the shape of the logarithmic spiral. l is a random number between [ 1 , 1 ] .
In nature, humpback whales perform the process of encircling the prey and spiral movement simultaneously. To model this simultaneous behavior, there is a 50% chance of choosing between the prey encircling mechanism and the spiral model. The mathematical model is as follows:
X i t + 1 = X * A · D If p < 0.5 D · e b l · cos ( 2 π l ) + X * If p 0.5
This metaheuristic was also designed to solve continuous problems. Therefore, we must incorporate the two-step technique to solve binary combinatorial problems. This adaptation was proposed in [65], and Algorithm 5 shows its behavior.
Algorithm 5 Binary Whale Optimization Algorithm.
 Input: The population X = { X 1 , X 2 , , X i }
 Output: The updated population X = { X 1 , X 2 , , X i } and X *
 1:
Initialize binary random population X
 2:
Evaluate the objective function of each individual
 3:
Identify the best individual in the population ( X * )
 4:
for   i t e r a t i o n   ( t )   do
 5:
      for  s o l u t i o n ( i )  do
 6:
            Update a , A , C , l and p
 7:
            if  p < 0.5  then
 8:
                 if  | A | < 1  then
 9:
                       Apply Equation (14)
10:
               else
11:
                     Apply Equation (11)
12:
               end if
13:
            else
14:
               Apply Equation (15)
15:
            end if
16:
      end for
17:
      Binarization of population X using Algorithm 1
18:
      Evaluate the objective function of each individual
19:
      Update X * if there is a better solution
20:
end for
21:
Return ( X * )

5. Enhanced Prediction of IDH Through ML and Biomarker Analysis

5.1. Addressing the Imbalance in the Dataset

As mentioned in Section 3 and visible in the preliminary results in Table 2, data imbalance significantly harms the performance of the machine learning model for the minority class.
To address the dataset imbalance, several machine learning techniques were tested. These techniques were categorized into three main approaches: undersampling, oversampling, and hybrid. Undersampling techniques aim to remove samples from the majority class to bring them closer to the number of minority samples. Oversampling techniques aim to create synthetic samples from the minority class to balance it with the majority class. Finally, hybrid techniques mix both oversampling and undersampling techniques. Currently, Python 3.13.2 is a very popular programming language in machine learning due to its access to multiple libraries related to this field. Data imbalance techniques are no exception, and the Imbalanced-learn library, developed by [69], is the tool.
We have performed preliminary experiments with different algorithms for each type of swing. For undersampling, two methods were applied: (i) Random Undersampler, which reduces the majority class by randomly selecting a subset of its instances, and (ii) NearMiss, which selects instances from the majority class that are closest to the minority class based on distance measures. For oversampling, two techniques were used: (i) Random Oversampler, which increases the size of the minority class by randomly duplicating instances, and (ii) SMOTE (Synthetic Minority Oversampling Technique), which generates synthetic instances for the minority class. Additionally, a hybrid technique, SMOTEENN, was tested, which combines SMOTE for oversampling and Edited Nearest Neighbors (ENN) for undersampling. The results indicate that the best data balancing technique for our dataset is the Random Undersampler.

5.2. Construction of Objective Function

Our proposal incorporates metaheuristics for feature selection to improve the performance of classification models that seek to predict intradialytic hypotension during dialysis treatment.
As mentioned in Section 4, metaheuristics require an objective function to perform the optimization process. Although Equation (1) defines the objective function, obtaining the metrics incorporates several steps. First, we need to work with the data to maintain data balance, perform cross-validation to avoid data bias, and apply a classifier to obtain the performance metrics. Given this, Algorithm 6 shows how we calculate the objective function in our proposal.
Algorithm 6 Objective function.
 Input: Selected features and dataset
 Output: Objective function and performance metrics
 1:
Filter the original dataset with the selected features
 2:
Apply KFold technique for cross-validation
 3:
for each fold do
 4:
      Apply balancing technique for training fold
 5:
      Normalize training fold and testing fold using MinMax technique
 6:
      Train the model with the classification algorithm
 7:
      Predict with the trained model from the testing fold
 8:
      Compare predicted observations with test fold
 9:
      Save performance metrics from the fold
10:
end for
11:
Calculate the average of the metrics across the KFold
12:
Apply Equation (1)
13:
Return average metrics and evaluate the objective function

5.3. Selection of Classifiers

In our proposal, we have considered three different classification algorithms for line 6 of Algorithm 6. The classifiers used were K-Nearest Neighbors (KNN), Random Forest (RF), and XGBoost (XG). KNN was chosen because of its wide use in the literature, as mentioned by [70], and because it is a distance-based classifier, while RF and XGBoost were chosen for their frequent use as decision tree- and boosting-type classifiers, respectively, as described by [56].

5.4. Metaheuristics for Feature Selection

The metaheuristics used in this study include Particle Swarm Optimization (PSO), Grey Wolf Optimizer (GWO), Pendulum Search Algorithm (PSA), and Whale Optimization Algorithm (WOA) as described in Section 4. As discussed in the articles by [57,66], the algorithms were selected for the following reasons: PSO for its strong performance in various optimization tasks, GWO for its growing popularity in recent years, WOA for its relevance in the literature, and PSA for being a parameter-free metaheuristic, simplifying the configuration process. In addition, we have used different metaheuristics, and not just one, based on the No Free Lunch Theorem ([71,72,73]), which tells us that there is no optimization algorithm capable of solving all existing optimization problems.
As mentioned in Section 4, the metaheuristics we have used must be binarized to solve binary combinatorial problems. In this case, we have used the two-step technique presented in Algorithm 1, where we have used the S4 function as the transfer function and the standard rule as the binarization rule. This selection was based on previous work in [67]. Thus, our complete proposal is summarized by Figure 8.

6. Results

This section presents the results of the experiments that evaluate the performance of the proposed approach. We describe the experimental setup, including the datasets, evaluation metrics, and metaheuristic configurations. Additionally, we detail the hardware used and the number of experiments performed. Finally, we compare and discuss the results obtained under different configurations.

6.1. Experiment Configuration

6.1.1. Sampling Parameters

Table 5 presents the classification results of three different classifiers (KNN, Random Forest, and XGBoost) using the Random Undersampler technique with varying sampling parameters. The classifiers are evaluated based on several performance metrics: macro F-score (f1_m), recall (r_m), and precision (p_m), as well as F-scores, recall, and precision for both the minority class (f1_min, r_min, p_min) and the majority class (f1_may, r_may, p_may). Additionally, the computation time for each configuration is reported.
From the results described in Table 5, it can be noted that across all parameters, the effect of the sampling parameter is that as the parameter decreases from 1.0 to 0.6, all classifiers show a slight increase in the macro F-scores. For instance, the f1_m of XGBoost increases from 0.725 at a parameter value of 1.0 to 0.745 at a parameter value of 0.6, suggesting an improvement in overall performance. However, this improvement comes at the expense of the minority class recall (r_min), which is critical for correctly predicting hypotensive episodes and reducing false positives. For XGBoost, r_min decreases from 0.766 at a parameter of 1.0 to 0.64 at a parameter of 0.6, indicating a significant reduction in the model’s ability to detect minority class instances. Similarly, Random Forest shows a considerable drop in r_min from 0.738 at 1.0 to 0.593 at 0.6. Therefore, although lower parameter values improve macro metrics, they negatively affect the minority class performance, going against the primary goal of accurately predicting hypotensive episodes.
Based on these findings, a classifier with a parameter value of 1.0 is recommended. This classifier provides the highest recall for the minority class, ensuring the best performance in minimizing false positive errors and correctly predicting hypotensive episodes.

6.1.2. Classifiers Parameters

In this proposal, we have used three classifiers: KNN, Random Forest, and XGBoost. For the first two algorithms, the scikit-learn library proposed by [74] was used, while for XGBoost, the XGBoost library proposed by [75] was implemented. Both libraries are coded in Python. As the main objective is to improve the performance of the classifiers, we have used the default parameters of each classifier provided by each library.

6.1.3. Metaheuristic Configuration

The parameters of each metaheuristic used in our experimentation were obtained from the recommendations made by the authors of the previously mentioned studies. The details of the parameters are shown in Table 6.

6.1.4. Objective Function

The objective function used to train the model is a weighted multi-objective function that aims to minimize the classification error and the number of selected features. The weighted multi-objective function can be defined as
minimize z = α · f 1 + ( 1 α ) · f 2
where
  • α is a weight parameter set to 0.99 in this experiment.
  • f 1 represents the classification error metric.
  • f 2 represents the proportion of selected features.
We define two specific objective functions to evaluate the performance of the model:
  • Objective Function 1 (OF1)
    minimize z 1 = 0.99 · ( 1 Recall Macro ) + 0.01 · NSF TNF
    Recall macro is the simple recall average for both classes, representing the ratio of false negatives overall, NSF is the number of selected features, and TNF is the total number of features.
  • Objective Function 2 (OF2)
    minimize z 2 = 0.99 · ( 1 Recall Minority ) + 0.01 · NSF TNF
    where recall minority represents the ratio of false negatives in the minority class, and as in Objective Function 1, NSF is the number of selected features, and TNF is the total number of features.

6.1.5. Experimentation Environment

A total of 24 experiments were conducted, accounting for combinations of metaheuristics (4), classifiers (3), and objective functions (2). Each experiment was repeated 31 times to ensure robust results and account for the metaheuristics’ stochastic nature. The results were averaged, and the best outcomes were compared across these repetitions. The different metrics used are described next.
The experiments were executed on a machine equipped with an Intel Core i7 processor and 16 GB of RAM. They were run on a Windows operating system, ensuring consistent trial performance. Regarding the coding of the experiments, the following Python libraries were used: NumPy==1.24.4, SciPy==1.15.2, Scikit-learn==1.1.1, Pandas==1.4.3, Matplotlib==3.5.2, Seaborn==0.11.2, XGBoost==1.7.3, and Imbalanced-learn==0.11.0.
To ensure a robust evaluation of the classifiers, we employed k-fold cross-validation in all experiments with ( k = 5 ).

6.2. Evaluation Criteria

For the evaluation, we used five types of metrics. The first one is related to the evaluation in the objective function, the second one is related to macro metrics, the third one is related to majority class metrics, the fourth one is related to minority class metrics, and the fifth one is related to the number of selected features. The mathematical formulations of each one are detailed below.

6.2.1. Macro Metrics

As defined in the scikit-learn libraries proposed by [74], macro metrics calculate metrics for each label and find their unweighted mean. This does not take label imbalance into account.
  • F-score (f1_m): The F-score is the harmonic mean of precision and recall, providing a single metric that balances both. It is calculated as
    f m = 2 · Precision Macro · Recall Macro Precision Macro + Recall Macro
  • Recall Macro (r_m): Recall macro is the average recall score over all classes, considering class imbalance by treating all classes equally. It is defined as
    r m = 1 n i = 1 n T P i T P i + F N i
    where T P i and F N i represent the true positives and false negatives for each class i, and n is the number of classes.
  • Precision Macro (p_m): Precision macro is the average precision score over all classes. It is defined as
    p m = 1 n i = 1 n T P i T P i + F P i
    where T P i and F P i are the true positives and false positives for each class i, and n is the number of classes.

6.2.2. Class-Specific Metrics

In imbalanced classification problems, it is important to evaluate the performance on both the minority and majority classes separately. The following metrics are calculated for each class:

Minority Class Metrics

  • F-score minority (f_min): The F-score for the minority class, calculated similarly to the macro F-score, focuses specifically on the performance of the minority class:
    f min = 2 · Precision Minority · Recall Minority Precision Minority + Recall Minority
  • Recall minority (r_min): Recall for the minority class measures how well the minority class is identified:
    r min = T P min T P min + F N min
  • Precision Minority (p_min): Precision for the minority class measures the accuracy of positive predictions for the minority class:
    p min = T P min T P min + F P min

Majority Class Metrics

  • F-score majority (f1_may): The F-score for the majority class, calculated similarly to the macro F-score, focuses on the majority class:
    f may = 2 · Precision Majority · Recall Majority Precision Majority + Recall Majority
  • Recall majority (r_may): Recall for the majority class measures how well the majority class is identified:
    r may = T P may T P may + F N may
  • Precision majority (p_may): Precision for the majority class measures the accuracy of positive predictions for the majority class:
    p may = T P may T P may + F P may

6.2.3. Total Features Selected (TFS)

Let T F S represent the number of selected features, which is defined as
T F S = i = 1 m I ( f i )
where m is the total number of features, and I ( f i ) is an indicator function such that
I ( f i ) = 1 if feature f i is selected 0 otherwise

6.3. Experiment Results

The following tables present the results obtained from the experiments using the different classifiers (KNN, Random Forest, XGBoost) and objective functions. The values reported in each table represent the average and standard deviation for the defined evaluation metrics across 31 runs.
Table 7 shows the experimental results related to the fitness obtained. The table shows the fitness obtained by both objective functions (OF 1 and OF 2) for each metaheuristic with its corresponding classifier. The best fitness obtained, the average, and the standard deviation of the 31 independent runs are presented. The best results for each objective function are highlighted in bold. On the other hand, Figure 9 shows the fitness distribution for each experiment using a box–whisker plot. The x-axis corresponds to the metaheuristic–classifier pair, and the y-axis corresponds to the fitness.
From Table 7 and Figure 9, it can be noted that PSO paired with XGBoost produces the best overall results in minimizing the fitness function for both objective functions. This suggests that PSO is the most effective metaheuristic for this task, with XGBoost being the best classifier for achieving lower fitness values. While the other metaheuristics (GWO, PSA, WOA) also perform reasonably well, they are outperformed by PSO, particularly when paired with XGBoost. Therefore, the combination of PSO and XGBoost is recommended for minimizing both objective functions in this problem.
Table 8 shows the experimental results related to the F1-macro (f1_m) obtained. The table shows the f1_m obtained by both objective functions (OF 1 and OF 2) for each metaheuristic with its corresponding classifier. The average and the standard deviation of the 31 independent runs are presented. The best average for each objective function is highlighted in bold. On the other hand, Figure 10 shows the f1_m distribution for each experiment using a box-whisker plot. The x-axis corresponds to the metaheuristic–classifier pair, and the y-axis corresponds to the f1_m.
Table 8 and Figure 10 show the f1_m results across the classifiers and metaheuristics. The Random Forest (RF) consistently achieved the highest F1-macro score under Objective Functions 1 and 2 (OF 1 and OF 2), averaging 0.732 across the GWO, PSA, PSO, and WOA metaheuristics. K-Nearest Neighbors (KNN) displayed the lowest performance, with average values around 0.671 to 0.672, highlighting that RF is the most effective classifier in terms of overall macro performance. At the same time, KNN struggles with feature selection methods. The standard deviation across all results remains small, indicating stable performance.
Table 9 shows the experimental results related to the F1-majority class (f1_may) obtained. The table shows the f1_may obtained by both objective functions (OF 1 and OF 2) for each metaheuristic with its corresponding classifier. The average and standard deviation of the 31 independent runs are presented. The best average for each objective function is highlighted in bold. On the other hand, Figure 11 shows the f1_may distribution for each experiment using a box–whisker plot. The x-axis corresponds to the metaheuristic–classifier pair, and the y-axis corresponds to the f1_may.
Like the F1-macro results, RF performs the best, achieving a peak F1-majority score of 0.820 for GWO and WOA under OF 1. XGBoost (XGB) performed slightly lower, with a peak of 0.808, while KNN remains the least effective.
Table 10 shows the experimental results related to the F1-minority class (f1_min) obtained. The table shows the f1_min obtained by both objective functions (OF 1 and OF 2) for each metaheuristic with its corresponding classifier. The average and the standard deviation of the 31 independent runs are presented. The best average for each objective function is highlighted in bold. On the other hand, Figure 12 shows the f1_min distribution for each experiment using a box–whisker plot. The x-axis corresponds to the metaheuristic–classifier pair, and the y-axis corresponds to the f1_min.
The highest score was achieved by WOA, with an F1-minority score of 0.645 under RF. The other metaheuristics (GWO, PSA, PSO) also performed well with RF, indicating that feature selection by metaheuristics plays an important role in identifying minority class instances. KNN consistently showed the lowest performance in both objective functions.
Table 11, Table 12 and Table 13 show the experimental results related to the precision macro (p_m), majority (p_may), and minority classes (p_min) obtained. The tables show the p_m, p_may, and p_min obtained by both objective functions (OF 1 and OF 2) for each metaheuristic with its corresponding classifier. The average and standard deviation of the 31 independent runs are presented. The best average for each objective function is highlighted in bold. On the other hand, Figure 13, Figure 14 and Figure 15 show the p_m, p_may, and p_min distributions for each experiment using a box–whisker plot. The x-axis corresponds to the metaheuristic–classifier pair, and the y-axis corresponds to the p_m, p_may, and p_min.
For macro precision (p_m), RF achieved the highest value of 0.724 across GWO, PSA, PSO, and WOA metaheuristics. Regarding majority class precision (p_may), RF and XGB reached 0.888, while KNN showed slightly lower performance. Regarding minority class precision (p_min), WOA led with a score of 0.566, followed closely by PSO and GWO.
Table 14, Table 15 and Table 16 show the experimental results related to the recall macro (r_m), majority (r_may), and minority classes (r_min) obtained. The tables show the r_m, r_may, and r_min obtained by both objective functions (OF 1 and OF 2) for each metaheuristic with its corresponding classifier. The average and the standard deviation of the 31 independent runs are presented. The best average for each objective function is highlighted in bold. On the other hand, Figure 16, Figure 17 and Figure 18 show the r_m, r_may, and r_min distributions for each experiment using a box–whisker plot. The x-axis corresponds to the metaheuristic–classifier pair, and the y-axis corresponds to the r_m, r_may, and r_min.
WOA achieved the highest recall for the minority class, with 0.770, while PSA and GWO performed well. For macro recall (r_m), RF scored 0.758, outperforming the other classifiers. In the majority class recall (r_may), WOA again took the lead, indicating its strong performance across multiple metrics.
Finally, Table 17 shows the total features selected (TFS) across the metaheuristics. The table shows the TFS obtained by both objective functions (OF 1 and OF 2) for each metaheuristic with its corresponding classifier. The average and the standard deviation of the 31 independent runs are presented. The best average for each objective function is highlighted in bold. On the other hand, Figure 19 shows the TFS distribution for each experiment using a box–whisker plot. The x-axis corresponds to the metaheuristic–classifier pair, and the y-axis corresponds to the TFS.
PSO demonstrated the most efficient feature selection, selecting the fewest features on average (36.5 for OF 1), while WOA selected the most features, reaching an average of 53.4 for XGB under OF 1. This suggests that while PSO is more aggressive in feature reduction, WOA maintains a more comprehensive feature set, which may contribute to its superior performance in minority class identification.
Overall, Random Forest emerged as the top-performing classifier across most metrics, particularly in the F1-macro, precision, and recall scores for both the majority and minority classes. WOA and PSO stood out as the most effective metaheuristics, with WOA excelling in minority class recall and PSO showing the best feature reduction capabilities. KNN consistently underperformed, suggesting it may not be suitable for this specific feature selection problem.

6.4. Statistical Test

A robust way to compare different algorithms is through statistical testing [62,76]. Our data does not follow a normal distribution, so we must perform a nonparametric test. Given this, we have applied the following comparative methodology:
  • Apply the Friedman test to determine if there is an overall statistical difference between all the algorithms.
  • If the Friedman test is positive (p-value < 0.05), the Neminyi post hoc test is applied to identify the pairs of algorithms with statistical differences.
  • Once the pairs have a statistical difference, the Wilcoxon signed-rank test will be applied to determine the directionality of the statistical difference, that is, to determine which algorithm is better than the other.
Table 18 shows the p-value obtained after applying the Friedman test for all experiments carried out considering Objective Functions 1 and 2. As can be seen, in both objective functions, a p-value lower than 0.05 is obtained; therefore, there is a global statistical difference, and we can apply the Neminyi post hoc test to determine the pairs with significant differences.
Table 19 and Table 20 show the p-values obtained after applying the Neminyi post hoc test considering the fitness of Objective Function 1 and Objective Function 2, respectively. Analyzing these tables, we observe the following:
  • From Table 19, 34 pairs with significant differences are observed.
  • From Table 20, 38 pairs with significant differences are observed.
The Wilcoxon signed-rank test was performed because the data did not follow a normal distribution, and each experiment was conducted independently. We apply this test to determine the directionality of the statistical difference detected between these pairs in the previous step. Furthermore, we considered a significance level of 0.05, where the hypotheses are as follows:
H 0 = O p t i m i z e r A O p t i m i z e r B H 1 = O p t i m i z e r A < O p t i m i z e r B
If the result of the statistical test is a p-value < 0.05 , it cannot be assumed that O p t i m i z e r A performs less well than O p t i m i z e r B , and rejects H 0 . This comparison is performed because our problem is a minimization problem.
We have used the scipy.stats.wilcoxon function from the Python scipy library. Among the configuration parameters for this function is the “alternative” parameter, which we defined with the value “less.” With this configuration, the applied test tells us whether O p t i m i z e r A is statistically less than O p t i m i z e r B , in other words, whether O p t i m i z e r A is better than O p t i m i z e r B , since we are working on a minimization problem.
Table 21 shows the number of times each algorithm was statistically better than another after applying the statistical test. The p-values for the statistical test are shown in Table 22 and Table 23. Table 21 shows that the algorithms using the KNN classifier are not statistically better than any algorithm for both objective functions. On the other hand, the algorithms incorporating the RF classifier are shown to be statistically better than other algorithms for both objective functions. A review of Table 22 and Table 23 shows that these triumphs are over algorithms that incorporate KNN. Finally, the best algorithms incorporate XGB as a classifier for both objective functions. Specifically, these algorithms are statistically better than ones that incorporate KNN and others that incorporate RF.
The algorithm that stands out the most is XGB_PSO, which triumphs 6 times in Objective Function 1 and 8 times in Objective Function 2. It is followed by the XGB_WOA algorithm, which triumphs 6 times in Objective Function 1 and 7 times in Objective Function 2. Closing the top 3, we find the XGB_PSA algorithm, which triumphs only 4 times for Objective Function 1 but 7 times for Objective Function 2.

7. Conclusions

This study demonstrated that employing metaheuristic techniques significantly reduced the features required to train machine learning models for predicting intradialytic hypotension (IDH) in dialysis patients. The XGBoost model, in particular, outperformed other classifiers such as Random Forest and K-Nearest Neighbors. Combined with feature selection techniques optimized by Particle Swarm Optimization (PSO), the XGBoost model achieved higher performance in minority class recall, minimizing false positives for hypotensive patients.
As mentioned in Section 1, our research hypothesis indicates that it is possible to detect an optimal combination of clinical and analytical parameters associated with the development of hypotension during dialysis. The results confirmed this hypothesis, as the optimized models could accurately predict hypotensive episodes with reduced features.
On the other hand, the main objective of improving the accuracy of IDH prediction while minimizing the number of required features was successfully achieved. The results of the study highlight the practical application of metaheuristics in healthcare, particularly in optimizing predictive models for critical conditions such as IDH.
These findings improve patient care by enabling early detection of hypotensive episodes, allowing for enhanced optimization of medical resources during dialysis sessions. In particular, it is observed that the XGBoost classifier outperforms both Random Forest and KNN. Furthermore, the incorporation of PSO into XGBoost delivers better experimental results than the rest, being statistically validated with tests.
While the results of this study demonstrate the effectiveness of metaheuristics in optimizing predictive models for intradialytic hypotension (IDH), several avenues for future research are worth exploring:
  • Algorithm optimization beyond feature selection: In addition to feature selection, future studies could explore the optimization of hyperparameters of machine learning models.
  • Application of deep learning models: Investigating the application of deep learning models, such as recurrent neural networks (RNNs) or long short-term memory networks (LSTMs), may provide further insight into complex relationships in the data and improve predictive accuracy.
  • Personalized prediction models: Future research could focus on creating personalized prediction models based on individual patient profiles and climatic or geographic characteristics.
  • Clinical validation and implementation: Validation through real-world clinical trials is necessary to bring the predictive model into clinical practice. This would involve integrating the model into dialysis machines or clinical decision support systems to assess its efficacy in a live healthcare setting. This gives rise to another future research oriented towards Explainable Artificial Intelligence. With this, models seek to be interpretable by professionals outside the field, such as medical professionals.
These future directions provide a framework to further the findings of this study, advance the role of predictive analytics in nephrology, and improve patient care through the integration of artificial intelligence.

Author Contributions

Conceptualization, F.C.-C., B.C. and J.M.G.-P.; methodology, F.C.-C., M.S.-L., J.B.-G., B.C., A.G.-J. and J.M.G.-P.; software, F.C.-C. and J.B.-G.; validation, M.S.-L., B.C., R.S., C.B.-A. and D.R.-P.; formal analysis, F.C.-C., M.S.-L. and J.B.-G.; investigation, F.C.-C., M.S.-L., J.B.-G., B.C., R.S., C.B.-A., A.G.-J., D.R.-P. and J.M.G.-P.; resources, F.C.-C., M.S.-L. and J.B.-G.; writing—original draft, F.C.-C., M.S.-L., J.B.-G., B.C., A.G.-J. and J.M.G.-P.; writing—review and editing, R.S., C.B.-A. and D.R.-P.; supervision, B.C., A.G.-J. and J.M.G.-P.; funding acquisition, B.C., R.S., A.G.-J., D.R.-P. and J.M.G.-P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

In our study, the dataset was provided by the Instituto de Salud Carlos III, which conducted all initial data extraction, analysis, processing, and rigorous anonymization procedures. We utilized the resulting anonymized data for our research on predicting intradialytic hypotension, ensuring that no personally identifiable information was accessed or retained. This approach not only safeguards patient privacy but also upholds data transparency and reproducibility standards, as the provenance and processing of the data are clearly documented and independently verified.

Acknowledgments

The authors would like to thank Pablo Herrera for his previous work on the analysis and debugging of the analytical and clinical databases used. Felipe Cisternas-Caneo is supported by the National Agency for Research and Development (ANID)/Scholarship Program/DOCTORADO NACIONAL/2023-21230203. José Barrera-García is supported by the National Agency for Research and Development (ANID)/Scholarship Program/DOCTORADO NACIONAL/2024-21242516. This work is part of the project “Prevention of serious pathological events in hemodialysis patients by non-invasive continuous monitoring of vital signs and analysis of circular biomarkers (ALLPREVENT)”, PMPTA23/00033, which has been funded by the Instituto de Salud Carlos III (ISCIII) within the program of R&D projects linked to personalized medicine and advanced therapies. The authors are supported by the grant Investiga sin Fronteras/VINCI/PUCV/2024.04.INV. COIL.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Sociedad Española de Nefrología. La Enfermedad Renal Crónica en España 2023. 2023. Available online: https://www.xn--diamundialdelrion-txb.com/wp-content/uploads/2023/03/SEN_dossier_DMR2023.pdf (accessed on 12 July 2024).
  2. Arrieta, J.; Rodriguez-Carmona, A.; Remon, C.; Pérez-Fontán, M.; Ortega, F.; Sanchez-Tomero, J.A.; Selgas, R. Cost comparison between haemodialysis and peritoneal dialysis outsourcing agreements. Nefrol. Publ. Of. Soc. Esp. Nefrol. 2012, 32, 247–248. [Google Scholar]
  3. Kuipers, J.; Verboom, L.M.; Ipema, K.J.; Paans, W.; Krijnen, W.P.; Gaillard, C.A.; Westerhuis, R.; Franssen, C.F. The prevalence of intradialytic hypotension in patients on conventional hemodialysis: A systematic review with meta-analysis. Am. J. Nephrol. 2019, 49, 497–506. [Google Scholar] [CrossRef] [PubMed]
  4. Flythe, J.E.; Chang, T.I.; Gallagher, M.P.; Lindley, E.; Madero, M.; Sarafidis, P.A.; Unruh, M.L.; Wang, A.Y.M.; Weiner, D.E.; Cheung, M.; et al. Blood pressure and volume management in dialysis: Conclusions from a Kidney Disease: Improving Global Outcomes (KDIGO) Controversies Conference. Kidney Int. 2020, 97, 861–876. [Google Scholar] [CrossRef]
  5. Furaz Czerpak, K.R.; Puente García, A.; Corchete Prats, E.; Moreno de la Higuera, M.; Gruss Vergara, E.; Martín-Hernández, R. Estrategias para el control de la hipotensión en hemodiálisis. Nefrología 2014, 6, 1–14. [Google Scholar]
  6. Agarwal, R. How can we prevent intradialytic hypotension? Curr. Opin. Nephrol. Hypertens. 2012, 21, 593–599. [Google Scholar] [CrossRef]
  7. Aceto, G.; Persico, V.; Pescapé, A. Industry 4.0 and health: Internet of things, big data, and cloud computing for healthcare 4.0. J. Ind. Inf. Integr. 2020, 18, 100129. [Google Scholar] [CrossRef]
  8. Gambhir, S.; Malik, S.K.; Kumar, Y. Role of soft computing approaches in healthcare domain: A mini review. J. Med Syst. 2016, 40, 1–20. [Google Scholar] [CrossRef]
  9. Rayan, Z.; Alfonse, M.; Salem, A.B.M. Machine learning approaches in smart health. Procedia Comput. Sci. 2019, 154, 361–368. [Google Scholar] [CrossRef]
  10. Slon Roblero, M.F.; Bajo Rubio, M.A.; Gonzalez-Moya, M.; Calvino Varela, J.; Perez Alba, A.; Villaro Gumpert, J.; Cigarran, S.; Vidau, P.; Garcia Marcos, S.; Abaigar Luquin, P.; et al. Experience in Spain with the first patients in home hemodialysis treated with low-flow dialysate monitors. Nefrologia 2022, 42, 460–470. [Google Scholar] [CrossRef]
  11. Gilissen, J.; Pivodic, L.; Unroe, K.T.; Van den Block, L. International COVID-19 palliative care guidance for nursing homes leaves key themes unaddressed. J. Pain Symptom Manag. 2020, 60, e56–e69. [Google Scholar] [CrossRef]
  12. Lim, C.; Kim, K.J.; Maglio, P.P. Smart cities with big data: Reference models, challenges, and considerations. Cities 2018, 82, 86–99. [Google Scholar] [CrossRef]
  13. Zadeh, L.A. Soft computing and fuzzy logic. IEEE Softw. 1994, 11, 48–56. [Google Scholar] [CrossRef]
  14. Ibrahim, D. An overview of soft computing. Procedia Comput. Sci. 2016, 102, 34–38. [Google Scholar] [CrossRef]
  15. Tamboli, S.; Bewoor, L. A review of soft computing technique for real-time data forecasting. In Proceedings of the International Conference on Communication and Information Processing (ICCIP), Chongqing, China, 15–17 November 2019. [Google Scholar]
  16. Binitha, S.; Sathya, S.S. A survey of bio inspired optimization algorithms. Int. J. Soft Comput. Eng. 2012, 2, 137–151. [Google Scholar]
  17. Sharma, D.; Chandra, P. A comparative analysis of soft computing techniques in software fault prediction model development. Int. J. Inf. Technol. 2019, 11, 37–46. [Google Scholar] [CrossRef]
  18. Santos, F.A.O.; de Jesus, G.S.; Botelho, G.A.; Macedo, H.T. Smart health: Using fuzzy logic in the monitoring of health-related indicators. In Proceedings of the 2016 8th Euro American Conference on Telematics and Information Systems (EATIS), Cartagena, Colombia, 28–29 April 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 1–4. [Google Scholar]
  19. Mehta, R. Multivariate Fuzzy Logic Based Smart Healthcare Monitoring for Risk Evaluation of Cardiac Patients. In Medical Informatics and Bioimaging Using Artificial Intelligence: Challenges, Issues, Innovations and Recent Developments; Springer: Cham, Switzerland, 2022; pp. 219–243. [Google Scholar]
  20. Alkeshuosh, A.H.; Moghadam, M.Z.; Al Mansoori, I.; Abdar, M. Using PSO algorithm for producing best rules in diagnosis of heart disease. In Proceedings of the 2017 International Conference on Computer and Applications (ICCA), Doha, Qatar, 6–7 September 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 306–311. [Google Scholar]
  21. Mansour, R.F. Evolutionary computing enriched computer-aided diagnosis system for diabetic retinopathy: A survey. IEEE Rev. Biomed. Eng. 2017, 10, 334–349. [Google Scholar] [CrossRef]
  22. Kumar, S.; Nayyar, A.; Paul, A. Swarm Intelligence and Evolutionary Algorithms in Healthcare and Drug Development; CRC Press: Boca Raton, FL, USA, 2019. [Google Scholar]
  23. Jasmine Gabriel, J.; Jani Anbarasi, L. Evolutionary computing-based feature selection for cardiovascular disease: A review. In International Virtual Conference on Industry 4.0: Select Proceedings of IVCI4.0 2020; Springer: Singapore, 2021; pp. 47–56. [Google Scholar]
  24. Kumar, T.A.; Rajmohan, R.; Pavithra, M.; Balamurugan, S. Evolutionary Intelligence for Healthcare Applications; CRC Press: Boca Raton, FL, USA, 2022. [Google Scholar]
  25. Di Biasi, L.; De Marco, F.; Auriemma Citarella, A.; Barra, P.; Piotto Piotto, S.; Tortora, G. Hybrid Approach for the Design of CNNs Using Genetic Algorithms for Melanoma Classification. In Proceedings of the Pattern Recognition, Computer Vision, and Image Processing. ICPR 2022 International Workshops and Challenges, Montreal, QC, Canada, 21–25 August 2022; Proceedings, Part I. Springer: Berlin/Heidelberg, Germany, 2022; pp. 514–528. [Google Scholar] [CrossRef]
  26. Gupta, S.; Sedamkar, R. Machine learning for healthcare: Introduction. In Machine Learning with Health Care Perspective: Machine Learning and Healthcare; Springer: Cham, Switzerland, 2020; pp. 1–25. [Google Scholar]
  27. Nayyar, A.; Gadhavi, L.; Zaman, N. Machine learning in healthcare: Review, opportunities and challenges. In Machine Learning and the Internet of Medical Things in Healthcare; Springer: Cham, Switzerland, 2021; pp. 23–45. [Google Scholar]
  28. Usmani, U.A.; Jaafar, J. Machine learning in healthcare: Current trends and the future. In Proceedings of the International Conference on Artificial Intelligence for Smart Community: AISC 2020, Universiti Teknologi Petronas, Seri Iskandar, Malaysia, 17–18 December 2020; Springer: Cham, Switzerland, 2022; pp. 659–675. [Google Scholar]
  29. Kaur, P.; Kumar, R.; Kumar, M. A healthcare monitoring system using random forest and internet of things (IoT). Multimed. Tools Appl. 2019, 78, 19905–19916. [Google Scholar] [CrossRef]
  30. Haglin, J.M.; Jimenez, G.; Eltorai, A.E. Artificial neural networks in medicine. Health Technol. 2019, 9, 1–6. [Google Scholar] [CrossRef]
  31. Lisboa, P.J.; Ifeachor, E.C.; Szczepaniak, P.S. Artificial Neural Networks in Biomedicine; Springer Science & Business Media: Berlin, Germany, 2000. [Google Scholar]
  32. Filist, S.; Al-Kasasbeh, R.T.; Shatalova, O.; Aikeyeva, A.; Korenevskiy, N.; Shaqadan, A.; Trifonov, A.; Ilyash, M. Developing neural network model for predicting cardiac and cardiovascular health using bioelectrical signal processing. Comput. Methods Biomech. Biomed. Eng. 2022, 25, 908–921. [Google Scholar] [CrossRef]
  33. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  34. Sujith, A.; Sajja, G.S.; Mahalakshmi, V.; Nuhmani, S.; Prasanalakshmi, B. Systematic review of smart health monitoring using deep learning and Artificial intelligence. Neurosci. Inform. 2022, 2, 100028. [Google Scholar] [CrossRef]
  35. Yang, S.; Zhu, F.; Ling, X.; Liu, Q.; Zhao, P. Intelligent health care: Applications of deep learning in computational medicine. Front. Genet. 2021, 12, 607471. [Google Scholar] [CrossRef] [PubMed]
  36. Iyortsuun, N.K.; Kim, S.H.; Jhon, M.; Yang, H.J.; Pant, S. A review of machine learning and deep learning approaches on mental health diagnosis. Healthcare 2023, 11, 285. [Google Scholar] [CrossRef] [PubMed]
  37. Singh, K.; Malhotra, J. Deep learning based smart health monitoring for automated prediction of epileptic seizures using spectral analysis of scalp EEG. Phys. Eng. Sci. Med. 2021, 44, 1161–1173. [Google Scholar] [CrossRef]
  38. Shafi, J.; Obaidat, M.S.; Krishna, P.V.; Sadoun, B.; Pounambal, M.; Gitanjali, J. Prediction of heart abnormalities using deep learning model and wearabledevices in smart health homes. Multimed. Tools Appl. 2022, 81, 543–557. [Google Scholar] [CrossRef]
  39. Zitzler, E.; Deb, K.; Thiele, L. Comparison of Multiobjective Evolutionary Algorithms: Empirical Results. Evol. Comput. 2000, 8, 173–195. [Google Scholar] [CrossRef]
  40. Haupt, R.L.; Haupt, S.E. Practical Genetic Algorithms; John Wiley & Sons: Hoboken, NJ, USA, 2004. [Google Scholar]
  41. Houssein, E.H.; Saber, E.; Ali, A.A.; Wazery, Y.M. Integrating metaheuristics and artificial intelligence for healthcare: Basics, challenging and future directions. Artif. Intell. Rev. 2024, 57, 205. [Google Scholar] [CrossRef]
  42. Kaur, S.; Kumar, Y.; Koul, A.; Kumar Kamboj, S. A systematic review on metaheuristic optimization techniques for feature selections in disease diagnosis: Open issues and challenges. Arch. Comput. Methods Eng. 2023, 30, 1863–1895. [Google Scholar] [CrossRef]
  43. Hong, D.; Chang, H.; He, X.; Zhan, Y.; Tong, R.; Wu, X.; Li, G. Construction of an Early Alert System for Intradialytic Hypotension before Initiating Hemodialysis Based on Machine Learning. Kidney Dis. 2023, 9, 433–442. [Google Scholar] [CrossRef]
  44. Gervasoni, F.; Bellocchio, F.; Rosenberger, J.; Arkossy, O.; Ion Titapiccolo, J.; Kovarova, V.; Larkin, J.; Nikam, M.; Stuard, S.; Tripepi, G.L.; et al. Development and validation of AI-based triage support algorithms for prevention of intradialytic hypotension. J. Nephrol. 2023, 36, 2001–2011. [Google Scholar] [CrossRef]
  45. Moeinzadeh, F.; Sattari, M. Proposed Method for Predicting COVID-19 Severity in Chronic Kidney Disease Patients Based on Ant Colony Algorithm and CHAID. J. Adv. Med Biomed. Res. 2022, 30, 507–512. [Google Scholar] [CrossRef]
  46. Yang, X.; Zhao, D.; Yu, F.; Heidari, A.A.; Bano, Y.; Ibrohimov, A.; Liu, Y.; Cai, Z.; Chen, H.; Chen, X. An optimized machine learning framework for predicting intradialytic hypotension using indexes of chronic kidney disease-mineral and bone disorders. Comput. Biol. Med. 2022, 145, 105510. [Google Scholar] [CrossRef] [PubMed]
  47. Othman, M.; Elbasha, A.M.; Naga, Y.S.; Moussa, N.D. Early prediction of hemodialysis complications employing ensemble techniques. BioMedical Eng. Online 2022, 21, 74. [Google Scholar] [CrossRef] [PubMed]
  48. Nafisi, V.R.; Shahabi, M. Intradialytic hypotension related episodes identification based on the most effective features of photoplethysmography signal. Comput. Methods Programs Biomed. 2018, 157, 1–9. [Google Scholar] [CrossRef] [PubMed]
  49. Arienzo, G.; Citarella, A.A.; De Marco, F.; De Roberto, A.M.; Di Biasi, L.; Francese, R.; Tortora, G. Cardioview: A framework for detection premature ventricular contractions with explainable artificial intelligence. In Proceedings of the INI-DH 2024: Workshop on Innovative Interfaces in Digital Healthcare, in Conjunction with International Conference on Advanced Visual Interfaces, Genoa, Italy, 3–7 June 2024; pp. 3–7. [Google Scholar]
  50. De Marco, F.; Di Biasi, L.; Auriemma Citarella, A.; Tortora, G. Improving pvc detection in ecg signals: A recurrent neural network approach. In Proceedings of the Italian Workshop on Artificial Life and Evolutionary Computation, Venice, Italy, 6–8 September 2023; Springer: Berlin/Heidelberg, Germany, 2023; pp. 256–267. [Google Scholar]
  51. Mendoza-Pittí, L.; Gómez-Pulido, J.M.; Vargas-Lombardo, M.; Gómez-Pulido, J.A.; Polo-Luque, M.L.; Rodréguez-Puyol, D. Machine-learning model to predict the intradialytic hypotension based on clinical-analytical data. IEEE Access 2022, 10, 72065–72079. [Google Scholar] [CrossRef]
  52. Cheripurathu, K.G.; Kulkarni, S. Integrating Microservices and Microfrontends: A Comprehensive Literature Review on Architecture, Design Patterns, and Implementation Challenges. J. Sci. Res. Technol. 2024, 2, 1–12. [Google Scholar]
  53. Chowdhury, M.Z.I.; Turin, T.C. Variable selection strategies and its importance in clinical prediction modelling. Fam. Med. Community Health 2020, 8, e000262. [Google Scholar] [CrossRef]
  54. Steyerberg, E.W.; Vergouwe, Y. Towards better clinical prediction models: Seven steps for development and an ABCD for validation. Eur. Heart J. 2014, 35, 1925–1931. [Google Scholar] [CrossRef]
  55. Mehrpoor, G.; Azimzadeh, M.M.; Monfared, A. Data mining: A novel outlook to explore knowledge in health and medical sciences. Int. J. Travel Med. Glob. Health 2014, 2, 87–90. [Google Scholar]
  56. Barrera-García, J.; Cisternas-Caneo, F.; Crawford, B.; Gómez Sánchez, M.; Soto, R. Feature selection problem and metaheuristics: A systematic literature review about its formulation, evaluation and applications. Biomimetics 2023, 9, 9. [Google Scholar] [CrossRef]
  57. Rajwar, K.; Deep, K.; Das, S. An exhaustive review of the metaheuristic algorithms for search and optimization: Taxonomy, applications, and open challenges. Artif. Intell. Rev. 2023, 56, 13187–13257. [Google Scholar] [CrossRef]
  58. Becerra-Rozas, M.; Lemus-Romani, J.; Cisternas-Caneo, F.; Crawford, B.; Soto, R.; Astorga, G.; Castro, C.; García, J. Continuous metaheuristics for binary optimization problems: An updated systematic literature review. Mathematics 2022, 11, 129. [Google Scholar] [CrossRef]
  59. Kennedy, J.; Eberhart, R.C. A discrete binary version of the particle swarm algorithm. In Proceedings of the 1997 IEEE International Conference on Systems, Man, and Cybernetics. Computational Cybernetics and Simulation, Orlando, FL, USA, 12–15 October 1997; IEEE: Piscataway, NJ, USA, 1997; Volume 5, pp. 4104–4108. [Google Scholar] [CrossRef]
  60. Mirjalili, S.; Lewis, A. S-shaped versus V-shaped transfer functions for binary particle swarm optimization. Swarm Evol. Comput. 2013, 9, 1–14. [Google Scholar] [CrossRef]
  61. Crawford, B.; Soto, R.; Astorga, G.; García, J.; Castro, C.; Paredes, F. Putting continuous metaheuristics to work in binary search spaces. Complexity 2017, 2017, 8404231. [Google Scholar] [CrossRef]
  62. Lanza-Gutierrez, J.M.; Crawford, B.; Soto, R.; Berrios, N.; Gomez-Pulido, J.A.; Paredes, F. Analyzing the effects of binarization techniques when solving the set covering problem through swarm optimization. Expert Syst. Appl. 2017, 70, 67–82. [Google Scholar] [CrossRef]
  63. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95-International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995; IEEE: Piscataway, NJ, USA, 1995; Volume 4, pp. 1942–1948. [Google Scholar]
  64. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey wolf optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef]
  65. Lemus-Romani, J.; Becerra-Rozas, M.; Crawford, B.; Soto, R.; Cisternas-Caneo, F.; Vega, E.; Castillo, M.; Tapia, D.; Astorga, G.; Palma, W.; et al. A novel learning-based binarization scheme selector for swarm algorithms solving combinatorial problems. Mathematics 2021, 9, 2887. [Google Scholar] [CrossRef]
  66. Ab. Aziz, N.A.; Ab. Aziz, K. Pendulum Search Algorithm: An Optimization Algorithm Based on Simple Harmonic Motion and Its Application for a Vaccine Distribution Problem. Algorithms 2022, 15, 214. [Google Scholar] [CrossRef]
  67. Crawford, B.; Cisternas-Caneo, F.; Sepúlveda, K.; Soto, R.; Paz, Á.; Peña, A.; León de la Barra, C.; Rodriguez-Tello, E.; Astorga, G.; Castro, C.; et al. B-PSA: A Binary Pendulum Search Algorithm for the Feature Selection Problem. Computers 2023, 12, 249. [Google Scholar] [CrossRef]
  68. Mirjalili, S.; Lewis, A. The whale optimization algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  69. Lemaître, G.; Nogueira, F.; Aridas, C.K. Imbalanced-learn: A Python Toolbox to Tackle the Curse of Imbalanced Datasets in Machine Learning. J. Mach. Learn. Res. 2017, 18, 1–5. [Google Scholar]
  70. Agrawal, P.; Abutarboush, H.F.; Ganesh, T.; Mohamed, A.W. Metaheuristic Algorithms on Feature Selection: A Survey of One Decade of Research (2009–2019). IEEE Access 2021, 9, 26766–26791. [Google Scholar] [CrossRef]
  71. Igel, C. No Free Lunch Theorems: Limitations and Perspectives of Metaheuristics. In Theory and Principled Methods for the Design of Metaheuristics; Borenstein, Y., Moraglio, A., Eds.; Springer: Berlin/Heidelberg, Germany, 2014; pp. 1–23. [Google Scholar] [CrossRef]
  72. Ho, Y.C.; Pepyne, D.L. Simple explanation of the no-free-lunch theorem and its implications. J. Optim. Theory Appl. 2002, 115, 549–570. [Google Scholar] [CrossRef]
  73. Wolpert, D.; Macready, W. No free lunch theorems for optimization. IEEE Trans. Evol. Comput. 1997, 1, 67–82. [Google Scholar] [CrossRef]
  74. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  75. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 13–17 August 2016; KDD ’16. pp. 785–794. [Google Scholar] [CrossRef]
  76. Hays, W.L.; Winkler, R.L. Statistics: Probability, Inference, and Decision; Holt, Rinehart and Winston: Austin, TX, USA, 1970. [Google Scholar]
Figure 1. Taxonomy of Soft Computing techniques.
Figure 1. Taxonomy of Soft Computing techniques.
Biomimetics 10 00314 g001
Figure 2. Digitized clinical database of hemodialysis patients.
Figure 2. Digitized clinical database of hemodialysis patients.
Biomimetics 10 00314 g002
Figure 3. Predictive model at the start of the hemodialysis session.
Figure 3. Predictive model at the start of the hemodialysis session.
Biomimetics 10 00314 g003
Figure 4. HDI prediction system mockup.
Figure 4. HDI prediction system mockup.
Biomimetics 10 00314 g004
Figure 5. Binary representation of vectors S and D.
Figure 5. Binary representation of vectors S and D.
Biomimetics 10 00314 g005
Figure 6. Two−step technique.
Figure 6. Two−step technique.
Biomimetics 10 00314 g006
Figure 7. S-shaped and V-shaped transfer functions.
Figure 7. S-shaped and V-shaped transfer functions.
Biomimetics 10 00314 g007
Figure 8. Enhanced prediction of IDH through ML and biomarker analysis.
Figure 8. Enhanced prediction of IDH through ML and biomarker analysis.
Biomimetics 10 00314 g008
Figure 9. Fitness boxplots for each metaheuristic, classifier, and objective function used.
Figure 9. Fitness boxplots for each metaheuristic, classifier, and objective function used.
Biomimetics 10 00314 g009
Figure 10. Boxplots of f1_m for each metaheuristic, classifier, and objective function used.
Figure 10. Boxplots of f1_m for each metaheuristic, classifier, and objective function used.
Biomimetics 10 00314 g010
Figure 11. Boxplots of f1_may for each metaheuristic, classifier, and objective function used.
Figure 11. Boxplots of f1_may for each metaheuristic, classifier, and objective function used.
Biomimetics 10 00314 g011
Figure 12. Boxplots of f1_min for each metaheuristic, classifier, and objective function used.
Figure 12. Boxplots of f1_min for each metaheuristic, classifier, and objective function used.
Biomimetics 10 00314 g012
Figure 13. Boxplots of p_m for each metaheuristic, classifier, and objective function used.
Figure 13. Boxplots of p_m for each metaheuristic, classifier, and objective function used.
Biomimetics 10 00314 g013
Figure 14. Boxplots of p_may for each metaheuristic, classifier, and objective function used.
Figure 14. Boxplots of p_may for each metaheuristic, classifier, and objective function used.
Biomimetics 10 00314 g014
Figure 15. Boxplots of p_min for each metaheuristic, classifier, and objective function used.
Figure 15. Boxplots of p_min for each metaheuristic, classifier, and objective function used.
Biomimetics 10 00314 g015
Figure 16. Boxplots of r_m for each metaheuristic, classifier, and objective function used.
Figure 16. Boxplots of r_m for each metaheuristic, classifier, and objective function used.
Biomimetics 10 00314 g016
Figure 17. Boxplots of r_may for each metaheuristic, classifier, and objective function used.
Figure 17. Boxplots of r_may for each metaheuristic, classifier, and objective function used.
Biomimetics 10 00314 g017
Figure 18. Boxplots of r_min for each metaheuristic, classifier, and objective function used.
Figure 18. Boxplots of r_min for each metaheuristic, classifier, and objective function used.
Biomimetics 10 00314 g018
Figure 19. Boxplots of TFS for each metaheuristic, classifier, and objective function used.
Figure 19. Boxplots of TFS for each metaheuristic, classifier, and objective function used.
Biomimetics 10 00314 g019
Table 1. IDH dataset.
Table 1. IDH dataset.
LabelsFeatures
Instances Non Hypotensive Hypotensive Pre. Post.
IDH dataset68,57448,764 (71.11%)19,810 (28.89%)7187
Table 2. Classification results without optimization of the prediction model.
Table 2. Classification results without optimization of the prediction model.
Classifierf1_mr_mp_mfs_minr_minp_minfs_mayr_mayp_may
XGBoost0.7370.7210.7690.6070.5290.7110.8670.9130.827
RF0.7170.6980.7630.5700.4740.7150.8640.9230.812
KNN0.6510.6410.6770.4740.4050.5710.8280.8760.784
Table 3. S-shaped and V-shaped transfer functions.
Table 3. S-shaped and V-shaped transfer functions.
S-ShapedV-Shaped
NameEquationNameEquation
S1 T ( d i j ) = 1 1 + e 2 d i j V1 T ( d i j ) = e r f π 2 d i j
S2 T ( d i j ) = 1 1 + e d i j V2 T ( d i j ) = t a n h ( d i j )
S3 T ( d i j ) = 1 1 + e d i j 2 V3 T ( d i j ) = d i j 1 + ( d i j ) 2
S4 T ( d i j ) = 1 1 + e d i j 3 V4 T ( d i j ) = 2 π a r c t a n π 2 d i j
Table 4. Binarization rules.
Table 4. Binarization rules.
TypeBinarization Rules
Standard X new j = 1 if rand T ( d i j ) 0 else .
Complement X new j = Complement ( X w j ) if rand T ( d i j ) 0 else .
Static Probability X n e w j = 0 i f T ( d i j ) α X w j i f α < T ( d i j ) 1 2 ( 1 + α ) 1 i f T ( d i j ) 1 2 ( 1 + α )
Elitist X n e w j = X B e s t j i f r a n d < T ( d i j ) 0 e l s e .
Roulette Elitist X n e w j = P [ X n e w j = ζ j ] = f ( ζ ) δ Q g f ( δ ) if rand T ( d i j ) P [ X n e w j = 0 ] = 1 e l s e .
Table 5. Results of classifiers with RandomUnderSampler and varying parameters.
Table 5. Results of classifiers with RandomUnderSampler and varying parameters.
ClassifierParameterf1_mr_mp_mf1_minr_minp_minf1_mayr_mayp_may
KNN1.00.6320.6590.6340.5270.6520.4430.7370.6660.825
Random Forest1.00.720.7460.7120.6290.7380.5490.810.7530.876
XGBoost1.00.7250.7560.7180.640.7660.5490.810.7450.887
KNN0.90.640.660.6380.5260.6220.4550.7540.6980.82
Random Forest0.90.7270.7450.7180.6310.7090.5680.8220.7810.869
XGBoost0.90.7340.7580.7250.6450.7450.5690.8230.7710.881
KNN0.80.6460.660.6420.5230.590.470.7690.7290.814
Random Forest0.80.7320.7430.7250.630.6770.590.8340.8090.86
XGBoost0.80.7390.7560.7310.6460.7180.5870.8330.7950.874
KNN0.70.6520.6590.6480.5190.5530.4880.7860.7650.808
Random Forest0.70.7340.7370.7320.6250.6360.6130.8430.8370.85
XGBoost0.70.7440.7530.7370.6450.6860.6090.8430.8210.866
KNN0.60.6550.6550.6550.5090.5080.5090.80.8010.8
Random Forest0.60.7360.730.7430.6180.5930.6460.8540.8680.84
XGBoost0.60.7450.7450.7450.6380.640.6360.8520.8510.853
Table 6. Setting parameters for each MH.
Table 6. Setting parameters for each MH.
MHParameterValue
PSO V M a x 5000
c 1 2
c 2 2
w m a x 0.9
w m i n 0.2
GWOadecreases linearly from 2 to 0
WOAadecreases linearly from 2 to 0
b1
PSAfree parameters
Table 7. Results by evaluation metric: fitness (best, average, and standard deviation).
Table 7. Results by evaluation metric: fitness (best, average, and standard deviation).
OF 1OF 2
KNN RF XGB KNN RF XGB
Fitness Best Avg. Std. Best Avg. Std. Best Avg. Std. Best Avg. Std. Best Avg. Std. Best Avg. Std.
GWO0.2980.3020.0030.2380.2420.0020.2360.2380.0010.2960.3020.0030.2350.2400.0030.2110.2150.002
PSA0.2980.3030.0020.2380.2420.0020.2360.2380.0010.2970.3020.0030.2360.2410.0030.2130.2150.001
PSO0.2950.2990.0030.2360.2400.0020.2350.2370.0010.2910.2980.0030.2370.2420.0020.2120.2140.001
WOA0.2950.3010.0030.2350.2400.0020.2360.2380.0010.2950.3010.0030.2350.2410.0030.2100.2150.002
Table 8. Results by evaluation metric: f1_m (average and standard deviation).
Table 8. Results by evaluation metric: f1_m (average and standard deviation).
OF 1OF 2
KNN RF XGB KNN RF XGB
f1_m Avg. Std. Avg. Std. Avg. Std. Avg. Std. Avg. Std. Avg. Std.
GWO0.6710.0020.7320.0020.7240.0010.6680.0040.7300.0030.7220.002
PSA0.6690.0030.7310.0020.7240.0010.6690.0040.7290.0020.7210.002
PSO0.6720.0030.7320.0020.7240.0010.6710.0040.7290.0030.7220.003
WOA0.6710.0030.7320.0010.7240.0020.6680.0040.7290.0030.7220.002
Table 9. Results by evaluation metric: f1_may (average and standard deviation).
Table 9. Results by evaluation metric: f1_may (average and standard deviation).
OF 1OF 2
KNN RF XGB KNN RF XGB
f1_may Avg. Std. Avg. Std. Avg. Std. Avg. Std. Avg. Std. Avg. Std.
GWO0.7700.0020.8200.0010.8090.0010.7680.0030.8170.0020.8060.002
PSA0.7690.0030.8190.0020.8080.0010.7680.0030.8170.0020.8050.002
PSO0.7710.0030.8200.0020.8080.0010.7700.0030.8170.0020.8060.002
WOA0.7700.0020.8200.0010.8080.0020.7680.0040.8180.0020.8060.002
Table 10. Results by evaluation metric: f1_min (average and standard deviation).
Table 10. Results by evaluation metric: f1_min (average and standard deviation).
OF 1OF 2
KNN RF XGB KNN RF XGB
f1_min Avg. Std. Avg. Std. Avg. Std. Avg. Std. Avg. Std. Avg. Std.
GWO0.5720.0030.6440.0020.6400.0010.5690.0040.6420.0030.6380.002
PSA0.5700.0030.6430.0020.6390.0010.5690.0040.6410.0030.6370.002
PSO0.5730.0040.6440.0020.6400.0010.5730.0040.6410.0030.6380.003
WOA0.5720.0030.6450.0010.6400.0020.5690.0050.6420.0030.6380.002
Table 11. Results by evaluation metric: p_m (average and standard deviation).
Table 11. Results by evaluation metric: p_m (average and standard deviation).
OF 1OF 2
KNN RF XGB KNN RF XGB
p_m Avg. Std. Avg. Std. Avg. Std. Avg. Std. Avg. Std. Avg. Std.
GWO0.6680.0020.7240.0010.7180.0010.6660.0030.7210.0020.7160.002
PSA0.6670.0030.7230.0020.7170.0010.6660.0030.7210.0020.7160.001
PSO0.6690.0030.7240.0020.7180.0010.6690.0030.7210.0020.7160.002
WOA0.6680.0030.7240.0010.7180.0020.6660.0040.7220.0030.7160.002
Table 12. Results by evaluation metric: p_may (average and standard deviation).
Table 12. Results by evaluation metric: p_may (average and standard deviation).
OF 1OF 2
KNN RF XGB KNN RF XGB
p_may Avg. Std. Avg. Std. Avg. Std. Avg. Std. Avg. Std. Avg. Std.
GWO0.8490.0020.8820.0010.8870.0010.8470.0020.8820.0010.8880.001
PSA0.8480.0020.8820.0010.8870.0010.8480.0020.8820.0010.8870.001
PSO0.8500.0020.8830.0010.8880.0010.8500.0020.8820.0010.8880.001
WOA0.8490.0020.8830.0010.8870.0010.8480.0020.8820.0010.8880.001
Table 13. Results by evaluation metric: p_min (average and standard deviation).
Table 13. Results by evaluation metric: p_min (average and standard deviation).
OF 1OF 2
KNN RF XGB KNN RF XGB
p_min Avg. Std. Avg. Std. Avg. Std. Avg. Std. Avg. Std. Avg. Std.
GWO0.4880.0030.5640.0020.5470.0020.4840.0040.5610.0040.5450.003
PSA0.4850.0040.5640.0030.5480.0020.4850.0050.5600.0030.5430.002
PSO0.4890.0040.5650.0030.5480.0020.4880.0050.5600.0040.5450.003
WOA0.4870.0030.5660.0010.5480.0030.4840.0050.5610.0040.5440.003
Table 14. Results by evaluation metric: r_m (average and standard deviation).
Table 14. Results by evaluation metric: r_m (average and standard deviation).
OF 1OF 2
KNN RF XGB KNN RF XGB
r_m Avg. Std. Avg. Std. Avg. Std. Avg. Std. Avg. Std. Avg. Std.
GWO0.6980.0030.7570.0020.7550.0010.6960.0030.7560.0020.7540.002
PSA0.6970.0030.7570.0020.7550.0010.6960.0040.7550.0020.7540.001
PSO0.6990.0030.7570.0020.7550.0010.6990.0040.7550.0020.7540.002
WOA0.6980.0030.7580.0010.7550.0010.6960.0040.7560.0020.7540.001
Table 15. Results by evaluation metric: r_may (average and standard deviation).
Table 15. Results by evaluation metric: r_may (average and standard deviation).
OF 1OF 2
KNN RF XGB KNN RF XGB
r_may Avg. Std. Avg. Std. Avg. Std. Avg. Std. Avg. Std. Avg. Std.
GWO0.7060.0020.7650.0020.7420.0020.7020.0040.7610.0030.7380.003
PSA0.7030.0040.7650.0030.7420.0020.7020.0040.7610.0030.7370.002
PSO0.7060.0040.7650.0030.7420.0020.7050.0040.7610.0040.7390.003
WOA0.7040.0020.7660.0010.7420.0030.7010.0050.7620.0030.7380.003
Table 16. Results by evaluation metric: r_min (average and standard deviation).
Table 16. Results by evaluation metric: r_min (average and standard deviation).
OF 1OF 2
KNN RF XGB KNN RF XGB
r_min Avg. Std. Avg. Std. Avg. Std. Avg. Std. Avg. Std. Avg. Std.
GWO0.6910.0030.7490.0020.7680.0020.6890.0030.7500.0020.7700.002
PSA0.6900.0030.7490.0020.7680.0020.6900.0030.7490.0010.7700.002
PSO0.6930.0040.7490.0010.7690.0020.6930.0040.7490.0020.7690.002
WOA0.6910.0040.7500.0020.7680.0020.6890.0040.7490.0020.7700.002
Table 17. Results by evaluation metric: total features selected (TFS) (average and standard deviation).
Table 17. Results by evaluation metric: total features selected (TFS) (average and standard deviation).
OF 1OF 2
KNN RF XGB KNN RF XGB
TFS Avg. Std. Avg. Std. Avg. Std. Avg. Std. Avg. Std. Avg. Std.
GWO39.42.242.60.846.84.239.63.343.44.248.13.6
PSA38.22.842.64.649.24.139.33.543.65.645.73.7
PSO36.53.439.92.748.54.837.53.743.63.546.04.5
WOA37.04.241.03.953.47.737.93.940.42.448.76.1
Table 18. Friedman test for objective functions OF 1 and OF 2.
Table 18. Friedman test for objective functions OF 1 and OF 2.
OF1OF2
p-value 1.51 × 10 34 5.09 × 10 35
Table 19. Neminyi post hoc test considering OF 1.
Table 19. Neminyi post hoc test considering OF 1.
AlgorithmKNN_GWOKNN_PSAKNN_PSOKNN_WOARF_GWORF_PSARF_PSORF_WOAXGB_GWOXGB_PSAXGB_PSOXGB_WOA
KNN_GWOX--- 4.69 × 10 02 4.38 × 10 02 1.53 × 10 03 5.83 × 10 04 7.04 × 10 10 1.55 × 10 08 3.81 × 10 10 1.74 × 10 09
KNN_PSA-X-- 2.43 × 10 03 2.22 × 10 03 3.30 × 10 05 1.04 × 10 05 1.83 × 10 12 5.81 × 10 11 9.25 × 10 13 5.02 × 10 12
KNN_PSO--X--- 1.23 × 10 02 5.39 × 10 03 2.36 × 10 08 4.09 × 10 07 1.34 × 10 08 5.45 × 10 08
KNN_WOA---X-- 3.48 × 10 03 1.39 × 10 03 2.71 × 10 09 5.45 × 10 08 1.50 × 10 09 6.53 × 10 09
RF_GWO 4.69 × 10 02 2.43 × 10 03 --X--- 2.29 × 10 02 - 1.69 × 10 02 3.55 × 10 02
RF_PSA 4.38 × 10 02 2.22 × 10 03 ---X-- 2.47 × 10 02 - 1.82 × 10 02 3.81 × 10 02
RF_PSO 1.53 × 10 03 3.30 × 10 05 1.23 × 10 02 3.48 × 10 03 --X-----
RF_WOA 5.83 × 10 04 1.04 × 10 05 5.39 × 10 03 1.39 × 10 03 ---X----
XGB_GWO 7.04 × 10 10 1.83 × 10 12 2.36 × 10 08 2.71 × 10 09 2.29 × 10 02 2.47 × 10 02 --X---
XGB_PSA 1.55 × 10 08 5.81 × 10 11 4.09 × 10 07 5.45 × 10 08 -----X--
XGB_PSO 3.81 × 10 10 9.25 × 10 13 1.34 × 10 08 1.50 × 10 09 1.69 × 10 02 1.82 × 10 02 ----X-
XGB_WOA 1.74 × 10 09 5.02 × 10 12 5.45 × 10 08 6.53 × 10 09 3.55 × 10 02 3.81 × 10 02 -----X
Table 20. Neminyi post hoc test considering OF 2.
Table 20. Neminyi post hoc test considering OF 2.
AlgorithmKNN_GWOKNN_PSAKNN_PSOKNN_WOARF_GWORF_PSARF_PSORF_WOAXGB_GWOXGB_PSAXGB_PSOXGB_WOA
KNN_GWOX--- 2.73 × 10 03 1.02 × 10 02 1.42 × 10 02 1.81 × 10 02 6.16 × 10 10 1.07 × 10 10 6.63 × 10 13 9.09 × 10 11
KNN_PSA-X-- 7.30 × 10 03 2.46 × 10 02 3.32 × 10 02 4.13 × 10 02 3.34 × 10 09 6.16 × 10 10 4.50 × 10 12 5.27 × 10 10
KNN_PSO--X----- 6.34 × 10 07 1.43 × 10 07 1.82 × 10 09 1.25 × 10 07
KNN_WOA---X 3.00 × 10 03 1.11 × 10 02 1.54 × 10 02 1.95 × 10 02 7.20 × 10 10 1.26 × 10 10 7.91 × 10 13 1.07 × 10 10
RF_GWO 2.73 × 10 03 7.30 × 10 03 - 3.00 × 10 03 X----- 1.31 × 10 02 -
RF_PSA 1.02 × 10 02 2.46 × 10 02 - 1.11 × 10 02 -X--- 4.13 × 10 02 3.60 × 10 03 3.84 × 10 02
RF_PSO 1.42 × 10 02 3.32 × 10 02 - 1.54 × 10 02 --X-- 3.09 × 10 02 2.49 × 10 03 2.86 × 10 02
RF_WOA 1.81 × 10 02 4.13 × 10 02 - 1.95 × 10 02 ---X- 2.46 × 10 02 1.88 × 10 03 2.28 × 10 02
XGB_GWO 6.16 × 10 10 3.34 × 10 09 6.34 × 10 07 7.20 × 10 10 ----X---
XGB_PSA 1.07 × 10 10 6.16 × 10 10 1.43 × 10 07 1.26 × 10 10 - 4.13 × 10 02 3.09 × 10 02 2.46 × 10 02 -X--
XGB_PSO 6.63 × 10 13 4.50 × 10 12 1.82 × 10 09 7.91 × 10 13 1.31 × 10 02 3.60 × 10 03 2.49 × 10 03 1.88 × 10 03 --X-
XGB_WOA 9.09 × 10 11 5.27 × 10 10 1.25 × 10 07 1.07 × 10 10 - 3.84 × 10 02 2.86 × 10 02 2.28 × 10 02 ---X
Table 21. Summary Wilcoxon signed-rank test.
Table 21. Summary Wilcoxon signed-rank test.
AlgorithmOF1OF2
KNN_GWO00
KNN_PSA00
KNN_PSO00
KNN_WOA00
RF_GWO23
RF_PSA23
RF_PSO43
RF_WOA43
XGB_GWO64
XGB_PSA47
XGB_PSO68
XGB_WOA67
Table 22. Wilcoxon signed-rank test for pairs with significant difference from Table 19.
Table 22. Wilcoxon signed-rank test for pairs with significant difference from Table 19.
Comparisonp-ValueConclusion
RF_GWO v/s KNN_GWO0.0RF_GWO is better than KNN_GWO
RF_GWO v/s KNN_PSA0.0RF_GWO is better than KNN_PSA
RF_PSA v/s KNN_GWO0.0RF_PSA is better than KNN_GWO
RF_PSA v/s KNN_PSA0.0RF_PSA is better than KNN_PSA
RF_PSO v/s KNN_GWO0.0RF_PSO is better than KNN_GWO
RF_PSO v/s KNN_PSA0.0RF_PSO is better than KNN_PSA
RF_PSO v/s KNN_PSO0.0RF_PSO is better than KNN_PSO
RF_PSO v/s KNN_WOA0.0RF_PSO is better than KNN_WOA
RF_WOA v/s KNN_GWO0.0RF_WOA is better than KNN_GWO
RF_WOA v/s KNN_PSA0.0RF_WOA is better than KNN_PSA
RF_WOA v/s KNN_PSO0.0RF_WOA is better than KNN_PSO
RF_WOA v/s KNN_WOA0.0RF_WOA is better than KNN_WOA
XGB_GWO v/s KNN_GWO0.0XGB_GWO is better than KNN_GWO
XGB_GWO v/s KNN_PSA0.0XGB_GWO is better than KNN_PSA
XGB_GWO v/s KNN_PSO0.0XGB_GWO is better than KNN_PSO
XGB_GWO v/s KNN_WOA0.0XGB_GWO is better than KNN_WOA
XGB_GWO v/s RF_GWO0.0001XGB_GWO is better than RF_GWO
XGB_GWO v/s RF_PSA0.0XGB_GWO is better than RF_PSA
XGB_PSA v/s KNN_GWO0.0XGB_PSA is better than KNN_GWO
XGB_PSA v/s KNN_PSA0.0XGB_PSA is better than KNN_PSA
XGB_PSA v/s KNN_PSO0.0XGB_PSA is better than KNN_PSO
XGB_PSA v/s KNN_WOA0.0XGB_PSA is better than KNN_WOA
XGB_PSO v/s KNN_GWO0.0XGB_PSO is better than KNN_GWO
XGB_PSO v/s KNN_PSA0.0XGB_PSO is better than KNN_PSA
XGB_PSO v/s KNN_PSO0.0XGB_PSO is better than KNN_PSO
XGB_PSO v/s KNN_WOA0.0XGB_PSO is better than KNN_WOA
XGB_PSO v/s RF_GWO0.0001XGB_PSO is better than RF_GWO
XGB_PSO v/s RF_PSA0.0001XGB_PSO is better than RF_PSA
XGB_WOA v/s KNN_GWO0.0XGB_WOA is better than KNN_GWO
XGB_WOA v/s KNN_PSA0.0XGB_WOA is better than KNN_PSA
XGB_WOA v/s KNN_PSO0.0XGB_WOA is better than KNN_PSO
XGB_WOA v/s KNN_WOA0.0XGB_WOA is better than KNN_WOA
XGB_WOA v/s RF_GWO0.0001XGB_WOA is better than RF_GWO
XGB_WOA v/s RF_PSA0.0XGB_WOA is better than RF_PSA
Table 23. Wilcoxon signed-rank test for pairs with significant difference from Table 20.
Table 23. Wilcoxon signed-rank test for pairs with significant difference from Table 20.
Comparisonp-ValueConclusion
RF_GWO v/s KNN_GWO0.0001RF_GWO is better than KNN_GWO
RF_GWO v/s KNN_PSA0.0001RF_GWO is better than KNN_PSA
RF_GWO v/s KNN_WOA0.0001RF_GWO is better than KNN_WOA
RF_PSA v/s KNN_GWO0.0001RF_PSA is better than KNN_GWO
RF_PSA v/s KNN_PSA0.0001RF_PSA is better than KNN_PSA
RF_PSA v/s KNN_WOA0.0001RF_PSA is better than KNN_WOA
RF_PSO v/s KNN_GWO0.0001RF_PSO is better than KNN_GWO
RF_PSO v/s KNN_PSA0.0001RF_PSO is better than KNN_PSA
RF_PSO v/s KNN_WOA0.0001RF_PSO is better than KNN_WOA
RF_WOA v/s KNN_GWO0.0001RF_WOA is better than KNN_GWO
RF_WOA v/s KNN_PSA0.0001RF_WOA is better than KNN_PSA
RF_WOA v/s KNN_WOA0.0001RF_WOA is better than KNN_WOA
XGB_GWO v/s KNN_GWO0.0001XGB_GWO is better than KNN_GWO
XGB_GWO v/s KNN_PSA0.0001XGB_GWO is better than KNN_PSA
XGB_GWO v/s KNN_PSO0.0001XGB_GWO is better than KNN_PSO
XGB_GWO v/s KNN_WOA0.0001XGB_GWO is better than KNN_WOA
XGB_PSA v/s KNN_GWO0.0001XGB_PSA is better than KNN_GWO
XGB_PSA v/s KNN_PSA0.0001XGB_PSA is better than KNN_PSA
XGB_PSA v/s KNN_PSO0.0001XGB_PSA is better than KNN_PSO
XGB_PSA v/s KNN_WOA0.0001XGB_PSA is better than KNN_WOA
XGB_PSA v/s RF_PSA0.0001XGB_PSA is better than RF_PSA
XGB_PSA v/s RF_PSO0.0001XGB_PSA is better than RF_PSO
XGB_PSA v/s RF_WOA0.0001XGB_PSA is better than RF_WOA
XGB_PSO v/s KNN_GWO0.0001XGB_PSO is better than KNN_GWO
XGB_PSO v/s KNN_PSA0.0001XGB_PSO is better than KNN_PSA
XGB_PSO v/s KNN_PSO0.0001XGB_PSO is better than KNN_PSO
XGB_PSO v/s KNN_WOA0.0001XGB_PSO is better than KNN_WOA
XGB_PSO v/s RF_GWO0.0001XGB_PSO is better than RF_GWO
XGB_PSO v/s RF_PSA0.0001XGB_PSO is better than RF_PSA
XGB_PSO v/s RF_PSO0.0001XGB_PSO is better than RF_PSO
XGB_PSO v/s RF_WOA0.0001XGB_PSO is better than RF_WOA
XGB_WOA v/s KNN_GWO0.0001XGB_WOA is better than KNN_GWO
XGB_WOA v/s KNN_PSA0.0001XGB_WOA is better than KNN_PSA
XGB_WOA v/s KNN_PSO0.0001XGB_WOA is better than KNN_PSO
XGB_WOA v/s KNN_WOA0.0001XGB_WOA is better than KNN_WOA
XGB_WOA v/s RF_PSA0.0001XGB_WOA is better than RF_PSA
XGB_WOA v/s RF_PSO0.0001XGB_WOA is better than RF_PSO
XGB_WOA v/s RF_WOA0.0001XGB_WOA is better than RF_WOA
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cisternas-Caneo, F.; Santamera-Lastras, M.; Barrera-Garcia, J.; Crawford, B.; Soto, R.; Brante-Aguilera, C.; Garcés-Jiménez, A.; Rodriguez-Puyol, D.; Gómez-Pulido, J.M. Application of Metaheuristics for Optimizing Predictive Models in iHealth: A Case Study on Hypotension Prediction in Dialysis Patients. Biomimetics 2025, 10, 314. https://doi.org/10.3390/biomimetics10050314

AMA Style

Cisternas-Caneo F, Santamera-Lastras M, Barrera-Garcia J, Crawford B, Soto R, Brante-Aguilera C, Garcés-Jiménez A, Rodriguez-Puyol D, Gómez-Pulido JM. Application of Metaheuristics for Optimizing Predictive Models in iHealth: A Case Study on Hypotension Prediction in Dialysis Patients. Biomimetics. 2025; 10(5):314. https://doi.org/10.3390/biomimetics10050314

Chicago/Turabian Style

Cisternas-Caneo, Felipe, María Santamera-Lastras, José Barrera-Garcia, Broderick Crawford, Ricardo Soto, Cristóbal Brante-Aguilera, Alberto Garcés-Jiménez, Diego Rodriguez-Puyol, and José Manuel Gómez-Pulido. 2025. "Application of Metaheuristics for Optimizing Predictive Models in iHealth: A Case Study on Hypotension Prediction in Dialysis Patients" Biomimetics 10, no. 5: 314. https://doi.org/10.3390/biomimetics10050314

APA Style

Cisternas-Caneo, F., Santamera-Lastras, M., Barrera-Garcia, J., Crawford, B., Soto, R., Brante-Aguilera, C., Garcés-Jiménez, A., Rodriguez-Puyol, D., & Gómez-Pulido, J. M. (2025). Application of Metaheuristics for Optimizing Predictive Models in iHealth: A Case Study on Hypotension Prediction in Dialysis Patients. Biomimetics, 10(5), 314. https://doi.org/10.3390/biomimetics10050314

Article Metrics

Back to TopTop