Next Article in Journal
Advanced Sensors Technologies Applied in Mobile Robot
Next Article in Special Issue
Enzymatic Glucose Fiber Sensor for Glucose Concentration Measurement with a Heterodyne Interferometry
Previous Article in Journal
Multi-Target Tracking Based on a Combined Attention Mechanism and Occlusion Sensing in a Behavior-Analysis System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Federated Learning-Inspired Evolutionary Algorithm: Application to Glucose Prediction

1
ICAR-National Research Council of Italy, Via P. Castellino, 80131 Naples, Italy
2
Natural Computation Lab, DIEM, University of Salerno, Via Giovanni Paolo II 132, 84084 Fisciano, Italy
3
Department of Computer Science and Engineering, New Technologies for Information Society, University of West Bohemia, Technicka 18, 330 01 Pilsen, Czech Republic
4
Department of Computer Science and Engineering, University of West Bohemia, Technicka 18, 330 01 Pilsen, Czech Republic
5
Diabetology Center, First Department of Internal Medicine, University Hospital Pilsen, Alej Svobody 923/80, 323 00 Pilsen, Czech Republic
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(6), 2957; https://doi.org/10.3390/s23062957
Submission received: 2 February 2023 / Revised: 17 February 2023 / Accepted: 4 March 2023 / Published: 8 March 2023
(This article belongs to the Special Issue Advanced Sensing Technology in Blood Glucose Monitoring)

Abstract

:
In this paper, we propose an innovative Federated Learning-inspired evolutionary framework. Its main novelty is that this is the first time that an Evolutionary Algorithm is employed on its own to directly perform Federated Learning activity. A further novelty resides in the fact that, differently from the other Federated Learning frameworks in the literature, ours can efficiently deal at the same time with two relevant issues in Machine Learning, i.e., data privacy and interpretability of the solutions. Our framework consists of a master/slave approach in which each slave contains local data, protecting sensible private data, and exploits an evolutionary algorithm to generate prediction models. The master shares through the slaves the locally learned models that emerge on each slave. Sharing these local models results in global models. Being that data privacy and interpretability are very significant in the medical domain, the algorithm is tested to forecast future glucose values for diabetic patients by exploiting a Grammatical Evolution algorithm. The effectiveness of this knowledge-sharing process is assessed experimentally by comparing the proposed framework with another where no exchange of local models occurs. The results show that the performance of the proposed approach is better and demonstrate the validity of its sharing process for the emergence of local models for personal diabetes management, usable as efficient global models. When further subjects not involved in the learning process are considered, the models discovered by our framework show higher generalization capability than those achieved without knowledge sharing: the improvement provided by knowledge sharing is equal to about 3.03% for precision, 1.56% for recall, 3.17% for F 1 , and 1.56% for accuracy. Moreover, statistical analysis reveals the statistical superiority of model exchange with respect to the case of no exchange taking place.

1. Introduction

In Machine Learning (ML) [1], the problem of data privacy, i.e., the existence of data private to the owning subject, has become relevant in many application fields in these last years, as well shown in the recent survey in [2].
In ML, this data privacy issue was explicitly tackled for the first time in 2015 with the introduction of the concept of Federated Learning (FL) [3,4]; in the basic form of this approach, a server starts its execution by creating a random solution under the form of a model that is sent to a set of clients, one for each set of data that should be kept private. Learning on any local set of private data only takes place on the local client associated with those specific data, so it is never sent elsewhere. Learning on a node results in the model being modified locally to best adhere to the local data. The locally modified model is sent back to the server at the end of the local learning. Once the latter has received all the modified models, it aggregates them to create a new model that considers all the local learning. After the aggregation phase, this latter model is sent again to the clients, and the process continues until it reaches a termination criterion on the server. A good survey on recent advances in FL can be found in [5].
Not all the existing ML techniques can be used within FL, because only some can undergo a meaningful aggregation process. For example, there is no such aggregation process available for methods based on Deep Neural Networks (DNNs) [6], logistic regression [7] and Radial Basis Functions [8].
Another important issue relevant to ML is the interpretability of the proposed solutions, as well evidenced in [9]. This means that a proposed solution should be understandable by any user. This is clearly a problem with the recent and numerically well-performing ML techniques, which DNNs [10,11] are. The latter build an internal model that, although capable of excellent numerical performance, is unintelligible to the user, be it a physician or a patient, which is why they are called black boxes.
To somehow get rid of this problem, in 2017, Explainable Artificial Intelligence (XAI) (https://sites.google.com/view/fl-tutorial/?pli=1 accessed on 8 February 2023) [12] was introduced, which tries to endow DNNs with mechanisms allowing the creation of an external model that can somehow explain the behavior of the algorithm in making its decisions [13]. The problem with this is that there cannot be any guarantee that the external model is the same as the internal one over the data domain, which could lead to serious errors that could even be fatal in the medical domain.
This paper proposes a general ML framework that can satisfy the issues related to data privacy and interpretability. Namely, our approach is based on the use of Evolutionary Algorithms (EAs), a widely used class of ML methodologies [14,15,16], and consists of the use of a distributed version of an EA (dEA) [17,18].
The proposed methodology is close to the classical FL approach, yet it is simultaneously different from it. It is similar because it allows each client to work on local data only and because global knowledge of the problem is obtained by aggregating the different local knowledge. The difference with the classical FL is that this aggregation is performed implicitly rather than explicitly, as in the typical FL scheme. A thorough explanation of this difference is given later in the paper.
From an architectural viewpoint, the dEA framework we put forward contains both a master acting as the server and, thus, managing the algorithm, and a set of nodes, each of which represents a client and only contains local data to be kept private. Grammatical Evolution (GE) [19,20] is used as the specific EA: in it, each proposed solution is a model constituted by an expression linking (some of) the problem variables, so it represents explicit knowledge that users can immediately understand. Given this choice, we make use of a distributed GE scheme.
In our proposed framework, learning takes place locally on each node, on which good local knowledge specific to the local data is gained in terms of a local model. Moreover, at given times, these good local models are sent to the master. This latter evaluates the global quality of each of these models over all the data, thanks to the help of all the nodes, without any need to transmit data from any node; then, it sends to all of the nodes all these models. These arrived solutions enter the local learning process; in this way, the local knowledge in each node can be augmented thanks to that arriving from other nodes. As a consequence of this exchange of information, better global knowledge can be obtained.
This kind of implicit aggregation process ties local models into a global one. Said otherwise, during the execution of our algorithm, global knowledge emerges from the data contained in the various local sets without any need to physically exchange or make them visible. At the end of the execution, the model performing the best globally, i.e., over all the local sets of private data, is obtained as desired. Moreover, as a very interesting byproduct of our framework, apart from the global knowledge of the problem, for each set of private data, personalized knowledge is obtained that is specific to each of them.
The proposed framework does not deal with the security issue in the current implementation. For a thorough description of the problems of privacy guarantees for users and detection against possible attacks, interested readers can refer to [21,22].
The framework is applicable in different domains [23,24]: healthcare, bank loans, advertising, financial fraud, and insurance, among others. In this paper, we focus our attention on the medical field, where there is a high need for data privacy and interpretability of the solutions.
Regarding privacy, medical data are highly sensitive and strictly personal to the patient, so they should not be disclosed to anybody else, meaning both any other patient participating in the study and any person involved in the handling of the data or the learning process. In the European Union, this issue is regulated by the General Data Protection Regulation (GDPR) (2016/679 law) (https://eur-lex.europa.eu/eli/reg/2016/679/oj accessed on 3 February 2023) [25] that concerns data protection and privacy within the Union.
For interpretability, in medicine, a solution should be understandable by any subject participating in the study. This holds for a physician wishing to evaluate from a medical viewpoint the soundness and the interest of the knowledge proposed by the ML system, or a patient wishing to receive a diagnosis that is clear and well explains the reasons for that decision. This also follows the GDPR law, specifically Article 12, which states that the data controller gives information to the ‘data subject in a concise, transparent, intelligible and easily accessible form, using clear and plain language’. Moreover, Article 25 recognizes subjects’ right to contest any automated decision making that was solely algorithmic.
Within the medical field, we chose to take into account diabetes [26] disease, with specific reference to the prediction of future glucose values for subjects suffering from Type-1 diabetes mellitus (T1DM). Diabetes is a chronic disease, and its T1DM version is characterized by the fact that the subject’s pancreas produces practically no insulin, which calls for a life-lasting treatment consisting in the daily administration of amounts of insulin. In fact, if not treated, diabetes determines hyperglycemia, a condition of increased blood glucose values that with time may yield relevant damage to several parts of the body, among which are the eyes, kidneys, nerves, heart, lower limbs, and blood vessels [27].
As the data set to conduct our experiments, we avail ourselves of the well-known and publicly available Ohio T1DM data set [28]. In the experiments, rather than attempting to predict the exact future glucose values, as  would be the case in multivariable regression, we treat prediction as a classification problem. This is an approach already taken in the scientific literature through various methods [29]. To follow this way of operating, we divide the glucose range into seven intervals, and for each future value, we aim at predicting the interval it lies within. This is a good way to predict if a future glucose value will lie in high-risk intervals, such as those associated with very low or very high values. In this case, immediate recovery actions can be taken to eliminate or reduce risks to the subject’s health.
In clinical practice, Time-in-Range represents the time spent within a safe glucose-level range [30]. Within the safe range, the patient may avoid unnecessary actions to correct the blood glucose level, which may accidentally trigger an undesired outcome. In principle, the patient needs to know whether he/she is staying within the safe range or deviating from it. The proposed prediction addresses this need, while relieving the patient from the stress of operating with exact glucose levels, which may lead to diabetes burnout [31]. As a side effect, the resulting models can be simpler, thus reducing the computational complexity for lower-power devices and for possible cloud processing for thousands of patients. Moreover, the proposed method has the future potential to be applied as a watchdog over an insulin pump’s controller activity. As it is a prediction method, it could detect the controller’s failure to keep glucose levels in the safe range ahead of time.
As the outcome of our experiments, we expect to obtain an explicit global model able to perform generalization. This means that such a model should perform acceptably well on all the subjects involved in this learning process and on others not involved in creating the model. This would be highly important in real-world situations where we have to start monitoring diabetic subjects for which we do not have specific knowledge. We would need a general model to use on them to predict their future glucose values, and we could use the one obtained through our framework.
To evaluate the effectiveness of the mechanism of information exchange among nodes, our algorithm is experimentally compared against a distributed EA differing only in the absence of exchange. This comparison is effected both in terms of numerical performance achieved in the classification and from the statistical analysis perspective.
The rest of this paper is structured as follows. Section 2 presents a brief state-of-the-art review. Section 3 describes the proposed collaborative approach and the data set used. The experimental framework and the discussion about findings are reported in Section 4. In the same section, the results of the statistical analysis test, performed over the twelve subjects of the complete Ohio T1DM data set, are outlined. The last section exposes the conclusions and provides some indications on future work.

2. State of the Art

An important issue when dealing with ML applications is data privacy related to the protection of sensible personal information. This issue is increasing with the usage of online platforms collecting private data to provide services. A privacy-preservation framework must ensure high protection to let individuals share their information. FL represents the most employed technology to accomplish the privacy task [3,4,32]. This federated technique facilitates distributed collaborative learning by multiple clients under the coordination of a server. Data privacy is assured by training a prediction model through decentralized data, locally associated with different clients and not exchanged or transferred. Federated Learning is applied to support privacy-sensitive applications in several fields [24].
Another important issue of ML lies in its ability to discover underlying explanatory structures. The most performing techniques, i.e., deep learning neural networks, can be regarded as black boxes lacking an explicit knowledge representation. Utilizing black box learning models involves difficulty in understanding what model inputs drive the decisions (explainability) and, above all, prevents specialists from understanding the reason for a prediction (interpretability) [9,33]. The demand for transparent decisions pushes towards explainable and interpretable systems [34]. Explainable systems are black box learning models endowed with external XAI tools, without guarantee that these external tools allow capturing the internal model behavior. Interpretable models are models able to explain themselves by providing explicit models. From now on, the term interpretability is employed with the above meaning.
Both the above issues assume noticeable importance in the medical domain, e.g., diabetes management. Several techniques have been investigated to discover data-driven glucose forecasting models, ranging from approaches based on regression [35,36,37,38,39] to those that handle the prediction as a classification problem [29,40,41,42]. These techniques can be classified as explainable or interpretable based on the techniques employed for discovering the learning model.
Leaving aside the regression-based models, a brief literature survey on the state-of-the-art works on diabetes classification using data-driven ML models is conducted for the explainable and interpretable models described above. The review is related to recent articles that explore different techniques for dealing with glucose prediction formulated as a classification problem.
The first category concerns explainable techniques, most based on neural models, that exhibit outstanding performance at the expense of the difficulty of comprehending the aspects that can explain the decision, even when enriched with external XAI explanation tools [43,44,45,46,47,48,49,50,51,52,53,54,55]. As already illustrated in the introduction, the lack of explanation could yield the usage of these classification techniques problematic in the medical domain [56]. In fact, these learning models’ inner workings are too complicated to understand for physicians.
The second category includes interpretable models characterized by explicit prediction models. Most of these models rely on decision trees [57,58,59,60,61,62,63,64,65,66,67,68,69]. Although the methods based on these trees [70] could provide explicit knowledge, in many cases, it is challenging to linearize the resulting acyclic decision graphs into simple decision rules. Other attempts have been carried out to make predictions through classification rules based on if–then–else conditions induced by an evolutionary approach [71,72].
Independently of the belonging category, none of the above-examined approaches consider the problem of data privacy, which remains a critical concern when handling sensitive information such as diabetic data [73]. FL technology has been utilized in the medical domain to train a prediction model through decentralized data for dealing with different problems [74,75,76,77].
Only some recent papers contemplate the problem of implementing privacy-protected diabetes prediction systems relying on FL approaches and encryption with different training processes [78,79]. However, instead of employing data related to a single patient, the training concerns data collected in each hospital [78] or grouped by defining cohorts associated with diabetes-related complications [79]. Therefore, while ensuring data protection, these approaches do not permit the development of personalized models that are important from the point of view of precision medicine.
This review makes us confident that, at least in the recent scientific literature, data privacy has not been considered for tuning interpretable glucose forecasting models for diabetic patients. Indeed, most reviewed predictive models rely on centralized training data, or refer to decentralized training clients associated with data not referred to single patients, and thus are unable to allow personal disease treatment and prevention strategies. Table 1 summarizes the results of the review. As evinced from this table, the limitation of the current FL approaches is related to solution interpretability.
We aim to overcome this limitation by dealing with a data-privacy paradigm able to discover interpretable Machine Learning models for glucose prediction. This paradigm is based on some collaborative concepts inspired by FL. Specifically, this collaboration is pursued by training a Federated Learning-inspired global model, relying on a dEA that evolves multiple decentralized clients, each representing a single patient, holding local data samples without exchanging them. The collaborative training consists in sharing the model discovered by each local patient to be aggregated in a global model. The personalized models emerged at the end of evolution can be exploited for personal diabetes treatment or aggregated and used as global models.
Regarding interpretability, we concentrated on a grammar-based Evolutionary Algorithm to discover explicit classification models.
The following section illustrates the devised framework and its specific application to the glucose forecasting problem.

3. Methods and Materials

3.1. The Proposed Approach

This paper introduces a novel Federated Learning-inspired Evolutionary Algorithm (FLEA). The proposed methodology is a master/slave dEA [18,80] in which each slave runs a canonical sequential EA, and individuals, i.e., predictive models, can synchronously migrate between populations with a given frequency [81]. Specifically, each slave evolves a population of predictive models using learning data from a specific slave exclusively. In particular, the proposed method works as follows. During the evolution:
  • at specific instants of time (migration interval), the best model evolved so far on each slave is sent to the master node;
  • the master node returns the collected models to all the slaves, which evaluate them on the local data, thus preserving data privacy;
  • each slave uses the immigrant models within its population by replacing as many local individuals as possible with the lowest performance if better. In this way, the evolved models on each slave node receive information about the evolving models on the other slave nodes. Thanks to the mechanisms of selection, replacement, and genetic variation, the slave nodes can integrate the incoming information into their own population.
The above steps are graphically outlined in Figure 1.
At the end of evolution, each slave sends the best predictive model found to the master node that collects all of them. Then, the master node sends to each slave the list of all the best local models just received. Each slave evaluates such models on its local data and sends the list of their performance to the master node. This last step is particularly important when the proposed method is inserted in a system that continuously optimizes models on local data. In fact, when new local data are added to the system, the master node could provide the initial population of a new slave node with the predictive models coming from the other individuals, thus allowing a boost in the search for the specific local model on these new data.
Algorithms 1 and 2 report the pseudocode for the master and slave, respectively. It is worth noting that the proposed methodology is very close to an FL approach [5]. The differences lie in the fact that, in the federated approach, the integration of patterns is explicitly performed by the master node, and the communication of learned patterns is direct. In contrast, in the proposed approach, the integration of patterns is implicitly performed by the slave nodes, in that it is demanded of the mechanisms of selection and genetic variation, i. e., crossover [82], that, eventually, perform the integration. The communication is indirectly effected through the master node.

3.2. The Data Set

The FLEA framework is investigated to forecast future glycemic trends for T1DM patients. The experiments are conducted on the Ohio T1DM data set, released in 2020 [28], which gathers data of T1DM patients. This data set was collected at the Ohio University and contains data related to twelve subjects, each of whom was monitored with a Continuous Glucose Monitoring (CGM) system for about eight weeks while being on insulin pump therapy. Given the availability of data in the dataset related to, among others, measured subcutaneous glucose, injected insulin (basal plus boluses), and carbohydrate ingested during the day (time and estimated size of all meals), future glucose values can be predicted on the basis of the sets of current and recent values available for these three parameters. The sampling interval of glucose measurements achieved by the CGM system is equal to Δ t = 5 min. Each slave only contains the private data associated with a single patient.
To allow supervised learning, the data series of each patient are partitioned into training and testing sets, respectively, used to extract the model during the learning phase and assess its quality over unseen samples. The supervised learning phase is carried out on the six patients related to the data set released in 2018 [83], while a successive validation phase is performed on the testing sets of the remaining six patients added to the data set in 2020. The information about the number of training and testing samples for each patient is reported in [28].
Algorithm 1 Pseudocode of FLEA on the master node
  • set global stopping condition to FALSE
  • while not global stopping condition do
  •     for each slave do
  •         receive the best local model
  •     end for
  •     for each slave do
  •         send the list of the received best local models
  •     end for
  •     for each slave do
  •         receive the local stopping condition
  •     end for
  •     if all local stopping conditions are TRUE then
  •         set global stopping condition to TRUE
  •     end if
  • end while
  • for each slave do
  •     receive the final best local model
  • end for
  • for each slave do
  •     send the list of the final best local models
  •     receive the list of the fitness of the final best local models
  • end for
Algorithm 2 Pseudocode of FLEA on a slave node
  • randomly generate an initial population of n models
  • evaluate the fitness of each model
  • set the maximum number of generations g max
  • set current generation g equal to 0
  • set local stop condition to FALSE
  • while  g < g max  do
  •     if migration time is TRUE then
  •         send the best local model to the master node
  •         receive from the master node the best local models of the other slaves (immigrants)
  •         evaluate on the local data the best local models received
  •         replace the worst local models with the immigrants if better
  •     end if
  •     generate the new population according to the chosen EA
  •     evaluate the fitness of the models in the new population
  •     if some local conditions hold then
  •         set local stop condition to TRUE
  •     end if
  •     send to the master node the local stop condition
  •      g = g + 1
  • end while
  • send the final best local model to the master node
  • receive from the master node the list of final best local models of the other slaves
  • evaluate the fitness of each final best local model on local data
  • send to the master node the list of the evaluated fitness of the final best local models

Data Preprocessing

As regards the data preprocessing, we performed the following arrangement:
  • Samples with missing glucose readings in training and testing sets are thrown away to avoid that the predicting model can be the result of artificial observations;
  • Insulin and carbohydrates data were aligned to the closest CGM glucose reading time;
  • No outlier detection and no data normalization were effected.
It is pointed out that the discrete signals of administered insulin, i. e., insulin boluses plus insulin basal, and the assumed carbohydrates are to convert into continuous signals to estimate their impact on the glucose values over time. The Hovorka model [84], simulating the absorption rate of the injected insulin through a two-compartment chain, is employed for preprocessing the injected insulin boluses. This model permits adding the signal delineating the absorption rate of the boluses to the signal representing the absorption rate of subcutaneously administered long-acting insulin.
Let us assume that the glucose level G ( t ) , the injected insulin U ( t ) , and the consumed carbohydrates D g ( t ) are available. The model for insulin absorption is
d S 1 d t = U ( t ) S 1 t max I
d S 2 d t = S 1 S 2 t max I
in which S 1 and S 2 are the two compartments making up the chain for modeling the absorption of subcutaneously infused short-acting insulin, U ( t ) [mU min 1 ] is the amount of injected insulin, t max I = 55 [min] is the constant indicating the time-to-maximum insulin absorption, and S 1 ( t ) [mU] and S 2 ( t ) [mU] are the amounts of insulin in the two compartments. Then, the plasma insulin concentration I [mU L 1 ] is described as
d I d t = S 2 V I · t max I k e · I
where k e = 0.138 [min 1 ] is the fractional elimination rate of the insulin from plasma and V I = 0.12 [L kg 1 ] is the insulin distribution volume. The constant values are derived from Hovorka’s model [85].
Regarding the carbohydrate intake, in the presence of a meal, the gut absorption rate is modeled in accord with [84] as
C ( t ) = D g · A g · t · e t / t max t max 2
where t max = 40 [min] is the time-of-maximum appearance rate of glucose in the accessible compartment, D g is the amount of digested carbohydrates, and A g = 0.8 is the carbohydrates bioavailability [86]. This function rapidly increases after the meal and then lowers to 0 in 2–3 h. Outside such a period, the values of missing carbohydrate are filled with zeroes.
At the end of the preprocessing, by integrating Equation (3) and exploiting Equation (4), we have two signals, discretized every Δ t minutes, for the absorbed insulin and carbohydrates, i.e., I ( t ) and C ( t ) , respectively. More specifically, when at time t there is an insulin release or carbohydrate intake event, their absorbed quantities are propagated over time through Equations (3) and (4) from the current time t ahead and, if needed, summed to the residual quantity evaluated by the compartment model in the past. Typically, the variation range of G ( t ) is about [ 2 ÷ 25 ] [ mmol L 1 ] , of I ( t ) is about [ 0 ÷ 10 ] [ mU L 1 ] , and of C ( t ) is [ 0 ÷ 3 ] [ g ] .

3.3. FLEA to Forecast Future Glycemic Trends

Forecasting future glycemic trends for T1DM patients can be regarded as a multivariate time series regression problem, falling within the learning of data-driven models exploiting information extracted from CGM systems. To apply the FLEA framework to the above problem, we exploit the capability of the Grammatical Evolution [82] to automatically evolve interpretable regression models.
Moreover, differently from other EAs, GE explicitly makes use of the context-free grammars that are able to design a specific form for the evolved models.
To do this, we need to define a suitable grammar and a fitness function. Moreover, rather than attempting to predict the exact future glucose values, we transform the time series regression into a classification problem.

3.3.1. The Grammar

The context-free grammar in Figure 2 depicts the syntax of the GE-based expressions evolved on each slave, where g l u c represents the glucose levels in the past, i n s and c h o indicate insulin and carbohydrates to be absorbed in the future, and d g is the difference between the current and past glucose levels, respectively. In our grammar, the protected psqrt and plog functions return the square root of the absolute value of the argument, the logarithm of the summation of 1, and the absolute value of the argument, respectively, while aq stands for the protected analytic quotient operator [87]. Table 2 outlines the protected functions utilized in the grammar.
By considering the values of G ( t ) every Δ t minutes in a time window of k Δ t minutes before the current instant t, as well as the values of I ( t ) and C ( t ) every Δ t minutes in a time window of h Δ t minutes after the current instant t, we search for an explicit regression model to predict the future glucose value G ^ ( t + h Δ t ) at a forecasting horizon h Δ t :
G ^ ( t + h Δ t ) = ( Γ G ( t ) , G ( t Δ t ) , , G ( t k Δ t ) Θ ( I ( t ) , I ( t + Δ t ) , , I ( t + h Δ t ) ) + Ω C ( t ) , C ( t + Δ t ) , , C ( t + h Δ t ) ) Φ d G ( t , t Δ t ) , , d G ( t , t k Δ t )
where the symbol ◊ represents an algebraic operation in the set: { + , , · } , and Γ , Θ , Ω , and Φ are expressions on G, I, C, and d G , respectively.

3.3.2. From Regression to Classification

Glucose prediction is typically performed through multiseries regression to predict glucose values as accurately as possible. Nonetheless, in the literature, the use of classification to forecast glucose ranges rather than exact values is becoming more and more popular [47,48,49,88]. This is of great help when high-risk situations such as hyperglycemic events or, even more crucially, hypoglycemic ones should be forecasted with good advance. In these cases, it is more important to predict the occurrence of such an event than the precise glucose values.
To perform classification, the continuous glucose values were mapped into seven intervals, leading to a seven-class problem. More precisely, two classes make reference to hypoglycemia, three relate to euglycemia (normal values), and two refer to hyperglycemia.
The decision to consider two hypo- and two hyperclasses is based on the outcome of the international consensus held in 2017 and reported in [89]. Following that consensus, we used the same bounds as in that document; the corresponding bounds are reported in Table 3.
As concerns the euglycemic range, unlike [89], we prefer instead to consider a division into three classes, as reported in Table 3. The rationale for this is that if we had just one normal/target range, then we would not be able to track possibly dangerous, out-of-target-range deviating glucose development: we would directly pass from a series of normal values to the occurrence of a hypoglycemic event, without any warning. Instead, by using three classes, the middle one being larger and the two border zones towards hypo- and hyper-being ‘thin’, we would obtain warnings before a hypo- or a hyperglycemic event took place. In fact, for hypoglycemia, we would have a series of normal values, followed by a (series of) normal-closing-to-hypovalues, followed by hypovalues.
In Table 3, for each class, the ID we assigned to it is displayed, the corresponding glucose value range is shown both in mmol/L and in mg/dL, and the action(s) required during the monitoring are reported.
In this way, the problem is transformed into a classification task, and the aim is to predict the class of any glucose value in the future, starting from the available values for glucose, absorbed insulin, and carbohydrates, as expressed in Equation (5).
Figure 3 shows, for the testing set of each subject, the transformation of the continuous glucose signal into the corresponding set of items for the seven-class classification task considered in this paper.
Table 4 reports the number of samples in the training (Tr) and testing (Ts) sets for each patient and class. From the table, it can be easily seen that, for all six subjects, the three classes related to normal values and the two for the hyperglycemic values are much more populated than the two corresponding to ’very low’ and ’low’ glucose values. The latter two often only contain few values, and it is worth noting that for subjects 563, 570, and 588, the ’very low’ class is even empty. This means that all the six data sets are highly unbalanced, which is a complication in classification [90].

3.3.3. Fitness Function

To evaluate the quality of any solution proposed, a suitable fitness function should make reference to the specific metrics typically used for this kind of problem, as, e.g., accuracy, sensitivity, specificity, area under the ROC curve, F 1 score, Matthews correlation coefficient, and so on.
We decided to use the F1 score, since the data sets corresponding to each of the six subjects investigated here are highly unbalanced; especially, their classes 0 (very low glucose values), 1 (low glucose values), and, for some subjects, 6 (very high glucose values) contain very few items with respect to the other four classes.
It is well known that, whenever a data set is highly unbalanced in terms of number of items contained in the different classes, as it is the case here, metrics such as accuracy, sensitivity, and specificity are not suitable: good performance on the most populated class(es) could lead to numerically good results without actual learning taking place on the least populated class(es). This could mean that every time an item belonging to one of the minority classes has to be classified, the algorithm could wrongly assign it to one of the majority class(es).
For unbalanced data sets, instead, metrics such as F 1 score or Matthews correlation coefficient can more effectively take this problem into account.
For a two-class problem in which we have a positive class and a negative one, F 1 score is computed as
F 1 = t p t p + 0.5 · ( f p + f n )
where:
  • tp: the number of true positives, i.e., the items in the positive class that are correctly assigned to that class;
  • fp: the number of false positives, i.e., the items in the negative class that are incorrectly assigned to the positive class;
  • fn: the number of false negatives, i.e., the items in the positive class that are incorrectly assigned to the negative class.
When, instead, there are more than two classes, as in this case, the definition of F 1 score can be generalized in several ways. Within this paper, we used the method of weighted averaging, This means that the resulting F 1 score value accounts for the contribution of the F 1 score computed for each class and weighted by the number of items of that given class. In formula:
F 1 = n = 1 n c p i · F 1 i n c
where nc is the number of classes in the data set, p i is the percentage of items in the i-th class, and F 1 i is the F 1 score value computed on the i-th class.
The admissible range for F 1 score is [0.0–1.0], and higher values represent better classifications. By choosing this metric, the classification problem becomes a maximization one.

4. Experimental Results

4.1. Experimental Framework Setting

Our approach was implemented by exploiting PonyGE2, a freely downloadable and patent-free GE implementation in Python [82]. PonyGE2 has a number of GE-specific parameters to set, the meaning of which can be found in [82]. After a preliminary tuning, the parameters used for all the experiments were set as follows: population size and maximum generations equal to 200 and 500, respectively; codon size equal to 100,000, tournament selection with size 4, mutation probability equal to 10%, one-point crossover probability equal to 90%, int flip per codon mutation with one mutation event, and Position Independent Grow method for the individual initialization. For the slaves, a single local stopping conditionwas considered and set to the fulfillment of the maximum number of generations.
We set the communication between the master and the slaves to take place every 100 generations. This value was chosen because of two motivations.
The first reason comes from the field of dEAs: it is known that any subpopulation should not receive immigrating individuals too frequently, because this would perturb the local evolution at each communication time. The local search must be given sufficient time to suitably integrate the arrived individuals into the local subpopulation, so as to exploit their good features.
The second reason is related to the FL principles themselves in terms of security: an FL algorithm should involve the least possible amount of information being transmitted, because any possible communication could be attacked, possibly resulting in a subject’s relevant information being disclosed or in an injection of fake data by attackers, which could lead to totally wrong learning.
Hence, based on our experience, we feel the value of 100 implies the lowest amount of communication that allows improvement in the learning process while, at the same time, not excessively exposing the process to external attacks. In fact, this value of 100, together with the number of generations being set to 500, means that, during the whole execution, only five communication phases between the master and the slaves take place.
The forecasting horizon is h Δ ( t ) = 30 min, because the forecasting accuracy becomes worse and less reliable as the prediction horizon augments [91,92]. A horizon longer than 30 min, e.g., 2 or 4 h, is only practical for time spans that refer to almost steady-state situations as nocturnal predictions when sleeping. Any external event can cause a substantial and unpredictable glucose-level variation during these long intervals. The considered past time window is k Δ ( t ) = 60 min for the historical samples leveraged for the forecasting. The time span for the historical data is chosen considering that 30-min data in the past are enough to perform an effective prediction [93]. Given that the values are taken at 5-min intervals, the values for k and h are equal to 12 and 6, respectively. This implies that both in the grammar and in Equation (5), at each time t, we consider twelve glucose values in the past and six insulin and carbohydrate values in the future.
To assess the effectiveness of the proposed approach, we conducted two experiments:
  • In the first experiment, we used a non-FL approach consisting of FLEA with no communication between the master and the slave nodes during the evolution. In other words, we executed a separate optimization for all the patients, thus obtaining for each of them a personalized model. At the end of the executions, we collected all the models and selected among all of them the model with the best average performance on all the patients;
  • In the second experiment, we used FLEA. The average outcomes for each run were evaluated at the end of the evolution by considering all the best local models received by the master node from all the slaves and measuring their performance on all the patients to evaluate how they perform on average when adopted as global models.
For each patient, indicated with the identifier ID, twenty runs were carried out to reduce the randomness in the GE algorithm initialization. The evaluation is performed on all instances for which a glucose measurement is available over the prediction period.

4.2. Findings and Discussion

Table 5 and Table 6 report the F 1 score of the best models on each slave of both experiments, and the last row and column show their averages and standard deviations. By looking at Table 5, it can be evidenced that, when adopted as a global model, each personalized model exhibits F 1 score values that are quite different on a specific patient (rows), and the same is true also when the performance of the models is measured on a specific patient (columns). On the contrary, inspecting the results of Table 6 related to our approach, the scenario changes. Independently of the adopted global model, the average F 1 score of the evolved models is always better than the previous case, except for subject 570, and very close to each other (rows). A similar consideration also holds for all the models on a specific patient (columns), thus evidencing that the proposed approach can evolve generalized models exhibiting better performance. Moreover, if we look at the best models evolved on local data (diagonals in the tables), it is evident that communication helps improve the performance on local nodes too.
Analogous reflections can be made by comparing the corresponding panes of Figure 4 and Figure 5 reporting the confusion matrices on the testing set for both experiments. This comparison documents that FLEA frequently increases the number of items correctly assigned to the classes. This can be verified by looking at the cells along the diagonals, which in most cases contain higher values.
Once we found the two global models proposed by the two different approaches, we wished to investigate their generalization capability to ascertain whether or not they have similar performance for this issue. To this aim, we executed them on six more patients in the Ohio data set. Table 7 reports the corresponding results in terms of F 1 score for the two models over the testing sets of these six additional patients. The last column in the table reports the average and the standard deviation of these F 1 values.
The table reports that, over all six subjects, the model obtained by FLEA always performs better than that achieved by the non-FL approach. This is very important, because it allows concluding that the model provided by our FLEA framework is more general; therefore, it can be used for new subjects not participating in the learning process.
The confusion matrices corresponding to these experiments are reported in Figure 6 and Figure 7 for the two algorithms without and with communication, respectively.
The comparison between the corresponding panes of the two figures evidences that the items are more frequently assigned correctly when FLEA is considered: the cells along the diagonals contain higher values when communication takes place. This is of crucial importance for the two classes corresponding to hypoglycemic events. In fact, the latter are high-risk situations, therefore correctly predicting them well in advance is a major issue for subjects’ health. Moreover, in this case, the results confirm that the model produced by the FLEA algorithm is more general and can be useful when new subjects are to be monitored from scratch.
Table 8 reports a comprehensive view of the six subjects’ numerical scores. This table further confirms that the model achieved by FLEA, on average, performs better than that obtained by the non-FL approach. Specifically, the former shows an improvement of about 3.03% for precision, 1.56% for recall, 3.17% for F 1 , and 1.56% for accuracy.

4.3. Statistical Analysis

A statistical analysis test was executed to assess whether or not the best model proposed by the FLEA algorithm performs better than that obtained by the non-FL one over the complete set of the twelve subjects making up the Ohio T1DM data set. This analysis was carried out on the online web platform ‘Statistical Tests for Algorithms Comparison’ [94] (STAC) (https://tec.citius.usc.es/stac/ accessed on 15 January 2023). From among the several statistical tests available, the Quade test was chosen because it considers the higher difficulty of some problems and the larger differences that may be shown by the various algorithms over them; hence, the Quade test is different from the Friedman and Aligned Friedman ones, in which all the problems are considered to be of equal importance. For more details on the statistical analysis shown here in general, and on the Quade test in particular, interested readers can refer to the widely cited paper by [95].
Before running a statistical test, a null hypothesis H 0 must be chosen; we set it as the fact that the two models proposed by the two algorithms are statistically equivalent. Moreover, a significance level α must be chosen; we set its value as 0.05, which means that, if the null hypothesis is rejected by the test, there is a 5 % probability of incorrectly rejecting it.
Table 9 reports the results of this test: the first column contains the algorithms compared, and the second the corresponding rank value; better algorithms are characterized by lower ranking values.
The table shows that FLEA performs better than non-FL on this test. Yet, as the computed p-value is 0.16680, which is higher than 0.05, this test cannot exclude the statistical equivalence between the two algorithms.
To further investigate this issue, we must make reference to post hoc procedures, also described in [95]. Table 10 reports the results, in terms of adjusted p-values, for the complete set of post hoc procedures available in STAC, i.e., Bonferroni–Dunn, Holm, Hochberg, Finner, and Li. The FLEA algorithm was chosen as the control method because it is the algorithm with the lowest ranking value in the Quade test.
To understand the results in the table, each post hoc procedure returns an adjusted p-value for non-FL. For all the procedures, this value is lower than the significance level 0.05. This means that the null hypothesis H 0 of equivalence is rejected by all of them. Hence, FLEA is statistically better than non-FL.

5. Conclusions and Future Work

To suitably deal with the issues of data privacy and interpretability, in this paper, we proposed a distributed framework that constitutes an innovative approach to Federated Learning. The framework consists in a master process and a set of slaves: On each of the latter, a Grammatical Evolution algorithm is run that only learns from local private data and generates explicit models that humans can interpret. At given times, a migration process occurs between the slaves through the master, the result of which is that each slave receives the local best models found by the other slaves. This results in an exchange of knowledge between the different nodes merging locally gained knowledge. This process of knowledge exchange allows obtaining local models that can work effectively over all the local sets of private data, i.e., they can be used as global models.
As data privacy and interpretability are highly relevant to the medical field, we applied this framework to a medical problem, i.e., that of the prediction of future glucose values for T1DM patients. This problem is transformed into a seven-class classification task. To assess the importance of this process of knowledge exchange, the framework was experimentally compared with another that only differs in the fact that knowledge exchange does not take place between the slaves.
The results show the importance of this exchange process to create a set of personalized models, each of which can be used as the global model. Moreover, the model obtained by our federated approach showed higher generalization capability than that achieved by the non-FL approach when the two were applied to the data of subjects who did not participate in the learning process. A statistical analysis evidenced the superiority of the FLEA algorithm.
In our future work, the results and behavior shown by our framework on this specific problem must be further investigated on other data sets from the medical domain in which data privacy and interpretable solutions are hard constraints.
Another important step to take in our future work is to perform an experimental investigation to evaluate if an optimal frequency for information exchange exists that allows improving the results without causing too many risks in terms of security for the whole framework. This analysis can be crucial considering that it is well in the field of dEAs that the numerical quality of the results may depend, even more highly, on communication frequency. This could lead to improvement in the results provided by our approach, which could, in this way, perform much better than the model without communication.
Regarding security, it should be evidenced that, currently, the framework we proposed does not deal with the data security problem during the phases of information exchange. As we exploit a distributed approach to evolve a global model, it is highly appropriate to discuss data transmission security. The proposed approach does not propagate patient data (e.g., his/her measured glucose) but rather a compression of his/her metabolic responses in the form of a best-suited model at the given time. Solely, this does not raise a concern; nonetheless, in connection with the known treatment setup, it may give an opportunity to use such representation to build a targeted attack. To avoid an eavesdropping attack, an encrypted communication link between the master and slave nodes has to be established.
Data transmission security is not a concern of this paper and will be a subject of future work. In our future work, we will have to suitably address this problem to define a complete Federated Learning framework that could help in real-world medical trials where data privacy must be guaranteed.

Author Contributions

Conceptualization, A.D.C., I.D.F., U.S. and T.K.; methodology, A.D.C., I.D.F., M.U. and U.S.; software, A.D.C. and U.S.; validation T.K., M.K., M.U. and E.T.; formal analysis T.K., U.S., M.K. and E.T.; writing—original draft preparation, I.D.F., A.D.C. and E.T.; writing—review and editing I.D.F., A.D.C., T.K., M.K, M.U., U.S. and E.T.; visualization, A.D.C., I.D.F. and U.S. All authors have read and agreed to the published version of the manuscript.

Funding

This publication was partially supported by University of West Bohemia specific research project SGS-2022-015. We acknowledge financial support from: PNRR MUR project PE0000013-FAIR.

Informed Consent Statement

Patient consent was waived due to the Data Use Agreement with the Ohio University, which collected and apriori anonymized the data with NIH 1R21EB022356 grant support.

Data Availability Statement

The authors do not have permission to share the data.

Conflicts of Interest

No conflict of interest declared.

Abbreviations

The following abbreviations are used in this manuscript:
CGMContinuous Glucose Monitoring
DNNDeep Neural Network
EAEvolutionary Algorithms
FLFederated Learning
FLEAFederated Learning-inspired Evolutionary Algorithm
GDPRGeneral Data Protection Regulation
GEGrammatical Evolution
MLMachine Learning
T1DMType 1 Diabetes Mellitus
XAIExplainable Artificial Intelligence

References

  1. Mitchell, T.M. Machine Learning; McGraw-Hill: New York, NY, USA, 1997; Volume 1. [Google Scholar]
  2. Liu, B.; Ding, M.; Shaham, S.; Rahayu, W.; Farokhi, F.; Lin, Z. When machine learning meets privacy: A survey and outlook. ACM Comput. Surv. (CSUR) 2021, 54, 1–36. [Google Scholar] [CrossRef]
  3. Konečnỳ, J.; McMahan, B.; Ramage, D. Federated optimization: Distributed optimization beyond the datacenter. arXiv 2015, arXiv:1511.03575. [Google Scholar]
  4. Konečnỳ, J.; McMahan, H.B.; Ramage, D.; Richtárik, P. Federated optimization: Distributed machine learning for on-device intelligence. arXiv 2016, arXiv:1610.02527. [Google Scholar]
  5. Li, Q.; Wen, Z.; Wu, Z.; Hu, S.; Wang, N.; Li, Y.; Liu, X.; He, B. A survey on federated learning systems: Vision, hype and reality for data privacy and protection. IEEE Trans. Knowl. Data Eng. 2023, 35, 3347–3366. [Google Scholar] [CrossRef]
  6. McMahan, H.B.; Moore, E.; Ramage, D.; Agüera y Arcas, B. Federated learning of deep networks using model averaging. arXiv 2016, arXiv:1602.05629. [Google Scholar]
  7. Yang, Q.; Liu, Y.; Chen, T.; Tong, Y. Federated machine learning: Concept and applications. ACM Trans. Intell. Syst. Technol. (TIST) 2019, 10, 1–19. [Google Scholar] [CrossRef]
  8. Xu, J.; Jin, Y.; Du, W.; Gu, S. A federated data-driven evolutionary algorithm. Knowl.-Based Syst. 2021, 233, 107532. [Google Scholar] [CrossRef]
  9. Rudin, C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 2019, 1, 206–215. [Google Scholar] [CrossRef] [Green Version]
  10. Bengio, Y. Learning deep architectures for AI. Found. Trends® Mach. Learn. 2009, 2, 1–127. [Google Scholar] [CrossRef]
  11. Schmidhuber, J. Deep learning in neural networks: An overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef] [Green Version]
  12. Gunning, D. Explainable artificial intelligence (xai). Def. Adv. Res. Proj. Agency (DARPA) Nd Web 2017, 2, 1. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Montavon, G.; Samek, W.; Müller, K.R. Methods for interpreting and understanding deep neural networks. Digit. Signal Process. 2018, 73, 1–15. [Google Scholar] [CrossRef]
  14. Back, T. Evolutionary Algorithms in Theory and Practice: Evolution Strategies, Evolutionary Programming, Genetic Algorithms; Oxford University Press: Oxford, UK, 1996. [Google Scholar]
  15. Back, T.; Fogel, D.B.; Michalewicz, Z. Handbook of Evolutionary Computation; IOP Publishing Ltd.: Bristol, UK, 1997. [Google Scholar]
  16. Whitley, D. An overview of evolutionary algorithms: Practical issues and common pitfalls. Inf. Softw. Technol. 2001, 43, 817–831. [Google Scholar] [CrossRef]
  17. Tomassini, M. Parallel and distributed evolutionary algorithms: A review. In Evolutionary Algorithms in Engineering and Computer Science; John Wiley & Sons: Chichester, UK, 1999; pp. 113–133. [Google Scholar]
  18. Gong, Y.J.; Chen, W.N.; Zhan, Z.H.; Zhang, J.; Li, Y.; Zhang, Q.; Li, J.J. Distributed evolutionary algorithms and their models: A survey of the state-of-the-art. Appl. Soft Comput. 2015, 34, 286–300. [Google Scholar] [CrossRef] [Green Version]
  19. Ryan, C.; Collins, J.J.; O’Neill, M. Grammatical evolution: Evolving programs for an arbitrary language. In Proceedings of the European Conference on Genetic Programming, Paris, France, 14–15 April 1998; pp. 83–96. [Google Scholar]
  20. O’Neill, M.; Ryan, C. Grammatical evolution. IEEE Trans. Evol. Comput. 2001, 5, 349–358. [Google Scholar] [CrossRef] [Green Version]
  21. Blanco-Justicia, A.; Domingo-Ferrer, J.; Martínez, S.; Sánchez, D.; Flanagan, A.; Eeik Tan, K. Achieving security and privacy in federated learning systems: Survey, research challenges and future directions. Eng. Appl. Artif. Intell. 2021, 106, 104468. [Google Scholar] [CrossRef]
  22. Mothukuri, V.; Parizi, R.M.; Pouriyeh, S.; Huang, Y.; Dehghantanha, A.; Srivastava, G. A survey on security and privacy of federated learning. Future Gener. Comput. Syst. 2021, 115, 619–640. [Google Scholar] [CrossRef]
  23. Li, L.; Fan, Y.; Tse, M.; Lin, K.Y. A review of applications in federated learning. Comput. Ind. Eng. 2020, 149, 106854. [Google Scholar] [CrossRef]
  24. Banabilah, S.; Aloqaily, M.; Alsayed, E.; Malik, N.; Jararweh, Y. Federated learning review: Fundamentals, enabling technologies, and future applications. Inf. Process. Manag. 2022, 59, 103061. [Google Scholar] [CrossRef]
  25. Regulation, P. General data protection regulation. Intouch 2018, 25, 1–5. [Google Scholar]
  26. Sun, H.; Saeedi, P.; Karuranga, S.; Pinkepank, M.; Ogurtsova, K.; Duncan, B.B.; Stein, C.; Basit, A.; Chan, J.C.; Mbanya, J.C.; et al. IDF Diabetes Atlas: Global, regional and country-level diabetes prevalence estimates for 2021 and projections for 2045. Diabetes Res. Clin. Pract. 2022, 183, 109119. [Google Scholar] [CrossRef]
  27. Papatheodorou, K.; Banach, M.; Bekiari, E.; Rizzo, M.; Edmonds, M. Complications of diabetes 2017. J. Diabetes Res. 2018, 2018, 3086167. [Google Scholar] [CrossRef]
  28. Marling, C.; Bunescu, R. The OhioT1DM dataset for blood glucose level prediction: Update 2020. NIH Public Access 2020, 2675, 71. [Google Scholar]
  29. Woldaregay, A.; Arsand, E.; Botsis, T.; Albers, D.; Mamykina, L.; Hartvigsen, G. Data-driven blood glucose pattern classification and anomalies detection: Machine-learning applications in type 1 diabetes. J. Med. Internet Res. 2019, 21, e11030. [Google Scholar] [CrossRef] [Green Version]
  30. Battelino, T.; Danne, T.; Bergenstal, R.M.; Amiel, S.A.; Beck, R.; Biester, T.; Bosi, E.; Buckingham, B.A.; Cefalu, W.T.; Close, K.L.; et al. Clinical targets for continuous glucose monitoring data interpretation: Recommendations from the international consensus on time in range. Diabetes Care 2019, 42, 1593–1603. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Abdoli, S.; Hessler, D.; Smither, B.; Miller-Bains, K.; Burr, E.M.; Stuckey, H.L. New insights into diabetes burnout and its distinction from diabetes distress and depressive symptoms: A qualitative study. Diabetes Res. Clin. Pract. 2020, 169, 108446. [Google Scholar] [CrossRef]
  32. Rodríguez-Barroso, N.; Jiménez-López, D.; Luzón, M.V.; Herrera, F.; Martínez-Cámara, E. Survey on federated learning threats: Concepts, taxonomy on attacks and defences, experimental study and challenges. Inf. Fusion 2023, 90, 148–173. [Google Scholar] [CrossRef]
  33. Rudin, C.; Chen, C.; Chen, Z.; Huang, H.; Semenova, L.; Zhong, C. Interpretable machine learning: Fundamental principles and 10 grand challenges. Stat. Surv. 2022, 16, 1–85. [Google Scholar] [CrossRef]
  34. Mengnan, D.; Ninghao, L.; Xia, H. Techniques for interpretable machine learning. Commun. ACM 2020, 63, 68–77. [Google Scholar]
  35. De Falco, I.; Della Cioppa, A.; Koutny, T.; Krcma, M.; Scafuri, U.; Tarantino, E. Genetic programming-based induction of a glucose-dynamics model for telemedicine. J. Netw. Comput. Appl. 2018, 119, 1–13. [Google Scholar] [CrossRef]
  36. De Falco, I.; Della Cioppa, A.; Giugliano, A.; Marcelli, A.; Koutny, T.; Krcma, M.; Scafuri, U.; Tarantino, E. A genetic programming-based regression for extrapolating a blood glucose-dynamics model from interstitial glucose measurements and their first derivatives. Appl. Soft Comput. 2019, 77, 316–328. [Google Scholar] [CrossRef]
  37. Tyler, N.S.; Jacobs, P.G. Artificial intelligence in decision support systems for type 1 diabetes. Sensors 2020, 20, 3214. [Google Scholar] [CrossRef]
  38. Felizardo, V.; Garcia, N.M.; Pombo, N.; Megdiche, I. Data-based algorithms and models using diabetics real data for blood glucose and hypoglycaemia prediction—A systematic literature review. Artif. Intell. Med. 2021, 118, 1–15. [Google Scholar] [CrossRef]
  39. Della Cioppa, A.; De Falco, I.; Koutny, T.; Ubl, M.; Scafuri, U.; Tarantino, E. Reducing high-risk glucose forecasting errors by evolving interpretable models for type 1 diabetes. Appl. Soft Comput. 2023, 134, 110012. [Google Scholar] [CrossRef]
  40. Maniruzzaman, M.; Rahman, M.; Ahammed, B. Classification and prediction of diabetes disease using machine learning paradigm. Health Inf. Sci. Syst. 2020, 8, 1–14. [Google Scholar] [CrossRef]
  41. Sharma, T.; Shah, M. A comprehensive review of machine learning techniques on diabetes detection. Vis. Comput. Ind. Biomed. Art 2021, 4, 30. [Google Scholar] [CrossRef]
  42. Rastogi, R.; Bansal, M. Diabetes prediction model using data mining techniques. Meas. Sensors 2023, 25, 100605. [Google Scholar] [CrossRef]
  43. Cuesta-Frau, D.; Miró–Martínez, P.; Oltra–Crespo, S.; Jordán–Núñez, J.; Vargas, B.; Vigil, L. Classification of glucose records from patients at diabetes risk using a combined permutation entropy algorithm. Comput. Methods Programs Biomed. 2018, 165, 197–204. [Google Scholar] [CrossRef] [PubMed]
  44. Lekha, S.; Suchetha, M. Real-time non-invasive detection and classification of diabetes using modified convolution neural network. IEEE J. Biomed. Health Inform. 2018, 22, 1630–1636. [Google Scholar] [CrossRef] [PubMed]
  45. Wan, S.; Liang, Y.; Zhang, Y. Deep convolutional neural networks for diabetic retinopathy detection by image classification. Comput. Electr. Eng. 2018, 72, 274–282. [Google Scholar] [CrossRef]
  46. Kannadasan, K.; Damodar Reddy, E. andVenkatanareshbabu, K. Type 2 diabetes data classification using stacked autoencoders in deep neural networks. Clin. Epidemiol. Glob. Health 2019, 7, 530–535. [Google Scholar] [CrossRef] [Green Version]
  47. Zhu, T.; Li, K.; Herrero, P.; Chen, J.; Georgiou, P. A Deep Learning Algorithm for Personalized Blood Glucose Prediction. In Proceedings of the 3rd International Workshop on Knowledge Discovery in Healthcare Data (KDH), within the 27th International Joint Conference on Artificial Intelligence and the 23rd European Conference on Artificial Intelligence (IJCAI-ECAI 2018), CEUR Workshop Proceedings, Stockholm, Schweden, 8–12 July 2018; pp. 64–78. [Google Scholar]
  48. Bevan, R.; Coenen, F. Experiments in non-personalized future blood glucose level prediction. In Proceedings of the 5th International Workshop on Knowledge Discovery in Healthcare Data (KDH), within the 24th European Conference on Artificial Intelligence (ECAI2020), CEUR Workshop Proceedings, Santiago de Compostela, Spain, 29–30 August 2020; Volume 2675, pp. 100–104. [Google Scholar]
  49. Mayo, M.; Koutny, T. Neural multi-class classification approach to blood glucose level forecasting with prediction uncertainty visualisation. In Proceedings of the 5th International Workshop on Knowledge Discovery in Healthcare Data (KDH), within the 24th European Conference on Artificial Intelligence (ECAI2020), CEUR Workshop Proceedings, Santiago de Compostela, Spain, 29–30 August 2020; Volume 2675, pp. 80–84. [Google Scholar]
  50. Liu, Y.; Liu, W.; Chen, H.; Cai, X.; Zhang, R.; An, Z.; Shi, D.; Ji, L. Graph convolutional network enabled two-stream learning architecture for diabetes classification based on flash glucose monitoring data. Biomed. Signal Process. Control 2021, 69, 102896. [Google Scholar] [CrossRef]
  51. Yin, H.; Mukadam, B.; Dai, X.; Jha, N.K. DiabDeep: Pervasive diabetes diagnosis based on wearable medical sensors and efficient neural networks. IEEE Trans. Emerg. Top. Comput. 2021, 9, 1139–1150. [Google Scholar] [CrossRef]
  52. Kumari, S.; Kumar, D.; Mittal, M. An ensemble approach for classification and prediction of diabetes mellitus using soft voting classifier. Int. J. Cogn. Comput. Eng. 2021, 2, 40–46. [Google Scholar] [CrossRef]
  53. Ganie, S.; Malik, M. An ensemble machine learning approach for predicting type-II diabetes mellitus based on lifestyle indicators. Healthc. Anal. 2022, 2, 100092. [Google Scholar] [CrossRef]
  54. Gupta, H.; Varshney, H.; Sharma, T.K.; Pachauri, N.; Verma, O.P. Comparative performance analysis of quantum machine learning with deep learning for diabetes prediction. Complex Intell. Syst. 2022, 8, 3073–3087. [Google Scholar] [CrossRef]
  55. Joseph, L.P.; Joseph, E.A.; Ramendra, P. Explainable diabetes classification using hybrid Bayesian-optimized TabNet architecture. Comput. Biol. Med. 2022, 151, 106178. [Google Scholar] [CrossRef]
  56. Lundberg, S.M.; Nair, B.; Vavilala, M.S.; Horibe, M.; Eisses, M.J.; Adams, T.; Liston, D.E.; Low, D.K.W.; Newman, S.F.; Kim, J.; et al. Explainable machine-learning predictions for the prevention of hypoxaemia during surgery. Nat. Biomed. Eng. 2018, 2, 749–760. [Google Scholar] [CrossRef] [PubMed]
  57. Sisodia, D.; Sisodia, D.S. Prediction of diabetes using classification algorithms. Procedia Comput. Sci. 2018, 132, 1578–1585. [Google Scholar] [CrossRef]
  58. Cuesta, H.A.; Coffman, D.L.; Branas, C.; Murphy, H.M. Using decision trees to understand the influence of individual- and neighborhood-level factors on urban diabetes and asthma. Health Place 2019, 58, 102119. [Google Scholar] [CrossRef] [PubMed]
  59. Sneha, C.; Gangil, T. Analysis of diabetes mellitus for early prediction using optimal features selection. J. Big Data 2019, 6, 13. [Google Scholar] [CrossRef]
  60. Hasan, K.; Alam, A.; Das, D.; Hossain, E.; Hasan, M. Diabetes prediction using ensembling of different machine learning classifiers. IEEE Access 2020, 8, 76516–76531. [Google Scholar] [CrossRef]
  61. Choubey, D.K.; Kumar, P.; Tripathi, S.; Kumar, S. Performance evaluation of classification methods with PCA and PSO for diabetes. Netw. Model. Anal. Health Inform. Bioinform. 2020, 9, 1–30. [Google Scholar] [CrossRef]
  62. Özmen, E.P.; Özcan, T. Diagnosis of diabetes mellitus using artificial neural network and classification and regression tree optimized with genetic algorithm. J. Forecast. 2020, 39, 661–670. [Google Scholar] [CrossRef]
  63. Naveen Kishore, G.; Rajesh, V.; Vamsi Akki Reddy, A.; Sumedh, K.; Rajesh Sai Reddy, T. Prediction of diabetes using machine learning classification algorithms. Int. J. Sci. Technol. Res. 2020, 1, 1805–1808. [Google Scholar]
  64. Torkey, H.; Ibrahim, E.; El-Din Hemdan, E.; El-Sayed, A.; Shouman, M.A. Diabetes classification application with efficient missing and outliers data handling algorithms. Complex Intell. Syst. 2022, 8, 237–253. [Google Scholar] [CrossRef]
  65. Hassan, M.M.; Mollick, S.; Yasmin, F. An unsupervised cluster-based feature grouping model for early diabetes detection. Healthc. Anal. 2022, 2, 10o112. [Google Scholar]
  66. Ahmedi, U.; Issa, G.F.; Khan, M.A.; Aftab, S.; Khang, M.F.; Said, R.A.T.; Ghazal, T.M.; Ahmad, M. Prediction of diabetes empowered with fused machine learning. IEEE Access 2022, 10, 8529–8538. [Google Scholar] [CrossRef]
  67. Chang, V.; Ganatra, M.A.; Hall, K.; Golightly, L.; Xu, Q.A. An assessment of machine learning models and algorithms for early prediction and diagnosis of diabetes using health indicators. Healthc. Anal. 2022, 2, 100118. [Google Scholar] [CrossRef]
  68. Thotad, P.N.; Bharamagoudar, G.R.; Anami, B.S. Diabetes disease detection and classification on Indian demographic and health survey data using machine learning methods. Diabetes Metab. Syndr. Clin. Res. Rev. 2023, 17, 102690. [Google Scholar] [CrossRef]
  69. Theerthagiri, P.; Ruby, A.U.; Vidya, J. Diagnosis and classification of the diabetes using machine learning algorithms. SN Comput. Sci. 2023, 4, 1–10. [Google Scholar] [CrossRef]
  70. Fürnkranz, J. Decision Tree. In Encyclopedia of Machine Learning; Springer: Boston, MA, USA, 2010; pp. 263–267. [Google Scholar]
  71. Alvarado, J.; Velasco, J.M.; Hidalgo, J.I.; Fernández de Vega, F. Blood Glucose Prediction Using a Two Phase TSK Fuzzy Rule Based System. In Proceedings of the IEEE Congress on Evolutionary Computation (CEC), Kraków, Poland, 28 June–1 July 2021; pp. 1–8. [Google Scholar]
  72. De La Cruz, M.; Cervigón, C.; Alvarado, J.; Botella-Serrano, M.; Hidalgo, J.I. Evolving Classification Rules for Predicting Hypoglycemia Events. In Proceedings of the IEEE Congress on Evolutionary Computation (CEC), Padua, Italy, 18–23 July 2022; pp. 1–8. [Google Scholar]
  73. Sun, C.; Ippel, L.; Dekker, A.; Dumontier, M.; van Soes, J. A systematic review on privacy-preserving distributed data mining. Data Sci. 2021, 4, 121–150. [Google Scholar] [CrossRef]
  74. Xu, J.; Glicksberg, B.S.; Su, C.; Walker, P.; Bian, J.; Wang, F. Federated learning for healthcare informatics. J. Healthc. Inform. Res. 2021, 5, 1–19. [Google Scholar] [CrossRef]
  75. Rahman, A.; Hossain, M.S.; Muhammad, G.; Kundu, D.; Debnath, T.; Rahman, M.; Khan, M.S.I.; Tiwari, P.; Band, S.S. Federated learning-based AI approaches in smart healthcare: Concepts, taxonomies, challenges and open issues. Cluster Comput. 2022, 1–41. [Google Scholar] [CrossRef]
  76. Wang, W.; Li, X.; Qiu, X.; Zhang, X.; Brusic, V.; Zhao, J. A privacy preserving framework for federated learning in smart healthcare systems. Inf. Process. Manag. 2023, 60, 103167. [Google Scholar] [CrossRef]
  77. Lai, J.; Song, X.; Wang, R.; Li, X. Edge intelligent collaborative privacy protection solution for smart medical. Cyber Secur. Appl. 2023, 1, 100010. [Google Scholar] [CrossRef]
  78. Liu, J.; Xi, L.; Yang, H.; Zhuang, L. A diabetes prediction system based on federated learning. In Proceedings of the IEEE International Conference on Big Data, Information and Computer Network (BDICN), Sanya, China, 20–22 January 2022; pp. 486–491. [Google Scholar]
  79. Islam, H.; Mosa, A.; Famia. A federated mining approach on predicting diabetes-related complications: Demonstration using real-world clinical data. In Proceedings of the AMIA Annual Symposium Proceedings, Washington, DC, USA, 5–9 November 2022; pp. 556–564. [Google Scholar]
  80. Dubreuil, M.; Gagn e, C.; Parizeau, M. Analysis of a master-slave architecture for distributed evolutionary computations. IEEE Trans. Syst. Man Cybern.-Part B 2001, 36, 229–235. [Google Scholar] [CrossRef] [PubMed]
  81. Cantú-Paz, E. Migration policies, selection pressure, and parallel evolutionary algorithms. J. Heuristics 2001, 7, 311–334. [Google Scholar] [CrossRef]
  82. O’Neill, M.; Ryan, C. Grammatical Evolution: Evolutionary Automatic Programming in an Arbitrary Language; Kluwer Academic Publishers: New York, NY, USA, 2003. [Google Scholar]
  83. Marling, C.; Bunescu, R. The OhioT1DM dataset for blood glucose level prediction. In Proceedings of the 3rd International Workshop on Knowledge Discovery in Healthcare Data (KDH), Stockholm, Sweden, 8–12 July 2018; pp. 60–63. [Google Scholar]
  84. Hovorka, R.; Canonico, V.; Chassin, L.J.; Haueter, U.; Massi-Benedetti, M.; Orsini Federici, M.; Pieber, T.R.; Schaller, H.C.; Schaupp, L.; Vering, T.; et al. Nonlinear model predictive control of glucose concentration in subjects with Type 1 diabetes. Physiol. Meas. 2004, 25, 905–920. [Google Scholar] [CrossRef] [Green Version]
  85. Hovorka, R.; Powrie, J.K.; Smith, G.D.; Sönksen, P.H.; Carson, E.R.; Jones, R.H. Five-compartment model of insulin kinetics and its use to investigate action of chloroquine in NIDDM. Am. J. Physiol. 1993, 265, 162–1752. [Google Scholar] [CrossRef]
  86. Livesey, G.; Wilson, P.D.; Dainty, J.R.; Brown, J.C.; Faulks, R.M.; Roe, M.A.; Newman, T.A.; Eagles, J.; Mellon, F.A.; Greenwood, R.H. Simultaneous time-varying systemic appearance of oral and hepatic glucose in adults monitored with stable isotopes. Am. J. Physiol. 1998, 275, 717–728. [Google Scholar] [CrossRef]
  87. Ni, J.; Drieberg, R.; Rockett, P. The use of an analytic quotient operator in Genetic Programming. IEEE Trans. Evol. Comput. 2013, 17, 146–152. [Google Scholar] [CrossRef]
  88. Varela Lorenzo, A.; Delgado Gutierrez, A. Glucose Classification and Prediction System with Neural Networks. Ph.D. Thesis, University of Madrid, Madrid, Spain, 2020. [Google Scholar]
  89. Danne, T.; Nimri, R.; Battelino, T.; Bergenstal, R.M.; Close, K.L.; DeVries, J.H.; Garg, S.; Heinemann, L.; Hirsch, I.; Amiel, S.A.; et al. International consensus on use of continuous glucose monitoring. Diabetes Care 2017, 40, 1631–1640. [Google Scholar] [CrossRef] [Green Version]
  90. Sun, Y.; Wong, A.K.; Kamel, M. Classification of imbalanced data: A review. Int. J. Pattern Recognit. Artif. Intell. 2009, 23, 687–719. [Google Scholar] [CrossRef]
  91. Hidalgo, J.I.; Colmenar, J.M.; Kronberger, G.; Winkler, S.M.; Garnica, O.; Lanchares, J. Data based prediction of blood glucose concentrations using evolutionary methods. J. Med. Syst. 2017, 41, 1–20. [Google Scholar] [CrossRef] [PubMed]
  92. Contador, S.; Colmenar, J.M.; Garnica, O.; Velasco, J.M.; Hidalgo, J.I. Blood glucose prediction using multi-objective grammatical evolution: Analysis of the “agnostic” and “what-if” scenarios. Genet. Program. Evolvable Mach. 2022, 23, 161–192. [Google Scholar] [CrossRef]
  93. De Falco, I.; Della Cioppa, A.; Koutny, T.; Krcma, M.; Scafuri, U.; Tarantino, E. A Grammatical Evolution approach for estimating blood glucose levels. In Proceedings of the 11th IEEE Global Communications Conf.-Int. Workshop on AI-driven Smart Healthcare (AIdSH), Taipei, Taiwan, 7–11 December 2020; pp. 1–6. [Google Scholar]
  94. Rodríguez-Fdez, I.; Canosa, A.; Mucientes, M.; Bugarín, A. STAC: A web platform for the comparison of algorithms using statistical tests. In Proceedings of the IEEE international conference on fuzzy systems (FUZZ-IEEE), Istanbul, Turkey, 2–5 August 2015; pp. 1–8. [Google Scholar]
  95. Derrac, J.; García, S.; Molina, D.; Herrera, F. A practical tutorial on the use of nonparametric statistical tests as a methodology for comparing evolutionary and swarm intelligence algorithms. Swarm Evol. Comput. 2011, 1, 3–18. [Google Scholar] [CrossRef]
Figure 1. FLEA migration. The left side of the figure sketches the migration from the slaves to the master, while the right one traces the migration from the master to the slaves. The largest circles show single slaves with the corresponding local individuals, i.e., prediction models ( m 1 , , m n ). In particular, the individuals internal to the left-side circle have the same color because the models are all related only to local data. Differently, the individuals inside the right-side circle indicate that each slave, after communication with the master, can integrate the information from the immigrants by exploiting the mechanism of selection, replacement, and genetic operators, as evidenced by the color overlapping of the single individuals.
Figure 1. FLEA migration. The left side of the figure sketches the migration from the slaves to the master, while the right one traces the migration from the master to the slaves. The largest circles show single slaves with the corresponding local individuals, i.e., prediction models ( m 1 , , m n ). In particular, the individuals internal to the left-side circle have the same color because the models are all related only to local data. Differently, the individuals inside the right-side circle indicate that each slave, after communication with the master, can integrate the information from the immigrants by exploiting the mechanism of selection, replacement, and genetic operators, as evidenced by the color overlapping of the single individuals.
Sensors 23 02957 g001
Figure 2. The grammar for the glucose forecasting model (Equation (5)).
Figure 2. The grammar for the glucose forecasting model (Equation (5)).
Sensors 23 02957 g002
Figure 3. Classification bands on the testing set for each patient in the Ohio data set [83].
Figure 3. Classification bands on the testing set for each patient in the Ohio data set [83].
Sensors 23 02957 g003
Figure 4. Confusion matrices on the testing set of the best model (model 570) evolved by the non-FL approach on each patient.
Figure 4. Confusion matrices on the testing set of the best model (model 570) evolved by the non-FL approach on each patient.
Sensors 23 02957 g004aSensors 23 02957 g004b
Figure 5. Confusion matrices on the testing set of each patient for the best model (model 575) evolved by FLEA algorithm.
Figure 5. Confusion matrices on the testing set of each patient for the best model (model 575) evolved by FLEA algorithm.
Sensors 23 02957 g005
Figure 6. Confusion matrices on the testing set of new subjects not participating in the learning process achieved by the best model (model 570) evolved by non-FL approach.
Figure 6. Confusion matrices on the testing set of new subjects not participating in the learning process achieved by the best model (model 570) evolved by non-FL approach.
Sensors 23 02957 g006
Figure 7. Confusion matrices on the testing set of new subjects not participating in the learning process achieved by the best model (model 575) evolved by FLEA algorithm.
Figure 7. Confusion matrices on the testing set of new subjects not participating in the learning process achieved by the best model (model 575) evolved by FLEA algorithm.
Sensors 23 02957 g007
Table 1. Review summary and positioning of our paper in the literature.
Table 1. Review summary and positioning of our paper in the literature.
Existing ML SystemsExamples Refs.StrengthsLimitations
Non-FL[43,44,45,46,47,48,49,50,51,52,53,54,55]May have goodProblems of data privacy -
[57,58,59,60,61,62,63,64,65,66,67,68,69,71,72]Numerical performanceNot usable in many fields
Current-FL[74,75,76,77]Solve dataSolutions are not
[78,79]Privacy problemsInterpretable
Proposed approachThis paperSolve data
Privacy problems -
Solutions are
Interpretable
Table 2. Protected functions used in the grammar.
Table 2. Protected functions used in the grammar.
FunctionProtected Function
plog ( x ) log ( 1 + | x | )
psqrt ( x ) | x |
aq ( x , y ) x 1 + y 2
Table 3. The seven classes used for the glucose classification problem.
Table 3. The seven classes used for the glucose classification problem.
ClassClass IDRange (mmol/L)Range (mg/dL)Action Required
very low0<3.0<54immediate action
low1[3.0–3.9[[54–70[hypoalert and monitor
normal-closing-to-hypo2[3.9–5[[70–90[‘towards hypo’ warning
normal3[5.0–7.8[[90–140[none
normal-closing-to-hyper4[7.8–10.0[[140–180[‘towards hyper’ warning
high5[10.0–13.9[[180–250[alert and monitor
very high6≥13.9≥250immediate action
Table 4. The number of samples in the training/testing (Tr/Ts) sets for each patient ID and class ( C 0 ÷ C 6 ) .
Table 4. The number of samples in the training/testing (Tr/Ts) sets for each patient ID and class ( C 0 ÷ C 6 ) .
C0C1C2C3C4C5C6
IDTr/TsTr/TsTr/TsTr/TsTr/TsTr/TsTr/Ts
55960/31307/471452/2311453/3081710/442348/5682131/438
56332/0270/191558/1202043/3212712/6502590/8891053/337
57014/0214/12730/1361092/1771544/2623325/5163460/1323
575215/35759/1021631/3922032/4342335/5051815/513978/355
58825/0113/4837/1201734/2263341/6153920/10142046/444
591110/14307/1281199/3831495/5612365/4742358/7061662/185
Table 5. Results of the non-FL approach. The best average F 1 score is reported in bold.
Table 5. Results of the non-FL approach. The best average F 1 score is reported in bold.
PatientF1 Score
Best Model559563570575588591Avg (StdDev)
model 5590.70060.57310.68030.58780.68230.54860.6288 (0.0604)
model 5630.68200.70720.82940.56790.69750.56350.6746 (0.0907)
model 5700.72700.72330.82690.62910.73120.60610.7073 (0.0729)
model 5750.70520.67260.76570.64230.69350.58660.6777 (0.0553)
model 5880.64660.53950.71670.49960.73100.51340.6078 (0.0947)
model 5910.65770.69130.75790.56370.67610.57960.6544 (0.0663)
Avg F1 score0.68650.65110.76280.58170.70190.5663
(StdDev)(0.0278)(0.0695)(0.0540)(0.0469)(0.0218)(0.0297)
Table 6. Results of the FLEA algorithm with a communication frequency equal to 100. The best average F 1 score is reported in bold.
Table 6. Results of the FLEA algorithm with a communication frequency equal to 100. The best average F 1 score is reported in bold.
PatientF1 Score
Best Model559563570575588591Avg (StdDev)
model 5590.73040.72100.80470.63370.73200.60150.7039 (0.0675)
model 5630.72780.71410.79470.61130.71930.60430.6953 (0.0673)
model 5700.71940.73120.83140.59860.72350.59520.6999 (0.0821)
model 5750.72340.71770.80160.64720.73390.60520.7048 (0.0632)
model 5880.72600.72870.82060.62300.72160.60200.7037 (0.0730)
model 5910.72600.72140.81800.61990.72910.59320.7013 (0.0750)
Avg F1 score0.72550.72230.81180.62230.72660.6002
(StdDev)(0.0035)(0.0059)(0.0126)(0.0155)(0.0054)(0.0045)
Table 7. Results obtained by the two global models over the testing sets of six new subjects.
Table 7. Results obtained by the two global models over the testing sets of six new subjects.
Patient F 1 Score
Model540544552567584596Avg (StdDev)
non-FL0.60750.70240.57480.61930.65040.65240.6345 (0.0402)
FLEA0.62610.70360.60770.63120.66670.66810.6506 (0.0321)
Table 8. Overall numerical scores over the testing sets of six new patients.
Table 8. Overall numerical scores over the testing sets of six new patients.
IDPrecisionRecall F 1 Accuracy
FLEA5400.640.630.630.63
5440.720.700.700.70
5520.650.600.610.60
5670.640.630.630.63
5840.690.670.670.67
5960.710.660.670.66
Avg 0.680.650.650.65
non-FL5400.620.610.610.61
5440.720.700.700.70
5520.620.580.570.58
5670.620.620.620.62
5840.670.660.650.66
5960.690.650.650.65
Avg 0.660.640.630.64
Table 9. Quade ranks test over the twelve subjects of the Ohio T1DM data set.
Table 9. Quade ranks test over the twelve subjects of the Ohio T1DM data set.
RankAlgorithm
FLEA1.26923
Non-FL1.73077
Statistic: 2.19188p-value: 0.16680
Table 10. Post hoc procedures for Quade ranks test over the twelve subjects of the Ohio T1DM data set. FLEA was chosen as the control method.
Table 10. Post hoc procedures for Quade ranks test over the twelve subjects of the Ohio T1DM data set. FLEA was chosen as the control method.
StatisticBonferroni-DunnHolmFinnerHochbergLi
FLEA vs. non-FL1.996920.045830.045830.045830.045830.04583
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

De Falco, I.; Della Cioppa, A.; Koutny, T.; Ubl, M.; Krcma, M.; Scafuri, U.; Tarantino, E. A Federated Learning-Inspired Evolutionary Algorithm: Application to Glucose Prediction. Sensors 2023, 23, 2957. https://doi.org/10.3390/s23062957

AMA Style

De Falco I, Della Cioppa A, Koutny T, Ubl M, Krcma M, Scafuri U, Tarantino E. A Federated Learning-Inspired Evolutionary Algorithm: Application to Glucose Prediction. Sensors. 2023; 23(6):2957. https://doi.org/10.3390/s23062957

Chicago/Turabian Style

De Falco, Ivanoe, Antonio Della Cioppa, Tomas Koutny, Martin Ubl, Michal Krcma, Umberto Scafuri, and Ernesto Tarantino. 2023. "A Federated Learning-Inspired Evolutionary Algorithm: Application to Glucose Prediction" Sensors 23, no. 6: 2957. https://doi.org/10.3390/s23062957

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop