You are currently viewing a new version of our website. To view the old version click .
Big Data and Cognitive Computing
  • Editor’s Choice
  • Article
  • Open Access

31 March 2025

A Verifiable, Privacy-Preserving, and Poisoning Attack-Resilient Federated Learning Framework

,
,
and
Secure Cyber Systems Research Group (CSCRG), WMG, University of Warwick, Coventry CV4 7AL, UK
*
Author to whom correspondence should be addressed.

Abstract

Federated learning is the on-device, collaborative training of a global model that can be utilized to support the privacy preservation of participants’ local data. In federated learning, there are challenges to model training regarding privacy preservation, security, resilience, and integrity. For example, a malicious server can indirectly obtain sensitive information through shared gradients. On the other hand, the correctness of the global model can be corrupted through poisoning attacks from malicious clients using carefully manipulated updates. Many related works on secure aggregation and poisoning attack detection have been proposed and applied in various scenarios to address these two issues. Nevertheless, existing works are based on the trust confidence that the server will return correctly aggregated results to the participants. However, a malicious server may return false aggregated results to participants. It is still an open problem to simultaneously preserve users’ privacy and defend against poisoning attacks while enabling participants to verify the correctness of aggregated results from the server. In this paper, we propose a privacy-preserving and poisoning attack-resilient federated learning framework that supports the verification of aggregated results from the server. Specifically, we design a zero-trust dual-server architectural framework instead of a traditional single-server scheme based on trust. We exploit additive secret sharing to eliminate the single point of exposure of the training data and implement a weight selection and filtering strategy to enhance robustness to poisoning attacks while supporting the verification of aggregated results from the servers. Theoretical analysis and extensive experiments conducted on real-world data demonstrate the practicability of our proposed framework.

1. Introduction

Data breaches are a general concern, and various governments have created regulations to safeguard users’ data. Some of such established regulations include the European Union General Data Protection Regulation (GDPR) [1], the California Consumer Privacy Act (CCPA) in the United States [2], and Singapore’s Personal Data Protection Act (PDPA) [3]. Organizations face high penalties for breaching these regulations, and under the above circumstances, preserving the users’ private data is paramount. Traditional machine learning (ML) approaches typically employ a centralized architecture where data are stored and processed centrally. In addition to the high cost and unacceptable latency associated with such practices, data privacy and security remain the major concerns. With central data storage and processing, sensitive data are exposed to leaks and attacks. An alternative solution that can preserve data privacy is a distributed and on-device ML technique known as federated learning (FL).
FL is a concept introduced by Google that can be used to support the privacy preservation of data. FL allows for the distributed and on-device collaborative training of devices or servers without sharing raw data. The privacy and security issues associated with centralizing data in a single location are solved using FL because it allows individual participants to process data locally and only share model weights. As a distributed ML technique, it can operate with a coordinating central server or without a server orchestrating the training in a fully distributed approach [4]. FL enables decentralized computing by utilizing edge devices’ computational power, reducing reliance on centralized servers. Moreover, there is a reduction in bandwidth usage and minimization of network load because only model updates are transmitted. Due to its numerous advantages, it has been applied in different areas, such as healthcare [5], finance [6], vehicle networks [7], and cyberattack detection [8]. A standard FL framework typically comprises a coordinating (central) server, participating devices, and a communication–computation framework to train the model [9]. The operating technique is such that the central server sends a pre-trained or global model to participating devices for global model training. The participating devices upload weights generated from local training data to the central server for aggregation. The central server aggregates these weights to generate an updated global model. To achieve model convergence or attain a specific number of rounds, there is a continuous client–server interaction [10].
Recent studies have identified specific challenges around the FL approach to model training regarding privacy preservation, security, and integrity [3,4,5,6,7,8]. A malicious server can indirectly obtain sensitive information such as local training data [11], class representatives [12], and labels [13] through shared weights. On the other hand, poisoning attacks by malicious clients through manipulated updates can corrupt the correctness of the global model [14,15,16]. Also, a malicious server may return incorrectly aggregated results to participants [17], precipitating the need to verify the correctness of the aggregated model. To tackle the above challenges, several works focusing on privacy preservation, security, and integrity of aggregation in FL have been proposed. To address the issue of privacy preservation in FL, Refs. [18,19] proposed using secure multiparty computation to enable secure aggregation. To tackle poisoning attacks, Refs. [20,21] implemented a statistical approach using Euclidean distance. Additionally, to prove the correctness of model aggregation, Ref. [22] proposed using a homomorphic hash function.
These research works either solved the problem of privacy preservation or poisoning attacks but failed to tackle either two or all the three issues identified simultaneously. It is particularly challenging to preserve privacy in FL and ensure robustness against poisoning attacks because the techniques employed for privacy preservation hide the exchanged weights, making it difficult to query them for anomaly [23]. To simultaneously preserve user privacy while defending against poisoning attacks, Ref. [24] proposed using two servers to detect poisoning attacks based on Euclidean distance and secure multiparty computation (SMC) for privacy preservation. Similarly, Ref. [25] adopted K-nearest weight to detect poisoning attacks and secret sharing for privacy preservation. Moreso, [26] proposed a distributed trust architecture for poisoning attack resilience and malformed update defense. Nonetheless, none of the solutions mentioned support the verification of the correctness of aggregated results from the server. The tendency to compromise participants’ privacy increases once a malicious server can manipulate the data returned to participants. Research carried out in [17] proposed using a double-masking protocol for privacy preservation and a homomorphic hash function to verify aggregated results from the server. However, they failed to provide a mechanism to defend against poisoning attacks from the participants. None of these solutions implemented a framework for simultaneously preserving privacy and security while supporting the verification of aggregated results. In addition, as mentioned earlier, these solutions have an element of trust in their framework. As stated in [27], a zero-trust architectural framework can transform the FL network into a robust defense system that will prevent an authorized party or compromised clients from acting. In an FL environment, there is a need for simultaneous preservation of privacy, security, and zero-trust verification of aggregated results.
This paper proposes a simultaneous privacy-preserving, poisoning attack-resilient, and zero-trust verification FL framework. Our privacy preservation technique and prevention of poisoning attacks while supporting the verification of aggregated results enhance the confidentiality, integrity, and authentication of device–server communication in FL. Our contributions can be summarized as follows:
  • To eliminate the single point of exposure of the training data and ensure that all parties are equally blind to the complete training data, we exploit additive secret sharing with an oblivious transfer protocol by splitting the training data between the two servers. This minimizes the risk of exposing the training data, enhances protection against reconstruction attacks, and enables privacy preservation even if one server is compromised.
  • To enhance robustness against poisoning attacks by malicious participants, we implement a weight selection and filtering strategy that identifies the coherent cores of benign updates and filters out boosted updates based on their L2 norm. The weight selection, filtering, and cross-validation across servers achieve a synergy that creates a resilient barrier against poisoning attacks.
  • To reduce reliance on trusting a single party and ensure that all participant contributions are proportionally represented, we utilize our dual-server architecture to exploit the additive secret-sharing scheme. Specifically, this allows each server to provide proof of aggregation and enables the participants to verify the correctness of the aggregated results returned by the servers. This helps to foster a zero-trust architecture and maintain data integrity and confidentiality throughout the training process.
The rest of this paper is organized as follows. In Section 2, we review related studies. Section 3 comprises preliminaries. Section 4 shows the system architecture, threat model, design goals, and system interpretation. In Section 5, we present our secret-sharing schemes and a method for detecting poisoning attacks. Section 6 shows the security analysis of our framework. Section 7 outlines the experimental setup and discusses the results. We conclude our work in Section 8.

3. Preliminaries

(1)
FL System
For this study, our framework is based on a scalable FL system, and we adopted the Horizontal FL (HFL) method in which participants with the same data structure train an ML model using a parameter.
A brief description of the significant steps involved in the HFL is as follows:
  • The server randomly selects participants for model training and distributes the global model for training.
  • Participants selected for model training use their local data to train the model, compute their model weights, and individually send their updates to the servers for aggregation.
  • The servers use the agreed protocol to collectively aggregate the updates to form a new global model without knowing the participants’ private data.
  • The aggregator server sends the aggregated new global model to the participants for another round of training.
(A)
Local training dataset ( D x ) : Within an FL process, assume that there are M participants and each participant P i holds a local training dataset:
D x = U y x , V y x y d x
where U y x = I n p u t data at index y in P i dataset;
V y x = Label or ground truth for the input U y x ;
d x = The total number of data samples available in the local training dataset D x of P i ;
D x = Local training dataset;
(B)
Global dataset (D): This is the union of all individual datasets from M participants consisting of the training data from all the participants in an FL setting. Let the global dataset be:
D = U x = 1 M D x
(C)
Neural Network (NN) and Loss Function ( L F ): For an NN function defined as f(U, N) where the model input is still U and the trainable parameters (weight and biases) are represented as N . As a Mean Squared Error (MSE), the loss function L F D , N can be defined as:
L F D , N = 1 D U y x , V y x D V y x f U y x , N 2
where f U y x , N = the predicted output of the NN;
V y x f U y x , N 2 = the squared error for each data point.
Here, the goal is to find the optimal parameters to minimize the loss function. If N is updated using SGD optimizer, iteratively, each parameter is calculated as:
N i + 1 = N i η Δ L F D i , N i
where
N i = ith iterative parameter;
η = Learning rate (predefined);
D i = A random subset of D.
(2)
Proof of Aggregation (PoA) [51]
PoA is considered a vital component in FL systems, providing a need to authenticate, validate, and ensure the integrity of the aggregated process and result [52]. Conventional FL relies on the trust assumption that the central server will perform its aggregation function effectively and faithfully. This may not be true because a malicious server, in reality, could manipulate aggregated results by carefully injecting false participants into the training process or replacing participant’s training models. Therefore, the aggregator server should provide proof of aggregation for each round to demonstrate to the participants that the aggregation process was executed faithfully and to ensure the correctness of the aggregated result [53]. In other words, it demonstrates the certainty of a result to another party without exposing any data beyond the result’s authenticity [54].
(3)
Zero-Trust Architecture
Participants in FL involve devices with few security measures and could pose weak points of entry to critical assets. In the same vein, the privacy of participants’ local updates is closely related to the correctness of results returned by the server. Manipulated results returned to participants by a malicious server tend to increase the risk of participants’ privacy being compromised. Consequently, carefully manipulated results in white-box attacks could incite participants to unleash sensitive information [55]. A zero-trust architecture is the assumption that every entity within an FL framework could be at some point malicious or compromised by external attackers, and amid an attack, the initial point of contact is often not the intended target [56].

4. System Architecture

This section describes the components of the system’s architectural framework, a description of the threat model, a design goal, and a detailed interpretation of the system architecture.
A.
Description of components
Server S a : This server is tasked with the responsibilities of generating the global model, the distribution of the model, the prevention of poisoning attacks by a malicious participant’s update, and the aggregation of combined results in collaboration with the server S b . Furthermore, it provides PoA after every iteration and sends the results to S b and the participants.
Server S b : Similarly to all the responsibilities associated with S a , it validates the correctness of the aggregated results from S a and also provides PoA to all entities within the framework.
Participants: They are multiple devices or organizations P 1   t o   P n that communicate with the servers for model computations and training.
B.
Threat Model
All entities in the framework are considered to be honest-but-curious entities [17], which implies that the entities could use an agreed arrangement to execute the protocol but may try to infer other entities’ private data. Similarly to related schemes [24,25], we consider the following threat model in this framework:
  • The two servers S a   a n d   S b will not collude with each other and will use the agreed arrangement to execute the protocol but may independently try to infer the data privacy of the participants.
  • Poison attacks from malicious participants corrupt the global model through the exchange of malicious local updates. We consider a strong threat model with diverse attack rates (10% to 40%) of participants in the FL process to be malicious. The assumption is that the participants may launch poisoning attacks such as label flipping, sign flipping, gradient manipulation, backdoor, and backdoor embedding attacks.
  • Modifying aggregated results and forging proof of aggregated results from any server to deceive other entities within the framework.
Specifically, there is no collusion between any entities in the framework.
C.
Design Goal
Our design goal is to preserve the privacy of local data and prevent poisoning attacks from malicious participants while supporting the correctness of the aggregated model. Consequently, our framework should simultaneously achieve privacy preservation, robustness, and proof of the correctness of the aggregated model and enable a zero-trust architectural framework. Specifically, our scheme guarantees the following design goal:
  • Privacy: The scheme should prevent an attacker or a malicious server from interfering with the privacy of the exchanged gradients and further ensure that there is no privacy leakage of the information used to detect poisoning attacks.
  • Fidelity: The accuracy of the global model should not be compromised due to the robustness of the framework. As a baseline, we use FedAVG [57], a standard algorithm resilient to poisoning attacks with no privacy preservation method. Therefore, the framework should achieve a level of accuracy comparable to FedAvg.
  • Robustness: The framework should ensure the robustness of the global model by preventing malicious participants from poisoning the global model. Specifically, the framework should be resilient to poisoning attacks from malicious participants.
  • Efficiency: Our framework should minimize the computation and communication overheads by enabling the preservation of the privacy of participants with limited computational resources.
In addition, the framework should ensure a zero-trust architectural framework capable of transforming the FL network into a solid defense system where no malicious entity can successfully act.
D.
System Interpretation
Figure 1 depicts a distributed ML framework. Similarly to [24,25,26], our framework utilizes a dual-server architecture S a   a n d   S b . This dualistic approach provides a robust, secure, decentralized data processing, and zero-trust framework. The server ( S a ) initiates the training process by randomly selecting participants for model training. The global model is sent to selected participants for model training through a secured channel. It is worth noting that both servers can initiate the model training.
Figure 1. System architecture.
On receiving the global model, the participants individually perform local computations using their local data to train the model. They try to solve the loss function using optimal parameters. After local computation and training of the model with their local data, each participant splits their updates to obtain W i a   a n d   W i b , and sends it to the two servers S a   a n d   S b , respectively, using additive secret sharing. No server receives the full participants’ update, but the splitting is at the discretion of individual participants. The splitting is such that W i a   a n d     W i b add up to give a complete update of each participant. In this framework, the oblivious transfer (OT) protocol transfers the split weights, and there is no collusion between entities within the framework. The OT protocol is explained in Section 5.
On receiving the updates, the servers collaboratively check for malicious updates to prevent poisoning of the global model. Servers S a   and   S b individually aggregate their results to compute a weighted average and combine their aggregated results to obtain a complete model update. Each server individually sends its partial aggregated result and complete model update to all the entities within the framework to show PoA and ensure a zero-trust architecture. Participants reject a round if the aggregated results from both servers do not add up to give the complete model update without splitting the updates.

5. Secret Sharing Schemes

5.1. Secret Sharing for Privacy Preservation

Using Lagrange polynomial interpolation over finite fields, Shamir’s Secret Sharing (SSS) [58] is an optimal and ideal t ,   n threshold method where t is the threshold, and n is the number of participants in our framework. SSS is a protocol that can be used to achieve dropout robustness [59]. In such a method, the objective is to split a confidential secret S into fragments of shares S 1 ,   S 2 ,   S 3 , ,   S i (where i = 1 , 2, 3, …, n) in such a manner that the following applies:
  • The learning knowledge of t or more shares S i permits the computation of the original confidential secrets S . That is, the complete reconstruction of the secret S has a combination of at least t shares.
  • The learning knowledge of t 1 or fewer shares S i   , makes S completely unknown. That is, the potential value of the secret S remains unknown when aware of up to t 1 shares, just as there would be knowledge of zero shares. It is vital to note that the reconstruction of the secret S requires the presence of at least t shares. If n = t , then the complete shares are required for the reconstruction of the secret S .

5.2. Mathematical Formulation of SSS

Assume a value S which is secret-shared, a vector of n ,   t . This value S is not known to anyone and is individually shared with each participant in the training process. As an element of the finite field A p , let S be represented as a 0 . A p is an array of large prime p where 0 < t n < p .
Picking a random polynomial A x with a degree of t (where t 1 ) and a constant term of S , the secret holder randomly picks a 1 ,   ,   a t 1 from A p where = = S . This polynomial can be defined as:
A x a 0 + a 1 x + a 2 x 2 + a 3 x 3 + a t 1 x t 1 m o d   p
A x = r P s , t
The evaluations are happening at α 1 ,   α 2 ,   ,   α n which are distinct, non-zero, and publicly known elements. These evaluations are fixed for all instances of the sharing method. A x is evaluated at x = α 1 ,   a 1 ,   α 2 ,   a 2 ,   α 3 ,   a 3 ,   ,   α n , a n . The value of S is unknown, and no participant has the full vector. Instead, component-wise, each participant has the corresponding component of this vector.
To share the secret, the secret holder computes S i = A i   f o r   i     1,2 , 3 , , n and each participant receives i ,   S i .
To reconstruct the secret, any subset of t of i ,   S i pairs will be able to reconstruct the secret using the Lagrange interpolation formula:
a 0 = S = j = 1 t S j . m = 0 , m j t x m x m x j m o d   p

5.3. Additive Secret Sharing (ASS)

ASS [60] is an n-out-of-n scheme that requires all shares to reconstruct the secret. It is typically an SMC technology with additive homomorphism, as illustrated in Algorithm 1. ASS is a simple secret-sharing scheme, and our framework considers an ASS of 2-out-of-2. The recovered value would be random if there is a lack of any share. Without any collusion, it supports linear computation and protocol. Consequently, a secret w i Z L is additively shared between two servers in a manner that S a and S b , respectively, possess the random shares w i a   a n d   w i b such that w i a + w i b = W i .
As a privacy-preserving protocol in FL, it naturally aids the addition of the weights as W 1 + W 2 = W 1 1 + W 2 1 + W 1 2 + W 2 2 . It is simple, practicable, and the communication cost of clients in multiple-server architectures is much less than in a single-server architecture [61]. In ASS, the privacy of participants’ updates is guaranteed as long as these two conditions are satisfied. Firstly, the servers do not collude with each other, or in the case of multiple servers, at least one server does not collude with other servers. This means that if the two servers collude in a dual-server architecture, the secret shares can be reconstructed because each server can access one share. In multiple-server architectures, the secret reconstruction will be impossible because the secret share of the non-colluding server will be missing. Secondly, at least two participants do not collude with the rest. Suppose all participants collude with each other except one; the colluding participants can compute the weight of the non-colluding participants by subtracting its weight from the total weight of all participants (colluding and non-colluding). The computation will reveal the original weight of the non-colluding participant.
Nevertheless, ASS could introduce rounding errors and compromise precision [62]. Moreover, it could lead to information loss during splitting because there may exist more significant digits in the original weight than the split can accommodate [63]. Additionally, finite bit representation may result in rounding errors during weight reconstruction. However, these vulnerabilities are prevalent in single-server architecture.
Algorithm 1: Additive Secret Sharing
Given W x Z L
Split W x into W i 1   ,   W i 2
Such that W i 1 + W i 2 = W   m o d   L
Apply Oblivious Transfer
W i 1   Z L
W i 1 W i W i 2 m o d   L
Send W i 1   to   S a and W i 2   to   S b
Given two secret values W x 1   and   W y 1
Shares of Z W x 1 + W x 2 m o d   L can be computed locally by each server
S a setting Z a W x 1 + W y 1   m o d   L
S b setting Z b W x 2 + W y 2   m o d   L
Where L = 2

5.4. Oblivious Transfer (OT)

An OT protocol [64] is a very important cryptographic primitive that involves sending data only one part of which will arrive, but the sender is unaware of the part that will arrive. The receiver selects what they can read. The most basic protocol, and the one we considered for this work, is 1-out-of-2 OT [65]. The sender (OTS) has two private inputs w 0 .   w 1 and the receiver (OTR) has a selection bit z   0 ,   1 . After the execution of the protocol by both parties, the OTR obtains the message w z from messages held by the OTS, with the latter receiving no output.
The security execution of the protocol requires that the OTS learns nothing about the choice of the OTR while the OTR learns only its choice. The practical implementation is in the privacy preservation of data and has been widely used in secure two-party computation. Assume participant P1  Pn as the sender and Server S a as the receiver. P1 splits its weight into two and then interacts with S a . As a result, S a only acquires the weight selected and nothing more, while P1 learns nothing about S a ’s chosen weight. In a 1-out-of-2 OT protocol, P1 takes W 1 1 ,   W 1 2   0 ,   1 as an input and S a takes x y y = 1 ,   2 as an input. As a result of running the protocol, S a receives a challenged weight W x y and P 1 is unaware of S a ’s choice. As such, the protocol satisfies the privacy properties of S a , the privacy of P 1 , and the correctness of OT. Other variants of the OT protocol include Cut-and-Choose OT (CCOT) [66] and Outsourced OT (OOT) [67]. CCOT has proven to be secure in the presence of malicious adversaries [68]. In CCOT, the receiver receives the sender’s output values differently. Similarly, OOT has been used in cloud-assisted models involving three entities known as the sender, receiver, and cloud. The cloud obtains the receiver’s choice output values, and this protocol allows participants to preserve their privacy while outsourcing their work to the cloud. A high level of system interpretation is illustrated with Algorithm 2.
Algorithm 2: Dual Server Secure Aggregation and PoA Protocol
  • Input: Servers Sa and Sb, Participants Pi = ( P 1 , P 2 , , P n ), Global model WG, Local data LD, number of participants n, update W, number of training rounds K, Additive Secret-Shared (ASS) update of current round W i 1 , W i 2 , Shamir’s Secret Sharing SSS, Selected Participants SP, Proof of Aggregation PoA.
  • Output: Aggregated final model W.
1  Initialization N:
2  Procedure
3  for k = 1 ≤ k:
4     (I) Participants
5     for Pi Є Pn do:
6        (a) Split LD between Pi Є Pn using SSS
7  (b) Train WG using LD
8        // Weight Splitting
9  (c) Split each participant W using ASS into W 1 and W 2
10     Participants P i i   1 , n splits W i into W i a   and   W i b subject to W i a + W i b = W i
11     Such that W 1 + W 2 = W 1 1 + W 2 1 + W 1 2 + W 2 2
12  // OT Protocol
13         (d) Use OT protocol to send W i a to Sa and W i b to Sb
14      S a holds the weight W i a but does not know of W i 1 a
15      S b holds W i b and no knowledge of W i 1 b
16     end
17     // Poisoning attack detection
18     (II) Sa and Sb
19     for Sa do:
20  (a) Compute W x 1 i = 1 n W i 1
21  (b) Send W x 1 to Sb
22     end
23     for Sb do:
24  (a) Compute W x 2 i = 1 n W i 2
25  (d) Send W x 2 to Sa
26     end
27     // Both servers recover weight W i 12
28     Sa, Sb performs W i 12 = 1 2 n W x 1 + W x 2
29     Sb sends the computation of 1 2 W i 2 W i 12 to Sa
30     Sa sends the computation of 1 2 W i 1 W i 12 to Sb
31     // Distance (D) computation on both servers
32     for all w i w  do
33       Sa via D i = 1 2 W i 1 + 1 2 W i 2 W i 12
34       Sb via D i = 1 2 W i 2 + 1 2 W i 1 W i 12
35        D i D i
36     end
37     // Determination of median (M)
38     Sa and Sb gets the median M of D i via D i I = D W 1 , D W 2 ,   ,   D W n
39     for all W i W do:
40       Sa and Sb computes D i I I = D W i M
41     end
42     // Participants N with the smallest similarity are selected
43      S P i S P 1 ,   S P 2 ,   ,   S P N
44     // Secure aggregation and proof of PoA
45     Sa computes W 1 = 1 2 N i = 1 N W i 1 ,   P i S P
46     Sb computes W 2 = 1 2 N i = 1 N W i 2 ,   P i S P
47     Sa sends W 1 to Sb and SP; Sb sends W 2 to Sa and SP
48     Sa and Sb computes W = W 1 + W 2
49     Sa updates global model and sends W to Sb and (W, W 2 ) to SP for PoA
50     Sb updates its global model and sends W to Sa and (W, W 1 ) to SP for PoA
51     if W W 1 + W 2 :
52       reject PoA
53     else:
54       accept PoA
55  end
56  return Aggregated final model W

5.5. Poisoning Attack Detection

Our focus is on data poisoning [69] and model poisoning attacks [68] that are implemented in the training phase. In this scenario, attackers intentionally modify the weight parameters to poison the global model. We adopt the strategy deployed in [25] with modifications. We mitigated poisoning attacks using a weight selection and filtering scheme. In our approach, we computed the median value of the distances between the individual participants’ models and the average model after transforming it from a vector to a weight value. The difference between the weighted distance and the median is utilized as a similarity judging criterion. This process is used to identify poisoned updates based on their distance from their nearest neighbors, and it uses distance metrics. Considering a dataset d = U 1 , V 1 , U 2 , V 2 , , U n , V n where U i and V i are the feature vector and corresponding label, respectively. The process is as follows:
Distance metrics calculation: This involves the use of Euclidean distance to calculate the distance between test point U and all other training points. It also measures the deviation of each participant’s update from some central tendency measure (such as the mean or median) of all updates. A participant’s update is considered an anomaly if the update is too far from the central tendency.
Assume we have n participants, and each participant provides an updated vector W i for i 1,2 , , n
i.
Central Tendency Calculation
Using the median (M) as a central tendency measure, the median update vector ( ) can be computed as M = x n 2   + 1 for an odd number of observations and M = x n 2 +   x n 2   + 1 for an even number of observations.
ii.
Euclidean Distance Calculation
The Euclidean distance (D) of each update from the median could be calculated as:
D U , U i = j = 1 q U j U i j 2
where q = the number of features.
iii.
Malicious Threshold
A threshold τ needs to be determined. This threshold could be a fixed value or function of the distances beyond which an update is considered malicious. The threshold τ could also be determined by a percentile.
iv.
Malicious Update Detection
An update is considered malicious if D i > τ .
A participant’s update, whose distance from the median is greater than this threshold, would be considered an anomaly that could be investigated further or potentially discarded. For each participant update, a similarity measure using Euclidean distance is computed. This computation is the distance between the participant’s update and the median update from other participants. The purpose is to determine the proximity of each update to the central update that represents the median of all updates. This central update invariably represents the consensus of honest participants. Moreover, a baseline threshold is defined based on initial runs, signifying the highest permissible deviation from the central update. A similarity measure of the threshold for each participant update is carried out to determine potentially honest or suspicious participants. However, this does not truly represent the criteria for judging similarity, as outliers from large weights may exist. The basic criterion for judging similarity is explained in the next section. In high-dimensional data settings, iteratively, the dynamic adjustment of the threshold can be carried out over time. That is to say, if too few or too many updates are labeled malicious, the threshold can be adjusted based on the latest distribution.
Locate Nearest Weights: Identify the closest K training points to U based on the distance metrics.
(1)
Classification: Classify U as a poisoned update if a significant proportion of outliers within its K nearest weights exist.

5.6. Application in Federated Learning

To demonstrate its application in FL, let us consider a scenario where multiple participants train a global model with their local data. We aim to detect poisoned updates from the participants before aggregation.
Let W represent the global model.
Let i index each participant,
where P i Є 1 , 2 , , n .
Let the local model updates sent by each participant be represented by w . This is a vector that represents the change they propose to W.
Let the median of the updates be represented as M .
Let the Euclidean distance for each update be represented as D i .
For the local updates W i collection, poisoning attack detection, and aggregation by the server, the following steps ensued:
Step 1: Collection of local updates with privacy preserved.
Each participant trains the model locally and individually sends their updates W i to the servers using the ASS protocol.
Let the computed weight at S a be w x 1 = i = 1 n w i 1 and S b be w x 2 = i = 1 n w i 2 , respectively.
S a sends w x 1 to S b while S b sends w x 2 to S a .
To recover the weight w i 12 , both S a and S b each perform computation w i 12 = 1 2 n w x 1 + w x 2 .
S b sends the computation of 1 2 w i 2 w i 12 to S a and S a sends the computation 1 2 w i 1 w i 12 to S b .
Step 2: Poisoning attack detection.
Distance (D) computation on both servers.
S a via D i = 1 2 w i 1 + 1 2 w i 2 w i 12
S b via D i = 1 2 w i 2 + 1 2 w i 1 w i 12
Selecting a baseline model (m), the similarity between participants’ updates w i can be calculated by each server [ D w i , m =   w i m ] . To filter out boosted weight using the L2 norm, each participant provides a weight D i n d , with d as the dimension of the model parameter space. The two servers can compute the L2 norm of each participant’s update as D i 2 = 1 2 n k = 1 d D i , k 2 , where k is the index from 1 to d. This computed L2 norm gives a scalar measure of the overall magnitude of the updates. To define a suitable threshold ( τ ) that distinguishes normal updates from outliers, we computed the median (M) as M = D i 2   : i = 1,2 , , n and median deviation ( M d ) as M d = D i 2 M   : i = 1,2 , , n . Thus, τ = M + C . M d , where C is a constant C > 0 . This represents the true criterion to judge similarity and updates larger than τ are flagged as malicious. The threshold ( τ ) is also used to obtain the deviation of each participant’s update from the average model with new benign participants selected for secure aggregation.
Step 3: Secure aggregation and PoA
Let new selected participants (SP) be S P i = S P 1 , S P 2 , , S P N .
S a   c o m p u t e s     W 1 = 1 2 N i = 1 N w i 1 , P i S P
S b   c o m p u t e s   W 2 = 1 2 N i = 1 N w i 2 , P i S P
S a sends W 1 to S b and SP while S b sends W 2 to S a and SP. S a and S b compute the new aggregated weight W = W 1 + W 2 .
S a updates the global model and sends W to S b , and (W, W 2 ) to SP for PoA. Similarly, S b updates its global model and sends W to S a , and (W, W 1 ) to SP for PoA. The participants reject the result if W   W 1 + W 2   .

6. Security Analysis

In this section, we analyze the security of our framework and showcase that it meets the security goals and design objectives.
Theorem 1. 
Servers S a   a n d   S b know no information about the local data used to train the model.
Proof of Theorem 1. 
To securely share data, we adopt a dual-server architecture S y   y   a ,   b . For the weight W i and dataset D i , each participant P i i   1 , n splits W i into W i a   and   W i b subject to W i a + W i b = W i . P i sends W i y ,   D i to the y-th server. Server S a is required to compute W i a = i = 1 n W i a D i i = 1 n D i and server S b is required to compute W i b = i = 1 n W i b D i i = 1 n D i . Consequently, S a holds the weight W i a but does not know of W i 1 a while S b holds W i b and no knowledge of W i 1 b . Suppose S a is malicious, it sees only W i a : i = 1 ,   2 , n . Knowing of W i a alone does not reveal W i because, for any given W i b , there is a different W i . The confidentiality or integrity cannot be broken if one server is compromised. During poisoning attack detection, the weight at both servers = 1 2 n W i a + W 1 b . Distance computation at both servers reveals that each server obtains 1 2 W i a + 1 2 W i b W i a b . Without collision, no server has the full weight of any participant. □
Theorem 2. 
Our framework shows robustness against malicious participants.
Proof of Theorem 2. 
Each participant splits their weight vector W i into W i a   and   W i b such that W i a + W i b = W i . Each server would ideally compute partial aggregation. Server S a computes W 1 = i = 1 n W i a D i i = 1 n D i and server S b computes W 2 = i = 1 n W i b D i i = 1 n D i . The final global weight W G is reconstructed as W G =   W 1 + W 2 . To influence W G , a malicious participant aims to upload the poisoned weight W P . W P is split as W P a   a n d   W P b such that W P a + W P b = W P . The visibility of the full poisoned vector at any single server is reduced because of the splitting of the poisoned weights. Selecting a baseline model m , each server computes the nearest weight based on Euclidean distance [ D w i , m =   w i m ] .
For legitimate participants, D W i a , m D W i b , m , ensures coherent weights across servers. For malicious participants, D W P a ,   m D W i a , m , suggests a deviation. Furthermore, the L2 norm of each participant’s update at both servers is computed as D i 2 = 1 2 n k = 1 d D i , k 2 . A threshold ( τ ) is computed as τ = M + C . M d , with median (M) as M = D i 2   : i = 1,2 , , n , median deviation ( M d ) as M d = D i 2 M   : i = 1,2 , , n , and C as a constant C > 0 . The threshold ( τ ) represents the criterion for judging similarity and the deviation of each participant’s update from the average model. At both servers, new participants are selected S P i 1 ,   N , and outliers excluded from aggregation as: W i a = i P D i W i a i P D i and W i b = i P D i W i b i P D i . The influence of W P is entirely removed, and our framework shows robustness to poisoned updates. □
Theorem 3. 
Our framework satisfies PoA.
Proof of Theorem 3. 
After filtering malicious updates, a new participant S P i 1 ,   N is selected for model training. During training, each participant splits their weight W i subject to W i a + W i b = W i . S a would aggregate weight as follows: W 1 = i = 1 N W i a D i i = 1 N D i and S b as W 2 = i = 1 N W i b D i i = 1 N D i . The final global aggregation would be W G =   W 1 + W 2 . The computation would result in W G = i = 1 N W i a D i i = 1 N D i + i = 1 N W i b D i i = 1 N D i which equates to W G = i = 1 N W i D i i = 1 N D i . □
For PoA, let δ i ε 0,1 be the proportion that determines the split such that:
  W i a = δ i w i and   W i b = 1 δ i w i
  W G = i = 1 N W i a D i i = 1 N D i + i = 1 N W i b D i i = 1 N D i
Substitution would result in:
W G = i = 1 N D i δ i w i i = 1 N D i + i = 1 N D i 1 δ i w i i = 1 N D i
W G = i = 1 N D i δ i w i + i = 1 N D i 1 δ i w i i = 1 N D i
W G = i = 1 N D i δ i + 1 δ i w i i = 1 N D i
W G = i = 1 N W i D i i = 1 N D i
A participant would reject the result if W G   W 1 + W 2 . If all entities within the framework correctly execute the protocol, each selected participant will receive W G , partial aggregation at S a   W 1 and S b   W 2 . This will provide the participants with all specific and reconfirmation details to investigate the correctness of aggregation.
Theorem 4. 
Our framework exhibits a zero-trust architecture.
Proof of Theorem 4. 
As previously stated, a zero-trust architecture ensures no malicious entity acts within a federated learning environment. We have established that partial aggregation from S a is W 1 = i = 1 N W i a D i i = 1 N D i , at S b is W 2 = i = 1 N W i b D i i = 1 N D i , and the global aggregated model is W G =   W 1 + W 2 . No single entity is trusted within the framework, and each entity is treated as potentially compromised. Through structural guarantees, security is achieved, and no server has full information. Consequently, S a holds only W i a and S b only holds W i b . The zero-trust principle can be seen as a requirement that a subset of the framework cannot compute the final. Hence, the final global model is a function of F W i = W G . Let us define G 1 W i a = W i a , G 2 W i b = W i b . Then, F W i =   G 1 W i a +   G 2 W i b . Without both servers’ output, there is no transformation of H 1 or H 2 that permits H 1 G 1 W i a = W G or H 2 G 2 W i b = W G . No single entity within the framework is privileged to control the global result independently. To reinforce zero trust and prevent undetected attacks during the training process, each participant’s update must be audited for anomalies during each round of training. No incentive is given for performance, and equal scrutiny is applied to all participants. Moreover, the aggregated results from the two servers should be added to give the total aggregated weight (W). Furthermore, the protocol should ensure that each server sends its weight and the corresponding weight of the other server to the participants for verification and reconfirmation. A round is terminated if there is a variation in the weights received from the servers or if the weights were not added to give the total aggregated weight. □

7. Evaluation

In this section, we experimentally evaluate our framework to verify that it meets the design goals. Specifically, we have evaluated our framework for fidelity, robustness, and efficiency.
A.
Implementation
All experiments were run on a High-Performance Computing (HPC) cluster configured with Linux CentOS. A Dell C6420 PowerEdge with compute nodes of 2 Intel Xeon Platinum, 48 cores per node, 8640 compute cores, 2.9 GHz 24-core processor, 180 nodes, 192 GB RAM per node, and 4 GB RAM per core. The experiment is implemented in Python 3.12 using TensorFlow Federated (TFF) libraries.
B.
Experimental Setup
(1)
Datasets: We evaluated our framework with two different domain benchmark datasets, which are MNIST [70] and CIFAR [71].
  • MNIST dataset: The MNIST dataset is among the most popular datasets used in ML and computer vision. As a standardized dataset, it consists of handwritten digits 0–9 as grayscale images sized 28 × 28 pixels. It has training and test sets of 60,000 and 10,000 images, respectively. We utilize CNN and MLP models to experiment with the IID and non-IID datasets.
  • CIFAR-10 dataset: The CIFAR-10 dataset is a 32 × 32-pixel image consisting of 10 different classes of color images of objects with a training set of 50,000 and a test set of 10,000 images. It has a total of 3072 pixels per image (32 × 32 × 3) with integer values of 0–9 representing the object class. We also use CNN and MLP models
(2)
Data Splits: We experimented by splitting each participant’s dataset into IID and non-IID. For IID, the total number of MNIST and CIFAR dataset samples were shuffled and divided evenly among participants. For non-IID, the data distribution is based on class where each participant may possess data from a few classes. To create class-specific data distributions, we utilized a stratified sampling method based on class labels.
(3)
Baselines: As benchmarks, we utilized a standard aggregation rule, a poisoning attack detection scheme, an SMC protocol with filtering and clipping method, and a secret-sharing secure aggregation scheme resilient to poisoning attacks.
  • FedAvg [57]: This standard algorithm is used in FL and is resistant to poisoning attacks without any privacy preservation method.
  • Krum [72]: A poisoning attack detection method in FL that optimally selects an update from returned post-trained models and uses the updates’ Euclidean distance as a baseline of comparison with other received updates.
  • EPPRFL [24]: A resistant poisoning attack privacy preservation scheme that utilizes SMC protocols for filtering and clipping poisoned updates.
  • LSFL [25]: An ASS scheme for privacy preservation as well as a K-nearest weight selection aggregation method that is robust to Byzantine participants.
(4)
Hyper-Parameters: In our setup, we conducted experiments to test our framework using the following cases: MNIST CNN IID, MNIST CNN non-IID, MNIST MLP IID, MNIST MLP non-IID, CIFAR CNN IID, CIFAR CNN non-IID, CIFAR MLP IID, and CIFAR MLP non-IID. The learning rate for local training was set at 0.0001, an epoch as 5, and a batch size of 32 was chosen while the server learning rate was set to 0.5. For each round of training in FL, we had at least 70% of selected participants N for model training.
(5)
Poisoning Attack: We considered the following poisoning attacks such as sign-flipping, label-flipping, backdoor, backdoor embedding, and gradient manipulation attacks. We carried out experiments to demonstrate the resilience and robustness of our framework to these attacks and conducted a comparison with some SOTA research
C.
Experimental Results
We performed experiments using the hyperparameters listed above to evaluate our framework in terms of fidelity, robustness, and efficiency.
(1)
Fidelity Evaluation: We compared FedAvg and our framework in terms of accuracy and loss to evaluate the fidelity goal when there are no poisoning attackers. In this experiment, the number of participants and communication rounds was set to 100, respectively, with the evaluation carried out using the MNIST and CIFAR-10 datasets over the CNN and MLP models in both IID and non-IID settings. The convergence of our framework and the FedAvg are shown in Figure 2 and Figure 3, using the MNIST and CIFAR-10 datasets over the CNN and MLP models in both IID and non-IID settings, respectively. The two figures illustrate the comparable convergence performance of our framework and a standard FedAvg in terms of training loss and the number of training rounds using both the MNIST and CIFAR-10 datasets over the CNN and MLP models in IID and non-IID settings. In terms of convergence, our framework has no impact on the model’s performance and, as observed, it yields approximately the same result as a standard FedAVG without any privacy-preserving mechanisms or poisoning attack-resilient methods. The results show that our framework does not impact model performance but performs equally well as a standard FedAvg offering additional benefits such as stability and robustness.
Figure 2. The convergence of our framework and the FedAvg on MNIST (CNN and MLP) in IID and non-IID settings. (a) MNIST CNN IID, (b) MNIST CNN non-IID, (c) MNIST MLP IID, and (d) MNIST MLP non-IID.
Figure 3. The convergence of our framework and the FedAVG on CIFAR-10 (CNN and MLP) in IID and non-IID settings. (a) CIFAR-10 CNN IID, (b) CIFAR-10 CNN non-IID, (c) CIFAR-10 MLP IID, and (d) CIFAR-10 MLP non-IID.
Similarly, we demonstrate that the accuracy of our framework is approximately consistent with that of FedAVG, as shown in Figure 4 and Figure 5. using the MNIST and CIFAR-10 datasets over CNN and MLP models in both IID and non-IID settings, respectively. Our framework has no effect on accuracy, as it is consistent with that of FedAVG on both the MNIST (CNN and MLP) and CIFAR10 (CNN and MLP) datasets across IID and non-IID settings with both methods converging to the same accuracy. These figures demonstrate that our framework has no impact on the model’s accuracy and effectively preserves participants’ privacy.
Figure 4. Accuracy curves of our framework and FedAvg on MNIST CNN (IID and non-IID) and MNIST MLP (IID and non-IID). (a) MNIST CNN IID, (b) MNIST CNN non-IID, (c) MNIST MLP IID, and (d) MNIST MLP non-IID.
Figure 5. Accuracy curves of our framework and FedAvg on CIFAR10 CNN (IID and non-IID) and MLP (IID and non-IID). (a) CIFAR10 CNN IID, (b) CIFAR10 CNN non-IID, (c) CIFAR10 MLP IID, and (d) CIFAR10 MLP non-IID.
(2)
Robustness Evaluation: We considered label flipping, sign flipping, backdoor, backdoor embedding, and gradient manipulation attacks on the MNIST and CIFAR-10 datasets in both IID and non-IID settings. The effectiveness of these attacks was evaluated by their impact on the model’s accuracy. First, we considered the attacks in situations where 10%, 20%, 30%, and 40% of the participants were poisoning attackers and evaluated the impact on the model’s accuracy. Figure 6, Figure 7, Figure 8, Figure 9 and Figure 10 illustrate the accuracy of our framework with varying numbers of label flipping, sign flipping, backdoor, backdoor embedding, and gradient manipulation attacks, respectively, using the MNIST and CIFAR-10 datasets in both IID and non-IID settings. Our framework demonstrated graceful degradation in accuracy as the attack fraction increased, maintaining an accuracy level of above 60% on all attacks, even at a 40% attack fraction, suggesting that the model is resilient to data heterogeneity and robust to poisoning attacks.
Figure 6. Accuracy with label flipping attack. (a) Label flipping attack MNIST CNN. (b) Label flipping attack MNIST MLP. (c) Label flipping attack CIFAR10 CNN (d) Label flipping attack CIFAR10 MLP.
Figure 7. Accuracy with sign flipping attack. (a) Sign flipping attack MNIST CNN (b) Sign flipping attack MNIST MLP (c) Sign flipping attack CIFAR10 CNN (d) Sign flipping attack CIFAR10 MLP.
Figure 8. Accuracy with backdoor attack. (a) Backdoor attack MNIST CNN (b) Backdoor attack MNIST MLP (c) Backdoor attack CIFAR10 CNN (d) Backdoor attack CIFAR10 MLP.
Figure 9. Accuracy with backdoor embedding attack. (a) Backdoor embedding MNIST CNN; (b) backdoor embedding MNIST MLP; (c) backdoor embedding CIFAR10 CNN; (d) backdoor embedding CIFAR10 MLP.
Figure 10. Accuracy with gradient manipulation attack. (a) Gradient manipulation MNIST CNN; (b) gradient manipulation MNIST MLP; (c) gradient manipulation CIFAR10 CNN; and (d) gradient manipulation CIFAR10 MLP.
Secondly, we compared our framework with schemes such as EPPRFL, LSFL, and KRUM using various attack fractions, as previously stated. We maintained the number of participants at 100 and the number of training rounds at 100. The results show that our framework achieves higher accuracy in most instances, even in the presence of varying attack fractions, as shown in Figure 11, Figure 12, Figure 13, Figure 14, Figure 15 and Figure 16.
Figure 11. Label flipping attack over MNIST dataset in IID and non-IID settings. (a) MNIST CNN (IID), (b) MNIST CNN (non-IID), (c) MNIST MLP (IID), and (d) MNIST MLP (non-IID).
Figure 12. Label flipping attack over CIFAR-10 dataset in IID and non-IID settings. (a) CIFAR10 CNN (IID), (b) CIFAR10 CNN (non-IID), (c) CIFAR10 MLP (IID), and (d) CIFAR10 MLP (non-IID).
Figure 13. Sign flipping attack over MNIST dataset in IID and non-IID settings. (a) MNIST CNN (IID), (b) MNIST CNN (non-IID), (c) MNIST MLP (IID), and (d) MNIST MLP (non-IID).
Figure 14. Sign flipping attack over CIFAR-10 dataset in IID and non-IID settings. (a) CIFAR-10 CNN (IID), (b) CIFAR-10 CNN (non-IID), (c) CIFAR-10 MLP (IID), and (d) CIFAR-10 MLP (non-IID).
Figure 15. Backdoor attack over MNIST dataset in IID and non-IID settings. (a) Backdoor CNN (IID), (b) Backdoor CNN (non-IID), (c) Backdoor MLP (IID), and (d) Backdoor MLP (non-IID).
Figure 16. Backdoor attack over CIFAR-10 dataset (a) CNN (IID), (b) CNN (non-IID), (c) MLP (IID), and (d) MLP (non-IID).
The experimental results regarding label flipping, sign flipping, and backdoor attacks over MNIST and CIFAR-10 datasets in IID and non-IID settings are shown in Figure 11, Figure 12, Figure 13, Figure 14, Figure 15 and Figure 16. We compared our results with LSFL, KRUM, and EPPRFL in cases where 10%, 20%, 30%, and 40% of participants were subjected to poisoning attacks, respectively. The experimental results indicate that our framework outperforms LSFL, KRUM, and EPPRFL in certain cases while maintaining relatively stable model accuracy in situations involving high levels of poisoning attacks. The performance of our framework is attributed to the use of the median and median deviation as metric information, which is robust to outliers. Both metrics were used to compute a suitable threshold that distinguishes normal updates from outliers. This represents the true criterion for judging similarity and updates larger than the threshold are flagged as malicious. The results indicate that our framework can detect varying fractions of poisoning attacks in both IID and non-IID settings.
(3)
Efficiency Evaluation: To evaluate the efficiency of our framework, we measured and compared the latency of FedAVG and our framework for various participants, providing a real-world efficiency comparison. Therefore, we tested FedAVG and our framework with varying numbers of participants (10, 50, 100, 150, and 200), and the result showed no difference despite the robustness of our framework. Figure 17 shows latency iteration over a varying number of participants. The average latency per round is the ratio of the total time to the number of rounds. This is conducted using MNIST and CIFAR10 datasets over the CNN model in an IID setting with the number of rounds set at 20. Moreover, we demonstrate that the communication overheads of our framework do not result in a drastic increase as the number of participants increases. Instead, it linearly increases with an increased number of participants, as shown in Figure 18. This implies that our framework would work with large-scale deployments. Finally, we show our framework’s computation cost per round in Figure 19 to reflect real-world performance.
Figure 17. Latency comparison between FedAvg and our framework.
Figure 18. Communication costs for different numbers of participants.
Figure 19. Computation cost per round.
Additionally, participant diversity has no significant impact on our framework performance regarding privacy preservation and poisoning attack resilience. Our framework performed optimally when compared to FedAvg and other SOTA schemes in terms of privacy and robustness against diverse poisoning attacks.

8. Conclusions

In this paper, we propose a framework for privacy preservation, defense against poisoning attacks, and verification of the correctness of aggregation in the federated learning environment. The framework’s robustness tends to ensure that no malicious entity acts successfully, thereby satisfying a zero-trust architectural framework. The privacy of the local data used to train the model is preserved by additive secret sharing that splits the weights between the two servers. The splitting is carried out using oblivious transfer protocol, which guarantees that a malicious server cannot infer the private data used to train the model. Our poisoning attack detection method is based on a weight selection and filtering scheme that uses distance metrics and L2 defense. We provided a means of verifying the correctness of aggregated results from the servers. Participants are to reject and terminate a round if there is a discrepancy in the aggregated results returned by the servers, guaranteeing a zero-trust architecture. Theoretical analysis and experimental results show that our framework can achieve an accuracy level almost the same as FedAvg with the privacy of the local data preserved and not incurring additional computational burden. Therefore, our framework demonstrates a robust defense system that prevents malicious parties from acting, enhancing confidentiality, integrity, and the authentication of device–server communication in a federated learning environment. Future directions would be to preserve privacy and improve poisoning attack resilience in the potential collusion of two entities within the framework.

Author Contributions

Conceptualization, W.E.M.; methodology, W.E.M.; software, W.E.M.; validation, W.E.M., C.M. and G.E.; formal analysis, W.E.M.; investigation, W.E.M.; resources, W.E.M.; data curation, W.E.M.; writing—original draft preparation, W.E.M.; writing—review and editing, W.E.M., C.M. and G.E.; supervision, C.M., G.E. and C.P.; funding acquisition, C.M. and G.E. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Hub for Edge AI EP/Y028813/1. Academic Centre of Excellence in Cyber Security Research—University of Warwick EP/R007195/1.

Data Availability Statement

washington 24/Machine-Learning-Codes.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Abbas, Z.; Ahmad, S.F.; Syed, M.H.; Anjum, A.; Rehman, S. Exploring Deep Federated Learning for the Internet of Things: A GDPR-Compliant Architecture. IEEE Access 2024, 12, 10548–10574. [Google Scholar] [CrossRef]
  2. Zhang, M.; Meng, W.; Zhou, Y.; Ren, K. CSCHECKER: Revisiting GDPR and CCPA Compliance of Cookie Banners on the Web. In Proceedings of the IEEE/ACM 46th International Conference on Software Engineering, Lisbon, Portugal, 14–20 April 2024; IEEE Computer Society: Washington, DC, USA, 2024; pp. 2147–2158. [Google Scholar] [CrossRef]
  3. Baskaran, H.; Yussof, S.; Abdul Rahim, F.; Abu Bakar, A. Blockchain and The Personal Data Protection Act 2010 (PDPA) in Malaysia. In Proceedings of the 2020 8th International Conference on Information Technology and Multimedia (ICIMU), Selangor, Malaysia, 24–26 August 2020; IEEE: Piscataway, NJ, USA, 2020. [Google Scholar]
  4. Yuan, L.; Wang, Z.; Sun, L.; Yu, P.S.; Brinton, C.G. Decentralized Federated Learning: A Survey and Perspective. IEEE Internet Things J. 2024, 11, 34617–34638. [Google Scholar] [CrossRef]
  5. Rauniyar, A.; Hagos, D.H.; Jha, D.; Håkegård, J.E.; Bagci, U.; Rawat, D.B.; Vlassov, V. Federated Learning for Medical Applications: A Taxonomy, Current Trends, Challenges, and Future Research Directions. IEEE Internet Things J. 2024, 11, 7374–7398. [Google Scholar] [CrossRef]
  6. Tang, Y.; Liu, Z. A Credit Card Fraud Detection Algorithm Based on SDT and Federated Learning. IEEE Access 2024, 12, 182547–182560. [Google Scholar] [CrossRef]
  7. Parekh, R.; Patel, N.; Gupta, R.; Jadav, N.K.; Tanwar, S.; Alharbi, A.; Tolba, A.; Neagu, B.C.; Raboaca, M.S. GeFL: Gradient Encryption-Aided Privacy Preserved Federated Learning for Autonomous Vehicles. IEEE Access 2023, 11, 1825–1839. [Google Scholar] [CrossRef]
  8. Mothukuri, V.; Khare, P.; Parizi, R.M.; Pouriyeh, S.; Dehghantanha, A.; Srivastava, G. Federated-Learning-Based Anomaly Detection for IoT Security Attacks. IEEE Internet Things J. 2022, 9, 2545–2554. [Google Scholar] [CrossRef]
  9. Li, Q.; Wen, Z.; Wu, Z.; Hu, S.; Wang, N.; Li, Y.; Liu, X.; He, B. A Survey on Federated Learning Systems: Vision, Hype and Reality for Data Privacy and Protection. IEEE Trans. Knowl. Data Eng. 2023, 35, 3347–3366. [Google Scholar] [CrossRef]
  10. Jawadur Rahman, K.M.; Ahmed, F.; Akhter, N.; Hasan, M.; Amin, R.; Aziz, K.E.; Islam, A.M.; Mukta, M.S.H.; Islam, A.N. Challenges, Applications and Design Aspects of Federated Learning: A Survey. IEEE Access 2021, 9, 124682–124700. [Google Scholar] [CrossRef]
  11. Xu, G.; Li, H.; Zhang, Y.; Xu, S.; Ning, J.; Deng, R.H. Privacy-Preserving Federated Deep Learning with Irregular Users. IEEE Trans. Dependable Secure Comput. 2022, 19, 1364–1381. [Google Scholar] [CrossRef]
  12. Wang, Z.; Song, M.; Zhang, Z.; Song, Y.; Wang, Q.; Qi, H. Beyond Inferring Class Representatives: User-Level Privacy Leakage from Federated Learning. arXiv 2018, arXiv:1812.00535. [Google Scholar]
  13. Zhao, B.; Mopuri, K.R.; Bilen, H. iDLG: Improved Deep Leakage from Gradients. arXiv 2020, arXiv:2001.02610. [Google Scholar]
  14. Zhang, J.; Chen, B.; Cheng, X.; Binh, H.T.T.; Yu, S. PoisonGAN: Generative Poisoning Attacks against Federated Learning in Edge Computing Systems. IEEE Internet Things J. 2021, 8, 3310–3322. [Google Scholar] [CrossRef]
  15. Xia, G.; Chen, J.; Yu, C.; Ma, J. Poisoning Attacks in Federated Learning: A Survey. IEEE Access 2023, 11, 10708–10722. [Google Scholar] [CrossRef]
  16. Sun, G.; Cong, Y.; Dong, J.; Wang, Q.; Lyu, L.; Liu, J. Data Poisoning Attacks on Federated Machine Learning. IEEE Internet Things J. 2022, 9, 11365–11375. [Google Scholar] [CrossRef]
  17. Xu, G.; Li, H.; Liu, S.; Yang, K.; Lin, X. VerifyNet: Secure and Verifiable Federated Learning. IEEE Trans. Inf. Forensics Secur. 2020, 15, 911–926. [Google Scholar] [CrossRef]
  18. Sav, S.; Pyrgelis, A.; Troncoso-Pastoriza, J.R.; Froelicher, D.; Bossuat, J.P.; Sousa, J.S.; Hubaux, J.P. POSEIDON: Privacy-Preserving Federated Neural Network Learning. arXiv 2020, arXiv:2009.00349. [Google Scholar]
  19. Byrd, D.; Polychroniadou, A. Differentially private secure multi-party computation for federated learning in financial applications. In Proceedings of the ICAIF 2020—1st ACM International Conference on AI in Finance, New York, NY, USA, 15–16 October 2020; Association for Computing Machinery, Inc.: New York, NY, USA, 2021. [Google Scholar] [CrossRef]
  20. Cao, D.; Chang, S.; Lin, Z.; Liu, G.; Sun, D. Understanding distributed poisoning attack in federated learning. In Proceedings of the International Conference on Parallel and Distributed Systems—ICPADS, Tianjin, China, 4–6 December 2019; IEEE Computer Society: Washington, DC, USA, 2019; pp. 233–239. [Google Scholar] [CrossRef]
  21. Cao, X.; Fang, M.; Liu, J.; Gong, N.Z. FLTrust: Byzantine-robust Federated Learning via Trust Bootstrapping. In Proceedings of the 28th Annual Network and Distributed System Security Symposium, NDSS 2021, Virtual, 21–25 February 2021; The Internet Society: Reston, VA, USA, 2021. [Google Scholar] [CrossRef]
  22. Fu, A.; Zhang, X.; Xiong, N.; Gao, Y.; Wang, H.; Zhang, J. VFL: A Verifiable Federated Learning with Privacy-Preserving for Big Data in Industrial IoT. IEEE Trans. Ind. Inform. 2022, 18, 3316–3326. [Google Scholar] [CrossRef]
  23. Bagdasaryan, E.; Veit, A.; Hua, Y.; Estrin, D.; Shmatikov, V. How To Backdoor Federated Learning. arXiv 2018, arXiv:1807.00459. [Google Scholar]
  24. Li, X.; Yang, X.; Zhou, Z.; Lu, R. Efficiently Achieving Privacy Preservation and Poisoning Attack Resistance in Federated Learning. IEEE Trans. Inf. Forensics Secur. 2024, 19, 4358–4373. [Google Scholar] [CrossRef]
  25. Zhang, Z.; Wu, L.; Ma, C.; Li, J.; Wang, J.; Wang, Q.; Yu, S. LSFL: A Lightweight and Secure Federated Learning Scheme for Edge Computing. IEEE Trans. Inf. Forensics Secur. 2023, 18, 365–379. [Google Scholar] [CrossRef]
  26. Rathee, M.; Shen, C.; Wagh, S.; Popa, R.A. ELSA: Secure Aggregation for Federated Learning with Malicious Actors. In Proceedings of the 2023 IEEE Symposium on Security and Privacy (SP), San Francisco, CA, USA, 21–25 May 2023. [Google Scholar] [CrossRef]
  27. Wahab, O.A.; Mourad, A.; Otrok, H.; Taleb, T. Federated Machine Learning: Survey, Multi-Level Classification, Desirable Criteria and Future Directions in Communication and Networking Systems. IEEE Commun. Surv. Tutor. 2021, 23, 1342–1397. [Google Scholar] [CrossRef]
  28. Zhao, Z.; Liu, D.; Cao, Y.; Chen, T.; Zhang, S.; Tang, H. U-shaped Split Federated Learning: An Efficient Cross-Device Learning Framework with Enhanced Privacy-preserving. In Proceedings of the 2023 9th International Conference on Computer and Communications, ICCC 2023, Chengdu, China, 8–11 December 2023; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2023; pp. 2182–2186. [Google Scholar] [CrossRef]
  29. Li, B.; Shi, Y.; Guo, Y.; Kong, Q.; Jiang, Y. Incentive and Knowledge Distillation Based Federated Learning for Cross-Silo Applications. In Proceedings of the INFOCOM WKSHPS 2022—IEEE Conference on Computer Communications Workshops, New York, NY, USA, 2–5 May 2022; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2022. [Google Scholar] [CrossRef]
  30. Iqbal, M.; Tariq, A.; Adnan, M.; Din, I.U.; Qayyum, T. FL-ODP: An Optimized Differential Privacy Enabled Privacy Preserving Federated Learning. IEEE Access 2023, 11, 116674–116683. [Google Scholar] [CrossRef]
  31. Zhang, G.; Liu, B.; Zhu, T.; Ding, M.; Zhou, W. PPFed: A Privacy-Preserving and Personalized Federated Learning Framework. IEEE Internet Things J. 2024, 11, 19380–19393. [Google Scholar] [CrossRef]
  32. Ye, D.; Shen, S.; Zhu, T.; Liu, B.; Zhou, W. One Parameter Defense—Defending Against Data Inference Attacks via Differential Privacy. IEEE Trans. Inf. Forensics Secur. 2022, 17, 1466–1480. [Google Scholar] [CrossRef]
  33. Hijazi, N.M.; Aloqaily, M.; Guizani, M.; Ouni, B.; Karray, F. Secure Federated Learning with Fully Homomorphic Encryption for IoT Communications. IEEE Internet Things J. 2024, 11, 4289–4300. [Google Scholar] [CrossRef]
  34. Zhang, L.; Xu, J.; Vijayakumar, P.; Sharma, P.K.; Ghosh, U. Homomorphic Encryption-Based Privacy-Preserving Federated Learning in IoT-Enabled Healthcare System. IEEE Trans. Netw. Sci. Eng. 2023, 10, 2864–2880. [Google Scholar] [CrossRef]
  35. Cai, Y.; Ding, W.; Xiao, Y.; Yan, Z.; Liu, X.; Wan, Z. SecFed: A Secure and Efficient Federated Learning Based on Multi-Key Homomorphic Encryption. IEEE Trans. Dependable Secure Comput. 2024, 21, 3817–3833. [Google Scholar] [CrossRef]
  36. Cao, X.K.; Wang, C.D.; Lai, J.H.; Huang, Q.; Chen, C.L.P. Multiparty Secure Broad Learning System for Privacy Preserving. IEEE Trans. Cybern. 2023, 53, 6636–6648. [Google Scholar] [CrossRef] [PubMed]
  37. Sotthiwat, E.; Zhen, L.; Li, Z.; Zhang, C. Partially encrypted multi-party computation for federated learning. In Proceedings of the 21st IEEE/ACM International Symposium on Cluster, Cloud and Internet Computing, CCGrid 2021, Melbourne, Australia, 10–13 May 2021; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2021; pp. 828–835. [Google Scholar] [CrossRef]
  38. Mbonu, W.E.; Maple, C.; Epiphaniou, G. An End-Process Blockchain-Based Secure Aggregation Mechanism Using Federated Machine Learning. Electronics 2023, 12, 4543. [Google Scholar] [CrossRef]
  39. Dragos, L.; Togan, M. Privacy-Preserving Machine Learning Using Federated Learning and Secure Aggregation. In Proceedings of the 12th International Conference on Electronics, Computers and Artificial Intelligence—ECAI-2020, Bucharest, Romania, 25–27 June 2020; IEEE: Piscataway, NJ, USA, 2020. [Google Scholar]
  40. Ma, Z.; Ma, J.; Miao, Y.; Li, Y.; Deng, R.H. ShieldFL: Mitigating Model Poisoning Attacks in Privacy-Preserving Federated Learning. IEEE Trans. Inf. Forensics Secur. 2022, 17, 1639–1654. [Google Scholar] [CrossRef]
  41. Liu, J.; Li, X.; Liu, X.; Zhang, H.; Miao, Y.; Deng, R.H. DefendFL: A Privacy-Preserving Federated Learning Scheme Against Poisoning Attacks. IEEE Trans. Neural Netw. Learn. Syst. 2024, 1–14. [Google Scholar] [CrossRef]
  42. Khazbak, Y.; Tan, T.; Cao, G. MLGuard: Mitigating Poisoning Attacks in Privacy Preserving Distributed Collaborative Learning. In Proceedings of the ICCCN 2020: The 29th International Conference on Computer Communication and Networks: Final Program, Honolulu, HI, USA, 3–6 August 2020; IEEE: Piscataway, NJ, USA, 2020. [Google Scholar]
  43. Liu, X.; Li, H.; Xu, G.; Chen, Z.; Huang, X.; Lu, R. Privacy-Enhanced Federated Learning against Poisoning Adversaries. IEEE Trans. Inf. Forensics Secur. 2021, 16, 4574–4588. [Google Scholar] [CrossRef]
  44. Li, X.; Wen, M.; He, S.; Lu, R.; Wang, L. A Privacy-Preserving Federated Learning Scheme Against Poisoning Attacks in Smart Grid. IEEE Internet Things J. 2024, 11, 16805–16816. [Google Scholar] [CrossRef]
  45. Sakazi, I.; Grolman, E.; Elovici, Y.; Shabtai, A. STFL: Utilizing a Semi-Supervised, Transfer-Learning, Federated-Learning Approach to Detect Phishing URL Attacks. In Proceedings of the 2024 International Joint Conference on Neural Networks (IJCNN), Yokohama, Japan, 30 June–5 July 2024; Institute of Electrical and Electronics Engineers (IEEE): Piscataway, NJ, USA, 2024; pp. 1–10. [Google Scholar] [CrossRef]
  46. Liu, Y.; Yuan, X.; Zhao, R.; Wang, C.; Niyato, D.; Zheng, Y. Poisoning Semi-supervised Federated Learning via Unlabeled Data: Attacks and Defenses. arXiv 2020, arXiv:2012.04432. [Google Scholar]
  47. Xu, R.; Li, B.; Li, C.; Joshi, J.B.; Ma, S.; Li, J. TAPFed: Threshold Secure Aggregation for Privacy-Preserving Federated Learning. IEEE Trans. Dependable Secure Comput. 2024, 21, 4309–4323. [Google Scholar] [CrossRef]
  48. Yazdinejad, A.; Dehghantanha, A.; Karimipour, H.; Srivastava, G.; Parizi, R.M. A Robust Privacy-Preserving Federated Learning Model Against Model Poisoning Attacks. IEEE Trans. Inf. Forensics Secur. 2024, 19, 6693–6708. [Google Scholar] [CrossRef]
  49. Hao, M.; Li, H.; Xu, G.; Chen, H.; Zhang, T. Efficient, Private and Robust Federated Learning. In Proceedings of the ACM International Conference Proceeding Series, ACSAC’21: Annual Computer Security Applications Conference, Virtual Event, 6–10 December 2021; Association for Computing Machinery: New York, NY, USA, 2021; pp. 45–60. [Google Scholar] [CrossRef]
  50. Dong, Y.; Chen, X.; Li, K.; Wang, D.; Zeng, S. FLOD: Oblivious Defender for Private Byzantine-Robust Federated Learning with Dishonest-Majority. In Computer Security–ESORICS 2021, Proceedings of the 26th European Symposium on Research in Computer Security, Darmstadt, Germany, 4–8 October 2021; Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics; Springer Science and Business Media Deutschland GmbH: Cham, Switzerland, 2021; pp. 497–518. [Google Scholar] [CrossRef]
  51. Kumar, S.S.; Joshith, T.S.; Lokesh, D.D.; Jahnavi, D.; Mahato, G.K.; Chakraborty, S.K. Privacy-Preserving and Verifiable Decentralized Federated Learning. In Proceedings of the 5th International Conference on Energy, Power, and Environment: Towards Flexible Green Energy Technologies, ICEPE 2023, Shillong, India, 15–17 June 2023; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2023. [Google Scholar] [CrossRef]
  52. Liu, T.; Xie, X.; Zhang, Y. ZkCNN: Zero Knowledge Proofs for Convolutional Neural Network Predictions and Accuracy. In Proceedings of the ACM Conference on Computer and Communications Security, Virtual Event, 15–19 November 2021; Association for Computing Machinery: New York, NY, USA, 2021; pp. 2968–2985. [Google Scholar] [CrossRef]
  53. Wang, Z.; Dong, N.; Sun, J.; Knottenbelt, W.; Guo, Y. zkFL: Zero-Knowledge Proof-based Gradient Aggregation for Federated Learning. IEEE Trans. Big Data 2024, 11, 447–460. [Google Scholar] [CrossRef]
  54. Sun, X.; Yu, F.R.; Zhang, P.; Sun, Z.; Xie, W.; Peng, X. A Survey on Zero-Knowledge Proof in Blockchain. IEEE Netw. 2021, 35, 198–205. [Google Scholar] [CrossRef]
  55. Nasr, M.; Shokri, R.; Houmansadr, A. Comprehensive privacy analysis of deep learning: Passive and active white-box inference attacks against centralized and federated learning. In Proceedings of the IEEE Symposium on Security and Privacy, San Francisco, CA, USA, 19–23 May 2019; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2019; pp. 739–753. [Google Scholar] [CrossRef]
  56. Akhtar, N.; Mian, A.; Kardan, N.; Shah, M. Advances in adversarial attacks and defenses in computer vision: A survey. arXiv 2021, arXiv:2108.00401. [Google Scholar]
  57. Brendan McMahan, H.; Moore, E.; Ramage, D.; Hampson, S.; Agüera y Arcas, B. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, AISTATS 2017, Fort Lauderdale, FL, USA, 20–22 April 2017; PMLR: Cambridge, MA, USA, 2017. [Google Scholar]
  58. Tan, L.; Yu, K.; Yang, C.; Bashir, A.K. A blockchain-based Shamir’s threshold cryptography for data protection in industrial internet of things of smart city. In Proceedings of the 6G-ABS 2021—Proceedings of the 1st ACM Workshop on Artificial Intelligence and Blockchain Technologies for Smart Cities with 6G, Part of ACM MobiCom 2021, New Orleans, LA, USA, 25–29 October 2021; Association for Computing Machinery, Inc.: New York, NY, USA, 2021; pp. 13–18. [Google Scholar] [CrossRef]
  59. Tang, J.; Xu, H.; Wang, M.; Tang, T.; Peng, C.; Liao, H. A Flexible and Scalable Malicious Secure Aggregation Protocol for Federated Learning. IEEE Trans. Inf. Forensics Secur. 2024, 19, 4174–4187. [Google Scholar] [CrossRef]
  60. Zhang, L.; Xia, R.; Tian, W.; Cheng, Z.; Yan, Z.; Tang, P. FLSIR: Secure Image Retrieval Based on Federated Learning and Additive Secret Sharing. IEEE Access 2022, 10, 64028–64042. [Google Scholar] [CrossRef]
  61. Fazli Khojir, H.; Alhadidi, D.; Rouhani, S.; Mohammed, N. FedShare: Secure Aggregation based on Additive Secret Sharing in Federated Learning. In ACM International Conference Proceeding Series, Proceedings of the IDEAS’23: International Database Engineered Applications Symposium, Heraklion, Greece, 5–7 May 2023; Association for Computing Machinery: New York, NY, USA, 2023; pp. 25–33. [Google Scholar] [CrossRef]
  62. More, Y.; Ramachandran, P.; Panda, P.; Mondal, A.; Virk, H.; Gupta, D. SCOTCH: An Efficient Secure Computation Framework for Secure Aggregation. arXiv 2022, arXiv:2201.07730. [Google Scholar]
  63. Asare, B.A.; Branco, P.; Kiringa, I.; Yeap, T. AddShare+: Efficient Selective Additive Secret Sharing Approach for Private Federated Learning. In Proceedings of the 2024 IEEE 11th International Conference on Data Science and Advanced Analytics (DSAA), San Diego, CA, USA, 6–10 October 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1–10. [Google Scholar] [CrossRef]
  64. Wang, X.; Kuang, X.; Li, J.; Li, J.; Chen, X.; Liu, Z. Oblivious Transfer for Privacy-Preserving in VANET’s Feature Matching. IEEE Trans. Intell. Transp. Syst. 2021, 22, 4359–4366. [Google Scholar] [CrossRef]
  65. Long, H.; Zhang, S.; Zhang, Y.; Zhang, L.; Wang, J. A Privacy-Preserving Method Based on Server-Aided Reverse Oblivious Transfer Protocol in MCS. IEEE Access 2019, 7, 164667–164681. [Google Scholar] [CrossRef]
  66. Wei, X.; Xu, L.; Wang, H.; Zheng, Z. Permutable Cut-and-Choose Oblivious Transfer and Its Application. IEEE Access 2020, 8, 17378–17389. [Google Scholar] [CrossRef]
  67. Traugott, S. Why Order Matters: Turing Equivalence in Automated Systems Administration. In Proceedings of the Sixteenth Systems Administration Conference: (LISA XVI), Philadelphia, PA, USA, 3–8 November 2000; USENIX Association: Berkeley, CA, USA, 2002. [Google Scholar]
  68. Wei, X.; Jiang, H.; Zhao, C.; Zhao, M.; Xu, Q. Fast Cut-and-Choose Bilateral Oblivious Transfer for Malicious Adversaries. In Proceedings of the 2016 IEEE Trustcom/BigDataSE/ISPA, Tianjin, China, 23–26 August 2016. [Google Scholar] [CrossRef]
  69. Fan, J.; Yan, Q.; Li, M.; Qu, G.; Xiao, Y. A Survey on Data Poisoning Attacks and Defenses. In Proceedings of the 2022 7th IEEE International Conference on Data Science in Cyberspace, DSC 2022, Guilin, China, 11–13 July 2022; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2022; pp. 48–55. [Google Scholar] [CrossRef]
  70. Aly, S.; Almotairi, S. Deep Convolutional Self-Organizing Map Network for Robust Handwritten Digit Recognition. IEEE Access 2020, 8, 107035–107045. [Google Scholar] [CrossRef]
  71. Kundroo, M.; Kim, T. Demystifying Impact of Key Hyper-Parameters in Federated Learning: A Case Study on CIFAR-10 and FashionMNIST. IEEE Access 2024, 12, 120570–120583. [Google Scholar] [CrossRef]
  72. Blanchard, P.; El Mhamdi, E.M.; Guerraoui, R.; Stainer, J. Machine Learning with Adversaries: Byzantine Tolerant Gradient Descent. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.