Previous Article in Journal
A Capacity-Constrained Weighted Clustering Algorithm for UAV Self-Organizing Networks Under Interference
Previous Article in Special Issue
Performance Analysis of Blockchain Consensus Algorithm in Unmanned Aerial Vehicle Ad Hoc Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Lightweight Robust Training Method for Defending Model Poisoning Attacks in Federated Learning Assisted UAV Networks

1
State Key Laboratory of Massive Personalized Customization System and Technology and COSMOPlat IoT Technology Co., Ltd., Qingdao 266101, China
2
College of Computer Science and Technology, Nanjing University of Aeronautics and Astronautics, Nanjing 211106, China
3
Qingdao Penghai Software Co., Ltd., Qingdao 266071, China
4
COSMOPlat lnstitute of Industrial Intelligence (Qingdao) Co., Ltd., Qingdao 266000, China
*
Author to whom correspondence should be addressed.
Drones 2025, 9(8), 528; https://doi.org/10.3390/drones9080528 (registering DOI)
Submission received: 12 June 2025 / Revised: 15 July 2025 / Accepted: 16 July 2025 / Published: 28 July 2025
(This article belongs to the Special Issue IoT-Enabled UAV Networks for Secure Communication)

Abstract

The integration of unmanned aerial vehicles (UAVs) into next-generation wireless networks greatly enhances the flexibility and efficiency of communication and distributed computation for ground mobile devices. Federated learning (FL) provides a privacy-preserving paradigm for device collaboration but remains highly vulnerable to poisoning attacks and is further challenged by the resource constraints and heterogeneous data common to UAV-assisted systems. Existing robust aggregation and anomaly detection methods often degrade in efficiency and reliability under these realistic adversarial and non-IID settings. To bridge these gaps, we propose FedULite, a lightweight and robust federated learning framework specifically designed for UAV-assisted environments. FedULite features unsupervised local representation learning optimized for unlabeled, non-IID data. Moreover, FedULite leverages a robust, adaptive server-side aggregation strategy that uses cosine similarity-based update filtering and dimension-wise adaptive learning rates to neutralize sophisticated data and model poisoning attacks. Extensive experiments across diverse datasets and adversarial scenarios demonstrate that FedULite reduces the attack success rate (ASR) from over 90% in undefended scenarios to below 5%, while maintaining the main task accuracy loss within 2%. Moreover, it introduces negligible computational overhead compared to standard FedAvg, with approximately 7% additional training time.

1. Introduction

The rapid deployment of fifth-generation (5G) and beyond wireless networks has positioned unmanned aerial vehicles (UAVs) as a transformative platform for delivering communication and computation services to ground mobile devices. UAVs are increasingly utilized in diverse scenarios, including sports stadiums, outdoor events, traffic hotspots, and remote regions, where they function as aerial base stations (BSs) or edge servers [1,2,3]. Leveraging their agility, mobility, and favorable line-of-sight (LoS) propagation, UAV-enabled networks facilitate rapid-response wireless communications, distributed data collection, artificial intelligence (AI) model training, and dynamic coverage enhancement [1]. These capabilities have spurred significant research into UAV-assisted computation and communication for next-generation wireless networks [3].
Despite their advantages, UAVs face challenges in processing and training large-scale datasets due to inherent limitations in energy, storage, and computational capacity [1,2]. Moreover, transmitting raw data from numerous ground devices to UAV servers is often impractical due to privacy concerns and constrained communication resources [4]. Federated learning (FL) [5,6] offers a privacy-preserving solution by enabling devices to train AI models locally without sharing raw data. In FL, devices perform local model updates using gradient descent optimization [7] and share only model parameters with a central server for aggregation, thereby reducing data transmission and enhancing privacy. Integrating FL into UAV-assisted networks enables distributed AI tasks without reliance on centralized BSs. Ground devices train models locally and upload model parameters to an FL-enabled UAV server, which aggregates them to update a global model. As shown in Figure 1, the UAV broadcasts the updated global model back to the devices for further refinement. This iterative process continues until the desired learning accuracy is achieved, minimizing network congestion and preserving data privacy. Compared to traditional cloud-centric frameworks, FL enhances the efficiency of AI model training in UAV-assisted wireless networks.
However, the distributed nature of FL renders it susceptible to poisoning attacks, where adversaries compromise local training data or model parameters [8]. Poisoning attacks are categorized as non-targeted, which degrade overall model performance [9], or targeted, which manipulate predictions for specific inputs [10]. A key vulnerability in FL systems arises from the potential untrustworthiness of participating clients, some of which may be malicious or compromised. For instance, adversaries may use projected gradient descent to craft local models that mislead the global model’s convergence [11]. Additionally, surrogate models mimicking target model behavior pose significant security risks [12]. These threats underscore the need for robust detection and defense mechanisms against poisoning attacks in FL environments.
The widely adopted Federated Averaging (FedAvg) algorithm is particularly vulnerable, as even rudimentary attack strategies can disrupt its convergence [13]. Existing defenses include robust aggregation algorithms [14,15] and anomaly detection techniques [16]. Robust aggregation methods employ weighting or filtering to mitigate anomalous updates but falter against numerous malicious clients. Anomaly detection approaches analyze update patterns to identify malicious clients, yet they often suffer from high false-positive and negative rates due to the complexity of attack strategies [17]. Gradient-based filtering methods struggle with coordinated attacks, while geometric median or gradient analysis approaches offer partial mitigation but compromise training efficiency and stability against sophisticated attacks.
Despite notable progress in designing secure federated learning frameworks, existing approaches exhibit critical limitations, particularly in UAV-assisted networks where resource constraints and data heterogeneity are pronounced. Many robust aggregation strategies, while effective against certain classes of attack, often suffer substantial performance degradation when confronted with highly coordinated or stealthy poisoning strategies or when the proportion of malicious clients increases. Anomaly detection techniques also tend to exhibit limited reliability under non-IID data distributions, producing high rates of false positives and negatives that undermine both the efficiency and trustworthiness of the learning process. Furthermore, the majority of prior proposals are challenged by the computational limitations of UAV hardware, as they frequently involve heavyweight model architectures or multiple training iterations, resulting in excessive local computation or communication overhead. These issues are compounded in practical deployments where bandwidth is severely limited, data is unlabeled, and sophisticated attackers can dynamically adapt to existing defense mechanisms. Model compression techniques, such as pruning and quantization, can reduce communication overhead in FL. However, these methods are ineffective against sophisticated model poisoning attacks, which manipulate the direction and magnitude of model parameter updates in a targeted manner. Pruning and quantization cannot reliably identify the affected neurons, and the information loss introduced by these techniques may even exacerbate model vulnerability, leading to a degradation in the main task accuracy.
To address these challenges, we propose FedULite, a federated learning framework specifically designed for UAV-assisted networks facing non-IID data and poisoning attacks. FedULite introduces a lightweight and robust local training pipeline that leverages unsupervised representation learning on unlabeled data, making it well-suited for resource-constrained UAV platforms. To further defend against model poisoning attacks, FedULite employs a robust server-side aggregation mechanism that adaptively filters and aggregates client updates based on update consistency and cosine similarity thresholds. Although techniques such as contrastive learning and cosine similarity-based filtering have been independently investigated in prior research, the core innovation of FedULite lies in its synergistic defense mechanism. At the client side, we employ lightweight unsupervised contrastive learning as the first line of defense. This approach not only addresses the challenges of unlabeled and non-IID data prevalent in UAV scenarios but also significantly enhances the robustness of local model feature representations before model updates are uploaded, thereby mitigating the impact of potential poisoning attacks at the source. On the server side, robust adaptive aggregation serves as the second line of defense. The local training in the first stage ensures that updates from benign clients exhibit greater consistency, substantially improving the accuracy of server-side cosine similarity-based filtering. This enables more effective identification of malicious models that deviate from the dominant update patterns. The two-stage design collectively forms a lightweight yet highly efficient defense framework, specifically tailored for resource-constrained environments such as UAVs. The holistic integration and synergistic effect of this approach represent the primary distinction and advantage over frameworks that rely solely on individual defense strategies.
The main contributions of this work are summarized as follows.
  • We introduce a computationally efficient FL framework optimized for UAV-assisted networks, leveraging unsupervised contrastive learning and lightweight architectures to enable robust representation learning on resource-constrained clients, addressing the computational and storage limitations of UAVs.
  • We develop a robust, adaptive aggregation method at the server, which combines cosine similarity-based update filtering and dimension-wise aggregation with adaptive learning rates, effectively countering both traditional and adaptive model poisoning strategies, including stealthy and coordinated attacks.
  • We provide extensive experimental validation, demonstrating that FedULite not only significantly improves robustness and efficiency in UAV-assisted federated learning but also achieves reliable convergence and strong resistance to adversarial disruptions under diverse real-world conditions.
The remainder of this paper is organized as follows. Section 2 reviews related work on federated learning assisted UAV networks and backdoor attacks, as well as defense mechanisms in federated learning. Section 3 presents the proposed FedULite method in detail. Section 4 evaluates and analyzes the experimental results. Finally, Section 5 concludes the paper and discusses future research directions.

2. Related Work

2.1. Federated Learning Assisted UAV Networks

In recent years, FL has garnered significant attention for its implementation in wireless networks, with researchers extensively exploring methods to enhance learning efficiency [18,19,20]. Tran et al. [18] formulated the FL framework in wireless networks as an optimization problem, targeting the minimization of aggregation latency and total device energy consumption. Further studies [19,20] refined global aggregation frequency under various resource constraints, including device CPU performance, transmission delays, and model accuracy, highlighting the need for efficient resource management in dynamic settings.
In the context of UAV-assisted communication, research has focused on leveraging FL to improve AI model training efficiency [4,21,22,23,24,25,26,27]. For instance, Studies [21,23] developed FL frameworks for UAV swarms, proposing joint optimization of power allocation and flight trajectories to enhance learning performance. Lim et al. [24] introduced an FL-aware collaborative learning approach tailored for vehicular networks, while Ng et al. [25] utilized UAVs as relays to improve FL accuracy through enhanced communication reliability. Notably, Shiri et al. [22] innovatively combined FL with mean-field game theory to address large-scale UAV path planning, offering a scalable solution for complex scenarios. Additionally, [4] systematically analyzed the prospects and challenges of FL in UAV-assisted networks, whereas [25,26] proposed a two-tier FL algorithm for heterogeneous computing architectures and a blockchain-based secure collaboration scheme, respectively.
Significant progress has also been made in joint trajectory and resource management for multi-UAV networks [28,29,30,31,32,33]. Early works [28,29] investigated the joint optimization of user association, resource allocation, and UAV deployment, yet traditional optimization methods struggled with dynamic and complex environments. To address this, reinforcement learning (RL) and deep reinforcement learning (DRL) approaches have been adopted [30,31,32,33,34,35]. Studies [30,31] employed multi-agent DRL algorithms to optimize UAV trajectories and transmit power, achieving robust performance in dynamic settings. Zhu et al. [32] proposed an Actor–Critic-based algorithm for joint optimization of three-dimensional UAV flight paths and transmission control in continuous action spaces. Similarly, Ref. [33] applied an Actor–Critic framework to dynamically optimize device association, resource allocation, and trajectories. While these advancements enhance adaptability, they often overlook security vulnerabilities in FL, such as poisoning attacks, which our work aims to address.

2.2. Poisoning Attacks in Federated Learning

Poisoning attacks in FL systems primarily target the data collection [36] and model training [37] phases, characterized by low implementation barriers and high cost-effectiveness. These attacks are classified into non-targeted attacks, which aim to degrade overall model performance, and targeted attacks, which induce incorrect predictions for specific inputs. From a mechanistic perspective, data poisoning attacks [38,39,40] manipulate training samples, such as through label-flipping techniques, while model poisoning attacks [41,42,43] directly alter local model parameters. A prominent example is the model replacement attack, where malicious updates are crafted to supplant global model updates. Research indicates [44] that model poisoning attacks pose a severe threat, achieving significant disruption even with few attackers and limited training rounds. Liu et al. [45] enhanced the efficacy of model replacement (MR) attacks by precisely evaluating global parameters and injecting malicious updates with weighted factors. Similarly, Wang et al. [46] proposed a Projected Gradient Descent (PGD) attack, employing iterative parameter perturbations for stealthier manipulation.
Notably, traditional defense mechanisms relying on specific metrics face emerging challenges. Huang et al. [47] introduced the Cosine Constraint Attack (CCA), which incorporates a cosine distance term into the loss function to align malicious gradients with global gradients, effectively evading cosine similarity-based detection. More critically, even multi-metric defense strategies struggle against sophisticated attacks that mimic normal gradient characteristics, underscoring the limitations of existing defenses.

2.3. Poisoning Defenses in Federated Learning

Research on defense mechanisms against poisoning attacks in FL primarily focuses on server-side strategies [48], predicated on the assumption of basic trust between clients and the server, as well as among clients. These approaches aim to mitigate the adverse impact of malicious updates on the global model by employing specific defense mechanisms after receiving client updates. Huang et al. [47] proposed the Scope method, which utilizes gradient normalization to counteract the masking effect of benign gradients on backdoor gradients, enhances anomalous gradient features through differential privacy, and employs cosine distance to measure gradient direction similarity, followed by dominant gradient clustering to detect and filter backdoor gradients. Blanchard et al. [49] introduced the Krum scheme, which selects the update with the smallest Euclidean distance to others as the global model parameter. However, its reliance on a single update may introduce model bias. The improved Multi-Krum variant aggregates the mean of multiple updates with minimal Euclidean distances, enhancing aggregation robustness and convergence efficiency, though its effectiveness is limited by the reduced discriminative power of Euclidean distance in high-dimensional spaces and its inability to detect diverse malicious updates.
To address these limitations, Zhang et al. [50] developed the DWAMA algorithm, which integrates Mahalanobis distance with an adaptive threshold mechanism. By capturing complex correlations among data features and dynamically adjusting detection thresholds and aggregation weights, DWAMA significantly improves update stability. Huang et al. [51] proposed the Multi-Metrics algorithm, combining Manhattan distance, Euclidean distance, and cosine similarity to implement a multi-dimensional scoring mechanism for selecting high-quality updates for aggregation. Jebreel et al. [52] introduced the FL-Defender algorithm, which identifies potential malicious updates based on cosine similarity of the last-layer gradients, incorporating PCA dimensionality reduction and reweighting strategies to suppress malicious client influence. Pillutla et al. [53] proposed the Robust Federated Aggregation (RFA) scheme, which replaces traditional mean aggregation with geometric median estimation to reduce the impact of anomalous clients through robust statistical methods. Liu et al. [54] developed the DefendFL algorithm, which employs collinearity masks to protect gradient privacy and implements a cosine similarity-based mask verification mechanism to counter poisoning attacks. While these methods effectively mitigate poisoning threats in specific scenarios, their performance diminishes in settings with large-scale clients or a high proportion of malicious clients, highlighting the need for more robust and scalable defense strategies. In addition, certain personalized federated learning techniques, such as those proposed in [55,56], achieve model personalization through partial model-sharing methods. These approaches aggregate only specific parts of the model at the server while retaining or locally learning additional modules. Such strategies inherently offer some resistance to poisoning attacks by limiting the scope of shared parameters. However, these methods primarily aim to improve efficiency or address data heterogeneity. In the face of sophisticated poisoning attacks, their performance is limited. In contrast, the core objective of FedULite is to ensure model security and robustness in adversarial environments, addressing a distinct research dimension.

3. Threat Model

Attack model: We adopt the attack model established in prior works [45,46,57]. Specifically, the adversary controls one or more malicious clients, which may include fabricated clients injected into the system. The adversary does not compromise the server. During each iteration of the FL training process, malicious clients can send arbitrary local model updates to the server.
Attacker’s goal: Poisoning attacks in FL are generally categorized into untargeted poisoning attacks, which degrade overall model performance, and targeted poisoning attacks, which induce erroneous predictions on specific target samples while maintaining high overall model accuracy to evade detection. This study focuses on targeted poisoning attacks due to their stealthier nature.
Attacker’s capability: The adversary possesses full control over the training process of compromised clients, enabling direct manipulation of their model parameters in each attack cycle. The adversary has access to the global model updates in each training round and knowledge of the local data and model parameters of controlled clients. However, the adversary cannot access the server’s aggregation algorithm, defense mechanisms, or the data and training strategies of benign clients.
Defender’s goal: Existing poisoning defense methods are predominantly server-side, assuming a trusted server, but their performance degrades sharply as the proportion of malicious clients increases. This work aims to design a lightweight, localized poisoning defense method to achieve the following objectives: (1) Utility: In the absence of attacks, the method should not compromise the classification accuracy of the global model, achieving performance comparable to the widely used FedAvg algorithm in non-adversarial settings. (2) Robustness: In the presence of strong targeted poisoning attacks, the global model trained with our method should not predict attacker-specified labels for attacker-chosen target test samples. (3) Lightweight: Given the resource-constrained nature of clients in UAV-assisted FL systems, the method should impose minimal additional computational workload.

4. Method

4.1. Overview of FedULite

Figure 2 presents FedULite, a federated learning framework tailored for unmanned aerial vehicle (UAV)-assisted scenarios, addressing the challenges of non-independent and identically distributed (non-IID) data and poisoning attacks. FedULite leverages unlabeled data to train a robust global encoder f g , enabling effective representation learning in resource-constrained and dynamic network environments. To accommodate the computational limitations of UAV clients, the framework employs lightweight local training with optimized architectures and minimal computational overhead. At the server, a robust aggregation mechanism mitigates model poisoning attacks, such as backdoor and model replacement attacks [23,24,25], by prioritizing consistent client updates. A streamlined bidirectional communication protocol facilitates the exchange of online encoders f g k and predictors p g k between clients and the server, with the server distributing aggregated models f g and p g . The FedULite framework operates through three core stages:
  • Local Training: Each client performs unsupervised representation learning with a lightweight contrastive loss, training an online encoder f g k , predictor p g k , and target encoder f t k , with anomaly detection to counter data poisoning.
  • Model Aggregation: The server aggregates client models using a robust strategy that evaluates update consistency, producing updated global models f g and p g .
  • Model Update: The server distributes f g and p g , which clients adopt to update their local models.
Algorithm 1 encapsulates FedULite, detailing its lightweight training, robust aggregation, and update mechanisms.

4.2. Local Robust Training

In FedULite, local training employs a dual-network architecture comprising an online network, consisting of an online encoder f g k with parameters θ and a predictor p g k , and a target network, represented by a target encoder f t k with parameters ξ . This architecture facilitates unsupervised representation learning, optimized for the resource-constrained environment of UAV clients while ensuring robustness against data poisoning attacks, such as label flipping or adversarial samples.
To mitigate the impact of potentially malicious encoders in federated learning, FedULite employs a strategic initialization and update policy for local models. In the first training round, benign clients initialize both their online encoder f g k and target encoder f t k with the global encoder f g , ensuring a clean starting point free from malicious influence. In subsequent rounds, the online encoder f g k and predictor p g k are set to the global encoder f g and predictor p g , respectively, while the target encoder f t k adopts the parameters of the previous round’s local online encoder f g k . This approach leverages the historical, presumably benign, local encoder as the target to stabilize training and counteract potential malicious updates aggregated into the global encoder. By using the previous round’s local encoder as a reference, FedULite disperses the influence of malicious clusters, enhancing the robustness of the global model against backdoor attacks.
Algorithm 1: Algorithm of FedULite process
  Data: Number of clients N, communication rounds T, local epochs E = 1
  Result: Global encoder f g , global predictor p g
  // Initialize global and local modelsDrones 09 00528 i001
Training commences with the generation of two augmentations, t and t , from an input image x, employing computationally efficient transformations, such as random cropping and horizontal flipping, to minimize preprocessing overhead. Both the online and target encoders are implemented as lightweight convolutional neural networks, designed to reduce model complexity while preserving expressive capacity. The predictor p g k is a compact multilayer perceptron (MLP), further reducing computational demands.
The online network is optimized using a contrastive loss that aligns the normalized outputs of the online and target networks:
L θ ξ p g k ( f g k ( t ) ) p g k ( f g k ( t ) ) 2 f t k ( t ) f t k ( t ) 2 2 2 ,
where p g k ( f g k ( t ) ) represents the output of the online network, and f t k ( t ) serves as the regression target provided by the target network. The parameters of the target network, ξ , are updated via an exponential moving average (EMA) of the online encoder’s parameters θ :
ξ m ξ + ( 1 m ) θ ,
where the decay rate m [ 0 , 1 ] is dynamically adjusted based on training dynamics, such as loss variance or communication rounds, to balance stability and adaptability in non-IID settings. To enhance computational efficiency, each client performs a single training iteration per communication round on a small batch of samples, significantly reducing overhead compared to multi-iteration training.
To leverage labeled data available at each client and align with the primary objectives of federated learning, FedULite incorporates supervised learning when labels are present. The online encoder f g k is optimized with a cross-entropy loss over labeled samples:
θ f g k t = arg min θ f g k t ( x , y ) D k c = 1 C y c · log f g k ( x ) c
where D k denotes the local labeled dataset of client k, and θ f g k t represents the parameters of f g k at round t. This supervised objective complements the unsupervised contrastive loss, enhancing the encoder’s ability to capture task-specific features while maintaining robustness in non-IID settings.
The communication protocol is designed for efficiency: clients upload only the online encoder f g k and predictor p g k , as f g k encapsulates the most recent data representations, while the target encoder f t k remains local to provide stable regression targets. This approach minimizes uplink communication overhead, a critical consideration for bandwidth-constrained UAV networks.

4.3. Robust Adaptive Aggregation

To enhance the robustness of the global model against poisoning attacks, FedULite employs a sophisticated aggregation strategy at the server, supplanting traditional weighted averaging. It is crucial to note that our robust adaptive aggregation strategy fundamentally relies on the server’s ability to access and inspect the individual model update, Δ t k , from each client prior to aggregation. This access is essential for computing the cosine similarities used in the filtering stage (Equation (4)) and for assessing the sign consistency of each parameter dimension (Equation (5)). We therefore must clearly state that this requirement makes FedULite, in its current form, incompatible with standard secure aggregation protocols that are designed to provide server-side privacy by hiding these very updates.
For each client k, model updates are computed as Δ t k = θ f g k θ f g for the online encoder and Δ p g k = θ p g k θ p g for the predictor, where θ f g and θ p g represent the global parameters from the previous communication round. The aggregation process leverages a two-stage mechanism to evaluate update consistency, effectively neutralizing malicious contributions while preserving model convergence in non-IID settings.
In the first stage, updates are pre-filtered based on their alignment with the mean update across clients:
S t = k S t cos Δ t k , 1 | S t | j S t Δ t j τ ,
where τ is a cosine similarity threshold. This filtering excludes updates that significantly deviate from the majority, thereby mitigating stealthy attacks, such as cosine-constrained poisoning attempts [29]. In the second stage, the server assesses the consistency of updates along each parameter dimension i by computing the sum of update signs:
sign _ sum = k S t sgn ( Δ t , i k ) .
A dimension-specific learning rate η θ , i is then assigned:
η θ , i = η if | sign _ sum i |   θ , η otherwise ,
where η is the base learning rate, and θ is a threshold dynamically adjusted based on the number of participating clients to accommodate the variable client participation inherent in UAV networks. The global encoder and predictor parameters are updated as follows:
θ f g t + 1 = θ f g t + η θ k S t n k · Δ t k k S t n k ,
θ p g t + 1 = θ p g t + η θ k S t n k · Δ p g k k S t n k ,
where η θ = [ η θ , 1 , η θ , 2 , , η θ , d ] , and ⊙ denotes element-wise multiplication. The vector n k represents the data volume of client k, ensuring weighted contributions proportional to local dataset sizes. This mechanism suppresses the influence of inconsistent updates, effectively mitigating poisoning attacks while maintaining robust convergence.
Upon aggregation, the server distributes the updated global models f g and p g to all clients, who synchronize their local models by setting f g k f g and p g k p g . This streamlined update process ensures model consistency across clients, enhancing robustness in heterogeneous, non-IID data environments characteristic of UAV-assisted FL.

5. Evaluation

5.1. Experimental Setup

5.1.1. Datasets and Models

To evaluate the defensive capabilities of the FedULite framework, we conducted experiments across four widely adopted public datasets: MNIST, Fashion-MNIST (F-MNIST), EMNIST, and CIFAR-10. Notably, EMNIST was specifically employed to simulate non-IID settings, reflecting the heterogeneous data distributions common in UAV-assisted federated learning scenarios. The MNIST dataset comprises 70,000 grayscale handwritten digit images (0–9), with 60,000 training samples and 10,000 test samples, all normalized to a resolution of 28 × 28 pixels. Fashion-MNIST mirrors MNIST in data size and scale, containing 70,000 grayscale images of 10 fashion item categories (e.g., clothing, footwear), but presents greater complexity due to its diverse visual patterns. The CIFAR-10 dataset consists of 60,000 color images across 10 classes (e.g., airplane, automobile, bird), with 50,000 training samples and 10,000 test samples, each normalized to 32 × 32 pixels with three RGB channels during preprocessing. These datasets were selected to encompass a range of data modalities (grayscale vs. color), complexities, and distribution characteristics, enabling a comprehensive assessment of FedULite’s robustness against poisoning attacks under varied conditions.
The experimental setup incorporated tailored model architectures to align with the distinct characteristics of each dataset, ensuring both applicability and fairness in evaluating defensive performance. For MNIST and Fashion-MNIST, we adopted the LeNet-5 architecture, a lightweight convolutional neural network well-suited for grayscale image classification. LeNet-5 features two convolutional layers (with 6 and 16 feature maps, respectively, using 5 × 5 kernels), followed by three fully connected layers, with ReLU activation functions and standard max-pooling layers to enhance feature extraction and reduce computational overhead. For CIFAR-10, which involves more complex color images, we employed the ResNet-18 model, a deeper architecture with residual connections that facilitate training and improve generalization on challenging visual tasks. This differentiated design accounts for the varying computational demands and feature complexities of the datasets, aligning with FedULite’s lightweight local training paradigm optimized for resource-constrained UAV environments.

5.1.2. Federated Learning Settings

To rigorously assess the defensive performance of the FedULite framework, we conducted experiments with an FL setup comprising N = 100 clients, reflecting a realistic scale for UAV-assisted networks. In each communication round, the server randomly selected M = 10 clients to contribute updates to the global model, corresponding to a participation ratio of 10%, a common configuration in FL studies to balance computational efficiency and model convergence. To emulate the non-IID data distributions prevalent in such scenarios, we employed Dirichlet sampling [58] to partition the datasets across clients.
For each selected benign client, local training was performed using a stochastic gradient descent (SGD) optimizer with a learning rate of η = 0.01 , a standard choice ensuring stable convergence across diverse datasets. Clients executed two local training epochs consistent with FedULite’s lightweight design, which employs single-iteration contrastive and supervised training to minimize computational overhead on resource-constrained devices. A batch size of 64 was adopted, striking a balance between gradient estimation stability and memory efficiency, particularly suitable for the constrained computational capabilities of UAV clients. The local training process incorporated anomaly detection to discard samples with excessively high contrastive losses, enhancing resilience against data poisoning.

5.1.3. Attacks and Defenses Settings

In terms of attack settings, we conduct experimental evaluations against four distinct poisoning attack strategies. For the model replacement attack, also known as the constrain-and-scale (CSC) attack, the adversary amplifies its malicious update by a factor of N / K , where N denotes the total number of clients and K is the number of clients participating in each training round. This amplification is designed to ensure that the attacker’s update can fully override the global model, effectively hijacking the aggregation process. In the projected gradient descent (PGD) attack, we impose an L 2 norm constraint by projecting the adversarial gradients onto a norm ball with a radius of 2. The specific hyperparameter settings for this attack are as follows: each gradient update uses a step size of 0.02, the model parameters are projected every 10 updates, and Gaussian noise with a standard deviation of 0.025 is added to further obfuscate the update’s malicious intent. For the cosine constraint attack (CCA), we balance the cross-entropy loss and cosine loss with equal weights of 0.5, thereby ensuring that the direction of the attacker’s update aligns closely with that of the global model, making malicious contributions harder to distinguish during aggregation. Finally, for the distributed backdoor attack (DBA), we adopt the approach in [57], wherein multiple local triggers are collaboratively combined into a single global trigger. We configure three independent local triggers, thereby enhancing the attack’s stealthiness and effectiveness through the cooperation among multiple adversarial clients.
These attack strategies are evaluated in conjunction with four state-of-the-art defense mechanisms: FoolsGold [59], Multi-Krum [49], RLR [60], and FL-Defender [52]. For both attack implementations and defense algorithms, we adhere strictly to the original official implementations to ensure the fairness and reproducibility of our experiments.

5.1.4. Evaluation Metrics

Drawing on prior research, we evaluated the performance of the FedULite framework using two key metrics: Attack Success Rate (ASR) and Accuracy on the main classification task (ACC). ASR is defined as the proportion of backdoor inputs misclassified as the attacker’s target label by the compromised global model, providing a direct measure of the poisoning attack’s effectiveness. In contrast, ACC quantifies the model’s performance on benign, clean test samples, serving as an indicator of the defense mechanism’s ability to maintain overall model utility while countering malicious interference. In the context of FL, particularly within UAV-assisted networks, benign clients aim to maximize ACC to ensure robust task performance, whereas attackers strive to maximize ASR while concealing their presence by ensuring ACC remains largely unaffected. An effective defense, such as FedULite, seeks to minimize ASR, ideally approaching zero, while preserving high ACC, thereby achieving a balance between security and functionality in the face of sophisticated poisoning attacks.

5.2. Experimental Results

5.2.1. Comparison

Table 1 presents a comparative analysis of FedULite and four benchmark backdoor defense methods, with all approaches evaluated under the same configurations as their original works. The “No-Defense” column denotes the baseline scenario where no defensive mechanisms are applied, and the best results in each category are highlighted in bold. Experimental results indicate that FedULite consistently reduces the ASR of all four evaluated attacks to as low as 4% across three benchmark datasets, outperforming all other baseline methods in terms of attack mitigation.
Overall, while Multi-Krum and RLR exhibit a certain degree of effectiveness against a range of poisoning attacks under balanced data distributions, this often comes at the cost of a substantial decrease in the main task’s accuracy. FoolsGold, on the other hand, fails to defend against DBA, reflecting its limitations in more sophisticated adversarial scenarios. FL-Defender demonstrates inconsistent performance, showing effectiveness on some datasets such as F-MNIST but reduced robustness on others, suggesting its sensitivity to data characteristics. In contrast, FedULite achieves robust resistance against poisoning attacks by combining local mitigation techniques with server-side filtering and robust aggregation. This integrated defense paradigm effectively balances model accuracy and security, demonstrating superior adaptability across various data distributions and attack types. Moreover, the observed drops are attributed to the inherent trade-off of the defense mechanism, where filtering malicious updates occasionally affects benign updates that deviate significantly from the mainstream pattern. This defensive overhead is necessary to ensure robustness but is minimized in FedULite ’s design, as evidenced by the limited accuracy degradation compared to other methods.
Moreover, we visualize the distribution in the feature space of the global model under the attacks with/without FedULite to further understand the advantages of our proposed method. Specifically, we poisoned part of the test data in CIFAR-10, where the number of each poisoned class is the same. Then, the poisoned test data are used for t-SNE visualization. From Figure 3a, we can see that the feature representations of benign samples from the same category form an individual cluster, while the poisoned samples form a new cluster (in grey color). However, the clusters of poisoned samples in Figure 3b are absolutely damaged, and all the poisoned samples are assembled with the benign samples from the same category. This means that FedULite can effectively break the cluster of backdoor features and further prevent the establishment of the connection between the targeted label and the trigger through contrastive training.

5.2.2. Different FL and Attack Settings

Non-IID data distributions. To investigate the impact of data heterogeneity on backdoor defense effectiveness, we evaluate the performance of FedULite under varying levels of statistical heterogeneity and compare its robustness against baseline methods. In this analysis, we manipulate the Dirichlet distribution parameter α { 0.05 , 0.1 , 10 } to control the degree of non-IID data allocation among clients, with a particular focus on DBA, as its distributed triggering mechanism poses substantial challenges under heterogeneous conditions. To visualize the effect of data heterogeneity, we randomly select 10 clients from a total of 100 and display the number of samples per class for each client, as illustrated in Figure 4. Here, α = 10 approximates an IID setting, while smaller values, such as α = 0.05 , indicate higher heterogeneity with increasingly skewed class distributions across clients.
The experimental results reported in Table 2, corresponding to the non-IID settings, demonstrate that data heterogeneity has a pronounced influence on both ASR and ACC, where the best results in each category are highlighted in bold. As the value of α decreases, representing increased heterogeneity, both ASR and ACC tend to decline. This phenomenon reflects the dual effects of non-IID distributions: on one hand, elevated heterogeneity restricts the global model’s generalization ability due to some clients possessing samples predominantly from a few classes while having few or no samples from others, leading to diminished overall model performance; on the other hand, heterogeneity impedes the propagation of malicious triggers, thereby reducing the effectiveness of backdoor injection attacks. Notably, data heterogeneity presents additional challenges for defense techniques based on detection, such as FoolsGold and Multi-Krum, as these methods rely on similar client updates to distinguish adversarial behavior.
By contrast, FedULite exhibits exceptional resilience regardless of the degree of data heterogeneity. Even under highly non-IID settings ( α = 0.05 ), FedULite consistently reduces the DBA ASR to below 3% for all tested α values. Although there is a modest decline in ACC owing to the extreme skewness of class distributions, FedULite still outperforms all baseline methods, underscoring its robustness and adaptability to heterogeneous FL environments.
Different ratios of malicious clients. To assess the robustness of FedULite in defending against poisoning attacks, we systematically investigate its performance under varying proportions of malicious participants. The proportion of adversarial clients is a critical factor in practical FL deployments, as the prevalence of malicious actors may fluctuate over time. In our experiments, the fraction of malicious clients is varied within the set {0.1, 0.2, 0.4, 0.6, 0.8, 0.9}, enabling a comprehensive evaluation of the impact on both ASR and ACC.
The trends observed in Figure 5 reveal that as the proportion of malicious clients increases, ACC for all evaluated methods exhibits a consistent decline, while ASR rises sharply. This behavior is intuitive: higher proportions of adversarial clients introduce a greater number of backdoor triggers into the global model during the aggregation process, thereby undermining its ability to learn benign patterns and resulting in degraded classification performance. Concurrently, the amount of poisoned data utilized during local training by malicious clients grows as their proportion increases, strengthening the efficacy of the injected triggers and further elevating the ASR. Notably, when the malicious client ratio exceeds 40%, most baseline defense methods experience a considerable performance drop. These results highlight the vulnerability of server-side defense approaches, particularly those based on detection, to high proportions of adversarial participation.
By contrast, FedULite demonstrates remarkable robustness even as the proportion of malicious clients increases. Its design, which integrates local mitigation strategies with server-side filtering and aggregation techniques, enables it to maintain lower ASR values and relatively stable ACC, thereby providing a reliable defense against poisoning attacks in FL scenarios with varying degrees of adversarial presence.
Number of clients. The number of clients is a critical parameter in FL scenarios, as it directly influences system dynamics and the effectiveness of defense mechanisms. To validate the applicability of FedULite, we examine its performance under varying numbers of participating clients. Additionally, to rigorously assess FedULite’s resilience against hybrid poisoning attacks, we combine the DBA and the CSC attack, and further vary the scaling factor S in the CSC attack from 10 to 50 in increments of 20. The experimental results shown in Figure 6 indicate that FedULite maintains robust defense efficacy across different client population settings. Notably, as the proportion of clients selected per round increases, the ASR exhibits a decreasing trend. This effect is mainly attributed to the dilution of malicious influence: with a larger number of benign clients participating in the aggregation process, the relative weight and impact of malicious updates are diminished, thereby weakening the ability of adversaries to implant effective backdoor triggers. Furthermore, FedULite demonstrates a strong immunity to variations in attack intensity, maintaining stable performance even as the attacks’ strength escalates. This highlights the method’s adaptability and robustness in practical FL environments with dynamic client populations and evolving adversarial threats.
Different model structures. To ensure the generalizability of the FedULite framework across diverse model architectures, which inherently learn distinct feature representations, we conducted an extensive evaluation of its robustness against poisoning attacks using six widely adopted image classification architectures: ResNet18, ResNet34, VGG11, VGG19, MobileNet-V2, and LeNet-5. These architectures were selected to span a range of complexities, depths, and computational requirements, from lightweight models like MobileNet-V2, suitable for resource-constrained UAV clients, to deeper networks like VGG19, which are more computationally intensive. All experiments maintained consistency with the default configuration. As shown in Figure 7, although ACC exhibits some variation depending on the choice of encoder architecture—likely due to differences in local contrastive training resulting from architectural dissimilarities—our method consistently demonstrates effective performance across these models. This consistency highlights the adaptability of our approach and suggests that it can be reliably applied to diverse neural network architectures in practical federated learning settings.

5.3. Ablation and Parameter Sensitivity Analysis

5.3.1. Robust Adaptive Aggregation

In this section, we investigate the impact of Robust Adaptive Aggregation (RAA) on the effectiveness of our method. Specifically, we consider three ablation scenarios: removing only the Filter Updates component while retaining Robust Aggregation (denoted as RAA-DFU), removing only Robust Aggregation while retaining Filter Updates (RAA-DRA), and eliminating all RAA components (RAA-DALL). We evaluate the performance of these configurations under the DBA attack across three datasets. For scenarios without RAA, we adopt the standard FedAvg aggregation algorithm by default. The experimental results, summarized in Table 3, demonstrate that each RAA component plays a critical role. When all server-side operations are removed, both ASR and ACC show significant degradation, indicating that although local defenses provide some level of protection, they are insufficient on their own. The Robust Aggregation module proves essential for mitigating advanced poisoning strategies, while Filter Updates effectively filters out anomalous gradients at an early stage. These findings highlight the necessity of both server-side mechanisms in achieving comprehensive and resilient defense against backdoor attacks in federated learning.
To further elucidate the impact of RAA, we quantified the learning rate inversion ratio across different dimensions, defined as the proportion of participants exhibiting learning rate inversion in a specific dimension relative to the total number of participants. In our experiments, we fixed the number of malicious participants at three and selected eight dimensions for analysis, with the first five corresponding to the encoder and the remaining three to the predictor. We computed the inversion ratio over ten communication rounds and reported the average values. Additionally, we examined the inversion ratio under varying degrees of data heterogeneity to assess its sensitivity to non-IID data distributions. The experimental results shown in Table 4 reveal that RAA effectively induces learning rate inversion across different dimensions, with the observed ratio closely aligning with the proportion of malicious participants, approximately 30%. However, the degree of data heterogeneity significantly amplifies this ratio, indicating that non-uniform data distributions exacerbate the challenges posed by malicious participants. This finding underscores the interplay between data heterogeneity and the efficacy of RAA in mitigating adaptive attacks, highlighting its robustness in detecting and countering malicious behavior under diverse conditions.

5.3.2. Impact of τ

τ represents a cosine similarity threshold, which is employed to filter client model updates during the aggregation process. Specifically, an update from client k is included in the set S t for subsequent aggregation only if its cosine similarity meets or exceeds τ . By tuning τ , the algorithm can exclude malicious updates that significantly deviate from the predominant update direction. A smaller value of τ allows a broader range of updates to be included—including those that might arise from benign sources of noise or normal deviations due to non-IID data distribution—whereas a larger value of τ enforces stricter filtering, potentially increasing robustness but also risking the exclusion of legitimate diversity among client updates.
In this section, we examine the effects of different τ settings on our method under varying levels of data heterogeneity. As illustrated in Figure 8, our experimental results align with expectations: the choice of τ is closely related to the degree of data heterogeneity. In scenarios characterized by high degrees of non-IID data, opting for a smaller τ is more appropriate, given the inherently lower similarity among local models. This nuanced adjustment enables the aggregation scheme to better accommodate natural variability while still providing a degree of resilience against adversarial updates. These findings highlight the nuanced interplay between τ , data heterogeneity, and defense efficacy. In non-IID scenarios, the diversity of legitimate updates necessitates a relaxed τ to preserve model utility, as overly strict thresholds risk undermining the global model’s ability to generalize across skewed distributions.

5.3.3. Computational Overhead

To quantify the additional computational overhead introduced by FedULite, we measured the average time and peak memory usage required for a single round of local training on the CIFAR-10 dataset, using models of varying scales on a single client. The experiments were conducted on an NVIDIA RTX 3090 GPU. We compared FedULite against FedAvg, which lacks any defense mechanisms, and a representative defense method, FL-Defender. The results are presented in Table 5.
As shown in Table 5, compared to the baseline FedAvg, FedULite incurs minimal additional time overhead (approximately 7%) and memory overhead (approximately 1%). This efficiency is primarily attributed to our lightweight contrastive learning approach, which requires only a single forward and backward propagation. In contrast to other defense methods, such as FL-Defender, which demand additional computations or complex analyses, FedULite achieves robust performance with negligible computational burden on clients. These results strongly validate its practicality in resource-constrained UAV environments.

5.3.4. Impact of Local Epoch

In this section, we explore the impact of local training epochs on the performance of our proposed method. The experimental results shown in Table 6 demonstrate that increasing the number of local training epochs enhances the defense against backdoor attacks while simultaneously improving the accuracy of the primary task. This improvement is attributed to the extended local optimization, which strengthens the model’s resilience to malicious perturbations. However, our findings indicate that this enhancement is subject to diminishing returns, with performance gains plateauing after four local training epochs. Beyond this threshold, additional epochs yield negligible improvements, suggesting an upper limit to the benefits of extended local training. Notably, our method achieves satisfactory performance with just a single training epoch, a result that can be attributed to the synergistic integration of our chosen contrastive learning paradigm with cross-entropy loss. This combination effectively balances robustness and efficiency, enabling the model to maintain strong defensive capabilities even with minimal local training. In scenarios where computational resources are constrained, this lightweight design represents a practical trade-off, where marginal performance sacrifices are acceptable to ensure resource efficiency and scalability.

5.4. Adaptive Attack

In this section, we investigate the robustness of FedULite against adaptive attacks, where attackers are aware of the filtering threshold τ and constrain the model updates to ensure the model differences remain below τ during the attack process. To evaluate the resilience of FedULite, we conducted experiments across three distinct datasets, reporting standard deviations across multiple random seeds to ensure statistical reliability. The results shown in Table 7 demonstrate that constraining model differences inherently exerts a suppressive effect on the efficacy of such attacks. However, this approach fails to fully circumvent our defense mechanism, as attackers cannot simultaneously achieve a successful backdoor attack while maintaining normal behavior across all feature dimensions. This limitation arises because the constrained updates disrupt the attacker’s ability to embed malicious behavior without exceeding the detectable threshold. Furthermore, FedULite incorporates a two-stage defense strategy, where the contrastive learning component, integrated into the local training process, introduces an additional layer of resistance against backdoor attacks. This mechanism effectively disrupts the attacker’s ability to manipulate the model covertly, reinforcing the overall robustness of the system.

6. Conclusions and Future Work

In this paper, we present FedULite, a lightweight and robust federated learning framework tailored for UAV-assisted wireless networks. Specifically, FedULite enables clients to perform efficient unsupervised local representation learning. To enhance model robustness, FedULite incorporates a two-stage adaptive aggregation strategy at the server side, leveraging cosine similarity-based update filtering and dimension-wise adaptive learning rates to effectively suppress both data and model poisoning attacks. Compared with conventional defense approaches, FedULite achieves superior robustness and efficiency in resource-constrained, adversarial UAV environments. Extensive experiments demonstrate that FedULite ensures reliable learning performance and strong resistance to advanced adversarial threats. Looking ahead, we plan to extend the research on FedULite in several key directions. First, regarding deployment in more realistic UAV environments, our framework already exhibits inherent adaptability. For instance, Intermittent Connectivity and Mobility: The federated learning paradigm of FedULite is designed to tolerate dynamic changes in participant availability. Client mobility or temporary disconnections are treated as temporary non-participation in a given aggregation round, and our robust aggregation algorithm remains insensitive to variations in the number of participating clients. Energy Constraints: The lightweight design of our framework is a core strength. Local training requires only a single iteration, significantly reducing energy consumption for UAVs or ground devices, thereby ensuring feasibility in energy-constrained environments. Second, the assumption of a fully trusted server represents a limitation of the current work, and we envision a clear path toward extending FedULite to decentralized or semi-trusted settings. FedULite can be integrated with blockchain technology. Model updates and aggregation rules can be deployed as smart contracts on the blockchain, ensuring transparency, immutability, and auditability of the process. This enables reliable collaboration in semi-trusted environments. In addition, we will explore effective integration schemes of FedULite with privacy-enhancing technologies such as differential privacy and secure aggregation to build a comprehensive framework that is both secure, robust, and privacy-preserving. Finally, we provide a strict theoretical convergence guarantee for our robust aggregation algorithm, which is a key step in pushing this method from empirical validity to theoretical reliability.

Author Contributions

Conceptualization, L.C. and W.Z.; methodology, L.C.; software, X.B.; validation, M.S. and C.Z.; formal analysis, W.Z.; data curation, X.B.; writing—original draft preparation, L.C.; writing—review and editing, M.S.; visualization, C.Z.; supervision, C.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the State Key Laboratory of Massive Personalized Customization System and Technology (No. H&C-MPC-2023-02-05).

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

Author Lucheng Chen was employed by the State Key Laboratory of Massive Personalized Customization System and Technology and COSMOPlat IoT Technology Co., Ltd. Authors Weiwei Zhai and Xiangfeng Bu were employed by the Qingdao Penghai Software Co., Ltd. Authors Ming Sun and Chenglin Zhu were employed by the COSMOPlat Institute of Industrial Intelligence (Qingdao) Co., Ltd. All authors declared that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Lu, X.; Xiao, L.; Dai, C.; Dai, H. UAV-aided cellular communications with deep reinforcement learning against jamming. IEEE Wirel. Commun. 2020, 27, 48–53. [Google Scholar] [CrossRef]
  2. Qi, F.; Zhu, X.; Mang, G.; Kadoch, M.; Li, W. UAV network and IoT in the sky for future smart cities. IEEE Netw. 2019, 33, 96–101. [Google Scholar] [CrossRef]
  3. Zeng, Y.; Wu, Q.; Zhang, R. Accessing from the sky: A tutorial on UAV communications for 5G and beyond. Proc. IEEE 2019, 107, 2327–2375. [Google Scholar] [CrossRef]
  4. Brik, B.; Ksentini, A.; Bouaziz, M. Federated learning for UAVs-enabled wireless networks: Use cases, challenges, and open problems. IEEE Access 2020, 8, 53841–53849. [Google Scholar] [CrossRef]
  5. Konečnỳ, J.; McMahan, H.B.; Yu, F.X.; Richtárik, P.; Suresh, A.T.; Bacon, D. Federated learning: Strategies for improving communication efficiency. arXiv 2016, arXiv:1610.05492. [Google Scholar]
  6. McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the Artificial Intelligence and Statistics, Fort Lauderdale, FL, USA, 20–22 April 2017; PMLR. pp. 1273–1282. [Google Scholar]
  7. Ruder, S. An overview of gradient descent optimization algorithms. arXiv 2016, arXiv:1609.04747. [Google Scholar]
  8. Zhu, C.; Zhang, J.; Sun, X.; Chen, B.; Meng, W. ADFL: Defending backdoor attacks in federated learning via adversarial distillation. Comput. Secur. 2023, 132, 103366. [Google Scholar] [CrossRef]
  9. Gong, X.; Chen, Y.; Wang, Q.; Kong, W. Backdoor attacks and defenses in federated learning: State-of-the-art, taxonomy, and future directions. IEEE Wirel. Commun. 2022, 30, 114–121. [Google Scholar] [CrossRef]
  10. Xia, G.; Chen, J.; Yu, C.; Ma, J. Poisoning attacks in federated learning: A survey. IEEE Access 2023, 11, 10708–10722. [Google Scholar] [CrossRef]
  11. Ma, C.; Chen, L.; Yong, J.H. Simulating unknown target models for query-efficient black-box attacks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 11835–11844. [Google Scholar]
  12. Bagdasaryan, E.; Veit, A.; Hua, Y.; Estrin, D.; Shmatikov, V. How to backdoor federated learning. In Proceedings of the International Conference on Artificial Intelligence and Statistics, Online, 26–28 August 2020; PMLR. pp. 2938–2948. [Google Scholar]
  13. Nguyen, T.D.; Nguyen, T.; Le Nguyen, P.; Pham, H.H.; Doan, K.D.; Wong, K.S. Backdoor attacks and defenses in federated learning: Survey, challenges and future research directions. Eng. Appl. Artif. Intell. 2024, 127, 107166. [Google Scholar] [CrossRef]
  14. Chen, H.; Chen, X.; Peng, L.; Ma, R. Flram: Robust aggregation technique for defense against byzantine poisoning attacks in federated learning. Electronics 2023, 12, 4463. [Google Scholar] [CrossRef]
  15. Xu, S.; Xia, H.; Zhang, R.; Liu, P.; Fu, Y. FedNor: A robust training framework for federated learning based on normal aggregation. Inf. Sci. 2024, 684, 121274. [Google Scholar] [CrossRef]
  16. Zhang, C.; Yang, S.; Mao, L.; Ning, H. Anomaly detection and defense techniques in federated learning: A comprehensive review. Artif. Intell. Rev. 2024, 57, 150. [Google Scholar] [CrossRef]
  17. Yin, D.; Chen, Y.; Kannan, R.; Bartlett, P. Byzantine-robust distributed learning: Towards optimal statistical rates. In Proceedings of the International Conference on Machine Learning, Pmlr, Stockholm, Sweden, 10–15 July 2018; pp. 5650–5659. [Google Scholar]
  18. Tran, N.H.; Bao, W.; Zomaya, A.; Nguyen, M.N.; Hong, C.S. Federated learning over wireless networks: Optimization model design and analysis. In Proceedings of the IEEE INFOCOM 2019-IEEE Conference on Computer Communications, Paris, France, 29 April–2 May 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1387–1395. [Google Scholar]
  19. Wang, S.; Tuor, T.; Salonidis, T.; Leung, K.K.; Makaya, C.; He, T.; Chan, K. Adaptive federated learning in resource constrained edge computing systems. IEEE J. Sel. Areas Commun. 2019, 37, 1205–1221. [Google Scholar] [CrossRef]
  20. Wang, S.; Tuor, T.; Salonidis, T.; Leung, K.K.; Makaya, C.; He, T.; Chan, K. When edge meets learning: Adaptive control for resource-constrained distributed machine learning. In Proceedings of the IEEE INFOCOM 2018-IEEE Conference on Computer Communications, Honolulu, HI, USA, 16–19 April 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 63–71. [Google Scholar]
  21. Zeng, T.; Semiari, O.; Mozaffari, M.; Chen, M.; Saad, W.; Bennis, M. Federated learning in the sky: Joint power allocation and scheduling with UAV swarms. In Proceedings of the ICC 2020–2020 IEEE International Conference on Communications (ICC), Dublin, Ireland, 7–11 June 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–6. [Google Scholar]
  22. Shiri, H.; Park, J.; Bennis, M. Communication-efficient massive UAV online path control: Federated learning meets mean-field game theory. IEEE Trans. Commun. 2020, 68, 6840–6857. [Google Scholar] [CrossRef]
  23. Liu, Y.; Nie, J.; Li, X.; Ahmed, S.H.; Lim, W.Y.B.; Miao, C. Federated learning in the sky: Aerial-ground air quality sensing framework with UAV swarms. IEEE Internet Things J. 2020, 8, 9827–9837. [Google Scholar] [CrossRef]
  24. Lim, W.Y.B.; Huang, J.; Xiong, Z.; Kang, J.; Niyato, D.; Hua, X.S.; Leung, C.; Miao, C. Towards federated learning in UAV-enabled Internet of Vehicles: A multi-dimensional contract-matching approach. IEEE Trans. Intell. Transp. Syst. 2021, 22, 5140–5154. [Google Scholar] [CrossRef]
  25. Ng, J.S.; Lim, W.Y.B.; Dai, H.N.; Xiong, Z.; Huang, J.; Niyato, D.; Hua, X.S.; Leung, C.; Miao, C. Joint auction-coalition formation framework for communication-efficient federated learning in UAV-enabled internet of vehicles. IEEE Trans. Intell. Transp. Syst. 2020, 22, 2326–2344. [Google Scholar] [CrossRef]
  26. Fadlullah, Z.M.; Kato, N. HCP: Heterogeneous computing platform for federated learning based collaborative content caching towards 6G networks. IEEE Trans. Emerg. Top. Comput. 2020, 10, 112–123. [Google Scholar] [CrossRef]
  27. Wang, Y.; Su, Z.; Zhang, N.; Benslimane, A. Learning in the air: Secure federated learning for UAV-assisted crowdsensing. IEEE Trans. Netw. Sci. Eng. 2020, 8, 1055–1069. [Google Scholar] [CrossRef]
  28. Wu, Q.; Zeng, Y.; Zhang, R. Joint trajectory and communication design for multi-UAV enabled wireless networks. IEEE Trans. Wirel. Commun. 2018, 17, 2109–2121. [Google Scholar] [CrossRef]
  29. Yin, S.; Li, L.; Yu, F.R. Resource allocation and basestation placement in downlink cellular networks assisted by multiple wireless powered UAVs. IEEE Trans. Veh. Technol. 2019, 69, 2171–2184. [Google Scholar] [CrossRef]
  30. Zhang, Y.; Mou, Z.; Gao, F.; Jiang, J.; Ding, R.; Han, Z. UAV-enabled secure communications by multi-agent deep reinforcement learning. IEEE Trans. Veh. Technol. 2020, 69, 11599–11611. [Google Scholar] [CrossRef]
  31. Zhao, N.; Cheng, Y.; Pei, Y.; Liang, Y.C.; Niyato, D. Deep reinforcement learning for trajectory design and power allocation in UAV networks. In Proceedings of the ICC 2020–2020 IEEE International Conference on Communications (ICC), Dublin, Ireland, 7–11 June 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–6. [Google Scholar]
  32. Zhu, M.; Liu, X.Y.; Walid, A. Deep reinforcement learning for unmanned aerial vehicle-assisted vehicular networks. arXiv 2019, arXiv:1906.05015. [Google Scholar]
  33. Wang, L.; Wang, K.; Pan, C.; Xu, W.; Aslam, N.; Nallanathan, A. Deep reinforcement learning based dynamic trajectory control for UAV-assisted mobile edge computing. IEEE Trans. Mob. Comput. 2021, 21, 3536–3550. [Google Scholar] [CrossRef]
  34. Fang, D.; Qiang, J.; Ouyang, X.; Zhu, Y.; Yuan, Y.; Li, Y. Collaborative Document Simplification Using Multi-Agent Systems. In Proceedings of the 31st International Conference on Computational Linguistics, Dhabi, UAE, 19–24 January 2025; pp. 897–912. [Google Scholar]
  35. Fang, D.; Qiang, J.; Zhu, Y.; Yuan, Y.; Li, W.; Liu, Y. Progressive Document-level Text Simplification via Large Language Models. arXiv 2025, arXiv:2501.03857. [Google Scholar]
  36. Tolpegin, V.; Truex, S.; Gursoy, M.E.; Liu, L. Data poisoning attacks against federated learning systems. In Proceedings of the Computer Security–ESORICs 2020: 25th European Symposium on Research in Computer Security, ESORICs 2020, Guildford, UK, 14–18 September 2020; Proceedings, part i 25. Springer: Berlin/Heidelberg, Germany, 2020; pp. 480–501. [Google Scholar]
  37. Jagielski, M.; Oprea, A.; Biggio, B.; Liu, C.; Nita-Rotaru, C.; Li, B. Manipulating machine learning: Poisoning attacks and countermeasures for regression learning. In Proceedings of the 2018 IEEE Symposium on Security and Privacy (SP), San Francisco, CA, USA, 21–23 May 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 19–35. [Google Scholar]
  38. Kumar, K.N.; Mohan, C.K.; Machiry, A. Precision Guided Approach to Mitigate Data Poisoning Attacks in Federated Learning. In Proceedings of the Fourteenth ACM Conference on Data and Application Security and Privacy, Porto, Portugal, 19–21 June 2024; pp. 233–244. [Google Scholar]
  39. Sun, G.; Cong, Y.; Dong, J.; Wang, Q.; Lyu, L.; Liu, J. Data poisoning attacks on federated machine learning. IEEE Internet Things J. 2021, 9, 11365–11375. [Google Scholar] [CrossRef]
  40. Alharbi, E.; Marcolino, L.S.; Gouglidis, A.; Ni, Q. Robust federated learning method against data and model poisoning attacks with heterogeneous data distribution. In ECAI 2023; IOS Press: Amsterdam, The Netherlands, 2023; pp. 85–92. [Google Scholar]
  41. Cao, X.; Gong, N.Z. Mpaf: Model poisoning attacks to federated learning based on fake clients. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 3396–3404. [Google Scholar]
  42. Zhang, J.; Zhu, C.; Ge, C.; Ma, C.; Zhao, Y.; Sun, X.; Chen, B. Badcleaner: Defending backdoor attacks in federated learning via attention-based multi-teacher distillation. IEEE Trans. Dependable Secur. Comput. 2024, 21, 4559–4573. [Google Scholar] [CrossRef]
  43. Sun, J.; Li, A.; DiValentin, L.; Hassanzadeh, A.; Chen, Y.; Li, H. Fl-wbc: Enhancing robustness against model poisoning attacks in federated learning from a client perspective. Adv. Neural Inf. Process. Syst. 2021, 34, 12613–12624. [Google Scholar]
  44. Sun, S.; Sugrim, S.; Stavrou, A.; Wang, H. Partner in Crime: Boosting Targeted Poisoning Attacks against Federated Learning. IEEE Trans. Inf. Forensics Secur. 2025, 20, 4152–4166. [Google Scholar] [CrossRef]
  45. Liu, T.; Zhang, Y.; Feng, Z.; Yang, Z.; Xu, C.; Man, D.; Yang, W. Beyond traditional threats: A persistent backdoor attack on federated learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 26–27 February 2024; Volume 38, pp. 21359–21367. [Google Scholar]
  46. Wang, H.; Sreenivasan, K.; Rajput, S.; Vishwakarma, H.; Agarwal, S.; Sohn, J.Y.; Lee, K.; Papailiopoulos, D. Attack of the tails: Yes, you really can backdoor federated learning. Adv. Neural Inf. Process. Syst. 2020, 33, 16070–16084. [Google Scholar]
  47. Huang, S.; Li, Y.; Yan, X.; Gao, Y.; Chen, C.; Shi, L.; Chen, B.; Ng, W.W. Scope: On Detecting Constrained Backdoor Attacks in Federated Learning. IEEE Trans. Inf. Forensics Secur. 2025, 20, 3302–3315. [Google Scholar] [CrossRef]
  48. Wu, J.; Jin, J.; Wu, C. Challenges and countermeasures of federated learning data poisoning attack situation prediction. Mathematics 2024, 12, 901. [Google Scholar] [CrossRef]
  49. Blanchard, P.; El Mhamdi, E.M.; Guerraoui, R.; Stainer, J. Machine learning with adversaries: Byzantine tolerant gradient descent. Adv. Neural Inf. Process. Syst. 2017, 119–129. [Google Scholar]
  50. Zhang, G.; Liu, H.; Yang, B.; Feng, S. Dwama: Dynamic weight-adjusted mahalanobis defense algorithm for mitigating poisoning attacks in federated learning. Peer-to-Peer Netw. Appl. 2024, 17, 3750–3764. [Google Scholar] [CrossRef]
  51. Huang, S.; Li, Y.; Chen, C.; Shi, L.; Gao, Y. Multi-metrics adaptively identifies backdoors in federated learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 4652–4662. [Google Scholar]
  52. Jebreel, N.M.; Domingo-Ferrer, J. Fl-defender: Combating targeted attacks in federated learning. Knowl.-Based Syst. 2023, 260, 110178. [Google Scholar] [CrossRef]
  53. Pillutla, K.; Kakade, S.M.; Harchaoui, Z. Robust aggregation for federated learning. IEEE Trans. Signal Process. 2022, 70, 1142–1154. [Google Scholar] [CrossRef]
  54. Liu, J.; Li, X.; Liu, X.; Zhang, H.; Miao, Y.; Deng, R.H. DefendFL: A privacy-preserving federated learning scheme against poisoning attacks. IEEE Trans. Neural Netw. Learn. Syst. 2024, 36, 9098–9111. [Google Scholar] [CrossRef] [PubMed]
  55. Chen, H.Y.; Chao, W.L. On Bridging Generic and Personalized Federated Learning for Image Classification. In Proceedings of the International Conference on Learning Representations, Vienna, Austria, 4 May 2021. [Google Scholar]
  56. Chen, C.; Ye, T.; Wang, L.; Gao, M. Learning to generalize in heterogeneous federated networks. In Proceedings of the 31st ACM International Conference on Information & Knowledge Management, Atlanta, GA, USA, 17–21 October 2022; pp. 159–168. [Google Scholar]
  57. Xie, C.; Huang, K.; Chen, P.Y.; Li, B. Dba: Distributed backdoor attacks against federated learning. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  58. Hsu, T.M.H.; Qi, H.; Brown, M. Measuring the effects of non-identical data distribution for federated visual classification. arXiv 2019, arXiv:1909.06335. [Google Scholar]
  59. Fung, C.; Yoon, C.J.M.; Beschastnikh, I. The Limitations of Federated Learning in Sybil Settings. In Proceedings of the 23rd International Symposium on Research in Attacks, Intrusions and Defenses, RAID 2020, San Sebastian, Spain, 14–15 October 2020; USENIX Association: Berkeley, CA, USA, 2020; pp. 301–316. [Google Scholar]
  60. Ozdayi, M.S.; Kantarcioglu, M.; Gel, Y.R. Defending against backdoors in federated learning with robust learning rate. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 19–21 May 2021; Volume 35, pp. 9268–9276. [Google Scholar]
Figure 1. Federated learning-assisted UAV networks.
Figure 1. Federated learning-assisted UAV networks.
Drones 09 00528 g001
Figure 2. Framework of the proposed FedULite.
Figure 2. Framework of the proposed FedULite.
Drones 09 00528 g002
Figure 3. Statistical heterogeneity visualization with different values.
Figure 3. Statistical heterogeneity visualization with different values.
Drones 09 00528 g003
Figure 4. Statistical heterogeneity visualization with different values.
Figure 4. Statistical heterogeneity visualization with different values.
Drones 09 00528 g004
Figure 5. Comparison of different defenses under different ratios of malicious clients. (a: ASR on CSC attack–b: ACC on CSC attack–c: ASR on PGD attack–d: ACC on PGD attack–e: ASR on CCA attack–f: ACC on CCA attack–g: ASR on DBA attack–h: ACC on DBA attack).
Figure 5. Comparison of different defenses under different ratios of malicious clients. (a: ASR on CSC attack–b: ACC on CSC attack–c: ASR on PGD attack–d: ACC on PGD attack–e: ASR on CCA attack–f: ACC on CCA attack–g: ASR on DBA attack–h: ACC on DBA attack).
Drones 09 00528 g005
Figure 6. Comparison of different defenses under different number of clients. (a: ASR on MNIST dataset–b: ACC on MNIST dataset–c: ASR on F-MNIST dataset–d: ACC on F-MNIST dataset–e: ASR on CIFAR-10 dataset–f: ACC on CIFAR-10 dataset).
Figure 6. Comparison of different defenses under different number of clients. (a: ASR on MNIST dataset–b: ACC on MNIST dataset–c: ASR on F-MNIST dataset–d: ACC on F-MNIST dataset–e: ASR on CIFAR-10 dataset–f: ACC on CIFAR-10 dataset).
Drones 09 00528 g006
Figure 7. Performance of different FedULite model structures.
Figure 7. Performance of different FedULite model structures.
Drones 09 00528 g007
Figure 8. Performance of FedULite under varying values of τ . (a: ASR–b: ACC).
Figure 8. Performance of FedULite under varying values of τ . (a: ASR–b: ACC).
Drones 09 00528 g008
Table 1. Performance comparison between FedULite and four SOTA defense methods.
Table 1. Performance comparison between FedULite and four SOTA defense methods.
DatasetsAttackNo-DefenseFoolsGoldMulti-KrumRLRFL-DefenderFedULite
ASRACCASRACCASRACCASRACCASRACCASRACC
MNISTCSC87.3382.088.4580.264.7180.336.7282.4975.5479.581.9792.08
PGD89.8891.9413.0682.374.5888.460.4684.2173.1386.812.4193.66
CCA91.5992.5812.8782.294.1287.494.1688.7586.1282.550.6790.43
DBA99.190.2796.9278.446.9188.085.687.9687.4882.864.3390.2
F-MNISTCSC79.0682.546.881.63.8980.345.181.444.3580.160.2590.74
PGD90.5490.1712.1379.352.288.682.6986.572.9682.442.491.12
CCA90.5694.349.581.422.8689.174.6884.562.5784.631.6388.16
DBA99.1189.9597.779.417.4286.327.3283.8410.6982.823.2689.7
CIFAR-10CSC77.3487.9310.9778.392.1481.3710.181.5314.980.251.7283.96
PGD88.586.89.0876.726.0177.766.1671.764.8877.962.3488.92
CCA94.8783.278.880.462.5382.517.5981.0366.8279.681.1186.42
DBA93.7478.7594.3670.078.1470.679.7478.3480.3268.842.9186.91
Table 2. Performance comparison between FedULite and four SOTA defenses on non-IID settings.
Table 2. Performance comparison between FedULite and four SOTA defenses on non-IID settings.
DatasetsSettingNo-DefenseFoolsGoldMulti-KrumRLRFL-DefenderFedULite
ASRACCASRACCASRACCASRACCASRACCASRACC
MNIST α = 0.0585.0785.6281.2675.4140.9680.8642.8878.0412.5379.111.1882.29
α = 0.189.5887.2287.5579.3622.5981.9627.2481.428.2979.73.4488.47
α = 1099.190.2796.9278.446.9188.085.687.9687.4882.864.3390.2
E-MNIST α = 0.0581.566.4680.4564.2635.3160.3739.6662.777.2762.391.5465.86
α = 0.186.5474.382.3367.6937.2563.6639.1364.6310.2663.145.4873.25
α = 1095.2575.0988.2368.8217.6772.512.9472.796.1970.613.4873.7
Table 3. Performance of FedULite using different components.
Table 3. Performance of FedULite using different components.
DatasetRAA-DFURAA-DRARAA-DALLFedULite
ASRACCASRACCASRACCASRACC
MNIST12.0889.1326.7885.4234.1781.044.3390.2
F-MNIST14.3786.6132.0284.4543.5580.133.2689.7
CIFAR-1013.7483.4839.4584.346.0672.262.9186.91
Table 4. The inversion ratio of the learning rate across different dimensions.
Table 4. The inversion ratio of the learning rate across different dimensions.
SettingEncoderPredictor
12345678
a = 0.050.480.510.460.520.490.370.540.29
a = 0.10.370.410.420.460.360.280.360.26
a = 100.360.390.350.370.320.260.310.25
Table 5. Computational overhead of FedULite.
Table 5. Computational overhead of FedULite.
MethodMobileNetV2ResNet50
Average Training TimeMemory UsageAverage Training TimeMemory Usage
FedAvg0.85 (seconds/epoch)20 MB3.71 (seconds/epoch)90 MB
FL-Defender1.32 (seconds/epoch)28 MB5.82 (seconds/epoch)135 MB
FedULite0.91 (seconds/epoch)22 MB4.05 (seconds/epoch)96 MB
Table 6. Performance of FedULite under different local epochs.
Table 6. Performance of FedULite under different local epochs.
AttackMetricNo-DefenseE = 1E = 2E = 4E = 6E = 8
CSCASR77.341.721.541.131.520.96
ACC87.9383.9684.1486.7187.1587.24
PGDASR88.52.342.561.791.621.57
ACC86.888.9288.7689.0189.2189.35
CCAASR94.871.111.040.5300
ACC83.2786.4286.7287.5887.4287.73
DBAASR93.742.912.181.981.761.05
ACC78.7586.9185.7486.9587.1387.42
Table 7. Performance of FedULite under adaptive attack.
Table 7. Performance of FedULite under adaptive attack.
Adaptive AttackDatasetNo-DefenseFedULite
ASRACCASRACC
MNIST90.63 ± 0.9890.18 ± 0.725.42 ± 1.3290.72 ± 2.71
F-MNIST88.92 ± 1.7389.72 ± 1.893.58 ± 2.2190.18 ± 1.98
CIFAR-1089.15 ± 2.1583.24 ± 5.736.72 ± 1.5488.74 ± 3.46
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, L.; Zhai, W.; Bu, X.; Sun, M.; Zhu, C. A Lightweight Robust Training Method for Defending Model Poisoning Attacks in Federated Learning Assisted UAV Networks. Drones 2025, 9, 528. https://doi.org/10.3390/drones9080528

AMA Style

Chen L, Zhai W, Bu X, Sun M, Zhu C. A Lightweight Robust Training Method for Defending Model Poisoning Attacks in Federated Learning Assisted UAV Networks. Drones. 2025; 9(8):528. https://doi.org/10.3390/drones9080528

Chicago/Turabian Style

Chen, Lucheng, Weiwei Zhai, Xiangfeng Bu, Ming Sun, and Chenglin Zhu. 2025. "A Lightweight Robust Training Method for Defending Model Poisoning Attacks in Federated Learning Assisted UAV Networks" Drones 9, no. 8: 528. https://doi.org/10.3390/drones9080528

APA Style

Chen, L., Zhai, W., Bu, X., Sun, M., & Zhu, C. (2025). A Lightweight Robust Training Method for Defending Model Poisoning Attacks in Federated Learning Assisted UAV Networks. Drones, 9(8), 528. https://doi.org/10.3390/drones9080528

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop