Next Article in Journal
A Federated Learning Based Intrusion Detection Method with Multi-Scale Parallel Convolution and Adaptive Soft Prediction Clustering
Next Article in Special Issue
The Manhattan δ-Corridor: A Maximal Connectivity-Preserving Framework for Scalable Robot Navigation
Previous Article in Journal
Automatic Labeling of Real-World PMU Data: A Weakly Supervised Learning Approach
Previous Article in Special Issue
Energy Efficiency Optimization in Heterogeneous 5G Networks Using DUDe
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

HyFLM: A Hypernetwork-Based Federated Learning with Multidimensional Trajectory Optimization on Diffusion Paths

1
Department of Computer Science and Engineering, Pohang University of Science and Technology (POSTECH), Pohang 37673, Republic of Korea
2
Graduate School of Artificial Intelligence, Pohang University of Science and Technology (POSTECH), Pohang 37673, Republic of Korea
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(23), 4704; https://doi.org/10.3390/electronics14234704
Submission received: 22 October 2025 / Revised: 17 November 2025 / Accepted: 21 November 2025 / Published: 28 November 2025
(This article belongs to the Special Issue Feature Papers in Networks: 2025–2026 Edition)

Abstract

The effective training of large-scale distributed deep learning models has become an active and emerging research area in recent years. Federated learning (FL) can address those challenges by training global models through parameter exchange of client models rather than raw data sharing, thereby preserving security and communication efficiency. However, conventional linear aggregation approaches in FL neglect heterogeneous client models and non-IID data. This often results in inter-layer information imbalance and feature-space misalignment, leading to low overall accuracy and unstable training. To overcome these limitations, we propose HyFLM, a personalized federated learning framework that maximizes performance with Multidimensional Trajectory Optimization theory (MTO) on diffusion paths. HyFLM extends a diffusion-based FL framework by encoding client–parameter dependencies with a diffusion model and precisely controlling dimension-specific paths, thereby generating personalized weights that reflect both the data complexity and the resource constraints of each client. In addition, a lightweight hypernetwork generates client-specific adapters or weights to further enhance personalization. Extensive experiments on multiple benchmarks demonstrate that HyFLM consistently outperforms major baselines in terms of both accuracy and communication efficiency, achieving faster convergence and higher accuracy. Furthermore, ablation studies verify the contribution of MAC to convergence acceleration, confirming that HyFLM is an effective and practical personalized FL paradigm for heterogeneous client models.

1. Introduction

Federated learning (FL) is a distributed machine learning paradigm that enables multiple clients to collaboratively train a shared model without exchanging raw data [1]. In typical FL frameworks, such as FedAvg [2,3,4,5,6,7], each client updates gradients or model parameters from its local dataset to a central server, without sharing their own data. This approach is being studied in various fields, including finance, healthcare, and business. For example, FL can implement intelligent functions on smartphones and IoT devices by learning user data on the device without uploading personal data to the cloud. In healthcare, FL enables hospitals to jointly train diagnostic models based on distributed patient datasets, enabling collaborative medical AI research without compromising patient privacy. Furthermore, in large-scale video surveillance networks such as military perimeter monitoring systems, federated learning (FL) serves as an effective approach to share locally trained model parameters, thus reducing overall training costs and improving the performance of distributed AI models [8,9]. Despite its potential, existing FL faces several fundamental challenges that can degrade performance. One of the most significant issues in FL is the non-Independent and Identically Distributed (non-IID) environment [10], as illustrated in Figure 1. In practical federated learning scenarios, models are often deployed under non-IID conditions, where client data distributions differ significantly [1]. This exposes the problem with simple aggregation methods and leads to inconsistent or biased global models [11]. For example, the well-known Federated Averaging algorithm, FedAvg [2], linearly averages client models as if all client updates were drawn from the same distribution, which can lead to low-probability regions of the parameter space and information collapse [12]. Moreover, averaging two well-optimized client models trained on different data can produce a global model that performs poorly on both client tasks. This feature space collapse or misalignment arises because simple linear combinations do not reconcile the knowledge of different clients. Second, the heterogeneity of the model presents additional limitations. Standard FL frameworks assume the same model architecture for all clients, making it difficult to accommodate small models or unique user-defined models. This uniform approach can lead to misalignment of learned representations and does not account for the various computational resources of the clients [1]. These challenges highlight the need for more sophisticated FL techniques to handle a heterogeneous and non-IID environment, prevent representation collapse, and adapt to heterogeneous client environments.
To overcome these limitations, recent FL research has focused on implementing more advanced aggregation mechanisms for each client and addressing data heterogeneity. One promising solution is personalized Federated Learning (pFL) [12], which addresses the limitations of a single global model by tailoring the learned model to the unique data distribution of each client. Various pFL methods have been proposed, notable examples of which include FedPer [13], which shares a common feature extractor but trains a private output layer for each client; FedRep [14], which trains a federated representation and allows each client to train its own classifier; and FedBABU [15], which fine-tunes a globally pre-trained model for each client without a final aggregation step. These methods often achieve improvements by identifying client-specific requirements and mitigating the negative impact of data heterogeneity. Furthermore, to overcome the limits of simple averaging, recent work has leveraged diffusion models for data-driven aggregation in FL. pFedGPA [12] employs a server-side diffusion model that learns the distribution of client parameters and, via parameter inversion, generates personalized weights for each client, thus encoding client-specific data characteristics. Similarly, FedDiff [16] integrates a diffusion model into the FL loop for multi-modal data fusion, achieving higher accuracy and communication efficiency. By framing model fusion as a generative task, these approaches better capture the complex, high-dimensional structure of parameter space than linear methods.
Building upon diffusion-based FL frameworks, we propose HyFLM, an enhanced FL framework that optimizes model training through inference trajectory optimization. HyFLM shows reliable convergence on heterogeneous data sets. It shows a rapidly decreasing training loss and reaches an accuracy approaching 95% in only a few rounds, indicating both fast convergence and high final accuracy compared to FL baselines. In summary, the contributions of this work are as follows:
  • We propose a novel federated learning framework that combines diffusion-based aggregation and hypernetwork-driven personalization under the theory of inference trajectory optimization.
  • Our empirical results show that the proposed approach achieves superior generalization performance on diverse benchmark datasets compared to well-recognized baselines.
  • We demonstrate that HyFLM yields stable convergence under non-IID data and heterogeneous client models, reduced communication cost, and preserved security.
Section 2 reviews the background and related work on federated learning (FL), hypernetwork-based personalization, and diffusion-based FL, and explains why our work focuses on the intersection of these areas. Section 3 introduces our diffusion-space inference and the Multidimensional Adaptive Coefficients (MAC). Section 4 presents the hypernetwork-driven personalization and its integration with MAC. Section 5 reports experiments result, and Section 6 discusses limitations and concludes.

2. Related Work

This section reviews research most relevant to our contribution, beginning with the foundations of federated learning and personalized FL under non-IID client distributions, then progressing to hypernetwork-based approaches that condition on client descriptors to produce lightweight adapters for client-specific models, and culminating with diffusion-based FL that leverages denoising/score models for robust aggregation and trajectory control.

2.1. Federated Learning

Federated learning has emerged as a prominent paradigm for training models on decentralized AI environment. The landmark FedAvg [2] algorithm showed that simply averaging the parameters of the client model can produce a viable global model, even in unbalanced non-IID data distributions. However, subsequent studies have shown that data heterogeneity remains a fundamental bottleneck. The so-called client drift arises because each client often converges to a local optimum that does not coincide with the global optimum, causing local updates to deviate and degrade the accuracy of the global model [1]. To combat this, numerous FL frameworks have been proposed to stabilize training on non-IID data. For example, FedProx [17] adds a proximal regularization term to keep local model updates closer to global weights, improving convergence under statistical heterogeneity. Similarly, SCAFFOLD [18] uses control variates (server and client update correction vectors) to explicitly mitigate client drift during local training. Another research strand is personalized FL [13] which seeks to optimize each client’s own performance rather than relying on a single global model. Another approach is cluster-based FL [19,20], which partitions clients into groups with similar data distributions and performs model averaging within each cluster [10]. This reduces cross-client divergence, but producing high-quality personalized models and efficiently sharing knowledge between clusters remains an open challenge. Recent works [1,21] explore fine-grained personalization, but achieving both high accuracy and fairness in large-scale, highly non-IID settings continues to be actively researched.

2.2. Hypernetwork-Based FL

A hypernetwork is a neural network that generates (all or part of) the weights of another target network. This concept has gained traction as a tool for tackling model heterogeneity and personalization in FL. Early federated studies used hypernetwork mainly to provide better initial weights for each client model [22], but recent work conditions weight generation on richer signals such as client metadata, update gradients, or graph-based relationships between clients. For example, hFedF [23] (federated fusion based on the hypernetwork) embeds each client’s model update into a latent state and uses a server-side hypernetwork to fuse models in a non-linear fashion, producing a robust global model that outperforms linear FedAvg under the non-IID condition. HyperFedNet [24] addresses the communication overhead and privacy risks of conventional FL by transmitting a group of hypernetwork parameters. Meanwhile, FedAGHN [25] employs an attentive graph hypernetwork to personalize models. It treats clients as nodes in a graph and uses attention-weighted edges to learn dynamic collaboration weights. As a result, the hypernetwork can produce a customized initial model for each client by aggregating parameters from neighboring clients. This approach maximizes cooperation among statistically similar clients while preserving individual performance. Likewise, FedSheafHN [26] leverages advanced graph signal processing. It constructs a server-side sheaf diffusion process over a client collaboration graph to derive high-order client embeddings (capturing complex inter-client relationships), then feeds those embeddings into a hypernetwork that generates personalized Graph Neural Network parameters for each client. Hypernetwork generation has shown strong gains on graph-structured data and unseen clients. Recent work enriches conditioning inputs (e.g., graphs, contrastive embeddings) and employs auxiliary objectives such as adversarial or contrastive losses to improve robustness under heterogeneity. In general, advances in hypernetwork-based FL yield higher accuracy and faster convergence, marking it as a promising direction to improve personalization and communication efficiency.

2.3. Diffusion Based Federated Learning

Diffusion models have attracted substantial interest in federated learning. With a solid theoretical foundation and a simple yet effective loss formulation [27], these models have the ability to generate high-quality data for a wide range of applications. Recently, the synergy between diffusion generative models and FL has been actively explored to mitigate non-IID challenges and improve generalization [28,29]. Denoising diffusion models (DDPMs [27]) can synthesize realistic data by iteratively refining noise, and several studies have leveraged this capability within FL. One line of work uses diffusion for synthetic data augmentation [30], where local diffusion models generate auxiliary samples to balance skewed distributions and improve global convergence. Frameworks such as FedDDPM [31] and FedDrip [32] have demonstrated consistent gains across benchmarks, while additional studies showed that diffusion-based augmentation boosts performance on highly imbalanced datasets, alleviating client drift and mode collapse. Another direction applies diffusion to model fusion, replacing simple averaging with generative aggregation. For example, pFedGPA [12] employs parameter inversion and denoising sampling to generate refined personalized weights, achieving better performance over FedAvg [2]. FedDiff [16] extends this idea to multi-modal settings, coordinating diffusion branches across modalities (e.g., hyperspectral vs. LiDAR) and enabling privacy-preserving feature fusion. Finally, diffusion has been explored as an enhanced training signal, where large diffusion models act as auxiliary teachers to provide generative semantic guidance and consistency regularization, improving alignment of heterogeneous client representations. Such diffusion-based approaches are emerging as versatile tools in FL, enabling balanced data generation, advanced parameter aggregation, and richer training signals. Furthermore, recent advances in diffusion-based FL highlight that optimizing inference trajectories is becoming a critical challenge. The present work introduces a novel methodology to address this issue.
Classical FL with FedAvg-style aggregation degrades under non-IID partitions. In personalization, hypernetwork offers a streamlined way to produce client-conditioned weights from compact server-side models. Diffusion-based FL improves robustness and exploration, but typically lacks explicit trajectory control and a coherent link to per-client adaptation. Building on these observations, HyFLM integrates hypernetwork-based FL with multidimensional trajectory optimization (MTO) on diffusion paths to provide time-varying trajectory control while the hypernetwork injects client specific signals yielding a robust, communication-aware personalization framework.

3. Preliminaries

This section briefly reviews the two theoretical pillars used in HyFLM: Denoising Diffusion Probabilistic Models (DDPM) [27] and the more recent Multidimensional Adaptive Coefficient (MAC) [33]. Understanding these concepts clarifies how HyFLM leverages diffusion-based generative modeling and adaptive control of inference trajectories.

3.1. Design Diffusion-Based Inference Space

In a federated learning environment with non-IID data and heterogeneous architectures, naïve parameter averaging neglects geometric coupling across the multiple layers of the weight space and can degrade accuracy. To address this, we construct a diffusion-based inference space that models the empirical distribution of client weights via a forward noising and reverse denoising process. This enables the server to synthesize a global anchor while preserving the curvature and scale of the layer. Coupling this space with Multidimensional Adaptive Coefficients (MAC) provides trajectory-level control. DDPMs can be used in FL training. They define a forward diffusion process that adds noise to the data step by setp and a reverse denoising process that learns to invert this perturbation [33]. The generative model is parameterized as a Markov chain with latent variables { x 1 , , x T } and observed data x 0 . We denote the forward process by q ( x t x t 1 ) and the learned reverse chain by p θ ( x t 1 x t ) , with a fixed Gaussian prior p ( x T ) = N ( 0 , I ) . The reverse process is formulated as
p θ ( x 0 : T ) = p θ ( x 0 ) t = 1 T p θ ( x t 1 x t ) ,
where p θ ( x t 1 x t ) denotes Gaussian transitions with learned mean and covariance μ θ ( x t , t ) and Σ θ ( x t , t )  [27]. The approximate posterior (also called the forward or diffusion process) is fixed to a Markov chain that gradually corrupts data by adding Gaussian noise at each step [27]:
q ( x t x t 1 ) = N ( x t ; 1 β t x t 1 , β t I ) ,
where β t ( 0 , 1 ) is a small variance schedule. Following the formulation in [27], repeated application of this transition gradually drives the data distribution toward a standard normal, enabling sampling to begin from pure noise, as illustrated in Figure 2.
To synthesize data, DDPMs train a neural network to model the reverse transition p θ ( x t 1 x t ) , thus learning to iteratively denoise perturbed samples. The training objective minimizes a variational lower bound on the negative log-likelihood, which simplifies to a standard denoising loss function.
A specific parameterization reveals equivalence with denoising score matching and annealed Langevin dynamics, and the sampling procedure resembles a progressive decoding mechanism. These properties make diffusion models powerful building blocks for FL frameworks that can handle diverse client data and produce high-quality generative output. These methods are incorporated into HyFLM to construct client-specific diffusion trajectories while preserving global generative capability.

3.2. Multidimensional Adaptive Coefficient (MAC) for Inference Trajectory Optimization

In conventional diffusion or flow models, a single global scaling factor is applied to the entire trajectory, limiting the flexibility of the model. To overcome this restriction, we introduce the Multidimensional Adaptive Coefficient (MAC) [33], which enables different components of the model update to scale adaptively over time, thus improving trajectory expressiveness and optimization efficiency, as illustrated in Figure 3.
Flow and diffusion-based models typically employ one-dimensional scalar coefficients to control inference trajectories, limiting adaptability across different datasets or tasks. This scalar formulation imposes two main constraints. The Multidimensional Adaptive Coefficient (MAC) addresses this limitation by extending the conventional unidimensional coefficient to higher dimensions. Given the trajectory x ( t ) = γ 0 ( t ) x 0 + γ 1 ( t ) x 1 , x 0 , x 1 R d and γ ( t ) C 1 ( [ 0 , T ] , R d × d × 2 ), the multi-dimensional coefficient is defined as:
γ ( t ) = [ γ 0 ( t ) , γ 1 ( t ) ] : [ 0 , T ] R d × d × 2
The Multidimensional Adaptive Coefficient (MAC) is defined as γ ϕ , which dynamically adjusts to each inference trajectory x θ , ϕ ( t ) defined over discrete inference times τ = { t 0 , , t N } . Formally, the MAC is defined as
γ ϕ ( t , x θ , ϕ S ) = [ γ 0 , ϕ ( t , x θ , ϕ S ) , γ 1 , ϕ ( t , x θ , ϕ S ) ] : [ 0 , T ] × R l × d R d × d × 2 ,
where S = { t ( 1 ) , , t ( l ) } [ 0 , T ] represents an arbitrary subset of inference times and ϕ parameterizes the adaptive behavior. Boundary and smoothness properties are preserved, allowing the MAC to flexibly model sample- and time-dependent trajectories during diffusion-based inference.
To handle high-dimensional datasets, formulate the trajectory optimization as a GAN-based min–max problem:
min θ , ϕ max ψ E x 0 log D ψ ( x 0 ) + E x T log 1 D ψ ( G θ , ϕ ( τ , x T ) ) ,
Here, D ψ acts as a discriminator that distinguishes real samples x 0 from generated samples G θ , ϕ ( τ , x T ) , while G θ , ϕ jointly optimizes the vector field θ and the MAC parameters ϕ . This adversarial formulation enables implicit divergence minimization and improves transport quality in diffusion-based high-dimensional inference.
In summary, DDPM provides a principled probabilistic foundation for modeling generative processes via forward diffusion and learned reverse denoising. MAC extends this framework by introducing learnable multidimensional coefficients that control the trajectory of the client weighty adaptively. These preliminaries form the conceptual basis of HyFLM, where a personalized hypernetwork leverages MAC-modulated diffusion paths for flexible and high-quality federated learning.

4. Methodology

HyFLM is proposed as a robust and efficient federated learning framework. It effectively learns client-specific weights through adaptive coefficient optimization in a diffusion-based environment while simultaneously generating personalized client models via a server-side hypernetwork. As illustrated in Figure 4 HyFLM includes two main components: (1) client-side training with an advanced MAC (which adaptively scales update directions at each step and gates residual connections and message signals); and (2) a lightweight server-side hypernetwork for personalization. The advanced MAC regulates how the global model evolves. This enables HyFLM to capture client-specific parameter dependencies and deliver optimal and more personalized aggregation trajectories than previous approaches. Another important element of HyFLM is a lightweight hypernetwork that enhances personalization for resource-constrained or biased clients. The hypernetwork conditions the diffusion-based aggregator on client-specific signals, adjusting the generated parameters to reflect data biases or resource constraints. However, client-specific weight generation is achieved through a hypernetwork on the server side. By integrating and extending these components, HyFLM can minimize the distortion of the non-IID environment, address model heterogeneity, and provide efficient communication personalization across diverse clients. This section details each module and outlines the overall process of integrating them; the detailed procedure of HyFLM is summarized in Algorithm 1.
Algorithm 1: HyFLM: A Hypernetwork-Based Federated Learning with Multidimensional Trajectory Optimization
Electronics 14 04704 i001

4.1. Research Methodology Overview

The research focuses on personalized federated learning in realistic heterogeneous environments, where many clients train on non-IID data under constrained communication and computation costs. In addition, the research aims to improve the average accuracy per-client in statistically biased partitions, lower the total communication required to reach a given target accuracy, and strengthen convergence stability and robustness among diverse client populations. A final objective is to design a federated learning framework that scales gracefully with the number of clients and the available system resources, so that high performance remains attainable even in large-scale deployments. These aims motivate several concrete research problems: whether the proposed method can surpass strong existing baselines in both early and final training stages under the simulation environment; how much of the observed improvement is attributable to diffusion-path trajectory control via the MAC module and how much to hypernetwork-based personalization, individually and in combination; how the communication–computation trade-offs evolve as key design parameters such as the number of function evaluations, adapter capacity, and update frequency vary; and how stable the training dynamics remain under severe non-IID conditions and large client pools.
To investigate these problems, the work adopts a controlled empirical methodology with fixed data splits and equalized training budgets across all compared methods. The experimental pipeline is implemented in PyTorch with a simulated server–client FL environment that orchestrates local training, aggregation, and personalization. The proposed technique integrates MAC-gated diffusion sampling on the server with lightweight hypernetwork adapters that generate client-specific parameter adjustments from learned embeddings. Multiple random seeds are used and results are reported with mean and standard deviation to quantify variance. Communication cost is modeled as the product of the number of communication rounds and the payload per round, while computational effort is approximated through FLOPs-based proxies. Independent variables include the number of clients and their participation ratio, the Dirichlet concentration parameter controlling the degree of non-IID partitioning, the number of function evaluations for diffusion, the rank or bitwidth of the personalization adapters, and the regeneration interval for global updates. Dependent variables comprise personalized test accuracy, the number of rounds required to reach a performance plateau or a fixed accuracy fraction, payload per round, total data volume exchanged, and stability indicators; together these metrics support a systematic analysis of accuracy–efficiency trade-offs under identical resource constraints.

4.2. Diffusion-Based Training and Trajectory Optimization with MAC

In federated learning (FL) systems characterized by irregular data distributions and heterogeneous client features, diffusion-based FL approaches leverage the generative capabilities of diffusion models to capture the data distribution of each client and produce personalized parameters accordingly. The Multidimensional Adaptive Coefficient (MAC), denoted γ ϕ , is a novel coefficient introduced to optimize inference trajectories in flow and diffusion models. In this study, MAC is applied within FL to optimize the inference trajectories of temporally aligned flow and diffusion models between the global and client models, thereby generating client-specific personalized parameters tailored to each client’s characteristics. The entire framework operates through a compact iterative loop of pretraining and local update → aggregation and correction. The core idea of this approach is to integrate powerful diffusion-based generative modeling into federated learning, enabling the learning of the multi-dimensional distribution of client model parameters within the parameter space and improving both adaptation and personalization across heterogeneous environments.
Before global model training, each client performs a brief pre-training phase using randomly sampled MAC coefficients to adapt H θ for a stable response to non-linear (curved) trajectories. In each communication round, the server broadcasts the current model ( θ , φ ) to a subset of clients. Each client k updates both θ k and φ k by minimizing the local objective:
θ k θ η θ L k ( θ , φ ) , φ k φ η φ L k ( θ , φ ) ,
where η is the learning rate. This procedure jointly optimizes the vector field  ( θ ) and the trajectory plan ( φ ) , adapting both to the data distribution of each client and the resource constraints.
In normalized time t [ 0 , 1 ] , HyFLM adopts an advanced Multidimensional Adaptive Coefficient based on the theory of the Multidimensional Adaptive Coefficient (MAC), which dynamically balances exploration and stability during diffusion-based optimization. Specifically, these gates regulate the message and residual paths:
  • γ msg ( t ) : Scales the magnitude of the diffusion update (drift or score term);
  • γ res ( t ) : Controls the residual pull toward an anchor (e.g., the previous global model).
The server-side model updates govern the global diffusion process, where the server integrates information from client models to reconstruct and refine the global parameter distribution. Each reverse step applies MAC-modulated scaling to both the diffusion score and the residual path, thereby adapting the generative trajectory to the global dynamics of heterogeneous client updates. The update rule is formally expressed as
x k + 1 = x k + h k γ msg ( t k ) s ϕ ( x k , t k ) + h k γ res ( t k ) ( a x k ) + σ k z k ,
where t k = k / ( N 1 ) , h k denotes the size of the integration step, a is the anchor representing the global reference model, and z k N ( 0 , I ) introduces stochastic regularization through the variance schedule σ k .
The client-side local update applies MAC gating to control both the gradient descent magnitude and the retention of prior knowledge during local optimization. This mechanism enables each client to follow an individualized optimization trajectory that reflects its data distribution and computational characteristics, preventing unstable divergence in heterogeneous settings. The update process is defined as:
g τ = γ msg ( t ) L w ( τ ) , t = τ max ( 1 , T loc 1 ) , w ˜ ( τ + 1 ) = w ( τ ) η g τ , w ( τ + 1 ) = w ˜ ( τ + 1 ) + α γ res ( t ) w ( 0 ) w ˜ ( τ + 1 ) ,
where η denotes the learning rate and α represents the residual coefficient. In this formulation, γ msg ( t ) adaptively scales the gradient to regulate the intensity of learning at each step, effectively serving as a temporal controller of the learning rate, while γ res ( t ) enforces a smooth proximal correction towards the initial model w ( 0 ) . These gates ensure that client updates remain both locally responsive and globally consistent, thus improving convergence stability under non-IID and resource-constrained environments.
To achieve per-dimension and time-adaptive control, the gates are implemented as vector-valued functions (or per-group vectors) with smoothing to prevent overfitting. Each gate is parameterized by a truncated low-pass Fourier expansion.
γ ( t ) = Π [ γ min , γ max ] b + k = 1 K a k cos 2 π k t + b k sin 2 π k t e σ 2 k 2 ,
where { a k , b k , b } are learnable coefficients (optionally vector-valued), σ > 0 denotes the bandwidth, and Π is the projection operator that clamps outputs to [ γ min , γ max ] . We parameterize γ ( t ) with a truncated low-pass Fourier expansion (Equation (9)), projected to [ γ min , γ max ] . This guarantees smooth, bounded gates, enables multi-timescale control of the diffusion/local update, and has O ( K d ) per-step evaluation cost suitable for FL. With the spectral envelope e σ 2 k 2 and projection Π [ γ min , γ max ] , γ ( t ) is bounded and differentiable, and γ ( t ) k = 1 K 2 π k c k e σ 2 k 2 . Hence MAC-modulated increments in Equations (7) and (8) remain well-behaved. When γ msg 1 and γ res 0 , the method reduces to standard sampling/gradient descent. When γ msg 1 and γ res 0 , this formulation reduces to standard sampling or gradient descent.

4.3. Hypernetwork-Based Personalization

A hypernetwork is a meta-generator network that produces the parameters of another target network, enabling flexible adaptation between tasks or clients. In the proposed framework, we employ a hypernetwork H ψ that maps client embeddings to personalized weight adjustments. Given the embedding matrix X ^ (or individual embedding e i ), the hypernetwork generates client-specific parameters:
Ω ˜ = { ω ˜ 1 , ω ˜ 2 , , ω ˜ K } , with ω ˜ i = H ψ ( e i ) .
Each ω ˜ i represents an adaptation of the shared model θ for client i, associated with an embedding vector e i in our implementation. This embedding encodes client-specific information—such as data statistics or learned latent features—and is utilized both to modulate the diffusion process through the MAC and to serve as input to the hypernetwork. This design ensures that personalization is consistently guided by client characteristics, yielding coherent and data-informed parameter generation across clients.
w ˜ i = θ Δ w i ,
where ⊕ denotes the composition or addition of a delta. This enables efficient personalization without retraining the denoising network ϵ ϕ on each client. The same client embedding used in MAC gating is reused here, ensuring consistent characterization of client heterogeneity across both local trajectory modulation and global parameter generation. In this way, the hypernetwork serves as a conditional generator that maps heterogeneous client embeddings to individualized model configurations, thereby enabling scalable and communication-efficient personalization within the federated learning framework. HyFLM integrates MAC gating into both client-side optimization and server-side diffusion to modulate update magnitudes and trajectories across dimensions. As a result, it better captures individual client characteristics and improves stability, personalization, and robustness under heterogeneous data. The vectorized, Fourier-parameterized gates enable per-parameter and time-adaptive control over learning and diffusion. The hypernetwork personalizes the final model for each client, yielding scalable personalization. Collectively, these innovations ensure stability, improved convergence, and robust performance in non-IID federated environments.

5. Experiments

5.1. Experimental Environment

All experiments were conducted on a Linux workstation with 4× NVIDIA RTX A6000 GPUs (48 GB each; NVIDIA Corporation, Santa Clara, CA, USA), CUDA 12.4, and NVIDIA driver 550.54.14. The system ran Ubuntu 22.04 with Python 3.10 and PyTorch 2.1. Our federated learning simulation was built using PyTorch and Flower/PySyft primitives, and diffusion components were adapted from Hugging Face’s diffusers.Resource usage was recorded with Weights & Biases. All experiments were containerized and run with fixed seeds for reproducibility.

5.2. Dataset

HyFLM was evaluated on three standard vision benchmarks EMNIST, Fashion-MNIST, and CIFAR-10 under varying numbers of clients and baseline settings. Each data set was first treated as a single global pool following the standard torchvision train/test split. Subsequently, the partitions of non-IID clients were constructed to simulate the label-skew and quantity-skew scenarios for M { 10 , 20 , 100 } clients, as described in Table 1.
  • EMNIST. Handwritten English letters (26 classes), grayscale 28 × 28 , which reflects on-device handwriting variability and serves as a classic benchmark.
  • Fashion-MNIST. Grayscale clothing images (10 classes) at 28 × 28 . It provides more shape/texture diversity than EMNIST and is suitable for lightweight CNNs.
  • CIFAR-10. Natural RGB images (10 classes) at 32 × 32 . Stresses color/texture cues and tests FL robustness under higher visual complexity.

5.3. Experiment

HyFLM is compared with representative personalized FL baselines, including global-only, partial-sharing, meta-initialization, hypernetwork, and diffusion-based aggregation methods. Baselines were chosen primarily from methods that are commonly adopted and reported in related work, to enhance the objectivity and reliability of our empirical evaluation. All models are trained in an identical computational and communication environment.

5.3.1. Baseline Methods

  • Local—Each client independently trains its own model on local data without any communication or parameter sharing.
  • FedAvg—The standard federated averaging algorithm with a single global model and aggregated each client model.
  • FedPer—Partial personalization that shares a common feature extractor while each client keeps a private classification head.
  • LG-FedAvg—Layer-grouped FedAvg updates lower layers globally while upper layers remain local, allowing architectural heterogeneity with minimal communication.
  • pFedHN—Hypernetwork-based personalization where a server-side hypernetwork generates (parts of) client weights or adapters conditioned on a client embedding.
  • pFedGPA—Diffusion-based generative parameter aggregation; the server learns a diffusion model over client parameters, performs parameter inversion, and synthesizes personalized weights by denoising sampling.

5.3.2. Implementation Details

For EMNIST and Fashion-MNIST we use a lightweight CNN (2conv + 1fc); for CIFAR-10 we use a CIFAR-CNN (3conv + 2fc). All baselines share the same backbone on a given dataset. CIFAR-10 uses standard augmentation.
Experiments are conducted with M 10 , 20 , 100 clients. In each round, only a part of clients is sampled (50% of clients when M 20 and 20% when M = 100 ). Unless otherwise stated, training runs for 300 rounds in EMNIST, Fashion-MNIST, and CIFAR-10. Local epochs are set to E 1 , 5 , 10 depending on the dataset size (default E = 5 ). We used a fixed batch size of 64 for all experiments.
All baselines are implemented using Python with unified optimizer configurations to ensure consistency across datasets. Unless otherwise required, clients adjust a learning rate of 0.05 for EMNIST/Fashion-MNIST and 0.01 for CIFAR-10, momentum 0.5 , and weight decay 5 × 10 4 . For CIFAR-10, a cosine learning rate decay is employed throughout the training. Overfitting is mitigated by early stopping with a patience of 10 based on the validation accuracy. All reported results correspond to experimental data over five runs with different initialization seeds.
HyFLM extends diffusion-based aggregation with MAC-guided trajectory control and a lightweight personalization head. We use s = 20 denoising steps per client and the following MAC settings (also used in ablations): Fourier order K = 6 , bounds γ min = 0 , γ max = 2 , low-pass coefficient σ = 0.35 , and residual pull α res = 0.05 . The progression-to-time mapping is linear within each local update t = u / ( U 1 ) . During local training, MAC scales the message (gradient) by γ msg ( t ) and applies a residual pull toward the start weights of the round with γ res ( t ) .

6. Results and Discussion

6.1. Experimental Result

The experiments are designed to comprehensively evaluate the proposed HyFLM framework across multiple dimensions of federated learning performance. In particular, we focus on three key aspects: (1) personalized accuracy across heterogeneous clients, (2) communication efficiency measured by convergence speed and payload cost, and (3) scalability with respect to the number of participating clients and dataset complexity. Through these evaluations, we aim to verify both the effectiveness and practicality of HyFLM in real-world federated settings.

6.1.1. Personalization Performance Comparison

We compare HyFLM with three benchmark datasets, EMNIST, Fashion-MNIST (FMNIST), and CIFAR-10, and representative FL baselines: FedAvg, FedPer, LG-FedAvg, pFedHN, and pFedGPA. As shown in Table 2, HyFLM achieves the highest accuracy in each of the nine experimental settings (three datasets × three client counts). Its overall mean accuracy is approximately 86.8% across all configurations, substantially outperforming the strongest baseline.
HyFLM attains 78.4%, 80.0%, and 82.9% accuracy with 10, 20, and 100 clients on EMNIST. These results surpass pFedGPA’s 78.4%, 80.0%, and 82.9% under the same conditions. Other strong personalized baselines remain lower; for example, pFedHN tops out at ∼82.2% and FedPer at ∼78.7% (in the 100-client EMNIST case), both well below HyFLM’s 88.5%. Similarly, on Fashion-MNIST, HyFLM achieves 89.0%, 89.8%, and 91.1% for 10, 20, and 100 clients, outperforming FedAvg (82.23%, 84.96%, 84.15%) by approximately 6–9 percentage points in each case. In the more challenging CIFAR-10 dataset, all methods obtain lower accuracies; HyFLM reaches 74.7%, 76.7%, and 78.1% for 10, 20, and 100 clients, maintaining a slight edge over pFedGPA (70.10%, 71.90%, 74.20%). Other personalized approaches also struggle on CIFAR-10 (for example, FedAvg peaks at ∼70.55% with 100 clients), underscoring the challenge of this dataset and the competitiveness of HyFLM’s result.
Increasing the number of clients generally boosts performance for most methods, but HyFLM stands out in its scalability and robustness. HyFLM’s accuracy consistently improves as more clients participate, indicating effective utilization of larger distributed data and resilience to heightened data heterogeneity. For example, on CIFAR-10, HyFLM improves from 74.7% (with 10 clients) to 78.6% (100 clients), and on EMNIST from 84.5% to 88.5%, demonstrating solid gains with more clients. Although baselines like FedAvg and pFedGPA also improve with scale (e.g., FedAvg from ∼65% to ∼66.2% in CIFAR-10), their gains are more modest or plateau earlier. In fact, beyond 20 clients, pFedGPA shows diminishing returns on CIFAR-10 (e.g., 70.1% at 20 clients vs. 74.9% at 100). Moreover, pFedHN’s performance actually degrades at larger scales (on EMNIST it drops from 78.12% at 20 clients to 77.2% at 100). In contrast, HyFLM continues to improve and maintains high accuracy even with 100 clients. This underscores HyFLM’s robust personalization capability in large federations.
The characteristics of the data set further influence performance gaps. All methods achieve higher absolute accuracy on the simpler character/handwritten digit tasks (EMNIST and FMNIST) than on the complex CIFAR-10 images. For example, FedAvg reaches 82.23% on FMNIST (20 clients) but only 65.27% on CIFAR-10; similarly, pFedGPA achieves 86.6% versus 70.1%. HyFLM follows this trend, obtaining 89.0% on FMNIST but 74.7% on CIFAR-10 (100 clients). Importantly, HyFLM’s improvement over the best baseline is more pronounced in the easiest data sets (e.g., approximately +8–9% on FMNIST), while in CIFAR-10 the margin narrows to less than 1%. This suggests that while HyFLM excels across the board, its relative advantage is especially significant when task complexity is moderate.
HyFLM attains a higher precision and comparable or faster convergence than the diffusion-based baseline (pFedGPA) in all six comparisons from Figure 5. In EMNIST, HyFLM reaches 84.5% (10 clients, round 161), 86.0% (20 clients, round 127) and 88.5% (100 clients, round 270), outperforming pFedGPA’ 78.4%, 80.0%, and 82.9%, respectively – gains of roughly +6.1, +6.0, and +5.6 percentage points. In Fashion-MNIST, HyFLM achieves 88.95% (10 clients, round 156), 89.78% (20 clients, round 276), and 91.12% (100 clients, round 299), exceeding pFedGPA’s 85.6%, 87.0%, and 85.6% by approximately +3.3, +2.8, and +5.5 points. In all settings, the curve of HyFLM stays above the baseline early in the training and maintains a higher asymptote, indicating improved sample efficiency (fewer rounds to surpass the baseline plateau) and better final precision. In particular, the margin widens at larger client counts (100-client EMNIST/FMNIST panels), suggesting stronger robustness to heightened statistical heterogeneity. Therefore, the graphs corroborate the consistent accuracy advantage, stable convergence, and communication efficiency of HyFLM with increasing federation size.

6.1.2. Communication Efficiency and Scalability

In Table 3, we summarize accuracy and communication cost under practical settings. Acc@Last denotes the final personalized test accuracy, and R@95% represents the number of rounds required to reach 95% of this value. The communication cost is quantified by the payload per round, defined as the total traffic from server broadcast and client uploads. The total communication volume (Total MB) is computed as R @ 95 % × Payload / round . HyFLM shows better communication efficiency by reducing the payload per round compared to pFedGPA in Figure 5. Due to its hierarchical model-sharing strategy, HyFLM can transmit fewer or more compact parameters each round, so that when pFedGPA sends a full model update, HyFLM often needs only a smaller embedded update (as reflected by the lower Payload/round in Table 3). As a result, the total communication volume required for HyFLM to reach near-top accuracy is significantly lower. For example, on CIFAR-10 with 100 clients, The combination of faster convergence and reduced per-round payload yields approximately a 20–30% reduction in total data transfer compared to pFedGPA. Similar trends are observed on EMNIST and FMNIST, where HyFLM achieves smaller R@95% and Total MB, indicating more efficient bandwidth usage without sacrificing accuracy.
Both methods exhibit robust scaling as the number of clients increases from 10 to 100, albeit with some differences in behavior on simpler vs. more complex tasks. In general, adding more clients (which introduces more data diversity but also more heterogeneity) does not degrade final accuracy in these experiments—in fact, the EMNIST and FMNIST results show a slight increase in accuracy at 100 clients relative to 10. For example, EMNIST accuracy improves from the low– 80 % range with 10 clients to about 88.5 % with 100 clients for both algorithms, suggesting that a broader client participation helped the models generalize better. HyFLM handles this scaling gracefully: its accuracy scales at least as well as pFedGPA’s, while requiring fewer rounds to converge.
On the more challenging CIFAR-10 dataset (which has more complex data distributions), overall accuracy is lower for both methods, and the benefit of more clients is less pronounced. At 100 clients, pFedGPA and HyFLM reach high– 70 % accuracy on CIFAR-10, which is an expected drop compared to the easier datasets. Importantly, HyFLM’s performance relative to other baselines improves with scale and complexity. HyFLM consistently maintains the highest accuracy across all client scales on CIFAR-10, and its performance gap relative to strong baselines becomes more evident as the number of clients increases. This suggests that HyFLM is particularly effective in large-scale, heterogeneous environment.
In terms of communication scaling, as the number of clients increases, the number of rounds required to achieve R@95 tends to increase substantially. However, HyFLM can mitigates this effect. Even with 100 clients, HyFLM maintains a low R@95% and moderate communication cost, demonstrating better scalability. Overall, the results in Table 3 show that HyFLM not only matches or exceeds pFedGPA in accuracy across different workloads, but does so with greater communication efficiency and resilience as the federated network scales up.

6.2. Discussion

Interpretation of Key Results: Our method (HyFLM) consistently shows superior personalization accuracy across varying degrees of non-IID data distribution. Notably, it outperforms strong baselines not only in final accuracy but even from the early rounds of training. This suggests that the diffusion-based inference trajectory optimization and hypernetwork-driven updates in HyFLM enable faster initial alignment to each client’s data while preserving global consistency, maintaining a balance that baseline methods struggle to achieve so early. Beyond accuracy, communication efficiency is markedly improved. HyFLM reaches convergence in fewer communication rounds thanks to the accelerated learning provided by the MAC module and hypernetwork. Moreover, the updates exchanged each round are smaller and more modular: instead of sending full model parameters, clients and server share compact hypernetwork outputs or anchor adjustments, significantly reducing the payload. Equally important, HyFLM demonstrates stability at scale. In experiments with up to 100 clients, a regime where many federated methods degrade due to statistical heterogeneity and client drift, HyFLM maintained stable and high performance. We attribute this robustness to the globally guided diffusion process, which MAC drives client updates and mitigates divergent drifts even as the client population grows.
Relation to Literature: HyFLM effectively integrates personalized federated learning with diffusion-based aggregation by addressing key limitations in prior approaches. Traditional personalized FL method focus on local adaptation but lack global coordination, often resulting in client drift. HyFLM mitigates this by using diffusion-based inference trajectory optimization, offering globally informed yet client-specific updates. Unlike prior diffusion FL methods, which apply uniform trajectories, the HyFLM MAC module enables per-timestep, per-client modulation, enhancing personalization precision. Additionally, compared to hypernetwork-based methods, HyFLM offers improved efficiency and adaptability through its lightweight hypernetwork coupled with evolving diffusion anchors. This unified design achieves scalable and fine-grained personalization with better alignment between clients.
Overall, We have highlighted why HyFLM achieves better personalization (through globally anchored yet client-specific updates), under what conditions it yields the greatest benefits (moderate heterogeneity and network size), and how it connects to and improves upon prior art in federated learning. These insights clarify the contexts in which the proposed model excels, as well as its limitations, guiding practitioners on when HyFLM is the most appropriate choice and where future refinements could further enhance its practicality.

6.3. Future Work

While HyFLM has shown strong scalability and effectiveness across a wide range of standard benchmark conditions, additional research is necessary to extend its practicality to ultra-large federated networks involving thousands or even tens of thousands of clients. In such extreme-scale scenarios, personalized weight generation via a server-side hypernetwork may lead to increased computational load and communication overhead, especially when handling high-dimensional client embeddings or supporting complex personalization adapters.
To address this, we will consider to incorporate clustering or grouping techniques that partition clients based on similarity in data distribution, resource profile, or model updates. Such stratification would enable the hypernetwork to generate shared personalization templates per group, thus amortizing the cost of per-client weight generation and reducing redundant computation. This approach has the potential to preserve the quality of personalization while significantly improving communication and computational efficiency. In addition, our goal is to investigate hierarchical personalization strategies, where coarse-grained anchors are produced at the group level and refined locally with lightweight updates.

7. Conclusions

This paper proposes HyFLM, a personalized federated learning framework that combines diffusion-based model aggregation with hypernetwork-driven personalization to address challenges posed by non-IID data and client heterogeneity. We hypothesized that federated learning performance could be enhanced through trajectory optimization theory applied within diffusion-based model aggregation. To validate this hypothesis, we implemented a novel framework incorporating multidimensional trajectory control. Extensive experiments confirmed that this approach improves both personalization accuracy and communication efficiency under non-IID conditions. Empirical evaluations demonstrated that HyFLM outperforms strong baselines across varying federation sizes and data skews. HyFLM achieved higher personalized accuracy, faster convergence, and reduced communication volume, particularly excelling in large-scale or highly skewed settings. These results establish HyFLM as a robust and scalable solution for a federated learning environment. Future directions include scaling HyFLM to diverse modalities and larger federations and investigating theoretical guaranties, with the underlying principles of the framework offering a foundation for future advances in personalized and communication-efficient FL systems.

Author Contributions

H.-j.S. designed and conducted the study, performed the experiments, and wrote the manuscript. Y.-J.S. supervised the research, provided guidance throughout the study, and revised the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

National Research Foundation of Korea (NRF): RS-2022-NR070870; Artificial Intelligence Graduate School Program: RS-2019-II191906; RISE (Regional Innovation System & Education) project: B0080527002599.

Data Availability Statement

The original contributions presented in the study are included in the article. Further inquiries can be directed to the corresponding author.

Acknowledgments

This work was supported by the National Research Foundation of Korea (NRF) under Grant RS-2022-NR070870, the Institute of Information & Communications Technology Planning & Evaluation (IITP) AI Graduate School Program (POSTECH) under Grant RS-2019-II191906, and the Gyeongsangbuk-do RISE project (B0080527002599). The authors would like to thank the members of the AIoT Lab for their helpful discussions and feedback.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analysis, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial Intelligence
DLDeep Learning
DDPMDenoising Diffusion Probabilistic Model
FedAvgFederated Averaging
FLFederated Learning
GPUGraphics Processing Unit
ITOInference Trajectory Optimization
MACMultidimensional Adaptive Coefficient
MLMachine Learning
MTOMultidimensional Trajectory Optimization
IIDIndependent and Identically Distributed
non-IIDnon Independent and Identically Distributed
pFLPersonalized Federated Learning
NFENumber of Function Evaluations
NLPNatural Language Processing
CNNConvolutional Neural Network
SGDStochastic Gradient Descent

References

  1. Lu, Z.; Pan, H.; Dai, Y.; Si, X.; Zhang, Y. Federated Learning with Non-IID Data: A Survey. IEEE Internet Things J. 2024, 11, 19188–19209. [Google Scholar] [CrossRef]
  2. McMahan, H.B.; Moore, E.; Ramage, D.; Hampson, S.; Arcas, B.A.Y. Communication-Efficient Learning of Deep Networks from Decentralized Data. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics (AISTATS), Fort Lauderdale, FL, USA, 20–22 April 2017. [Google Scholar]
  3. Bonawitz, K.; Eichner, H.; Grieskamp, W.; Huba, D.; Ingerman, A.; Ivanov, V.; Kiddon, C.; Konečný, J.; Mazzocchi, S.; McMahan, H.B.; et al. Towards federated learning at scale: System design. arXiv 2019. [Google Scholar] [CrossRef]
  4. Phong, L.T.; Aono, Y.; Hayashi, T.; Wang, L.; Moriai, S. Privacy-preserving deep learning via additively homomorphic encryption. IEEE Trans. Inf. Forensics Secur. 2018, 13, 1333–1345. [Google Scholar] [CrossRef]
  5. Bonawitz, K.; Ivanov, V.; Kreuter, B.; Marcedone, A.; McMahan, H.B.; Patel, S.; Ramage, D.; Segal, A.; Seth, K. Practical secure aggregation for privacy-preserving machine learning. In Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security (CCS ’17), Dallas, TX, USA, 30 October–3 November 2017; Association for Computing Machinery: New York, NY, USA, 2017; pp. 1175–1191. [Google Scholar] [CrossRef]
  6. McMahan, B.; Ramage, D. Federated Learning: Collaborative Machine Learning Without Centralized Training Data. 2017. Available online: https://research.google/blog/federated-learning-collaborative-machine-learning-without-centralized-training-data/ (accessed on 24 October 2024).
  7. McMahan, H.B.; Ramage, D.; Talwar, K.; Zhang, L. Learning differentially private recurrent language models. arXiv 2018. [Google Scholar] [CrossRef]
  8. Li, Z.; Xu, P.; Chang, X.; Yang, L.; Zhang, Y.; Yao, L.; Chen, X. Feature Fusion for Object Detection Knowledge Distillation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  9. Su, S.; Li, B.; Zhang, C.; Yang, M.; Xue, X. Cross-domain Federated Object Detection. In Proceedings of the IEEE International Conference on Multimedia and Expo (ICME), Brisbane, Australia, 10–14 July 2023. [Google Scholar]
  10. Lazaros, K.; Koumadorakis, D.E.; Vrahatis, A.G.; Kotsiantis, S. Federated learning: Navigating the landscape of collaborative intelligence. Electronics 2024, 13, 4744. [Google Scholar] [CrossRef]
  11. Li, T.; Sahu, A.K.; Zaheer, M.; Sanjabi, M.; Talwalkar, A.; Smith, V. Federated Optimization in Heterogeneous Networks. arXiv 2018, arXiv:1812.06127. [Google Scholar]
  12. Lai, J.; Li, J.; Xu, J.; Wu, Y.; Tang, B.; Chen, S.; Li, Y. pFedGPA: Diffusion-based Generative Parameter Aggregation for Personalized Federated Learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Philadelphia, PA, USA, 25 February–5 March 2025; Volume 39, pp. 17999–18007. [Google Scholar]
  13. Arivazhagan, M.G.; Aggarwal, V.; Singh, A.K.; Choudhary, S. Federated learning with personalization layers. arXiv 2019, arXiv:1912.00818. [Google Scholar] [CrossRef]
  14. Husnoo, M.A.; Anwar, A.; Hosseinzadeh, N.; Islam, S.N.; Mahmood, A.N.; Doss, R. Fedrep: Towards horizontal federated load forecasting for retail energy providers. In Proceedings of the IEEE PES 14th Asia-Pacific Power and Energy Engineering Conference (APPEEC), Melbourne, Australia, 20–23 November 2022; pp. 1–6. [Google Scholar]
  15. Oh, J.; Kim, S.; Yun, S.Y. Fedbabu: Towards enhanced representation for federated image classification. arXiv 2021, arXiv:2106.06042. [Google Scholar]
  16. Li, D.; Xie, W.; Wang, Z.; Lu, Y.; Li, Y.; Fang, L. Feddiff: Diffusion model driven federated learning for multi-modal and multi-clients. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 10353–10367. [Google Scholar] [CrossRef]
  17. Mathew, C.; Asha, P. FedProx: FedSplit Algorithm based Federated Learning for Statistical and System Heterogeneity in Medical Data Communication. J. Internet Serv. Inf. Secur. 2024, 14, 353–370. [Google Scholar] [CrossRef]
  18. Karimireddy, S.P.; Kale, S.; Mohri, M.; Reddi, S.; Stich, S.; Suresh, A.T. SCAFFOLD: Stochastic Controlled Averaging for Federated Learning. In Proceedings of the 37th International Conference on Machine Learning (ICML), Online, 13–18 July 2020; Volume 119, pp. 5132–5143. Available online: https://proceedings.mlr.press/v119/karimireddy20a.html (accessed on 20 November 2025).
  19. Tsouvalas, V.; Saeed, A.; Ozcelebi, T.; Meratnia, N. Communication-Efficient Federated Learning through Adaptive Weight Clustering and Server-Side Distillation. arXiv 2024, arXiv:2401.14211. [Google Scholar]
  20. Fallah, A.; Mokhtari, A.; Ozdaglar, A. Personalized Federated Learning: A Meta-Learning Approach. arXiv 2020, arXiv:2002.07948. [Google Scholar] [CrossRef]
  21. Chen, H.Y.; Chao, W.L. FedBE: Making Bayesian Model Ensemble Applicable to Federated Learning. In Proceedings of the International Conference on Learning Representations (ICLR), Virtual, 3–7 May 2021. [Google Scholar]
  22. Shamsian, A.; Navon, A.; Fetaya, E.; Chechik, G. Personalized Federated Learning using Hypernetworks. In Proceedings of the 38th International Conference on Machine Learning (ICML), Virtual, 18–24 July 2021; Volume 139, pp. 9489–9502. Available online: https://proceedings.mlr.press/v139/shamsian21a.html (accessed on 20 November 2025).
  23. Bartholet, M.; Kim, T.; Beuret, A.; Yun, S.; Buhmann, J. Hypernetwork-Driven Model Fusion for Federated Domain Generalization. In Proceedings of the ICML, Vienna, Austria, 21–27 July 2024. [Google Scholar]
  24. Chen, X.; Huang, Y.; Xie, Z.; Pang, J. HyperFedNet: Communication-efficient personalized federated learning via hypernetwork. arXiv 2024, arXiv:2402.18445. [Google Scholar]
  25. Song, J.; Shen, Y.; Hou, C.; Wang, P.; Wang, J.; Tang, K.; Lv, H. FedAGHN: Personalized Federated Learning with Attentive Graph HyperNetworks. arXiv 2025, arXiv:2501.16379. [Google Scholar] [CrossRef]
  26. Liang, W.; Zhao, Y.; She, R.; Li, Y.; Tay, W.P. FedSheafHN: Personalized Federated Learning on Graph-structured Data. arXiv 2024, arXiv:2405.16056. [Google Scholar]
  27. Ho, J.; Jain, A.; Abbeel, P. Denoising diffusion probabilistic models. Adv. Neural Inf. Process. Syst. (NeurIPS) 2020, 33, 6840–6851. [Google Scholar]
  28. Bartosh, G.; Vetrov, D.; Naesseth, C.A. Neural Flow Diffusion Models: Learnable Forward Process for Improved Diffusion Modelling. Adv. Neural Inf. Process. Syst. 2024, 37, 73952–73985. [Google Scholar]
  29. Sasaki, H.; Willcocks, C.G.; Breckon, T.P. Unit-ddpm: Unpaired image translation with denoising diffusion probabilistic models. arXiv 2021, arXiv:2104.05358. [Google Scholar] [CrossRef]
  30. Yoon, T.; Shin, S.; Hwang, S.J.; Yang, E. FedMix: Approximation of mixup under mean augmented federated learning. In Proceedings of the International Conference on Learning Representations (ICLR), Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
  31. Peng, Z.; Wang, X.; Chen, S.; Rao, H.; Shen, C.; Jiang, J. Federated Learning for Diffusion Models. IEEE Trans. Cogn. Commun. Netw. 2025; Early Access. [Google Scholar] [CrossRef]
  32. Huangsuwan, K.; Liu, T.; See, S.; Ng, A.B.; Vateekul, P. FedDrip: Federated Learning With Diffusion-Generated Synthetic Image. IEEE Access 2025, 13, 10111–10125. [Google Scholar] [CrossRef]
  33. Lee, D.; Park, J.; Kim, J.; Lee, K. Multidimensional Trajectory Optimization for Flow and Diffusion. In Proceedings of the International Conference on Learning Representations (ICLR), Singapore, 24–28 April 2025. [Google Scholar]
  34. Wang, C.; Sun, W. Controllable reference guided diffusion with local global fusion for real world remote sensing image super resolution. arXiv 2025, arXiv:2506.23801. [Google Scholar] [CrossRef]
Figure 1. FL scenario in a non-IID environment. Local models are trained on heterogeneous data may degrade overall performance due to accuracy imbalance and representation distortion across clients.
Figure 1. FL scenario in a non-IID environment. Local models are trained on heterogeneous data may degrade overall performance due to accuracy imbalance and representation distortion across clients.
Electronics 14 04704 g001
Figure 2. In the forward process (blue), a clean sample x 0 is gradually noised via q ( x t x t 1 ) until x T N ( 0 , I ) . In the reverse process (green), a neural network learns p θ ( x t 1 x t ) by predicting the noise or score, so sampling starts from p ( x T ) N ( 0 , I ) and iteratively denoise back to x0 [34].
Figure 2. In the forward process (blue), a clean sample x 0 is gradually noised via q ( x t x t 1 ) until x T N ( 0 , I ) . In the reverse process (green), a neural network learns p θ ( x t 1 x t ) by predicting the noise or score, so sampling starts from p ( x T ) N ( 0 , I ) and iteratively denoise back to x0 [34].
Electronics 14 04704 g002
Figure 3. Inference Trajectory Optimization Using MAC (a) Without MAC, the trajectory follows a fixed direction with uniform step sizes, which can yield inaccurate endpoints (red dot). (b) MAC introduces time-varying coefficients that adapt step size and direction at each reverse step, producing stable, personalized endpoints X ^ 0 , θ , ϕ under heterogeneous clients. (c) Example Inference Trajectory Optimization path with MAC illustrating early exploration and late stabilization; colored segments denote adaptive update magnitude and direction.
Figure 3. Inference Trajectory Optimization Using MAC (a) Without MAC, the trajectory follows a fixed direction with uniform step sizes, which can yield inaccurate endpoints (red dot). (b) MAC introduces time-varying coefficients that adapt step size and direction at each reverse step, producing stable, personalized endpoints X ^ 0 , θ , ϕ under heterogeneous clients. (c) Example Inference Trajectory Optimization path with MAC illustrating early exploration and late stabilization; colored segments denote adaptive update magnitude and direction.
Electronics 14 04704 g003
Figure 4. Overview of HyFLM (a Hypernetwork-based Personalized Federated Learning Framework with Multidimensional Coefficients) on diffusion paths.
Figure 4. Overview of HyFLM (a Hypernetwork-based Personalized Federated Learning Framework with Multidimensional Coefficients) on diffusion paths.
Electronics 14 04704 g004
Figure 5. Experimental comparison of accuracy and convergence between HyFLM and the diffusion-based baseline FedGPA (pFedGPA) on EMNIST, Fashion-MNIST, and CIFAR-10 under different numbers of participating clients (10, 20, and 100).
Figure 5. Experimental comparison of accuracy and convergence between HyFLM and the diffusion-based baseline FedGPA (pFedGPA) on EMNIST, Fashion-MNIST, and CIFAR-10 under different numbers of participating clients (10, 20, and 100).
Electronics 14 04704 g005
Table 1. Summary of dataset properties for MNIST, EMNIST, and CIFAR-10.
Table 1. Summary of dataset properties for MNIST, EMNIST, and CIFAR-10.
DatasetImage SizeChannels# ClassesClients
MNIST 28 × 28 1 (grayscale)1010/20/100
EMNIST 28 × 28 1 (grayscale)2610/20/100
CIFAR-10 32 × 32 3 (RGB)1010/20/100
Table 2. Accuracy comparison of HyFLM and baseline methods on EMNIST, Fashion-MNIST, and CIFAR-10 with varying numbers of clients.
Table 2. Accuracy comparison of HyFLM and baseline methods on EMNIST, Fashion-MNIST, and CIFAR-10 with varying numbers of clients.
MethodEMNISTFashion-MNISTCIFAR-10
# Clients102010010201001020100
Local-only71.7571.2674.8486.2485.2486.1465.4565.2166.12
FedAvg72.4373.4475.2582.2384.9684.1565.2768.4770.55
FedPer75.2477.4578.6588.2489.4188.1564.5464.8870.21
LG-FedAvg73.4271.2176.5486.1286.4584.4565.1565.6566.24
pFedHN77.3478.1277.1586.8086.9086.2571.6675.1277.10
pFedGPA78.4080.0082.9085.6087.0089.6070.1071.9074.20
HyFLM84.5286.0088.5089.0089.8091.1074.6776.0878.10
Table 3. Accuracy and communication result in the experiment environment.
Table 3. Accuracy and communication result in the experiment environment.
Dataset#ClientsMethodAcc@Last (%)R@95%Payload/Round (MB)Total MB
EMNIST10pFedGPA78.41452.37236.6
HyFLM83.9282.37165.6
20pFedGPA81.01604.52542.0
HyFLM85.8324.52361.3
100pFedGPA83.424021.723910.2
HyFLM88.14521.722606.8
FMNIST10pFedGPA88.71220.3731.8
HyFLM95.3240.3722.4
20pFedGPA89.41500.7174.9
HyFLM97.9190.7153.5
100pFedGPA91.01843.43514.9
HyFLM98.8143.43377.6
CIFAR-1010pFedGPA72.52603.12436.4
HyFLM74.91233.12374.0
20pFedGPA72.51815.951071.1
HyFLM75.11585.95892.6
100pFedGPA73.823028.626582.5
HyFLM79.719528.625437.7
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Song, H.-j.; Suh, Y.-J. HyFLM: A Hypernetwork-Based Federated Learning with Multidimensional Trajectory Optimization on Diffusion Paths. Electronics 2025, 14, 4704. https://doi.org/10.3390/electronics14234704

AMA Style

Song H-j, Suh Y-J. HyFLM: A Hypernetwork-Based Federated Learning with Multidimensional Trajectory Optimization on Diffusion Paths. Electronics. 2025; 14(23):4704. https://doi.org/10.3390/electronics14234704

Chicago/Turabian Style

Song, Ho-jun, and Young-Joo Suh. 2025. "HyFLM: A Hypernetwork-Based Federated Learning with Multidimensional Trajectory Optimization on Diffusion Paths" Electronics 14, no. 23: 4704. https://doi.org/10.3390/electronics14234704

APA Style

Song, H.-j., & Suh, Y.-J. (2025). HyFLM: A Hypernetwork-Based Federated Learning with Multidimensional Trajectory Optimization on Diffusion Paths. Electronics, 14(23), 4704. https://doi.org/10.3390/electronics14234704

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop