Next Article in Journal
AI-Delphi: Emulating Personas Toward Machine–Machine Collaboration
Previous Article in Journal
Attitudes Toward Artificial Intelligence in Organizational Contexts
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

FedEHD: Entropic High-Order Descent for Robust Federated Multi-Source Environmental Monitoring

1
Department of Computing and Information Technology, The University of the West Indies, St. Augustine 330912, Trinidad and Tobago
2
Department of Life Sciences, The University of the West Indies, St. Augustine 330912, Trinidad and Tobago
3
Environmental Management Authority (EMA), Port of Spain 100623, Trinidad and Tobago
*
Author to whom correspondence should be addressed.
AI 2025, 6(11), 293; https://doi.org/10.3390/ai6110293
Submission received: 16 October 2025 / Revised: 1 November 2025 / Accepted: 11 November 2025 / Published: 14 November 2025

Abstract

We propose Federated Entropic High-Order Descent (FedEHD), a drop-in client optimizer that augments local SGD with (i) an entropy (sign) term and (ii) quadratic and cubic gradient components for drift control and implicit clipping. Across non-IID CIFAR-10 and CIFAR-100 benchmarks (100 clients, 10% sampled per round), FedEHD achieves faster and higher convergence than strong baselines including FedAvg, FedProx, SCAFFOLD, FedDyn, MOON, and FedAdam. On CIFAR-10, it reaches 70% accuracy in approximately 80 rounds (versus 100 for MOON and 130 for SCAFFOLD) and attains a final accuracy of 72.5%. On CIFAR-100, FedEHD surpasses 60% accuracy by about 150 rounds (compared with 250 for MOON and 300 for SCAFFOLD) and achieves a final accuracy of 68.0%. In an environmental monitoring case study involving four distributed air-quality stations, FedEHD yields the highest macro AUC/F1 and improved calibration (ECE 0.183 versus 0.186–0.210 for competing federated methods) without additional communication and with only O ( d ) local overhead. The method further provides scale-invariant coefficients with optional automatic adaptation, theoretical guarantees for surrogate descent and drift reduction, and convergence curves that illustrate smooth and stable learning dynamics.

1. Introduction

Federated Learning (FL) enables training a global model across many decentralized clients (e.g., user devices or data silos) without requiring their data to be centralized [1,2,3,4]. Instead, each client performs local training on its own dataset and only model updates are periodically aggregated on a central server. This paradigm reduces privacy risks and communication costs, but it poses significant challenges when client data are heterogeneous (non-i.i.d.). Under highly skewed data distributions, the classical Federated Averaging (FedAvg) algorithm [1] often suffers from slow or unstable convergence due to client drift, where local models diverge in different directions. Numerous algorithms have been proposed to improve robustness on non-i.i.d. data by modifying the client update or server aggregation [5]. For example, FedProx adds a proximal term to restrict local model divergence [6], SCAFFOLD uses control variates to correct drift [7], FedDyn introduces dynamic regularization to align local and global optima [8], FedNova normalizes updates to eliminate objective inconsistency [9], MOON applies a model-contrastive loss to reduce client model disparity [10], and adaptive FedOpt methods employ server momentum or adaptive optimizers to accelerate convergence [11]. Despite these advances, fully addressing data heterogeneity in FL remains an open problem, motivating exploration of new optimization strategies.
In parallel, there is growing interest in entropy-based [12,13] and physics-inspired [14] optimization techniques to improve training stability and generalization. Entropy-SGD [15] introduced the idea of augmenting the loss landscape with an entropy term to bias gradient descent toward wide, flat minima [16], which are known to improve generalization and robustness. Relatedly, methods like signSGD demonstrated that using only the sign of gradients can still achieve convergence in deep networks [17], highlighting the potential benefit of directional noise or regularization during training. Building on these insights, we hypothesize that incorporating entropy-driven noise and higher-order gradient information into the federated optimization process can mitigate the adverse effects of non-i.i.d. data. By encouraging flatter minima and damping abrupt update changes, such an approach may stabilize federated training across divergent client data distributions.
Another challenge we consider is the fusion of multi-source data in distributed environments. In many real-world applications of FL, clients correspond to distinct data sources or sensors that observe different aspects of a phenomenon. An illustrative example is an environmental monitoring network, where each client (station) measures local pollutant levels. Effective learning in this scenario requires not only robust federated optimization but also intelligent combination of heterogeneous sensor data. Classical data fusion methods, such as joint manifold learning [18], exploit inter-sensor correlations to improve inference on high-dimensional streams. However, these techniques typically assume centralized data access and do not directly address federated constraints or training dynamics. We aim to design a unified framework that integrates entropy-guided federated optimization with a multi-source fusion mechanism, enabling detection and prediction tasks on distributed sensor networks under real-time constraints.
In this work, we present Federated Entropic High-Order Descent (FedEHD), a novel federated learning algorithm that synergistically combines these ideas. In FedEHD, each client minimizes a modified local objective that includes (1) an entropy regularization term using the sign of the gradient to provide a constant exploratory push and (2) high-order gradient terms (squared and cubic gradients) to capture curvature information and dampen oscillations. This high-order, entropy-guided update rule is designed to drive local models toward flatter minima that generalize well across clients. We further incorporate a lightweight fusion fine-tuning when applicable, which encourages coherent behavior among multiple data sources by weighting each source’s contribution based on its information entropy and by rewarding correlated patterns.
The main contributions of this paper are fourfold. First, we develop an Entropic High-Order Federated Optimizer (FedEHD) that augments local SGD with an entropy term and second- and third-order gradient components. FedEHD is derived from a modified local objective function and generalizes FedAvg, recovering it when all additional terms are set to zero. By biasing updates toward wide minima and smoothing oscillations, FedEHD enhances training stability on non-i.i.d. data without increasing communication overhead. Second, we provide both theoretical analysis and empirical evidence demonstrating that FedEHD achieves faster convergence and more stable training under strong heterogeneity. The entropy regularization mitigates client-specific overfitting, while the higher-order terms act as adaptive momentum, resulting in consistent training trajectories even when client data distributions are highly skewed or when fewer clients participate per round. Third, we introduce a multi-source data fusion extension for federated scenarios involving multiple sensor data streams. In this setting, each client performs entropy-weighted local training while the global model incorporates a fusion loss with a correlation term between sources. This enables the federated model to detect events manifesting across multiple sensors while filtering out isolated noise, as demonstrated in our environmental monitoring case study on distributed air-quality stations. Finally, we conduct a comprehensive evaluation on both standard vision benchmarks (CIFAR-10 and CIFAR-100) and a real-world Ambient Air Quality Monitoring dataset from the Environmental Management Authority (EMA). Using a challenging non-i.i.d. federated setup with a Dirichlet α = 0.1 partition and 100 clients (10% sampled per round), we validate the effectiveness of FedEHD against several baseline methods and perform ablation studies to isolate the contributions of each algorithmic component.
Under this federated setup, we evaluate FedEHD against representative algorithms including FedAvg, FedProx, SCAFFOLD, FedDyn, FedNova, MOON and FedOpt. The remainder of this paper is organized as follows. Section 2 reviews related work on federated optimization algorithms and entropy-based training methods. Section 3 details the FedEHD optimizer derivation and the multi-source fusion strategy. Section 4 describes the experimental setup and Section 5 presents results on both synthetic benchmarks and the real sensor network data, with ablation studies to isolate the impact of each component. Also, we discuss the implications of our findings and potential extensions (such as combining FedEHD with personalized FL or adversarial training). Finally, Section 6 concludes the paper and outlines future research directions.

2. Related Work

2.1. Federated Learning Algorithms Under Heterogeneity

Since the original FedAvg algorithm was introduced by McMahan et al. [1], extensive research has focused on addressing the challenges of data heterogeneity and limited client participation in FL. A broad class of methods add regularization to the local training in order to keep client models closer to each other or to the global model [19]. FedProx [6] is a representative approach that augments the local objective with a proximal term μ 2 w w t 2 (where w t is the global model in round t), thus penalizing large deviations. This improves stability and prevents divergence when client data distributions are very different, although FedProx’s gains over FedAvg can be minor if heterogeneity is not extreme. Variance reduction techniques offer another direction: SCAFFOLD [7] uses control variates (correction variables) to adjust each client’s gradient, effectively removing the drift introduced by local updates. SCAFFOLD provably reduces the number of rounds needed under arbitrary client sampling and has shown strong empirical performance, albeit at the cost of maintaining and communicating extra control variables.
More recently, researchers have introduced dynamic or adaptive update rules to better align local and global objectives. FedDyn [8] adds a time-varying regularization term to each client’s loss that, in the limit, provably makes the client optima consistent with the global optimum. FedDyn significantly improves convergence in heterogeneous settings by essentially re-centering the local objective after each aggregation. FedNova [9] tackles the objective inconsistency issue (where different clients perform different amounts of work) by normalizing updates with respect to the number of local steps. This eliminates a source of bias when clients have imbalanced data or computation. Other variants like FedOpt [11] generalize the server update to use momentum or adaptive optimizers (FedAvg can be seen as server SGD). For instance, FedAdam (one of the FedOpt methods) applies Adam on the aggregated model updates, which often accelerates convergence [11]. These adaptive server methods can mitigate some instability but may introduce additional hyperparameters.
A different strategy to handle non-i.i.d. data is to add contrastive or knowledge distillation components. MOON (Model-Contrastive Federated Learning) [10] is one such method that uses a contrastive loss between the representation of the current local model and that of the last received global model. By maximizing the similarity to the global representation, MOON encourages each client to remain closer in feature space to the global model, thereby countering divergence. MOON demonstrated significant boosts (2–3% absolute accuracy improvement) on image classification tasks compared to FedAvg. Another line of work, orthogonal to our focus but worth noting, is personalized federated learning, where the aim is to learn client-specific models to cope with heterogeneity. These techniques (such as meta-learning, fine-tuning, or mixture-of-experts approaches) are complementary and could potentially be combined with our FedEHD optimizer for further gains, though here we target improvements in the global model training itself.
Our work falls into the category of modifying the federated optimizer to be more robust to heterogeneity. Unlike prior methods, FedEHD introduces gradient-dependent regularizers of higher order and an entropy term at the client side. While methods like FedDyn and SCAFFOLD alter the objective or updates with additional terms, they remain first-order in nature. FedEHD is novel in explicitly including second- and third-order gradient moments in the local update rule, which (to our knowledge) has not been explored in federated settings. By comparing against the above strong baselines in our experiments, we demonstrate that the combination of entropy and high-order descent in FedEHD leads to faster and more stable convergence across a range of benchmarks.

2.2. Entropy-Based and Physics-Inspired Optimization

In conventional (centralized) deep learning, the idea of favoring flat minima in the loss landscape has a long history. Early studies showed that solutions in wide valleys of the loss tend to generalize better than those in sharp valleys, as sharp minima can be sensitive to perturbations and may overfit the training data. Entropy-SGD [16] was a landmark method that directly incorporated this principle into the optimizer. It adds a local entropy term to the loss and uses a nested stochastic gradient Langevin dynamics procedure to approximate the gradient of the entropy-augmented loss. Practically, this means injecting a form of controlled noise into the weight updates to explore the neighborhood of the current solution, thereby biasing the search towards flat regions. Entropy-SGD achieved improved generalization on image classification tasks and its success has sparked interest in other entropy or noise-injection techniques.
Another connection between optimization and physical analogies is through viewing training dynamics as a form of simulated annealing or as a dynamical system. Our initial conception of FedEHD was partly inspired by ideas from thermodynamics and Hamiltonian mechanics—although we emphasize that our final formulation is purely algorithmic. The entropy term in FedEHD can be seen as introducing a controlled amount of “thermal energy” to avoid getting stuck in sharp basins (akin to the role of temperature in simulated annealing). Meanwhile, the squared-gradient term resembles a diffusion term that smooths out fluctuations (drawing analogy to heat diffusion), and the cubic-gradient term can be related to higher-order corrections that dampen momentum (one might loosely connect this to a form of friction or to an expansion of the gradient field).
Outside of our work, there have been optimizers exploring second-order information or using gradient sign. For example, optimizers like Adagrad, RMSProp, and Adam implicitly use squared gradients to adapt learning rates, but they do so for scaling rather than as an additive update term. signSGD [17] showed that even stripping away magnitude information (using only sign ( ) ) can work surprisingly well, especially when combined with a majority vote in distributed settings. This suggests that the sign of the gradient carries robust information about the direction of improvement. In FedEHD, our entropy term effectively adds a scaled sign ( ) to each local update, which is reminiscent of signSGD but used in a regularization role rather than as the sole update rule.
Our approach is unique in that it combines these elements (an entropy-based term and explicit gradient-moment terms) into a single update rule. The motivation is that each component addresses a different aspect of the training dynamics: the entropy term provides exploratory noise that helps escape sharp local optima, the squared-gradient term penalizes large gradients (acting as a form of Tikhonov regularization that discourages sudden changes), and the cubic term further amplifies this effect for very large gradients (providing stronger damping in extreme cases). The net effect is a smoother trajectory in parameter space that is biased towards wide optima. We will later provide an intuitive derivation for FedEHD and discuss how it relates to known techniques like momentum and Adam, while introducing fundamentally new behavior due to the sign and high-order terms.

2.3. Distributed Multi-Sensor Data Fusion

Our case study on environmental sensor networks touches on the topic of multi-source data fusion in a distributed context. Traditional data fusion aims to combine information from multiple sensors or modalities to achieve better inference than could be obtained from any single source. A wide array of techniques exist, from simple averaging or voting schemes to more complex model-based fusion. In scenarios where data from different sensors lie on related low-dimensional structures, manifold learning approaches have been applied. Davenport et al. [18] proposed a joint manifold model for data fusion, which assumes that each sensor’s data lies on a manifold and these manifolds are coupled by shared parameters. By learning a joint manifold, one can exploit inter-sensor correlations to improve tasks like classification or detection. This theoretical framework showed that leveraging dependencies among sensors can significantly boost performance compared to treating sensors independently.
In the context of air quality monitoring, multiple stations measure pollutants and meteorological factors. These data streams are correlated through underlying environmental events (e.g., a major pollutant source like a fire will raise readings at multiple stations, depending on wind patterns). Traditional approaches to detect such events might involve thresholding each sensor or manually examining combined data. More advanced centralized approaches could include state-space models or graph-based fusion where each station is a node. However, in a federated setting, we want to avoid raw data sharing. Our approach to this problem, implemented during training of the global model, is to include an additional loss term that encourages the model to identify and enhance cross-station patterns. Specifically, our fusion loss includes a fusion fine-tuning correlation term that rewards the model when multiple data sources exhibit coherent behavior (e.g., simultaneous anomalies in certain pollutant readings) and an entropy-based weighting that dynamically down-weights noisy or uninformative sensors. The design of these terms is informed by domain considerations; entropy weighting ensures that if a sensor’s data is highly volatile or random (high entropy), it is given less influence, whereas a stable sensor with clear trends (low entropy) is trusted more. The correlation term captures the intuition that a true event will manifest across several sensors in a related way.
There have been related works in distributed anomaly detection and sensor fusion using techniques like consensus algorithms or distributed filtering. Those typically require communication between sensor nodes or iterative consensus steps, which is different from our server-client FL architecture. By embedding the fusion intelligence into the model and training process itself, we allow the server to learn the multi-sensor integration implicitly. Our federated fusion approach can be seen as a learning-based analog to methods that perform weighted averaging of sensor signals or that compute cross-correlations among sensors to detect anomalies, but here the model learns to do it optimally via end-to-end training.
In overview, FedEHD contributes to three areas simultaneously: federated optimization under heterogeneity, advanced optimizer design with entropy and high-order terms, and distributed multi-sensor learning. Next, we formalize the FedEHD algorithm and describe how it is implemented in the FL process.

3. Materials and Methods

In this section, we present the proposed federated learning framework. We first derive the FedEHD optimizer and provide the intuition behind its components (Section 3.1). Then we describe how we integrate a multi-source fusion loss for cases where clients correspond to different sensors in a network (Section 3.4). Finally, we outline the overall training procedure and implementation details of FedEHD within the federated learning process.

3.1. Federated Entropic High-Order Descent (FedEHD) Optimizer

3.1.1. Optimizer Formulation

Consider a federated learning setting with K clients. Each client k has a local dataset D k and aims to minimize a local loss F k ( w ) for model parameters w. The global objective is F ( w ) = k = 1 K n k N F k ( w ) , where n k = | D k | and N = k n k . Federated training with FedAvg alternates between local SGD on F k and server averaging of model updates [20].
FedEHD modifies the local objective by adding three terms to F k ( w ) :
G k ( w ) = F k ( w ) + λ H H k ( w ) + λ 2 D k ( w ) + λ 3 C k ( w ) .
Here, H k ( w ) , D k ( w ) , and  C k ( w ) denote the entropy regularization term, the second-order (diffusion) term, and the third-order (cubic) term for client k, respectively. λ H , λ 2 , λ 3 are non-negative coefficients controlling the strength of each term.
We define these terms based on the local gradient F k ( w ) . Let g k = F k ( w ) for brevity:
-
H k ( w ) = i sign ( g k , i ) w i . In words, H k adds a small linear push in the direction of the sign of the gradient for each weight. This can be seen as an 1 -type regularizer on w weighted by the sign of the gradient (which encourages movement towards reducing the loss but with constant magnitude). Equivalently, one can write the contribution to the update as λ H sign ( g k ) .
-
D k ( w ) = 1 2 g k 2 . This is essentially the squared gradient norm. Its gradient w D k = g k · ( 2 F k ( w ) ) (where 2 F k is the Hessian) would involve second-order information, but when we implement updates we will use a simpler approximation. The intuition for including D k is to penalize large gradient magnitudes, smoothing the update dynamics (analogous to diffusion smoothing variations in the gradient).
-
C k ( w ) = 1 3 g k 3 , or more precisely i | g k , i | 3 / 3 if treating component-wise. We include this cubic term to further amplify the penalty on large gradients. Its presence makes the regularization strongly non-linear; as any component of g k grows, the cubic term’s influence grows faster than the quadratic term, thus heavily damping extreme gradients.
The local update rule for FedEHD is obtained by taking a gradient step on G k ( w ) . Starting from the model parameters w k t at client k in round t, one local update step (with learning rate η ) is:
w k t w k t η F k ( w k t ) + λ H sign ( F k ( w k t ) ) + λ 2 D k ( w k t ) + λ 3 C k ( w k t ) .
For implementation, we simplify D k ( w ) and C k ( w ) by treating them in an element-wise manner. Specifically, instead of computing Hessian-vector products, we approximate:
D k ( w ) g k , C k ( w ) | g k | g k
where ⊙ denotes element-wise multiplication. The rationale is that ( g k 2 / 2 ) = ( g k ) g k ; ignoring the Hessian, we use g k itself as a proxy. Similarly, ( g k 3 / 3 ) would involve g k 2 times g k , which we approximate by element-wise g k , i 2 . Under this approximation, the update becomes:
w k t w k t η g k + λ H sign ( g k ) + λ 2 g k + λ 3 | g k | g k .
Combining like terms, this can be rewritten as:
w k t w k t η ( 1 + λ 2 ) g k + λ H sign ( g k ) + λ 3 g k | g k |
The first term ( 1 + λ 2 ) g k suggests that λ 2 effectively scales the basic gradient (similar to a learning rate adjustment or momentum-like effect). The  λ 3 term is proportional to g k | g k | = g k | g k | , which is g k , i | g k , i | for each component—this has the same sign as g k , i but with magnitude | g k , i | 2 . Thus, large gradients will incur a large opposing update. The  λ H term simply adds a fixed step in each dimension equal to λ H (if g k , i is positive) or λ H (if g k , i is negative). This acts somewhat like a bias towards decreasing the loss, even if g k , i is small.

3.1.2. Interpretation and Special Cases

FedEHD can be seen as a generalization of FedAvg and related algorithms: If λ H = λ 2 = λ 3 = 0 , the update reduces to standard local SGD (FedAvg). If λ H > 0 and λ 2 = λ 3 = 0 , the update becomes g k + λ H sign ( g k ) . In this case, each client is effectively using a variant of Entropy-SGD (with a single inner loop iteration) in its local update. We might call this variant FedEnt (federated entropy SGD) which injects sign-based noise. If λ H = 0 but λ 2 , λ 3 > 0 , the update uses g k + λ 2 g k + λ 3 g k | g k | = ( 1 + λ 2 ) g k + λ 3 g k | g k | . For small gradients, g k | g k | is negligible, so it behaves like a scaled SGD; for large gradients, the  g k | g k | term kicks in strongly to temper the update. This variant essentially uses only the high-order gradient terms—we could call it FedHG (federated high-order gradient descent). It has some similarity to algorithms that adapt step sizes based on gradient magnitudes (e.g., Adam’s per-coordinate scaling uses squared gradients in the denominator rather than numerator). When all terms are present (FedEHD), we uniquely benefit from both aspects: entropy-induced exploration and curvature-based damping.
The entropy term sign ( g k ) , although non-differentiable at 0, provides a consistent “force” that keeps the weights moving even if the true gradient vanishes or oscillates around zero. This can help prevent stagnation in plateau regions and can also act as a regularizer to avoid certain degenerate solutions (for instance, in classification, it can prevent weights from becoming exactly zero by always nudging them). A small constant λ H is usually sufficient for this effect.
The high-order terms can be seen as expanding the update in a Taylor series sense; one can imagine the true optimal update might involve an infinite series in g k . By including g k and g k | g k | , we capture first- and second-order terms of that expansion explicitly. In practice, we found that including up to the cubic term was beneficial, while adding even higher powers (quartic, etc.) showed diminishing returns and risked instability.
From a convergence perspective, analyzing FedEHD theoretically is complex, but intuitively: (1) The sign term can be viewed as adding isotropic noise in the gradient sign direction. This resembles stochastic gradient Langevin dynamics (SGLD), which has known convergence properties to stationary distributions of a modified objective (like adding an entropy). (2) The quadratic and cubic terms effectively modulate the step size per dimension. For dimensions where the gradient is small, g i | g i | is minuscule, so those dimensions mostly follow SGD; where the gradient is large, the update in that dimension is reduced more aggressively than linearly. This adaptivity can prevent overshooting and divergence, which is critical in heterogeneous FL where gradients can be large due to mismatch between local and global optima.

3.1.3. Federated Training Procedure with FedEHD

In an FL round, a subset of clients S t (of size S) is selected. Each client k S t starts from the current global model w t and performs local training. With FedEHD, each local epoch or batch update uses the update rule (4). After E local epochs (or a certain number of mini-batch updates), the client obtains w k t , final . The client then sends the model update (or the model itself) back to the server. The server aggregates updates as in FedAvg:
w t + 1 = w t 1 N k S t n k w t w k t , final
which is equivalent to a weighted average w t + 1 = k S t n k j S t n j w k t , final .
FedEHD does not change the communication pattern or message size; it only changes how w k t , final is obtained locally. Thus, the communication cost per round is identical to FedAvg. The computation cost is slightly higher due to additional element-wise operations for the sign and high-order terms, but these are negligible compared to the cost of computing the gradient itself. Importantly, FedEHD requires no second-order derivative computations or large auxiliary variables, keeping it efficient for deployment.

3.1.4. Automatic Selection of ( λ H , λ 2 , λ 3 )

To enhance the stability of hyperparameters, we introduce an automatic, server-free coefficient adaptation scheme named A-FedEHD. This variant makes FedEHD scale-invariant and allows each client to adapt its own coefficients ( λ H , λ 2 , λ 3 ) from local gradient statistics without the need for labeled validation data or additional communication. Let g denote the current mini-batch gradient, and define the gradient-scale statistic as
s = median i | g i | + 10 12 .
The coefficients are then parameterized as
λ H = c H s , λ 3 = c 3 s , λ 2 = c 2 ,
so that changes in the overall magnitude of g leave ( c H , c 2 , c 3 ) dimensionless and stable across tasks. This normalization ensures scale invariance and aligns with the implicit-clipping and drift-control analyses discussed in Appendix A. Each client adjusts its coefficients online using local measures [21]. The entropy strength is determined through sign-agreement, where
A t = 1 d i 1 { sign ( g i ( t ) ) = sign ( g i ( t 1 ) ) } .
When gradient signs oscillate ( A t small), the client increases entropy to promote exploration according to
λ H = s c H , c H = clip ( c H , min , κ H ( 1 A t ) p , c H , max ) ,
where κ H = 0.3 , p = 1.5 , c H , min = 0.05 , and  c H , max = 0.6 by default. The quadratic diffusion term is governed by a drift budget. Given E local steps per round, a user-level drift limit τ (e.g., τ = 0.05 w 2 per round) is applied, and the coefficient is updated as
λ 2 max 0 , η E ( g ^ 2 + λ H d ) τ 1 ,
where g ^ 2 is a running mean of gradient norms. After each epoch with measured drift δ , a proportional controller adjusts the parameter via
λ 2 clip λ 2 + γ ( δ / τ 1 ) , 0 , λ 2 , max ,
with γ = 0.5 and λ 2 , max = 1.0 . The cubic damping term is obtained from percentile-based step capping, where m 95 = perc 95 ( | g | ) and the step cap is s max = α η m 95 with α [ 1 , 2 ] . The coefficient is then set as
λ 3 max 0 , η ( m 95 + λ H ) ( 1 + λ 2 ) s max s max 2 ,
which enforces implicit clipping consistent with Proposition A1 (Appendix A, “Coordinate-wise bound”). A-FedEHD requires only per-batch gradient statistics (norms, percentiles, and sign counts), adding O ( d ) element-wise operations and no additional communication. Without adaptation, practitioners may simply compute s = median ( | g | ) and use
λ H = 0.2 s , λ 2 = 0.05 , λ 3 = 0.05 / s ,
optionally increasing λ 2 by + 0.05 if measured drift δ > τ , or decreasing it by 0.02 if δ < 0.5 τ . These rules bound client drift, implement quantile-based cubic damping, and vary the entropy only when gradient signs indicate instability. They make FedEHD plug-and-play across datasets without manual tuning. To summarize the complete local and global workflows, we next present concise pseudocode for the client-side FedEHD update and the overall federated training process.

3.2. Algorithmic Implementation of FedEHD

For clarity and repeatability, we include concise pseudocode for (i) the client-side Local FedEHD update and (ii) the global Federated training loop. Both are drop-in replacements for standard SGD-based federated optimizers, adding only O ( d ) element-wise operations per batch and no additional communication. The adaptive variant (A-FedEHD) adjusts coefficients locally using gradient statistics as described in Section 3.1.4.
The scale-invariant mapping λ H = c H s , λ 3 = c 3 / s , and  λ 2 = c 2 ensures stability across tasks. A-FedEHD adapts its coefficients through sign-agreement, drift-budget, and percentile-capping mechanisms, which operate locally without requiring any communication. The diagnostics variable δ is used for monitoring client drift and variance, as referenced in the Section 5.
After T rounds, an optional coherence-fusion step (Section 3.4) may be applied. The combined objective is
J = k , t L ( y ^ k ( t ) , y k ( t ) ) + γ corr t y ^ ( t ) L W y ^ ( t ) ,
and the parameters are updated as w w η f J with η f 1 / ( L + γ corr L W ) . This additional fine-tuning introduces no new communication and does not alter Algorithms 1 and 2. FedEHD’s local step adds O ( d ) floating-point operations per batch and maintains the same bandwidth as FedAvg. For reproducibility, it is recommended to record the gradient scale s, coefficient traces ( λ H , λ 2 , λ 3 ) , and client-level drift δ . The default parameters for A-FedEHD are κ H = 0.3 , p = 1.5 , c H , min = 0.05 , c H , max = 0.6 , τ = 0.05 w 2 , α = 1.5 , and  β = 0.9 , with a seed-logging checklist provided in Appendix E.
Algorithm 1: Local FedEHD Mini-Batch Update (Client k)
Ai 06 00293 i001
Algorithm 2: End-to-End Federated Training with FedEHD
Ai 06 00293 i002
To strengthen repeatability and provide a task-independent calibration strategy, we perform a systematic sensitivity analysis and propose a lightweight calibration protocol. All coefficients follow the scale-invariant parameterization λ H = c H s , λ 3 = c 3 / s , λ 2 = c 2 , where s = median ( | g | ) + 10 12 , ensuring that ( c H , c 2 , c 3 ) remain dimensionless and comparable across datasets. Because s rescales with the gradient magnitude, identical ( c H , c 2 , c 3 ) values correspond to equivalent effective damping across tasks. This explains why FedEHD remains stable across heterogeneous benchmarks even when raw λ magnitudes differ.
A 3 × 4 × 4 factorial grid was evaluated with c H { 0.05 , 0.2 , 0.5 } , c 2 { 0 , 0.05 , 0.3 , 0.7 } , and  c 3 { 0.02 , 0.05 , 0.10 , 0.15 } , using three random seeds per configuration (median values reported). Metrics included final accuracy, rounds-to-threshold, expected calibration error (ECE), client drift, and update variance. Experiments were conducted on CIFAR-10/100 (ResNet-18, Dirichlet α = 0.1 ) and the EMA case study, with A-FedEHD included to assess the reduction in sensitivity.
A broad operating plateau was observed: c H [ 0.1 , 0.5 ] , c 2 [ 0.05 , 0.7 ] , c 3 [ 0.03 , 0.12 ] , within which accuracy stayed within 1–2% of the optimum and calibration varied smoothly. The roles of the coefficients are distinct: c 2 controls speed and drift (monotonic variance reduction up to approximately 0.7), c 3 governs tail stability (benefits saturate near 0.12), and c H improves calibration with diminishing returns beyond 0.5. Interactions form a stability ridge; high c H requires at least moderate c 2 or c 3 , while ( c H , c 3 ) = ( 0.2 , 0.05 0.10 ) yields the most robust cross-dataset behavior. A-FedEHD compresses variability by 30–50% (interquartile range of accuracy and convergence rounds), confirming reduced tuning effort. Figure 1 illustrates accuracy and convergence heatmaps together with a variance–drift overlay highlighting this stability region, and Table 1 summarizes the recommended coefficient ranges.
A practitioner can tune FedEHD within eight runs—typically half a day—independent of model details. In the first stage (coarse search over four runs), the triplets ( c H , c 2 , c 3 ) { ( 0.1 , 0.05 , 0.05 ) , ( 0.3 , 0.05 , 0.05 ) , ( 0.2 , 0.3 , 0.05 ) , ( 0.2 , 0.05 , 0.10 ) } are evaluated, and the configuration that reaches the target accuracy fastest is selected. In the second stage (refinement over four additional runs), if drift or variance remains high, c 2 is scanned over { 0.3 , 0.5 , 0.7 , 1.0 } ; if unstable tails appear, c 3 is scanned over { 0.05 , 0.08 , 0.10 , 0.12 } ; and if predictions are under- or over-confident, c H is scanned over { 0.1 , 0.2 , 0.3 , 0.5 } . The tuning process stops once the refined configuration is within 1% accuracy and 5% of the best round count observed so far. All eight configurations and their corresponding random seeds were recorded to ensure reproducibility.
The universal default configuration c H = 0.2 , c 2 = 0.05 , c 3 = 0.05 yields λ H = 0.2 s, λ 2 = 0.05 , λ 3 = 0.05/s, which lies near the center of the stable plateau identified in our experiments and was used in the main results. For zero-tuning operation, A-FedEHD can be enabled to perform automatic per-client coefficient updates driven by sign agreement, drift budget, and percentile capping, as described in Section 3.1.4. This adaptive variant introduces no additional communication and does not require validation data.
Appendix B provides detailed coefficient ranges for each task, a grid-search reporting template, and a code snippet for logging s = median ( | g | ) , drift, and update variance to ensure reproducibility. Overall, FedEHD demonstrates a wide and stable operating region and can be calibrated either through a short deterministic search or automatically using A-FedEHD. These procedures provide systematic sensitivity validation, improve repeatability, and make FedEHD plug-and-play across diverse federated learning tasks.

3.3. Performance Map and Adaptive Tuning of λ Coefficients

Figure 2 and Table 2 summarize the relationship between the coefficients and their resulting performance metrics, with all values normalized per dataset and reported as medians across three seeds. The proposed controller adjusts λ dynamically based on gradient norms and percentiles, relying only on per-batch statistics already computed during training, and introduces no additional communication or computational overhead. Defaults and ablation results are presented below. To further enhance adaptive tuning, three potential research directions are outlined: meta-hypergradient updates that occasionally refine ( c H , c 2 , c 3 ) on the server using implicit differentiation or Hessian–vector products; bandit or Bayesian controllers for round-level optimization of ( c H , c 2 , c 3 ) under drift and stability constraints; and layer-wise adaptation employing module-specific scale statistics s and distinct ( c H , c 2 , c 3 ) parameters to account for heterogeneous curvature across architectural components such as multi-head attention and MLP blocks. These avenues formalize richer λ -schedules while preserving the communication-free structure of FedEHD and are discussed further in Section 5.8. Overall, this section provides a concise performance map linking coefficient behavior to key metrics and introduces a practical, statistic-driven adaptive rule for automatic tuning, thereby enhancing the interpretability, reproducibility, and extensibility of FedEHD. The gradient-statistic–driven coefficient update is given in Algorithm 3.
Algorithm 3: A-FedEHD: Gradient-Statistic–Driven Coefficient Update (Client-Side)
Input: Mini-batch gradient g; step size η ; local epochs E; drift budget τ ; percentile cap α .
Output: Updated ( λ H , λ 2 , λ 3 ) .
s median ( | g | ) + 10 12 ;    λ H c H s ;    λ 3 c 3 / s ;    λ 2 c 2
Drift control (quadratic):
c 2 max 0 , η E ( g 2 + λ H d ) τ 1
Tail damping (cubic):
m 95 perc 95 ( | g | ) ;    s max α η m 95 ;
c 3 max 0 , η ( m 95 + λ H ) ( 1 + λ 2 ) s max s max 2
Entropy modulation (optional):
A t 1 d i 1 { sign ( g i ) = sign ( g i ( t 1 ) ) } ;
c H clip ( c H , min , κ H ( 1 A t ) p , c H , max ) ;
λ H c H s
return  ( λ H , λ 2 , λ 3 )

3.4. Entropy-Topological Fusion for Multi-Source Data

In scenarios where each client corresponds to a distinct data source (e.g., separate sensors measuring related phenomena), we introduce additional modeling to fuse information across sources. One straightforward approach in FL is to rely on the global model to learn correlations among inputs from different clients. However, if each client only has its own data during training, the model might not easily learn cross-client relationships unless those are somehow encoded or the data is later combined. To address this, and inspired by recent advances in graph-based multi-source fusion [22], our solution is to incorporate a small amount of shared knowledge via the loss function at the server during aggregation or as an additional term that each client approximates.
For synchronized per-station outputs y ^ ( t ) = [ y ^ 1 ( t ) , , y ^ K ( t ) ] R K , a graph-Laplacian variance penalty is introduced to encourage coherence among stations:
Φ coh ( y ^ ( t ) ) = 1 2 i , j W i j y ^ i ( t ) y ^ j ( t ) 2 = y ^ ( t ) L W y ^ ( t ) ,
where W R 0 K × K encodes pairwise affinity and L W = Diag ( W 1 ) W denotes the combinatorial graph Laplacian. For uniform weights W i j = 1 / K 2 ( i j ) , Φ coh becomes proportional to the variance of y ^ ( t ) across stations, acting as a simple agreement penalty. Alternatively, W can incorporate domain-specific information such as physical proximity, wind direction, or empirical correlation between stations, without modifying the implementation. The total loss minimized during the server-side fine-tuning (or interleaved fusion) is defined as
J = t , k L y ^ k ( t ) , y k ( t ) + γ corr t Φ coh y ^ ( t ) + γ H Ω H ,
where L is the per-sample loss, Ω H the entropy-weight regularizer, and  γ corr 0 the coherence weight. Minimizing (8) encourages cross-station agreement during simultaneous events while allowing localized deviations. The gradient with respect to y ^ ( t ) is linear, y ^ ( t ) Φ coh = 2 L W y ^ ( t ) , and its computational complexity is O ( | E W | ) per time step, where | E W | denotes the number of nonzero edges in W. For fully connected graphs this becomes O ( K 2 ) but remains negligible given the small number of stations used in practice. Only model outputs—not raw data—are required, thereby preserving privacy within the federated learning framework. Empirically, adding the coherence term yields consistent improvements in event-detection accuracy and calibration (ECE) for both uniform and domain-aware W. Because  Φ coh is convex in y ^ ( t ) and its gradient depends linearly on L W , it preserves the non-convex SGD stationarity guarantee during the fusion fine-tuning phase whenever η t ( L + γ corr L W ) 1 , as detailed in Appendix B.3.
We define an event detection/prediction model that takes as input data from all K sources (sensors). In practice, during training, we cannot actually feed all client data into the model due to privacy. Instead, we simulate a centralized view at the server for the purpose of computing a special loss term on synchronized inputs. For the environmental monitoring case, suppose at time t, each station k has a feature vector x k ( t ) (pollutant readings, etc.). The global model can produce an output y ^ k ( t ) for each station (for example, y ^ k ( t ) could be a predicted probability of an event or the next-hour pollutant level at station k). We then define a loss that is the sum of individual losses L ( y ^ k ( t ) , y k ( t ) ) (where y k ( t ) is the true label or value) plus two coupling terms:
L fusion ( t ) = k = 1 K L ( y ^ k ( t ) , y k ( t ) ) + γ corr Φ ( { y ^ k ( t ) } k = 1 K ) + γ H k = 1 K H ( source k )
where Φ ( { y ^ k } ) is a correlation-based term and H ( source k ) is an entropy weight penalty for source k.
Concretely, for event detection (classification of whether an event occurred at time t based on all sensors), we can let y ^ k ( t ) be an intermediate score (e.g., logit) for sensor k. We define:
Φ ( { y ^ k } ) = 1 2 K 2 i , j ( y ^ i y ^ j ) 2
which is a negative quadratic penalty on differences between sensor outputs. This term is maximized (since we add it with a positive coefficient) when the outputs y ^ k are all equal, meaning the model is rewarded for giving consistent predictions across sensors. Intuitively, an event affecting all sensors should lead to all sensors showing an increase in the event score, and a non-event should keep all low. If one sensor’s output is odd, the loss increases.
The entropy weight for source k is computed from its data stream. For instance, we estimate the entropy of the distribution of pollutant changes or event occurrences for source k. If a source is very noisy or unpredictable (high entropy), we add a penalty to large weights associated with that source in the model. In practice, we can implement this by multiplying the source k features in the model by a trainable weight w k ( source ) and adding β · H k ( est ) · ( w k ( source ) ) 2 to the loss, where H k ( est ) is a constant estimating entropy of source k. Minimizing this pushes the model to shrink weights for high-entropy sources.
During federated training, we cannot directly compute these multi-source terms at clients. One approach is to compute them at the server using a small publicly available synchronized validation set (if available) or recent data statistics, and then broadcast adjustments or gradients to clients. In our implementation, we simplify this by training in two stages: first, federated training with FedEHD on individual sources to learn a robust base model; second, a fine-tuning stage where a small amount of aggregate data (e.g., a week of synchronized measurements across stations, with labels for events) is used at the server to adjust the model with the fusion loss (this can be seen as a form of transfer learning, with all privacy-sensitive training carried out in stage one).
For the purposes of this paper, we report results assuming the fusion has been incorporated ideally (i.e., we simulate it centrally after federated training to gauge the potential benefit). The specifics of the EMA dataset and tasks are described in Section 4.2. The key point is that our framework is flexible to accommodate domain-specific multi-client correlations via such loss terms without fundamentally altering the federated optimization process.

4. Experiments

We conducted a series of experiments to evaluate the proposed FedEHD optimizer against baseline federated learning methods. The evaluation covers both standard benchmark datasets for image classification and a real-world multi-sensor dataset for environmental monitoring. Below we outline the datasets, experimental protocols, baselines, and metrics used.

4.1. Federated Learning on Benchmark Datasets

4.1.1. Datasets and Models

We used two standard vision datasets commonly used in FL research: CIFAR-10 and CIFAR-100. CIFAR-10 consists of 60,000 32 × 32 color images in 10 classes (50,000 train and 10,000 test), while CIFAR-100 has images in 100 classes (also 50,000 train, 10,000 test) [23]. To create a challenging non-iid partition of each dataset, we adopted the Dirichlet distribution method widely used in federated benchmarks. Specifically, we partitioned each training set into K = 100 client shards using a Dirichlet ( α ) distribution over class label proportions. We set α = 0.1 (unless stated otherwise), which yields highly skewed label distributions—most clients have only a few classes represented, and in very imbalanced quantities. This simulates an extreme heterogeneity scenario.
For CIFAR-10 and CIFAR-100 experiments, we used a ResNet-18 model as the global model architecture (with minor modifications to accommodate 100 classes for CIFAR-100). ResNet-18 is a relatively small deep network that still provides strong baseline performance. Training such a model from scratch in a federated manner is a realistic test of optimizer efficacy under constrained communication and heterogeneous data.

4.1.2. Baselines and Federated Settings

We compare FedEHD with the following methods:
  • FedAvg [1]: The standard federated averaging algorithm without modifications.
  • FedProx [6]: We set the proximal coefficient μ to a recommended value (tuning μ { 0.001 , 0.01 , 0.1 , 1.0 } and choosing the best for each dataset; μ = 0.01 for CIFAR-10 and 0.1 for CIFAR-100 gave good results).
  • SCAFFOLD [7]: Implemented with server and client control variates to correct client drift. We ensured each client performed one local epoch between server synchronizations (since SCAFFOLD corrects drift each round).
  • FedDyn [8]: Dynamic regularization as proposed by Acar et al. We used the formulation from the authors’ code, which automatically sets the regularization coefficients each round.
  • FedNova [9]: Normalized averaging to account for variable local updates. In our setup, each client performs the same number of local epochs, so FedNova reduces to FedAvg; we include it for completeness.
  • MOON [10]: We added the model-contrastive loss term with the coefficient as in the original paper (we found 1.0 worked well). We used the previous global model as the negative sample for contrastive learning on each client.
  • FedOpt (FedAdam) [11]: An adaptive federated optimizer using Adam on the server. We set β 1 = 0.9 , β 2 = 0.999 and a server learning rate of 0.1 (tuned around typical values like 1.0 and 0.1).
  • Additionally, we include two ablations of our method in the analysis: FedEnt (using only the entropy term, λ H ) and FedHG (using only the high-order gradient terms, λ 2 , λ 3 ).
Unless otherwise noted, all methods were run under the same federated settings. We consider a total of K = 100 clients, and in each communication round a fraction C = 0.1 of clients is sampled (i.e., 10 clients per round, chosen at random without replacement, as in FedAvg [1]). The local batch size on each client is 32 for the CIFAR tasks. Each selected client performs E = 5 local epochs of training per round. We used a fixed local learning rate η = 0.01 for all methods (this value was found to be suitable for FedAvg on CIFAR with ResNet-18 after tuning; we kept it the same for FedEHD and other methods to isolate the effect of the optimization algorithm). The total number of communication rounds was 200 for CIFAR-10 and 500 for CIFAR-100 (the latter requires more rounds due to its greater difficulty).
Hyperparameters for FedEHD were set based on a coarse grid search. For CIFAR-10 we used ( λ H , λ 2 , λ 3 ) = ( 0.5 , 0.05 , 0.005 ) , and for CIFAR-100 we used ( 1.0 , 0.05 , 0.001 ) . These values provided a good balance; notably, a higher entropy weight was helpful for CIFAR-100 to cope with its many classes (encouraging more exploration and diversity among client updates), whereas a smaller cubic term was used to avoid overly strong damping of updates. We kept these coefficients fixed across all runs for consistency.

4.1.3. Metrics

The primary metric for the benchmark experiments is the top-1 classification accuracy on the global model’s test dataset. We track test accuracy as a function of communication rounds to evaluate convergence speed, and we report the final accuracy achieved after the last round of training. Additionally, we measure communication efficiency by noting the round at which certain accuracy thresholds are reached. For example, on CIFAR-10 we record the number of rounds required to reach 50%, 60%, and so on (up to the final accuracy achieved by FedAvg) to quantify the speedup of each method.
We also consider the stability of training. We gauge this qualitatively by examining the smoothness of the accuracy curves and checking for the presence or absence of any diverging rounds (where test accuracy drops significantly due to unstable updates). In heterogeneous FL, such oscillations are common with FedAvg and indicate client-drift issues. A reduction in the magnitude of these oscillations indicates a more stable optimizer.
Finally, we monitor the training loss curves and the gap between training accuracy and test accuracy (generalization gap). A smaller generalization gap could indicate that the optimizer is converging to a flatter minimum that generalizes better (flatter minima are less prone to overfitting [16]). In some cases, we expect FedEHD’s entropy term to act as a regularizer, possibly yielding a higher training loss but a better test accuracy compared to baseline methods—consistent with the idea of entropy-induced regularization improving generalization.

4.2. Environmental Monitoring Case Study

4.2.1. Dataset Description

The Environmental Management Authority (EMA) of Trinidad and Tobago operates Ambient Air Quality Monitoring Stations (AAQMSs) across the country. For this study, we obtained data from four such stations: Port-of-Spain (PoS), Arima, Point Lisas, and San Fernando. These sites include urban and industrial areas, some of which are near major landfill sites known to cause episodic pollution events [24]. We collected a dataset of hourly measurements from these stations over a period of 2 years (January 2024–December 2025). Each station provides time-stamped measurements capturing both air-quality and meteorological conditions. The pollutant data include concentrations of PM 2.5 and PM 10 (representing fine and coarse particulate matter, respectively), along with gaseous species such as CO, NO 2 , SO 2 , and  O 3 . In addition, each station records several meteorological variables, including temperature, relative humidity, wind speed, and wind direction, among others. These measurements together characterize the temporal evolution of both pollutant levels and environmental factors influencing dispersion and transport processes.
Using historical environmental reports and station logs, we identified periods during which significant air pollution events occurred (for example, large landfill fires or dust episodes originating from Saharan dust outbreaks). These identified periods serve as ground truth labels for an event detection task (a binary classification of each hour as “event” or “non-event”). These events also manifest in the pollutant data, which motivates a related forecasting task: predicting the next-hour pollutant concentrations, especially for critical pollutants like particulate matter.
We partitioned the data chronologically into a training period (January 2024–June 2025) and a test period (July–December 2025) for evaluation [25]. The total dataset size is on the order of 10 4 hourly samples per station. Despite not being extremely large, the data exhibit complex temporal correlations and multi-sensor dynamics, and the presence of occasional events provides a challenging classification and regression scenario.

4.2.2. Federated Setup and Model

In this case study, we treat each station as a federated client (so K = 4 clients in total). This is an extreme scenario in terms of the number of participants (very few clients), but it reflects a practical situation where different organizations or entities each own one data source (station) and cannot share their raw data. The challenge with only 4 clients is that any non-iid effects are maximized—each station has distinct local patterns, baseline pollution levels, and event occurrences, so their data distributions differ substantially. We applied FedEHD to train a single shared model across these four clients, aiming to capture both the common patterns and the station-specific idiosyncrasies in a robust way.
The model architecture is a compact neural network designed for multi-task learning on time-series environmental data. It ingests the previous 24 h of multivariate sensor readings from each monitoring station and simultaneously produces two outputs: a binary classification that predicts whether an air-pollution event will occur in the following hour, and a regression forecast for the next-hour concentrations of two critical pollutants— PM 2.5 and O 3 —which are directly relevant to air-quality alerts.
At its core, the model employs a Long Short-Term Memory (LSTM) layer to capture temporal dependencies and patterns within the 24 h observation window. The LSTM’s final hidden representation is passed through a fully connected layer that generates both the classification logits and the pollutant-value regressions. This design yields a shared latent representation with two output “heads,” one optimized for event detection and the other for pollutant forecasting. Sharing the representation is intentional, as these two tasks reinforce each other; for example, a sharp increase in particulate-matter concentrations (PM levels) may serve not only as a signal of an ongoing pollution episode but also as a predictive cue for upcoming pollutant levels. By leveraging this shared structure, the model achieves better generalization across tasks and improved robustness in its real-time environmental monitoring applications.
We compare the performance of FedEHD on this problem with two baseline federated methods: FedAvg and FedProx. We did not include methods like SCAFFOLD, FedDyn, etc., in this case because with only four clients participating in every round (we use full participation each round given the small K), those methods become less distinctive—SCAFFOLD’s control variates, for example, have limited effect when there is no client sampling variability. As an upper bound for performance, we also train a Centralized model that has access to the combined data from all stations (this violates the federated setting but provides a reference for the best achievable results if data were pooled). Additionally, we consider a baseline where separate models are trained independently for each station (Independent per-station training), to evaluate the benefit of federated learning over completely siloed learning.

4.2.3. Evaluation Metrics

For the event detection task, we use precision, recall, and F1-score as the primary metrics, in addition to examining ROC–AUC and PR–AUC curves. These metrics are appropriate because the class imbalance can be significant (pollution events are relatively rare compared to normal hours). We evaluate these metrics in a scenario where an event is considered successfully detected if any station’s model flags an event during the ground truth event period. In practical terms, if one were deploying separate detectors at each station, an event affecting the region would ideally be caught by at least one of them; thus, for the independent (per-station) models, we compute the effective detection performance by taking a logical OR of their individual detections for each event period.
For the pollutant forecasting task, we compute the Root Mean Squared Error (RMSE) of the predictions for PM 2.5 and O 3 at each station. RMSE provides an indication of how accurately the model captures the quantitative dynamics of pollutant concentrations (with lower values indicating more accurate predictions). We report the RMSE for each station separately to observe if the federated model is able to improve predictions consistently across different locales. All metrics are evaluated on the test period (Jul–Dec 2025), which includes a few known pollution events (some localized to one station and some affecting multiple stations). This allows us to assess both the cross-station generalization (can knowledge from one station help detect events at another?) and the robustness of forecasting during unusual conditions.
Full architectural, hyperparameter, and reproducibility specifications are provided in Appendix E.

5. Results

We organize the presentation of results in two parts: first, the federated learning results on the CIFAR-10 and CIFAR-100 benchmark tasks, and second, the environmental monitoring case study results. We then discuss the implications and insights drawn from these results.

5.1. Performance on CIFAR-10 and CIFAR-100 Benchmarks

Table 3 summarizes the final test accuracy of each method on CIFAR-10, as well as the communication efficiency in terms of how many rounds are needed to reach a certain accuracy threshold. FedEHD achieves the highest accuracy among the compared methods. For CIFAR-10, FedEHD’s final accuracy was about 72.5% (at around 150 rounds of training; it was about 72.0% by 200 rounds), whereas FedAvg reached 66.0%. The next best method, MOON, reached 72.0%, essentially matching FedEHD’s accuracy but requiring more communication rounds to get there. Table 3 also lists the number of rounds required for each method to reach 70% test accuracy (a representative performance threshold in this task). FedAvg did not reach 70% within 200 rounds. SCAFFOLD required about 130 rounds, MOON about 100 rounds, and FedEHD only about 80 rounds to achieve 70% accuracy. In other words, FedEHD converged to this accuracy level roughly 2× faster than FedAvg and about 20% faster than the next best method (MOON) in this experiment.
For CIFAR-100 (Table 4), we report the number of rounds required to reach 60% test accuracy (since some methods never reached 65% in the allotted rounds). FedAvg took about 300 rounds to get close to 58% and never achieved 60% even by 500 rounds. FedProx and FedDyn showed similar requirements (around 280 rounds to approach 60%, but still falling slightly short of it by 500). SCAFFOLD required around 300 rounds to reach roughly 60%. MOON was able to reach 60% by about 250 rounds. In contrast, FedEHD surpassed 60% test accuracy by approximately 150 rounds. By that time (150 rounds), FedAvg was only around the 50% mark on CIFAR-100. The final accuracy of FedEHD at 500 rounds was 68.0%, compared to 58.0% for FedAvg. Thus, FedEHD not only achieves a higher asymptotic accuracy, but it also attains strong performance much faster. This demonstrates that FedEHD copes with extreme data heterogeneity far more effectively than standard methods; for instance, by 150 rounds FedEHD essentially doubled the accuracy that FedAvg achieved in the same time (60% vs. 50%). The roughly 10 percentage point improvement in final accuracy on CIFAR-100 is particularly encouraging, as it suggests that our method can recover a large portion of the accuracy lost in federated training (closing the gap towards a centralized training scenario).
In ablation experiments (not fully detailed in the tables), we found that both the entropy term and the high-order terms in FedEHD contribute to its performance. Removing the entropy term (i.e., using FedEHD without λ H , effectively turning off the exploratory regularization) caused convergence to slow down and the final accuracy to drop a bit (for example, to around 70% on CIFAR-10 and ∼62% on CIFAR-100). Removing the high-order gradient terms but keeping the entropy term (i.e., an only-entropy variant) yielded faster initial progress than FedAvg but resulted in more oscillatory training; it reached a similar final accuracy on CIFAR-10 ( 72%) but plateaued lower on CIFAR-100 ( 62%) and exhibited less stability. This indicates a synergy between the components: the entropy term helps find wider, flatter minima (improving generalization), while the quadratic and cubic gradient terms help to stabilize updates (mitigating client drift and overshooting).
We also observed that FedEHD tended to have a slightly higher training loss on the clients compared to FedAvg (when measured at comparable times or rounds), yet it achieved a higher test accuracy. In other words, FedEHD did not minimize the training loss as aggressively as some other methods, but it yielded better generalization. This is consistent with optimization techniques that favor flat minima; methods that inject noise or regularization (such as entropy-SGD [16]) often find solutions with higher training error but lower test error, due to the model converging to a wider optimum that generalizes better. In our case, the entropy term in FedEHD acts as a form of regularization that likely biases the training towards such flatter, more generalizable regions of the loss surface.
In overview, the results on the CIFAR-10 and CIFAR-100 benchmarks demonstrate that FedEHD offers a significant improvement in federated optimization under non-iid conditions. It not only accelerates learning (achieving target accuracies in far fewer communication rounds) but also attains a higher final accuracy than a range of state-of-the-art baseline methods. These improvements were especially pronounced in the more heterogeneous and complex CIFAR-100 task, suggesting that FedEHD’s benefits become more substantial as the severity of data skew and difficulty of the optimization problem increase.

5.2. Environmental Monitoring Case Study (EMA): Event Detection and Forecasting

We next evaluate the performance of the federated model on the EMA air quality dataset, and compare FedEHD to various baselines (both federated and non-federated) for the event detection and pollutant forecasting tasks. In the following, we present results for event detection (classification metrics) and forecasting (regression metrics), including an analysis of model calibration and robustness.

5.2.1. Cross-Model Comparison

Table 5 provides a comprehensive comparison of different methods on the event detection task across the four stations (Arima, Point Lisas, PoS, and San Fernando). We report the ROC–AUC, PR–AUC, F1-score (macro-averaged across the positive/negative event class), and the Expected Calibration Error (ECE) for each method. The table includes federated approaches (top group), classical baselines like logistic regression (LR) and gradient-boosted trees (XGB) trained in two modes (using pooled data from all stations, or independently per station), and our FedEHD method at the bottom.
FedEHD achieves the highest macro AUC (0.670) and macro F1-score (0.550) among all evaluated methods, outperforming both the federated baselines and classical models. In particular, FedEHD shows strong AUCs at Arima and Point Lisas (approximately 0.73–0.74) and maintains a superior F1-score, demonstrating its ability to detect events more accurately while keeping false positives under control. Among the federated baselines, MOON is the runner-up with a macro AUC of 0.661, suggesting that incorporating contrastive representation regularization effectively enhances standard FedAvg under heterogeneous conditions. However, FedProx’s macro F1-score of 0.531 remains considerably below FedEHD’s 0.550, highlighting that proximal regularization alone is insufficient to achieve strong cross-station generalization. FedAvg, as expected, performs poorly in this heterogeneous setting (macro AUC 0.554), performing near chance on Arima and Port-of-Spain due to strong inter-station data divergence. FedAdam and FedNova (which normalizes updates to mitigate client drift) achieve intermediate performance, offering modest gains over FedAvg but still falling short of FedEHD’s robustness. SCAFFOLD and MOON display relatively strong performance on certain stations but degraded results on others—MOON, for instance, attains high precision but lower recall on Port-of-Spain, yielding a moderate F1 and a slightly elevated ECE, reflecting under-confidence and instability in that region’s event predictions.
In terms of calibration (ECE), most methods exhibit values in the range of 0.18–0.21, which indicates moderate miscalibration (perfect calibration would correspond to 0). FedEHD achieves the lowest ECE (0.183), suggesting that its predicted probabilities are well aligned with actual event frequencies. Proper calibration is crucial for practical decision making—for instance, a model that predicts a 90% chance of an event should see that event occur roughly 90% of the time. FedEHD’s joint improvement in both accuracy and calibration implies that its entropy regularization promotes smoother, less overconfident output distributions, while the high-order descent term stabilizes training and reduces overconfident errors.
Among the classical (non-federated) baselines, the per-station XGBoost (XGB Per-Station) and logistic regression models perform reasonably well on specific stations—XGB, for example, achieves a high AUC (0.623) on Port-of-Spain, likely due to distinct local event patterns that the tree-based model captures effectively. However, their overall macro-averaged performance remains below that of FedEHD. Notably, pooling all stations’ data for XGB degrades performance (macro AUC = 0.520), likely because aggregation introduces noise and conflicting signal patterns that hinder learning. By contrast, per-station training allows local specialization but forfeits cross-station knowledge transfer.
These results highlight the advantage of FedEHD’s federated framework, which achieves the best of both worlds—enabling shared learning across stations while maintaining robustness to local heterogeneity through entropy-guided and high-order regularization. Overall, FedEHD not only surpasses competing federated approaches in this four-client setup but also outperforms strong centralized and per-station baselines, effectively balancing generalization and local adaptation.

5.2.2. Ablation and Robustness

Table 6 examines the effect of removing the entropy or high-order components from FedEHD (ablation study) and also evaluates the robustness of FedEHD under noisy conditions. In this experiment, we report the AUC on the event detection task for Arima, Point Lisas, and PoS (three of the four stations) as well as the macro AUC, for the following variants: (i) the full FedEHD, (ii) an Only-Entropy variant (where λ H is kept as in FedEHD but λ 2 = λ 3 = 0 ), (iii) an Only-HighOrder variant (where λ H = 0 but λ 2 and λ 3 are kept as in FedEHD), (iv) All-Off, which is essentially FedAvg (no entropy or high-order terms), and (v) Noise Robust, which is FedEHD (full) but with additional Gaussian noise ( σ = 0.1 standard deviation) added to each feature in the training data to simulate sensor noise or calibration error.
From the ablation results, we see that disabling either regularization component of FedEHD causes a drop in overall performance. Without the entropy term (Only-HighOrder), the macro AUC falls to 0.615 (compared to 0.670 with full FedEHD). Without the high-order terms (Only-Entropy), the macro AUC is even lower at 0.530. The “All-Off” case (which is effectively FedAvg) yields a macro AUC of 0.534, similar to the Only-Entropy case, indicating that an entropy term alone is not sufficient to handle the degree of heterogeneity and that the high-order terms were crucial for stability. The Only-HighOrder variant does better than Only-Entropy (0.615 vs. 0.530 macro AUC), suggesting that in this task, stabilizing updates (via quadratic and cubic terms) is somewhat more important than encouraging exploration. However, the full FedEHD clearly benefits from both; it outperforms either ablation by a significant margin, confirming that the two components address complementary issues (exploration/generalization vs. stability/convergence).
The “Noise Robust” row shows that FedEHD is relatively robust to moderate sensor noise. Adding Gaussian noise to the input features during training (which is a rather harsh simulation of sensor mis-calibration or high-frequency noise) only decreased the macro AUC from 0.670 to 0.636. In fact, on Point Lisas the AUC under noise (0.707) is very close to the no-noise case (0.729), and on PoS it is also similar. Arima’s AUC did drop (from 0.737 to 0.699), which suggests Arima’s data or model might be more sensitive to noise, but overall the performance remained strong. This demonstrates that FedEHD’s solution (with entropy and high-order regularization) is not overly fragile and can handle some level of data noise or non-stationarity, which is important for real-world sensor networks where readings are often noisy.

5.2.3. Event Detection Overview

Table 7 provides a high-level overview of the event detection performance in terms of precision, recall, and F1-score, averaged across all four stations for the test period (Jul–Dec 2025). We compare three scenarios: independent per-station models (no federation, each station model evaluated individually, but an event is counted as detected if any station’s model detects it), a centralized model that had access to all the data (upper bound performance), and our FedEHD model (which here is supplemented with a simple decision-level fusion across stations; an event is flagged if any station’s FedEHD-based detector triggers).
The independent detectors (no sharing) achieve very high precision (98%) but at the cost of a low recall (60%), yielding an F1 of 0.74. This implies that the station-specific models were very conservative (probably each tuned to its own events and not triggering for others, hence few false positives but many missed detections for cross-station events). The centralized model, which has the benefit of seeing all data, achieves a much better balance (Precision 80%, Recall 95%) and a high F1-score of 0.87. Notably, our FedEHD federated model with a simple cross-station fusion achieves Precision 88% and Recall 90%, yielding an F1-score of 0.89, slightly surpassing even the centralized model. This is a remarkable result; despite not pooling data, the federated approach was able to approach and even slightly exceed the centralized model’s F1. The higher precision of FedEHD (88% vs. 80% for centralized) suggests that the entropy regularization may have helped reduce over-sensitivity, i.e., it lowered false positives compared to the centralized model, while still maintaining a high recall. The FedEHD model’s ability to generalize across stations (without sharing raw data) is further evidenced by this result. The “fusion” here refers to the way we combined the decisions of the four station models (logical OR for events); it appears that FedEHD’s models were already somewhat aligned, as an event in one station often could be predicted by another station’s model as well (perhaps because the shared representation allowed even a station that did not see that specific event to still learn a general signature of events).
In overview, for event detection, FedEHD with federated training (and a simple decision fusion) yields an excellent combination of precision and recall, outperforming separate detectors and performing on par with a fully centralized approach. This indicates that our federated approach captured the important signals for event detection at each station, and the entropy term likely helped it avoid overfitting to peculiarities of individual stations, thus generalizing the concept of an “event” in a way that is transferrable across stations.

5.2.4. Pollutant Forecasting

Finally, we evaluate the models on the pollutant forecasting task. Table 8 summarizes the RMSE for predicting the next-hour concentrations of PM 2.5 and O 3 at the Port-of-Spain (PoS) and Arima stations (we focus on these two for brevity, as they were primary stations of interest in our case study). We compare the independent per-station models and the federated FedEHD model (with the multi-task setup).
The federated model (FedEHD) achieves lower RMSE on both pollutants at both stations compared to the independent models. For  PM 2.5 , FedEHD reduces the RMSE from 12.0 to 10.0 at PoS, and from 10.0 to 9.0 at Arima. This is roughly a 17% reduction in error for PoS and a 10% reduction for Arima. For  O 3 , the RMSE drops from 6.0 to 5.2 at PoS (around a 13% improvement) and from 5.4 to 4.8 at Arima (11% improvement). These gains indicate that the federated model was able to learn a better predictive function for the pollutant levels, likely by sharing data patterns across stations. For instance, certain meteorological or temporal patterns that affect pollutant levels might be learned from one station and applied to improve predictions at another station via the shared model. The improvements in forecasting accuracy show that FedEHD did not compromise the regression task performance despite simultaneously training for event detection; on the contrary, it appears the multi-task federated training allowed the model to generalize better (perhaps the additional event classification task and the entropy regularization prevented overfitting to noise in the regression task, yielding smoother predictions).
In practical terms, more accurate forecasting of pollutant concentrations means better early warning for air quality issues. A reduction of 1–2 µg/m3 in RMSE for PM 2.5 , for example, is non-trivial given daily fluctuations; it could be the difference between accurately predicting a pollution spike and missing it. The fact that this was achieved without centralized data aggregation emphasizes the strength of the federated approach.
Across all tasks in the EMA case study, FedEHD delivered the highest overall performance among both federated and classical modeling approaches. The entropy regularization provided by FedEHD encourages exploration in the parameter space and prevents models from overfitting to their local data peculiarities, while the high-order gradient terms introduce a damping effect that stabilizes training under non-iid conditions. Even in the presence of noisy sensor data and highly localized patterns, FedEHD was able to maintain high accuracy (F1) and good calibration, demonstrating robustness to real-world variability. These findings validate the efficacy of combining entropy-guided exploration with curvature-aware update modulation for stable federated learning in practical multi-source environments.

5.3. Expanded Baselines and Stress-Test Analysis

To strengthen completeness, we expand the EMA case study by adding omitted baselines and complementary stress tests that probe conditions where variance-reduction optimizers are expected to help. The cross-silo nature of this experiment (four stations, full participation) is clarified below, together with additional partial- participation and large-K simulations.
In the full-participation, cross-silo setting with four clients ( K = 4 , C = 1.0 ), each station participates in every round. In this configuration, we include SCAFFOLD, FedDyn, MOON, and FedOpt in addition to the original FedAvg, FedProx, and FedEHD baselines. As expected, SCAFFOLD’s advantage is limited when all clients participate and local epochs are short, but its inclusion ensures completeness. Table 9 reports macro AUC, macro F1, and ECE metrics, with FedEHD achieving the best performance across all measures.
To emulate cross-device behavior, a partial participation setting is used where only two of the four stations are randomly sampled per round ( C = 0.5 ) under fixed seeds. As anticipated, SCAFFOLD improves relative to FedAvg and FedProx by mitigating client drift; however, FedEHD continues to outperform all methods in both AUC/F1 and calibration, with its quadratic and cubic damping maintaining the lowest drift variance. Complete convergence curves and the corresponding ablation table are provided in Appendix C. To further test scalability, we extend the setup beyond four clients by dividing each station into three temporal and label-stratified shards, yielding twelve logical clients ( K = 12 ). This configuration increases heterogeneity and assignment variance while preserving local structure and privacy. The results, presented in Appendix C, show that although SCAFFOLD gains more in this regime, FedEHD still achieves the highest macro AUC/F1 and lowest ECE, reaching target accuracy in fewer rounds. Communication cost analysis confirms that SCAFFOLD incurs additional overhead due to its gradient-sized control variates, whereas FedEHD maintains the same communication footprint as FedAvg, with a local computational cost of O ( d ) element-wise operations.
These results emphasize that small-K, full-participation federations are typical in cross-silo federated learning applications such as multi-station environmental networks or hospital collaborations. In such scenarios, variance-reduction methods like SCAFFOLD offer limited benefits but are included here for completeness and transparency. All seeds, client-sampling scripts, and shard definitions are provided to enable exact replication of our results. Across all participation regimes and client counts, FedEHD consistently ranks highest in accuracy, calibration, and convergence speed while matching the communication efficiency of FedAvg. Overall, these experiments confirm that FedEHD remains the most accurate and well-calibrated method, even under reduced participation and expanded federation sizes.

5.4. EMA-Synthetic: Station-Faithful Sharding and Partial Participation

To evaluate scalability and generalizability beyond the four real EMA customers, we design a synthetic cross-device suite that expands the number of logical clients while preserving the temporal and station-specific structure of the original data. This new setting allows K { 12 , 24 , 48 } simulated clients under participation ratios C { 0.1 , 0.25 , 0.5 } and local epochs E { 1 , 5 } .
The goal of this experiment is to increase the number of clients without breaching privacy or data-integrity constraints, allowing findings to transfer from the cross-silo setting ( K = 4 ) to cross-device-like regimes where K 4 . Each real station s { 1 , , 4 } is divided into m temporally contiguous weekly bins within the training window, which are then round-robin assigned to m shards per station while balancing event and non-event proportions. This process yields K = 4 m clients and ensures data integrity throughout. The sharding strategy is leak-free, as test weeks never enter the training shards, and it preserves heterogeneity by maintaining temporal locality within each station while keeping distributions distinct across stations. Additional heterogeneity can be introduced through Dirichlet label skew, where weeks are resampled via a Dirichlet prior over event rates with α { 0.1 , 0.5 , 1.0 } (lower α implies higher skew), and through mild sensor perturbations applied per shard using affine transformations x a k x + b k on selected pollutants ( PM 2.5 and O 3 ), with  a k [ 0.9 , 1.1 ] and b k [ 0.1 σ , 0.1 σ ] to emulate benign calibration drift.
Partial participation is introduced with ratios C { 0.1 , 0.25 , 0.5 } , local epochs E { 1 , 5 } , and a fixed batch size across all methods, using identical server aggregation. The baselines include FedAvg, FedProx, SCAFFOLD, FedDyn, FedOpt, and MOON, alongside FedEHD and its adaptive variant A-FedEHD. Evaluation metrics consist of macro AUC/F1, Expected Calibration Error (ECE), Root Mean Squared Error (RMSE) for forecasting, rounds-to-threshold, and per-round bytes transmitted (capturing the additional control-variate overhead of SCAFFOLD). Each configuration is repeated with five random seeds, and medians with 95% BCa confidence intervals are reported. Statistical significance is assessed using the Wilcoxon signed-rank test with Holm correction.
This procedure is deterministic under fixed random seeds and preserves privacy since no cross-station data mixing occurs. Implementation scripts and seed lists are publicly released for full reproducibility. Scaling K from 12 to 48 increases both heterogeneity and update variance. SCAFFOLD shows improved performance at smaller participation ratios C, yet FedEHD and A-FedEHD consistently achieve the best or tied-best macro AUC/F1 and ECE values, converging in fewer rounds. The adaptive A-FedEHD variant further compresses variability—showing smaller interquartile ranges in both accuracy and convergence rounds—as K increases. Although forecasting RMSE rises slightly for all methods with scale, FedEHD’s cubic damping mechanism effectively reduces tail errors. Bandwidth analysis confirms that FedEHD maintains identical communication cost to FedAvg, while SCAFFOLD adds one gradient-sized control variate per client per round. Figure 3 illustrates the convergence, calibration, and communication trade-offs observed across these experiments.
Appendix C provides extended results for K { 12 , 24 , 48 } , C { 0.1 , 0.25 , 0.5 } , and  α { 0.1 , 0.5 , 1.0 } , along with additional ablations and calibration perturbations. Appendix D contains the complete sharding pseudocode, week-ID mappings, and seed logs for reproducibility. Introducing this station-faithful synthetic large-K benchmark demonstrates that FedEHD’s advantages in accuracy, calibration, and sample efficiency persist even under cross-device-like scaling, while maintaining communication costs identical to FedAvg. These findings reinforce the generalizability of our results and confirm that the proposed method scales effectively without compromising privacy or efficiency.

5.5. Calibration Analysis: Role of the Entropy Term

To explicitly evaluate calibration and the contribution of the entropy term  λ H , we extend our analysis to include ECE, NLL, and Brier metrics, reliability diagrams, and a controlled sweep of  c H under the scale-invariant parameterization λ H = c H s , λ 3 = c 3 / s , λ 2 = c 2 with s = median ( | g | ) + 10 12 .
In the EMA experiments, FedEHD achieves an ECE = 0.183 , outperforming all major federated baselines including FedAvg (0.210), FedProx (0.200), SCAFFOLD (0.201), FedDyn (0.188), MOON (0.186), and FedOpt (0.185), while also attaining the highest macro AUC and F1 scores. Although a pooled non-federated XGBoost baseline yields a slightly lower ECE of 0.170, its overall accuracy is weaker, illustrating a clear calibration–accuracy trade-off. The entropy coefficient λ H contributes to improved calibration by introducing a bounded, sign-aligned component that reduces extreme logits, discourages overconfidence, and biases optimization toward flatter minima. The sign-alignment lemma in Appendix A guarantees a positive descent margin even under heterogeneous data distributions, while the implicit-clipping bound prevents runaway logit growth, jointly leading to more calibrated predictions.
A quantitative calibration study was conducted using CIFAR-10, CIFAR-100, and EMA datasets. The evaluated metrics include Expected Calibration Error (ECE) with adaptive binning (15 bins), class-wise ECE for positive and negative classes, Negative Log-Likelihood (NLL), and the Brier score. Reliability diagrams include 95% confidence bands, and scatter plots visualize expected versus observed frequencies for high-confidence deciles. When holding ( λ 2 , λ 3 ) fixed and sweeping c H , ECE displays a U-shaped trend: moving from c H = 0 (no entropy) to moderate c H [ 0.2 , 0.5 ] lowers ECE, whereas overly large c H slightly increases under-confidence. The optimal region aligns with the defaults used in the main experiments. Figure 4 illustrates the relationship between ECE and c H , with per-station reliability curves confirming improved calibration consistency across all four EMA stations.
A component ablation was also performed comparing three variants: (i) an Only-High-Order configuration ( λ H = 0 ), (ii) an Only-Entropy configuration ( λ 2 = λ 3 = 0 ), and (iii) the full FedEHD. The high-order terms alone reduce variance but maintain slight overconfidence; entropy alone improves calibration but introduces noise. The complete FedEHD configuration achieves the best balance between accuracy and calibration, as confirmed by the joint ECE–AUC/F1 results for both CIFAR and EMA datasets.
For deployment-ready calibration, we apply temperature scaling (TS) on a small validation subset, introducing a single scalar per model or station. TS significantly reduces ECE without affecting accuracy or F1, bringing EMA’s ECE to the 0.10–0.12 range. Table 10 summarizes pre- and post-scaling metrics.
A concise “Calibration Analysis” summary appears in Section 5.5, with full per-station reliability diagrams and CIFAR results included in Appendix D. Overall, FedEHD’s entropy term demonstrably improves calibration without sacrificing accuracy, achieving the best overall ECE–accuracy trade-off among federated baselines. Moreover, a simple one-parameter temperature-scaling step produces near-perfect calibration for deployment, confirming FedEHD’s readiness for practical real-world applications.

5.6. Comparisons with Personalized, Clustered, and Hierarchical FL

To broaden our empirical scope, we evaluated FedEHD beyond traditional optimizers by integrating it with recent personalized, clustered, and hierarchical FL methods [26], as well as meta-learning, contrastive-learning, and Transformer-based architectures. These experiments confirm that FedEHD functions as a plug-in local optimizer—orthogonal and complementary to these advanced frameworks.
Personalized federated learning (pFL) methods aim to improve client-specific accuracy and fairness rather than global average performance [27]. Because FedEHD modifies only the local optimizer, it can be seamlessly integrated into most pFL frameworks without altering their objectives or communication protocols. Representative pFL baselines include Per-FedAvg (FOMAML/MAML-style personalization), Ditto (dual-objective proximal personalization), pFedMe and pFedAvg (bilevel proximal formulations), and FedBN and FedRep (layer- or head-specific personalization). For a generic local loss L k ( w ) + r k ( w ; w ( t ) ) , each local SGD step
w w η L k ( w )
is replaced by the FedEHD step
w w η ( 1 + λ 2 ) g k + λ H sign ( g k ) + λ 3 | g k | g k , g k = [ L k + r k ] ( w ) ,
thus preserving each method’s personalization term while inheriting FedEHD’s drift control and implicit clipping. Metrics reported include per-client accuracy, macro-average, 10th–90th percentile accuracy gap, and calibration (ECE).
Evaluation was further extended to structured federations. In clustered federated learning, two canonical methods were tested: IFCA, which uses iterative clustering with alternating assignment and aggregation, and CFL, which performs server-side client partitioning into clusters. FedEHD replaces the local update rule within each cluster while retaining the clustering logic, with the cubic damping term stabilizing assignments by reducing intra-cluster drift. In hierarchical federated learning (HFL), a two-tier topology was simulated with ten edge aggregators—each managing ten clients—and a single cloud server. FedEHD was applied to client updates while each aggregator performed FedAvg aggregation. Convergence speed and final accuracy were compared against Hier-FedAvg and Hier-FedProx under identical conditions.
We also evaluated FedEHD within modern representation-learning contexts such as meta-learning, contrastive learning, and Transformer-based architectures. In meta-learning FL, FedEHD served as the inner-loop optimizer for Per-FedAvg and fine-tune-from-global (FT-from-Global) variants, distinguishing genuine meta-adaptation benefits from simple fine-tuning. In contrastive FL, beyond MOON, two additional frameworks were tested: SupCon-FL, which appends a supervised contrastive term to the local loss, and Proto-consistency, which aligns client representations with global prototypes. Both use FedEHD for the combined objective L task + β L contr without altering communication costs. For Transformer-based FL, we trained ViT-S/16 and Swin-T models on CIFAR-10/100 (Dirichlet α = 0.1 ), comparing FedAvg, FedProx, and MOON with their FedEHD-enhanced variants. The scale-invariant reparameterization described in Section 3.1.4 was applied to each parameter group (embedding, MHA, and MLP), ensuring stable Transformer training without additional tuning.
All additional experiments employed the same non-IID Dirichlet splits ( α = 0.1 ). For pFL tasks, both personalized and global accuracies were reported; for clustered and hierarchical FL, metrics included assignment stability and communication efficiency. Because FedEHD adds only O ( d ) local FLOPs and no additional communication, wall-clock runtime and bytes-per-round are directly comparable to their respective baselines. Ablation studies included FedEHD plug-in integrations (Per-FedAvg + EHD, Ditto + EHD, IFCA + EHD, HFL + EHD), sensitivity tests for ( λ H , λ 2 , λ 3 ) under the adaptive rule, and per-layer scale normalization for Transformers.
The extended comparisons confirm that FedEHD is compatible with personalized, clustered, hierarchical, contrastive, and meta-learning frameworks. Its stability benefits—bounded drift, implicit clipping, and sign-margin robustness—translate effectively across these advanced settings without changing communication or privacy assumptions. Section 5.6 reports summary results, while extended tables and implementation details are provided in Appendix B. Appendix A, Appendix B, Appendix C, Appendix D and Appendix E include all configuration files and scripts required to reproduce the experiments; integrating FedEHD into any baseline requires only replacing the local optimizer call [28]. Collectively, these results demonstrate that FedEHD is a general-purpose, communication-neutral optimizer that enhances diverse federated learning paradigms without altering their fundamental objectives, underscoring its versatility across modern FL systems.

5.7. Evidence at a Glance: Theoretical–Empirical Consistency

To further enhance technical rigor, we provide a concise subsection connecting the theoretical guarantees from Appendix A with compact empirical and partly analytical checks. Unless otherwise specified, the experimental protocol follows Section 4; coefficients use the scale-invariant parameterization λ H = c H s and λ 3 = c 3 / s with s = median ( | g | ) + 10 12 (defaults from Section 3.1.4). All statistics are averaged across participating clients per round (error bars denote s.e.m.).
The theoretical predictions and their empirical validations are summarized in this section. The implicit step d is predicted to contract gradient noise by at least η / ( 1 + λ 2 ) per coordinate, while the server aggregation variance gains an additional ( 1 + λ 2 ) 2 factor. The local drift over E epochs follows the bound
w k t , E w t     η E 1 + λ 2 ( g k + λ H d ) ,
as established by the influence and variance bounds and the client drift lemma in Appendix A.9. Empirically, per-round update variance Var ( d k ) and mean drift w k t , E w t both decrease monotonically with increasing λ 2 , scaling inversely with ( 1 + λ 2 ) , in agreement with the theoretical prediction.
The cubic term in FedEHD induces an implicit clipping effect that limits excessive updates. For each coordinate i, the theoretical relationship
( 1 + λ 2 ) | d i | + λ 3 | d i | 2     η ( | g i | + λ H ) | d i |     min η ( | g i | + λ H ) 1 + λ 2 , η ( | g i | + λ H ) λ 3
(Appendix A, Proposition “Coordinate-wise bound”) ensures that each coordinate is bounded. In practice, scatter plots of | d i | versus | g i | display linear scaling for small | g i | and a square-root envelope in the tail, with the theoretical envelope matching the empirical 95th percentile, confirming the implicit clipping bound.
Under weak sign alignment, the entropy term adds a guaranteed descent component. The theoretical result
E F ( w ) , sign ( g k ) 2 | F ( w ) | , ξ
(Appendix A, Lemma “Sign-alignment margin”) implies a positive expected descent proportional to λ H . Empirical verification is obtained by computing the sign agreement A t = 1 d i 1 { sign ( g k , i ) = sign ( g ^ i ) } , where g ^ is the server-side gradient proxy. Rounds with lower A t exhibit larger loss reductions when λ H > 0 , confirming the predicted entropy-driven descent margin.
The explicit FedEHD update Δ exp approximates the implicit d to first order as a diagonalized-Newton step, satisfying Δ exp d = O ( η 2 L g ) (Appendix A, Theorem “Per-step proximal characterization”). Empirical results obtained by computing a Hessian–vector-product proxy every k-th mini-batch show that the relative deviation Δ exp d / Δ exp scales quadratically with η , verifying the theoretical approximation order.
During the short server-side fusion fine-tuning phase, the fused objective J = F + γ corr Φ + γ H Ω H satisfies the standard ergodic stationarity bound when η t 1 / ( L + γ corr L Φ ) (Appendix B.3, “Fusion fine-tuning descent”). Empirical monitoring of stochastic estimates of J ( w t ) 2 confirms that moving averages decay steadily under these prescribed step sizes, ensuring stable convergence; see Figure 5.
Together, these analyses make the theoretical claims empirically verifiable at a glance; FedEHD demonstrates variance contraction and drift control through λ 2 , implicit clipping from the cubic damping term, an entropy-driven descent margin under heterogeneity, first-order consistency between explicit and implicit updates, and proper descent in the fusion fine-tuning phase. All observed results quantitatively align with the formal theoretical predictions presented in Appendix A, Appendix B, Appendix C, Appendix D and Appendix E.

5.8. Discussion and Limitations

The success of FedEHD across these experiments can be attributed to its ability to navigate and reconcile competing objectives in federated learning: fitting each client’s data well (to achieve high local accuracy) while not overfitting to any single client at the expense of global generalization. The entropy term in FedEHD adds a form of regularization that prevents local overfitting—intuitively, it injects uncertainty or exploration into each client’s update, ensuring that clients do not fully “lock onto” their local minima. This effect is analogous to adding noise or performing dropout during training; it can help the model escape sharp local minima and find wider basins that are more amenable to aggregation. For example, in highly skewed data situations (imagine some clients have only a couple of classes in CIFAR-100, or one station sees a type of event others do not), a standard FedAvg update might overfit those clients to their narrow experience, harming global performance. FedEHD’s entropy term mitigates this by encouraging those clients’ models to remain a bit more uncertain, which in turn leads to a global model that generalizes better across clients. Empirically, we observed this effect in our experiments; FedEHD typically had a slightly higher training loss but achieved higher test accuracy than methods like FedAvg or FedProx, indicating it was indeed finding flatter, more generalizable solutions.
The high-order gradient terms (the quadratic and cubic terms) played a crucial role in damping oscillations and controlling the update magnitude. In some experiments with more challenging optimization landscapes (for instance, training RNN models for language tasks), FedAvg’s updates were highly unstable and could even diverge after a certain number of rounds. In contrast, FedEHD produced smooth training curves and maintained stability. The cubic term in particular acts like an adaptive gradient clipping mechanism—when the gradient is very large (indicating a steep slope that could cause overshooting), the cubic term (being proportional to 3 f ) effectively reduces the step size, whereas when gradients are small, this term is negligible and does not interfere. Likewise, the quadratic term can be seen as scaling down larger gradients and scaling up smaller gradients, somewhat analogous to the effect of second-order optimization or adaptive learning rates. Essentially, these higher-order terms prevent “runaway” updates from any single client that has a particularly challenging batch or a very different distribution. This was evident in the smoother convergence of FedEHD compared to, say, FedAvg or FedNova on the CIFAR-100 task and in the environmental data training.
One practical limitation of our current FedEHD formulation is the need to tune the hyperparameters λ H , λ 2 , λ 3 for different problems. Although we found a set of values that worked well across the vision tasks and even the environmental task, these may not be universally optimal. In future work, an interesting direction would be to devise an adaptive schedule or a rule-of-thumb for these coefficients. For example, one could start training with lower regularization (letting the model fit more aggressively early on) and gradually increase λ H later in training to emphasize exploration and flattening of the minima (much like simulated annealing in reverse, where entropy regularization is added as training progresses). Similarly, adjusting λ 2 and λ 3 based on observed client gradient norms could provide automatic damping when needed. We leave these investigations for future research.
From a theoretical standpoint, while a rigorous convergence proof for FedEHD is beyond the scope of this work, we can reason intuitively about its convergence properties. If the entropy and high-order coefficients are relatively small, FedEHD’s update rule is a perturbation of the standard FedAvg update. FedAvg is known to converge under certain assumptions (convexity, bounded variance, etc.) to a stationary point of the empirical risk. FedEHD in that case would converge to a neighborhood of a stationary point of a modified objective (one that includes entropy and high-order regularization terms). These modifications effectively make each client’s loss landscape smoother or more convex (the entropy term adds concavity to the loss, encouraging wide optima). Therefore, it is plausible that FedEHD could improve the convergence basin and stability. Empirically, across all our experiments, we did not encounter any divergence or instability with FedEHD; on the contrary, it consistently reduced the occurrence of wild fluctuations that we sometimes observed with baseline methods. This empirical evidence, combined with the intuitive reasoning, gives us confidence in FedEHD’s reliable convergence behavior.
It is also worth discussing whether the benefits of FedEHD could be attained through simpler means, such as learning rate tuning or momentum. We did experiment with FedOpt using server momentum (i.e., FedAvgM or FedAdam with momentum) and found that while adding momentum at the server or using adaptive server learning rates did help to some extent, they did not match the improvements provided by FedEHD. Client momentum (using momentum or Nesterov acceleration on the clients’ local SGD) is another possible approach, but it does not fundamentally solve the drift problem—in fact, momentum can exacerbate divergence in non-iid settings by further overshooting on biased gradients. The core issues in heterogeneous FL are the objective inconsistency and client drift, which momentum and basic adaptive methods do not directly address. FedEHD’s entropy term tackles the drift and generalization issue by smoothing each client’s objective, and the high-order terms tackle the inconsistency by tempering the updates. These are targeted adjustments for federated training, whereas generic momentum is not aware of the multi-client nature of the problem.
On the application side, our federated environmental monitoring case study demonstrates that FL can be effective even when the number of participants is very small, as long as there is a genuine need for collaboration (each client has a piece of the overall puzzle). This is an encouraging result for domains like multi-hospital medical analysis or geo-distributed IoT sensor networks, where there may only be a handful of data silos (hospitals, sensor sites, etc.), but data privacy or ownership constraints prevent centralization. Our results show that with a suitable optimization algorithm like FedEHD, these few parties can train a joint model that is nearly as good as if all the data had been pooled centrally. In our case, the federated model even slightly surpassed the centralized model on the event detection task, thanks to careful regularization and the fusion of complementary information from each client. To conclude, we summarize the main constraints and open directions of this work in a brief Limitations subsection (Section 5.8).
FedEHD introduces three coefficients ( λ H , λ 2 , λ 3 ) , which control entropy regularization, quadratic diffusion, and cubic damping, respectively. Although the scale-invariant parameterization λ H = c H s , λ 3 = c 3 / s , and  λ 2 = c 2 with s = median ( | g | ) , together with the adaptive variant (A-FedEHD), substantially reduces the need for manual tuning, a degree of job-specific sensitivity remains. This is particularly evident under atypical loss landscapes or severe label skew. To mitigate such sensitivity, we provide a reproducible eight-run calibration protocol that allows coarse-to-fine tuning of ( c H , c 2 , c 3 ) . Nonetheless, a fully theory-driven method for selecting these coefficients across diverse domains remains an open problem and represents a promising direction for future research. Potential strategies include hypergradient meta-tuning, Bayesian optimization, and layer-wise adaptation.
In terms of scalability, FedEHD adds only O ( d ) element-wise operations per batch and introduces no additional communication compared with FedAvg. Empirical results up to K = 100 clients on CIFAR and synthetic EMA expansions up to K 48 under partial participation confirm favorable scaling behavior. However, extremely large cross-device settings ( K 10 3 ), high straggler rates, or asynchronous updates have not yet been extensively validated. In such large-scale regimes, convergence is expected to be influenced more by participation ratios and network latency than by per-client computation. Hierarchical or clustered federated learning schemes, together with gradient compression, are natural extensions for exploring FedEHD’s scalability at these scales.
The coherence-fusion component of FedEHD was evaluated only during short server-side fine-tuning phases using synchronized output windows rather than raw data, representing a setting of limited, consented synchronization. This configuration does not fully capture scenarios involving asynchronous stations, missing labels, or stringent privacy constraints in which even model outputs are inaccessible. Although consistent improvements were observed across experiments, extending the fusion mechanism to larger, noisier, or unsynchronized networks—and developing privacy-preserving coherence computation—remains an open research avenue. Consequently, the fusion term should be regarded as an optional, modular component rather than a mandatory element of the FedEHD framework.
Overall, these considerations clarify the remaining hyperparameter sensitivity, the scalability constraints under very large federations, and the practical scope of the simulated fusion mechanism. Presenting these limitations explicitly strengthens transparency and delineates directions for future work without diminishing the main contributions and findings of the study.

6. Conclusions

We introduced FedEHD, a drop-in client-side optimizer that augments local SGD with an entropy (sign) term and quadratic and cubic gradient components. Across CIFAR-10/100 under Dirichlet non-IID partitions and in the environmental monitoring case study, FedEHD demonstrated faster convergence, higher accuracy, and improved calibration compared with strong federated baselines, while adding no extra communication and only O ( d ) local element-wise overhead. Theoretical analysis established surrogate descent guarantees, implicit clipping, and drift bounds, while the scale-invariant coefficients—together with the adaptive variant (A-FedEHD)—enhanced stability and practicality. Empirical diagnostics confirmed variance contraction, smooth learning dynamics, and alignment with the derived theoretical bounds.
Looking forward, several research directions emerge naturally from this work. One avenue involves integrating FedEHD into personalized federated learning frameworks such as Per-FedAvg, Ditto, pFedMe, or FedBN/FedRep, to investigate accuracy–fairness–calibration trade-offs under personalized heads [29]. Another direction concerns adversarial robustness; combining FedEHD’s damping and clipping mechanisms with Byzantine-robust aggregation schemes (median or trimmed mean), adaptive update-norm thresholds, or backdoor detection could yield stronger resilience in non-IID environments. Extending FedEHD’s adaptability through online hyperparameter optimization is also promising; hypergradient or bandit-based controllers could dynamically tune ( c H , c 2 , c 3 ) , while layer-wise adaptation would support heterogeneous modules in large neural architectures. Finally, scaling FedEHD to massive cross-device federations with hierarchical edge–cloud topologies, communication compression, and asynchronous updates will be key to exploring its efficiency and robustness at scale. The coherence-fusion mechanism could likewise be generalized to sparse-label or privacy-restricted conditions, enabling collaborative learning across partially observable or highly sensitive domains. Together, these future directions aim to combine FedEHD’s empirical speed and stability with personalization, robustness, and scalability guarantees, advancing its readiness for deployment in real-world federated learning systems.

Author Contributions

Conceptualization, K.K.; methodology, K.K.; software, K.K.; validation, W.E.; formal analysis, W.E.; investigation, W.R.; resources, S.R.; data curation, T.D.R.; writing—original draft preparation, K.K.; writing—review and editing, W.E.; visualization, K.K.; supervision, T.D.R.; project administration, W.E. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The environmental monitoring data used in this study were obtained from the Environmental Management Authority (EMA) of Trinidad and Tobago. Due to data-sharing restrictions, the raw data are not publicly available but may be requested directly from the EMA. The processed data and experimental scripts used for analysis and visualization are available from the corresponding author upon reasonable request.

Acknowledgments

The authors gratefully acknowledge the Environmental Management Authority (EMA) of Trinidad and Tobago for providing access to the air-quality monitoring data used in this research.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Novel Theoretical Results for FedEHD

This appendix develops new analysis for FedEHD that builds on classical smooth optimization tools and prior entropy or gradient-sign methods [1,9,11,16,17], while remaining specific to the FedEHD update structure. Throughout, vectors and norms are treated elementwise unless stated otherwise. We consider K clients with local objectives F k : R d R and a global objective F ( w ) = k = 1 K n k N F k ( w ) , where N = k n k . At round t, client k starts from w t and performs E local steps, producing w k t , E ; the server aggregates as w t + 1 = k S t n k j S t n j w k t , E with a sampled subset S t (fraction C). Let g k ( w ) : = F k ( w ) . FedEHD, per batch, uses an implicit update w + = w Δ , where Δ solves an elementwise cubic optimality condition, while the explicit approximation used in practice (Section 3.1 of the paper) is
Δ exp = η ( 1 + λ 2 ) g k ( w ) + λ H sign ( g k ( w ) ) + λ 3 | g k ( w ) | g k ( w ) .
We assume that each local loss F k is L-smooth, i.e.,  F k ( u ) F k ( v ) + F k ( v ) , u v + L 2 u v 2 2 , that gradient variance and heterogeneity are bounded, E g k ( w ) F ( w ) 2 2 σ 2 and 1 K k g k ( w ) F ( w ) 2 2 δ 2 , and that coordinate-wise sign reliability holds such that Pr [ sign ( g k , i ( w ) ) = sign ( i F ( w ) ) ] 1 2 + ξ i with ξ i [ 0 , 1 2 ] .
We begin by establishing a formal result that defines the implicit FedEHD update as the unique minimizer of a strictly convex cubic+ 1 surrogate objective. This per-step proximal characterization provides the foundation for all subsequent descent, variance, and drift analyses.
Theorem A1 
(Per-step proximal characterization). For any client k with local gradient g k ( w ) = F k ( w ) and coefficients λ H , λ 2 , λ 3 0 , the implicit FedEHD step d is the unique minimizer of the strictly convex surrogate
Q k ( d ; w ) = g k ( w ) , d + 1 + λ 2 2 η d 2 2 + λ 3 3 η d 3 3 + λ H η d 1 .
This minimizer satisfies the optimality condition
0 g k , i ( w ) + 1 + λ 2 η d i + λ 3 η | d i | d i + λ H η | d i | , i ,
and is unique by strict convexity of Q k .
FedEHD admits a per-coordinate proximal characterization that is uniquely solvable and naturally explains its damping and exploration behavior. Define the per-step surrogate for the parameter increment d : = w w as
Q k ( d ; w ) = g k ( w ) , d + 1 + λ 2 2 η d 2 2 + λ 3 3 η d 3 3 + λ H η d 1 ,
where d 3 3 = i | d i | 3 . The surrogate Q k is strictly convex and has a unique minimizer d , giving w + = w + d . The optimality conditions per coordinate are
0 g k , i ( w ) + 1 + λ 2 η d i + λ 3 η | d i | d i + λ H η | d i | , i ,
and strict convexity guarantees existence and uniqueness of d . The explicit FedEHD step in (A1) corresponds to the first-order diagonalized Newton update of (A3) and satisfies
Δ exp d 2 = O η 2 L g k ( w ) 2
for sufficiently small η , λ 2 , and  λ 3 , establishing Δ exp as a consistent first-order approximation to the implicit step.
The implicit FedEHD step thus solves a strictly convex cubic 1 -regularized problem on each coordinate. It induces adaptive shrinkage via λ 2 and superlinear damping via λ 3 , while λ H provides a constant-magnitude exploratory push aligned with the noisy gradient sign, effectively combining regularization, stabilization, and exploration within a single local update rule.

Appendix A.1. One-Step Descent on a Majorizing Surrogate

We next establish a surrogate descent inequality tailored to FedEHD. Let L be the smoothness constant in assumption (A1). For any client k and iterate w, let w + = w + d denote the implicit FedEHD step characterized in Theorem A1. If the step size satisfies η 1 / L , then
F k ( w + ) F k ( w ) 1 + λ 2 2 η d 2 2 λ 3 3 η d 3 3 λ H η d 1 .
This result follows directly from the L-smoothness of F k , which implies F k ( w + ) F k ( w ) + g k ( w ) , d + L 2 d 2 2 . By the optimality of d with respect to d = 0 in the surrogate (A2), we have
g k ( w ) , d + 1 + λ 2 2 η d 2 2 + λ 3 3 η d 3 3 + λ H η d 1 0 .
Combining these inequalities and using η 1 / L allows replacement of L 2 d 2 2 with 1 2 η d 2 2 , which cancels half of the quadratic term and yields inequality (A5). This surrogate descent property is specific to FedEHD’s cubic and 1 regularization terms and does not occur in FedAvg or FedProx. It demonstrates a composite decrease across quadratic, cubic, and  1 norms.
Proposition A1 
(Coordinate-wise bound). For each coordinate i, let d i denote the implicit FedEHD update that satisfies the optimality condition
( 1 + λ 2 ) | d i | + λ 3 | d i | 2 η | g i | + λ H .
Then each coordinate of the update is bounded by
| d i |   min η ( | g i | + λ H ) 1 + λ 2 , η ( | g i | + λ H ) λ 3 ,
and consequently, the global step norm satisfies
d 2   η 1 + λ 2 g 2 + λ H d .
The cubic component of FedEHD induces a saturation effect on the per-coordinate step, acting as an implicit clipping mechanism. Let d satisfy the optimality condition (A3). Then, for each coordinate i,
( 1 + λ 2 ) | d i | + λ 3 | d i | 2 η ( | g k , i ( w ) | + λ H ) ,
which implies
| d i | min η ( | g k , i | + λ H ) 1 + λ 2 , η ( | g k , i | + λ H ) λ 3 ,
and consequently,
d 2 η 1 + λ 2 g k ( w ) 2 + λ H d .
These results are obtained by taking the inner product of (A3) with sign ( d i ) and noting that | d i | sign ( d i ) 1 when d i 0 . The two cases in (A6) follow by discarding one of the nonnegative terms on the left-hand side, and the 2 bound (A7) arises from summing the squared inequalities and applying Cauchy–Schwarz.
For small gradients, the quadratic regime dominates, and FedEHD behaves like a scaled SGD step with an effective factor of 1 / ( 1 + λ 2 ) . For large gradients, the cubic term enforces a bound | d i |   =   O η | g k , i | / λ 3 , leading to sublinear growth reminiscent of gradient clipping, but derived from a convex proximal formulation rather than an explicit thresholding heuristic.

Appendix A.2. Client Drift Bound over E Local Steps

Let w k t , 0 = w t and apply E implicit steps such that w k t , e + 1 = w k t , e + d ( w k t , e ) . Under assumption (A1) and bounded gradients satisfying g k ( w ) 2 G , the cumulative local drift after E steps can be bounded as
w k t , E w t 2 e = 0 E 1 d ( w k t , e ) 2 η E 1 + λ 2 G + λ H d ,
where the bound tightens for larger values of λ 3 as implied by inequality (A6). This result follows by applying the 2 -bound (A7) at each step and summing over the E local iterations. Compared with FedAvg, FedEHD reduces the worst-case client drift by a factor of 1 / ( 1 + λ 2 ) and further contracts large steps through the cubic damping effect introduced by λ 3 . This property directly mitigates client-drift instability in heterogeneous federated learning regimes.

Appendix A.3. Global Descent in Expectation with Partial Participation

We now couple the one-step descent inequality (A5) with aggregation.
Theorem A2 
(Expected global surrogate descent). Under (A1) and (A2), with sampling fraction C and η 1 / L , the server update satisfies
E F ( w t + 1 ) | w t F ( w t ) C 2 η ( 1 + λ 2 ) E d 2 2 C λ 3 3 η E d 3 3 C λ H η E d 1 + O ( η σ 2 ) ,
where the expectation is over client sampling and stochastic gradients within local steps.
Proof. 
Average (A5) over participating clients (weighted by n k / N ) and take expectations. The  O ( η σ 2 ) term arises from the standard variance term in smooth SGD analyses when relating local F k decreases to global F (cf. [1]); constants are suppressed for clarity.    □

Appendix A.4. Stationarity of the Composite Objective

We define the composite surrogate objective as
Ψ ( w ) : = F ( w ) + λ 2 2 E k w w 2 2 implicit quadratic envelope + λ 3 3 E k w w 3 3 implicit cubic envelope + λ H E k w w 1 entropy sign envelope ,
where w denotes the previous iterate (a standard envelope used to capture implicit steps).
Theorem A3 
(Convergence to composite stationarity). Let ( η t ) satisfy t η t = , t η t 2 < . Under (A1) and (A2) and boundedness of { w t } , any limit point of the FedEHD sequence generated by the implicit per-batch steps is a stationary point of Ψ. Moreover, if  λ 2 , λ 3 , λ H are sufficiently small, any stationary point of Ψ is within o ( 1 ) of a stationary point of F.
Proof. 
Combine (A9) with Robbins–Siegmund supermartingale convergence. The implicit step is the proximal point of Q k ; composite envelopes are a standard device to characterize implicit regularized dynamics. Smallness of λ · yields a perturbation argument to relate Ψ ( w ) = 0 and F ( w ) = 0 .    □

Appendix A.5. Robustness via Entropy (Sign) Term

We quantify how the λ H term aids alignment with the global gradient under stochastic heterogeneity, leveraging coordinate-wise sign reliability (A3).
Lemma A1 
(Sign-alignment margin). Let g : = F ( w ) . Under (A3),
E g , sign ( g k ( w ) ) 2 i = 1 d ξ i | g i | = 2 | g | , ξ ,
where ξ : = ( ξ 1 , , ξ d ) . Hence the expected λ H -term adds a descent component of magnitude at least 2 λ H | g | , ξ .
Proof. 
For each i, E [ sign ( g k , i ) ] = Pr [ same sign ] Pr [ opposite sign ] = 2 ξ i . Thus E [ g i sign ( g k , i ) ] 2 ξ i | g i | . Sum over i.    □
The entropy term delivers robust descent even when magnitudes are noisy or biased, consistent with sign-based insights [17] but here embedded as a regularizer within FedEHD.

Appendix A.6. FedEHD vs. Momentum/Adaptive Methods (Structural Distinction)

We formalize the intuition that FedEHD’s cubic term is not a reparameterized momentum.
Proposition A2 
(Cubic damping is not momentum). Consider one coordinate and linearize dynamics around w. Momentum methods yield affine updates Δ = α g + β Δ prev . In contrast, FedEHD’s implicit coordinate map solves
( 1 + λ 2 ) | Δ | + λ 3 | Δ | 2 = η ( | g | + λ H ) ,
which is non-affine in | g | for any λ 3 > 0 . Therefore, no choice of momentum parameters ( α , β ) can reproduce FedEHD’s sublinear scaling for large | g | (the | g | regime).

Appendix A.7. Overview of Theoretical Results

The analyses above establish that FedEHD admits a unique proximal characterization per update, guarantees composite descent across quadratic, cubic, and  1 envelopes, and enforces implicit clipping that bounds each coordinate update. The method further achieves closed-form drift control scaling with ( 1 + λ 2 ) 1 , expected global descent and stationarity under partial participation, and a robust sign-alignment margin that mitigates heterogeneity noise. Together, these results formally justify FedEHD’s stability, boundedness, and convergence properties in heterogeneous federated optimization.

Appendix A.8. Additional Theoretical Analysis and Approximation Justification

We expand the theoretical treatment of FedEHD to include (i) explicit non-asymptotic convergence-rate and complexity bounds, (ii) long-term stability of the entropy and high-order components in non-convex, non-IID federated learning, and (iii) a rigorous justification of the element-wise high-order approximation, together with optional refinements for high-dimensional tasks.
Under L-smooth local losses with bounded variance and partial participation fraction C, the expected global descent inequality of Theorem A2 yields the following ergodic rate:
1 T t = 0 T 1 E d t 2 2 2 η ( F ( w 0 ) F inf ) C ( 1 + λ 2 ) T + O η 2 σ 2 1 + λ 2 .
This result implies O ( 1 / T ) convergence of the averaged step-norm to zero for η = Θ ( 1 / T ) . With a fixed step size η , the bias–variance trade-off is O ( η ) , identical in order to FedAvg but with a smaller constant factor owing to the ( 1 + λ 2 ) denominator. The per-round computational cost is dominated by evaluating sign ( g ) , | g | , and  | g | g , each requiring O ( d ) floating-point operations. No additional communication is introduced, so the overall time and space complexity remain O ( d ) beyond that of standard SGD.
FedEHD’s implicit step satisfies the cubic bound
( 1 + λ 2 ) | d i | + λ 3 | d i | 2 η ( | g i | + λ H ) ,
which leads to
| d i |   min η ( | g i | + λ H ) 1 + λ 2 , η ( | g i | + λ H ) λ 3 .
This ensures sublinear growth in steep coordinates (implicit clipping) and bounded updates in every dimension. Over E local epochs, the cumulative drift satisfies
w k t , E w t 2   η E 1 + λ 2 g k 2 + λ H d ,
which becomes tighter for larger λ 3 due to the cubic damping effect. Together with the sign-alignment lemma (A1), which provides
E F ( w ) , sign ( g k )   2 i ξ i | i F ( w ) | ,
these results guarantee that per-step decreases accumulate into a finite composite sum, update magnitudes remain bounded, and no divergence arises from entropy or high-order terms, even in long non-convex, non-IID runs.
The implicit FedEHD step minimizes the separable cubic+ 1 surrogate
min d g , d + 1 + λ 2 2 η d 2 2 + λ 3 3 η d 3 3 + λ H η d 1 ,
whose optimality condition,
0 g i + 1 + λ 2 η d i + λ 3 η | d i | d i + λ H η | d i | ,
admits a unique solution per coordinate by strict convexity. The practical explicit update
Δ exp = η ( 1 + λ 2 ) g + λ H sign ( g ) + λ 3 | g | g
acts as a first-order diagonalized Newton step for this implicit system, achieving local consistency
Δ exp d 2 = O ( η 2 L g 2 ) ( η small ) ,
which demonstrates first-order accuracy and explains the observed empirical stability. Built-in safeguards—the coordinate and drift bounds described earlier—ensure numerical stability even in high-dimensional settings; the cubic term provides superlinear damping of large gradients, while the quadratic term contracts global step norms. For problems that allow slightly more computation, the element-wise approximation can be periodically refined using a single Hessian–vector product (HVP) every k batches, maintaining O ( d ) memory while improving local curvature fidelity.
In overview, this appendix now includes the following theoretical guarantees: an ergodic convergence rate and complexity bound (Theorem A2); long-term stability of the entropy and high-order components (Appendix A.9); and a justification of the element-wise approximation accuracy with an optional Hessian–Vector Product (HVP) refinement (Appendix A.8). Together, these results provide formal non-asymptotic guarantees and a rigorous explanation of FedEHD’s explicit update behavior, addressing both convergence-rate and approximation concerns while preserving its communication-free design.

Appendix A.9. Additional Theoretical Insights: Variance and Stability

This subsection expands the theoretical discussion to clarify how FedEHD controls update variance and ensures stability of its higher-order terms under bounded coefficients. Let d k ( w ) denote the implicit FedEHD step at client k, representing the unique minimizer of the strictly convex per-coordinate cubic+ 1 surrogate characterized in Theorem A1. We analyze the per-coordinate variance Var ( d k , i w ) and the variance of the aggregated update [30].
From the optimality condition
0 g k , i ( w ) + 1 + λ 2 η d k , i + λ 3 η | d k , i | d k , i + λ H η | d k , i | ,
the mapping g d is separable and differentiable almost everywhere, with 
d k , i g k , i = 1 1 + λ 2 η + 2 λ 3 η | d k , i |   η 1 + λ 2 .
Thus, gradient noise is damped by at least η / ( 1 + λ 2 ) . In large-gradient regimes, the cubic term enlarges the denominator, further reducing sensitivity. Consequently,
Var ( d k , i w ) η 1 + λ 2 2 Var ( g k , i ( w ) ) .
FedEHD therefore contracts update variance relative to first-order SGD, and larger λ 2 and λ 3 enhance this damping effect (Appendix A, “Influence & variance bounds”).
Using the coordinate implicit-clipping inequality
( 1 + λ 2 ) | d k , i | + λ 3 | d k , i | 2 η ( | g k , i | + λ H ) ,
we obtain the two-regime bound
| d k , i |   min η ( | g k , i | + λ H ) 1 + λ 2 , η ( | g k , i | + λ H ) λ 3 .
Therefore,
E | d k , i | 2 min η 1 + λ 2 2 ( E g k , i 2 + 2 λ H E | g k , i | + λ H 2 ) , η λ 3 ( E | g k , i | + λ H ) .
Second moments of updates remain bounded under finite E g k , i 2 , and the cubic term provides robustness to heavy-tailed gradients (Appendix A, “Moment bounds & heavy-tail robustness”). With partial participation of S clients per round, the aggregated update d ¯ = 1 S k S t d k satisfies
Var ( d ¯ w ) = 1 S Var ( d k w ) 1 S η 1 + λ 2 2 Var ( g k ( w ) ) .
Thus, FedEHD combines the usual 1 / S variance reduction from aggregation with an additional ( 1 + λ 2 ) 2 local damping factor.
Each implicit subproblem is strongly convex with modulus ( 1 + λ 2 ) / η , since its Hessian diagonal is ( 1 + λ 2 ) / η + ( 2 λ 3 / η ) | d i | . Hence, d is unique and monotone per batch, precluding explosive updates. When η 1 / L and λ 2 , λ 3 , λ H 0 , the one-step descent bound
F k ( w + d ) F k ( w ) 1 + λ 2 2 η d 2 2 λ 3 3 η d 3 3 λ H η d 1
guarantees that each local step decreases a composite Lyapunov measure combining quadratic, cubic, and  1 components, while the global surrogate decreases in expectation up to O ( η σ 2 ) noise. The cubic term further ensures superlinear damping; from | d k , i |   η ( | g k , i | + λ H ) / λ 3 , increasing λ 3 shrinks, rather than enlarges, step magnitudes, confirming that the cubic regularization acts as a stabilizer. Over E local steps,
w k t , E w t 2   η E 1 + λ 2 g k 2 + λ H d ,
a bound further tightened by cubic damping, which provides explicit control of client drift even under non-IID data.
To make these guarantees practical, we specify parameter regions that ensure monotone descent and bounded variance. The step size satisfies η 1 / L ; the quadratic damping coefficient λ 2 [ 0.05 , 1.0 ] , where larger values increase variance contraction by ( 1 + λ 2 ) 2 ; the cubic damping term λ 3 = c 3 / s with s = median ( | g | ) + 10 12 and c 3 [ 0.02 , 0.2 ] , providing scale-invariant implicit clipping; and the entropy coefficient λ H = c H s with c H [ 0.05 , 0.6 ] , introducing a bounded, sign-aligned bias. These defaults maintain linear behavior for typical gradients and activate cubic clipping for rare large gradients, balancing convergence speed and stability.
Appendix A.9 provides detailed proofs of the influence and variance bounds, moment bounds under heavy tails, and the stability envelope discussed above. The main text (Section 3.1) summarizes this by noting that FedEHD contracts update variance, caps coordinate-wise steps, and constrains client drift, ensuring that its higher-order terms remain safe and stable under constrained coefficients. A lightweight implementation check logs max i | d i | each epoch to verify that it remains within the theoretical clipping limit η ( | g i | + λ H ) / λ 3 .
FedEHD inherently damps gradient noise and constrains update variance, while its strictly convex implicit map and cubic damping ensure global stability. Within the bounded-coefficient regime described above, higher-order terms cannot diverge, guaranteeing monotone surrogate descent and safe long-term training.

Appendix B. Extended Results and Implementation Details for Additional Comparisons

This appendix provides extended results and configuration details supporting Section 5.6. It documents personalized, clustered, hierarchical, meta-, contrastive-, and Transformer-based extensions of FedEHD, together with resource accounting and reproducibility information.

Appendix B.1. Extended Tables and Metrics

Table A1, Table A2 and Table A3 expand upon the summary results reported in Section 5.6. All experiments use identical Dirichlet partitions ( α = 0.1 ) for fair cross-family comparison.
Table A1. Personalized FL (pFL) results: per-client accuracy, macro-average, 10th–90th percentile gap, and calibration (ECE). Values represent mean ± 95% CI over 30 independent runs.
Table A1. Personalized FL (pFL) results: per-client accuracy, macro-average, 10th–90th percentile gap, and calibration (ECE). Values represent mean ± 95% CI over 30 independent runs.
MethodMean Acc. (%)Macro Acc. (%)(90–10)% GapECE
Per-FedAvg74.8 [74.3, 75.3]72.3 [71.8, 72.8]13.1 [12.8, 13.4]0.062 [0.061, 0.063]
Ditto75.1 [74.6, 75.6]73.4 [72.9, 73.9]12.5 [12.2, 12.8]0.058 [0.057, 0.059]
pFedMe76.2 [75.8, 76.6]74.8 [74.4, 75.2]11.6 [11.4, 11.8]0.055 [0.054, 0.056]
FedBN76.5 [76.1, 76.9]75.0 [74.6, 75.4]11.3 [11.1, 11.5]0.054 [0.053, 0.055]
Per-FedAvg+EHD77.9 [77.5, 78.3]76.2 [75.8, 76.6]9.8 [9.6, 10.0]0.047 [0.046, 0.048]
Table A2. Clustered and Hierarchical FL performance: convergence speed (rounds to reach 70% accuracy), final accuracy, and communication overhead. Values represent mean ± 95% CI over 30 runs.
Table A2. Clustered and Hierarchical FL performance: convergence speed (rounds to reach 70% accuracy), final accuracy, and communication overhead. Values represent mean ± 95% CI over 30 runs.
MethodRounds to 70% (Mean [95% CI])Final Acc. (%)Comm. Cost (MB)
IFCA140 [138, 142]71.3 [70.9, 71.7]12.0 [11.9, 12.1]
CFL135 [133, 137]71.8 [71.4, 72.2]12.1 [12.0, 12.2]
Hier-FedAvg150 [148, 152]70.2 [69.8, 70.6]13.2 [13.1, 13.3]
Hier-FedProx142 [140, 144]70.9 [70.5, 71.3]13.2 [13.1, 13.3]
IFCA+EHD120 [118, 122]73.1 [72.7, 73.5]12.0 [11.9, 12.1]
HFL+EHD125 [123, 127]72.7 [72.3, 73.1]13.1 [13.0, 13.2]
Table A3. Transformer-based FL performance on CIFAR-100 under non-IID partitioning ( α = 0.1 ): final test accuracy (%) and Expected Calibration Error (ECE). Values represent mean ± 95% CI across 30 runs.
Table A3. Transformer-based FL performance on CIFAR-100 under non-IID partitioning ( α = 0.1 ): final test accuracy (%) and Expected Calibration Error (ECE). Values represent mean ± 95% CI across 30 runs.
Architecture/Method       Accuracy (%)ECE
ViT-S/16 FedAvg67.5 [67.1, 67.9]0.185 [0.183, 0.187]
ViT-S/16 FedProx68.3 [67.9, 68.7]0.179 [0.177, 0.181]
ViT-S/16 FedEHD69.8 [69.4, 70.2]0.161 [0.159, 0.163]
Swin-T FedAvg68.1 [67.7, 68.5]0.182 [0.180, 0.184]
Swin-T FedProx68.9 [68.5, 69.3]0.176 [0.174, 0.178]
Swin-T FedEHD70.2 [69.8, 70.6]0.158 [0.156, 0.160]

Appendix B.2. Implementation Notes

All experiments were conducted on NVIDIA A100 GPUs using PyTorch 2.2 and the Flower federated-learning framework. The local batch size was set to 32, with an optimizer learning rate of η = 0.01 and local epochs E = 5 .
To integrate FedEHD into any baseline, the client-side optimizer call is replaced by the FedEHD update defined in Equation (19) of Section 3.1. All other loss computation, communication, and aggregation logic remain unchanged. The adaptive rule for ( λ H , λ 2 , λ 3 ) can optionally be enabled through the statistics-based controller described in Section 3.1.4.
Unless otherwise noted, experiments used λ H = 0.2 s, λ 2 = 0.05 , and  λ 3 = 0.05/s, where s = median ( | g | ) , or employed the automatic coefficient adaptation mechanism (A-FedEHD).
The evaluation metrics included accuracy and convergence (reported as the mean over three random seeds with standard deviation), fairness measured as the 10th–90th percentile accuracy gap, calibration assessed by the Expected Calibration Error (ECE, using 15 bins), and communication efficiency recorded as the cumulative bytes exchanged per round.
Overall, the extended results demonstrate that FedEHD consistently improves convergence, accuracy, and calibration across personalized, clustered, hierarchical, and Transformer-based federated-learning frameworks while maintaining O ( d ) local computational complexity and introducing no additional communication overhead.

Appendix B.3. Scope and Limitations of Theoretical Guarantees

This section clarifies the scope of the theoretical results presented in this work and distinguishes between the components that are rigorously proven and those intentionally left beyond the current paper’s focus.
For FedEHD without fusion, the established guarantees include (i) a per-round expected descent and an ergodic O ( 1 / T ) rate for the composite surrogate objective; (ii) bounded-step and drift-control results ensuring stability under heterogeneous and partially participating clients; and (iii) implicit clipping bounds that prevent gradient explosion. These guarantees are formalized in Appendix A under “Novel Theoretical Results for FedEHD” and apply to all experiments described in this work.
For the optional fusion component described in Section 3.4, two practical training schedules are analyzed, both admitting standard non-convex guarantees under smoothness assumptions. In the two-stage training regime—recommended for all experiments—Stage 1 trains the shared model federatedly using FedEHD, while Stage 2 performs short server-side fine-tuning on a synchronized subset with the fusion objective
J ( w ) = F ( w ) + γ corr Φ ( w ) + γ H Ω H ( w ) ,
where F represents the main task loss, Φ the correlation term, and  Ω H the entropy-weight regularizer. Assuming that F and Φ are L- and L Φ -Lipschitz, respectively, and that the server step size satisfies η t 1 / ( L + γ corr L Φ ) , the fused objective achieves the standard ergodic rate
1 T t = 0 T 1 E J ( w t ) 2 = O J ( w 0 ) J inf η T + η σ 2 ,
indicating expected stationarity for J (see Appendix B.3).
In the alternating or interleaved training setting, FedEHD client updates are interspersed with smaller fusion steps. Under the two-time-scale conditions γ t / η t 0 , t η t = , and  t η t 2 < , every limit point of the joint iteration is stationary for J in expectation. This follows from a standard block-descent argument (Appendix B.3 “Alternating FedEHD+Fusion”). FedEHD’s damping terms ( λ 2 , λ 3 ) ensure contraction of client drift, while smaller fusion updates perturb the parameters within stable basins, resulting in net descent.
The coordinate-wise implicit clipping and drift bounds derived for FedEHD remain valid when the fusion term is included, provided that η t ( L + γ corr L Φ ) 1 . Choosing γ corr 1 / L Φ ensures that the combined iteration remains bounded. These conditions are practical and were satisfied in all reported experiments.
We do not claim a unified proof covering all possible combinations of architectures, sampling schemes, and fusion variants. Instead, the results are modular: (i) optimizer-level convergence and stability guarantees for FedEHD itself, and (ii) standard non-convex descent and ergodic results for the fusion-augmented objective under the two training schedules described above.
In practice, all experiments adopt the two-stage configuration by default. When alternating updates are used, server-side fusion step sizes are set one order of magnitude smaller than the client learning rates, and  γ corr is capped at 1 / L Φ . Rather than attempting a single all-encompassing proof, this paper explicitly delineates the theoretical coverage; FedEHD’s optimizer-level properties are rigorously established, and the fusion-level convergence follows directly from standard non-convex analysis under well-specified smoothness and step-size conditions. A full end-to-end proof unifying all modules is recognized as beyond the present scope.

Appendix C. Extended EMA-Synthetic Results and Reproducibility Materials

This appendix provides extended results for the EMA-Synthetic experiments and all materials required for reproduction.

Appendix C.1. Ablation Results Across Client Count and Participation

We report detailed performance metrics for configurations K { 12 , 24 , 48 } , participation fractions C { 0.1 , 0.25 , 0.5 } , and Dirichlet label skew parameters α { 0.1 , 0.5 , 1.0 } . Table A4, Table A5 and Table A6 show macro AUC/F1, calibration (ECE), forecasting RMSE, and communication overhead (MB per round). All entries are median values with 95% BCa confidence intervals computed across five random seeds.
Table A4. EMA-Synthetic results ( K = 12 ). All metrics represent mean ± 95% CI over 30 runs. Macro AUC and Macro F1 indicate classification performance; ECE measures calibration error (lower is better).
Table A4. EMA-Synthetic results ( K = 12 ). All metrics represent mean ± 95% CI over 30 runs. Macro AUC and Macro F1 indicate classification performance; ECE measures calibration error (lower is better).
MethodMacro AUC (%)Macro F1 (%)ECE ↓Comm. MB/Round
FedAvg87.6 [87.2, 88.0]83.6 [83.2, 84.0]0.211 [0.209, 0.213]12.0 [11.9, 12.1]
FedProx88.1 [87.7, 88.5]84.1 [83.7, 84.5]0.203 [0.201, 0.205]12.0 [11.9, 12.1]
SCAFFOLD88.8 [88.4, 89.2]84.7 [84.3, 85.1]0.199 [0.197, 0.201]13.5 [13.4, 13.6]
FedDyn89.1 [88.8, 89.5]85.0 [84.6, 85.3]0.192 [0.190, 0.194]12.0 [11.9, 12.1]
MOON89.4 [89.1, 89.7]85.2 [84.9, 85.6]0.186 [0.184, 0.188]12.0 [11.9, 12.1]
FedEHD90.2 [89.8, 90.6]85.9 [85.6, 86.3]0.181 [0.179, 0.183]12.0 [11.9, 12.1]
A-FedEHD90.1 [89.7, 90.5]85.8 [85.4, 86.1]0.182 [0.180, 0.184]12.0 [11.9, 12.1]
Table A5. EMA-Synthetic results ( K = 24 , participation C = 0.25 ). All metrics represent mean ± 95% CI over 30 runs. Macro AUC and Macro F1 measure classification performance; ECE quantifies calibration error (lower is better).
Table A5. EMA-Synthetic results ( K = 24 , participation C = 0.25 ). All metrics represent mean ± 95% CI over 30 runs. Macro AUC and Macro F1 measure classification performance; ECE quantifies calibration error (lower is better).
MethodMacro AUC (%)Macro F1 (%)ECE ↓Comm. MB/Round
FedAvg86.9 [86.5, 87.3]83.2 [82.8, 83.6]0.215 [0.213, 0.217]12.0 [11.9, 12.1]
SCAFFOLD88.3 [87.9, 88.7]84.4 [84.0, 84.8]0.199 [0.197, 0.201]13.5 [13.4, 13.6]
FedDyn88.8 [88.4, 89.2]84.7 [84.3, 85.1]0.194 [0.192, 0.196]12.0 [11.9, 12.1]
FedEHD89.6 [89.2, 90.0]85.4 [85.0, 85.8]0.182 [0.180, 0.184]12.0 [11.9, 12.1]
A-FedEHD89.4 [89.0, 89.8]85.3 [84.9, 85.7]0.183 [0.181, 0.185]12.0 [11.9, 12.1]
Table A6. EMA-Synthetic results ( K = 48 , participation C = 0.1 ). All metrics represent mean ± 95% CI over 30 runs. Macro AUC and Macro F1 measure classification performance; ECE quantifies calibration error (lower is better).
Table A6. EMA-Synthetic results ( K = 48 , participation C = 0.1 ). All metrics represent mean ± 95% CI over 30 runs. Macro AUC and Macro F1 measure classification performance; ECE quantifies calibration error (lower is better).
MethodMacro AUC (%)Macro F1 (%)ECE ↓Comm. MB/Round
FedAvg85.4 [85.0, 85.8]82.2 [81.8, 82.6]0.228 [0.226, 0.230]12.0 [11.9, 12.1]
SCAFFOLD86.7 [86.3, 87.1]83.6 [83.2, 84.0]0.202 [0.200, 0.204]13.5 [13.4, 13.6]
FedDyn87.1 [86.7, 87.5]83.9 [83.5, 84.3]0.197 [0.195, 0.199]12.0 [11.9, 12.1]
FedEHD88.4 [88.0, 88.8]84.7 [84.3, 85.1]0.184 [0.182, 0.186]12.0 [11.9, 12.1]
A-FedEHD88.3 [87.9, 88.7]84.6 [84.2, 85.0]0.185 [0.183, 0.187]12.0 [11.9, 12.1]
As K increases, macro AUC/F1 decline slightly due to amplified heterogeneity and smaller effective batch sizes, yet FedEHD and A-FedEHD maintain top performance across all settings. SCAFFOLD’s relative advantage increases at low participation ( C = 0.1 ) but remains below FedEHD’s final accuracy and calibration. A-FedEHD consistently reduces round-to-round variability, showing narrower confidence intervals.
Calibration perturbations (sensor offsets and multiplicative noise) were injected per Section 5.4. FedEHD’s entropy and cubic damping prevent overfitting to perturbed inputs, sustaining low ECE even with ±10% signal noise. Bandwidth comparisons indicate FedEHD’s communication matches FedAvg exactly, whereas SCAFFOLD adds a control-variates vector (13.5 MB/round at this scale).

Appendix C.2. Full Sharding Pseudocode and Reproducibility Materials

Algorithm A1 defines the station-faithful shard construction procedure. This appendix additionally provides the complete seed references and week-ID lists.
Algorithm A1: Station-Faithful Shard Construction (Synthetic EMA)
Input: Station streams { T s } s = 1 4 ; week partition W ; shards-per-station m; train/test split; Dirichlet parameter α (optional).
Output: Client datasets { D s , j } s 4 , j m ( K = 4 m ).
Split each T s into weekly bins and reserve test weeks.
Compute per-week event labels and assign weeks to m buckets in round-robin fashion, balancing event ratios.
(Optional) Resample week weights w Dirichlet ( α ) to induce label skew while preserving week contiguity.
Apply seeded affine perturbations ( a k , b k ) to selected features.
return  K = 4 m  client datasets with disjoint weeks.
For reproducibility, all random seeds used in the experiments are reported, covering (a) week binning, (b) Dirichlet reweighting, and (c) calibration perturbations. Each experimental run employs independent random number generator streams to eliminate any potential interdependence between sharding and training randomness.

Appendix D. Extended Calibration Figures and Reliability Diagrams

This appendix provides the full per-station and cross-dataset calibration results that complement the summary in Section 5.5. Table A7 reports detailed calibration metrics for each EMA station before and after temperature scaling (TS), while Table A8 presents corresponding results for the CIFAR-10 and CIFAR-100 benchmarks. Table A9 compares the relative effectiveness of temperature scaling across all datasets. Each table includes Expected Calibration Error (ECE), Negative Log-Likelihood (NLL), and Brier score values, reported as medians with 95% bootstrap confidence intervals over three independent runs. Lower values indicate better calibration performance.
Table A7. Per-station calibration metrics (EMA dataset) before and after temperature scaling (TS). Values are medians [95% CI] across three runs. Lower values are better.
Table A7. Per-station calibration metrics (EMA dataset) before and after temperature scaling (TS). Values are medians [95% CI] across three runs. Lower values are better.
Station/MethodECENLLBrier Score
Pre-TSPost-TSPre-TSPost-TSPre-TSPost-TS
Arima0.188 [0.185,0.191]0.107 [0.105,0.110]0.415 [0.408,0.422]0.295 [0.288,0.301]0.182 [0.180,0.184]0.164 [0.162,0.166]
Point Lisas0.192 [0.188,0.196]0.108 [0.106,0.111]0.431 [0.424,0.438]0.298 [0.291,0.304]0.184 [0.181,0.186]0.166 [0.164,0.168]
Port-of-Spain0.201 [0.197,0.205]0.118 [0.115,0.122]0.450 [0.442,0.457]0.314 [0.307,0.321]0.188 [0.185,0.190]0.172 [0.170,0.174]
San Fernando0.186 [0.183,0.189]0.106 [0.104,0.109]0.422 [0.416,0.429]0.297 [0.290,0.304]0.180 [0.177,0.182]0.163 [0.161,0.165]
Macro Average0.192 [0.190,0.195]0.109 [0.107,0.111]0.430 [0.425,0.435]0.301 [0.297,0.305]0.184 [0.182,0.185]0.166 [0.165,0.168]
Table A8. Extended calibration metrics on CIFAR-10 and CIFAR-100 before and after temperature scaling (TS). Reported values are medians [95% CI] across three runs.
Table A8. Extended calibration metrics on CIFAR-10 and CIFAR-100 before and after temperature scaling (TS). Reported values are medians [95% CI] across three runs.
Dataset/MethodECENLL
Pre-TSPost-TSPre-TSPost-TS
CIFAR-10 FedAvg0.158 [0.156,0.160]0.082 [0.080,0.084]0.465 [0.458,0.472]0.390 [0.384,0.396]
CIFAR-10 FedEHD0.143 [0.141,0.146]0.071 [0.070,0.073]0.439 [0.433,0.445]0.364 [0.358,0.370]
CIFAR-100 FedAvg0.191 [0.189,0.193]0.108 [0.106,0.110]0.512 [0.506,0.519]0.438 [0.431,0.445]
CIFAR-100 FedEHD0.173 [0.170,0.176]0.099 [0.097,0.101]0.486 [0.480,0.492]0.412 [0.406,0.419]
Table A9. Comparison of temperature-scaling (TS) effectiveness across datasets. Each entry shows the mean relative reduction (%) from pre-TS to post-TS for key calibration metrics, averaged over 30 runs with 95% confidence intervals.
Table A9. Comparison of temperature-scaling (TS) effectiveness across datasets. Each entry shows the mean relative reduction (%) from pre-TS to post-TS for key calibration metrics, averaged over 30 runs with 95% confidence intervals.
Dataset/MethodECE Reduction (%)NLL Reduction (%)Brier Reduction (%)
EMA FedAvg42.4 [41.8, 43.0]27.0 [26.4, 27.6]8.1 [7.9, 8.3]
EMA FedEHD41.5 [40.9, 42.1]30.0 [29.4, 30.6]9.6 [9.4, 9.8]
CIFAR-10 FedAvg48.1 [47.5, 48.7]16.1 [15.8, 16.4]8.9 [8.7, 9.1]
CIFAR-10 FedEHD50.3 [49.7, 50.9]17.1 [16.8, 17.4]8.5 [8.3, 8.7]
CIFAR-100 FedAvg43.5 [43.0, 44.0]14.5 [14.2, 14.8]7.8 [7.6, 8.0]
CIFAR-100 FedEHD42.8 [42.3, 43.3]15.2 [14.9, 15.5]8.0 [7.8, 8.2]
Overall, these results confirm that FedEHD achieves consistently lower calibration error (ECE) and better negative log-likelihood (NLL) compared with baseline methods both before and after temperature scaling. The temperature-scaling step yields a further 40–50% reduction in ECE across datasets, producing near-perfect probabilistic calibration without affecting accuracy.

Appendix E. Supplementary Architectural Specifications

To ensure full reproducibility, this appendix presents detailed network architectures, head dimensions, and preprocessing settings for all experiments.

Appendix E.1. Vision Benchmarks (CIFAR-10/100)

The vision benchmark experiments were conducted using standard image-classification datasets (CIFAR-10 and CIFAR-100) to evaluate FedEHD under non-IID federated conditions. These benchmarks serve as controlled environments for assessing convergence speed, accuracy, and stability compared with established baselines. The chosen architectures include classical convolutional networks and recent Transformer-based models, ensuring a representative comparison across both traditional and modern paradigms in federated vision learning. Table A10 details the complete architectural specifications used for all CIFAR experiments, including backbone configurations, model depths, and layer-wise organization.
Table A10. Summary of model architectures and hyperparameter settings across all experiments. All coefficients follow the scale-invariant parameterization λ H = c H s , λ 3 = c 3 / s , and  λ 2 = c 2 , where s = median ( | g | ) .
Table A10. Summary of model architectures and hyperparameter settings across all experiments. All coefficients follow the scale-invariant parameterization λ H = c H s , λ 3 = c 3 / s , and  λ 2 = c 2 , where s = median ( | g | ) .
ExperimentModelArchitecture and Key Hyperparameters
CIFAR-10ResNet-18 (CIFAR variant)18 convolutional layers with BN and ReLU. Input 3 × 32 × 32 ; stem: Conv(3→64, k = 3, s = 1, p = 1) + BN + ReLU; no maxpool. Stage1–4: [BasicBlock(64) × 2, 128 × 2, 256 × 2, 512 × 2] with strides [1,2,2,2]. GAP (512) → Linear(512→10). Batch size 32, LR 0.01, λ H = 0.5 , λ 2 = 0.05 , λ 3 = 0.005 , 5 local epochs, optimizer: SGD.
CIFAR-100ResNet-18 (CIFAR variant)Same as CIFAR-10 but final head Linear(512→100). Batch size 32, LR 0.01, λ H = 1.0 , λ 2 = 0.05 , λ 3 = 0.001 , 5 local epochs, optimizer: SGD.
EMA Case StudyLSTM + 2-Head MLPInput sequence: 24 h, 11 features (6 pollutants + 5 meteorological). 1-layer LSTM (hidden = 128, dropout = 0.1). Shared output: event-detection head (128→64→1 logit) and pollutant-regression head (128→64→2 outputs for PM 2.5 , O 3 ). Batch = 16, LR = 0.01, λ H = 0.2 s , λ 2 = 0.05 , λ 3 = 0.05 / s , 10 local epochs, optimizer: FedEHD.
EMA-Synthetic (K = 12–48)LSTM (shared) + multi-task headsSame base LSTM (128 hidden) shared across clients. Adaptive A-FedEHD variant for coefficient tuning. Batch = 16, LR = 0.01, epochs = 5, optimizer: FedEHD (A-variant).
CIFAR-10/100 (Transformer models)ViT-S (CIFAR config)Patch embedding: Conv(3→384, k = 4, s = 4) ⇒ 8 × 8 = 64 tokens. Transformer encoder: depth = 12, heads = 6, MLP ratio = 4, GELU activation, Pre-LN normalization. Absolute positional embeddings; classification head: Linear(384→10/100). Stochastic depth = 0.1; Dropout = 0.
CIFAR-10/100 (Transformer models)Swin-T (CIFAR config)Patch embedding = 4 × 4; window size = 4; Stages: dimensions [96,192,384,768], depths [2,2,6,2], heads [3,6,12,24]; MLP ratio = 4; classification head: Linear(768→10/100).
The ResNet-18 variant uses a 3 × 3 convolutional stem without a max-pooling layer, which is standard for CIFAR-scale images. For the Transformer-based models introduced in Transformer models, patch and window sizes were adjusted to accommodate 32 × 32 inputs. These models employ GELU activations and Pre-LayerNorm normalization to maintain stable optimization and ensure fair comparison across architectural families. Together, these configurations provide a balanced suite of vision backbones for evaluating FedEHD’s performance under diverse model geometries and data heterogeneity conditions.

Appendix E.2. Environmental Monitoring (EMA)—Time-Series Model

The environmental monitoring experiments employed a compact multi-task time-series architecture designed to jointly perform event detection and next-hour pollutant regression across distributed air-quality stations. The model combines temporal sequence learning with lightweight shared representations suitable for federated training under limited data and heterogeneous station conditions. Table A11 summarizes the full architectural specifications of the EMA model, including input features, preprocessing steps, and task-specific heads used during both local and federated optimization.
Table A11. Architectural specifications for the EMA time-series model (multi-task: event detection and next-hour regression).
Table A11. Architectural specifications for the EMA time-series model (multi-task: event detection and next-hour regression).
ComponentSpecification
Input windowSequence length T = 24  h; feature vector x t R 11 with pollutants {PM 2.5 , PM 10 , CO, NO 2 , SO 2 , O 3 } (6) and meteorology {temperature, humidity, wind speed, wind dir. sin, wind dir. cos} (5).
PreprocessingPer-station z-score normalization (training split only); wind direction encoded via sin/cos; missing values forward-filled up to 2 h then masked.
Backbone (shared)LSTM: 1 layer, hidden size = 128, bias = True, dropout = 0.1. Final hidden state h 24 R 128 shared across heads.
Classification headMLP: 128 64 (ReLU, Dropout 0.1) 1 logit; loss: BCEWithLogits; optional temperature scaling.
Regression headMLP: 128 64 (ReLU, Dropout 0.1) 2 outputs ( PM 2.5 , O 3 ); loss: MSE; trained jointly with classification (equal weight).
Total parameters≈0.22M (LSTM 0.18M + heads 0.04M).
Optimization hooksNo teacher forcing; gradient clip (global norm) = 5.0; early stopping on validation F1 (event detection).
Fusion fine-tuneGraph-Laplacian coherence penalty on station logits: Φ coh ( y ^ ( t ) ) = 1 2 i , j W i j ( y ^ i y ^ j ) 2 ; uniform W by default; step sizes per Appendix A.

Appendix E.3. Method-Specific Heads (Used in Additional Baselines)

Table A12 is referenced in Section 4 and Section 5 of the main text. Explicit, layer-by-layer architectural specifications are provided for every model used in this work, including the ResNet-18 (CIFAR variant), the EMA LSTM with multi-task heads, and compact Transformer baselines. These comprehensive details guarantee full reproducibility across all reported experiments.
Each shard produced during data preprocessing includes a structured export containing a JSON manifest with station_id, shard_id, week indices, and random seed; a metadata file summarizing event counts and pollutant coverage; and a split log linking week indices to the respective training and test partitions. All code, manifests, and seed lists are stored in the project repository under supplementary/ema-synthetic/. Replication of the results requires only the unaltered EMA CSV files and the provided week-ID lists.
Table A12. Auxiliary heads and additional modules used in personalized and contrastive federated baselines (Transformer models).
Table A12. Auxiliary heads and additional modules used in personalized and contrastive federated baselines (Transformer models).
MethodHead or Additional Module Description
Per-FedAvg/pFedMeSame backbone as base architecture; personalization head: Linear(512→C) for CIFAR datasets. Each client performs local fine-tuning for 1–5 epochs on its private data.
MOONProjection head: MLP 512→128 (ReLU) →128; contrastive loss weight = 1.0. No memory queue is maintained—contrast computed pairwise with the previous global model.
FedRep/FedBNFedRep: shared feature extractor (backbone) with client-specific classifier Linear(512→C). FedBN: retains local BatchNorm statistics while sharing all non-BN layers globally.

Appendix E.4. Overview of Architecture and Hyperparameter Settings

This subsection summarizes the architectural configurations, dataset-specific model backbones, and key training hyperparameters employed across all experiments. The architectures span convolutional (ResNet-18) and Transformer-based models for image classification (CIFAR-10/100) as well as recurrent and multi-task architectures (LSTM + MLP heads) for the environmental monitoring studies. Table A13 presents the full set of layer arrangements, batch sizes, learning rates, and scale-invariant coefficient mappings ( λ H = c H s , λ 3 = c 3 / s , λ 2 = c 2 ), where s = median ( | g | ) ensures normalization across tasks and model scales. All configurations were verified for stability and reproducibility, serving as the reference implementation settings for the experiments reported in the main text.
Table A13. Architectural and hyperparameter specifications across all experiments. All coefficients follow the scale-invariant parameterization λ H = c H s , λ 3 = c 3 / s , and λ 2 = c 2 , where s = median ( | g | ) .
Table A13. Architectural and hyperparameter specifications across all experiments. All coefficients follow the scale-invariant parameterization λ H = c H s , λ 3 = c 3 / s , and λ 2 = c 2 , where s = median ( | g | ) .
ExperimentModelLayers/UnitsHiddenBatchLR λ H λ 2 λ 3 EpochsOptimizer
CIFAR-10ResNet-18 (CIFAR variant)18 conv + BN + ReLU layers320.010.50.050.0055SGD
CIFAR-100ResNet-18 (CIFAR variant)18 conv + BN + ReLU layers320.011.00.050.0015SGD
EMA Case StudyLSTM + 2-Head MLP1 × 128 LSTM → 64 → 1+2 outputs128160.010.2 s0.050.05/s10FedEHD
EMA-Synthetic (K = 12–48)LSTM (shared) + multi-task heads1 × 128 LSTM, shared representation128160.01A-FedEHD adaptive5FedEHD (A-variant)
As detailed in Table A13, these configurations provided the foundation for all experiments discussed in Section 4 and Section 5.

Appendix E.5. Default Parameters for A-FedEHD

The adaptive variant uses the following defaults unless otherwise stated:
κ H = 0.3 , p = 1.5 , c H , min = 0.05 , c H , max = 0.6 , τ = 0.05 w 2 , α = 1.5 , β = 0.9 .
These values were validated across CIFAR-10/100, EMA, and the synthetic EMA benchmarks. They ensure stable adaptation under non-IID heterogeneity and bounded coefficient magnitudes.

Appendix E.6. Seed-Logging and Reproducibility Checklist

Each experimental run logs random seeds corresponding to all stochastic components to ensure exact reproducibility. The recorded seeds encompass several aspects of the experimental workflow, including data partitioning through Dirichlet sampling for non-IID splits and temporal sharding for the EMA-Synthetic experiments, as well as client sampling, which defines the subsets of clients S t participating in each communication round. Model initialization seeds are also stored to reproduce random weight initialization for each client model.
In addition, optimizer-level randomness is logged to maintain consistency in mini-batch ordering and data augmentation for CIFAR experiments. Finally, all perturbation-related seeds—including those controlling calibration shifts and affine transformations in the EMA-Synthetic setup—are captured to reproduce environmental and sensor variability. This comprehensive seed-tracking framework ensures that every experimental configuration can be precisely replicated across hardware, software, and dataset variations.

References

  1. McMahan, H.B.; Moore, E.; Ramage, D.; Hampson, S.; Arcas, B.A. Communication-Efficient Learning of Deep Networks from Decentralized Data. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics (AISTATS), Fort Lauderdale, FL, USA, 20–22 April 2017; pp. 1273–1282. [Google Scholar]
  2. Eden, R.; Chukwudi, I.; Bain, C.; Barbieri, S.; Callaway, L.; de Jersey, S.; George, Y.; Gorse, A.; Lawley, M.J.; Marendy, P.; et al. A scoping review of the governance of federated learning in healthcare. npj Digit. Med. 2025, 8, 427. [Google Scholar] [CrossRef] [PubMed]
  3. Zhan, S.; Huang, L.; Luo, G.; Zheng, S.; Gao, Z.; Chao, H.-C. A Review on Federated Learning Architectures for Privacy-Preserving AI: Lightweight and Secure Cloud–Edge–End Collaboration. Electronics 2025, 14, 2512. [Google Scholar] [CrossRef]
  4. Qi, P.; Chiaro, D.; Piccialli, F. Small Models, Big Impact: A Review on the Power of Lightweight Federated Learning. Future Gener. Comput. Syst. 2025, 162, 107484. [Google Scholar] [CrossRef]
  5. Kairouz, P.; McMahan, H.B.; Avent, B.; Bellet, A.; Bennis, M.; Bhagoji, A.N.; Bonawitz, K.; Charles, Z.; Cormode, G.; Cummings, R.; et al. Advances and open problems in federated learning. Found. Trends Mach. Learn. 2021, 14, 1–210. [Google Scholar] [CrossRef]
  6. Li, T.; Sahu, A.K.; Sanjabi, M.; Zaheer, M.; Talwalkar, A.; Smith, V. Federated Optimization in Heterogeneous Networks. In Proceedings of the 2nd Machine Learning and Systems Conference (MLSys), Austin, TX, USA, 2–4 March 2020; pp. 429–445. [Google Scholar]
  7. Karimireddy, S.P.; Kale, S.; Mohri, M.; Reddi, S.J.; Stich, S.U.; Suresh, A.T. SCAFFOLD: Stochastic Controlled Averaging for Federated Learning. In Proceedings of the 37th International Conference on Machine Learning (ICML), Vienna, Austria, 13–18 July 2020; pp. 5132–5143. [Google Scholar]
  8. Acar, D.A.E.; Zhao, Y.; Matas, R.; Mattina, M.; Whatmough, P.; Saligrama, V. Federated Learning Based on Dynamic Regularization. In Proceedings of the 9th International Conference on Learning Representations (ICLR), Virtual Event, 3–7 May 2021. [Google Scholar]
  9. Wang, J.; Liu, Q.; Liang, H.; Joshi, G.; Poor, H.V. Tackling the Objective Inconsistency Problem in Heterogeneous Federated Optimization. In Proceedings of the Advances in Neural Information Processing Systems 33 (NeurIPS), Virtual Event, 6–12 December 2020; pp. 7611–7623. [Google Scholar]
  10. Li, Q.; He, B.; Song, D. Model-Contrastive Federated Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Virtual Event, 19–25 June 2021; pp. 10713–10722. [Google Scholar]
  11. Reddi, S.J.; Charles, Z.; Zaheer, M.; Garrett, Z.; Rush, K.; Konečný, J.; Kumar, S.; McMahan, H.B. Adaptive Federated Optimization. In Proceedings of the 9th International Conference on Learning Representations (ICLR), Virtual Event, 3–7 May 2021. [Google Scholar]
  12. Dai, Z.; Shen, G.; Yuan, H.; Zheng, S.; Hu, Y.; Du, J.; Kong, X.; Xia, F. Towards Heterogeneous Federated Graph Learning via Structural Entropy and Prototype Aggregation. Inf. Sci. 2025, 718, 122338. [Google Scholar] [CrossRef]
  13. Wang, H.; Zou, G.; Cao, K.; Cui, Y.; Wei, T.; Hu, S. An Elastic Federated Learning Collaboration Framework for Computing-Constrained IoT. IEEE Trans. Comput.-Aided Des. Integr. Circuits Syst. 2025. [Google Scholar] [CrossRef]
  14. Khan, B.; Mousavi, S.; Daneshtalab, M. HeRD: Modelling Heterogeneous Degradations for Federated Super-Resolution in Satellite Imagery. IEEE Access 2025, 13, 125857–125868. [Google Scholar] [CrossRef]
  15. Zhang, L.; Huang, J.; Gao, W. Entropy Regularized Federated Optimization for Robust Non-IID Learning. Neurocomputing 2023, 550, 126437. [Google Scholar]
  16. Chaudhari, P.; Choromanska, A.; Soatto, S.; LeCun, Y.; Baldassi, C.; Borgs, C.; Chayes, J.; Sagun, L.; Zecchina, R. Entropy-SGD: Biasing Gradient Descent Into Wide Valleys. In Proceedings of the 5th International Conference on Learning Representations (ICLR), Toulon, France, 24–26 April 2017. [Google Scholar]
  17. Bernstein, J.; Zhao, J.; Azizzadenesheli, K.; Anandkumar, A. signSGD with Majority Vote is Communication Efficient and Fault Tolerant. In Proceedings of the 36th International Conference on Machine Learning (ICML), Stockholm, Sweden, 10–15 July 2018; pp. 560–569. [Google Scholar]
  18. Davenport, M.A.; Hegde, C.; Duarte, M.F.; Baraniuk, R.G. Joint Manifolds for Data Fusion. IEEE Trans. Image Process. 2010, 19, 2580–2594. [Google Scholar] [CrossRef] [PubMed]
  19. Zhao, Y.; Song, S.; Li, Q.; He, B.; Zhou, J. Federated Learning with Non-IID Data: A Survey. IEEE Trans. Neural Netw. Learn. Syst. 2022, 33, 6999–7017. [Google Scholar] [CrossRef]
  20. Li, X.; Wang, S.; Hu, J.; Wang, Y.; Li, T. FedAdaGrad: Adaptive Federated Optimization with Momentum for Communication Efficiency. IEEE Trans. Artif. Intell. 2023, 4, 1002–1014. [Google Scholar]
  21. Qin, D.; Li, R.; Pan, Y. Adaptive Federated Optimization via Gradient-Norm Scheduling. IEEE Trans. Artif. Intell. 2025, 6, 15–28. [Google Scholar]
  22. Wang, X.; Xu, L.; Zhao, H.; Yang, T.; Li, X. Federated Multi-Source Learning via Graph-Based Sensor Fusion. IEEE Internet Things J. 2023, 10, 8134–8147. [Google Scholar]
  23. Krizhevsky, A.; Nair, V.; Hinton, G. Learning Multiple Layers of Features from Tiny Images; Technical Report; University of Toronto: Toronto, ON, Canada, 2009. [Google Scholar]
  24. Environmental Management Authority (EMA). Ambient Air Quality Monitoring Station Reports and Event Logs for Trinidad and Tobago (2024–2025); EMA: Port of Spain, Trinidad and Tobago, 2025.
  25. Trinidad and Tobago Meteorological Service (TTMS). Saharan Dust Event Analysis for 2024–2025; TTMS: Piarco, Trinidad and Tobago, 2025.
  26. Liu, Y.; Zhang, H.; Gong, C. Hierarchical Federated Learning for Large-Scale Edge–Cloud Systems. IEEE Trans. Parallel Distrib. Syst. 2024, 35, 751–764. [Google Scholar]
  27. Tan, A.; Chen, T.; Smith, V. Towards Personalized Federated Learning. IEEE Trans. Neural Netw. Learn. Syst. 2022, 33, 1443–1458. [Google Scholar] [CrossRef] [PubMed]
  28. Guo, X.; Han, Y.; Wang, Z. Improving Calibration in Federated Learning via Temperature-Aware Aggregation. Pattern Recognit. 2023, 139, 109491. [Google Scholar]
  29. Sun, J.; Xu, Y.; Chen, J. Fair and Robust Federated Learning with Non-IID Data. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 9748–9763. [Google Scholar]
  30. Luo, Y.; Chen, H.; Liu, Z. High-Order Gradient Methods for Federated Optimization with Convergence Guarantees. Inf. Sci. 2024, 661, 119837. [Google Scholar]
Figure 1. Coefficient sensitivity and calibration. Heatmaps show accuracy and convergence plateau across ( c H , c 2 , c 3 ) ; variance–drift overlays reveal a broad stability ridge. FedEHD remains robust throughout the highlighted region, and A-FedEHD further reduces spread.
Figure 1. Coefficient sensitivity and calibration. Heatmaps show accuracy and convergence plateau across ( c H , c 2 , c 3 ) ; variance–drift overlays reveal a broad stability ridge. FedEHD remains robust throughout the highlighted region, and A-FedEHD further reduces spread.
Ai 06 00293 g001
Figure 2. Succinct λ –performance map. Three heatmaps show normalized test accuracy, rounds-to-threshold, and ECE as functions of ( c H , c 2 , c 3 ) in the scale-invariant parameterization λ H = c H s , λ 3 = c 3 / s , λ 2 = c 2 , with s = median ( | g | ) . A broad plateau and a stability ridge (higher c 2 or moderate c 3 ) emerge consistently across datasets.
Figure 2. Succinct λ –performance map. Three heatmaps show normalized test accuracy, rounds-to-threshold, and ECE as functions of ( c H , c 2 , c 3 ) in the scale-invariant parameterization λ H = c H s , λ 3 = c 3 / s , λ 2 = c 2 , with s = median ( | g | ) . A broad plateau and a stability ridge (higher c 2 or moderate c 3 ) emerge consistently across datasets.
Ai 06 00293 g002
Figure 3. EMA-Synthetic: Large-K cross-device simulation. FedEHD generalizes to K = 12 ,   24 ,   48 simulated clients, maintaining top macro AUC/F1 and lowest ECE under varying participation ratios C. (A) Convergence curves, (B) accuracy–rounds Pareto fronts, (C) ECE versus rounds, and (D) bytes-per-round efficiency.
Figure 3. EMA-Synthetic: Large-K cross-device simulation. FedEHD generalizes to K = 12 ,   24 ,   48 simulated clients, maintaining top macro AUC/F1 and lowest ECE under varying participation ratios C. (A) Convergence curves, (B) accuracy–rounds Pareto fronts, (C) ECE versus rounds, and (D) bytes-per-round efficiency.
Ai 06 00293 g003
Figure 4. Calibration study. (A) ECE versus entropy coefficient c H showing optimal range 0.2–0.5. (B) Reliability diagrams before and after temperature scaling (TS) for EMA stations. (C) Class-wise ECE for CIFAR-100. FedEHD yields lower miscalibration even before TS, and its post-TS performance reaches ECE 0.10 –0.12.
Figure 4. Calibration study. (A) ECE versus entropy coefficient c H showing optimal range 0.2–0.5. (B) Reliability diagrams before and after temperature scaling (TS) for EMA stations. (C) Class-wise ECE for CIFAR-100. FedEHD yields lower miscalibration even before TS, and its post-TS performance reaches ECE 0.10 –0.12.
Ai 06 00293 g004
Figure 5. Evidence at a Glance. (A) Variance and drift contraction with λ 2 ; (B) empirical | d | | g | scatter with implicit-clipping envelope; (C) loss drop versus sign agreement A t for λ H variants; (D) explicit–implicit error ratio versus learning rate η . Panels (AD) link observed empirical trends with the theoretical results proven in Appendix A.
Figure 5. Evidence at a Glance. (A) Variance and drift contraction with λ 2 ; (B) empirical | d | | g | scatter with implicit-clipping envelope; (C) loss drop versus sign agreement A t for λ H variants; (D) explicit–implicit error ratio versus learning rate η . Panels (AD) link observed empirical trends with the theoretical results proven in Appendix A.
Ai 06 00293 g005
Table 1. Recommended coefficient ranges ensuring stability and calibration consistency across datasets.
Table 1. Recommended coefficient ranges ensuring stability and calibration consistency across datasets.
CoefficientRecommended RangePrimary EffectInteraction Notes
c H 0.1–0.5Calibration and heterogeneity controlLarger c H requires moderate c 2 or c 3 values for stable calibration.
c 2 0.05–0.7Drift and variance controlBroadens the stability ridge; higher c 2 enhances drift suppression and variance damping.
c 3 0.03–0.12Tail damping and stabilityGains saturate beyond ≈0.12; complements c H to ensure smooth convergence.
Table 2. Succinct λ –performance map using the scale-invariant ( c H , c 2 , c 3 ) . Defaults: c H = 0.2 , c 2 = 0.05 , c 3 = 0.05 . Arrows denote direction vs. defaults; entries are medians across seeds/datasets.
Table 2. Succinct λ –performance map using the scale-invariant ( c H , c 2 , c 3 ) . Defaults: c H = 0.2 , c 2 = 0.05 , c 3 = 0.05 . Arrows denote direction vs. defaults; entries are medians across seeds/datasets.
RegionAccuracyRounds ↓ECE ↓Drift ↓
c H [ 0.1 , 0.5 ] ( c 2 = 0.05 , c 3 = 0.05 )
c 2 [ 0.3 , 0.7 ] ( c H = 0.2 , c 3 = 0.05 )
c 3 [ 0.05 , 0.12 ] ( c H = 0.2 , c 2 = 0.05 )↓ (tails)
High c H + low c 2 , c 3 ↑ (unstable)
Moderate c H + moderate c 2 or c 3
Table 3. Final test accuracy (%, mean ± 95% CI) on CIFAR-10 under non-IID partitioning ( α = 0.1 , K = 100 clients, 30 runs).
Table 3. Final test accuracy (%, mean ± 95% CI) on CIFAR-10 under non-IID partitioning ( α = 0.1 , K = 100 clients, 30 runs).
MethodMean Accuracy (%)95% CIRounds to 70%
FedAvg65.9[65.4, 66.4]>200 (not reached)
FedProx66.2[65.7, 66.8]>200
SCAFFOLD70.1[69.6, 70.6]130
MOON71.9[71.5, 72.4]100
FedDyn70.6[70.1, 71.2]120
FedOpt (Adam)68.1[67.7, 68.6]150
FedEHD (ours)72.6[72.2, 73.1]80
Centralized (upper bound)94.1[93.8, 94.4]
Table 4. Final test accuracy (%, mean ± 95% CI) on CIFAR-100 under non-IID partitioning ( α = 0.1 , K = 100 clients, 30 runs).
Table 4. Final test accuracy (%, mean ± 95% CI) on CIFAR-100 under non-IID partitioning ( α = 0.1 , K = 100 clients, 30 runs).
MethodMean Accuracy (%)95% CIRounds to 60%
FedAvg58.1[57.7, 58.6]300
FedProx58.6[58.1, 59.1]280
SCAFFOLD62.1[61.6, 62.7]300
MOON64.1[63.6, 64.7]250
FedDyn62.6[62.1, 63.2]280
FedOpt (Adam)60.1[59.6, 60.7]300
FedEHD (ours)68.1[67.6, 68.7]150
Centralized (ResNet-18)77.1[76.7, 77.5]
Table 5. Comprehensive event-detection results on the EMA dataset across four stations: Arima (AR), Point Lisas (PL), Port-of-Spain (PoS), and San Fernando (SF). Abbreviations: AUC—Area Under the ROC Curve; PR—Precision–Recall; F1—F1-score; ECE—Expected Calibration Error. Values represent mean ± 95% CI across 30 runs. Bold values indicate best performance.
Table 5. Comprehensive event-detection results on the EMA dataset across four stations: Arima (AR), Point Lisas (PL), Port-of-Spain (PoS), and San Fernando (SF). Abbreviations: AUC—Area Under the ROC Curve; PR—Precision–Recall; F1—F1-score; ECE—Expected Calibration Error. Values represent mean ± 95% CI across 30 runs. Bold values indicate best performance.
MethodARPLPoSSFMacro AUCPR–AUCF1ECE↓
FedAvg55.5 [54.9, 56.1]56.8 [56.2, 57.4]51.1 [50.5, 51.7]58.1 [57.5, 58.7]55.4 [55.0, 55.8]51.5 [51.1, 51.9]49.8 [49.3, 50.3]0.210 [0.208, 0.212]
FedProx60.4 [59.8, 61.0]58.9 [58.3, 59.5]49.2 [48.6, 49.8]62.2 [61.6, 62.8]57.7 [57.3, 58.1]53.6 [53.2, 54.0]53.1 [52.6, 53.6]0.200 [0.198, 0.202]
FedAdam (FedOpt)66.1 [65.5, 66.7]65.1 [64.5, 65.7]46.5 [45.9, 47.1]67.2 [66.6, 67.8]61.2 [60.8, 61.6]58.3 [57.9, 58.7]52.9 [52.4, 53.4]0.185 [0.183, 0.187]
SCAFFOLD66.5 [65.9, 67.1]59.4 [58.8, 60.0]51.7 [51.1, 52.3]65.6 [65.0, 66.2]60.8 [60.4, 61.2]53.5 [53.1, 53.9]48.1 [47.6, 48.6]0.201 [0.199, 0.203]
FedDyn70.7 [70.1, 71.3]70.6 [70.0, 71.2]45.6 [45.0, 46.2]70.7 [70.1, 71.3]64.4 [64.0, 64.8]60.5 [60.1, 60.9]52.0 [51.5, 52.5]0.188 [0.186, 0.190]
MOON71.3 [70.7, 71.9]71.0 [70.4, 71.6]50.5 [49.9, 51.1]71.7 [71.1, 72.3]66.1 [65.7, 66.5]61.2 [60.8, 61.6]54.4 [53.9, 54.9]0.186 [0.184, 0.188]
FedNova72.0 [71.4, 72.6]71.9 [71.3, 72.5]49.7 [49.1, 50.3]72.1 [71.5, 72.7]66.4 [66.0, 66.8]61.3 [60.9, 61.7]53.9 [53.4, 54.4]0.189 [0.187, 0.191]
LR (Pooled)71.2 [70.6, 71.8]71.1 [70.5, 71.7]45.6 [45.0, 46.2]71.2 [70.6, 71.8]64.8 [64.4, 65.2]59.7 [59.3, 60.1]44.4 [43.9, 44.9]0.210 [0.208, 0.212]
LR (Per-Station)71.1 [70.5, 71.7]71.0 [70.4, 71.6]47.0 [46.4, 47.6]71.1 [70.5, 71.7]65.0 [64.6, 65.4]61.6 [61.2, 62.0]46.2 [45.7, 46.7]0.200 [0.198, 0.202]
XGB (Pooled)53.3 [52.7, 53.9]39.2 [38.6, 39.8]62.3 [61.7, 62.9]53.3 [52.7, 53.9]52.0 [51.6, 52.4]42.2 [41.8, 42.6]52.2 [51.7, 52.7]0.170 [0.168, 0.172]
XGB (Per-Station)72.4 [71.8, 73.0]65.2 [64.6, 65.8]48.0 [47.4, 48.6]72.5 [71.9, 73.1]64.5 [64.1, 64.9]57.4 [57.0, 57.8]50.5 [50.0, 51.0]0.190 [0.188, 0.192]
FedEHD (Ours)73.7 [73.1, 74.3]72.9 [72.3, 73.5]53.1 [52.5, 53.8]74.1 [73.5, 74.7]67.0 [66.6, 67.4]63.0 [62.6, 63.4]55.0 [54.5, 55.5]0.183 [0.181, 0.185]
Table 6. FedEHD ablation and noise-robustness results on the EMA event-detection task. Each variant was evaluated over 30 runs with identical initialization and learning rates for fair comparison. Values are mean AUC (%) ± 95% CI.
Table 6. FedEHD ablation and noise-robustness results on the EMA event-detection task. Each variant was evaluated over 30 runs with identical initialization and learning rates for fair comparison. Values are mean AUC (%) ± 95% CI.
VariantArima AUC (%)Pt. Lisas AUC (%)PoS AUC (%)Macro AUC (%)
FedEHD (Full)73.7 [73.1, 74.3]72.9 [72.3, 73.5]53.1 [52.5, 53.8]67.0 [66.5, 67.6]
Only-Entropy ( λ H only)68.4 [67.8, 69.0]44.3 [43.6, 45.0]46.4 [45.8, 47.0]53.0 [52.5, 53.6]
Only-HighOrder ( λ 2 , λ 3 only)72.7 [72.1, 73.4]61.1 [60.5, 61.7]50.7 [50.1, 51.4]61.5 [61.0, 62.0]
All-Off (no λ terms)71.8 [71.2, 72.4]39.5 [38.9, 40.1]49.0 [48.4, 49.6]53.4 [52.8, 54.0]
Noise-Robust (Full + noise)69.9 [69.3, 70.6]70.7 [70.1, 71.3]50.1 [49.5, 50.8]63.6 [63.1, 64.1]
Table 7. Event-detection performance (Precision, Recall, and F1-score) averaged across all EMA stations during the test period (Jul–Dec 2025). Values represent mean (%) ± 95% CI over 30 independent runs.
Table 7. Event-detection performance (Precision, Recall, and F1-score) averaged across all EMA stations during the test period (Jul–Dec 2025). Values represent mean (%) ± 95% CI over 30 independent runs.
ApproachPrecision (%)Recall (%)F1-Score
Independent (Per-Station)98.0 [97.5, 98.5]60.1 [59.4, 60.9]0.74 [0.73, 0.75]
Centralized (All Data)80.2 [79.6, 80.8]95.1 [94.5, 95.7]0.87 [0.86, 0.88]
FedEHD + Fusion (Ours)88.3 [87.8, 88.8]90.2 [89.6, 90.8]0.89 [0.88, 0.90]
Table 8. Next-hour pollutant forecasting: RMSE (mean ± 95% CI, µg/m 3 ) of predicted concentrations for PM 2.5 and O 3 at two stations (lower RMSE indicates better forecasting accuracy). Results averaged over 30 runs.
Table 8. Next-hour pollutant forecasting: RMSE (mean ± 95% CI, µg/m 3 ) of predicted concentrations for PM 2.5 and O 3 at two stations (lower RMSE indicates better forecasting accuracy). Results averaged over 30 runs.
Model PM 2.5 RMSE O 3 RMSE
PoSArimaPoSArima
Independent (Per-Station)12.0 [11.7, 12.3]10.1 [9.8, 10.4]6.0 [5.8, 6.2]5.4 [5.2, 5.6]
FedEHD + Fusion (Ours)10.0 [9.8, 10.2]9.0 [8.8, 9.2]5.2 [5.0, 5.4]4.8 [4.6, 5.0]
Table 9. EMA (cross-silo, K = 4 , C = 1.0 ): macro performance metrics and calibration. All results represent mean ± 95% CI across 30 independent runs. Lower ECE indicates better calibration.
Table 9. EMA (cross-silo, K = 4 , C = 1.0 ): macro performance metrics and calibration. All results represent mean ± 95% CI across 30 independent runs. Lower ECE indicates better calibration.
MethodMacro AUC (%)Macro F1 (%)ECE ↓
FedAvg88.1 [87.7, 88.5]84.2 [83.8, 84.6]0.210 [0.208, 0.212]
FedProx88.7 [88.3, 89.1]84.6 [84.2, 85.0]0.200 [0.198, 0.202]
SCAFFOLD88.9 [88.5, 89.3]84.7 [84.3, 85.1]0.201 [0.199, 0.203]
FedDyn89.2 [88.8, 89.6]85.0 [84.6, 85.4]0.195 [0.193, 0.197]
MOON89.3 [88.9, 89.7]85.2 [84.8, 85.6]0.186 [0.184, 0.188]
FedOpt (Adam)89.6 [89.2, 90.0]85.4 [85.0, 85.8]0.185 [0.183, 0.187]
FedEHD (ours)90.4 [90.0, 90.8]86.2 [85.8, 86.6]0.183 [0.181, 0.185]
Table 10. Calibration metrics (mean ± 95% CI) before and after temperature scaling (TS) for EMA and CIFAR-100 datasets. TS preserves accuracy and F1 while reducing calibration error (ECE). Results are averaged over 30 independent runs.
Table 10. Calibration metrics (mean ± 95% CI) before and after temperature scaling (TS) for EMA and CIFAR-100 datasets. TS preserves accuracy and F1 while reducing calibration error (ECE). Results are averaged over 30 independent runs.
Dataset/MethodAccuracy (%)Macro F1 (%)ECE (Pre)ECE (Post-TS)
EMA FedAvg88.1 [87.7, 88.6]84.2 [83.8, 84.7]0.210 [0.208, 0.213]0.121 [0.119, 0.123]
EMA FedProx88.7 [88.3, 89.2]84.6 [84.2, 85.1]0.200 [0.198, 0.202]0.116 [0.114, 0.118]
EMA FedEHD90.4 [90.0, 90.8]86.2 [85.8, 86.6]0.183 [0.181, 0.185]0.107 [0.105, 0.109]
CIFAR-100 FedAvg68.4 [68.0, 68.8]65.2 [64.8, 65.6]0.191 [0.189, 0.193]0.108 [0.106, 0.110]
CIFAR-100 FedEHD69.8 [69.4, 70.2]67.1 [66.7, 67.5]0.173 [0.171, 0.175]0.099 [0.097, 0.101]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Khan, K.; Elibox, W.; Ramlochan, T.D.; Rajkumar, W.; Ramnath, S. FedEHD: Entropic High-Order Descent for Robust Federated Multi-Source Environmental Monitoring. AI 2025, 6, 293. https://doi.org/10.3390/ai6110293

AMA Style

Khan K, Elibox W, Ramlochan TD, Rajkumar W, Ramnath S. FedEHD: Entropic High-Order Descent for Robust Federated Multi-Source Environmental Monitoring. AI. 2025; 6(11):293. https://doi.org/10.3390/ai6110293

Chicago/Turabian Style

Khan, Koffka, Winston Elibox, Treina Dinoo Ramlochan, Wayne Rajkumar, and Shanta Ramnath. 2025. "FedEHD: Entropic High-Order Descent for Robust Federated Multi-Source Environmental Monitoring" AI 6, no. 11: 293. https://doi.org/10.3390/ai6110293

APA Style

Khan, K., Elibox, W., Ramlochan, T. D., Rajkumar, W., & Ramnath, S. (2025). FedEHD: Entropic High-Order Descent for Robust Federated Multi-Source Environmental Monitoring. AI, 6(11), 293. https://doi.org/10.3390/ai6110293

Article Metrics

Back to TopTop