Convergence Analysis and Optimisation of Privacy-Preserving Federated Learning for Hierarchical Graph Neural Networks in Distributed Cloud Anomaly Detection
Abstract
1. Introduction
1.1. Challenges in Distributed Cloud Network Monitoring
1.2. Limitations of Existing Federated Learning Approaches
1.3. Research Contributions
2. Related Work
2.1. Federated Learning in Distributed Networks
2.2. Cross-Jurisdictional Federated Learning
2.3. Federated Graph Neural Networks
2.4. Privacy-Preserving Network Monitoring
3. Privacy-Preserving Federated Learning Framework
3.1. System Architecture
- Component Level: Detects individual service anomalies (e.g., CPU spikes, memory leaks).
- System Level: Identifies multi-component interaction failures, including cascading service failures.
- Infrastructure Level: Captures data centre-wide issues like network partitions and power events.
3.1.1. Network-Centric Design Philosophy
- Geographical Distribution Awareness: The framework explicitly models the geographical distribution of federated sites, accounting for network latency, bandwidth limitations, and connectivity reliability.
- Communication-Efficient Protocols: All framework components are designed to minimise communication overhead while maintaining learning effectiveness.
- Network Resilience: The framework gracefully handles network disruptions, client dropouts, and varying connectivity conditions.
3.1.2. Framework Components
- Regional Client Nodes: Each geographical region operates an autonomous client that maintains local cloud infrastructure data and performs local model training using hierarchical GraphSAGE.
- Global Coordination Server: A central coordinator manages the federated learning process, performing layer-wise model aggregation, privacy preservation, and global model distribution.
- Privacy Protection Module: Implements differential privacy mechanisms with layer-specific noise calibration, recognising that different layers have varying sensitivity requirements.
- Network-Aware Aggregation Module: Performs layer-wise federated averaging that preserves hierarchical structure while adapting to network conditions and client availability patterns.
- Regional Personalisation Engine: Each region personalises the model to its specific network characteristics and infrastructure patterns after global model distribution.
3.1.3. Model Architecture Specifications
- Embedding layers: ~1.2 M parameters for feature transformation
- Graph convolution layers: ~1.8 M parameters for hierarchical message passing
- Classification heads: ~0.4 M parameters for multi-level anomaly detection
3.2. Layer-Wise Federated Aggregation
3.2.1. Limitations of Standard Federated Averaging
3.2.2. Layer-Wise Aggregation Mechanism
- Input embeddings . Higher weights preserve feature representations critical for downstream layers, based on information theory, showing that 60% of discriminating information resides in initial layers
- Graph convolutions Standard weight maintains structural relationships without bias
- Classification heads Lower weight enables regional specialisation while maintaining global coherence
3.2.3. Convergence Analysis
3.3. Privacy Enhancement with Differential Privacy
3.3.1. Threat Model and Privacy Objectives
3.3.2. Layer-Specific Differential Privacy
3.3.3. Privacy Composition Analysis
3.4. Network-Aware Personalisation Strategy
3.4.1. Meta-Learning Foundation
3.4.2. Regional Personalisation Process
3.4.3. Communication-Efficient Personalisation
4. Experimental Setup and Methodology
4.1. FEDGEN Testbed: A Real Distributed Cloud Network
4.1.1. Dataset Characteristics
- Data Volume: 57,663 total log entries across all sites
- Temporal Span: 6 months (January–June 2024)
- Log Types: System metrics, network traffic, application logs
- Graph Statistics (per site after balancing):
- (a)
- AfeBabalola: 6200 training nodes, 260,914 edges
- (b)
- LandMark: 13,454 training nodes, 577,736 edges
- (c)
- DRC_Congo: 11,532 training nodes, 494,362 edges
- (d)
- Features: 128-dimensional vectors
- Hierarchical Levels: 5 levels with 32 unique module classes at Level 5
- Model Complexity: 3,394,105 trainable parameters
- Training Configuration: 15 federated rounds, ~107 min total training time
- Anomaly Distribution: 3.2% positive samples (naturally imbalanced)
4.1.2. Geographical Distribution and Network Characteristics
- Site Locations:
- Landmark University (Omu-Aran, Kwara State, Nigeria): 8.1388° N, 5.0964° E
- Afe Babalola University (Ado-Ekiti, Nigeria): 7.6124° N, 5.3046° E
- DRC Congo Node (Kinshasa, DRC): 4.4419° S, 15.2663° E
- Global Coordinator: Covenant University (Ota, Nigeria): 6.6718° N, 3.1581° E
- Network Constraints:
- Maximum distance: 2780 km (between Nigeria and DRC)
- Inter-site latency: 45–120 ms (varying by route and time)
- Available bandwidth: 100 Mbps shared across all sites
- Connection reliability: 94.2% uptime (measured over 6-month period)
4.1.3. Cross-Jurisdictional Regulatory Environment
4.2. Federated Learning Configuration
4.3. Evaluation Metrics
- Communication Efficiency:
- Total communication volume (bytes transmitted)
- Communication rounds for convergence
- Bandwidth utilisation percentage
- Communication cost per performance unit
- Privacy Metrics:
- Membership inference attack success rate
- Property inference resistance
- Model inversion robustness
- Privacy budget accounting
- Network Resilience:
- Performance under client dropouts
- Latency sensitivity analysis
- Bandwidth adaptation capability
- Recovery time from disruptions
5. Results and Analysis
5.1. Federated Learning Effectiveness
5.1.1. Layer-Wise vs. Standard Aggregation
5.1.2. Regional Performance Analysis
5.2. Privacy–Utility Trade-Off Analysis
5.2.1. Privacy Budget Sensitivity
5.2.2. Privacy Attack Resistance
5.3. Communication Efficiency Analysis
5.3.1. Communication Overhead Comparison
5.3.2. Communication Breakdown Analysis
- Communication Components per Round (Proposed Approach):
- Model Parameters: 8.2 MB (80.4%)
- Privacy Noise: 1.1 MB (10.8%)
- Metadata: 0.6 MB (5.9%)
- Compression Overhead: 0.3 MB (2.9%)
5.4. Network Resilience and Robustness Analysis
5.4.1. Client Dropout Resilience
5.4.2. Network Condition Sensitivity
5.5. Convergence Patterns in Federated Training
- Key Observations:
- Rapid early convergence: Average client loss dropped from approximately 0.82 to 0.08 by round 3.
- Stable convergence: Loss values reached 0.0363 by the final round
- Regional variations: AfeBabalola consistently maintained slightly higher loss values
- Global model compromise: Higher stable loss (0.28–0.29), accommodating regional heterogeneity.
5.6. Energy Efficiency Analysis
- Lowest Communication Energy: 0.23 kWh (26% less than FedAvg, 30% less than FedProx), critical for bandwidth-constrained international networks
- Best Absolute Performance: 97.75% F1-score justifies the 1.98 kWh training energy investment, providing 12.89% higher accuracy than FedAvg
- Privacy–Performance Balance: Achieves 44.25 F1/kWh while maintaining differential privacy guarantees (ε = 5.83, δ = 10−5), whereas FedAvg and FedProx offer no privacy protection
- Operational Efficiency: Completes training in 15 rounds versus 22–24 for baselines, reducing coordination overhead and network disruption risk
- Energy–Performance Trade-off Analysis:
- Performance-Energy Relationship: Our 30% increase in training energy (1.98 kWh vs. 1.52 kWh for FedAvg) yields a 12.89% F1-score improvement (97.75% vs. 84.86%), demonstrating favourable returns on energy investment.
- CommunicationEnergy Synergy: The 93.3% reduction in communication volume (Table 9) translates to 0.23 kWh communication energy—the lowest among all methods despite higher training costs. This is particularly valuable for our 2780 km international deployment, where network transmission costs are significant.
- Round Efficiency Impact: Completing training in 15 rounds versus 22–24 for baselines (Table 3) reduces total coordination overhead, offsetting the higher per-round training energy through fewer synchronisation cycles.
- Privacy–Energy Trade-off: The differential privacy mechanisms add computational overhead, yet the framework maintains competitive efficiency at 44.25 F1-points/kWh while providing formal privacy guarantees (ε = 5.83, δ = 10−5) that FedAvg and FedProx lack.
5.7. Multi-Objective Optimisation Analysis and Pareto-Analysis
6. Deployment Considerations and Practical Guidelines
6.1. Infrastructure Requirements and Network Architecture
- Minimum Hardware Configuration (per regional site):
- CPU: 8 cores, 2.4 GHz+ (for graph construction and local training)
- Memory: 32 GB RAM (large graph representation)
- GPU: 16 GB VRAM (T4 or better for efficient training)
- Storage: 1 TB NVMe SSD (fast I/O for graph processing)
- Network: 10 Mbps sustained (federated communication)
- Production Deployment Configuration:
- Federated Network Topology:
| # Privacy-Preserving Federated Learning Production Configuration federation: coordinator: location: “Primary datacenter” hardware: “24 cores, 64 GB RAM, V100 GPU” redundancy: “Active-passive failover” regional_sites: - name: “Site A” location: “Geographic Region 1” hardware: “16 cores, 32 GB RAM, T4 GPU” connectivity: “100 Mbps primary, 20 Mbps backup” - name: “Site B” location: “Geographic Region 2” hardware: “16 cores, 32 GB RAM, T4 GPU” connectivity: “100 Mbps primary, 20 Mbps backup” security: encryption: “TLS 1.3 for all communication” authentication: “Mutual certificate authentication” privacy: “ε = 1.0, δ = 10−5 differential privacy” monitoring: metrics: “Real-time performance dashboards” alerts: “Automated failure detection” logging: “Complete audit trail” |
6.2. Production Deployment Configuration
- Key deployment configurations include:
- Load Balancing: Distribute client connections across multiple coordination servers
- Failover Strategy: Implement active–passive redundancy for the global coordinator
- Security Hardening: Enable TLS 1.3, mutual authentication, and audit logging
- Performance Tuning: Optimise batch sizes based on available GPU memory
- Monitoring Integration: Connect to existing APM and logging infrastructure
6.3. Operational Guidelines
- Deployment Phases:
- Infrastructure Setup (Weeks 1–2): Hardware installation and network configuration
- Data Pipeline Configuration (Weeks 3–4): Log collection and privacy setup
- Model Training and Validation (Weeks 5–6): Initial training and performance validation
- Production Deployment (Weeks 7–8): Gradual rollout with monitoring
- Monitoring Requirements:
- Client participation rates (target: >90%)
- Communication latency (alert if >200 ms)
- Privacy budget utilisation tracking
- Model performance metrics by region
7. Conclusions and Future Work
- Key Achievements
- Technical Innovations: Our layer-wise federated aggregation mechanism preserves hierarchical structure during distributed model updates, contributing 12.89% improvement in hierarchical F1-score over standard FedAvg. The privacy-preserving framework provides formal ( = 5.83, = 10−5) differential privacy guarantees with minimal utility loss (0.37% F1 degradation). Network-aware personalisation achieves substantial performance improvements (up to 55.66% in path accuracy) while maintaining communication efficiency.
- Operational Viability: The framework reduces communication overhead by 93.3% compared to raw data sharing and achieves 20–29% better efficiency than standard federated learning approaches. Comprehensive robustness analysis shows graceful degradation under realistic network constraints, maintaining effectiveness with 67% client participation and various network limitations.
- Economic Impact: Cost–benefit analysis demonstrates a 66% reduction in total annual operational costs with an 8.2-month payback period, making the proposed framework compelling for enterprise deployment across multiple jurisdictions.
- Limitations and Practical Considerations
- Scale Limitations: Our experiments involved three sites across 2780 km. While successful, scaling beyond this to tens or hundreds of sites will require architectural considerations not tested in our current setup.
- Platform Homogeneity: The FEDGEN testbed uses similar infrastructure across sites (NVIDIA T4 GPUs, Intel Xeon processors). Deployments with heterogeneous hardware may experience different convergence patterns.
- Temporal Patterns: Our 15-round training treats each round independently. Incorporating temporal dependencies between rounds could potentially improve convergence speed.
- Practical Deployment Significance
- Privacy Compliance: Our differential privacy implementation (ε = 5.83, δ = 10−5) provides formal guarantees suitable for regulatory compliance, as validated through resistance to membership inference (52.3% success rate) and model inversion attacks (4.9% success rate) shown in Table 7.
- Regional Adaptation: The 16.32% F1 improvement for DRC Congo through personalisation (Table 4) demonstrates the framework’s ability to adapt to regional characteristics while maintaining global model coherence.
- Operational Resilience: Maintaining 94.2% F1-score with only 67% client participation (Table 10) provides confidence for deployment in unreliable network environments.
- Future Research Directions
- Cross-Platform Federation: Extending the proposed framework to support federated learning across different cloud platforms (OpenStack, Kubernetes, proprietary systems) would address the reality of heterogeneous enterprise environments.
- Advanced Privacy Mechanisms: Investigating hierarchical split learning where different components of the model are split between clients and servers based on privacy sensitivity, and integration with homomorphic encryption for stronger privacy guarantees in high-security environments.
- Intelligent Network Optimisation: Developing federated learning protocols that dynamically adapt to varying network conditions and bandwidth availability, and incorporating network topology information into the aggregation process for optimal communication efficiency.
- Real-time Learning Capabilities: Implementing online learning capabilities for continuous model adaptation to evolving network conditions and attack patterns, and developing predictive capabilities for anomaly prevention rather than just detection.
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Aazam, M.; Huh, E.N.; St-Hilaire, M.; Lung, C.H.; Lambadaris, I. Cloud of things: Integration of IoT with cloud computing. In Studies in Systems, Decision and Control; Springer: Berlin/Heidelberg, Germany, 2016; Volume 36, pp. 77–94. [Google Scholar] [CrossRef]
- Vaquero, L.M.; Rodero-Merino, L.; Caceres, J.; Lindner, M. A break in the clouds: Towards a cloud definition. ACM SIGCOMM Comput. Commun. Rev. 2009, 39, 50–55. [Google Scholar] [CrossRef]
- Zhang, K.; Ni, J.; Yang, K.; Liang, X.; Ren, J.; Shen, X.S. Security and privacy in smart city applications: Challenges and solutions. IEEE Commun. Mag. 2017, 55, 122–129. [Google Scholar] [CrossRef]
- McMahan, B.; Ramage, D. Federated Learning: Collaborative Machine Learning Without Centralised Training Data. Google Research Blog, April 2017. Available online: https://research.google/blog/federated-learning-collaborative-machine-learning-without-centralized-training-data/ (accessed on 18 July 2025).
- Pearson, S.; Benameur, A. Privacy, security and trust issues arising from cloud computing. In Proceedings of the 2nd IEEE International Conference Cloud Computing Technology and Science (CloudCom), Indianapolis, IN, USA, 30 November–3 December 2010; pp. 693–702. [Google Scholar] [CrossRef]
- European Parliament. Regulation (EU) 2016/679 of the European Parliament and of the Council of 27 April 2016 on the Protection of Natural Persons with Regard to the Processing of Personal Data and on the Free Movement of Such Data, and Repealing Directive 95/46/EC (General Data Protection Regulation) (Text with EEA Relevance). EUR-Lex, April 2016. Available online: https://eur-lex.europa.eu/eli/reg/2016/679/oj/eng (accessed on 18 July 2025).
- Hsieh, K.; Ananthanarayanan, G.; Bodik, P.; Venkataraman, S.; Stoica, I. Gaia: Geo-distributed machine learning approaching LAN speeds. In Proceedings of the 14th USENIX Symposium on Networked Systems Design and Implementation, Boston, MA, USA, 27–29 March 2017; pp. 629–647. [Google Scholar]
- Konečný, J.; McMahan, H.B.; Yu, F.X.; Richtárik, P.; Suresh, A.T.; Bacon, D. Federated learning: Strategies for improving communication efficiency. In Proceedings of the NIPS Workshop on Private Multi-Party Machine Learning, Barcelona, Spain, 9 December 2016. [Google Scholar]
- McMahan, H.B.; Moore, E.; Ramage, D.; Hampson, S.; Arcas, B.A.Y. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the 20th International Conference Artificial Intelligence and Statistics (AISTATS), Fort Lauderdale, FL, USA, 20–22 April 2017; pp. 1273–1282. [Google Scholar] [CrossRef]
- Wu, C.; Wu, F.; Cao, Y.; Huang, Y.; Xie, X. FedGNN: Federated graph neural network for privacy-preserving recommendation. arXiv 2021, arXiv:2102.04925. [Google Scholar] [CrossRef]
- Xia, Q.; Ye, W.; Tao, Z.; Wu, J.; Li, Q. A survey of federated learning for edge computing: Research problems and solutions. High Confid. Comput. 2021, 1, 100008. [Google Scholar] [CrossRef]
- Shokri, R.; Shmatikov, V. Privacy-preserving deep learning. In Proceedings of the 53rd Annual Allerton Conference Communication, Control, and Computing (Allerton), Monticello, IL, USA, 29 September–2 October 2015; pp. 909–910. [Google Scholar] [CrossRef]
- Belenguer, A.; Navaridas, J.; Pascual, J.A. A review of federated learning applications in intrusion detection systems. Comput. Netw. 2025, 258, 111023. [Google Scholar] [CrossRef]
- Yang, Q.; Liu, Y.; Chen, T.; Tong, Y. Federated machine learning: Concept and applications. ACM Trans. Intell. Syst. Technol. 2019, 10, 1–19. [Google Scholar] [CrossRef]
- Liu, Y.; Yu, J.J.Q.; Kang, J.; Niyato, D.; Zhang, S. Privacy-preserving traffic flow prediction: A federated learning approach. IEEE Internet Things J. 2020, 7, 7751–7763. [Google Scholar] [CrossRef]
- Wang, S.; Tuor, T.; Salonidis, T.; Leung, K.K.; Makaya, C.; He, T. Adaptive federated learning in resource constrained edge computing systems. IEEE J. Sel. Areas Commun. 2019, 37, 1205–1221. [Google Scholar] [CrossRef]
- Kairouz, P.; McMahan, H.B.; Avent, B.; Bennis, M.; Bhagoji, A.N.; Bonawitz, K.; Charles, Z.; Cormode, G.; Cummings, R.; D’ Oliveira, R.G.; et al. Advances and open problems in federated learning. Found. Trends Mach. Learn. 2021, 14, 1–210. [Google Scholar] [CrossRef]
- Hamilton, W.L.; Ying, R.; Leskovec, J. Inductive representation learning on large graphs. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; pp. 1024–1034. [Google Scholar] [CrossRef]
- Sprenkamp, K.; Fernández, J.D.; Eckhardt, S.; Zavolokina, L. Overcoming intergovernmental data sharing challenges with federated learning. Data Policy 2024, 6, e27. [Google Scholar] [CrossRef]
- Xu, H.; Pan, M.; Huang, X.; Shen, A. Federated Learning in Autonomous Vehicles Using Cross-Jurisdictional Training. NVIDIA Technical Blog. 2024. Available online: https://developer.nvidia.com/blog/federated-learning-in-autonomous-vehicles-using-cross-jurisdictional-training/ (accessed on 18 July 2025).
- Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. In Proceedings of the 5th International Conference Learning Representations (ICLR), Toulon, France, 24–26 April 2017. [Google Scholar] [CrossRef]
- Zhang, K.; Yang, C.; Li, X.; Sun, L.; Yiu, S.M. Subgraph federated learning with missing neighbor generation. arXiv 2021, arXiv:2106.13430. [Google Scholar] [CrossRef]
- He, C.; Balasubramanian, K.; Ceyani, E.; Yang, C.; Xie, H.; Sun, L.; He, L.; Yang, L.; Yu, P.S.; Rong, Y.; et al. FedGraphNN: A federated learning system and benchmark for graph neural networks. arXiv 2021, arXiv:2104.07145. [Google Scholar] [CrossRef]
- Kellaris, G.; Papadopoulos, S.; Xiao, X.; Papadias, D. Differentially private event sequences over infinite streams. Proc. VLDB Endow. 2014, 7, 1155–1166. [Google Scholar] [CrossRef]
- Dwork, C.; McSherry, F.; Nissim, K.; Smith, A. Calibrating noise to sensitivity in private data analysis. In Proceedings of the 3rd Theory of Cryptography Conference (TCC), New York, NY, USA, 4–7 March 2006; pp. 265–284. [Google Scholar] [CrossRef]
- Wei, W.; Li, J.; Ding, M.; Ma, C.; Yang, H.H.; Farokhi, F.; Jin, S.; Quek, T.Q.S.; Poor, H.V. Federated learning with differential privacy: Algorithms and performance analysis. IEEE Trans. Inf. Forensics Secur. 2020, 15, 3454–3469. [Google Scholar] [CrossRef]
- Truex, S.; Baracaldo, N.; Anwar, A.; Steinke, T.; Ludwig, H.; Zhang, R.; Zhou, Y. A hybrid approach to privacy-preserving federated learning. In Proceedings of the 12th ACM Workshop on Artificial Intelligence and Security (AISec), London, UK, 15 November 2019; pp. 1–11. [Google Scholar] [CrossRef]
- Ma, X.; Wu, J.; Xue, S.; Yang, J.; Zhou, C.; Sheng, Q.Z.; Xiong, H.; Akoglu, L. A comprehensive survey on graph anomaly detection with deep learning. IEEE Trans. Knowl. Data Eng. 2023, 35, 12012–12038. [Google Scholar] [CrossRef]
- Akoglu, L.; Tong, H.; Koutra, D. Graph-based anomaly detection and description: A survey. Data Min. Knowl. Discov. 2015, 29, 626–688. [Google Scholar] [CrossRef]
- Li, T.; Sahu, A.K.; Zaheer, M.; Sanjabi, M.; Talwalkar, A.; Smith, V. Federated optimization in heterogeneous networks. In Proceedings of the Machine Learning and Systems 2 (MLSys), Austin, TX, USA, 2–4 March 2020; pp. 429–450. Available online: https://proceedings.mlsys.org/paper_files/paper/2020/file/1f5fe83998a09396ebe6477d9475ba0c-Paper.pdf (accessed on 18 July 2025).
- Dwork, C.; Roth, A. The algorithmic foundations of differential privacy. Found. Trends Theor. Comput. Sci. 2014, 9, 211–407. [Google Scholar] [CrossRef]
- Abadi, M.; Chu, A.; Goodfellow, I.; McMahan, H.B.; Mironov, I.; Talwar, K.; Zhang, L. Deep learning with differential privacy. In Proceedings of the 2016 ACM SIGSAC Conference Computer and Communications Security (CCS), Vienna, Austria, 24–28 October 2016; pp. 308–318. [Google Scholar] [CrossRef]
- Finn, C.; Abbeel, P.; Levine, S. Model-agnostic meta-learning for fast adaptation of deep networks. In Proceedings of the 34th International Conference Machine Learning (ICML), Sydney, Australia, 6–11 August 2017; pp. 1126–1135. [Google Scholar] [CrossRef]
- Zhao, Y.; Li, M.; Lai, L.; Suda, N.; Civin, D.; Chandra, V. Federated learning with non-IID data. arXiv 2018, arXiv:1806.00582. [Google Scholar] [CrossRef]
- Mohri, M.; Sivek, G.; Suresh, A.T. Agnostic federated learning. In Proceedings of the 36th International Conference Machine Learning (ICML), Long Beach, CA, USA, 10–15 June 2019; pp. 4615–4625. [Google Scholar] [CrossRef]
- Valiant, L.G. A theory of the learnable. Commun. ACM 1984, 27, 1134–1142. [Google Scholar] [CrossRef]
- Fallah, A.; Mokhtari, A.; Ozdaglar, A. Personalized federated learning: A meta-learning approach. arXiv 2020, arXiv:2002.07948. [Google Scholar] [CrossRef]
- Geiping, J.; Bauermeister, H.; Dröge, H.; Moeller, M. Inverting gradients—How easy is it to break privacy in federated learning? In Proceedings of the Advances in Neural Information Processing Systems 33 (NeurIPS 2020), Virtual Conference, 6–12 December 2020; pp. 16937–16947. Available online: https://proceedings.neurips.cc/paper/2020/hash/c4ede56bbd98819ae6112b20ac6bf145-Abstract.html (accessed on 15 August 2025).
- Nasr, M.; Shokri, R.; Houmansadr, A. Comprehensive privacy analysis of deep learning: Passive and active white-box inference attacks against centralized and federated learning. In Proceedings of the 2019 IEEE Symposium on Security and Privacy (SP), San Francisco, CA, USA, 20–22 May 2019; pp. 739–753. [Google Scholar] [CrossRef]
- He, X.; Jia, J.; Backes, M.; Gong, N.Z.; Zhang, Y. Stealing links from graph neural networks. In Proceedings of the 30th USENIX Security Symposium, Virtual Conference, 11–13 August 2021; pp. 2669–2686. Available online: https://www.usenix.org/conference/usenixsecurity21/presentation/he-xinlei (accessed on 15 August 2025).
- Mughole, K.D.; Badejo, J.; Okokpujie, K.O.; Adetiba, E.; Mbasa, M.; Vingi, N.; Oyeyemi, G.; Takenga, C. SecFedMDM-1: A Federated Learning-Based Malware Detection Model for Interconnected Cloud Infrastructures. IEEE Access 2025, 13, 101246–101261. [Google Scholar] [CrossRef]
- Adewale, A.O.; Obiazi, O.; Okokpujie, K.; Koto, O. Hybridization of the Q-learning and honey bee foraging algorithms for load balancing in cloud environments. Int. J. Electr. Comput. Eng. 2024, 14, 4602–4615. [Google Scholar] [CrossRef]
- Ewejobi, P.; Okokpujie, K.; Adetiba, E.; Alao, B. Homomorphic Encryption for Genomics Data Storage on a Federated Cloud: A Mini Review. In Proceedings of the 2024 International Conference on Science, Engineering and Business for Driving Sustainable Development Goals (SEB4SDG), Omu-Aran, Nigeria, 2–4 April 2024; IEEE: New York, NY, USA, 2024; pp. 1–13. [Google Scholar] [CrossRef]





| Framework | Layer-Preserving | Privacy Mechanism | Dynamic Graphs | Hierarchical Support | Communication Efficiency |
|---|---|---|---|---|---|
| FedGraphNN [23] | No | Basic Encryption | No | Limited | baseline |
| FedSage+ [22] | Partial | None | No | No | 85% |
| FedGNN [10] | No | Local DP | No | No | 78% |
| Proposed | Yes | Layer-Specific DP | Adaptive | Full | 93.3% |
| Property | Guarantee | Section |
|---|---|---|
| Convergence Rate | with bounded gradients | Section 3.2.3 |
| Privacy Budget | per round | Section 3.3.2 |
| Communication | = model size | Section 3.4.3 |
| Personalisation | steps | Section 3.4.1 |
| Method | Hierarchical F1 | Path Accuracy | Communication Rounds | Training Time |
|---|---|---|---|---|
| FedAvg | 84.86 ± 1.48% | 48.09 ± 2.15% | 22 | 78 min |
| FedProx | 86.06 ± 1.36% | 50.43 ± 1.96% | 24 | 82 min |
| Proposed Framework | 97.75 ± 0.32% | 90.04 ± 0.55% | 15 | 107 min |
| Region | Global F1 | Personalised F1 | Improvement | Global Path Acc | Personalised Path Acc | Improvement |
|---|---|---|---|---|---|---|
| Afe Babalola | 87.77% | 95.1% | +7.34% | 53.30% | 84.18% | +30.88% |
| Landmark | 83.43% | 98.9% | +15.46% | 39.41% | 95.07% | +55.66% |
| DRC Congo | 82.94% | 99.3% | +16.32% | 43.03% | 96.87% | +53.84% |
| Method | Privacy Guarantee | Hierarchical Preservation | Communication Efficiency | Convergence Rate |
|---|---|---|---|---|
| FedAvg [8] | None | No | Baseline | |
| FedProx [19] | None | No | Baseline | |
| DP-FL [12] | No | 85% | ||
| Proposed Method | Yes | 93.3% |
| Privacy Budget () | Hierarchical F1 | Path Accuracy | Privacy Level | Utility Loss |
|---|---|---|---|---|
| ∞ (No Privacy) | 98.12% | 91.24% | None | - |
| 10.0 | 97.89% (−0.23%) | 90.87% (−0.37%) | Low | Minimal |
| 5.83 | 97.75% (−0.37%) | 90.04% (−1.20%) | Moderate | Low |
| 3.0 | 96.18% (−1.94%) | 88.31% (−2.93%) | High | Moderate |
| 1.0 | 93.84% (−4.28%) | 84.73% (−6.51%) | Very High | Significant |
| Attack Type | Success Rate (No Privacy) | Success Rate ( = 5.83) | Protection Effectiveness |
|---|---|---|---|
| Membership Inference | 78.4% | 52.3% | 33% reduction |
| Property Inference | 89.2% | 74.0% | 17% reduction |
| Model Inversion | 65.1% | 4.9% | 92% reduction |
| Attribute Inference | 71.3% | 43.8% | 39% reduction |
| Attack Type | Success Rate (No Privacy) | Success Rate ( = 5.83) | Latest Methods Tested |
|---|---|---|---|
| Gradient Inversion [38] | 82.3% | 8.7% | Inverting Gradients (2020) |
| Feature Reconstruction [39] | 76.5% | 12.4% | ML Privacy Meter (2019) |
| Graph Structure Inference [40] | 69.2% | 15.3% | Link Stealing (2021) |
| Metric | Raw Data Sharing | FedAvg | FedProx | Proposed Framework |
|---|---|---|---|---|
| Data per Round | 2.3 GB | 8.7 MB | 8.9 MB | 10.2 MB |
| Total Rounds | 1 | 22 | 24 | 15 |
| Total Communication | 2.3 GB | 191 MB | 214 MB | 153 MB |
| Reduction vs. Raw | - | 92.0% | 90.7% | 93.3% |
| Dropout Scenario | Participating Clients | Hierarchical F1 | Path Accuracy | Recovery Time |
|---|---|---|---|---|
| No Dropout | 3/3 (100%) | 97.75% | 90.04% | - |
| Single Dropout | 2/3 (67%) | 0.92%) | 88.21% (−1.83%) | 2 rounds |
| Intermittent | ~2.4/3 (80%) | 1.84%) | 86.45% (−3.59%) | 4 rounds |
| DRC Isolation | 2/3 (67%) | 94.22% (−3.53%) | 83.17% (−6.87%) | 6 rounds |
| Network Condition | Latency | Bandwidth | Hierarchical F1 | Convergence Rounds |
|---|---|---|---|---|
| Optimal | 45 ms | 100 Mbps | 97.75% | 15 |
| High Latency | 200 ms | 100 Mbps | 97.41% (−0.34%) | 17 |
| Low Bandwidth | 45 ms | 20 Mbps | 97.52% (−0.23%) | 19 |
| Constrained | 200 ms | 20 Mbps | 96.89% (−0.86%) | 22 |
| Method | Training Energy | Communication Energy | Total Energy | Efficiency (F1/kWh) |
|---|---|---|---|---|
| Centralised | 1.42 kWh | 0.95 kWh | 2.37 kWh | 41.43 |
| FedAvg | 1.52 kWh | 0.31 kWh | 1.83 kWh | 46.36 |
| FedProx | 1.60 kWh | 0.33 kWh | 1.93 kWh | 44.59 |
| Proposed Framework | 1.98 kWh | 0.23 kWh | 2.21 kWh | 44.25 |
| Configuration | Privacy (ε) | F1-Score | Communication (MB) | Pareto Score |
|---|---|---|---|---|
| High Privacy | 1.0 | 93.84% | 153 | 0.72 |
| Balanced | 5.83 | 97.75% | 153 | 0.89 |
| High Accuracy | 10.0 | 97.89% | 167 | 0.81 |
| Low Communication | 5.83 | 96.12% | 112 | 0.85 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Lawal, C.; Olaniyan, O.M.; Okokpujie, K.; Adetiba, E. Convergence Analysis and Optimisation of Privacy-Preserving Federated Learning for Hierarchical Graph Neural Networks in Distributed Cloud Anomaly Detection. Computation 2025, 13, 283. https://doi.org/10.3390/computation13120283
Lawal C, Olaniyan OM, Okokpujie K, Adetiba E. Convergence Analysis and Optimisation of Privacy-Preserving Federated Learning for Hierarchical Graph Neural Networks in Distributed Cloud Anomaly Detection. Computation. 2025; 13(12):283. https://doi.org/10.3390/computation13120283
Chicago/Turabian StyleLawal, Comfort, Olatayo M. Olaniyan, Kennedy Okokpujie, and Emmanuel Adetiba. 2025. "Convergence Analysis and Optimisation of Privacy-Preserving Federated Learning for Hierarchical Graph Neural Networks in Distributed Cloud Anomaly Detection" Computation 13, no. 12: 283. https://doi.org/10.3390/computation13120283
APA StyleLawal, C., Olaniyan, O. M., Okokpujie, K., & Adetiba, E. (2025). Convergence Analysis and Optimisation of Privacy-Preserving Federated Learning for Hierarchical Graph Neural Networks in Distributed Cloud Anomaly Detection. Computation, 13(12), 283. https://doi.org/10.3390/computation13120283

