A Novel Multi-Server Federated Learning Framework in Vehicular Edge Computing
Abstract
1. Introduction
- We propose a novel multi-server framework at the RSU level, enabling peer servers to share model updates among neighboring servers. By leveraging inter-RSU collaboration, our framework accelerates convergence and enhances robustness in highly mobile vehicular networks;
- We develop and evaluate server-level, performance-based aggregation strategies whereby each FL server first assesses incoming peer models’ accuracy and loss against its own validation data and then selectively incorporates these updates. Empirical results demonstrate that these performance-driven methods consistently outperform standard baselines. We also employ a statistical model to penalize outliers and reduce the impact of contributions from servers whose data or model updates deviate significantly from the expected behavior. These deviations, or “outliers”, may indicate malicious activity, faulty data, or other anomalous behaviors;
- We propose an inter-server handover mechanism for continuous FL that preserves updates that would otherwise be lost as vehicles traverse multiple servers, leveraging migrating clients to accelerate global convergence. Our approach also incorporates a server-side evaluation module that assesses and weights each newly arrived client’s update—taking into account its origin server—before integrating it into the global model aggregation;
- We show via experiments that our solution provides remarkable performance gains compared to (1) baseline FL at each server, where servers train and aggregate only their own local updates (no inter-server collaboration), referred to as “single-server FL” in this paper; (2) hierarchical FL with cloud synchronization, in which servers periodically send models to—and receive updates from—a central cloud server (incurring a high communication overhead), referred to as “cloud-based FL” in this paper; and (3) methods that do not take model handover between servers into account.
2. Related Work
2.1. Multi-Server Federated Learning
2.2. Multi-Server Federated Learning in Vehicular Networks
2.3. Research Gap
3. System Model and Problem Definition
3.1. System Model
- A set of RSUs., each hosting an FL server.
- A set of vehicles across RSUs.
3.2. Problem Definition
- Ensuring robust aggregation, despite heterogeneous updates;
- Respecting vehicles’ mobility, which causes frequent handovers and variable participation;
- Minimizing communication overhead, given limited edge-network bandwidth.
4. Proposed MS-FL Framework
4.1. Inter-Server Model Aggregation
- Sequential Aggregation (SA)
Algorithm 1: Sequential Aggregation (SA) |
Input: server model , peer server model (, aggregation parameter ( Output: updated server model |
- Binary Aggregation (BA)
- Dynamic Weighted Averaging Aggregation (DWAA)
- Statistical Performance-Aware Aggregation (SPAA)
Algorithm 2: FL server operation in MS-FL using SPAA/DWAA |
Input: servers’ models (, vehicles’ models (). Output: updated server model |
#Initialize the model at server |
4.2. Intra-Server Training Phase
4.2.1. Local Training on Vehicles
4.2.2. Vehicles’ Models’ Aggregation at Server
4.3. Clients’ Movements and Handover
- be the set of vehicles that both started and completed their local updates entirely within server ’s coverage during round t;
- be the set of vehicles that began training under server but crossed into ’s coverage before uploading their update.
5. Transmission Latency Analysis
5.1. MS-FL
- Client-to-RSU Communication: For each client in round , the uploading time to its associated server is given by the following:
- RSU-to-RSU Communication: Although RSUs typically use high-speed backhaul links [31], we include the following general mode for completeness. The latency to exchange a model between two servers at round t is as follows:
- Overall Transmission Latency: Assuming that training must proceed for rounds to reach the target performance, the total transmission latency is simply the sum of the per-round latencies:
5.2. Single-Server FL
5.3. Cloud-Based FL
- Edge aggregation: Exactly as in MS-FL and single-server FL, clients exchange updates with their local server. The per-round “edge” delay is therefore the same defined in Section 5.1.
- Global aggregation: After edge aggregation, the cloud server aggregates the models from the various edge servers. The transmission latency for global aggregation is expressed as follows:
- Overall Transmission Latency: The end-to-end transmission time to achieve the targeted accuracy is the sum of the latencies over all edge and global aggregation rounds. Assuming the two layer federated learning process requires global aggregation rounds to reach the specified performance level, the total transmission latency is given by the following:
5.4. Handover Latency
6. Experimental Results
6.1. Simulation Setup
6.2. Dataset
6.3. Comparision Schemes
- Cloud-based FL (hierarchical FL)—following the layered design in [10], edge servers serve as intermediaries between vehicles and a single cloud server. Each client trains locally for epochs and uploads its parameters to its server; the server aggregates these updates, then forwards the result to the cloud. The cloud performs a second-level aggregation across all servers and broadcasts the global model back. Crucially, servers cooperate only through the cloud—no peer-to-peer exchange occurs;
- Isolated single-server FL at each RSU—here, every RSU acts as an independent FL server [5,6]. It trains a model using only the clients in its own coverage area and never exchanges parameters with either peer servers or a cloud node. This baseline captures the performance of a non-collaborative multi-server deployment.
6.4. Simulation Results
- True positives (TP)—images that truly belong to a class and are labeled as such by the model (e.g., a “speed-limit-30 km h” sign or a handwritten ‘8’ identified correctly);
- True negatives (TN)—images that do not belong to the class and are correctly rejected;
- False positives (FP)—images that are erroneously assigned to the class (a “speed-limit-50” labeled as “30”; a ‘3’ labeled as ‘8’);
- False negatives (FN)—images that should have been assigned to the class but were missed by the model.
- Precision—the proportion of predicted positives that are indeed correct. High precision shows that the model rarely confuses one traffic sign (or digit) for another;
- Recall—the proportion of actual positives retrieved by the model. A high recall means that very few relevant signs or digits are overlooked;
- F1-score—the harmonic balance between precision and recall, providing a single, robust measure of overall classification effectiveness.
6.4.1. Performance Evaluation of the Proposed Server-Level Aggregation Methods
6.4.2. Comparative Experiments: MS-FL vs. Baselines
6.4.3. Convergence Speed Under Latency Constraints
6.4.4. Mobility and Handover Effects
7. Discussion and Future Directions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Grigorescu, S.; Trasnea, B.; Cocias, T.; Macesanu, G. A survey of deep learning techniques for autonomous driving. J. Field Robot. 2020, 37, 362–386. [Google Scholar] [CrossRef]
- Brown, T.B.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language models are few-shot learners. arXiv 2020, arXiv:2005.14165. [Google Scholar] [CrossRef]
- Lu, Y.; Huang, X.; Zhang, K.; Maharjan, S.; Zhang, Y. Blockchain empowered asynchronous federated learning for secure data sharing in Internet of Vehicles. IEEE Trans. Veh. Technol. 2020, 69, 4298–4311. [Google Scholar] [CrossRef]
- Häfner, B.; Jiru, J.; Schepker, H.; Schmitt, G.A.; Ott, J. Proposing cooperative maneuvers among automated vehicles using machine learning. In Proceedings of the 24th International ACM Conference on Modeling, Analysis and Simulation of Wireless and Mobile Systems (MSWiM ’21), Alicante, Spain, 22–26 November 2021; Association for Computing Machinery: New York, NY, USA, 2021; pp. 173–180. [Google Scholar]
- McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics (AISTATS), Fort Lauderdale, FL, USA, 20–22 March 2017; pp. 1273–1282. [Google Scholar]
- Kairouz, P.; McMahan, H.B.; Avent, B.; Bellet, A.; Bennis, M.; Bhagoji, A.N.; Bonawit, K.; Charles, Z.; Cormode, G.; Cummings, R.; et al. Advances and Open Problems in Federated Learning. arXiv 2021, arXiv:1912.04977. [Google Scholar] [CrossRef]
- Moon, S.; Lim, Y. Client selection for federated learning in vehicular edge computing: A deep reinforcement learning approach. IEEE Access, 2024; in press. [Google Scholar]
- Cha, N.; Chang, L. Multi-objective distributed client selection in federated learning-assisted Internet of Vehicles. Sensors 2024, 24, 4180. [Google Scholar] [CrossRef] [PubMed]
- Albelaihi, R. Mobility prediction and resource-aware client selection for federated learning in IoT. Future Internet 2025, 17, 3. [Google Scholar] [CrossRef]
- Zhou, X.; Liang, W.; She, J.; Yan, Z.; Wang, K.I.-K. Two-layer federated learning with heterogeneous model aggregation for 6G supported Internet of Vehicles. IEEE Trans. Veh. Technol. 2021, 70, 5308–5317. [Google Scholar] [CrossRef]
- Lim, W.Y.B.; Luong, N.C.; Hoang, D.T.; Jiao, Y.; Liang, Y.-C.; Yang, Q.; Niyato, D.; Miao, C. Federated learning in mobile edge networks: A comprehensive survey. IEEE Commun. Surv. Tutor. 2020, 22, 2031–2063. [Google Scholar] [CrossRef]
- Liu, L.; Zhang, J.; Song, S.; Letaief, K.B. Client-edge-cloud hierarchical federated learning. In Proceedings of the IEEE International Conference on Communications (ICC), Dublin, Ireland, 7–11 June 2020; pp. 1–6. [Google Scholar]
- Wang, J.; Wang, S.; Chen, R.-R.; Ji, M. Local averaging helps: Hierarchical federated learning and convergence analysis. arXiv 2020, arXiv:2010.12998. [Google Scholar]
- Xie, M.; Long, G.; Shen, T.; Zhou, T.; Wang, X.; Jiang, J. Multicenter federated learning. arXiv 2020, arXiv:2005.01026. [Google Scholar]
- Lee, J.-W.; Oh, J.; Lim, S.; Yun, S.-Y.; Lee, J.-G. Tornadoaggregate: Accurate and scalable federated learning via the ring-based architecture. arXiv 2020, arXiv:2012.03214. [Google Scholar]
- Duan, M.; Liu, D.; Chen, X.; Liu, R.; Tan, Y.; Liang, L. Self-balancing federated learning with global imbalanced data in mobile systems. IEEE Trans. Parallel Distrib. Syst. 2020, 32, 59–71. [Google Scholar] [CrossRef]
- Taïk, A.; Mlika, Z.; Cherkaoui, S. Clustered vehicular federated learning: Process and optimization. IEEE Trans. Intell. Transp. Syst. 2022, 23, 25371–25383. [Google Scholar] [CrossRef]
- Han, D.-J.; Choi, M.; Park, J.; Moon, J. FedMes: Speeding up federated learning with multiple edge servers. IEEE J. Sel. Areas Commun. 2021, 39, 3870–3885. [Google Scholar] [CrossRef]
- Qu, Z.; Li, X.; Xu, J.; Tang, B.; Lu, Z.; Liu, Y. On the convergence of multi-server federated learning with overlapping area. IEEE Trans. Mob. Comput. 2023, 22, 6647–6662. [Google Scholar] [CrossRef]
- Rjoub, G.; Wahab, O.A.; Bentahar, J.; Bataineh, A. Trust-driven reinforcement selection strategy for federated learning on IoT devices. Computing, 2022; in press. [Google Scholar]
- Mazloomi, F.; Shah Heydari, S.; El-Khatib, K. Trust-based knowledge sharing among federated learning servers in vehicular edge computing. In Proceedings of the International ACM Symposium on Design and Analysis of Intelligent Vehicular Networks and Applications, Montreal, QC, Canada, 30 October–3 November 2023; Association for Computing Machinery: New York, NY, USA, 2023; pp. 9–15. [Google Scholar]
- Liu, L.; Chen, C.; Pei, Q.; Maharjan, S.; Zhang, Y. Vehicular edge computing and networking: A survey. arXiv 2019, arXiv:1908.06849. [Google Scholar] [CrossRef]
- Liu, S.; Yu, J.; Deng, X.; Wan, S. FedCPF: An efficient-communication federated learning approach for vehicular edge computing in 6G communication networks. IEEE Trans. Intell. Transp. Syst. 2021, 23, 1616–1629. [Google Scholar] [CrossRef]
- 3GPP. Evolved Universal Terrestrial Radio Access (E-UTRA); Radio Resource Control (RRC); Protocol Specification (Release 16), 3GPP TS 36.331 V16.5.0; 3GPP: Sophia Antipolis, France, 2021; Available online: https://www.3gpp.org/ftp/Specs/archive/36_series/36.331/ (accessed on 15 May 2025).
- Han, J.; Wu, B. Handover in the 3GPP Long Term Evolution (LTE) systems. In Proceedings of the 2010 Global Mobile Congress (GMC), Shanghai, China, 18–19 October 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 1–6. [Google Scholar]
- Ekiz, N.; Salih, T.; Küçüköner, S.; Fidanboylu, K. An overview of handoff techniques in cellular networks. Database 2005, 318, 6069. [Google Scholar]
- Zhou, Y.; Ye, Q.; Lv, J. Communication-efficient federated learning with compensated Overlap-FedAvg. IEEE Trans. Parallel Distrib. Syst. 2022, 33, 192–205. [Google Scholar] [CrossRef]
- Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
- Xie, S.; Girshick, R.; Dollar, P.; Tu, Z.; He, K. Aggregated residual transformations for deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1492–1500. [Google Scholar]
- Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
- Vieira, E.; Almeida, J.; Ferreira, J.; Dias, T.; Vieira Silva, A.; Moura, L. A roadside and cloud-based vehicular communications framework for the provision of C-ITS services. Information 2023, 14, 153. [Google Scholar] [CrossRef]
- Tran, T.X.; Pompili, D. Joint task offloading and resource allocation for multi-server mobile-edge computing networks. IEEE Trans. Veh. Technol. 2018, 68, 856–868. [Google Scholar] [CrossRef]
- Karimireddy, S.P.; Jaggi, M.; Kale, S.; Mohri, M.; Reddi, S.; Stich, S.U.; Suresh, A.T. Breaking the centralized barrier for cross-device federated learning. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Red Hook, NY, USA, 2021; Volume 34. [Google Scholar]
- Li, X.; Huang, K.; Yang, W.; Wang, S.; Zhang, Z. On the convergence of FedAvg on non-IID data. In Proceedings of the International Conference on Learning Representations (ICLR 2019), New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
- Karimireddy, S.P.; Kale, S.; Mohri, M.; Reddi, S.; Stich, S.U.; Suresh, A.T. SCAFFOLD: Stochastic controlled averaging for federated learning. In Proceedings of the 37th International Conference on Machine Learning (ICML 2020), Virtual, 13–18 July 2020; pp. 5132–5143. [Google Scholar]
- 3GPP. NR. Radio Resource Control (RRC); Protocol specification (Release 15), 3GPP TS 38.331 V15.6.0; 3GPP: Sophia Antipolis, France, 2020; Available online: https://www.3gpp.org/ftp/Specs/archive/38_series/38.331/ (accessed on 15 May 2025).
- Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Chintala, S. PyTorch: An imperative style, high-performance deep learning library. In Advances in Neural Information Processing Systems 32 (NeurIPS ’19); Curran Associates: Red Hook, NY, USA, 2019; pp. 8024–8035. [Google Scholar]
- Ashwin, R.J. Federated Learning PyTorch, GitHub Repository, Commit 26eaec40fa8beb5A6777feb89756f6401c28c4736. 2021. Available online: https://github.com/AshwinRJ/Federated-Learning PyTorch/tree/26eaec40fa8beb5A6777feb89756f6401c28c4736 (accessed on 23 May 2025).
- Stallkamp, J.; Schlipsing, M.; Salmen, J.; Igel, C. Man vs. computer: Benchmarking machine learning algorithms for traffic sign recognition. Neural Netw. 2012, 32, 323–332. [Google Scholar] [CrossRef] [PubMed]
- Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
- Reisizadeh, A.; Mokhtari, A.; Hassani, H.; Jadbabaie, A.; Pedarsani, R. FedPAQ: A communication-efficient federated learning method with periodic averaging and quantization. In Proceedings of the International Conference on Artificial Intelligence and Statistics (AISTATS 2020), Vienna, Austria, 26–28 August 2020; pp. 2021–2031. [Google Scholar]
Method | Latency | Handover | Pre-Aggregation Evaluation | Server-Level Aggregation | Strengths | Weaknesses |
---|---|---|---|---|---|---|
Single-server FL (Vanilla FL) [6] | Edge-only | No | No | No | Lower communication overhead compared to cloud-based. | Poor performance under non-IID data; no inter-server collaboration. |
Cloud-based FL (Hierarchical) [10] | Edge + Cloud | No | No | Yes | Global aggregation improves IID performance | High end-to-end latency. |
FedMes [18] | Edge-only | No | No | No | Exploits overlapping coverage, faster convergence due to aggregation at clients. | Depends on clients in overlap regions for convergence—insufficient overlap can degrade performance. |
Trust-based [21] | Edge-only | No | Yes | No | Considers clients’ preferences to select the next model. | Requires feedback from clients, high communication overhead. |
MS-FL (ours) | Edge-only | Yes | Yes | Yes | Quality-aware model aggregation among servers, continuous training for moving vehicles. | Slightly higher compute per round for evaluation steps. |
Proposed Aggregation Method | Core Mechanism | Strengths | Weaknesses |
---|---|---|---|
BA | Evaluates all peer models on the local validation set and selects the single best-performing model to multicast. | No need for client feedback or fine-grained weighting. | Discards all but one update—loses potentially useful information. Can be brittle if the “best” model is an outlier. |
SA | Iteratively incorporates each peer model: computes a weighted average with the current model, accepts the update only if it improves validation accuracy, then moves on to the next peer. | Guards against degradations—only beneficial updates are adopted. | Order-dependent: later peers may never be applied if an early update saturates performance. Can be slow, since each candidate requires a full validation pass. |
DWAA | Assigns each peer model a weight proportional to its validation accuracy and performs a single weighted average over all peer updates. | Utilizes information from every peer—more robust than picking just one. | No explicit outlier filtering—poor or malicious models can still influence the result. |
SPAA | Extends DWAA by computing a z-score on validation loss (or accuracy), converting large deviations into a sigmoid “penalty,” and then weighting each model. | Guards against outliers or malicious updates. Yields faster convergence, especially when model quality varies widely. | In highly non-IID settings, legitimately diverse models may be over-penalized. Slight overhead from computing z-scores and penalties. |
Differed Aggregation Methods at Server Level | Performance Evaluation Parameters | ||||
---|---|---|---|---|---|
Recall | Test Accuracy | Precision | F1-Score | Total Run Time (ms) | |
SPAA | 75% | 84% | 78% | 76% | 36,211 |
DWAA | 74% | 83% | 77% | 75% | 36,211 |
SA | 74% | 82.5% | 75.5% | 74.3% | 36,175 |
BA | 72.7% | 82.9% | 73.5% | 71.5% | 36,148 |
WA | 49% | 62.6% | 38% | 41% | 28,042 |
Different FL Frameworks | Performance Evaluation Parameters | |||
---|---|---|---|---|
Test Accuracy | Precision | Recall | F1-Score | |
MS-FL DWAA | 0.732858 | 0.623197 | 0.622598 | 0.604465 |
Cloud-based FL | 0.729137 | 0.600695 | 0.62328 | 0.595817 |
Single-server FL | 0.658802 | 0.549469 | 0.560228 | 0.523089 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Mazloomi, F.; Shah Heydari, S.; El-Khatib, K. A Novel Multi-Server Federated Learning Framework in Vehicular Edge Computing. Future Internet 2025, 17, 315. https://doi.org/10.3390/fi17070315
Mazloomi F, Shah Heydari S, El-Khatib K. A Novel Multi-Server Federated Learning Framework in Vehicular Edge Computing. Future Internet. 2025; 17(7):315. https://doi.org/10.3390/fi17070315
Chicago/Turabian StyleMazloomi, Fateme, Shahram Shah Heydari, and Khalil El-Khatib. 2025. "A Novel Multi-Server Federated Learning Framework in Vehicular Edge Computing" Future Internet 17, no. 7: 315. https://doi.org/10.3390/fi17070315
APA StyleMazloomi, F., Shah Heydari, S., & El-Khatib, K. (2025). A Novel Multi-Server Federated Learning Framework in Vehicular Edge Computing. Future Internet, 17(7), 315. https://doi.org/10.3390/fi17070315