Unified Distributed Machine Learning for 6G Intelligent Transportation Systems: A Hierarchical Approach for Terrestrial and Non-Terrestrial Networks
Abstract
1. Introduction
1.1. Gaps in the Literature and Motivations
- 1.
- Under-utilization of multilayer NTN architectures: Many prior frameworks assume single-tier or purely terrestrial deployments and therefore do not exploit the potential benefits of coordinated processing across RSUs, HAPs, and satellites. This omission reduces opportunities for latency-aware placement, load balancing across tiers, and resilience in coverage-challenged regions—all critical for large-scale ITS.
- 2.
- Limited onboard and in-space data processing: Existing works rarely address efficient processing of EO data within non-terrestrial platforms or the trade-offs introduced by performing model tasks in space (e.g., at HAPs/LEO nodes). Without concrete strategies for in-space inference/aggregation, systems face increased communication cost or delayed decision-making for time-sensitive ITS applications.
- 3.
- Insufficient integration of DML paradigms for resource-constrained, heterogeneous clients: While FL, SL, and TL have been studied individually (and some combinations proposed), there is a lack of comprehensive frameworks that jointly leverage model partitioning, transfer initialization, and hierarchical aggregation to simultaneously address privacy, compute constraints, and converge quickly in highly heterogeneous ITS fleets.
Ref. | Contributions | Differences With Our Work |
---|---|---|
[6] | Multitask FL with hybrid client selection/aggregation in edge networks. | Pure FL (no SL/TL); no multilayer hierarchy or NTN integration; not tailored to ITS. |
[8] | Evaluates and optimizes DML techniques for IoT; benchmarking focus. | Not ITS-centric; not hierarchical FL/SL over T/NTNs nor real-time ITS latency study. |
[9] | Wireless Distributed Learning: hybrid split+federated approach. | No T/NTN tiering, no EO/ITS integration, and no multilayer ITS latency/accuracy study. |
[11] | Federated Transfer Learning for cross-domain sensing with resource efficiency. | No ITS; no SL or multilayer NTN; narrower modality than our EO/vehicular perception. |
[15] | 6G-enabled advanced transportation systems; network capabilities and use cases. | No DML pipeline or a multilayer T/NTN latency/accuracy quantification. |
[16] | Explores the potential of NTNs in next-gen ITS; positioning paper. | Lacks concrete DML training/aggregation across layers and empirical ITS evaluation. |
[17] | Personalized FSL for ITS-distributed training: single-layer training. | No hierarchy, no EO integration, and no unified FL/SL/TL orchestration across T/NTNs. |
[18] | Feasibility of SL/TL/FL for security in ITS; SL outperforms FL/TL baselines; security focus. | No ITS; no hierarchical T/NTN training or EO-ITS fusion; no multi-tier aggregation. |
[19] | FSL with data/label privacy in vehicular networks. | No multilayer architecture and no joint FL+SL+TL orchestration. |
1.2. Key Contributions
- 1.
- New hybrid learning frameworks (FSTL and GFSTL): We propose FSTL and its generalized version (GFSTL). FSTL integrates FL, SL, and TL into a single pipeline tailored for resource-constrained edge nodes. GFSTL extends FSTL by supporting multiple independent FL/SL server placements and hierarchical aggregation across Road-Side Units (RSUs) and HAPs, enabling flexible deployment across Terrestrial and Non-Terrestrial Network layers (see Section 3 and Section 4).
- 2.
- Exploitation of multilayer T/NTN architectures for scalable DML: We explicitly design the frameworks to exploit the multilayer structure of NTNs (RSUs, UAVs, HAPs, Low/Medium Earth Orbit (LEO/MEO) satellites) so that computation and aggregation are performed at the most appropriate network tier (see the multilayer architectures in Section 5). This hierarchical design reduces wall-clock training time, improves scalability, and increases the number of feasible participants in federated training (see Section 6 and the latency analysis in Section 5.4).
- 3.
- Efficient Data Processing for Real-Time EO/ITS Applications: The paper proposes novel strategies for real-time data processing in space to optimize the utilization of EO/ITS data. By leveraging GFSTL’s flexible architecture, we address the inefficiencies in current EO/ITS systems, enabling more effective data transmission and model updates, even in highly dynamic environments with limited resources (Section 5).
- 4.
- Application to Diverse Use Cases and Comprehensive Evaluation: The proposed methodologies, system architecture, training process, added benefits, and challenges are introduced and evaluated across three distinct use cases: vehicular scenarios (Section 5.1), EO (Section 5.2), and their integration into a unified ITS scenario (Section 5.3). This work also provides a comprehensive latency analysis for a multilayer 6G ITS environment (Section 5.4), showcasing the practical benefits of our approach in large-scale, real-world deployments.
- 5.
- Enhanced Latency, Accuracy, and Privacy in DML: Our proposed methods significantly reduce latency and enhance training accuracy compared to traditional FL-based frameworks. The introduction of flexible server configurations in GFSTL further improves the number of participants in the FL process while ensuring robust privacy protection. These advances are demonstrated through simulations in typical 6G ITS scenarios, where the proposed methodologies outperform traditional FL techniques in terms of convergence rates, model accuracy (Section 6.2 and Section 6.3), and overall system latency (Section 6.4).
- 6.
- Practical split/transfer parameterization for ITS and EO tasks: We provide a concrete parameterization (model cut-points, smashed-representation sizing, and YOLO/ResNet choices) that maps model components to node capabilities (RSU/HAP/edge) (Section 6.1), balancing accuracy, computation, and communication. This parameterization is validated empirically in Section 6 and supports real-time ITS perception tasks using compact intermediate representations.
- 7.
- Comprehensive performance and latency study under unified experimental settings: We evaluate FL, SL, FSL, FSTL, and GFSTL on a representative ITS scenario using the same hardware/network assumptions to ensure a fair comparison. This study quantifies trade-offs in accuracy, convergence speed, per-round latency, and communication volume and demonstrates our proposed frameworks’ improvements in these dimensions (results and discussion in Section 6 and Section 7).
1.3. Organization of the Paper
2. Distributed Machine Learning Frameworks
2.1. Federated Learning
- Data Privacy: FL allows users to keep their raw data local, making it ideal for privacy-sensitive scenarios.
- Reduced Communication Overhead: Only the model parameters are exchanged, not the raw data, saving bandwidth.
- Adaptability: FL can continuously improve models with real-time data from users, making it responsive to dynamic environments.
- Communication Latency: Multiple rounds of communication are required, increasing the overall latency, especially in bandwidth-constrained networks.
- Convergence Speed: Achieving convergence may require many iterations, resulting in high computational and communication costs for resource-constrained users.
2.2. Split Learning
- 1.
- Each user processes its local data using the client-side model , which generates intermediate activations:These intermediate activations represent compressed feature representations and are transmitted to the server.
- 2.
- Upon receiving , the server performs the forward pass using its merge model to produce predictions :This completes the forward pass for the entire model using both client- and server-side components.
- 3.
- Next, the server begins the backward pass on the server-side model to compute the gradients of its model parameters based on the loss function, typically using a method such as stochastic gradient descent. The server then calculates the gradient of the intermediate activations , denoted as , and transmits this gradient back to the user.
- 4.
- Finally, the user completes the backward pass locally on its split model using . This enables each client to update their local model parameters for based on its unique data, thus maintaining data privacy and security.
- Reduced Client-Side Load: Clients only need to compute part of the model, reducing their computational overhead.
- Privacy Preservation: Only smashed data (intermediate activations) is transmitted, protecting the privacy of raw data.
- Lower Communication Cost: Fewer data need to be exchanged between clients and the server, as only intermediate activations are shared.
- Sequential Processing: The split model introduces latency because clients and the server must alternate between steps.
- Reduced Model Expressiveness: Splitting the model might reduce its capacity, potentially leading to performance degradation.
2.3. Federated Split Learning
- Hybrid Approach: FSL benefits from both FL and SL, enabling DML with lower communication costs and enhanced privacy.
- Scalability: The federated averaging mechanism allows the framework to scale effectively across many clients without increasing the communication burden.
- Reduced Client-Side Load: Like SL, FSL reduces the computational burden on clients by only requiring them to compute the lower layers of the model.
- Slow Convergence: The nature of SL can result in slower convergence, particularly in large-scale networks.
- High Local Training Requirements: Clients still need sufficient local computational resources to handle their portion of the model, which may not always be feasible in resource-constrained environments.
2.4. Transfer Learning
- Accelerated Training: TL reduces the training time required for model convergence, which is particularly advantageous in real-time or resource-constrained applications.
- Improved Model Accuracy: Leveraging learned knowledge from a related domain improves model performance, especially when training data is limited or heterogeneous.
- Enhanced Generalization: Pre-trained models offer better adaptability across diverse client datasets, as they can transfer features to new tasks effectively.
3. Federated Split Transfer Learning
3.1. FSTL Architecture
- 1.
- is the portion up to layer k, which is deployed on each user for local processing, and
- 2.
- represents the remaining layers, which are managed by the server in the SL paradigm.
- Higher Convergence Rate: By enabling parallel, simultaneous updates from all clients, the model can learn from a diverse set of data within each training round. This avoids the slow, one-by-one sequential processing inherent in standard SL, leading to a more efficient path to model convergence.
- Reduced Communication Bottlenecks: In a sequential learning setup, the server can only communicate with one client at a time, creating a significant bottleneck that scales with the number of users. The FSTL structure removes this limitation by allowing the central server to handle communications with all participating clients in parallel, thereby improving network efficiency.
- Balanced Computational Loads: The framework design inherently distributes the workload. While each client handles the initial, less intensive part of the model, the server manages the more computationally demanding tasks and the aggregation of updates from all clients. This parallel structure ensures a more balanced and efficient use of computational resources throughout the network.
3.2. FSTL Training Process
- 1.
- Initialization:
- (a)
- The global model parameters are initialized using pre-trained weights from a TL model.
- (b)
- A specific layer index k is chosen, determining the split between the client and server computations. The layers before k are hosted on the client side, while the remaining layers are hosted on the server.
- 2.
- Iterative Training: For each training iteration , the following steps occur:
- (a)
- The initial global model parameters are distributed to all participating users.
- (b)
- Each user i processes its local data through its client-side split model to generate intermediate representations:
- (c)
- The user sends these intermediate representations to the central server.
- (d)
- The central server aggregates the intermediate representations from all users by applying the following merge function :
- (e)
- Using , the server performs a forward–backward pass to compute an updated version of the global model parameters:Here, L denotes the loss function, while is the learning rate. This forward–backward process, which leverages federated averaging, is conducted in a way that preserves privacy, since only aggregated updates are calculated.
- (f)
- The central server updates by setting
- (g)
- Each user i then receives the updated global model parameters from the server and uses them to update their own server-side model parameters as follows:
- (h)
- Finally, each user i updates its client-side model parameters as follows:
3.3. FSTL Advantages
- 1.
- Data Privacy and Security: With the retention of raw data on the user’s side, FSTL guarantees that sensitive personal information is not exposed or communicated outward. This addresses the fundamental privacy issues associated with collaborative learning frameworks.
- 2.
- Efficient Communication: The transmission involves only the intermediate data representations between clients and the central server, thus significantly reducing communication overhead in the process.
- 3.
- Accelerated Convergence: By employing models that have been pre-trained, convergence is more rapidly achieved since these models utilize existing knowledge instead of starting the learning process from the beginning.
- 4.
- Improved Performance: The server’s collaborative learning approach fosters improvements in the global model across all users, facilitating the adaptability of the framework to ever-changing environments and diverse ITS scenarios.
3.4. FSTL in ITS Scenarios
3.5. Benefits of FSTL in ITSs
3.6. FSTL Training Process
Algorithm 1 FSTL iterative algorithm for ITSs |
Input: |
Output: Updated local and server model parameters and |
|
4. Generalized Federated Split Transfer Learning
4.1. GFSTL Architecture
4.2. GFSTL Training Process
Algorithm 2 GFSTL iterative algorithm |
Input: |
Output: |
|
4.3. GFSTL Advantages
- 1.
- Scalability: By distributing clients into multiple groups, the architecture significantly reduces the computational load on any single server. This group-based structure enables scalable training across large numbers of users, each contributing to the global model.
- 2.
- Privacy Preservation: GFSTL ensures that raw data never leave the local client devices. Only smashed data, which contain no identifiable information, are transmitted to the SL server, thus preserving the privacy of sensitive user data.
- 3.
- Efficient Resource Utilization: The parallelization of groups allows efficient use of computational resources. The training process within each group is sequential, but the groups themselves operate in parallel, which reduces the overall training time and improves performance in large-scale systems.
- 4.
- Improved Model Accuracy: By enabling the use of TL with pre-trained models, GFSTL accelerates the convergence of the global model. Moreover, the aggregation of submodels across groups ensures that the global model is more robust, with improved accuracy due to contributions from diverse groups of users.
- 5.
- Low Communication Overhead: Since only model parameters are exchanged between users and SL servers, the communication overhead is minimal. This is particularly advantageous in NTNs, where communication links can be intermittent or costly.
4.4. Summary of Advantages and Disadvantages of DML Approaches
5. Application of GFSTL in 6G
5.1. Use Case 1: GFSTL in Vehicular Aerial–Ground Integrated Network
5.1.1. System Architecture and Methodology
5.1.2. Advantages of Integrating GFSTL with Aerial Networks
5.1.3. GFSTL Workflow in Vehicular Scenarios
- 1.
- The RSUs first act as local FSTL servers, managing the training and model aggregation for their respective groups of VUs.
- 2.
- Each VU performs a local computation and forwards its model updates to its corresponding RSU, which aggregates these updates.
- 3.
- Once the FSTL process is completed for all groups of VUs managed by the RSUs, the fully trained FSTL models are transmitted from the RSUs to the HAP.
- 4.
- The HAP, which functions as the main FL server, aggregates the complete set of model parameters from all RSUs, thereby finalizing the global model.
5.2. Use Case 2: GFSTL in NTN-Based EO
5.2.1. Motivation and Proposed Framework
5.2.2. Framework Components and Advantages
- 1.
- Increased Accuracy: UAVs can focus on capturing precise data in specific regions, especially in areas where satellite imagery might be less effective (e.g., dense forests). This localized data collection ensures higher accuracy and relevance of the EO data.
- 2.
- Resource Efficiency: HAPs serve as intermediaries between UAVs and LEO satellites, efficiently distributing computational tasks and enabling seamless data processing across layers.
- 3.
- Privacy and Security: The FSL approach ensures that raw data remains on local devices (UAVs), preserving privacy while transmitting only intermediate data (smashed data) to HAP servers. This minimizes the risk of data breaches and ensures compliance with privacy regulations.
- 4.
- Reduced Latency: By using TL to initialize the training process with pre-trained models, the framework reduces the total training time, allowing faster convergence of the global model.
5.2.3. GFSTL Training Process for EO over NTNs
5.2.4. Added Benefits and Challenges
5.3. Unified Multilayer 6G ITS Architecture
5.3.1. Architecture and Components
- 1.
- Users (VUs, UAVs, and Cameras): VUs are vehicles equipped with various sensors and communication modules that gather real-time traffic and environmental data. UAVs provide additional data, particularly in areas that are difficult to monitor or reach from the ground, providing support in vehicle navigation, surveillance, and emergency responses. Street cameras positioned at critical points such as intersections and crossroads monitor traffic flow, detect accidents, and feed real-time data into the network.
- 2.
- FSL Servers (RSUs and Base Stations (BSs)): RSUs and BSs act as FSL servers. They aggregate and process data received from nearby VUs, UAVs, and cameras, performing preliminary computations.
- 3.
- Main FL Server (HAP): The HAP serves as the main FL server. It receives and aggregates the models trained by the FSL servers (RSUs/BSs), conducting the final stage of training, which is aggregation to a global model that integrates data from all layers of the ITS.
5.3.2. Training Process and Iterative Algorithm
- 1.
- Initialization: Each client (VUs, UAVs, and cameras) is initialized with a pre-trained model that is specific to its task, with parameters denoted as for the client model and for the server model.
- 2.
- Local Model Update: Clients perform forward propagation on their local models using the collected data, producing intermediate results (smashed data) that are sent to the nearest FSL server (RSU or BS).
- 3.
- FSL Server Aggregation: RSUs and BSs aggregate the intermediate results, perform forward and backward propagation on their server models, and send the corresponding gradients back to the clients to update their local models.
- 4.
- Main FL Server Aggregation: Once all FSL servers complete their local training, they send the updated models to the main FL server (HAP), aggregating these models into a unified global model.
- 5.
- Global Model Update: The global model is redistributed to the clients through the FSL servers for the next training iteration.
5.3.3. Added Benefits and Challenges
- Scalability: The multilayer GFSTL is inherently scalable, allowing it to accommodate the increasing number of VUs, UAVs, and cameras as the ITS infrastructure grows. Its distributed nature ensures that the computational load is balanced throughout the system.
- Privacy Preservation: Since raw data remain on the client side (VUs, UAVs, and cameras), the model leverages FSL to ensure privacy-preserving model training, which is crucial for sensitive traffic data and user privacy.
- Reduced Latency: By distributing computational tasks across FSL servers (RSUs/BSs) and the main FL server (HAP), GFSTL minimizes communication overhead and reduces the time needed to update the global model.
- Resilience: The integration of multiple hierarchical layers ensures redundancy, making the system more resilient to failures in individual components, such as RSUs or UAVs.
- Resource Management: Efficiently managing network and computational resources across VUs, UAVs, cameras, RSUs, and the HAP is critical. Bandwidth allocation, processing power, and memory management are areas that need careful optimization.
- Data Heterogeneity: Data collected from various components (VUs, UAVs, and cameras) may differ in format and quality. Ensuring that the global model can generalize across heterogeneous data sources is a challenging task.
- Model Convergence: Although TL and SL help to accelerate the training process, achieving convergence in a system as complex as the 6G ITS requires fine-tuning of hyperparameters, especially with regard to the learning rate and aggregation intervals.
5.4. Latency Analysis for a Multilayer 6G ITS Scenario
6. Simulations and Performance Evaluations
6.1. Simulation, Hardware, and Network Parameters
6.2. Model Convergence and Accuracy Versus Rounds
6.3. User Diversity and Model Accuracy
6.4. Latency and Communication Efficiency
- GFSTL (purple curve) is effectively flat and lowest at scale. GFSTL exhibits a nearly constant latency (≈14–22 s across the sweep) and shows only a minor increase even at 10 batches. This behavior follows directly from the hierarchical, parallel aggregation used by GFSTL: RSUs aggregate their connected clients locally in parallel, and only compressed intermediate representations (the smashed tensors) are exchanged upward. In the latency equation, this reduces per client and keeps dominated by the slowest RSU rather than the sum of all clients, therefore avoiding the explosive growth seen in serial schemes.
- FSTL (red curve) shows moderate growth but remains substantially lower than FL/FSL/SL at high loads. FSTL starts around 21 s and rises to ≈31 s at 10 batches. Compared to GFSTL, FSTL lacks the same level of hierarchical parallelism or dynamic placement of FL/SL servers, so its and terms increase more with the user count. However, because FSTL still transmits compact representations rather than entire models, it avoids the large spikes observed for full-FL-style approaches.
- FL and FSL (blue and green curves) remain low on a small scale but spike dramatically at 10 batches. Both FL and FSL are roughly in the 12–22 s range for small-to-moderate batch counts, then jump to ≈61 s (FL) and ≈71 s (FSL) at 10 batches. This non-linear escalation is explained by two mechanisms in the latency model: (i) the upload/download of full model parameters or large gradient vectors causes to grow with the number of clients, and (ii) global aggregation in FL requires communication with a centralized server (HAP or cloud) that becomes a bottleneck when many clients simultaneously upload large models. In FSL, the extra split/label-handling complexity can further amplify communication and aggregation overhead, hence the higher spike compared to plain FL.
- SL (orange curve) is highest overall and grows rapidly at large scale. SL demonstrates high latency even at small batch counts (≈33 s) and reaches ≈80 s at 10 batches. This is expected because classic Split Learning operates in a sequential (or partly sequential) manner where server-side processing often waits for client-by-client smashed uploads and sequential forward/backward passes. In terms of latency decomposition, SL’s effectively becomes a sum of per-client server processing times, causing linear (or worse) scaling with the number of clients.
6.5. Summary of Key Results
7. Discussion, Limitations, and Future Directions
8. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Tang, F.; Kawamoto, Y.; Kato, N.; Liu, J. Future Intelligent and Secure Vehicular Network Toward 6G: Machine-Learning Approaches. Proc. IEEE 2020, 108, 292–307. [Google Scholar] [CrossRef]
- Brincat, A.A.; Pacifici, F.; Martinaglia, S.; Mazzola, F. The Internet of Things for Intelligent Transportation Systems in Real Smart Cities Scenarios. In Proceedings of the 2019 IEEE 5th World Forum on Internet of Things (WF-IoT), Limerick, Ireland, 15–18 April 2019; pp. 128–132. [Google Scholar] [CrossRef]
- Gubbi, J.; Buyya, R.; Marusic, S.; Palaniswami, M. Internet of Things (IoT): A vision, architectural elements, and future directions. Future Gener. Comput. Syst. 2013, 29, 1645–1660. [Google Scholar] [CrossRef]
- Darwish, T.S.J.; Abu Bakar, K. Fog Based Intelligent Transportation Big Data Analytics in The Internet of Vehicles Environment: Motivations, Architecture, Challenges, and Critical Issues. IEEE Access 2018, 6, 15679–15701. [Google Scholar] [CrossRef]
- Yurdem, B.; Kuzlu, M.; Gullu, M.K.; Catak, F.O.; Tabassum, M. Federated learning: Overview, strategies, applications, tools and future directions. Heliyon 2024, 10, e38137. [Google Scholar] [CrossRef]
- Hamood, M.; Albaseer, A.; Abdallah, M.; Al-Fuqaha, A.; Mohamed, A. Optimized Federated Multitask Learning in Mobile Edge Networks: A Hybrid Client Selection and Model Aggregation Approach. IEEE Trans. Veh. Technol. 2024, 73, 17613–17629. [Google Scholar] [CrossRef]
- Chen, H.; Zhu, T.; Zhang, T.; Zhou, W.; Yu, P.S. Privacy and Fairness in Federated Learning: On the Perspective of Tradeoff. ACM Comput. Surv. 2023, 56, 1–37. [Google Scholar] [CrossRef]
- Gao, Y.; Kim, M.; Thapa, C.; Abuadbba, A.; Zhang, Z.; Camtepe, S.; Kim, H.; Nepal, S. Evaluation and Optimization of Distributed Machine Learning Techniques for Internet of Things. IEEE Trans. Comput. 2022, 71, 2538–2552. [Google Scholar] [CrossRef]
- Liu, X.; Deng, Y.; Mahmoodi, T. Wireless Distributed Learning: A New Hybrid Split and Federated Learning Approach. IEEE Trans. Wirel. Commun. 2023, 22, 2650–2665. [Google Scholar] [CrossRef]
- Vilalta, R.; Meskhi, M.M. Transfer of Knowledge Across Tasks. In Metalearning: Applications to Automated Machine Learning and Data Mining; Springer International Publishing: Cham, Switzerland, 2022; pp. 219–236. [Google Scholar] [CrossRef]
- Qi, W.; Zhang, R.; Zhou, J.; Zhang, H.; Xie, Y.; Jing, X. A Resource-Efficient Cross-Domain Sensing Method for Device-Free Gesture Recognition With Federated Transfer Learning. Trans. Green Commun. Netw. 2023, 7, 393–400. [Google Scholar] [CrossRef]
- Leo, M.D.; Minghini, M.; Kóňa, A.; Kotsev, A.; Dusart, J.; Lumnitz, S.; Ilie, C.M.; Kilsedar, C.E.; Tzotsos, A. Digital earth observation infrastructures and initiatives: A review framework based on open principles. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2023, 48, 33–40. [Google Scholar] [CrossRef]
- Salcedo-Sanz, S.; Ghamisi, P.; Piles, M.; Werner, M.; Cuadra, L.; Moreno-Martínez, A.; Izquierdo-Verdiguier, E.; Muñoz-Marí, J.; Mosavi, A.; Camps-Valls, G. Machine learning information fusion in Earth observation: A comprehensive review of methods, applications and data sources. Inf. Fusion 2020, 63, 256–272. [Google Scholar] [CrossRef]
- Blasch, E.; Pham, T.; Chong, C.Y.; Koch, W.; Leung, H.; Braines, D.; Abdelzaher, T. Machine Learning/Artificial Intelligence for Sensor Data Fusion–Opportunities and Challenges. IEEE Aerosp. Electron. Syst. Mag. 2021, 36, 80–93. [Google Scholar] [CrossRef]
- Liu, R.; Hua, M.; Guan, K.; Wang, X.; Zhang, L.; Mao, T.; Zhang, D.; Wu, Q.; Jamalipour, A. 6G Enabled Advanced Transportation Systems. IEEE Trans. Intell. Transp. Syst. 2024, 25, 10564–10580. [Google Scholar] [CrossRef]
- Djihane, M.; Abdelkrim, H.; Mohamed, B. Exploring the potential of Non Terrestrial Networks in next generation Intelligent Transport Systems. In Proceedings of the 2024 8th International Conference on Image and Signal Processing and Their Applications (ISPA), Biskra, Algeria, 21–22 April 2024; pp. 1–7. [Google Scholar] [CrossRef]
- Dai, C.; Zhu, T.; Xiang, S.; Xie, L.; Garg, S.; Hossain, M.S. PSFL: Personalized Split Federated Learning Framework for Distributed Model Training in Intelligent Transportation Systems. IEEE Trans. Intell. Transp. Syst. 2025, 1–10, early access. [Google Scholar] [CrossRef]
- Otoum, S.; Guizani, N.; Mouftah, H. On the Feasibility of Split Learning, Transfer Learning and Federated Learning for Preserving Security in ITS Systems. IEEE Trans. Intell. Transp. Syst. 2023, 24, 7462–7470. [Google Scholar] [CrossRef]
- Wu, M.; Cheng, G.; Ye, D.; Kang, J.; Yu, R.; Wu, Y.; Pan, M. Federated Split Learning With Data and Label Privacy Preservation in Vehicular Networks. IEEE Trans. Veh. Technol. 2024, 73, 1223–1238. [Google Scholar] [CrossRef]
- Naseh, D.; Shinde, S.S.; Tarchi, D. Enabling Intelligent Vehicular Networks Through Distributed Learning in the Non-Terrestrial Networks 6G Vision. In Proceedings of the European Wireless 2023 (EW 2023), Rome, Italy, 2–4 October 2023. [Google Scholar] [CrossRef]
- Kosasi, T.; Sihombing, Z.A.L.; Husein, A.M. YOLO-Based Vehicle Detection: Literature Review. J. Comput. Networks, Archit. High Perform. Comput. 2024, 6, 1384–1389. [Google Scholar] [CrossRef]
- Naseh, D.; Shinde, S.S.; Tarchi, D. Network Sliced Distributed Learning-as-a-Service for Internet of Vehicles Applications in 6G Non-Terrestrial Network Scenarios. J. Sens. Actuator Netw. 2024, 13, 14. [Google Scholar] [CrossRef]
- Ridolfi, L.; Naseh, D.; Shinde, S.S.; Tarchi, D. Implementation and Evaluation of a Federated Learning Framework on Raspberry PI Platforms for IoT 6G Applications. Future Internet 2023, 15, 358. [Google Scholar] [CrossRef]
- Qi, P.; Chiaro, D.; Guzzo, A.; Ianni, M.; Fortino, G.; Piccialli, F. Model aggregation techniques in federated learning: A comprehensive survey. Future Gener. Comput. Syst. 2024, 150, 272–293. [Google Scholar] [CrossRef]
- Alsharif, M.H.; Kannadasan, R.; Wei, W.; Nisar, K.S.; Abdel-Aty, A.H. A contemporary survey of recent advances in federated learning: Taxonomies, applications, and challenges. Internet Things 2024, 27, 101251. [Google Scholar] [CrossRef]
- Jin, Y.; Zhu, H.; Xu, J.; Chen, Y. Federated Learning: Fundamentals and Advances; Machine Learning: Foundations, Methodologies, and Applications; Springer: Singapore, 2023. [Google Scholar] [CrossRef]
- Nguyen, D.C.; Ding, M.; Pathirana, P.N.; Seneviratne, A.; Li, J.; Vincent Poor, H. Federated Learning for Internet of Things: A Comprehensive Survey. IEEE Commun. Surv. Tutor. 2021, 23, 1622–1658. [Google Scholar] [CrossRef]
- Lin, Z.; Qu, G.; Chen, X.; Huang, K. Split Learning in 6G Edge Networks. IEEE Wirel. Commun. 2024, 31, 170–176. [Google Scholar] [CrossRef]
- Wu, W.; Li, M.; Qu, K.; Zhou, C.; Shen, X.; Zhuang, W.; Li, X.; Shi, W. Split Learning Over Wireless Networks: Parallel Design and Resource Management. IEEE J. Sel. Areas Commun. 2023, 41, 1051–1066. [Google Scholar] [CrossRef]
- Vepakomma, P.; Raskar, R. Split Learning: A Resource Efficient Model and Data Parallel Approach for Distributed Deep Learning. In Federated Learning: A Comprehensive Overview of Methods and Applications; Ludwig, H., Baracaldo, N., Eds.; Springer International Publishing: Cham, Switzerland, 2022; pp. 439–451. [Google Scholar] [CrossRef]
- Turina, V.; Zhang, Z.; Esposito, F.; Matta, I. Federated or Split? A Performance and Privacy Analysis of Hybrid Split and Federated Learning Architectures. In Proceedings of the 2021 IEEE 14th International Conference on Cloud Computing (CLOUD), Chicago, IL, USA, 5–10 September 2021; pp. 250–260. [Google Scholar] [CrossRef]
- Liang, Y.; Kortoçi, P.; Zhou, P.; Lee, L.H.; Mehrabi, A.; Hui, P.; Tarkoma, S.; Crowcroft, J. Federated split GANs for collaborative training with heterogeneous devices. Softw. Impacts 2022, 14, 100436. [Google Scholar] [CrossRef]
- Naseh, D.; Abdollahpour, M.; Tarchi, D. Real-World Implementation and Performance Analysis of Distributed Learning Frameworks for 6G IoT Applications. Information 2024, 15, 190. [Google Scholar] [CrossRef]
- Zhao, Z.; Alzubaidi, L.; Zhang, J.; Duan, Y.; Gu, Y. A comparison review of transfer learning and self-supervised learning: Definitions, applications, advantages and limitations. Expert Syst. Appl. 2024, 242, 122807. [Google Scholar] [CrossRef]
- Naseh, D.; Shinde, S.S.; Tarchi, D. Multi-Layer Distributed Learning for Intelligent Transportation Systems in 6G Aerial-Ground Integrated Networks. In Proceedings of the 2024 Joint European Conference on Networks and Communications & 6G Summit (EuCNC/6G Summit), Antwerp, Belgium, 3–6 June 2024; pp. 711–716. [Google Scholar] [CrossRef]
- Naseh, D.; Shinde, S.S.; Tarchi, D. Distributed Learning Framework for Earth Observation on Multilayer Non-Terrestrial Networks. In Proceedings of the 2024 IEEE International Conference on Machine Learning for Communication and Networking (ICMLCN), Stockholm, Sweden, 5–8 May 2024; pp. 1–2. [Google Scholar] [CrossRef]
- Price, E. Federated Learning for Privacy-Preserving Edge Intelligence: A Scalable Systems Perspective. J. Comput. Sci. Softw. Appl. 2025, 5. [Google Scholar] [CrossRef]
- Jia, Y.; Liu, B.; Zhang, X.; Dai, F.; Khan, A.; Qi, L.; Dou, W. Model Pruning-enabled Federated Split Learning for Resource-constrained Devices in Artificial Intelligence Empowered Edge Computing Environment. ACM Trans. Sens. Netw. 2024. [Google Scholar] [CrossRef]
- Zhuang, F.; Qi, Z.; Duan, K.; Xi, D.; Zhu, Y.; Zhu, H.; Xiong, H.; He, Q. A Comprehensive Survey on Transfer Learning. Proc. IEEE 2021, 109, 43–76. [Google Scholar] [CrossRef]
- Niu, Z.; Shen, X.S.; Zhang, Q.; Tang, Y. Space-air-ground integrated vehicular network for connected and automated vehicles: Challenges and solutions. Intell. Converg. Netw. 2020, 1, 142–169. [Google Scholar] [CrossRef]
- Naseh, D.; Shinde, S.S.; Tarchi, D.; DeCola, T. Distributed Intelligent Framework for Remote Area Observation on Multilayer Non-Terrestrial Networks. In Proceedings of the 2024 IEEE International Mediterranean Conference on Communications and Networking (MeditCom), Madrid, Spain, 8–11 July 2024; pp. 1–6. [Google Scholar] [CrossRef]
- Jocher, G.; Chaurasia, A.; Stoken, A.; Borovec, J.; Kwon, Y.; Michael, K.; Fang, J.; Yifu, Z.; Wong, C.; Montes, D.; et al. ultralytics/yolov5: V7.0—YOLOv5 SOTA Realtime Instance Segmentation (v7.0). Zenodo 2022. [Google Scholar] [CrossRef]
- Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T.; Enzweiler, M.; Benenson, R.; Franke, U.; Roth, S.; Schiele, B. The Cityscapes Dataset for Semantic Urban Scene Understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 3213–3223. [Google Scholar] [CrossRef]
- Khanam, R.; Hussain, M. What is YOLOv5: A deep look into the internal features of the popular object detector. arXiv 2024, arXiv:2407.20892. [Google Scholar] [CrossRef]
- Desislavov, R.; Martínez-Plumed, F.; Hernández-Orallo, J. Trends in AI inference energy consumption: Beyond the performance-vs-parameter laws of deep learning. Sustain. Comput. Inform. Syst. 2023, 38, 100857. [Google Scholar] [CrossRef]
Technique | Scalability | Privacy | Computational Efficiency | Communication Overhead |
---|---|---|---|---|
FL | + High scalability due to parallel model training across clients. − Limited by full model update aggregation on the server. | + Data remains on local devices (enhances privacy). − Vulnerable to model inversion and poisoning attacks. | − High computation requirements due to full model training on each client. | − High communication overhead with full model parameter exchange. |
SL | + Suitable for scenarios with fewer clients. − Limited scalability due to serialized client training. | + Raw data remains private (smashed data shared only). − Some data leakage can occur via smashed data. | + Reduced computational burden on the client’s side. − Server-side computation burden remains high. | + Lower communication overhead (only smashed data is sent). − High latency due to serial model updates. |
FSL | + More scalable than SL with parallelized FL elements. − Limited scalability compared to pure FL. | + Enhanced privacy through combined FL and SL benefits. − Privacy concerns due to server–client split. | + Reduced client-side computational load. − Server-side computational load remains significant. | + Lower communication overhead compared to FL (smashed data only). − Higher latency than FL due to split training. |
TL | + Applicable for diverse domains; increases scalability in heterogeneous data settings. | + Pre-trained models can mask some sensitive information. | + Reduces computational load by leveraging pre-trained models. − Limited by domain compatibility of source and target tasks. | + Lower communication overhead due to pre-trained model usage. |
FSTL | + More scalable than FSL due to faster model convergence using TL. | + High data privacy with Transfer Learning masking and combined FL-SL protections. | + Enhanced efficiency by utilizing pre-trained model splits. | + Lower communication overhead due to smaller model splits. + Improved latency by combining parallel FL updates and TL. |
GFSTL | + Highly scalable due to the hierarchical multi-server structure with multiple client groups. | + High data privacy preservation via group separation and SL protection. | + Efficient model convergence with pre-trained models and reduced client computation. | + Reduced latency by parallel updates within groups and lower bandwidth requirements per user. |
Learning Method | Training + Aggregation Time | Communications per User/Server | Total Communications/Server | Total Communication Time | Total Latency |
---|---|---|---|---|---|
FL | |||||
SL | T | ||||
FSL | |||||
FSTL | |||||
GFSTL |
Parameter | Value/Type |
---|---|
YOLOv5 variant | YOLOv5-m (21.2 M params, 49 GFLOPs @ 640 × 640). |
Input image size | 640 × 640 px. |
Pre-trained weights | YOLOv5 (Ultralytics) pre-trained then fine-tuned on Cityscapes. |
Cityscapes images (fine annotations) | Train: 2975; Val: 500; Test: 1525. |
Per-batch composition | 5 VUs, 2 UAVs, and 3 static cameras (per batch). |
Max batches per RSU | 1–10 (experiments sweep). |
GPU | NVIDIA Tesla T4, with 16 GB of GDDR6. |
RSU compute | 1 × Tesla T4 (per RSU experiment). |
HAP compute (main FL server) | 2 × Tesla T4 (simulated higher-tier server). |
CPU (host) | Intel Xeon 8 cores @ 2.3 GHz (typical server host). |
Per-client memory (simulated) | 4 GB (typical VU/UAV/camera edge device budget). |
Batch size (per client) | 8 images. |
Optimizer | Adam, initial LR = 1 × 10−4, and weight decay = 1 × 10−4. |
Training epochs per local update | 1 (local)/global rounds up to 200 (early stopping). |
Communications link rate (R) | 100 Mbps uplink/downlink. |
Smashed representation (cut layer) | 256-dim float vector per image (256 × 4 bytes = 1.0 KB). |
Estimated smashed payload per batch | 1.0 KB × batch size (8) × number of clients per batch (10). |
Method | Final Accuracy | Rounds to Converge | Latency per Round | Comm. vol. per Round |
---|---|---|---|---|
FL | 94.2% | 5 | 420 ms | 4.5 MB |
SL | 97.6% | 10 | 1500 ms | 0.9 MB |
FSL | 94.5% | 8 | 380 ms | 0.8 MB |
FSTL | 94.9% | 3 | 220 ms | 0.8 MB |
GFSTL | 99.8% | 2 | 160 ms | 0.6 MB |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Naseh, D.; Bozorgchenani, A.; Shinde, S.S.; Tarchi, D. Unified Distributed Machine Learning for 6G Intelligent Transportation Systems: A Hierarchical Approach for Terrestrial and Non-Terrestrial Networks. Network 2025, 5, 41. https://doi.org/10.3390/network5030041
Naseh D, Bozorgchenani A, Shinde SS, Tarchi D. Unified Distributed Machine Learning for 6G Intelligent Transportation Systems: A Hierarchical Approach for Terrestrial and Non-Terrestrial Networks. Network. 2025; 5(3):41. https://doi.org/10.3390/network5030041
Chicago/Turabian StyleNaseh, David, Arash Bozorgchenani, Swapnil Sadashiv Shinde, and Daniele Tarchi. 2025. "Unified Distributed Machine Learning for 6G Intelligent Transportation Systems: A Hierarchical Approach for Terrestrial and Non-Terrestrial Networks" Network 5, no. 3: 41. https://doi.org/10.3390/network5030041
APA StyleNaseh, D., Bozorgchenani, A., Shinde, S. S., & Tarchi, D. (2025). Unified Distributed Machine Learning for 6G Intelligent Transportation Systems: A Hierarchical Approach for Terrestrial and Non-Terrestrial Networks. Network, 5(3), 41. https://doi.org/10.3390/network5030041