Next Article in Journal
Adaptive and Sustainable Smart Environments Using Predictive Reasoning and Context-Aware Reinforcement Learning
Next Article in Special Issue
DACCA: Distributed Adaptive Cloud Continuum Architecture
Previous Article in Journal / Special Issue
Multi-Attribute Physical-Layer Authentication Against Jamming and Battery-Depletion Attacks in LoRaWAN
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Architecting Multi-Cluster Layer-2 Connectivity for Cloud-Native Network Slicing

Department of Telematic Engineering, Escuela Politécnica Superior, Universidad Carlos III de Madrid (UC3M), Avda. de la Universidad, 30, 28911 Leganés, Spain
*
Author to whom correspondence should be addressed.
Future Internet 2026, 18(1), 39; https://doi.org/10.3390/fi18010039
Submission received: 20 November 2025 / Revised: 29 December 2025 / Accepted: 31 December 2025 / Published: 8 January 2026

Abstract

Connecting distributed applications across multiple cloud-native domains is growing in complexity. Applications have become containerized and fragmented across heterogeneous infrastructures, such as public clouds, edge nodes, and private data centers, including emerging IoT-driven environments. Existing networking solutions like CNI plugins and service meshes have proven insufficient for providing isolated, low-latency and secure multi-cluster communication. By combining SDN control with Kubernetes abstractions, we present L2S-CES, a Kubernetes-native solution for multi-cluster layer-2 network slicing that offers flexible isolated connectivity for microservices while maintaining performance and automation. In this work, we detail the design and implementation of L2S-CES, outlining its architecture and operational workflow. We experimentally validate against state-of-the-art alternatives and show superior isolation, reduced setup time, native support for broadcast and multicast, and minimal performance overhead. By addressing the current lack of native link-layer networking capabilities across multiple Kubernetes domains, L2S-CES provides a unified and practical foundation for deploying scalable, multi-tenant, and latency-sensitive cloud-native applications.

1. Introduction

From submarine cables laid across the ocean floor to satellite-based communication systems, the infrastructure underlying global Internet data flows is diverse and complex. As millions of users rely on these networks daily, the Internet faces increasing demands and intricate technical challenges, including network interoperability, routing within highly dynamic topologies, and ensuring security and resilience across administrative domains [1].
These challenges are largely addressed by virtualization, a technology that allows the partitioning of physical resources into multiple independent virtual environments [2,3]. More precisely, in data center communications, by abstracting hardware components, virtualization allows them to manage computing and networking resources more efficiently. It enables multiple users to share the same physical infrastructure while maintaining secure, isolated, and customizable environments.
Building on these virtualization paradigms, containerization is one technique that enables applications to be deployed in lightweight isolated computing environments, known as containers [4]. Modern software applications are increasingly decomposed into multiple containers [5] given that microservice architectures enable teams to be more productive, and improve the scalability and maintainability of services [6]. Those are reliant on cloud-native platforms as the elasticity and on-demand provisioning eases its implementation, so containerized microservices need automated management through orchestration. Container orchestration platforms, such as Kubernetes [7], have evolved well beyond their initial purpose, supporting diverse workloads and operational paradigms [8], while their networking subsystem, in particular, remains comparatively constrained. The abstractions that once sufficed for container orchestration now face limitations in scalability, flexibility, and integration with emerging network paradigms. Pods, which encapsulate containers, are the smallest units of computation in Kubernetes and lack network isolation between one another. Additionally, when exposing a pod to the Internet, tools for securing them rely on the application implementation, which leaves room for errors.
Network isolation can be achieved with layer-2 mechanisms using software-defined networking (SDN) and network function virtualization (NFV). SDN logically separates the control and data planes, exposing open southbound interfaces that allow operators to program forwarding behavior at runtime [9]. Complementarily, NFV relocates packet-processing functions—firewalls, load balancers, and virtual switches—onto commodity servers, decoupling them from proprietary middleboxes [10]. NFV therefore offers the elastic scaling and rapid instantiation needed when Kubernetes clusters burst or shrink. Combining SDN’s holistic visibility with NFV’s agile function placement yields a modular toolkit that (i) enforces network isolation end-to-end, (ii) minimizes hardware lock-in, and (iii) shortens service deployment cycles from weeks to seconds—all of which are indispensable for layer-2 virtual network setup.
Building on the capabilities of SDN and NFV, network slicing is the solution that enables the provision of service instances, network functions, and the necessary resources—aligned with the specific requirements of those services—within logically isolated segments of a shared infrastructure [11]. A slice is defined as the set of network functions and the resources to run these network functions, providing a logical infrastructure suited to the user’s needs. Each slice can be tailored to optimize characteristics such as bandwidth, latency, security, or reliability, effectively ensuring that different workloads and tenants can coexist without negatively impacting each other. The fine-grained control granted by SDN is essential for network slicing as each slice may demand distinct forwarding policies, quality of service (QoS) profiles, or traffic-engineering objectives that can change on short notice.
In this context, the primary goal of this paper is to provide a homogeneous platform that facilitates the creation of network slices which provision isolated communication and make it scalable and adaptable to current edge-cloud solutions, as depicted in Figure 1. By creating a multi-domain slice spanning multiple Kubernetes clusters, multiple virtual networks are formed, interconnecting containerized microservices across these clusters. This approach not only ensures isolated and secure communication channels but also facilitates observability and metric gathering. As a result, operators gain visibility into network performance, allowing them to monitor resource usage, detect anomalies, and more effectively manage network policies.
Despite the theoretical benefits of slicing, realizing strict isolation in cloud and edge platforms remains a challenge. While several solutions have been proposed, such as Kubeslice [12] and KubeOVN [13], they predominantly rely on network layer-3 mechanisms. These approaches are inherently constrained by architectural limitations that hinder robust service isolation. Specifically, layer-3 routing struggles to enforce strict separation across disparate Kubernetes clusters and lacks the necessary flexibility regarding network protocols. These shortcomings highlight a critical gap in the current landscape: the need for robust layer-2 mechanisms capable of delivering seamless, agnostic, and scalable connectivity in multi-cluster environments.
This paper continues some research that has been performed on layer-2 networking solutions for cloud-native environments, like L2S-M [14] and the Inter-Domain Connectivity Controller (IDCO) [15]. Our approach integrates concepts from both solutions and extends them by providing connectivity across multiple Kubernetes clusters and introducing a user-friendly API that integrates with Kubernetes. The end result is Link Layer Secure ConnEctivity slicES (L2S-CES) for microservices platforms, a network slicing solution that, by connecting multiple cloud-native domains, tackles key challenges in deploying edge cloud solutions, such as securing service exposure, expanding business opportunities, and managing data flow complexity.
This solution is evaluated and validated within a real-world architecture using a practical use-case scenario, and we demonstrate its capabilities by highlighting the benefits of incorporating network slicing into multi-tenant Kubernetes environments and by showcasing its potential for broad applicability. All code and artifacts used in this study—including those required to reproduce the validation results, the proposed scenario, documentation, and setup scripts—are available in the main branch of our official GitHub repository [16].
The rest of this paper is structured as follows. Section 2 presents the foundational concepts and technologies that underpin the proposed approach. Section 3 details the design architecture, followed by Section 4, which details the technologies and tools utilized in the solution. Section 5 reports on the validation results and evaluates the effectiveness of the proposed method. Finally, Section 6 summarizes the main conclusions and outlines potential directions for future research.

2. Related Work

Connecting application workloads across diverse cloud and edge environments has been addressed through various methods, each with distinct limitations and advantages. This section provides an overview of existing solutions, introducing network slicing techniques used in infrastructure-as-a-service (IaaS) projects and then focusing on how slicing and inter-cluster connectivity is currently being carried out in cloud-native platforms, such as Kubernetes, concluding with a comparison table of main functionalities that may be expected in a network slicing solution.

2.1. Cloud Platform Solutions for Network Slicing Orchestration

Cloud platforms incorporate slicing through varied network-provisioning techniques that enable isolation, segmentation, and resource allocation in the provided network. Many of these solutions are based on SDN because of the flexibility these architectures offer, while NFV uses virtualization to replace specialized hardware appliances with software-based network functions, improving flexibility and scalability.
As explained earlier, SDN is a network architecture that separates the control plane from the data plane, enabling more flexible and dynamic network management. Rather than configuring devices like routers and switches individually, SDN centralizes network logic into a dedicated controller, streamlining configuration and policy enforcement. This approach is commonly implemented using protocols such as OpenFlow [17].
NFV virtualizes network functions, replacing specialized hardware with software-based applications. This reduces reliance on proprietary equipment, improving network scalability, flexibility, and efficiency.
A notable example of the practical integration of SDN and NFV technologies is Open Source MANO (OSM) [18], an open-source implementation aligned closely with the ETSI NFV Management and Orchestration (MANO) framework. OSM primarily focuses on providing a robust orchestration platform for virtual network functions (VNFs), effectively managing lifecycle operations, including instantiation, scaling, and termination of VNFs. By adhering closely to the ETSI NFV architectural framework, OSM ensures compatibility and streamlined integration within multi-vendor cloud and edge deployments, enhancing the flexibility and scalability of network slicing solutions.
A widely used VIM with which OSM integrates is OpenStack [19]. As an open-source cloud operating system, OpenStack provides IaaS by managing computing, storage, and networking resources through well-defined, standardized APIs. An essential element within the OpenStack architecture is the support for SDN and NFV, facilitating highly flexible and dynamic network slicing capabilities. Particularly, OpenStack employs components such as Neutron, a networking project within OpenStack that provides network connectivity as a service between interface devices managed by other OpenStack services. Neutron integrates closely with Open vSwitch (OvS), a multilayer software switch that can act as an SDN-managed virtual bridge [20], enabling the creation of virtual isolated networks. This capability is crucial for network slicing as it provides the necessary isolation, segmentation, and performance management required by modern network workloads.
Amazon Web Services and Google Cloud both offer commercial, managed cloud IaaS solutions, EC2 [21] and Compute Engine [22], respectively. Virtual networking is provisioned in both solutions, relying on SDN to provide functionalities like on-demand virtual private networks and cross-cloud networks.
The Cloud Provider solutions detailed in this section serve as foundational technologies and methodologies essential for understanding modern network slicing orchestration. By analyzing these platforms, it is shown how combining SDN and NFV principles, isolation, segmentation, and dynamic resource management are achieved in IaaS scenarios. This analysis sets a necessary baseline for appreciating the evolution of network slicing in cloud-native platforms, like Kubernetes.

2.2. Kubernetes Network Virtualization Solutions

The functionality and applicability of Kubernetes [7] has significantly evolved over recent years. Initially, Kubernetes was developed as a cloud platform intended to deploy containerized web applications. Over time, Kubernetes has expanded to support increasingly sophisticated scenarios, including fine-grained access control, complex networking, and deployments involving constrained devices. While basic networking between containers remains straightforward, networking solutions have needed continuous adaptation to support emerging requirements. To address these evolving needs, the container networking interface (CNI) framework [23] was introduced.

2.2.1. CNI Plugins

CNI is a framework utilized by developers to create network plugins in Linux environments. The CNI manages the life-cycle of network interfaces for containers, including interface creation, deletion, and IP address assignment. Kubernetes CNI plugins are implementations of the CNI framework that provide connectivity between pods and communicate those to external networks using Linux kernel utilities. Flannel [24] serves as a representative example of a CNI plugin, implementing VXLAN tunnels among the cluster nodes and network bridges for connecting pods to this VXLAN overlay, providing basic communication and IP address management.
As mentioned in the introduction, Kubernetes’ scope has increasingly expanded to edge cloud solutions, IoT scenarios, and multi-tenant solutions. These CNI plugins, in their basic form, are not sufficient for supporting secure inter-cluster communication, observability, and advanced traffic management. Some CNI plugins have introduced extensions that partially address these limitations.
A notable example of this approach is Calico [25], one of the most popular Kubernetes CNI plugins. It provides versatility in network solutions by enabling the administrator to choose which data plane should be used. Calico can optionally establish communication between pods residing on the same host, forwarding packets using eBPF, a technology that enables software to be loaded into the kernel, as programs directly attached to the pods’ virtual interfaces, thereby improving intra-cluster communication efficiency as it can bypass many of the packet processing layers in the Linux network stack. Calico has addressed multi-cluster connectivity for industry use, with a federated approach, through Tigera’s commercial solution, Calico Enterprise [26]. A multi-cluster mesh is established, with one cluster serving as the service host, responsible for providing endpoints for the pods accessible from all clusters in the mesh. Despite being an open-source project, the cluster–mesh solutions provided by Calico are marketed primarily as enterprise solutions and are not freely available for general public use.
Another emerging CNI plugin that achieves higher throughput and lower latency compared to Calico is Cilium [27], primarily through a more extensive use of eBPF [28]. Cilium supports multi-cluster networking through ClusterMesh, which deploys a dedicated API server per cluster to manage service discovery and encryption keys, enabling distributed networking. Packet routing is performed using packet headers or via hybrid routing mode enabling direct pod-to-pod communication.
Other solutions more targeted on security aspects, like Kube-OVN [13], solve the inter-cluster connectivity with their component OVN interconnection (OVN-IC) by creating a single equal cost multiple path (ECMP) tunnel between two clusters with gateways on both sides [29]. While Kube-OVN offers enhanced connectivity and security through IPSec, it currently supports direct connectivity only between two clusters. Extending communication to multiple clusters increases complexity, requiring manual route configuration. Moreover, while traffic is isolated inside the cluster, OVN-IC uses layer-3 routing mechanisms for inter-cluster solutions, meaning that traffic between clusters is not inherently isolated at layer-2, leaving it more susceptible to certain attacks unless additional security measures are applied.
With CNI plugins, certain aspects of network slicing can be supported. Calico and Cilium both advertise the usage of network policies alongside a cluster mesh, allowing multiple pods to connect from one cluster to another. However, it remains limited in scope, particularly to some layer-7 protocols, such as HTTP and HTTPS, and only offers functionality in these specific protocols. Kube-OVN is not able to provide this autonomously; it relies on other networking solutions, such as Submariner [30], which will be further discussed later. While Multus can support multiple interfaces per pod, it does not natively manage inter-cluster connectivity or network slicing. It can serve as a complementary approach when the network administrator has in-depth knowledge and full control over the nodes in each cluster; however, the setup is complex and lacks any form of automation.
As many of these plugins rely on additional abstractions to enforce fine-grained policies and enhance observability, developers and system administrators often adopt service meshes as a complementary solution.

2.2.2. Kubernetes Service Meshes

CNI plugin solutions, such as Calico and Cilium, often rely on service meshes to enforce network policies, enhance security, and provide observability. In addition, there are standalone service mesh solutions developed independently from CNI plugins, offering inter-cluster connectivity for pods. A service mesh is an abstraction layer that defines how pods are connected by using the proxy structural pattern. Traffic is routed through a proxy that impersonates the pod, enforces layer-7 rules, and then forwards the request to the pod. For instance, a service mesh can enforce access control policies by acting as the pod’s proxy and then enforcing rules based on HTTP headers.
Service meshes typically support standardized protocols such as gRPC and HTTP. However, reliance on these protocols can lead to increased latency [31]. Another critical limitation is the absence of strict network isolation since all workloads share the same virtual network. This poses security risks in multi-tenant environments, delegating security responsibilities to the application layer, thereby increasing application complexity.
Regarding network slicing, service meshes can facilitate multi-cluster service discovery. However, as their core focus is not slicing, they typically do not incorporate dedicated operators or controllers capable of dynamically managing network slices. This implies limited granular control of the network, insufficient isolation, and limited options when designing network scenarios.
Two prominent service mesh solutions are Istio [32] and Linkerd [33]. Both provide comprehensive documentation and built-in mechanisms for managing application exposure across clusters. They achieve this by sharing service states and configurations across clusters and extending the Kubernetes API into a unified, shared operational domain. Both solutions have ways to provide multi-cluster networking, by writing on the Kubernetes Service API and scaling it to a multi-cluster level. Similarly to how CNI plugins are limited in terms of protocols, these solutions offer limited configuration flexibility.

2.2.3. Evolution of Kubernetes Virtual Networking

Alternative solutions that are designed with a focus on providing more options when utilizing multi-cluster deployments in Kubernetes are currently under active development. These are designed specifically to facilitate integration with other CNI plugins and interoperate with service meshes as well.
One of these solutions is L2S-M [14]. L2S-M is an open-source solution developed by Universidad Carlos III de Madrid (UC3M) and Telefónica Investigación y Desarrollo (TID) currently present in multiple European-funded projects [14], and it forms the basis of this paper’s design. With the capability of working alongside other CNI plugins, L2S-M enables the creation of virtual networks that deliver isolated layer-2 communication, providing consistent connectivity between application workloads. Traffic between workloads within the same virtual network is completely isolated from other networks and can be encrypted using standard IPSec technologies. However, while L2S-M effectively provides programmable link-layer connectivity, its architecture does not natively support the orchestration required for multi-cluster environments. It lacks the automated mechanisms necessary for dynamic slice creation and management across distributed domains. Architecturally extending L2S-M, L2S-CES incorporates a hierarchical control plane capable of managing connectivity across distinct Kubernetes domains, moving plain network connectivity to provide full network slicing capabilities, offering superior isolation, broadcast and multicast support across clusters, and a unified API for edge-cloud integration.
In contrast, as a multi-cluster networking tool for Kubernetes clusters, Kubeslice [12] aims to provide a framework to build slices. It defines a slice by the set of clusters that are associated with one another. It utilizes network service mesh (NSM) technology [34], distinct from traditional service mesh implementations. A network service mesh is an abstraction layer that connects multiple microservices or pods by providing a proxy for each so that their connection can be manipulated with layer-3 and layer-4 mechanisms. The NSM connects two containers forming a vWire, a set of layer-3 rules that enable packets to move from one of the pods to the other as if they were directly connected. Kubeslice uses this by setting up virtual private network (VPN) gateways to secure the communication and a virtual router per slice. The pods are connected directly to the virtual routers using the vWire, and the router will forward packets to other clusters in a programmatic way. A notable limitation of this solution is its complex setup. It achieves isolation by deploying multiple components with multiple configurations, so deploying a slice offers limited flexibility. Additionally, if multiple isolated environments were to be created, required for multi-tenancy and security compliance support, multiple slices would be required, which introduces considerable computational overhead.
The most stable solution for multi-cluster Kubernetes networking is Submariner [30]. It is widely adopted, and although it does not offer slicing-specific features such as network segmentation, it ensures secure inter-cluster connectivity for workloads. In Submariner a ClusterSet is defined as a group of two or more clusters with a high degree of mutual trust. In a ClusterSet, Kubernetes services, which act as layer-7 proxies of pods, can be exported and imported, forwarding traffic between its Gateway Engine, a component that creates a VXLAN tunnel with other clusters’ Gateway Engine. For plain inter-cluster connectivity, Submariner is a widely adopted and effective tool. While it ensures secure (IPSec-encrypted) inter-cluster connectivity, it lacks native support for network segmentation or tenant isolation.

2.3. Features Comparison of Existing Solutions

As discussed throughout this section, combining multiple solutions enables partial support for network slicing, primarily in the form of segmented connectivity and policy enforcement. Nevertheless, certain aspects such as layer-2 isolation requirements in multi-cluster environments remain an open challenge in current implementations.
The presented solution, as will be demonstrated in the subsequent sections, is maintained as a homogeneous solution for creating slices among multiple clusters, with virtual networks that can isolate traffic among workloads. This enables a variety of deployment scenarios and flexibility in designing IoT and Cloud solutions, with multiple tenants and technologies involved.
In Table 1 all the discussed solutions are presented and compared, with a detailed explanation for each below. Open-source solutions are compared functionality-wise, alongside our proposed solution, labeled as L2S-CES. A further analysis is available in Section 5 where a use case scenario is compared both in Submariner and our solution.
  • Isolated communication: The possibility of isolating traffic for security reasons in such a way that a pod can be isolated from the rest of the network. L2S-M and L2S-CES support multiple isolated virtual networks with the core functionality provided. Kubeslice and Submariner have similar features through a virtualization of some services like using NSM in Kubeslice and route agents in Submariner, which have certain isolation but do not contemplate scenarios with multiple micro-segments, similar to how L2S-CES allows multiple virtual networks at the same time. Using manual setup, Multus, and Kube-OVN can provide this kind of functionality, but it requires that the user changes the configuration inside the host.
  • Encryption: Encrypted traffic thanks to technologies like IPSec. Submariner has an implementation with strongSwan [35]. L2S-CES inherently allows for the use of any encryption mechanism by supporting layer-2 communications. Other solutions, such as service meshes and CNI plugins, provide certain encryption thanks to mutual TLS (mTLS), which authenticates using TLS from both ends, but this is limited to certain protocols and is derived from the user.
  • Low-performance impact: Low impact on the network performance. Standard in most solutions and a key feature in Calico and Cilium as they use eBPF which allows efficiency and high performance in the network. By contrast, service meshes do not comply in general in terms of efficiency as they commonly use sidecar containers, which act as middlemen in the communication, producing additional overhead. L2S-CES-inherited tunnels from L2S-M do not negatively impact the inherent network performance [36].
  • Traffic engineering: Unique feature in L2S-CES as it allows custom path creation by specifying the desired path in the network topology for each virtual network. This can be used for multiple purposes, (e.g., quality of service and security).
  • Multicast support: Support for a variety of multicast and broadcast protocols. Being a layer-2 solution, both L2S-M and L2S-CES can provide scenarios with this type of traffic. The current alternative in other scenarios would be a complex set-up with Multus and a VPN in the cluster.
  • Service discovery: Capability to expose microservices and make them accessible automatically without manual intervention. This is commonly integrated within the Kubernetes Service API by CNI plugins and service meshes. L2S-CES has this feature through its custom multi-domain DNS implementation.
  • Kubernetes integration: Degree to which the solution is Kubernetes-native. L2S-CES exposes slice lifecycle and policy as Kubernetes resources; Submariner and Kubeslice integrate via Kubernetes operators atop the primary CNI; and CNI plugins (e.g., Calico, Cilium, and Kube-OVN) are natively integrated as the cluster network layer.
  • Component interoperability: The component’s compatibility with other components. Submariner, Kubeslice, and L2S-CES are agnostic in terms of the primary CNI plugin, so they integrate well while giving more characteristics. Service meshes are fully integrated with the base CNI plugin as well.
  • Multiplatform support: Support for multiple platforms to work together, i.e., public and private clouds, k3s cluster with bare metal cluster, etc. L2S-CES allows the connection of one virtual or physical machine to a microservice if required, thanks to SDN.
  • Automated configuration: The automation in initially configuring the solution. One of the improvements of L2S-CES is its automation in slice management and configuration, whereas with Kubeslice a deep understanding of the tool is required. Submariner has a CLI with many automated features. Other solutions require a much more complex setup as they are not natively built into the tool, Calico and Cilium, for instance, which do not have much documentation available, and Kube-OVN requires ECMP tunnels to be set up, which is not an easy task.
  • Metric exposure: Common feature with service meshes, where performance metrics are exposed. L2S-CES has a built-in tool for latency, bandwidth, and jitter measurements.
  • Slice management: Ability to define, instantiate, modify, and delete slices dynamically at runtime. L2S-CES offers fully dynamic slice management with per-slice path selection; other solutions provide coarser or manual slice-like constructs with limited slice segmentation.

3. Design

3.1. Key Design Concepts

The proposed networking solution primarily manages network slices and services across multiple cloud domains. It exposes an accessible API and ensures consistent configuration and isolated communication across clusters. The main key technical objective of the solution is providing flexible slicing with the following capabilities:
  • Heterogeneous platform support: Providing a user-friendly API and command-line interface (CLI) to create isolated environments in Kubernetes clusters, enabling users to easily establish network slices that host virtual networks. Multiple tenants can dynamically join these slices and communicate without concern for the heterogeneity of the underlying infrastructure.
  • Isolated network communication: Ensuring secure, isolated communication between containers within and across Kubernetes clusters, spanning multiple domains and cluster types. A single slice may contain one or multiple isolated networks, which are accessible on-demand.
  • Platform-agnostic deployment: Providing scalability and adaptability as part of the solution.
  • Flexible provisioning: Enabling the provisioning of slices attending to underlying resources, like deciding the topology of interconnection between the multiple clusters and computing nodes.
This section details the main architectural components and key concepts that underpin the solution.
Figure 1 illustrates how the solution interconnects multiple Kubernetes clusters through network slices. A set of Internet-connected clusters (clusters A, B, and C) are connected via two different slices. Within each slice, a specific network topology is defined, describing how compute nodes within and across clusters interconnect. This topology is configured and maintained by the Slice Manager, enabling the creation of multiple virtual networks within a single slice. These virtual networks interconnect pods located in separate clusters, enabling developers to deploy and interconnect workloads across clusters without needing to manage or understand the underlying virtual infrastructure.
For instance, as shown in Figure 1, Slice 1 contains three distinct virtual networks: the first connects pods in Cluster A with ones in Cluster B, the second connects two pods within Cluster B, and the third one connects pods in Cluster B with Cluster C. Additionally, Slice 2, which omits Cluster B, demonstrates another virtual network that directly connects pods in Cluster A with pods in Cluster C.
The architecture consists of three operational layers, each managed by different actors with specific roles. Starting from the bottom, these layers include the Cloud Infrastructure Layer, the Slice Administration Layer, and the Service Provisioning Layer.
Cloud Providers—such as Amazon Web Services [37], Google Cloud [38], or private Cloud Providers—establish a service-level agreement (SLA) with Slice Administrators. These agreements enable Slice Administrators to connect multiple Kubernetes clusters across various cloud platforms, facilitating the creation of interconnected network slices.
Slice Administrators use the Slice Manager to create and manage slices. There is no design or technical limit on the number of slices that can be provisioned. Upon slice creation, they define the clusters involved and specify detailed configurations, such as which nodes within each cluster have connectivity to nodes in other clusters. Figure 1 illustrates these connections for each slice. For example, Slice 1 connects Cluster A, Cluster B, and Cluster C using multiple gateway nodes: Cluster A has two gateways connecting directly to Cluster B’s gateways, while an additional gateway node in Cluster B connects onward to Cluster C. Conversely, Slice 2 provides a simpler topology, directly linking Cluster A and Cluster C through single gateway nodes, bypassing Cluster B entirely.
At the top layer, Service Providers manage individual virtual networks, providing isolated layer-2 communication channels specifically tailored to connect pods or containers. Each virtual network provides an isolated layer-2 channel for its pods or containers inside. When handling service deployment requests—especially for complex, distributed services running across multiple clusters—Service Providers create virtual networks within slices, explicitly defining which pods should be interconnected and the number of virtual networks needed. To achieve this, the Service Provider deploys the necessary Kubernetes resources and the slice’s custom resource definitions (CRDs).
We designed the system with a Kubernetes-native declarative interface to streamline slice and network management. By treating virtual networks and edge devices as CRDs, the system achieves consistent orchestration across multi-cluster environments. This architecture inherently supports horizontal scalability; it avoids arbitrary design limits on slice count, instead allowing system capacity to expand organically alongside the provisioning of additional computational and networking resources.
To ensure resilience, the system leverages Kubernetes reconciliation loops to automatically restart failed components and restore the state from CRDs. Simultaneously, the SDN Controller utilizes active Link Layer Discovery Protocol (LLDP) monitoring to dynamically recalculate paths around topology faults.

3.2. Actors

This solution defines three distinct types of actors, each differentiated by their specific roles and responsibilities within the slicing workflow: Service Provider, Slice Administrators, and Cloud Providers.
Cloud Providers are the companies or organizations that provide cloud computing services. For the purpose of this paper, they are narrowed down to their IaaS, where they provide a cloud environment and a Kubernetes cluster as Platform as a Service (PaaS) is deployed on it. Slicing will be built upon different infrastructures, providing a uniform solution.
Slice Administrators utilize platforms provided by Cloud Providers, managing the slice resources. They request and configure Kubernetes custom resources to establish, manipulate, and maintain network slices across clusters.
Service Providers, through established SLAs with Slice Administrators, deploy and manage applications or workloads that span multiple Kubernetes clusters. They ensure isolated and secure communication by creating dedicated virtual networks within existing slices.
A typical slice workflow and communication of these users is depicted in Figure 2, including a fourth actor, the Service Requester, which initiates the workflow after requesting the deployment of a multi-domain service.
Upon receiving a request for a multi-domain service, the Service Provider has two options: either utilize an existing slice—provided it meets the computational and networking requirements and can be properly associated with the Service Requester—or initiate the creation of a new slice.
When a new slice is requested, the Slice Administrator coordinates with Cluster Administrators to agree upon and configure necessary parameters. Subsequently, the Slice Administrator deploys the appropriate Kubernetes resources across each involved cluster to finalize slice creation.
Once the slice becomes operational, the Service Provider deploys the requested multi-domain service, establishing multi-domain virtual networks within the configured Kubernetes clusters.
When defining slices, administrators should select nodes and inter-node connections based on available resources, including CPU, memory, storage, and physical location, to ensure optimal performance and resource efficiency. For example, resource-constrained edge devices, such as Raspberry Pi devices, should not serve as central hubs in network topologies. Instead, they should be connected directly to central nodes that have sufficient computational and networking resources.

4. Implementation

This section gives an overview of the technologies and tools adapted to implement the proposed architecture.
Slices are provisioned as SDN-managed overlay networks, run on Kubernetes clusters, meaning both Service Providers and Slice Administrators manage the Kubernetes clusters to varying degrees. The requirement of a close integration into the Kubernetes platform serves as the rationale for a Kubernetes-based implementation. Logically the actors’ entry component for L2S-CES is the Kubernetes controller-manager framework, which provides an API for CRDs. Specifically, the L2Network, Overlay, and NetworkEdgeDevice CRDs are defined. Actors define their desired configurations using standard YAML files and can manage and monitor these resources via the Kubernetes command-line client (kubectl).

4.1. Software Components

Figure 3 illustrates the primary software components responsible for managing and enabling network slices within this solution. Each component contributes to either intra- or inter-cluster connectivity and resource management. All components are deployed as Kubernetes pods: the Operator, SDN Controller, Slice Packet Switches (SPSs), Network Edge Devices (NEDs), Slice Connectivity Orchestrator (SCO) Provider, and the Slice Manager. The architecture differentiates between two types of clusters: the control plane cluster, which hosts the SCO Provider and the Slice Manager, and the managed clusters, each containing operational components such as the operator, SDN Controller, SPSs, and NEDs. Managed clusters refer specifically to Kubernetes clusters actively managed and interconnected by these components.
Inside the control plane of the managed clusters, the operator and internal SDN Controller are running. The operator acts as the intermediary between Kubernetes and the solution’s resources. It interprets user-defined configurations sent through the northbound interface—NBI. The NBI exposes high-level intents and policies to the control layer, implemented with CRDs, which are cached by the Kubernetes API server upon its reception using the standard Kubernetes CRD managing flow [39] so the operator can ensure that the cluster consistently reflects the desired operational state. Within each managed cluster, the operator communicates both with the SDN Controller and SCO Provider, translating these configurations, and communicates back with the API Server, launching Kubernetes resources in the cluster, and reporting the status of the CRDs. The SDN Controller enforces network traffic policies across internal nodes via an OpenFlow interface, directly controlling SPSs. The SPSs are custom implementations based on OvS, deployed on every node within a managed cluster. They collectively establish and maintain the internal overlay network topology specified by the user. The NEDs are unique services, deployed in the gateway nodes of the managed clusters. Their primary function is to bridge internal cluster networks with other clusters by establishing direct interconnections with remote NEDs.
The SCO Provider, residing externally in the control plane cluster, functions as an inter-cluster SDN Controller. It coordinates with NEDs across multiple clusters, managing the overarching inter-cluster overlay network topology, in a similar fashion to how the SPS and SDN Controller inside the cluster communicate and behave. To facilitate the multi-domain infrastructure management, the Slice Manager is presented as the sub-component that uses the NBI of each through which users can simultaneously deploy and manage custom resources across multiple clusters from a single access point.
Below, further definitions and technical operations are described for every component.

4.1.1. Operator

The operator component within each cluster serves as the interface connecting SDN capabilities with Kubernetes-native resource management.
As part of the Kubernetes controller-manager paradigm, it performs Kubernetes-style reconciliation, actively monitoring custom resources (L2Network, NetworkEdgeDevices, and Overlays) to align the actual cluster state with the user’s desired configuration, responding promptly to creation, modification, or deletion events. It owns all resources of these kinds and provides the logic that makes sure that the user desired specification is fulfilled while being responsive to changes and Kubernetes events, such as modifications, deletions, and creations.
In addition to managing custom resources, the operator also intervenes in the pod’s lifecycle if it is possible to use this solution through the definition of a label l2sm: true.
For Overlay custom resources, the operator deploys and configures both the SDN Controller and associated SPSs according to the topology specified by the user. Users define the nodes involved, their interconnections, and initial scaling parameters, ensuring the network’s precise configuration. The user can create a topology, indicating the nodes in the cluster that are going to be connected, and the links between those that shall be instantiated. Additionally, an interface number is specified, where the user indicates an approximation of how many pods per node are going to be deployed, helping L2S-CES scale resources initially; a switch template, which is a configuration of the container that will run the SPS; and a provider, that references the internal SDN Controller. Aside from the topology field, the fields are set to default to use the standard solution integrated with L2S-CES. This implementation defines a custom SDN Controller and SPS implementation, alongside 10 initial interfaces per node. These fields can be modified for more flexible solutions where the user may want a different implementation or scalability.
Similarly, creating a NED custom resource results in the deployment of a pod specifically configured to act as a gateway node. Users specify the gateway node’s location, other clusters to connect with, and the configuration details necessary to communicate with remote SCO Providers. The user can indicate which node in the cluster should be used as the gateway, a list of other clusters this NED should be connected to, by specifying their domain location, the location and configuration of the remote SCO Provider to be used, and a switch template, which specifies the Docker [40] image and container configuration of the inter-domain OvS solution.
L2Network describes isolated layer-2 network segments within clusters. Each L2Network resource corresponds directly to a single virtual network managed by the slice’s dedicated SDN Controller, represented as policy intents enforced by the SDN Controllers, connecting specific pods as defined by the user. The operator uses the SDN Controller REST API to create the network and specify which pods are being connected and from which node. There are no solution-imposed limits beyond cluster and controller capacity on the number of pods that can join a virtual network nor on the quantity of virtual networks established within a given slice. These virtual networks are managed by the SDN Controller of the slice, a unique SDN Provider that enforces rules on the slice to provide this isolated communication.
Optionally, L2Networks support layer-3 configurations when the user provides a NetworkCIDR. In this scenario, each pod receives an IP address assigned by the IPAM CNI plugin, and an integrated DNS service may be deployed to facilitate network name resolution across inter-domain networks.
To attach pods to an L2Network, each pod must explicitly include the label l2sm: true and an annotation l2sm/networks, specifying the network names. The operator then utilizes Multus to dynamically provision the required network interfaces and inform the SDN Controller of each new pod joining the network.
As a Kubernetes-native component, the operator continuously provides detailed resource status information to users. For example, it reports network availability, pod connections, IP address assignments, and ensures the correct functioning and communication of switches and controllers within Overlays and NEDs.

4.1.2. SDN Controller

The SDN Controller is the central orchestrator of the network fabric, implemented as a distributed Java application that extends the base ONOS platform, offering a dedicated REST API. Building upon the foundations presented in [15], it maintains the logic required to instantiate flexible layer-2 services within each cluster and coordinates inter-domain connectivity.
At a high level, the operator communicates directly with the SDN Controller through its REST API to create or update layer-2 ‘intents.’ These intents explicitly define the SPSs, associated ports, and network flows required for each layer-2 network. Specifically, when the operator requests a new L2Network through the solution’s northbound interface, the SDN Controller receives the following:
  • A unique Network ID for identifying the slice or network instance to be deployed.
  • A list of devices (SPSs) and associated OpenFlow identifiers (DPIDs) to be attached.
  • A specification of the ports (e.g., OpenFlow port numbers) that each device will use to exchange traffic within the network.
Upon receiving these configurations, the SDN Controller dynamically installs precise OpenFlow rules into each SPS, enforcing isolated connectivity through unique tunnel identifiers, managing traffic forwarding, and optimizing multicast or broadcast traffic flows.
To enhance network resilience, the SDN Controller continuously monitors network topology by directing each switch to transmit Link Layer Discovery Protocol (LLDP) messages. In case of device or link failures, the SDN Controller rapidly recalculates optimal paths and re-installs flows to maintain consistent connectivity and traffic isolation. This approach allows the solution to adapt to changes in real-time and avoids the need for spanning tree protocols since loop-free trees are enforced explicitly through the Controller’s flow management.
Additionally, the SDN Controller dynamically learns and distributes endpoint MAC addresses, simplifying network operations by reducing unnecessary traffic and proactively managing endpoint discovery.

4.1.3. Slice Packet Switches (SPSs)

SPSs are lightweight, Kubernetes-native instances of OvS deployed as replica sets, the resource that manages a set of pods, on every cluster node. They are connected to pods and external NEDs, with a custom Go Library that wraps OvS command-line utilities (e.g., ovs-vsctl and ovs-ofctl) to program flow rules, manage ports, and establish VXLAN tunnels [41].
Upon creation of an Overlay custom resource, the operator automatically generates a config-map, the Kubernetes resource used for configuration files, containing two critical configurations provided by the user: the network topology and the specific switch settings. The SPS then uses this config-map during startup to initialize communication with the SDN Controller and to configure its network interfaces. The microservice will load its configuration from this config-map and execute its entry program, initializing the communication with the SDN Controller and connecting several desired ports.
These interfaces will be later attached to pods with the Multus solution, using a layer-2 bridge CNI Plugin. This way, the pod has a network interface connected to the bridge in the host node, connected simultaneously to the SPS pod, which will receive incoming traffic from the pod and forward it according to the SDN Controller flows. If no flow is installed, the pod will be isolated from the network overlay. This way, multiple pods can be connected to the overlay, and each can be in its desired networks, relying on the layer-2 isolation given by the SDN paradigm.

4.1.4. NEDs

NEDs are specialized switches designed to connect Kubernetes clusters to other Kubernetes clusters or other SDN domains (e.g., an OpenStack-managed data center).
The NEDs are represented by the NetworkEdgeDevice custom resource, a programmable switch with its basic configuration required to connect an overlay in one cluster to an overlay in another cluster.
They act as the gateway from the cluster that enables the use of L2S-CES to be expanded to a broader selection of scenarios and solutions. They constitute part of the data plane of an inter-cluster network topology, managed by the Slice Connectivity Orchestrator (SCO), an external SDN Controller that acts from the central control plane cluster.
Unlike standard switches, NEDs operate as standalone Kubernetes pods featuring a custom gRPC-based interface. This design allows Slice Administrators to dynamically configure network interfaces, perform hot-plug operations for pods, and quickly establish or remove inter-cluster VXLAN tunnels as needed.

4.1.5. SCO Provider

The SCO Provider is the Slice Connectivity Orchestrator, the component responsible for the multi-domain virtual networks inside the slice.
As a provider, it enables virtual network isolation, in a similar fashion to the SDN Controller, by isolating traffic in the NEDs, identified by the incoming port and MAC address. Additionally, it includes a microservice capable of providing DNS services, the inter-domain DNS. As a result, inside a slice, multiple virtual networks are created and have the capability to support the basic layer-3 set-up with IP address management and DNS resolution inside each virtual network. Then, a pod inside a virtual network can communicate either via layer-2 or automate for simpler scenarios with the use of DNS to connect to other pods.
Because of this component’s similarity to the SDN Controller, its interface with the NEDs is implemented like the SDN Controller–SPS interface, using the OpenFlow Protocol.
On the other side, its interaction with each cluster’s operator is inherently different. Virtual networks can be created and managed from different clusters in parallel, and their operation process differs. When a virtual network is created inside a cluster, an L2Network resource is created inside that cluster, and the SCO Provider initially processes it, attaching one port in the cluster’s NED to the L2Network. Later on, other clusters can create that same resource, and the SCO Provider will confirm its existence, assigning one port of each of the interested cluster’s NED to the virtual network. This approach enables the unification of the management of a single resource at a central point.
Inside a cluster, the SDN Controller will forward traffic from the network to the NED using one of the SPS’s ports and forwarding any traffic through it.
The SCO Provider manages DNS resolution for inter-domain networks using a customized CoreDNS implementation. It dynamically updates DNS entries based on pod information provided through a dedicated gRPC interface, ensuring that pods across clusters can reliably communicate by resolving addresses of the form
<pod-name>.<virtual-network-name>.inter.l2sces

4.1.6. Slice Manager

The Slice Manager is implemented as a Kubernetes client designed to centrally manage multi-domain resources. Provided a list of managed clusters, it can create the necessary L2S-CES custom resources (L2Network, Overlays, and NEDs) and configure them in such a way that enables the creation of slices and virtual networks more easily, from a single interface.
The Slice Manager operates in two modes: it can function as a gRPC server accessible to other microservices or as a CLI tool, the L2S-CES Client, enabling flexible interactions for creating and managing Kubernetes resources across clusters. It will be given a list of clusters on which to operate, with bearer tokens and API endpoints for each and the operation that is requested, with the least amount of info, and as an output it will make the necessary calls to the Kubernetes API server of each cluster, creating the Kubernetes resources.
For managing a slice, the Slice Manager needs as input the list of clusters. A cluster is configured by determining the identifier of the slice, the nodes on it that will be part of the layer-2 overlay, which of these nodes will be acting as gateways, and optionally the network topology. If no topology is chosen, it will be full-mesh by default, where every node is connected to every node. To simplify slice creation, the Slice Manager automatically translates high-level user-defined cluster and topology configurations into the appropriate Kubernetes custom resources (NEDs and Overlays), executing the necessary API calls to deploy these configurations to each targeted cluster.
To add an additional cluster to an existing slice, the Slice Manager requires the slice identifier, new cluster details, and connectivity information (such as IP addresses) of existing clusters. This ensures that the appropriate inter-cluster VXLAN tunnels are correctly established through NEDs.
For creating a virtual network, the Slice Manager needs the name of the network, the list of clusters that will have this network created, and which SCO Provider is available, including its endpoint.

5. Use Case

In order to validate L2S-CES, we chose a multi-tenant emergency-response scenario and compared it with the same environment using Submariner instead. The scenario comprises three tenants, the park fire department, the firefighters, and the police, each operating a dedicated Kubernetes cluster, joined in a single slice with three virtual networks as shown in Figure 4. The fire brigade supervises a park by partitioning it into sensorized blocks (temperature, smoke, wind, and humidity), producing continuous telemetry for a real-time risk map. Under non-emergency conditions, only an internal virtual network is active. Upon anomaly detection, tenants programmatically instantiate a shared slice and multiple virtual networks via the Kubernetes API to exchange telemetry and application data.
If a fire is too large or it is spreading at an abnormally fast pace, the brigade must alert the second tenant, the firefighters, so they can reach the scene quickly with more equipment and personnel. If the firefighters and the brigade have an SLA with a Network Administrator and want to create networks dynamically, they can create a slice between their Cloud Providers or local servers and share data securely and quickly. In this way, the firefighters promptly receive information about where the fire is located and how it has spread and can connect to their own analysis applications. An additional tenant is included, the police, who can dynamically enter the scenario for additional support.
The presented scenario is used as an example to validate L2S-CES due to how its technical requirements demonstrate its usefulness: the setup must be quick and efficient in order to avoid unnecessary delays; the flexibility of the network for different multicast/broadcast protocols such as the Real-Time Transport Protocol (RTP) over User Datagram Protocol (UDP) [42] is fundamental for communication redundancy; network isolation is essential in order to avoid attacks or accidental exposure to unintended data; and good computing performance is needed. In addition to L2S-CES, we will also deploy the Submariner alternative because it is currently the most advanced tool with similar capabilities and can provide a strong point of comparison.
Likewise, the following services could be configured in the system:
  • A call application using the Session Initiation Protocol (SIP) [43] among the three entities;
  • An RTP application for streaming real-time video data of the fire status;
  • A service that delivers sensor data to the firefighters and the private network;
  • Fire simulation and prevention software in the private network, with outputs shared dynamically.

5.1. Hypotheses

In order to perform the different comparisons, we will consider the following hypotheses:
H1. 
L2S-CES achieves low slice instantiation time.
H2. 
L2S-CES incurs no greater CPU overhead at steady state.
H3. 
L2S-CES enforces stronger isolation under active probing conditions.
H4. 
L2S-CES supports multicast traffic.

5.2. Experimental Setup

Given the technical requirements that this use case supposes, unique tests were performed to analyze and compare each of them. In order to make the tests replicable, the test repository provided [44] includes all tests, results, plotting code, and configuration files to automatically deploy the same testbed, including all Kubernetes clusters with their configurations. A combination of Terraform [45] and Ansible [46] was used to launch the necessary infrastructure for the tests, with a fixed configuration to ensure that the data remained consistent and hardware homogeneous.
Particularly, the deployment consisted of 1 computing node with 2 CPUs, 4 GB of RAM, and Ubuntu 20.04, running Prometheus (scraping/storage) and Grafana [47] (visualization) to collect and present metrics for CPU-usage measurement and 14 computing nodes, each with 2 CPUs, 4 GB of RAM, Ubuntu 20.04, and Kubernetes v1.32.5, which are shared across 6 clusters as presented in Table 2. Control clusters in Submariner and L2S-CES act as central control plane clusters of the communication for each solution, and the managed ones are the clusters that host the applications and workloads for the tests. Each node has a cAdvisor container [48] that exports to the monitoring cluster the resource utilization of each container in the node and a Node Exporter [49] that exports general node measurements and diagnostics for performance evaluations.
The repository contains documentation for using automated scripts, the provisioning of this infrastructure, and all requisites for the tools.

5.3. Evaluation and Results

5.3.1. H1—L2S-CES Achieves Low Slice Instantiation Time

Both L2S-CES and Submariner have a low setup time of the network segments, the L2Networks in L2S-CES and the service exports in Submariner, measured by deploying a workload and measuring 50 times the time it takes to expose them to the other clusters, recording less than 100 milliseconds in L2S-CES. As both solutions come with a CLI, installation and management are automated. Submariner requires the specification of the cluster’s Kubernetes configuration files (kubeconfigs) and a cluster ID, as well as specifying a gateway node. L2S-CES requires the same specification through a YAML file, which can use a bearer token as an alternative to the kubeconfigs. L2S-CES also lets the network administrator choose the interface to be used for communication, in case it is not the default one.

5.3.2. H2—L2S-CES Incurs No Greater CPU Overhead at Steady State

As shown in Figure 5, we report per-application CPU utilization (in core-equivalents) obtained by aggregating each application’s container usage as exposed by cAdvisor. Measurements consist of non-overlapping 5 min windows collected dispersively over a three-day period while a representative multi-cluster application generated background traffic. In total, 400 windows were analyzed in each cluster.
Both solutions impose negligible CPU usage on managed clusters of μ 0.055 cores, which amounts to less than 3% of CPU overhead (0.055/2 × 100 = 2.75%). The control plane has a lower CPU usage, with L2S-CES at μ 0.024 cores and Submariner at μ 0.002 cores. All measurements have a standard deviation of σ < 0.04 .

5.3.3. H3—L2S-CES Enforces Stronger Isolation Under Active Probing Conditions

The network isolation in the two solutions is compared in Figure 6, with an emitter pod deployed in the managed-1 cluster of each solution, which is connected to a pod in the managed-2 cluster. A third pod is deployed in the managed-2 cluster that takes no part in the communication. It is noted that the figure shows all packets received by the pods, with the goal of achieving maximum visibility of traffic observed in the two solutions. In Submariner a headless exposed service for each pod is used, and in L2S-CES two virtual networks are used, one attached to the first and second pod and the other for exposing the third pod. No firewalls or additional security measures are implemented to showcase the tool’s inherent network isolation.
The experiment has two stages; in the first one an emitter pod is deployed in the managed-1 cluster of each solution that performs a ping, sending Internet Control Message Protocol (ICMP) to the receiver pod in the managed-2 cluster. During the second stage, the pod in managed-1 cluster runs nmap v7.95 [50] to scan all the exposed pods in the managed-2 cluster.
Both Submariner and L2S-CES show traffic from the management components. Submariner has recurrent Address Resolution Protocol (ARP) packets from the cluster’s operator, which manages its exposure to other nodes, but does not go beyond the cluster’s nodes. L2S-CES, on the other hand, has Link Layer Discovery Protocol (LLDP) packets for host discovery in the layer-2 infrastructure, which can be observed as background traffic in the figure.
During the first stage, in a non-isolated environment, the third pod would receive ARP traffic from the moment the emitter wants to begin communication with the other pod. As in Submariner, ARP traffic is managed from the operator, these packets do not get through, and no packet is seen in that moment. This is also the case for L2S-CES as well as this third pod is exposed but isolated from that communication.
During the second stage, in a non-isolated environment, the third pod would receive traffic from the scan, thereby revealing its existence to other pods in the other clusters. This is the case of Submariner, where both the second and third pod receive an ICMP packet and answer it. In the case of the second pod, this is an expected behavior, contrary to the third pod, which means its insecure exposure. On L2S-CES, the second pod receives ARP traffic from the nmap and answers to its ICMP packet, as expected, and the third pod receives neither ARP traffic nor ICMP, so the pod is isolated completely.

5.3.4. H4—L2S-CES Supports Multicast Traffic

Multicast and broadcast traffic were tested in the two solutions as shown in Figure 7. An emitter pod is deployed in the managed-1 cluster of each solution that sends traffic to two connected pods in the managed-2 cluster. In Submariner a headless exposed service for each pod is used, and in L2S-CES a virtual network is used, connecting the three pods. No additional VPN tunnels are implemented to showcase the tool’s inherent support for these traffic types.
First some broadcast traffic is sent from the pod in the managed-1 cluster using the ping broadcast command. Submariner’s pod does not receive any packet back as neither the other pods receive the traffic, and on L2S-CES, the emitter receives replies from both pods.
The resulting packet rates, as depicted in the figure, underscore the contrast in handling multicast traffic: Submariner fails to deliver packets to the destination pods, whereas L2S-CES ensures successful delivery.
To further validate this capability within an application context, we conducted a Session Initiation Protocol (SIP) experiment based on the use case defined in Figure 4. An L2S-CES emitter was configured within the multicast group to host a SIP server instance using SIPp, an open-source tool for traffic generation and benchmarking [51]. A corresponding SIP client, deployed on one of the receivers directed traffic to the multicast group address at a target rate of 1000 calls per second (CPS) for a duration of 100 s.
The results, presented in Figure 8 and Figure 9, demonstrate the stability and efficiency of multicast SIP communication. Figure 8 confirms a consistent transmission rate of 1000 CPS, adhering strictly to the configured parameters. This reliability is further corroborated by Figure 9, which indicates a 100% call completion rate (zero failures) with response times consistently remaining below 10 ms.

5.4. Performance Summary

These results show that the communication is isolated among pods in each network, supporting diverse solution methodologies with a quick setup time and good CPU performance. It is thus verified that L2S-CES is dynamically responsive, provides secure communication, and does not significantly impact the cluster capabilities, which is required by the explained scenario. In addition, its native layer-2 support facilitates deployment in more complex topologies and with legacy protocols under operator-defined policies.

6. Conclusions

Virtualization technology has significantly advanced cloud computing networks, fostering the development of numerous innovative solutions and tools to address diverse requirements from both providers and users globally. As applications grow increasingly complex and expand across heterogeneous domains, there is a need for developers, cloud operators, and service providers to deploy flexible, dynamic, and secure solutions within cloud-native platforms.
In this paper, we presented the design and implementation of a slicing solution for cloud-native platforms that distinguishes itself through its flexibility in network protocols and underlying platforms by offering layer-2 connectivity, with robust, secure traffic isolation. Experimental validation demonstrates the solution’s efficiency and scalability, achieving network slice instantiation times of under 100 milliseconds while imposing minimal resource consumption, with a steady-state CPU overhead of less than 3% (approximately 0.055 cores) on managed clusters. Furthermore, L2S-CES effectively bridges cloud infrastructures with native support for broadcast and multicast traffic, as evidenced by its ability to sustain 1000 SIP calls per second with a 100% completion rate and response times consistently below 10 ms. Unlike existing layer-3 alternatives, this approach enforces robust security through complete network isolation during active probing, providing a verifiable foundation for deploying multi-tenant applications in distributed edge-cloud environments.
Future work will further explore the scalability and performance implications of this solution in larger, heterogeneous deployments as well as extend the slicing capabilities to support more fine-grained flexible traffic engineering mechanisms and enhanced observability mechanisms.

Author Contributions

Conceptualization, A.T.d.C.B., I.V., and F.V.; methodology, A.T.d.C.B.; software, A.T.d.C.B.; validation, A.T.d.C.B., I.V., and F.V.; investigation, I.V. and F.V.; resources, I.V. and F.V.; writing—original draft preparation, A.T.d.C.B.; writing—review and editing, A.T.d.C.B., I.V., and F.V.; visualization, A.T.d.C.B.; supervision, I.V. and F.V.; project administration, I.V. and F.V.; funding acquisition, I.V. and F.V. All authors have read and agreed to the published version of the manuscript.

Funding

This article has partially been supported by the 6G-INSPIRE (Enabling Native-AI Secure Deterministic 6G Networks for Hyper-connected Environments) project, funded by the Spanish State Research Agency with reference PID2022-137329OB-C42; the NEMO (Next Generation Meta Operating System) European project, funded by the Horizon Europe programme with grant agreement 101070118; and by the CyberNEMO (End-to-end Cybersecurity to NEMO meta-OS) European project, funded by the Horizon Europe programme with grant agreement 101168182.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The experimental results and replication scripts are available at https://github.com/Tjaarda1/paper-slices-2025 (accessed on 30 October 2025).

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
APIApplication Programming Interface
CIDRClassless Inter-Domain Routing
CLICommand-Line Interface
CNIContainer Network Interface
CRDCustom Resource Definition
DNSDomain Name System
EC2Elastic Compute Cloud
ECMPEqual-Cost Multi-Path
eBPFExtended Berkeley Packet Filter
ETSIEuropean Telecommunications Standards Institute
gRPCRemote Procedure Call framework
IaaSInfrastructure as a Service
IDCOInter-Domain Connectivity Controller
IPAMIP Address Management
LLDPLink Layer Discovery Protocol
L2S-CESLink Layer Secure ConnEctivity slicES
mTLSMutual Transport Layer Security
NBINorthbound Interface
NEDNetwork Edge Device
NFVNetwork Function Virtualization
OvSOpen vSwitch
RTPReal-Time Transport Protocol
SCOSlice Connectivity Orchestrator
SDNSoftware-Defined Networking
SIPSession Initiation Protocol
SPSSlice Packet Switch
TTLTime To Live
UDPUser Datagram Protocol
VIMVirtualized Infrastructure Manager
VNFVirtual Network Function
VPNVirtual Private Network

References

  1. Ammar, S.; Lau, C.P.; Shihada, B. An in-depth survey on virtualization technologies in 6G integrated terrestrial and non-terrestrial networks. IEEE Open J. Commun. Soc. 2024, 5, 3690–3734. [Google Scholar] [CrossRef]
  2. SubOptic Spectrum Sharing Working Group. Spectrum Sharing Working Group White Paper; White Paper; SubOptic Association: London, UK, 2021; Available online: https://www.suboptic.org/papers-presentations/spectrum-sharing-working-group (accessed on 30 October 2025).
  3. Bari, M.F.; Boutaba, R.; Esteves, R.; Granville, L.Z.; Podlesny, M.; Rabbani, M.G.; Zhang, Q.; Zhani, M.F. Data center network virtualization: A survey. IEEE Commun. Surv. Tutorials 2012, 15, 909–928. [Google Scholar] [CrossRef]
  4. Bentaleb, O.; Belloum, A.; Sebaa, A.; El-Maouhab, A. Containerization technologies: Taxonomies, applications and challenges. J. Supercomput. 2022, 78, 1144–1181. [Google Scholar] [CrossRef]
  5. Velepucha, V.; Flores, P. A Survey on Microservices Architecture: Principles, Patterns and Migration Challenges. IEEE Access 2023, 11, 88339–88358. [Google Scholar] [CrossRef]
  6. Alshuqayran, N.; Ali, N.; Evans, R. A systematic mapping study in microservice architecture. In Proceedings of the 2016 IEEE 9th International Conference on Service-Oriented Computing and Applications (SOCA), Macau, China, 4–6 November 2016; pp. 44–51. [Google Scholar]
  7. Kubernetes. Kubernetes Overview. 2025. Available online: https://kubernetes.io/docs/concepts/overview/ (accessed on 30 October 2025).
  8. Deng, S.; Zhao, H.; Huang, B.; Zhang, C.; Chen, F.; Deng, Y.; Yin, J.; Dustdar, S.; Zomaya, A.Y. Cloud-native computing: A survey from the perspective of services. Proc. IEEE 2024, 112, 12–46. [Google Scholar] [CrossRef]
  9. McKeown, N. Software-defined Networking. INFOCOM Keynote Talk 2009, 17, 30–32. [Google Scholar]
  10. Chiosi, M.; Clarke, D.; Willis, P.; Reid, A.; Feger, J.; Bugenhagen, M.; Khan, W.; Fargano, M.; Cui, C.; Deng, H.; et al. Network Functions Virtualisation: An Introduction, Benefits, Enablers, Challenges & Call for Action. In Proceedings of the SDN and OpenFlow World Congress, Darmstadt, Germany, 22–24 October 2012; White Paper; ETSI: Sophia Antipolis, France, 2012; Available online: https://portal.etsi.org/NFV/NFV_White_Paper.pdf (accessed on 30 October 2025).
  11. NGMN Alliance. Description of Network Slicing Concept. Technical report, Next Generation Mobile Networks (NGMN) Alliance, 2016. NGMN 5G P 2016, 1, 1–11. [Google Scholar]
  12. Avesha, I. Kubeslice. 2025. Available online: https://github.com/kubeslice (accessed on 30 October 2025).
  13. Liu, M. Kube-OVN: Bring OpenStack Network Infra into Kubernetes. Presented at KubeCon + CloudNativeCon China 2019, Shanghai, China, 24–26 June 2019; Available online: https://www.youtube.com/watch?v=7gQJ4ZJ9Z8E (accessed on 30 October 2025).
  14. Networks-it-uc3m. L2S-M: Link-Layer Secure Connectivity for Microservice Platforms. 2023. Available online: https://github.com/Networks-it-uc3m/L2S-M (accessed on 30 October 2025).
  15. Martin, R.; Vidal, I.; Valera, F. A software-defined connectivity service for multi-cluster cloud native applications. Comput. Netw. 2024, 248, 110479. [Google Scholar] [CrossRef]
  16. Networks-it-uc3m. L2S-M MD: L2S-M Multi Domain. 2024. Available online: https://github.com/Networks-it-uc3m/l2sm-md (accessed on 30 October 2025).
  17. Peterson, L.; Cascone, C.; Davie, B. Software-Defined Networks a Systems Approach; Systems Approach LLC: Tucson, AZ, USA, 2021. [Google Scholar]
  18. Open Source MANO. What Is OSM? 2025. Available online: https://osm.etsi.org/ (accessed on 30 October 2025).
  19. OpenStack. OpenStack Software Features. Available online: https://www.openstack.org/software/ (accessed on 30 October 2025).
  20. Open vSwitch. What Is Open vSwitch? Available online: https://www.openvswitch.org/ (accessed on 30 October 2025).
  21. Amazon Web Services. Amazon Elastic Compute Cloud (EC2) Documentation. 2025. Available online: https://docs.aws.amazon.com/ec2/ (accessed on 30 October 2025).
  22. Google Cloud. Compute Engine Documentation. 2025. Available online: https://cloud.google.com/compute/docs (accessed on 30 October 2025).
  23. Hausenblas, M. Container Networking; O’Reilly Media, Incorporated: Newton, MA, USA, 2018. [Google Scholar]
  24. Flannle. Flannel: A Network Fabric for Containers, Designed for Kubernetes. Available online: https://github.com/flannel-io/flannel (accessed on 30 October 2025).
  25. TIGERA. What Is Calico. 2025. Available online: https://docs.tigera.io/calico/latest/about (accessed on 30 October 2025).
  26. TIGERA. About Calico Enterprise. 2025. Available online: https://docs.tigera.io/calico-enterprise/latest/about/ (accessed on 30 October 2025).
  27. Cilium. Cilium Documentation. Available online: https://docs.cilium.io/ (accessed on 30 October 2025).
  28. Qi, S.; Kulkarni, S.G.; Ramakrishnan, K. Understanding container network interface plugins: Design considerations and performance. In Proceedings of the 2020 IEEE International Symposium on Local and Metropolitan Area Networks (LANMAN), Orlando, FL, USA, 13–15 July 2020; pp. 1–6. [Google Scholar]
  29. Kube-OVN Project. Cluster Inter-Connection with OVN-IC. 2025. Available online: https://kubeovn.github.io/docs/v1.13.x/en/advance/with-ovn-ic/ (accessed on 30 October 2025).
  30. Submariner. 2025. Available online: https://submariner.io/ (accessed on 30 October 2025).
  31. Li, W.; Lemieux, Y.; Gao, J.; Zhao, Z.; Han, Y. Service mesh: Challenges, state of the art, and future research opportunities. In Proceedings of the 2019 IEEE International Conference on Service-Oriented System Engineering (SOSE), San Francisco, CA, USA, 4–9 April 2019; pp. 122–125. [Google Scholar]
  32. Istio. Istio Documentation. Available online: https://istio.io/latest/docs/ (accessed on 30 October 2025).
  33. Linkerd. 2025. Available online: https://linkerd.io/ (accessed on 30 October 2025).
  34. Network Service Mesh. Network Service Mesh Documentation. Available online: https://networkservicemesh.io/ (accessed on 30 October 2025).
  35. Institute for Internet Technologies and Applications. Advanced Features of Linux strongSwan: The OpenSource VPN Solution. Proceedings of LinuxTag 2005, Karlsruhe, Germany, 22–25 June 2005; University of Applied Sciences Rapperswil (HSR): Rapperswil-Jona, Switzerland, 2005. Available online: https://docs.strongswan.org/docs/5.9/_attachments/LinuxTag_Karlsruhe_2005.pdf (accessed on 30 October 2025).
  36. Gonzalez, L.F.; Vidal, I.; Valera, F.; Lopez, D.R. Link layer connectivity as a service for ad-hoc microservice platforms. IEEE Netw. 2022, 36, 10–17. [Google Scholar] [CrossRef]
  37. Amazon Web Services. Available online: https://aws.amazon.com/ (accessed on 30 October 2025).
  38. Google Cloud. Available online: https://cloud.google.com/ (accessed on 30 October 2025).
  39. Kubebuilder. Architecture Concept Diagram. Available online: https://book.kubebuilder.io/architecture (accessed on 30 October 2025).
  40. Docker Docs. Available online: https://docs.docker.com/ (accessed on 30 October 2025).
  41. Networks-it-uc3m. L2S-M Switch. 2025. Available online: https://github.com/Networks-it-uc3m/l2sm-switch (accessed on 30 October 2025).
  42. Schulzrinne, H.; Casner, S.; Frederick, R.; Jacobson, V. RTP: A Transport Protocol for Real-Time Applications; RFC 3550; RFC Editor: Fremont, CA, USA, 2003. [Google Scholar]
  43. Handley, M.; Schulzrinne, H.; Schooler, E.; Rosenberg, J. SIP: Session Initiation Protocol; RFC 2543; RFC Editor: Fremont, CA, USA, 1999. [Google Scholar]
  44. Tjaarda1. Architecting Multi-Cluster Layer 2 Connectivity for Cloud Native Network Slicing Test Repository. 2025. Available online: https://github.com/Tjaarda1/paper-slices-2025 (accessed on 30 October 2025).
  45. HashiCorp. Terraform. 2025. Available online: https://github.com/hashicorp/terraform (accessed on 30 October 2025).
  46. Ansible. 2025. Available online: https://github.com/ansible/ansible (accessed on 30 October 2025).
  47. Grafana. Grafana: The Open-Source Platform for Monitoring and Observability. 2025. Available online: https://github.com/grafana/grafana (accessed on 30 October 2025).
  48. Google. cAdvisor. 2025. Available online: https://github.com/google/cadvisor (accessed on 30 October 2025).
  49. Prometheus. Node Exporter. 2025. Available online: https://github.com/prometheus/node_exporter (accessed on 30 October 2025).
  50. Nmap. Nmap: The Network Mapper. 2025. Available online: https://github.com/nmap/nmap (accessed on 30 October 2025).
  51. Gayraud, R.; Jacques, O. SIPp Reference Manual; SourceForge: San Diego, CA, USA, 2004; Available online: https://sipp.sourceforge.net/doc/reference.html (accessed on 30 October 2025).
Figure 1. Overview of the network slice management solution.
Figure 1. Overview of the network slice management solution.
Futureinternet 18 00039 g001
Figure 2. Slice operation workflow.
Figure 2. Slice operation workflow.
Futureinternet 18 00039 g002
Figure 3. Architectural overview.
Figure 3. Architectural overview.
Futureinternet 18 00039 g003
Figure 4. Use case scenario where each cluster represents an entity.
Figure 4. Use case scenario where each cluster represents an entity.
Futureinternet 18 00039 g004
Figure 5. Mean CPU usage (cores) by cluster.
Figure 5. Mean CPU usage (cores) by cluster.
Futureinternet 18 00039 g005
Figure 6. Packets per second during the isolation test in Submariner vs. L2S-CES.
Figure 6. Packets per second during the isolation test in Submariner vs. L2S-CES.
Futureinternet 18 00039 g006
Figure 7. Packets per second in Submariner vs. L2S-CES.
Figure 7. Packets per second in Submariner vs. L2S-CES.
Futureinternet 18 00039 g007
Figure 8. Call rate over time.
Figure 8. Call rate over time.
Futureinternet 18 00039 g008
Figure 9. Response time distribution.
Figure 9. Response time distribution.
Futureinternet 18 00039 g009
Table 1. Comparison with other networking solutions.
Table 1. Comparison with other networking solutions.
Network ToolTypeFeatures
Isolated communicationEncryptionLow performance impactCustom path selectionMulticast supportService discoveryKubernetes integrationComponent interop.Multi-platform supportAutomated configurationMetric exposureSlice management
Calico CNI plugin----
Cilium †*CNI plugin w/service mesh-----
Multus †*CNI plugin------
Kube-OVN †*CNI plugin--
Istio †*Service mesh----
Linkerd †*Service mesh-----
Submariner †*L3 networking tool---
Kubeslice †*L3 slicing tool----
L2S-M †*L2 networking solution---
L2S-CES †*L2 slicing solution
⬤ = provides property; ◐ = partially provides property; - = does not provide property; has an academic publication or white paper; * end-user tool available.
Table 2. Cluster roles and components used in the evaluation testbed. Managed clusters have one control plane node and two worker nodes; the gateway is hosted on one of the workers.
Table 2. Cluster roles and components used in the evaluation testbed. Managed clusters have one control plane node and two worker nodes; the gateway is hosted on one of the workers.
ClusterNodesControl Plane
Components
Per-Node
Components
Gateway-Node
Components
l2sces-control1l2sces-client, sco-provider, l2sces-dns
l2sces-managed-13L2S-CES Operator, SDN ControllerSPS per nodeNED on 1 worker
l2sces-managed-23L2S-CES Operator, SDN ControllerSPS per nodeNED on 1 worker
sub-control1Submariner broker
sub-managed-13Lighthouse, Globalnet, Submariner operatorRoute agent per nodeSubmariner gateway on 1 worker
sub-managed-23Lighthouse, Globalnet, Submariner operatorRoute agent per nodeSubmariner gateway on 1 worker
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Buning, A.T.d.C.; Vidal, I.; Valera, F. Architecting Multi-Cluster Layer-2 Connectivity for Cloud-Native Network Slicing. Future Internet 2026, 18, 39. https://doi.org/10.3390/fi18010039

AMA Style

Buning ATdC, Vidal I, Valera F. Architecting Multi-Cluster Layer-2 Connectivity for Cloud-Native Network Slicing. Future Internet. 2026; 18(1):39. https://doi.org/10.3390/fi18010039

Chicago/Turabian Style

Buning, Alex T. de Cock, Ivan Vidal, and Francisco Valera. 2026. "Architecting Multi-Cluster Layer-2 Connectivity for Cloud-Native Network Slicing" Future Internet 18, no. 1: 39. https://doi.org/10.3390/fi18010039

APA Style

Buning, A. T. d. C., Vidal, I., & Valera, F. (2026). Architecting Multi-Cluster Layer-2 Connectivity for Cloud-Native Network Slicing. Future Internet, 18(1), 39. https://doi.org/10.3390/fi18010039

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop