You are currently viewing a new version of our website. To view the old version click .
Electronics
  • Article
  • Open Access

6 February 2019

Regulating Scheduler (RSC): A Novel Solution for IEEE 802.1 Time Sensitive Network (TSN)

Department of Human Intelligence Information Engineering, Sangmyung University, Seoul 03016, Korea
This article belongs to the Special Issue Emerging Trends in Industrial Communication

Abstract

Emerging applications such as industrial automation, in-vehicle, professional audio-video, and wide area electrical utility networks require strict bounds on the end-to-end network delay. Solutions so far to such a requirement are either impractical or ineffective. Flow based schedulers suggested in a traditional integrated services (IntServ) framework are O(N) or O(log N), where N is the number of flows in the scheduler, which can grow to tens of thousands in a core router. Due to such complexity, class-based schedulers are adopted in real deployments. The class-based systems, however, cannot provide bounded delays in networks with cycle, since the maximum burst grows infinitely along the cycled path. Attaching a regulator in front of a scheduler to limit the maximum burst is considered as a viable solution. International standards, such as IEEE 802.1 time sensitive network (TSN) and IETF deterministic network (DetNet) are adopting this approach as a standard. The regulator in TSN and DetNet, however, requires flow state information, therefore contradicts to the simple class-based schedulers. This paper suggests non-work conserving fair schedulers, called ‘regulating schedulers’ (RSC), which function as a regulator and a scheduler at the same time. A deficit round-robin (DRR) based RSC, called nw-DRR, is devised and proved to be both a fair scheduler and a regulator. Despite the lower complexity, the input port-based nw-DRR is shown to perform better than the current TSN approach, and to bind the end-to-end delay within a few milliseconds in realistic network scenarios.

1. Introduction

Various emerging applications such as industrial automation, building automation, inter-vehicle network, in-vehicle network, professional audio video (AV) network, and wide-area electrical utility network require strict delay bounds, which range from a few milliseconds to a few seconds. IEEE 802.1 time sensitive network (TSN) and IETF deterministic network (DetNet) are representative standard groups for dealing such delay requirements on various networks. TSN aims guarantee end-to-end delay and zero packet loss on an Ethernet data plane [1]. IETF’s DetNet has a similar scope with TSN, but is biased on network layer issues [2]. DetNet can be seen as an extension of TSN into routed networks, with IP and MPLS data plane. DetNet claims its scope is limited to a single administrative domain such as campus-area network or private WAN, but does not restrict itself from expanding into wider area networks. DetNet defines the electrical utility networks using time division multiplexing currently to be one of the key use cases [2]. For example, the utility network in Quebec, Canada, which includes 60 generating stations, 513 substations, and 10,500 km of optical fiber, carries instantaneous electrical information (current, voltage, active power, etc.) and real-time commands such as trip, open/close relay, etc. This network has a mixed Layer 2 and Layer 3 topology, and requires the deterministic behavior from the network such as bounded delay/jitter and precise timing. By applying DetNet technology the utility networks can increase the electrical grid’s reliability and efficiency.
TSN and DetNet both cover a wide area of technological aspects, including packet forwarding, time synchronization, path selection, resource reservation, reliable transmission, etc. This paper focuses on the packet forwarding technology. The packet forwarding function includes switching, queueing, and scheduling of packets. The TSN task group (TG) was initiated as the residential Ethernet TG in the IEEE 802.1 working group (WG) around the year 2005 and expanded to the audio video bridge (AVB) TG. Both TGs considered the professional AV traffic transport among nodes in close proximity as the main target. In its early stage, the TGs set a clear performance goal to bind the delay within ”2 ms in seven hops”. This clear goal helped the TGs got the attention from the beginning. The AVB TG in 2012 evolved into the TSN TG with the extended scope for various industrial applications. The residential Ethernet and the AVB TG’s target, the AV applications have a small number of large bandwidth flows. A flow is defined to be a set of packets of the same source-destination processes pair, generated within a relatively limited amount of time. Meanwhile in TSN, the target applications also generate numerous flows with relatively small bandwidth with nodes in a complex, larger area network, such as industrial sensor-actuator networks. For example, IEC/IEEE 60802 TSN Profile for Industrial Automation, a joint project of IEC SC65C/MT9 and IEEE 802 to define TSN profiles for industrial automation, has defined the traffic types of industrial automation and control systems, at the November 2018 meeting. According to this definition, isochronous traffic with 100 byte maximum packet length requires less than 2 ms end-to-end delay, while cyclic traffic with 1000 byte maximum packet length requires 2 ms to 20 ms end-to-end delay. These end-to-end delay requirements will be the primary target of this paper.
There has been a tremendous amount of research activities regarding the end-to-end delay bound in large scale networks, since early 1990s [3,4,5,6,7,8,9,10,11]. While various approaches for such requirements have their pros and cons, meeting delay requirement and implementation simplicity have been incompatible with one another so far. We propose a novel solution that can perform comparatively with the state-of-the-art solutions that are considered as the standard for the TSN and the DetNet, but with a far less complexity.
Our key contribution is schedulers that also perform the regulating function at the same time. This set of schedulers, called the ‘regulating scheduler’ (RSC), are devised by giving a non-work conserving characteristics to fair schedulers. Then a deficit round robin (DRR) based RSC, called nw-DRR, is devised and proved to be both a fair scheduler and a regulator. Despite the lower complexity, the input port-based nw-DRR is shown to perform better than the current TSN approach, and to bind the end-to-end delay within a few milliseconds in realistic network scenarios. In seven-hop networks with a few crossing flows, the maximum end-to-end delay can be bound within 2 ms, if the packet length is limited to be less than 400 bit.
This paper is comprised as follows. The related works to the problem of delay bound are reviewed in Section 2. In Section 3 the RSC and its realization with DRR, nw-DRR, are introduced and formally analyzed. The performance of the nw-DRR is investigated through numerical analysis with an exemplary network topologies in the Section 4. The last section discusses the result, the remaining issues, and possible follow up research activities.

3. Proposed Scheduler and Its Analysis

3.1. Proposed Approach

This paper proposes a simpler alternative for end-to-end delay guarantee, the regulating scheduler (RSC). The RSC can be thought as a non-work conserving version of fair schedulers. In order to build an RSC, a fair scheduler is modified to make their queues always filled with a virtual packet if there is no actual packet waits for the service. The virtual packet receives the service but it just disappear after the service rather than be transferred to the next node. When an actual packet arrives, the virtual packet in that queue, if existing, is removed even if it was in the middle of a service. By this modification all the queues look backlogged and the queue under observation will be served at a steady pace. Therefore, the traffic served out from the queue under observation is strictly regulated. Intuitively, a burst increases at a node as follows. At a certain moment, a single queue is full of traffic while all the other queues are empty. The incoming burst to the queue arrives right before the queued burst’s service time. They merged together and served instantly, and becomes a larger burst. The RSC prevents the instant service of a large amount of traffic from a single queue by having the other backlogged queues.
We carefully devised a pure per input port RSC for the TSN environment. Its operation can be described as the following. The whole traffic is divided into high priority and low priority traffic. The high priority traffic includes CDT, Class A, and Class B. The low priority traffic includes the rest. At an output port of a relaying node such as switch and router the high priority traffic is divided based on their input ports and allocated to different queues. The low priority traffic is allocated to another queue at the output port. Therefore, if there are N input ports, (N+1) queues will be established. These queues are scheduled with a DRR scheduler [8] but based on the RSC principle. The algorithm detail will follow in Section 3.3. This scheduler is the main contribution of this paper, which is called nw-DRR (the non-work conserving DRR).
Figure 4 depicts the proposed input-port based nw-DRR’s architecture. It replaces the whole system depicted in Figure 3. One can easily see the simplicity of the proposed system’s architecture. We analyze and prove that the nw-DRR is a scheduler and a regulator at the same time in Section 3.3. We also prove that nw-DRR is a fair scheduler with the same {latency, service rate} parameters with the original DRR before modification. It is also shown that the input-port based nw-DRR can guarantee end-to-end delay within the required bounds of a few milliseconds.
Figure 4. Architecture of the proposed scheduler at an output port. The scheduler works as a regulator as well.

3.2. DRR Scheduler

DRR’s complexity depends heavily on its parameter, deficit value. Larger the deficit easier to implement, but with larger latency. DRR works as the follows [8].
  • Each queue is assigned a deficit value (δ) and a quantum value (φ).
  • Deficit value is variable but initialized to zero. Quantum value is fixed and determined to be proportional to the arrival rate of the traffic to the queue. For example, if arrival rates to two queues are 1 Mbps and 2 Mbps then quantum value of the latter queue should be twice to the one of the former. The quantum values of two queues can be set to be 10 byte and 20 byte, or 50 byte and 100 byte, respectively as well. It is shown that a smaller quantum gives a smaller latency [10]. A smaller quantum makes the operation of the scheduler slower, however.
  • Each backlogged queue takes a turn, and a complete cycle of turns of backlogged queues is called a frame ( F ).
  • At the start of each turn a queue increases the deficit value as much as the quantum value.
  • If the deficit value is larger or equal to the packet length at the head of the queue, then serve the packet. After service of the packet, the deficit is reduced as much as the packet length. Repeat the Step 5 as long as the condition holds.
  • If the queue becomes empty then reset deficit to zero.
  • Proceed to the next backlogged queue, and go to the Step 4.
The latency of a DRR server, in case the deficit value is smaller than the max packet length, is given as the follows [22].
Θ i D R R = 1 r [ ( F φ i ) ( 1 + L i φ i ) + n = 1 N L n ]
where F is the sum of all φ i of active flows in the server, and N is the number of active flows. The significant term in (6) is F and n = 1 N L n . F can be reduced by reducing deficit values locally, but sum of max packet lengths of each queue is not easily controlled.

3.3. DRR Scheduler as a Regulator

From now on, we will show that a DRR scheduler can work as a regulator, with a proper modification in line with the RSC principle. The following theorems show that a general DRR scheduler bounds the work given to a flow within a certain limit.
Theorem 1.
Assume that flow i is continuously backlogged during time period (a,b]. Assume that the flow is served k turns during (a,b]. The work given to the flow during this period is bounded by the following inequality.
k φ i δ i k W i ( a , b ) k φ i + δ i 0 ,
where δ i k is the deficit value at the finish time of kth round during (a,b].
Proof. 
See the proof of Theorem 4.2 in [8]. □
Theorem 2 generalizes the result of Theorem 1 without using k and the deficit value terms.
Theorem 2.
During any time period ( a , b ] within a flow i backlog period, the work given to the flow is bounded by the following inequality.
W i ( a , b ) ρ i F m a x f ( b a ) + φ i + L i ,
where f is the sum of quantum values of the flows that have been always backlogged during (a,b], and F m a x is the sum of quantum values of all the flows in the scheduler, in an imaginary situation where the sum of arrival rates of all the flows equals to the link capacity.
Proof. 
Let t k be the finish time of kth round, counting from a . Let t 0 = a . Duration of a round t k t k 1 = 1 r j B k ( φ j + δ j k 1 δ j k ) , where B k is the set of flows that are continuously backlogged during ( t k 1 , t k . During ( t k 1 , t k the elements of B k are not changed. Let B be the set of flows that are continuously backlogged during ( t 0 , t k . For any k, B B k . The following equation holds. t k t k 1 = 1 r j B ( φ j + δ j k 1 δ j k ) . By summing up the equation for all k, we get t k t 0 k f r + 1 r j B ( δ j 0 δ j k ) f r ( k 1 ) , since j B δ j 0 0 and j B δ j k f . Therefore k r f ( t k t 0 ) + 1 . From Theorem 1, during ( t 0 , t k .
W i ( t 0 , t k ) k φ i + δ i 0 r f ( t k t 0 ) φ i + φ i + δ i 0 ρ i F m a x f ( t k t 0 ) + φ i + L i ,
because L i δ i 0 and ρ i / r = φ i / F m a x . In other words, the deficit values cannot be larger than the maximum packet length L i , at all the round finish time instants including t 0 . Furthermore, the flow arrival rate is proportional to the quantum value of the flow, and F m a x is so to the link capacity r. Therefore, for arbitrary time instant b between t k and t k + 1 ,
W i ( a , b ) = W i ( a , t k ) ρ i F m a x f ( t k a ) + φ i + L i ρ i F m a x f ( b a ) + φ i + L i ,
and the theorem follows. □
From Theorem 2, if a DRR scheduler satisfies two conditions 1) the sum of arrival rates of flows is equal to the link capacity and 2) all the queues in the scheduler are always backlogged, then the work given to the scheduler is bounded as the following inequality.
W i ( a , b ) ρ i ( b a ) + φ i + L i ,
which is the inequality for the amount of work of a regulator with maximum burst size φ i + L i , or an arrival curve of a flow { ρ i , φ i + L i } where φ i + L i is the max burst size.
As the next step, we will introduce a modified DRR that satisfies the two conditions given above, and show that it indeed works as a regulator with inequality (11). Consider a DRR scheduler that is modified to work with the following algorithm.
  • Assign a queue for a virtual flow v , which has the arrival rate equals to the difference of the link capacity and the sum of the arrival rates of all the flows in the scheduler. ( ρ v = r i H ρ i .) Flow v will serve the low priority traffic, whose arrival rates are not predefined.
  • If a queue, including the queue for v , is empty then place a virtual packet in the queue immediately, so that all the queues are never empty. The length of the virtual packet is set to the quantum value of the queue. In other words, the virtual packet is sized such that it can be served in a single turn.
  • The virtual packet is served the same way with the actual packets. After the service virtual packets do not leave the scheduler, though.
  • Consider an actual packet arrives to the queue with a virtual packet. If the virtual packet was not in a service then remove the virtual packet. If it was during a service then immediately stop serving the virtual packet, set the deficit value to 0, and proceed to the next queue.
The above scheduler may serve a virtual packet even when actual packets are waiting in the other queues, therefore is non-work conserving. This is called the nw-DRR scheduler.

3.4. Properties of the nw-DRR

The following theorem proves the nw-DRR gives the same service curve, i.e.,
W i n w D R R ( t 0 , t ) m a x 0 , ρ i ( t t 0 Θ i D R R ) ,
with the original DRR scheduler that does not generate virtual packets.
Theorem 3.
The nw-DRR is an LR server with the same service rate and latency with the DRR server that does not generate any virtual packets.
Proof. 
Assume N flows arrive to an nw-DRR, thus there are N queues. It is sufficient to prove that for all the packets in an arbitrary flow i , i { 1 , , N } , there exists a work conserving DRR scheduler that serves later than or equal to the nw-DRR would. □
Let all the other flows, except the flow under observation i , have the same arrival rates with those with the nw-DRR but sufficiently large maximum burst sizes to make their queues always be backlogged during the flow i backlog period. Recall that the max burst size does not affect the latency of a flow with a DRR scheduler. Furthermore upon the start of the flow i backlogged period, let the service of the (i+1)th queue start. Let us call this work conserving server DRR*. Intuitively, DRR* gives the worst case DRR service to i in terms of packet service finish time.
Now we will come back to the nw-DRR scheduler. Let us call p k , k = 0 , 1 , , K , the packets arrive during a flow i backlog period. p 0 arrives to a queue with a virtual packet, but nw-DRR removes the virtual packet just before the p 0 ’s arrival. This is true even if that virtual packet was under service, in which case the (i+1)th queue’s service starts upon the p 0 ’s arrival. Therefore, p 0 does not see that it is non-work conserving. The subsequent packets p k , k = 1 , , K , arrive to a backlogged queue and do not see that it is non-work conserving, as well. Therefore, the service times of p k , k = 0 , 1 , , K , of the nw-DRR are always earlier than or equal to those of the DRR*. The theorem follows.
By Theorem 3, the nw-DRR is an LR server with the same {latency, service rate} parameters with the ordinary work-conserving DRR. Let i be a set of packets assigned to the i th queue in the nw-DRR. We will call i the flow aggregate. The maximum delay experienced by packets in a flow-aggregate i in an nw-DRR’s is given as
D i σ i L i ρ i + Θ i n w D R R .
Here, Θ i n w D R R is the latency of the nw-DRR server and equals to a work conserving DRR scheduler, given in (6).
Consider a node architecture with nw-DRR at the output ports, as depicted in Figure 5. Let us calculate the maximum burst of i q D + 1 . Let us denote it with σ q . i q D + 1 is the set of flows from the input port q to the output port under observation of the node (D+1). We argue that σ q σ I D , and σ I D p ( φ p + L p ) , therefore
σ q p ( φ p + L p ) .
Similarly, we can argue that σ p o ( φ o + L o ) , if the input port p is connected to an nw-DRR scheduler in the upstream node (D-1), with o representing the input ports to the output port under observation in the node (D-1).
Figure 5. Architecture of a relaying node with the nw-DRR scheduler at the output ports.
The first argument above ( σ q σ I D ) holds. For any t and t 0 , t > t 0 ,
A I D ( t 0 , t ) = q A i q D + 1 ( t 0 , t ) = q ( σ i q D + 1 + ρ i q D + 1 ( t t 0 ) ) = q σ i q D + 1 + q ρ i q D + 1 ( t t 0 ) = σ I D + ρ I D ( t t 0 ) ,   and q ρ i q D + 1 ( t t 0 ) = ρ I D ( t t 0 ) .
Therefore q σ i q D + 1 = σ I D , and σ q σ i q D + 1 σ I D , for any q. The second argument ( σ I D p ( φ p + L p ) ) also holds because of (11).
We can now calculate the delay of packets in i p D at the node D. The delay of i p D is bounded by (13), where ρ i , L i , and Θ i n w D R R can be easily obtained by letting i p D be the aggregate of all the flows traverse from the input port p to the output port under observation at the node D. The maximum burst of i p D , σ p , can be obtained similarly with (14).

4. Case Study

4.1. Case of a Network with a Cycle

In this subsection we investigate the performance of the nw-DRR with an exemplary network depicted in Figure 6, which was suggested in [19] for the study of the performance of the TSN asynchronous forwarding module in Figure 3. By adopting the already studied network topology, we can clearly compare the performance of the two solutions head to head. The flow under observation, f 1 , starts from the host 1, traverses the node 1, 2, 3, 4, and then sinks to the host 4. All the other flows from f 2 to f 5 have the same parameters with f 1 . Low priority traffic also exists and occupies the low priority traffic queue in each nw-DRR.
Figure 6. Network topology with a cycle, which is used for the case study in this paper and in [19].
We calculated the maximum end-to-end delay while varying the maximum packet length that applies the same to both high and low priority traffic, the max input burst size, the flow input data rate, and the quantum size, and fixing the link capacity, as in Table 2.
Table 2. Parameter values used for the case study.
The delay at a node is calculated with (13). At the first node f 1 and f 2 share a queue, since they enter the node with the same port. The maximum burst at the first node is equal to the max packet length. Therefore the delay of f 1 at the first node satisfies the inequality D f 1 1 Θ 1 1 , where Θ 1 1 = 1 r [ ( r ρ i 2 φ i ) ( 1 + L 2 φ i ) + 2 L ] from (6). The delay of f 1 at the nodes from 2 to 4 satisfies the inequality
D f 1 j σ 1 j L ρ i + Θ 1 j ,
where Θ 1 j = 1 r [ ( r ρ i φ i ) ( 1 + L φ i ) + 3 L ] with j is the node identifier that j = 2, 3, or 4. σ 1 j in (15) is the max burst into the queue of the output port of node j, to which f 1 belongs. σ 1 j is calculated with (14), and in our case
σ 1 j 2 ( φ i + L ) ,
for all j = 2, 3, or 4, since there are always two flows in the queue in the previous node j-1.
Figure 7 shows the maximum end-to-end delay of f 1 with varying max packet size from 400 bit to 3200 bit, flow arrival rate from 10 Mbps to 40 Mbps, and a fixed quantum value at 80 bit. It is shown that the max packet length and the input data rate are both significant parameters for the delay. The input data rate is inversely proportional to the max delay. One can consider to increase the input data rate during the connection setup phase, while actual amount of traffic remains unchanged, in order to reduce the max delay.
Figure 7. Max end-to-end delay with variable max packet size, variable flow arrival rate, and fixed quantum value 80 bit.
The Table 3 shows the exact values of the max end-to-end delay at sample parameter value sets.
Table 3. End-to-end delay bound at sample parameter sets (ms) with the fixed quantum value at 80 bit.
As in Table 3, with the quantum value 80 bit and flow arrival rate 10 Mbps, the maximum end-to-end delay is around 2 ms even when the max packet length is 400 Byte. If we can limit the max packet length to 50 Byte, the max delay reduces to 0.364 ms.
Next, we investigated how the max end-to-end delay changes when the quantum size varies. The smaller quantum size implies more computational burden to the scheduler.
Figure 8 shows that the quantum size does not significantly affects the delay. When the max packet length becomes larger, the effect of the quantum value becomes even less significant. One can choose to use a relatively large quantum value in case the complexity is one of key design issues, without sacrificing the delay performance too much.
Figure 8. Max end-to-end delay with variable max packet size, variable quantum value, and fixed flow arrival rate at 20 Mbps.
Next, we compare our performance with the ATS system in [19]. It is revealed that the ATS system shows the max end-to-end delay of 0.70 ms when the max packet length 1000 bit, flow arrival rate 20 Mbps [19]. The same parameter set gives 0.431 ms with quantum value 80 bit and 0.575 ms with quantum value 400 bit in our result, which can be found at Table 4. This result is summarized in Table 5.
Table 4. End-to-end delay bound at sample parameter sets (ms) with the flow arrival rate at 20 Mbps.
Table 5. End-to-end delay bound comparison with [19] (ms) with the maximum packet size 1000bit and the flow arrival rate at 20Mbps.
The performances of two different approaches are remarkably similar, while ours is slightly better. This is encouraging considering the simplicity of the proposed architecture.

4.2. Case of a Seven-Hop Network

In this subsection a network with a flow that traverses seven hops is considered. The early residential Ethernet TG defined the “2 ms delay bound in seven hops” as an ultimate performance goal for the strictest applications. We will investigate if this goal can be met with the proposed RSC. Consider the following topology depicted partly in Figure 9. There are 6 relaying nodes between the source and the destination of the flow, which makes up seven hops between the source and the destination. All the nodes have multiples input and output ports. The flow under observation enters the input port 1 and then departs from the output port 1, in every bridge node. A flow with the same specification with the flow under observation comes in at every other input port and then leaves at the output port 1. The set of such flows from all the input ports is called crossing flows. The crossing flows in a node shares the output ports with the flow under observation, but are separated to different output ports at the next node. This pattern continues at the subsequent nodes. This type of tandem networks is suitable for the investigation of the delay performance of schedulers with flow aggregation [23]. Although not specified in the figure, there exists low priority traffic at every output port. This setup of flows makes the network quite busy and large-scale.
Figure 9. Network topology of a series of relaying nodes with the presence of crossing flows.
As in Section 4.1, we set the parameters within a predefined range specified in Table 6. We fixed the values of ρ i , φ i . The number of crossing flows, N -1, varies from 1 to 8, which makes the number of ports, N , from 2 to 9. The number of queues in an RSC scheduler is then N +1, because of the low priority traffic queue.
Table 6. Parameter values used for the case study.
Let us call f 1 the flow under observation. The delay of f 1 at the nodes from 1 to 6 satisfies the inequality
D f 1 j σ 1 j L ρ i + Θ 1 j ,
where Θ 1 j = 1 r [ ( r ρ i φ i ) ( 1 + L φ i ) + ( N + 1 ) L ] with j is the node identifier that j = 1, …, 6. σ 1 j in (17) is the max burst into the queue of the output port of node j, to which f 1 belongs. σ 1 j is calculated with (14), and in our case
σ 1 j N ( φ i + L ) ,
for j = 2, …, 6, since there are always N high priority queues in the scheduler in the previous node j-1. The maximum burst at the first node is equal to the max initial burst, which is same with the packet length in our case. Therefore σ 1 1 = σ 1 = L .
Figure 10 shows the maximum end-to-end delay of f 1 , j = 1 6 D f 1 j with varying max packet size from 400 bit to 1600 bit, number of crossing flows from 1 to 8, fixed flow arrival rate at 10 Mbps, and a fixed quantum value at 80 bit.
Figure 10. Max end-to-end delay of a seven-hop network with variable max packet size, variable number of crossing flows, and a fixed arrival rate at 10 Mbps.
As depicted in Figure 10, even when the number of crossing flows is eight, which represents quite a busy network, the end-to-end delay bound can be near 2 ms if we set the maximum packet length to be less than 400 bit.
Table 7 shows that with the maximum packet size 400 bit, or 50 byte, even if the number of crossing flows at every node is 8, the maximum delay is 2.175 ms, which is around the IEEE TSN’s ultimate performance goal of the “2 ms with seven hops”.
Table 7. End-to-end delay bound (mc) at sample parameter sets with the fixed arrival rate at 10 Mbps.

5. Discussion

A method of building a regulating scheduler (RSC) from a fair scheduler is proposed in this paper, which has non-work conserving characteristic. Possibility of the RSC as a delay bounding module is investigated. Especially, the DRR scheduler is modified to be an RSC, the nw-DRR. The nw-DRR is proved to be an LR server, with the same {latency, service rate} parameters with the original DRR server. The nw-DRR is then applied to work as an input-port based scheduler. At an output port of a relaying node the high priority traffic is divided based on their input ports and allocated to different queues. The low priority traffic is allocated to another queue at the output port. Therefore, if there are N input ports, (N+1) queues will be established.
The input-port based scheduling prevents the ill-behaving flows from overwhelming the network, if applied at the boundary of a network. As such it can guarantee the fair resource sharing among all the flows in the high priority class, without per-flow scheduling. For example, if the first Ethernet switch from a flow’s source host employs an input port-based RSC at the output ports, then automatically the flow is regulated, therefore shares only a fair amount of network resources.
Low priority traffic also benefits from an RSC since they do not suffer from the exhaustion, the bandwidth monopoly by high priority traffic. Low priority traffic is assigned a single queue in an RSC and guaranteed to receive the remaining available bandwidth without any resource reservation process. This property fits well to the “best-effort” traffic class, literally, and is in line with the available bit rate (ABR) service paradigm defined in the ATM networks [24].
The performance of the nw-DRR was investigated with realistic network topologies. The first network examined has a cycle, so the class-based scheduler would not meet any delay bound requirement. It was shown that in over various network parameters range, the nw-DRR bounds the delay within a few milliseconds. The network topology was adopted from [19] to compare the performance of the nw-DRR with that of the ATS approach of TSN. While the two perform similarly, ours was slightly better in the given parameter set. Moreover, our architecture is simpler in two ways. First, it is purely input port-based. Second, the regulator and the scheduler are in one piece. Despite the simplicity, the performance was acceptable. It is shown that the max packet length and the input data rate are both significant parameters for the delay. The input data rate is inversely proportional to the max delay. One can consider to increase the input data rate during the connection setup phase, while actual amount of traffic remains unchanged, in order to reduce the max delay. The quantum size does not significantly affect the delay. When the max packet length becomes larger, the effect of the quantum value becomes even less significant. One can choose to use a relatively large quantum value in case the complexity is one of key design issues, without sacrificing the delay performance too much. In the second network examined, a flow under observation traverses seven-hops with a number of crossing flows. Even with eight crossing flows at every node, which represents quite a busy network, if we set the maximum packet length to be less than 400 bit, then the maximum end-to-end delay bound is around the IEEE TSN’s ultimate performance goal of “2 ms with seven hops”.
The non-work conserving nature of the proposed scheduler may increase the average delay. The guarantee on the delay bound, however, is crucial in emerging applications. Therefore, the RSC will play a key role in such application networks. Various fair schedulers can be modified to be an RSC. The DRR’s latency depends on the sum of max packet lengths of each queues ( n = 1 N L n ), which makes it too large in situations with many input ports, and thus many queues in the nw-DRR. The virtual-finish time based LR servers’ latency is usually dominated by the max packet length, not their sum, therefore will be more suitable to TSN. In future work, the possibility of modifying a virtual finish time based LR server to an RSC will be investigated. Although they are more complex, input port-based implementation will mitigate the complexity greatly.

Funding

This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (NRF-2016R1D1A1B03932066).

Conflicts of Interest

The author declares no conflict of interest.

References

  1. IEEE 802.1 Time-Sensitive Networking Task Group Home Page. Available online: http://www.ieee802.org/1/pages/tsn.html (accessed on 13 December 2018).
  2. IETF DetNet Working Group Home Page. Available online: https://datatracker.ietf.org/wg/detnet/documents/ (accessed on 13 December 2018).
  3. Braden, R.; Clark, D.; Shenker, S. Integrated Services in the Internet Architecture: An Overview. Available online: https://tools.ietf.org/html/rfc1633 (accessed on 13 December 2018).
  4. Parekh, A.; Gallagher, R.G. A generalized processor sharing approach to flow control in integrated services networks: The multiple node case. IEEE/ACM Trans. Netw. 1994, 2, 137–150. [Google Scholar] [CrossRef]
  5. Golestani, S.J. A Self-Clocked Fair Queueing Scheme for High Speed Applications. In Proceedings of the IEEE Conference on Computer Communications and Networking (INFOCOM), Toronto, ON, Canada, 12–16 June 1994. [Google Scholar]
  6. Zhang, H.; Ferrari, D. Rate-controlled service disciplines. J. High Speed Netw. 1994, 3, 389–412. [Google Scholar]
  7. Xie, G.; Lam, S. Delay guarantee of virtual clock server. IEEE/ACM Trans. Netw. 1995, 3, 683–689. [Google Scholar] [CrossRef]
  8. Shreedhar, M.; Varghese, G. Efficient fair queueing using deficit round-robin. IEEE/ACM Trans. Netw. 1996, 4, 375–385. [Google Scholar] [CrossRef]
  9. Georgiadis, L.; Guerin, R.; Peris, V.; Sivarajan, K.N. The effect of traffic shaping in efficiently providing end-to-end performance guarantees. Telecommun. Syst. 1996, 5, 71–83. [Google Scholar] [CrossRef]
  10. Stiliadis, D.; Varma, A. Latency-Rate Servers: A General Model for Analysis of Traffic Scheduling Algorithms. IEEE/ACM Trans. Netw. 1998, 6, 611–624. [Google Scholar] [CrossRef]
  11. Joung, J.; Choe, B.-S.; Jeong, H.; Ryu, H. Effect of Flow Aggregation on the Maximum End-to-End Delay. In Proceedings of the High Performance Computing and Communications, Munich, Germany, 13–15 September 2006; Volume 4208, pp. 426–435. [Google Scholar]
  12. IEEE 802.1 Time-Sensitive Networking Task Group P802.1Qch—Cyclic Queuing and Forwarding. Available online: https://1.ieee802.org/tsn/802-1qch/ (accessed on 13 December 2018).
  13. Dürr, F.; Nayak, N.G. No-wait Packet Scheduling for IEEE Time-sensitive Networks (TSN). In Proceedings of the 24th International Conference on Real-Time Networks and Systems, Brest, France, 19–21 October 2016; pp. 203–212. [Google Scholar] [CrossRef]
  14. Specht, J.; Samii, S. Urgency-Based Scheduler for Time-Sensitive Switched Ethernet Networks. In Proceedings of the 28th Euromicro Conference on Real-Time Systems (ECRTS), Toulouse, France, 5–8 July 2016; pp. 75–85. [Google Scholar]
  15. Blake, S.; Black, D.; Carlson, M.; Davies, E.; Wang, Z.; Weiss, W. An Architecture for Differentiated Service, IETF RFC 2475. Available online: https://tools.ietf.org/html/rfc2475 (accessed on 13 December 2018).
  16. Le Boudec, J.-Y. Some properties of variable length packet shapers. IEEE/ACM Trans. Netw. 2002, 10, 329–337. [Google Scholar] [CrossRef]
  17. Gao, F.; Qian, H. Efficient, real-world token bucket configuration for residential gateways. IEEE/ACM Trans. Netw. 2016, 24, 462–475. [Google Scholar] [CrossRef]
  18. Thangamuthu, S.; Concer, N.; Cuijpers, P.J.L.; Lukkien, J.J. Analysis of Ethernet-switch traffic shapers for in-vehicle networking applications. In Proceedings of the Design, Automation & Test in Europe Conference & Exhibition (DATE), Grenoble, France, 9–13 March 2015; pp. 55–60. [Google Scholar]
  19. Mohammadpour, E.; Stai, E.; Mohiuddin, M.; Le Boudec, J.-Y. Latency and Backlog Bounds in Time-Sensitive Networking with Credit Based Shapers and Asynchronous Traffic Shaping. In Proceedings of the 30th International Teletraffic Congress (ITC 30), Vienna, Austria, 4–7 September 2018. [Google Scholar]
  20. Zhao, L.; Pop, P.; Zheng, Z.; Li, Q. Timing analysis of AVB traffic in TSN networks using network calculus. In Proceedings of the 2018 IEEE Real-Time and Embedded Technology and Applications Symposium (RTAS), Porto, Portugal, 11–13 April 2018; pp. 25–36. [Google Scholar]
  21. De Azua, J.A.R.; Boyer, M. Complete modelling of AVB in network calculus framework. In Proceedings of the 22nd International Conference on Real-Time Networks and Systems, Versaille, France, 8–10 October 2014; pp. 55–64. [Google Scholar]
  22. Lenzini, L.; Mingozzi, E.; Stea, G. Tradeoffs between low complexity, low latency, and fairness with deficit round-robin schedulers. IEEE/ACM Trans. Netw. 2004, 12, 681–693. [Google Scholar] [CrossRef]
  23. Bouillard, A.; Stea, G. Exact worst-case delay for FIFO-multiplexing tandems. In Proceedings of the 6th International Conference on Performance Evaluation Methodologies and Tools (VALUETOOLS), Cargese, France, 9–12 October 2012; pp. 158–167. [Google Scholar]
  24. Imer, O.C.; Compans, S.; Basar, T.; Srikant, R. Available bit rate congestion control in ATM networks. IEEE Control Syst. Mag. 2001, 21, 38–56. [Google Scholar]

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.