TDTS: Three-Dimensional Trafﬁc Scheduling in Optical Fronthaul Networks with Conv-LSTM

: Given the more intensive deployments of emerging Internet of Things applications with beyond-ﬁfth-generation communication, the access network becomes bandwidth-hungry to support more kinds of services, requiring higher resource utilization of the optical fronthaul network. To enhance resource utilization, this study novelly proposed a three-dimensional trafﬁc scheduling (TDTS) scheme in the optical fronthaul network. Speciﬁcally, large and mixed trafﬁc data with multiple different requirements were ﬁrstly divided according to three-dimensions parameters of trafﬁc requests, i.e., arriving time, transmission tolerance delay, and bandwidth requirements, forming eight types of trafﬁc model. Then, historical trafﬁc data with division results were put into convolutional-long short-term memory (Conv-LSTM) strategy for trafﬁc prediction, obtaining a clear trafﬁc pattern. Next, the trafﬁc processing order was supported by a priority evaluation factor that was measured by trafﬁc status of the link and network characteristics comprehensively. Finally, following the priority, the proposed TDTS scheme assigned the resource to trafﬁc requests according to their results of trafﬁc division, prediction, and processing order with the shortest path routing and ﬁrst-ﬁt spectrum allocation policies. Simulation results demonstrated that the proposed TDTS scheme, on the premise of accurate trafﬁc prediction, could outperform conventional resource-allocation schemes in terms of blocking probability and resource utilization. their results of trafﬁc division, prediction, and processing priority. Simulation results demonstrated that the proposed TDTS scheme provided a lower blocking probability and higher resource utilization in the fronthaul network compared with conventional resource-allocation schemes.


Introduction
With the evolution of the fifth generation (5G) or even beyond 5G (B5G) of mobile communication, the number of machine-to-machine (M2M) applications is growing within world-wide IoT development, which contributes to the increase in device-access connections in an exponential way in a network [1][2][3]. According to the Ericsson Mobility Reports, the number of mobile terminals in the world will reach 8.3 billion by 2024, and 45% of cellular traffics in this case are expected to be created by Internet of Things (IoT) devices [4][5][6], with the penetration of IoT applications and services in our daily life, as well as in industrial scenarios [7,8]. In this case, it is a significant situation that the access traffics from various devices naturally form diverse traffic types with different requirements, where various connection services spring up with the requirement of delay, bandwidth, high-speed access, and so on [9][10][11]. However, given the situation with more intensive deployments [12], even B5G would find it hard to support current bandwidth-hungry network scenarios if there is not flexible resource allocation for services to avoid spectrum wastage.
To cope with the flexibility of B5G traffic, access networks evolve to a new fronthaul network architecture including a centralized unit (CU), distributed unit (DU), and active antenna unit (AAU) [13][14][15]. In this architecture, the traditional BBU baseband portion is split into two logical network elements, i.e., the CU and DU, and part of the baseband of arriving time determining the required resources of traffic was allocated to this traffic immediately or reserved, and delay-sensitive traffics were processed in DUs while CUs were selected to serve delay-tolerable traffics. Then, historical traffic data with division results were put into an artificial neural network, i.e., Conv-LSTM strategy, for traffic prediction to obtain a clear traffic pattern in the TDTS scheme. Based on this, the routing and spectrum assignment of TDTS was processed according to a priority-evaluation factor, which measured comprehensively the current status of link and traffic characteristics with the shortest path routing and first-fit spectrum allocation policies. Simulation results indicated that the proposed TDTS scheme could provide lower blocking probability and higher resource utilization, compared with conventional resource-allocation schemes.
This paper is organized as follows. Section 2 introduces the B5G fronthaul network and artificial-intelligence traffic prediction approach. In Section 3, The three-dimensional traffic scheduling scheme is proposed with the three-dimensional traffic division and traffic priority definition. The results of simulation and discussion are presented in Section 4. Finally, we conclude the paper in Section 5.

System Model Design
This section introduces the system model and artificial-intelligence traffic prediction approach that were used in this study.

Fronthaul Network Architecture Design
The new access network, architecture for the fronthaul network, consisted of three stand-alone devices, i.e., AAU, DU, and CU [13]. The distributed DU unit and centralized CU pool had several DUs and CUs. Each DU could control one or more AAUs and the CU pool was responsible for one or more DUs. This separate deployment architecture is more adaptable to future network needs as the DU is responsible for completing protocol stack-processing functions with high real-time requirements, and the CU is responsible for completing the protocol stack-processing functions with low real-time requirements [34,35]. Figure 1 shows the architecture of the fronthaul network. The access network system contained the centralized CU pool, distributed DU unit, and locally deployed AAUs. Each AAU has a logic connection with DUs, and multiple DU connects with a CU. As shown in Figure 1, DUs are responsible for dealing with high real-time requirements while CUs are responsible for completing low real-time requirements. Apart from default signaling intercommunication and data transmission between DUs and AAUs, some signaling messages need encapsulating and sending to CUs. In our study, the proposed TDTS scheme focused on complicated traffic processing, and different traffic flows were sent to DUs or CUs instead of sending all traffic to DUs. In this way, the fronthaul network was simplified, and network processing efficiency was increased. requests, where large and mixed traffic data with multiple different requirements from AAUs were combed according to three-dimensional characters, i.e., arriving time, transmission tolerance delay, and bandwidth requirements. In the traffic division progress, the character of arriving time determining the required resources of traffic was allocated to this traffic immediately or reserved, and delay-sensitive traffics were processed in DUs while CUs were selected to serve delay-tolerable traffics. Then, historical traffic data with division results were put into an artificial neural network, i.e., Conv-LSTM strategy, for traffic prediction to obtain a clear traffic pattern in the TDTS scheme. Based on this, the routing and spectrum assignment of TDTS was processed according to a priority-evaluation factor, which measured comprehensively the current status of link and traffic characteristics with the shortest path routing and first-fit spectrum allocation policies. Simulation results indicated that the proposed TDTS scheme could provide lower blocking probability and higher resource utilization, compared with conventional resource-allocation schemes.
This paper is organized as follows. Section 2 introduces the B5G fronthaul network and artificial-intelligence traffic prediction approach. In Section 3, The three-dimensional traffic scheduling scheme is proposed with the three-dimensional traffic division and traffic priority definition. The results of simulation and discussion are presented in Section 4. Finally, we conclude the paper in Section 5.

System Model Design
This section introduces the system model and artificial-intelligence traffic prediction approach that were used in this study.

Fronthaul Network Architecture Design
The new access network, architecture for the fronthaul network, consisted of three stand-alone devices, i.e., AAU, DU, and CU [13]. The distributed DU unit and centralized CU pool had several DUs and CUs. Each DU could control one or more AAUs and the CU pool was responsible for one or more DUs. This separate deployment architecture is more adaptable to future network needs as the DU is responsible for completing protocol stackprocessing functions with high real-time requirements, and the CU is responsible for completing the protocol stack-processing functions with low real-time requirements [34,35]. Figure 1 shows the architecture of the fronthaul network. The access network system contained the centralized CU pool, distributed DU unit, and locally deployed AAUs. Each AAU has a logic connection with DUs, and multiple DU connects with a CU. As shown in Figure 1, DUs are responsible for dealing with high real-time requirements while CUs are responsible for completing low real-time requirements. Apart from default signaling intercommunication and data transmission between DUs and AAUs, some signaling messages need encapsulating and sending to CUs. In our study, the proposed TDTS scheme focused on complicated traffic processing, and different traffic flows were sent to DUs or CUs instead of sending all traffic to DUs. In this way, the fronthaul network was simplified, and network processing efficiency was increased.

Conv-LSTM Approach
For general sequence modeling, CNN and LSTM, as traditional deep-learning methods, have been applied [21,36]. CNN is a non-linear activation function applied to the results of convolutional operation, where a full connection layer is used after the pooling operation for classification [37]. The core of convolutional operation is called filtering, i.e., kernel function, which completes feature extraction by sliding from top to bottom and from left to right in the original matrix. LSTM is a special recurrent neural network (RNN) structure that is stable and powerful for natural language processing tasks. Given the door mechanism and cell state for information storage, LSTM consists of three gates, i.e., forget gate, input gate, and output gate. The forget gate decides the information deleted from the cell state and the input gate inputs information to update the cell state. Then, the output gate decides the final output.
Conv-LSTM is the combination of CNN and LSTM, the core essence of which is similar to LSTM, taking the output of the previous layer as the input of the next layer. Note that to extract the feature of inter-relationship among the trained data, a data-convolution process is required, and, hence, the CNN was adopted in this study to capture the data-correlation character. However, after introducing the convolution operation, Conv-LSTM could not only obtain the temporal data relationship, but also extract the spatial features of traffic, such as the convolution layer, at the same time. In this way, the interaction between each state pair in Conv-LSTM was replaced by convolution operation, where the Conv-LSTM process is shown as Equations (1)- (5).
where σ is the Sigmoid activation function, * is convolution operation, and is the Hadamard product. i t , f t , and o t , are the input gate, forget gate, and output gate, respectively. c t is the storage unit, which could accumulate the unit status and update the status in real time.

Three-Dimensional Traffic Scheduling Scheme
For complicated traffic processing and resource allocation for traffic, this study proposed the three-dimensional traffic scheduling (TDTS) scheme to analyze the characteristics of different traffics and allocate resources for traffic. Figure 2 shows the supporting network structure of the proposed TDTS scheme. An internal switch could redirect traffic from AAUs, the processor could conduct three-dimensional traffic division, and the centralized controller could schedule network resources according to three-dimensional traffic-division results [38]. We assumed that the traffic from AAUs was sent separately to different DUs or CUs after three-dimensional traffic division. As DUs and CUs can synchronously process different types of traffic, the resource-allocation efficiency and connection response latency could be improved by using the proposed TDTS scheme.

Three-Dimensional Traffic Division
In our proposed TDTS scheme, a large amount of traffic was divided into eight types with three-dimensional parameters including arriving time, delay, and spectrum consumption. In terms of arriving time, we classified the traffic via two patterns as traffic that arrives immediately and the other that arrives after a while. In this case, resources were determined to allocate resources immediately or reserve resources according to the arriving time of traffic. Generally, different traffics have different tolerances to processing delay such as video, text, and satellite remote sensing. Therefore, the second characteristic of traffic was the tolerance difference on transmission delay, and the traffic was classified as delay-sensitive traffic or delay-tolerant traffic. In this case, traffic could be processed in DUs or CUs according to sensitivity, as DUs handle traffic that is sensitive to latency while CUs handle traffic tolerant to delay. The third characteristic was bandwidth requirements of traffic, where different traffic requires different sizes of spectrum resource. Figure 3 depicts the traffic division with three-dimension parameters for eight types of traffic.

Conv-LSTM Based Traffic Prediction
In the proposed TDTS scheme, the main content of prediction included the processing of raw data, the segmenting of training data, the training of the model, and the prediction of future traffic. For the sake of a high accurate prediction, the convolutional layer and LSTM were combined to generate a Conv-LSTM model for sequence modeling, which could fully fuse the temporal characteristics and the spatial characteristics of traffic flow in the adjacent region of the prediction point [21]. The applied Conv-LSTM model is shown in Figure 4, where a long period of historical traffic with the parameter of access node, arriving time, transmission tolerance delay, and bandwidth requirements was the input of the Conv-LSTM traffic prediction model, and the forecasting traffic feature for the next period was

Three-Dimensional Traffic Division
In our proposed TDTS scheme, a large amount of traffic was divided into eight types with three-dimensional parameters including arriving time, delay, and spectrum consumption. In terms of arriving time, we classified the traffic via two patterns as traffic that arrives immediately and the other that arrives after a while. In this case, resources were determined to allocate resources immediately or reserve resources according to the arriving time of traffic. Generally, different traffics have different tolerances to processing delay such as video, text, and satellite remote sensing. Therefore, the second characteristic of traffic was the tolerance difference on transmission delay, and the traffic was classified as delay-sensitive traffic or delay-tolerant traffic. In this case, traffic could be processed in DUs or CUs according to sensitivity, as DUs handle traffic that is sensitive to latency while CUs handle traffic tolerant to delay. The third characteristic was bandwidth requirements of traffic, where different traffic requires different sizes of spectrum resource. Figure 3 depicts the traffic division with three-dimension parameters for eight types of traffic.

Three-Dimensional Traffic Division
In our proposed TDTS scheme, a large amount of traffic was divided into eight types with three-dimensional parameters including arriving time, delay, and spectrum consumption. In terms of arriving time, we classified the traffic via two patterns as traffic that arrives immediately and the other that arrives after a while. In this case, resources were determined to allocate resources immediately or reserve resources according to the arriving time of traffic. Generally, different traffics have different tolerances to processing delay such as video, text, and satellite remote sensing. Therefore, the second characteristic of traffic was the tolerance difference on transmission delay, and the traffic was classified as delay-sensitive traffic or delay-tolerant traffic. In this case, traffic could be processed in DUs or CUs according to sensitivity, as DUs handle traffic that is sensitive to latency while CUs handle traffic tolerant to delay. The third characteristic was bandwidth requirements of traffic, where different traffic requires different sizes of spectrum resource. Figure 3 depicts the traffic division with three-dimension parameters for eight types of traffic.

Conv-LSTM Based Traffic Prediction
In the proposed TDTS scheme, the main content of prediction included the processing of raw data, the segmenting of training data, the training of the model, and the prediction of future traffic. For the sake of a high accurate prediction, the convolutional layer and LSTM were combined to generate a Conv-LSTM model for sequence modeling, which could fully fuse the temporal characteristics and the spatial characteristics of traffic flow in the adjacent region of the prediction point [21]. The applied Conv-LSTM model is shown in Figure 4, where a long period of historical traffic with the parameter of access node, arriving time, transmission tolerance delay, and bandwidth requirements was the input of the Conv-LSTM traffic prediction model, and the forecasting traffic feature for the next period was

Conv-LSTM Based Traffic Prediction
In the proposed TDTS scheme, the main content of prediction included the processing of raw data, the segmenting of training data, the training of the model, and the prediction of future traffic. For the sake of a high accurate prediction, the convolutional layer and LSTM were combined to generate a Conv-LSTM model for sequence modeling, which could fully fuse the temporal characteristics and the spatial characteristics of traffic flow in the adjacent region of the prediction point [21]. The applied Conv-LSTM model is shown in Figure 4, where a long period of historical traffic with the parameter of access node, arriving time, transmission tolerance delay, and bandwidth requirements was the input of the Conv-LSTM traffic prediction model, and the forecasting traffic feature for the next period was the output of this model. In this case, it was sufficient to obtain the timing relationship, and it was also possible to extract features including delay and spectrum like a convolutional layer. To optimize the Conv-LSTM model, the Adam algorithm was adopted in the TDTS scheme, which could adjust the learning rate adaptively based on the first-moment mean value and make full use of the second-order-moment mean of the gradient. the output of this model. In this case, it was sufficient to obtain the timing relationship, and it was also possible to extract features including delay and spectrum like a convolutional layer. To optimize the Conv-LSTM model, the Adam algorithm was adopted in the TDTS scheme, which could adjust the learning rate adaptively based on the first-moment mean value and make full use of the second-order-moment mean of the gradient.  (6) and (7).
If there is no special explanation, all { }treated in the following part of this paper are simplified as { }. Then, in the process of prediction, both training sequence forward and backward were two independent Conv-LSTMs that were connected to an output layer, which jointly provided complete past-and future-context information as an input sequence for each point in the output layer.
The proposed TDTS was to predict the mean value of traffic in the next period of time, which had the length of T. Assuming that previous traffic within the length of , the total traffic needed to be segmented to D traffic data sections, , , … , , 1,2, … , . In Conv-LSTM, represents the output value of the network for the jth pair of samples, and represents the actual value of previous traffic. Taking the input value and derived the prediction of future value . Finally, the accurate value of T was taken to obtain a predicted sequence, calculated by Equation (8).
, , , where the average of was regarded as the prediction, denoted by P, for the mean value of the traffic within the next period, which was calculated by Equation (9).
In the prediction process, the latest data were added to continue training the Conv-LSTM model, and in this case, it could meet the real-time requirements of the network. First, a long period of historical traffic {x t }, t = 1, 2, . . . , N needed to be transformed into a series of operable values for the Conv-LSTM model, where the process can be depicted as shown in Equations (6) and (7).
If there is no special explanation, all {x i } treated in the following part of this paper are simplified as {x i }. Then, in the process of prediction, both training sequence forward and backward were two independent Conv-LSTMs that were connected to an output layer, which jointly provided complete past-and future-context information as an input sequence for each point in the output layer.
The proposed TDTS was to predict the mean value of traffic in the next period of time, which had the length of T. Assuming that previous traffic within the length of D × T, the total traffic needed to be segmented to D traffic data sections, S j = x j , x j+T , . . . , x j+D×T , j = 1, 2, . . . , T. In Conv-LSTM, O j represents the output value of the network for the jth pair of samples, and x j+D×T represents the actual value of previous traffic. Taking the input value S j and O j derived the prediction of future value x j+(D+1)×T . Finally, the accurate value of T was taken to obtain a predicted sequence, calculated by Equation (8).
where the average of S T was regarded as the prediction, denoted by P, for the mean value of the traffic within the next period, which was calculated by Equation (9).
In the prediction process, the latest data were added to continue training the Conv-LSTM model, and in this case, it could meet the real-time requirements of the network. In the resource-allocation process, as a new request arrived, the proposed TDTS scheme established a connection and assigned a proper amount of resource for traffic with the priority of different types of traffic to improve resource utilization and reduce the blocking probability.
In terms of the traffic order of waiting requests, we novelly defined a concept of a request queue priority (RQP) model to represent the urgent degree of traffic, including two features as a link-request queue priority (LRQP) and network-request queue priority (NRQP). The NRQP of predicted traffic decided the order of the traffic entering the link, while the LRQP of traffic in the network link decided processing order. Table 1 presents all notations and definitions for LRQP and NRQP in this paper.

Q(V, E)
A network topology is presented by Q(V, E), V presents the node set in the network topology, and E presents the link set. M The number of links in the network.
λ max , λ min The maximum occupied wavelength and minimum occupied wavelength in the link, of which ID = i is presented by λ i max , λ i min , respectively. K The number of network requests which exist between spectrum blocks being occupied by services in the selected link, of which ID = i is presented by K i .
The connection of which ID = j occupies B j spectral resources, and contains H j links, respectively. N The number of connections that exist in the network at the moment.
The arriving time and leaving time of request. t s The time sensitivity of the request. p The spectrum resources required to process the request. α LRQP , α NRQP The values of LRQP and NRQP, respectively. Based on the above notation, the value of LRQP (α LRQP ) and NRQP (α NRQP ) can be calculated by Equations (10) and (11), respectively.
Note that the values of α LRQP and α NRQP presented the priorities of requests in the waiting line, where a larger value meant greater priority of traffic request in the waiting line. The traffic request needed to be processed when the value of either LRQP or NRQP exceeded that of another request or a pre-determined threshold (TH) in the network, where the threshold was determined by the average value of ROP among the requests processed in CU in historical and future traffic via the prediction result. For example, if LRQP was larger than the others or a concerted threshold, the request would be processed in DU later; otherwise, it would be processed in CU. Figure 5 shows an outline of the whole procedure. In the process of responding to the request with priority, connections were ordered according to a priority level. Connections with top priority were the first to be checked if they could be shifted to another wavelength, and the bottom-priority connections were the last to be checked. The priority was decided by either bandwidth or path length. In the priority process, we needed to make sure that any existing request had one and only one chance to be checked if it could be shifted to another wavelength. In this way, the cost of calculation remained the same as common processes.

Priority Model for Traffic Request
In fronthaul networks, connections between AAUs and DUs have some fixed statistical properties. In terms of latency sensitivity of traffic, the service volume and traffic request sensitivity of each area are typically constant. In terms of time, it is the probability contribution of holding time and arriving interval time of connections. In terms of wavelength, the probability contributions are not likely to be fluctuated. Therefore, it was practical that we died some work with these properties of connections of the networks. Despite the complexity of networks that makes direct evaluation of the result of priority opaque for traffic prediction, we focused on the general statistical results that some networks with connections of certain properties yield, i.e., resource utilization and network blocking probability.
The proposed TDTS scheme took advantage of centralized and automated control and management of the EON data plane, where DUs and CUs interacted to obtain realtime messages of network-state, and required resource allocation. As shown in Figure 6, DUs and CUs collected traffic-requests messages from AAUs, and path lengths, link-spectrum utilization, and other features were measured to execute the TDTS scheme. For traffic from AAUs, DUs and CUs collected information on traffic priority, resource utilization, and topology abstraction to generate state-data for the centralized network controller. Then, according to traffic priority, as shown in Figure 6, T1, T3, T5, and T7 were sent to DUs, while T2, T4, T6, and T8 were sent to CUs. Traffics with a solid line, meaning traffic requests, were processed immediately, while traffics with a dotted line meant reserved resources for traffic request.  Figure 5. The procedure of the priority model for traffic requests.
In the process of responding to the request with priority, connections were ordered according to a priority level. Connections with top priority were the first to be checked if they could be shifted to another wavelength, and the bottom-priority connections were the last to be checked. The priority was decided by either bandwidth or path length. In the priority process, we needed to make sure that any existing request had one and only one chance to be checked if it could be shifted to another wavelength. In this way, the cost of calculation remained the same as common processes.

Priority Model for Traffic Request
In fronthaul networks, connections between AAUs and DUs have some fixed statistical properties. In terms of latency sensitivity of traffic, the service volume and traffic request sensitivity of each area are typically constant. In terms of time, it is the probability contribution of holding time and arriving interval time of connections. In terms of wavelength, the probability contributions are not likely to be fluctuated. Therefore, it was practical that we died some work with these properties of connections of the networks. Despite the complexity of networks that makes direct evaluation of the result of priority opaque for traffic prediction, we focused on the general statistical results that some networks with connections of certain properties yield, i.e., resource utilization and network blocking probability.
The proposed TDTS scheme took advantage of centralized and automated control and management of the EON data plane, where DUs and CUs interacted to obtain realtime messages of network-state, and required resource allocation. As shown in Figure 6, DUs and CUs collected traffic-requests messages from AAUs, and path lengths, linkspectrum utilization, and other features were measured to execute the TDTS scheme. For traffic from AAUs, DUs and CUs collected information on traffic priority, resource utilization, and topology abstraction to generate state-data for the centralized network controller. Then, according to traffic priority, as shown in Figure 6, T1, T3, T5, and T7 were sent to DUs, while T2, T4, T6, and T8 were sent to CUs. Traffics with a solid line, meaning traffic requests, were processed immediately, while traffics with a dotted line meant reserved resources for traffic request. The proposed TDTS scheme took path lengths, link-spectrum utilization, and other features into account to realize dynamic RMSA strategy with the shortest path routing and first-fit spectrum allocation (SPFF). The procedure of the proposed priority-based TDTS scheme is given in Algorithm 1. The complexity to search the shortest path for all requested traffics, assuming adopting the common Dijkstra's algorithm, was O(|N|·|V|·log|V|+|N|·|E|), where the V and E are the set of the node link in network, respectively, and N is the set of all requested traffics. Then, the complexity to check whether sufficient resource was available for the requested traffic in the shortest path was O(|N|·|C|·|V|·log|V|+|N|·|C|·|E|), where C is the set of usable resource in each link. The complexity to divide traffic into different types was O(|D|), where D is the set of traffic division-type candidates. The complexity to compare priority among all requested traffics was O(|N| 2 ). Thus, the overall complexity of the proposed resource-allocation strategy was O(|V|·|C|·|T|·|N| 3 ·log|V|+|E|·|C|·|T|·|N| 3 ). According to the probability of solution, the centralized network controller set up the corresponding lightpaths for corresponding traffic. Search the shortest path for { } .

5.
If there are enough resources in the path then 6.
The proposed TDTS scheme took path lengths, link-spectrum utilization, and other features into account to realize dynamic RMSA strategy with the shortest path routing and firstfit spectrum allocation (SPFF). The procedure of the proposed priority-based TDTS scheme is given in Algorithm 1. The complexity to search the shortest path for all requested traffics, assuming adopting the common Dijkstra's algorithm, was O(|N|·|V|·log|V|+|N|·|E|), where the V and E are the set of the node link in network, respectively, and N is the set of all requested traffics. Then, the complexity to check whether sufficient resource was available for the requested traffic in the shortest path was O(|N|·|C|·|V|·log|V|+|N|·|C|·|E|), where C is the set of usable resource in each link. The complexity to divide traffic into different types was O(|D|), where D is the set of traffic division-type candidates. The complexity to compare priority among all requested traffics was O(|N| 2 ). Thus, the overall complexity of the proposed resource-allocation strategy wasO(|V|·|C|·|T|·|N| 3 ·log|V|+|E|·|C|·|T|·|N| 3 ). According to the probability of solution, the centralized network controller set up the corresponding lightpaths for corresponding traffic.

Simulation and Result Analysis
This section estimates the effectiveness of traffic prediction by using different neural networks for traffic prediction, including the common CNN, LSTM, CNN-LSTM, and the presented Conv-LSTM. Then, the performances of the proposed TDTS scheme are evaluated in terms of blocking probability and resource utilization, compared with the conventional DNN-RMSA [39], shortest path routing, and first-fit RMSA (SPFF-RMSA) [26], and Knapsack [40]. The experiment platform was a multi-core server with 122.5 GHz Intel Core i5-7200U CPU cores, two NVIDA TITAN GPU cores, and 32 GB RAM to guarantee meeting the high requirements of the network. We ran Ubuntu 16.04 and code in the Theano framework using Python 3.5.

Traffic Prediction Verification
For traffic prediction, we used a Conv-LSTM model to achieve our object of 3D traffic combing and traffic prediction, where the data were collected by the State Key Laboratory data center in Beijing China [41,42]. We experimented the Conv-LSTM module with other neural networks, i.e., CNN, LSTM, and CNN-LSTM, in traffic prediction, as shown in Figure 7. All traffic prediction models were trained with 200 epochs, where 1, 1.13, 0.94, 0.89 were the execution time with normalization when using the Conv-LSTM, CNN-LSTM, LSTM, and CNN, respectively. CNN has the ability to represent learning and can perform shift-invariant classification on input traffics according to their hierarchical structure. LSTM shows good performance in solving long-term dependence problems. However, CNN and LSTM are one-sided for our research questions. CNN-LSTM comprises two divided modules to extract the temporal feature and spatial feature. Conv-LSTM is a combined module to extract temporal-spatial features of traffic-flow data. Figure 7 compares results of root mean square error (RMSE) and mean absolute error (MSE), showing that our applied Conv-LSTM was better than others in traffic prediction.

Simulation and Result Analysis
This section estimates the effectiveness of traffic prediction by using different neural networks for traffic prediction, including the common CNN, LSTM, CNN-LSTM, and the presented Conv-LSTM. Then, the performances of the proposed TDTS scheme are evaluated in terms of blocking probability and resource utilization, compared with the conventional DNN-RMSA [39], shortest path routing, and first-fit RMSA (SPFF-RMSA) [26], and Knapsack [40]. The experiment platform was a multi-core server with 122.5 GHz Intel Core i5-7200U CPU cores, two NVIDA TITAN GPU cores, and 32 GB RAM to guarantee meeting the high requirements of the network. We ran Ubuntu 16.04 and code in the Theano framework using Python 3.5.

Traffic Prediction Verification
For traffic prediction, we used a Conv-LSTM model to achieve our object of 3D traffic combing and traffic prediction, where the data were collected by the State Key Laboratory data center in Beijing China [41,42]. LSTM shows good performance in solving long-term dependence problems. However, CNN and LSTM are one-sided for our research questions. CNN-LSTM comprises two divided modules to extract the temporal feature and spatial feature. Conv-LSTM is a combined module to extract temporal-spatial features of traffic-flow data. Figure 7 compares results of root mean square error (RMSE) and mean absolute error (MSE), showing that our applied Conv-LSTM was better than others in traffic prediction.

Performance Comparisons of the Proposed TDTS Scheme in a Static Network Scenario
For a fixed number of traffic requests in resource-allocation strategy, we compared the performance of the proposed TDTS scheme with other RMSA schemes based on DNN strategy and SPFF-RMSA strategy, where NSFNET was adopted as the simulated network topology [43], as shown in Figure 8. Besides, traffic request, resource occupation, and link

Performance Comparisons of the Proposed TDTS Scheme in a Static Network Scenario
For a fixed number of traffic requests in resource-allocation strategy, we compared the performance of the proposed TDTS scheme with other RMSA schemes based on DNN strategy and SPFF-RMSA strategy, where NSFNET was adopted as the simulated network topology [43], as shown in Figure 8. Besides, traffic request, resource occupation, and link information are known in advance. RMSA based on DNN can perceive complex system status. By accumulating operational experience from repeated interactions with the target system, and by enhancing operations that bring higher returns, DNN can learn successful strategies. The SPFF-RMSA calculates resource allocation for all traffic requests with the shortest length of path with the first available spectrum resource in each used link, which are contiguous and continuous in links.
information are known in advance. RMSA based on DNN can perceive complex system status. By accumulating operational experience from repeated interactions with the target system, and by enhancing operations that bring higher returns, DNN can learn successful strategies. The SPFF-RMSA calculates resource allocation for all traffic requests with the shortest length of path with the first available spectrum resource in each used link, which are contiguous and continuous in links. We simulated these three schemes and compared their performance in terms of blocking probability that indicates the probability of traffic blocking on the link, as shown in Figure 9. We can see that the proposed TDTS scheme and the conventional DNN-based RMSA scheme performed similarly at the beginning of simulation. The benchmark SPFF-RMSA scheme may easily be blocked due to limited resources by using inefficient resource allocation, while the proposed TDTS scheme and the conventional DNN-based RMSA scheme could more effectively access resource-utilization state and allocate resources due to their adaptability. Obviously, the intelligent schemes, i.e., TDTS and DNN-RMSA, were not suitable for static resource-allocation strategies due to its static characteristics when the number of traffic requests was small. This is because a certain amount of data is required to train the prediction model in the beginning of the simulation. After that, as much of the available spectrum resources as possible will be left for subsequent service requests when adopting the two intelligent schemes. Based on this, more traffic requests will be effectively allocated. When the number of traffic requests exceeded 500,000, the proposed TDTS scheme was better than the conventional DNN-RMSA scheme and they were both better than benchmark SPFF-RMSA scheme with traffic requests rising. This is because the proposed TDTS scheme had a more accurate prediction in resource allocation than that of the conventional DNN-RMSA scheme.  We simulated these three schemes and compared their performance in terms of blocking probability that indicates the probability of traffic blocking on the link, as shown in Figure 9. We can see that the proposed TDTS scheme and the conventional DNN-based RMSA scheme performed similarly at the beginning of simulation. The benchmark SPFF-RMSA scheme may easily be blocked due to limited resources by using inefficient resource allocation, while the proposed TDTS scheme and the conventional DNN-based RMSA scheme could more effectively access resource-utilization state and allocate resources due to their adaptability. Obviously, the intelligent schemes, i.e., TDTS and DNN-RMSA, were not suitable for static resource-allocation strategies due to its static characteristics when the number of traffic requests was small. This is because a certain amount of data is required to train the prediction model in the beginning of the simulation. After that, as much of the available spectrum resources as possible will be left for subsequent service requests when adopting the two intelligent schemes. Based on this, more traffic requests will be effectively allocated. When the number of traffic requests exceeded 500,000, the proposed TDTS scheme was better than the conventional DNN-RMSA scheme and they were both better than benchmark SPFF-RMSA scheme with traffic requests rising. This is because the proposed TDTS scheme had a more accurate prediction in resource allocation than that of the conventional DNN-RMSA scheme. information are known in advance. RMSA based on DNN can perceive complex system status. By accumulating operational experience from repeated interactions with the target system, and by enhancing operations that bring higher returns, DNN can learn successful strategies. The SPFF-RMSA calculates resource allocation for all traffic requests with the shortest length of path with the first available spectrum resource in each used link, which are contiguous and continuous in links. We simulated these three schemes and compared their performance in terms of blocking probability that indicates the probability of traffic blocking on the link, as shown in Figure 9. We can see that the proposed TDTS scheme and the conventional DNN-based RMSA scheme performed similarly at the beginning of simulation. The benchmark SPFF-RMSA scheme may easily be blocked due to limited resources by using inefficient resource allocation, while the proposed TDTS scheme and the conventional DNN-based RMSA scheme could more effectively access resource-utilization state and allocate resources due to their adaptability. Obviously, the intelligent schemes, i.e., TDTS and DNN-RMSA, were not suitable for static resource-allocation strategies due to its static characteristics when the number of traffic requests was small. This is because a certain amount of data is required to train the prediction model in the beginning of the simulation. After that, as much of the available spectrum resources as possible will be left for subsequent service requests when adopting the two intelligent schemes. Based on this, more traffic requests will be effectively allocated. When the number of traffic requests exceeded 500,000, the proposed TDTS scheme was better than the conventional DNN-RMSA scheme and they were both better than benchmark SPFF-RMSA scheme with traffic requests rising. This is because the proposed TDTS scheme had a more accurate prediction in resource allocation than that of the conventional DNN-RMSA scheme.  Furthermore, we compared the resource utilization of two intelligent schemes, i.e., the proposed TDTS and the conventional DNN-based RMSA. According to Figure 10, it is evident that the proposed TDTS scheme outperformed the conventional DNN-RMSA scheme in the aspect of resource utilization. As RNN was used in the proposed TDTS scheme, it could reserve resources for traffic with low delay tolerance, which was more suitable in our situation. In resource allocation, spectrum resources were divided into smaller spectrum blocks, and when traffic requests were combing, time and spectrum resources could be more fully utilized, and more traffic could be allocated according to combing results in the network. Furthermore, we compared the resource utilization of two intelligent schemes, i.e., the proposed TDTS and the conventional DNN-based RMSA. According to Figure 10, it is evident that the proposed TDTS scheme outperformed the conventional DNN-RMSA scheme in the aspect of resource utilization. As RNN was used in the proposed TDTS scheme, it could reserve resources for traffic with low delay tolerance, which was more suitable in our situation. In resource allocation, spectrum resources were divided into smaller spectrum blocks, and when traffic requests were combing, time and spectrum resources could be more fully utilized, and more traffic could be allocated according to combing results in the network.

Performance Comparisons of the Proposed TDTS Scheme in a Dynamic Network Scenario
For dynamic-network resource allocation in fronthaul networks, we compared the performances of the proposed TDTS scheme with two other benchmark schemes. The first one was based on a Knapsack formulation to model the resource-assignment problem without traffic prediction, which used a mixed-integer linear problem (MILP) for realtime fluctuating traffic of mobile users. The second one was the SPFF-RMSA scheme. Based on this, the performances of the proposed TDTS scheme were evaluated in terms of blocking probability and resource utilization, shown in Figures 11 and 12, respectively.

Performance Comparisons of the Proposed TDTS Scheme in a Dynamic Network Scenario
For dynamic-network resource allocation in fronthaul networks, we compared the performances of the proposed TDTS scheme with two other benchmark schemes. The first one was based on a Knapsack formulation to model the resource-assignment problem without traffic prediction, which used a mixed-integer linear problem (MILP) for real-time fluctuating traffic of mobile users. The second one was the SPFF-RMSA scheme. Based on this, the performances of the proposed TDTS scheme were evaluated in terms of blocking probability and resource utilization, shown in Figures 11 and 12, respectively. Furthermore, we compared the resource utilization of two intelligent schemes, i.e., the proposed TDTS and the conventional DNN-based RMSA. According to Figure 10, it is evident that the proposed TDTS scheme outperformed the conventional DNN-RMSA scheme in the aspect of resource utilization. As RNN was used in the proposed TDTS scheme, it could reserve resources for traffic with low delay tolerance, which was more suitable in our situation. In resource allocation, spectrum resources were divided into smaller spectrum blocks, and when traffic requests were combing, time and spectrum resources could be more fully utilized, and more traffic could be allocated according to combing results in the network.

Performance Comparisons of the Proposed TDTS Scheme in a Dynamic Network Scenario
For dynamic-network resource allocation in fronthaul networks, we compared the performances of the proposed TDTS scheme with two other benchmark schemes. The first one was based on a Knapsack formulation to model the resource-assignment problem without traffic prediction, which used a mixed-integer linear problem (MILP) for realtime fluctuating traffic of mobile users. The second one was the SPFF-RMSA scheme. Based on this, the performances of the proposed TDTS scheme were evaluated in terms of blocking probability and resource utilization, shown in Figures 11 and 12, respectively.  We observe in Figure 11 that the proposed TDTS scheme always provided the lowest blocking probability compared with the benchmarks, i.e., Knapsack and SPFF-RMSA. This was because, in the fronthaul network, our TDTS could effectively predict traffic requests to reserve resources for future incoming traffic requests dynamically. Unfortunately, the SPFF-RMSA scheme reconfigures and virtualizes DU, and includes a centralized controller taking network change into account, presenting a comprehensive but complex study on the optimization problem. The Knapsack-based scheme is based on the result of a mixed-linear integer problem to allocate resources without virtualizing resources. In this case, the two involved benchmark resource-allocation schemes, i.e., Knapsack and SPFF-RMSA, had a higher blocking probability, since the resources with traffic features were out of consideration in resource allocation. On the contrary, during the virtual optical network mapping process, the proposed TDTS scheme considered both time and spectrum features of traffic requests. The blocking of links was minimized when allocating resources according to traffic type and link state. If the link was blocking, the centralized controller responded to traffic requests with other available link resources. Thus, the proposed TDTS scheme could reserve as many resources as possible for future incoming requests, which could effectively decrease blocking probability.
We also compared the performance in terms of resource utilization, as shown in Figure 12. It is obvious that the highest resource utilization was obtained when the proposed TDTS scheme was adopted. This resulted from the fact that resource reservation for future incoming traffic requests could efficiently use the available but limited spectrum resource in the network. A simple and untargeted resource allocation, such as Knapsack and SPFF-RMSA, causes a deterioration for spectrum usage efficiency.

Conclusions
To enhance resource utilization in the B5G scenario, this study proposed a three-dimensional traffic scheduling (TDTS) scheme for the optical fronthaul network, which was based on future-traffic data prediction and traffic division. In this scheme, large and mixed traffic data were firstly divided into eight types according to three-dimensional parameters including arriving time, transmission tolerance delay, and bandwidth requirements. Then, historical traffic data with division results were put into an artificial neural network for traffic prediction to obtain a clear traffic pattern, where convolutional-long short-term memory strategy was adopted. Next, based on the traffic division result, the future traffic processing order was supported by a priority-evaluation factor, which was measured by the current traffic status of link and network characteristics comprehensively. Finally, the proposed TDTS scheme assigned spectrum resource to traffic requests according to their We observe in Figure 11 that the proposed TDTS scheme always provided the lowest blocking probability compared with the benchmarks, i.e., Knapsack and SPFF-RMSA. This was because, in the fronthaul network, our TDTS could effectively predict traffic requests to reserve resources for future incoming traffic requests dynamically. Unfortunately, the SPFF-RMSA scheme reconfigures and virtualizes DU, and includes a centralized controller taking network change into account, presenting a comprehensive but complex study on the optimization problem. The Knapsack-based scheme is based on the result of a mixed-linear integer problem to allocate resources without virtualizing resources. In this case, the two involved benchmark resource-allocation schemes, i.e., Knapsack and SPFF-RMSA, had a higher blocking probability, since the resources with traffic features were out of consideration in resource allocation. On the contrary, during the virtual optical network mapping process, the proposed TDTS scheme considered both time and spectrum features of traffic requests. The blocking of links was minimized when allocating resources according to traffic type and link state. If the link was blocking, the centralized controller responded to traffic requests with other available link resources. Thus, the proposed TDTS scheme could reserve as many resources as possible for future incoming requests, which could effectively decrease blocking probability.
We also compared the performance in terms of resource utilization, as shown in Figure 12. It is obvious that the highest resource utilization was obtained when the proposed TDTS scheme was adopted. This resulted from the fact that resource reservation for future incoming traffic requests could efficiently use the available but limited spectrum resource in the network. A simple and untargeted resource allocation, such as Knapsack and SPFF-RMSA, causes a deterioration for spectrum usage efficiency.

Conclusions
To enhance resource utilization in the B5G scenario, this study proposed a threedimensional traffic scheduling (TDTS) scheme for the optical fronthaul network, which was based on future-traffic data prediction and traffic division. In this scheme, large and mixed traffic data were firstly divided into eight types according to three-dimensional parameters including arriving time, transmission tolerance delay, and bandwidth requirements. Then, historical traffic data with division results were put into an artificial neural network for traffic prediction to obtain a clear traffic pattern, where convolutional-long short-term memory strategy was adopted. Next, based on the traffic division result, the future traffic processing order was supported by a priority-evaluation factor, which was measured by the current traffic status of link and network characteristics comprehensively. Finally, the proposed TDTS scheme assigned spectrum resource to traffic requests according to their results of traffic division, prediction, and processing priority. Simulation results demonstrated that the proposed TDTS scheme provided a lower blocking probability and higher resource utilization in the fronthaul network compared with conventional resource-allocation schemes.