1. Introduction
Wireless sensor networks have played a crucial role in communication in the past few decades due to their unique features (e.g., mobility and ease of connection), which make them a major carrier of data across networks [
1,
2,
3]. Consequently, the communication revolution embraces various standard wireless networks. Numerous types of wireless networks exist, and mobile wireless communication is one of the most widespread because it offers high data bandwidth. Mobile wireless communication utilizes a transport layer to transfer data and launches a specific protocol for data transmission. The transport layer protocol uses a congestion control mechanism to ensure good network resource allocation, and the network operates optimally even when demands reach the maximum network resources and size. One of the most critical events of the transport layer protocol is congestion control over wireless networks. If this event is used in a mobile wireless communication platform such as 5G, and does not satisfy the requirements of the platform, then the performance of the wireless network will deteriorate.
Congestion control shortage is one of the main causes of the decrease in the lifetime and energy of wireless sensor network (WSN) nodes [
4,
5]. This decrease leads to many other issues, such as delay, packet loss, and bandwidth degradation [
6]. Network resources include buffer size, queue size, throughput, link bandwidth, and packet loss of intercommunication between nodes [
7]. Congestion control influences diverse applications, such as event-based, continuous sensing, query-driven, and hybrid applications [
8]. Congestion control is applied in different protocols in the transport layer, such as the transmission control protocol (TCP) and stream control transmission protocol (SCTP) [
9]. Many conditions, such as packet collisions, interference, channel contention, and buffer overflow, require congestion control [
4], whereas slow links, buffer size shortage, and slow processors do not need congestion control [
7].
Many attempts have been made to enhance—while increasing the pace of—congestion control mechanisms, such as the proposal of a scalable distributed cluster-based congestion control scheme. In this scheme, the nodes are grouped in many clusters, where each cluster estimates the optimal path through message exchange within the cluster. However, this process may result in extra overhead along with energy consumption for head nodes within clusters [
10]. An optimal model for network load level was proposed based on 3-bit representation; the model produced good convergence results and effective bandwidth allocation [
11]. Measures have also been proposed to address issues in resource allocation, bandwidth delay, queue size, and high throughput [
12], and a set of parameters based on channel load has been developed for distributed congestion control [
13]. A previous study proposed a new algorithm to maximize the utilization of a link in data networking as well as overcoming data loss when congestion occurs [
14,
15]. A machine learning approach was used to improve congestion control, and the approach produced a promising result for the practical and oracular measurements of path properties [
16,
17]. Another study developed a novel method (expert framework) to estimate the end-to-end round-trip time using a machine learning approach. The simulation of this expert framework based on machine learning could adapt to changes in round-trip time, thereby reducing the number of transmitted packets while increasing the throughput [
18]. As a part of machine learning, clustering was used in congestion control for a vehicular ad-hoc network that included three parts, namely, detecting congestion, clustering messages, and controlling data congestion. Machine learning plays a vital role in improving congestion control [
19,
20].
The 5G environment is applied in many fields and utilized for many purposes, such as smart cities [
21], e-Health [
22], and environmental surveillance [
23]. Many approaches have been applied to the 5G environment in order to enhance the network performance [
24]. The machine learning approach is effective and widely applicable. Thus, it can be implemented with the conventional congestion control mechanism to satisfy the demands of 5G Internet of Things (IoT) nodes/sensors networks. The aim of applying the machine learning method to congestion control is to predict the optimal parametric setting and path to send/receive data. Finding the optimal node reduces data traffic and packet loss and improves throughput. In this work, a machine learning approach is utilized for prediction based on support vector regression using two parameters, namely, file transfer history and the measurement of simple path properties. This study presents a new methodology for improving the congestion control mechanism based on the machine learning decision tree (DT) approach.
The rest of this paper is organized as follows.
Section 2 presents related work on machine learning utilization for network optimization and prediction, congestion control mechanisms, and studies regarding congestion control improvement.
Section 3 explains the mathematical model of the proposed congestion control method.
Section 4 discusses the experimental results and demonstrates the effect of specific performance factors and the use of DT prediction.
Section 5 presents the conclusions and directions for future work.
2. Related Work
Regarding the approaches to congestion control in WSN IoT highlighted by many studies, Sangeetha et al. [
25] proposed a reduction of energy and data loss because of congestion across the network. Mainly, the topology of sensor nodes is attuned periodically at a regular time interval and node degree in order to improve the power consumption of sensor nodes, the interference, and provide an energy-efficient congestion-aware routing technique for WSNs—namely, survivable path routing (SPR). IoT applications use this protocol in networks with high traffic, since multiple sources are simultaneously trying to send their packets to a destination, as shown by Elappila et al. [
26]. Singh et al. [
27] developed a new congestion control algorithm for WSNs, where the conventional algorithms had high power usage and high complexity, and obtained the optimal rate by retransmission with congestion control, using a simple Poisson process. The routing protocol to select the optimized route proposed by Shelke et al. [
28], based on opportunistic theory and by integrating suitable sleep scheduling mechanisms in order to decrease congestion in the network, increases individual node life, the entire network lifetime, and decreases division in the network. Godoy et al. [
29] investigated and analyzed the environments that lead to congestion of the communication channel based on node configuration parameters: transmission time intervals, data packet generation rate, and transmitter output power level.
The congestion control mechanism of the SCTP determines its performance in various environments. Many studies have mentioned that conventional SCTP congestion control algorithms over Long Term Evolution-Advanced (LTE-A) present poor performance in terms of various factors, such as congestion window size (cwnd), throughput (throu), and send/receive/lose packets. Thus, Najm et al. [
30] proposed an improvement mechanism called IMP-SCTP for 4G networks based on multi-criteria decision making (MCDM). Meanwhile, Najm et al. [
31] described the conventional SCTP congestion control mechanism over a high-bandwidth network such as 4G as having weak performance through a hands-on and systemic review related to SCTP congestion control. New techniques must be adopted to enhance the congestion control scheme. One of the most vital studies involved the improvement of the congestion control of TCP in wired/wireless networks by applying a supervised machine learning technique; the authors built an automatic loss classifier from a database obtained by simulations of random network topologies [
32]. However, the simulation scenario was poor and simple. Furthermore, a fair and efficient distributed congestion control algorithm for tree-based communication in WSNs that seeks to adaptively assign a transmission rate to each node was developed as stated by Brahma et al. [
33]. However, the validation process did not compare the proposed algorithm to a standard/traditional method or a related previous study. In research published by Jagannathan and Almeroth [
34], the authors proposed a new model for multicast congestion control called TopoSense. However, a multicast sub-tree needed to be developed, and high congestion control at high rates was caused by dropping. Moreover, the link capacity calculation was poor, the interval size needed to be calculated, bursty traffic occurred, and control traffic needed to be minimized. Sunny et al. [
5] proposed a generic controller that ensures the fair and efficient operation of Wireless Local Area Network (WLANs) with multiple co-channel access points, long-lived performance, TCP transfers in multi-AP WLAN, and improves performance issues of long-lived TCP transfers in multi-AP WLANs.
Subsequently, many studies have adopted the DT concept in networks for various applications. For example, Katuwal et al. [
35] proposed new multiple-classifier systems for multi-class classification problems. The proposed model is a combination of DTs (random forest and oblique random forest) with an efficient neural network (random vector functional link network). The result of model evaluation on 65 multi-class datasets was better than those on medium- and large-sized datasets. The model experiment on the multi-class datasets produced good results, and the model could perform similarly on large- and medium-sized datasets. The model considers the problem of multi-class classification and provides the same training datasets for two-class datasets. The result of the proposed ensemble of classifiers was better than those of other classifiers. Furthermore, Gómez et al. [
36] compared the performance of various ensemble algorithms based on DT and presented a new classifier. The computational capacity of small network devices was not a crucial limitation.
Hasan et al. [
37] conducted a review of a machine-to-machine communication technology related to the system model architecture anticipated by diverse standards-developing organizations and prospective machine-to-machine communication challenges and their intended state-of-the-art solutions. Leng et al. [
38] proposed a new system for solving the flow table congestion problem in a software-defined network (SDN) based on a DT algorithm (C4.5). C4.5 was used to compress the flow entries in the data flow table in order to reduce the flow matching cost and time. This machine-learning-based online process in the SDN that is the first model to utilize a DT with a flow table. Then, a new trend of localization based on adaptive cluster splitting and AP reselection was proposed; Liang et al. [
39] presented a new fingerprint-based indoor wireless LAN that adopts adaptive cluster splitting and access point reflection. However, only received signal strength (RSS) RSS was considered for fingerprinting, a large area of wireless technologies was not applied, and many other factors were not considered (e.g., packet loss, information of channel state, queue size, throughput). Following this trend, research published by Liu and Wu [
40] used the machine learning approach (random forest algorithm) to predict traffic congestion. Variables such as weather conditions, road quality, time period, and type of day were used to build the model. However, many other factors, such as time and type of day (holiday), influence traffic congestion, make the prediction easy, and eliminate the need for prediction.
Following the trend of models based on Bayesian neural networks, Park et al. [
41] created a model based on a Bayesian neural network with gradient-boosted DTs to predict secondary incident occurrences. The model aims to reduce potential accidents by alerting drivers about events that cause accidents. However, similar to many other models, this model cannot be applied in real-time systems that predict secondary incidents because the model needs real parameters to train and build the prediction results. Following the trend of implementing methods for link quality estimation based on DT and support vector machine (SVM) to propose an improved route of data delivery over WSNs, Shu et al. [
42] utilized two parameters for estimation—namely, receive signal strength and link quality indicator. DT was combined with SVM in this model because SVM can only handle problems with binary classification. For energy-saving solutions for data centers and network infrastructures, a solution that is not a base technique was proposed in [
43]. Following the trend of using DTs, Soltani and Mutka [
44] presented a scheme using DT cognitive routing to determine the optimal path for sending data over a mesh cognitive radio network. The model allows a node to select the best candidate after analyzing the result tree and eliminate choices that decrease the node gain. Mesh cognitive networks and cognitive routing fields require analysis and intensive research to determine the critical points.
Furthermore, Soltani and Mutka [
44] used the DT approach to translate cognitive video routing in a dynamic radio network. Background induction was used from the ending branches to find the optimal path to the root node in order to receive and construct the sent video. Following the trend of using machine learning DT techniques, Adi et al. [
45] built a model to prevent denial of service over the hypertext transfer protocol (HTTP2). SVM, DT, naïve Bayes, and JRip DT were used for feature selection to determine important factors that influence false alarms. Using machine learning, the results demonstrated that the model is better than other models. Stimpfling et al. [
46] proposed a method based on the DT algorithm in order to overcome the drawback of memory access as well as the size of the data structure. DT is considered as the optimizing point in the model because it reduces the searching time. Furthermore, Singh et al. [
47] presented a new model to predict the noise of vehicular traffic. Four methods were used to build the model: DT, neural network, random forest, and the generalized linear model. The authors found that the random forest DT approach produced better prediction results than the other methods. To improve decryption efficiency over vehicular ad-hoc networks, Xia et al. [
48] proposed a delegation scheme (CP-ABE) based on DT. DT was used to optimize the joined factors utilized for estimating the overhead of vehicle decryption. The results showed that DT had higher accuracy and precision than SVM and K-nearest neighbor and was thus the best choice for prediction.
The authors of the botnet attacker detector in SDNs [
49] deployed a machine learning method regardless of the packet payload reading, but the stats plan required extensive computation. Wu et al. [
50] presented a scalable machine learning (ML) approach that detects and predicts video freezes online by using features extracted from network-based monitoring data; the proposed scheme determines discriminative features based on multi-scale windows by using the criterion of information gain (IG), but requests of audio segments are ignored, user interactivity is restricted, the cache between the delivery nodes and the video player is ignored, and the segment length is fixed. Chen et al. [
51] presented an online data model for traffic condition prediction that is based on online open data by using DT and SVM; however, the model selected to predict a 30 min traffic condition and the former two services exceeded the set value. Abar et al. [
52] predicted the quality of experience based on full-reference parameters Structural Similarity Index (SSIM) and Video Quality Model. (VQM), but they did not explain why k = 9 in a random forest was the best.
Many studies have focused on network security by using the DT concept. For example, Pham and Yeo [
53] proposed a secure and flexible framework for vehicles to manage trust and privacy, but the applied flexible recognition was unclear. Following the trends of reducing the amount of data transfer from smart meters, Mohamed et al. [
54] presented an adaptive framework for minimizing the amount of data transfer from smart meters, but the proposed framework did not clarify the effect of the number of model updates. Next, Kong, Zang, and Ma [
17] proposed two machine learning schemes based on TCP congestion control for under-buffered bottleneck links in wired networks with a compromise loss predictor and reinforcement to achieve a better tradeoff of delay. However, the environment used was wired. Taherkhani and Pierre [
19] proposed a localized and centralized data congestion control approach to control data congestion employed in roadside units at intersections. The proposed approach comprises three parts for controlling data congestion, detecting congestion, and clustering messages. Thus, the K-means machine learning method was proposed to detect data congestion in channels over a VANET. In research published by Fadlullah et al. [
20], the authors emphasized and demonstrated the necessity of overviewing the scattering mechanisms on deep learning applications for several network traffic control features, the state-of-the-art deep learning architectures and algorithms related to network traffic control systems, and also the potential enablers of deep learning for network systems. However, this was without a perspective of 5G environment utilization. Furthermore, authors in [
45,
55] proposed techniques for detecting previously unknown attacks on networks by identifying attack features. However, the relevance of window size and the attacks were not presented clearly.
Most of the previous studies frequently used decision tree machine learning to improve the network status without considering congestion control mechanisms over the recent environments (specifically the 5G IoT environment), even though DT presents a tree-based graph for prediction and classification. Therefore, DT should be developed for the prediction of the optimal congestion control mechanism among current congestion control transport protocol mechanisms in the 5G IoT environment. Despite these proposed methods, most studies only implemented machine learning capabilities in classifiers to improve network status without considering the current mechanisms required to adapt to new environments, such as advanced 4G and 5G, which could improve throughput and reduce end-to-end delay. Thus, 5G and the Internet of Things (IoT) represent a new era of applications and infrastructures that require the use of current mechanisms and protocols. Adapting to new environments is challenging in 5G, IoT, and current transport protocol mechanisms because transported protocols comprise the core of networks, which have unique characteristics compared with other protocols in mobile communication. The SCTP was introduced as the next-generation transport protocol to facilitate traffic in a way superior to conventional TCP and User Datagram Protocol (UDP). However, standard SCTP, TCP, and other transport protocols must be modified to suit the 5G environment. For these reasons, rapid response and reduced latency should be considered when improving a congestion control mechanism within 5G IoT networks. A demonstration of new approaches to overcome the limitation of current mechanisms should be considered to score the evaluation metrics of prediction and select the optimal techniques for congestion control as a means of achieving improved WSN performance in adjusting transport protocols for the improved support of 5G IoT environments.
The purpose of this work was to develop a new method for the prediction of congestion control mechanisms to adapt to 5G and evaluate the performance factors according to the machine learning capabilities for the SCTP. An analysis of the literature indicated several parameters for assessment/evaluation or utilization as evaluation metrics. However, in this research, we used the DT concept of machine learning capabilities during the development phase to predict the best I-cwnd for selecting the optimal threshold to adapt the SCTP to 5G IoT.
4. Performance Evaluation
In this section, we discuss the experiments we conducted using the millimeter-wave (mmWave) model in the network simulator ns-3 [
79,
80] to evaluate the performance of the proposed DT in predicting the optimal parametric settings of sensors’ transport protocols over 5G IoT. In the evaluation of the improved SCTP IMP-SCTP versus ORG-SCTP, we specifically focused on two parts, namely, the setup for machine learning and the simulation results. We also compared the findings with those of state-of-the-art DTs and predictions of optimal sensors’ SCTP congestion control mechanisms over 5G IoT. We conducted a paired-sample
t-test to show the significant variance between standard mechanism results and those of improved congestion control approaches.
4.1. Simulation Setup
We employed the mmWave ns–3 module and mmWave protocols to evaluate network protocols, such as TCP and SCTP throughout the 5G mmWave environment. The attributes were configured accordingly as needed in the simulation. The details of attributes concerning the mmWave are found in the file named “mmWaveAttributesList”, whereas the objects are illustrated in the file named “MmWaveHelper”, where the method is provided [
79,
80]. mmWave was stacked through ns-three endpoints that included “UEs” and “eNBs”. The UEs were connected to E-UTRAN Node B (eNB), where both TCP and SCTP generated traffic throughout connections. The bandwidth uplink was 500 Mbps, and the downlink was 1Gpbs where the “UEs” delay was 1 ms. Then, the message transmission unit (MTU) size was 1500 bytes and the chunk size was 1468 bytes. Next, each eNB was connected to a group of 20–30 UEs. The eNBs were linked to Serving Gateway (SGW) and the core gateway Packet Data Network Gateway (PGW), the bandwidth of the upload link was 5 Gbps, with 5 ms delay. Both bandwidth and delay for the download link had the same upload link. PGW was employed as an HTTP cache server. It was connected to the remote host over a bidirectional connection. The bandwidth connection was 10 Gbps with a 10 ms delay. The drop-tail was used for the queue management scheme and active queue management. The mmWave carrier frequency was 28 GHz, and the mmWave bandwidth was 1 GHz, as presented in
Figure 4.
4.2. Machine Learning Implementation
The model was implemented and tested using the Weka 3.8 tool. Weka 3.8 is an open-source machine learning tool that can be used to test and implement data mining tasks, such as prediction, analysis, and classification. Weka 3.8 offers wide flexibility in applying and listing predicted values. It enables the implementation of nearly all algorithms of data mining functions, such as clustering, association rules, regression, classification, and data pre-processing. The C4.5 DT algorithm is one of the commonly used algorithms for prediction and classification. C4.5 (J48) is used in Weka 3.8 for prediction and has proven accurate in different disciplines, such as education, medicine, sports, and economics. In our model, C4.5 was used to predict the optimal node in a 5G environment to help in congestion control. The first step in the model evaluation was DT/ML evaluation to determine if it covered all the training datasets and if the reader could visually determine the optimal node.
Figure 5 shows the optimum alternatives whereas
Figure 6 shows the result of the DT of congestion control based on the training dataset with predefined optimal nodes. The tree graph can provide the optimal node based on four factors (cwnd, que, throu, and pkt loss).
4.3. Effect of DT Prediction
The prediction concept is based on substituting a specific I-cwnd into the improved approaches. The selection of the optimal I is based on the conduct of the performance factors, where maximum cwnd, throu, and que represent the ideal outcome, in addition to a limited packet loss. Each I represents one of the I-alternative congestion control parameters. The branches below each node can define the conditions to determine the optimal and non-optimal nodes, where N refers to non-optimal and O refers to optimal in
Figure 5.
The DT prediction in
Figure 6 presents a tree graph after applying the J48 (C4.5) DT algorithm on the enhanced SCTP congestion control alternative. The root node is cwnd, and the arches represent the conditions of value comparison to determine the optimal node. The N leaf represents a non-optimal node, and optimal nodes are represented by “O”. If cwnd was less than or equal to 133, then the node was considered as N. Besides, the N represents the number of nodes that match the condition. If cwnd was greater than 133 and the queue size was less than or equal to 50,639, then six cases that matching this condition were considered as N nodes. If que was greater than 50,639 and throu was less than or equal to 9,881,687, then 70 cases were considered as N nodes. If throu was greater than 9,881,687 and pkt loss was less than or equal to 222, then the nodes would be considered as O.
Fourteen cases satisfied this condition, and two cases were misclassified. In the final leaf branch, when the packet loss was less than 22, eight cases were considered non-optimal, and one case was misclassified. Based on this road map of DT and the IF–THEN statement used to find the optimal node, the best optimal nodes among all nodes were 3 and 4. I = 3 had a cwnd of 145, a throu of 9,905,462, a que of 53,788, and a pkt loss of 24. I = 4 had a cwnd of 200, a throu of 10,120,778, a que of 57,674, and a pkt loss of 27. These nodes satisfied all conditions and achieved the optimal factors that determine the optimal node among all nodes. The roadmap in the figure is helpful for manually detecting the optimal nodes by dropping the parameter readings on the tree.
After 80% of the dataset was divided into a training set and the remaining 20% as test set, the model was reevaluated to find the prediction margin in order to prove the accuracy of the DT prediction model.
Figure 7 shows the margin score, which is a number that measures the accuracy of the predicted alternative, the prediction margin of each predicted node, and the accuracy measure of the prediction model. Margin scores range between 0 and 1; when the score is near 1 it means that the prediction is 100% accurate and when it nears 0 it indicates an incorrectly predicted alternative. Actual alternative (AA) represents whether an alternative is optimal or non-optimal. Predicted alternative (PA) represents the predicted case of the alternative after applying the DT algorithm. The figure demonstrates the accuracy of the DT prediction model, where the
y-axis shows the margin scope of the prediction margin, and the
x-axis represents the test alternatives by which the alternatives are shown as the actual alternatives and the predicted alternatives above them. These actual cases of alternatives were taken from the test set by which the alternatives were either O or N. The predicted alternatives represent the result of DT prediction by which the model predicted the alternative as either N or O. The prediction margin is a value that falls between 1 and −1. If the prediction value is 1, it means that the predicted value is 100% correct. When the value reaches 1, the prediction accuracy becomes perfect, and when the values reaches −1, it means that the predicted value is incorrect. The DT prediction model predicted the optimal node with 100% accuracy. The non-optimal alternatives were predicted with an average rate of 96%.
Table 3 illustrates a comparison among three selected machine learning decision tree algorithms based on six performance criteria (TP rate, FP rate, precision, recall, receiver operating characteristic (ROC), and precision–recall curve (PRC)). The performance results of decision tree algorithms (C4.5, RepTree, and random tree) are presented. Additionally, in the clustering field of machine learning, the performance results simple K means and hierarchical clustering algorithms are listed, where the final section of the table presents the performance results of stacking, where many classifiers can be combined to predict the results.
The true positive (TP) rate represents the percentage of correct predictions as positive cases, and the false positive (FP) rate represents the percentage of incorrect predictions as positive cases of different selected classifiers and machine learning algorithms. Precision is the result of all correctly predicted cases divided by all cases. Recall or sensitivity is the result of correct predictions divided by the total positive cases. The area under the receiver operating characteristic curve (ROC) and area under the precision–recall curve (PRC) represent the general performance of the machine learning algorithm. While the TP rate is considered as optimal when its value is equal to 1, the optimal FP rate is reached when it is equal to zero. Moreover, the ideal value of precision and recall is 1.0, the ideal value of recall and precision is 1.0, where the worst value is 0. The area under the ROC curve represents the relationship between recall and specificity, where specificity represents the precision of truly predicted cases as negative when they are negative. Area under the PRC is where the relationship between recall and precision is estimated. Higher rates of ROC and PRC represent a better classifier.
The congestion control dataset was divided into 10% of instances as a test dataset and 90% of all data as a training dataset. The experimental congestion control dataset was comprised of 13% optimal class and 87% non-optimal class. The extraordinary outcome of DT (C4.5) observed from
Table 3 shows that C4.5 was the best among the selected algorithms. The optimal TP rate is 1.0, and the TP rate of DT C4.5 was near 0.924—better than both RepTree and random tree. The closer the FP rate is to zero, the more optimal the result. The C4.5 DT FP rate satisfied this condition. Precision and recall values were optimistic because they were close to the optimal value of 1.0. In respect of the extraordinary DT C4.5 performances observed in
Table 3, the goal of this research to enhance the network performance by utilizing DT C4.5 was achieved. Additionally, the tree graph of the decision tree allows characterization with simplicity and ease, making it the best choice for analyzing prediction results. Furthermore, ROC was used to clarify the classifier accuracy. Some datasets may achieve a prediction of 92% of instances and result in a non-preferable classifier. The ROC curve can determine whether or not the classifier is appropriate for prediction. The nearer the curve is to 1.0, the better the classifier is.
Figure 8 shows that the model was suitable and reliable for prediction. The area under the ROC curve was 0.889, which is a reasonable and optimal value.
4.4. Effect of Delay
Simulations using network topology were conducted to clarify the effect of delay on SCTP congestion control. Delay represents how long it can take for a bit of data to be transmitted across the network from one node to the destination over a specific data link. Delay is presented in multiples or fractions of seconds.
Figure 9 displays the delay up to 50 s. The favorable performance of the proposed congestion control mechanism is clearly reflected in
Figure 9. The simulation time was set to 50 s. Given the lack of a congestion control mechanism in ORG-SCTP, the delay was prominent in the high-traffic-load environment. The total delay rates of SCTPs differed significantly within the period of 50 s. The performance of ORG-SCTP deteriorated (high delay indicates poor performance). The proposed congestion control of IMP-SCTP performed 14.5% better than that of ORG-SCTP. The performance of ORG-SCTP showed consistent degradation (
Figure 9). ORG-SCTP and IMP-SCTP experienced a slight increase in performance from 1 to 10 s. Beyond 10 s, the delay for ORG-SCTP declined continuously, whereas the delay for IMP-SCTP was less significant. Furthermore, the throughput for ORG-SCTP was minimized from 14 to 26 s and showed the highest drawback at 33–38 s. The delay of ORG-SCTP increased again and continued to fluctuate until the end of the experiment. Thus, IMP-SCTP outperformed ORG-SCTP in terms of delay from 11 to 35 s. IMP-SCTP showed a high delay rate at 45 and 47 s; nevertheless, the delay of IMP-SCTP was less than that of ORG-SCTP. The data in
Figure 9 were measured at a time step of 0.05 s. This period was derived from the overall simulation time. Furthermore, the sample was taken from one connection between a specific UE and an eNB across a high-load-traffic bottleneck. The result of the paired sample
t-test demonstrates the significant difference between the delays of ORG-SCTP and IMP-SCTP. The mean of the delay for ORG-SCTP (m = 1118.023) was greater than that for IMP-SCTP (m = 971.849). Again, the equation t(df) = t state,
p ≤ 0.05, was used to determine the significance of the difference between the two SCTPs, t(1275) = 2.93,
p ≤ 0.05. The
p-value was 0.003 at 50 s. Given that 0.003 ≤ 0.5, the difference between the delays of ORG-SCTP and IMP-SCTP was significant. Overall, the proposed enhancement improved the delay of IMP-SCTP.
4.5. Effect of Jitter
This part shows effect of jitter on the proposed SCTP congestion control. Jitter can be defined as the difference in delay of the received packets. The transmitting side sends packets in a continuous flow and spaces them equally as a variation in the delay of received packets. Due to congestion control, the received packets are either received or dropped. Jitter is used as an impacted utilization to evaluate the performance of the enhanced congestion control mechanism in the post-simulation stage. The average rate of jitter for each flow can represent the router resource management.
Figure 10 shows the difference between ORG-SCTP and IMP-SCTP.
This performance parameter was assessed for 180 s. The scale of milliseconds was used as a rate for evaluation. The two SCTPs were started at 0.3 s, and then jitter of ORG-SCTP was maximized from 0.4 to 9 s. It was observed that the performance of IMP-SCTP oscillated slightly but did not reach a higher level than ORG-SCTP. It was observed that although the performance of IMP-SCTP was unstable throughout the simulation period, it was better than that of ORG-SCTP by 64%. Although the performance of IMP-SCTP declined (the rising level in jitter indicates degrading performance), IMP-SCTP was still better than ORG-SCTP. The performance degradations of the latter were evident in the following periods: 13–18 s, 61–69 s, 133–143 s, and 163–172 s. Meanwhile, the lowest performance degradations were observed at 11, 20, 35, 46, 59, 70, 82, 92, 106, 118, 144, 161, 162, and 173 s. Moreover, the jitter utilization of IMP-SCTP increased at 3, 43, 52, 100, 108, and 178–180 s. The performance of IMP-SCTP was better and more stable than that of ORG-SCTP from 10 s to the end of the simulation period, and rarely reached the maximum level. Throughout the 180 s, the behaviors were independent of elapsed time. Since the behaviors were independent of time, the result records are reported in the 180 s. A paired-sample t-test was implemented to calculate the significant difference between jitter levels of IMP-SCTP and ORG-SCTP. The mean of jitter for IMP-SCTP (mean = 0.0017) was less than the mean of ORG-SCTP (mean = 0.005). To determine the significant difference, the equation t(df) = t state (p ≤ 0.05) was used again between two SCTPs, t(703) = 11.65, p ≤ 0.05. The p-values reached 0.0 at 180 s, and since 0.0 ≤ 0.5, the jitter level of IMP-SCTP and ORG-SCTP differed significantly. Therefore, the proposed enhancements improved IMP-SCTP’s jitter.
The improved congestion control based on DT C4.5 machine learning presented superior outcomes to original SCTP in the 5G environment. The new mathematical approaches that were incorporated for the prediction were developed for an improved congestion control approach to acquire the optimal values from the proposed equations.
Figure 11 provides the outcomes showing that the throughput, packet loss, cwnd, and queue size were enhanced by 8.5%, 16%, 75%, and 93%, respectively, as compared to conventional SCTP variants. In addition, the jitter was improved, being decreased by 64%, delay was decreased by 14.5%, and the packet send/receive was increased by 20%. It can be concluded that the proposed approach is efficient for wide-bandwidth networks such as 5G IoT.