You are currently on the new version of our website. Access the old version .
ElectronicsElectronics
  • Article
  • Open Access

13 April 2022

Machine-Learning-Based Uplink Throughput Prediction from Physical Layer Measurements

,
and
1
Electrical and Electronics Engineering, Faculty of Engineering and Arts, Nevsehir Haci Bektas Veli University, Nevsehir 50300, Turkey
2
Collage of Engineering and Technology, American University of Middle East, Egaila 54200, Kuwait
3
Electrical and Electronics Engineering, Faculty of Engineering, Sirnak University, Sirnak 73000, Turkey
4
Computer Engineering and Sciences, Florida Institute of Technology, Melbourne, FL 32901, USA
This article belongs to the Section Networks

Abstract

The uplink (UL) throughput prediction is indispensable for a sustainable and reliable cellular network due to the enormous amounts of mobile data used by interconnecting devices, cloud services, and social media. Therefore, network service providers implement highly complex mobile network systems with a large number of parameters and feature add-ons. In addition to the increased complexity, old-fashioned methods have become insufficient for network management, requiring an autonomous calibration to minimize utilization of the system parameter and the processing time. Many machine learning algorithms utilize the Long-Term Evolution (LTE) parameters for channel throughput prediction, mainly in favor of downlink (DL). However, these algorithms have not achieved the desired results because UL traffic prediction has become more critical due to the channel asymmetry in favor of DL throughput closing rapidly. The environment (urban, suburban, rural areas) affect should also be taken into account to improve the accuracy of the machine learning algorithm. Thus, in this research, we propose a machine learning-based UL data rate prediction solution by comparing several machine learning algorithms for three locations (Houston, Texas, Melbourne, Florida, and Batman, Turkey) and determine the best accuracy among all. We first performed an extensive LTE data collection in proposed locations and determined the LTE lower layer parameters correlated with UL throughput. The selected LTE parameters, which are highly correlated with UL throughput (RSRP, RSRQ, and SNR), are trained in five different learning algorithms for estimating UL data rates. The results show that decision tree and k-nearest neighbor algorithms outperform the other algorithms at throughput estimation. The prediction accuracy with the R2 determination coefficient of 92%, 85%, and 69% is obtained from Melbourne, Florida, Batman, Turkey, and Houston, Texas, respectively.

1. Introduction

Broadband data usage is facing a massive demand with the availability of the fourth-generation (4G) and evolving fifth-generation (5G) wireless technology. These technologies promise higher spectral efficiency, low latency, high peak data rates, and flexibility in frequency and bandwidth to the end-user [1]. Hence, it is crucial to maintain the Quality of Service (QoS), while guaranteeing reliable and robust networks with achieving higher data rates. In wireless communications, data rates have increased exponentially from second generation (2G/GSM) (data speed up to 9.6 Kbps) to 4G/LTE (data speed up to 300 Mbps). LTE employs 20 MHz bandwidth with different modulation and multiplexing techniques such as Orthogonal Frequency Division Multiple Access (OFDM), 64 QAM, and 4 × 4 spatial multiplexing to achieve higher data rates [2,3]. Higher-resolution video and audio content have become widespread with reliable and fast mobile internet access. The demand for cellular data doubles every two years and is expected to reach 77.5 exabytes by 2022 [4]. Under the umbrella of the Internet of Things (IoT), massive mobile applications at an unprecedented pace lead to reliable networks and higher peak data rates. In November 2019, COVID-19 was identified in Wuhan, China, and became a pandemic in a short time. Online education, working from home, and online businesses have become widespread as the world has experienced lockdowns due to the pandemic. This new normal has led the world to comprehend the necessity of UL throughput and innovative cellular technologies eminently. According to Lutu et al., pandemics such as the SARS-CoV-2 virus can drastically impact the network data traffic in favor of UL demand. Their results confirm an increase of 10% in UL mobile data usage and a decrease of 20% in DL mobile data usage [5].
As stated by cellular standards, the DL contains more measurement elements and traceable features than the UL [6]. Naturally, carriers tend to focus more on the DL and deploy an asymmetric channel capacity between DL and UL channels [7]. Due to this asymmetric deployment, the DL channel has a higher transmission capacity than the UL. However, the need for UL data has increased considerably with the rapid increase in the use of applications that require UL throughput, such as AV control, machine type communication devices (MTDs) smart body area networks (SmartBAN), IoT, wireless sensor networks, video conferencing, file sharing, VoIP, surveillance cameras, peer-to-peer (P2P) and cloud services [8,9,10,11,12,13,14]. Although many studies have focused on DL channel prediction due to the much higher demand for DL data rate before, UL traffic estimation has gained significance since the channel asymmetry in favor of DL throughput is closing rapidly [15]. MTD applications that rely on vast amounts of small data packets transmitted via the UL channel can place extremely heavy congestion on the channel [16]. This congestion will cause significant difficulties for cellular networks, and if optimizations are not made, it becomes inevitable to use 5G network systems that occupy a large amount of traffic in the UL direction [17]. SmartBANs, which are networks of sensors embedded in the human body that provide beneficial health care monitoring such as EEG, EKG, blood pressure, etc., also mostly use UL transmission. Therefore, possible channel congestion may delay the vital feedback about the patient’s health [18]. Carson and Lundvall state that around 20% of end-users are dissatisfied with their UL data rate [19]. The complaints will require data collection and optimization of UL channels in cellular networks. However, the outdated rule-based techniques cannot bear even current system requirements due to the unprecedented growth of the number of parameters, carrier features, and counters. Machine learning (ML) algorithms used in the most complex tasks in every field [20] are great for designing and optimizing mobile networks.
ML’s primary concept is to create autonomous systems that learn from data throughput by identifying patterns and making decisions based on results [21]. The learning process and its implementation are illustrated in Figure 1.
Figure 1. Illustration of the learning process.
There are several commonly used ML algorithms, such as decision trees, linear regression, artificial neural networks (ANN), and K-nearest neighbor. After calculating the cost, weights are adjusted through a gradient descent algorithm and repeated until gradient descent converges to the optimum value. This convergence is called “the learning process” and is valid for almost every machine learning algorithm.
Recently, 4G has become the dominant technology in cellular networks globally and soon will be introduced in leading radio mechanisms, 5G systems. While using these technologies, it is critical to provide a stable and high UL throughput in emergencies. For instance, it is essential to use AI-aided autonomous ambulance drones to speed up emergency responses, prevent deaths, and accelerate recovery dramatically [22]. Thus, a robust UL throughput predicting mechanism is required to ensure the highest quality of service.
LTE has various radio measurements that can be a key input for machine learning to model and predict UL throughput. Using ML algorithms, this research evaluates the lower layer LTE radio metrics for UL throughput prediction. This study utilizes three datasets collected from Melbourne, Florida, Houston, Texas, and Batman, Turkey. It is also one of the first in a series to initiate model development and create a framework from a comprehensive data set. The paper continues with the following sections: literature review, understanding of machine learning algorithms, data collection, and prediction results and analysis.

2. Literature Review

Minimizing the latency, geopositioning of systems, and keeping the throughput at a premium are necessary for modern telecommunication systems such as LTE and upcoming 5G technologies. As mobile applications need more content to be delivered wirelessly, e.g., video streaming and 3D point clouds, modeling and optimizing the throughput become indispensable in terms of QoS. Cellular mobile networks utilize throughput models that contain history-based, computational-based, and ML-based predictions for design and optimization. History-based predictions mainly extrapolate archived data, which contains various parameters, including time, place, and density of the traffic, to decide. On the other hand, computation-based predictions evaluate a mathematical model as a function of several variables to derive a decision. Nevertheless, there is still a lack of accuracy in both approaches, especially on the UL side.
An experiment conducted by Yue et al. benefits from stationary LTE measurements taken from regional routes, highways, and pedestrian lanes to forecast DL throughput. The study performs ML algorithms on two US cellular operators and achieves an error rate between 4 and 17% [23]. Even though the study has high accuracy in DL prediction, it still needs to be improved to eliminate error fluctuation. In his study, Jomrich et al. seek a reliable cellular data connection since mobility causes significant oscillation in QoS. Thus, they investigated an automated bandwidth prediction model aided by ML algorithms. They stated that their machine learning algorithms were lightly affected by additional oscillations caused by vehicle velocity [24]. Bojovic et al. researched the Self-Organizing Network (SON) to reduce the time and expenditure of the network processes. Learning-based SON uses key performance indicators (KPI’s) as estimation parameters and can adapt dynamically to the environment for QoS improvement purposes. The researchers used LTE parameters such as dynamic frequency and bandwidth assignment with 95% of optimal network performance [25]. Finally, Oussakel et al. investigated a machine learning solution to manage the immense demand for UL transmission to increase the QoS. As stated in their work, the main supervised learning algorithms are utilized to predict UL bandwidth in the range of 88% to 94% with different scenarios [26].

3. Understanding Machine Learning Algorithms

Machine learning evolved from the field of data analytics, including computational learning and pattern recognition. ML comprises many mathematical models to make predictions for specific tasks. Since it has a variety of applications such as e-mail filtering, anomaly detection, and image recognition, it is a popular tool for researchers, engineers, and data scientists [27,28,29]. This section will explain some of the important machine learning algorithms in detail.

3.1. Linear Regression

Linear regression is an ML algorithm that seeks a linear relationship between the number of independent variables ( x ^ ) and the dependent variable ( y ^ ). The equation that is used to predict a linear regression line is indicated in Equation (1).
y ^ i = β ^ 0 + β ^ 1 x i + ϵ ^ i
The intuition behind the linear regression algorithm is to determine the best values for coefficients β ^ 0 and β ^ 1 . Cost function and slope are essential key concepts to derive results for coefficients β ^ 0 and β ^ 1 . Equation (2) represents the cost function and investigates the minimum value of the cost function.
m i n i m i z e ( 1 n i = 1 n ( y i y ¯ ) 2 )

3.2. Gradient Descent

Gradient descent is one of the first-order methods which basically begin with an initial point and repeatedly move its steps in the direction of the gradient descent that minimize the f ( x ) function as seen in Algorithm 1 [30]:
Algorithm 1 The Gradient Descent Algorithm
for k = 0 , 1 , 2 , do
    g k f ( x k )
    x k + 1 x k t k g k
end for
Where k is the number of iterations, g k is the gradient of the dependent variable, and x k is the independent variable. The gradient descent iterative action is shown in Figure 2. According to Equation (1), there are two coefficients as β ^ 0 and β ^ 1 presented as contour levels. The starting point of gradient descent, x 0 , moves to x 1 , then from x 1 to x 2 , until it converges to the optimum value. The ideal optimum value is the middle of the circles, which indicates the global minimum. Figure 3 also shows the local minimum, which may cause deceptive predictions. Gradient descent does an iterative update to find the best coefficients leading the global minimum.
Figure 2. Demonstration of gradient descent.
Figure 3. Representation of a DT.
By definition, the gradient descent algorithm has several criteria to find its direction during the iterative process, such as the initial point x 0 , step size t k , and stopping condition. Since these criteria are linked to each other, a wrong choice may end up with a local minimum instead of the global minimum. Besides, the technique called early stopping, a form of regularization, should be applied to a gradient descent algorithm to avoid the over-fitting problem. However, implementing such techniques increases the learning performance with each iteration, which leads to a higher generalization error. Thus, it is wise to skip the early stopping and keep f ( x k ) as close as possible to the global minimum. The following convergence bounds rule should be applied to determine the minimum number of steps:
k k f ( f ( x 0 ) f ( x * ) ) g 1 ε
where k is the number of iterations, k f is a feature estimated from the independent variables of the function, g is a function dependent on the independent variable of the function, ε is the error of confidence, and f ( x * ) is the true minimum.

3.3. Gradient Boosting Regression

Boosting is a technique that is applied to base learners of machine learning tools such as regression and classification. Boosting is a doping factor for poor iterative models and a way to combine multiple weak models into one single model. Different boosting techniques reveal the quantity of the misclassification and pick the appropriate option for the next iteration. This mainly decreases the bias due to the focus on misclassified cases. In this research, we employ Gradient Boosting Regression (GBR) as a boosting technique.
The GBR uses gradient terms for prediction since it utilizes the gradient descent algorithm to minimize the cost function. Even though there is a slight difference between The GBR and linear regression, GBR has a notable increase in accuracy. The GBR is generally used with decision trees (DT) of fixed-size base learners. The GBR determines the error between the current prediction and actual values. This error is called residual. The GBR trains weak models, which maps features to this residual. After acquiring the residual, it combines them with the input of the existing model. This iterative process pushes the model in the direction of the actual value and improves the overall model prediction [31]. The GBR is represented in Algorithm 2.
Algorithm 2 The GBR Algorithm
1. Initialize model with a constant value:
F 0 ( x ) = a r g m i n i = 1 n L o s s ( y i , γ )

2. i=1:M 1. Compute the so-called pseudo-residuals:
r i m = L o s s ( y i , F ( x i ) ) F ( x i ) F ( x ) = F m 1 ( x ) for   i = 1 , , n .

3. Fit a weak learner h m ( x ) to pseudo-residuals, i.e., train it using the training set { ( x i , r i m ) } i = 1 n
4. Compute multiplier γ m by solving the following one-dimensional optimization problem:
γ m = arg min γ i = 1 n L o s s y i , F m 1 ( x i ) + γ h m ( x i )

5. Update the model:
F m ( x ) = F m 1 ( x ) + γ m h m ( x )

6. Output F M ( x )
where F 0 ( x ) is an intended function. γ represents the learning rate.

3.4. Decision Tree Regression (DTR)

DTR is a prediction model based on recursive partitioning of features into a tree-like structure. It is like a hierarchical flowchart with internal nodes, and every node has an attribute to test and classify an outcome. Based on probability, the nodes are assigned to a related class. Thus, even slight variations may end up with different DTs [32]. Figure 3 shows a flowchart representation of DTs. According to the flowchart, the data stream goes through the outlook and checks the RSRP and RSRQ classes. If the class comprises low SNR, then it picks RSRP, else, it picks RSRQ. If RSRP and RSRQ branches determine the throughput under or over the predefined threshold, then it makes a decision accordingly.

3.5. K-Nearest Neighbors (KNN)

The KNN algorithm is a supervised learning algorithm meant to solve classification, pattern recognition, and regression problems. It is also a non-parametric technique that can be used in various applications. The KNN algorithm considers that similar attributes should be close to the selected data points [33]. K demonstrates the number of nearest neighbors, which is crucial for prediction accuracy. The distance vectors ( D K N N ( x i , y i ) ) are generally determined by Manhattan distance formula as seen in Equation (8) [34].
D K N N ( x , y ) = i = 1 n x i y i
Root Mean Square Prediction Error (RMSPE) is used with a cross-validation method to choose the best k number. Then, the target prediction is made by taking the average of k nearest data points. The mathematical formula of RMSPE is indicated in Equation (9) [35].
R M S P E = 1 n i = 1 n ( y i y ^ i ) 2

5. Prediction Results and Analysis

Three data sets collected over 10 days were divided into training and test sets where 90% of data were randomly assigned as a training set and fed into ML algorithms. The remaining data were used for the test set, which was used to assess the model prediction accuracy. The trained model was deployed on the test measurement subset to predict the UL throughput. Predicted UL throughputs were compared with test labels via the coefficient of determination R 2 , as seen in Figure 10 and Table 2.
Figure 10. Prediction Algorithm Comparison.
Table 2. R 2 and MSE [Mbps] Analysis of ML Algorithms for Melbourne, FL, Batman, Turkey, and Houston, TX.
According to Table 2 and Figure 10, DTR and KNN have the highest prediction accuracy among all ML algorithms. This highest prediction accuracy is likely because DTR and KNN both use classification while the others build mathematical models. Linear regression and gradient descent algorithms form a linear relationship between the LTE parameters and UL throughput. However, more intricate rules are needed to deduce this relationship instead of simple linearity and increase the prediction rate. GBR improves the prediction accuracy by implementing the residuals to the gradient descent method, and it provides almost as good RMSE as DTR and KNN algorithms. Moreover, we see a significant decrease in R 2 values from suburban to metropolitan areas. This decrease can be linked to scattering and fading effects in dense building areas. Additionally, Houston data might be affected by possible rain conditions, which might cause more estimation errors in this data set. Following the ML training and prediction, the variable’s importance of each parameter was investigated to rank features according to the impact on prediction accuracy. Calculated importance is illustrated in the bar chart in Figure 11.
Figure 11. Variable Importance.
It confirms that the RSRP metric is the main contributor (over 60%) to the prediction algorithm, while baseband SNR and RSRQ have almost equal importance (between 9 to 20%). According to DTR results, the prediction feature importance levels are drawn using Python sckit-learn library.

6. Limitation and Discussion

The gradual closure of the channel capacity asymmetry between the UL and the DL with developing and widespread technologies has shown that intensive research on the DL channel basis should also be performed in the UL direction. This article aimed to make channel estimation with a robust and high accuracy prediction rate using as few LTE parameters as possible. In the system, physical layer parameters must be obtained via receivers or scanners located in the experiment area to estimate the UL throughput of a specific location. However, receivers and scanners may not represent the exact measurements as phones because of the differences in antenna gain. Therefore, the proposed method should also be performed by mobile phones to see whether there is a significant change in measurements. Moreover, regarding sustainable and high-quality network services, the proposed method is essential in providing adequate bandwidth and higher UL throughput rates, which may be needed at shopping malls, concerts, fairs, and crowded areas. The proposed method will estimate the required UL throughput and may be used to adjust the DL-UL channel bandwidth allocation asymmetry in favor of UL. In other words, the allocated DL channel bandwidth can be shifted to the UL channel when needed. Finally, it should be emphasized that most of the studies focused on DL channel prediction, and a few of the studies have performed UL throughput prediction with low prediction accuracy (83–94%). The measurements of these studies are collected from the same environment, which may limit their general usage in different environments.

7. Conclusions

This research presented an ML-based UL throughput prediction model, which applies to 4G and possibly, 5G mobile networks. The data sets were collected through drive tests on currently deployed 4G LTE mobile networks in different locations (Melbourne, FL, Batman, Turkey, Houston, TX), and the performance of ML algorithms such as linear regression, gradient descent, gradient boosting regression, decision tree regression, and k-nearest neighbor models were tested on UL throughput prediction. An enhanced ML traffic management is modeled based on minimal feature sets such as RSRP, RSRQ, and SNR. The highest correlation is observed with RSRP 0.85 in Melbourne, FL, RSRQ 0.46 in Batman, Turkey, and SNR 0.53 in Melbourne, FL. The determination coefficient ( R 2 ) values were calculated along with RMSE values. The observed coefficient of determination was 0.92 for both DTR and KNN algorithms. It is also seen that it is likely to observe a relationship between UL throughput prediction accuracy and the environment’s type since R 2 values decrease from 0.92 to 0.69 from suburban areas to metropolitan areas. However, more research needs to be conducted with additional data to support this result.

Author Contributions

Conceptualization, E.E. and Y.E.; methodology, E.E. and Y.E.; software, E.E.; validation, Y.E. and J.Z.; formal analysis, E.E. and Y.E.; investigation, E.E.; resources, E.E.; data curation, E.E. and J.Z.; writing—original draft preparation, E.E. and Y.E. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
LTELong Term Evolution
RSRPReference Signal Received Power
RSRQReference Signal Received Quality
SNRSignal to Noise Ratio
ULUplink
DLDownlink
KNNK-Nearest Neighbor
4G4th Generation
5G5th Generation
GSMGlobal System for Mobile Communications
OFDMOrthogonal Frequency Division Multiple Access
QAMQuadrature Amplitude Modulation
IotInternet of Things
MLMachine Learning
ANNArtificail Neural Network
QoSQuality of Service
SONSelf Organizing Network
DTDecision Trees
GBRGradient Boosting Regression

References

  1. Kim, Y.; Kim, Y.; Oh, J.; Ji, H.; Yeo, J.; Choi, S.; Ryu, H.; Noh, H.; Kim, T.; Lee, J.; et al. New Radio (NR) and its Evolution toward 5G-Advanced. IEEE Wirel. Commun. 2019, 26, 2–7. [Google Scholar] [CrossRef]
  2. Hajlaoui, E.; Khlifi, A.; Zaier, A.; Ghodhbane, J.; Hamed, M.B.; Sbita, L. Performance Evaluation of LTE Physical Layer. In Proceedings of the 2019 International Conference on Internet of Things, Embedded Systems and Communications (IINTEC), Tunis, Tunisia, 20–22 December 2019; pp. 106–111. [Google Scholar] [CrossRef]
  3. Singh, H.; Prasad, R.; Bonev, B. The Studies of Millimeter Waves at 60 GHz in Outdoor Environments for IMT Applications: A State of Art. Wireless Pers. Commun. 2018, 100, 463–474. [Google Scholar] [CrossRef]
  4. Isyaku, B.; Mohd Zahid, M.S.; Bte Kamat, M.; Abu Bakar, K.; Ghaleb, F.A. Software Defined Networking Flow Table Management of OpenFlow Switches Performance and Security Challenges: A Survey. Future Internet 2020, 12, 147. [Google Scholar] [CrossRef]
  5. Lutu, A.; Perino, D.; Bagnulo, M.; Frias-Martinez, E.; Khangosstar, J. A Characterization of the COVID-19 Pandemic Impact on a Mobile Network Operator Traffic. In Proceedings of the IMC ’20: ACM Internet Measurement Conference, Virtual Event, USA, 27–29 October 2020. [Google Scholar]
  6. Edler, G.; Wang, L.; Horiuchi, A. Special Subframe Configuration for Latency Reduction. U.S. Patent Application No. 16/089,279, 26 October 2021. [Google Scholar]
  7. Rayal, F. LTE in a Nutshell. 2020. Available online: https://home.zhaw.ch/kunr/NTM1/literatur/LTE%20in%20a%20Nutshell%20-%20Physical%20Layer.pdf (accessed on 24 October 2020).
  8. Teng, Y.; Yan, M.; Liu, D.; Han, Z.; Song, M. Distributed Learning Solution for Uplink Traffic Control in Energy Harvesting Massive Machine-Type Communications. IEEE Wirel. Commun. Lett. 2020, 9, 485–489. [Google Scholar] [CrossRef]
  9. Kim, T.; Jung, B.C. Performance Analysis of Grant-Free Multiple Access for Supporting Sporadic Traffic in Massive IoT Networks. IEEE Access 2019, 7, 166648–166656. [Google Scholar] [CrossRef]
  10. Kim, T.; Song, T.; Kim, W.; Pack, S. Phase-Divided MAC Protocol for Integrated Uplink and Downlink Multiuser MIMO WLANs. IEEE Trans. Veh. Technol. 2018, 67, 3172–3185. [Google Scholar] [CrossRef]
  11. Xu, C.; Wu, M.; Xu, Y.; Xu, Y. Shortest Uplink Scheduling for NOMA-Based Industrial Wireless Networks. IEEE Syst. J. 2020, 14, 5384–5395. [Google Scholar] [CrossRef]
  12. Ma, Z.; Feng, L.; Wang, Z. Supporting Asymmetric Transmission for Full-Duplex Smart-Home Networks. IEEE Access 2019, 7, 34807–34822. [Google Scholar] [CrossRef]
  13. Sun, K.; Wu, J.; Huang, W.; Zhang, H.; Hsieh, H.-Y.; Leung, V.C.M. Uplink Performance Improvement for Downlink-Uplink Decoupled HetNets with Non-Uniform User Distribution. IEEE Trans. Veh. Technol. 2020, 69, 7518–7530. [Google Scholar] [CrossRef]
  14. Jiménez, L.R.; Solera, M.; Toril, M.; Luna-Ramírez, S.; Bejarano-Luque, J.L. The Upstream Matters: Impact of Uplink Performance on YouTube 360° Live Video Streaming in LTE. IEEE Access 2021, 9, 123245–123259. [Google Scholar] [CrossRef]
  15. Homssi, B.A.; Al-Hourani, A. Modeling Uplink Coverage Performance in Hybrid Satellite-Terrestrial Networks. IEEE Commun. Lett. 2021, 25, 3239–32431. [Google Scholar] [CrossRef]
  16. Ali, S.; Rajatheva, N.; Saad, W. Fast Uplink Grant for Machine Type Communications: Challenges and Opportunities. IEEE Commun. Mag. 2019, 57, 97–103. [Google Scholar] [CrossRef]
  17. Shen, H.; Ye, Q.; Zhuang, W.; Shi, W.; Bai, G.; Yang, G. Drone-Small-Cell-Assisted Resource Slicing for 5G Uplink Radio Access Networks. IEEE Trans. Veh. Technol. 2021, 70, 7071–7086. [Google Scholar] [CrossRef]
  18. Ruan, L.; Dias, M.P.I.; Wong, E. SmartBAN With Periodic Monitoring Traffic: A Performance Study on Low Delay and High Energy Efficiency. IEEE J. Biomed. Health Inform. 2018, 22, 471–482. [Google Scholar] [CrossRef]
  19. Carson, S.; Lundvall, A. Mobility on The Pulse of The Networked Society; Ericsson: Stockholm, Sweden, 2016; pp. 1–36. [Google Scholar]
  20. Kato, N.; Mao, B.; Tang, F.; Kawamoto, Y.; Liu, J. Ten Challenges in Advancing Machine Learning Technologies toward 6G. IEEE Wirel. Commun. 2020, 27, 96–103. [Google Scholar] [CrossRef]
  21. Egi, Y.; Otero, C.E. Machine-Learning and 3D Point-Cloud Based Signal Power Path Loss Model for the Deployment of Wireless Communication Systems. IEEE Access 2019, 7, 42507–42517. [Google Scholar] [CrossRef]
  22. Ray, P.P.; Nguyen, K. A Review on Blockchain for Medical Delivery Drones in 5G-IoT Era: Progress and Challenges. In Proceedings of the 2020 IEEE/CIC International Conference on Communications in China (ICCC Workshops), Chongqing, China, 9–11 August 2020; pp. 29–34. [Google Scholar] [CrossRef]
  23. Yue, C.; Jin, R.; Suh, K.; Qin, Y.; Wang, B.; Wei, W. LinkForecast: Cellular Link Bandwidth Prediction in LTE Networks. IEEE Trans. Mob. Comput. 2018, 17, 1582–1594. [Google Scholar] [CrossRef]
  24. Jomrich, F.; Herzberger, A.; Meuser, T.; Richerzhagen, B.; Steinmetz, R.; Wille, C. Cellular bandwidth prediction for highly automated driving evaluation of machine learning approaches based on real-world data. In Proceedings of the VEHITS 2018—International Conference on Vehicle Technology and Intelligent Transport Systems, Funchal-Madeira, Portugal, 16–18 March 2018; pp. 121–132. [Google Scholar]
  25. Bojovic, B.; Meshkova, E.; Baldo, N.; Riihijarvi, J.; Petrova, M. Machine learning-based dynamic frequency and bandwidth allocation in self-organized LTE dense small cell deployments. Eurasip J. Wirel. Commun. Netw. 2016, 2016, 1–16. [Google Scholar] [CrossRef]
  26. Oussakel, I.; Owezarski, P.; Berthou, P. Experimental Estimation of LTE-A Performance. In Proceedings of the 2019 15th International Conference on Network and Service Management (CNSM), Halifax, NS, Canada, 21–25 October 2019. [Google Scholar]
  27. Awad, W.A.; ELseuofi, S.M. Machine Learning methods for E-mail Classification. Int. J. Comput. Appl. 2011, 16, 39–45. [Google Scholar] [CrossRef][Green Version]
  28. Hasan, M.; Islam, M.M.; Zarif, M.I.I.; Hashem, M.M.A. Attack and anomaly detection in IoT sensors in IoT sites using machine learning approaches. Internet Things 2019, 7, 100059. [Google Scholar] [CrossRef]
  29. Olukan, T.A.; Chiou, Y.C.; Chiu, C.H.; Lai, C.Y.; Santos, S.; Chiesa, M. Predicting the suitability of lateritic soil type for low cost sustainable housing with image recognition and machine learning techniques. J. Build. Eng. 2000, 29, 101175. [Google Scholar] [CrossRef]
  30. Ketkar, N. Stochastic Gradient Descent. In Deep Learning with Python; Apress: Berkeley, CA, USA, 2017; pp. 113–132. [Google Scholar]
  31. Li, C. A Gentle Introduction to Gradient Boosting. 2016. Available online: http://www.ccs.neu.edu/home/vip/teach/MLcourse/4boosting/slides/gradient-boosting.pdf (accessed on 5 November 2021).
  32. Wang, F.; Wang, Q.; Nie, F.; Li, Z.; Yu, W.; Ren, F. A linear multivariate binary decision tree classifier based on K-means splitting. Pattern Recognit. 2020, 107, 107521. [Google Scholar] [CrossRef]
  33. Kramer, O. K-Nearest Neighbors. In Dimensionality Reduction with Unsupervised Nearest Neighbors. In Intelligent Systems Reference Library; Springer: Berlin/Heidelberg, 2013; Volume 51. [Google Scholar] [CrossRef]
  34. SinghAn, A. K-Nearest Neighbors Algorithm: KNN Regression Python. 2020. Available online: https://www.analyticsvidhya.com/blog/2018/08/k-nearest-neighbor-introduction-regression-python/ (accessed on 4 November 2020).
  35. Christodoulou, C.; Moorby, J.M.; Tsiplakou, E.; Kantas, D.; Foskolos, A. Evaluation of nitrogen excretion equations for ryegrass pasture-fed dairy cows. Animal 2021, 15, 100311. [Google Scholar] [CrossRef] [PubMed]
  36. Egi, Y.; Eyceyurt, E.; Kostanic, I.; Otero, C.E. An Efficient Approach for Evaluating Performance in LTE Wireless Networks. In Proceedings of the International Conference on Wireless Networks (ICWN); The Steering Committee of the World Congress in Computer Science, Computer Engineering and Applied Computing (WorldComp): Las Vegas, NV, USA, 2017; pp. 48–54. [Google Scholar]
  37. Mehta, D.S.; Chen, S. A spearman correlation based star pattern recognition. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.