Next Article in Journal
Unjumbling Procedure in the Algorithmic Analysis of Biomechanical Torques Induced by Electrical Stimulation: Case Study of the Lower Limb
Next Article in Special Issue
Special Issue on Advanced Information Processing Methods and Their Applications
Previous Article in Journal
Early Warning of the Construction Safety Risk of a Subway Station Based on the LSSVM Optimized by QPSO
Previous Article in Special Issue
Patch-Wise Infrared and Visible Image Fusion Using Spatial Adaptive Weights
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Development of a Multilayer Perceptron Neural Network for Optimal Predictive Modeling in Urban Microcellular Radio Environments

1
Department of Physics, Federal University Lokoja, Lokoja 260101, Nigeria
2
Department of Electrical and Electronics Engineering, Faculty of Engineering, University of Lagos, Akoka, Lagos 100213, Nigeria
3
Department of Electrical Engineering and Information Technology, Institute of Digital Communication, Ruhr University, 44801 Bochum, Germany
4
Department of Electrical and Computer Engineering, College of Engineering, Anderson University, Anderson, SC 29621, USA
5
College of Arts and Sciences, Anderson University, Anderson, SC 29621, USA
6
Department of Technology Education, Chungnam National University, Daejeon 34134, Korea
7
Research and Development Center for Physical Education, Health and Information Technology, Department of Library and Information Science, Fu Jen Catholic University, New Taipei City 24205, Taiwan
8
Department of Computer Science and Information Engineering, Asia University, Taichung City 41354, Taiwan
9
Department of Information Management, Tainan University of Technology, 529 Zhongzheng Road, Tainan City 710, Taiwan
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2022, 12(11), 5713; https://doi.org/10.3390/app12115713
Submission received: 12 April 2022 / Revised: 10 May 2022 / Accepted: 2 June 2022 / Published: 3 June 2022
(This article belongs to the Special Issue Advanced Information Processing Methods and Their Applications)

Abstract

:
Modern cellular communication networks are already being perturbed by large and steadily increasing mobile subscribers in high demand for better service quality. To constantly and reliably deploy and optimally manage such mobile cellular networks, the radio signal attenuation loss between the path lengths of a base transmitter and the mobile station receiver must be appropriately estimated. Although many log-distance-based linear models for path loss prediction in wireless cellular networks exist, radio frequency planning requires advanced non-linear models for more accurate predictive path loss estimation, particularly for complex microcellular environments. The precision of the conventional models on path loss prediction has been reported in several works, generally ranging from 8–12 dB in terms of Root Mean Square Error (RMSE), which is too high compared to the acceptable error limit between 0 and 6 dB. Toward this end, the need for near-precise machine learning-based path loss prediction models becomes imperative. This work develops a distinctive multi-layer perception (MLP) neural network-based path loss model with well-structured implementation network architecture, empowered with the grid search-based hyperparameter tuning method. The proposed model is designed for optimal path loss approximation between mobile station and base station. The hyperparameters examined include the neuron number, learning rate and hidden layers number. In detail, the developed MLP model prediction accuracy level using different learning and training algorithms with the tuned best values of the hyperparameters have been applied for extensive path loss experimental datasets. The experimental path loss data is acquired via a field drive test conducted over an operational 4G LTE network in an urban microcellular environment. The results were assessed using several first-order statistical performance indicators. The results show that prediction errors of the proposed MLP model compared favourably with measured data and were better than those obtained using conventional log-distance-based path loss models.

1. Introduction

Path loss models are unique prediction models employed by telecom network engineers to estimate the signal coverage area being served by a given transmitter during networking and management [1,2,3]. However, developing these signal path loss models with the optimal accuracy it deserves is a complex and significant problem in the planning of telecommunication networks. The conventional log-distance-based statistical models available in the literature, such as the cluster factor model, COST 234 Hata Model, free space model, Hata model and Lee models, lack accuracy for realistic path loss prediction applications in cellular mobile networks environments [4,5,6,7,8,9,10,11]. The aforementioned fundamental limitation of the conventional models is usually very pronounced when the respective models have been applied in cellular radio environments other than the developed and designed environment [12,13]. This scenario is mainly due to dissimilarities and variations in environmental formations (hilly, mountainous or quasi-plain), weather conditions, soil electrical properties and terrain type (open, rural, suburban or urban) that exist in different radio propagation locations, cities and countries [14,15,16,17,18,19,20,21,22,23]. For example, the Hata loss model was developed based on extensive practical measurements carried out in Japan at transmission frequency ranges of 150 to 1920 MHz and macrocellular communication distances of 1 to 100 km, with the mobile station and base station antenna heights of 2 to 3 and 30 to 1000 m, respectively [24]. This model, including other aforementioned conventional ones, is also generally limited in capturing the non-linear relationship between the independent variable (e.g., path loss) and dependent variable (e.g., distance) [25]. The precision of these conventional models on path loss prediction has been reported in many previous works to generally range from 8–12 dB in terms of Root Mean Square Error (RMSE), which is exceptionally higher than the acceptable values [9,14,15].
Recently, an Artificial Neural Network (ANN), a unique artificial intelligence soft computing and modelling technique, has been acknowledged and proven to solve function approximation and pattern classification problems [26]. Some ANN models exist in the literature; the key ones are Radial Basis Function models, Multilayer Perception Models, Generalized Neural Network models, etc. Among these models, the MLP ANN models have stood out most recently because they are very robust and popular for learning, function approximation, and pattern classification [27,28,29,30]. The MLP ANN possess many robust algorithms that can be explored to carry out more proficient adaptive nonlinear statistical modelling over the classical logistic regression methods [14,15,16,17,18,19,20,21,22,23] that are frequently engaged in developing predictive models. This robustness can be ascribed to their acknowledged special ability to learn, predict, and classify non-linear data using experience and preceding samples introduced to the network model. Huang [31] also noted that the MLP is characteristically good for input-output data mapping. Generally, a clear-cut underlining capability of ANN-based models over the conventional log-distance-based models is their large degrees of freedom structure which provides means for fitting many datasets with non-linear or linear correlation patterns. The concept of intelligent-based ANN models for optimal and adaptive prognostic estimations of path losses was introduced to surmount the limitations of existing empirically and deterministically developed log-distance models [32,33]. In the paper, the availability of manifold resourceful training algorithms and hypermeters of MLP ANN that can be tuned to further boost its extrapolative data analysis is worth exploring in this paper for optimal predictive modelling. Hence, the “Development of a Multilayer Perception Neural Network for Optimal Predictive Modeling in Urban Microcellular Radio Environments is self-evident.” Other key robust advantages of the general ANNs are highlighted in Section 2.3.
Though several ANN models exist in the literature, a critical, challenging task remains in developing and using them appropriately through the correct selection of its network structural design with the required input elements and hyperparameters to solve peculiar predictive mapping and functional problems. The quest to address this issue is the leading motivation for this research paper. However, one primary challenge in using the MLP model is correctly selecting its network architecture with the required input elements (hyperparameters) to solve a particular mapping problem [34]. Another critical challenge with neural network models is the problem of determining the input data variables that must correlate with the target variables [35,36].
This paper develops a distinctive MLP-based path loss model with well-structured implementation network architecture, empowered with the grid search-based hyperparameter tuning method for optimal path loss approximation between mobile-station and base-station path lengths. The hyperparameters include the neuron number, learning rate and hidden layers number. In detail, the developed MLP model prediction accuracy level using different learning and training algorithms with the tuned best values of the hyperparameters have been applied for extensive path loss experimental datasets. The datasets were acquired via field drive tests conducted in Long Term Evolution (LTE) in urban microcellular radio networks. For the development and implementation of the MLP-ANNs model, we utilized version 2018a of the MATLAB neural networks toolbox. The toolbox provides the required user interface, algorithm and platform to train, test, validate, visualize and simulate networks with the desired number of layers, neurons and activation functions.
In particular, the contributions of this paper are summarized as follows:
  • ▪ A distinctive MLP neural network-based path loss model with well-structured implementation network architecture, embedded with the precise transfer function, neuron number, learning algorithm and hidden layers, is developed for optimal path loss approximation between mobile-station and base-station path lengths.
  • ▪ The developed MLP neural network model was tested and validated for realistic path loss prediction using extensive experimental signal attenuation loss datasets acquired via field drive test conducted over LTE networks in urban microcellular radio environments and tested using first-order statistical performance indicators.
  • ▪ Optimization of the projected model via hyperparameter tuning leveraging the grid search algorithm analyses of the experimental path loss data.
  • ▪ Optimal prediction efficacy of the developed MLP neural network model compared with well-structured implementation architecture over standard log-distance models using several first-order statistical performance indicators.
The remainder of this work is structured in the following manner. Section 2 outlines the background information, such as radio propagation mechanisms, log-distance-based path loss prediction models, and the basis of artificial neural networks (ANN). Section 3 presents the methodology detailing the neural network implementation for predictive modelling. Section 4 provides the results, analysis, evaluation of the introduced neural network model at different study locations, comparison of the developed neural network model with log-distance models, and discussions. Finally, the conclusion is given in Section 5.

2. Theoretical Background

The theoretical background covers the radio propagation mechanism, log-distance-based path loss prediction models and artificial neural network systems.

2.1. Radio Propagation Mechanism

When radio signals travel, which are a form of electromagnetic waves, they interact with the media and objects they travel through. In the sequence of their interaction, the radio signals become weaker owing to refraction, reflection, diffraction, absorption and other propagation phenomena. The resultant effect of all the phenomena on propagated signals is signal propagation loss. The characteristics of the pathway or medium through which the radio signals travel determine the amount of propagation loss and the quality of the received signal that is attainable at the receiving terminal. Radio propagation loss is also governed by other sundry elements, particularly the transmitter power, receiver sensitivity and general antenna parameters such as antenna gain, antenna height and receiver location [1,2,37,38].
The prominent factors that influence the number of signal path losses in a medium include diffraction, reflection, refraction, scattering and absorption, to mention a few. For example, diffraction arises when radio waves collide with huge obstacles compared to the propagating signal wavelengths. Moreover, diffraction occurs when radio signals bend around objects, especially those with sharp edges. This alteration often empowers the received radio signal energy to spread around the boundaries of the obstructing object [39,40]. Diffraction is also influenced by the phase, amplitude, pathway and frequency of the transmitted waves.
The environment in which the radio frequency signals travel (or are propagated) will undoubtedly negatively impact the signal. For example, radio wave signals and propagation loss vary extensively in correspondence to the terrain landscape, building structures and population density. Marshy, damp and sandy terrain also attenuate radio signals, primarily propagated low-frequency signals. In other words, signals travel faster over conducive terrain than in sandy and marshy or damp terrains.

2.2. Log-Distance-Based Path Loss Prediction Models

Generally, path loss models are a set of mathematical models, expressions, resources and algorithms used for signal attenuation loss prediction between the paths of a base transmitter and the mobile station receiver. These models are helpful planning tools that assist the radio network designers of cellular telecommunication systems in optimally positioning base station transmitters to meet the desired signal coverage level and service quality requirements of the networks.
The predictive performance of any path loss model is determined by the resultant prediction accuracy with actual field measured loss data.
The log-distance-based path loss models are models whose average power loss logarithmically depends on distance (transmission path length) intertwined with a propagation exponent modelling parameter. The propagation exponent is usually employed to account for a specific radio propagation environment. They can also be described as simplified models that attempt to model variations, fluctuations and attenuations in the received signal power. Examples of log-distance-based models include the Walficsh–Ikegami, Walficsh Bertoni, cluster factor, COST 234 Hata, Hata Okumura, SUI model, Lee model, Egli Model and others [41]. Though these models have varying frequency validity thresholds, different correction factors have been applied to ease their applicability at the tested frequency band. Detailed descriptions of these models are contained in [14,17,42].

2.3. Artificial Neural Networks (ANNs)

ANNs, also popularly referred to as artificial neural systems, are efficient computing systems or relatively simple computational models founded on the neural organization of the brain with functional changing parameters to process information effectively. ANNs are distinctive and robust non-linear statistical data modeling networks wherein reasonably simple connections between inputs and outputs nodes alignments are established. According to Robert Hecht-Nielsen, the first inventor of the neurocomputer, a neural network can be defined as “a computing system made up of several simple, highly interconnected processing elements, which process information by their dynamic state response external inputs”. The processing elements are called neurons. The neuron is the special mathematical function that captures and organizes information according to the neural network architecture.
Some of the essential features or advantages of ANNs are [31,34,35,36].
  • High-speed computation adeptness
  • Global interconnection of network processing elements
  • Robust distributed and asynchronous parallel processing
  • High adaptability to non-linear input-data mapping
  • Robust noise tolerance
  • High fault tolerance utilizing Redundant Information Coding
  • Robust in providing real-time operations
  • High potential for hardware application
  • Capable of deriving meaning from the imprecise or complicated dataset
  • High capacity to learn, recall and generalize input data training pattern

3. Methodology

As mentioned earlier, the ANN model possesses many robust training algorithms and hyperparameters that can be explored to conduct proficient adaptive nonlinear statistical modelling over the classical logistic regression methods. This section contains the materials and method explored to develop the proposed MLPANN-based model with well-structured implementation network architecture, empowered with the right hyperparameter tuning algorithm for optimal predictive analysis of practical path loss data. The stepwise exploratory method explored to develop the proposed MLPANN-based model is highlighted as follows:
  • Acquire the path loss datasets
  • Preprocessing the dataset and splitting
  • Obtain the MLP neural network model.
  • Identify the adaptive learning algorithms for MLP neural network model training and testing
  • Identify the modelling hyperparameters
  • Select a hyperparameter tuning algorithm for the model (e.g., Bayesian optimization search, grid search, etc.).
  • Obtain a set of the tuned best hyperparameter values.
  • Train the model to obtain the best hyperparameter values combination.
  • Appraise and validate the accuracy of the training process.
  • Repeat the process to optimal configuration and best-desired results for the model.
  • Engage the model with well-structured implementation network architecture, learning algorithm and set of the tuned best hyperparameter values to conductive predictive path loss modelling.

3.1. Data Collection

The field measurement was conducted to acquire live signal data around three Long Term Evolution (LTE) transceiver base station antennas for one year (i.e., 12 months). The measurement took one year to cater to the study locations’ seasonal variations and three LTE transceiver base station antennas operating at 2600 MHz with 10 MHz bandwidth [43]. The transceiver base station antennas (called NodeBs) are sectorized with 17.5 dBi gain and 43 dBm transmit power. The LTE network belongs to one of the major GSM/WCDMA/HSPA/LTE telecom service providers operating across major towns, villages and cities in Nigeria. The measurements were performed with field test tools with TEMS application tools for radio spectrum analysis. The test tools, some of which are displayed in Figure 1, include a Rover car, scanner, two Samsung mobile phones, and an HP lap, were explored to assess the performance of eNode B over the LTE radio air interface by connecting mobile phones directly to the Node B transmitters. To obtain the eNode B locations and delineate measurement data locations/information, the Global Positioning System (GPS) equipment was employed. The path loss data to be predicted are related to the acquired radio signal data by the measured path loss data where P L m e a ( dB ) , values have been obtained from the measured signal, RSRP (dBm) by Equation (1):
P L m e a ( dB ) = E I R P + G A R S R P m e a s
With EIRP calculated as Equation (2):
EIRP = PTX + GTXCLTX
where GTX and G A are the base station (BS) transmit antenna gain and receiver (MS) antenna gain, respectively, PTX is the transmitted power, and CLTX, denotes transmission cable loss, all in dB. Table 1 reveals some of the key BS antenna site parameters acquired during the field drive test for calculation.

3.2. The MLP Neural Network Model

The first step towards effectively engaging neural networks for predictive modelling is to know the exact type you want to use and determine network architecture. This paper considers the most robust and special type of neural network: the multi-layer perceptrons (MLP). A single perceptron (LP) has limitations in terms of input-desired output mapping capability. This limitation is because it only contains a single neuron per adaptable synaptic weights and bias; thus, it is only proficient in catering to ridge-like function, notwithstanding the type activation function explored [42]. The above limitation can be catered to using an MLP neural network with more source nodes with data input and output layers sandwiched with hidden layers nodes. Multiple layers of neurons in the MLP network provide enhanced input-desired output mapping capability.
Figure 2 displays a structure of a classic feedforward MLP network model composed of g 1 , g 2 ,… g I , inputs, and predicted output, ( y 1 , y 2 , , y N ) with kh hidden nodes, h number. The respective weights connecting the input and hidden layer, as well as the weights connecting the hidden layer and the output layer, are designated by w i j 1 and w j n 2 , while C j indicates the hidden nodes thresholds. The network learns the correlation between input datasets and predicted output feedback by varying weight and bias values. Accordingly, the MLP network predicted output in correspondence to jth neurons with the kth node could be articulated as (3):
y ^ n t = j = 1 k h w j 2 F w i j 1 g i ( t ) + c j
for 1 n m , 1 j k h , w j , j = 0 , 1 , , k h , w i j , i = 0 , 1 , , m ; j = 0 , 1 , , k h
where:
m, h and kh indicate the input node number, hidden node and hidden node number, respectively; i designate input to j hidden layer neuron.
The F (·) in Equation (1) denotes the sigmoid activation function, an import function usually utilized in the MLP network. It can be defined by Equation (4):
F ( a ) = 1 1 + e a
where F(x) is at all times in the range [−1, 1], with F(a) being a set of real numbers.
The weights w i j 1 and w j n 2 , including the threshold C j , are unknown and thus can be chosen to update and reduce the error during prediction. The prediction error can be expressed by employing the expression (5):
ε n = 1 2 n = 1 y n y ^ n 2
where,
y n and y ^ n represent the target (i.e., actual) data and their predicted output; and n = 1…, N, with N indicating the actual data sample number.
In MLP training, the error verve for assessing the network learning improvement related to convergence speed is the generalized aggregate error values. It is often computed using mean square error (MSE). The MSE can be obtained from the least square formation of Equation (6).
M S E = 1 N n = 1 N y n y ^ n 2
In this work, the feedforward MLP network model explored for path loss predictive modelling is displayed in Figure 3.

3.3. MLP Modelling Parameters and Search Space

Hyperparameters are a special set of regulating parameters that the NN model utilizes for the adaptive learning process in data training and testing. The special parameters may be categorical, continuous or integer variables whose values range are usually lower and upper bounded. Thus, there exists several MLPs directly impacting the predictive modelling. They include the hidden layers number, neurons number in the hidden layer, transfer function, etc. A summarised description of the transfer function is given in the following subsection.

3.3.1. The Hidden Layers

Deciding the number of the hidden layers is one of the most important issues while investigating the neural network architecture for predictive modelling and data mining. Using too many hidden layers can result in poor generalization and complex neural network training. According to authors in [44,45,46,47], two hidden layers, combined with m output neurons, are adequate for a neural network to learn N data samples and produce negligibly minor errors.
Previous studies have examined the suitability of several machine models for path loss predictions as contained in [48,49,50]. The need to overcome the problems of empirical models when used for path loss predictions led to an artificial neural network [49]. ANN path loss prediction models were also more efficient and easier to deploy than deterministic models [51]. In [52,53], analyses of empirical models with different propagation features were performed, and the model with the lowest RMSE value was then compared with the prediction from ANN. The ANN-based path loss prediction model produced a much lower value of MSE upon validation. In [54], a multi-layer perceptron neural network was introduced for path loss prediction. The MLP network was then trained with a backpropagation algorithm. The MLP-based prediction was compared with predictions from analytical models, and the results indicated the former to be efficient for radio network planning and optimization.
ANN was also used for path loss prediction in urban areas [55]. The work explored the effect of the various input parameters and the environmental terrains on the robustness of the path loss prediction. One key finding from the study is that the accuracy of the signal prediction model increases with more input parameters: the greater the number of features, the greater the system’s accuracy. This trend is because machine learning algorithms thrive well with the availability of large datasets. The model is trained with the help of the data and, in this case, the input features. An ANN-based path loss model at 800 MHz and 1800 MHz introduced in [47,48] were input for longitude, latitude, distance, elevation, clutter height and elevation. The ANN method in [56] outperforms the Support Vector Machine (SVM)- and the Random Forest (RF)-based predictions.
In [57], an artificial neural network was used for path loss prediction in a smart campus environment at 1800 MHz. There were two hidden layers for this network, and the performance of the network outperforms the prediction made by using RF. Moreover, in [58,59,60], several machine learning-based prediction models were introduced for signal predictions for wireless sensor networks. The machine learning-based prediction model in [61,62] also produced the lowest values of RMSE when compared to the other analytical models in a wireless sensor network.

3.3.2. Neurons Number in the Hidden Layers

Determining the neurons number in the hidden layers remains an integral part of the inclusive neural network architecture. An inadequate neurons number in the hidden layers can lead to an underfitting problem. Underfitting arises once there are insufficient neurons number in the hidden layers to learn or detect the signals satisfactorily, especially in a multifaceted dataset.
On the other hand, using too many neurons can lead to overfitting problems. Overfitting occurs once the neural network contains too much information processing capacity problem. It can also result in excessive time increase during neural network training. The amount of training time can increase to the point that it is impossible to train the neural network adequately. It is evident that some give and take must be grasped between too few and too many neurons number in the hidden layers.

3.3.3. Transfer Function

The transfer function is a singular, monotonically increasing and differentiable function used for translating the input data signals to produce the final output signals of a neuron. The transfer function is fundamental to the concrete concept of neural networks mainly for two key reasons. First, without activation functions, the entire organization of the neural network will be similar to a typical linear function that cannot learn non-linear relationships. Second, transfer function styles and graces the main computation accomplished by neural networks.

3.3.4. Learning Rate

The learning rate is another vital hyperparameter that regulates or fine tunes the weights of NN in relation to the loss gradient. Its value must be cautiously chosen to support both optimization and generalization robustly. A too-large learning rate value can cause the entire learning process to jump over minima. Similarly, a too-small learning rate value can make the entire learning process too long to converge, resulting in it being trapped in negative and spurious local minima.

3.4. Hyperparameter Tuning

Hyperparameter tuning or optimization expresses the robust procedure of identifying and finding the best feasible values of hyperparameters for a machine learning model to attain the desired resultant modeling outcome. Popular hyperparameter tuning algorithms in the literature include random search, grid search and Bayesian optimization search. In this paper, the last two methods are considered. Grid search is a standard hyperparameter optimization technique wherein a list of critical parameters is selected with attached feasible values for each parameter, followed by training the model for every single blend and then choosing the values that yield the most desired resultant outcome. The Bayesian optimization method is a special sequential model-based optimization (also known as Bayesian optimization), utilizing the ‘Bayes Theorem’ to conduct an automatic hyperparameter search. Particularly, the Bayesian optimization search algorithm utilizes the upshot from the preceding iteration to select the next hyperparameter values.

3.5. MLP Learning Algorithms

The training algorithm used for a neural network system to learn and solve a problem is essential. The correct training algorithm from the available sundry types depends on diverse factors, including data sample size, task type, training time constraints, precision/accuracy requirements, etc. It is demanding to find out which training algorithm will produce the most satisfactory results. For example, suppose it is a predictive modelling task with function approximation. In that case, the dominant or most common ones are backpropagation training algorithms, which involve carrying out computations backwards over the network to fine-tune the weights and minimize performance error.

3.6. MLP Network Model Implementation Process

MATLAB is a distinctive programming language with a multi-exemplar numerical computing environment and a user interface. It provides easy matrix manipulation, graphical multi-domain simulation, figurative computing, creative functions plotting, excellent data mining, easy algorithms implementation, etc. MATLAB allows access to optional toolbox uses. The neural network toolbox has special tools for model-based design, implementation, visualization and simulation of neural networks. MATLAB is employed to encode the script files for the MLP network model predictive training, testing and quantitative evaluation in this work. The program code for conventional path loss calculation and assessment is also explored. The proposed MLP neural model consists of five input nodes and one output node. Flowchart for executing proposed predictive modelling with ANN while training and testing are shown in Figure 4.
As mentioned earlier, for practical and optimal application of MLP network for predictive modelling purposes, the right choice of the learning algorithm, and selection of the network processing elements such as the number of neurons, number of network layers and transfer functions, are crucial. For example, a network with insufficient neurons number in a hidden layer may fail to capture complex links between target output and input variables. Conversely, if the number of neurons allotted in the network hidden layer is too many, the network would likely follow the latent noise in the dataset owing to over-parameterization, and this, in turn, can lead to awkward generalization and poor predictive modelling of the original data [63,64]. Therefore, the determination of the hidden layers number and their number of neurons is performed by trial and random selection. However, for conciseness and the need to attain optimal neural network training and testing, the search for the required number of neurons and hidden layers in the network layers were narrowed down to 2–50 and 1–3, respectively.
Generally, if a particular algorithm performs well during the dataset training but flops in the aspect of generalization, we refer to it as overfitting. To improve generalization (or prevent overfitting) during the path loss data training with each of the NN algorithms, we employed input/target transformations and early stopping techniques. Thus, the inputs and targets datasets were scaled to reside in the range [−1, 1] to enhance training and testing speed. Moreover, the early stopping measures were engaged for training and testing to avoid overtraining, eliminate contemptuous impact stimulated by the initial values, and develop robust adaptive predictive ability. Although many learning algorithms are available for MLP neural network training and testing in MATLAB software, it is demanding to identify which algorithm works best for a given predictive modelling problem concerning convergence speed and accuracy [16]. Therefore, an exhaustive search is employed in this study to accelerate the convergence and evaluate the impact of all the available learning algorithms during network training. The respective learning algorithms assessed to develop an optimal MLP network predictive model and their weight adaptation techniques are listed in Table 2. The target of the weight adaptation is to determine the optimum weight update for the input-output data array pair during training.

4. Results and Discussions

As mentioned earlier, many factors directly impact the development of an excellent back-propagation neural network predictive model, especially with a trial and error method, as adopted in this work. They include training algorithm, hidden layers number, neurons number in the hidden layer, transfer function, momentum term, learning rate, etc. Here, the concentration is on the training algorithm, hidden layers number and neurons number in the hidden layer. To obtain the predictive path loss modelling results, first we divided the path loss data into portions and employed the grid search-based hyperparameter tuning method to generate configurations for the parted data chunks for training and testing. The performance results of the proposed method were evaluated and reported for each machine learning algorithm described in Table 1, using Mean Absolute Error (MAE), Mean Square Error (MSE), Root Mean Square Error (RMSE), Coefficient of Efficiency (COE), Correlation Coefficient (R) and Standard Deviation Error (STD) [65].
Secondly, the predictive path loss modelling was conducted for three study locations using the standard Bayesian optimization for hyperparameter tuning. The results were compared to our first results using the grid search-based hyperparameter tuning method.

4.1. Neurons Number Impact

Determining the neurons number in the hidden layers also remains integral in the inclusive neural network architecture. An inadequate neurons number in the hidden layers can lead to an underfitting problem. Underfitting arises once there is insufficient neuron number in the hidden layers to satisfactorily learn or detect the signals, especially in a multifaceted dataset. On the other hand, making use of too many neurons can lead to overfitting problems. Overfitting takes place once the neural network contains too much information processing. It can also result in excessive time increase during neural network training. The amount of training time can increase to the point that it is impossible to adequately train the neural network. It is very clear at this point that some give and take must be grasped between too few and too many neurons number in the hidden layers. Accordingly, by starting with the fastest training algorithm, which is Levenberg–Marquardt (lm), the network was trained and tested with 2, 5, 10, 15, 20, 25, 30, 35, 40, 45 and 50 neurons. This is to ascertain their incremental impact on its performance. Table 3, Figure 5 and Figure 6 display the detailed overall predictive performance of each neurons number using different error statistics. As seen in Table 3, it was expected that a continuing increase in neurons number per layer would result in an upturn in the resolution of the neural network prediction fitting pattern to the measured loss data. At 30 neurons per layer, the neural network model had a good enough fit to the measured loss data with an MAE value of 2, RMSE value of 2.71, STD value of 1.82, R-value of 95 (%) and COE value of 90 (%). Increasing the neuron number to 50 showed no performance improvement, as seen in Figure 6.

4.2. Transfer Function Impact

The transfer function is a singular monotonically increasing and differentiable function used for translating the input data signals to produce the final output signals of a neuron. The transfer function is fundamental to the concrete concept of neural networks mainly for three key reasons. Firstly, without activation functions, the entire organization of the neural network will be similar to an ordinary linear function that cannot learn non-linear relationships. Secondly, transfer function styles grace the main computation accomplished by neural networks. Thirdly, transfer functions possess the proclivity to boost the learning rate and formation patterns in datasets. Thus, the choice of the right transfer (activation) function also positively influences the performance of the NN training algorithm. Table 4 presents the list of sigmoid transfer functions employed in this study to ascertain the stability of the proposed neural network model. The performance of sigmoid transfer functions in terms of MAE, MSE, RMSE, R and STD are also displayed in Table 4. The standard deviation (STD) statistics with one, two and three layers of training are given in Figure 7, and the Root Mean Error performance statistics with one, two and three layers of training are shown in Figure 8.

4.3. Training Algorithm and Hidden Layers Number Impact

The learning algorithm and hidden layers number also significantly impact the success of the neural network in coming up with a healthy predictive model. Therefore, deciding on the number of the hidden layer is one of the most important issues that come up while investigating the neural network architecture for predictive modelling and data mining. Using too many hidden layers can result in poor generalization and complex neural network training. According to the work [31], two hidden layers, combined with m output neurons, are adequate for a neural network to learn N data samples and produce negligibly small errors.
Here, the impact of many learning algorithms has been studied with one, two and three hidden layers numbers. Interestingly, results reveal that the two-layered network is superior to one-layered and three-layered layer network for all the 12 learning algorithms investigated. Interestingly, results show that the neural network architecture trained using lm (i.e., the Levenberg–Marquardt training algorithm) with two hidden-layer sizes and logsig-transit transfer function gave the best performance. Table 5 displays detailed network training/testing error statistics results and hidden layer numbers for each learning algorithm.

4.4. Performance of Grid Search Algorithm and Bayesian Optimisation Search Algorithm for Hyperparameter Tuning

The hyperparameter tuning process has a weighty influence on neural network learning performance. Given the computational resources requisite, the hyperparameters of high relevance receive superior usage in the hyperparameter tuning process. Hyperparameters with a more robust influence on weights are more effective during neural network training. Thus, the appropriate choice of hyperparameters selected for neural network model training influences the network training and performance.
As displayed in Table 6, the proposed MLP neural network model results with grid search algorithm-based hyperparameter tuning (optimization) are compared with those obtained using the traditional Bayesian optimization search-based hyperparameter tuning approach for path loss data predictive analysis using location one as a case study. We have reported the results attained for lm and br learning for brevity. While the grid search-based hyperparameter tuning performs an in-depth and comprehensive search on the hyperparameters in a stepwise manner as set specified by users with limited search space, the Bayesian search-based hyperparameter tuning performs a sequential-based search on the hyperparameters via several trials, without the user having preliminary information of the hyper-parameters distribution. From the results in Table 6, it is clear that the proposed MLP neural network model with grid search algorithm-based hyperparameter tuning outperforms the ones obtained using Bayesian Optimization search-based hyperparameter tuning. Next, Section 4.5 is to apply the proposed MLP neural network model with grid search algorithm-based hyperparameter tuning for detailed predictive path loss analysis across the three study locations.

4.5. Evaluation of Proposed Neural Network Model at Different Locations

The evaluation results of the proposed neural network model at different study locations are presented as follows. Figure 9, Figure 10 and Figure 11 show the proposed neural network model prediction with measured path loss data configured with the Levenberg–Marquardt training algorithm, two hidden-layer sizes and logsig-transig transfer function—the prediction performance of the developed neural network model in terms of R and MSE values. The R-value measures the prediction correlation between outputs (predicted loss data) and targets (actual loss data). The closeness of the R-value to 1 corresponds to a high positive correction. Otherwise, it is poorly correlated. Figure 12, Figure 13 and Figure 14 are plotted R values between the predicted loss data and the actual loss values for sites 1 to 3 during training, validation and testing with neural networks. The R-values obtained from the plots are 0.97, 0.93 and 0.94 for site 1, 0.92, 0.93 and 0.94 for site 2 and 0.91, 0.93, 0.96, 0.94 for site 3. The performance plots in Figure 15, Figure 16 and Figure 17 indicate that the MSE becomes smaller with the epoch number (one complete training/testing/validation cycle). The word ‘epoch’ is used here to mean a special hyperparameter term that defines the number of times (in terms of iteration) that the NN algorithms undergo during the entire data training duration. The error of test and validation display similar characteristics while predicting the measured loss across sites 1 to 3. Specifically, the validation MSE error shows that the proposed neural network model would not generalize well or fit the measured loss data well if trained further than 4, 8 and 8 epochs. The mean prediction error along measurement data points in sites 1, 2 and 3 are presented in Figure 18, Figure 19 and Figure 20.

4.6. Comparison of Prediction Accuracy of Proposed Neural Network Model with Log-Distance Models

Detailed prediction capabilities of all the log-distance models and the proposed model on measured path loss data are provided in the plotted graphs of Figure 21, Figure 22 and Figure 23 in terms of MAE, RMSE and STD. The graphs show that the neural network model achieved the best predictions with marginal errors. The COST 213 (W/I) made the closest prediction to measured loss, but in terms of accuracy, the proposed neural network model achieved the best performance by 20%, 15% and 25%, respectively, across study sites. For example, while COST 213 (W/I) reached 3.34, 2.35 and 4.23 dB in terms of RMSE, the proposed neural network model attained 1.73, 2.11 and 1.45 dB across study sites. Generally, models which predict the path loss in the tested areas with RMSEs higher than the acceptable range of up to 6 dB are not selected as most suitable. However, such models could be further optimized for improved performance. The lower the RMSE value towards zero, the better the model. Regarding standard deviation error, COST 213 (W/I) achieved 1.73, 2.11 and 1.45 dB, while the proposed neural network model achieved 1.73, 2.11 and 1.45 dB, respectively. The poor predictions made by the log-distance-based models can be ascribed to due to dissimilarities and variations in environmental formations (hilly, mountainous or quasi-plain), weather conditions, soil electrical properties and terrain types that exist in different radio propagation environments.

5. Conclusions

The growing demand for mobile and fixed cellular telecommunication services have given substantial weight to the limited available radio frequency spectrum. Proper modelling and precise signal coverage predictions are crucial to utilizing this scarce resource effectively. Reliable predictive modeling of signal path loss aids in controlling the load on base station transmitters and assists in designing efficient radio network channels with less interference and coverage hole problems. The conventional log-distance-based statistical models for path loss prediction comprising the clustering factor, COST 234 Hata, free space, Hata, Lee models, etc., are generally limited for predicting signal attenuation losses, especially when employed in different environments other than the environment for which they have been designed.
The main objective of this paper was to develop a distinctive MLP-based path loss model with well-structured implementation network architecture, empowered with the grid search-based hyperparameter tuning method for optimal path loss approximation between mobile-station and base-station path lengths. The degree of prediction accuracy with the developed MLP network model over eight conventional log-distance-based path loss models is also clearly provided using first-order statistics. In summary, this research paper has revealed that:
  • MLPANN-based path loss model with well-structured implementation network architecture, empowered with the right hyperparameter tuning algorithm is better than the standard long-distance path loss models
  • the choice of both MLP-ANN modelling structure and selection of training algorithms do have a clear impact on the quality of its prediction proficiency. Specifically, in terms of MAE, RMSE and STD statistical values, the proposed model yielded up to 50% performance prediction accuracies improvement over the standard models on the acquired LTE path loss datasets.
  • The selection of adaptive learning tuning hyperparameters of MLP-ANN and the tuning algorithm both have an impact on its overall predictive modelling capacity.
Future work would consider more hyperparameter selection techniques to optimize MLP model prediction accuracy during NN training. We also intend to explore more super layered training capacity of deep neural networks such as the long-short memory (LSTM) network model for predictive modelling of path loss data in our work.

Author Contributions

The manuscript was written through the contributions of all authors. J.I. was responsible for the conceptualization of the topic; article gathering and sorting were carried out by J.I., A.L.I. and S.O.; manuscript writing, original drafting and formal analysis were carried out by J.I., A.L.I., S.O., O.K., Y.K., C.-C.L. and C.-T.L.; writing of reviews and editing was carried out by J.I., A.L.I., S.O., O.K., Y.K., C.-C.L. and C.-T.L.; and J.I. led the overall research activity. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Research Foundation of Korea (NRF) grant funded by the Korean government (MSIT) (No. 2020R1G1A1099559).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that supports the findings of this paper are available from the corresponding author upon reasonable request.

Acknowledgments

The work of Agbotiname Lucky Imoize is supported in part by the Nigerian Petroleum Technology Development Fund (PTDF) and in part by the German Academic Exchange Service (DAAD) through the Nigerian–German Postgraduate Program under grant 57473408.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Molisch, A.F. Wireless Communications, 2nd ed.; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2012; ISBN 9780470741870. [Google Scholar]
  2. Goldsmith, A.J. Wireless Communications; Cambridge University Press: Cambridge, UK, 2005. [Google Scholar]
  3. Isabona, J.; Imoize, A.L.; Ojo, S.; Lee, C.-C.; Li, C.-T. Atmospheric Propagation Modelling for Terrestrial Radio Frequency Communication Links in a Tropical Wet and Dry Savanna Climate. Information 2022, 13, 141. [Google Scholar] [CrossRef]
  4. Isabona, J.; Konyeha, C.C.; Chinule, C.B.; Isaiah, G.P. Radio field strength propagation data and pathloss calculation methods in UMTS network. Adv. Phys. Theor. Appl. 2013, 21, 54–68. [Google Scholar]
  5. Nawrocki, M.; Aghvami, H.; Dohler, M. Understanding UMTS Radio Network Modelling, Planning and Automated Optimisation: Theory and Practice; John Wiley & Sons: Hoboken, NJ, USA, 2006. [Google Scholar]
  6. Nawrocki, M.J.; Dohler, M.; Aghvami, A.H. Modern approaches to radio network modelling and planning. In Understanding UMTS Radio Network Modelling, Planning and Automated Optimisation: Theory and Practice; John Wiley & Sons: Hoboken, NJ, USA, 2006; p. 3. [Google Scholar]
  7. Sharma, P.K.; Singh, R.K. Comparative analysis of propagation path loss models with field measured data. Int. J. Eng. Sci. Technol. 2010, 2, 2008–2013. [Google Scholar]
  8. Ajose, S.O.; Imoize, A.L. Propagation measurements and modelling at 1800 MHz in Lagos Nigeria. Int. J. Wirel. Mob. Comput. 2013, 6, 165–174. [Google Scholar] [CrossRef]
  9. Ojo, S.; Imoize, A.; Alienyi, D. Radial basis function neural network path loss prediction model for LTE networks in multitransmitter signal propagation environments. Int. J. Commun. Syst. 2021, 34, e4680. [Google Scholar] [CrossRef]
  10. Imoize, A.L.; Ibhaze, A.E.; Nwosu, P.O.; Ajose, S.O. Determination of Best-fit Propagation Models for Pathloss Prediction of a 4G LTE Network in Suburban and Urban Areas of Lagos, Nigeria. West Indian J. Eng. 2019, 41, 13–21. [Google Scholar]
  11. Ibhaze, A.E.; Imoize, A.L.; Ajose, S.O.; John, S.N.; Ndujiuba, C.U.; Idachaba, F.E. An Empirical Propagation Model for Path Loss Prediction at 2100 MHz in a Dense Urban Environment. Indian J. Sci. Technol. 2017, 10, 1–9. [Google Scholar] [CrossRef]
  12. Rathore, M.M.; Paul, A.; Rho, S.; Khan, M.; Vimal, S.; Shah, S.A. Smart traffic control: Identifying driving-violations using fog devices with vehicular cameras in smart cities. Sustain. Cities Soc. 2021, 71, 102986. [Google Scholar] [CrossRef]
  13. Imoize, A.L.; Tofade, S.O.; Ughegbe, G.U.; Anyasi, F.I.; Isabona, J. Updating analysis of key performance indicators of 4G LTE network with the prediction of missing values of critical network parameters based on experimental data from a dense urban environment. Data Br. 2022, 42, 108240. [Google Scholar] [CrossRef]
  14. Fujimoto, K. Mobile Antenna Systems Handbook; Artech House: Norwood, MA, USA, 2008; ISBN 1596931272. [Google Scholar]
  15. Tataria, H.; Haneda, K.; Molisch, A.F.; Shafi, M.; Tufvesson, F. Standardization of Propagation Models for Terrestrial Cellular Systems: A Historical Perspective. Int. J. Wirel. Inf. Networks 2021, 28, 20–44. [Google Scholar] [CrossRef]
  16. Hanci, B.Y.; Cavdar, I.H. Mobile radio propagation measurements and tuning the path loss model in urban areas at GSM-900 band in Istanbul-Turkey. In Proceedings of the IEEE 60th Vehicular Technology Conference, 2004, VTC2004-Fall, Los Angeles, CA, USA, 26–29 September 2004; Volume 1, pp. 139–143. [Google Scholar]
  17. Ekpenyong, M.; Isabona, J.; Ekong, E. On Propagation Path Loss Models For 3-G Based Wireless Networks: A Comparative Analysis. Comput. Sci. Telecommun. 2010, 25, 74–84. [Google Scholar]
  18. Isabona, J. Wavelet Generalized Regression Neural Network Approach for Robust Field Strength Prediction. Wirel. Pers. Commun. 2020, 114, 3635–3653. [Google Scholar] [CrossRef]
  19. Imoize, A.L.; Oseni, A.I. Investigation and pathloss modeling of fourth generation long term evolution network along major highways in Lagos Nigeria. Ife J. Sci. 2019, 21, 39–60. [Google Scholar] [CrossRef]
  20. Imoize, A.L.; Ogunfuwa, T.E. Propagation measurements of a 4G LTE network in Lagoon environment. Niger. J. Technol. Dev. 2019, 16, 1–9. [Google Scholar] [CrossRef] [Green Version]
  21. Imoize, A.L.; Dosunmu, A.I. Path Loss Characterization of Long Term Evolution Network for. Jordan J. Electr. Eng. 2018, 4, 114–128. [Google Scholar]
  22. Ekpenyong, M.E.; Robinson, S.; Isabona, J. Macrocellular propagation prediction for wireless communications in urban environments. J. Comput. Sci. Technol. 2010, 10, 130–136. [Google Scholar]
  23. Nadir, Z. Empirical pathloss characterization for Oman. In Proceedings of the 2012 Computing, Communications and Applications Conference, Tamilnadu, India, 22–24 February 2012; pp. 133–137. [Google Scholar] [CrossRef]
  24. Hunter, D.; Yu, H.; Pukish III, M.S.; Kolbusz, J.; Wilamowski, B.M. Selection of proper neural network sizes and architectures—A comparative study. IEEE Trans. Ind. Informatics 2012, 8, 228–240. [Google Scholar] [CrossRef]
  25. Imoize, A.L.; Ibhaze, A.E.; Atayero, A.A.; Kavitha, K.V.N. Standard Propagation Channel Models for MIMO Communication Systems. Wirel. Commun. Mob. Comput. 2021, 2021, 8838792. [Google Scholar] [CrossRef]
  26. Khan, N.; Haq, I.U.; Khan, S.U.; Rho, S.; Lee, M.Y.; Baik, S.W. DB-Net: A novel dilated CNN based multi-step forecasting model for power consumption in integrated local energy systems. Int. J. Electr. Power Energy Syst. 2021, 133, 107023. [Google Scholar] [CrossRef]
  27. Jo, H.-S.; Park, C.; Lee, E.; Choi, H.K.; Park, J. Path loss prediction based on machine learning techniques: Principal component analysis, artificial neural network, and Gaussian process. Sensors 2020, 20, 1927. [Google Scholar] [CrossRef] [Green Version]
  28. Guo, D.; Zhang, Y.; Xiang, Q.; Li, Z. Improved radio frequency identification indoor localization method via radial basis function neural network. Math. Probl. Eng. 2014, 2014, 420482. [Google Scholar] [CrossRef] [Green Version]
  29. Guo, Y.; Liu, Y.; Li, S. Modeling and Simulation of Terahertz Indoor Wireless Channel Based on Radial Basis Function Neural Network. In Proceedings of the 2021 International Conference on Microwave and Millimeter Wave Technology (ICMMT), Nanjing, China, 23–26 May 2021; pp. 1–3. [Google Scholar]
  30. Annepu, V.; Rajesh, A.; Bagadi, K. Radial basis function-based node localization for unmanned aerial vehicle-assisted 5G wireless sensor networks. Neural Comput. Appl. 2021, 33, 12333–12346. [Google Scholar] [CrossRef]
  31. Huang, G.-B. Learning capability and storage capacity of two-hidden-layer feedforward networks. IEEE Trans. Neural Netw. 2003, 14, 274–281. [Google Scholar] [CrossRef] [Green Version]
  32. Abhayawardhana, V.S.; Wassellt, I.J.; Crosby, D.; Sellars, M.P.; Brown, M.G. Comparison of empirical propagation path loss models for fixed wireless access systems. In Proceedings of the IEEE Vehicular Technology Conference, Stockholm, Sweden, 30 May–1 June 2005; Volume 61, pp. 73–77. [Google Scholar]
  33. Hinga, S.K.; Atayero, A.A. Deterministic 5G mmWave Large-Scale 3D Path Loss Model for Lagos Island, Nigeria. IEEE Access 2021, 9, 134270–134288. [Google Scholar] [CrossRef]
  34. Haykin, S. Neural networks: A guided tour. Soft Comput. Intell. Syst. theory Appl. 1999, 71, 71–80. [Google Scholar]
  35. Isabona, J.; Osaigbovo, A.I. Investigating predictive capabilities of RBFNN, MLPNN and GRNN models for LTE cellular network radio signal power datasets. FUOYE J. Eng. Technol. 2019, 4, 155–159. [Google Scholar] [CrossRef]
  36. Sotiroudis, S.P.; Goudos, S.K.; Gotsis, K.A.; Siakavara, K.; Sahalos, J.N. Modeling by optimal artificial neural networks the prediction of propagation path loss in urban environments. In Proceedings of the 2013 IEEE-APS Topical Conference on Antennas and Propagation in Wireless Communications (APWC), Turin, Italy, 9–13 September 2013; pp. 599–602. [Google Scholar]
  37. Isabona, J.; Imoize, A.L. Terrain-based adaption of propagation model loss parameters using non-linear square regression. J. Eng. Appl. Sci. 2021, 68, 33. [Google Scholar] [CrossRef]
  38. Phillips, C.; Sicker, D.; Grunwald, D. A Survey of Wireless Path Loss Prediction and Coverage Mapping Methods. IEEE Commun. Surv. Tutorials 2013, 15, 255–270. [Google Scholar] [CrossRef]
  39. Imoize, A.L.; Oyedare, T.; Ezekafor, C.G.; Shetty, S. Deployment of an Energy Efficient Routing Protocol for Wireless Sensor Networks Operating in a Resource Constrained Environment. Trans. Networks Commun. 2019, 7, 34–50. [Google Scholar] [CrossRef] [Green Version]
  40. Imoize, A.L.; Ajibola, O.A.; Oyedare, T.R.; Ogbebor, J.O.; Ajose, S.O. Development of an Energy-Efficient Wireless Sensor Network Model for Perimeter Surveillance. Int. J. Electr. Eng. Appl. Sci. 2021, 4, 1–17. [Google Scholar]
  41. Imoize, A.L. Analysis of Propagation Models for Mobile Radio Reception at 1800MHz. Master’s Thesis, University of Lagos, Lagos, Nigeria, 2011. [Google Scholar]
  42. Isabona, J.; Ojuh, D. Adaptation of Propagation Model Parameters toward Efficient Cellular Network Planning using Robust LAD Algorithm. Int. J. Wirel. Microw. Technol. 2020, 10, 13–24. [Google Scholar] [CrossRef]
  43. Imoize, A.L.; Orolu, K.; Atayero, A.A.-A. Analysis of key performance indicators of a 4G LTE network based on experimental data obtained from a densely populated smart city. Data Br. 2020, 29, 105304. [Google Scholar] [CrossRef] [PubMed]
  44. Ojo, S.; Akkaya, M.; Sopuru, J.C. An ensemble machine learning approach for enhanced path loss predictions for 4G LTE wireless networks. Int. J. Commun. Syst. 2022, 35, e5101. [Google Scholar] [CrossRef]
  45. Ebhota, V.C.; Isabona, J.; Srivastava, V.M. Improved adaptive signal power loss prediction using combined vector statistics based smoothing and neural network approach. Prog. Electromagn. Res. C 2018, 82, 155–169. [Google Scholar] [CrossRef] [Green Version]
  46. Ebhota, V.C.; Isabona, J.; Srivastava, V.M. Base line knowledge on propagation modelling and prediction techniques in wireless communication networks. J. Eng. Appl. Sci. 2018, 13, 1919–1934. [Google Scholar]
  47. Coskun, N.; Yildirim, T. The effects of training algorithms in MLP network on image classification. In Proceedings of the International Joint Conference on Neural Networks, Portland, OR, USA, 20–24 July 2003; Volume 2, pp. 1223–1226. [Google Scholar]
  48. Salman, M.A.; Popoola, S.I.; Faruk, N.; Surajudeen-Bakinde, N.T.; Oloyede, A.A.; Olawoyin, L.A. Adaptive Neuro-Fuzzy model for path loss prediction in the VHF band. In Proceedings of the 2017 International Conference on Computing Networking and Informatics (ICCNI), Lagos, Nigeria, 29–31 October 2017; pp. 1–6. [Google Scholar]
  49. Aldossari, S.; Chen, K.-C. Predicting the path loss of wireless channel models using machine learning techniques in mmwave urban communications. In Proceedings of the 2019 22nd International Symposium on Wireless Personal Multimedia Communications (WPMC), Lisbon, Portugal, 24–27 November 2019; pp. 1–6. [Google Scholar]
  50. Ahmadien, O.; Ates, H.F.; Baykas, T.; Gunturk, B.K. Predicting Path Loss Distribution of an Area From Satellite Images Using Deep Learning. IEEE Access 2020, 8, 64982–64991. [Google Scholar] [CrossRef]
  51. Turan, B.; Coleri, S. Machine learning based channel modeling for vehicular visible light communication. IEEE Trans. Veh. Technol. 2021, 70, 9659–9672. [Google Scholar] [CrossRef]
  52. Ates, H.F.; Hashir, S.M.; Baykas, T.; Gunturk, B.K. Path loss exponent and shadowing factor prediction from satellite images using deep learning. IEEE Access 2019, 7, 101366–101375. [Google Scholar] [CrossRef]
  53. Huang, J.; Cao, Y.; Raimundo, X.; Cheema, A.; Salous, S. Rain statistics investigation and rain attenuation modeling for millimeter wave short-range fixed links. IEEE Access 2019, 7, 156110–156120. [Google Scholar] [CrossRef]
  54. Zhang, Y.; Wen, J.; Yang, G.; He, Z.; Wang, J. Path loss prediction based on machine learning: Principle, method, and data expansion. Appl. Sci. 2019, 9, 1908. [Google Scholar] [CrossRef] [Green Version]
  55. Nguyen, D.C.; Cheng, P.; Ding, M.; Lopez-Perez, D.; Pathirana, P.N.; Li, J.; Seneviratne, A.; Li, Y.; Poor, H.V. Enabling AI in future wireless networks: A data life cycle perspective. IEEE Commun. Surv. Tutor. 2020, 23, 553–595. [Google Scholar] [CrossRef]
  56. Ferreira, G.P.; Matos, L.J.; Silva, J.M.M. Improvement of outdoor signal strength prediction in UHF band by artificial neural network. IEEE Trans. Antennas Propag. 2016, 64, 5404–5410. [Google Scholar] [CrossRef]
  57. Singh, H.; Gupta, S.; Dhawan, C.; Mishra, A. Path loss prediction in smart campus environment: Machine learning-based approaches. In Proceedings of the 2020 IEEE 91st Vehicular Technology Conference (VTC2020-Spring), Antwerp, Belgium, 25–28 May 2020; pp. 1–5. [Google Scholar]
  58. Thrane, J.; Zibar, D.; Christiansen, H.L. Model-aided deep learning method for path loss prediction in mobile communication systems at 2.6 GHz. IEEE Access 2020, 8, 7925–7936. [Google Scholar] [CrossRef]
  59. Shrestha, A.; Mahmood, A. Review of deep learning algorithms and architectures. IEEE Access 2019, 7, 53040–53065. [Google Scholar] [CrossRef]
  60. Nguyen, C.; Cheema, A.A. A deep neural network-based multi-frequency path loss prediction model from 0.8 GHz to 70 GHz. Sensors 2021, 21, 5100. [Google Scholar] [CrossRef]
  61. Challita, U.; Dong, L.; Saad, W. Deep learning for proactive resource allocation in LTE-U networks. In Proceedings of the European Wireless Technology Conference, Dresden, Germany, 15 May 2017. [Google Scholar]
  62. Song, W.; Zeng, F.; Hu, J.; Wang, Z.; Mao, X. An unsupervised-learning-based method for multi-hop wireless broadcast relay selection in urban vehicular networks. In Proceedings of the 2017 IEEE 85th Vehicular Technology Conference (VTC Spring), Sydney, NSW, Australia, 4–7 June 2017; pp. 1–5. [Google Scholar]
  63. Ebhota, V.C.; Isabona, J.; Srivastava, V.M. Environment-Adaptation Based Hybrid Neural Network Predictor for Signal Propagation Loss Prediction in Cluttered and Open Urban Microcells. Wirel. Pers. Commun. 2019, 104, 935–948. [Google Scholar] [CrossRef]
  64. Isabona, J.; Babalola, M. Statistical Tuning of Walfisch-Bertoni Pathloss Model based on Building and Street Geometry Parameters in Built-up Terrains. Am. J. Phys. Appl. 2013, 1, 10–16. [Google Scholar]
  65. Bird, J. Engineering Mathematics, 5th ed.; Newness: Los Angeles, CA, USA, 2010; ISBN 9780750681520. [Google Scholar]
Figure 1. Illustration of the TEMS Drive Test Measurement System.
Figure 1. Illustration of the TEMS Drive Test Measurement System.
Applsci 12 05713 g001
Figure 2. Scheme of a three-layered ANN multi-layer perceptron.
Figure 2. Scheme of a three-layered ANN multi-layer perceptron.
Applsci 12 05713 g002
Figure 3. ANN MLP Model Structure for Path Loss Prediction.
Figure 3. ANN MLP Model Structure for Path Loss Prediction.
Applsci 12 05713 g003
Figure 4. Flowchart for the execution of Proposed Predictive Modelling with MLP Neural Network.
Figure 4. Flowchart for the execution of Proposed Predictive Modelling with MLP Neural Network.
Applsci 12 05713 g004
Figure 5. Overall Performance Error Statistics with MAE, STD and RMSE.
Figure 5. Overall Performance Error Statistics with MAE, STD and RMSE.
Applsci 12 05713 g005
Figure 6. Overall Performance Error Statistics with R (%) and COE (%).
Figure 6. Overall Performance Error Statistics with R (%) and COE (%).
Applsci 12 05713 g006
Figure 7. Standard deviation (STD) statistics with one, two and three layers of training.
Figure 7. Standard deviation (STD) statistics with one, two and three layers of training.
Applsci 12 05713 g007
Figure 8. Root Mean Error performance statistics with one, two and three layers of training.
Figure 8. Root Mean Error performance statistics with one, two and three layers of training.
Applsci 12 05713 g008
Figure 9. Comparison between measured loss and the prediction ANN model in site 1.
Figure 9. Comparison between measured loss and the prediction ANN model in site 1.
Applsci 12 05713 g009
Figure 10. Comparison between measured loss and the prediction ANN model in site 2.
Figure 10. Comparison between measured loss and the prediction ANN model in site 2.
Applsci 12 05713 g010
Figure 11. Comparison between measured loss and the prediction ANN model in site 3.
Figure 11. Comparison between measured loss and the prediction ANN model in site 3.
Applsci 12 05713 g011
Figure 12. Prediction performance with correlation coefficient in site 1.
Figure 12. Prediction performance with correlation coefficient in site 1.
Applsci 12 05713 g012
Figure 13. Prediction performance with correlation coefficient in site 2.
Figure 13. Prediction performance with correlation coefficient in site 2.
Applsci 12 05713 g013
Figure 14. Prediction performance with correlation coefficient in site 3.
Figure 14. Prediction performance with correlation coefficient in site 3.
Applsci 12 05713 g014
Figure 15. Network training cycles in site 1.
Figure 15. Network training cycles in site 1.
Applsci 12 05713 g015
Figure 16. Network training cycles in site 2.
Figure 16. Network training cycles in site 2.
Applsci 12 05713 g016
Figure 17. Network training cycles in site 3.
Figure 17. Network training cycles in site 3.
Applsci 12 05713 g017
Figure 18. Mean prediction error statistics along with Data points in site 1.
Figure 18. Mean prediction error statistics along with Data points in site 1.
Applsci 12 05713 g018
Figure 19. Mean prediction error statistics along with Data points in site 2.
Figure 19. Mean prediction error statistics along with Data points in site 2.
Applsci 12 05713 g019
Figure 20. Mean prediction error statistics along with Data points in site 3.
Figure 20. Mean prediction error statistics along with Data points in site 3.
Applsci 12 05713 g020
Figure 21. A comparison of mean absolute error statistics between the proposed ANN model and log-distance models on measured path loss in site locations 1, 2 and 3.
Figure 21. A comparison of mean absolute error statistics between the proposed ANN model and log-distance models on measured path loss in site locations 1, 2 and 3.
Applsci 12 05713 g021
Figure 22. A comparison of root mean absolute error statistics between the proposed ANN model and log-distance models on measured path loss in site locations 1, 2 and 3.
Figure 22. A comparison of root mean absolute error statistics between the proposed ANN model and log-distance models on measured path loss in site locations 1, 2 and 3.
Applsci 12 05713 g022
Figure 23. A comparison of standard deviation error statistics between the proposed ANN and log-distance models on measured path loss in site locations 1, 2 and 3.
Figure 23. A comparison of standard deviation error statistics between the proposed ANN and log-distance models on measured path loss in site locations 1, 2 and 3.
Applsci 12 05713 g023
Table 1. Measurement Path Loss Computation Parameters.
Table 1. Measurement Path Loss Computation Parameters.
ParametersSite 1Site 2Site 3
BS Operation Transmitting Frequency (MHz)260026002600
BS Antenna Height (m)283045
MS Antenna Height (m)1.51.51.5
BS antenna gain (dBi)17.517.517.5
MS antenna gain (dBi)000
MS Transmit power (dB)303030
BS Transmit power (dB)434343
Transmitter cable loss (dB)0.50.50.5
Feeder Loss (dB)333
Table 2. Respective MLP Network Learning Algorithm.
Table 2. Respective MLP Network Learning Algorithm.
Learning AlgorithmWeight AdaptationTraining Acronym
Levenberg–Marquardt (lm) The weight of the lm algorithm is updated via:
w q + 1 = w q ( H + η I ) 1 ε q
where the Hessian matrix, H = J T J .
trainlm
Bayesian Regularization (br) The br approach involves the modification of the performance function, ε q plus regularisation term
F ( w ) = β ε q + α ε w
where β & α are special function parameters and a regularisation term, ε w = w 2
trainbr
Polak–Ribiere Conjugate Gradient (cgp)The weight of cgp algorithm is updated by:
w q + 1 = w q + a q p q
With p q = g q + b q p q 1 and b q = g q 1 T g q g q 1 T g q 1
traincgp
Fletcher-Powell Conjugate Gradient (cgf)The weight of scg algorithm is updated by:
w q + 1 = w q + a q p q
With p q = g q + b q p q 1 and b q = g q 1 T g q g q 1 T g q 1
traincgf
Scaled Conjugate Gradient (scg)The weight of scg algorithm is updated by:
w q + 1 = w q + a q p q
trainscg
Resilient Backpropagation (rp)With RP algorithm, weight update is by
w q + 1 = w q s i g n ε q w q . ( q )
where ( q ) = individual step size for weight adaptation
trainrp
BFGS Quasi-Newton (bfg)bfg weight update is accomplished via
w q + 1 = w q H 1 g q
where H q 1 indicates the Hessian matrix inversion on iteration q.
trainbfg
Conjugate gradient with Powell/Beale Restarts (cgb)The cgb employs update search direction by:
g q 1 g q = 0.2 g q 2
traincgb
Gradient Descent with Adaptive Learning Rate (gda)gda weight update is accomplish via w q + 1 = w q + ε η q + 1 w q
where α indicates the user-defined momentum factor and it ranges from 0 to 1
taingda
Gradient Descent Variable Learning Rate (gdx)With gdx algorithm, weight update is by
w q + 1 = w q η q + 1 s i g n ε q w q . ( q )
where ( q ) = individual step size for weight adaptation and η q = variable learning rate
traingdx
One Step Secant (oss)With oss algorithm, weight update is by
w q + 1 = w q H q 1 g q
where H q = Hessian matrix (2nd derivatives)
trainoss
Gradient Descent with Momentum (gdm)gdm weight update is accomplish via w q + 1 = w q η g q + α ε w q 1
where α indicates the user-defined momentum factor and it ranges from 0 to 1
traingdm
Gradient Descent (gd)With gm algorithm, weight update is by
w q + 1 = w q η g q
where η = learning rate
traingd
Table 3. Neurons Number impact on MLP Network Predictive Modelling with LM.
Table 3. Neurons Number impact on MLP Network Predictive Modelling with LM.
TrainingTestingOverall Performance
No of NeuronsMSERMSERMAESTDRMSERCOE (%)
212.910.887423.220.79132.962.413.810.863775
512.220.864311.930.92052.922.403.780.863775
107.320.93208.010.89002.962.383.800.925186
157.950.922915.720.83392.411.923.080.911783
208.590.924713.790.91862.341.953.050.914884
257.830.929653.080.48652.472.963.850.863575
305.080.959117.060.85582.001.822.710.949990
354.880.953122.450.84821.952.062.840.925586
407.290.936014.310.79732.162.363.200.913284
457.950.931511.060.82572.232.073.040.914384
503.620.967321.070.76242.133.664.240.849872
Table 4. Transfer Functions used for network training/testing and their performance.
Table 4. Transfer Functions used for network training/testing and their performance.
TrainingTestingOverall Performance
Transfer FunctionMSERMSERMAESTDRMSER
purelin19.740.861418.290.86923.362.654.280.8636
tansig6.820.936626.070.78652.192.453.290.9073
logsig4.620.959423.090.78122.152.303.150.9106
purelin-purelin20.330.871513.090.89643.592.754.520.8637
purelin-tagsig7.830.932512.200.85392.381.802.990.9173
purelin-logsig9.180.917320.240.87552.522.023.240.9026
tansig-tansig3.180.969720.170.86431.821.892.630.9365
tansig-logsig3.230.969723.650.80941.822.182.840.9266
tansig-purelin6.390.762210.280.69072.041.822.730.9315
logsig-logsig4.820.951110.310.93821.761.632.400.9290
logsig-tansig3.970.966912.960.84671.601.702.340.9499
logsig-pureline6.010.943016.450.84712.181.912.900.9222
purelin-purelin-purelin13.070.880418.340.84132.932.393.780.8637
tansig-tangsig-tansig5.870.94878.890.91811.881.782.590.9283
tansig-tansig-purelin6.380.933717.890.86072.172.062.760.9200
tangsig-tansig-logsig5.110.95688.980.89071.921.932.720.9318
logsig-logsig-logsig7.180.935813.860.91872.151.852.830.9285
logsig-logsig-purelin9.120.923898.480.63842.914.034.970.8049
logsig-tansig-tansig5.340.94897.520.94292.181.922.900.9245
logsig-tansig-logsig6.900.94283.970.94671.981.752.670.9340
tansig-logsig-tansig6.530.94333.680.94681.971.562.470.9445
logsic-tansig-purelin6.920.939211.340.90932.092.022.900.9222
purelin-logsig-tansig25.570.858325.940.87073.283.584.860.8597
Table 5. Learning Algorithm and Hidden Layers Number impact on MLP Network.
Table 5. Learning Algorithm and Hidden Layers Number impact on MLP Network.
TrainingTestingOverall Performance
Training AlgorithmNo of Hidden LayersMSERMSERMAESTDMSER
lm17.650.935911.550.90332.201.822.860.9248
23.970.966912.960.84671.601.702.340.9499
33.940.96329.310.93111.711.702.410.9473
Br13.720.966026.590.75601.791.992.680.9340
20.570.995347.070.75621.052.552.760.9408
30.920.991925.680.30751.736.096.330.7557
Bfg17.880.925214.880.90702.561.873.170.9076
29.120.909212.920.88692.421.813.030.9149
310.040.915910.930.84252.452.043.200.9045
Rp17.680.93165.030.96012.091.752.730.9323
24.380.933917.020.91231.901.092.600.9456
35.030.905228.590.88162.692.433.630.8755
Scg17.850.928211.080.86082.261.912.970.9185
26.690.937313.520.75312.261.862.920.9209
311.160.89519.650.89512.661.943.290.9005
Cgb19.010.916814.750.88822.491.943.310.9089
28.300.92938.890.92332.361.741.950.9202
316.360.846029.720.88143.262.754.270.8510
Cgf18.440.918612.650.90792.421.783.000.9165
29.150.91528.930.92042.331.882.990.9168
317.840.827421.140.85963.392.724.350.8449
Cgp118.220.849623.490.78183.572.364.280.8426
217.190.847211.260.90393.052.614.010.8482
338.020.746336.790.70993.062.614.000.7486
Oss19.530.916712.550.87032.701.883.280.9015
29.060.91619.770.91162.491.982.490.9055
310.160.910116.900.80842.5920.053.300.8984
Gdx111.820.880515.020.85842.912.113.590.8794
29.590.906018.150.84302.542.093.290.8982
318.600.828612.070.82523.503.094.660.8113
Gdm11.42 × 1030.80081.26 × 1030.704130.2621.0236.840.7870
2447.600.7801571.300.538516.9712.3721.010.8034
31.22 × 1030.43401.5 × 1031.5 × 10326.6723.6136.380.4504
Gd12.94 × 1030.72383.08 × 1030.588253.8116.0456.150.6286
2655.140.8583724.470.826825.677.1426.350.8124
3468.860.7885433.890.758953.7243.1168.880.7890
Table 6. Comparison of Hyperparameter Tuning algorithm performance with Grid Search and Bayesian Optimisation search.
Table 6. Comparison of Hyperparameter Tuning algorithm performance with Grid Search and Bayesian Optimisation search.
TrainingTestingOverall Performance
Hyperparmeter Tuning AlgorithmMLP AlgorithmMSERMSERMAESTDMSER
Grid Searchlm3.970.966912.960.84671.601.702.340.9499
br0.570.995347.070.75621.052.552.760.9408
Bayessian Optimisation searchlm5.200.956516.730.83552.294.90
br3.520.966918.840.85673.404.24
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Isabona, J.; Imoize, A.L.; Ojo, S.; Karunwi, O.; Kim, Y.; Lee, C.-C.; Li, C.-T. Development of a Multilayer Perceptron Neural Network for Optimal Predictive Modeling in Urban Microcellular Radio Environments. Appl. Sci. 2022, 12, 5713. https://doi.org/10.3390/app12115713

AMA Style

Isabona J, Imoize AL, Ojo S, Karunwi O, Kim Y, Lee C-C, Li C-T. Development of a Multilayer Perceptron Neural Network for Optimal Predictive Modeling in Urban Microcellular Radio Environments. Applied Sciences. 2022; 12(11):5713. https://doi.org/10.3390/app12115713

Chicago/Turabian Style

Isabona, Joseph, Agbotiname Lucky Imoize, Stephen Ojo, Olukayode Karunwi, Yongsung Kim, Cheng-Chi Lee, and Chun-Ta Li. 2022. "Development of a Multilayer Perceptron Neural Network for Optimal Predictive Modeling in Urban Microcellular Radio Environments" Applied Sciences 12, no. 11: 5713. https://doi.org/10.3390/app12115713

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop