Skip Content
You are currently on the new version of our website. Access the old version .
Applied SciencesApplied Sciences
  • Article
  • Open Access

30 July 2022

Improving Path Loss Prediction Using Environmental Feature Extraction from Satellite Images: Hand-Crafted vs. Convolutional Neural Network

,
and
1
School of Digital Science, Universiti Brunei Darussalam, Tungku Link, Gadong BE1410, Brunei
2
Institute of Applied Data Analytics, Universiti Brunei Darussalam, Tungku Link, Gadong BE1410, Brunei
*
Author to whom correspondence should be addressed.
This article belongs to the Section Electrical, Electronics and Communications Engineering

Abstract

There is an increased exploration of the potential of wireless communication networks in the automation of daily human tasks via the Internet of Things. Such implementations are only possible with the proper design of networks. Path loss prediction is a key factor in the design of networks with parameters such as cell radius, antenna heights, and the number of cell sites that can be set. As path loss is affected by the environment, satellite images of network locations are used in developing path loss prediction models such that environmental effects are captured. We developed a path loss model based on the Extreme Gradient Boosting (XGBoost) algorithm, whose inputs are numeric (non-image) features that influence path loss and features extracted from images composed of four tiled satellite images of points along the transmitter to receiver path. The model can predict path loss for multiple frequencies, antenna heights, and environments such that it can be incorporated into Radio Planning Tools. Various feature extraction methods that included CNN and hand-crafted and their combinations were applied to the images in order to determine the best input features, which, when combined with non-image features, will result in the best XGBoost model. Although hand-crafted features have the advantage of not requiring a large volume of data as no training is involved in them, they failed in this application as their use led to a reduction in accuracy. However, the best model was obtained when image features extracted using CNN and GLCM were combined with the non-image features, resulting in an RMSE improvement of 9.4272% against a model with non-image features only without satellite images. The XGBoost model performed better than Random Forest (RF), Extreme Learning Trees (ET), Gradient Boosting, and K Nearest Neighbor (KNN) based on the combination of CNN, GLCM, and non-image features. Further analysis using the Shapley Additive Explanations (SHAP) revealed that features extracted from the satellite images using CNN had the highest contribution toward the XGBoost model’s output. The variation in values of features with output path loss values was presented using SHAP summary plots. Interactions were also observed between some features based on their dependence plots from the computed SHAP values. This information, when further explored, could serve as the basis for the development of an explainable/glass box path loss model.

1. Introduction

Accurate path loss prediction is important in wireless networks. Power distribution within the coverage area can be accurately ascertained during Link Budget, leading to a network with a better quality of service if the predicted values are effectively used during design, expansion, or maintenance [1,2,3]. Deterministic path loss models provide a reasonable accuracy but are computationally expensive and require details that include a 3D model of the environment, building materials, and so on. Empirical models are much easier to use but are prone to errors. Meanwhile, machine learning (ML) models are more accurate than empirical models and sometimes exceed the accuracy of deterministic models. Such ML models were developed with algorithms such as Multiple Layer Perceptron (MLP), Support Vector Regression (SVR), Random Forest (RF), and XGBoost [4,5,6]. Furthermore, a Convolutional Neural Network (CNN) is employed to extract features from images that describe the effects of environment due to the dependence of path loss on the propagation path/environment. Such images include satellite images for a wide area [7], pointwise satellite images [8], segmented images with shadings that represent heights of buildings in an urban environment [9], segmented images based on some parameters, with building height inclusive [10], and pseudoimages representing the fusion of various data types [11].
Most of the models developed are for fixed frequency, antenna height, and environment. However, since path loss varies with the environment, such model prediction will be inaccurate in other scenarios. Considering the fact that path loss prediction is performed before networks are deployed, it is reasonable to develop models that can be used to make accurate predictions for varieties of environments and network parameters. Otherwise, the model will only be suitable for the existing network it was designed for or a network with almost the same parameters in a similar environment. Although ML models make better predictions, empirical models have the advantage of being versatile as they consider a range of frequencies, distances, and antenna heights over which these models are valid. Some of these models have parameters tailored to a specific type of environment, such as COST 231 Hata [12], and Stanford University Interim (SUI) model [13]. Therefore, there is a need to have machine learning path loss models valid for multiple parameters and environments so that they can provide accurate predictions during design and prior to network deployment. This will eliminate the cost of taking measurements under certain network parameters from individual environments and developing individual models specific to the environments and selected network parameters.
Ayadi et al. [14] developed an MLP model for the prediction of a number of frequencies in the Ultra-High-Frequency (UHF) band. The study considered the fact that the propagation path from the transmitter to receiver position varies such that some portions can be classified as either rural, urban, or suburban environments. Thus, distances covered in each type of environment were included as features for model development. The problem with such a model is that it does not capture the micro variations of the environment types along the propagation path; rather, it is developed based on distances covered in each environment type without a detailed description. Nguyen et al. [15] developed a model valid at frequencies between 0.8 and 70 GHz. The model was developed based on the features distance, frequency, propagation scenario, and environment type. The propagation scenario could be either “above rooftop” or “below rooftop”. Above rooftop is a scenario in which one of the antenna heights is above the surrounding rooftops, and below rooftop is when all antenna heights are below the surrounding rooftops. The model does not capture environment details, unlike that in [16], where features extracted from satellite images are included during model development to capture environmental details.
Sani et al. [17] also developed models for multiple frequencies, multiple antenna heights, and multiple environments. A clutter height feature was used in representing the various types of environments such that each type of environment had an integer value representing it. The problem with this method is that it does not take into consideration the micro differences within each type of environment, whereas [16] included satellite images to acknowledge the effects of these micro variations. The ResNet50V2 used in [16] takes time and memory space to train. This work proposes the use of a less complex CNN that will train faster and occupy less memory space.
Another issue to note is the dependence of a CNN’s accuracy on a large volume of data and a large amount of time it takes a CNN to get trained with the data, even though the prediction time is minimal. An alternative is to use hand-crafted feature extraction methods to deal with data limitation as well as the large training time. Sotiroudis et al. used the Segmentation-based Fractal Texture Analysis (SFTA) texture extraction method to extract features from images with grayscale shadings of building heights between the transmitter and receiver and developed three models with features based on RF, Extreme Gradient Boosting Machine (XGBoost), and Light Gradient Boosting Machine (LightGBM) [18]. However, a comparison between the performance of SFTA and CNN was not performed. To the best of our knowledge, their work is the first that incorporated texture feature extraction in a path loss problem. Predictions from a LightGBM model were more accurate, although the hyper-parameter optimization that was carried out in this work was not carried out in theirs to ascertain the optimal hyper-parameters of the algorithms. In other domains, comparisons were made between the performance of hand-crafted features and CNN, in which the CNN-extracted features performed better [19,20]. Combining CNN features and hand-crafted features improved the accuracy [21,22].
The aim of this work is to develop a single ML-based path loss model valid for some network parameters (UHF band and antenna heights) and environments (rural, suburban, urban, and urban highrise) with the effect of terrain/elevation taken into consideration. The essence is to have the ML path loss models valid for various values of network parameters and environment types, similar to existing empirical models. The ability of ML algorithms to model non-linear relationships improves performance over the existing empirical models. Numeric features combined with satellite image features were used as input to the model. Satellite images are used to capture micro and macro variations within the environments, unlike in existing empirical models, where such variations are not captured, resulting in poor predictions in new environments. The performance of features extracted using various methods (hand-crafted and CNN) and their combinations on the model’s output was evaluated. Hand-crafted features were investigated to ascertain the possibility of their use in this application, knowing that they need no training, unlike CNN, which requires training with a lot of data. However, since combining hand-crafted and CNN features was reported to improve accuracy in some applications, we also investigated it, and improvements in accuracy and training time were observed compared to [16]. In addition, the study explores feature importance and analysis. Feature importance was not conducted in all works cited in this paper, except in [17], where only non-image features were considered for model development. Our contributions are:
  • Review of path loss prediction based on image processing using CNN.
    Most works presented ML-based image processing approaches for path loss prediction, but to the best of our knowledge, no such comprehensive review of such existing works was supplied.
  • Investigation on the use of various hand-crafted feature extraction techniques and their combinations on 2D satellite images in path loss prediction.
  • Combining CNN features with hand-crafted features to improve the accuracy of path loss prediction.
    While this has been applied in computer vision tasks, to the best of our knowledge, this has not been applied to the problem of path loss prediction.
  • Interpretation of model prediction based on Shapley Additive Explanations (SHAP).
    Most ML-based path loss prediction works do not analyze the features deeply, including the comparison of different types of features.

3. Methodology

3.1. Dataset

The dataset in [16] is used in this work. It is composed of four datasets combined to form a single dataset, with measurements from different frequencies, environments, and antenna heights. Such a type of dataset combination was performed in [11,23]. Three of the datasets are available at [34,35,36] and have been used in [37,38,39], respectively. The fourth dataset is measurements taken by us. It should be noted that not all samples in [34] were used to maintain a balanced dataset in terms of environment type. The final dataset is made up of 12369 samples, consisting of 8 features and path loss values as the target. The features include elevation at receiver position (elevation), frequency, height of transmitting antenna (ht), height of receiving antenna (hr), latitude difference between transmitting and receiving antenna (distance_x), longitude difference between transmitting and receiving antenna (distance_y), elevation at transmitter position (tantennaelev), and distance between transmitter and receiver (distance). Table 1 presents a summary of the dataset, showing the different types of environments from which the measurements were taken and the parameters under which they were taken that include frequency, transmitting antenna height, and receiving antenna height.
Table 1. Dataset summary.

3.2. Image Data

The image used is a concatenation of four satellite images, as illustrated in Figure 1. The images represent various positions along the path from transmitter to receiver locationm as shown in Figure 1a with a sample in Figure 1b. The images were downloaded from Mapbox [40] based on the latitude and longitude values in the dataset.
Figure 1. (a) Format of image concatenation. (b) Final concatenated image sample.

3.3. Proposed Method

Figure 2 depicts the development process with three main phases; data preparation, model development and model testing.
Figure 2. Model development process with three main phases; data preparation, model development, and model testing.

3.3.1. Data Preparation

In the data preparation phase, the features are extracted from satellite images using an algorithm. The extracted features are combined with the numerical features to form a new dataset. The new dataset is then split into the training set and test set with a ratio of 80:20. As a preprocessing step, the two sets of data were scaled with a standard scaler to eliminate magnitude difference amongst features, thereby improving generalization and reducing training time.

3.3.2. Model Development

In the model development phase, hyper-parameter optimization (HPO) was performed on the training set based on XGBoost with hyperband algorithm [41] and then model training was performed with the optimized hyper-parameters.

3.3.3. Model Testing

In this phase, ten different samplings of training and test sets splits were used to train and test an XGBoost fitted with the optimized hyper-parameters. For each of the training and test sets, model performance based on the test set is noted, and the average of the ten samplings was computed.

3.4. Feature Extraction Methods

The following hand-crafted feature extraction methods were applied to the tiled images: Local Binary Pattern (LBP), Histogram of Oriented Gradients (HOG), Gray Level Co-occurrence Matrix (GLCM), and Discrete wavelet transform (DWT). Combinations of GLCM features with other hand-crafted features were considered, as well as a combination of the hand-crafted features with CNN features. The feature extraction methods used are described as follows:

3.4.1. Local Binary Pattern (LBP)

In LBP, a pixel’s intensity is compared with other pixel’s intensities within its local neighborhood, from which a histogram of binary patterns is obtained [42]. It is a less complex texture extraction method that is invariant to rotation. Differences in pixel intensities are presented using a binary encoding method described in Equation (1).
L B P ( P , R ) = p = 0 p 1 s ( g p g c ) 2 p
where g c is the center pixel, and g p is a neighborhood pixel, P stands for the number of neighbors, R is the neighborhood distance, and s ( x ) is computed in Equation (2) and stands for the binary encoding.
s ( x ) = 1 x 0 0 x < 0
LBP features are computed from histograms of the binary encodings based on Equation (3), where k = 0 , 1 , 2 , , K , K is the maximum value in the computed L B P ( P , R ) within the image with dimension ( M , N ) , and I ( x ) has a value of 1 when L B P ( P , R ) ) ( i , j ) = k , and 0 otherwise. There are various types of LBP [43,44]. A uniform LBP with a p value of 3 and an R-value of 3 was considered in this work.
H ( k ) = i = 1 M j = 1 N I ( L B P ( P , R ) ) ( i , j ) )

3.4.2. Gray Level Co-Occurrence Matrix (GLCM)

In GLCM, a matrix C ( i , j ) known as the co-occurrence matrix is computed by counting the co-occurrences in gray pixels that occur within a distance, d, and orientation, θ , which takes any value between 0 and 315 at 45 intervals. Features can be extracted from the co-occurrence matrix as a form of dimensionality reduction, with Energy, Inertia, Correlation, Difference Moment, Entropy [45], Contrast, and Homogeneity [46], as the most commonly computed features. Many other features can also be extracted from the co-occurrence matrix [47]. A d value of 3 and a θ value of 45 were used, with Contrast, Dissimilarity, Homogeneity, Angular Second Moment (ASM), Energy, and Correlation used as features.

3.4.3. Histogram of Oriented Gradients (HOG)

In HOG, each pixel’s magnitude and direction are computed from its x and y gradients. The image is then divided into cells, and histograms are computed for each cell based on gradient values. An aggregation of cells forms a block, and normalization is applied to each block. A combination of the normalized histograms from the blocks forms the HOG features [48]. In this work, the cells are 8 by 8 pixels, and tyebblocks are 2 by 2 cells. The number of orientation bins for the histogram is 9.

3.4.4. Discrete Wavelet Transform (DWT)

In DWT, data are decomposed into low pass L and high pass H components, yielding two sub-bands of L / H = l 1 , l 2 , , l ( N / 2 ) , h 1 , h 2 , , h 2 . For a bidimensional image, decomposition is performed in the vertical and horizontal directions, resulting in four sub-bands. The formed sub-bands are Approximation, Horizontal, Vertical, and Diagonal details and are due to the combination of L L , L H , H L , and H H filters, respectively. Statistical parameters can be computed from each sub-band to have a reduced feature set [43,47]. Methods such as the Line Measure Technique (LMT) [49] are also used to reduce the feature set. The mean, energy, standard deviation, and entropy of each sub-band were used as the extracted features in this work, summing to a total of 16 features.

3.4.5. Segmentation Fractal-Based Texture Analysis (SFTA)

SFTA is a two-staged feature extraction technique where, in the first stage, Two-Threshold Binary Decomposition (TTBD) is applied on a grayscale image to produce binary images. The Multilevel Otsu algorithm is used in TTBD to calculate some threshold values from the image’s gray level distribution. The algorithm first determines a threshold value that ensures that the image’s intra-class variance is minimized. This is applied to all regions of the image to obtain n t threshold values. Decomposition is carried out by the selection of pairs of threshold values and the application of a two-threshold segmentation, as shown in Equation (4) [50].
I b ( x , y ) = 1 t 1 < I ( x , y ) t u 0 o t h e r w i s e
I ( x , y ) and I b ( x , y ) are pixel intensities of grayscale and binary images, respectively, at ( x , y ) pixel position. t l and t u are the upper and lower threshold values. A total of ( n t 1 ) binary images are produced based on the procedure. In addition, simple thresholding is applied to produce n t binary images summing up to ( 2 n t 1 ) binary images. In the second step, three features are extracted from each binary image, namely, Area (A), Mean Gray Level (V), and Fractal Dimension (D). Area (A) represents the number of white pixels in the binary image, Mean Gray Level (V) is the average brightness of gray image pixel positions corresponding to white pixel positions in the binary image, and Fractal Dimension (D) is a measure of boundary complexity. A total of ( 6 n t 3 ) features are extracted for an input image in SFTA [18]. An n t value of 8 was used, thereby extracting 45 features from each input image.

3.4.6. Convolutional Neural Networks (CNN)

A CNN is used to process data that is in grid form, such as an image. It is made up of three main layers, the convolutional layer, the pooling layer, and the fully connected layer. Feature extraction is performed at the convolutional and pooling layers, whereas the fully connected layer performs mappings between the extracted features and target values. The convolutional layer is made up of low-dimensional filters that are multiplied by the image pixels by sliding it from left to right and top to bottom based on a stride value. A stride is the number of skips in pixel values for the filter to slide across the image. Each convolutional layer is followed by activation, and the Rectified Linear Unit (ReLU) is commonly used because it is fast and gives a better generalization. The image dimension is reduced when passed through a convolutional layer. The pooling layer is for further dimension reduction to decrease training time and increase immunity against noise, thereby preventing overfitting. Pooling could be carried out by summing the pixel values within a small grid (Sum Pooling), averaging (Average Pooling), or finding the maximum value (Max Pooling). Convolution layers and pooling layers are stacked for feature extraction. The deeper the network, the more sophisticated the process of extracting features. The first convolutional layer extracts features such as lines and edges, and deeper layers extract more detailed information. The fully connected layer is made up of neurons, as in the case of a Multiple Layer Perceptron, where two or more layers could be stacked, with the last layer forming the output layer. Regularization techniques, such as image augmentation, dropout, and batch normalization, are used to improve the accuracy of CNN [51,52,53]. The CNN used in this work is made up of five convolutional layers with batch normalization, and maximum pooling applied after each layer. The output of the last max pooling layer is passed to a fully connected layer with 177 neurons from which the features are extracted. The number of filters in the five consecutive convolutional layers are 24, 38, 83, 98, and 199, respectively, with square filters of sizes 2, 4, 4, 4, and 3, respectively. The maximum pooling layers used are of size 2. The complexity of the CNN is less compared to those of pretrained networks such as VGG-16, ResNet50V2, InceptionV3, EfficientNetB0, and so on. It, therefore, trains faster and occupies less memory space.

3.5. Extreme Gradient Boosting (XGBoost)

XGBoost is an advanced Gradient Boosting with improvements in speed, efficiency, and scalability. It has produced high accuracy in various applications such as prediction of Type 2 diabetes risk [54], hot spots at protein–protein interfaces [55], sales forecasting [56], and has been used in solving various problems in Kaggle competitions [57]. Like any boosting algorithm, it is made up of Classification and Regression Trees (CART) as base learners. Sometimes, linear functions are also used as base learners. The base learners are fitted to minimize a loss function. The loss function determines how well the model performs, and it is minimized during training [58,59]. Minimization of the loss function is achieved through gradient descent [60]. In XGBoost, a regularization term is added to the loss function, and Taylor Series Expansion is used in approximating the modified loss function. Optimization of the tree structure is performed using an approximate greedy algorithm. XGBoost uses column subsampling such that portions of features are collected to grow trees, thereby minimizing overfitting and improving training speed. In addition, it uses a presorting technique and storing mechanism whereby datasets are stored in blocks, and the search for the best split is performed by scanning columns in a block. It also uses techniques such as Block Compression and Block Sharding for out-of-core computations and when fetching data from multiple drives, respectively, thereby enhancing the execution time [58,61]. The boosting system can be composed of normal trees, as in the case of Gradient Boosting, or the use of Dropouts meet Multiple Additive Regression Trees (DART). DART is a process whereby multiple numbers of regression trees are fitted at a low learning rate, and trees fitted earlier are maintained, whereas those fitted later are dropped in order to prevent overfitting [62]. In this work, the number of trees, learning rate, and type of booster were ascertained based on hyperband for each of the combinations of numeric features with one or two feature extraction methods used on the input image, as shown in Table 2.
Table 2. Hyper-parameters of XGBoost for combinations of numeric features and features from one or two feature extraction method(s).

3.6. Performance Metrics

The performance metrics used in evaluating the performance of models include Mean Absolute Error (MAE), Root Mean Squared Error (RMSE), Mean Absolute Percentage Error (MAPE), and Coefficient of Determination ( R 2 ) [63,64].
M A E = 1 N i = 1 N | P m ( i ) P p ( i ) |
R M S E = i = 1 N ( P m ( i ) P p ( i ) ) 2 N
M A P E = 100 N i = 1 N | P m ( i ) P p ( i ) P m ( i ) |
R 2 = 1 i = 1 N ( P m ( i ) P p ( i ) ) 2 i = 1 N ( P m ( i ) P p ¯ ) 2
P m is the measured path loss, P p is the predicted path loss, P p ¯ is the mean of predicted path loss values, and N is the number of samples.

4. Result and Discussions

The performance metrics of each model resulting from a combination of non-image features merged with features from one or two feature extraction techniques applied to the images are shown in Table 3. Amongst the hand-crafted features, GLCM had the lowest error metrics and the highest coefficient of determination, but its accuracy is below the case in which no image features were used. The combination of GLCM features with other hand-crafted features further decreased the accuracy below that of GLCM features. The addition of hand-crafted features led to a reduction in accuracy when compared to using only non-image features, as shown by the performance metrics as well as the last column of Table 3, indicating a percentage improvement over using only non-image features. Negative values were observed as percentage improvements, indicating a reduction in performance. However, even with the reduction in performance in models with hand-crafted features, they are expected to perform better when tested with data from a site not used in model training. This is because the model trained with the hand-crafted features has some knowledge of the environments. The performance when CNN-extracted features were used was better than when features extracted based on any of the hand-crafted techniques were used. A combination of CNN features with any hand-crafted feature extraction methods decreased the accuracy compared to when CNN features were used alone, except in the case of LBP and GLCM, where an increase was observed. A combination of CNN-extracted features with GLCM features (CNN + GLCM) had the highest overall accuracy with an R 2 of 0.9372, seconded by a combination of CNN with LBP (CNN + LBP) with an R 2 of 0.9364 and were ahead of CNN features R 2 = 0.9357 . The RMSE value of 3.6682 dB when CNN + GLCM features were used is an indication that the model’s performance is acceptable as it is below the maximum allowed value. The acceptable RMSE for a path loss model is a value less than 7 dB in an urban environment and any value less than 15 dB in suburban and rural environments [65]. The RMSE value achieved is an improvement over the results in [11] with an RMSE of 8.27 dB, [17], 6.8944 dB, [12], 4.4286 dB, [29], 7.61 dB, and [66], 3.89 dB.
Table 3. XGBoost performance on combinations of numeric features and features from one or two feature extraction method(s).
The results indicated that the hand-crafted features and their combinations could not compete with CNN for feature extraction. As such, the requirement of training a CNN that needs a large volume of data and training time still holds as CNN-extracted features cannot be replaced by hand-crafted features due to the low accuracy of the hand-crafted features. However, the addition of LBP or GLCM features to the CNN features improved model accuracy over CNN features alone. However, since a combination of CNN and GLCM features had the highest improvement; they are recommended for use in this application. Meanwhile, the architecture of the CNN used resulted in an 18-fold speed over the ResNet50V2 used in [16] as it trained in 29.8242 min while the ResNet50V2 trained in 546.5685 min. In addition, a reduction in the RMSE value of 0.7604 dB was observed compared to the results in [16]. Compared to models based on diffraction that require the complete path profile from transmitter to receiver, such as the Walfisch Ikegami and Walfisch Bertoni, the proposed model is cost-effective and saves time as it does not require the additional measurement of parameters such as building heights and street width. In addition, ML models have accuracies better than such models, as seen in [6,67,68] because ML models do capture non-linear relationships.
Deterministic models, which also use the complete profile from transmitter to receiver, as well as the principles of wave transmission, such as reflection, refraction, diffraction, and shadowing, have high computational complexity and are very slow. With the proposed method, predictions are made instantly from the trained model, and the model can be used for similar environments. Meanwhile, in the case of deterministic models, each environment has to undergo the process of computation.
Figure 3 shows the percentage improvement calculated from RMSE values when either CNN, CNN + LBP, or CNN + GLCM features were used in combination with numeric features over when numeric features were used alone. The addition of CNN + GLCM features to the numeric features had the highest improvement of 9.4272%, followed by the addition of CNN + LBP, which resulted in 8.8667%, and lastly, the addition of CNN features improved it by 8.3802%, as shown in line Table 3.
Figure 3. Percentage improvement over non use of images for the top three feature extraction methods; CNN, CNN + LBP, and CNN + GLCM.

4.1. Comparison of XGBoost with Other Algorithms

The best features observed were a combination of the numeric features and features extracted using GLCM and CNN. Four other models were developed based on the features using different algorithms, and the performances of the models were compared with XGBoost. The algorithms considered are Random Forest (RF), Extreme Learning Trees (ET), Gradient Boosting (GB), and K Nearest Neighbor (KNN), with hyper-parameter optimization conducted in each case. The average RMSE of each algorithm is presented in Figure 4. It is observed that XGBoost (together with CNN and GLCM extracted features and numeric features) had the lowest RMSE, followed by GB, and ET had the highest value. XGBoost was better because of the regularization term added to its loss function and the column subsampling technique employed in it, which minimizes overfitting. ET had the least performance because it does not employ any optimization process for the determination of splits; rather it chooses the splits randomly.
Figure 4. RMSE values of the ML models for a combination of CNN + GLCM features and non-image features.

4.2. Feature Importance

Feature importance of the best model (with CNN + GLCM features), as well as the model that has only non-image features, were determined based on the Shapley Additive Explanations (SHAP). SHAP is a method for interpreting ML models, based on a combination of game theory, Shapley values, and some conditional expectations to compute the importance of each feature [69]. Figure 5, Figure 6, Figure 7 and Figure 8 present explanations of the model with non-image features only. Figure 5 shows the mean importance of each feature, with “tantennaelev” having the highest importance. Next to it is “distance”, as expected, because it is well established that path loss varies with distance. However, “tantennaelev” had the highest impact on the model output because studies have shown that apart from distance, the elevation of a position affects path loss at that position due to variations in height difference between the transmitting and receiving antennas. Thus, for a multiple antenna path loss model, it is wise to include the elevation of the transmitting and receiving antenna positions so that their height difference effect is captured. The “tantennaelev” had an importance that superseded that of “distance” because it was observed by Jimoh et al. [70] that elevation affects path loss to the extent that a position at a further distance from the transmitting antenna but at a high elevation could have lower path loss values than a position closer to the transmitting antenna but at a low elevation. The third and fourth most important features are distances in latitude and longitude, and then the height of the receiving antenna (hr) had the least importance.
Figure 5. Mean feature importances of the model with non-image features only, with “tantennaelev” having the highest importance and “hr” the least.
Figure 6. SHAP summary plot of the model with non-image features only, showing the variations in feature values with output path loss value.
Figure 7. Dependence plot of the feature “tantennaelev” in a model with non-image features only with the feature “distance_x”. Feature “tantennaelev” has the highest interaction index with “distance_x”.
Figure 8. Dependence plot of the feature “distance” in a model with non-image features only with feature “tantennaelev”. Feature “distance” has the highest interaction index with “tantennaelev”.
Figure 6 shows the detailed explanations of the features, with the position of the dots with respect to the horizontal axis showing SHAP values that determine whether the path loss is high or low, and the color indicating the feature value. Positive SHAP values mean that the feature has a positive impact on path loss and negative means it has a negative impact. The results show that for higher values of “tantennaelev”, the path loss value is low, but for lower feature values, the output is high, though with a few instances where the output is low at lower values of “tantennaelev”. This is because the higher the elevation of the transmitting antenna, the wider the signal coverage; therefore, path loss decreases for the distance concerned. Likewise, the lower the elevation, the smaller the coverage, and the higher the path loss values. The fewer points where the low output values were observed might be due to irregularities in elevations of the receiver position as well as clutter. In the case of “distance”, it is observed that at lower distances, the path loss values are low and high at higher distances. Some minor instances in which the linearity is violated are due to the effect of other features. However, not only does the difference between transmitting and receiving antenna altitude (height + elevation) affect the signal coverage. Other antenna parameters such as azimuth and tilt also affect the signal coverage, but the study of their effects is beyond the scope of this work.
The difference in latitude between the transmitting and receiving antenna and that of longitude (distance_x and distance_y) were observed to push path loss values high and moderate, respectively. Higher “frequency” values pushed the path loss values higher as expected, the moderately low values pushed it low, and afew instances in the lowest frequency values pushed the path loss high due to the effect of other factors, such as elevation and/or clutter. The features “elevation” and height of the receiving antenna (ht) were observed to push the path loss values low when they are high or extremely low, whereas the upper part of their low values increased path loss values. This could be due to situations when the transmitting antenna is at a higher elevation.
Dependence plots of the two most important features (tantennaelev and distance) were plotted. Figure 7 shows that “tantennaelev” has some interactions with “distance_x”. Therefore, the effect of “tantennaelev” on path loss is dependent on “distance_x” value. The realization that “distance_x” has interaction with “tantennaelev” is based on the computation of the approximate interaction index between the features, from which the value between “tantennaelev” and “distance_x” was the highest. Confirmation of the presence of interaction is based on the presence of some vertical color patterns. Conclusions about the interactions of these variables cannot be made based on the plot because the data used are not a representation of all latitude and longitude positions on earth. Figure 8 is the dependence plot of “distance”, showing how path loss increases with increasing distance. The presence of vertical color patterns within the plot around the start values of distance indicates an interaction between the features “distance” and “tantennaelev”. As mentioned earlier, path loss at a distance is dependent on elevation. Therefore, path loss at the same distance but different elevation values are unequal.
Figure 9 shows the 20 topmost features in terms of mean importance in the model that had non-image inputs as well as combinations of features extracted using CNN and GLCM from the images. Features named “f_x” are the features extracted by the CNN, where x takes values from 0 to 176. It is observed that the 144th feature ”f_143“ had the highest importance, followed by the third, “f_2”. Only two of the non-image features appeared in the top 20, with the height of the transmitting antenna (ht) appearing as the 13th and elevation as the 20th. The most important features are the image features extracted using CNN, and none of the GLCM features appeared in the top 20. Unlike in the case of the non-image input model, where “ht” and “elevation” were the last two, in this case, they were the first two because, in the non-image input model, the model was developed without consideration of environmental factors, and its explanation was similar to that of existing empirical models. The inclusion of features of the satellite images enhances the model’s ability to differentiate between environments. The model is well fitted to differentiate between the environments based on the importance it gives to the extracted features.
Figure 9. Mean feature importances of the top 20 features in the model with a combination of non-image features and CNN + GLCM features. “f_143”, which is a CNN-extracted feature, has the highest importance, while “elevation”, which stands for elevation at receiving antenna position, has the least importance among them.
Figure 10 is a summary plot of the 20 most important features. It was observed that higher values of the CNN features produce higher path loss values, and lower values produce lower path loss values, with the exceptions of “f_0” and “f_1”. However, the features were extracted by a CNN, and the reason for this behavior cannot be explained as the features are not physical quantities and were not generated by a known equation but a black box model, unlike other quantities such as “distance” and “elevation” that are physical quantities. However, it is clearly seen that predicted path loss is dependent on these extracted features. Similarly, dependence plots for features “f_143” and “f_2” are presented in Figure 11 and Figure 12, respectively. Figure 11 showed that “f_143” had the highest approximate interaction index with “f_167”, and it was observed that as “f_143” increased, “f_167” also increased, as well as path loss but in a stepwise fashion without interactions even though their partial interaction index is the highest. The pair of feature “f_2” as shown in Figure 12 is “correlation”, which is a GLCM feature. Interactions were found between the variables due to the presence of vertical color patterns, and different portions of the plots show different combination rules that determine path loss values. The vertical color patterns indicate that for the same “f_2” value, the path loss value differs depending on the value of “correlation”. More of such relationships between features could be observed if the dependence plot of the remaining features is plotted. The concept of feature interaction can be explained by Equation (9), where y depends on two variables x 1 and x 2
y = β 0 + β 1 x 1 + β 2 x 2 + β 3 x 1 x 2 + e r r o r
Figure 10. SHAP summary plot of the 20 most important features in the model developed with a combination of non-image features and CNN + GLCM features. It shows the variations in feature values with output path loss value.
Figure 11. Dependence plot of the feature “f_143” in the model developed with a combination of non-image features and CNN + GLCM features. Feature ”f_143“ has the highest interaction index with ”f_167“.
Figure 12. Dependence plot of the feature “f_2” in the model developed with a combination of non-image features and CNN + GLCM features. Feature ”f_2“ has the highest interaction index with the GLCM feature “correlation”.
β 0 is an average response, β 1 and β 2 define contributions of features x 1 and x 2 , respectively, and β 3 defines the combined contributions of x 1 and x 2 beyond their individual explanations alone. Random variations in the values unexplained by the remaining part of the equation are represented by the error term. Thus, the interaction between two features such as “f_2” and “correlation” can be combined and included in an explainable model similar to the β 3 x 1 x 2 term in Equation (9) by replacing x 1 and x 2 with ”f_2“ and ”correlation“, and the dependent variable y with path loss. Further analysis of interactions between the features could result in a glass box path loss model in a form similar to Equation (9) that gives an explanation of how the path loss values are calculated, unlike ML models that are black box models [71,72].

5. Conclusions

A path loss model with a good prediction accuracy is required for the determination of wireless network parameters such as transmitting power, cell radius, antenna height, number of cell sites, and their optimal position. The attainment of the required data rate is dependent on the proper setting of the parameters such that there is good signal coverage and no interference. The dependency of path loss on the environment led to the inclusion of satellite images of network location in developing machine learning path loss models such that the influence of the environment is learned. A path loss predictive model that handles multiple frequencies, multiple antenna heights, and multiple environments were developed in this work. The model takes in a combination of numeric features and features extracted from satellite images to make the path loss prediction. XGBoost was the algorithm used in model development and was selected based on its good performance in several applications. Several hand-crafted feature extraction methods, their combinations, or combinations of hand-crafted features and CNN features, were tested. The hand-crafted feature extraction methods included LBP, HOG, GLCM, DWT, and SFTA. None of the hand-crafted features improved accuracy compared with using numeric features alone; instead, the accuracy was reduced. Combining GLCM with any of the hand-crafted feature extraction methods decreased performance below that of GLCM features. When CNN features were combined with any of the hand-crafted feature extraction methods, combinations with LBP or GLCM resulted in accuracies better than that of CNN features. A combination of CNN and GLCM features from the images had the best performance overall. This resulted in a 9.4272% improvement in RMSE for combined environments over when no satellite image features were used in model development. A comparison was performed to evaluate the performance of other ML algorithms on the combination of numeric, GLCM, and CNN features. Random Forest, Extreme Learning Trees, Gradient Boosting, and K Nearest Neighbor were used for this purpose. XGBoost had the best performance metrics and also has an advantage of a fast training time over them. Features extracted using CNN were observed to have the highest impact on the model’s output based on SHAP. Although the model’s accuracy is good, there is a need for more training data from different environments and network parameters such that a single model can be developed and used for pre-network deployment path loss prediction. Meanwhile, the study demonstrated the possibility of developing such a model because an RMSE value of 3.6682 dB was obtained when combined data were used. This RMSE value is acceptable for a path loss model as it is below 7 dB. In our future work, the model’s performance for the different sub-datasets will be explored, and methods for improving the accuracy of sub-datasets with low accuracy will be addressed.

Author Contributions

Conceptualization, U.S.S., D.T.C.L. and O.A.M.; methodology, U.S.S., D.T.C.L. and O.A.M.; formal analysis, U.S.S.; data curation, U.S.S.; writing—original draft preparation, U.S.S.; writing—review and editing, U.S.S., D.T.C.L. and O.A.M.; supervision, D.T.C.L. and O.A.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by Universiti Brunei Darussalam, grant number RSCH/1.18/FICBF(a)/2022/004.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data used in this research are available on the github repository, “A Multiple Environment and Multiple Network Parameter Path Loss Dataset”.

Acknowledgments

We express our sincere gratitude to Universiti Brunei Darussalam for the provision of resources used in this research.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lukman, S.; Nazaruddin, Y.Y.; Ai, B.; He, R.; Joelianto, E. Estimation of received signal power for 5G-railway communication systems. In Proceedings of the 6th International Conference on Electric Vehicular Technology (ICEVT 2019), Bali, Indonesia, 18–21 November 2019; pp. 35–39. [Google Scholar]
  2. Jo, H.S.; Park, C.; Lee, E.; Choi, H.K.; Park, J. Path Loss Prediction Based on Machine Learning Techniques: Principal Component Analysis, Artificial Neural Network and Gaussian Process. Sensors 2020, 20, 1927. [Google Scholar] [CrossRef] [Green Version]
  3. Zhang, Y.; Wen, J.; Yang, G.; He, Z.; Wang, J. Path Loss Prediction Based on Machine Learning: Principle, Method, and Data Expansion. Appl. Sci. 2019, 9, 1908. [Google Scholar] [CrossRef] [Green Version]
  4. Sotiroudis, S.P.; Goudos, S.K.; Siakavara, K.F.I.A. Tool to Explain Radio Propagation and Reduce Model Complexity. Telecom 2020, 1, 114–125. [Google Scholar] [CrossRef]
  5. Moraitis, N.; Tsipi, L.; Vouyioukas, D.; Gkioni, A.; Louvros, S. Performance Evaluation of Machine Learning Methods for Path Loss Prediction in Rural Environment at 3.7GHz. Wirel. Netw. 2021, 27, 4169–4188. [Google Scholar] [CrossRef]
  6. Isabona, J.; Imoize, A.; Ojo, S.; Karunwi, O.; Kim, Y.; Lee, C.; Li, C. Development of a Multilayer Perceptron Neural Network for Optimal Predictive Modeling in Urban Microcellular Radio Environments. Appl. Sci. 2022, 12, 5713. [Google Scholar] [CrossRef]
  7. Ahmadien, O.; Ates, H.F.; Baykas, T.; Gunturk, B.K. Predicting Path Loss Distribution of an Area from Satellite Images Using Deep Learning. IEEE Access 2020, 8, 64982–64991. [Google Scholar] [CrossRef]
  8. Thrane, J.; Zibar, D.; Christiansen, H.L. Model-Aided Deep Learning Method for Path Loss Prediction in Mobile Communication Systems at 2.6 GHz. IEEE Access 2020, 8, 7925–7936. [Google Scholar] [CrossRef]
  9. Sotiroudis, S.P.; Goudos, S.K.; Siakavara, K. Deep Learning for Radio Propagation: Using Image-Driven Regression to Estimate Path Loss in Urban Areas. ICT Express 2020, 6, 160–165. [Google Scholar] [CrossRef]
  10. Lee, J.G.Y.; Kang, M.Y.; Kim, S.C. Path loss exponent prediction for outdoor millimeter wave channels through deep learning. In Proceedings of the IEEE Conference on Wireless Communications and Networking, Marrakech, Morocco, 15–18 April 2019. [Google Scholar]
  11. Sotiroudis, S.P.; Sarigiannidis, P.; Goudos, S.K.; Siakavara, K. Fusing Diverse Input Modalities for Path Loss Prediction: A Deep Learning Approach. IEEE Access 2021, 9, 30441–30451. [Google Scholar] [CrossRef]
  12. Omoze, E.L.; Edeko, F.O. Statistical Tuning of COST 231 Hata Model in Deployed 1800 MHz GSM Networks for a Rural Environment. Niger. J. Technol. 2021, 39, 1216–1222. [Google Scholar] [CrossRef]
  13. Cahyadi, M.B.; Sudiarta, P.K.; Hartawan, D.D.; Analisis, I.G.A.K. Perbandingan Nilai Shadow Fading Pada Model Propagasi Stanford University Interim ( Sui ) Dengan Metode Simulasi Dan Drive Test. J. Spektrum 2021, 8, 230–242. [Google Scholar]
  14. Ayadi, M.; Zineb, B.; Tabbane, A.; Uhf, S.A. Path Loss Model Using Learning Machine for Heterogeneous Networks. IEEE Trans. Antennas Propag. 2017, 65, 3675–3683. [Google Scholar] [CrossRef]
  15. Nguyen, C.; Cheema, A.A. A Deep Neural Network-Based Multi-Frequency Path Loss Prediction Model from 0.8 GHz to 70 GHz. Sensors 2021, 21, 5100. [Google Scholar] [CrossRef] [PubMed]
  16. Sani, U.S.; Lai, D.T.C.; Malik, O.A. A hybrid combination of a convolutional neural network with a regression model for path loss prediction using tiles of 2D satellite images. In Proceedings of the 8th International Conference on Intelligent and Advanced Systems (ICIAS), Kuching, Malaysia, 13–15 July 2021. [Google Scholar]
  17. Sani, U.S.; Lai, D.T.C.; Malik, O.A. Investigating Automated Hyper-Parameter Optimization for a Generalized Path Loss Model. In Proceedings of the 11th International Conference on Electronics Communications and Networks (CECNet), Beijing, China, 18–21 November 2021; pp. 283–291. [Google Scholar]
  18. Sotiroudis, S.P.; Siakavara, K.; Koudouridis, G.P.; Sarigiannidis, P.; Goudos, S.K. Enhancing Machine Learning Models for Path Loss Prediction Using Image Texture Techniques. IEEE Antennas Wirel. Propag. Lett. 2021, 20, 1443–1447. [Google Scholar] [CrossRef]
  19. Bodapati, J.D.; Suvarna, B.; Veeranjaneyulu, N. Role of Deep Neural Features vs Hand Crafted Features for Hand Written Digit Recognition. Int. J. Recent Technol. Eng 2019, 7, 147–152. [Google Scholar]
  20. Sumi, T.A.; Hossain, M.S.; Islam, R.U.; Andersson, K. Human Gender Detection from Facial Images Using Convolution Neural Network. In Proceedings of the Communications in Computer and Information Science, Nottingham, UK, 30–31 July 2021; Volume 1435, pp. 188–203. [Google Scholar]
  21. Lin, W.; Hasenstab, K.; Cunha, G.M.; Schwartzman, A. Comparison of Handcrafted Features and Convolutional Neural Networks for Liver MR Image Adequacy Assessment. Sci. Rep. 2020, 10, 20336. [Google Scholar] [CrossRef] [PubMed]
  22. Nanni, L.; Ghidoni, S.; Brahnam, S. Handcrafted vs. Non-Handcrafted Features for Computer Vision Classification. Pattern Recognit. 2017, 71, 158–172. [Google Scholar] [CrossRef]
  23. Thrane, J.; Sliwa, B.; Wietfeld, C.; Christiansen, H.L. Deep learning-based signal strength prediction using geographical images and expert Knowledge. In Proceedings of the 2020 IEEE Global Communications Conference, Taipei, Taiwan, 7–11 December 2020. [Google Scholar]
  24. Sliwa, B.; Geis, M.; Bektas, C.; Lop, M.; Mogensen, P.; Wietfeld, C. DRaGon: Mining latent radio channel information from geographical data leveraging deep learning. In Proceedings of the 2022 IEEE Wireless Communications and Networking Conference (WCNC), Austin, TX, USA, 10–13 April 2022. [Google Scholar]
  25. Nguyen, T.T.; Caromi, R.; Kallas, K.; Souryal, M.R. Deep learning for path loss prediction in the 3.5 GHz CBRS spectrum band. In In Proceedings of the 2022 IEEE Wireless Communications and Networking Conference (WCNC), Austin, TX, USA, 10–13 April 2022. [Google Scholar]
  26. Ates, H.F.; Hashir, S.M.; Baykas, T.; Gunturk, B.K. Path Loss Exponent and Shadowing Factor Prediction From Satellite Images Using Deep Learning. IEEE Access 2019, 7, 101366–101375. [Google Scholar] [CrossRef]
  27. Cheng, H.; Lee, H.; Cho, M. Millimeter Wave Path Loss Modeling for 5G Communications Using Deep Learning With Dilated Convolution and Attention. IEEE Access 2021, 9, 62867–62879. [Google Scholar] [CrossRef]
  28. Kim, H.; Jin, W.; Lee, H. MmWave Path Loss Modeling for Urban Scenarios Based on 3D-Convolutional Neural Networks. In Proceedings of the 2022 International Conference on Information Networking (ICOIN), Jeju, Korea, 12–15 January 2022. [Google Scholar]
  29. Wu, L.; He, D.; Ai, B.; Wang, J.; Liu, D.; Zhu, F. Enhanced Path Loss Model by Image-Based Environmental Characterization. IEEE Antennas Wirel. Propag. Lett. 2022, 21, 903–907. [Google Scholar] [CrossRef]
  30. Ratnam, V.V.; Chen, H.; Pawar, S.; Zhang, B.; Zhang, C.J.; Kim, Y.J.; Lee, S.; Cho, M.; Yoon, S.R. FadeNet: Deep Learning-Based Mm-Wave Large-Scale Channel Fading Prediction and Its Applications. IEEE Access 2021, 9, 3278–3290. [Google Scholar] [CrossRef]
  31. Zhang, X.; Shu, X.; Zhang, B.; Ren, J.; Zhou, L.; Chen, X. Cellular network radio propagation modeling with deep convolutional neural networks. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Diego, CA, USA, 6–10 July 2020; pp. 2378–2386. [Google Scholar]
  32. Hayashi, T.; Nagao, T.; Ito, S.A. A Study on the variety and size of input data for radio propagation prediction using a deep neural network. In Proceedings of the 2020 14th European Conference on Antennas and Propagation (EuCAP), Copenhagen, Denmark, 15–20 March 2020. [Google Scholar]
  33. Inoue, K.; Ichige, K.; Nagao, T.; Hayashi, T. Radio Propagation Prediction Using Deep Neural Network and Building Occupancy Estimation. IEICE Commun. Express 2020, 9, 506–511. [Google Scholar] [CrossRef]
  34. SVR PATHLOSS. Available online: https://github.com/timotrob/SVR_PATHLOSS (accessed on 2 January 2022).
  35. LoRaWAN Measurement Campaigns in Lebanon. Available online: https://zenodo.org/record/1560654#.X-9B-VUzbIV (accessed on 2 January 2022).
  36. Path Loss Prediction. Available online: https://github.com/lamvng/Path-loss-prediction (accessed on 2 January 2022).
  37. Timoteo, R.D.A.; Cunha, D.C.; Cavalcanti, G.D.C. A Proposal for Path Loss Prediction in Urban Environments Using Support Vector Regression. In Proceedings of the The Tenth Advanced International Conference on Telecommunications, Paris, France, 20–24 July 2014; pp. 119–124. [Google Scholar]
  38. El Chall, R.; Lahoud, S.; El Helou, M. LoRaWAN Network Radio Propagation Models and Performance Evaluation in Various Environments in Lebanon. IEEE Internet Things J. 2019, 6, 2366–2378. [Google Scholar] [CrossRef]
  39. Popoola, S.I.; Atayero, A.A.; Arausi, O.D.; Matthews, V.O. Path Loss Dataset for Modeling Radio Wave Propagation in Smart Campus Environment. Data Br. 2018, 17, 1062–1073. [Google Scholar] [CrossRef] [PubMed]
  40. Mapbox. Static Maps. 2022. Available online: https://www.mapbox.com/static-maps (accessed on 18 January 2022).
  41. Li, L.; Jamieson, K.; DeSalvo, G.; Rostamizadeh, A.; Talwalkar, A. Hyperband: A Novel Bandit-Based Approach To Hyperparameter Optimization. J. Mach Learn. Res. 2018, 18, 6765–6816. [Google Scholar]
  42. Horn, Z.C.; Auret, L.; McCoy, J.T.; Aldrich, C.; Herbst, B.M. Performance of Convolutional Neural Networks for Feature Extraction in Froth Flotation Sensing. IFAC-PapersOnLine 2017, 50, 13–18. [Google Scholar] [CrossRef]
  43. Wang, S.; Han, K.; Jin, J. Review of Image Low-Level Feature Extraction Methods for Content-Based Image Retrieval. Sens. Rev. 2019, 39, 783–809. [Google Scholar] [CrossRef]
  44. Paramkusham, S.; Rao, K.M.M.; Rao, B.P. Comparison of Rotation Invariant Local Frequency, LBP and SFTA Methods for Breast Abnormality Classification. Int. J. Signal Imaging Syst. Eng. 2018, 11, 136–150. [Google Scholar] [CrossRef]
  45. Choras, R.S. Image Feature Extraction Techniques and Their Applications for CBIR and Biometrics Systems. Int. J. Biol. Biomed. Eng. 2007, 1, 6–15. [Google Scholar]
  46. Sukiman, T.S.A.; Suwilo, S.; Zarlis, M. Feature Extraction Method GLCM and LVQ in Digital Image-Based Face Recognition. SinkrOn 2019, 4, 1–4. [Google Scholar] [CrossRef]
  47. Casagrande, L.; Macarini, L.A.B.; Bitencourt, D.; Fröhlich, A.A.; de Araujo, G.M.A. New Feature Extraction Process Based on SFTA and DWT to Enhance Classification of Ceramic Tiles Quality. Mach. Vis. Appl. 2020, 31, 71. [Google Scholar] [CrossRef]
  48. Althnian, A.; Aloboud, N.; Alkharashi, N.; Alduwaish, F.; Alrshoud, M.; Kurdi, H. Face Gender Recognition in the Wild: An Extensive Performance Comparison of Deep-Learned, Hand-Crafted, and Fused Features with Deep and Traditional Models. Appl. Sci. 2021, 11, 89. [Google Scholar] [CrossRef]
  49. Ghazali, K.H.; Mansor, M.F.; Mustafa, M.M.; Hussain, A. Feature extraction technique using discrete wavelet transform for image classification. In Proceedings of the 2007 5th Student Conference on Research and Development, Selangor, Malaysia, 11–12 December 2007. [Google Scholar]
  50. Costa, A.F.; Humpire-Mamani, G.; Traina, A.J.M.H. An Efficient Algorithm for Fractal Analysis of Textures. In Proceedings of the Brazilian Symposium on Computer Graphics and Image Processing, Ouro Preto, Brazil, 22–25 August 2012; pp. 39–46. [Google Scholar]
  51. Yamashita, R.; Nishio, M.; Do, R.K.G.; Togashi, K. Convolutional Neural Networks: An Overview and Application in Radiology. Insights Imaging 2018, 9, 611–629. [Google Scholar] [CrossRef] [Green Version]
  52. Ranjbar, S.; Nejad, F.M.; Zakeri, H.; Gandomi, A.H. Computational intelligence for modeling of asphalt pavement surface distress. In New Materials in Civil Engineering; Elsevier: Oxford, UK, 2020. [Google Scholar]
  53. Weimer, D.; Scholz-reiter, B.; Shpitalni, M. Design of Deep Convolutional Neural Network Architectures for Automated Feature Extraction in Industrial Inspection. CIRP Ann. Manuf. Technol. 2016, 65, 417–420. [Google Scholar] [CrossRef]
  54. Wang, L.; Wang, X.; Chen, A.; Jin, X.; Che, H. Prediction of Type 2 Diabetes Risk and Its Effect Evaluation Based on the XGBoost Model. Healthcare 2020, 8, 247. [Google Scholar] [CrossRef] [PubMed]
  55. Wang, H.; Liu, C.; Deng, L. Enhanced Prediction of Hot Spots at Protein-Protein Interfaces Using Extreme Gradient Boosting. Sci. Rep. 2018, 8, 14285. [Google Scholar] [CrossRef]
  56. Niu, Y. Walmart Sales Forecasting Using XGBOOST Algorithm and Feature Engineering. In Proceedings of the 2020 International Conference on Big Data and Artificial Intelligence and Software Engineering (ICBASE), Bangkok, Thailand, 30 October–1 November 2020. [Google Scholar]
  57. Adebayo, S. How The Kaggle Winners Algorithm XGBoost Algorithm Works. Available online: https://dataaspirant.com/xgboost-algorithm/ (accessed on 16 February 2022).
  58. Xia, Y.; Liu, C.; Li, Y.Y.; Liu, N.A. Boosted Decision Tree Approach Using Bayesian Hyper-Parameter Optimization for Credit Scoring. Expert Syst. Appl. 2017, 78, 225–241. [Google Scholar] [CrossRef]
  59. Xgboost Developers. Release 1.5.0-Dev Xgboost Developers; Technical Report; Xgboost Developers: Seattle, WA, USA, 2021. [Google Scholar]
  60. Nagao, T.; Hayashi, T. Study on Radio Propagation Prediction by Machine Learning Using Urban Structure Maps. In Proceedings of the 14th European Conference on Antennas and Propagation (EuCAP2020), Copenhagen, Denmark, 15–20 March 2020. [Google Scholar]
  61. Chen, T.; Guestrin, C.X.A. Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016. [Google Scholar]
  62. Rashmi, K.V.; Gilad-Bachrach, R.D. Dropouts Meet Multiple Additive Regression Trees. J. Mach. Learn. Res. 2015, 38, 489–497. [Google Scholar]
  63. Ojo, S.; Imoize, A.; Alienyi, D. Radial Basis Function Neural Network Path Loss Prediction Model for LTE Networks in Multitransmitter Signal Propagation Environments. Int. J. Commun. Syst. 2021, 34, e4680. [Google Scholar] [CrossRef]
  64. Ebhota, V.C.; Isabona, J.; Srivastava, V.M. Investigating Signal Power Loss Prediction in A Metropolitan Island Using ADALINE and Multi-Layer Perceptron Back Propagation Networks. Int. J. Appl. Eng. Res. 2018, 13, 13409–13420. [Google Scholar]
  65. Moraitis, N.; Vouyioukas, D.; Gkioni, A.; Louvros, S. Measurements and Path Loss Models for a TD-LTE Network at 3.7 GHz in Rural Areas. Wirel. Netw. 2020, 26, 2891–2904. [Google Scholar] [CrossRef]
  66. Aldosary, A.M.; Aldossari, S.A.; Chen, K.C.; Mohamed, E.M.; Al-Saman, A. Predictive Wireless Channel Modeling of Mmwave Bands Using Machine Learning. Electronics 2021, 10, 3114. [Google Scholar] [CrossRef]
  67. Moraitis, N.; Tsipi, L.; Vouyioukas, D. Machine learning-based methods for path loss prediction in urban environment for LTE networks. In Proceedings of the International Conference on Wireless and Mobile Computing, Networking and Communications, Thessaloniki, Greece, 12–14 October 2020. [Google Scholar]
  68. Ojo, S.; Sari, A.; Ojo, T.P. Path Loss Modeling: A Machine Learning Based Approach Using Support Vector Regression and Radial Basis Function Models. Open J. Appl. Sci. 2022, 12, 990–1010. [Google Scholar] [CrossRef]
  69. Garcia, M.V. Interpretable Forecast of NO2 Concentration Through Deep SHAP. Master’s Thesis, Universidad NacionalL De Educacion a Distacia, Madrid, Spain, 2019. [Google Scholar]
  70. Jimoh, A.A.; Bakinde, N.T.; Faruk, N.; Bello, O.W.; Ayeni, A.A. Clutter Height Variation Effects on Frequency Dependent Path Loss Models at UHF Bands in Build-Up Areas. Sci. Technol. Arts Res. J. 2015, 4, 138–147. [Google Scholar] [CrossRef] [Green Version]
  71. Kuhn, M.; Johnson, K. Feature Engineering and Selection: A Practical Approach for Predictive Models; CRC Press: New York, NY, USA, 2020. [Google Scholar]
  72. Molnar, C. Interpretable Machine Learning: A Guide for Making Black Box Models Explainable; Lulu Press Inc.: Morrisville, NC, USA, 2022. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.