You are currently viewing a new version of our website. To view the old version click .
Signals
  • Article
  • Open Access

25 September 2023

Improved RSSI Indoor Localization in IoT Systems with Machine Learning Algorithms

,
and
1
Faculty of Graduate Studies, IIC University of Technology, Phnom Penh 121206, Cambodia
2
School of Electrical Engineering, Korea University, Seoul 02841, Republic of Korea
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue The 2022 7th International Conference on Intelligent Information Processing

Abstract

Recent developments in machine learning algorithms are playing a significant role in wireless communication and Internet of Things (IoT) systems. Location-based Internet of Things services (LBIoTS) are considered one of the primary applications among those IoT applications. The key information involved in LBIoTS is finding an object’s geographical location. The Global Positioning System (GPS) technique does not perform better in indoor environments due to multipath. Numerous methods have been investigated for indoor localization scenarios. However, the precise location estimation of a moving object in such an application is challenging due to the high signal fluctuations. Therefore, this paper presents machine learning algorithms to estimate the object’s location based on the Received Signal Strength Indicator (RSSI) values collected through Bluetooth low-energy (BLE)-based nodes. In this experiment, we utilize a publicly available RSSI dataset. The RSSI data are collected from different BLE ibeacon nodes installed in a complex indoor environment with labels. Then, the RSSI data are linearized using the weighted least-squares method and filtered using moving average filters. Moreover, machine learning algorithms are used for training and testing the dataset to estimate the precise location of the objects. All the proposed algorithms were tested and evaluated under their different hyperparameters. The tested models provided approximately 85% accuracy for KNN, 84% for SVM and 76% accuracy in FFNN.

1. Introduction

With the rapid development of IoT- and AI-integrated technologies, researchers and developers have been driven to create intelligent location-based IoT applications for smart cities, transportation, intelligent buildings, smart agriculture, military applications, etc. Location-based services (LBS) are among these applications, and they are regarded as crucial ones since LBS makes it possible to determine the precise location of moving objects in enclosed spaces. Many people spend a day in indoor environments because of work, school, shopping, etc. Therefore, LBIoTSs play a significant role in this era by locating objects in indoor environments and providing crucial information for many intelligent applications, for example, locating an older person who is living alone in a house, locating a human in a shopping mall, automatically taking students’ attendance on a smart campus, locating a person using a wearable sensor in a hospital, locating people inside a building during a fire, etc. [1,2].
Regarding localization, there are two environments: outdoor and indoor. GPS technology is well performed in outdoor environments; however, it is inappropriate for indoor localization due to several reasons [3,4]. The GPS will not work accurately in the indoor environment due to the multipath fading effect and less accuracy in estimation [5]. Due to the above reasons, researchers have developed indoor localization algorithms based on deterministic algorithms, probabilistic algorithms, and filters. Many widely used deterministic and probabilistic algorithms can be found in related works, such as trilateration, multilateration, the Levenberg–Marquardt algorithm, and Kalman filters [6]. The drawbacks of the mentioned algorithms are lower accuracy, inefficiency, complex hardware systems, and most of the time, they are impractical to deploy in robust indoor environments.
Similar works exist on using ML techniques for indoor localization applications. The existing works can be categorized as supervised regressor-based, classifier-based, deep learning-based, and unsupervised learning-based approaches. The works [7] used a supervised regressor to find the location. In contrast, different types of regressors, such as Decision Tree Regression, Random Forest Regression, Linear Regression, Polynomial Regression, Support Vector Regression, and Extra Tree Regression, have been employed. The results indicate that RFR, SVR, and ETR perform better than other algorithms. While various ML-based frameworks have been proposed, certain algorithms need to be carefully analyzed, taking into consideration the practicality of implementing them on real hardware devices. There is a high impact of RSSI data filtering on localization accuracy. In the literature, many works have proposed linear and non-linear filters to filter the collected data. The works [8,9] used moving average filters, and [10,11] used Kalman and extended Kalman filters when smoothing RSSI data. Though Kalman filters outperform moving average filters, the moving average filters are widely used due to their simplicity [12].
The works [12,13,14,15,16] have used supervised classifiers for indoor localization. Neural Networks (NN), Feed Forward Neural Networks (FFNNs), SVM, and kNN are used as ML algorithms. According to [16], NN is good in accuracy, but it requires high computational power. The kNN also gives a reasonable accuracy. MATLAB, a Python-based IDE, has been used to implement these algorithms. During the evaluation of performances, confusion matrices are used. One crucial observation is that a standard ML algorithm cannot be proposed for all indoor applications because the outperformed algorithm varies due to many reasons, such as wireless technology, signal measurement techniques, the types of filters used, the environment’s robustness, the computational power of devices, the dataset’s size, the dataset’s hyperparameter tuning aspects, etc.
This paper investigates how to use supervised classifiers for indoor localization, where locations are estimated as categorical data with improved RSSI linearization. We introduced a novel linearization and filtering mechanism with ML framework to improve the localization accuracy in robust indoor localization systems. For this experiment, we used a publicly available testbed and its dataset [17]. In this case, we consider the location estimation as a classification-type ML problem, and where the specific location (zone) of an existing person will be predicted. This testbed consists of 13 BLE ibeacon nodes implemented inside a library. In addition, high security, a low cost, and minimal power consumption are all benefits of Bluetooth positioning technology; yet, Bluetooth has poor signal reliability and a limited communication range. More details on the dataset and testbed will be discussed later in this paper. We used the RSSI dataset for pre-processing using a moving average filter and weighted least-squares method. Afterward, the dataset was trained by the supervised classifiers FFNN, SVM, and kNN, and we evaluated each algorithm’s performance on localization. However, this study was limited to one particular BLE dataset, and other wireless technologies were not considered.
This paper is organized as follows: firstly, the background theory of wireless indoor localization will be discussed in the background section. Secondly, related works in ML for indoor localization and the mathematical model will be discussed. Thirdly, our novel system model is proposed. In the fourth section, we elaborate on the steps of data pre-processing and the ML-model development and training. The fifth section discusses the model evaluation in detail, including the hyperparameter tuning for each model. Finally, we conclude the findings.
Wireless indoor localization is a technique that estimates the objects’ location by using distance measurements. Primarily, indoor localization systems can be categorized as device-free and device-based localization. Typically, these location localization techniques require distance measurements as input for the algorithm. The widely used distance measurement techniques in positioning systems are the Angle of Arrival (AoA), Time of Arrival (ToA), Time of Flight (ToF), Time Difference of Arrival (TDoA), Channel State Information (CSI), and Received Signal Strength Indicator (RSSI). When comparing the above distance measurement techniques, there are pros and cons. For instance, the AoA system uses a set of antennae to determine the signal propagation angle, since more hardware is needed and implementation cost is high. The AoA technology typically requires complex hardware such as antenna arrays calibrated to get the correct position. The RSSI level correlates the distance between the transmitter and the receiver. Additionally, the signal intensity weakens as it goes away from the transmitter; it is frequently used to calculate the separation between a transmitter and the receiver of the nodes. Because the transmitting signal is vulnerable to environmental noise and multipath fading, RSSIs generally lead to inaccurate values and errors in localization systems; therefore, strong signal processing techniques should apply before it is processed through ML algorithms. Once it has a distance measure, it can feed as the input for the algorithm. Before being fed into the algorithms, it requires strong signal filtering. Specific algorithms require a bare minimum number of anchor nodes. For instance, trilateration requires at least three anchor nodes in the system [17,18,19].

3. System Model

3.1. Machine Learning for Wireless Indoor Localization

Machine learning algorithms can be categorized into supervised, semi-supervised, and unsupervised learning algorithms. The supervised learning algorithms required a labeled dataset for predictions, and unsupervised algorithms used an unlabeled dataset for their predictions. The typical ML lifecycle has several stages: data pre-processing, model training, hyperparameter tuning, and model evaluation. ML algorithms can be used for classification or regression tasks. In classification, machine learning algorithms classify data into different categories, while regression learns from training data to predict continuous variables. The object’s location can be estimated as a numerical value (regression) or as a classification (classifiers). For example, consider a person living inside a house and obtaining the location through an IoT system. The location of the person or object can be estimated geographically with coordinates or can predict the location as an area such as the bedroom, kitchen, etc. In this work, we have selected three algorithms, and each algorithm’s brief details are given as follows.

3.1.1. Artificial Neural Networks

Artificial Neural Networks considers a powerful ML technique that can be applied to indoor localization systems. The basic working of an ANN is like the human neural network in our brain. The ANN follows the same structure of how brain neurons are interconnected. The ANN has three layers: the input layer, the hidden layer(s), and the output layer. The ANN has many advantages. Once the ANN is trained, the model will provide the output even with inadequate data. Moreover, ANN has a memory distribution, manages fault tolerance, and has a flexible network structure. The ANN performance is highly dependent on the hardware it runs.
A Feed-Forward Neural Network (FFNN) is considered a sub-branch of ANN. An FFNN consists of at least one layer of neurons and input and output layers. The network’s intensity can be observed based on the collective behavior of the connected neurons. The output is chosen by evaluating the network’s output in its input context. The main benefit of this network is that it learns to assess and identify input patterns. Generally speaking, an Artificial Neural Network has three layers, named the input layer, the hidden layer, and the output layer. Few works in the literature have used FFNN for indoor localization, giving considerably good results. However, it was observed that ANNs require a considerable computational time and capacity.

3.1.2. Support Vector Machine

The SVM is a machine learning algorithm that uses labeled training samples to learn the classification of data points. For example, one way to detect malicious behavior on a node is to use SVMs to investigate the temporal and spatial correlation of data. To explain this point, the WSN observation is used as a point in the element space, and the SVM divides it into multiple parts. These parts are separated by the widest possible margin (the separation gap), and the new readings are sorted according to which side of the gap they are on. An SVM algorithm uses linear constraints to optimize a quadratic function (the problem of building a set of hyperplanes). It provides an alternative to Multi-layer Neural Networks with non-convex, unconstrained optimization problems.

3.1.3. K-Nearest Neighbor Algorithm

K-Nearest Neighbor (kNN) is considered the simplest classifier used for indoor positioning systems. The kNN algorithm uses a feature-similarity approach for prediction. The kNN algorithm can be used for both classification- and regression-type problems. In this study, the nearest neighbors were those data points with a minimal distance in the feature space from the new data point. Additionally, the value k was the number of such data points we considered in implementing the algorithm. For example, a number k was selected by KNN algorithms as the nearest neighbor to the data point that needed to be categorized. For instance, if k is 3, it will search for the three closest neighbors to that data point, where k is the hyperparameter of the kNN algorithm. The value of k has an impact on the model accuracy. A prepared table that lists the relative strengths of two or more separate received signals from neighborhood routers is necessary for the kNN algorithm. The offline phase is when the predetermined table is built.

3.2. Testbed and Dataset

We used the publicly available dataset in [23] for our experiments. The dataset used in this experiment is available [8]. In this experimental setup, the dataset seen in Figure 2 was created using the RSSI readings from a set of 13 iBeacons on the first floor of the Waldo Library at Western Michigan University. The data were collected using an iPhone 6S, Apple USA. A labeled dataset (1420 samples) and an unlabeled dataset (1420 samples) were included in the dataset (5191 samples). In the original dataset in [8] authors have numbered the experiment area from A to U as rows and 1–18 as columns. In the experiment each cell has a unique identification number i.e B3001 as in Figure 2. When it comes to our experiment, the total experiment area divided in to four zones as A, B, C and D for simplicity and the ML models developed to estimate the location as one of these zones as a categorical problem. The recording was completed within the library’s regular operating hours. During the experiments, other wireless devices were switched off to reduce the disturbances from the different sources. More significant RSSI values showed a nearer proximity to a given iBeacon. For out-of-range iBeacons, the RSSI was shown by −200. Figure 2 shows the layout of the plan of iBeacons in green color circles. For the experimentation, we used only the labeled dataset given by the authors.
Figure 2. Arrangement of the sensor nodes in the testbed [8].

4. Model Training and Evaluation

In this work, we followed the machine learning life cycle to develop the models for location estimation. We started with data pre-processing, where row RSSI data were linearized and filtered. Additionally, we trained our models with the row RSSI dataset to understand the proposed filtering effect in location estimation accuracy. Then, the dataset was trained using ML models implemented using the scikit-learn ML library available in Python 3. All the models were developed on a Jupyter Notebook. Each stage of the implementation is explained as follows.

4.1. Data Pre-Processing

The RSSI dataset consisted of 5191 data and was filtered using the moving average filters. The moving average (MA) filter is the most common and widely used filter in digital signal processing because it is the most accessible digital filter in this kind of application. The row RSSI data in both the time and frequency domains are shown in Figure 3. Where we have only plotted the RSSI data received from three access points to understand the fluctuation and nature of the signals. The RSSI values received from access points one, two and three are shown in black, orange and blue color in the Figure 3.
Figure 3. RSSI data with frequency and time domain.
After the filtering, four columns were added to the dataset, as shown in Table 2, where zones A, B, C, and D were shown as per Figure 2. This modification was performed as we needed to train our ML models as a classification task. Table 2 shows the original dataset, where the location column denotes the location ID allocated for each location, and OO2, PO1, PO2 and PO3 are the RSSI levels received by the anchor nodes. Table 3 shows the dataset modified for classification purposes, where four new columns are added.
Table 2. A few data samples from the original dataset.
Table 3. A few samples from the modified dataset.

4.2. Model Training

The filtered RSSI dataset was trained using FFNN, SVM, and kNN, respectively. The proposed FFNN architecture comprised two fully connected layers and one output layer, as shown in Figure 4. This FFNN had three layers, where, thirteen inputs were presented at the input layer as it had thirteen beacon node inputs. And it had one hidden layer and the output layer. The first fully connected layer included twenty neurons, while the second fully connected layer had seventeen neurons. For all simulations, the network architecture had four completely linked levels and one categorization layer. Using the neural network toolkit in MATLAB 2020, the FFNN was implemented. Moreover, the FFNN was trained underneath three different hyper-parameters values of the learning rate (LR), batch size, and epochs. The activation function of the FFNN is very important as it is accountable for transforming the node’s summed weighted input into the node’s activation. Here, we used the ReLU activation function, which will output the input directly if it is positive; otherwise, it is zero. Moreover, ReLU is easy to train and provides better performance. To increase the model’s accuracy during the simulations, the ReLU activation function in fully connected layers and the SoftMax function in the final layer were both utilized. We re-trained our model data-interpolating techniques to generate new samples between existing data points. This did not increase the accuracy significantly, but we observed smoother representation of the signal variations. This information is mentioned in the manuscript.
Figure 4. Testing accuracy vs. number of epochs.
In SVM, the choice of kernel function determined the mapping of data into a higher-dimensional space; here, we used a linear kernel, and the regularization parameter(C) was set to 0.01. The most important hyperparameter to consider when making predictions in k-NN was the number of nearest neighbors. It determines the size of the neighborhood used to classify or regress new data points. We kept k as 5 during our experiment. Both the SVM and kNN were trained using a Jypiter notebook and Python 3 programming language. To build the model, the scikit learn and Pandas library were used. The models were built on an Intel (R) Core (TM) i5-10210U CPU @ 1.60 GHz 2.11 GHz personal computer. During training, the dataset of 5191 was split as 70% for training and 30% for testing. The linear kernel was used for SVM, and the number of vectors was set as 25 in the initial phase. Further, the number of vectors was increased in the model so we could observe the effect of the number of vectors on the accuracy. Initially, the k value of kNN was set to be 4, and we gradually increased the value of k to observe the changes in the accuracy. Further, we also happened to experiment with Logistic Regression (LR), Decision Tree Classifier (DTC), and Random Forest Classifier (RFC). LR yielded an accuracy of 54.2%, DTC demonstrated 76.4% accuracy, and RFC achieved 78.7% accuracy.
All the algorithms evaluated the accuracy, precision, sensitivity, and F1 score. In SVM, the number of support vectors changed from 25 to 75 as a hyperparameter tuning.

5. Results and Discussion

5.1. Results

For all the algorithms, FFNN, SVM and kNN, we calculated the performance evaluation matrices, precision, recall, and the f1-score, in each zone. Finally, we calculated the matrices of the entire dataset. The general definitions for each performance evaluation matrix are as follows:
A c c u r a c y = T P + T N T P + F P + F N + T N
P r e c i s i o n = T P T P + F P
  S e n s i t i v i t y = T P T P + F N
F 1 - S c o r e = 2 × ( S e n s i t i v i t y × P r e c i s i o n ) S e n s i t i v i t y + P r e s i t i o n
Without linearization and filtering, the model provided overall accuracies of approximately 77% for KNN, 70% for SVM and 42% for FFNN. Once the filters and linearization were applied, the precision of the FFNN varied from 0.50 to 0.71, the SVM from 0.75 to 0.88, and the kNN from 0.78 to 0.96. Additionally, for the recall, the FFNN was from 0.40 to 0.76, the SVM was from 0.74 to 0.90, and the kNN was from 0.81 to 0.96. For output-sensitive estimation such as localization, we required a high precision and recall. Overall, all the algorithms had over 83% precision and over 76% accuracy in location estimation. The zone-wise precision, recall and f1-score for the FFNN and SVM are presented in Table 4 and Table 5, respectively, where the linear kernel used in the SVM had a gamma value of 0.001 and regularization parameter of 0.01. Table 6 shows the estimation performance zone-wise for the KNN algorithm.
Table 4. FFNN.
Table 5. SVM.
Table 6. KNN.
Previous studies have not paid much attention to hyperparameter tuning. This may have led to low accuracy in the location estimation being provided. Related works show some low accuracies, such as 50–60% in algorithms, due to the number of samples in the dataset and the due unavailability of RSSI filtering [18,19,20]. During the experiments, we considered the hyperparameter tuning of the SVM, kNN, and FFNN algorithms. For the SVM, the number of vectors was increased from 25 to 75. It can be observed that accuracy improves when the number of vectors in the SVM increases. Table 7 denotes the summary of performance for each algorithm. When increasing the vectors, it was observed that there was a significant improvement in the accuracy and f1-score. Overall, the kNN gave 0.8511 accuracies, and it outperformed the other algorithms.
Table 7. Performance comparison of all the machine learning algorithms.
For FFNN, we adjusted the hyper-parameters, including the learning rate, batch size, and the number of epochs. It was observed that the training accuracy was consistently higher than the testing accuracy. Three distinct batch sizes, 10, 50, and 100, were used to test the training and testing accuracy. The epochs ranged from 0 to 100. Figure 4 and Figure 5 display the outcomes. The kNN operates based on its proximity to training examples. In cases where the decision boundary is not highly complex and can be approximated well by a set of local regions, kNN can be effective. The FFNN and SVM might introduce unnecessary complexity in such scenarios. The practicality of implementing machine learning algorithms like the FFNN, k-NN, and SVM on IoT edge devices depends on factors like model complexity, computational resources, memory constraints, and the specific use case. While lightweight algorithms like k-NN are generally well-suited for edge deployment, more complex models like FFNN and SVMs may require careful optimization and the consideration of resource limitations. Additionally, model quantization, pruning, and approximate algorithms can be used to improve the feasibility of these algorithms on IoT edge devices.
Figure 5. Training accuracy vs. number of epochs.

5.2. Training Time

Figure 6 shows the association between the number of samples and the training time for all three algorithms. When the number of samples increases, all the algorithms’ training times increase, where kNN shows the lowest training time while FFNN gives the highest training time. It was observed that from 200 samples to 1200 samples, the training time increased exponentially. Additionally, after 1200 samples, it showed stableness. When comparing the training time of FFNN, it was higher compared to the other algorithms tested. This could be due to the computational complexity of the neural network. To overcome this, the stochastic gradient descent can be proposed. Stochastic gradient descent works by randomly selecting a small number m, chosen haphazardly training inputs. The kNN showed less training time because it does not learn a discriminative function from the training data.
Figure 6. Training time vs. various machine learning algorithms.

5.3. Accuracy

Figure 7 shows the correlation between the number of samples versus the accuracy. The accuracy rose as the number of samples increased for all the algorithms. In an FFNN up to 1000 samples, the accuracy increased exponentially. Additionally, after 1000 samples, accuracy became sort of saturated. However, the FFNN provides the lowest accuracy compared to the other two algorithms. The smaller number of data points could be a reason for it showing the lowest accuracy in the FFNN. In the SVM and kNN, up to 200 samples, we could see a significant increment of accuracy. However, after 200 samples, we saw less of an increment. The reduction of model overfitting can be used to explain the accuracy versus the training dataset size curve. In general, as the amount of the training dataset increases, model overfitting considerably decreases.
Figure 7. Accuracy vs. various machine learning algorithms.

5.4. F1-Score

Figure 8 shows the relationship between the number of samples and the f1-score. By computing the harmonic mean of a classifier’s precision and recall, the f1-score assimilates both into a single metric. It is mainly used to compare the effectiveness of two classifiers. As per the results, the f1-score also rapidly upsurged as the number of samples increased. According to the results, the kNN gave the highest f1-score, and the FNN offered the lowest value. Generally, the accuracy was suitable for a balanced dataset. However, the utilized dataset in this work was an imbalanced dataset, where we had an unequal number of data samples for four Zones A, B, C, and D. Since the dataset had an imbalance, the matrix f1-score was more suitable for evaluating this model.
Figure 8. f1-score vs. various machine learning algorithms.

5.5. Accuracy with Number of Vectors

Figure 9 shows the relationship between the accuracy, number of samples, and vectors. When the number of samples increases, accuracy increases irrespective of the number of vectors. When the number of vectors increases, accuracy sees a significant reduction. We considered the linear kernel SVM in this experiment. Additionally, improving the classifier’s performance did not always result in the addition of more support vectors. The trade-off between accuracy and the number of support vectors that we previously considered true did not always hold in this specific situation. If there were already more than a particular number of support vectors—in this case, about 110—then adding another support vector resulted in declining marginal returns.
Figure 9. Accuracy vs. number of samples, with different number of vectors.

5.6. Accuracy vs. k Value in kNN

Figure 10 shows the effect of the k value in the kNN on the accuracy. When k increased from 2 to 6, we can observe an exponential increment in accuracy. Additionally, from 6 to 20 the accuracy increased. A low number of k indicates that noise will have a more significant impact on the outcome. A considerable value makes it computationally expensive and somewhat contradicts the fundamental tenets of KNN.
Figure 10. Accuracy vs. number of samples with different k values.

6. Conclusions

This paper presented a simplified localization method based on BLE and supervised ML classifiers that could apply in robust indoor localization applications. Since the proposed localization method is based on ML, it could be easily deployed in practical environments as the ML model could run on tiny embedded systems. The RSSI values were collected from thirteen different iBeacon nodes installed in an indoor environment utilized in this experiment. The RSSI data were linearized using the weighted least-squares method and filtered using moving average filters to remove the outliers. Afterward, RSSI data were trained using the SVM, kNN, and FFNN algorithms. During the experiments, each algorithm’s hyperparameters were tuned, and the impact on its accuracy was observed. It was observed that there were significant improvements in the accuracy once the sample size was increased. Further, training time was observed against the sample size and the number of epochs versus accuracy. The training time increased as the number of samples increased. Further, accuracy had a significant improvement once the epoch size increased. Overall, the kNN offers reasonably good accuracy in classifying the right zone of 85%, the FFNN provides 76% when the batch size is 100, under a learning rate of 0.01, and the SVM provides 84% with 75 support vectors.

Author Contributions

Conceptualization, M.W.P.M. and V.T.; methodology, M.W.P.M. and V.T.; software, M.W.P.M.; validation, M.W.P.M., V.T. and R.A.; formal analysis, M.W.P.M. and V.T.; investigation, V.T. and R.A.; resources, M.W.P.M. and V.T.; data curation, M.W.P.M. and V.T.; writing—original draft preparation, M.W.P.M., V.T. and R.A.; writing—review and editing, M.W.P.M. and V.T.; visualization, M.W.P.M. and V.T.; supervision, V.T. and R.A.; project administration, V.T. and R.A.; funding acquisition, V.T. and R.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare that there are no conflict of interest regarding the publication of this paper.

References

  1. Psychoula, I.; Chen, L.; Amft, O.; Van Laerhoven, K. Privacy Risk Awareness in Wearables and the Internet of Things. IEEE Pervasive Comput. 2020, 19, 60–66. [Google Scholar] [CrossRef]
  2. Langheinrich, M. Long Live the IoT. IEEE Pervasive Comput. 2020, 19, 4–7. [Google Scholar] [CrossRef]
  3. Yassin, M.; Rachid, E. A survey of positioning techniques and location based services in wireless networks. In Proceedings of the 2015 IEEE International Conference on Signal Processing, Informatics, Communication and Energy Systems (SPICES), Kozhikode, India, 19–21 February 2015; pp. 1–5. [Google Scholar] [CrossRef]
  4. Alqahtani, E.J.; Alshamrani, F.H.; Syed, H.F.; Alhaidari, F.A. Survey on Algorithms and Techniques for Indoor Navigation Systems. In Proceedings of the 2018 21st Saudi Computer Society National Computer Conference (NCC), Riyadh, Saudi Arabia, 25–26 April 2018; pp. 1–9. [Google Scholar] [CrossRef]
  5. Nessa, A.; Adhikari, B.; Hussain, F.; Fernando, X.N. A Survey of Machine Learning for Indoor Positioning. IEEE Access 2020, 8, 214945–214965. [Google Scholar] [CrossRef]
  6. Maduranga, M.W.P.; Abeysekara, R. Supervised Machine Learning for RSSI-based Indoor Localization in IoT Appli-cations. Int. J. Comput. Appl. 2021, 183, 26–32. [Google Scholar]
  7. Mohammadi, M.; Al-Fuqaha, A. Enabling Cognitive Smart Cities Using Big Data and Machine Learning: Approaches and Challenges. IEEE Commun. Mag. 2018, 56, 94–101. [Google Scholar] [CrossRef]
  8. Sadowski, S.; Spachos, P. RSSI-Based Indoor Localization With the Internet of Things. IEEE Access 2018, 6, 30149–30161. [Google Scholar] [CrossRef]
  9. Chen, L.; Zhou, X.; Chen, F.; Yang, L.-L.; Chen, R. Carrier Phase Ranging for Indoor Positioning With 5G NR Signals. IEEE Internet Things J. 2021, 9, 10908–10919. [Google Scholar] [CrossRef]
  10. Boyaci, O.; Narimani, M.R.; Davis, K.R.; Ismail, M.; Overbye, T.J.; Serpedin, E. Joint Detection and Localization of Stealth False Data Injection Attacks in Smart Grids Using Graph Neural Networks. IEEE Trans. Smart Grid 2021, 13, 807–819. [Google Scholar] [CrossRef]
  11. Mussina, A.; Aubakirov, S. Improving Indoor Positioning via Machine Learning. In Proceedings of the 8th International Conference on Data Science, Technology and Applications (DATA 2019), Prague, Czech Republic, 26–28 July 2019. [Google Scholar] [CrossRef]
  12. Maduraga, M.W.P.; Abeysekara, R. Comparison of supervised learning-based indoor localization techniques for smart building applications. In Proceedings of the 2021 International Research Conference on Smart Computing and Systems Engineering (SCSE), Colombo, Sri Lanka, 16 September 2021; pp. 145–148. [Google Scholar] [CrossRef]
  13. Gadhgadhi, A.; Hachaichi, Y.; Zairi, H. A Machine Learning based Indoor Localization. In Proceedings of the 2020 4th International Conference on Advanced Systems and Emergent Technologies (IC_ASET), Hammamet, Tunisia, 15–18 December 2020; pp. 33–38. [Google Scholar] [CrossRef]
  14. Jun, Z.G.; Xin, L.; Long, X.Z.; Chao, L.H. Weighted Least Square Localization Algorithm Based on RSSI Values. In Proceedings of the 2015 Fifth International Conference on Instrumentation and Measurement, Computer, Communication and Control (IMCCC), Qinhuangdao, China, 18–20 September 2015; pp. 1236–1239. [Google Scholar] [CrossRef]
  15. Ashraf, I.; Hur, S.; Park, S.; Park, Y. DeepLocate: Smartphone Based Indoor Localization with a Deep Neural Network Ensemble Classifier. Sensors 2019, 20, 133. [Google Scholar] [CrossRef] [PubMed]
  16. Maduranga, M.; Abeysekera, R. Bluetooth Low Energy (BLE) and Feed Forward Neural Network (FFNN) Based Indoor Positioning for Location-based IoT Applications. Int. J. Wirel. Microw. Technol. 2022, 12, 33–39. [Google Scholar] [CrossRef]
  17. Che, F.; Ahmed, A.; Ahmed, Q.Z.; Zaidi, S.A.R.; Shakir, M.Z. Machine Learning Based Approach for Indoor Localization Using Ultra-Wide Bandwidth (UWB) System for Industrial Internet of Things (IIoT). In Proceedings of the 2020 International Conference on UK-China Emerging Technologies (UCET), Glasgow, UK, 20–21 August 2020; pp. 1–4. [Google Scholar] [CrossRef]
  18. Koutris, A.; Siozos, T.; Kopsinis, Y.; Pikrakis, A.; Merk, T.; Mahlig, M.; Papaharalabos, S.; Karlsson, P. Deep Learning-Based Indoor Localization Using Multi-View BLE Signal. Sensors 2022, 22, 2759. [Google Scholar] [CrossRef] [PubMed]
  19. Paudel, K.; Kadel, R.; Guruge, D.B. Machine-Learning-Based Indoor Mobile Positioning Using Wireless Access Points with Dual SSIDs—An Experimental Study. J. Sens. Actuator Netw. 2022, 11, 42. [Google Scholar] [CrossRef]
  20. Singh, S.; Kumar, K.; Gupta, S. Machine Learning based Indoor Localization Techniques for Wireless Sensor Networks. In Proceedings of the 2020 2nd International Conference on Advances in Computing, Communication Control and Networking (ICACCCN), Greater Noida, India, 18–19 December 2020; pp. 373–380. [Google Scholar] [CrossRef]
  21. Kaur, J.; Shawky, M.; Mollel, M.S.; Popoola, O.R.; Imran, M.A.; Abbasi, Q.H.; Abbas, H.T. AI-enabled CSI fingerprinting for indoor localisation towards context-aware networking in 6G. In Proceedings of the 2023 IEEE Wireless Communications and Networking Conference (WCNC), Glasgow, UK, 26–29 March 2023; pp. 1–5. [Google Scholar] [CrossRef]
  22. Song, X.; Fan, X.; He, X.; Xiang, C.; Ye, Q.; Huang, X.; Fang, G.; Chen, L.L.; Qin, J.; Wang, Z. CNNLoc: Deep-Learning Based Indoor Localization with WiFi Fingerprinting. In Proceedings of the 2019 IEEE SmartWorld, Ubiquitous Intelligence & Computing, Advanced & Trusted Computing, Scalable Computing & Communications, Cloud & Big Data Computing, Internet of People and Smart City Innovation (SmartWorld/SCALCOM/UIC/ATC/CBDCom/IOP/SCI), Leicester, UK, 19–23 August 2019; pp. 589–595. [Google Scholar] [CrossRef]
  23. Mohammadi, M.; Al-Fuqaha, A.; Guizani, M.; Oh, J. Semi-supervised Deep Reinforcement Learning in Support of IoT and Smart City Services. IEEE Internet Things J. 2017, 5, 624–635. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.