You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

25 May 2021

LightGBM Indoor Positioning Method Based on Merged Wi-Fi and Image Fingerprints

and
1
Faculty of Information Technology, Beijing University of Technology, Beijing 100124, China
2
Engineering Research Center of Digital Community, Ministry of Education, Beijing 100124, China
3
Beijing Laboratory for Urban Mass Transit, Beijing 100124, China
*
Author to whom correspondence should be addressed.
This article belongs to the Section Navigation and Positioning

Abstract

Smartphones are increasingly becoming an efficient platform for solving indoor positioning problems. Fingerprint-based positioning methods are popular because of the wide deployment of wireless local area networks in indoor environments and the lack of model propagation paths. However, Wi-Fi fingerprint information is singular, and its positioning accuracy is typically 2–10 m; thus, it struggles to meet the requirements of high-precision indoor positioning. Therefore, this paper proposes a positioning algorithm that combines Wi-Fi fingerprints and visual information to generate fingerprints. The algorithm involves two steps: merged-fingerprint generation and fingerprint positioning. In the merged-fingerprint generation stage, the kernel principal component analysis feature of the Wi-Fi fingerprint and the local binary pattern features of the scene image are fused. In the fingerprint positioning stage, a light gradient boosting machine (LightGBM) is trained with mutually exclusive feature bundling and histogram optimization to obtain an accurate positioning model. The method is tested in an actual environment. The experimental results show that the positioning accuracy of the LightGBM method is 90% within a range of 1.53 m. Compared with the single-fingerprint positioning method, the accuracy is improved by more than 20%, and the performance is improved by more than 15% compared with other methods. The average locating error is 0.78 m.

1. Introduction

Location-based services deliver excellent research and commercial value and have become a common object of research. The Global Navigation Satellite System (GNSS) provides reliable location services outdoors. With the expansion of urban areas, human activities in indoor environments are becoming increasingly abundant, and the demand for indoor positioning services is increasing. Meanwhile, smartphones that integrate wireless, visual, and accelerometer sensors can facilitate indoor positioning services [1]. However, owing to the complex and variable natures of indoor environments, the large-scale application of indoor positioning solutions has yet to be achieved. Researchers have used a variety of indoor signals for positioning, including wireless local area network (WLAN) facilities widely distributed in indoor environments, cellular networks [2], Bluetooth [3], radio-frequency identification [4] and other radio frequency signals, microelectromechanical system gyroscopes [5], ultra-wideband (UWB) [6,7], laser ranging [8], and visual information [9]. Wi-Fi fingerprint positioning does not require the distances and angles to be known in advance; however, it is seriously affected by indoor multipath effects. Cellular networks are mainly used for smartphone positioning; however, their accuracy is generally low because of problems such as time synchronization; with the development of 5G communication network technology, this method is expected to achieve higher accuracies. Although image-based positioning technology offers high accuracy, it suffers from occlusion, illumination, and blur. Bluetooth positioning offers the advantages of low power consumption, close range, and wide applicability, though its stability is poor. The UWB signal has strong penetrability and high security, and it can integrate positioning and communication functions; however, owing to its high cost, it has not yet been widely implemented. Individual positioning methods struggle to meet the high-precision pedestrian self-positioning requirements, despite their portability, cost, and environmental adaptability [10]. The fusion positioning method exploits the complementary capacities of different sensors to obtain information and provide a richer description of indoor locations; thus, it has become a research hotspot in recent years. For example, the channel-state-information/magnetic-field-strength fusion positioning method proposed by Li et al. [11] and Simon et al. [12] uses a visual inertial positioning algorithm to automatically collect Bluetooth signal strength data and reduce the labor density of the earlier data-collection methods based on wireless signal positioning. Single-signal-source positioning technologies rely on the stability of this type of signal, which is extremely difficult to maintain in a complex indoor environment. The motivation for fusing visual and Wi-Fi signals for positioning applications is to improve the adaptability of the positioning system to indoor environments. This can prevent positioning failures caused by a single signal fluctuation. The present work proposes a new smartphone-based indoor positioning method that uses the scene image and received signal strength (RSS) value of the WLAN access point (AP) as the input of the merge positioning system to realize high-precision pedestrian self-positioning. This research can also be used for the auxiliary positioning of indoor service robots and in business advertising campaigns.
Scene images and Wi-Fi fingerprint positioning approaches must consider two key issues: (1) the extraction of key positioning information from the Wi-Fi fingerprints and scene images, and (2) the dimensional unification of image information and Wi-Fi fingerprints. The focus of this article is to propose a new combined fingerprint that describes location information using visual and Wi-Fi signals to realize combined-fingerprint positioning. The achieved positioning accuracy is more than 20% higher than that obtained using Wi-Fi or visual positioning methods alone; meanwhile, it achieves a faster running speed. It takes less than 2 s to obtain the predicted position coordinates from the feature extraction results; thus, this method can meet the demands of real-time positioning. The main contributions of this study are as follows:
  • We propose a merged location fingerprint based on Wi-Fi fingerprints and scene image features. Of these, Wi-Fi fingerprint features are obtained by extracting effective positioning information from the original Wi-Fi fingerprint using the kernel principal component analysis (KPCA) method; next, scene image features are extracted by local binary patterns (LBPs). The image data are transformed into structured data so that the scene information and Wi-Fi fingerprint can jointly describe the spatial location in the same dimension, which reduces the storage space occupied by the merge fingerprint library.
  • Based on the merged location fingerprint, a light gradient boosting machine (LightGBM), which can effectively process structured data, is used to construct an indoor positioning model. This positioning model can quickly and accurately obtain positioning results, and is easy to implement on smartphone platforms with limited computing resources. Our experiments prove that the proposed method is simple and effective.
The structure of this paper is as follows: Section 2 introduces the relevant research regarding the use of scene images and Wi-Fi fingerprints in indoor positioning; Section 3 introduces the merged fingerprint generation procedure and the construction of indoor positioning models to realize individual positioning; Section 4 presents the experimental results and analysis; finally, Section 5 summarizes the research and considers future research directions.

3. Merge Fingerprint LightGBM Indoor Positioning Algorithm

The merged fingerprint proposed in this study includes Wi-Fi fingerprint features and scene image features. Figure 1 presents a flow chart of the merged-fingerprint positioning system. The Wi-Fi fingerprint feature is obtained using the KPCA method, and the scene image feature is represented by the LBP feature histogram (i.e., the LBPH). Both Wi-Fi signals and scene images have their own uniqueness; that is, the merged fingerprint and location information correspond uniquely. The flow chart of the merged location fingerprint positioning system is shown in Figure 1; it is divided into offline and online stages.
Figure 1. Flow chart of merged location fingerprint positioning system.
First, in the offline stage, the RPs are divided into the experimental area P = { P 1 , P 2 , , P n } , where P denotes the set of RPs, and n is the total number of sampling points. The location feature is extracted from the scene image and Wi-Fi RSSI obtained at the RP, the merged fingerprint database is generated, and the LightGBM positioning model is trained. In the online stage, the same processing is performed on the data collected at the test point to generate a merged fingerprint of the point to be located, and the trained positioning model is used to predict the current position coordinates.

3.1. Extract the KPCA Features of Wi-Fi Fingerprints

The Wi-Fi fingerprint is collected by the mobile device, and it includes the media access control address of the access point and the corresponding RSSI. The dimension of the Wi-Fi fingerprint depends on the number of access points that can be received in the localization area; hence, the Wi-Fi fingerprint is a type of high-dimensional data. At the same time, Wi-Fi fingerprints are closely related to location, albeit non-linearly. Wi-Fi fingerprints are time-varying, high-dimensional nonlinear data. Directly using the original Wi-Fi fingerprint to identify the location is inefficient, and noise interference arises. In this research, KPCA was selected for Wi-Fi fingerprint characteristics, to reduce fingerprint dimensions and extract key positioning features.
KPCA represents a non-linear extension of PCA; it can realize the nonlinear dimensionality reduction of, and feature extraction from, data. The basic idea is to use the kernel function to map a linearly inseparable dataspace to a high-dimensional space, making it linearly separable in the high-dimensional space before performing PCA.
The Wi-Fi fingerprint sample set is Z = { ( R i q , P i ) q = 1 Q } i = 1 n , q = 1 , 2 , , Q . R i q denotes the q-th fingerprint at the i-th position, Q is the number of fingerprint samples at each RP, n is the number of RPs, and N is the total number of fingerprint samples, N = n × Q . R i q = [ R S S I i 1 , R S S I i 2 , , R S S I i k ] , R S S I i j ( j = 1 , , K ) indicates that the RSSI value from the j-th AP is received at the position of RP i. P i = ( x i , y i ) denotes the physical position coordinate of the fingerprint point i. Let W = ( R 11 , R 12 , , R n Q ) , where W is the standardized Wi-Fi fingerprint data. When mapping W to a high-dimensional space via the mapping function Φ, the mapping function is unknown. In the dataset W, each sample R i q is a K-dimensional column vector, and there are N samples in W. The space containing the K × N matrix W is referred to as the input space. The KPCA feature extraction process for WiFi fingerprints is as follows:
We use a nonlinear mapping Φ to map the vector w in W to the D-dimensional feature space F, as follows:
Φ ( W ) : R i q K R i q D , D K .
After mapping, a new D × N matrix Φ(W) in the feature space is obtained, as
Φ ( W ) = [ Φ ( R 11 ) , Φ ( R 12 ) , , Φ ( R n Q ) ] ,
where Φ ( R i q ) represents the mapping of R i q in the high-dimensional feature space. We perform PCA analysis on Φ ( W ) to obtain the KPCA features of the Wi-Fi fingerprint w = [ w 11 , w 12 , , w n Q ] .
Through the KPCA processing, the original position fingerprint space of the K × N order can be transformed into a K × m feature-position fingerprint space. Z = { ( w i q , P i ) q = 1 Q } i = 1 n is the Wi-Fi fingerprint KPCA feature dataset, where the dimension is m. The fingerprint feature dimension m has a larger influence on the model prediction accuracy. Therefore, the fingerprint feature dimension must be selected in the offline training stage to achieve the optimal positioning effect.

3.2. Extract Image LBP Features as Image Fingerprints

The location fingerprint can be any feature that facilitates location discrimination, and it can have diverse types. In this study, the LBPH is used as the image fingerprint to describe the location information together with the Wi-Fi fingerprint. Positioning methods based on visual information are often affected by illumination, occlusion, and shooting angles. LBP is a highly discriminative texture operator with significant advantages in terms of gray-level and rotation invariance. It is widely used in target detection to describe image texture features. Ojala et al. [37] proposed a uniform pattern to reduce the dimensionality of the LBP operator model. This is defined such that when the cyclic binary number corresponding to a certain LBP has at most two transitions (from 0 to 1 or from 1 to 0), the binary corresponding to the LBP is a uniform pattern class. This study uses the uniform-pattern-based rotation-invariant LBP feature to describe the scene contour and texture information, and it calculates it according to
U ( L B P l , r u ) = s ( g l 1 g c ) s ( g 0 g c ) + l = 1 L s ( g l g c ) s ( g l 1 g c ) ,
L B P l , r 2 = l = 0 L 1 s ( g l g c ) , i f   U ( L B P l , r u ) 2 L + 1 , o t h e r w i s e ,
s ( g ) = 1 , i f   g 0 0 , i f   g 0 ,
where L represents the number of pixels in the neighborhood, g c is the gray value of the central pixel, and g l is the gray value of the neighborhood pixel. s ( g l g c ) compares the sizes of g l and g c ; if g l exceeds g c , the value of s ( g l g c ) is 1; otherwise, it is 0. U ( L B P l , r u ) is the metric of the rotation-invariant uniform pattern, which represents the number of transitions from 0 to 1 or from 1 to 0. L B P l , r 2 calculates the unique label corresponding to the type of LBP rotation-invariant equivalent mode operator whose number of transitions (from 0 to 1 or from 1 to 0) does not exceed 2. s ( g ) is a symbolic function. This study uses L B P 8 , 1 2 to calculate the type of rotation-invariant uniform pattern. We perform histogram statistics on the number of pixels in the LBP rotation-invariant uniform pattern operator category of the entire image, to obtain a ten-dimensional LBP feature histogram vector. The uniform-pattern-based rotation-invariant LBP operator guarantees the stability of the image fingerprint. At the same time, the LBP image fingerprint is a type of structured data that can be combined with a Wi-Fi fingerprint to form a merged fingerprint.

3.3. Build a Merged Fingerprint

The first fusion of the image and Wi-Fi fingerprints involves direct stitching. The merged fingerprint is obtained by merging the KPCA features of the Wi-Fi fingerprint and image fingerprint after dimensionality reduction. Generally, Wi-Fi fingerprints are taken at the same location through multiple collections of the mean or median values. The merged fingerprint proposed in this study uses multiple scene images from the same RP position. Therefore, in the experiment, the same number of Wi-Fi fingerprint data as the scene image were also collected at the same RP position, and the corresponding scene image features and Wi-Fi fingerprint features were spliced together. The merge fingerprint database is shown in Figure 2; it contains the KPCA features of the Wi-Fi fingerprints, a total of six dimensions, and the scene image LBP features with a length of 10 dimensions. The form of the merge fingerprint is d i q = { ( w i q * , H i q ) q = 1 Q } i = 1 n , and the merge fingerprint data set is D = { ( d i q , P i ) q = 1 Q } i = 1 n , where w i q * is the Wi-Fi fingerprint feature of the location point i, H i q is the q-th image fingerprint at position i, and P i is the coordinate of the location point i. The merged fingerprint database is shown in Figure 2.
Figure 2. Merged fingerprint database.

3.4. Establish LightGBM Positioning Model

LightGBM is an ensemble learning framework based on gradient boosting decision trees that was developed by Microsoft Research. It uses decision trees as the base learner to continuously fit the residuals of the current learner, and it iteratively trains the model using a forward distribution algorithm. Each iteration seeks to minimize the loss function. LightGBM uses a histogram-based segmentation algorithm to replace the pre-sort traversal algorithm, and it reduces the number of samples and features through gradient-based, one-side sampling and exclusive feature bundling (EFB). This model offers the advantages of fast and high performance.
The second merge of the merge fingerprints is performed using the EFB algorithm in LightGBM. ‘Exclusive feature’ refers to the fact that some features rarely have non-zero values simultaneously, and these features are bundled together to form a new feature, which is used to reduce the number of features and improve training speeds. The EFB algorithm applies the idea of graph building, uses features as nodes, connects edges between non-mutually exclusive features, and then identifies all bundled feature sets from the graph. This problem is an NP-hard problem, and EFB uses a greedy strategy to solve it. This allows a small number of samples between features that are not mutually exclusive and sets a maximum conflict threshold K. The time complexity of the EFB algorithm is O(n2).
The process of establishing the fusion fingerprint LightGBM positioning model is as follows:
First, the merged fingerprint dataset D is used as the input, and the first boosted tree f 0 ( d i q ) is initialized as
f 0 ( d i q ) = arg min c i = 1 n q = 1 Q L ( P i , c ) ,
where P i represents the spatial position coordinates of the i-th collection point; c is the output value of the leaf node of the promotion tree, which is the value that minimizes the loss function (i.e., the predicted value of the position coordinates of the i-th collection point); and L is the loss function.
Suppose the decision tree obtained from the t−1 iteration is f t 1 ( d ) , and the loss function is L ( P i , f t 1 ( d ) ) . Then, the purpose of the t-th iteration is to identify the base learner T ( d , θ t ) and minimize the loss function L ( P i , f t ( d ) ) = L ( y , f t 1 ( d ) + T ( d , θ t ) ) .
This article uses the mean square error loss function, which is
L ( y , f t 1 ( d ) + T ( d , θ t ) ) = 1 2 [ y f t 1 ( d ) T ( d , θ t ) ] 2 = [ τ T ( d , θ t ) ] 2 ,
where τ = P i f t 1 ( d ) is the residual. The decision tree fits the residual of the current learner at each iteration. Typically, the value of the negative gradient of the loss function in the current learner is used as an approximate value:
τ i q , t [ L ( P i , f ( d ) ) f ( d ) ] f ( d ) = f t 1 ( d ) .
The residual is taken as the new true value of the sample, and we use { ( d i q , τ i q , t ) q = 1 Q } i = 1 N as the training data to obtain the decision tree f t ( d ) , where the set of leaf nodes is C t j , j = 1 , 2 , , J . For each leaf node, we calculate the best-fit value c t j as follows:
c t j arg min c d i q C t j L ( P i , f t 1 ( d ) + c f t ( d ) ) .
We update positioning model f t ( d ) , as follows:
f t ( d ) = f t 1 ( d ) + j = 1 J c t j I ( d C t j ) .
We obtain the final positioning model as:
f T ( d ) = f 0 ( d ) + t = 1 T j = 1 J c t j I ( d C t j ) .
The process of training a regression tree in LightGBM is shown in Figure 3.
Figure 3. Process of training a regression tree in LightGBM.
The pseudocode of the algorithm that identifies the optimal split point mentioned in Figure 3 is shown in Algorithm 1. The time complexity of the histogram algorithm for calculating the split gain is O(bin*features), where “bin” denotes the number of bins for each feature. Compared with other decision tree algorithms that use pre-sorting algorithms, the time complexity of the pre-sorting algorithm is greatly reduced, being expressed as O(data*features). The bin is much smaller than the data.
Algorithm 1 Identifying the optimal splitting point algorithm of the histogram.
Algorithm 1 BestSplitByHistogram Algorithm
1Input: d: training data, max_depth
2Input:m: merger fingerprint dimension
3nodeSet = {0} #tree nodes in current level
4rowSet = {{0,1,2,...}} #data indices in tree nodes
5for i = 1 to max_depth
6 for node in nodeSet do
7 usedRows = rowSet[node]
8 for j = 1 to m do
9 H = new Histogram()
10 #Build histogram
11 for k in usedRows do
12 bin = d.s[j][k].bin
13 H[bin].g += d.g[j] #Sum of gradients in each bin
14 H[bin].n += 1 #Sum of samples in each bin
15 Find the best split on histogram H.
16 Update rowSet and nodeSet according to the best split points
In this study, we constructed the LightGBM positioning model for merged fingerprint dataset D according to the X and Y coordinates. The establishment of the fusion fingerprint LightGBM model proceeds as shown in Algorithm 2.
Algorithm 2 Establishing the merged fingerprint LightGBM model.
Algorithm 2 LightGBM localization algorithm based on merged fingerprint
1Input: imgSet, wifiFingerprintSet, Rpnum
2wifi_KPCA = [[]]
3imgFingerprint = [[]]
4mergeFP = [[]]
5wifi_KPCA = KPCA(wifiFingerprintSet)
6for n = 1 to Rpnum
7 for q to n do
8 imgFingerprint[n][q] = LBP(imgSet[n][q])
9 mergeFP[n][q] = [wifi_KPCA[n][q], imgFingerprint[n][q]]
10XpreModel = LightGBM.train(mergeFP,Xcoordinates)
11YpreModel = LightGBM.train(mergeFP,Ycoordinates)

4. Experimental Results and Analysis

4.1. Experimental Setup

This study verifies the positioning performance of the proposed algorithm in a teaching facility environment. The experiment was conducted in the corridor and elevator room on the tenth floor of the Science Building of Beijing University of Technology. The area of the experimental environment was 10 m × 7 m. A partial plan view is presented in Figure 4. In the figure, the distance between two adjacent points in the X direction is 0.85 m, and the distance in the Y direction is 0.7 m; the 60 dots represent RPs, and the 20 cross-shaped dots represent test points. The experiment used self-developed RSSI signal acquisition software to collect 69 APs deployed in the teaching area. A total of 20 merged fingerprints at each RP were collected in the southern, eastern, northern, and western directions. The Wi-Fi fingerprint and image acquisition device was a Mi 10 Ultra smartphone, for which the parameters are listed in Table 1. The shooting height of the scene image was ~1.5 m (the experimenter’s height was 1.7 m).
Figure 4. Plan view of the experimental area.
Table 1. Experimental equipment parameters.

4.2. Data Preprocessing

Figure 5a denotes the RSS value of the same AP received at Positions 6 and 27. It can be observed that the Wi-Fi signal exhibits severe volatility. However, the same RSS value may appear in different positions, which causes difficulties in position discrimination. Position 6 is represented by a red dot in Figure 4, and Position 27 is represented by a blue dot. Figure 5b shows the scene images in Positions 6 and 27 at the top and bottom, respectively. Though the Wi-Fi signal strengths may appear identical, the difference between the scene images is very large, providing a multi-angle description of the position information.
Figure 5. Comparison of Wi-Fi fingerprints and scene images in different locations: (a) The RSS value for the same AP, received at Positions 6 and 27. (b) The upper two pictures are the scene images at Position 6, and the lower two pictures are the scene images at Position 27.
The Wi-Fi fingerprint was preprocessed, the signal strengths of the AP not collected at the collection point was set to −100, and the Wi-Fi fingerprint data after KPCA dimension reduction were normalized using
w i q * = w i q w i q , min w i q , max w i q , min ,
where w i q , min refers to the minimum value in the sample data, w i q , max refers to the maximum value in the sample data, and w i q * is the normalized Wi-Fi fingerprint data.
We normalize the LBP feature histogram, as follows:
h i q , k = b k k = 0 K + 1 b k .
Here, h i q , k is the normalized result of the K-th type of rotation-invariant uniform pattern LBP operator, and b k denotes the number of pixels belonging to the k-th rotation-invariant uniform pattern operator class; there are ten types in total; hence, H i q = [ h i q , 1 , h i q , 2 , , h i q , 10 ] . b is the normalized LBP feature vector of the i-th scene image.
In this study, the calculation formulas of the positioning error e i and average positioning error ave are as follows:
e i = ( x p r e x i ) 2 + ( y p r e y i ) 2 ,
a v e = 1 M i = 1 M e i .
( x p r e , y p r e ) are the predicted coordinates of the positioning algorithm, ( x i , y i ) are the real coordinates of the test point, and M is the total number of samples in the test set.
Accuracy was also used to evaluate the positioning results in this study [38]; it corresponds to the error distribution of the distance between the predicted and true positions. The cumulative distribution function is typically used to measure accuracy. For an indoor positioning algorithm with identical accuracy, the faster the cumulative distribution function curve reaches the peak, the better the method performance. In practice, a percentage is generally used to calculate accuracy. For example, if the accuracy of a positioning method within 1.5 m is 90%, the cumulative distribution function of the positioning error is less than 90% within 1.5 m.

4.3. The Influence of KPCA Location Feature Extraction on Location Accuracy

Through a selection experiment, this study verifies the ability of KPCA to reduce the noise interference of the original Wi-Fi fingerprint whilst retaining the fingerprint dimension; as shown in Figure 6, we compare the obtained results with those of the PCA algorithm, to verify the effectiveness of KPCA for processing the sparse data of Wi-Fi fingerprints. The PCA algorithm’s mapping from high- to low-dimensional spaces is linear; thus, it is difficult to effectively process Wi-Fi fingerprint information, and the positioning results of the KPCA-LGBM algorithm are significantly better than those of the PCA-LGBM one. To summarize, when the Wi-Fi fingerprint dimension k increases, the positioning error first decreases and then increases. When the fingerprint dimension m = 6, the average positioning error is 0.78 m. Taking this as the inflection point, when the fingerprint dimension is too low, the features pertaining to positioning are also lost; hence, the positioning error is relatively large. When the fingerprint dimension is too high, the noise in the data cannot be effectively removed; this influences the positioning accuracy.
Figure 6. Influence of fingerprint dimension on average positioning error.

4.4. Analysis of Single Fingerprint Positioning Error

For comparison, Table 2 and Figure 7 show the average positioning error of the LightGBM positioning algorithm using the original Wi-Fi fingerprint alone and the image fingerprint alone. The average positioning error of the image fingerprint is 0.97 m, less than the average positioning error of the Wi-Fi fingerprint (2.30 m). Owing to the time variability of Wi-Fi fingerprints and the possible similarities of the teaching building scene, the positioning accuracy of the single-fingerprint dataset was lower than that obtained using a merged one. It can be seen that the complementarity between the merged features is essential for improving the positioning accuracy. When Wi-Fi fingerprints in different locations are similar, these locations are distinguished by image features; conversely, when it is difficult to determine the location for the scene, Wi-Fi fingerprints can provide identification information.
Table 2. Comparison of average positioning error for a single-fingerprint dataset.
Figure 7. Cumulative probability distribution of positioning error for a single-fingerprint dataset.

4.5. Error Comparison between the Merged-Fingerprint LightGBM Algorithm and Other Positioning Algorithms

This study compared the average positioning error and running time of the merged-fingerprint LightGBM positioning algorithm to verify its effectiveness. The merge-fingerprint LightGBM positioning method was compared with Adaptive Boosting Algorithm (AdaBoost), Decision Tree (DT), Gradient Boosting Decision Tree (GBDT), Random Forest (RF), and Support Vector Regression (SVR). The comparison experiments all use the merged fingerprint dataset with a Wi-Fi fingerprint dimension of 6.
Figure 8 shows the cumulative probability distribution of the positioning errors for the six algorithms after tuning the grid search parameters to reflect the positioning accuracy. It can be seen from Figure 8 that the cumulative probability distribution of the fusion fingerprint LightGBM positioning algorithm in each error range exceeded that of the other algorithms, and a positioning accuracy of 90% was achieved within 1.53 m for all samples.
Figure 8. Cumulative probability distribution of the average errors for the six algorithms.
Table 3 compares the average positioning errors and running times of the six algorithms when using 50, 80, 90, and 100% of the samples. The final average positioning error of the fusion fingerprint LightGBM algorithm was 0.78 m, more than 15% more accurate than the other five algorithms. The merged fingerprint data proposed in this paper are a type of structured data, and the DT-based model performed better in terms of positioning error. Among the three high-positioning-accuracy algorithms (i.e., LightGBM, GBDT, and RF), the LightGBM positioning model ran fastest, at 16.75 ms. The DT-based ensemble learning model was slower than the DT model, though it fitted the data better. The running time of the proposed algorithm was primarily consumed in the feature extraction stage. The total running time (for extracting WiFi fingerprint KPCA features and image LBP features) was 1.71 s.
Table 3. Comparison of average positioning errors and running times of the six algorithms.

4.6. The Influence of the Maximum Depth of the Classification Regression Tree on Positioning Accuracy

The LightGBM base learner is a classification regression tree. To increase the generalizability of the model and prevent it from overfitting, the maximum depth of the classification regression tree must be limited. As can be seen in Figure 9, the average positioning error first decreases and then increases when the maximum depth of the classification regression tree is increased. When the maximum depth was 8, the curve reached its lowest point, and the average positioning error was 0.78 m. By continuing to increase the maximum depth, the model appeared to over-fit, the generalizability was weakened, and the positioning error increased. The model learning rate was 0.08, the number of classification regression tree was 85, and the maximum number of leaves was 17.
Figure 9. Influence of the maximum depth of the classification regression tree on the average positioning error.

4.7. Comparison with Other Algorithms

Table 4 presents a comparison between the proposed algorithm and those reported in other studies. It can be seen from the table that the error of the method proposed in this study is relatively small. The positioning error of the Wi-Vi method (proposed by Huang et al.) after a large number of experiments is very small, and the images taken by this method are limited to exit signs and surrounding scenes. When the number of exit signs is small, Wi-Fi positioning is used. This study considers the effects of illumination and shooting angle on the image. We chose as many angles and lighting conditions as possible when shooting the scene images of the training set. Meanwhile, the images in the test set of their experiment were very different from those in the training set. The time interval between the Wi-Fi fingerprint training set and the test set collection in the experiment was two weeks, and the experimental area was a corridor. The flow of people had a greater impact on the Wi-Fi signal strength, though the proposed algorithm was still effective. Because the experimental environments in different studies differ significantly, only the experimental results given in the literature are listed for comparison.
Table 4. Comparison of the proposed algorithm with established algorithms.

4.8. Threats to Validity and Limitations

Internal validity: The main threat to internal validity arises from factors that may affect positioning performance. The factors that affect the experimental results include the dimensions of the Wi-Fi fingerprint KPCA feature, the parameter settings of LightGBM, and the quantity of scene-image texture information.
External validity: In this study, all experimental data were collected in static mode, and restrictions were placed on the height and angle of image capture. The density of access points in the environment and the use of different equipment to collect data may affect the results of the experiment.
However, this does not mean that this research can only be applied in static mode. In future work, we will conduct experiments under dynamic conditions. This research requires a large quantity of data to be collected during the model-establishment stage, and the quantity and quality of the data are closely related to the experimental results. Therefore, reducing the fingerprint-collection workload is a key consideration.

5. Conclusions

A fusion location fingerprint combining Wi-Fi and image features was proposed, and a LightGBM regression positioning model was established. The algorithm first extracts the KPCA function from Wi-Fi information to eliminate noise. Experiments show that, compared with PCA, KPCA extracts more Wi-Fi fingerprint features and reduces the positioning error by more than 0.5 m. Second, the algorithm extracts LBP rotation-invariant unified pattern features from the scene image and stitches these two features together to form a merged fingerprint. Next, it uses LightGBM to build a regression positioning model and construct a mapping relationship between the merged fingerprint and spatial position coordinates to predict the position coordinates of the points to be measured. We chose to transform the image data into structured data to achieve fusion with Wi-Fi fingerprints in the same dimension; thus, we did not need to calculate large amounts of image data and could reduce the algorithm execution time to within 2 s.
This article compares and analyzes the selection of Wi-Fi positioning functions and positioning algorithms. The experimental results showed that the proposed LightGBM fingerprint fusion positioning algorithm exhibited less error and better environmental adaptability compared with the single-fingerprint positioning one. Compared with the traditional fingerprint positioning algorithm, the average error was reduced by 20%, and the model ran faster than other positioning algorithms. Thus, it represents a simple and effective positioning method. Compared with other similar studies, our model achieves a smaller average positioning error of 0.78 m.
This work is not just suitable for the automatic positioning of pedestrians: it can also be combined with other positioning methods to be implemented in robots. However, the data collection workload is relatively large and experimental scenarios are scarce. Future research directions include the rapid generation of fingerprint databases and the development of adaptive positioning systems.

Author Contributions

Conceptualization, H.Z. and Y.L.; methodology, Y.L.; software, Y.L.; validation, Y.L.; formal analysis, Y.L.; resources, H.Z.; data curation, Y.L.; writing—original draft preparation, Y.L.; writing—review and editing, H.Z.; visualization, Y.L.; supervision, H.Z.; project administration, H.Z.; funding acquisition, H.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Major Science and Technology Program for Water Pollution Control and Treatment of China, grant number 2018ZX07111005.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The experiment uses an internal data set. The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shu, Y.; Huang, Y.; Zhang, J.; Coué, P.; Cheng, P.; Chen, J.; Shin, K.G. Gradient-Based Fingerprinting for Indoor Localization and Tracking. IEEE Trans. Ind. Electron. 2016, 63, 2424–2433. [Google Scholar] [CrossRef]
  2. Rizk, H.; Torki, M.; Youssef, M. CellinDeep: Robust and Accurate Cellular-Based Indoor Localization via Deep Learning. IEEE Sens. J. 2019, 19, 2305–2312. [Google Scholar] [CrossRef]
  3. Sun, D.; Wei, E.; Ma, Z.; Wu, C.; Xu, S. Optimized CNNs to Indoor Localization through BLE Sensors Using Improved PSO. Sensors 2021, 21, 1995. [Google Scholar] [CrossRef] [PubMed]
  4. Liu, M.; Wang, H.; Yang, Y.; Zhang, Y.; Ma, L.; Wang, N. RFID 3-D Indoor Localization for Tag and Tag-Free Target Based on Interference. IEEE Trans. Instrum. Meas. 2019, 68, 3718–3732. [Google Scholar] [CrossRef]
  5. Segers, L.; Tiete, J.; Braeken, A.; Touhafi, A. Ultrasonic Multiple-Access Ranging System Using Spread Spectrum and MEMS Technology for Indoor Localization. Sensors 2014, 14, 3172–3187. [Google Scholar] [CrossRef]
  6. Bartosz, J.; Damian, D.; Wlodek, K. Customization of UWB 3D-RTLS Based on the New Uncertainty Model of the AoA Ranging Technique. Sensors 2017, 17, 227. [Google Scholar] [CrossRef]
  7. Krapež, P.; Vidmar, M.; Munih, M. Distance Measurements in UWB-Radio Localization Systems Corrected with a Feedforward Neural Network Model. Sensors 2021, 21, 2294. [Google Scholar] [CrossRef]
  8. Tang, J.; Chen, Y.; Jaakkola, A.; Liu, J. NAVIS-An UGV Indoor Positioning System Using Laser Scan Matching for Large-Area Real-Time Applications. Sensors 2014, 14, 11805–11824. [Google Scholar] [CrossRef]
  9. Li, B.; Muñoz, J.P.; Rong, X.; Chen, Q.; Xiao, J.; Tian, Y.; Arditi, A.; Yousuf, M. Vision-Based Mobile Indoor Assistive Navigation Aid for Blind People. IEEE Trans. Mob. Comput. 2019, 18, 702–714. [Google Scholar] [CrossRef]
  10. Lu, G.; Yan, Y.; Sebe, N.; Kambhamettu, C. Indoor Localization via Multi-View Images and Videos. Comput. Vis. Image Underst. 2017, 161. [Google Scholar] [CrossRef]
  11. Li, P.; Yang, X.; Yin, Y.; Gao, S.; Niu, Q. Smartphone-Based Indoor Localization With Integrated Fingerprint Signal. IEEE Access 2020, 8, 33178–33187. [Google Scholar] [CrossRef]
  12. Tomažič, S.; Škrjanc, I. An Automated Indoor Localization System for Online Bluetooth Signal Strength Modeling Using Visual-Inertial SLAM. Sensors 2021, 21, 2857. [Google Scholar] [CrossRef]
  13. Sy Au, A.W.; Feng, C.; Valaee, S.; Reyes, S.; Sorour, S.; Markowitz, S.N.; Gold, D.; Gordon, K.; Eizenman, M. Indoor Tracking and Navigation Using Received Signal Strength and Compressive Sensing on a Mobile Device. IEEE Trans. Mob. Comput. 2013, 12, 2050–2062. [Google Scholar] [CrossRef]
  14. Edwards, J.S. A survey on wireless indoor localization from the device perspective. Comput. Rev. 2017, 58, 109–110. [Google Scholar]
  15. He, S.; Chan, S. Wi-Fi Fingerprint-Based Indoor Positioning: Recent Advances and Comparisons. IEEE Commun. Surv. Tutor. 2017, 18, 466–490. [Google Scholar] [CrossRef]
  16. Huang, G.; Hu, Z.; Wu, J.; Xiao, H.; Zhang, F. WiFi and Vision Integrated Fingerprint for Smartphone-Based Self-Localization in Public Indoor Scenes. IEEE Internet Things J. 2020, 7, 6748–6761. [Google Scholar] [CrossRef]
  17. AL-Madani, B.; Orujov, F.; Maskeliūnas, R.; Damaševičius, R.; Venčkauskas, A. Fuzzy Logic Type-2 Based Wireless Indoor Localization System for Navigation of Visually Impaired People in Buildings. Sensors 2019, 19, 2114. [Google Scholar] [CrossRef]
  18. Zeng, Q.; Wang, J.; Meng, Q.; Zhang, X.; Zeng, S. Seamless Pedestrian Navigation Methodology Optimized for Indoor/Outdoor Detection. IEEE Sens. J. 2018, 18, 363–374. [Google Scholar] [CrossRef]
  19. Dong, J.; Noreikis, M.; Xiao, Y.; Ylä-Jääski, A. ViNav: A Vision-Based Indoor Navigation System for Smartphones. IEEE Trans. Mob. Comput. 2019, 18, 1461–1475. [Google Scholar] [CrossRef]
  20. Lu, J.; Chen, K.; Li, B.; Dai, M. Hybrid Navigation Method of INS/PDR Based on Action Recognition. IEEE Sens. J. 2018, 18, 8541–8548. [Google Scholar] [CrossRef]
  21. Zayets, A.; Steinbach, E. Robust WiFi-based indoor localization using multipath component analysis. In Proceedings of the International Conference on Indoor Positioning & Indoor Navigation, Sapporo, Japan, 18–21 September 2017; IEEE: Sapporo, Japan, 2017; pp. 1–8. [Google Scholar]
  22. Mathivannan, S.; Srinath, S.; Shashank, R.; Aravindh, R.; Balasubramanian, V. A Dynamic Weighted Trilateration Algorithm for Indoor Localization Using Dual-Band WiFi. In Proceedings of the W2GIS: International Symposium on Web and Wireless Geographical Information Systems, Kyoto, Japan, 16–17 May 2019; Springer: Kyoto, Japan, 2019. [Google Scholar]
  23. Husen, M.N.; Lee, S. Indoor Human Localization with Orientation using WiFi Fingerprinting. In Proceedings of the ACM 8th International Conference on Ubiquitous Information Management and Communication (ACM IMCOM), Siem Reap, Cambodia, 9–14 January 2014. [Google Scholar]
  24. Guo, X.; Li, L.; Ansari, N.; Liao, B. Accurate WiFi Localization by Fusing a Group of Fingerprints via Global Fusion Profile. IEEE Trans. Veh. Technol. 2018, 7314–7325. [Google Scholar] [CrossRef]
  25. He, S.; Chan, S. INTRI: Contour-Based Trilateration for Indoor Fingerprint-Based Localization. IEEE Trans. Mob. Comput. 2017, 16, 1676–1690. [Google Scholar] [CrossRef]
  26. Brumitt, B.; Meyers, B.; Krumm, J.; Kern, A.; Shafer, S.A. EasyLiving: Technologies for Intelligent Environments. In Proceedings of the Handheld and Ubiquitous Computing, Second International Symposium, HUC 2000, Bristol, UK, 25–27 September 2000. [Google Scholar]
  27. Vedadi, F.; Valaee, S. Automatic Visual Fingerprinting for Indoor Image-Based Localization Applications. IEEE Trans. Syst. Man Cybern. Syst. 2017, 1–13. [Google Scholar] [CrossRef]
  28. Walch, F.; Hazirbas, C.; Leal-Taixe, L.; Sattler, T.; Cremers, D. Image-based localization using LSTMs for structured feature correlation. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  29. Papaioannou, S.; Wen, H.; Markham, A.; Trigoni, N. Fusion of Radio and Camera Sensor Data for Accurate Indoor Positioning. In Proceedings of the 2014 IEEE 11th International Conference on Mobile Ad Hoc and Sensor Systems, Philadelphia, PA, USA, 28–30 October 2014; pp. 109–117. [Google Scholar]
  30. Jiao, J.; Deng, Z.; Xu, L.; Li, F. A Hybrid of Smartphone Camera and Basestation Wide-area Indoor Positioning Method. Ksii Trans. Internet Inf. Syst. 2016, 10. [Google Scholar] [CrossRef]
  31. Ruiz-Ruiz, A.J.; Lopez-de-Teruel, P.E.; Canovas, O. A multisensor LBS using SIFT-based 3D models. In Proceedings of the 2012 International Conference on Indoor Positioning and Indoor Navigation (IPIN), Sydney, Australia, 13–15 November 2012; pp. 1–10. [Google Scholar]
  32. Alahi, A.; Haque, A.; Li, F.F. RGB-W: When Vision Meets Wireless. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015. [Google Scholar]
  33. Jia, J.; Wang, X.; Deng, Z. Build a Robust Learning Feature Descriptor by Using a New Image Visualization Method for Indoor Scenario Recognition. Sensors 2017, 17, 1569. [Google Scholar] [CrossRef] [PubMed]
  34. Hu, Z.; Gang, H.; Hu, Y.; Zhe, Y. Wi-Vi Fingerprint: WiFi and Vision Integrated Fingerprint for Smartphone-based Indoor Self-Localization. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP 2017), Beijing, China, 17–20 September 2017. [Google Scholar]
  35. Redzic, M.; Laoudias, C.; Kyriakides, I. Image and WLAN Bimodal Integration for Indoor User Localization. IEEE Trans. Mob. Comput. 2019, 19, 1109–1122. [Google Scholar] [CrossRef]
  36. Jiao, J.; Deng, Z.; Arain, Q.A.; Li, F. Smart Fusion of Multi-sensor Ubiquitous Signals of Mobile Device for Localization in GNSS-Denied Scenarios. Wirel. Pers. Commun. 2018, 116, 1507–1523. [Google Scholar] [CrossRef]
  37. Ojala, T.; Pietikainen, M.; Maenpaa, T. Multiresolution Gray-Scale and Rotation Invariant Texture Classification with Local Binary Patterns. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 971–987. [Google Scholar] [CrossRef]
  38. Xia, S.; Liu, Y.; Yuan, G.; Zhu, M.; Wang, Z. Indoor Fingerprint Positioning Based on Wi-Fi: An Overview. ISPRS Int. J. Geo-Inf. 2017, 6, 135. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.