You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

20 January 2022

A Semi-Supervised 3D Indoor Localization Using Multi-Kernel Learning for WiFi Networks

,
and
1
Department of Computer Science and Information Engineering, National Taipei University, No. 151, University Rd., San Shia District, New Taipei City 237, Taiwan
2
Department of Information Management, Shih Hsin University, No. 1, Ln. 17, Sec. 1, Muzha Rd., Wenshan District, Taipei City 116, Taiwan
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Multi-Radio and/or Multi-Sensor Integrated Navigation System

Abstract

Indoor localization is an important issue for indoor location-based services. As opposed to the other indoor localization approaches, the radio frequency (RF) based approaches are low-energy solutions with simple implementation. The kernel learning has been used for the RF-based indoor localization in 2D environment. However, the kernel learning has not been used in 3D environment. Hence, this paper proposes a multi-kernel learning scheme for 3D indoor localization. Based on the signals collected in the area of interest, the WiFi signals with better quality and closer to the user are selected so as to reduce the multipath effect and the external interference. Through the construction of multi-kernel, the localization accuracy can be improved as opposed to the localization based on the single kernel. We build multiple kernels to get the user’s location by collecting wireless received signal strengths (RSS) and signal-to-noise ratios (SNR). The kernel learning maps data to high dimension space and uses the optimization process to find the surface where the data are mapped. By multi-kernel training, the surface is fine-tuned and eventually converges to form the location database during the mapping process. The proposed localization scheme is verified by the real RSS and SNR collected from multiple wireless access points (AP) in a building. The experimental results verify that the proposed multi-kernel learning scheme performs better than the multi-DNN scheme and the existing kernel-based localization schemes in terms of localization accuracy and error in 3D indoor environment.

1. Introduction

Recently, the location-based service (LBS) for internet of things (IoT) becomes popular in many fields such as office building, shopping mall, and community building. Since the GPS system cannot catch satellite signals in an indoor environment, many indoor localization systems based on different RF signals have been presented, such as RFID [1,2,3], Bluetooth, and LoRa [4] based localization systems.
With the popularity of wireless LANs and wireless access devices, WiFi-based indoor localization has become more and more popular in recent years. Based on the WiFi localization systems, fingerprint schemes [5,6,7,8,9,10] show great advantages and accuracy in indoor localization with RSS and SNR signals. In the offline phase, select multiple physical locations in the region of interest and the signals received from APs built in indoor environment are defined as labeled data in these selected locations. Therefore, the obtained position and the corresponding labeled data are defined as the fingerprint database. In the training phase, the labeled data and the corresponding RSS and SNR are obtained. In the online phase, the RSS and SNR received by devices determine the specific location based on the radio map. Consequently, this kind of scheme shows high potential in indoor localization.
A few 3D indoor localization schemes have been proposed. Qi et al. [11] proposed a hybrid RSS-AOA measurement in non-cooperative and cooperative 3D wireless sensor networks. By using approximate error expressions for the RSS and angle of arrival (AOA) measurement models, new estimation based on the proposed method can be obtained. Cramariuc et al. [12] proposed a Penalized Logarithmic Gaussian Distance metric, which can improve the comparability of different clustering schemes in 3D positioning.
The existing localization schemes either need additional hardware and measurements or need to collect numerous samples for learning or training. To reduce the costs of hardware, collecting, and measurements, a novel semi-supervised localization scheme is proposed, which only needs to collect a few labeled signals of RSS and SNR and hence reduces the costs (only 25 % of the collected signals are labeled). However, the accuracy of the existing localization schemes using only a few labeled signals of RSS and SNR are not satisfactory. To improve these schemes, the multi-kernel learning scheme is proposed and adopted in our semi-supervised localization scheme so as to improve the localization accuracy and alleviate the multipath effect and signal deviating problem. The kernel model is adopted because the kernel model can learn the distribution of data according to the kernel function selected for the problem, and judge the result according to the distribution of the data. During localization, the kernel model can learn the distribution of signals in the positioning area, and determine where the user may be based on the signal distribution.
The main contributions of the proposed approach are summarized as follow:
  • A semi-supervised localization scheme, which only needs to collect a few labeled signals of RSS and SNR, is proposed in this paper.
  • A multi-kernel learning scheme with weight adjustment and optimization for 3D indoor localization in WiFi networks is proposed in this paper so as to further improve the accuracy of localization.
  • Experimental results demonstrate that the proposed localization scheme performs better than the multi-DNN scheme and the existing kernel-based localization schemes in terms of localization accuracy and error.
The rest of this paper is organized as follows. Section 2 describes the related works. Section 3 describes the system model and the problem formulation of the proposed scheme. Section 4 describes the proposed semi-supervised multi-kernel learning scheme for 3D indoor localization. The experimental results are shown in Section 5, and the conclusions are summarized in Section 6.

3. Preliminaries

This section introduces the system architecture, problem formulation, and the comparison with the single kernel scheme.

3.1. System Architecture

The WiFi 3D indoor localization architecture is shown in Figure 1. There are many APs deployed in the 3D indoor environment with multiple floors. The mobile devices are placed in the localization area. The mobile device can retrieve the RSSI, SNR and the corresponding MAC address within the transmission range of the AP as the identifier. Let D denote the data which is retrieved by a mobile device on the location . D can be represented as D = ( ( r 1 , s 1 , m 1 ) , ( r 2 , s 2 , m 2 ) , , ( r n , s n , m n ) , ) , where r , s , m denote RSSI, SNR and MAC address, respectively, n is the number of APs which can be heard by the mobile device within the transmission range. All the data are saved in MongoDB for the training phase.
Figure 1. System architecture.
When the dataset D is collected, the next step is to categorize the dataset into two categories, labeled data and unlabeled data based on whether or not the data are collected at known locations. In the training phase, each labeled data from the source domain is correspond to the true location . In order to find the mapping relationship between the labeled data and unlabeled data, the kernel optimization learning is used to form the location database. Each kernel learns the distribution of its own region. In the localization stage, the signal distribution obtained by the mobile device is sent to the server. The server then calculates the expected value of the related kernels and obtains the estimated location. Note that, the training, learning, and location estimation are all performed in the server.
In the proposed scheme, an anchor point located at the corner of the building, whose latitude, longitude, and altitude have been obtained through GPS, is regarded as the point of origin (0, 0, 0) in our coordinate system. The Cartesian coordinate system is adopted in the proposed scheme and a meter is regarded as the basic unit in our coordinate system. We get the estimated relative coordinate (x, y, z) of the target object in our coordinate system first and then we use the anchor point’s position to derive the target object’s estimated location. The reason the GPS coordinates are used because we can use the relative coordinate of the target object to get the possible latitude, longitude, and altitude of the target object. We get the relative coordinate of the labeled data by combining the blueprint of the building and a distance measuring instrument (e.g., laser ranger or ruler). The measuring error is less than 10 cm. To get the relative coordinates of all the labeled nodes indeed consumes a lot of manpower. Hence, we propose a semi-supervised localization scheme.
Figure 2 shows the difference between the RBF kernel and the chi-squared kernel. The data are mapped to high dimension by RBF kernel and chi-squared kernel. In RBF kernel, the signal obtained by users with short distance movement is not obviously different after mapped to high-dimension by RBF kernel which leads to higher localization error as opposed to the chi-squared kernel. Hence, the chi-squared kernel is adopted in the proposed multi-kernel learning scheme for 3D indoor localization.
Figure 2. The comparison with RBF kernel and chi-squared kernel.

3.2. Problem Formulation

Transfer kernel learning [15,17] has been studied widely in recent years and has been applied successfully in indoor localization. However, when the location field is too large resulting in lots of data groups, the accuracy of transfer kernel learning is decreased and becomes unsuitable. Therefore, the multi-kernel optimization learning method is proposed to solve the above problem.
Assume that p APs are installed in an indoor environment, and the signals received by mobile devices from these APs in three dimension position can be represented by a matrix X, which includes q samples. The received signal x i = ( r i , s i ) represents the i-th RSSI and SNR, respectively. X can be represented as X = [ x 1 , x 2 , , x q ] .
The signal collected at the offline calibration is named as the source domain, which is denoted as ( x i s , l i s ) . The signal collected in the online localization is named as the target domain and is denoted as ( x i T ) , which is the data needed to be estimated its location. The goal of the proposed scheme is to estimate the location ( T ) of mobile device by constructing a localization model using multi-kernel optimization learning.
Multi-kernel optimization learning improves the accuracy based on labeled data. There are two domains D S and D T , where D S = { x i S , i S } i = 1 q indicates the data in the source domain and the corresponding labeled data i , and the target domain D T = { x i T } i = 1 n indicates the unlabeled data. In order to find the relationship between the source domain and the target domain, the cross domain is used. The cross domain is the intersection with the source domain and the target domain and can be represented as D S T D S D T . By using the cross domain data, the extrapolated kernel is built. The detailed methodology of building extrapolated kernel is introduced in the next section. During the training phase, in order to minimize the difference between the source domain and the target domain, the squared loss optimization problem can be used, which is shown as follows:
arg min * Λ K ¯ M K M = arg min * Φ ¯ S i Λ T i Φ S i T K M . s u b j e c t t o Λ T = d i a g { λ 1 , λ 2 , , λ l } λ T Q i λ 2 r i T λ + s i 0 , i Z + λ i 0
where K M denotes the source domain D S mapped to high dimension through the kernel function and K ¯ M represents the hyperplane found by Mercer’s theorem, Φ ¯ M is the eigenvector matrix of K ¯ M and Φ ¯ M T is the transpose matrix of Φ ¯ M . The restriction condition is defined as the quadratic constrained quadratic programming (QCQP) optimization problem.
The QCQP is an optimization problem in which both the objective function and the constraints are quadratic functions. It has the following form.
min ( 1 2 x T P 0 x + q 0 T x ) . s u b j e c t t o 1 2 x T P i x + q i T x + r i 0 , i = 1 , 2 , , m A x = b
where P 0 , P 1 , , P m are n-by-n matrices, q 0 , q 1 , , q m are positive semidefinite matrices and x R n is the optimization variable.
The variable Λ is defined as the eigenvalue in the training phase. The purpose is to compute the domain-invariant kernel K A and estimate the location of mobile device. The goal of the above equation is to minimize the difference of the domain location error between the source and target domain in raw data.

3.3. Comparison with the Single Kernel Scheme

Figure 3 shows the difference between the single kernel learning and the multi-kernel learning. The figure is divided into two parts. The upper part represents the single kernel learning; the bottom part is the proposed multi-kernel learning. The proposed scheme is designed for 3D indoor environment compared with the single kernel learning which is designed for 2D plane. By multi-kernel learning, each kernel learns the distribution of own region. According to the signal distribution obtained by mobile device, the expected kernel is the estimated location. The expected kernel is derived from multiple kernels during the multi-kernel optimization phase, which is shown in Figure 4. By using QCQP optimization, each kernel learns the distribution of its own region and fuse all the models to get the distribution of the area of interest.
Figure 3. A comparison with (a) single kernel learning based on SVM and (b) the multi-kernel based semi-supervised learning.
Figure 4. The multi-kernel learning.

4. A 3D Multi-Kernel Learning Approach for 3D Indoor Localization

This section proposes a 3D indoor localization scheme based on semi-supervised multi-kernel optimization learning in WiFi network for predicting the unlabeled target with cross-domain data in 3D indoor environment. The notations are defined in Table 1. The system architecture is shown in Figure 5 with six phases. The data acquisition is shown in Figure 4 and divided into the source domain data D S and the target domain data D T , which represent the labeled data and the unlabeled data, respectively. The six phases of the proposed scheme are briefly described as follows:
Table 1. Notation table.
Figure 5. The 3D iterative multi-kernel optimization.
(1)
3D data collection phase: During the data collection stage, the data collected in the 3D environment is divided into D S , D T and D S T . The D S indicates that the data are in source domain and D T indicates the unlabeled data. D S T is the intersection of D S and D T and can be represented as D S T D S D T . The collected data are used to choose multiple APs in the next phase.
(2)
Multi-AP selection phase: In order to learn the signal distribution generated by each AP, the multi-AP selection phase is proposed. In this phase, according to the signal obtained by the user at the area of interest, select multiple APs near the user. The selected APs form the kernels in the next phase.
(3)
3D Multi-kernel construction phase: In order to find the corresponding signal plane between multiple APs, the multi-kernel construction phase is proposed to solve this problem. In this phase, the selected APs are used to form the kernel model. The data, include D S , D T and D S T , which located in the area of the intersection between multiple APs is input to the kernel function to form the high-dimensional data. By Mercer’s theorem, the plane between multiple APs is built. The plane is adjusted by the next phase.
(4)
Weight adjustment phase: After constructing multiple kernels, the kernel closer to the user is more accurate, therefore, the weight adjustment phase is proposed to determine the weight of each kernel. In this phase, the distance and the SNR between the kernel and the user is measured. If the kernel is closer to the user, the kernel gains a larger weight. The weight of the kernel is optimized in the next phase.
(5)
The multi-kernel optimization phase: After giving the weight to each kernel, the area responsible for each kernel is combined according to the weight to form an expected kernel K ¯ S . In order to reduce the distribution between the source kernel K S and K ¯ S , the multi-kernel optimization phase is proposed to solve this problem.
(6)
Location estimation phase: After multi-kernel optimization phase, the expected kernel K ¯ A is constructed and used for location estimation.

4.1. 3D data Collection Phase

The data collection in the 3D environment is shown in Figure 6. The data are collected at a multi-floor building. Each floor is divided into several cubes and each cube collect training data, which includes RSS, SNR and the corresponding MAC address. The detail of the data collection phase is described as follow:
Figure 6. The 3D Data collection phase.
S1.
The data collected in each cube is separated into labeled data ( D S ), unlabeled data ( D T ) and the cross-domain data ( D S T ). D S = { r t , s t , a i , j , i } t = 1 p denotes the labeled data which is collected at the eight corners of the cube with R S S t , S N R t and the signal collected at the j-th AP on the i-th floor and the corresponding labeled data t , respectively.
S2.
D T = { r t , s t , a i , j } t = 1 q denotes the unlabeled data which is collected randomly at each cube with R S S t , S N R t and the signal collected at the j-th AP on the i-th floor, respectively.
S3.
D S T = { r t , s t , a i , j , t } t = 1 r indicates the cross-domain data with R S S t , S N R t and the signal collected at the j-th AP on the i-th floor, which is the intersection of D S and D T and can be represented as D S T D S D T .

4.2. Multi-AP Selection Phase

The multi-AP selection phase is shown in Figure 7. Inspired by the existing works [12,20], the ideas of filtering and clustering are adopted in the multi-AP selection phase so as to reduce the training cost and improve the accuracy of localization. In the localization area, the features of the signal including r i and s i often have the data inaccuracy and offset problems which have a considerable impact on the subsequent training of the entire data set. Before multi-AP selection phase, the data with too high deviation value are filtered. The collected data are first sorted according to the deviation of the data in an ascending order. The first k data with the lowest deviation is keep the other data are filtered, where k is the number of data set candidates. The mean values of r i and s i are computed after removing the outliers. The remaining data are exported for the next step.
Figure 7. The multi-AP selection phase.
The user receives signals at the area of interest. After removing the outliers, the received signal is arranged in descending order according to the value of SNR. Choose the first n signals to form the kernel model. The details of the multi-AP selection phase is described as follow:
S1.
The collected signals are arranged in descending order according to the value of S N R and stored into Q. Q can be represented as Q = { [ r 1 , s 1 , a 1 ] , [ r 2 , s 2 , a 2 ] , [ r 3 , s 3 , a 3 ] , , [ r k , s k , a k ] } , where k is the number of data set candidates which can be collected at the reference point.
S2.
According to the signals stored in Q, choose the first n signals from Q, where Q = { [ r ^ 1 , s ^ 1 , a ^ 1 ] , [ r ^ 2 , s ^ 2 , a ^ 2 ] , , [ r ^ n , s ^ n , a ^ n ] } is the n best signals which can be heard at the reference point. The intersection between each AP’s transmission range have different signal distribution. In order to learn the signal distribution generated by each AP, each element in Q is permutation to select multiple APs whose signal is within the n best signals and stored in cluster C.
S3.
The cluster C can be represented as C = { [ a ^ 1 , a ^ 2 ] , C 2 n , [ a ^ 1 , a ^ 2 , a ^ 3 ] , C 3 n , , [ a ^ 1 , a ^ 2 , , ] , , a ^ m C m n } , where m represents the number of AP groups. In the following phase, each component of the cluster is operated separately to form the kernel.

4.3. Multi-Kernel Construction Phase

In the multi-kernel construction phase, in order to form the kernel, the data, including D S , D T and D S T , is mapped to high dimension by kernel function. The following lemma proves that the data do not disappear when it is mapped to high dimension and can form the kernel space by Mercer’s theorem.
Mercer’s Theorem 
[21]:If k ( x , y ) is a kernel function, then k ( x , y ) must satisfy the Mercer’s condition, that is i = 1 n j = 1 n k ( x i , y i ) 0 , x y R .
Lemma 1.
The RBF kernel is K R ( x , y ) = e x p ( d ( x , y ) 2 2 l 2 ) , where l > 0 is the parameter and d ( x , y ) is the Euclidean distance between x and y in high dimension space. The chi-squared kernel is K C = e x p ( γ i ( x [ i ] y [ i ] ) 2 x [ i ] + y [ i ] ) , where γ [ 0 , 1 ] is the parameter. Let data { [ r 1 , s 1 , a 1 ] , [ r 2 , s 2 , a 2 ] , , [ r m , s m , a m ] , [ r m + 1 , s m + 1 ] , [ r m + 2 , s m + 2 ] , , [ r n , s n ] , [ r n + 1 , s n + 1 , a n + 1 ] [ r n + 2 , s n + 2 , a n + 2 ] , , [ r k , s k , a k ] } denote the D S , D T , and D S T , respectively, where D S R m , D T R n m and D S T R k n . The D S , D T and D S T can be mapped to high dimension by RBF kernel to form the kernel and can also be mapped by chi-squared kernel to form the kernel.
Proof 
According to the definition of kernel function, there is a function that satisfies k ( x , y ) = Φ ( x ) , Φ ( y ) for all data, the k ( x , y ) is a kernel function. By Mercer’s theorem, if k ( x , y ) is a kernel function, then k ( x , y ) must satisfy the Mercer’s condition, that is i = 1 n j = 1 n k ( x i , y j ) 0 , x y R . The RBF kernel is K R ( x , y ) = e x p ( d ( x , y ) 2 2 l 2 ) , by Taylor series e x = 1 + n = 1 1 + x + x 2 2 ! + x 3 3 ! + + x n n ! = lim x ( 1 + x n ) n 2 . 7 x . The RBF kernel can be rewrite as K R ( x , y ) = 2 . 7 d ( x , y ) 2 2 l 2 , where K R ( x , y ) > 0 , x y R , and satisfies the Mercer’s condition. If K R ( x , y ) > 0 , x y R , this condition means that the data do not disappear x y R when it is mapped to high dimension and can form the kernel. The chi-squared kernel is K C = e x p ( γ i ( x [ i ] y [ i ] ) 2 x [ i ] + y [ i ] ) and can be rewrite as K C = 2 . 7 γ i ( x [ i ] y [ i ] ) 2 x [ i ] + y [ i ] . Under the condition of chi-squared kernel, the K c ( x , y ) > 0 , x y R + and satisfies the Mercer’s condition. If K c ( x , y ) > 0 , x y R + , this condition means that the data do not disappear x y R + when it is mapped to high dimension and can form the kernel space. Because K R ( x , y ) > 0 and K c ( x , y ) > 0 , the D S , D T and D S T can be mapped to high dimensions by RBF kernel to form the kernel, and can also be mapped by chi-squared kernel to form the kernel without lost data.    □
The kernel learning represents the domain-invariant kernel by matching the source distribution and the target distribution to generate the kernel Hilbert space. In multi-AP selection phase, each group of the cluster’s data, including D S , D T and D S T , represent the intersection of the area formed by multiple APs. For example, if the intersection area is formed by two APs a ¯ 1 and a ¯ 2 , the data must have signal heard by a ¯ 1 , a ¯ 2 . Each group of cluster’s data is operated separately to form the kernel in this phase. The detailed of multi-kernel construction phase is shown as Figure 8 and is described as follow:
Figure 8. The multi-kernel construction phase.
S1.
The first step is to calculate the source kernel K S and the target kernel K T using the kernel function, such as chi-squared kernel K ( x , y ) = e x p ( γ i ( x [ i ] y [ i ] ) 2 x [ i ] + y [ i ] ) . The cross-domain kernel K S T , from the common area between K S and K T , are also using the chi-squared kernel.
S2.
In order to evaluate the distribution difference between K S and K T in the space of Hilbert, the second step is to compute the difference of distribution between the source kernel and the target kernel. However, since K S and K T have distinct dimensions, that is K ¯ S R l S × l S and K T R l T × l T , the difference between K S and K T cannot be directly estimated. In order to solve the problem of different dimensions, an extrapolated kernel K ¯ S R l S × l S is constructed using the eigensystem { Φ T , λ T } . The target eigenvector matrix Φ T and the target eigenvalue matrix λ T can be constructed using the standard problem of eigenvalue K T Φ T = Φ T λ T .
S3.
After calculated eigensystem and the cross-domain kernel K S T , the extrapolated source kernel K S of the eigenvector matrix is calculated by the Mercer’s theorem as Φ S ¯ K S T Φ T λ T 1 .
S4.
After S3, the new source kernel K ¯ S is constructed from the eigenvectors of the target kernel, where K ¯ S = Φ ¯ S λ T Φ S 1 . In the next phase, each kernel is given a weight according to the distance and the SNR value between the user and the kernel.
The pseudo code of the multi-kernel construction phase is shown in Algorithm 1.
Algorithm 1: The multi-kernel construction phase
Input: 
The dataset of D S i , D T i and D S T i derived in the multi-AP selection phase
D S i = [ r ^ 1 , s ^ 1 , a ^ 1 , 1 ] , [ r ^ 2 , s ^ 2 , a ^ 2 , 2 ] , , [ r m , s m , a ^ m , m ] i
D T i = [ r ^ 1 , s ^ 1 , a ^ 1 ] , [ r ^ 2 , s ^ 2 , a ^ 2 ] , , [ r ^ n , s ^ n , a ^ n ] i
D S T i = [ r ^ 1 , s ^ 1 , a ^ 1 , 1 ] , [ r ^ 2 , s ^ 2 , a ^ 2 , 2 ] , , [ r ^ d , s ^ d , a ^ d , d ] i
Output: 
The multiple domain-invariant kernel K = { K ¯ S 1 , K ¯ S 2 , , K ¯ S u } , each K ¯ S i represent the kernel model form by D S i , D T i and D S T i , where u = j = 3 m C j n
  • K=
  • for j = 3 to mdo
Sensors 22 00776 i001

4.4. The Weight Adjustment Phase

The weight adjustment phase is shown in Figure 9. After constructing multiple kernels, the kernel closer to the user have higher accuracy. In order to find the relationship between the kernel and the user, the distance is measured between the kernel and the user. If the kernel is closer to the user, the kernel has a larger weight. According to the signal that user received, the distance between each AP can be calculated by the signal strength, the detailed of the weight adjustment phase is described as follow:
Figure 9. The weight adjustment phase.
S1.
The distance between the user and AP can be calculated by d ( r , a ^ ) = 10 P r 10 n , where P is the measured power from a ^ , n is the environment factor.
S2.
The distance between the user and the kernel can be calculated by f d ( R , k ) , where R represent the corresponding RSS to form the kernel k. For example, if kernel k is formed by a ^ 1 , a ^ 2 , a ^ 3 , then f d ( R , k ) = d ( r 1 , a ^ 1 ) + d ( r 2 , a ^ 2 ) + d ( r 3 , a ^ 3 ) 3 , where r 1 , r 2 and r 3 denotes the RSS that the user received from a ^ 1 , a ^ 2 and a ^ 3 , respectively.
S3.
The SNR signal strength can be calculated by w t =   w t = 0 , t = 1 w t 1 + f s ( S , K ) T m i = 1 n S i t , t > 1 , where T denotes the length of time, f s ( S , K ) is the sum of the corresponding SNR signal to form the kernel model k, and m represents the number of AP. For example, if kernel model k is formed by a ¯ 1 , a ¯ 2 , a ¯ 3 , then w t = w t 1 + s 1 t + s 2 t + s 3 t T m i = 1 n S i t , where s 1 t , s 2 t and s 3 t denotes the SNR that the user received from a ^ 1 , a ^ 2 and a ^ 3 .
S4.
The weight of each kernel can be determined by α w t + ( 1 α ) f ( R , k i ) 1 i = 1 C m n f ( R , k i ) 1 , where α [ 0 , 1 ] . If the kernel is closer to the user, the kernel has a larger weight and if the kernel is farther to the user, the kernel has a smaller weight. In order to formalize the distribution discrepancy between the new source kernel K ¯ S and the ground truth source kernel K S , the multi-kernel optimization is used in the next phase to solve this problem.
The pseudo code of the weight adjustment phase is shown in Algorithm 2.
Algorithm 2: The weight adjustment phase
Input: 
The multiple domain-invariant kernel K = { K ¯ S 1 , K ¯ S 2 , , K ¯ S u } construct by Algorithm 1
Output: 
The weight of each domain-invariant kernel P = { β 1 K ¯ S 1 , β 2 K ¯ S 2 , , β u K ¯ S u }
  • P=
  • foreachkernel K ¯ S i in Kdo
Sensors 22 00776 i002

4.5. The Multi-Kernel Optimization Learning Phase

After the weight adjustment phase, each kernel model has its own weight. In multi-kernel optimization phase, each kernel is adjusted by the QCQP optimization as shown in Figure 10. K C is the source domain and K ¯ C is constructed by K C using Mercer’s theorem, where C = S 1 , S 2 , , S u . Each kernel can be represented as the distribution of the localization area, and is combined according to the weight to form a combined kernel K ¯ S , where K ¯ S = β 1 K ¯ S 1 + β 2 K ¯ S 2 + + β u K ¯ S u . The K S is the source domain combined with the weight which is denoted as K S = β 1 K S 1 + β 2 K S 2 + + β u K S u . In order to minimize the difference between K ¯ S and K S , the QCQP optimization is used. The detail of the multi-kernel optimization phase is described as follow:
Figure 10. The multi-kernel optimization phase.
S1.
In order to minimize the difference between K S and K ¯ S , the QCQP optimization is used which is described as min λ i λ ( K ¯ S K S ) 2 , where λ = { λ 1 , λ 2 , , λ n } are the n nonnegative eigen-spectrum parameters.
S2.
The QCQP optimization is used to derive the intersection between K S and K ¯ S in the high dimension. The crossover point in K S K ¯ S is λ found by QCQP optimization, where i 1 , 2 , , n .
S3.
The goal of the multi-kernel optimization phase is to find a λ i that can minimize the difference between K S and K ¯ S . In this step, K C is calculated by K C = Φ C λ i Φ C T , where Φ C is the eigenvector matrix and λ i is the eigenvalue matrix. K ¯ A is constructed with λ i which has the minimized distribution of K S K ¯ S .

4.6. Location Estimation Phase

After the multi-kernel optimization learning, λ i is found and forms K ¯ A with the minimized difference between K S and K ¯ S . Finally, K ¯ A is used to estimate the location with the signal received by the user.
S1.
After K ¯ A is constructed in multi-kernel optimization learning, the user’s location t can be estimated by K ¯ A according to the signal received by user.
S2.
The user’s location t can be denoted as t = K ¯ A ( { [ r 1 , s 1 , a 1 ] , [ r 2 , s 2 , a 2 ] , , [ r n , s n , a n ] } ) , where { [ r 1 , s 1 , a 1 ] , [ r 2 , s 2 , a 2 ] , , [ r n , s n , a n ] } represents the n best signals received by the user.

5. Experiment Results

The experiments are performed in the Elytone building of National Taipei University. 8 APs are deployed in a 20 m × 20 m plane with the height of 2 m to construct the multiple kernels on each floor. In total, there are 24 APs deployed in three consecutive floors. User can receive signals from APs on the upper and lower floors. The cube size is 1 m × 1 m × 1 m. More than 1000 samples are collected at each cube. After filtering, 800 samples with lower deviation are used for training. The proportion of the labeled data to unlabeled data and cross-domain data is 1:2:1. Hence, the proportion of the labeled data is 25%. To verify the efficiency of the proposed scheme, seven schemes, namely, the proposed multi-kernel scheme with chi-squared as the kernel function (denoted as CMK) and RBF as the kernel function (denoted as RMK), the single-kernel scheme with chi-squared as the kernel function (denoted as CSK) and RBF as the kernel function (denoted as RSK), the hybrid-kernel scheme with the hybrid of RBF and chi-squared (denoted as RBF + chi-squared) and the hybrid of RBF and polynomial (denoted as RBF + polynomial), and the scheme using multiple DNN models (denoted as multi-DNN) are experimented and compared in this paper. The performance metrics to be observed are defined as follows.
  • Localization error: The average difference between the real location and the estimated location.
  • Localization accuracy: The ratio that the estimated location is located at the same cube as the real location.
  • Learning time: The time it takes to form each kernel.

5.1. Localization Error

The experiment results of the localization error are shown in Figure 11. As the epoch increases the localization error decreases. The localization error of the CMK scheme is the lowest, followed by the multi-DNN, RMK, RBF + chi-squared, CSK, RBF+polynomial, and the RSK schemes. Overall, the schemes with chi-squared as the kernel function performs better than the schemes with RBF as the kernel function and the proposed multi-kernel schemes performs better than the hybrid-kernel and single-kernel schemes because we create multiple kernels and try to get the optimal expected kernel and thus the proposed multi-kernel scheme can reduce the localization error and enhance the localization accuracy.
Figure 11. (a) Localization error vs. epoch.(b) Localization error vs. iteration. (c) Localization error vs. number of selected APs with epoch 80. (d) CDF vs. localization error.
The experiment results of the localization error under each iterations and epoch are shown in Figure 11a,b. There are two parameters in the experimental results, the C is the regularization parameter. The intensity of normalization is inversely proportional to C. ζ is the error caused by how much data are allowed. The greater the value is, the greater the error is allowed. In the single kernel part, the chi-squared single kernel (CSK) scheme is better than the RBF single kernel (RSK) scheme, because of the different distribution in the RBF kernel and chi-squared kernel. The distribution of the RBF kernel and chi-squared kernel are shown in Figure 8. The color with dark part is the place where the data are gathered. In the RBF kernel, the place where the data are gathered is more separated than the chi-squared kernel according to the similarity of the signal received by data collection in 3D environment, which leads to higher localization error compared with the chi-squared kernel. In the multi-kernel part, the hybrid kernel (RBF+polynomial, RBF+chi-squared) schemes generally are better than the single kernel schemes, in order to reduce the localization error, the 3D multi-kernel learning scheme is proposed which improves the shortcomings of the singe kernel and hybrid kernel schemes. The results show that the chi-squared multi-kernel (CMK) scheme is better than the RBF multi-kernel (RMK) and the single kernel schemes in the 3D localization environment. In multi-DNN, by creating multiple DNN models to cover the localization area, the location of the user can be calculated by using probability that the phone is located at a given location. Since multi-DNN needs to adjust the weight of each neuron, the convergence speed is slower than that of the multi-kernel scheme. Figure 11c presents the localization error with number of APs which are selected in the multi-AP selection phase. Basically, the more APs are chosen, the lower localization error will be. However, if too many APs are selected, the complexity of the 3D multi-kernel construction will increase. Figure 11d presents the CDF of the localization error.

5.2. Localization Accuracy

The experiment results of the localization accuracy are shown in Figure 12. As the epoch increases the localization accuracy also increases. The localization accuracy of the CMK scheme is the highest, followed by the multi-DNN, RMK, RBF+chi-squared, CSK, RBF+polynomial, and the RSK schemes.
Figure 12. (a) Localization accuracy vs. epoch. (b) Localization accuracy vs. iteration. (c) Localization error vs number of selected APs with epoch 80.
The experiment results of the localization accuracy under each iterations and epoch are shown in Figure 12a,b. In the single kernel part, the accuracy of CSK is better than RSK because of the different distribution in RBF kernel and chi-squared kernel in Figure 8. In the multi-kernel part, the accuracy of CMK and hybrid kernel is also better than RMK.

5.3. Time Cost

The experiment results of the learning time under different number of APs are shown in Figure 13. As the number of APs increases, the time to form the kernel models also increases. The learning time of RSK and CSK schemes is the lowest because the single-kernel schemes only need to form a single kernel. The hybrid kernel schemes take more time than RSK and CSK, because the hybrid kernel schemes need to form two kernels. The learning time of RMK and CMK schemes is the highest because the multi-kernel schemes need to form multiple kernels.
Figure 13. Time cost vs. number of selected APs.
The experiment results of time cost under each number of APs are shown in Figure 13. Basically, if too many APs are selected, the time to form kernel model will increase. In the single kernel part, the time cost of RSK and CSK are the least because they use a single kernel to cover the localization area. In the multi-kernel part, the hybrid kernel takes more times than RSK and CSK, because the hybrid kernel uses two kernels to cover the localization area. In RMK and CMK, the time cost to form multiple kernels are the most, because the multi-kernel scheme uses multiple kernels to cover the localization area.

5.4. Discussions

According to the above experiment results, the proposed multi-kernel scheme can achieve the lowest localization error and the highest localization accuracy because through the multi-AP selection, 3D multi-kernel construction, weight adjustment, and the multi-kernel optimization phases, the proposed multi-kernel scheme can derive the best expected kernel through multiple kernels constructed by the signals collected from the selected APs and thus the proposed localization scheme can estimate the location more accurately and hence reduce the localization error. The CMK scheme performs better than the RBF scheme because the signal obtained by users with short distance movement is not obviously different after mapped to high-dimension by RBF kernel which leads to higher localization error as opposed to the chi-squared kernel.
The multi-DNN scheme performs the second best in terms of localization error and localization accuracy. However, the multi-DNN scheme needs to adjust the weight of each neuron, the convergence speed is slower than that of the multi-kernel scheme. Besides, the multi-DNN scheme needs to collect more sample data.
The time cost of the proposed multi-kernel scheme is higher than the other kernel-based scheme, because proposed multi-kernel scheme needs to construct multiple kernels, adjust weight for each kernel, and optimize the expected kernel and hence the time cost of the proposed scheme is the highest. Basically, the more the kernels are constructed, the more the time cost becomes. The time cost is also proportional to the number of selected APs.

6. Conclusions

In this paper, we present a new 3D localization scheme which is able to construct multiple kernels for 3D indoor localization. Compared with the traditional kernel learning scheme, we create multiple kernels to improve the localization accuracy. The proposed scheme considers multiple parameters and learns the signal’s distribution generated by each AP in each cube from the source domain and transfer to the target domain. Finally, we provide the experiment results and show that the proposed scheme performs better than the multi-DNN scheme and the existing kernel-based localization schemes in terms of localization accuracy and error in the 3D indoor environment.

Author Contributions

Conceptualization, Y.-S.C.; methodology, Y.-S.C.; C.-S.H. and R.-S.C.; software, Y.-S.C. and R.-S.C.; validation, Y.-S.C. and C.-S.H.; formal analysis, Y.-S.C. and C.-S.H.; resources, Y.-S.C.; data collection and experiments, R.-S.C.; writing—original draft preparation, R.-S.C.; writing—review and editing, Y.-S.C. and C.-S.H.; visualization, Y.-S.C. and C.-S.H.; supervision, Y.-S.C. and C.-S.H.; project administration, Y.-S.C.; funding acquisition, Y.-S.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Ministry of Science and Technology of Taiwan under Grants MOST 109-2221-E-305-004-MY3.

Data Availability Statement

Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Tzitzis, A.; Megalou, S.; Siachalou, S.; Tsardoulias, E.; Yioultsis, T. 3D Localization of RFID Tags with a Single Antenna by a Moving Robot and Phase ReLock. In Proceedings of the IEEE International Conference on RFID Technology and Applications, Pisa, Italy, 25–27 September 2019. [Google Scholar]
  2. Wu, J.; Zhu, M.; Xiao, B.; Qiu, Y. The Improved Fingerprint-Based Indoor Localization with RFID/PDR/MM Technologies. In Proceedings of the IEEE 24th International Conference on Parallel and Distributed Systems (ICPADS 2018), Singapore, 11–13 December 2018. [Google Scholar]
  3. Cheng, S.; Wang, S.; Guan, W.; Xu, H.; Li, P. 3DLRA: An RFID 3D Indoor Localization Method Based on Deep Learning. Sensors 2020, 20, 2731. [Google Scholar] [CrossRef] [PubMed]
  4. Ha, G.Y.; Seo, S.B.; Oh, H.S.; Jeon, W.S. LoRa ToA-Based Localization Using Fingerprint Method. In Proceedings of the International Conference on Information and Communication Technology Convergence (ICTC 2019), Jeju Island, Korea, 16–18 October 2019. [Google Scholar]
  5. Abbas, M.; Elhamshary, M.; Rizk, H. WiDeep: WiFi-based Accurate and Robust Indoor Localization System using Deep Learning. In Proceedings of the IEEE International Conference on Pervasive Computing and Communications (PerCom 2019), Kyoto, Japan, 11–15 March 2019. [Google Scholar]
  6. Anzum, N.; Afroze, S.F.; Rahman, A. Zone-Based Indoor Localization Using Neural Networks: A View from a Real Testbed. In Proceedings of the IEEE International Conference on Communications (ICC 2018), Kansas City, MO, USA, 20–24 May 2018. [Google Scholar]
  7. Chang, R.Y.; Liu, S.J.; Cheng, Y.K. Device-Free Indoor Localization Using Wi-Fi Channel State Information for Internet of Things. In Proceedings of the IEEE Global Communications Conference (GLOBECOM 2018), Abu Dhabi, United Arab Emirates, 9–13 December 2018. [Google Scholar]
  8. Jiang, H.; Peng, C.; Sun, J. Deep Belief Network for Fingerprinting-Based RFID Indoor Localization. In Proceedings of the IEEE International Conference on Communications (ICC 2019), Shanghai, China, 20–24 May 2019. [Google Scholar]
  9. Zhang, L.; Wang, H. Fingerprinting-based Indoor Localization with Relation Learning Network. In Proceedings of the IEEE/CIC International Conference on Communications in China (ICCC 2019), Changchun, China, 11–13 August 2019. [Google Scholar]
  10. Hsu, C.S.; Chen, Y.S.; Juang, T.Y.; Wu, Y.T. An Adaptive Wi-Fi Indoor Localization Scheme using Deep Learning. In Proceedings of the IEEE Asia-Pacific Conference on Antennas and Propagation (APCAP 2018), Auckland, New Zealand, 5–8 August 2018. [Google Scholar]
  11. Qi, Q.; Li, Y.; Wu, Y.; Wang, Y.; Yue, Y.; Wang, X. RSS-AOA-Based Localization via Mixed Semi-Definite and Second-Order Cone Relaxation in 3-D Wireless Sensor Networks. IEEE Access 2019, 7, 117768–117779. [Google Scholar] [CrossRef]
  12. Cramariuc, A.; Huttunen, H.; Lohan, E.S. Clustering benefits in mobile-centric WiFi positioning in multi-floor buildings. In Proceedings of the International Conference on Localization and GNSS (ICL-GNSS 2016), Barcelona, Spain, 28–30 June 2016. [Google Scholar]
  13. Zanca, G.; Zorzi, F.; Zanella, A.; Zorzi, M. Experimental comparison of RSSI-based localization algorithms for indoor wireless sensor networks. In Proceedings of the Workshop on Real-World Wireless Sensor Networks, Glasgow, UK, 1 April 2008; pp. 1–5. [Google Scholar]
  14. Wu, G.; Tseng, P. A Deep Neural Network Based Indoor Positioning Method Using Channel State Information. In Proceedings of the International Conference on Computing, Networking and Communications (ICCNC 2018), Maui, HI, USA, 5–8 March 2018; pp. 290–294. [Google Scholar]
  15. Yan, J.; Zhao, L.; Tang, J.; Chen, Y. Hybrid Kernel Based Machine Learning Using Received Signal Strength Measurements for Indoor Localization. IEEE Trans. Veh. Technol. 2018, 67, 2824–2829. [Google Scholar] [CrossRef]
  16. Mari, S.K.; Kiong, L.C.; Loong, H.K. A Hybrid Trilateration and Fingerprinting Approach for Indoor Localization Based on WiFi. In Proceedings of the Fourth International Conference on Advances in Computing, Communication and Automation (ICACCA 2018), Subang Jaya, Malaysia, 26–28 October 2018; pp. 1–6. [Google Scholar]
  17. Zou, H.; Zhou, Y.; Jiang, H.; Huang, B.; Xie, L.; Spanos, C. Adaptive Localization in Dynamic Indoor Environments by Transfer Kernel Learning. In Proceedings of the IEEE Wireless Communications and Networking Conference (WCNC 2017), San Francisco, CA, USA, 19–22 March 2017. [Google Scholar]
  18. Wen, F.; Liu, P.; Wei, H. Joint Azimuth, Elevation, and Delay Estimation for 3-D Indoor Localization. IEEE Trans. Veh. Technol. 2018, 67, 4248–4261. [Google Scholar] [CrossRef]
  19. Wang, W.; Zhang, Y.; Tian, L. TOA-based NLOS error mitigation algorithm for 3D indoor localization. China Commun. 2020, 17, 63–72. [Google Scholar] [CrossRef]
  20. Marques, N.; Meneses, F.; Moreira, A. Combining similarity functions and majority rules for multi-building, multi-floor, WiFi positioning. In Proceedings of the International Conference on Indoor Positioning and Indoor Navigation (IPIN), Sydney, Australia, 13–15 November 2012; pp. 1–6. [Google Scholar]
  21. Chen, D.G.; Wang, H.Y.; Tsang, E.C. Generalized Mercer theorem and its application to feature space related to indefinite kernels. In Proceedings of the International Conference on Machine Learning and Cybernetics, Kunming, China, 12–15 July 2008; pp. 1–6. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.