Next Article in Journal
Improvement of Noise Uncertainty and Signal-To-Noise Ratio Wall in Spectrum Sensing Based on Optimal Stochastic Resonance
Next Article in Special Issue
A Human Activity Recognition Algorithm Based on Stacking Denoising Autoencoder and LightGBM
Previous Article in Journal
Real-Time Estimation for Roll Angle of Spinning Projectile Based on Phase-Locked Loop on Signals from Single-Axis Magnetometer
Previous Article in Special Issue
Wireless Fingerprinting Uncertainty Prediction Based on Machine Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Pedestrian Stride-Length Estimation Based on LSTM and Denoising Autoencoders

1
School of Information and Communication Engineering, Beijing University of Posts and Telecommunication, Beijing 100876, China
2
Beijing Key Laboratory of Mobile Computing and Pervasive Device, Institute of Computing Technology Chinese Academy of Sciences, Beijing 100190, China
3
School of Software Engineering, Beijing University of Posts and Telecommunication, Beijing 100876, China
4
State Key Laboratory of Advanced Optical Communication Systems and Networks, Peking University, Beijing 100871, China
*
Authors to whom correspondence should be addressed.
Sensors 2019, 19(4), 840; https://doi.org/10.3390/s19040840
Submission received: 15 December 2018 / Revised: 2 February 2019 / Accepted: 15 February 2019 / Published: 18 February 2019

Abstract

:
Accurate stride-length estimation is a fundamental component in numerous applications, such as pedestrian dead reckoning, gait analysis, and human activity recognition. The existing stride-length estimation algorithms work relatively well in cases of walking a straight line at normal speed, but their error overgrows in complex scenes. Inaccurate walking-distance estimation leads to huge accumulative positioning errors of pedestrian dead reckoning. This paper proposes TapeLine, an adaptive stride-length estimation algorithm that automatically estimates a pedestrian’s stride-length and walking-distance using the low-cost inertial-sensor embedded in a smartphone. TapeLine consists of a Long Short-Term Memory module and Denoising Autoencoders that aim to sanitize the noise in raw inertial-sensor data. In addition to accelerometer and gyroscope readings during stride interval, extracted higher-level features based on excellent early studies were also fed to proposed network model for stride-length estimation. To train the model and evaluate its performance, we designed a platform to collect inertial-sensor measurements from a smartphone as training data, pedestrian step events, actual stride-length, and cumulative walking-distance from a foot-mounted inertial navigation system module as training labels at the same time. We conducted elaborate experiments to verify the performance of the proposed algorithm and compared it with the state-of-the-art SLE algorithms. The experimental results demonstrated that the proposed algorithm outperformed the existing methods and achieves good estimation accuracy, with a stride-length error rate of 4.63% and a walking-distance error rate of 1.43% using inertial-sensor embedded in smartphone without depending on any additional infrastructure or pre-collected database when a pedestrian is walking in both indoor and outdoor complex environments (stairs, spiral stairs, escalators and elevators) with natural motion patterns (fast walking, normal walking, slow walking, running, jumping).

1. Introduction

Accurate and pervasive indoor positioning significantly improves our daily life [1]. The demand for accurate and practical location-based services anywhere using portable devices, such as smartphones, is quickly increasing in various applications, including asset and personnel tracking, health monitoring, precision advertising, and location-specific push notifications. A recent market report predicted that the global indoor-location market size is expected to grow from $7.11 billion in 2017 to $40.99 billion by 2022, at a Compound Annual Growth Rate of 42.0% during the forecast period [2].
To meet this explosive demand, various indoor positioning approaches have recently been developed, including RFID [3], Wi-Fi [4,5], UWB [6], BLE [7], magnetic [1,8,9,10], visible light [11,12] and visual methods [13]. Indoor localization techniques are classified into propagation model-based, fingerprint, and dead reckoning methods. Positioning performance of propagation model-based methods depends on the deployment density of the reference points. However, these methods are ineffective when the radio signal is weak or not available in many scenarios, such as underground parking lots. The accuracy of fingerprint-based approaches is affected by device orientation, pedestrians, and layout changes of indoor environments (e.g., removing or adding furniture). It is important that fingerprint- or infrastructure-based positioning techniques are not available for emergency scenarios, such as anti-terrorism action, emergency rescues and exploration missions.
Inertial Measurement Unit (IMU, consisting of gyroscopes and accelerometers)-based navigation algorithms overcome the limitations of radio-frequency signal fluctuations and blockage since they provide real-time locations of a pedestrian or object given an initial position, as well as not relying on any additional infrastructure or pre-collected database. To date, the previous studies on IMU-based navigation algorithms fall into two typical mechanizations: strap-down inertial navigation systems (INS) and pedestrian dead reckoning (PDR) [14].
INS mechanization provides more high-frequency and richer navigation information, including pedestrians’ 3D position, velocity, and attitude by integrating raw data from accelerometers and gyroscopes. However, positioning errors rapidly increase with time due to the drift characteristics of IMU sensors. To constrain the accumulative distance error of double integration, a zero-velocity-aided inertial navigation system uses periodic zero-velocity updates [15] by mounting an IMU module to the foot (such as toes, instep, or heel) of pedestrians. To restrain the accumulation errors of heading, zero angular rate updates [16] and magnetic angular rate updates [17] have been used to eliminate gyroscope bias. To date, the best foot-mounted INS (FM-INS) controls walking-distance estimation error in 0.3% of the entire travel distance [18]. However, FM-INS needs an additional foot-mounted inertial module as the platform, as well as special protection solutions (e.g., shockproofing, waterproofing, and pressure prevention) are necessary to protect the module from damages. Therefore, it is challenging to implement an FM-INS for extensive use in the consumer market [14].
The smartphone-based PDR mainly benefits from the extensive use of smartphones—most consumers always carry smartphones that have integrated inertial-sensor. PDR without the restriction of only using foot-mounted sensors estimates the relative location of a pedestrian by joining heading estimation, step detection, and SLE (stride-length estimation). As shown in Figure 1, strides are defined by the positions of two consecutive footfalls of the same foot, while the steps are defined by the positions of opposite feet. More details of the gait description can be found in [19]. The AHRS [20,21] provides robust and high-precision heading estimation for PDR by fusing the information from IMU sensors and magnetometers. An ocean of studies about step detection and step count have been done, including peak detection [22], zero-crossing [23], auto-correlation [24], and neural networks [25]. All of them achieve superior step-count accuracy. Stride-length is simultaneously computed using the generalized formulas related to the magnitudes of the accelerometers and gyroscopes, step frequency, and height of the pedestrian. SLE has a significant impact on the performance of PDR systems. Moreover, a pedestrian’s stride-length provides beneficial information for human gait analysis, sports-activity monitoring, prediction the status of human health, or energy-consumption assessment [26]. Nowadays, to develop PDR systems with smartphone embedded low-cost sensors, accurate SLE over a long period is a crucial problem for the following reasons [27]: (1) Stride-length differs with height, gender, age and weight; (2) Even the same person’s stride-length significantly varies depending on environmental differences and walking patterns (fast walking, normal walking, slow walking, running and jumping); (3) It is difficult to keep the stride-length constant even for the same person walking in a single pattern, in a single environment. The methods for estimating pedestrian step length are summarized as two categories: one is direct methods based on the integration of the acceleration; the other is indirect methods that leverage a model or assumption to compute step length.
The double integration of the acceleration component of the sagittal axis, is, in theory, the best method to compute the stride-length of pedestrians because it does not rely on any model or assumption, and it does not require training phases or individual information (leg length, height, weight) [28]. Kourogi [29] leveraged the correlation between walking velocity and vertical acceleration to estimate walking speed, then calculated step length by multiplying walking speed by step interval. Kang et al. [30] simultaneously measured the inertial sensor and global positioning system (GPS) position while walking outdoors with a reliable GPS fix, and regarded the velocity from the GPS as labels to train a hybrid multiscale convolutional and recurrent neural network model. After that, Kang leveraged the prediction velocity and moving time to estimate the traveled distance. However, it is challenging to obtain accurate labels since GPS contains a positional error. Moreover, it is difficult to obtain the acceleration component in the forward direction from the sensor’s measurements, as well as constantly maintain the sensor heading parallel to the pedestrian’s walking direction [31].
Considerable studies based on model or assumption have been conducted to improve the accuracy of stride-length estimation (SLE), and are summarized as empirical relationships [32,33], linear models [34], nonlinear models [35,36,37,38], biomechanical models [35,38], context-based [39,40] regression-based [41,42,43,44] and neural network [45,46]. A general pointer [47] had systematic compared of different methods for estimating short linear displacements of the body. Weinberg [36] leveraged the maximum and minimum acceleration values on the Z-axis in each stride to estimate walk distance. Kim [37] developed an empirical method that utilized the average of the acceleration amplitude in each step during walking to calculate the movement distance. To accurately estimate the traveled distance of a pedestrian, Ladetto [34] leveraged the linear relationship between step length and frequency, and the local variance of acceleration amplitude, to calculate motion distance. The methods mentioned above are only based on the maximum and minimum of acceleration amplitude, without considering the orientation and position of the smartphone. Gao [39] and Martinelli [40] proposed context-based step length estimation method However, the limited context classification cannot reflect the diversity of pedestrian walking patterns and environments. Zihajehzadeh and Park [43,44] leveraged Lasso regression to fit the linear model by minimizing a penalized version of the least squares loss function. However, it is difficult for handcraft-extracted features or models to reflect the walking patterns of a pedestrian. To solve this problem, Cho [46] utilized the neural-network method for step length estimation. Hannink [45] described a novel approach to stride-length estimation with deep convolutional neural networks and achieved better accuracy. However, this method relied on the stride-specific inertial-sensor that were attached to the pedestrian’s feet. Therefore, this method is unsuitable for smartphone-based applications.
These SLE algorithms present good performance in the case of walking a straight line at normal speed. Most of them however suffer from various limitations, such as unsuitability for smartphones, limited attitude, and relying on individual information or spatial constraints. However, a pedestrian may walk arbitrarily in different directions and may stop from time to time. Moreover, real paths including turns, stairs, sidesteps, varying speed or actions performed by the subject result in unacceptable SLE accuracy.
Deep learning has the ability to automatically learn features with high-level abstractions by using multiple nonlinear transformations [48]. Deep learning has been proven to be powerful for natural language processing and speech-recognition, without additional processing. Convolutional neural networks and Recurrent Neural Networks show excellent potential to exploit and analyze the collected data without handcraft feature extraction. Motivated by the fact that speech recognition based on deep learning outperforms other existing traditional speech recognition methods, this paper proposed a stride-length estimation method based on Long Short-Term Memory (LSTM) and Denoising Autoencoders (DAE), termed TapeLine, to address the challenge of PDR systematic errors caused by motion distance estimation from a natural walk in complex conditions. TapeLine estimated pedestrian’s stride-length and walking-distance with inertial-sensor embedded in smartphones. The key contributions of our study are as follows:
  • We propose a training frame for combining LSTM and DAE to deal with sequential data to extract the temporal feature meanwhile denoise, and a stride-length estimation model based on the training frame. Since the inertial-sensor measurements are time series data, we leveraged LSTM to excavate the temporal dependencies and extract significant features vectors from noisy inertial-sensor measurements. Denoising Autoencoders were adopted to automatically sanitize the inherent noise and obtain denoised feature vectors. A regression module was employed to map the denoised feature vectors to the resulting stride length.
  • We trained the proposed model with walking information from a smartphone, and the ground truth of stride-length from a foot-mounted IMU module, to predict an adaptive stride-length. In addition to the raw inertial-sensor data, the high-level stride-length features based on the excellent early studies are directly fed to the merge layer of networks. The proposed method is free of the zero-velocity assumption that double-integration methods need to reinitialize the integration process and eliminate accumulative errors.
  • In addition to evaluate the robustness performance of the proposed TapeLine under different operation conditions, we compared TapeLine with the existing commonly-used stride-length estimation methods in both single-stride and complex paths. Whether stride-length estimation or walking-distance estimation in complex environments with natural walking patterns, our proposed method outperformed commonly-used stride-length estimation methods and achieved a superior performance, with a stride-length error rate of 4.63% and a walking-distance error rate of 1.43%.
  • We established a benchmark dataset with ground truth for training step counting and stride-length estimation. A foot-mounted IMU module (x-IMU [49] controls motion distance errors in 0.3% of the entire travel distance) was attached to a pedestrian’s shoes that provided precise heel strike times and actual stride-length. Training data was generated from smartphone, and the annotated data were generated by a foot-mounted IMU module. In addition to model training and performance evaluation of step counting and stride-length estimation, the dataset is applied to explore the optimal parameters.
The rest of the paper is organized as follows: in Section 2, we detail the solution of the proposed TapeLine. In Section 3, we describe the benchmark dataset collection in detail and evaluate the proposed scheme. In Section 4, we provide a conclusion that summarizes the importance of our proposed work.

2. Materials and Methods

In this section, an overview of the system is depicted. Then, the key modules and important algorithms are described, which include data pre-processing based on LSTM-DAE and adaptive stride-length estimation.

2.1. System Architecture

Figure 2 illustrates the overall structure of TapeLine. A smartphone is used to collect the inertial data that is split into segments, with each segment representing the data for one stride. We utilized the inertial-sensor measurements from the smartphone as training data, and the corresponding motion distances from the FM-INS module (x-IMU [49] with a three-axis accelerometer (range ±16 g, resolution 490 μg), and a three-axis gyroscope (range ±2000 deg/sec, resolution 0.06 deg/sec) from x-io technologies) as labels to train a predictive model in the offline phase. In the online predicting phase, we leveraged the real-time sensor data and trained model to adaptively estimate the length of each stride.

2.2. Benchmark Dataset

The lack of benchmarking datasets for pedestrian stride-length and walking-distance makes it hard to pinpoint the differences of the published methods. The existing datasets either lack the ground-truth of each stride or are limited to small spaces with a single-scene or motion pattern. It is natural to think of using laser rangefinder to measure the motion distance of each stride. However, it is laborious or impossible to measure the stride-length of numerous strides using a laser rangefinder. To thoroughly evaluate the performance of stride-length and walking-distance estimation, we used our benchmark dataset (https://github.com/Archeries/StrideLengthEstimation/ tree/master/Benchmark-Dataset-for-Adaptive-Stride-Length-Estimation) for natural pedestrian dead reckoning using a smartphone (Huawei Mate 9 with an 8-core 2.4 GHz processor) and an FM-INS module (x-IMU [49] with a three-axis accelerometer (range ±16 g), and a three-axis gyroscope (range ±2000 deg/sec) from x-io technologies), as shown in Figure 3. The smartphone inertial-sensors (a three-axis accelerometer (range ±8 g), and a three-axis gyroscope (ICM-20690, range ±2000 deg/sec) from InvenSense (Sunnyvale, CA, USA) were sampled at 100 Hz. The FM-INS module performs zero-velocity updates algorithm to detect heel strike moment of foot attached FM-INS module, and then calculates motion distance of the foot by acceleration integration, and sends motion distance to smartphone via Bluetooth. The x-IMU module only sends motion distance to mobile phone, which guarantee the real-time performance of Bluetooth communication. x-IMU controls walking-distance estimation error in 0.3% of the entire travel distance, while the mean of stride (two steps) length is 1.36 m. The single stride error of x-IMU is less than 0.4 cm (1.36 m *0.3% = 0.00408 m ≈ 0.4 cm). Therefore, the FM-INS module is precise enough to serve as ground truth. Once a stride is detected, the sampling application on smartphone will automatically associate the collected sensor data with motion distance. The datasets were gathered by five volunteers (three males and two females, aged between 23 and 32, height between 152 and 196 cm, and weight between 45 and 80 kg) with natural motion patterns (including, fast walking, normal walking, slow walking, running, jumping). Throughout the datasets, the users hold the phone in their right hand in front of their chest. The datasets contain more than 14 km and 10,000 strides of gait measurements. The datasets contain indoor and outdoor cases, including stairs, escalators, elevators, office environments, shopping malls, streets, and metro stations. To maximize the compatibility, all data were published in open and simple file formats. The sample hold nine degree-of-freedom sensor data from smartphone embedded sensors and the corresponding stride count, stride-length, and cumulative walking-distance from foot-mounted module attached to the instep of the right foot of the pedestrian. To make it easier for readers to replicate experiment, we shared the sampling software in GitHub (https://github.com/Archeries/StrideLengthEstimation/tree/master/SampleAPP) More detailed info about the dataset and the sample app can be found in the readme.md file (https://github.com/ Archeries/StrideLengthEstimation) of GitHub.

2.3. Data Preprocessing and High-Level Feature Extraction

Before data were fed to the LSTM network, we performed a series of preprocessing operations, including data segmentation, extraction of annotated strides, and length alignment. First, we split the inertial sensor data according to the step event. For each segment, we extracted the sensor data and corresponding ground truth of the stride-length to generate the training data and labels. We normalized the readings from the accelerometer and gyroscope with respect to the respective sensor ranges. Figure 4a shows the histograms of the sensor-reading number over each stride. From the figure, we can see that the mean and deviation of the stride-length are 122.2 and 15.6, respectively. We inferred that the average time for each stride was 1.22 s since the sampling rate was set to 100 Hz. Figure 4b demonstrates that the sensor-reading number of the 99th percentile of each stride was less than 200. Therefore, 300 samples (3 s) of each stride were enough to cover different walking speeds except static. To ensure equally scaled and fixed-size input to the network, we infinity-padded or intercepted the sensor samples per stride to a fixed length of 300. Preprocessed data xi,j contain i = 1, …, 300 samples and j = 1, …, 6 channels.
After data preprocessing, we extracted higher-level features based on excellent early studies including Weinberg [36], Kim [37], Ladetto [34] and Scarlett [50] according to the following equations:
W e i n b e r g = K × a max a min 4
K i m = K × i = 1 N | a i | N 3
L a d e t t o = α × f + β × v + γ
S c a r l e t t = K × i = 1 N | a i | N a min a max a min
where K denotes the calibration coefficient that is obtained by the ratio of the real distance and the estimated distance. a max and a min denote the maximum and minimum acceleration values on the Z-axis in each stride, respectively. a i represents the measured acceleration value of the i t h sample in each stride. f and v denote step frequency and acceleration variance, respectively. α and β denote the weighting factors of step frequency and acceleration variance, respectively. γ represents a constant.
As shown in Figure 5, we construct stride data using the six channels of raw inertial information (the silver numbers) over a stride as well as the four extracted high-level features (the yellow numbers). The red number is the motion distance of each stride from the FM-INS module. After that, the constructed stride data will be fed to the neural network described later.

2.4. Stride-Length Estimation Model

Our proposed solution for stride-length estimation is composed of three parts for three steps, as shown in Figure 6: (1) A pure LSTM model which contains two LSTM layers and four fully connected layers, which is first trained to get a relatively optimal weights distribution of LSTM. (2) A DAE model including the two LSTM layers in (1), keeping them untrainable, and a Dropout-Encoder-Decoder module, which is trained to get a noise-robust encoder. (3) The final regression model built by adding a regression module with three fully connected layers to the encoder we got, of which all the layers are fine-tuned.

2.4.1. Temporal Feature Extraction based on Long Short-Term Memory

The measurements of sensors (i.e., accelerometer, gyroscope) over a stride (two consecutive steps) are expressed as follows:
a i = [ a x a y a z ] T
g i = [ g x g y g z ] T
A c c i = [ a 1 a 2 a T ]
G y r i = [ g 1 g 2 g T ]
D = { ( [ A c c i , G y r i ] , y i ) , i = 1 , , N }
where the segment size T is fixed constant. ax, ay and az indicates the tri-axis accelerometer. gx, gy and gz indicates the tri-axis gyroscope.
In a LSTM cell, xt, ht and Ct represent the input vector, the recurrent hidden state, and the long-term state which keeps the long-term memory of cell by gates (forget, input and output), at time-step t, respectively. LSTM takes the acceleration measurements in stride as the input sequences, and feeds to the cell sequentially. The outputs of gates at time-step t are first calculated as Equations (10)–(12). For t = 0, both the previous cell state and hidden layer are assigned to zero vectors:
f t = σ ( W f · [ h t 1 , x t ] + b f )
i t = σ ( W i · [ h t 1 , x t ] + b i )
o t = σ ( W o · [ h t 1 , x t ] + b o )
where W is M × ( M + 3 ) weight matrix. b is a M × 1 bias matrix. σ ( · ) is a sigmoid function providing nonlinearity for gates.
Then the candidate state value C ˜ t is calculated as (13), and the long-term state C t is updated as (14):
C ˜ t = tanh ( W C · [ h t 1 , x t ] + b C )
C t = f t C t 1 + i t C ˜ t
The output of current time-step is calculated as (16), and so as the final output, which should be an embedding of the input sequence in temporal feature space:
h t = o t tanh ( C t )
H ( x ) = h T = o T tanh ( C T )
The architecture and parameters of layers of the pure LSTM model are shown in Figure 7. As we can see, the LSTM model takes accelerometer and gyroscope measurements as well as the high-level features mentioned above. The hidden Dense layers have Rectified Linear Unit (ReLU) as their activation function, and the prediction Dense layer has no activation function cause it’s a regression task. The LSTM layers map the input sequences to an embedding in temporal feature space, while the high-level features provide expert knowledge for the model. A concatenate layer is added to merge these information. Then, four cascading fully connected layers are added to build a map between feature space and target value (i.e., the stride-length).

2.4.2. Noise sanitization based on Denoising Autoencoders

Due to the fact the raw sensor readings from accelerometers and gyroscopes inevitably contain noise, it is necessary to perform a filtering operation. Unlike the traditional filtering techniques (e.g., low-pass or median filter), the denoising autoencoders approach is highly efficient in learning signal features and predicting stride-length with much better accuracy. In order to force the hidden layer to extract more robust features and prevent it from simply learning the identity, we trained a DAE (Denoising Autoencoders) to reconstruct the merged feature from a corrupted version of it.
As shown in Figure 8, the DAE contains three parts: a Dropout, an Encoder and a Decoder. The network output maps the hidden representation of corrupted h back to a reconstruction h ^ . Our DAE model takes the merged feature h as the input and corrupts it by a Dropout layer Drop(·), then the Encoder Enc(·) maps it to a lower dimension space and Decoder Dec(·) maps it back to merged feature space. After that, the hidden features learned from autoencoders are utilized as inputs for the regression model for stride-length estimation:
h = [ H ( A c c ) H ( G y r ) F e a t s ]
h ^ = Dec ( Enc ( Drop ( h ) ) )
Given a merged feature h , the DAE trains the Encoder and Decoder to minimize the reconstruction error, which corresponds to minimize the following objective function:
J D A E ( h , h ^ ) = 1 2 M i = 1 M ( h i h ^ i ) 2

2.4.3. Stride-Length Regression

As shown in Figure 9, once the DAE is built, the actual stride-length is utilized to train the supervised regression layer that estimates the stride-length of pedestrian. The global objective of the regression layer is to minimize the error loss function J(D,G) between the actual stride-length and the estimated value:
y ^ = G ( Enc ( Drop ( h ) ) )
J ( D , G ) = 1 2 N i = 1 N ( y i y ^ i ) 2
where yi is the actual stride-length corresponding to the input D. y ^ i is the stride-length estimation from the regression layer. G(·) is the regression module in the final regression model. The minimization is realized by the RMS propagation which often used in the training of neural networks.
Algorithm 1 depicts the complete procedures of the proposed stride-length estimation method. The algorithm takes a set of training samples with the corresponding actual stride-length as input to train the network. The accelerometer readings and gyroscope readings are divided into segments by stride event. The fixed length of stride data is fed to LSTM-DAE network. The actual stride-length is used to train the regression layer on the top of the network. Once the training is done, the LSTM-DAE model will be used to predict the stride-length of pedestrian.
Algorithm 1. adaptive stride-length estimation based on LSTM-DAE
1Input: training data with actual stride-length T = { t i , i = 1 , 2 , , n } , test data without actual stride-length
2Output: stride-length estimation of pedestrian
3// Data preprocessing
4Split the inertial sensor data according to the stride event.
5For each stride do
6   Extract sensor data and corresponding ground truth to generate the training data and labels
7   Extract high-level feature
8   Infinity-pad or intercept the sensor samples of per stride to a fixed length
9   Construct Stride data as shown in Figure 5
10End for
11// Model training
12build and train the pure LSTM model
13build the DAE model and initialize the weights of LSTM layers by the pure LSTM model, set the LSTM layers to be untrainable and train DAE model
14build the final regression model and initialize the weights of layers before Decoder, set all layers to be trainable and train to fine-tune
15//Testing
16Leverage trained model to predict stride-length of pedestrian

2.5. Parameter Set and Network Performance

Our proposed TapeLine algorithm was implemented using Keras (https://keras.io/) with pandas for data management in a Windows environment. Optimization was performed with the RMS propagation (RMSprop) algorithm [51]. Table 1 summarizes the hyperparameter values of TapeLine.
To prevent over-fitting, we employed a dynamic stop criterion for the model training. System automatically stop the iteration when the loss function does not drop within 50 epochs. Figure 10 illustrates the variation in loss function with respect to training and validation. As the figure shows, the loss function for training data tends to decrease with the iterations, whereas that for validation decreases to a certain iteration and then slightly increases again. Stride length estimation based on pure LSTM needed 436 epochs to reach a stable regime on the entire training dataset, whereas stride length estimation based on LSTM-DAE only need 221 epochs. Stride length estimation based on LSTM-DAE converges faster than that based on pure LSTM.

2.6. Walking-Distance Estimation

To improve the robustness of the step counting, we utilized the step counting algorithm proposed by Kourosh [52], which solves the overcounting problem caused by false walking (e.g., when users use their phones for playing games in a still state). The cumulative walking-distance of pedestrians is calculated with the number of steps and estimated stride-length in every valid step event. Denoting N as the cumulative number of strides walked during the experiment, we established walking-distance CD by summing the adaptive stride-length of each stride as follows:
C D = i = 1 N L i
where Li represents the length of the i-th stride.

2.7. Evaluation Metrics

We leveraged the stride-length error rate and walking-distance error rate to measure the proposed method. The relative error rate of the stride-length is calculated with the following formula:
E s = 1 N i = 1 N ( | L e i L t i | L t i × 100 % )
where L e i , L t i denote the estimated stride-length and the actual stride-length of the i-th stride, respectively.
The relative error rate of walking-distance is calculated with the following formula:
E c d = | i = 1 M L e i i = 1 M L t i | i = 1 M L t i × 100 %
where L e i , L t i denote the estimated stride-length and the actual stride-length of the i-th stride, respectively.

3. Experimentation and Evaluation

To understand the effectiveness and limitations of the proposed TapeLine algorithm, we implemented and evaluated TapeLine in both indoor and outdoor complex environments (stairs, spiral stairs, escalators and elevators) with natural motion patterns (fast walking, normal walking, slow walking, running, jumping).

3.1. Experimental Setup

We collected test data using an Android smartphone (Huawei Mate 9 with an 8-core 2.4 GHz processor), which equipped with a three-axis accelerometer (range ±8 g), and a three-axis gyroscope (range ±2000 deg/sec) from InvenSense (ICM-20690). Both single stride-length error and accumulative stride-length error are essential indicator that used to evaluate the accuracy of stride-length algorithms. In Section 3.2, we first utilize the single stride-length error to compare the proposed TapeLine with the state-of-the-art SLE methods and verify the robustness of the proposed TapeLine among different scenarios, heterogeneous devices and different pedestrians using the control variable method, respectively. In Section 3.3, we utilized accumulative stride-length error to evaluate the accuracy of TapeLine in an indoor-outdoor complicated path and an indoor closure path.

3.2. Experiment Results of Stride-Length Estimation

We leveraged the trained SLE model and learnable parameters (e.g., weights and biases) to predict stride-length. In addition to comparing the proposed stride-length estimation with the state-of-the-art SLE methods, we also conducted extensive experiments to verify the robustness of the proposed TapeLine in different scenarios, with heterogeneous devices, and different pedestrians using the variable-controlling approach, respectively.

3.2.1. Comparison of Stride-Length Estimation using LSTM and LSTM-DAE

To explore how much performance improvement was gained from DAE, we trained two SLE models (LSTM-based, and LSTM and DAE-based) using the same training dataset and test data. Table 2 summarizes stride-length estimation comparison of LSTM and LSTM–DAE. 80% of the stride-length error and error rate were 0.063 m and 4.63%, respectively. Compared to LSTM, mean error rate reduced from 3.75% to 3.16%. In another word, the mean error rate of LSTM-DAE was reduced by 15.7% ((3.75%–3.16%)/3.75%). The experiment results demonstrated that DAE help improved stride-length estimation accuracy.

3.2.2. Comparison with Other Methods

We also compared the proposed stride-length estimation method with the state-of-the-art SLE methods (Kim [37], Ladetto [34] and Weinberg [36]). To clearly illustrate the distribution of SLE errors, we used box plots and CDF to compare the statistics of single SLE errors (total 1000 strides with natural walking). From Figure 11, we can see that the relative error of the proposed algorithm is smaller than those achieved by the Kim, Weinberg, Ladetto algorithms. From Figure 12, we can see that the median, the lower and the upper quartiles of the proposed algorithm are smaller than those of the Kim, Weinberg, Ladetto algorithms.

3.2.3. Robustness among Typical Scenarios

To verify the performance and practicality of our proposed method, we conducted experiments with six different settings that covering a wide range of typical scenarios (offices, shopping malls, streets, metro stations, underground parking lots, streets, and footpaths). Figure 13 shows the comparison of estimated stride-length and actual stride-length.
The CDF (cumulative distribution function) of the SLE error for all scenarios is plotted in Figure 14. All results demonstrate that it achieves promising performance, where 80% of the stride-length estimation errors occur within 0.071, 0.076, 0.075, 0.058, 0.082 and 0.088 m, and 80% of the error rates occur within 5.22%, 5.59%, 5.51%, 4.26%, 6.03%, and 6.47%, respectively.

3.2.4. Robustness among Heterogeneous Devices

Device heterogeneity is a long-standing common challenge of indoor localization techniques. Even under identical environment, different devices would observe different inertial-sensor signals due to the hardware diversity and inevitable sensors noise. To examine TapeLine’s robustness, we conducted experiments using a data acquisition device (Huawei Mate 9) and another three heterogeneous devices (Samsung Galaxy S6, Huawei Mate 20, and Huawei P 9). The CDF of stride-length estimation error for four devices is plotted in Figure 15. From the figure, we can see that the CDF curves are very close, and they are consistent with the experimental results of data acquisition device. This demonstrates the robustness and practicality of TapeLine.

3.2.5. Robustness among Different Pedestrians

The above experiment results were performed by two of our authors. To evaluate the robustness of TapeLine with respect to different pedestrians that have different walk patterns, we invited five other volunteers. Naturally, their gait patterns were different. To accurately record the actual stride-length of pedestrians, an FM-INS module was attached to the volunteers’ instep of the right foot. All the volunteers were assigned to walk along the same path (about 400 m) in the office environment. The CDF of the SLE error for all five volunteers is plotted in Figure 16. From the figure, we can see that the five CDF curves are very similar, and they are consistent with the experiment results from our own walks. The experiment results demonstrate the robustness and practicality of TapeLine.

3.3. Walking-Distance Estimation in Complex Paths

To evaluate the accuracy of TapeLine under various conditions with natural walking patterns, we started walking from an indoor office (the seventh floor of the Institute of Computing Technology, Chinese Academy of Sciences). After walking for about 100 meters, we reached the stairs. We walked downstairs from the seventh floor to the ground floor. Then, we exited the office and walked along streets to the youth apartment of the Chinese Academy of Sciences. The path length was about 1265 meters and 907 strides, including a pedestrian skyway. Figure 17a illustrates the entire walking path. We also invited a new volunteer using a new smartphone to conduct similar experiment along an indoor closure path (see Figure 17b). The comparison of walking-distance estimation is shown in Table 3.
Whether the proposed or compared methods, the walking-distance error rate is significantly less than the stride-length error rate. To explore the reason, we visualized the error distribution of stride-length estimation as Figure 18.
From the figure, we can see that the stride estimation error follows a Gaussian distribution. Therefore, the cumulative stride-length error is less than the sum of the single stride-length error.

3.4. Time Complexity Analysis

The most time-consuming procedures of TapeLine are the training data collection and neural-network training. However, both of these two procedures are performed in the offline phase, which means they do not consume any time during the online prediction phase. The proposed method was implemented in Python and performed on a personal computer equipped with an Intel Core i5-4460 CPU at 3.20 GHz and 16 GB of DDR4 RAM. Table 4 reports the training and test time of TapeLine with different network models on the whole training dataset and test data. The time consumption of the training data collection equals walking time. For a 14 km and 10,000 strides data set, the time consumption of training is 3 h 01 min 26 s. From our test, the running time of prediction was less than 2.7 ms (2.369 s/888) for each stride.

4. Discussion and Conclusions

Pedestrian stride-length provides beneficial information for human gait analysis, sports-activity monitoring, and the prediction status of human health. Especially, SLE is one of the most critical factors in PDR-based indoor positioning. Accurate SLE of pedestrians is a challenging research topic, due to the various walking patterns. Since LSTM examined the temporal dependencies and extracted significant features vectors from noisy inertial-sensor measurements, and Denoising Autoencoders automatically sanitized the inherent noise and obtained denoised feature vectors, the proposed model adapts to the characteristics of different pedestrians and their walking patterns. In addition to single stride-length estimation, we also conducted walking-distance estimation experiments in challenging scenarios, including indoor and outdoor environments with natural walking patterns. The proposed method achieved a superior performance, with a single stride-length error rate of 4.63% and a walking-distance error rate of 1.43%. Generally, whether for single stride-length or walking-distance estimation, our proposed method outperforms commonly-used stride-length estimation methods. The performance of the Weinberg model and the Kim model were similar and worse than our proposed stride estimation algorithm. Both of them take the acceleration and the K parameter as input, and do not consider walking speeds or phone attitudes. By considering pedestrian’s acceleration and step frequency, the Ladetto model is more robust against different walking speeds and phone attitudes than the Weinberg model and the Kim model.
Detecting heel strike moment and splitting the inertial-sensors data are the key point for accurate stride-length estimation. In this paper, the training data is split by a foot-mounted IMU module. The training data segmentation is affected by the time delay of acceleration integration and Bluetooth communication. In practical applications, it is extremely difficult to strictly obtain sensor data at every stride by only using the smartphone built-in inertial sensor. In the offline phase, we estimate the offset between the peak point of acceleration and the heel strike moment detected by foot-mounted IMU module. In the online phase, we utilize the offset and peak point of acceleration to split the sensor data of each stride. However, the online sensor data segmentation for each stride is not accurate enough, thus resulting in inaccurate stride-length estimation. The problem of inertial-sensor data splitting and segmentation is still an open issue in all stride-length estimation methods. In our future work, we will explore more accurate tool (e.g., a high speed camera) to obtain more accurate ground truth, segment and align the sensor data of each stride.
Since humans are flexible structures, it is difficult to ensure that the movement of mobile phones equals the movement of pedestrians. Extra actions (standing still or swing hands, playing games, calling, reading, etc.) results in inaccurate stride-length estimation. Different smartphone carrying methods will have an obvious influence on stride-length estimation accuracy. Therefore, we reiterate here that pedestrians should hold their phone horizontally with the hand in front of their chest. The proposed approach was tested on a few healthy adults and devices. However, the trained model may be not suitable for children, elderly and non-healthy adults (e.g., Parkinson’s patients). In the future, we will investigate how to automatically obtain training data by crowdsourcing, then train a personalized SLE model in the form of online learning. The personalized model will undoubtedly further enhance the performance of the proposed method.

Author Contributions

Q.W., H.L. and A.M. conceived the framework and designed the algorithm and experiments; Q.W. wrote the paper; L.Y. and Y.H. performed some experiments and analyzed the data; F.Z. and Y.Z. guided the paper writing and reviewed the paper. All authors read and approved the final manuscript.

Funding

This work was supported in part by the National Key Research and Development Program (2018YFB0505200), the BUPT Excellent Ph.D. Students Foundation (CX2018102), the National Natural Science Foundation of China (61872046, 61671264 and 61671077) and the Open Project of the Beijing Key Laboratory of Mobile Computing and Pervasive Device.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shu, Y.; Bo, C.; Shen, G.; Zhao, C.; Li, L.; Zhao, F. Magicol: Indoor Localization Using Pervasive Magnetic Field and Opportunistic WiFi Sensing. IEEE J. Sel. Areas Commun. 2015, 33, 1443–1457. [Google Scholar] [CrossRef]
  2. Indoor Location Market Worth 40.99 Billion USD by 2022. Available online: https://www.marketsand markets.com/PressReleases/indoor-location.asp (accessed on 1 December 2018).
  3. Jimenez Ruiz, A.R.; Seco Granja, F.; Prieto Honorato, J.C.; Guevara Rosas, J.I. Accurate Pedestrian Indoor Navigation by Tightly Coupling Foot-Mounted IMU and RFID Measurements. IEEE Trans. Instrum. Meas. 2012, 61, 178–189. [Google Scholar] [CrossRef] [Green Version]
  4. Fox, D.; Burgard, W.; Thrun, S. Markov Localization for Mobile Robots in Dynamic Environments. J. Artif. Intell. Res. 1999, 11, 391–427. [Google Scholar] [CrossRef]
  5. Liu, H.-H.; Liu, C. Implementation of Wi-Fi Signal Sampling on an Android Smartphone for Indoor Positioning Systems. Sensors 2017, 18, 3. [Google Scholar] [CrossRef] [PubMed]
  6. Tiemann, J.; Pillmann, J.; Wietfeld, C. Ultra-Wideband Antenna-Induced Error Prediction Using Deep Learning on Channel Response Data. In Proceedings of the 2017 IEEE 85th Vehicular Technology Conference (VTC Spring), Sydney, NSW, Australia, 4–7 June 2017; pp. 1–5. [Google Scholar]
  7. Yohan, A.; Lo, N.-W.; Winata, D. An Indoor Positioning-Based Mobile Payment System Using Bluetooth Low Energy Technology. Sensors 2018, 18, 974. [Google Scholar] [CrossRef]
  8. Wang, Q.; Luo, H.; Zhao, F.; Shao, W. An indoor self-localization algorithm using the calibration of the online magnetic fingerprints and indoor landmarks. In Proceedings of the 2016 International Conference on Indoor Positioning and Indoor Navigation (IPIN), Alcala de Henares, Spain, 4–7 October 2016; pp. 1–8. [Google Scholar]
  9. Shao, W.; Zhao, F.; Wang, C.; Luo, H.; Muhammad Zahid, T.; Wang, Q.; Li, D. Location Fingerprint Extraction for Magnetic Field Magnitude Based Indoor Positioning. J. Sens. 2016, 2016, 1–16. [Google Scholar] [CrossRef]
  10. Xie, H.; Gu, T.; Tao, X.; Ye, H.; Lu, J. A Reliability-Augmented Particle Filter for Magnetic Fingerprinting Based Indoor Localization on Smartphone. IEEE Trans. Mob. Comput. 2016, 15, 1877–1892. [Google Scholar] [CrossRef]
  11. Wang, Q.; Luo, H.; Men, A.; Zhao, F.; Huang, Y. An Infrastructure-Free Indoor Localization Algorithm for Smartphones. Sensors 2018, 18, 3317. [Google Scholar] [CrossRef]
  12. Wang, Q.; Luo, H.; Men, A.; Zhao, F.; Gao, X.; Wei, J.; Zhang, Y.; Huang, Y. Light positioning: A high-accuracy visible light indoor positioning system based on attitude identification and propagation model. Int. J. Distrib. Sens. Networks 2018, 14, 155014771875826. [Google Scholar] [CrossRef]
  13. Liu, T.; Zhang, X.; Li, Q.; Fang, Z. A Visual-Based Approach for Indoor Radio Map Construction Using Smartphones. Sensors 2017, 17, 1790. [Google Scholar] [CrossRef] [PubMed]
  14. Kuang, J.; Niu, X.; Chen, X. Robust Pedestrian Dead Reckoning Based on MEMS-IMU for Smartphones. Sensors 2018, 18, 1391. [Google Scholar] [CrossRef] [PubMed]
  15. Foxlin, E. Pedestrian Tracking with Shoe-Mounted Inertial Sensors. IEEE Comput. Graph. Appl. 2005, 25, 38–46. [Google Scholar] [CrossRef] [PubMed]
  16. Rajagopal, S. Personal Dead Reckoning System with Shoe Mounted Inertial Sensors. Master’s Thesis, Royal Institute of Technology, Stockholm, Sweden, 2008. [Google Scholar]
  17. Zampella, F.; Khider, M.; Robertson, P.; Jimenez, A. Unscented Kalman filter and Magnetic Angular Rate Update (MARU) for an improved Pedestrian Dead-Reckoning. In Proceedings of the 2012 IEEE/ION Position, Location and Navigation Symposium, Myrtle Beach, SC, USA, 23–26 April 2012; pp. 129–139. [Google Scholar]
  18. Gu, Y.; Song, Q.; Li, Y.; Ma, M. Foot-mounted Pedestrian Navigation based on Particle Filter with an Adaptive Weight Updating Strategy. J. Navig. 2015, 68, 23–38. [Google Scholar] [CrossRef]
  19. Del Din, S.; Hickey, A.; Ladha, C.; Stuart, S.; Bourke, A.K.; Esser, P.; Rochester, L.; Godfrey, A. Instrumented gait assessment with a single wearable: An introductory tutorial. F1000Research 2016, 5, 2323. [Google Scholar] [CrossRef]
  20. Yadav, N.; Bleakley, C. Accurate Orientation Estimation Using AHRS under Conditions of Magnetic Distortion. Sensors 2014, 14, 20008–20024. [Google Scholar] [CrossRef] [Green Version]
  21. Del Rosario, M.B.; Lovell, N.H.; Redmond, S.J. Quaternion-Based Complementary Filter for Attitude Determination of a Smartphone. IEEE Sens. J. 2016, 16, 6008–6017. [Google Scholar] [CrossRef]
  22. Qian, J.; Ma, J.; Ying, R.; Liu, P.; Pei, L. An improved indoor localization method using smartphone inertial sensors. In Proceedings of the International Conference on Indoor Positioning and Indoor Navigation, Montbeliard-Belfort, France, 28–31 October 2013; pp. 1–7. [Google Scholar]
  23. Goyal, P.; Ribeiro, V.J.; Saran, H.; Kumar, A. Strap-down Pedestrian Dead-Reckoning system. In Proceedings of the 2011 International Conference on Indoor Positioning and Indoor Navigation, Guimaraes, Portugal, 21–23 September 2011; pp. 1–7. [Google Scholar]
  24. Rai, A.; Chintalapudi, K.K.; Padmanabhan, V.N.; Sen, R. Zee: Zero-effort crowdsourcing for indoor localization. In Proceedings of the 18th Annual International Conference on Mobile Computing and Networking, Istanbul, Turkey, 22–26 August 2012. [Google Scholar]
  25. Shao, W.; Luo, H.; Zhao, F.; Wang, C.; Crivello, A.; Tunio, M.Z. DePedo: Anti Periodic Negative-Step Movement Pedometer with Deep Convolutional Neural Networks. In Proceedings of the 2018 IEEE International Conference on Communications (ICC), Kansas City, MO, USA, 20–24 May 2018; pp. 1–6. [Google Scholar]
  26. Beauchet, O.; Allali, G.; Sekhon, H.; Verghese, J.; Guilain, S.; Steinmetz, J.-P.; Kressig, R.W.; Barden, J.M.; Szturm, T.; Launay, C.P.; et al. Guidelines for Assessment of Gait and Reference Values for Spatiotemporal Gait Parameters in Older Adults: The Biomathics and Canadian Gait Consortiums Initiative. Front. Hum. Neurosci. 2017, 11, 353. [Google Scholar] [CrossRef] [PubMed]
  27. Edel, M.; Koppe, E. An advanced method for pedestrian dead reckoning using BLSTM-RNNs. In Proceedings of the 2015 International Conference on Indoor Positioning and Indoor Navigation (IPIN), Banff, AB, Canada, 13–16 October 2015; pp. 1–6. [Google Scholar]
  28. Diez, L.E.; Bahillo, A.; Otegui, J.; Otim, T. Step Length Estimation Methods Based on Inertial Sensors: A Review. IEEE Sens. J. 2018, 18, 6908–6926. [Google Scholar] [CrossRef]
  29. Kourogi, M.; Kurata, T. A wearable augmented reality system with personal positioning based on walking locomotion analysis. In Proceedings of the Second IEEE and ACM International Symposium on Mixed and Augmented Reality (ISMAR ’03), Tokyo, Japan, 7–10 October 2003; IEEE Computer Society: Washington, DC, USA, 2003; pp. 342–343. [Google Scholar]
  30. Kang, J.; Lee, J.; Eom, D.-S. Smartphone-Based Traveled Distance Estimation Using Individual Walking Patterns for Indoor Localization. Sensors 2018, 18, 3149. [Google Scholar] [CrossRef] [PubMed]
  31. Combettes, C.; Renaudin, V. Comparison of misalignment estimation techniques between handheld device and walking directions. In Proceedings of the 2015 International Conference on Indoor Positioning and Indoor Navigation (IPIN), Banff, AB, Canada, 13–16 October 2015; pp. 1–8. [Google Scholar]
  32. Ho, N.-H.; Truong, P.; Jeong, G.-M. Step-Detection and Adaptive Step-Length Estimation for Pedestrian Dead-Reckoning at Various Walking Speeds Using a Smartphone. Sensors 2016, 16, 1423. [Google Scholar] [CrossRef]
  33. Jahn, J.; Batzer, U.; Seitz, J.; Patino-Studencka, L.; Gutierrez Boronat, J. Comparison and evaluation of acceleration based step length estimators for handheld devices. In Proceedings of the 2010 International Conference on Indoor Positioning and Indoor Navigation, Zurich, Switzerland, 15–17 September 2010; pp. 1–6. [Google Scholar]
  34. Ladetto, Q. On foot navigation: Continuous step calibration using both complementary recursive prediction and adaptive Kalman filtering. In Proceedings of the 13th International Technical Meeting of the Satellite Division of The Institute of Navigation, Salt Lake City, UT, USA, 19–22 September 2000; Volume 2000, pp. 1735–1740. [Google Scholar]
  35. Huang, B.; Qi, G.; Yang, X.; Zhao, L.; Zou, H. Exploiting cyclic features of walking for pedestrian dead reckoning with unconstrained smartphones. In Proceedings of the 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing, Heidelberg, Germany, 12–16 September 2016; ACM Press: New York, NY, USA, 2016; pp. 374–385. [Google Scholar]
  36. Weinberg, H. Using the ADXL202 in Pedometer and Personal Navigation Applications. Available online: http://www.bdtic.com/DownLoad/ADI/AN-602.pdf (accessed on 17 February 2019).
  37. Kim, J.W.; Jang, H.J.; Hwang, D.-H.; Park, C. A Step, Stride and Heading Determination for the Pedestrian Navigation System. J. Glob. Position. Syst. 2004, 3, 273–279. [Google Scholar] [CrossRef] [Green Version]
  38. Miyazaki, S. Long-term unrestrained measurement of stride length and walking velocity utilizing a piezoelectric gyroscope. IEEE Trans. Biomed. Eng. 1997, 44, 753–759. [Google Scholar] [CrossRef] [PubMed]
  39. Gao, H.; Groves, P.D. Context Determination for Adaptive Navigation using Multiple Sensors on a Smartphone. In Proceedings of the 29th International Technical Meeting of The Satellite Division of the Institute of Navigation (ION GNSS+ 2016), Portland, OR, USA, 12–16 September 2016; pp. 12–16. [Google Scholar]
  40. Martinelli, A.; Gao, H.; Groves, P.D.; Morosi, S. Probabilistic Context-Aware Step Length Estimation for Pedestrian Dead Reckoning. IEEE Sens. J. 2018, 18, 1600–1611. [Google Scholar] [CrossRef] [Green Version]
  41. Munoz Diaz, E. Inertial Pocket Navigation System: Unaided 3D Positioning. Sensors 2015, 15, 9156–9178. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  42. Diaz, E.M.; Gonzalez, A.L.M. Step detector and step length estimator for an inertial pocket navigation system. In Proceedings of the 2014 International Conference on Indoor Positioning and Indoor Navigation (IPIN), Busan, South Korea, 27–30 October 2014; pp. 105–110. [Google Scholar] [CrossRef]
  43. Zihajehzadeh, S.; Park, E.J. Experimental evaluation of regression model-based walking speed estimation using lower body-mounted IMU. In Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBS), Orlando, FL, USA, 16–20 August 2016; pp. 243–246. [Google Scholar]
  44. Zihajehzadeh, S.; Park, E.J. Regression model-based walking speed estimation using wrist-worn inertial sensor. PLoS ONE 2016, 11, e0165211. [Google Scholar] [CrossRef] [PubMed]
  45. Hannink, J.; Kautz, T.; Pasluosta, C.F.; Barth, J.; Sch, S. Mobile Stride Length Estimation with Deep Convolutional Neural Networks. IEEE J. Biomed. Health Inform. 2018, 22, 354–362. [Google Scholar] [CrossRef] [PubMed]
  46. Cho, S.Y.; Park, C.G. MEMS Based Pedestrian Navigation System. J. Navig. 2005, 59, 135. [Google Scholar] [CrossRef]
  47. Alvarez, J.; Álvarez, D.; López, A. Accelerometry-Based Distance Estimation for Ambulatory Human Motion Analysis. Sensors 2018, 18, 4441. [Google Scholar] [CrossRef] [PubMed]
  48. Lecun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436. [Google Scholar] [CrossRef] [PubMed]
  49. x-IMU Sensor Board. Available online: http://x-io.co.uk/x-imu/ (accessed on 17 February 2019).
  50. Hayashitani, M.; Kasahara, T.; Ishii, D.; Arakawa, Y.; Okamoto, S.; Yamanaka, N.; Takezawa, N.; Nashimoto, K. 10ns High-speed PLZT optical content distribution system having slot-switch and GMPLS controller. IEICE Electron. Express 2008, 5, 181–186. [Google Scholar] [CrossRef] [Green Version]
  51. Tieleman, T.; Hinton, G.E.; Srivastava, N.; Swersky, K. Lecture 6.5-rmsprop: Divide the gradient by a running average of its recent magnitude. COURSERA Neural Networks Mach. Learn. 2012, 4, 26–31. [Google Scholar]
  52. Gu, F.; Khoshelham, K.; Shang, J.; Yu, F.; Wei, Z. Robust and Accurate Smartphone-Based Step Counting for Indoor Localization. IEEE Sens. J. 2017, 17, 3453–3460. [Google Scholar] [CrossRef]
Figure 1. Strides are defined by the positions of two consecutive footfalls of the same foot, while the steps are defined by the positions of opposite feet.
Figure 1. Strides are defined by the positions of two consecutive footfalls of the same foot, while the steps are defined by the positions of opposite feet.
Sensors 19 00840 g001
Figure 2. System architecture of the proposed TapeLine.
Figure 2. System architecture of the proposed TapeLine.
Sensors 19 00840 g002
Figure 3. Training data collection system.
Figure 3. Training data collection system.
Sensors 19 00840 g003
Figure 4. Distribution of sensor-reading number over each stride. (a) histogram; (b) cumulative distribution.
Figure 4. Distribution of sensor-reading number over each stride. (a) histogram; (b) cumulative distribution.
Sensors 19 00840 g004
Figure 5. Input example for a stride-length estimation network.
Figure 5. Input example for a stride-length estimation network.
Sensors 19 00840 g005
Figure 6. The proposed training frame.
Figure 6. The proposed training frame.
Sensors 19 00840 g006
Figure 7. The architecture and parameters of the pure LSTM model.
Figure 7. The architecture and parameters of the pure LSTM model.
Sensors 19 00840 g007
Figure 8. The architecture and parameters of DAE model.
Figure 8. The architecture and parameters of DAE model.
Sensors 19 00840 g008
Figure 9. The architecture and parameters of final regression model.
Figure 9. The architecture and parameters of final regression model.
Sensors 19 00840 g009
Figure 10. Cost function with respect to training and the test data set.
Figure 10. Cost function with respect to training and the test data set.
Sensors 19 00840 g010
Figure 11. The estimation error of proposed model and other methods.
Figure 11. The estimation error of proposed model and other methods.
Sensors 19 00840 g011
Figure 12. Box plot of stride-length estimation error. In this box plot, the whiskers represent the 99.3% coverage. On each box, the central (red) mark is the median, the edges of the box are the 25th and 75th percentiles.
Figure 12. Box plot of stride-length estimation error. In this box plot, the whiskers represent the 99.3% coverage. On each box, the central (red) mark is the median, the edges of the box are the 25th and 75th percentiles.
Sensors 19 00840 g012
Figure 13. Comparison of estimated stride-length and true stride-length in typical scenarios. (a) offices; (b) shopping malls; (c) metro stations; (d) underground parking lots; (e) streets; (f) footpath.
Figure 13. Comparison of estimated stride-length and true stride-length in typical scenarios. (a) offices; (b) shopping malls; (c) metro stations; (d) underground parking lots; (e) streets; (f) footpath.
Sensors 19 00840 g013aSensors 19 00840 g013b
Figure 14. Estimation error of stride-length in typical scenarios.
Figure 14. Estimation error of stride-length in typical scenarios.
Sensors 19 00840 g014
Figure 15. Robustness among different devises.
Figure 15. Robustness among different devises.
Sensors 19 00840 g015
Figure 16. Robustness among different pedestrians.
Figure 16. Robustness among different pedestrians.
Sensors 19 00840 g016
Figure 17. Walking path description. The volunteers were asked to travel the highlighted path with a natural pattern. Volunteer 1 (172 cm, 66 kg) walked along path (a). Volunteer 2 (180 cm, 80 kg) walked along path (b).
Figure 17. Walking path description. The volunteers were asked to travel the highlighted path with a natural pattern. Volunteer 1 (172 cm, 66 kg) walked along path (a). Volunteer 2 (180 cm, 80 kg) walked along path (b).
Sensors 19 00840 g017
Figure 18. Estimation error distribution of stride-length.
Figure 18. Estimation error distribution of stride-length.
Sensors 19 00840 g018
Table 1. List of hyperparameter values for network models.
Table 1. List of hyperparameter values for network models.
ParameterLSTMDAEFinal Model
Batch size128128128
Hidden layers32-16-8-132-16332-16-8-1
ActivationReLUSigmoid/LinearReLU
OptimizerRMSprop [51]RMSpropRMSprop
Learning rate0.0010.0010.001
Epochs50050500
Early stopping50/50
Loss functionMSEMSEMSE
Table 2. Comparison of stride-length estimation using LSTM and LSTM–DAE.
Table 2. Comparison of stride-length estimation using LSTM and LSTM–DAE.
AttributesLSTMLSTM-DAE
ErrorError Rate 1ErrorError Rate
Mean0.0513.75%0.0433.16%
Std0.037-0.036-
25%0.0251.83%0.0171.25%
50%0.0453.31%0.0362.64%
75%0.0685.00%0.0594.34%
min4.38 × 10−405.67 × 10−50
max0.34025.00%0.23917.57%
1 According to Equation (23).
Table 3. Comparison of walking-distance estimation in complex paths.
Table 3. Comparison of walking-distance estimation in complex paths.
PathAttributesRealProposedLadettoWeinberg Kim
aTotal distance (m)1267.821249.671238.201223.401219.74
Error (m)-18.1529.6245.4248.08
Error rate 2-1.43%2.34%3.50%3.80%
bTotal distance (m)94.4393.0191.7597.5797.69
Error (m)-1.422.683.143.26
Error rate-1.50%2.83%3.32%3.45%
2 According to Equation (24).
Table 4. Comparison of time complexity.
Table 4. Comparison of time complexity.
ModelsTraining Dataset Size Test Dataset SizeTrainable ParametersTraining Time Test Time
LSTM6571 strides888 strides407372 h 11 min 34 s2.158 s
LSTM-DAE92101 (40737 + 10627 + 40737)3 h 01 min 26 s2.369 s

Share and Cite

MDPI and ACS Style

Wang, Q.; Ye, L.; Luo, H.; Men, A.; Zhao, F.; Huang, Y. Pedestrian Stride-Length Estimation Based on LSTM and Denoising Autoencoders. Sensors 2019, 19, 840. https://doi.org/10.3390/s19040840

AMA Style

Wang Q, Ye L, Luo H, Men A, Zhao F, Huang Y. Pedestrian Stride-Length Estimation Based on LSTM and Denoising Autoencoders. Sensors. 2019; 19(4):840. https://doi.org/10.3390/s19040840

Chicago/Turabian Style

Wang, Qu, Langlang Ye, Haiyong Luo, Aidong Men, Fang Zhao, and Yan Huang. 2019. "Pedestrian Stride-Length Estimation Based on LSTM and Denoising Autoencoders" Sensors 19, no. 4: 840. https://doi.org/10.3390/s19040840

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop