Next Article in Journal
The Use of Artificial Neural Networks for Forecasting of Air Temperature inside a Heated Foil Tunnel
Next Article in Special Issue
High-Specificity Digital Architecture for Real-Time Recognition of Loss of Balance Inducing Fall
Previous Article in Journal
The Impact of Low Latency Satellite Sounder Observations on Local Severe Storm Forecasts in Regional NWP
Previous Article in Special Issue
The Coefficient of Variation of Step Time Can Overestimate Gait Abnormality: Test-Retest Reliability of Gait-Related Parameters Obtained with a Tri-Axial Accelerometer in Healthy Subjects
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Indoor Trajectory Reconstruction of Walking, Jogging, and Running Activities Based on a Foot-Mounted Inertial Pedestrian Dead-Reckoning System

1
Telematics Engineering Research Group, Telematics Department, Universidad Del Cauca (Unicauca), Popayán 190002, Colombia
2
Machine Learning and Data Analytics Lab, Computer Science Department, Friedrich-Alexander University Erlangen-Nürnberg (FAU), 91052 Erlangen, Germany
*
Authors to whom correspondence should be addressed.
Sensors 2020, 20(3), 651; https://doi.org/10.3390/s20030651
Submission received: 27 September 2019 / Revised: 22 October 2019 / Accepted: 1 November 2019 / Published: 24 January 2020

Abstract

:
The evaluation of trajectory reconstruction of the human body obtained by foot-mounted Inertial Pedestrian Dead-Reckoning (IPDR) methods has usually been carried out in controlled environments, with very few participants and limited to walking. In this study, a pipeline for trajectory reconstruction using a foot-mounted IPDR system is proposed and evaluated in two large datasets containing activities that involve walking, jogging, and running, as well as movements such as side and backward strides, sitting, and standing. First, stride segmentation is addressed using a multi-subsequence Dynamic Time Warping method. Then, detection of Toe-Off and Mid-Stance is performed by using two new algorithms. Finally, stride length and orientation estimation are performed using a Zero Velocity Update algorithm empowered by a complementary Kalman filter. As a result, the Toe-Off detection algorithm reached an F-score between 90% and 100% for activities that do not involve stopping, and between 71% and 78% otherwise. Resulting return position errors were in the range of 0.5% to 8.8% for non-stopping activities and 8.8% to 27.4% otherwise. The proposed pipeline is able to reconstruct indoor trajectories of people performing activities that involve walking, jogging, running, side and backward walking, sitting, and standing.

1. Introduction

Indoor positioning systems (IPS) enable the provision of several location-based services such as home monitoring, rehabilitation, navigation for blind and visual impaired people, and finding and rescuing people/firefighters in emergencies. IPSs can be divided into two approaches: infrastructure-based and infrastructure-free [1,2]. Infrastructure-based IPS require the deployment of devices in the indoor environment to calculate the position of the person. Among the technologies used by this type of IPS are Wi-Fi [3], radio frequency identification (RFID) [4], Bluetooth [5], ultra-wide band (UWB) [6], infrared [7], and video cameras [4]. Infrastructure-free IPS do not need the deployment of devices and mainly use dead-reckoning algorithms. Those systems are called inertial pedestrian dead-reckoning (IPDR) because they use body movement information measured by inertial measurement units (IMU) to estimate a person’s position changes based on a previously estimated or known position [2]. The sum of these changes of position allows the reconstruction of the person’s trajectory [2]. An IMU usually consists of a triaxial accelerometer and gyroscope. Although some IMUs also incorporate a triaxial magnetometer, alterations of the magnetic field indoors make it unreliable for indoor positioning [8].
The advantages of IPDR systems over infrastructure-based systems are generally lower cost, data privacy, and ease of deployment. However, IPDR systems without correction suffer from severe drift, as person displacement is often calculated by integrating acceleration data from the accelerometer twice and integrating the rotational angle from the gyroscope. In consequence, intrinsic errors and IMU noise are raised to the third power, making a person’s trajectory reconstruction by direct integration without correction impractical [9,10,11].
The literature review done in this study is aimed at foot-mounted IMU IPDR systems that only use the accelerometer and/or gyroscope. Foot-mounted IPDRs, together with a zero velocity update (ZUPT) algorithm, have been the most widely and successful method used to mitigate the drift in trajectory reconstruction [9]. We use only the accelerometer and gyroscope because in indoor environments, different sources might produce alterations in the magnetic field that make the magnetometer readings unreliable for trajectory reconstruction [8]. Most of the foot-mounted IPDR systems that only use accelerometer and gyroscope data are based on trajectory reconstruction during normal walking. Natural movements like avoiding obstacles, sitting, swinging legs, stopping, or performing activities like jumping, jogging, or running have rarely been considered [9,10]. In consequence, the literature review is focused on the foot-mounted IPDR systems that have reconstructed the trajectory of walking, jogging, and/or running activities. Thus, only six studies met the inclusion criteria and are part of the literature review. The foot-mounted IPDR systems are usually evaluated in closed-loop trajectories by measuring the return position error (RPE). The RPE indicates the distance between the final position of the person obtained by the system and the actual physical final position of the person at the end of the trial [8].
Threshold-based and machine learning-based foot-mounted IPDR approaches have been proposed to deal with walking and running activities [12,13,14,15,16]. Li et al. [12,13] proposed a threshold-based stance-phase detector that consists of one footstep detector and two zero velocity detectors, one for walking and another for running. The evaluation of the system was done with one pedestrian who followed two closed-loop trajectories while walking and running. For the square-shape path (195.7 m), the RPE was 0.24% for walking and 0.42% for running. For the eight-shape path (292.1 m), the RPE was 0.2% for walking and 1.01% for running. An adaptive zero-velocity detector that selects an optimal threshold for zero-velocity detection depending on the movement (walking or running) of the person was proposed by Wagstaff et al. [15]. This system was evaluated by five people who walked and ran a distance of 130 m in an “L” shaped path. The RPE reported were 1% for walking and 3.24% for running.
Considering that zero-velocity detection using machine learning-based IPDR systems is free of threshold-tuning, Wagstaff et al. proposed a method for zero-velocity detection by using a long short-term memory neural network (LSTM) [16]. Five people walked and ran a 220-m “L” shaped path. The RPE in walking was 0.49% and running 0.93%. Similarly, Ren et al. proposed a zero-velocity detection algorithm based on HMM [14]. The system was evaluated by one person in an oval-shaped sports field of 422 m. The RPE when walking and running was 0.6% and 1.61%, respectively.
The described works have obtained very high precision in the trajectory reconstruction of walking and running. However, the systems were evaluated with very few participants, and the evaluated trajectories involved continuous walking and running activities. Currently, trajectory reconstruction methods in realistic scenarios—with several people, and considering walking, jogging, and running strides—are still missing.
Physical activity classification and gait event detection are key components of the trajectory reconstruction process using IPDR. Machine learning has played an important role in both topics. In [17] it is shown how different machine learning-based algorithms are able to classify different physical activities, including standing, sitting, walking, and running. Gait event detection has been performed by using several machine learning algorithms such as deep learning [18], hidden Markov models (HMM) [19,20] and neural networks [21,22].
The aim of the present work was to propose a pipeline for trajectory reconstruction using a foot-mounted IPDR system able to reconstruct the trajectories of activities that involve walking, jogging, and running strides as well as natural movements like stopping, standing, sitting, and side-walking.
This paper contributes to foot-mounted IPDR systems by (1) comprehensively evaluating the trajectory reconstruction of activities that involve walking, jogging, and running strides including the discrimination of natural activities such as stopping, sitting, and side-walking; and (2) evaluating two algorithms for Toe-off and Mid-Stance detection during walking, jogging, and running strides adapted from the ones proposed by Barth et al. [23].
The proposed pipeline is able to recognize walking, jogging and running strides and detect the Toe-off and Mid-Stance events in each of them. With this information, a foot-mounted IPDR system is able to reconstruct the person’s trajectory regardless of their gait speed. This allows the development of new ambient assisted living applications in which indoor tracking is a ground technology as well as the development of new applications for indoor sports.

2. Datasets

2.1. Unicauca Dataset

The objective of the Unicauca dataset was to evaluate the trajectory reconstruction of walking, jogging, and running in similar settings as the state-of-the-art methods, which are usually evaluated in close-loop trajectories and the activities performed by the participants include continuous walking, jogging, or running. This dataset was collected at the University of Cauca, Popayán, Colombia. Ten participants (mean age: 30 ± 3 years) walked, jogged, and ran a closed-loop P-shaped path of approximately 150 m (Figure 1) with an IMU attached to the lateral side of the left shoe with a Velcro strap (Figure 2).
The IMU was a Shimmer3 GSR+ (Shimmer Sensing, Dublin, Ireland). Acceleration (range: ±16 g) and angular velocity (range: ±2000 dps) data were collected at a frequency of 200 Hz. Accelerometer calibration consisted in leaving the sensor still for a few seconds lying on each of its 6 sides on a flat surface. For gyroscope calibration, the sensor is rotated around the three axes. At the beginning of each trial, the participant was asked to remain standing without moving the IMU for at least 10 s for gyroscope bias calculation.

2.2. FAU Dataset

The FAU dataset is based on a previous study evaluating a method for smart labeling of cyclic activities [24] and is publicly available at www.activitynet.org. The dataset provides gait data in a relatively natural setting, and its protocol consisted in the execution of 12 different task-driven activities performed in random order for each participant. It includes data from 80 healthy participants with a mean age of 27 ± 6 years. Data were collected from 56 participants at the Friedrich-Alexander University Erlangen-Nürnberg (Germany) and from 24 participants at the University of Ljubljana (Slovenia). In this study, data collected at Slovenia from 20 of the 24 participants (mean age of 28 years) was used as training dataset [25] and data collected in Germany from the 56 participants were used as evaluation dataset. Only the data collected from the IMU worn on the left foot was used for trajectory reconstruction of ten activities (Table 1). Sensor placement and axis alignment are the same used in the Unicauca dataset (Figure 2). The acceleration (range: ±8 g) and angular velocity (range: ±2000 dps) were collected at a frequency of 200 Hz. The on-ground and off-ground phases of each stride are labeled. The accelerometer was calibrated using six static positions and the gyroscope was calibrated using a complete rotation about each of the three axes. Data were acquired in an indoor environment which including chairs and tables (Figure 3). Jogging was described to the participants as “if one would jog for exercise in the evening” and running as “if one is late for a bus”. These instructions were the same used in the Unicauca dataset.

3. Methods

A trajectory reconstruction pipeline was carried out separately for each activity of both datasets (Figure 4). This pipeline is based on previous work by Hannink et al. [26]. A type of activity classification step was included. Toe-Off and Mid-Stance algorithms were modified in order to deal with non-walking strides as well as a complementary filter added for stride length and orientation estimation.

3.1. Stride Segmentation

As shown by Zrenner et al., a threshold-based stride segmentation and a double integration with the ZUPT algorithm performed better than other approaches based on stride time, foot acceleration, and deep learning for calculating stride length in running using a foot-mounted IMU [27]. Thus, multi-dimensional subsequence dynamic time warping (msDTW) and a double integration with ZUPT were used as the stride segmentation and stride length and orientation estimation methods, respectively, in this study [23].
msDTW is used to find a subsequence of continuous signal sequences similar to a given reference pattern. In the context of stride segmentation, that pattern consists of a template of one stride. The stride start was set to the negative peak before the swing phase and stride end to the negative peak at the end of the stance phase (Figure 5a), according to the definition of stride given in [20]. Using that template, msDTW looks for similarities in a movement sequence. msDTW has been shown to be a robust method to segment strides from healthy, geriatric, and Parkinson’s patients using foot-mounted IMUs [28].

3.1.1. Template Generation

A MatLab script was developed for template generation. It included two steps: interpolation and averaging. Interpolation consisted of taking each stride and interpolating it to a fixed duration of 200 samples. After interpolation, the template was obtained by averaging, sample by sample, all the strides. The templates for walking, jogging, and running were built using the 8724, 1688, and 1360 walking, jogging, and running strides, respectively, of the training dataset. Unlike other studies, which used only straight strides for building templates [23,28,29], the three templates were built with all the strides of the activities. Thus, both straight and non-straight strides were included in the templates.
The swing-phase starts when the foot leaves the ground (Toe-Off) and ends when the heel strikes the ground (Heel Strike). The portion of the gyroscope z-signal after Heel Strike (HS) describes the stance-phase. A Mid-Stance (MS) event is defined as the part of the stance-phase when the signal energy is zero [30].

3.1.2. Classification of Walking, Jogging, and Running Activities

In order to automatically select the walking, jogging, or running template that will be used in the stride segmentation process, the machine learning algorithms included in the Matlab Classification Learner app were trained using the activities of the training dataset. A window size of 200 samples (1 s of data) and an overlap of 100 samples were used for feature extraction. The features extracted were velocity (by integrating accelerometer readings), angular velocity (by integrating gyroscope readings) and energy of accelerometer and gyroscope axes. The most frequent value in the result was chosen as the final classification. The evaluation was performed using ten-fold cross-validation. As a result, the highest accuracy (98.1%) was achieved by the SVM classifier with a polynomial kernel function of third-order.

3.1.3. Multi-Subsequence Dynamic Time Warping Implementation

The output of the stride segmentation based on msDTW is a set of segments [31]. Each segment describes a possible stride. One issue using these resulting segments for trajectory reconstruction is that often the end of a segment does not coincide with the start of the next segment even for consecutive strides (Figure 6a). The solution to this issue is based on the Toe-Off (TO) detection, which is described in the next section. Using the templates (Figure 5a), the first event detected in each stride is TO. For this reason, TO was defined as the beginning of a stride. For consecutive strides, the end of the stride corresponds with the beginning of the next stride (next TO), resulting in a stride segmentation without “holes” (Figure 6b).
The precision and sensitivity of the stride segmentation using msDTW can be tuned using a threshold. The threshold needed to detect a stride indicates the similarity between that stride and the template used, that is, a large threshold indicates a large difference between the template and the segmented stride [23]. Therefore, with a very small threshold, the number of false negatives strides would increase, and a very large threshold would generate false positives strides. Thresholds from 0 to 100 in steps of 5 were tested on the training dataset. As a result, it was found that a fixed threshold of 65 maximizes the F-score of the stride segmentation in walking, jogging, and running activities (Figure 7).

3.2. Toe-Off and Mid-Stance Detection

The previous algorithms for TO and MS detection [31] were modified in order to improve detection accuracy in jogging and running. These modifications are described in this section. Both previous and proposed algorithms use the signal of the gyroscope z-axis for TO and MS detection.

3.2.1. To Detection

At TO, the gyroscope z-axis describes a zero-crossing because of the ankle joint changes from plantar flexion to a dorsal extension position in the sagittal plane [23]. The algorithm included in [31] for TO detection consists of detecting the first zero-crossing in the gyroscope z-axis. Due to the abrupt movements in jogging and running strides, in a few cases, a peak located at the beginning of the stride causes a zero crossing. This would lead to a wrong TO detection (red circle in Figure 8). Consequently, the adapted algorithm for TO detection (Algorithm 1) find the maximum peak of the signal and then find the nearest zero crossing before it (blue circle in Figure 8). After the detection of all the TOs that belong to the activity, all the portions corresponding from TO to TO are considered as strides (Figure 6b). Considering that the stride time of walking strides is around one second [24], if one TO to TO portion is greater than 2 s (400 samples), only the signal until 1.5 s was taken into account. This often happens because the participant is standing still or sitting.
Algorithm 1: Toe-off (TO) detection algorithm.
1: xMP ← getMaximumPeak(stride)
2: xZC ← getZeroCrossings(stride(1 : xMP))
3: TO ← getNearestZCtoMP(xZC, xMP)

3.2.2. Mid-Stance Detection

At Mid-Stance (MS) we define that the foot is entirely stationary on the ground [23,28] and its velocity is zero. The gyroscope z-signal is minimal at that moment. As the speed of movement increases from walking to running, the stance-phase time decreases (Figure 5a) making MS detection more difficult [10]. The previous algorithm for MS detection in walking strides consists of calculating the middle of the window with the lowest energy in the full stride’s gyroscope z-signal [23,28,31]. For jogging and running strides, the MS is often confused with other parts of the signal like the valley just before the HS or the peak before the next TO (red square in Figure 8).
The adaptation of the MS detection algorithm (Algorithm 2) consisted of (1) taking only the stride portion from HS to 80% of the stride—this portion was chosen taking into account that the stance-phase of walking strides is approximately the last 60% of the stride and for jogging and running strides it is approximately the last 40% of the stride [25]; (2) calculating the middle of the window with the lowest energy within that portion—to this end, a window size of 20 samples (100 ms) and a window overlap of 10 samples (Blue square in Figure 8) are used.
Algorithm 2: Mid-Stance (MS) detection algorithm.
1: windowSize ← 20
2: overlap ← 10
3: stride ← interpolateStrideTo200Samples(stride)
4: xMP ← getMaximumPeak(stride)
5: stride ← stride(xMP : 160)
6: xHS ← getMinimumPeak(stride)
7: stride ← stride(xHS : end)
8: MS getMinimumEnergy(stride, windowSize, overlap)

3.2.3. Stride Length and Orientation Estimation

The biggest challenge to adequately estimate stride length using IMU data is the significant bias derived from the use of IMUs, which leads to large drifts after the double-integration process. For that reason, the ZUPT method was used. Zero-velocity detection was done by evaluating a threshold on the magnitude of the gyroscope rate of turn of each measurement. If the measurement is less than a threshold of 0.6 dps, that measurement is considered as a zero-velocity measurement. It has been proved that this simple approach works properly in walking strides [11,30]. However, this approach does not work correctly in jogging and running strides due to the abrupt signal variations. The solution to this problem is the use of the MS detected previously. Taking into account that the average stance-phase time in running strides is around 100 ms (20 samples), it was empirically found that taking 5 samples to each side of the MS (which corresponds to 50 ms with the sampling frequency used) leads to better zero-velocity detection in jogging and running strides.
After zero-velocity detection, a complementary Kalman filter (CF) was used in order to model the error in velocity and position estimates using the ZUPTs as measurements (see Appendix A for details). When zero-velocity is detected, but the estimated velocity is different to zero, the CF adjusts the velocity and the corresponding displacement. The CF used in this work is based on the proposed work by Fischer et al. [11]. Three main parameters have to be set up for CF initialization: accelerometer and gyroscope noise ( σ a and σ w ) and the ZUPT detection noise ( σ v ). Accelerometer and gyroscope noise were set to equal value in both datasets ( σ a = 0.01   m / s 2 and σ w = 0.01   r a d / s ). ZUPT detection noise depends on the velocity of the participant. That parameter was established by evaluating from σ v = 0.001   m / s to σ v = 0.05   m / s in steps of 0.001   m / s for each trajectory performed. The σ v chosen was the one that produced the least error in the final distance evaluated. The stride length and orientation estimation are obtained using the position increments in each MS event. Stride length, where P k is the position increment from stride k-1 to stride k, is calculated as follows:
S L k =   P k ( x ) 2 +   P k ( y ) 2 ,

4. Results

4.1. Unicauca Dataset

4.1.1. Classification of the Type of Activity

The accuracy in the activity classification was 90%. There were only three misclassifications: two running activities were classified as jogging activities and one jogging activity was classified as a running activity (Figure 9).

4.1.2. Toe-Off and Mid-Stance Detection

In this dataset, TO and MS were manually labeled. A TO/MS is considered as a true positive (TP) if it is located within 15% of the total number of samples of the stride to the right and left of the TO/MS ground truth. A false positive (FP) occurs when a TO/MS is detected outside this range. A false negative (FN) indicates that a TO/MS for a stride was not detected. Having in mind that 40% and 60% of the stride corresponds to the stance-phase of walking and running strides, respectively [25], the TO detection performance was evaluated in the training dataset using error ranges from 5% to 21% of the total stride in steps of 3% (Figure 10). As a result, 15% was chosen as an acceptable error range for TP calculation.
Results of the evaluation of the TO and MS detection using the previous and proposed algorithms are shown in Table 2 and Table 3, respectively.
A perfect F-score was obtained for TO and MS detection in walking strides. Very few mistakes occurred for jogging and running, but the F-score remains high.

4.1.3. Trajectory Reconstruction

Two evaluation measures were used. (1) Return position error (RPE): the distance between the coordinates of the actual final point of the activity and the coordinates of the participant’s final stride of the corresponding activity. (2) Strides out of trajectory (SOT): All strides of the reconstructed trajectory should be within the boundaries of the corridors represented by black dotted lines (Figure 11). Otherwise, those strides will be counted as out of trajectory.
Higher velocity corresponds to more SOT and RPE. Although, on average, 5.7 % of the strides are out of trajectory in the running trial, the RPE remains less than 1.0% (Table 4). Trajectories of the three trials are mostly within the boundaries (Figure 11).

4.2. FAU Dataset

4.2.1. Classification of the Type of Activity

The accuracy obtained by the SVM classifier was 93%. Most of the misclassifications occurred when classifying between running and jogging (Figure 12).

4.2.2. Toe-Off Detection

The last sample of the on-ground phase of each stride was used as ground truth for the evaluation of the TO detection algorithm (Table 5). The same criteria used in the Unicauca dataset for TP, FP, and FN calculations were used. The evaluation was carried out on the data collected from the 56 participants at the Friedrich-Alexander University Erlangen-Nürnberg (Germany) of FAU dataset.

4.2.3. Body Trajectory Reconstruction

For RPE estimation in FAU dataset (Table 6), it is important to note that the start/end activity positions were defined by chairs in the indoor environment. For that reason, the actual positions where the participants started and finished the activities were not precisely the same as the chairs’ positions since participants began each activity near the corresponding chair and did not necessarily return to the exact point where they started the activity. Based on the videos of the data collection, participants started and finished the activities within a radius of 1.5 m around the chairs. Light blue and gray rectangles in Figure 13 and Figure 14, respectively, indicate the path where all the strides related to a certain activity should take place. If a stride is out of this path, it is considered as a Stride Out of Trajectory (SOT). A SOT can be caused by the accumulative error of stride lengths and angle calculation of previous strides. These zones were defined taking into account the coordinates of the chairs and tables and the boundaries of the indoor environment.
Most of the trajectories were inside the zones (Figure 13 and Figure 14). The trajectory reconstruction of activities W-20, J-20, and R-20 describes two straight trajectories, joined by a 180-degree turn. The trajectory reconstruction of W-Slalom allows sight of the area where the tables are located. The W-Posters activity includes non-straight strides, which are well described in the trajectory obtained. Regarding the circuit activities, although most of the strides are inside the activity zones, some trajectories lead towards the outer part of the activity zone. Others lead towards the internal part of the circuit (Figure 13).

5. Discussion

We have proposed a pipeline for indoor trajectory reconstruction of walking, jogging, and running activities. The proposed pipeline was evaluated with two datasets. The results showed that it is able to reconstruct a person’s trajectory regardless of their gait speed.

5.1. Classification of the Type of Activity

It was found that the classification model obtained with the SVM algorithm is able to classify the three types of activities performed: walking, jogging, and running. The classification between jogging and running is the one in which the classifier made more mistakes. This is possibly due to the jogging and running speeds of some participants being similar. The use of personal models to avoid this problem could be promising.

5.2. TO and MS Detection

Previous studies focused on the reconstruction of the trajectory during walking and running and do not show results of segmentation or detection of strides [18,19,20,21,22]. The two datasets used in this study allow TO evaluation. In the case of MS detection, ground truth information was not available in the FAU dataset. Therefore, it was not possible to evaluate MS detection in that dataset. However, a high F-score was obtained in the detection of MS in the Unicauca dataset.
While the F-score obtained for the proposed TO and MS detection algorithms is similar to that obtained for the previous algorithms for walking activities, the F-score achieved for the proposed TO and MS detection algorithms outperformed that achieved for the previous algorithms for all jogging and running activities. That suggests that the proposed algorithms can detect those gait events in walking, jogging, and running strides. The number of false positives (FP) was always higher than the number of false negatives (FN). This could indicate that the threshold used for stride segmentation with msDTW might have been overestimated, since stride segmentation using a large threshold implies that there is a large difference between the template used and the segmented strides, leading to the detection of FP strides. However, it was checked that by reducing that threshold, the number of FN increased, causing a decrease in the F-score. Threshold-free methods based on machine learning techniques such as those used by Ren [20] and Wagstaff [22] would make the stride segmentation process straightforward by avoiding setting any threshold.
The lowest F-scores are obtained for three walking activities: W-Posters, W-Tables, and W-Cards, which might be due to the fact that those activities involve non-stride movements such as stopping, sitting, lateral and backward steps. This could be because the signal generated for those foot movements is different from the walking/running templates. This could be accounted for by using templates generated by those specific movements, as previously demonstrated in [29], where specific templates were generated for each specific activity such as ascending and descending stairs. Unfortunately, the wide range of possible natural foot movements makes this alternative hard to implement. A hierarchical hidden Markov model (hHMM) approach has proved to be a robust method for stride segmentation of walking activities that include non-stride movements in Parkinson’s patients [14] and for stride segmentation of jogging activities [15]. Furthermore, hHMM is a threshold-free approach, therefore it should be explored in order to improve the results obtained for the walking activities that include non-stride movements such as W-Posters, W-Tables, and W-Cards, as well as for stride segmentation of jogging and running activities.

5.3. Trajectory Reconstruction

Usually, the foot-mounted IPDR systems have been evaluated in closed-loop trajectories and by measuring the Return Position Error (RPE) [18,19,20,21,22]. The purpose of the Unicauca dataset was, therefore, to provide a starting point to allow a fair comparison with the state-of-the-art papers.
Sometimes the RPE is small, although the reconstructed trajectory does not fit the actual trajectory performed by the person. That is why we proposed the number of strides out of the trajectory as an additional evaluation metric. The RPEs obtained with the pipeline proposed in this paper for the three trials collected in the Unicauca dataset are less than 1%. The results obtained by the works described in the literature review section are also lower than 1%.
As a result of the better detection of TO and MS obtained by using the algorithms proposed in this study, there is also a better trajectory reconstruction since there were fewer strides out of trajectory (SOT) and shorter RPE for jogging and running activities. This demonstrates two things. The first is the importance of performing a correct detection of TO and MS for trajectory reconstruction. The second is that if the complementary filter does not have precise data to perform the ZVUs, it is not capable of modeling errors in speed on its own, even if its parameters were tuned. It has also been demonstrated that by properly detecting TO and MS, the complementary filter is capable of modeling errors in walking, jogging, and running strides.
RPE obtained for trajectories in the FAU dataset are higher than for the Unicauca dataset. It is important to highlight two limitations that the FAU dataset has for trajectory reconstruction. Firstly, the position of the participants at the beginning and end of the activities is not exactly the same. When analyzing the videos of the FAU dataset collection, it was concluded that these positions vary approximately in a radius of one and a half meters, taking as reference the chairs that indicated the start and end of the activities. Therefore, the RPEs calculated have an error of ±1.5 m. This fact should be taken into account for the preparation of the protocol for the collection of a future dataset. Secondly, it was not possible to subtract the gyroscope bias in all activities performed in the FAU dataset, because the activities were performed continuously. A prerequisite for bias computation is that the person stands still for a few seconds for the calculation of the mean of the gyroscope readings and then subtracting it from the entire movement sequence.
The number of strides out of trajectory is directly related to the RPE obtained; the more strides out of the acceptable path range, the higher the RPE. When observing the trajectory reconstruction of the activities W-20, J-20, R-20, and W-Circuit, J-Circuit, R-Circuit, it appears that the difficulty in trajectory reconstruction increases with stride velocity (from walking to jogging and running). This also occurred in the five papers described in the literature review section [18,19,20,21,22]. In those papers, the evaluation was performed with very few people. From our study, we can confirm that there is still a gap in trajectory reconstruction using foot-mounted IPDR systems of jogging/running activities regarding the trajectory reconstruction of walking activities.
The RPE of the trajectory reconstruction of W-Cards, W-Tables, and W-Posters activities are particularly high, due to the bad detection of TOs. These activities should be treated with special care in future works since they describe movements of daily living activities that happen frequently.
The trajectories obtained have a very well-defined shape and could be used for mapping an indoor environment.
One important recommendation for future work in the field of trajectory reconstruction using IPDR systems is that the datasets collected for evaluation are labeled at activity and stride/step levels, as the FAU dataset used in this paper. Additionally, the participants of the data collection process must start and end precisely at the indicated coordinates.

6. Conclusions

In this paper, we have proposed and evaluated a pipeline for trajectory reconstruction of walking, jogging, and running activities using a foot-mounted inertial pedestrian dead-reckoning system. The dynamic time warping method was adapted within this paper to segment walking, jogging, and running strides. Stride length and orientation estimation were performed using a zero velocity update algorithm adapted for walking, jogging, and running strides and empowered by a complementary Kalman filter.
The presented results showed that the proposed pipeline provides good trajectory estimations during walking, jogging, and running. TO detection algorithm reached an F-score between 92% and 100% for activities that do not involve stopping, and between 67% and 70% otherwise. Resulting return distance errors were in the range of 0.51% to 8.67% for non-stopping activities and 8.79% to 27.36% otherwise.
To the best of the authors’ knowledge, this is the most comprehensive evaluation of a foot-mounted IPDR system regarding the type and number of activities and quantity of people included in the datasets and can serve as a baseline for the comparison of future systems. Future work will be focused on using hidden Markov models in order to improve stride segmentation and fusing symbolic location from an RSSI signal to update the indoor localization when possible.

Author Contributions

Conceptualization, J.D.C., C.F.M., D.M.L., F.K. and B.M.E.; Formal analysis, J.D.C. and D.M.L.; Resources, C.F.M., F.K. and B.M.E.; Software, J.D.C.; Writing—original draft, J.D.C. and D.M.L.; Writing—review & editing, C.F.M., F.K. and B.M.E. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Departamento Administrativo de Ciencia, Tecnología e Innovación (COLCIENCIAS), (Call 727-2015).

Acknowledgments

Jesus Ceron gratefully acknowledges the support of the Departamento Administrativo de Ciencia, Tecnología e Innovación (COLCIENCIAS) within the national doctoral grants, call 727-2015. Bjoern Eskofier gratefully acknowledges the support of the German Research Foundation (DFG) within the framework of the Heisenberg professorship programme (grant number ES 434/8-1).

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Appendix A. Complementary Filter

The initialization of the Complementary Filter (CF) implies to establish a series of matrices. First, the state of the CF includes the errors in orientation, position, and velocity. (A1) shows the state in an array representation. Each array element is a 1 × 3 array containing the errors in the three-axis.
E = [ E o   E p   E v ]
The error covariance matrix accumulates the error in orientation, position, and velocity produced in each sample k:
P k = [ 0 9 x 9 ]
The state transition function is a matrix that is multiplied with the previous state to get the next state, as shown in (A7). ‘S’ is the Skew-symmetric cross-product operator matrix formed from the n-frame accelerations and is the time step equals to 0.005 s, which results from dividing 1 s between the IMU data collection frequency (200Hz).
F k =   I 3 X 3 0 3 x 3 0 3 x 3 0 3 x 3 I 3 X 3 I 3 X 3 t S t 0 3 x 3 I 3 X 3
The process noise covariance matrix is calculated for each sample by multiplying the accelerometer and gyroscope noise by:
Q k =   [ ( σ w x   σ w y   σ w z   0   0   0   σ a x   σ a y   σ a z ) t ]
The uncertainty in velocity during each ZUPT is represented using the measurement noise covariance matrix (A5). It is a diagonal matrix because no correlation in velocity is supposed to exist between axes.
R =   σ v x 2 0 0 0 σ v y 2 0 0 0 σ v z 2
The measurement function matrix is used to move from the state variables space to the measurement variables states. In this implementation, the measurements are the ZUPTs that is when velocity is supposed to be zero. That way, the measurement function has to contain an identity matrix in the position of the velocity error state as follows:
H k =   [ ( 0 3 x 3   0 3 x 3   I 3 x 3 ) ]
Before running the CF, the gyroscope bias has to be removed. Gyroscope bias is obtained by calculating the mean of the gyroscope readings while IMU is not moving just before the beginning of the activity. The resulting value is subtracted to all gyroscope signals.
After gyroscope bias subtraction, the CF is executed. It has two phases: Prediction and update. In the prediction phase, the error covariance matrix ( P k ) is propagated using (A7):
P k =   F k P k 1 F k T + Q k
Only when a sample k is a ZUPT, the Update phase comes into play. In this case, the Kalman gain is calculated with (A8), and with that gain, the error is obtained using (A9).
K k =   P k H T ( H P k H T + R ) 1
E = [ E o   E p   E v ] =   K k V k
Finally, the velocity and position estimates are corrected as well as Pk:
V k =   V k   E v
P o s k =   P o s k   E p
P k = ( I 9 x 9 K k H ) P k

References

  1. Zheng, L.; Zhou, W.; Tang, W.; Zheng, X.; Peng, A.; Zheng, H. A 3D indoor positioning system based on low-cost MEMS sensors. Simul. Model. Pract. Theory 2016, 65, 45–56. [Google Scholar] [CrossRef]
  2. Susanti, R.M.; Adhinugraha, K.M.; Alamri, S.; Barolli, L.; Taniar, D. Indoor Trajectory Reconstruction Using Mobile Devices. In Proceedings of the IEEE 32nd International Conference on Advanced Information Networking and Applications (AINA), Krakow, Poland, 16–18 May 2018. [Google Scholar]
  3. Alarifi, A.; Al-Salman, A.; Alsaleh, M.; Alnafessah, A.; Al-Hadhrami, S.; Al-Ammar, M.A.; Al-Khalifa, H.S. Ultra Wideband Indoor Positioning Technologies: Analysis and Recent Advances. Sensors 2016, 16, 707. [Google Scholar] [CrossRef] [PubMed]
  4. Leong, C.Y.; Perumal, T.; Peng, K.W.; Yaakob, R. Enabling Indoor Localization with Internet of Things (IoT). In Proceedings of the IEEE 7th Global Conference on Consumer Electronics (GCCE), Nara, Japan, 9–12 October 2018; pp. 571–573. [Google Scholar]
  5. Correa, A.; Barcelo, M.; Morell, A.; Vicario, J.L. A Review of Pedestrian Indoor Positioning Systems for Mass Market Applications. Sensors 2017, 17, 1927. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Farid, Z.; Nordin, R.; Ismail, M. Recent Advances in Wireless Indoor Localization Techniques and System. J. Comput. Netw. Commun. 2013, 2013, 185138. [Google Scholar] [CrossRef]
  7. Mainetti, L.; Patrono, L.; Sergi, I. A survey on indoor positioning systems. In Proceedings of the 22nd International Conference on Software, Telecommunications and Computer Networks (SoftCOM), Split, Croatia, 17–19 September 2014; pp. 111–120. [Google Scholar]
  8. Muhammad, M.N.; Salcic, Z.; Wang, K.I.-K. Indoor Pedestrian Tracking Using Consumer-Grade Inertial Sensors with PZTD Heading Correction. IEEE Sens. J. 2018, 18, 5164–5172. [Google Scholar] [CrossRef]
  9. Harle, R. A Survey of Indoor Inertial Positioning Systems for Pedestrians. IEEE Commun. Surv. Tutor. 2013, 15, 1281–1293. [Google Scholar] [CrossRef]
  10. Wu, Y.; Zhu, H.B.; Du, Q.X.; Tang, S.M. A Survey of the Research Status of Pedestrian Dead Reckoning Systems Based on Inertial Sensors. Int. J. Autom. Comput. 2019, 16, 65–83. [Google Scholar] [CrossRef]
  11. Fischer, C.; Sukumar, P.T.; Hazas, M. Tutorial: Implementing a pedestrian tracker using inertial sensors. IEEE Pervasive Comput. 2012, 12, 17–27. [Google Scholar] [CrossRef]
  12. Li, Y.; Wang, J.J. A robust pedestrian navigation algorithm with low cost IMU. In Proceedings of the International Conference on Indoor Positioning and Indoor Navigation (IPIN), Sydney, Australia, 13–15 November 2012; pp. 1–7. [Google Scholar]
  13. Li, Y.; Wang, J.J. A Pedestrian Navigation System Based on Low Cost IMU. J. Navig. 2014, 67, 929–949. [Google Scholar] [CrossRef] [Green Version]
  14. Ren, M.; Pan, K.; Liu, Y.; Guo, H.; Zhang, X.; Wang, P. A Novel Pedestrian Navigation Algorithm for a Foot-Mounted Inertial-Sensor-Based System. Sensors 2016, 16, 139. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Wagstaff, B.; Peretroukhin, V.; Kelly, J. Improving foot-mounted inertial navigation through real-time motion classification. In Proceedings of the International Conference on Indoor Positioning and Indoor Navigation (IPIN), Sapporo, Japan, 18–21 September 2017; pp. 1–8. [Google Scholar]
  16. Wagstaff, B.; Kelly, J. LSTM-Based Zero-Velocity Detection for Robust Inertial Navigation. In Proceedings of the International Conference on Indoor Positioning and Indoor Navigation (IPIN), Nantes, France, 24–27 September 2018; pp. 1–8. [Google Scholar]
  17. Mannini, A.; Sabatini, A.M. Machine Learning Methods for Classifying Human Physical Activity from On-Body Accelerometers. Sensors 2010, 10, 1154–1175. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Hannink, J.; Kautz, T.; Pasluosta, C.F.; Gasmann, K.G.; Klucken, J.; Eskofier, B.M. Sensor-Based Gait Parameter Extraction with Deep Convolutional Neural Networks. IEEE J. Biomed. Health Inform. 2016, 21, 85–93. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Ghassemi, N.H.; Hannink, J.; Martindale, C.F.; Gaßner, H.; Muller, M.; Klucken, J.; Eskofier, B.M. Segmentation of Gait Sequences in Sensor-Based Movement Analysis: A Comparison of Methods in Parkinson’s Disease. Sensors 2018, 18, 145. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Mannini, A.; Sabatini, A.M. Gait phase detection and discrimination between walking–jogging activities using hidden Markov models applied to foot motion data from a gyroscope. Gait Posture 2012, 36, 657–661. [Google Scholar] [CrossRef] [PubMed]
  21. Stetter, B.J.; Ringhof, S.; Krafft, F.C.; Sell, S.; Stein, T. Estimation of Knee Joint Forces in Sport Movements Using Wearable Sensors and Machine Learning. Sensors 2019, 19, 3690. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Wouda, F.J.; Giuberti, M.; Bellusci, G.; Maartens, E.; Reenalda, J.; Van Beijnum, B.-J.F.; Veltink, P.H. Estimation of Vertical Ground Reaction Forces and Sagittal Knee Kinematics During Running Using Three Inertial Sensors. Front. Physiol. 2018, 9, 218. [Google Scholar] [CrossRef] [PubMed]
  23. Barth, J.; Oberndorfer, C.; Kugler, P.; Schuldhaus, D.; Winkler, J.; Klucken, J.; Eskofier, B.; Barth, J. Subsequence dynamic time warping as a method for robust step segmentation using gyroscope signals of daily life activities. In Proceedings of the 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Osaka, Japan, 3–7 July 2013; Volume 2013, pp. 6744–6747. [Google Scholar]
  24. Martindale, C.F.; Sprager, S.; Eskofier, B.M. Hidden Markov Model-Based Smart Annotation for Benchmark Cyclic Activity Recognition Database Using Wearables. Sensors 2019, 19, 1820. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Martindale, C.F.; Roth, N.; Hannink, J.; Sprager, S.; Eskofier, B.M. Smart Annotation Tool for Multi-sensor Gait-based Daily Activity Data. In Proceedings of the IEEE International Conference on Pervasive Computing and Communications Workshops (PerCom Workshops), Athens, Greece, 19–23 March 2018; Institute of Electrical and Electronics Engineers (IEEE): Piscataway, NJ, USA, 2018; pp. 549–554. [Google Scholar]
  26. Hannink, J.; Ollenschläger, M.; Kluge, F.; Roth, N.; Klucken, J.; Eskofier, B.M. Benchmarking Foot Trajectory Estimation Methods for Mobile Gait Analysis. Sensors 2017, 17, 1940. [Google Scholar] [CrossRef] [PubMed]
  27. Zrenner, M.; Gradl, S.; Jensen, U.; Ullrich, M.; Eskofier, B.M. Comparison of Different Algorithms for Calculating Velocity and Stride Length in Running Using Inertial Measurement Units. Sensors 2018, 18, 4194. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Barth, J.; Oberndorfer, C.; Pasluosta, C.; Schülein, S.; Gassner, H.; Reinfelder, S.; Kugler, P.; Schuldhaus, D.; Winkler, J.; Klucken, J.; et al. Stride Segmentation during Free Walk Movements Using Multi-Dimensional Subsequence Dynamic Time Warping on Inertial Sensor Data. Sensors 2015, 15, 6419–6440. [Google Scholar] [CrossRef] [PubMed]
  29. Leutheuser, H.; Doelfel, S.; Schuldhaus, D.; Reinfelder, S.; Eskofier, B.M. Performance Comparison of Two Step Segmentation Algorithms Using Different Step Activities. In Proceedings of the 11th International Conference on Wearable and Implantable Body Sensor Networks, Zurich, Switzerland, 16–19 June 2014; pp. 143–148. [Google Scholar]
  30. Skog, I.; Nilsson, J.-O.; Händel, P. Evaluation of zero-velocity detectors for foot-mounted inertial navigation systems. In Proceedings of the International Conference on Indoor Positioning and Indoor Navigation, Zurich, Switzerland, 15–17 September 2010; pp. 1–6. [Google Scholar]
  31. Hannink, J. Mobile Gait Analysis: From Prototype towards Clinical Grade Wearable; FAU University Press: Erlangen, Germany, 2019. [Google Scholar]
Figure 1. Illustration of the path used for walking, jogging and running in the Unicauca dataset. It is a “P” shaped path. The dotted red line represents the trajectory followed by one person, dotted black lines show outer edges (walls) of the path, and the blue square shows the start and end point of the trajectory.
Figure 1. Illustration of the path used for walking, jogging and running in the Unicauca dataset. It is a “P” shaped path. The dotted red line represents the trajectory followed by one person, dotted black lines show outer edges (walls) of the path, and the blue square shows the start and end point of the trajectory.
Sensors 20 00651 g001
Figure 2. IMU sensor placement and axis alignment. (a) Accelerometer. (b) Gyroscope.
Figure 2. IMU sensor placement and axis alignment. (a) Accelerometer. (b) Gyroscope.
Sensors 20 00651 g002
Figure 3. Map of the indoor environment used for collecting the FAU dataset. Blue squares represent chairs that denote start/end positions of activities. Black rectangles represent tables, and dotted red lines represent the possible trajectories followed by participants in each activity.
Figure 3. Map of the indoor environment used for collecting the FAU dataset. Blue squares represent chairs that denote start/end positions of activities. Black rectangles represent tables, and dotted red lines represent the possible trajectories followed by participants in each activity.
Sensors 20 00651 g003
Figure 4. Pipeline for trajectory reconstruction for each activity.
Figure 4. Pipeline for trajectory reconstruction for each activity.
Sensors 20 00651 g004
Figure 5. (a) Walking, jogging, and running templates (gyroscope z-axis). (b) Running stride example (gyroscope z-axis).
Figure 5. (a) Walking, jogging, and running templates (gyroscope z-axis). (b) Running stride example (gyroscope z-axis).
Sensors 20 00651 g005
Figure 6. (a) Result of stride segmentation with msDTW. (b) Final stride segmentation with TO detection. Blue vertical lines depict TOs. Light blue rectangles are segmented strides.
Figure 6. (a) Result of stride segmentation with msDTW. (b) Final stride segmentation with TO detection. Blue vertical lines depict TOs. Light blue rectangles are segmented strides.
Sensors 20 00651 g006
Figure 7. Threshold choice for stride segmentation of walking, jogging, and running strides using msDTW.
Figure 7. Threshold choice for stride segmentation of walking, jogging, and running strides using msDTW.
Sensors 20 00651 g007
Figure 8. Example of TO and MS detection. The red circle and square show a wrong TO and MS detection, respectively, using the previous TO detection algorithm. The blue circle and square show an adequate TO and MS detection, respectively, using the proposed algorithms.
Figure 8. Example of TO and MS detection. The red circle and square show a wrong TO and MS detection, respectively, using the previous TO detection algorithm. The blue circle and square show an adequate TO and MS detection, respectively, using the proposed algorithms.
Sensors 20 00651 g008
Figure 9. Confusion matrix of the classification of the type of activity in the Unicauca dataset.
Figure 9. Confusion matrix of the classification of the type of activity in the Unicauca dataset.
Sensors 20 00651 g009
Figure 10. TO performance evaluation using error ranges from 5% to 21% in steps of 3%.
Figure 10. TO performance evaluation using error ranges from 5% to 21% in steps of 3%.
Sensors 20 00651 g010
Figure 11. Trajectory reconstruction for the ten participants of the Unicauca dataset in a P shaped path. Black dotted lines show outer edges (walls) of the possible path. Gray lines are the trajectories reconstructed of the ten participants by using the proposed pipeline.
Figure 11. Trajectory reconstruction for the ten participants of the Unicauca dataset in a P shaped path. Black dotted lines show outer edges (walls) of the possible path. Gray lines are the trajectories reconstructed of the ten participants by using the proposed pipeline.
Sensors 20 00651 g011
Figure 12. Confusion matrix of the classification of the type of activity classification in the FAU dataset activities.
Figure 12. Confusion matrix of the classification of the type of activity classification in the FAU dataset activities.
Sensors 20 00651 g012
Figure 13. Trajectory reconstruction of non-circuit activities for all 56 participants of the FAU dataset. Black, blue and orange lines denote R-20, J-20, and W-20, respectively. Red, green, violet and light green lines represent W-Cards, W-Slalom, W-Posters, and W-Tables, respectively. Gray rectangles represent zones where all the strides related to certain activity should take place.
Figure 13. Trajectory reconstruction of non-circuit activities for all 56 participants of the FAU dataset. Black, blue and orange lines denote R-20, J-20, and W-20, respectively. Red, green, violet and light green lines represent W-Cards, W-Slalom, W-Posters, and W-Tables, respectively. Gray rectangles represent zones where all the strides related to certain activity should take place.
Sensors 20 00651 g013
Figure 14. Trajectory reconstruction of circuit activities for all 56 participants of the FAU dataset. Black lines denote the trajectory follows by the participants. Gray zones represent the zone where all the strides should take place.
Figure 14. Trajectory reconstruction of circuit activities for all 56 participants of the FAU dataset. Black lines denote the trajectory follows by the participants. Gray zones represent the zone where all the strides should take place.
Sensors 20 00651 g014
Table 1. Activity descriptions and abbreviations, shown with their relevant start and end points as labeled in Figure 3 as well as approximated distances.
Table 1. Activity descriptions and abbreviations, shown with their relevant start and end points as labeled in Figure 3 as well as approximated distances.
ActivityDescriptionStart/End PositionApproximated Distance (m)
W-SlalomWalk slalom through 3 tablesB→B31
W-PostersSign name on 5 posters on the wallC→G21
W-TablesPerform task at 3 different tables while sittingD→D20
W-CardsPerform task on a table while standingE→E6
W, J, R-20Walk, jog, run 2 times 20 mA→A40
W, J, R-CircuitWalk, jog and run half a circuit eachF, G→G,F43
Table 2. Averaged results of TO and MS detection for the 10 participants in the Unicauca dataset using the previous TO and MS detection algorithms.
Table 2. Averaged results of TO and MS detection for the 10 participants in the Unicauca dataset using the previous TO and MS detection algorithms.
Toe-OffMid-Stance
ActivityTO GTTPFPFNF-Score (%)MS GTTPFPFNF-Score (%)
Walking105.5105.40.10.199.9104.5104.40.10.199.9
Jogging75.439.437.136.251.574.441.234.233.354.9
Running59.621.737.537.136.458.625.831.530.745.3
TO GT: ground truth TO rate. MS GT: ground truth MS rate. TP: true-positive rate. FP: false-positive rate. FN: false-negative rate.
Table 3. Averaged results of TO and MS detection for the 10 participants in the Unicauca dataset using the proposed TO and MS detection algorithms.
Table 3. Averaged results of TO and MS detection for the 10 participants in the Unicauca dataset using the proposed TO and MS detection algorithms.
Toe-OffMid-Stance
ActivityTO GTTPFPFNF-Score (%)MS GTTPFPFNF-Score (%)
Walking105.5105.500100104.5104.500100
Jogging75.475.20.10.299.874.474.400100
Running59.659.30.30.299.758.658.50.10.199.8
TO GT: ground truth TO rate. MS GT: ground truth MS rate. TP: true-positive rate. FP: false-positive rate. FN: false-negative rate.
Table 4. Average results of trajectory reconstruction for each type of activity performed by the 10 participants using the previous and the proposed TO and MS detection algorithms.
Table 4. Average results of trajectory reconstruction for each type of activity performed by the 10 participants using the previous and the proposed TO and MS detection algorithms.
ActivitySOTRPE
[31]New A[31]New A
#%#%meters%meters%
Walking1.71.61.7 1.60.80.50.8 0.5
Jogging6.68.62.9 3.82.21.41.4 0.9
Running5.39.23.3 5.72.61.61.4 0.9
SOT: strides out of trajectory, RPE: return position error, [31]: previous TO and MS detection algorithms, New A: proposed TO and MS detection algorithms.
Table 5. Average results of TO detection for each type of activity performed by the 56 participants in the FAU dataset using the previous and the proposed TO detection algorithms.
Table 5. Average results of TO detection for each type of activity performed by the 56 participants in the FAU dataset using the previous and the proposed TO detection algorithms.
ActivityTOTPFPFNF-Score (%)
[31]New A.[31]New A.[31]New A.[31]New A.
W-Slalom21.521.221.20.80.80.40.39797
W-Posters13.010.610.63.63.52.32.37778
W-Tables11.99.59.53.73.72.42.47575
W-Cards4.333.73.71.61.62.40.67171
W-2028.428.228.21.31.00.40.29998
J-2022.313.721.69.71.18.60.75696
R-2018.48.117.012.32.110.61.34890
W-Circuit28.227.627.70.70.70.40.39898
J-Circuit21.911.821.310.80.7100.54997
R-Circuit17.77.617.310.70.810.20.44096
TO: toe-off rate, TP: true positives rate, FP: false positives rate, FN: false negatives rate, [31]: previous TO and MS detection algorithms, New A: proposed TO and MS detection algorithms.
Table 6. Averaged results of trajectory reconstruction of activities performed by the 56 participants in the FAU dataset using the previous and the proposed TO and MS detection algorithms.
Table 6. Averaged results of trajectory reconstruction of activities performed by the 56 participants in the FAU dataset using the previous and the proposed TO and MS detection algorithms.
ActivityActivity distanceSOTRPE
[31]New A.[31]New A.
meters#%#%meters%meters%
W-Slalom311.15.21.15.21.75.51.75.5
W-Posters211.08.01.08.01.99.01.88.8
W-Tables203.125.93.125.92.814.12.814.1
W-Cards61.330.51.330.51.627.41.627.4
W-20403.713.13.713.11.74.21.74.2
J-20407.534.23.917.95.514.22.05.1
R-20404.122.53.117.05.213.92.56.0
W-Circuit434.114.44.114.42.96.73.06.7
J-Circuit436.430.24.520.414.533.73.68.8
R-Circuit435.129.93.922.416.237.73.78.7
SOT: strides out of trajectory, RPE: return position error, [31]: previous TO and MS detection algorithms, New A: proposed TO and MS detection algorithms.

Share and Cite

MDPI and ACS Style

Ceron, J.D.; Martindale, C.F.; López, D.M.; Kluge, F.; Eskofier, B.M. Indoor Trajectory Reconstruction of Walking, Jogging, and Running Activities Based on a Foot-Mounted Inertial Pedestrian Dead-Reckoning System. Sensors 2020, 20, 651. https://doi.org/10.3390/s20030651

AMA Style

Ceron JD, Martindale CF, López DM, Kluge F, Eskofier BM. Indoor Trajectory Reconstruction of Walking, Jogging, and Running Activities Based on a Foot-Mounted Inertial Pedestrian Dead-Reckoning System. Sensors. 2020; 20(3):651. https://doi.org/10.3390/s20030651

Chicago/Turabian Style

Ceron, Jesus D., Christine F. Martindale, Diego M. López, Felix Kluge, and Bjoern M. Eskofier. 2020. "Indoor Trajectory Reconstruction of Walking, Jogging, and Running Activities Based on a Foot-Mounted Inertial Pedestrian Dead-Reckoning System" Sensors 20, no. 3: 651. https://doi.org/10.3390/s20030651

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop