Next Article in Journal
Floating Epoch Length Improves the Accuracy of Accelerometry-Based Estimation of Coincident Oxygen Consumption
Previous Article in Journal
Enhancing Data Freshness in Air-Ground Collaborative Heterogeneous Networks through Contract Theory and Generative Diffusion-Based Mobile Edge Computing
Previous Article in Special Issue
Validity of AI-Based Gait Analysis for Simultaneous Measurement of Bilateral Lower Limb Kinematics Using a Single Video Camera
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Systematic Evaluation of Feature Encoding Techniques for Gait Analysis Using Multimodal Sensory Data

by
Rimsha Fatima
1,
Muhammad Hassan Khan
1,*,
Muhammad Adeel Nisar
2,
Rafał Doniec
3,
Muhammad Shahid Farid
1 and
Marcin Grzegorzek
4
1
Department of Computer Science, University of the Punjab, Lahore 54590, Pakistan
2
Department of Information Technology, University of the Punjab, Lahore 54000, Pakistan
3
Faculty of Biomedical Engineering, The Silesian University of Technology, 44-100 Gliwice, Poland
4
Institute of Medical Informatics, University of Lübeck, 23562 Lübeck, Germany
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(1), 75; https://doi.org/10.3390/s24010075
Submission received: 15 October 2023 / Revised: 9 December 2023 / Accepted: 19 December 2023 / Published: 22 December 2023

Abstract

:
This paper addresses the problem of feature encoding for gait analysis using multimodal time series sensory data. In recent years, the dramatic increase in the use of numerous sensors, e.g., inertial measurement unit (IMU), in our daily wearable devices has gained the interest of the research community to collect kinematic and kinetic data to analyze the gait. The most crucial step for gait analysis is to find the set of appropriate features from continuous time series data to accurately represent human locomotion. This paper presents a systematic assessment of numerous feature extraction techniques. In particular, three different feature encoding techniques are presented to encode multimodal time series sensory data. In the first technique, we utilized eighteen different handcrafted features which are extracted directly from the raw sensory data. The second technique follows the Bag-of-Visual-Words model; the raw sensory data are encoded using a pre-computed codebook and a locality-constrained linear encoding (LLC)-based feature encoding technique. We evaluated two different machine learning algorithms to assess the effectiveness of the proposed features in the encoding of raw sensory data. In the third feature encoding technique, we proposed two end-to-end deep learning models to automatically extract the features from raw sensory data. A thorough experimental evaluation is conducted on four large sensory datasets and their outcomes are compared. A comparison of the recognition results with current state-of-the-art methods demonstrates the computational efficiency and high efficacy of the proposed feature encoding method. The robustness of the proposed feature encoding technique is also evaluated to recognize human daily activities. Additionally, this paper also presents a new dataset consisting of the gait patterns of 42 individuals, gathered using IMU sensors.

1. Introduction

Gait refers to the movement patterns of an individual’s walk. It encompasses the rhythm, speed, and style of movement which require a strong coordination of the upper and lower limbs. The process of gait analysis involves evaluating an individual’s walking pattern to assess their biomechanics and identify any abnormalities or inefficiencies in their movement [1]. It has been an active research area over the last few years due to its utilization in numerous real-world applications, e.g., clinical assessment and rehabilitation, robotics, gaming, entertainment, etc. [2,3]. The quantitative gait analysis has also been explored as a biometric modality to identify a person [4]. The gait analysis has several advantages over other existing modalities; in particular, it is unobtrusive and difficult to steal or falsify. Gait analysis is a challenging task because it involves the complex coordination of human skeletal, muscular, and nervous systems. Additionally, the gait can be affected by a wide range of factors, such as age, injury, disease, and environmental conditions. This results in intra-personal variations, which are always greater than inter-personal variations.
The gait data can be gathered via a variety of sensing modalities that can be broadly divided into two groups: using sensing modalities [5] and using visual cameras [3]. Figure 1 illustrates these sensing modalities that can be used for data gathering. Each of these modalities have their own strengths and limitations, and researchers choose to use single or a combination of several modalities to capture gait data based on their research questions and goals. Vision-based systems have been extensively used in the gait analysis due to their higher precisions however, their use raises confidentiality and privacy concerns [6]. Conversely, the digital sensors such as inertial measurement units (IMU), and pressure sensors have been also used to collect the gait data. These tiny sensors can easily be embedded in our environment, including walking floor, wearable devices, and clothes. Furthermore, a few of such sensors are already embedded in our daily used digital gadgets, e.g., smart phones, smart watches, and smart glasses [7]. These devices generate data in the form of pressure signals, velocity, acceleration, and positions which can be used to represent the gait.
Analyzing gait requires the use of sophisticated tools and techniques. Since the gait comprises continuous time series data, we need to extract a set of relevant features from these raw data to represent gait patterns. There are several approaches in the literature to extract features, and they can be categorized into three groups: (1) handcrafted [8], (2) codebook-based [9], and (3) automatic deep learning-based [10] feature extraction approaches. The handcrafted features are usually a few statistical quantities, e.g., mean, variance, skewness, kurtosis, etc., that are extracted from raw data based on the prior knowledge of experts in the application domain. Later, these features are fed to machine learning algorithms for classification. Typically, they are simple to implement and are computationally efficient; however, they are designed to solve a specific problem and are not robust [7]. The codebook-based feature learning techniques follow the channel of Bag-of-Visual-Words (BOVWs) approach [11] to compute gait features. Specifically, they employ clustering algorithms, e.g., k-means, to build a dictionary (also known as codebook) using the gait sub-sequences from raw data. The data are grouped based on their underlying similarities, and the clusters’ centers are known as “codewords”. Later, a histogram-based representation is computed for other sequences by tying them to the closest-related codeword. Codebook-based techniques are proven to be robust as they can capture more complex patterns in the gait data; however, they are computationally expensive [12]. Deep learning-based techniques automatically compute discriminative features directly from the input data using artificial neural networks (ANNs). A deep network usually comprises several layers where each layer consists of artificial neurons. After obtaining input, a specific feature map is computed by the neurons in each layer and then forwarded to the next layer for further processing, and so forth. Finally, the network’s last layers generate highly abstract feature representations of the raw sensory data. A few examples of deep learning approaches for gait analysis are convolutional neural networks (CNNs) [13], recurrent neural networks (RNNs) [14], and long short-term memory networks (LSTMs) [15]. Deep learning-based approaches can automatically learn complex features from raw data and are adaptable to new datasets and scenarios; however, they are computationally much more expensive and require a huge amount of labeled instances to optimally choose the hyperparameters’ values of the deep network. The entire process is greatly hampered by the lack of prior information necessary to encode the appropriate features in the application domain and the choice of the best parameters for the machine learning algorithms [7].
This paper presents a systematic evaluation of numerous feature learning techniques to encode multimodal time series sensory data for gait analysis. Furthermore, the proposed feature encoding techniques also explored to recognize the human daily activities which were recorded using sensory data too.Specifically, we presented three different encoding techniques to recognize the walking styles. In the first technique of handcrafted features, we computed eighteen different statistical quantities from the raw sensory data, and they are fused together in a single feature vector to obtain the high-level representation. In the second feature learning technique, we build a codebook using k-means clustering algorithm, and the high-level feature representation is obtained using Locality-constraint linear encoding (LLC) [16]. In both of the aforementioned techniques, we explored the effectiveness of Support Vector Machines (SVMs) and Random Forest (RF) as machine learning algorithms to recognize different walking styles. Third, we presented two deep learning models, a CNN and an LSTM, which employed raw sensory data as input to recognize walking sequences. The effectiveness of the proposed features was assessed on four datasets, namely Gait-IMU, MHEALTH [17], WISDM [18], and UCI-HAR [19]. An extensive experimental evaluation of different feature encoding techniques was carried out along with different values of hyperparameters, and the recognition scores were compared with recent state-of-the-art approaches. The proposed framework establishes a solid foundation and generic framework to encode any time series multimodal sensory data. We believe that the raw time series sensory data of other sensing modalities such as SensFloor, skin temperature sensors, and electrocardiograms (ECGs) can also be encoded using the proposed approaches to perform recognition tasks. The proposed framework can be employed in numerous emerging fields and technologies, including healthcare and rehabilitation, sports, assistive living, and others. The major contributions of the proposed manuscript are as follows:
  • A comparative study of three different feature encoding techniques is presented.
  • A comprehensive review of the existing techniques is presented, and their benefits and drawbacks are discussed.
  • The effectiveness of various machine learning methods is evaluated in order to classify the multimodal sensory data.
  • The computational analysis of different feature encoding techniques is presented.
  • The robustness of the proposed technique is assessed on several applications, including walking styles, human activities, etc.
  • A rigorous evaluation of all the feature encoding techniques is carried out on four large datasets.
  • A large gait dataset that is collected using IMU sensors is proposed.

2. Related Work

Gait analysis is a fundamental tool in biomechanics that allows for both quantitative and qualitative assessment of human movement. It involves the analysis of the spatiotemporal parameters, kinematics, and kinetics of gait, which are important indicators of the locomotion function [20,21,22]. This paper mainly emphasizes on gait analysis using multimodal time series sensory data. The existing techniques employed either a pressure sensor [23,24,25] or an IMU [26,27] to encode walking patterns. In recent years, a large number of algorithms has been presented to investigate the movement of human body parts for clinical and behavioural assessment. They can be broadly divided into three groups, as depicted in Figure 2. In the following, a brief summary of a few techniques from each group is presented.

2.1. Handcrafted Feature-Based Techniques

These set of techniques either compute several statistical measurements on input data (e.g., average, variance, skewness) or extract more complex gait characteristics, which may include stride length, joint angles, and other related features. In the context of machine learning, handcrafted features refer to manually designed features that are extracted from raw data and used as input to a classifier [8,28,29,30]. For instance, the technique proposed in [31] extracted several statistical quantities on input data (e.g., mean, median, mode, standard deviation, skewness, and kurtosis) to show the gait fluctuation in a patient with Parkinson’s. They employed Fisher Discriminant Ratio to determine the most discriminatory statistical feature. A comparative study of different handcrafted features is proposed in [28] for the early detection of traumatic brain injury. The authors employed the location and accelerometer sensory data of a smartphone to extract nine gait features, including coefficient of variance, step count, cadence, regularity in step, stride, etc. Similarly, the authors of [32] extracted standard deviation, skewness, kurtosis, and bandwidth frequency features from the accelerometer data of an IMU sensor mounted on the subject’s lower back to distinguish between normal and stroke gait patterns. The study presented in [33] extracted thirty-eight statistical quantities, including maximum, minimum, average, spectral energy, etc., to monitor and quantify various human physical activities using a smartphone’s IMU sensory data. Similarly, the technique proposed in [34] employed frequency domain features to assess gait accelerometer signals. The approach proposed in [35] analyzed the gait sensory data using adaptive one-dimensional time invariant features.
These techniques appear simple in implementation; however, they are highly based on expert knowledge in the application domain. Additionally, they are designed to solve a specific problem and are not robust.

2.2. Codebook-Based Approaches

These techniques follow the work-flow of the BOVWs approach to encode the raw sensory input data into its compact but high-level representation. In particular, the input data are clustered based on similar patterns using a clustering algorithm (e.g., k-means) to create a dictionary. The cluster centers in the dictionary are known as codewords which describe the underlying variability in the gait data. Later, a final representation of the raw gait input data is achieved by estimating the distribution of these codewords in the gait sequence. This results in histogram-like features (where a histogram bin represents the occurrences of a specific codeword) which are fed to a classification algorithm for further analysis. For instance, the authors of [7,36] employed sensory information from commonly used wearable devices to identify human activities. They built a codebook using a k-means clustering algorithm, and the final gait representation is achieved using a simple histogram-binning approach. Similarly, the technique proposed in [37] employed IMU sensory data from smartphones and smartwatches to recognize Parkinson’s tremor using a codebook-based approach. In [38], research was carried out to recognize the human gait phase using two sensors: an accelerometer and a gyroscope. A codebook-based approach was explored to extract gait features from the raw sensory data. Similarly, the authors of [39] presented an approach to determine the best working positions for various movement phases and to guide the performer on how to keep them while performing physical activities. They explored a k-means clustering codebook to encode the different working postures for each phase of movement. In [40], a separate codebook is constructed to each sensor modality, including accelerometer, gyroscope, and magnetometer. The resulting features were concatenated to form a compact and high-level feature vector to classify human daily activities using SVM with a radial basis function (RBF) kernel. The technique in [41] presented a visual inertial system to recognize daily activities using two RGB-D (red, green, blue, and depth) detectors with a wearable inertial movement unit. They employed codebook-based technique on different sensing modalities to compute the desired features. In [12], a detailed comparison of different codebook-based feature encoding approaches is presented to recognize gait.
These techniques are proven to be robust, to some extent, as they can capture more complex patterns in the gait data; however, they are computationally expensive [12]. Additionally, this process of codebook computation may need to be performed again if new classes are added in the dataset [42].

2.3. Deep Learning-Based Approaches

Lately, automatic deep learning-based feature extraction approaches have been largely explored in different classification tasks due to their robustness in automatic feature extraction, generalization capabilities, and convincing recognition results. A deep network is typically an end-to-end architecture that can learn complicated patterns and relationships in input data using a fully automated feature extraction process. In particular, it comprises many layers of interconnected neurons. The network’s intricate and iterative structure allows for the learning of high-level features from input data as it passes it through (i.e., the weight adjustments of neurons and back-propagation of errors [43]). Numerous automatic feature learning-based techniques on gait analysis have been proposed in the past; convolutional neural networks (CNNs) [44] and long short-term memory (LSTM) [45] are a few examples to quote. For instance, the authors of [44] presented an IMU-based spectrogram technique to categorize the gait characteristics using a deep CNN. The authors of [46] explored a CNN to extract the appropriate high-level feature representation from pre-processed time series input data. They turned the input data into two-dimensional (2D) images where the y-axis represents various features and the x-axis represents the time.
Since a CNN is typically designed to process imagery data, a few other networks have also been explored to analyze gait [47,48,49]. For instance, the authors of [45] employed window-based data segmentation to classify gait sequences using a multi-model long short-term memory (MM-LSTM) network. To compute features from each window, the MM-LSTM network accepts input gait data that have been segmented into several overlapping windows of equal length. The authors of [50] employed recurrent neural networks (RNNs) for gait analysis. The method for calculating gait mechanics using an artificial neural network (ANN) with measured and simulated inertial measurement unit (IMU) data is suggested in [51]. The authors concluded that more precise estimations of gait mechanics can be obtained by combining the ANN with both simulated and actual data. In [52], a comprehensive overview of different deep learning models is presented to monitor the human gait.
The studies show that deep learning-based feature encoding techniques are effective tools for analyzing gait and have the potential to explore the underlying mechanics of gait. They provide automatic end-to-end learning; however, they also require complex computational resources and a huge amount of training data.

3. Overview of the Proposed Method

This paper presents a systematic evaluation of three different feature learning techniques to recognize the different walking styles using sensory data. The gait sequences were recorded using a set of IMU sensors. In the first method, we present eighteen different handcrafted features to represent a gait sequence. Secondly, we built a codebook using a k-means clustering algorithm, and the high-level gait features are obtained using locality-constrained linear coding (LLC). We evaluated two machine learning algorithms, namely SVM and RF, to validate the effectiveness of the aforementioned proposed features. Third, we employed two deep neural networks, namely CNN and LSTM, to automatically compute the discriminative gait features from raw sensory data. The proposed feature learning techniques were evaluated on four large datasets, and their results were compared with existing state-of-the-art techniques.

4. Proposed Feature Learning Techniques

This paper aims to present a comparative study of several feature extraction approaches to encode multimodal time series data for gait analysis. This section presents three different feature learning techniques to recognize walk styles using IMU sensory data. In the following, we briefly describe each technique.

4.1. Handcrafted Features

These features are computed either using a few simple statistical quantities (e.g., variance, skewness, etc.) or complex frequency domain-based features directly from raw sensory data [53]. We computed eighteen different handcrafted features and fused them together (via simple concatenation) to form a single vector representation. In the following, a short description of each feature is explained.
  • Maximum: Let M be a set of values. M a x ( M ) will result in m i s . t . Then m i M is the maximum value of that set.
  • Minimum: Similarly, M i n ( M ) will result in m i s . t . Then m i M is the minimum value of the set.
  • Average: Suppose a dataset M comprises n numbers, the average can then be computed as:
    A v e r a g e ( M ) = μ = i = 1 n m i n
  • Standard deviation: It is the average deviation of data points from the distribution center (i.e., μ ) and can be calculated as:
    S t d D e v ( M ) = σ = i = 1 n m i μ 2 n 1
  • Zero crossing: This feature determines how many times the specified set of values have zero crossings in the data [53]. Specifically, it can be considered a location on a mathematical function’s graph where the axes intercept, that is, when the graph crosses zero in either direction.
  • Percentiles: This feature represents the number below a specific percentage of values in data. Specifically, a qth percentile would be a value in the dataset such that, at most, (100 × q)% of the data points fall below this value and 100 × (1 − q)% of the values fall above. That is, the 25th percentile (also known as the first quartile) reveals a feature whose value is greater than 25% and less than 75% in the dataset. Similarly, the 50th and 75th percentiles are represented by second and third quartiles, respectively. In the proposed technique, we computed three distinct percentiles features: (1) percentile 20, (2) percentile 50, and (3) percentile 80.
  • Interquartile range: The first quartile value is subtracted from the third quartile value to obtain the interquartile range.
  • Kurtosis: This feature is a measure to quantify the variations in the tails of a distribution from a normal distribution [54]. A large value represents a higher extremity of deviation, i.e., outliers. It can be formulated as:
    K u r t o s i s = ( m i μ ) 4 n σ 4
  • Skewness: Skewness describes the dataset’s divergence from the normal distribution. That is, it measures the asymmetry of normal distribution in either direction.
    s k e w n e s s ( M ) = μ ˜ 3 = i = 1 n m i μ 3 n σ 3
  • Auto-correlation: It is a numerical quantity to measure the similarity between the data at time t 1 and a lagged version of data in a temporal direction (at time t 2 ). Conceptually, it estimates the correlation between the current data and previous values [55]. It can be formulated as:
    R k = i = 1 n k ( m i μ ) ( m i + k μ ) i = 1 n ( m i μ ) 2
  • Order mean values: To compute order mean values, the data are arranged in an ascending order. The smallest value in the sorted dataset corresponds to the first-order statistic. The second-order statistic is the next smallest number, and so on. In the proposed technique, we computed four distinct features using order mean values: (1) First-Order Mean (FOM), (2) Norm of FOM, (3) Second-Order Mean (SOM), and (4) Norm of SOM. We employed two of the mostly used norm techniques: L 1 (i.e., Manhattan distance) and L 2 (i.e., Euclidean distance).
  • Spectral entropy: The spectral entropy (SEP), which gauges a signal’s spectral power distribution, is based on the Shannon entropy idea. The time-based signal was transformed into its frequency spectrum using the Fourier transform. The standardized power distribution in the frequency domain is taken into account as a probability density function to calculate the signal’s Shannon entropy:
    S E P = i = 1 n F ^ ( n ) × l o g F ^ ( n )
  • Spectral energy: Since various sensory data can be used to assess walking patterns in the recorded data, they may be considered as a function whose amplitude varies with time. Similar to SEP, the time-based signal was transformed into its frequency spectrum using the Fourier transform, and the signal energy distribution over the frequency was calculated using the spectral energy formulation:
    S E = i = 1 n F ( n ) 2
    where F ( n ) is the amplitude of frequency content. It may be formulated using normalized frequency spectra:
    F ^ ( n ) = F ( n ) i = 1 n F ( n )
    All the abovementioned feature quantities were computed on each walk pattern and they were fused together in a single row to form a feature vector representation.

4.2. Codebook-Based Feature Encoding

The codebook-based feature encoding technique has shown excellent recognition results in numerous applications of gait recognition [9], image classification [56], and activity analysis [57]. It follows the BOVWs channel, which has two main steps: (1) codebook computation and (2) feature encoding. The complete process is illustrated in Figure 3, an and explanation of each task is outlined in the following.

4.2.1. Codebook Computation

In the first step, the input data are distributed into groups based on their underlying similar patterns using a clustering algorithm to create a dictionary. Usually, the codebook is constructed using unlabeled data. Each of the cluster center in the dictionary is known as a codeword, which describes the underlying variability in the gait data. We divided the continuous time series sensory data into equal-length sub-sequences using a window-sliding approach. In particular, a window of size w is moved at every l th sensory data point with a small overlap between neighboring sub-sequences. We grouped all the segmented sub-sequences into different k clusters using a k-means clustering algorithm. The clustering method was run ten times to choose the best clusters with the minimum sum of Euclidean distances between the data and the centers of their respective clusters. Assuming that { x m , m = 1 , , M } are the sub-sequence, c k is the set of k clusters, and r m k { 0 , 1 } , the distance is computed as:
m i n J ( r m k , c k ) = m = 1 M k = 1 K r m k x m c k 2 2
That is, the objective in Equation (9) is to minimize the cost function J on r m k and c k values. The optimal number of clusters k is chosen empirically.

4.2.2. Feature Encoding

This explains the process to encode the raw sensory data into its final representation, which is usually computed by estimating the distribution of these codewords in the gait input sequence. To this end, we employed locality-constrained linear coding (LLC) to encode the gait sub-sequences into a single high-level representation. LLC belongs to the category of reconstruction-based encoding techniques [12]. It emphasizes the decoding process’ characteristics such that the feature v is enforced in order to reassemble the sub-sequence x [16]. Rather than sparsity, LLC uses the locality constraint to map each gait sub-sequence x into its own local coordinate system. Specifically, it projects each sub-sequence x into its local coordinate system using the locality constraint [58]. This constraint can be applied by reducing the distance between x and the codes used to reconstruct it. In order to calculate the coding coefficient, an encoded vector v is generated for x such that the gait sub-sequence x is identical to the product of C and v (i.e., C v ). It thus solves the following optimization problem:
v = a r g min v x C v 2 2 + λ d v 2 2 , s . t . 1 M v = 1 ,
where x represents gait sub-sequence, C = { C k , k = 1 , , K } is codebook, λ is the regularizer parameter whose value is set to 0.01 [12], and ⊙ explains the member-wise product. The term d = ( e x p ( d i s t ( x , C ) / σ ) ) in Equation (10) indicates the locality adaptor where σ is applied to adjust the speed of weight decay d. Additionally, the d i s t ( x , C ) can be computed as: x c 1 2 , x c 2 2 , , x c k 2 . Since C is the k closest cluster to sub-sequence x, a computational efficient LLC solution can be employed by selecting the k-nearest basis vectors of sub-sequence x to minimize the term x c k 2 2 in Equation (10). To this end, we set the value of k to 5 empirically.

4.3. Classification

In the framework of handcrafted features and BOVWs, the final step is the classification of gait sequences. We evaluated the performance of two well-known and mostly used machine learning algorithms for classification: Support Vector Machine (SVM) and Random Forest (RF). In the following, a brief summary of each classification tool is outlined.

4.3.1. Support Vector Machine (SVM)

We employed simple linear SVM implementation to recognize the aforementioned computed feature representation. SVM has been used as a classification tool in numerous recognition applications [57]. SVM follows the principle of margin maximization; it learns a decision boundary (e.g., hyperplane) between the different classes. The data points closest to the hyperplane are known as support vectors and have the greatest influence on the adjustment of the hyperplane’s location. SVM learns the following optimization problem to learn an optimal hyperplane:
min w 1 2 w 2 + C i = 1 N ξ i s . t . y i ( w T x i + b ) 1 ξ i , ξ i 0 , i
where w is the weight vector, C > 0 is a penalty to misclassification, and ξ i = m a x ( 0 , 1 y i w T x i ) 2 is a loss function. The term y i = { 1 , + 1 } is the label associated with training instance x i . The objective of the SVM training is margin maximization, which is approximately equal to minimizing the regularization term w 2 . To optimally select the hyperparameter C, a 10-fold cross-validation is performed. The proposed technique employed the implementation of the LIBLINEAR SVM library [59]. For multi-class classification, it employed the one-vs.-rest strategy.

4.3.2. Random Forest (RF)

We also evaluated the effectiveness of RF [60] to recognize the high-level feature representation of gait sequences. RF has also been proven to be an efficient classification tool which can handle multi-dimensional features. It does not require any cross-validation, runs efficiently on large datasets, and is capable to work with missing data (to some extent). It is based on the principle of decision trees and generates ensemble learning, which involves a lot of classification trees to solve complex problems. For classification, the computed feature representation is passed in each of the forest’s tree. At the end of every tree, it produces a classification score, known as “vote”. The predicted classification with more votes is the object class. The complete process is illustrated in Figure 4. The classification accuracy of Random Forest depends on the correlation of any two trees and the strength of each individual tree.

4.4. Deep Learning-Based Feature Extraction Techniques

In recent studies, deep learning-based automatic feature extraction techniques have been explored to analyze gait patterns due to their robustness and efficient feature extraction, generalization power, and convincing recognition accuracies. These techniques present an end-to-end deep architecture to learn high-level characteristics from raw sensory data using a fully automated feature extraction process. They compute a high-level representation of input data using the complex structure of the deep artificial neural networks. In this study, we presented two end-to-end deep networks for gait analysis and activity recognition: (1) convolutional neural network (CNN) and (2) long short-term memory (LSTM). Each technique is briefly described in the following subsections.

4.4.1. Convolutional Neural Network (CNN)

A CNN is a deep network to learn high-level features directly from input data. Although they are primarily designed to extract the high-level semantics in images or videos to recognize objects, they are also proven to be effective in the classification of audio and time series data. A CNN usually consists of several layers of artificial neurons, which are mathematical functions that calculate the weighted sum of multiple inputs and output an activation value (similar to biological neurons). Each layer of CNN accepts input from the previous layer, computes a specific feature from it, and forwards it to the following layer as input [61].
Typically, a CNN consists of three different types of layers: (1) convolutional, (2) pooling, and (3) fully connected layers. The convolutional layer is considered an essential block of CNN which performs convolutional operations using a certain number of filters with their specified sizes. It convolves the filters on input data, extracts the useful features from raw data, and passes the result to the next layer. The pooling layer plays a vital role in reducing the dimensionality of input data by eliminating the number of feature points in the output from the previous convolution layer. Specifically, the convolutional layer identifies some important region of the input data, whereas the pooling layer conceals its exact location and keeps the most important information. Finally, a fully connected layer maps the extracted features to the final output, such as classification [62].
The proposed CNN follows the sequential layered approach with inputs as raw sensory sequences. The network configuration is shown in Figure 5. More specifically, it consists of three convolutional, two sub-sampling, one flattened, and one dense layer. The input layer is followed by three Conv2D layers with 32, 64, and 128 filters, respectively. Each Conv2D layer uses a kernel size of (3, 2), (2, 1), and (2, 1), respectively, and a ReLU activation function. Two MaxPooling2D layers follow the first and second Conv2D layers with a pool size of (2, 1), which reduces the spatial dimensions of the feature maps. The third Conv2D layer is followed by a flattened layer. The output from the preceding layer is flattened in this layer into a one-dimensional (1D) vector that can be fed into a dense layer. Finally, a probability distribution across the available classes is produced by the dense layer with Softmax activation function. The number of epochs is 27 with a batch size of 5.

4.4.2. Long Short-Term Memory (LSTM)

The LSTM is a type of recurrent neural network (RNN) which is designed to learn long-term dependencies, particularly in sequence prediction scenarios of time series data. LSTM differs from other neural networks mainly due to the way it deals with information over time. The data processing method used by traditional deep networks is feed-forward. Contrarily, LSTM-based deep networks process data in a recurrent manner. That is, they accept input data at one timestamp and produce an output at the following timestamp. A typical RNN is unable to learn longer-term dependencies due to the “vanishing gradient” or “exploding gradient” problem in back-propagation training. Contrarily, LSTM-based deep networks use additional gates to limit this issue. These gates enable the network to learn longer-term dependencies by controlling what the information in the hidden cell is exported as output and to the next hidden state. LSTM-based deep networks employ feedback connections to process the entire data sequence and have shown convincing recognition results in the domain of speech recognition, activity classification, etc.
An LSTM architecture consists of four main components: (1) memory cell, (2) input gate, (3) forget gate, and (4) output gate. The memory cell is the memory of the network which is used to memorize the information from an earlier timestamp and the current one. The input gate controls how the cell state is updated with the information from the current timestamp. The output gate regulates what information is output from the cell state at the current timestamp, while the forget gate controls what information from the previous timestamp is maintained in the present timestamp, as depicted in Figure 6. All these gates work on input x at timestamp t (i.e., x t ) and the Y t 1 by following the equations:
f t = σ ( W f · [ Y t 1 , x t ] + b f )
i t = σ ( W f · [ Y t 1 , x t ] + b i )
C ˜ t = t a n h ( W C · [ Y t 1 , x t ] + b C )
C t = f t C t 1 + i t C ˜ t
o t = σ ( W o · [ Y t 1 , x t ] + b o )
h t = o t t a n h ( C t )
where x t represents the input of the current timestamp, Y t 1 denotes the hidden state of previous timestamp, W is the weight matrix associated with the hidden state, b are the bias vectors for gates, C t is the state of the cell at time t, σ indicates the activation functions to squeeze the input and output of the cell, and ⊗ represents the point-wise multiplication operations. The initial part determines whether the information from the earlier timestamp should be remembered or (if it is unnecessary and) can be ignored. Using a sigmoid function, the forget gate determines which data from the previous cell state need to be forgotten. It considers the values of Y t 1 and x t , and generates a value between 0 and 1. The cell attempts to learn new knowledge in the second section using the input provided to it. With the help of point-wise multiplication operations of “sigmoid” and “tanh”, the input gate regulates the information flow to the present cell state.
First, the input gate layer selects the values that will be updated. Next, a vector of potential new values C ˜ t (that could be added to the state) is created using a tanh layer. Later, these two values are combined to update the cell state. The equation of C ˜ t = t a n h ( W C · [ Y t 1 , x t ] + b C ) will be used to update the old cell state C t 1 into the new cell state C t . Specifically, the old state is multiplied with f t to forget, and the product of i t with C ˜ t is added. This is the updated value, scaled by the amount by which we decided to alter each state value. Finally, the output gate determines which data should be sent to the following hidden state. Similarly to a simple neural network, the LSTM cells are organized in layers. The output of the cell in each layer is forwarded to the cell in the next successive layer. Finally, dense and Softmax layers are used in the network to accept the output of the last layer for recognition.

5. Experiments and Results

This paper mainly emphasizes on a systematic evaluation of numerous feature learning techniques to encode multimodal time series sensory data for gait analysis. However, the proposed feature encoding techniques were also explored to determine whether they could recognize human daily activities, which were also recorded using sensory data. Therefore, we explored several datasets to assess the proposed features on both gait and daily activities.

5.1. Dataset Description

The proposed feature encoding techniques are evaluated on four large sensory datasets, and their details are briefly explained in the following:
  • Gait-IMU dataset: The first dataset is collected in our APPS lab, located at the University of Lübeck, Germany. The IMU data are collected using LPMS-B2 Series (Advanced Realtime Tracking GmbH & Co. KG, Weilheim i.OB, Germany) devices (https://www.lp-research.com, accessed on 20 August 2020) along with the SensFloor sensor to analyze the gait patterns. A preliminary study on the analysis of SensFloor sensory data for gait analysis has been published in [63]; however, IMU data were not explored. Since an IMU consists of several sensors (e.g., accelerometer, gyroscope, magnetometer), the collected multimodal sensory data can be used to analyze gait in a more effective way. We sampled the IMU data at a fixed rate of 50 Hz. A total of 42 individuals participated in the gait collection. The four IMUs were attached to different body parts as depicted in Figure 7. The first IMU was attached at the sternum (chest), the second was at the lower abdomen (belt buckle), the third was at the left ankle, and the forth was at the right ankle. The participants performed six different walking styles: normal, slow, fast, blindfolded, dual task, and post-UHR by walking back and forth 10 times. The SensFloor data were used to identify the turning locations, and from there, a cut mark was placed to indicate the start and stop times of a single trip. More detail about the SensFloor data can be found in [63]. During data collection, we also observed a few missing data points (i.e., NAN values) in recorded data perhaps due to sensor malfunctions, which were covered using interpolation. However, we dropped the sequences having more than 30% NAN data points in the recorded sequences. This pre-processing makes the raw data suitable to be used with the feature encoding techniques. A total of 238 cleaned gait sequences are saved and freely available to the research community at https://drive.google.com/drive/folders/15sQTn3P2x3M1Em5o8yz1U784tomXDsJW (accessed on 20 August 2020).
  • MHEALTH dataset: The second dataset we used to assess the effectiveness of the proposed feature encoding techniques is the MHEALTH (Mobile Health) dataset [17]. The dataset contains data on the body motions of ten volunteers of diverse profiles engaged in twelve different physical activities (e.g., walking, jogging, running, jumping, etc.). The recordings were made in a laboratory using four wearable sensors from Shimmer2 (https://shimmersensing.com, accessed on 20 August 2020). The sensors were attached on the chest, right wrist, and left ankle using elastic bands. The authors captured the dataset at a fixed sample rate of 50 Hz for all the modalities. These sensors capture the motion information of the human body in different aspects, namely acceleration, magnetic field orientation, and the rate of turn, which can effectively record the dynamics of the body. This dataset is quite useful to not only analyze the walking styles, but to also recognize the subjects’ daily activities.
  • WISDM-AR dataset: The third dataset we used in this study was the Wireless Sensor Data Mining (WISDM-AR) dataset [18]. The data of twenty-nine volunteers were collected using a smartphone. While performing different daily activities, the subjects were instructed to carry their Android smartphones in their front leg pockets. The subjects performed different sets of activities, including walking, jogging, ascend and descend stairs, etc., for a certain period of time. The authors captured the dataset at a fixed sample rate of 20 Hz for all the modalities. Similar to the MHEALTH dataset, this dataset is also used to analyze the walking styles and the subjects’ daily activities.
  • UCI-HAR dataset: This dataset comprises the sequences of activities of daily living [19]. The data were collected using a Samsung Galaxy S II sensor, which was attached to the waist of the participants during recording. Specifically, the data from the smartphone’s gyroscope and accelerometer sensors were collected at a frequency of 50 Hz. The dataset consists of a total of 748,406 sample points from 30 participants ranging in age from 19 to 48 years. The activity sequences are grouped into six classes, namely sitting, standing, walking, walking downstairs, walking upstairs, and lying down.
We performed different experiments to assess the performance of the proposed approaches. Similar to [64,65,66,67], we employed an overlapped and fixed-size sliding-window approach in all the experiments. The window size is set to 128 with a 50 percent overlap (i.e., 128 data points/window is segmented) to compute the features. Details of the different experiments on each of the dataset are explained below.

5.2. Analysis of Gait-IMU Dataset

The time series data of each walk were extracted using all the four IMU sensors. The recorded gait sequences contain three-dimensional (3D) data for each sensor, i.e., accelerometer, magnetometer, and gyroscope. Due to limited instances in the dataset, we employed only handcrafted and codebook-based feature encoding techniques. It is important to mention that the proposed techniques were evaluated on the data of each IMU sensor to assess their effectiveness individually.

5.2.1. Using Handcrafted Features

We extracted eighteen handcrafted features (Section 4.1) from each walking sequence and they were fused together using feature-level fusion [57], i.e., through simple concatenation. The features were computed from each of the sensor, i.e., accelerometer, gyroscope, and magnetometer. Since each of the sensor generates 3D data, the final length of the feature vector was 3 × 3 × 18 . The feature vectors were divided into training and testing sets with the ratio of 80 and 20, respectively. We trained the classifiers (SVM and RF) using training instances, and the testing sequences were fed to classifiers. The recognition results are summarized in Table 1. It can be shown that the proposed approach achieved a 100% recognition rate on multiple IMUs, which reveals the effectiveness of the proposed feature encoding technique. We further concluded that the best outcomes can be obtained even with just one IMU on an ankle.

5.2.2. Using Codebook Approach

We computed a codebook for each type of sensory datum using its 3D sub-sequences. In particular, we segmented a sub-sequence with a window of size w, representing a 3 w -dimensional vector. That is, the first, second, and third w represents the x, y, and z axes of sensory data, respectively. The intuitive reason behind this segmentation approach is to capture the correlation among three axes. The same sub-sequence segmentation approach was used in the encoding of the final feature vector. We empirically set the value of w to 64, and l to 8. It can be noted that we set a large overlapping with a small size of l, considering the fact that a sufficient number of sub-sequences can be collected to capture the temporal correlation in the sub-sequences. The number of clusters in codebook (i.e., k ) was empirically set to 32. The high-level feature representation of the segmented sub-sequences was obtained using LLC encoding technique (Section 4.2.2) and it was fed to classifiers (SVM and RF) for walking style recognition. The recognition results are summarized in Table 1.
It can be observed from the recognition results that handcrafted features using RF as classifier performs consistently better than the codebook-based features. Since the dataset consists of only a few gait sequences, we cannot evaluate the deep learning-based techniques on this dataset.

5.3. Analysis of MHEALTH Dataset

We performed three different types of experiments on the MHEALTH dataset to evaluate the performance of the proposed techniques. Following the recommendation in [17], all the IMU sensor’s data were fused together (i.e., descriptor level fusion [11]) and used as input to feature learning algorithms. Details of the different experiments on each of the dataset are explained in the following.

5.3.1. Using Handcrafted Features

To compute the handcrafted features, we followed the recommendation in [17] and used the raw data of all the sensors with a window size of 20 × 23 . To capture the body dynamics comprehensively, the subject’s chest, right wrist, and left ankle were equipped with sensors using elastic straps. By employing multiple sensors, we were able to measure various aspects of motion, including acceleration, rate of turn, and magnetic field orientation. Therefore, they jointly provided a more comprehensive understanding of the body’s movement. All the extracted handcrafted features were fused together, indicating the final representation of the walk or activity. Finally, the features were fed to a classifier for recognition. The recognition results are summarized in Table 2. It can be noticed that the proposed handcrafted feature achieved the highest recognition rate of 99% using RF as classifier.

5.3.2. Using Codebook Approach

Similar to the Gait-IMU dataset, we computed a codebook using a k-means clustering algorithm, and the high-level feature representation was obtained using an LLC encoding technique. We recall that all the IMU sensor’s data were fused and used together as input for codebook computation and feature encoding. A recognition rate of 85% was achieved, as depicted in Table 2.

5.3.3. Using Deep Learning Approaches

We recall that two end-to-end deep learning models are presented to automatically extract features from raw sensory data. In the following, the implementation details of each model are presented.
CNN: We recall that the proposed 2D-CNN consists of one input layer, three convolution layers, two sub-sampling layers, one flattened layer, and one dense layer. The first convolution layer used 32 filters of size 3 × 2 , the second layer used 64 filters of size 2 × 1 , and the third layer used 128 filters of size 2 × 1 . For each sub-sampling layer, the stride factor was set to 2 × 1 to reduce the spatial dimensionality of the data. The convolution layer’s bias term was set to true, the number of epochs was 27 with a batch size of 5, and Rectified Linear Unit (ReLU) was used as an activation function. The network is build using Keras framework [68] with tensor flow as back end [69].
To determine the best values for the hyperparameters of the presented deep network (e.g., the number of layers, the activation function, etc.), we employed a multi-resolution search strategy [70]. In particular, it chooses parameter values from a wider range in the first phase and selects a few optimum configurations. The best values are then chosen from a small search space near these values. For instance, we evaluated the performance of the proposed deep network in a an increasing number of layers (conv2D and MaxPooling2D) and stopped when the performance on the validation data reached at its peak. It was empirically concluded that increasing the number of layers beyond five did not improve the performance. We used the EarlyStopping callback mechanism to monitor the validation loss, and the best weights were learned. We achieved a 99% recognition rate on MHEALTH dataset, and the results are summarized in Table 2.
LSTM: The proposed LSTM-based deep network consists of four LSTM layers, one flattened layer, and one dense layer. The model accepts an input tensor X with the shapes of [batch_size, segment_time_size, no_of_features], which were set to 32, 200, and 3, respectively. The model starts by initializing the weight W and bias b of the hidden layer and output layer using random normal distributions. The input tensor X is then transposed and reshaped to a 2D tensor of size (batch_size * segment_time_size, no_of_features), which is fed to the hidden layer. The hidden layer applies the ReLU activation function to the input tensor, and then splits the output into segment_time_size tensors.
Similar to CNN, the network is built using Keras framework [68] with tensor flow as back end [69]. We stacked two LSTM cells on top of each other to form a multi-layered LSTM. Specifically, the tf.contrib.rnn.BasicLSTMCell function is used to create each LSTM cell with a hidden state size of 50 neurons and a forget bias term of 1.0. The tf.contrib.rnn.MultiRNNCell function is used to create a multi-layered LSTM using two LSTM cells. The tf.contrib.rnn.static_rnn function is used to run the LSTM model over the input sequence. The output for the last time step of the LSTM is extracted from the output’s tensor using indexing and fed to the output layer. The output layer applies a linear transformation to the input tensor, and the final output is obtained by multiplying the result with W and adding b. A multi-resolution search [70] approach is used to learn the optimal values for the hyperparameters of the proposed deep network (i.e., number of layers, activation function, etc.). We used the EarlyStopping callback mechanism to monitor the validation loss, and the best weights were learned. We achieved a 96.78% recognition rate on MHEALTH dataset, and the results are summarized in Table 2.
The performances of the proposed feature encoding techniques are also assessed with the existing state-of-the-art techniques and their accuracies are outlined in Table 3. The recognition results reveal that the proposed feature encoding techniques outperform the existing state-of-the-art techniques.

5.4. Analysis of WISDM-AR Dataset

The third dataset we used to assess the effectiveness of the proposed feature encoding techniques is the WISDM-AR dataset [17]. A short description of each of the feature encoding techniques is summarized below.

5.4.1. Using Handcrafted Features

Following the recommendation in [18], the subsets of each activity are used to extract the handcrafted features in order to make a fair comparison. We computed the aforementioned eighteen handcrafted features and they were fed to a classifier (after fusion) for recognition. The recognition results are summarized in Table 2. It can be observed that the proposed handcrafted feature encoidng tehcnique achieved a recognition rate of 94%.

5.4.2. Using Codebook Approach

We computed the codebook using a k-means clustering algorithm for the WISDM-AR dataset, and the activity sequences were transformed to high-level descriptors using the LLC encoding technique. It achieved a recognition rate of 79%, as depicted in Table 2.

5.4.3. Using Deep Learning Approaches

We employed the same implementation of CNN- and LSTM-based deep networks as explained in Section 5.3.3. The CNN achieved a recognition rate of 97%, whereas LSTM achieved a recognition rate of 94%.
The performance of the proposed feature encoding technique is also compared with the existing state-of-the-art techniques and their recognition scores are outlined in Table 4. The proposed feature encoding technique outperforms the current state-of-the-art techniques.

5.5. Analysis of UCI-HAR Dataset

The forth dataset we used to assess the effectiveness of the proposed feature encoding techniques is the UCI-HAR dataset [19]. We employed the leave-one-subject-out cross-validation technique to validate the generalization power of the proposed techniques. A short description of each feature encoding technique is presented below.

5.5.1. Using Handcrafted Features

Similar to previous experiments, we extracted eighteen handcrafted features from the raw sensory data. All the features are concatenated to form a vector representation and fed to classifier for recognition. The recognition results are summarized in Table 2. It can be observed that the proposed handcrafted feature encoding technique achieved a recognition rate of 95.9%.

5.5.2. Using Codebook Approach

We computed the codebook using a k-means clustering algorithm for the UCI-HAR dataset, and the sequences were transformed to high-level descriptors using the LLC encoding technique. It achieved a recognition rate of 92.3%, as depicted in Table 2.

5.5.3. Using Deep Learning Approaches

We employed the same implementation of CNN- and LSTM-based deep networks as explained in Section 5.3.3. The CNN achieved a recognition rate of 96%, whereas LSTM achieved a recognition rate of 94%.
The performance of the proposed feature encoding technique is also compared with the existing state-of-the-art techniques and their recognition scores are outlined in Table 5. The proposed feature encoding technique outperforms the current state-of-the-art techniques.

5.6. Discussion and Computational Analysis

The main aim of this research is to present a comparative study of several feature extraction approaches to encode multimodal time series data for gait analysis. To fulfill this task, we gathered the gait sequences of 42 individuals in our APPS lab (located at the University of Lübeck, Germany) using IMU sensor LPMS-B2 Series devices. We presented three feature learning techniques to encode multimodal time series sensory data for gait analysis. Among them, the proposed handcrafted feature extraction technique demonstrated excellent recognition results on the collected dataset. To further verify the robustness of the proposed feature encoding techniques, we performed another experiment to recognize the activities of daily living. We assessed the performance of the proposed techniques on four large activity datasets that are captured using different sensing modalities. A short description of different sensing modalities in each of the dataset is summarized in Table 6, and their recognition results are depicted in Figure 8. The results demonstrate that the proposed feature encoding techniques can encode the raw data of different sensing modalities. It establishes a solid foundation and generic framework to encode any time series multimodal sensory data. We empirically concluded that the combination of accelerometer, gyroscope, and magnetometer work well together to recognize human activities. The recognition scores of different feature encoding techniques in Table 1 and Table 2 have confirmed that handcrafted feature encoding technique using RF and CNN-based feature learning techniques outperformed the other remaining techniques. Furthermore, they also prove to be superior in comparison with existing state-of-the-art techniques, as shown in Table 3, Table 4 and Table 5. The excellent results demonstrate the applicability of the proposed techniques to work with different modalities. We believe that the proposed feature encoding techniques establish a solid foundation and generic framework to encode any time series multimodal sensory data. That is, it is capable to work with different sensing modalities such as SensFloor, skin temperature sensors, and electrocardiograms (ECGs); however, the classification of such data is beyond the scope of this research and can be explored in future work.
It is worth mentioning that the proposed technique is computationally inexpensive and does not require any expensive specialized hardware (e.g., GPU) for execution. All the experiments are performed on a machine with an Intel i5-4310U CPU with 16 GB RAM (Dell, Siegen, Germany). We also analyzed the computational complexity of different feature encoding techniques. Specifically, the average time for the encoding and classification of one activity/gait sequence is recorded. The results are reported in Table 7. The deep CNN- and LTSM-based features require an average time of 12.93 and 11.95 milliseconds (ms), respectively. The codebook-based feature encoding techniques seem to be more efficient and have an average of 10.5 ms for both feature encoding and classification; however, they require the prior computation of a codebook (once), which is quite a complex process, and require an additional 7.1 min to compute it. The proposed handcrafted feature encoding techniques require an average of 13 ms for feature encoding and classification, and they do not require any prior computation. It can be observed that the proposed techniques can classify the activity sequences almost in real time without any need for an expensive GPU. Among the three proposed feature encoding techniques, the handcrafted feature representation is computationally efficient; however, these features lack exploratory capabilities, rely on the domain knowledge of the researchers, and are application dependent [81]. It is quite difficult to build a generic framework based on these features. Additionally, they may not be scalable when used with different datasets and might show partiality toward certain features. The deep learning-based approaches perform poorly and suffer with overfitting (and underfitting) due to the limited number of instances in the dataset.

6. Conclusions

This paper presents three different feature encoding techniques to recognize different human walking styles using multimodal time series sensory data. The first technique extracts eighteen handcrafted features directly from the raw sensory data. The second technique follows the channel of Bag-of-Visual-Words (BOVWs) approach. The feature representation of sensory data is obtained using codebook- and LLC-based feature encoding techniques. For both of these feature representations, the performance of two different machine learning algorithms is assessed. In the third feature encoding technique, two different end-to-end deep learning models are presented to automatically extract the features from raw sensory data. An extensive experimental evaluation of all the feature encoding techniques is carried out on four large sensory datasets and their recognitions results are compared. The experimental evaluation reveals that the proposed features are quite robust and can be used to recognize human daily activities as well. A comparison of the proposed encoded feature with the existing state-of-the-art techniques reveals its superiority and effectiveness. This research also collected a large gait dataset using IMU sensors that we have made available to the research community. In future, we plan to explore more deep learning models (especially hybrid models) to automatically extract the features from raw sensory data.

Author Contributions

Conceptualization, R.F. and M.H.K.; methodology, R.F. and M.H.K.; software, R.F. and M.H.K.; validation, M.H.K. and M.S.F.; investigation, R.F., M.H.K., M.S.F., R.D. and M.A.N.; writing—original draft preparation, M.H.K.; writing—review and editing, M.H.K., R.D. and M.G.; supervision, M.H.K., M.S.F., R.D., M.A.N. and M.G.; project administration, M.H.K. and M.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially supported by the Higher Education Pakistan within the project “HumCare: Human Activity Analysis in Healthcare” (grant number: 15041).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset of 238 gait sequences are freely available to the research community at https://drive.google.com/drive/folders/15sQTn3P2x3M1Em5o8yz1U784tomXDsJW (accessed on 14 October 2023).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Whittle, M.W. Gait analysis. In The Soft Tissues; Elsevier: Amsterdam, The Netherlands, 1993; pp. 187–199. [Google Scholar]
  2. Ghent, F.; Mobbs, R.J.; Mobbs, R.R.; Sy, L.; Betteridge, C.; Choy, W.J. Assessment and post-intervention recovery after surgery for lumbar disk herniation based on objective gait metrics from wearable devices using the gait posture index. World Neurosurg. 2020, 142, e111–e116. [Google Scholar] [CrossRef] [PubMed]
  3. Khan, M.H.; Farid, M.S.; Grzegorzek, M. Vision-based approaches towards person identification using gait. Comput. Sci. Rev. 2023, 42, 100432. [Google Scholar] [CrossRef]
  4. Khan, M.H.; Farid, M.S.; Grzegorzek, M. Spatiotemporal features of human motion for gait recognition. Signal Image Video Process. 2019, 13, 369–377. [Google Scholar] [CrossRef]
  5. Ahad, M.A.R.; Ngo, T.T.; Antar, A.D.; Ahmed, M.; Hossain, T.; Muramatsu, D.; Makihara, Y.; Inoue, S.; Yagi, Y. Wearable sensor-based gait analysis for age and gender estimation. Sensors 2020, 20, 2424. [Google Scholar] [CrossRef] [PubMed]
  6. Kolokas, N.; Krinidis, S.; Drosou, A.; Ioannidis, D.; Tzovaras, D. Gait matching by mapping wearable to camera privacy-preserving recordings: Experimental comparison of multiple settings. In Proceedings of the 2019 6th International Conference on Control, Decision and Information Technologies (CoDIT), Paris, France, 23–26 April 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 338–343. [Google Scholar]
  7. Amjad, F.; Khan, M.H.; Nisar, M.A.; Farid, M.S.; Grzegorzek, M. A comparative study of feature selection approaches for human activity recognition using multimodal sensory data. Sensors 2021, 21, 2368. [Google Scholar] [CrossRef] [PubMed]
  8. Rani, V.; Kumar, M.; Singh, B. Handcrafted Features for Human Gait Recognition: CASIA-A Dataset. In Proceedings of the Artificial Intelligence and Data Science: First International Conference, ICAIDS 2021, Hyderabad, India, 17–18 December 2021; Revised Selected Papers. Springer: Berlin/Heidelberg, Germany, 2022; pp. 77–88. [Google Scholar]
  9. Khan, M.H.; Farid, M.S.; Grzegorzek, M. A generic codebook based approach for gait recognition. Multimed. Tools Appl. 2019, 78, 35689–35712. [Google Scholar] [CrossRef]
  10. Zhang, Z.; He, T.; Zhu, M.; Sun, Z.; Shi, Q.; Zhu, J.; Dong, B.; Yuce, M.R.; Lee, C. Deep learning-enabled triboelectric smart socks for IoT-based gait analysis and VR applications. Npj Flex. Electron. 2020, 4, 29. [Google Scholar] [CrossRef]
  11. Peng, X.; Wang, L.; Wang, X.; Qiao, Y. Bag of visual words and fusion methods for action recognition: Comprehensive study and good practice. Comput. Vis. Image Underst. 2016, 150, 109–125. [Google Scholar] [CrossRef]
  12. Khan, M.H.; Farid, M.S.; Grzegorzek, M. A comprehensive study on codebook-based feature fusion for gait recognition. Inf. Fusion 2023, 92, 216–230. [Google Scholar] [CrossRef]
  13. Zhang, Y.; Huang, Y.; Wang, L.; Yu, S. A comprehensive study on gait biometrics using a joint CNN-based method. Pattern Recognit. 2019, 93, 228–236. [Google Scholar] [CrossRef]
  14. Martindale, C.F.; Christlein, V.; Klumpp, P.; Eskofier, B.M. Wearables-based multi-task gait and activity segmentation using recurrent neural networks. Neurocomputing 2021, 432, 250–261. [Google Scholar] [CrossRef]
  15. Sarshar, M.; Polturi, S.; Schega, L. Gait phase estimation by using LSTM in IMU-based gait analysis—Proof of concept. Sensors 2021, 21, 5749. [Google Scholar] [CrossRef] [PubMed]
  16. Liu, L.; Wang, L.; Liu, X. In defense of soft-assignment coding. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 2486–2493. [Google Scholar]
  17. Banos, O.; Garcia, R.; Holgado-Terriza, J.A.; Damas, M.; Pomares, H.; Rojas, I.; Saez, A.; Villalonga, C. mHealthDroid: A novel framework for agile development of mobile health applications. In Proceedings of the Ambient Assisted Living and Daily Activities: 6th International Work-Conference, IWAAL 2014, Belfast, UK, 2–5 December 2014; Proceedings 6. Springer: Berlin/Heidelberg, Germany, 2014; pp. 91–98. [Google Scholar]
  18. Kwapisz, J.R.; Weiss, G.M.; Moore, S.A. Activity recognition using cell phone accelerometers. ACM SigKDD Explor. Newsl. 2011, 12, 74–82. [Google Scholar] [CrossRef]
  19. Reyes, J.; Anguita, D.; Ghio, A.; Oneto, L.; Parra, X. Human Activity Recognition Using Smartphones; UCI Machine Learning Repository. 2012. Available online: https://archive.ics.uci.edu/dataset/240/human+activity+recognition+using+smartphones (accessed on 14 October 2023).
  20. Dorschky, E.; Nitschke, M.; Seifer, A.K.; van den Bogert, A.J.; Eskofier, B.M. Estimation of gait kinematics and kinetics from inertial sensor data using optimal control of musculoskeletal models. J. Biomech. 2019, 95, 109278. [Google Scholar] [CrossRef] [PubMed]
  21. Zhong, Q.; Ali, N.; Gao, Y.; Wu, H.; Wu, X.; Sun, C.; Ma, J.; Thabane, L.; Xiao, M.; Zhou, Q.; et al. Gait kinematic and kinetic characteristics of older adults with mild cognitive impairment and subjective cognitive decline: A cross-sectional study. Front. Aging Neurosci. 2021, 13, 664558. [Google Scholar] [CrossRef] [PubMed]
  22. Ahmad, H.N.; Barbosa, T.M. The effects of backpack carriage on gait kinematics and kinetics of schoolchildren. Sci. Rep. 2019, 9, 3364. [Google Scholar] [CrossRef] [PubMed]
  23. Zheng, S.; Huang, K.; Tan, T. Evaluation framework on translation-invariant representation for cumulative foot pressure image. In Proceedings of the 2011 18th IEEE International Conference on Image Processing, Brussels, Belgium, 11–14 September 2011; pp. 201–204. [Google Scholar]
  24. McDonough, A.L.; Batavia, M.; Chen, F.C.; Kwon, S.; Ziai, J. The validity and reliability of the GAITRite system’s measurements: A preliminary evaluation. Arch. Phys. Med. Rehabil. 2001, 82, 419–425. [Google Scholar] [CrossRef]
  25. Bilney, B.; Morris, M.; Webster, K. Concurrent related validity of the GAITRite® walkway system for quantification of the spatial and temporal parameters of gait. Gait Posture 2003, 17, 68–74. [Google Scholar] [CrossRef]
  26. Leder, R.S.; Azcarate, G.; Savage, R.; Savage, S.; Sucar, L.E.; Reinkensmeyer, D.; Toxtli, C.; Roth, E.; Molina, A. Nintendo Wii remote for computer simulated arm and wrist therapy in stroke survivors with upper extremity hemipariesis. In Proceedings of the 2008 Virtual Rehabilitation, Vancouver, BC, Canada, 25–27 August 2008; p. 74. [Google Scholar]
  27. Han, Y.C.; Wong, K.I.; Murray, I. Gait phase detection for normal and abnormal gaits using IMU. IEEE Sens. J. 2019, 19, 3439–3448. [Google Scholar] [CrossRef]
  28. Patel, B.; Srikanthan, S.; Asanit, F.; Agu, E. Machine learning prediction of tbi from mobility, gait and balance patterns. In Proceedings of the 2021 IEEE/ACM Conference on Connected Health: Applications, Systems and Engineering Technologies (CHASE), Washington, DC, USA, 16–18 December 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 11–22. [Google Scholar]
  29. Saba, T.; Mohamed, A.S.; El-Affendi, M.; Amin, J.; Sharif, M. Brain tumor detection using fusion of hand crafted and deep learning features. Cogn. Syst. Res. 2020, 59, 221–230. [Google Scholar] [CrossRef]
  30. Schonberger, J.L.; Hardmeier, H.; Sattler, T.; Pollefeys, M. Comparative evaluation of hand-crafted and learned local features. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1482–1491. [Google Scholar]
  31. Nandy, A. Statistical methods for analysis of Parkinson’s disease gait pattern and classification. Multimed. Tools Appl. 2019, 78, 19697–19734. [Google Scholar] [CrossRef]
  32. Hsu, W.C.; Sugiarto, T.; Liao, Y.Y.; Lin, Y.J.; Yang, F.C.; Hueng, D.Y.; Sun, C.T.; Chou, K.N. Can trunk acceleration differentiate stroke patient gait patterns using time-and frequency-domain features? Appl. Sci. 2021, 11, 1541. [Google Scholar] [CrossRef]
  33. Huang, J.; Kaewunruen, S.; Ning, J. AI-based quantification of fitness activities using smartphones. Sustainability 2022, 14, 690. [Google Scholar] [CrossRef]
  34. Sejdic, E.; Lowry, K.A.; Bellanca, J.; Redfern, M.S.; Brach, J.S. A comprehensive assessment of gait accelerometry signals in time, frequency and time-frequency domains. IEEE Trans. Neural Syst. Rehabil. Eng. 2013, 22, 603–612. [Google Scholar] [CrossRef] [PubMed]
  35. Permatasari, J.; Connie, T.; Ong, T.S.; Teoh, A.B.J. Adaptive 1-dimensional time invariant learning for inertial sensor-based gait authentication. Neural Comput. Appl. 2023, 35, 2737–2753. [Google Scholar] [CrossRef]
  36. Köping, L.; Shirahama, K.; Grzegorzek, M. A general framework for sensor-based human activity recognition. Comput. Biol. Med. 2018, 95, 248–260. [Google Scholar] [CrossRef] [PubMed]
  37. Papadopoulos, A.; Kyritsis, K.; Klingelhoefer, L.; Bostanjopoulou, S.; Chaudhuri, K.R.; Delopoulos, A. Detecting Parkinsonian Tremor From IMU Data Collected in-the-Wild Using Deep Multiple-Instance Learning. IEEE J. Biomed. Health Inform. 2020, 24, 2559–2569. [Google Scholar] [CrossRef] [PubMed]
  38. Liu, Z. Human Gait Phase Recognition in Embedded Sensor System. Master’s Thesis, KTH Royal Institute of Technology, Stockholm, Sweden, 2021. [Google Scholar]
  39. Ryu, J.; McFarland, T.; Haas, C.T.; Abdel-Rahman, E. Automatic clustering of proper working postures for phases of movement. Autom. Constr. 2022, 138, 104223. [Google Scholar] [CrossRef]
  40. Calvo, A.F.; Holguin, G.A.; Medeiros, H. Human activity recognition using multi-modal data fusion. In Proceedings of the Progress in Pattern Recognition, Image Analysis, Computer Vision, and Applications: 23rd Iberoamerican Congress, CIARP 2018, Madrid, Spain, 19–22 November 2018; Proceedings 23. Springer: Berlin/Heidelberg, Germany, 2019; pp. 946–953. [Google Scholar]
  41. Clapés, A.; Pardo, À.; Pujol Vila, O.; Escalera, S. Action detection fusing multiple Kinects and a WIMU: An application to in-home assistive technology for the elderly. Mach. Vis. Appl. 2018, 29, 765–788. [Google Scholar] [CrossRef]
  42. Khan, M.H.; Farid, M.S.; Grzegorzek, M. Person identification using spatiotemporal motion characteristics. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 166–170. [Google Scholar]
  43. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  44. Nguyen, M.D.; Mun, K.R.; Jung, D.; Han, J.; Park, M.; Kim, J.; Kim, J. IMU-based spectrogram approach with deep convolutional neural networks for gait classification. In Proceedings of the 2020 IEEE International Conference on Consumer Electronics (ICCE), Las Vegas, NV, USA, 4–6 January 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–6. [Google Scholar]
  45. Tran, L.; Hoang, T.; Nguyen, T.; Kim, H.; Choi, D. Multi-model long short-term memory network for gait recognition using window-based data segment. IEEE Access 2021, 9, 23826–23839. [Google Scholar] [CrossRef]
  46. Zhao, B.; Lu, H.; Chen, S.; Liu, J.; Wu, D. Convolutional neural networks for time series classification. J. Syst. Eng. Electron. 2017, 28, 162–169. [Google Scholar] [CrossRef]
  47. Connor, J.T.; Martin, R.D.; Atlas, L.E. Recurrent neural networks and robust time series prediction. IEEE Trans. Neural Netw. 1994, 5, 240–254. [Google Scholar] [CrossRef] [PubMed]
  48. Feng, Y.; Li, Y.; Luo, J. Learning effective gait features using LSTM. In Proceedings of the 2016 23rd International Conference on Pattern Recognition (ICPR), Cancun, Mexico, 4–8 December 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 325–330. [Google Scholar]
  49. Khan, M.H.; Farid, M.S.; Grzegorzek, M. A non-linear view transformations model for cross-view gait recognition. Neurocomputing 2020, 402, 100–111. [Google Scholar] [CrossRef]
  50. Giorgi, G.; Saracino, A.; Martinelli, F. Using recurrent neural networks for continuous authentication through gait analysis. Pattern Recognit. Lett. 2021, 147, 157–163. [Google Scholar] [CrossRef]
  51. Mundt, M.; Koeppe, A.; David, S.; Witter, T.; Bamer, F.; Potthast, W.; Markert, B. Estimation of gait mechanics based on simulated and measured IMU data using an artificial neural network. Front. Bioeng. Biotechnol. 2020, 8, 41. [Google Scholar] [CrossRef]
  52. Alharthi, A.S.; Yunas, S.U.; Ozanyan, K.B. Deep learning for monitoring of human gait: A review. IEEE Sens. J. 2019, 19, 9575–9591. [Google Scholar] [CrossRef]
  53. Cook, D.J.; Krishnan, N.C. Activity Learning: Discovering, Recognizing, and Predicting Human Behavior from Sensor Data; John Wiley & Sons: Hoboken, NJ, USA, 2015. [Google Scholar]
  54. Esmael, B.; Arnaout, A.; Fruhwirth, R.; Thonhauser, G. A statistical feature-based approach for operations recognition in drilling time series. Int. J. Comput. Inf. Syst. Ind. Manag. Appl. 2012, 4, 100–108. [Google Scholar]
  55. Bartholomew, D.J. Time Series Analysis Forecasting and Control. Oper. Res. Q. 1971, 22, 199–201. [Google Scholar] [CrossRef]
  56. Liu, J.; Huang, Y.; Wang, L.; Wu, S. Hierarchical feature coding for image classification. Neurocomputing 2014, 144, 509–515. [Google Scholar] [CrossRef]
  57. Khan, M.H. Human Activity Analysis in Visual Surveillance and Healthcare; Logos Verlag Berlin GmbH: Berlin, Germany, 2018; Volume 45. [Google Scholar]
  58. Wang, J.; Yang, J.; Yu, K.; Lv, F.; Huang, T.; Gong, Y. Locality-constrained linear coding for image classification. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 3360–3367. [Google Scholar]
  59. Fan, R.E.; Chang, K.W.; Hsieh, C.J.; Wang, X.R.; Lin, C.J. LIBLINEAR: A library for large linear classification. J. Mach. Learn. Res. 2008, 9, 1871–1874. [Google Scholar]
  60. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  61. Li, Z.; Liu, F.; Yang, W.; Peng, S.; Zhou, J. A survey of convolutional neural networks: Analysis, applications, and prospects. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 6999–7019. [Google Scholar] [CrossRef] [PubMed]
  62. Albawi, S.; Mohammed, T.A.; Al-Zawi, S. Understanding of a convolutional neural network. In Proceedings of the 2017 International Conference on Engineering and Technology (ICET), Antalya, Turkey, 21–23 August 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1–6. [Google Scholar]
  63. Hoffmann, R.; Brodowski, H.; Steinhage, A.; Grzegorzek, M. Detecting walking challenges in gait patterns using a capacitive sensor floor and recurrent neural networks. Sensors 2021, 21, 1086. [Google Scholar] [CrossRef] [PubMed]
  64. Xia, K.; Huang, J.; Wang, H. LSTM-CNN architecture for human activity recognition. IEEE Access 2020, 8, 56855–56866. [Google Scholar] [CrossRef]
  65. Nematallah, H.; Rajan, S. Comparative study of time series-based human activity recognition using convolutional neural networks. In Proceedings of the 2020 IEEE International Instrumentation and Measurement Technology Conference (I2MTC), Dubrovnik, Croatia, 25–28 May 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–6. [Google Scholar]
  66. Yin, X.; Liu, Z.; Liu, D.; Ren, X. A Novel CNN-based Bi-LSTM parallel model with attention mechanism for human activity recognition with noisy data. Sci. Rep. 2022, 12, 7878. [Google Scholar] [CrossRef]
  67. Challa, S.K.; Kumar, A.; Semwal, V.B. A multibranch CNN-BiLSTM model for human activity recognition using wearable sensor data. Vis. Comput. 2022, 38, 4095–4109. [Google Scholar] [CrossRef]
  68. Chollet, F. Keras. 2015. Available online: https://github.com/keras-team/keras (accessed on 14 October 2023).
  69. Abadi, M.; Barham, P.; Chen, J.; Chen, Z.; Davis, A.; Dean, J.; Devin, M.; Ghemawat, S.; Irving, G.; Isard, M.; et al. {TensorFlow}: A system for {Large-Scale} machine learning. In Proceedings of the 12th USENIX Symposium on Operating Systems Design and Implementation (OSDI 16), Savannah, GA, USA, 2–4 November 2016; pp. 265–283. [Google Scholar]
  70. Bengio, Y. Practical recommendations for gradient-based training of deep architectures. In Neural Networks: Tricks of the Trade, 2nd ed.; Springer: Berlin/Heidelberg, Germany, 2012; pp. 437–478. [Google Scholar]
  71. O’Halloran, J.; Curry, E. A Comparison of Deep Learning Models in Human Activity Recognition and Behavioural Prediction on the MHEALTH Dataset. In Proceedings of the AICS, Wuhan, China, 12–13 July 2019; pp. 212–223. [Google Scholar]
  72. Khatun, M.A.; Yousuf, M.A.; Ahmed, S.; Uddin, M.Z.; Alyami, S.A.; Al-Ashhab, S.; Akhdar, H.F.; Khan, A.; Azad, A.; Moni, M.A. Deep CNN-LSTM with self-attention model for human activity recognition using wearable sensor. IEEE J. Transl. Eng. Health Med. 2022, 10, 1–16. [Google Scholar] [CrossRef]
  73. Davidashvilly, S.; Hssayeni, M.; Chi, C.; Jimenez-Shahed, J.; Ghoraani, B. Activity recognition in parkinson’s patients from motion data using a cnn model trained by healthy subjects. In Proceedings of the 2022 44th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Glasgow, UK, 11–15 July 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 3199–3202. [Google Scholar]
  74. Yatbaz, H.Y.; Ever, E.; Yazici, A. Activity recognition and anomaly detection in E-health applications using color-coded representation and lightweight CNN architectures. IEEE Sens. J. 2021, 21, 14191–14202. [Google Scholar] [CrossRef]
  75. Semwal, V.B.; Gaud, N.; Lalwani, P.; Bijalwan, V.; Alok, A.K. Pattern identification of different human joints for different human walking styles using inertial measurement unit (IMU) sensor. Artif. Intell. Rev. 2022, 55, 1149–1169. [Google Scholar] [CrossRef]
  76. Xu, W.; Pang, Y.; Yang, Y.; Liu, Y. Human activity recognition based on convolutional neural network. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 165–170. [Google Scholar]
  77. Khan, Z.N.; Ahmad, J. Attention induced multi-head convolutional neural network for human activity recognition. Appl. Soft Comput. 2021, 110, 107671. [Google Scholar] [CrossRef]
  78. Tong, L.; Ma, H.; Lin, Q.; He, J.; Peng, L. A novel deep learning Bi-GRU-I model for real-time human activity recognition using inertial sensors. IEEE Sens. J. 2022, 22, 6164–6174. [Google Scholar] [CrossRef]
  79. Perez-Gamboa, S.; Sun, Q.; Zhang, Y. Improved sensor based human activity recognition via hybrid convolutional and recurrent neural networks. In Proceedings of the 2021 IEEE International Symposium on Inertial Sensors and Systems (INERTIAL), Virtual, 22–25 March 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–4. [Google Scholar]
  80. Kolkar, R.; Geetha, V. Human activity recognition in smart home using deep learning techniques. In Proceedings of the 2021 13th International Conference on Information & Communication Technology and System (ICTS), Online, 20–21 October 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 230–234. [Google Scholar]
  81. Wang, J.; Chen, Y.; Hao, S.; Peng, X.; Hu, L. Deep learning for sensor-based activity recognition: A survey. Pattern Recognit. Lett. 2019, 119, 3–11. [Google Scholar] [CrossRef]
Figure 1. A set of sensing modalities that can be used for gait data collection.
Figure 1. A set of sensing modalities that can be used for gait data collection.
Sensors 24 00075 g001
Figure 2. Classification of existing feature encoding techniques using sensory data based on their underlying computing methods.
Figure 2. Classification of existing feature encoding techniques using sensory data based on their underlying computing methods.
Sensors 24 00075 g002
Figure 3. An illustration of codebook-based feature encoding technique.
Figure 3. An illustration of codebook-based feature encoding technique.
Sensors 24 00075 g003
Figure 4. An illustration of the Random Forest classifier.
Figure 4. An illustration of the Random Forest classifier.
Sensors 24 00075 g004
Figure 5. The proposed deep CNN consists of three convolutional, two sub-sampling, one flattened, and one dense layer.
Figure 5. The proposed deep CNN consists of three convolutional, two sub-sampling, one flattened, and one dense layer.
Sensors 24 00075 g005
Figure 6. An illustration of the proposed LSTM-based deep network to classify the multimodal sensory data.
Figure 6. An illustration of the proposed LSTM-based deep network to classify the multimodal sensory data.
Sensors 24 00075 g006
Figure 7. An illustration of IMU placement in collected dataset to capture the motion information of different body parts.
Figure 7. An illustration of IMU placement in collected dataset to capture the motion information of different body parts.
Sensors 24 00075 g007
Figure 8. A summary of the recognition results of four datasets using all the encoding methods. The terms HC, CB, DL, SVM, RF, CNN, and LSTM represent handcrafted features, codebook features, deep learning features, Support Vector Machine, Random Forest, convolutional neural network, and long short-term memory, respectively.
Figure 8. A summary of the recognition results of four datasets using all the encoding methods. The terms HC, CB, DL, SVM, RF, CNN, and LSTM represent handcrafted features, codebook features, deep learning features, Support Vector Machine, Random Forest, convolutional neural network, and long short-term memory, respectively.
Sensors 24 00075 g008
Table 1. Recognition accuracies (%) of all IMU sensors with eighteen handcrafted features and codebook-based features. The terms SVM and RF represent Support Vector Machine and Random Forest, respectively. The best results are marked in bold.
Table 1. Recognition accuracies (%) of all IMU sensors with eighteen handcrafted features and codebook-based features. The terms SVM and RF represent Support Vector Machine and Random Forest, respectively. The best results are marked in bold.
IMUsHandcrafted FeaturesCodebook-Based Features
SVMRFSVMRF
IMU-192.692.681.588.9
IMU-296.3100.088.996.3
IMU-396.396.392.685.2
IMU-4100.0100.091.795.8
Average96.397.288.791.6
Table 2. Recognition accuracies (%) of different feature encoding techniques on MHEALTH and WISDM-AR datasets. The terms SVM, RF, CNN, and LSTM represent Support Vector Machine, Random Forest, convolutional neural network, and long short-term memory, respectively. The best results are marked in bold.
Table 2. Recognition accuracies (%) of different feature encoding techniques on MHEALTH and WISDM-AR datasets. The terms SVM, RF, CNN, and LSTM represent Support Vector Machine, Random Forest, convolutional neural network, and long short-term memory, respectively. The best results are marked in bold.
HandcraftedCodebookDeep Learning
SVMRFSVMRFCNNLSTM
MHEALTH dataset98.099.084.085.099.096.0
WISDM-AR dataset89.094.077.079.097.094.0
UCI-HAR dataset92.895.990.592.396.092.7
Table 3. Comparison of recognition results (%) of the proposed technique with the state-of-the-art techniques on MHEALTH dataset. The best results are marked in bold.
Table 3. Comparison of recognition results (%) of the proposed technique with the state-of-the-art techniques on MHEALTH dataset. The best results are marked in bold.
MethodsYearAccuracy
Halloran et al. [71]201983
Khatun et al. [72]202293
Davidashvilly et al. [73]202287
Yatbaz et al. [74]202197
Nematallahet al. [65]202093
Proposed handcrafted features with RF202399
Proposed deep CNN202399
Table 4. Comparison of recognition results (%) of the proposed technique with the state-of-the-art techniques on WISDM dataset. The best results are marked in bold.
Table 4. Comparison of recognition results (%) of the proposed technique with the state-of-the-art techniques on WISDM dataset. The best results are marked in bold.
MethodsYearAccuracy
Xia et al. [64]202095
Semwal et al. [75]202290
Challa et al. [67]202296
Xu et al. [76]201891
Yin et al. [66]202296
Proposed deep CNN202397
Table 5. Comparison of recognition results (%) of the proposed technique with the state-of-the-art techniques on UCI-HAR dataset. The best results are marked in bold.
Table 5. Comparison of recognition results (%) of the proposed technique with the state-of-the-art techniques on UCI-HAR dataset. The best results are marked in bold.
MethodsYearAccuracy
Khan et al. [77]202195.4
Xia et al. [64]202095.8
Tong et al. [78]202295.4
Perez et al. [79]202194.7
Kolkar et al. [80]202193.1
Proposed handcrafted features202395.9
Proposed deep CNN202396.0
Table 6. A description of the different sensing modalities that were used to gather the datasets and the obtained recognition results.
Table 6. A description of the different sensing modalities that were used to gather the datasets and the obtained recognition results.
DatasetData CollectionCollection RateSensing ModalitiesAccuracy
Gait-IMUThe gait data are collected using IMU sensor LPMS-B2 Series devices.50 HzAccelerometer, magnetometer, and gyroscope97.2
MHEALTHThe activity data are collected using wearable IMU sensors from Shimmer2.50 HzAccelerometer, magnetometer, and gyroscope99.0
WISDM-ARThe activity data are collected using a smartphone.20 HzAccelerometer97.0
UCI-HARThe activity data are collected using a smartphone.50 HzAccelerometer, and gyroscope96.0
Table 7. Analysis of computational time in feature encoding and classification. The average computation time for each of the encoding techniques is presented. The average feature encoding and classification time is reported for a gait/activity sequence on a CPU machine. The time is computed in milliseconds (ms).
Table 7. Analysis of computational time in feature encoding and classification. The average computation time for each of the encoding techniques is presented. The average feature encoding and classification time is reported for a gait/activity sequence on a CPU machine. The time is computed in milliseconds (ms).
Feature Encoding (ms)Classification Time (ms)
Handcrafted using SVM5.078.98
Handcrafted using RF5.079.44
Codebook using SVM5.944.51
Codebook using RF5.944.51
CNN-12.93
LSTM-11.95
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Fatima, R.; Khan, M.H.; Nisar, M.A.; Doniec, R.; Farid, M.S.; Grzegorzek, M. A Systematic Evaluation of Feature Encoding Techniques for Gait Analysis Using Multimodal Sensory Data. Sensors 2024, 24, 75. https://doi.org/10.3390/s24010075

AMA Style

Fatima R, Khan MH, Nisar MA, Doniec R, Farid MS, Grzegorzek M. A Systematic Evaluation of Feature Encoding Techniques for Gait Analysis Using Multimodal Sensory Data. Sensors. 2024; 24(1):75. https://doi.org/10.3390/s24010075

Chicago/Turabian Style

Fatima, Rimsha, Muhammad Hassan Khan, Muhammad Adeel Nisar, Rafał Doniec, Muhammad Shahid Farid, and Marcin Grzegorzek. 2024. "A Systematic Evaluation of Feature Encoding Techniques for Gait Analysis Using Multimodal Sensory Data" Sensors 24, no. 1: 75. https://doi.org/10.3390/s24010075

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop