Next Article in Journal
Wearable Bluetooth Triage Healthcare Monitoring System
Previous Article in Journal
Development and Experimental Validation of an Intelligent Camera Model for Automated Driving
Previous Article in Special Issue
Optimization and Security of Hazardous Waste Incineration Plants with the Use of a Heuristic Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Human Gait Recognition: A Single Stream Optimal Deep Learning Features Fusion

1
Department of Computer Science, HITEC University Taxila, Taxila 47080, Pakistan
2
College of Computer Science and Engineering, University of Ha’il, Ha’il 55211, Saudi Arabia
3
College of Computer Engineering and Science, Prince Sattam Bin Abdulaziz University, Al-Kharaj 11942, Saudi Arabia
4
Department of Electrical Engineering, College of Engineering, Jouf University, Sakakah 72388, Saudi Arabia
5
Department of Applied Artificial Intelligence, Ajou University, Suwon 16499, Korea
6
Faculty of Applied Computing and Technology, Noroff University College, 4608 Kristiansand, Norway
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(22), 7584; https://doi.org/10.3390/s21227584
Submission received: 16 September 2021 / Revised: 8 November 2021 / Accepted: 12 November 2021 / Published: 15 November 2021

Abstract

:
Human Gait Recognition (HGR) is a biometric technique that has been utilized for security purposes for the last decade. The performance of gait recognition can be influenced by various factors such as wearing clothes, carrying a bag, and the walking surfaces. Furthermore, identification from differing views is a significant difficulty in HGR. Many techniques have been introduced in the literature for HGR using conventional and deep learning techniques. However, the traditional methods are not suitable for large datasets. Therefore, a new framework is proposed for human gait recognition using deep learning and best feature selection. The proposed framework includes data augmentation, feature extraction, feature selection, feature fusion, and classification. In the augmentation step, three flip operations were used. In the feature extraction step, two pre-trained models were employed, Inception-ResNet-V2 and NASNet Mobile. Both models were fine-tuned and trained using transfer learning on the CASIA B gait dataset. The features of the selected deep models were optimized using a modified three-step whale optimization algorithm and the best features were chosen. The selected best features were fused using the modified mean absolute deviation extended serial fusion (MDeSF) approach. Then, the final classification was performed using several classification algorithms. The experimental process was conducted on the entire CASIA B dataset and achieved an average accuracy of 89.0. Comparison with existing techniques showed an improvement in accuracy, recall rate, and computational time.

1. Introduction

Human gait recognition has become an active research area in the past decade in computer vision [1,2]. Biometrics is an important application of gait recognition, employed in many industrial areas such surveillance and healthcare systems [3,4]. Despite the specific characteristics of gait features, several factors influence gait identification, including camera perspectives, load holding, lighting conditions, clothing variance, walking speed, and shadows under feet [5]. As a result, designing a robust enough framework to solve these obstacles is crucial for accurate gait classification [6]. Moreover, a few biological signals are also useful for gait analysis such as electromyography (EMG) [7,8,9], inertial sensors [10,11], and plantar pressure [12]. Through these methods, the human gait can be analyzed, for example, through muscle movement while walking. Gait recognition algorithms have progressed to the point that they can now be used in a wide range of “real-world” applications, such as video monitoring, crime prevention, and forensic detection [13,14].
The two broad categories of HGR are model-based and model-free approaches [3]. Model-based approaches, such as human body structure and motion models, are used to classify human gait. They use a structural model to characterize a human subject and use underlying mathematical structures such as stick-figure, intertwined pendulum, and ellipse fitting techniques to track different body parts and joint positions over time to define gait [15]. The features of the object silhouette are derived from the gait period in the model-free method, which is simple to implement due to lower computational costs [16].
Computer vision, a floor sensor, and a watch sensor are three common approaches for detecting gait depending on the background [14]. Using computer vision, cameras are used for video capturing to process the frames’ transformation. In the past two decades, researchers have performed the silhouette extraction process to extract the classical features for gait recognition [17,18]. They extracted the silhouette using GEI [19] and MHI based techniques [20,21]. Moreover, GFHI and Markov Hidden Model-based techniques have also been introduced for gait recognition [22]. These methods are not reliable once the size of the dataset is increased. Moreover, this type of process is time-consuming and takes up a lot of space, which later affects the entire system’s computational time. Further, the main limitation of this process is that if a human silhouette is incorrectly detected, irrelevant features are extracted [23,24]. Researchers have tried to resolve these issues using feature reduction techniques such as principle component analysis (PCA) [25,26], linear discriminant analysis (LDA) [27], and improved PCA [28], among others [29,30].
The most recent research in computer vision applied deep learning for human gait recognition with better recognition accuracy [31]. The researchers used deep learning performance in varying domains such as object identification, disease segmentation, and recognition from photographs and videos [32,33,34,35]. The convolution neural network (CNN) is a type of deep learning, which includes several layers that extract the deep features of each image [36,37]. In a CNN, a few layers are added as an activation function, and a few are utilized for the downsampling [38,39].
The extracted deep features from the raw images also extract some redundant and irrelevant features. These types of features need refinement for better classification accuracy. The researchers introduced feature selection techniques to handle irrelevant features for classification purposes [40]. Feature selection (FS) is an active research area in machine learning currently [41]. The purpose of the feature selection step is to refine the high dimensional feature vector, which contains duplicate and unnecessary features for classification. The primary goal of FS is to eliminate redundant and unnecessary characteristics from the original feature vector to reduce its dimensionality and to choose the single most vital point for higher classification accuracy and low computing time. Some of the latest feature selection techniques are an entropy-based approach, a Genetic Algorithm based approach, and a grasshopper based approach, among others [42,43].
Problem Statement: The key problems this research article addresses are:
  • Recognizing gait in various circumstances, such as changing view angles and walking styles while still wearing a coat and bag, is a difficult challenge.
  • Different subjects have their distinct gait; however, some of the subjects have a very similar gait. This issue misleads the correct classification and degrades the system’s performance.
  • Some researchers follow the two step process for gait recognition, such as subject detection and then classification. However, this first step is not a guarantee for accurate recognition of human gait due to incorrect subject detection. Also, two-step process increases the computational time.
  • Variation in the style of clothings such as a long coat, half shirt, skirt, regular pants, coats, and so on. In these styles, it is not easy to extract the rich features for further classification.
  • The irrelevant features information which is extracted from the original frames is an impact on the system accuracy and computational time.
Contributions: A new framework is proposed using deep learning and selection of optimal features for human gait recognition to resolve the above-listed issues. The significant contributions are as follows:
  • Modified two pre-trained deep learning models on CASIA B dataset and performed transfer learning. After the transfer learning, features are extracted from the average pool layer.
  • Features are fused using a modified mean absolute deviation extended serial fusion (MDeSF) approach.
  • The best features are selected using a three-step Improved Whale Optimization algorithm.
The rest of the manuscript is organized as follows: Recent gait recognition techniques are discussed in Section 2. An explanation of the dataset is provided in Section 3. Section 4 presents the proposed methodology. The results of the proposed method are discussed in Section 5. Finally, we conclude.

2. Related Work

Deep learning showed much success in the area of machine learning in the past two decades. Well known applications of deep learning include visual surveillance, biometrics, and medicine, among others [44,45]. Biometrics is an important application and much research attends to this field [46]. In the biometric field, gait recognition is an important research area and several techniques using deep learning have been introduced [47].
Davarzani et al. [48] presented a system for human gait detection based on deep learning. They implemented three models at the initial stage such as linear regression, artificial neural network (ANN), and LSTM. Later on, the performance of each model was combined for the final detection. The experimental process was conducted on each model using publically available datasets. The results showed that the ANN performed well as compared to linear regression and LSTM. Asif et al. [3] presented an end-to-end system based on pre-trained convolutional neural network features and best features selection. They used deep learning for the deep features extraction, and then best features were selected for the final classification. Three different angles of the CASIA B dataset were used in the experimental process, 18, 36, and 54. The results showed that the precision and recall rate of the presented method was better than the current state-of-the-art methods.
Anusha et al. [49] described a Modified Local Optimal Oriented Pattern Binary (MLOOPB) descriptor for clothing-invariant HGR. The MLOOPB descriptors were the extension of LOOP descriptors. From MLOOPB, different types of feature vectors such as the histogram and horizontal width variables were extracted. Then, the extracted features were reduced using a new approach, and classification was performed. The experimental process was conducted on the OU-ISIR B treadmill gait dataset and the CASIA B gait dataset and showed improved accuracy over the compared techniques. Habiba et al. [1] presented an integrated framework for HGR using deep features and a hybrid feature selection method. They employed pre-trained models for deep features extraction. The extracted features were selected using a fuzzy entropy controlled skewness (FEcS) approach. Four popular datasets were used for the experimental process, CASIA A, CASIA B, CASIA C, and AVAMVG and showed better performance than existing techniques.
Kooksung et al. [50] presented an automated feature extraction approach using a recurrent neural network (RNN)-based auto encoder (AE) for skeleton-based abnormal gait identification. They introduced two models named gated recurrent unit (GRU)-based AE (GRU AE) and an LSTM-based AE (LSTM AE). Using these two models, they extracted deep features and performed identification. Based on the experimental process, it was noted that the LSTM AE outperformed the GRU AE in terms of efficiency for gait identification. Anusha et al. [51] extracted low-dimensional features such as gradient, spatial, and texture for gait recognition. They used a nine-cell gait gradient magnitude image to compute a histogram of oriented gradients. After that, Haralick texture descriptors were extracted and concatenated for final recognition. The experiment was carried out on five gait datasets and obtained better accuracy than the existing compared techniques.
Khan et al. [5] presented a cross-view gait identification using a view-invariant gait representation model. They trained a deep fully connected neural network to convert gait descriptors from different points of view into a single canonical view. Later, the single canonical view descriptors were used for the experimental process. The experimental process was conducted on the OU-ISIR large population and CASIA-B datasets and showed that the presented cross-view gait recognition method had improved accuracy. Elharrouss et al. [52] presented a deep learning structure for HGR. They used multitask CNN architectures and extracted gait energy images (GEIs). Based on the GEI, features were extracted that were later classified using CNN. The experimental process was conducted on the CASIA B, OU-ISIR, and OU-MVLP datasets. The results on the selected datasets were better than current methods.
A few other studies were also conducted for HGR, such as PoseGait and the CNN model [13], a multichannel CNN architecture [38], GEI and CNN architecture [53], and RCNN based gait recognition [54], among others [55]. The methods mentioned focused on the middle step, for example, GEI, handcrafted features, and CNN. They did not focus on the important steps of the fusion of features from different models and selecting the best features for better computational time. This article proposes a new framework using deep learning features selection and multisource features fusion for HGR.

3. Dataset Detail

CASIA B Dataset: CASIA B [56] is commonly implemented to identify patterns in multiple viewing of various clothing conditions, also known as a multiview home dataset [57]. This database includes 124 headings with gait sequences registered from 11 different angles of 0, 18, 36, 54, 72, 90, 108, 126, 144, 162, and 180 degrees, with three different styles of walking sequences, including walking with a backpack, wearing a hat, and a regular walking sequence. Figure 1 shows sample frames from the CASIA B dataset. Each video of the gait sequence is captured at a rate of 25 frames per second with a resolution of 352 × 240 pixels. In this work, all angles were selected for the experimental process.

4. Materials and Methods

A new framework was proposed for human gait recognition, including many steps such as data augmentation, feature extraction, feature fusion, and feature selection. The architecture of the proposed framework is shown in Figure 2. In the augmentation step, three flip operations were used. In the feature extraction step, two pre-trained models were used named Inception-ResNet-V2 and NASNet Mobile. Both models were fine-tuned and trained using transfer learning on the CASIA B dataset (three classes; walk with bag, walk with coat, and basic walk for five different angles 18, 36, 54, 72, and 90, extracting the features). The features of the selected deep models were optimized using a modified three-step whale optimization algorithm and the best features were selected. The selected best features were fused using the modified mean absolute deviation extended serial fusion (MDeSF) approach. Finally, the fused vector was classified using several machine learning classifiers. The details of each substep is given below.

4.1. Data Augmentation

The process of expanding the quantity of data utilized for training a model is known as data augmentation [58]. When preparing a machine learning model, it works as a regularizer and helps to minimize overfitting. Some of the main techniques used in data augmentation in deep learning are: the flip operation, rotation operations, scale operation, crop operation, translation operation, and Gaussian noise. In this proposed framework, data augmentation was performed using translation operations such as simple translation, horizontal translation, and vertical translation. The primary purpose of data augmentation is to balance the data and improve accuracy. In this work, the CASIA B dataset was utilized for the evaluation process [56]. This dataset is available in the form of videos having three different classes—normal walk, walk carrying a bag, and walk wearing a coat. This dataset was captured from 11 different angles with a total of 124 subjects. The entire dataset was considered and split into training and testing, with a ratio of 50:50 instead of 70:30. This meant that 50% of the videos of each class were selected for training, and the remaining 50% were utilized for testing. Each angle of this dataset consisted of three classes, and each class contained a different number of video frames. Some classes had 2000 frames, and some had 2500. To balance the numbers of frames, translation operations were performed. By employing the translation operations, we increased the numbers of frames from 2500 to 6000 frames.

4.1.1. Left and Right Horizontal Translation

Mathematically, these operations were formulated as follows:
x = f ( z a )
where f ( z ) is the original function and a is a constant added or removed to create a horizontal change in the constant of function. If the function a is positive, it is moved to the right, and if the function a is negative, it is transferred to the left. To examine horizontal translations, the identical fundamental quadratic function is utilized. We have a basic quadratic equation:
x = z 2
By shifting the function to the left by two, the following equation is employed:
x = f ( z + a )
= ( x + 2 ) 2
Here, a is a constant. By shifting the function to the right by two, the following equation is employed:
x = f ( z a )
= ( x 2 ) 2
Vertical translation shifts the function up or down. It is defined as follows:
x = f ( z ) + a
where f ( z ) is the function and a is the constant of this function to enable translation. Consider the quadratic equation:
x = z 2

4.1.2. Up and Down Vertical Translation

Then, the vertical translation can be defined as follows:
x = z 2 + 2
x = z 2 2
Visually, the output of these operations is shown in Figure 3. The resultant images were later utilized to train the CNN models.

4.2. Convolutional Neural Networks (CNN)

The convolutional neural network (CNN) is a type of deep learning that includes multiple layers that surpasses classical methods such as object classification, among others [46,59]. The layers of CNN were formed by a convolutional operation followed by an activation layer (ReLu). Other layers are included in a CNN, such as pooling, normalization, fully connected, average pool, and Softmax. Figure 4 illustrates a simple CNN model for the classification task.

4.2.1. Convolutional Layer

Initially, the selected dataset frames were utilized as an input to the convolutional layer. A convolutional layer connects a set of M filters to a group of T channels and size A × B in comparison to a mini-batch of X Images with T channels and size H e i g h t × w i d h t h . The filter elements are denoted by E w , x , y , z and image elements denoted by F g , h , i , j   , then the formulation of the convolutional operation is defined as follows:
Y g , w , i , j = c = 1 T u = 1 R v = 1 S F g , h , i , j + y , j + z   E w , x , y , z
And the output of a whole image/filter combination can be written as:
Y g , w = c = 1 T F g , h   E w , x
where show the two-dimensional correlation.

4.2.2. ReLU Layer

The goal of ReLU is to augment the CNN’s nonlinearity. Using this layer, the negative weights values were converted to zero. The rest of the positive values will be considered the same for the next process.
R e L U ( A ) = m a x { 0 , a }
The weights were adjusted with the updating rule by employing the following equation:
W g t = w + δ γ F 2 γ w
where the new weight is W g t with current value w , and the learning rate is γ .

4.2.3. Pooling Layer

The pooling layer is mostly employed for the downsampling of a feature map. Normally, three types of pooling operations are employed: maximum pooling, average pooling, and minimum pooling. Mathematically, these operations are defined as follows:
P L i = M a x p o o l ( P L ) i 1
M a x p o o l represents the max pooling, and i is the input matrix.

4.2.4. Batch Normalization

Normalization methods can reduce the training time of the selected model by a large proportion [60]. Batch normalization is a technique that normalizes network activity over a defined mini-batch size. For the mini-batch size, mean and variance were computed. Then the mean was subtracted, and the feature was divided by its standard mini-batch deviation.
A 1 n i = 1 n x i
ω A 2 1 n i = 1 n ( x i A ) 2
where A represents the mini-batch mean, and ω A 2 is the mini-batch variance.

4.2.5. Fully Connected Layer

The CNN extracted feature was concatenated in the fully connected layer f f u l l = f C N N . Then, the classification result was produced using a softmax operation as described by the following formulation:
p = S o f t m a x ( f f u l l     w p + b p )
where p is the output sample, the weight matrix of the output layer is denoted by w p   and the output bias is denoted by b p . The fully connected Layer (FC) functions on a flattened input, where all neurons are linked to each input.

4.3. Transfer Learning

Transfer learning (TL) is a process of reusing a pre-trained model for another task. Originally, the pre-trained deep models were trained on the ImageNet dataset, which included 1000 object classes. Consider a domain B made up of the following components: a feature space W and a marginal probability distribution P ( W ) , where M = { m 1 , , m n } X . The next specific domain is W = M , P ( M ) . This task consists of two main components, label space L and objective predictive functions f :   A B , in the presence of a source domain X s and learning task Y s , a target domain X t and learning task Y t , where X s   X t and Y s     Y t . The goal of transfer learning is to aid in the learning of the target predictive function. This process is illustrated in Figure 5.
In this work, TL was carried out on two modified pre-trained models Inception-ResNet-V2 and NasNet-model. Both models were trained through TL on the CASIA B gait dataset, where several parameters were utilized, as shown in Table 1. These parameters were initialized based on the size of the dataset and performance of the Desktop computer. However, these parameters can be changed according to the learning environment. After training, these models were saved and utilized for features extraction.

4.4. Deep Features of Modified Inception-ResNetV2

Inception-ResNetV2 is a CNN model originally trained on the ImageNet dataset, consisting of 164 layers. The network is strong and has learned feature representations for a wide range of images types. This network accepts an input of 299 × 299 and hybridization of two recently developed networks, residual links, and the latest architectural version. There are series of filters such as 1 × 1 ,   3 × 3 ,   5 × 5 , etc., fused with each branch concatenation. The split-transformation-mixing structure of the initiation module is seen in its thick layers as a strong representation. Figure 6 depicts the inner architecture of Inception-ResNet-v2. The residual connections enable the training of increasingly deeper neural networks, resulting in an even more outstanding performance. In this work, this model was fine-tuned and trained using TL. The detail of TL is given in Section 4.3. Features were extracted from the GAP layer and a feature vector of dimension N × 1536   was obtained, where N represented the number of images used for the testing purpose.

4.5. NASNet-Mobile

NASNet-Mobile is a scalable CNN (built by neural architecture searches) architecture consisting of fundamental building blocks (cells) improved by reinforcement learning. A cell comprises a few operations (many separation convolutions and pooling) and is repeated numerous times depending on the needed network capacity. The Mobile version contains 12 cells with 5.3 million parameters and a multiply-accumulated total of 564 million (MACs) [61]. The fundamental element of NASNet incorporates a regularization technology from Scheduled DropPath, which substantially increases model generalization. In the fine-tuning process, new layers were added and the entire network was trained on the Gait dataset (CASIA B). After the training, features were extracted from the convolutional layer, and a feature vector of dimension N × 1056 was obtained.

4.6. Novelty 1: Feature Selection

This work used a modified whale optimizer algorithm (WOA) for the best features selection. The modified whale optimization algorithm works in three phases:
  • Phase of Exploitation (circle technique of attack of prey/bubble-net)
  • Phase of Exploration (searching prey)
  • Define a new activation function to check the redundant features

4.6.1. Phase of Exploitation (Circle Technique of Attack of Prey/Bubble-Net)

The following formulation depicts the movement of a whale near prey [62]:
A = | B · Z   ( T ) Z ( T ) |
Z ( T + 1 ) = Z ( T ) C A
where T shows the current iteration, Z shows the best solution obtained, Z shows the best solution, | | offers the absolute value, and (·) represents a multiplication symbol. The coefficient vectors A and C are discussed in the following equations:
C =   2   a ·   r a
A = 2 · r
where a decreases linearly from 2 to 0, and r is a random vector in the range [0,1]. According to the above equation, solutions are based on the position of the most known solution to update their locations.
a = 2 t 2 M a x I t e r
where T is the number of iterations, and M a x I t e r is the maximum number of iterations permitted. The current and best (leading) solutions are formulated as follows:
Z   ( T + 1 ) = K ´ · e d l · cos ( 2 π l ) + Z ( T )
where the distance is defined by ( A = |   Z   ( T ) Z ( T ) | ), d represents the spiral shape, and l determines the random number [ 1 , 1 ] .

4.6.2. Phase of Exploration (Searching Prey)

Instead of requiring solutions to search randomly based on the location of the best solution, a randomly chosen solution is utilized to update the position [62]. Consequently, a C vector is utilized to control a solution, which is deviating from the best-known search agent with random values greater or less than one. The following formulation can be utilized for this process.
C = | C   · Z   r a n d Z   |
Z ( T + 1 ) = Z   r a n d C · A
where Z   r a n d is a randomly selected whale from the existing population.

4.6.3. Phase of Final Selection

In the next step, the existing algorithm was modified in the final selection phase. The mean deviation was computed by Z and multiplied with the highest feature value of the entire selected vector. After that, we searched for the redundant features and each redundant feature was replaced with the product value multiplied by the i 1 feature. Mathematically, this process is defined as follows:
R d ( k ) = S e a r c h ( i , j ) R e p ( k ) = V a l × ( i 1 ) M A D = | ϕ ϕ ¯ | N H = M a x ( Z )
V a l = M A D × H
The selected best features are fused in a later step using the proposed fusion approach. In this work, the size of the final selected features was N × 560 and N × 766 ; however, this size can be updated according to the selected dataset dimension. The berief description is also added in Algorithm 1.
Algorithm 1 Modified Whale Optimization for Best Features Selection
1. 
Initial population generation x i ( i = 1 , 2 , , n )
2. 
Calculate each solution’s objective value
3. 
x * = best solution
4. 
w h i l e ( t < M a x _ I t e r a t i o n )
5. 
     f o r each solution
6. 
            U p d a t e   a ,   A ,   C ,   l ,   and   p
7. 
       i f   1 ( p < 0.5 )
8. 
                     i f   2 ( | A | < 1 )
9. 
      To update the current solution’s location, use Equation (20).
10. 
                     e l s e   i f   2 ( | A | < 1 )
11. 
       Choose a solution at random Z r a n d
12. 
          Use Equation (25)
13. 
                    e n d   i f   2
14. 
                    e l s e   i f   1 ( p 0.5 )
15. 
     Update the current search position using Equation (24)
16. 
                   e n d   i f   1
17. 
            e n d   f o r
18. 
Verify that some solution exceeds and modifies the search area.
19. 
Determine the fitness of each solution.
20. 
Update X * if a better solution becomes available.
21. 
            t = t + 1
22. 
     e n d   w h i l e
23. 
     r e t u r n   X *
24. 
Perform Search Function R d ( k )
25. 
     -Find Redundant Features
26. 
     -Replace with R e p ( k )
27. 
     -Again determine the fitness
28. 
   r e t u r n   X *

4.7. Novelty 2: Feature Fusion

Features fusion represents the combination of features from the different layers or branches [63,64]. In this work, a modified features fusion approach was employed, the mean absolute deviation extended serial fusion (MDeSF). This approach was implemented in two steps. In the first step, features of both vectors were sorted into the descending order and then serially fused in one matrix. In the second step, the mean absolute deviation was computed and a threshold function defined. Based on the threshold function, the final features were added in the fused vector.
Assume X and Y are two pattern sample space-defined feature spaces. The corresponding two feature vectors α X and β Y for an arbitrary sample were ȴ Ȗ . The feature vectors α and β were sorted into descending order using the following expression:
α ( i ) = D e s c ( α ) ,     β ( j ) = D e s c ( β )
The serial combination feature vector is described as:
Ƒ = ( α ( i ) β ( j ) )
If α is an a-dimensional feature vector and β is the b-dimensional, then the combined serial function ϕ is ( a + b ) [65]. In the next step, the mean absolute deviation was computed from the obtained feature vector Ƒ , and a threshold function was designed as follows:
M A D = | ϕ ϕ ¯ | N
T h d = { ψ ( k )           f o r       ϕ M A D I g n o r e ,             E l s e w h e r e
The above threshold function represented the features of the originally fused vector ϕ compared with the M A D value, those which are greater or higher are added in the final fused vector ψ ( k ) . The features, which did not fulfill this criterion, were removed from the fusion list. The resultant feature vector was finally classified using machine learning algorithms.

5. Results

5.1. Experimental Setup

The proposed method was evaluated on a publicly available dataset, CASIA B. Details of this dataset are given in Section 3. Seven experiments were performed in this work with two models: (i) Inception-Resnet-V2 and (ii) Nasnet Mobile. Initially, the experiments were conducted on deep individual models using a non-augmented dataset, and an augmented dataset was then employed. The main purpose of these experiments was to analyze the performance of the augmentation process. In the next experiment, a feature selection algorithm was applied on both deep models and finally, fusion was performed. The classifiers used in the experiments were C-SVM, Q-SVM, SD, MG-SVM, F-KNN, LD, L-SVM, S-KNN, W-KNN, M-KNN, and C-KNN. For each classifier, we included the recall rate, F1-Score, accuracy, and computational time. The proposed method was implemented inn MATLAB 2020b using a personal computer with a 64-bit operating system, 16 GB of RAM, and an 8 GB graphics card.

5.2. Results

The results of the proposed method are given in this section. The modified Inception-ResNet V2 model was trained on a non-augmented dataset, and deep features were extracted. The results are given in Table 2. This table shows that the maximum achieved accuracies of Cubic SVM on selected angles were 88.8, 84.6, 85.1, 76.4, and 78.9%. The F1-Score of Cubic SVM was also calculated, and the value of each angle was 88.8, 84.64, 85.1, 76.44, and 79.02%. The performance of Cubic SVM was also compared with some other classification methods such as QSVM, MG SVM, and SD and showed that the performance of Cubic SVM was better in terms of recall rate, F1-Score, and accuracy. Moreover, each classifier noted the computational time showing that the QSVM executed faster than other classifiers.
Table 3 presents the results of the modified NasNet Mobile on selected angles of the non-augmented CASIA B dataset. This table shows the maximum accuracies of 87.0%, 83.6%, 80.4%, 72.3%, and 73.0% for Cubic SVM. The F1-Score values of each angle for Cubic SVM were 86.97%, 83.66%, 80.39%, 72.34%, and 73.04%. Other classifiers were also implemented, such as QSVM, MG SVM, and SD, which showed that the Cubic SVM classifier performed better. However, the computational time of QSVM was better than the Cubic SVM, MGSVM, and SD. Moreover, the accuracy of Inception-ResNet V2 was better than NasNet Mobile, but the computationally modified NasNet Mobile was efficient.
Table 4 presents the results of the modified Inception-ResNet V2 on selected angles of the augmented CASIA B dataset. The Cubic SVM classifier had better classification accuracy of each angle, and the values were 95.6%, 926%, 92.7%, 86.2%, and 87.7%. Compared with the non-augmented results (shown in Table 2), the accuracy after the augmentation process was significantly improved. The values of the F1-Score of each angle for Cubic SVM were 95.58%, 92.62%, 92.7%, 86.16%, and 87.74%. Similarly, the performance of the rest of the classifiers such as QSVM, GSVM, and SD improved. The time was also noted for each classifier for this experiment. Based on the recorded time, the QSVM and Cubic SVM were both efficiently executed.
Table 5 presents the results of the modified NasNet Mobile on selected angles of the augmented CASIA B dataset. This table shows that the Cubic SVM obtained better accuracy than the other classifiers QSVM, MGSVM, and SD. The obtained accuracies of the Cubic SVM classifier were 94.4%, 92.1%, 90.9%, 83.5%, and 84.8%. Compared with the non-augmented results (shown in Table 3), the accuracy after the augmentation process significantly increased for each angle. The computational time was also noted for each classifier, and the QSVM and Cubic SVM were both efficiently executed.
Table 6 presents the results of the proposed best feature selection algorithm for Inception-ResNet V2 deep features. The augmented CASIA B dataset was utilized for the experimental process. This table presents the best accuracies obtained by Cubic SVM of 95.2%, 92.5%, 91.4%, 85.3%, and 86.4%. Compared with the originally extracted deep features results (shown in Table 4), the accuracy of each classifier for different angles was almost consistent; however, computationally, this step was more efficient. Moreover, the performance of the rest of the classifiers was improved after the employing of a modified WOA.
Similarly, Table 7 presents the results of the modified WOA for NasNet Mobile. This table shows the Cubic SVM obtained the best accuracies of 94.1%, 91.3%, 90.7%, 82.7%, and 84.0%. Moreover, the comparison of this experiment with Table 5 shows that the accuracies of this experiment were almost consistent; however, the computational time was significantly decreased. This shows the importance of this algorithm.
Finally, the selected best features of modified Inception-ResNet V2 and modified NasNet Mobile were fused using the proposed fusion method. The results of the features fusion are given in Table 8. This table shows the best accuracies were obtained by Cubic SVM with 97.3%, 96.0%, 95.3%, 86.6%, and 89.8%. The F1-Score was computed for each classifier and the best obtained values of 97.24%, 95.99%, 95.28%, 86.61%, and 88.06% were for Cubic SVM. The performance of the fusion process was compared with Table 3, Table 4, Table 5, Table 6 and Table 7; the fusion method improved the recognition accuracy. Moreover, the performance of the rest of the classifiers was also significantly improved. However, the limitation of this step was an increase in the computational time.
We briefly discuss the computational time and comparison of the proposed method results with the state-of-the-art (SETA) techniques. Figure 7 shows the time plot of the fusion process and that the Cubic SVM and QSVM executed faster than the MGSVM and SD. Figure 8 shows the computational time of Cubic SVM for all five experiments; the feature selection process consumed less time than the original feature vectors and fused vector.
Experiments on the Entire CASIA B Dataset: The proposed framework was finally evaluated on the entire CASIA B dataset and the results are presented in Table 9. Moreover, in this table, a detailed analysis of the proposed method was conducted based on standard error mean ( σ x ¯ ) and confidence interval (CI). The final fusion process was executed 100 times, and we obtained three values- minimum (Min), average (Avg), and maximum (Max). Based on these three values, standard deviation ( σ ), standard error mean, and CI were calculated. The CI was calculated for the confidence interval 95%, 1.960 σ x ¯ . On this confidence level, the proposed method accuracy was noted to be consistent after the initialized iterations.
Moreover, a comparison was conducted with the SOTA techniques, shown in Table 10. The references used in this table used the CASIA B dataset for the experimental process. Each method considered all 11 angles for the experimental process. The mean accuracy of each method was added instead of individual angle-based accuracy. The average accuracy was computed based on variations including the normal walk, walk carrying a bag, and walk wearing a coat. Based on the mean values given in Table 10, the proposed method’s accuracy was good.

6. Conclusions

A new framework was proposed for human gait recognition using deep learning. The proposed framework included data augmentation, feature extraction, feature fusion, and feature selection. In the augmentation step, three flip operations were used. The main reason behind data augmentation was to improve the training data, increasing the classification accuracy. In the feature extraction step, two pre-trained models were fine-tuned, Inception-ResNet-V2 and NASNet Mobile. Both models were trained using transfer learning on the CASIA B dataset. The accuracy of these original deep features was not sufficient for the classification; therefore, a feature fusion technique was proposed to improve the classification accuracy. However, this step increased the computational time. A feature selection method was proposed to resolve this problem, which increased the classification accuracy and lowered the computational cost. Overall, the proposed framework achieved better accuracy on the selected dataset as compared with recent techniques. However, the proposed framework also has a few limitations: (i) the modified whale optimization algorithm selected features in two steps; however, there is a chance of reduction in a few important features that later could decrease the classification accuracy, and (ii) the fusion of features using the MAD based threshold function could ignore some important features. In the future, the following steps could be considered:
  • Fusion can be performed using various threshold function and then select the best of them based on the accuracy.
  • Moreover, OU-MVLP, OU-LP-BAG, TUM-GAID datasets will also be considered for the experimental process.
  • Adopt a two-stream approach such as optical flow based and raw images steps for the more accurate recognition accuracy.
  • Extract deep features using latest deep models like Efficient Net.

Author Contributions

Conceptualization, F.S. and M.A.K.; methodology, F.S. and M.A.K.; software, M.A.K. and M.A.; validation, M.A., U.T. and A.A.; formal analysis, U.T. and F.A; investigation, U.T. and M.A.; resources, A.A. and F.A.; data curation, M.A.K. and J.-I.C.; writing—original draft preparation, M.A.K. and F.S.; writing—review and editing, S.K. and J.-I.C.; visualization, S.K. and J.-I.C.; supervision, S.K.; project administration, M.A.K. and J.-I.C.; funding acquisition, J.-I.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (2018R1D1A1B07042967).

Data Availability Statement

The CASIA b dataset was utilized for the experimental process. This dataset is publically available for research purposes at the Center for Biometrics and Security Research (ia.ac.cn).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Arshad, H.; Khan, M.A.; Sharif, M.I.; Yasmin, M.; Tavares, J.M.R.; Zhang, Y.D.; Satapathy, S.C. A multilevel paradigm for deep convolutional neural network features selection with an application to human gait recognition. Expert Syst. 2020, e12541. [Google Scholar] [CrossRef]
  2. Manssor, S.A.; Sun, S.; Elhassan, M.A. Real-Time Human Recognition at Night via Integrated Face and Gait Recognition Technologies. Sensors 2021, 21, 4323. [Google Scholar] [CrossRef]
  3. Mehmood, A.; Khan, M.A.; Sharif, M.; Khan, S.A.; Shaheen, M.; Saba, T.; Riaz, N.; Ashraf, I. Prosperous human gait recognition: An end-to-end system based on pre-trained CNN features selection. Multimed. Tools Appl. 2020, 1–21. [Google Scholar] [CrossRef]
  4. Wang, L.; Li, Y.; Xiong, F.; Zhang, W. Gait Recognition Using Optical Motion Capture: A Decision Fusion Based Method. Sensors 2021, 21, 3496. [Google Scholar] [CrossRef]
  5. Khan, M.H.; Farid, M.S.; Grzegorzek, M. A non-linear view transformations model for cross-view gait recognition. Neurocomputing 2020, 402, 100–111. [Google Scholar] [CrossRef]
  6. Sharif, M.; Attique, M.; Tahir, M.Z.; Yasmim, M.; Saba, T.; Tanik, U.J. A machine learning method with threshold based parallel feature fusion and feature selection for automated gait recognition. J. Organ. End User Comput. (JOEUC) 2020, 32, 67–92. [Google Scholar] [CrossRef]
  7. Fricke, C.; Alizadeh, J.; Zakhary, N.; Woost, T.B.; Bogdan, M.; Classen, J. Evaluation of three machine learning algorithms for the automatic classification of EMG patterns in gait disorders. Front. Neurol. 2021, 12, 666458. [Google Scholar] [CrossRef]
  8. Gao, F.; Tian, T.; Yao, T.; Zhang, Q. Human gait recognition based on multiple feature combination and parameter optimization algorithms. Comput. Intell. Neurosci. 2021, 2021, 6693206. [Google Scholar] [CrossRef]
  9. Yao, T.; Gao, F.; Zhang, Q.; Ma, Y. Multi-feature gait recognition with DNN based on sEMG signals. Math. Biosci. Eng. 2021, 18, 3521–3542. [Google Scholar] [CrossRef]
  10. Steinmetzer, T.; Wilberg, S.; Bönninger, I.; Travieso, C.M. Analyzing gait symmetry with automatically synchronized wearable sensors in daily life. Microprocess. Microsyst. 2020, 77, 103118. [Google Scholar] [CrossRef]
  11. Mei, C.; Gao, F.; Li, Y. A determination method for gait event based on acceleration sensors. Sensors 2019, 19, 5499. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Luo, R.; Sun, S.; Zhang, X.; Tang, Z.; Wang, W. A low-cost end-to-end sEMG-based gait sub-phase recognition system. IEEE Trans. Neural Syst. Rehabil. Eng. 2019, 28, 267–276. [Google Scholar] [CrossRef] [PubMed]
  13. Liao, R.; Yu, S.; An, W.; Huang, Y. A model-based gait recognition method with body pose and human prior knowledge. Pattern Recognit. 2020, 98, 107069. [Google Scholar] [CrossRef]
  14. Addabbo, P.; Bernardi, M.L.; Biondi, F.; Cimitile, M.; Clemente, C.; Orlando, D. Temporal convolutional neural networks for radar micro-Doppler based gait recognition. Sensors 2021, 21, 381. [Google Scholar] [CrossRef]
  15. Khan, M.A.; Khan, M.A.; Ahmed, F.; Mittal, M.; Goyal, L.M.; Hemanth, D.J.; Satapathy, S.C. Gastrointestinal diseases segmentation and classification based on duo-deep architectures. Pattern Recognit. Lett. 2020, 131, 193–204. [Google Scholar] [CrossRef]
  16. Khan, M.A.; Kadry, S.; Parwekar, P.; Damaševičius, R.; Mehmood, A.; Khan, J.A.; Naqvi, S.R. Human gait analysis for osteoarthritis prediction: A framework of deep learning and kernel extreme learning machine. Complex Intell. Syst. 2021, 1–19. [Google Scholar] [CrossRef]
  17. Arshad, H.; Khan, M.A.; Sharif, M.; Yasmin, M.; Javed, M.Y. Multi-level features fusion and selection for human gait recognition: An optimized framework of Bayesian model and binomial distribution. Int. J. Mach. Learn. Cybern. 2019, 10, 3601–3618. [Google Scholar] [CrossRef]
  18. Farnoosh, A.; Wang, Z.; Zhu, S.; Ostadabbas, S. A Bayesian Dynamical Approach for Human Action Recognition. Sensors 2021, 21, 5613. [Google Scholar] [CrossRef]
  19. Shiraga, K.; Makihara, Y.; Muramatsu, D.; Echigo, T.; Yagi, Y. Geinet: View-invariant gait recognition using a convolutional neural network. In Proceedings of the 2016 International Conference on Biometrics (ICB), Halmstad, Sweden, 13–16 June 2016; pp. 1–8. [Google Scholar]
  20. Liu, J.; Zheng, N. Gait history image: A novel temporal template for gait recognition. In Proceedings of the 2007 IEEE International Conference on Multimedia and Expo, Beijing, China, 2–5 July 2007; pp. 663–666. [Google Scholar]
  21. Lv, Z.; Xing, X.; Wang, K.; Guan, D. Class energy image analysis for video sensor-based gait recognition: A review. Sensors 2015, 15, 932–964. [Google Scholar] [CrossRef] [Green Version]
  22. Chen, K.; Wu, S.; Li, Z. Gait Recognition Based on GFHI and Combined Hidden Markov Model. In Proceedings of the 2020 13th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI), Chengdu, China, 17–19 October 2020; pp. 287–292. [Google Scholar]
  23. Khan, M.A.; Akram, T.; Sharif, M.; Muhammad, N.; Javed, M.Y.; Naqvi, S.R. Improved strategy for human action recognition; experiencing a cascaded design. IET Image Process. 2019, 14, 818–829. [Google Scholar] [CrossRef]
  24. Sharif, A.; Khan, M.A.; Javed, K.; Gulfam, H.; Iqbal, T.; Saba, T.; Ali, H.; Nisar, W. Intelligent human action recognition: A framework of optimal features selection based on Euclidean distance and strong correlation. J. Control Eng. Appl. Inform. 2019, 21, 3–11. [Google Scholar]
  25. Arulananth, T.; Balaji, L.; Baskar, M.; Anbarasu, V.; Rao, K.S. PCA based dimensional data reduction and segmentation for DICOM images. Neural Process. Lett. 2020, 1–15. [Google Scholar] [CrossRef]
  26. Maitra, S.; Yan, J. Principle component analysis and partial least squares: Two dimension reduction techniques for regression. Appl. Multivar. Stat. Models 2008, 79, 79–90. [Google Scholar]
  27. Laohakiat, S.; Phimoltares, S.; Lursinsap, C. A clustering algorithm for stream data with LDA-based unsupervised localized dimension reduction. Inf. Sci. 2017, 381, 104–123. [Google Scholar] [CrossRef]
  28. Khan, M.A.; Sharif, M.; Javed, M.Y.; Akram, T.; Yasmin, M.; Saba, T. License number plate recognition system using entropy-based features selection approach with SVM. IET Image Process. 2018, 12, 200–209. [Google Scholar] [CrossRef]
  29. Siddiqui, S.; Khan, M.A.; Bashir, K.; Sharif, M.; Azam, F.; Javed, M.Y. Human action recognition: A construction of codebook by discriminative features selection approach. Int. J. Appl. Pattern Recognit. 2018, 5, 206–228. [Google Scholar] [CrossRef]
  30. Khan, M.A.; Arshad, H.; Nisar, W.; Javed, M.Y.; Sharif, M. An Integrated Design of Fuzzy C-Means and NCA-Based Multi-properties Feature Reduction for Brain Tumor Recognition. In Signal and Image Processing Techniques for the Development of Intelligent Healthcare Systems; Springer: Berlin/Heidelberg, Germany, 2021; pp. 1–28. [Google Scholar]
  31. Jiang, X.; Zhang, Y.; Yang, Q.; Deng, B.; Wang, H. Millimeter-Wave Array Radar-Based Human Gait Recognition Using Multi-Channel Three-Dimensional Convolutional Neural Network. Sensors 2020, 20, 5466. [Google Scholar] [CrossRef]
  32. Khan, M.A.; Zhang, Y.-D.; Khan, S.A.; Attique, M.; Rehman, A.; Seo, S. A resource conscious human action recognition framework using 26-layered deep convolutional neural network. Multimed. Tools Appl. 2020, 1–23. [Google Scholar] [CrossRef]
  33. Lu, S.; Wang, S.-H.; Zhang, Y.-D. Detection of abnormal brain in MRI via improved AlexNet and ELM optimized by chaotic bat algorithm. Neural Comput. Appl. 2020, 33, 10799–10811. [Google Scholar] [CrossRef]
  34. Liu, S.; Wang, X.; Zhao, L.; Li, B.; Hu, W.; Yu, J.; Zhang, Y. 3DCANN: A Spatio-Temporal Convolution Attention Neural Network for EEG Emotion Recognition. IEEE J. Biomed. Health Inform. 2021. [Google Scholar] [CrossRef]
  35. Wang, S.; Celebi, M.E.; Zhang, Y.-D.; Yu, X.; Lu, S.; Yao, X.; Zhou, Q.; Miguel, M.-G.; Tian, Y.; Gorriz, J.M. Advances in Data Preprocessing for Biomedical Data Fusion: An Overview of the Methods, Challenges, and Prospects. Inf. Fusion 2021, 76, 376–421. [Google Scholar] [CrossRef]
  36. Liu, X.; Song, L.; Liu, S.; Zhang, Y. A review of deep-learning-based medical image segmentation methods. Sustainability 2021, 13, 1224. [Google Scholar] [CrossRef]
  37. Zhang, Y.-D.; Dong, Z.; Wang, S.-H.; Yu, X.; Yao, X.; Zhou, Q.; Hu, H.; Li, M.; Jiménez-Mesa, C.; Ramirez, J. Advances in multimodal data fusion in neuroimaging: Overview, challenges, and novel orientation. Inf. Fusion 2020, 64, 149–187. [Google Scholar] [CrossRef]
  38. Wang, X.; Zhang, J.; Yan, W.Q. Gait recognition using multichannel convolution neural networks. Neural Comput. Appl. 2019, 32, 14275–14285. [Google Scholar] [CrossRef]
  39. Khan, M.A.; Zhang, Y.-D.; Alhusseni, M.; Kadry, S.; Wang, S.-H.; Saba, T.; Iqbal, T. A Fused Heterogeneous Deep Neural Network and Robust Feature Selection Framework for Human Actions Recognition. Arab. J. Sci. Eng. 2021, 1–16. [Google Scholar] [CrossRef]
  40. Hussain, N.; Khan, M.A.; Kadry, S.; Tariq, U.; Mostafa, R.R.; Choi, J.-I.; Nam, Y. Intelligent Deep Learning and Improved Whale Optimization Algorithm Based Framework for Object Recognition. Hum. Cent. Comput. Inf. Sci. 2021, 11, 34. [Google Scholar]
  41. Hussain, N.; Khan, M.A.; Tariq, U.; Kadry, S.; Yar, M.A.E.; Mostafa, A.M.; Alnuaim, A.A.; Ahmad, S. Multiclass Cucumber Leaf Diseases Recognition Using Best Feature Selection. Comput. Mater. Contin. 2022, 70, 3281–3294. [Google Scholar] [CrossRef]
  42. Kiran, S.; Khan, M.A.; Javed, M.Y.; Alhaisoni, M.; Tariq, U.; Nam, Y.; Damaševičius, R.; Sharif, M. Multi-Layered Deep Learning Features Fusion for Human Action Recognition. Comput. Mater. Contin. 2021, 69, 4061–4075. [Google Scholar] [CrossRef]
  43. Majid, A.; Khan, M.A.; Nam, Y.; Tariq, U.; Roy, S.; Mostafa, R.R.; Sakr, R.H. COVID19 classification using CT images via ensembles of deep learning models. Comput. Mater. Contin. 2021, 69, 319–337. [Google Scholar] [CrossRef]
  44. Khan, M.A.; Alhaisoni, M.; Armghan, A.; Alenezi, F.; Tariq, U.; Nam, Y.; Akram, T. Video Analytics Framework for Human Action Recognition. CMC-Comput. Mater. Contin. 2021, 68, 3841–3859. [Google Scholar] [CrossRef]
  45. Afza, F.; Khan, M.A.; Sharif, M.; Kadry, S.; Manogaran, G.; Saba, T.; Ashraf, I.; Damaševičius, R. A framework of human action recognition using length control features fusion and weighted entropy-variances based feature selection. Image Vis. Comput. 2021, 106, 104090. [Google Scholar] [CrossRef]
  46. Hussain, N.; Khan, M.A.; Sharif, M.; Khan, S.A.; Albesher, A.A.; Saba, T.; Armaghan, A. A deep neural network and classical features based scheme for objects recognition: An application for machine inspection. Multimed. Tools Appl. 2020, 1–23. [Google Scholar] [CrossRef]
  47. Gul, S.; Malik, M.I.; Khan, G.M.; Shafait, F. Multi-view gait recognition system using spatio-temporal features and deep learning. Expert Syst. Appl. 2021, 179, 115057. [Google Scholar] [CrossRef]
  48. Davarzani, S.; Saucier, D.; Peranich, P.; Carroll, W.; Turner, A.; Parker, E.; Middleton, C.; Nguyen, P.; Robertson, P.; Smith, B. Closing the wearable gap—Part VI: Human gait recognition using deep learning methodologies. Electronics 2020, 9, 796. [Google Scholar] [CrossRef]
  49. Anusha, R.; Jaidhar, C. Clothing invariant human gait recognition using modified local optimal oriented pattern binary descriptor. Multimed. Tools Appl. 2020, 79, 2873–2896. [Google Scholar] [CrossRef]
  50. Jun, K.; Lee, D.-W.; Lee, K.; Lee, S.; Kim, M.S. Feature extraction using an RNN autoencoder for skeleton-based abnormal gait recognition. IEEE Access 2020, 8, 19196–19207. [Google Scholar] [CrossRef]
  51. Anusha, R.; Jaidhar, C. Human gait recognition based on histogram of oriented gradients and Haralick texture descriptor. Multimed. Tools Appl. 2020, 79, 8213–8234. [Google Scholar] [CrossRef]
  52. Elharrouss, O.; Almaadeed, N.; Al-Maadeed, S.; Bouridane, A. Gait recognition for person re-identification. J. Supercomput. 2020, 77, 3653–3672. [Google Scholar] [CrossRef]
  53. Nithyakani, P.; Shanthini, A.; Ponsam, G. Human gait recognition using deep convolutional neural network. In Proceedings of the 2019 3rd International Conference on Computing and Communications Technologies (ICCCT), Chennai, India, 21–22 February 2019; pp. 208–211. [Google Scholar]
  54. Hasan, M.M.; Mustafa, H.A. Multi-level feature fusion for robust pose-based gait recognition using RNN. Int. J. Comput. Sci. Inf. Secur. (IJCSIS) 2020, 18, 20–31. [Google Scholar]
  55. Bai, X.; Hui, Y.; Wang, L.; Zhou, F. Radar-based human gait recognition using dual-channel deep convolutional neural network. IEEE Trans. Geosci. Remote Sens. 2019, 57, 9767–9778. [Google Scholar] [CrossRef]
  56. Zheng, S.; Zhang, J.; Huang, K.; He, R.; Tan, T. Robust view transformation model for gait recognition. In Proceedings of the 2011 18th IEEE International Conference on Image Processing, Brussels, Belgium, 11–14 September 2011; pp. 2073–2076. [Google Scholar]
  57. Chao, H.; He, Y.; Zhang, J.; Feng, J. Gaitset: Regarding gait as a set for cross-view gait recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January 27–1 February 2019; pp. 8126–8133. [Google Scholar]
  58. Lavin, A.; Gray, S. Fast algorithms for convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4013–4021. [Google Scholar]
  59. Rashid, M.; Khan, M.A.; Alhaisoni, M.; Wang, S.-H.; Naqvi, S.R.; Rehman, A.; Saba, T. A sustainable deep learning framework for object recognition using multi-layers deep features fusion and selection. Sustainability 2020, 12, 5037. [Google Scholar] [CrossRef]
  60. Zou, Q.; Wang, Y.; Wang, Q.; Zhao, Y.; Li, Q. Deep learning-based gait recognition using smartphones in the wild. IEEE Trans. Inf. Forensics Secur. 2020, 15, 3197–3212. [Google Scholar] [CrossRef] [Green Version]
  61. Vedaldi, A.; Lenc, K. Matconvnet: Convolutional neural networks for matlab. In Proceedings of the 23rd ACM International Conference on Multimedia, Brisbane, Australia, 26–30 October 2015; pp. 689–692. [Google Scholar]
  62. Mafarja, M.; Mirjalili, S. Whale optimization approaches for wrapper feature selection. Appl. Soft Comput. 2018, 62, 441–453. [Google Scholar] [CrossRef]
  63. Adeel, A.; Khan, M.A.; Akram, T.; Sharif, A.; Yasmin, M.; Saba, T.; Javed, K. Entropy-controlled deep features selection framework for grape leaf diseases recognition. Expert Syst. 2020. [Google Scholar] [CrossRef]
  64. Khan, M.A.; Sharif, M.I.; Raza, M.; Anjum, A.; Saba, T.; Shad, S.A. Skin lesion segmentation and classification: A unified framework of deep neural network features fusion and selection. Expert Syst. 2019, e12497. [Google Scholar] [CrossRef]
  65. Yang, J.; Yang, J.-y.; Zhang, D.; Lu, J.-F. Feature fusion: Parallel strategy vs. serial strategy. Pattern Recognit. 2003, 36, 1369–1381. [Google Scholar] [CrossRef]
  66. Wu, Z.; Huang, Y.; Wang, L.; Wang, X.; Tan, T. A comprehensive study on cross-view gait based human identification with deep cnns. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 209–226. [Google Scholar] [CrossRef]
Figure 1. Sample video frames from the CASIA B dataset.
Figure 1. Sample video frames from the CASIA B dataset.
Sensors 21 07584 g001
Figure 2. The main architecture of the proposed methodology for HGR.
Figure 2. The main architecture of the proposed methodology for HGR.
Sensors 21 07584 g002
Figure 3. Output of the data augmentation step.
Figure 3. Output of the data augmentation step.
Sensors 21 07584 g003
Figure 4. Convolutional neural network for the classification task.
Figure 4. Convolutional neural network for the classification task.
Sensors 21 07584 g004
Figure 5. Process of transfer learning for HGR.
Figure 5. Process of transfer learning for HGR.
Sensors 21 07584 g005
Figure 6. Interior block models of inception-ResNet-V2.
Figure 6. Interior block models of inception-ResNet-V2.
Sensors 21 07584 g006
Figure 7. Computational time of the features fusion process using the augmented CASIA B dataset.
Figure 7. Computational time of the features fusion process using the augmented CASIA B dataset.
Sensors 21 07584 g007
Figure 8. Computational time of Cubic SVM for all five steps.
Figure 8. Computational time of Cubic SVM for all five steps.
Sensors 21 07584 g008
Table 1. Configuration parameters of the modified deep learning models.
Table 1. Configuration parameters of the modified deep learning models.
Epochs100
Learning Rate0.05
Mini-Batch Size64
Learning MethodSGD
WeightLearnRateFactor20
LearnRateDropPeriod6
Table 2. Gait recognition results using modified inception-ResNet V2 on selected angles of the non-augmented CASIA B dataset.
Table 2. Gait recognition results using modified inception-ResNet V2 on selected angles of the non-augmented CASIA B dataset.
ClassifierAnglesPerformance Measures
Cubic (SVM)1836547290Recall Rate (%)F1-ScoreAccuracy (%)Time
(s)
88.7388.8088.8134.74
84.784.6484.6219.09
85.185.185.1234.95
76.576.4476.4238.18
7979.0278.9207.65
Quadratic (QSVM) 8787.1287.1122.2
83.583.4483.4205.2
83.183.0683.0213.94
75.975.8675.8222.8
78.578.7278.4197.3
Medium Gaussian (MGSVM) 81.1681.4481.3150.09
78.278.1478.1261.9
77.2377.4777.2269.01
72.5672.5772.5268.82
75.576.4575.6246.48
Subspace Discriminant (SD) 81.1681.4481.3150.09
78.278.1478.1261.9
77.2377.4777.2269.01
72.5672.5772.5268.82
75.576.4575.6246.48
Table 3. Gait recognition results using modified NasNet-Mobile on selected angles of the non-augmented CASIA B dataset.
Table 3. Gait recognition results using modified NasNet-Mobile on selected angles of the non-augmented CASIA B dataset.
ClassifierAnglesPerformance Measures
Cubic (SVM)1836547290Recall Rate (%)F1-ScoreAccuracy (%)Time
(s)
86.9386.9787.0140.32
83.783.6683.6752.26
80.4380.3980.4165.36
72.3672.3472.3174.19
73.0373.0473.0161.92
Quadratic (SVM) 85.285.2985.3136.44
81.5681.4781.4147.41
79.2679.2079.2159.89
71.4371.4171.4162.87
73.2373.2773.2152.06
Medium Gaussian (SVM) 82.5682.7782.6179.14
78.9678.9478.9171.86
74.8374.8474.7189.51
69.669.6269.5195.77
70.170.7770.2188.1
Subspace Discriminant 81.881.8681.9244.1
75.375.2475.2231.83
74.774.5974.5246.87
69.2369.1769.1245.52
70.5370.5970.4358.05
Table 4. Gait recognition results using modified Inception-ResNet V2 on selected angles of the augmented CASIA B dataset.
Table 4. Gait recognition results using modified Inception-ResNet V2 on selected angles of the augmented CASIA B dataset.
ClassifierAnglesPerformance Measures
Cubic (SVM)1836547290Recall Rate (%)F1-ScoreAccuracy (%)Time
(s)
95.5695.5895.6774.94
92.692.6292.6416.79
92.792.792.7470.59
86.1686.1686.2533.7
87.7687.7487.7455.63
Quadratic (SVM) 93.7393.7493.7613.87
89.8689.8889.9405.87
90.6390.6190.6466.01
83.4683.4283.5546.39
84.2384.2184.2453.3
Medium Gaussian (SVM) 88.9389.0988.9877.84
85.5685.5985.6553.11
86.2686.3086.3617.03
79.9379.8979.9645.77
81.2381.2181.2553.25
Subspace Discriminant 89.189.1489.11101.9
86.486.486.4814.27
86.3686.3686.4696.59
79.5679.4979.6708.82
81.8381.7781.8860.23
Table 5. Gait recognition results using modified NasNet Mobile on selected angles of the augmented CASIA B dataset.
Table 5. Gait recognition results using modified NasNet Mobile on selected angles of the augmented CASIA B dataset.
ClassifierAnglesPerformance Measures
Cubic (SVM)1836547290Recall Rate (%)F1-ScoreAccuracy (%)Time
(s)
94.4394.4694.4312.57
92.1392.1392.1298.68
90.990.990.9341.38
83.583.583.5408.83
84.884.7884.8353.21
Quadratic (SVM) 92.4392.4392.4318.57
89.7689.7689.7311.06
87.8387.8387.8341.79
80.880.7480.8394.58
81.7381.7181.7336.79
Medium Gaussian (SVM) 90.6590.7090.6408.38
87.287.2187.2384.16
84.384.384.1424.26
77.677.677.6484.31
78.8378.8678.9408.07
Subspace Discriminant 82.982.9282.9372.4
80.1380.1480.2461.67
79.3679.3479.4392.91
77.6375.5473.6367.88
75.4675.4275.5345.69
Table 6. Gait recognition results using modified WOA for Inception-ResNet V2 on the augmented CASIA B dataset.
Table 6. Gait recognition results using modified WOA for Inception-ResNet V2 on the augmented CASIA B dataset.
ClassifierAnglesPerformance Measures
Cubic (SVM)1836547290Recall Rate (%)F1- ScoreAccuracy (%)Time
(s)
95.1695.1695.2318.24
92.5392.5392.5327.51
91.3691.3691.4220.06
85.385.385.3283.39
86.3386.3186.4288.9
Quadratic (SVM) 92.9392.9692.9319
89.4389.589.5327.05
88.988.8888.9213.95
83.0683.0283.1273.6
83.3683.3283.4284.15
Medium Gaussian (SVM) 88.888.9688.8419.4
8686.0186.0416.31
85.185.1485.1251.64
8079.9680.0309.81
80.7380.7380.7350.5
Subspace Discriminant 85.2385.2985.2435.82
83.6383.5983.6380.47
77.6677.6677.7153.97
75.775.5975.7221.19
78.8378.7778.8331.38
Table 7. Gait recognition results using modified WOA for NasNet Mobile on the augmented CASIA B dataset.
Table 7. Gait recognition results using modified WOA for NasNet Mobile on the augmented CASIA B dataset.
ClassifierAnglesPerformance Measures
Cubic (SVM)1836547290Recall Rate (%)F1- ScoreAccuracy (%)Time
(s)
94.194.194.1208.83
91.3391.3391.3176.67
90.790.6890.7347.54
82.6682.6982.7311.67
83.9683.9884.0270.72
Quadratic (SVM) 91.7391.7691.7213.91
88.2388.2488.2184.07
87.8387.8187.8341.09
79.3679.3279.4304.87
80.8380.8380.8254.24
Medium Gaussian (SVM) 9090.1190.0257.75
86.0686.0986.1198.9
83.683.6783.8415.73
77.177.177.1369.89
77.8377.877.8294.81
Subspace Discriminant 78.678.6278.6193.54
73.6673.6673.7146.74
78.278.278.2357.09
71.575.4271.5258.36
72.1372.1672.2205.73
Table 8. Proposed gait recognition results using the selected best features’ fusion.
Table 8. Proposed gait recognition results using the selected best features’ fusion.
ClassifierAnglesPerformance Measures
Cubic (SVM)1836547290Recall Rate (%)F1- ScoreAccuracy (%)Time
(s)
97.2397.2497.3604.06
96.0395.9996.0458.67
95.395.2895.3667.35
86.6386.6186.6858.23
86.4388.0689.8746.09
Quadratic (SVM) 95.9695.9896565.47
95.0695.0895.1417.4
93.493.4493.4631.4
84.184.0484.1841.61
87.3687.3487.4689.05
Medium Gaussian (SVM) 94.0394.1194.1882.62
93.0393.0393628.42
91.591.5391.5867.96
81.8681.8081.91089.3
85.7385.7385.7927.9
Subspace Discriminant 94.6694.6894.7719.94
93.1693.1693.1650.3
92.2392.2492.21356.5
82.882.7482.81411.7
87.2687.2387.21504.8
Table 9. Proposed gait recognition results using best-selected features fusion.
Table 9. Proposed gait recognition results using best-selected features fusion.
Angle ° Min (%)Avg (%)Max (%) σ σ x ¯ CI
071.472.373.20.900.636372.3 ± 1.247 (±1.73%)
1896.196.7097.30.600.424296.7 ± 0.832 (±0.86%)
3695.395.6596.00.350.247495.65 ± 0.485 (±0.51%)
5493.894.5595.30.750.530394.55 ± 1.039 (±1.10%)
7285.285.9086.60.700.494985.9 ± 0.97 (±1.13%)
9088.589.1589.80.650.459689.15 ± 0.901 (±1.10%)
10893.193.8594.60.750.530393.85 ±1.039 (±1.11%)
12691.892.893.81.00.707192.8 ± 1.386 (±1.49%)
14478.680.081.41.40.989980 ± 1.94 (±2.43%)
16288.489.690.81.20.848589.6 ± 1.663 (±1.86%)
18077.678.380.30.750.530378.35 ± 1.039 (±1.33%)
Table 10. Comparison of proposed method with STOA techniques.
Table 10. Comparison of proposed method with STOA techniques.
ReferenceAngleMean Accuracy (%)
Chao et al. [57], 2019All 11 angles with three variations84.20
Wu et al. [66], 2016All 11 angles with three variations73.49
Mehmood et al. [3], 202018, 36, 5494.26
ProposedAll 11 angles with three variations89.00
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Saleem, F.; Khan, M.A.; Alhaisoni, M.; Tariq, U.; Armghan, A.; Alenezi, F.; Choi, J.-I.; Kadry, S. Human Gait Recognition: A Single Stream Optimal Deep Learning Features Fusion. Sensors 2021, 21, 7584. https://doi.org/10.3390/s21227584

AMA Style

Saleem F, Khan MA, Alhaisoni M, Tariq U, Armghan A, Alenezi F, Choi J-I, Kadry S. Human Gait Recognition: A Single Stream Optimal Deep Learning Features Fusion. Sensors. 2021; 21(22):7584. https://doi.org/10.3390/s21227584

Chicago/Turabian Style

Saleem, Faizan, Muhammad Attique Khan, Majed Alhaisoni, Usman Tariq, Ammar Armghan, Fayadh Alenezi, Jung-In Choi, and Seifedine Kadry. 2021. "Human Gait Recognition: A Single Stream Optimal Deep Learning Features Fusion" Sensors 21, no. 22: 7584. https://doi.org/10.3390/s21227584

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop