Next Article in Journal
Active Claw-Shaped Dry Electrodes for EEG Measurement in Hair Areas
Previous Article in Journal
Correlation between Malocclusion and Mandibular Fractures: An Experimental Study Comparing Dynamic Finite Element Models and Clinical Case Studies
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Gait Recognition and Assistance Parameter Prediction Determination Based on Kinematic Information Measured by Inertial Measurement Units

1
Engineering Research Center of the Ministry of Education for Intelligent Rehabilitation Equipment and Detection Technologies, Hebei University of Technology, Tianjin 300401, China
2
The Hebei Key Laboratory of Robot Sensing and Human-Robot Interaction, Hebei University of Technology, Tianjin 300401, China
3
School of Mechanical Engineering, Hebei University of Technology, Tianjin 300401, China
*
Author to whom correspondence should be addressed.
Bioengineering 2024, 11(3), 275; https://doi.org/10.3390/bioengineering11030275
Submission received: 9 February 2024 / Revised: 6 March 2024 / Accepted: 10 March 2024 / Published: 13 March 2024

Abstract

:
The gait recognition of exoskeletons includes motion recognition and gait phase recognition under various road conditions. The recognition of gait phase is a prerequisite for predicting exoskeleton assistance time. The estimation of real-time assistance time is crucial for the safety and accurate control of lower-limb exoskeletons. To solve the problem of predicting exoskeleton assistance time, this paper proposes a gait recognition model based on inertial measurement units that combines the real-time motion state recognition of support vector machines and phase recognition of long short-term memory networks. A recognition validation experiment was conducted on 30 subjects to determine the reliability of the gait recognition model. The results showed that the accuracy of motion state and gait phase were 99.98% and 98.26%, respectively. Based on the proposed SVM-LSTM gait model, exoskeleton assistance time was predicted. A test was conducted on 10 subjects, and the results showed that using assistive therapy based on exercise status and gait stage can significantly improve gait movement and reduce metabolic costs by an average of more than 10%.

1. Introduction

Soft exoskeletons have been widely studied for their advantages of being lightweight, comfortable, and easy to wear and take off [1,2]. During walking, they obtain motion information through wearable sensors and apply assistance based on gait styles and phases estimated from the motion information. If the exoskeleton assistance does not match the intention of human movement, the exoskeleton will hinder human movement and even cause instability, leading to falls. Therefore, gait recognition is crucial for the control of a soft exoskeleton.
In terms of sensor system design, commonly used sensors include inertial sensors (IMUs) [3,4,5], plantar pressure sensors [6], surface electromyography electrodes [7,8], and electroencephalogram electrodes [9]. IMUs are widely used in gait motion recognition for their high portability, low cost, and environmental friendliness [10,11,12,13,14]. At present, many studies on gait recognition using different sensor configurations have been proposed. These studies are based on threshold methods for identifying gait events, which are predetermined based on experience. Because gait features are unique, each person’s threshold may vary. The threshold adjustment process is manual and time-consuming. With the development of statistical methods, various neural networks have been applied to gait recognition in an attempt to solve this problem.
There are two main types of exoskeleton recognition for human motion intention: one is the recognition of motion status under different road conditions, and the other is the recognition of gait phase. The motion state and gait phase are important bases for the time and magnitude of the auxiliary force applied by the exoskeleton [15]. By adjusting the timing and magnitude of the assistance force, the exoskeleton can match the gait characteristics of the wearer well, thereby improving adaptability and efficiency. However, the adjustment is usually based on experience, resulting in poor adaptability of the exoskeleton and poor auxiliary effects [16,17,18]. To solve this problem, human-in-the-loop (HIL) optimization [19,20], which is the method of optimizing control parameters for individualized assistance within soft exosuits, has attracted increasing attention. HIL optimization is an effective method for identifying optimal control parameters based on feedback from human gait information, and is particularly suitable for the personalized customization of auxiliary parameters for users [21,22,23]. The effectiveness of the HIL optimization strategy depends on the accuracy of human gait information recognition. At present, research mainly focuses on the accuracy of gait recognition and assistance prediction itself, with relatively little application in the control of lower-limb exoskeleton assistance force to analyze the application effect of recognition.
In order to predict the auxiliary time of an exoskeleton, we build a gait recognition model that can be applied to exoskeletons. The contributions of this paper are as follows: Firstly, a hip-assisted exoskeleton gait recognition method based on five inertial sensors is proposed, including SVM motion state recognition and LSTM phase recognition. Secondly, considering the coupling relationship between the two legs, the model proposes a rule-based gait event labeling method that can recognize both the left and right leg assist times. Finally, the gait recognition model is applied for the prediction of exoskeleton assisting time, and the effectiveness and reliability of the gait recognition model are proved by analyzing the assisting effect.

2. Related Works

In terms of motion state recognition, different recognition models were studied according to different applications, and certain recognition effects were achieved [24,25]. For example, Semwal et al. used a CNN to identify six different motion states [26], Weiland et al. used a classification artificial neural network (ANN) to recognize horizontal walking, uphill climbing, and uphill movement states [27], and Liu et al. placed four IMUs on the front of the thighs and the front of the calves of both legs and used an SVM to identify the current motion state [28]. The identifications in these studies are mostly typical of movement states, such as standing, horizontal walking, going up and down slopes, and going up and down stairs, while basic human movements include turning left and right, and so on. In addition, different motion states require different assistance planning techniques, and most of these motion recognition studies focus on recognition, and have not been well applied to the control of exoskeletons.
On the other hand, the phase recognition of the human gait is also a hot topic [29]. Zhang et al. proposed a multi-degree-of-freedom gait-detection method based on the kinematic information collected by IMUs on the dorsum of the foot, ankle, and thigh, and pursued the real-time recognition of the five phase features in a gait cycle [30]. Luo et al. identified the swinging and standing phases using a quadratic discriminant analysis classifier [31] based on the information of IMUs located on the thigh and shank. Pazar et al. proposed an LSTM model based on the information of IMUs and knee strains to recognize gait phases [32]. The above studies are focused on gait recognition itself; less attention has been paid to its application in the assistance force control of lower-limb exoskeletons.
There is an increasing amount of research on HIL. Koller et al. used a one-dimensional gradient descent method to determine the optimal assist parameters for ankle-assisted exoskeletons. At a walking speed of 1.2 m per second, the algorithm process for each subject took 50 min [33]. Kim et al. used the Covariance Matrix Adaptation Evolution Strategy (CMA-ES) to determine three optimal parameters: peak time, offset time, and force magnitude. Eight participants walked on the treadmill at a speed of 1.25 m/s, and the iterative optimization process took an average of two hours per participant. The research results indicate that metabolism significantly decreases under moderate intensity and optimal time [34]. Ye et al. proposed a Bayesian optimization algorithm that uses minimum metabolic cost as an indicator to conduct parameter optimization, minimizing the metabolic cost for humans walking on a 1.25 m/s treadmill [23]. The disadvantage of this method is that the devices used for optimizing parameters are not convenient to carry and travel with, so it is important to propose an assistance time recognition method based on a wearable sensor. In the above studies, the required time is relatively long and cannot meet the requirements of real-time online adjustment during walking.

3. Materials and Methods

This section will provide a detailed introduction to the proposed approach, including platform design, data acquisition and preprocessing, feature extraction, the gait recognition model, the experiment, and other modules.

3.1. The Soft Lower-Limb Exoskeleton Platform

We proposed a belt-type soft hip-assist exoskeleton, as shown in Figure 1. The exoskeleton consists of two actuation units, a control unit, a power supply unit, a waist brace, a knee brace, a pair of shoulder straps, and two belts for transmitting the assistant forces to the legs of the wearer. A load cell is fixed to each of the belts for measuring the assistant forces. The actuation units are tied to the front of the waist brace (each unit has a motor). The assist forces are transmitted by the belts from the actuation units to the knee joints, generating driving torques to the hip joint. Five IMUs are used to measure the motions of the lower limbs and the trunk of the wearer. The IMUs are 9-axis ones consisting of accelerometers, magnetometers, and gyroscopes. Only gyroscopes and accelerometers are used in this work, as magnetic sensors are susceptible to environmental influences. The placement diagram for the IMUs is shown in Figure 1. The IMUs are connected to the main control board inside the back shell of the exoskeleton by a long wire (transmission rate: 500 KHZ, sampling frequency: 200 Hz). All electronic devices of the exoskeleton are powered by a 0.55 kg power supply unit equipped with lithium-ion batteries (36 V, 3 Ah). The weight of the exoskeleton is approximately 2.7 kg.
The soft hip-assist exoskeleton we proposed provides hip flexion assistance force with two belts by imitating the function of the rectus femoris muscle, which is the main hip flexor in human walking. The force of the rectus femoris muscle has the form of Equation (1) [34]. Thus, it is reasonable for us to suppose that the profile of the assistive force takes a similar form. So, the assistance force is expressed as
F = A sin π T a t + α sin π T a t + f ,
where F and A represent the assistance force and its magnitude; t represents time; α , the phase shift factor; T a , the swing period; and f , the force offset to prevent the belts getting loose along the thigh of the swing leg without influencing hip extension.
Imitating the muscle force of the rectus femoris, we set the start time of the assistance force as the gait event of toe off, and the end time as the gait event of heel strike in a gait cycle. The time corresponding to the peak of the assistance is set as the gait event of hip max.

3.2. Data Collection and Preprocessing

We collected gait data from 25 healthy experimenters during standing (SD), level walking (LW), going up stairs (US), going down stairs (DS), going up slopes (USL), going down slopes (DSL), turning left (TL), turning right (TR), left steering (LS), and right steering (RS). Walking characteristics such as pace, stride length, and height were carried out in full accordance with the wishes of the experimenter for the whole walking process. After the exoskeleton powered on, the experimenter stood up first, and after one minute, began to complete 9 motion states: level walking, going up stairs, going down stairs, going up slopes, going down slopes, turning left, turning right, left steering, and right steering. Level walking is walking for approximately 100 m under the VICON motion capture system (Oxford Metrology Limited, Oxford, UK). Going up and down stairs involves walking the height of two floors in each direction. Walking up and down slopes occurs for about 50 m. The left and right turns were tested for 2 min, respectively, and 50 left and right turns were completed, respectively. Each of the 25 participants performed a complete experiment, and the entire experiment was performed 25 times. The details of the dataset are shown in Table 1. A total of 4,414,627 sample points were collected, including 299,922 SD samples, 497,373 LW samples, 595,492 US samples, 594,342 DS samples, 287,584 USL samples, 288,389 DSL samples, 584,770 TL samples, 583,571 TR samples, 343,127 LS samples, and 340,057 RS samples.
The human walk exhibits a periodic pattern known as a gait cycle [35]. Each gait cycle has a series of gait events that occur chronologically at specific time points, which are important for determining the starting point of applying assistance force. According to the exoskeleton assistance strategy, the gait times that need to be recognized during a gait cycle are left heel strike (L-HS), right toe off (R-TO), right hip max (R-HMax), right heel strike (R-HS), left toe off (L-TO), and left hip max (L-HMax) (Figure 2).
We selected a dataset of horizontal walking in various road conditions as the dataset for gait phase recognition. The verification of motion status is based on the division of each gait cycle by the VICON system. We used IMU information to identify gait events, but also used the VICON optical motion capture system to measure the motion of the human body during walking for calibration and verification. The IMU inertial system and the VICON system have their own hardware and software. They are synchronized based on velocity information. In the sagittal plane of the VICON system, the lowest vertical position of the heel is the HS event, the highest elevation of the toe from the lowest is the TO event, and the highest hip joint angle is the HMax event. The details of the dataset are shown in Table 2. A total of 497,373 sample points were collected, including 57,214 L-HS samples, 117,840 R-TO samples, 71,250 R-HMax samples, 56,790 R-HS samples, 119,590 L-TO samples, and 74,689 L-HMax samples.

3.3. Feature Extraction

In terms of input feature selection, features with different gait types are selected because, for example, the hip angle curve when descending stairs is different from other gaits, and the angle curve of the ankle joint differs significantly from other gaits when going uphill and downhill. The standard deviation of the combined acceleration of the center of mass is almost zero when people are at rest, but it changes greatly when they are moving. The center of mass acceleration standard deviation reflects the discrete path of the acceleration data.
The information from IMUs in different positions was used for the identification and comparison of the input feature quantities: 3 IMUs were based on the waist and foot, 3 IMUs were based on the waist and thigh, and 5 IMUs were based on the waist, foot, and thigh. After the experiment, it was found that the three IMUs of the waist and foot could not classify going up and going down. Based on the 3 IMUs on the waist and thigh, the recognition rate of level walking and turning was low (about 45%). Therefore, a total of 30 feature vectors, including the angular velocity and acceleration information of 5 IMUs of the waist, thigh, and foot, were used as input feature vectors for the gait recognition model.
For the feature extraction method of gait phase recognition, we consider the coupled relationship between legs, and propose a rule-based feature labeling method. HS: As shown in Figure 3A,D, when event HS occurs, the ankle reaches the maximum dorsiflexion angle. It can be detected by the zero-crossing detection rule. TO: As shown in Figure 3A,C,E, when event TO occurs, the ankle reaches the maximum rate of plantar flexion. It can be detected by a peak detection rule. HMax: As shown in Figure 3B,C,F, when event HMax occurs, the hip reaches the maximum flexion angle. It can be detected by the zero-crossing detection rule.

3.4. Gait Recognition Model

We use an SVM to recognize motion states from the motion information measured by the IMUs, as it requires fewer samples and exhibits strong generalization ability. A radial basis function (RBF) kernel is used to transform the low dimensional data into a high-dimensional space [36], and a gait classification hyperplane is constructed in the high-dimensional space. This hyperplane can separate non-linear and non-separable data into linearly separable data. The classification hyperplane is defined as
f ( x ) = ω T φ ( x ) + b ,
in which ω is the hyperplane normal vector, φ ( x ) is the input feature vector, and b is the intercept.
Then, gait classification can be transformed into a problem of finding the maximum distance between walking styles, and this distance is expressed as
max dis = max w   2 × y ( ω T φ ( x ) + b ) w 2 ,
where y i ( ω T φ ( x ) + b ) = 1 , y i 1 , 1 . Thus, the problem of solving the optimal gait classification surface via SVM can be reduced to a quadratic programming problem, as shown by Equation (4). Introduce slack variables ξ to transform the inequality constraints in Equation (4) to an equality constraint, as shown in Equation (5). We can build a Lagrange function, as shown in Equation (6).
min d i s = m i n w w 2 2 w = w 1 2 + w 2 2 + + w n 2 y ( ω T φ ( x ) + b ) 1 , i = 1,2 , , n ,
h i ( w , ξ i ) = 1 y i ( w T φ ( x ) + b ) + ξ i
L ( w , b , ξ , λ , μ ) = 1 2 w 2 + C i = 1 m ξ i + i = 1 n ( λ i h i μ i ξ i )
where λ and μ are Lagrange multipliers, w and b are optimization parameters, and C represents the penalty factor. Taking the derivatives of the optimization parameters in Equation (6), we obtain the gait classification hyperplane as
f ( x ) = s i g n ( i = 1 N w i ξ i K ( x i , x ) + b i )
By training the model using a multi-classification method with binary tree structures, we can recognize the walking styles. Figure 4 gives the framework of the recognition. The data collected by the IMUs are the accelerations and angular velocities of the waist and right and left legs, represented by a f r , v f r , a f l , v f l , a t r , v t r , a t l , v t l , a w , v w . The penalty factor C is 2. Choose Gaussian (RBF) as the kernel function based on the characteristics of the data, and gamma is set to scale. It automatically calculates the value of gamma based on the scale of the input data.
We take the kinematic data of walking motion as a periodic time series and employ long short-term memory (LSTM) to identify gait phase. The cell state of the LSTM allows for the retention or addition of input information to the cell state, enabling a better integration of historical information for the recognition of gait phase at the current time point [9,37,38]. The LSTM cell consists of oblivion gate f t , input gate i t , output gate o t , and cell state C t .
f t = σ ( W f h t 1 , x t + b f )
i t = σ ( W i h t 1 , x t + b i )
o t = σ ( W o h t 1 , x t + b o )
C t = f t C t 1 + i t tanh ( W c h t 1 , x t + b c )
h t = o t tanh ( C t )
where σ is the sigmoid function, W represents the weight matrices of each gate, and b , the bias vectors of each gate.
The network model used for gait phase recognition is shown in Figure 2. The input to the network is the time series X = x 1 , x 2 , , x n 1 , x n , where x n is the vector consisting of the standardized IMU inertial raw data and n is the sequence length, which is 5. The epochs of the model are 200, and the number of samples in each training batch is 64. The gait phase recognition model consists of an LSTM layer and a fully connected (FC) layer followed by softmax. The input feature dimension is 30, the LSTM layer has 128 neurons, and the FC layer has 32 neurons. The output represents the probabilities of 6 different gait phases. This model runs on a notebook PC. The notebook PC receives the gait data sent by the robot, uses a trained LSTM model to obtain gait classification results, and finally sends the results to the exoskeleton through Wi-Fi. The artificial neural network training is implemented using Python 3.7.1 and TensorFlow 2.5.0.

4. Experiment

4.1. Subjects

The dataset for the gait recognition experiments was obtained from 25 healthy adults (age = 27.5 ± 2.4 years, height = 173.6 ± 5.9 cm, and weight = 69.5 ± 3.6 kg). The experiment was approved by the Ethics Committee of Hebei University of Technology, and each subject was asked to read and provide written informed consent before the test (Ethics Committee name: the Biomedical Ethics Committee of Hebei University of Technology; approval code: HEBUThMEC2023017).

4.2. Protocol

4.2.1. Gait Classification and Gait Phase Recognition

For the recognition verification of multiple road conditions, the data of the first 20 people in the dataset in Table 1 were used to train the SVM model, and the data of the last 5 people were used to test the performance of the model. For gait phase recognition validation, the data of the first 22 individuals in the dataset in Table 2 were used to train the LSTM model, while the data of the last 3 individuals were used to test the performance of the model. In this paper, we use accuracy (ACC), F 1 scores, and the Matthews correlation coefficient (MCC) to evaluate the performance of the model. Different recognition models were trained offline using Python 3.7.1 and TensorFlow 2.5.0 software. The performance of the motion state and gait phase recognition algorithm was validated using a five-fold cross validation.

4.2.2. Assistance Effect Experiment

Ten subjects were randomly selected from the twenty-five subjects to test the assistance effect of the assist policy based on the proposed gait classification and gait phase recognition. The exoskeleton’s assist time predicts the time difference in the gait event from the previous step of gait recognition. In order to verify the effectiveness of the gait recognition model, a metabolic power experiment for the exoskeleton-assisted effect experiment based on BP neural network gait phase recognition was added.
Each subject wore specially designed sportswear under the soft exoskeleton so that reflective markers could be pasted to them for the VICON system (Oxford Metrology Limited, Oxford, UK) to capture walking motions. Figure 5A presents the locations of the 16 markers in total that were pasted symmetrically to the left and right of the human body [39,40]. Each subject participated in two different speed walking experiments on a treadmill. Before the test, each subject was instructed to familiarize themselves with treadmill walking. To confirm the effect of the assist, power-off (P-OFF) tests were also carried out before the test with the assist (P-ON) for each walking speed. Figure 6 gives the test conditions.
Besides measuring the kinematic parameters in the assisted walking test, change in metabolic cost was also estimated using a portable respiratory measurement device (Cosmed, K4b2, Rome, Italy) worn by the subjects (Figure 5B). All the subjects were asked not to consume food or beverages 2 h before the test so that the metabolic cost could be estimated accurately.

4.3. Data Collection

Data were collected at 100 Hz using the VICON Motion system (Oxford Metrics, Oxford, UK) with 10 cameras (model: VANTAGE-V5-VS-5299). Real marker trajectory data were filtered with a quintic spline filter based on the code by Herman Woltring [41]. The position information of each subject’s joints was directly obtained. The angles of the hip, knee, and ankle joints were obtained using the plug-in gait model in Vicon Nexus software v1.8.5 [42,43,44]. All the joint angles were calculated from the YXZ Cardan angles, derived by comparing the relative orientations of the two segments [42]. Using Vicon Polygon (Oxford Metrics Group, Oxford, UK) [42] software, kinematic data were extracted to a C3D file or an ASCII file, which was then input into MATLAB software (MATLAB 23.2.0) for post-processing.
The respiratory data, oxygen consumption ( V O 2 [L/min]) data, and respiratory quotient (RER) were averaged across the latter half of the one-minute intervals for each sampling parameter condition. The net metabolic cost was calculated by subtracting the resting metabolic cost from the walking metabolic cost, and then normalized according to the weight of each subject.

4.4. Statistical Analysis

Means and standard deviations for each test condition were calculated. Statistical analysis was performed using the SPSS statistical software system (SPSS Inc., Chicago, IL, USA; version 22.0). One-way repeated-measures analyses of variance with three conditions (No-Exo, P-Off, P-On) were used to verify the effect of the hip assistance on metabolic reduction. Pairwise comparisons with Bonferroni post hoc tests were conducted to identify the differences between conditions when a statistically significant main effect was identified with the one-way repeated-measures analyses of variance. A paired t-test was performed to assess the difference between the different conditions for metabolic reduction. p < 0.05 represented a significant difference.

5. Results

5.1. Accuracy of Motion State Recognition

We compared the accuracy of four motion recognition models—KNN, Random Forest, Decision Tree, and SVM—as shown in Table 3. It can be seen that SVM has the highest recognition rate of 99.98%. Table 4 shows the confusion matrix of the motion state recognition, which shows the recognition results of the ten walking styles. It can be seen from Table 4 that a small portion of standing periods were identified as steering. The reason is that there were a few similar features between the two walking styles. The recognition rate of level walking was high. This is worth noting because we are more interested in assisting in level walking for frail, older individuals, and thus in removing assistance for other walking styles for safety.

5.2. Accuracy of Gait Phase Recognition

Table 5 compares the recognition rates of the different gait phases in a gait cycle. Table 3 shows the performance indicators of the different phase recognition models. It can be seen that, compared with NN, RBF and BP, LSTM has the highest recognition rate, which is 98.26%. The confusion matrix displays the recognition results of six gait phases, where the sample between two consecutive gait events is marked as the starting event. From Table 6 and Figure 7, it can be seen that the recognition rate of HS gait events (L-HS: 0.975, R-HS: 0.970) is lower than those of TO (L-TO: 0.991, R-TO: 0.995) and HMax (L-Hax: 0.983, R-Hax: 0.991) events. The reason is that the feature used to mark HS gait events was the zero-crossing detection rule of the Z-axis angular velocity of the foot IMU. Before the HS event, the Z-axis angular velocity of the foot IMU rapidly decreased in a short time. There was a certain error in finding the zero point during this rapid descent process. Although the recognition rate of HS was relatively low, the recognition accuracy was sufficient to apply to the assistance of the exoskeleton.

5.3. Kinematic Effect of the Assistance

The accuracy of gait recognition and the prediction of assistance parameters directly affects the effectiveness of wearing exoskeleton assistance, and the direct representation of this on the human body is the changes in the joints of the lower limbs during walking. The highest hip joint swing angle value during exoskeleton-assisted walking was significantly higher than that without exoskeleton walking. At speeds of 3.6 km/h and 5.5 km/h, the maximum swing angle of the hip joint increased by 5–6 degrees (Figure 8A,B). This indicates that setting the peak assist at the moment of maximum hip joint swing angle is effective and reasonable. Exoskeleton assistance generates torque on the hip joint. Although there is no torque generated on the knee joint, the force of the exoskeleton acts on the knee joint, and the angle of the knee joint also changes significantly. As shown in Figure 8E,F, at speeds of 3.6 km/h and 5.5 km/h, wearing an exoskeleton for assistance increases the maximum knee joint swing angle by 9 degrees and 7 degrees, respectively, compared to not wearing an exoskeleton.
As shown in Figure 8G–I, in toe off (point F, Figure 8I), there was a significant change in ankle plantar flexion, with an increase of 3–5 degrees. At the end of the assistance, there was a significant change in ankle dorsiflexion and an increase of 3 degrees in the angle (point E, Figure 8H). The changes in ankle joint plantar flexion and dorsiflexion corresponded to toe rocker and heel rocker, with an increased angle indicating better rolling effect and a better performance of human ankle rolling. This indicates that starting assistance at TO and ending assistance at HS is effective and reasonable. The changes in ankle joint plantar flexion and dorsiflexion can affect the ground clearance position of the heel and toe, and can be used to further analyze the changes in foot ground clearance distance. The position of the heel and toe (Figure 9A,C,D) and the position of the toe (Figure 9A,B) when the foot swings to the neutral position are both elevated under the assistance of exoskeletons. The elevation of the heel and toe positions reduces the risk of tripping over obstacles, which is one of the main purposes of hip joint assistance.
The combined result of changes in the swing angles of the human hip joint, knee, and lower limbs is an increase in step size, as shown in Figure 10. At two different walking speeds, there was an increase of 5–10 cm under assisted conditions. At the walking speed of 5.5 km/h, the step length increases even more, by about 10 cm. When walking speed is high, the more obvious the assist effect.

5.4. Metabolic Effect of Assistance

Figure 11 shows the average metabolic cost of ten subjects walking on a treadmill at a walking speed of 3.6 km/h and 5.5 km/h. Compared to not wearing an exoskeleton, wearing an exoskeleton at two different walking speeds reduced the metabolic power of the ten individuals by 10.98% (0.405 w/kg) and 14.28% (0.621 w/kg), respectively. The higher the speed, the more the metabolic power decreased, indicating that the assistance effect of exoskeletons is significant. As can be seen from Figure 11, the metabolic power of gait phase recognition based on a BP neural network for assist time prediction decreased by 6.3% (0.213 w/kg) and 8%,2% (0.357 w/kg), respectively. The metabolic power reduction in the SVM and LSTM recognition models proposed in this paper was significantly more than in the BP neural network, indicating that the proposed gait recognition model can be applied to exoskeleton-assisted control. This demonstrates the effect of the gait event based on assistance policies. When the respiratory exchange rate (RER) is less than 1, the subject is in an aerobic breathing state. When RER 1, the subject is in a respiratory state in which aerobic and anaerobic respiration work together. An increase in RER to some extent reflects an increase in exercise intensity and changes in the respiratory substrates of the body’s energy supply system. The RER of the ten experimenters was less than 1 under all three conditions, and the RER was the lowest under the assistance condition.

6. Discussion

With regard to kinematic analysis, there are still some issues worth discussing. We believe the slight improvement in point B (Figure 8A,C) is the result of alternating leg assistance. When the left leg is at the hip joint angle feature, point B and has not yet entered the assist phase, and the right leg has just completed the swing phase. Before the left leg enters the swing phase, potential energy is converted into kinetic energy to drive the movement of the left leg. Therefore, although assistance has not yet begun at point B, due to the conversion between kinetic energy and potential energy, the hip joint angle will also increase.
It should also be noted that, in the test, with the assistance of the exoskeleton, the step length significantly increased. Because the test was conducted at a constant speed on a treadmill, as the step length increased, the cadence inevitably decreased. Another possibility is that the human body autonomously changes its walking style, increasing step length and reducing step frequency.
When walking at asynchronous speeds, the balance system of the human body adjusts the relationship between step length and cadence to ensure an efficient walk, i.e., from the viewpoint of an efficient walk, an optimum relationship exists between step length and cadence. This is similar to the walking ratio (the ratio of step length to cadence) proposed by Murakami [45,46]. That is why the concept of walk ratio is usually used when walk efficiency is discussed. The metabolic consumption test demonstrated that the assistance did bring metabolic reduction. From the metabolic consumption test, it can be seen that the wearer’s metabolic cost did indeed decrease under assist conditions, ruling out the possibility that the increase in step length may not be the assisting effect.
At present, we have identified and perceived the gait phase of horizontal walking based on an assistance strategy for horizontal walking. However, there are significant differences in the walking characteristics of the human body under different road conditions. For example, in uphill and downhill movements, the main movement of the lower-limb hip joints is flexion, and the flexion angle is relatively small compared to horizontal walking as a whole. During the process of going up stairs, the flexion angle of the hip joint varies greatly. The assistance strategy of exoskeletons varies under different road conditions, and the effective implementation of assistance strategies requires an accurate perception of human motion intentions. Therefore, our plan for our next work is to study the perception and assistance strategies of gait phase under different road conditions.

7. Conclusions

Firstly, this paper proposed a gait recognition method using the information measured by five IMUs fixed on a soft exoskeleton. It distinguishes walk styles with the algorithm of a support vector machine (SVM) and gait phases with long and short-term memory (LSTM). Tests on 12 subjects demonstrated high accuracies of 99.98% and 98.06%, respectively. Secondly, a method for determining assistance parameters for the soft exoskeleton based on gait event and gait phase was proposed. Tests were conducted on three subjects to confirm its effect. The results showed that the assistance could significantly improve gait motion and reduce metabolic cost by an average of over 10%.

Author Contributions

Q.X.: methodology, software, investigation, formal analysis, validation, data curation, writing—original draft preparation, and writing—review and editing. J.W.: methodology, writing—review and editing, project administration, and resources. Y.L.: writing—review and editing, investigation, and data curation. S.G.: conceptualization, methodology, writing—review and editing, project administration, and funding acquisition. L.L.: data curation. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by S&T Program of Hebei No. 22372001D and Hebei Natural Science Foundation No. F2023202047, in part by National Natural Science Foundation of China No. 62303155.

Institutional Review Board Statement

This study was conducted in accordance with the Declaration of Helsinki and approved by the Biomedical Ethics Committee of Hebei University of Technology in Tianjin, China (Ethics committee name: the Biomedical Ethics Committee of Hebei University of Technology; approval code: HEBUThMEC2023018).

Informed Consent Statement

Informed consent was obtained from all subjects involved in this study.

Data Availability Statement

The original data are available following reasonable request.

Acknowledgments

The authors are grateful to Yao Li for his technical assistance.

Conflicts of Interest

The authors declare no financial or other relationships that may cause, or are perceived to cause, conflicts of interest in relation to this study.

References

  1. Xing, L.Y.; Wang, M.L.; Zhang, J.; Chen, X.Y.; Ye, X.D. A Survey on Flexible Exoskeleton Robot. In Proceedings of the 4th IEEE Information Technology, Networking, Electronic and Automation Control Conference (ITNEC), Electr Network, Chongqing, China, 12–14 June 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 170–174. [Google Scholar]
  2. Asbeck, A.T.; Schmidt, K.; Walsh, C.J. Soft exosuit for hip assistance. Robot. Auton. Syst. 2015, 73, 102–110. [Google Scholar] [CrossRef]
  3. McCabe, M.V.; Van Citters, D.W.; Chapman, R.M. Hip Joint Angles and Moments during Stair Ascent Using Neural Networks and Wearable Sensors. Bioengineering 2023, 10, 784. [Google Scholar] [CrossRef]
  4. Dasgupta, A.; Sharma, R.; Mishra, C.; Nagaraja, V.H. Machine Learning for Optical Motion Capture-Driven Musculoskeletal Modelling from Inertial Motion Capture Data. Bioengineering 2023, 10, 510. [Google Scholar] [CrossRef]
  5. Zhang, M.Y.; Wang, Q.S.; Liu, D.; Zhao, B.Q.; Tang, J.Z.; Sun, J.W. Real-Time Gait Phase Recognition Based on Time Domain Features of Multi-MEMS Inertial Sensors. IEEE Trans. Instrum. Meas. 2021, 70, 12. [Google Scholar] [CrossRef]
  6. Huang, C.H.; Fukushi, K.; Wang, Z.W.; Kajitani, H.; Nihey, F.; Pokka, H.; Narasaki, H.; Nakano, H.; Nakahara, K. Foot-Healthcare Application Using Inertial Sensor: Estimating First Metatarsophalangeal Angle from Foot Motion during Walking. IEEE Sens. J. 2022, 22, 2835–2844. [Google Scholar] [CrossRef]
  7. Lee, S.W.; Yi, T.; Jung, J.W.; Bien, Z. Design of a Gait Phase Recognition System That Can Cope with EMG Electrode Location Variation. IEEE Trans. Autom. Sci. Eng. 2017, 14, 1429–1439. [Google Scholar] [CrossRef]
  8. Rani, G.J.; Hashmi, M.F.; Gupta, A. Surface electromyography and artificial intelligence for human activity recognition-A systematic review on methods, emerging trends applications, challenges, and future implementation. IEEE Access 2023, 8, 105140–105169. [Google Scholar] [CrossRef]
  9. Tortora, S.; Ghidoni, S.; Chisari, C.; Micera, S.; Artoni, F. Deep learning-based BCI for gait decoding from EEG with LSTM recurrent neural network. J. Neural Eng. 2020, 17, 19. [Google Scholar] [CrossRef] [PubMed]
  10. Gravina, R.; Alinia, P.; Ghasemzadeh, H.; Fortino, G. Multi-sensor fusion in body sensor networks: State-of-the-art and research challenges. Inf. Fusion 2017, 35, 68–80. [Google Scholar] [CrossRef]
  11. Fortino, G.; Galzarano, S.; Gravina, R.; Li, W.F. A framework for collaborative computing and multi-sensor data fusion in body sensor networks. Inf. Fusion 2015, 22, 50–70. [Google Scholar] [CrossRef]
  12. Ghasemzadeh, H.; Panuccio, P.; Trovato, S.; Fortino, G.; Jafari, R. Power-Aware Activity Monitoring Using Distributed Wearable Sensors. IEEE Trans. Hum.-Mach. Syst. 2014, 44, 537–544. [Google Scholar] [CrossRef]
  13. Fortino, G.; Di Fatta, G.; Pathan, M.; Vasilakos, A.V. Cloud-assisted body area networks: State-of-the-art and future challenges. Wirel. Netw. 2014, 20, 1925–1938. [Google Scholar] [CrossRef]
  14. Fortino, G.; Giannantonio, R.; Gravina, R.; Kuryloski, P.; Jafari, R. Enabling effective programming and flexible management of efficient body sensor network applications. IEEE Trans. Hum.-Mach. Syst. 2012, 43, 115–133. [Google Scholar] [CrossRef]
  15. Tucker, M.; Novoseller, E.; Kann, C.; Sui, Y.; Yue, Y.; Burdick, J.W.; Ames, A.D. Preference-based learning for exoskeleton gait optimization. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; IEEE: Piscataway, NJ, USA, 2000; pp. 2351–2357. [Google Scholar]
  16. Tu, Y.; Zhu, A.; Song, J.; Zhang, X.; Cao, G. Design and experimental evaluation of a lower-limb exoskeleton for assisting workers with motorized tuning of squat heights. IEEE Trans. Neural Syst. Rehabil. Eng. 2022, 30, 184–193. [Google Scholar] [CrossRef] [PubMed]
  17. Stroppa, F. Design optimizer for planar soft-growing robot manipulators. Eng. Appl. Artif. Intell. 2024, 130, 107693. [Google Scholar] [CrossRef]
  18. Han, Y.; Liu, C.; Xiu, H.; Li, Z.; Shan, S.; Wang, X.; Ren, L.; Ren, L. Trajectory Control of an Active and Passive Hybrid Hydraulic Ankle Prosthesis Using an Improved PSO-PID Controller. J. Intell. Robot. Syst. 2022, 105, 48. [Google Scholar] [CrossRef]
  19. Zhang, J.; Fiers, P.; Witte, K.A.; Jackson, R.W.; Poggensee, K.L.; Atkeson, C.G.; Collins, S.H. Human-in-the-loop optimization of exoskeleton assistance during walking. Science 2017, 356, 1280–1284. [Google Scholar] [CrossRef]
  20. Kim, M.; Ding, Y.; Malcolm, P.; Speeckaert, J.; Siviy, C.J.; Walsh, C.J.; Kuindersma, S. Human-in-the-loop Bayesian optimization of wearable device parameters. PLoS ONE 2017, 12, e0184054. [Google Scholar] [CrossRef]
  21. Witte, K.; Fiers, P.; Sheets-Singer, A.; Collins, S. Improving the energy economy of human running with powered and unpowered ankle exoskeleton assistance. Sci. Robot. 2020, 5, eaay9108. [Google Scholar] [CrossRef]
  22. Grimmer, M.; Zeiss, J.; Weigand, F.; Zhao, G. Exploring surface electromyography (EMG) as a feedback variable for the human-in-the-loop optimization of lower limb wearable robotics. Front. Neurorobot. 2022, 16, 948093. [Google Scholar] [CrossRef]
  23. Ding, Y.; Kim, M.; Kuindersma, S.; Walsh, C.J. Human-in-the-loop optimization of hip assistance with a soft exosuit during walking. Sci. Robot. 2018, 3, eaar5438. [Google Scholar] [CrossRef]
  24. Kim, S.-K.; Hong, S.; Kim, D. A walking motion imitation framework of a humanoid robot by human walking recognition from IMU motion data. In Proceedings of the 2009 9th IEEE-RAS International Conference on Humanoid Robots, Paris, France, 7–10 December 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 343–348. [Google Scholar]
  25. Wang, Y.S.; Cheng, X.; Jabban, L.; Sui, X.H.; Zhang, D.G. Motion Intention Prediction and Joint Trajectories Generation toward Lower Limb Prostheses Using EMG and IMU Signals. IEEE Sens. J. 2022, 22, 10719–10729. [Google Scholar] [CrossRef]
  26. Semwal, V.B.; Gaud, N.; Lalwani, P.; Bijalwan, V.; Alok, A.K. Pattern identification of different human joints for different human walking styles using inertial measurement unit (IMU) sensor. Artif. Intell. Rev. 2022, 55, 1149–1169. [Google Scholar] [CrossRef]
  27. Weigand, F.; Höhl, A.; Zeiss, J.; Konigorski, U.; Grimmer, M. Continuous locomotion mode recognition and gait phase estimation based on a shank-mounted IMU with artificial neural networks. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Kyoto, Japan, 23–27 October 2022; IEEE: Kyoto, Japan, 2022; pp. 12744–12751. [Google Scholar]
  28. Liu, X.H.; Zhou, Z.H.; Wang, Q.N. Real-Time Onboard Recognition of Gait Transitions for A Bionic Knee Exoskeleton in Transparent Mode. In Proceedings of the 40th Annual International Conference of the IEEE-Engineering-in-Medicine-and-Biology-Society (EMBC), Honolulu, HI, USA, 18–21 July 2018; IEEE: Honolulu, HI, USA, 2018; pp. 3202–3205. [Google Scholar]
  29. Attal, F.; Amirat, Y.; Chibani, A.; Mohammed, S. Automatic Recognition of Gait Phases Using a Multiple-Regression Hidden Markov Model. IEEE-ASME Trans. Mechatron. 2018, 23, 1597–1607. [Google Scholar] [CrossRef]
  30. Zhang, X.X.; Zhang, H.J.; Hu, J.H.; Zheng, J.; Wang, X.B.; Deng, J.R.; Wan, Z.H.; Wang, H.T.; Wang, Y.K. Gait Pattern Identification and Phase Estimation in Continuous Multilocomotion Mode Based on Inertial Measurement Units. IEEE Sens. J. 2022, 22, 16952–16962. [Google Scholar] [CrossRef]
  31. Lou, Y.; Wang, R.L.; Mai, J.G.; Wang, N.H.; Wang, Q.N. IMU-Based Gait Phase Recognition for Stroke Survivors: Preliminary Results. In Proceedings of the 8th IEEE Annual International Conference on Cyber Technology in Automation, Control, and Intelligent Systems (IEEE-CYBER), Tianjin, China, 19–23 July 2018; IEEE: Tianjin, China, 2018; pp. 802–806. [Google Scholar]
  32. Pazar, A.; Khalilbayli, F.; Ozlem, K.; Yilmaz, A.F.; Atalay, A.T.; Atalay, O.; İnce, G. Gait Phase Recognition using Textile-based Sensor. In Proceedings of the 2022 7th International Conference on Computer Science and Engineering (UBMK), Diyarbakir, Turkey, 14–16 September 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–6. [Google Scholar]
  33. Koller, J.R.; Gates, D.H.; Ferris, D.P.; Remy, C.D. ‘Body-in-the-Loop’ Optimization of Assistive Robotic Devices: A Validation Study. In Proceedings of the Robotics: Science and Systems, Ann Arbor, MI, USA, 18–22 June 2016; pp. 1–10. [Google Scholar]
  34. Kim, J.; Quinlivan, B.T.; Deprey, L.-A.; Arumukhom Revi, D.; Eckert-Erdheim, A.; Murphy, P.; Orzel, D.; Walsh, C.J. Reducing the energy cost of walking with low assistance levels through optimized hip flexion assistance from a soft exosuit. Sci. Rep. 2022, 12, 11004. [Google Scholar] [CrossRef] [PubMed]
  35. Perry, J.; Burnfield, J.M. Gait Analysis: Normal and Pathological Function, 2nd ed.; SLACK Incorporated: Thorofare, NJ, USA, 2010. [Google Scholar]
  36. Chandra, M.A.; Bedi, S. Survey on SVM and their application in image classification. Int. J. Inf. Technol. 2021, 13, 1–11. [Google Scholar] [CrossRef]
  37. Lu, L.J.; Liu, S.Y.; Zhou, Z.Y.; Sun, J.T.; Melendez-Calderon, A.; Guo, Z. Continuous Gait Phase Recognition and Prediction Using LSTM Network across Walking Speeds and Inclinations. In Proceedings of the 7th IEEE International Conference on Advanced Robotics and Mechatronics, Guilin, China, 9–11 July 2022; IEEE: Guilin, China, 2022; pp. 101–106. [Google Scholar]
  38. Kolaghassi, R.; Al-Hares, M.K.; Sirlantzis, K. Systematic Review of Intelligent Algorithms in Gait Analysis and Prediction for Lower Limb Robotic Systems. IEEE Access 2021, 9, 113788–113812. [Google Scholar] [CrossRef]
  39. Vicon Documentation. Available online: https://docs.vicon.com/display/Nexus25/Plug-in+Gait+kinematic+variables (accessed on 20 March 2020).
  40. Pfister, A.; West, A.M.; Bronner, S. Comparative Abilities of Microsoft Kinect and Vicon 3D Motion Capture for Gait Analysis. J. Med. Eng. Technol. 2014, 38, 274–280. [Google Scholar] [CrossRef]
  41. Van, D.B.T. Practical Guide to Data Smoothing and Filtering. Study Mater. 1996, pp. 1–6. Available online: http://isbweb.org/software/sigproc/bogert/filter.pdf (accessed on 31 October 1996).
  42. Schwartz, M.H.; Trost, J.P.; Wervey, R.A. Measurement and Management of Errors in Quantitative Gait Data. Gait Posture 2004, 20, 196–203. [Google Scholar] [CrossRef]
  43. Wright, C.; Seitz, A.L.; Arnold, B.L. Repeatability of Ankle Joint Kinematic Data at Heel Strike Using the Vicon Plug-In Gait Model. Available online: https://www.researchgate.net/publication/268380486 (accessed on 16 January 2011).
  44. Webber, J.T.; Raichlen, D.A. The Role of Plantigrady and Heel-Strike in the Mechanics and Energetics of Human Walking with Implications for the Evolution of the Human Foot. J. Exp. Biol. 2016, 219, 3729–3737. [Google Scholar] [CrossRef] [PubMed]
  45. Murakami, R.; Otaka, Y. Estimated lower speed boundary at which the walk ratio constancy is broken in healthy adults. J. Phys. Ther. Sci. 2017, 29, 722–725. [Google Scholar] [CrossRef] [PubMed]
  46. Rota, V.; Perucca, L.; Simone, A.; Tesio, L. Walk ratio (step length/cadence) as a summary index of neuromotor control of gait: Application to multiple sclerosis. Int. J. Rehabil. Res. 2011, 34, 265–269. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The soft lower-limb exoskeleton.
Figure 1. The soft lower-limb exoskeleton.
Bioengineering 11 00275 g001
Figure 2. Gait event division within a gait cycle.
Figure 2. Gait event division within a gait cycle.
Bioengineering 11 00275 g002
Figure 3. (A) The angular velocity of the Z-axis of the IMU located on the foot. (B) The angular velocity of the Z-axis of the IMU located on the thigh. (C) The specific force of the Z-axis of the IMU located on the right foot and thigh. (D) The vertical position of the heel obtained from VICON. (E) The vertical position of the toe obtained from VICON. (F) The hip angle obtained from VICON.
Figure 3. (A) The angular velocity of the Z-axis of the IMU located on the foot. (B) The angular velocity of the Z-axis of the IMU located on the thigh. (C) The specific force of the Z-axis of the IMU located on the right foot and thigh. (D) The vertical position of the heel obtained from VICON. (E) The vertical position of the toe obtained from VICON. (F) The hip angle obtained from VICON.
Bioengineering 11 00275 g003
Figure 4. The gait recognition framework. The red curve represents the turning, with the arrow direction clockwise indicating the turning left (TL) and counter clockwise indicating the turning right (TR); The red straight line represents the steering, with the arrow direction to the right indicating the left steering (LS) and the direction to the left indicating the right steering (RS).
Figure 4. The gait recognition framework. The red curve represents the turning, with the arrow direction clockwise indicating the turning left (TL) and counter clockwise indicating the turning right (TR); The red straight line represents the steering, with the arrow direction to the right indicating the left steering (LS) and the direction to the left indicating the right steering (RS).
Bioengineering 11 00275 g004
Figure 5. Testing assistance effect. (A) Kinematic test. (B) Metabolic test.
Figure 5. Testing assistance effect. (A) Kinematic test. (B) Metabolic test.
Bioengineering 11 00275 g005
Figure 6. Test conditions.
Figure 6. Test conditions.
Bioengineering 11 00275 g006
Figure 7. Actual data and recognition results for six gait cycles.
Figure 7. Actual data and recognition results for six gait cycles.
Bioengineering 11 00275 g007
Figure 8. (A) The typical gait characteristics of the hip angle. Point A represents the typical point with the highest hip angle and point B represents the typical point with the lowest hip angle. (B) Changes in point A angles of the hip joint. (C) Changes in point B angles of the hip joint. (D) The typical gait characteristics of the knee angle. Point C represents the typical point with the highest knee angle during the swing period and point D represents the typical point with the highest knee angle during the support period. (E) Changes in point C angles of the hip joint. (F) Changes in point D angles of the hip joint. (G) The typical gait characteristics of the ankle angle. Point E represents the typical point with the lowest ankle angle and point F represents the typical point with the lowest ankle angle. (H) Changes in point E angles of the ankle joint. (I) Changes in point F angles of the ankle joint.
Figure 8. (A) The typical gait characteristics of the hip angle. Point A represents the typical point with the highest hip angle and point B represents the typical point with the lowest hip angle. (B) Changes in point A angles of the hip joint. (C) Changes in point B angles of the hip joint. (D) The typical gait characteristics of the knee angle. Point C represents the typical point with the highest knee angle during the swing period and point D represents the typical point with the highest knee angle during the support period. (E) Changes in point C angles of the hip joint. (F) Changes in point D angles of the hip joint. (G) The typical gait characteristics of the ankle angle. Point E represents the typical point with the lowest ankle angle and point F represents the typical point with the lowest ankle angle. (H) Changes in point E angles of the ankle joint. (I) Changes in point F angles of the ankle joint.
Bioengineering 11 00275 g008
Figure 9. (A) The typical gait characteristics of the heel and toe during a gait cycle. Point G represents the highest point of the vertical position of the heel during the swing period. Point H and point K respectively represent the highest point of the vertical position of the toes during the swing period and before the foot lands on the ground. (B) Vertical position of point K at the toe under different conditions. (C) Vertical position of point G at the heel under different conditions. (D) Vertical position of point H at the toe under different conditions.
Figure 9. (A) The typical gait characteristics of the heel and toe during a gait cycle. Point G represents the highest point of the vertical position of the heel during the swing period. Point H and point K respectively represent the highest point of the vertical position of the toes during the swing period and before the foot lands on the ground. (B) Vertical position of point K at the toe under different conditions. (C) Vertical position of point G at the heel under different conditions. (D) Vertical position of point H at the toe under different conditions.
Bioengineering 11 00275 g009
Figure 10. Changes in step length under different conditions.
Figure 10. Changes in step length under different conditions.
Bioengineering 11 00275 g010
Figure 11. Changes in metabolic cost under different conditions.
Figure 11. Changes in metabolic cost under different conditions.
Bioengineering 11 00275 g011
Table 1. Dataset for different road conditions.
Table 1. Dataset for different road conditions.
TrialsSDLWUSDSUSLDSLTLTRLSRS
112,00019,85023,76823,81911,68511,62523,58422,92613,83512,244
212,00019,76023,81823,86211,32311,34723,69123,76812,84713,745
311,99819,86323,63323,78411,56911,57223,39623,29414,35412,961
411,99919,96823,89723,99011,35311,19023,28122,54314,24214,485
512,00019,85223,77823,77811,99211,67023,03023,84913,37213,656
611,99419,93123,89523,69211,50911,30923,57722,74114,56113,304
712,00019,93123,95123,58611,77611,74023,88823,64513,11214,472
812,00219,80323,74323,54111,45611,46623,13223,56012,99714,447
912,00019,98423,95323,90011,18711,27023,50323,86512,78312,652
1011,98619,81823,84523,59611,17811,85923,70122,74014,81514,413
1112,00019,87223,78823,89811,91211,62023,94223,89912,43512,133
1211,98519,90023,93423,73311,23811,55223,85223,30814,13214,659
1312,00019,91423,86923,67111,12311,39222,73722,58812,57913,929
1411,99519,94023,79323,83511,60711,92923,17622,97613,50212,503
1511,99819,80723,81123,88911,25511,97323,06223,56914,57714,488
1612,00019,86323,59123,76011,92811,35623,92422,87213,95414,751
1712,00019,94523,85123,57911,68911,24423,04023,84512,97513,747
1811,99719,94823,91623,85111,52411,27823,06023,92612,98612,885
1912,00219,89423,65523,95111,27911,72123,05222,98913,76214,419
2012,00019,93723,88923,91611,80711,19922,79323,95314,40912,288
2111,98619,88923,50023,55311,11311,82323,76723,39713,89213,534
2212,00019,97523,98223,67811,87311,21623,67923,26913,45814,566
2311,98519,88923,69323,79911,59511,76922,97623,75914,77613,311
2412,00019,97124,01223,70111,22411,43522,95122,83414,38013,768
2511,99519,86923,92723,98011,38911,83423,97623,45614,39212,697
Total299,922497,373595,492594,342287,584288,389584,770583,571343,127340,057
Table 2. Dataset of gait phase recognition for horizontal walking.
Table 2. Dataset of gait phase recognition for horizontal walking.
TrialsLWL-HSR-TOR-HMaxR-HSL-TOL-HMax
119,850250448172981220246152731
219,760212947022769235848762926
319,863239146762783209148953027
419,968221349562796231549832705
519,852231844842980225046403180
619,931224244652791238050542999
719,931245750602791227149522400
819,803222044532874221949513086
919,984218845992997216649533081
1019,818236745272776244450592645
1119,872246650842783220845782753
1219,900204447562787238748983028
1319,914211747983088225846343019
1419,940244045022892251546572934
1519,807235545722874220249762828
1619,863204845922782237748733191
1719,945231146442792211348543231
1819,948219950232793215646033174
1919,894259746272786229948632722
2019,937229046223091214246643128
2119,889208247902785236745923273
2219,975245950772696215546182970
2319,889204344392785221146373774
2419,971228050672896232145862821
2519,869245445082882238345793063
Total497,37357,214117,84071,25056,790119,59074,689
Table 3. Evaluation of motion state recognition models.
Table 3. Evaluation of motion state recognition models.
Method ModelACC 1 (%) F 1 ScoreMCC
KNN98.720.98780.9880
Random Forest99.310.99420.9935
Decision Tree98.950.98940.9901
SVM99.980.99850.9987
1 Accuracy refers to correctly determined motion state.
Table 4. Confusion matrix of the trained SVM.
Table 4. Confusion matrix of the trained SVM.
SDLWUSDSUSLDSLTLTRLSRS
SD0.99000000000
LW01.0000000000
US001.000000000
DS0001.00000000
USL00001.0000000
DSL000001.000000
TL0000001.00000
TR00000001.0000
LS0.0100000000.990
RS000000000.011.00
Table 5. Evaluation of gait phase recognition models.
Table 5. Evaluation of gait phase recognition models.
Method ModelACC 1 (%) F 1 ScoreMCC
NN96.360.96350.9612
RBF96.780.96750.9680
BP97.550.97520.9748
LSTM98.260.98210.9825
1 Accuracy refers to correctly determined gait phase.
Table 6. Confusion matrix of the trained LSTM.
Table 6. Confusion matrix of the trained LSTM.
L-HSR-TOR-HMaxR-HSL-TOL-HMax
L-HS0.97500000.017
R-TO0.0100.9950000
R-HMax00.0050.9910.01000
R-HS000.0090.97000
L-TO0000.0200.9910
L-HMax0.0150000.0090.983
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xiang, Q.; Wang, J.; Liu, Y.; Guo, S.; Liu, L. Gait Recognition and Assistance Parameter Prediction Determination Based on Kinematic Information Measured by Inertial Measurement Units. Bioengineering 2024, 11, 275. https://doi.org/10.3390/bioengineering11030275

AMA Style

Xiang Q, Wang J, Liu Y, Guo S, Liu L. Gait Recognition and Assistance Parameter Prediction Determination Based on Kinematic Information Measured by Inertial Measurement Units. Bioengineering. 2024; 11(3):275. https://doi.org/10.3390/bioengineering11030275

Chicago/Turabian Style

Xiang, Qian, Jiaxin Wang, Yong Liu, Shijie Guo, and Lei Liu. 2024. "Gait Recognition and Assistance Parameter Prediction Determination Based on Kinematic Information Measured by Inertial Measurement Units" Bioengineering 11, no. 3: 275. https://doi.org/10.3390/bioengineering11030275

APA Style

Xiang, Q., Wang, J., Liu, Y., Guo, S., & Liu, L. (2024). Gait Recognition and Assistance Parameter Prediction Determination Based on Kinematic Information Measured by Inertial Measurement Units. Bioengineering, 11(3), 275. https://doi.org/10.3390/bioengineering11030275

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop