Next Article in Journal
A Theoretical Study of Love Wave Sensors Based on ZnO–Glass Layered Structures for Application to Liquid Environments
Next Article in Special Issue
Design and Development of Non-Contact Bio-Potential Electrodes for Pervasive Health Monitoring Applications
Previous Article in Journal
Different Phases of Breast Cancer Cells: Raman Study of Immortalized, Transformed, and Invasive Cells
Previous Article in Special Issue
Wrist Pulse Rate Monitor Using Self-Injection-Locked Radar Technology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Real-Time Classification of Patients with Balance Disorders vs. Normal Subjects Using a Low-Cost Small Wireless Wearable Gait Sensor

1
The Department of Electrical & Computer Engineering, Texas Tech University (TTU), Lubbock, 79409 TX, USA
2
(On Leave) The Department Electrical & Electronic Engineering, Tokushima University, 770-8502 Tokushima, Japan
3
Boys Town National Research Hospital, Omaha, 68131 NE, USA
4
Texas Tech University Health Sciences Center (TTUHSC), Lubbock, 79439 TX, USA
*
Author to whom correspondence should be addressed.
Biosensors 2016, 6(4), 58; https://doi.org/10.3390/bios6040058
Submission received: 1 September 2016 / Revised: 14 November 2016 / Accepted: 17 November 2016 / Published: 29 November 2016
(This article belongs to the Special Issue Latest Wearable Biosensors)

Abstract

:
Gait analysis using wearable wireless sensors can be an economical, convenient and effective way to provide diagnostic and clinical information for various health-related issues. In this work, our custom designed low-cost wireless gait analysis sensor that contains a basic inertial measurement unit (IMU) was used to collect the gait data for four patients diagnosed with balance disorders and additionally three normal subjects, each performing the Dynamic Gait Index (DGI) tests while wearing the custom wireless gait analysis sensor (WGAS). The small WGAS includes a tri-axial accelerometer integrated circuit (IC), two gyroscopes ICs and a Texas Instruments (TI) MSP430 microcontroller and is worn by each subject at the T4 position during the DGI tests. The raw gait data are wirelessly transmitted from the WGAS to a near-by PC for real-time gait data collection and analysis. In order to perform successful classification of patients vs. normal subjects, we used several different classification algorithms, such as the back propagation artificial neural network (BP-ANN), support vector machine (SVM), k-nearest neighbors (KNN) and binary decision trees (BDT), based on features extracted from the raw gait data of the gyroscopes and accelerometers. When the range was used as the input feature, the overall classification accuracy obtained is 100% with BP-ANN, 98% with SVM, 96% with KNN and 94% using BDT. Similar high classification accuracy results were also achieved when the standard deviation or other values were used as input features to these classifiers. These results show that gait data collected from our very low-cost wearable wireless gait sensor can effectively differentiate patients with balance disorders from normal subjects in real time using various classifiers, the success of which may eventually lead to accurate and objective diagnosis of abnormal human gaits and their underlying etiologies in the future, as more patient data are being collected.

1. Introduction

Analysis of human gaits has long been an active area of research, and many systems have been proposed for observing and differentiating different gait patterns and their irregularities in the literature. Many of these existing systems use appearance-based approaches by extracting features and/or positions from the images captured from different video sequences using high-speed video cameras with frame rates of 50–200 Hz. Many studies have reported that image feature extraction using biomechanical models can allow quantitative analysis of specific gait characteristics, such as joint moments and powers (i.e., kinetic analysis), joint angles, angular velocities and angular accelerations (i.e., kinematic analysis) [1]. The testing protocol of these systems includes placing the optical markers near anatomical landmarks of the body and features related to various gait patterns are extracted from video sequences. Parametric models have been used extensively to describe a set of successive image observations. A vision-based three-dimensional (3D) modeling of motions of a human subject can be achieved by using volumetric bodies (e.g., using cylinders) that represent the flesh of the human subject and a simple collection of segments with joint angles representing the skeletal structure of the human body [2]. Alternatively, 2D models that represent the projection of 3D data onto an imaging plane can also be used (i.e., 2D/3D contour modeling [3]). However, since different kinetic and kinematic methods have been developed from these sophisticated and expensive visual gait analysis systems, it can be rather challenging to directly compare the gait analysis results from different systems/methods, as there is no standardization in the visual gait analysis methodology.
In addition, the variables that can be measured during gait analysis depend on the technique and sensors selected, which makes the direct comparison of gait analysis results derived from different gait sensing systems even much more difficult. The most commonly-reported gait measurement data include the temporospatial parameters, which include walking speed, body rotations, step time, step length and the durations of the stance phase and the swing phase [4]. A basic inertial measurement unit (IMU) that includes 3D gyroscopes and accelerometers can measure angular velocity and linear acceleration for each of the X/Y/Z axes, respectively, and these inexpensive IMUs have been used as wearable sensors that provide a powerful option for human gait analysis [4,5,6]. For example, Aminian et al. [4] and Selles et al. [5] reported methods of measuring both terminal contact (TC) that defines the beginning of the swing phase, as well as the initial contact (IC) that defines the beginning of the gait cycle timing information using those body-worn sensors. On the other hand, Yoshida et al. [6] used an accelerometer/IMU sensor attached to the patient’s waist and observed frequency peaks in the anterior plane to detect leg injury. Boutaayamou et al. [7] developed a signal processing algorithm to automatically extract, on a stride-by-stride basis, four consecutive fundamental events of walking, i.e., heel strike (HS), toe strike (TS), heel-off (HO) and toe-off (TO), from wireless accelerometers applied to the right and left foot. This accelerometer-based event identification was validated in seven healthy volunteers and a total of 247 trials against reference data provided by a force plate, a kinematic 3D analysis system and a video camera. An ambulatory monitoring method using an IMU sensor for patients with Parkinson’s disease has also been developed [8,9].
We have, therefore, designed a custom low-cost wireless gait analysis sensor (WGAS), which can be used for both fall detection and gait analysis. We reported to have demonstrated measured fall detection rates of 99% in classification accuracies among young volunteers using a similar WGAS with the BP-ANN and SVM classification algorithms [10,11,12]. This paper, however, studies and reports our new WGAS that is used specifically for gait analysis to detect patients with balance disorders among normal subjects while using various classification algorithms and input features to check the speed and accuracy of each classifier. This WGAS is placed and tested at the T4 position on the back of each subject, as shown in Figure 1. For the learning of the classification algorithms, we will first use the following six input features for the X/Y/Z axes extracted from the raw data taken from the WGAS (i.e., Rω: range of angular velocity; RA: range of acceleration, as shown in Equations (1) and (2)). They are used as one of the input features sets for these classifiers that yield excellent classification accuracies, and the details of the WGAS and our experimentation and analysis will be explained next.
R ω , x = max ( ω x ) min ( ω x ) , R ω , y = max ( ω y ) min ( ω y ) , R ω , z = max ( ω z ) min ( ω z )
R A , x = max ( A x ) min ( A x ) , R A , y = max ( A y ) min ( A y ) , R A , z = max ( A z ) min ( A z )

Wireless Gait Analysis Sensor

The custom-designed WGAS consists of a three-axis linear accelerometer IC, a single-axis gyroscope IC and a dual-axis gyroscope IC to measure 3D human body translations and rotations during a gait pattern; these ICs are designed with the help of micro-electrical and mechanical system (MEMS) sensors. Our WGAS measures the linear acceleration and angular rotations of the body movements, and there is no need for a magnetometer for our application of gait analysis and gait classification. Furthermore, the sensor has a dual- and single-axis gyroscope instead of a tri-axial gyroscope, because the particular analog MEMS tri-axial gyroscope was not available on the market during the design of the WGAS, but it is now. The future design of our WGAS can have a more compact IMU with a tri-axial single accelerometer IC and a single-gyroscope IC. However, we do not believe this will affect the sensing results whether one IC or two ICs are used. The details of these ICs with their manufacturers’ information can be found in [10,11]. This WGAS system is supported by a Texas Instruments (TI) MSP430 microcontroller (Texas Instruments, Dallas, TA, USA) and a wireless 2.4 GHz USB transceiver using the TI SimpliciTI™ protocol (Texas Instruments, Dallas, TA, USA) with a wireless communication range of ~12 meters (40 ft). An overall simplified system block diagram for the WGAS analysis system is shown in Figure 2.
The two AAA batteries used in our earlier wired sensor [13] were replaced by a single rechargeable Li-ion coin battery, providing a battery lifetime of ~40 h of continuous operation time with each recharge. The PCB, coin battery and the microcontroller are placed in a specially-designed 3D-printed plastic box (2.2′ × 1.5′ × 0.8′) with a total weight of 42 g. The design of the box was done with the 3D modeling software Rhinoceros (Rhino, Robert McNeel & Associates, WA, USA) and printed using a 3D printer with acrylonitrile butadiene styrene (ABS) plastic. The box has a sliding lid and is shown in Figure 1. The accelerometer data are sampled at 160 Hz and digitized to eight bits, with its output scaled to ±6 g at Δ V = ±6 g/VDD (VDD, Supply voltage = 3.6 V) for each axis. The gyroscope data are also sampled at 160 Hz and digitized to eight bits, with its output scaled to 300°/s (dps) at ΔV = ±300 dps. Multiplied by So (note the typical So (sensitivity) value is 3.752 mV/dps for the accelerometer and 3.33 mV/dps for the gyroscopes). The sensor orientation and position on the body during testing are also shown in Figure 1. The sensor is carefully secured to the subjects during testing to avoid artifacts. A cloth is used to secure the WGAS by tying it tightly at the T4 position of the patient’s back. The cloth is also attached to the shirt of the subject by using plastic tape. In addition, Velcro is attached to the sensor box that firmly attaches to the shirt, as well, to make sure that the sensor does not move around during the DGI tests shown in Table 1. The method described above proved that it significantly reduced the outliers in the measured datasets from the sensor. The microcontroller and the transceiver unit enable the real-time wireless transmission of the six-dimensional gait data to the nearby PC, where a LABVIEW™ program is used for designing the graphical user interface (GUI), as shown in Figure 3. The DC values for the six signals are not the same as the battery charges and discharges, and the calibration to make them the same DC level was not done to simplify the WGAS design, as we have ascertained that the detected AC signals from WGAS are really the ones that contain the most useful gait information, as will be shown in the remaining of this paper.

2. Experimental Section

All four patients and the three normal subjects performed the 7 Dynamic Gait Index (DGI) tests [10,14] with the details shown in Table 1. Therefore, a total of 28 dynamic gait tests were performed on 4 patients and 21 tests on three normal subjects. During all of the tests, the WGAS is placed at the T4 position (at back) for all of the subjects involved. Our WGAS sensor has been accepted by the Texas Tech University Health Sciences Center (TTUHSC) Internal Review Board (IRB) with study title “Fall risk identification and assessment using body worn sensors (CRI12-030 Fall study)”.
The six features of the raw data on range from the gyroscopes and the accelerometers from all of the testing subjects are important, and they form the inputs for training the classification algorithms. These include the range of angular velocity in the X/Y/Z direction (i.e., Rωx, from “GYRO X”; Rωy, from “GYRO Y”; and Rωz, from “GYRO Z”) and the range of acceleration in the X/Y/Z direction (i.e., RAx, from “ACC X”; RAy, from “ACC Y”; and RAz, from “ACC Z”). The data were obtained using a 1.7 GHz PC with 4 GB of RAM, Windows 8 OS, and using MATLAB R2015b. We compared the box plots of the four patients and three normal subjects to show several key test data statistics: the median, mean, range (highest to lowest values) and interquartile range (IQR) of DGI Tests 2, 3, 4 and 6, as shown in Figure 4, Figure 5, Figure 6 and Figure 7. These four DGI tests are selected and plotted here; they can show better visual comparisons than the other three DGI tests; some detailed statistics of all DGI tests are also listed later in this paper i.e., at the end of Section 2 (e.g., Figure 8 and Figure 9, etc.)
Moreover, the average of STDEV (i.e., standard deviation), Range, Mean, Median, and IQR of DGI Tests 2 and 7 are calculated and shown in Table 2 and Table 3.
As shown in Figure 4, the box plots of the patients’ data appear to be mostly larger than those of the normal subjects for DGI Test 2 for all extracted features, and they look significantly tighter for all of the 3-axial acceleration data for all DGI tests as shown in Figure 4, Figure 5, Figure 6 and Figure 7 (i.e., “ACC X”, “ACC Y” and “ACC Z”). The tighter the box plot distributions suggest the normal subjects are walking more steadily with less wobbling or sways than patients during the DGI tests.
Besides the larger box plots associated with larger STDEV and IQR for patients, in all box plots, we have also observed the median/mean values of the acceleration measured on the X and Z axes (i.e., “ACC X”, “ACC Z”) of the normal subjects are significantly larger from those of patients for all of the DGI tests, and we have shown this difference in Table 2 and Table 3 for DGI Test 2 and Test 7 as examples for better illustration. This contrast might be explained because walking gaits of normal subjects are considerably different from those of patients, where they walk freely with more accelerations along the X/Z direction from their center of mass (CoM), while patients of balance disorders typically walk with a decrease of speed, shortened stride length and other associated factors [4,5]. It is also interesting to notice that the median/mean values of ACC Y of the normal subjects are not so different from those of patients. This might be due to that during the DGI tests, the Y axis is parallel to the walking direction, while the X and Z axes are perpendicular to the walking direction, and therefore, we did not see much change in the median/mean values along the Y axis. We will need to look at these effects related to the median/mean values from accelerometers closely after more patient data can be collected to hopefully understand them better.
Moreover, as mentioned before, when the gait data’s STDEV is smaller, one would expect the walking to be more steady or stable. We can indeed see in the box plots and tables that the normal subjects’ STDEV values on Gyro X and Y and ACC X, Y and Z are all smaller than those of patients. However, the normal subjects’ STDEV data of Gyro Z are actually than that of the patients, suggesting that normal subjects may rotate their bodies around the Z axis naturally more and with more variation than patients of balance disorders; their faster walking speed than the patients may contribute to this effect, as well. Note we would really need to collect more patient gait data to improve the statistics and analysis details especially on the gyroscope data, as we can see that for all seven DGI tests, the STDEVs of Gyro Z data for Patients No. 1 and No. 2 are so different from those of Patients No. 3 and No. 4 (see Figure 4, Figure 5, Figure 6 and Figure 7). To see this better, we have also shown the values of the STDEV for normal subjects and patients for DGI Test 2 in Table 4. Therefore, one can see from Table 4 that the STDEVs of Gyro Z for normal subjects are actually slightly lower than those of “Patient 1” and “Patient 2”, but much greater than those of “patient “and “Patient 4”; therefore, the average STDEV of Gyro Z among normal subjects becomes larger than that of patients.
Finally, not surprisingly, Table 3 shows that the average range values of Gyro X, Y, Z and also ACC X, Y, Z for the patients’ gait data are all greater than those of the normal subjects’ for all of the DGI tests, except for the range of the Gyro Z data of the DGI Test 7, probably for the reasons stated before that the normal subjects may rotate their bodies around the Z axis naturally more with faster walking speeds and, therefore, with more variation and range difference than patients of balance disorders. Moreover, Figure 4, Figure 5, Figure 6 and Figure 7 show the box plots of DGI Tests 2, 3, 4 and 6 that group normal subjects as one group vs. patients as another group. From these points, the DGI Tests 2, 3, 4 and 6 present as better tests than the DGI Tests 1, 5 and 7 for differentiating patients from normal subjects. Having checked those basic data statistics, we have decided to first use the range values of the patients and normal subjects from all six sensor axes and for all 7 DGI tests taken as the single input features for the classification algorithms. We will later also use STDEV, etc., as input features to these classifiers and compare them, as well.
Finally, we have shown in Table 5 the STDEV, range, median, mean and IQR for normal subjects vs. patients for all seven DGI tests. To see the distributions on range better, the histograms of the data with normal distribution are now also shown for each of the six features for both patients and normal subjects in Figure 8 and Figure 9. The box plots of normal subjects and patients are also shown for these six features. It can be clearly seen that the features related to the acceleration data (i.e., the “ACC” data) look tighter in the distribution for the normal subjects than for the patients, especially for ACC X and ACC Z. After checking the raw gait data carefully to ensure the data integrity, we are now ready to show the classification algorithms and classification results next.

2.1. Classification Algorithms

Classification algorithms were used on the 6 features extracted from all 7 DGI tests and for all testing subjects to differentiate patients vs. normal subjects from all of the data collected.

2.1.1. Back Propagation Artificial Neural Network

An artificial neural network (ANN) can be seen as a machine that is designed to mimic how the brain performs a particular task or function of interest. Using ANN as a classifier has several advantages, such as: (1) neural networks are data-driven self-adaptive methods that can adjust themselves to the input data without using explicit mathematical functions; (2) it is a nonlinear model that can model most real-world problems; (3) neural networks are able to estimate Bayesian posterior probability, which can provide the basic estimation of classification rules [10]. To train the feed-forward ANN classifier in this work, back propagation (BP) was applied on the input features with scaled conjugate gradient (SCG) learning [10,11]. Similar to the neural network [10], our gait classification neural network also has three layers in which the input layer has six neurons that correspond to the six input feature values. The hidden layer can automatically extract the features of the input patterns. There is no definite rule used to determine the number of neurons in the hidden layer; it is a hit-and-trial method. Here, in our classification study using ANN, there is one hidden layer holding 10 hidden neurons, the number of which was optimized by adjusting the size of hidden neurons (from 1 to 15) as shown in Figure 10. For the hidden layer, a hyperbolic tangent sigmoid transfer function is used for each neuron. At the hidden layer, it is used to calculate network output from its input. The tangent hyperbolic function and its fast approximation are given by the following equation:
a i 1 = t a n s i g ( n i 1 ) = 2 exp ( 2 n i 1 ) + 1 1
where a i 1 is the i-th element of the a 1 vector containing the output from the hidden neurons, n i 1 is the i-th element of n 1 vector containing the net input going into the hidden units and n 1 is calculated by using the formula:
n 1 = w 10 p + b 1
where p is the input pattern, b 1 is the bias vector and w 10 is the weight matrix between the hidden layer and the output layer.
The output layer is designed based on the required output of the neural network. Here, we have used two output neurons corresponding to the two target classes the network needs to differentiate (i.e., the features of all patients are considered as Class 1, and all features of normal subjects are as Class 2). The pure linear activation function is selected for the output, given by the following equation:
a 2 = n 2
where a 2 is the column vector coming from the output layer and n 2 is the output net inputs going into the output layer, which can be calculated by using the following formula:
n 2 = w 21 a 1 + b 2
Note here that b2 is the bias at the second layer; w 21 is the synaptic weights at the hidden layer and output layer; and a1 is the column vector containing the outputs from the hidden layer.
Here, we have used the back propagation learning algorithm that consists of two paths; the forward path and the backward path. The forward path includes creating the feed-forward network, initializing weights, simulation and training the network. The network weights and biases are updated in the backward path. Here, we have used the scaled conjugate gradient (SCG) training algorithm that uses the gradient of the performance function to determine how to adjust weights to minimize the performance function. An iteration of this algorithm can be written as:
  x k + 1 = x k α k g k
where x k is the vector of current weights and biases, g k is the current gradient and α k is the learning rate.

2.1.2. Support Vector Machine

The support vector machine (SVM) method is popular for performing pattern recognition/classification on two categories of data with supervised learning. In our work, SVM was implemented similar to [10,15,16] to classify patients from normal subjects using the gait data from our WGAS system. A linear kernel was used for training the SVM classifier, which finds the maximum-margin hyper-plane from the given training dataset D, and it can be described as:
D = { ( x i , y i ) | x i i p , y i { 1 , 1 } } i = 1 n
where y i is either 1 or −1, and n is the number of training data. Each x i is a p-dimensional vector having the feature quantity R. Any hyper-plane can be written as:
w · x b = 0
where w is the vector to the hyper-plane. If the training data are linearly separable, the hyper-plane can be described as [15]:
w · x b = 1   and   w · x b = 1
The distance between these two hyper-planes is 2 / w , so the purpose is to minimize w .
In general, it is hard to separate the training data linearly. When the training data are not linearly separable, the hyper-plane can be described as:
M i n   1 2 w 2 + C i = 1 l ε i
where parameter C determines a trade-off between the error on the training set and the separation of the two classes. Here, ε is a set of slack variables. The dual problems lie in maximizing the following function with respect to the Lagrange multiplier α [15]:
M a x   i = 1 l α i   1 2 j = 1 l α i α j y i y j ( x i · x j )
subject to 0     α i   C   ( i = 1 , , l ) :
i = 0 l α i y i = 0

2.1.3. K-Nearest Neighbor

The K-nearest neighbors algorithm (KNN) is a simple, efficient non-parametric method used for classification and regression in pattern recognition, object recognition, etc. [17]. In both cases, the input consists of the K closest training examples in the feature space.
In KNN classification, the object or an unknown sample is classified by assigning to a test pattern the class label of its K nearest neighbors. The object is classified based on the category of its nearest neighbors through a voting procedure. Majority votes of its neighbors are considered during classification, with the object assigned to the most common class among its K nearest neighbors (K is a positive integer, typically small). If K = 1, then there is only one nearest neighbor, and the object is simply assigned to that class. Because of its simplicity and efficiency, the KNN-based algorithms are widely used in many applications.

2.1.4. Binary Decision Tree

Binary decision tree (BDT), also called as decision trees or classification trees, can predict responses to data as a classifier. To predict a response, one needs to follow the decisions in the tree from the root (beginning) node down to a leaf node, where the leaf node contains the response. BDT give responses that are nominal, such as “true” or “false”. Here, we have used BDT for the patients’ and normal subjects’ gait data classification. In data mining, BDT can also be used for regression and generalization of a given set of data [18]. Data come in records of the form:
( x , Y ) = ( x 1 , x 2 , x 3 , , x k , Y )
The dependent variable in Equation (14), Y, is the target variable that we are trying to predict, classify or generalize. The vector x consists of the input variables, x1, x2, x3, etc., that are used for classification.
The overall system flow for our real-time gait classification system is shown in Figure 11. We will now present the classification results from our WGAS system using these aforementioned classifiers next.

3. Results and Discussion

The training and testing datasets were divided into 70:30 in percentages for the BP-ANN algorithm. The rest of the three algorithms used the K-fold cross validation method with K = 6 [19]. The K-fold cross validation method generalizes the approach by segmenting the data into K equally-sized partitions. During each run, one of the partitions is chosen for testing, while the rest of them are used for training. The procedure is repeated K times so that each partition is used for testing exactly once. Again, the total error is found by summing up the errors for K runs. A parallel coordinate plot is shown in Figure 12, illustrating the six features for all 49 datasets (seven subjects with seven DGI tests). The blue lines refer to extracted features for patients, and the brown lines refer to the extracted features for normal subjects.
From Figure 12, we can clearly see that the features F4, F5 and F6 of the normal subjects (i.e., brown lines; range of ACC Z, range of ACC X and range of ACC Y, respectively) are almost concentrated at 0.1 Volts in the feature space, but the F4, F5 and F6 of the patients (i.e., blue lines) are varying from 0.2 to 1 in the feature space. Therefore, the linear acceleration features (i.e., F4, F5 and F6) from the accelerometer data can be used to differentiate patients from normal subjects. However, the gyroscopic features (i.e., F1, F2 and F3) vary rather differently, and they appear not that different for normal subjects vs. patients. Nevertheless, we found that the gyroscopic features are still important for the classification algorithms to possibly increase their classification accuracy.
Before we show the classification results, we would like to introduce some straight-forward graphical representation and figures-of-merit used to evaluate a classifier. For example, a “confusion matrix” is a useful and simple tool for analyzing/representing how well a classifier can recognize input features of different classes [20]. It tabulates the outcomes performed by a classifier for both correctly- and incorrectly-classified cases. Its definition is shown in Figure 13.
Additionally, popular performance metrics for biosensors, such as sensitivity, specificity, positive predictive value (PPV) or precision, F-measure and accuracy, are also used in this work as defined below [21], where TP = true positives, TN = true negatives, FP = false positives, FN = false negatives.
Sensitivity (recall): This measures the actual members of the class that are correctly identified as such. It is also referred to as the true positive rate (TPR). It is defined as the fraction of positive examples predicted correctly by the classification model.
Sensitivity = TP TP + FN
Classifiers with large sensitivity have very few positive examples misclassified as the negative class.
Specificity: This is also known as the true negative rate. It is defined as the fraction of total negative examples that are predicted correctly by the model/classifier.
Specificity =   TN TN + FP
Precision (positive predictive value): Precision determines the fraction of records that actually turns out to be positive in the group the classifier has declared as positive class.
Precision = TP TP + FP
The higher the precision is, the lower the number of false positive errors committed by the classifier.
Negative predictive value (NPV): This is the proportion of samples that do not belong to the class under consideration and that are correctly identified as non-members of the class.
NPV = TN TN + FN
F − measure: Precision and sensitivity are two widely-used metrics for evaluating the correctness of a classifier or a pattern recognition algorithm. Building a model that maximizes both precision and sensitivity is the key challenge for classification algorithms. Precision and sensitivity can be summarized into another metric known as the F-measure, which is the harmonic mean of precision and sensitivity, given by,
F measure = 2 × Precision × Sensitivity Precision + Sensitivity
Accuracy: Accuracy is used as a statistical measure of how well a binary classification test identifies or excludes a condition. It is a measure of the proportion of the true results as defined by Equation (20):
Accuracy = TP + TN TP + FP + TN + FN
Now, we are ready to discuss the classification results. Considering the importance of all of these six features from the range data, each classifier was trained with the inputs of all six features. The back propagation artificial neural network (BP-ANN) achieved an impressive 100% accuracy with the SCG learning. This SCG algorithm performs the search and chooses the step size by using the information from the second order error function from the neural network. The SCG training is optimized by the parameter sigma σ (which determines the change in weight for the second derivative approximation) and lambda λ (which regulates the indefiniteness of the Hessian). The values of σ and λ were taken as 5 e 5 and 5 e 7 , respectively.
The confusion matrix of the BP-ANN classifier is shown in Figure 14, which gives the TP, TN, FP and FN values.
Next, the SVM classifier with a linear kernel achieved 98% overall accuracy with only one misclassification from the 49 feature datasets. The misclassified data are shown in Figure 15, and the corresponding confusion matrix is shown in Figure 16. The only one misclassification is Case 7 of the first normal subject, which was misclassified as the patient class.
The KNN and BDT algorithms were also trained on the data collected and achieved 96% and 94% in accuracy, respectively. The parallel coordinates plots and the confusion matrices are shown in Figure 17, Figure 18, Figure 19 and Figure 20, respectively. For the KNN classifier, there are two misclassifications: Case 7 of the first normal subject and Case 1 of the second patient were misclassified. Somewhat similarly, for the BDT algorithm, there are three misclassifications: Case 7 of the first normal subject, Case 4 of the fourth patient and Case 6 of the first patient.
From the classification results of the trained algorithms, we can clearly see that our WGAS system is robust enough to differentiate patients and the normal subjects with the simple features extracted from the raw accelerometer and gyroscope data as each classifier exhibits 94%–100% high accuracy. The comparison and the performance of the algorithms are shown in the Table 6 as below, where the range data have been used as the input features to the classifiers.
Next, we also tried different statistical parameters to use as the input features to check the impact on classifier accuracy. When only STDEV is used as the input feature, Table 7 suggests the classification accuracies for all classifiers degraded, while SVM classification accuracy is now 94%, slightly better than the other classifiers. When both range and STDEV are used as the input features, the SVM’s accuracy improves to 98%, and all classifiers have higher than 94% accuracy. The best gait classification accuracy still occurs when only range data are used as input features; in that case, the BP-ANN outperformed all other classifiers with 100% accuracy.
Finally, we need to compare the running time of each of the classifiers to see which algorithm is the fastest to perform this gait classification in real time; we also compared the speed when different input features are used to train the classification algorithms, and the results are shown in Table 8. It is apparent that a simple, but very fast BP-ANN classifier appears to be the best classifier to differentiate patients of balance disorders vs. the normal subjects in real time. SVM also has achieved 100% precision and specificity, but 98% accuracy; however, it is significantly slower than BP-ANN, KNN or BDT. We are collecting more gait data from patients and normal subjects now to improve the data statistics and to ascertain if BP-ANN and SVM would still be the best algorithms for real-time patient classification.
Table 9 shows the results of this work compared with previously-published work in the literature. The Mannini et al. [22] used three IMUs featuring a tri-axial accelerometer and a tri-axial gyroscope and collected the raw data from the testing subjects. They have achieved 90.5% classification accuracy using the RBF (radial basis function) kernel SVM classification algorithm by including time domain features, like the mean value, STDEV, maximum, minimum and range in their feature extraction. The Tahir et al. [23] used both ANN and SVM classifier algorithms for the gait classification in Parkinson’s disease patients. They have used the SVM classifier with the RBF kernel in distinguishing normal and patients based on kinetic features. They have also used ANN with the Levenberg–Marquardt training algorithm and achieved 98.2% and 96.9% classification accuracy, respectively.
The Bregg et al. [24] applied an ANN and SVM for the automatic recognition of young-old gait types from their respective gait patterns. Minimum foot clearance (MFC) data of young and elderly participants were analyzed using a PEAK-2D motion analysis system during a 20-min continuous walk on a treadmill at a self-selected walking speed. Gait features extracted from Poincaré plot images were used to train the SVM and ANN. Cross-validation test results indicate that their generalization performance of the SVM was on average 83.3% (±2.9) to recognize young and elderly gait patterns, compared to a neural network’s accuracy of 75.0% (±5.0). The same research group of [24] used a synchronized PEAK 3D motion analysis system and a force platform during normal walking for young and elderly subjects and achieved 83.3% vs. 91.7% generalization performance for ANN and SVM, respectively, as reported in [25]. The Hasin et al. [26] used both SVM and ANN for gait recognition by extracting geometry and texture features from the frame sequence of the video when the person is walking. They have used polynomial SVM of order three and BP-ANN with SCG training and achieved overall accuracy of 98% for both of the classifiers. The Huang et al. [27] has built intelligent shoes for human identification under the framework of capturing and analyzing human gait. The data of that work were collected from different sensors, like a pressure sensor, a tilt angle sensor, three single-axis gyros, one tri-axial ACC and a bend sensor installed in the shoe. Principle component analysis (PCA) was used for feature generation, and SVM was applied for training and classifier generation. They were successful in achieving a 98% human identification rate. The Lugade et al. [28] used ANN to determine dynamic balance control, as defined by the interaction of the center of mass (CoM) with the base of support (BoS), during the gait in the elderly using clinical evaluation on gaits. Subjects were asked to walk at a self-selected comfortable speed across a 10 m walkway in that work. During ambulation, 29 retro reflective markers were placed on bony landmarks of the body, where 3D marker trajectories were captured with an eight-camera motion analysis system (Motion Analysis Corp, Santa Rosa, CA, USA). BP-ANN was able to correctly identify the interaction of the CoM with BoS of elderly subjects with an 89% accuracy rate.
The Muhammad et al. [29] used 25 reflective markers, which were placed on the body, and data were acquired from the Vicon Nexus 3D motion capture system to analyze the gait patterns, kinetic and kinematic parameters of the hip, knee and ankle joints of patients and normal subjects. They have used ANN to predict the gait patters with approximately 95% accuracy. The Ahlrichs et al. [30] used one tri-accelerometer device worn on the waist of the testing subjects used in detecting the freezing of gait (FOG) symptoms on the people suffering from Parkinson’s disease. The acceleration signals from a waist-mounted sensor are split into equally-sized windows (i.e., a sliding window is applied to the time series), and features are extracted from those windows and fed to an SVM for training or classification. The RBF kernel SVM achieved 98% accuracy in detecting the symptoms for the patients with Parkinson’s disease.
We have attempted to validate the measured WGAS data with the data taken from a video-based kinematic reference system; i.e., a Vicon motion capture system with cameras. However, so far, we have found it very difficult to validate the WGAS data directly this way, as the Vicon camera-based system measures the movements from the markers on the body, but not the exact force-based acceleration data, as measured from our WGAS. Therefore, a limitation of the current paper exists in the lack of a direct validation of the values of the extracted features to the measured data provided by another reference system (e.g., a kinematic system). However, we do plan to continue the next stage of the research by including this validation step against possibly another kinematic system and/or using a different validation method. In addition, to improve the data statistics of our work, we would like to add more data as we compared the performance of the classifiers with the increased normal data, as shown in Table 10 in the revised manuscript. The total dataset has four patients and twelve normal subjects, and the overall accuracies are still improved when compared with the results from Table 6.

4. Conclusions

The results presented in this work using our custom WGAS gait analysis system with artificial neural networks (ANN) and other classifiers, such as SVM, suggest that our low-cost system can successfully classify/detect patients with balance disorders from normal subject with very high accuracy and in almost no time. In this study, we used six simple features from our raw WGAS data collected during the DGI tests for seven subjects with BP-ANN SCG trained and/or linear SVM classifiers, and they achieved impressive 100%, 98% overall accuracies, respectively. In addition, KNN and binary decision tree (BDT) algorithms have obtained still pretty good 96% and 94% overall accuracies, respectively. We have also studied the performance of these classification algorithms when different input features (i.e., STDEV and range + STDEV) are used to train the classifiers, and the results are shown in Table 7 and Table 8, indicating that BP-ANN appears to have the best combined overall performance of classification accuracy and speed. We have also compared our gait classification results against prior works in the literature and confirmed that our work achieved very high classification accuracies, albeit with a limited subject size. We are in the process of collecting more data for patients of balance disorders at this moment. One of the goals of our study is to provide physicians and patients with a cost-effective means to identify dynamic balance issues and the possible risk of falls from data routinely collected in clinical examinations, such as the DGI tests reported in this work. In the future, we also plan to use our low-cost WGAS and classifiers, such as BP-ANN, to differentiate patients-specific gait issues to potentially form a powerful expert system, capable of real-time gait analysis to assist quantitative diagnosis, monitoring and assessment of fall risks and also to potentially help suggesting effective fall prevention schemes.

Acknowledgments

The authors would like to thank TTUHSC (Texas Tech University Health Sciences Center) Clinical Research Institute for their assistance in the preparation and management of the study. The authors would also like to thank the Texas Instruments (TI) Analog Systems and Design (ASD) program and the Keh-Shew Lu Regents Chair Endowment at Texas Tech University for funding this study.

Author Contributions

Donald Yu-Chun Lie, Jerry Lopez and Jerry Tsay conceived of and developed the idea of making the WGAS sensors and fabricated and debugged earlier generations of the gait sensors. Bhargava Teja Nukala designed a new generation WGAS sensor. Amanda Rodriguez collected the gait data from patients and normal subjects under the supervision and help of Steven Zupancic and Tam Nguyen. Taro Nakano helped in doing some of the data analysis. Bhargava Teja Nukala has done the feature extraction from the collected gait data and applied classification algorithms, as well as the actual paper writing with the help of Donald Yu-Chun Lie.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Harris, G.F.; Wertsch, J.J. Procedures for Gait Analysis. Arch Phys. Med. Biol. 1994, 75, 216–225. [Google Scholar]
  2. Hogg, D. Model based vision: A program to see a walking person. Image Vis. Comput. 1983, 1, 5–20. [Google Scholar] [CrossRef]
  3. Isard, M.; Blake, A. Condensation-Conditional density propagation for visual tracking. Int. J. Comput. Vis. 1998, 29, 5–28. [Google Scholar] [CrossRef]
  4. Aminian, K.; Rezakhanlou, K.; De Andres, E.; Fritsch, C.; Leyvraz, P.F.; Robert, P. Temporal feature estimation during walking using miniature accelerometers: An analysis of gait improvement after hip arthroplasty. Med. Biol. Eng. Comput. 1999, 37, 686–691. [Google Scholar] [CrossRef] [PubMed]
  5. Selles, R.W.; Formanoy, M.A.; Bussmann, J.B.; Janssens, P.J.; Stam, H.J. Automated estimation of initial and terminal contact timing using accelerometers; development and validation in transtibial amputees and controls. Neural Syst. Rehabil. Eng. 2005, 13, 81–88. [Google Scholar] [CrossRef] [PubMed]
  6. Yoshida, T.; Mizuno, F.; Hayasaka, T.; Tsubota, K.; Wada, S.; Yamaguchi, T. Gait Analysis for Detecting a leg accident with an accelerometer. In Proceedings of the 1st Distributed Diagnosis and Home Healthcare (D2H2) Conference, Arlington, VA, USA, 2–4 April 2006.
  7. Boutaayamou, M.; Schwartz, C.; Stamatakis, J.; Denoël, V.; Maquet, D.; Forthomme, B.; Croisier, J.L.; Macq, B.; Verly, J.G.; Garraux, G.; et al. Development and validation of an accelerometer-based method for quantifying gait events. Med. Eng. Phys. 2015, 37, 226–232. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Moore, S.T.; MacDougall, H.G.; Ondo, W.G. Ambulatory monitoring of freezing of gait in Parkinson’s disease. J. Neurosci. Methods 2008, 167, 340–348. [Google Scholar] [CrossRef] [PubMed]
  9. Salarian, A.; Russmann, H.; Vingerhoets, F.J.G.; Burkhard, P.R.; Aminian, K. Ambulatory monitoring of physical activities in patients with Parkinson’s disease. IEEE Trans. Biomed. Eng. 2007, 54, 2296–2299. [Google Scholar] [CrossRef] [PubMed]
  10. Nukala, B.T.; Shibuya, N.; Rodriguez, A.; Tsay, J.; Lopez, J.; Nguyen, T.; Zupancic, S.; Lie, D.Y.C. An Efficient and Robust Fall Detection System Using Wireless Gait Analysis Sensor with Artificial Neural Network (ANN) and Support Vector Machine (SVM) Algorithms. Open J. Appl. Biosens. 2014, 3, 29–39. [Google Scholar] [CrossRef]
  11. Nukala, B.T.; Shibuya, N.; Rodriguez, A.I.; Tsay, J.; Nguyen, T.Q.; Zupancic, S.; Lie, D.Y.C. A Wireless Gait Analysis Sensor for Real-Time Robust Fall Detection Using an Artificial Neural Network. In Proceedings of the IEEE Point-of-Care Technologies Conference, Seattle, WA, USA, 8–10 October 2014; pp. 219–222.
  12. Lie, D.Y.C.; Nukala, B.T.; Jacob, J.; Tsay, J.; Lie, P.E.; Shibuya, N.; Rodriguez, A.I.; Nguyen, T.Q.; Zupancic, S. The Design of Robust Real-Time Wearable Fall Detection Systems Aiming for Fall Prevention, Activities of Daily Living (ADL): Cultural Differences, Impacts of Disease and Long-Term Health Effects, Chapter 3. In Public Health in the 21st Century; Lively, S.T., Ed.; Nova Science Publishers, Inc.: New York, NY, USA, 2015; pp. 51–76. [Google Scholar]
  13. Jacob, J.; Nguyen, T.; Zupancic, S.; Lie, D.Y.C. A Fall Detection Study on the Sensor Placement Locations and the Development of a Threshold-Based Algorithm Using Both Accelerometer and Gyroscope. In Proceedings of IEEE International Conference on Fuzzy Logics, Taipei, Taiwan, 27–30 June 2011; pp. 666–671.
  14. Perel, K.L.; Nelson, A.; Goldman, R.L.; Luther, S.L.; Prieto-Lewis, N.; Rubenstein, L.Z. Fall Risk Assessment Measures: An Analytic Review. J. Gerontol. Ser. A Biol. Sci. Med. Sci. 2001, 56, 761–766. [Google Scholar] [CrossRef]
  15. Joachims, T. Making Large-Scale Support Vector Machine Learning Practical. In Advances in Kernel Methods—Support Vector Learning; Schölkopf, B., Burges, C.J.C., Smola, A.J., Eds.; MIT Press: Cambridge, UK, 1999; pp. 169–184. [Google Scholar]
  16. Taro, N.; Nukala, B.T.; Tsay, J.; Rodriguez, A.I.; Nguyen, T.Q.; Zupancic, S.; Lie, D.Y.C. Gaits Classification of Normal vs. Patients by Wireless Gait Sensor and Support Vector Machine (SVM) Classifier. Int. J. Softw. Innov. 2016, 4, 17–29. [Google Scholar]
  17. Altman, N.S. An introduction to kernel and nearest-neighbor nonparametric regression. Am. Stat. 1992, 46, 175–185. [Google Scholar]
  18. Cha, S.H.; Tappert, C. A genetic algorithm for constructing compact binary decision trees. J. Pattern Recogn. Res. 2009, 4, 1–3. [Google Scholar] [CrossRef]
  19. Kohavi, R. A study of cross-validation and bootstrap for accuracy estimation and model selection. In Proceeding of the International Joint Conference on Artificial Intelligence, Montréal, QC, Canada, 20–25 August 1995; pp. 1137–1145.
  20. Han, J.; Kamber, M.; Pei, J. Data Mining: Concepts and Techniques, 3rd ed.; Morgan Kaufmann: San Francisco, CA, USA, 2012. [Google Scholar]
  21. Tan, P.N.; Steinbach, M.; Kumar, V. Introduction to Data Mining; Addision-Wesley: Reading, MA, USA, 2005; pp. 487–559. [Google Scholar]
  22. Mannini, A.; Trojaniello, D.; Cereatti, A.; Sabatini, A.M. A Machine Learning Framework for Gait Classification Using Inertial Sensors: Application to Elderly, Post-Stroke and Huntington’s disease Patients. Sensors 2016, 16, 134. [Google Scholar] [CrossRef] [PubMed]
  23. Tahir, N.M.; Manap, H.H. Parkinson Disease Gait Classification based on Machine Learning Approach. J. Appl. Sci. 2012, 12, 180–185. [Google Scholar] [CrossRef]
  24. Begg, R.K.; Palaniswami, M.; Owen, B. Support Vector Machines for automated gait classification. IEEE Trans. Biomed. Eng. 2005, 52, 828–838. [Google Scholar] [CrossRef] [PubMed]
  25. Begg, R.K.; Kamruzzaman, J. A comparison of neural networks and Support Vector Machines for recognizing young-old gait patterns. In Proceedings of the Conference on Convergent Technologies for the Asia-Pacific Region (TENCON 2003), Bangalore, India, 15–17 October 2003; pp. 354–358.
  26. Hassin, A.A.; Hanon, A.; Talal, A.A.; Barges, E. Gait Recognition System Using Support Vector Machine and Neural Network. Basrah J. Agric. Sci. 2014, 40, 68–78. [Google Scholar]
  27. Huang, B.; Chen, M.; Ye, W.; Xu, Y. Intelligent Shoes for Human Identification. In Proceedings of the 2006 IEEE International Conference on Robotics and Biomimetics, Kunming, China, 17–20 December 2006; pp. 601–606.
  28. Lugade, V.; Lin, V.; Farley, A.; Chou, L.S. An Artificial Neural Network estimation of gait balance control in the elderly using clinical evaluations. PLoS ONE 2014, 9, e97595. [Google Scholar] [CrossRef] [PubMed]
  29. Muhammad, J.; Gibbs, S.; Abbound, R.; Anand, S.; Wang, W. Gait Pattern Recognition in Cerebral Palsy Patients Using Neural Network Modeling. J. Ayub Med. Coll. Abbottabad 2015, 27, 754–758. [Google Scholar] [PubMed]
  30. Ahlrichs, C.; Samà, A.; Lawo, M.; Cabestany, J.; Rodríguez-Martín, D.; Pérez-López, C.; Sweeney, D.; Quinlan, L.R.; Laighin, G.Ò.; Counihan, T.; et al. Detecting freezing of gait with a tri-axial accelerometer in Parkinson’s disease patients. Med. Biol. Eng. Comput. 2016, 54, 223–233. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Physical structure and dimension of our custom-designed wireless gait analysis sensor (WGAS) hardware and the orientation placed during tests at T4 (back).
Figure 1. Physical structure and dimension of our custom-designed wireless gait analysis sensor (WGAS) hardware and the orientation placed during tests at T4 (back).
Biosensors 06 00058 g001
Figure 2. A simplified block diagram of the gait analysis/classification system using our WGAS.
Figure 2. A simplified block diagram of the gait analysis/classification system using our WGAS.
Biosensors 06 00058 g002
Figure 3. An example output on the LABVIEW GUI of the data wirelessly transmitted from the WGAS to a nearby PC, showing the real-time data for the six monitored signals. For example, the “GYRO X” signal is the angular velocity measured from the gyroscope centered on the X axis, and the “ACC X” signal is the acceleration measured from the accelerometer along the X axis, respectively.
Figure 3. An example output on the LABVIEW GUI of the data wirelessly transmitted from the WGAS to a nearby PC, showing the real-time data for the six monitored signals. For example, the “GYRO X” signal is the angular velocity measured from the gyroscope centered on the X axis, and the “ACC X” signal is the acceleration measured from the accelerometer along the X axis, respectively.
Biosensors 06 00058 g003
Figure 4. Box plots of DGI Test 2: Data for normal subjects (left) vs. patients (right).
Figure 4. Box plots of DGI Test 2: Data for normal subjects (left) vs. patients (right).
Biosensors 06 00058 g004
Figure 5. Box plots of DGI Test 3: Data for normal subjects (left) vs. patients (right).
Figure 5. Box plots of DGI Test 3: Data for normal subjects (left) vs. patients (right).
Biosensors 06 00058 g005
Figure 6. Box plots of DGI Test 4: Data for normal subjects (left) vs. patients (right).
Figure 6. Box plots of DGI Test 4: Data for normal subjects (left) vs. patients (right).
Biosensors 06 00058 g006
Figure 7. Box plots of DGI Test 6: Data for normal subjects (left) vs. patients (right).
Figure 7. Box plots of DGI Test 6: Data for normal subjects (left) vs. patients (right).
Biosensors 06 00058 g007
Figure 8. Histograms with the Gaussian distribution and box plots for the six ranges combining all of the DGI tests from the four patients.
Figure 8. Histograms with the Gaussian distribution and box plots for the six ranges combining all of the DGI tests from the four patients.
Biosensors 06 00058 g008
Figure 9. Histograms with the Gaussian distribution and box plots for the six ranges combining all of the DGI tests from the three normal subjects.
Figure 9. Histograms with the Gaussian distribution and box plots for the six ranges combining all of the DGI tests from the three normal subjects.
Biosensors 06 00058 g009
Figure 10. The 3-layer back propagation artificial neural network (BP-ANN) topology used for this work.
Figure 10. The 3-layer back propagation artificial neural network (BP-ANN) topology used for this work.
Biosensors 06 00058 g010
Figure 11. Overall data acquisition, feature extraction and data classification analysis flow for the gait analysis system using our WGAS.
Figure 11. Overall data acquisition, feature extraction and data classification analysis flow for the gait analysis system using our WGAS.
Biosensors 06 00058 g011
Figure 12. Six extracted features from the WGAS data collected on range (i.e., the difference between maximum and minimum) for four patients and three normal subjects. A total of seven persons each had all seven DGI tests done, so there are 7 × 7 = 49 cases.
Figure 12. Six extracted features from the WGAS data collected on range (i.e., the difference between maximum and minimum) for four patients and three normal subjects. A total of seven persons each had all seven DGI tests done, so there are 7 × 7 = 49 cases.
Biosensors 06 00058 g012
Figure 13. The definition of the confusion matrix [20], which is a tool for analyzing how well a classifier can recognize input features of different classes by tabulating the classification results correctly/incorrectly as predicted by the classification model.
Figure 13. The definition of the confusion matrix [20], which is a tool for analyzing how well a classifier can recognize input features of different classes by tabulating the classification results correctly/incorrectly as predicted by the classification model.
Biosensors 06 00058 g013
Figure 14. The final confusion matrix showing the gait classification results using the BP-ANN classification model with the true positive rate (TPR) and false positive rate (FPR).
Figure 14. The final confusion matrix showing the gait classification results using the BP-ANN classification model with the true positive rate (TPR) and false positive rate (FPR).
Biosensors 06 00058 g014
Figure 15. SVM classification results: misclassification shown in the feature space with only one case misclassified (i.e., DGI Test 7 of the first normal subject is classified as the “patient” class).
Figure 15. SVM classification results: misclassification shown in the feature space with only one case misclassified (i.e., DGI Test 7 of the first normal subject is classified as the “patient” class).
Biosensors 06 00058 g015
Figure 16. The final confusion matrix showing the gait classification results using the SVM classification model with the true positive rate (TPR) and false positive rate (FPR).
Figure 16. The final confusion matrix showing the gait classification results using the SVM classification model with the true positive rate (TPR) and false positive rate (FPR).
Biosensors 06 00058 g016
Figure 17. K-nearest neighbor (KNN) classification results: two misclassifications shown in the feature space: Case 7 (DGI 7) of the first normal subject and Case 1 (DGI 1) of the second patient classified as patient and normal subject, respectively.
Figure 17. K-nearest neighbor (KNN) classification results: two misclassifications shown in the feature space: Case 7 (DGI 7) of the first normal subject and Case 1 (DGI 1) of the second patient classified as patient and normal subject, respectively.
Biosensors 06 00058 g017
Figure 18. The final Confusion matrix showing the gait classification results using the KNN classification model with the TPR and False Positive Rate (FPR).
Figure 18. The final Confusion matrix showing the gait classification results using the KNN classification model with the TPR and False Positive Rate (FPR).
Biosensors 06 00058 g018
Figure 19. Binary decision tree (BDT) classification results: three misclassifications shown in the feature space: Case 7 (DGI 7) of the first normal subject, Case 4 (DGI 4) of the fourth patient and Case 6 (DGI 6) of the first patient.
Figure 19. Binary decision tree (BDT) classification results: three misclassifications shown in the feature space: Case 7 (DGI 7) of the first normal subject, Case 4 (DGI 4) of the fourth patient and Case 6 (DGI 6) of the first patient.
Biosensors 06 00058 g019
Figure 20. Confusion matrix showing the gait classification results using the binary decision tree (BDT) classification model with the true positive rate (TPR) and false positive rate (FPR).
Figure 20. Confusion matrix showing the gait classification results using the binary decision tree (BDT) classification model with the true positive rate (TPR) and false positive rate (FPR).
Biosensors 06 00058 g020
Table 1. Dynamic Gait Index (DGI) tests performed for gait analysis.
Table 1. Dynamic Gait Index (DGI) tests performed for gait analysis.
Test/Case No.Description
1Gait level surface: walk with a normal speed up to the 20′ mark
2Change in gait speed: walk with a normal pace up to 5′; walk fast for the next 5′; walk slowly for the next 5′; and walk normally for the last 5′
3Gait with horizontal head turns: walk normally with horizontal head turns up to the 20′ mark
4Gait with vertical head turns: walk normally with vertical head turns up to the 20′ mark
5Gait and pivot turn: walk normally, but at the end, turn around with a pivot turn
6Step over obstacle: walk normally, and when you come across the obstacle, step over, but not around it
7Step around obstacles: walk normal, and when encountering the 1st obstacle, walk around the right-side; when encountering the 2nd obstacle, walk around the left-side
Table 2. The standard deviation (STDEV), range, mean, median and IQR (interquartile range) of DGI Test 2 data averaged for all normal subjects vs. patients.
Table 2. The standard deviation (STDEV), range, mean, median and IQR (interquartile range) of DGI Test 2 data averaged for all normal subjects vs. patients.
STDEV-DGI Test 2
Unit: in Voltage (V)Gyro ZGyro XGyro YACC ZACC XACC Y
Normal0.07830.02020.03690.00510.00740.0076
Patients0.05820.05700.04240.03750.03250.0270
Range-DGI Test 2
Normal0.41300.15130.18550.03020.04100.0312
Patients0.55590.22300.26230.21120.18250.1861
Mean-DGI Test 2
Normal1.60482.21491.27852.92672.20802.2218
Patients1.19071.85701.27952.01531.85512.4705
Median-DGI Test 2
Normal1.59262.27081.30762.97282.23252.2445
Patients1.18691.85281.28132.01461.85612.4686
IQR-DGI Test 2
Normal0.02580.09760.01950.02930.00480.0053
Patients0.03360.04840.02050.02550.02180.0260
Table 3. The STDEV, range, mean, median and IQR of DGI Test 7 data averaged for all normal subjects vs. patients.
Table 3. The STDEV, range, mean, median and IQR of DGI Test 7 data averaged for all normal subjects vs. patients.
STDEV-DGI Test 7
Unit: in Voltage (V)Gyro ZGyro XGyro YACC ZACC XACC Y
Normal0.10880.06060.03950.01580.01670.0197
Patients0.08600.05920.05310.02800.02640.0244
Range-DGI Test 7
Normal0.72070.20110.20990.06830.06730.0820
Patients0.40810.22280.28150.16780.14030.1311
Mean-DGI Test 7
Normal1.63002.24141.34582.94702.21832.2318
Patients1.20411.82531.31041.99821.83842.2458
Median-DGI Test 7
Normal1.63592.21751.34472.94942.22082.2308
Patients1.19371.81581.31041.99711.83802.4559
IQR-DGI Test 7
Normal0.05850.05510.02730.01310.01360.0156
Patients0.04670.00430.04130.01410.01060.0121
Table 4. STDEV for three normal subjects and four patients for DGI Test 2.
Table 4. STDEV for three normal subjects and four patients for DGI Test 2.
STDEV of GYRO Z for DGI Test 2 (in Voltage)
Normal 1Normal 2Normal 3Patient1Patient 2Patient 3Patient 4
0.08020.07460.08020.10210.09400.01550.0212
Table 5. STDEV, range, mean, median and IQR of all DGI Tests 1–7 data averaged for all normal subjects vs. patients.
Table 5. STDEV, range, mean, median and IQR of all DGI Tests 1–7 data averaged for all normal subjects vs. patients.
STDEV-DGI Tests 1–7
Unit: in Voltage (V)Gyro ZGyro XGyro YACC ZACC XACC Y
Normal0.10330.03230.03720.01210.01380.0131
Patients0.08090.05320.04680.03770.03820.0355
Range-DGI Tests 1–7
Normal0.6250.15160.22920.05940.07390.0637
Patients0.38540.23620.26490.22360.20040.1813
Mean-DGI Tests 1–7
Normal1.62622.24221.32782.94732.21372.2264
Patients1.22001.84261.28841.99021.83452.4505
Median-DGI Tests 1–7
Normal1.62312.24561.33122.95542.21862.2308
Patients1.22621.84151.28961.98901.83332.4492
IQR-DGI Tests 1–7
Normal0.04980.02260.02260.00570.00710.0071
Patients0.04610.03910.02640.02240.02330.0222
Table 6. Comparison of classification algorithms using performance metrics when the range data are used as the input features.
Table 6. Comparison of classification algorithms using performance metrics when the range data are used as the input features.
ClassifierSensitivitySpecificityPrecisionNPVF-MeasureAccuracy
BP-ANN100%100%100%100%1100%
SVM95.2%100%100%96.5%0.9798%
KNN95.2%96.4%95.2%96.4%0.9596%
BDT95.2%92.8%91%96.3%0.9394%
Table 7. Overall classification accuracy of different algorithms using different input features.
Table 7. Overall classification accuracy of different algorithms using different input features.
FeaturesOverall Classification Accuracy (Patients vs. Normal Subjects)
ANNSVMKNNBDT
Range100%98%96%94%
STDEV92%94%89.8%85.7%
Range + STDEV96%98%94%94%
Table 8. Summary of each classifier’s speed when different input features are used.
Table 8. Summary of each classifier’s speed when different input features are used.
FeaturesOverall Classifier Speed (s)
ANNSVMKNNBDT
Range0.0611.781.041.12
STDEV0.0521.921.882.06
Range + STDEV0.972.801.762.31
Table 9. Our real-time WGAS gait classification system vs. the literature.
Table 9. Our real-time WGAS gait classification system vs. the literature.
ReferenceAlgorithms (*)Data AcquisitionOverall Accuracy
[22]SVMThree IMUs (Opal, APDM, Inc., Portland, OR, USA) featuring a tri-axial accelerometers and a tri-axial gyroscope90.5%
[23]ANN, SVM37 markers were placed on body traced by an infrared camera during walking on the two embedded force plates96.9%, 98.2%
[24]ANN, SVM2 markers on the shoe and traced by cameras during walking on a treadmill using the PEAK MOTUS motion analysis system75%, 83.3%
[25]ANN, SVMSynchronized PEAK 3D motion analysis system and a force platform during normal walking83.3%, 91.7%
[26]ANN, SVMGait video sequence captured by a static camera during normal walking98%, 98%
[27]SVM with PCAIntegrated a pressure sensor, a tilt angle sensor, three single-axis gyroscopes, one tri-axial accelerometer and a bend sensor inside a small module in a shoe with an RF transmitter98%
[28]ANN29 retro reflective markers placed on the body, with 3D marker trajectories captured with an 8-camera motion analysis system during normal walking89%
[29]ANN25 reflective markers placed on the body and data acquired from the Vicon Nexus 3D motion capture system95%
[30]SVMOne tri-axial ACC worn on the subject waist while walking98%
This workANN, SVMtri-axial gyroscopes and tri-axial accelerometers integrated on a PCB forming a WGAS with an MSP430 microcontroller and an RF transmitter embedded on the PCB100%, 98%
* SVM: support vector machine; ANN: artificial neural network; PCA: principle component analysis.
Table 10. Comparison of the classification algorithms using performance metrics when the range data are used as the input features (4 patients and 12 normal subjects).
Table 10. Comparison of the classification algorithms using performance metrics when the range data are used as the input features (4 patients and 12 normal subjects).
ClassifierSensitivitySpecificityAccuracy
BP-ANN100%100%100%
SVM98.80%96.30%98.21%
KNN96.42%96.42%96.42%
BDT96.42%89.28%94.60%

Share and Cite

MDPI and ACS Style

Nukala, B.T.; Nakano, T.; Rodriguez, A.; Tsay, J.; Lopez, J.; Nguyen, T.Q.; Zupancic, S.; Lie, D.Y.C. Real-Time Classification of Patients with Balance Disorders vs. Normal Subjects Using a Low-Cost Small Wireless Wearable Gait Sensor. Biosensors 2016, 6, 58. https://doi.org/10.3390/bios6040058

AMA Style

Nukala BT, Nakano T, Rodriguez A, Tsay J, Lopez J, Nguyen TQ, Zupancic S, Lie DYC. Real-Time Classification of Patients with Balance Disorders vs. Normal Subjects Using a Low-Cost Small Wireless Wearable Gait Sensor. Biosensors. 2016; 6(4):58. https://doi.org/10.3390/bios6040058

Chicago/Turabian Style

Nukala, Bhargava Teja, Taro Nakano, Amanda Rodriguez, Jerry Tsay, Jerry Lopez, Tam Q. Nguyen, Steven Zupancic, and Donald Y. C. Lie. 2016. "Real-Time Classification of Patients with Balance Disorders vs. Normal Subjects Using a Low-Cost Small Wireless Wearable Gait Sensor" Biosensors 6, no. 4: 58. https://doi.org/10.3390/bios6040058

APA Style

Nukala, B. T., Nakano, T., Rodriguez, A., Tsay, J., Lopez, J., Nguyen, T. Q., Zupancic, S., & Lie, D. Y. C. (2016). Real-Time Classification of Patients with Balance Disorders vs. Normal Subjects Using a Low-Cost Small Wireless Wearable Gait Sensor. Biosensors, 6(4), 58. https://doi.org/10.3390/bios6040058

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop