Next Article in Journal
Triple Immunochromatographic System for Simultaneous Serodiagnosis of Bovine Brucellosis, Tuberculosis, and Leukemia
Next Article in Special Issue
Complexity Measures of Voice Recordings as a Discriminative Tool for Parkinson’s Disease
Previous Article in Journal
Influence of Graphene Oxide Concentration when Fabricating an Electrochemical Biosensor for DNA Detection
Previous Article in Special Issue
EMG-Based Characterization of Walking Asymmetry in Children with Mild Hemiplegic Cerebral Palsy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Label Self-Advised Support Vector Machine (LSA-SVM)—Automated Classification of Foot Drop Rehabilitation Case Study

by
Sahar Adil Abboud
1,2,†,
Saba Al-Wais
3,†,
Salma Hameedi Abdullah
2,†,
Fady Alnajjar
4,*,† and
Adel Al-Jumaily
2,*,†
1
Computer Engineering Department, University of Technology, Baghdad, Iraq
2
Faculty of Engineering and Information Technology, University of Technology, Broadway, Sydney NSW 2007, Australia
3
Bio-Medical Engineering Department, University of Technology, Baghdad, Iraq
4
Department of Computer Science and Software Engineering, College of Information Technology, UAE University, Al Ain 15551, UAE
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Biosensors 2019, 9(4), 114; https://doi.org/10.3390/bios9040114
Submission received: 14 July 2019 / Revised: 19 August 2019 / Accepted: 2 September 2019 / Published: 27 September 2019

Abstract

:
Stroke represents a major health problem in our society. One of the effects of stroke is foot drop. Foot drop (FD) is a weakness that occurs in specific muscles in the ankle and foot such as the anterior tibialis, gastrocnemius, plantaris and soleus muscles. Foot flexion and extension are normally generated by lower motor neurons (LMN). The affected muscles impact the ankle and foot in both downward and upward motions. One possible solution for FD is to investigate the movement based on the bio signal (myoelectric signal) of the muscles. Bio signal control systems like electromyography (EMG) are used for rehabilitation devices that include foot drop. One of these systems is function electrical stimulation (FES). This paper proposes new methods and algorithms to develop the performance of myoelectric pattern recognition (M-PR), to improve automated rehabilitation devices, to test these methodologies in offline and real-time experimental datasets. Label classifying is a predictive data mining application with multiple applications in the world, including automatic labeling of resources such as videos, music, images and texts. We combine the label classification method with the self-advised support vector machine (SA-SVM) to create an adapted and altered label classification method, named the label self-advised support vector machine (LSA-SVM). For the experimental data, we collected data from foot drop patients using the sEMG device, in the Metro Rehabilitation Hospital in Sydney, Australia using Ethical Approval (UTS HREC NO. ETH15-0152). The experimental results for the EMG dataset and benchmark datasets exhibit its benefits. Furthermore, the experimental results on UCI datasets indicate that LSA-SVM achieves the best performance when working together with SA-SVM and SVM. This paper describes the state-of-the-art procedures for M-PR and studies all the conceivable structures.

1. Introduction

Stroke represents a major health problem in today’s society. One of the consequences of stroke is foot drop. Foot drop is a weakness that appears in specific muscles in the ankle and foot such as: the anterior tibialis (AT), gastrocnemius (Gas.), plantaris and soleus muscles. Foot flexion and extension are normally generated by lower motor neuron (LMN). Lesions in the lower motor neuron (LMN) will cause foot drop. Foot drop manifests in 52% to 67% of patients with spinal upper motor neuron (UMN) [1]. Foot Drop (FD) is a common disorder without specifying to age and affects around 1% of women and 2.8% of men [2]. Bio signal control systems like electromyography (EMG) are used devices that target leg rehabilitation. These devices target various leg impairments, including foot drop. Electromyography (EMG), which records myoelectric signals from muscle activity, has been widely used to detect the user’s intended action [3]. The EMG electrodes are placed in the subject’s limb, either in an invasive or non-invasive way. A wide range of people prefer not to implant electrodes inside the body and instead prefer to use a surface EMG (sEMG). However, the sEMG has numerous disadvantages such as crosstalk from other muscles and robustness. Furthermore, it is hard to acquire a myoelectric signal from deeper muscles, so it is difficult to deal with myoelectric signal processing using EMG. Myoelectric pattern recognition (M-PR) methods can also be applied to artificial intelligent applications. This system contains multiple steps—first the sEMG data is filtered from the noisy signal, then it is applied to extract and reduce features, to minimize the large sample data. In addition, methodologies such as support vector machine (SVM), cSA-SVM, vSA-SVM and label classifications are studied in this paper. One of the popular machine learning methodologies is the support vector machine (SVM) method, which is used to classify data. Vapnik [3,4] proposed the support vector machine as an influential classification method, various forms of SVM were presented in the literature and applied in several different applications. SVM categorizes class problems into two; classes and multi-classes. The SVM method gives an ideal decision for the formation of a boundary between two or multiple classes. The margin created, separates the classes and the decision boundary is maximized. To theorize the binary classification by a training set of N samples, a vector for the input data should be considered, which depends on the ith sample. This is labeled related to its class. The purpose of SVM is to separate the binary labeled training data within the hyperplane that has a maximum distance from them. This is called a maximum margin hyperplane [3,5]. The standard SVM disregards the train data that is not separated linearly by the kernels through the training stage, which occurs during the outline of the tolerance parameters in the impartial function and restrictions. For this reason, it will be classified incorrectly if the data are similar or match the misclassified data that appears in the test set. This happens when the data that are close to the misclassified data are unspecified. This results in a misclassification that is not sensible and not controlled [6]. A non-repeating, self-advising method for SVM was adapted [7], which extracts consequent knowledge from the training phase without the addition of extra parameters. The misclassified data is supplied from two prospective sources. First, from outliers and second, from the data that has not been separated correctly [8]. Many researchers have adapted versions of SVM with the goal of raising the classification efficiency and performance for some applications. Use of label classification on experienced knowledge is one of the methods that has been used without increasing the cost since there is no addition of an extra parameter. For example, they proposed novel methodologies for texture analysis to improve the single-label classification of facial features [9,10]. Masood, A. et al. [6] suggested to enhance label classification with class classification techniques. In addition, they dealt with the problem of limited labeled data available, especially for histopathological images. They proposed a novel learning model, created on a deep belief neural network and semi-advised SVM to make effective use of labeled data along with unlabeled data for the training phase. It displayed improved performance when matched with different state-of-the-art approaches for skin cancer diagnosis. The proposed model was used for diagnosing skin cancer. Multi-label classifying is a predictive data mining application with multiple applications in the world, including automatic labeling of resources, including videos, music, images and texts. The multi-label data can be used as a learning tool and can be achieved by different methods, such as: the problem transformation method which has two common methods, the adaptation method, and ensembles of classifiers [3]. Several applications include research work with the use of label classification methods (LCM) to improve the progression of the search for related information on Twitter. Five different labels are defined to categorize tweets, including news. The system was proposed to analyze complex motion in events; it combines the tracking and multilabel hypergraphs of moving targets in video sequences [11]. The adaptation methodologies state that some classification models were primarily intended for resolving binary problems and then were expanded to solve multi-class problems. By contrast, other methods can easily work with several classes. In this case study, the novel recognition system, that is an integration of label classification methods with SA-SVM is used to get LSA-SVM for two-class classification, to overcome the problems and improve the reliability of the diagnosis process. It is important to develop computational tools for automated diagnosis that operates on a quantitative measure. Such tools can facilitate objective mathematical judgment complementary to that of medical experts and help them to identify the affected areas more efficiently with more accurate diagnosis and less wastage of time in treatment while trying not to lose the bounder for working in real time standard.

2. Related Work

2.1. The Standard SVM

The basic idea of the SVM, which simplifies the pair ( w , b ) is described in the hyperplane with the equation < w , x > + b = 0 . SVM can utilize to produce a non-linear decision function, by projecting the training data to a higher dimensional internal product space, known as feature space, by applying a non-linear map ϕ ( x ) : R n R d .
Although the optimal linear hyper-plane calculates in the feature space, by applying kernels it is capable to make necessary processes in the input space using k ( x i , x j ) = < ϕ ( x i ) , ϕ ( x j ) > which is an internal product in the feature space. In terms of these kernels, the decision function can be written as the following Equation (1):
f ( x ) = s i g n ( Σ α i > 0 y i α i k ( x , x i ) + b )
The decision value for each X of the test sets have either a negative or a positive value, depends on the situation of X and the hyperplane that has been clarified as Equation (2):
h ( x ) = Σ α i > 0 y i α i k ( x , x i ) + b
There are three mutual kernel functions in SVM—Radial Basis Function kernel (RBF), Polynomial kernel and Sigmoid kernel. This paper has used the RBF kernel as Equation (3):
k ( x i , x j ) = e γ | x i x j | 2

2.2. Self-Advised Support Vector Machine (SA-SVM)

Advised Weighted-SVM treats the neglect of SVM from the information that can be obtained from misclassified data. By the creation of advised weights, which depends on the distance between the misclassified train data and the classified train data. In addition, applying these weights together with decision values of SVM in the test phase, assist the procedure to reduce the outlier data [12,13]. The details of the Self-Advised SVM procedure is demonstrated in the following steps:
  • Classify the hyper-plane founded by applying the decision function is Equation (4):
    f ( x ) = s i g n ( Σ α I > 0 y i α i k ( x , x i ) + b )
  • Misclassified data that samples in the first train phase are recognized. The misclassified datasets (MD) in the training phase are calculated by Equation (5):
    M D = i = 1 N X i | y i s i g n ( Σ α j > 0 y i α j k ( x , x j ) + b )
    The MD set may be null but the empirical outcomes appear when the presence of misclassified data in the training phase is a communal existence. It should be recognized that trying any technique to benefit from misclassified data should have a control to affect the outlier data. When the misclassified data is included to resemble samples, the use of misclassified data improved the classification accuracy [14].
  • The algorithm indicates: If MD is null then go to the testing phase or else compute neighborhood length (NL) for each Xi of MD. Equation (6), defined NL.
    N L ( x i ) = m i n i m u m x i ( x i x j | y i y j )
    where X j = 1 , , N is the training data that does not belong to the MD set. If the training data is a map with a higher dimension, the distance between xi and xj can be evaluated in Equation (7) with reference to the related RBF kernel
    θ ( x i ) θ ( x j ) = ( k ( x i , x i ) + k ( x j , x j ) 2 k ( x i , x j ) ) 0.5
  • Calculating Advised Weight A W ( x k ) for each sample x k from the test set using Equation (8). These A W s represent the closest test data to the misclassified data.
    0 x i M D , x k x i > N L ( x i ) o r M D = N U , Σ 1 Σ x i x k x i Σ x i N L ( x i ) x i M D , x k x i N L ( x i )
  • The absolute value of the SVM decision values for each x k from the test set are considered and scaled in to [0, 1].
  • Finally, for each xk from the test set in Equation (9):
    If A W ( x k ) < decision value ( x k ) then:
    y k = s i g n ( Σ α j > 0 y j α j k ( x k , x j ) + b )
    which is identified with normal SVM, otherwise:
    y k = y i | ( x k x j ( x i ) , x i M D )

2.3. Label Classification

Label Classification (LC) contains two types of classifying—Single Label Classification (SLC) and Multi-Label Classification (MLC)—which are the supervised learning problems where sample data are connected to single or multiple labels. Applications that use the SLC and MLC have increased in different fields. For example, text classification, scene and video classification, bioinformatics and Biomedical Text Data [15]. Generally, Binary Relevance (BR) is a method used for Multi-Label Classification (MLC). Label Classification LC deliberates every label as an independent binary problem and its work depends on the lack of appearance for the non-direct modeling label correlations. Most of the existing methods contribute to complexity of model inter-dependencies between the labels. Another method used in Multi-Label Classification is to perform problem transformation, where a multi-label problem transformed into one or more single-labels like binary or multi-class problems. That activates single-label classifiers and is transformed by their single-label predictions into multi-label predictions. Problem transformation is used to describe both flexibility and scalability. They apply Support Vector Machines, Naive Bayes, k Nearest Neighbor methods and Perceptron [16]. Where X d R the input sample domain, the sample feature formed as a vector d. The sample input forms a vector of d feature X = [ x 1 , , x d ] , while the l output domain is L = 1 , , L . Each sample ( x ) related to a subset of these labels, which forms as L vector y = [ y 1 , , y L ] . Where y 1 = 1 , if and only if label j is related with sample x and 0 otherwise. They assume a set of training data D of N labeled patterns as D = ( x i , y i ) | i = 1 , , N . So that the researcher can write multi-label accuracy equation for a set of N test as follow in Equation (10):
A c c u r a c y = 1 N Σ i = 1 N | y i y ^ i | | y i y ^ i |
Ensemble Classifier Chains (ECC) signifies a vector of absolute outputs W ^ = [ w ^ 1 , , w ^ L ] R L , while W ^ j signifying the absolute for the j t h label. For prediction vectors y ^ 1 , , y ^ m from repeating 1 , , m , The absolute equation evaluate as in Equation (11):
W ^ j = 1 m Σ k = 1 m y ^ i k
Also, threshold function f apply w ^ j to get a bipartition of appropriate and inappropriate labels: y ^ = f t ( w ^ ) Softmax functions also offer single-label classification. Practically, it applies to the multi-label scenario by problem transformation. Softmax loss function can modify the multilabel scenario as shown in Equation (12) [17]:
l s o f t m a x = Σ y Y i log exp ( f y ( x i ) ) Σ j y x p ( f y ( x i ) )

3. Materials and Methods

3.1. Materials

The Myoelectric Pattern Recognition system consists of two main parts—the software and hardware that are further elaborated in the phases of the real-time M-PR system, as visible in Figure 1 and Figure 2. Overall, all phases of real-time pattern recognition based sEMG is displayed in Table 1. For the experimental stages, the system collects the data from healthy and unhealthy subjects. The collected data is then applied to train the system, resulting in the output of classification; the trained classifier, the OpenSim prediction and simulation for gait level.
The collected EMG signals were processed on a Personal Computer Intel Core i7, 2.8 GHz with 16 GB RAM and equipped with a Windows 10 operating system. A band-pass filter was used for filtering the signals in the frequency band (25–550 Hz). A notch filter was applied to remove the 50 Hz line nosiness. The EMG signals were downsized to 1000 Hz to minimize the size. To estimate the recommended mSA-SVM, LSA-SVM and ELM-LSA-SVM, the experiment embraced 13 datasets from the UCI machine learning repository [18,19] These databases were nominated from the most public benchmarks for classification and diagnosis. The diversity of these databases supports the authentication in this study. The number of instances and the attributes of each database shown in the tables in each chapter was used. It should be noted that for a dataset with multi-class we used 6 datasets while for two classes 7 datasets [20].

3.2. Procedure for Collecting sEMG Signal Data

The data wre collected from the hospital, based on the design procedure that received ethical approval to collect data from FD patients in Metro Rehabilitation Hospital in Sydney, Australia using Ethical Approval (UTS HREC NO. ETH15-0152). For the experiment, data were collected from 13 subjects. These 13 subjects involved in the offline experiment; consisted of 6 females and 7 males, aged between 18 to 84 years; the average age being 51 years. Ten of them were affected with Foot Drop. The other three were healthy with no muscle disorder. During the experiment the subjects were seated, so that the knee is in a fixed position as shown in Figure 3, to avoid the influence of position movements on EMG signals. A few digital filters were applied during data collection. The filters applied were a band pass filter between 25 and 550 Hz and a notch filter to remove the 50 Hz line noise. The EMG signals were reduced to 1080 Hz.
The classification was dependent on the sample data from which it was acquired. The data were collected under the Supervision of the Stroke Coordinator at the Metro-Rehab Hospital. The data was collected in three phases at each posture (flexion or extension) of the knee joint. The healthy and unhealthy subjects, went through the following:
  • First trial: To move (with help, if required) his/her lower limb at the knee joint. Flexion and extension (bend and straighten) from the actual position while sitting on a chair. Each set of trials took 3 s and a 5 s rest was given between two trials (two sets of trials were done for this case) as shown in Figure 4 and Figure 5.
  • Second trial: To move (with help, if required) his/her foot up and down as much as possible from rest position while sitting on the chair. Each set of trials took 3 s and a 5 s rest was given between two trials. (Five sets of trials were done for this case) as shown in Figure 6 and Figure 7.
  • Third trial: To move (with help, if required) his/her lower limb (foot and leg) at the knee joint, to flex or extend (bend and straighten) with Extension Plantar flexion and Flexion Dorsiflexion from the rest position while sitting on the chair. Each set of trials took 3 s and a 5 s rest was given between two trials. (Two sets of trials were done for this case) as shown in Figure 8 and Figure 9.
Data for 12 s was collected for each subject’s trial and data for 156 s were collected for all repetitions by the subject. Three-fold cross-validation was conducted with the offline classification. To measure the Dorsal Plantar Flexion range for the leg, we used the Goniometric Measurement such as Protractor (Angle Finder and Bevel Square Head) as shown in Figure 10. Table 2 presents the characteristics for sick and healthy subjects collected from Metro Rehabilitation Hospital.
For the Metro-Hospital dataset, we collected the Surface electromyography (EMG) signals from 13 subjects, from Rectus Femurs (RF), Gastrocnemius (Gas), Soule (Sol) and Tibias Anterior (TA). The OpenSim dataset for CG, provides Surface electromyography (EMG) signals which recorded from ten subjects. sEMG collects signals from the Medial Hamstrings (mH), Biceps Femurs long head (BF), Rectus Femurs (RF), Gastrocnemius (Gas) and Tibias Anterior (TA). Each trial was 5 s with 2 repetitions, which makes 10 s in total. Three trials were done for each subject = 30 s 1000 in each class = 30,000 and there were 4 classes, so data = 120,000/channel. There were 4 channels, which makes the entire sample data = 480,000. The data collected were divided into training data and test data using 3-fold cross-validation.

3.3. Method: Label Self-Advised Support Vector Machine (LSA-SVM)

The following concepts describe the LSA-SVM, where the misclassified data was based on calculating the neighborhood length using the label of data instead of the value for single classification method. This procedure would minimize the time process and large data can be processed.
  • Applying the decision function as in Equation (10), to classify hyperplane:
    f ( x ) = s i g n ( Σ α i > 0 y i α i k ( x , x i ) + b )
    where x i is the input vector for i t h sample labeled with y i related to its class, while α i is the non-negative Lagrange multiplier, which conflicts with standard SVM training.
  • Misclassified data samples in the first train phase are recognized. The misclassified data sets (MD) in the training phase are calculated by Equation (11):
    M D = i = 1 N X i | y i s i g n ( Σ α j > 0 y i α i k ( x i , x j ) + b )
  • The algorithm indicates that: If MD is empty, then go to the testing phase, or else calculate neighborhood length (NL) for each yi of label MD. Equation (12) defined NL.
    N L ( y i ) = m i n i m u m y i ( y i y j | x i j )
    where y j , j = 1 , , N is the label of training data that do not belong to the label of MD set. The label of the training data is mapped to a higher dimension, the distance between y o u and y j is computed according to the following Equations (13) and (14) with reference to the related RBF kernel.
    ϕ ( y i ) ϕ ( y j ) = ( k ( y i , y i ) + k ( y j , y j ) 2 k k ( y i , y j ) ) 0.5
    that is, RBF will be:
    k ( y i , y j ) = e γ | y i y j | 2
  • For each label y k from test data, the Lab Advised Weight LAW ( y k ) figures out as Equation (15). These LAWs represent how close the label test data are to the label of misclassified data.
    L A W ( y k ) = 0 , x i M D , y k y i > N L ( y i ) o r M D = N U L , Σ 1 Σ y i y k y i Σ y i N L ( y i ) , x i M D y k y i N L ( y i )
    The absolute value of the SVM decision values for each xk from the test set is calculated and scaled to [ 0 , 1 ] .
  • For each y k from the label of the test set,
    If ( L A W ( y k ) < decision value ( y k ) then y k = s i g n ( Σ α j > 0 y i α j k ( x k , x i ) + b ) which is compatible with normal SVM labeling, otherwise y k = y i | ( y k y i N L ( y i ) a n d x i M D ) . Figure 11, explain the flow chart for the steps above.

4. Experiments and Results

The experiments check the performance of LSA-SVM for a single class as conducted through the state of the art pattern recognition system shown in Figure 12. This Flowchart describes each of the methods that were applied to get the output; classifying subjects as healthy and unhealthy using the novel.

4.1. Experiments on Hospital Datasets

Various experiments were done to test the performance of LSA-SVM in the Myoelectric pattern recognition. First, we adjusted the parameters of C and g from the range of (2-9, 2-8, …, 29,210), as shown in Table 3, then we examined the performance of accuracy while we apply v-SVM, c-SVM, v-SA-SVM, c-SA-SVM and LSA-SVM. We compare all the classifiers. Some analysis is given in each experiment; in the first experiment we used the first data set FD and the second data set CG to estimate accuracy for each classifier. The second experiment applies the third datasets from UCI.
Figure 13 and Table 4 specify that the accuracy of LSA-SVM was higher than SA-SVM and other classifiers in all types of dataset. The average accuracy for vLSA-SVM as compared with all other five classifiers is a little higher—equal to 99.06%—while in vSA-SVM it was equal to 98.75% for the FD Hospital dataset. However, for the CG OpenSim dataset vLSA-SVM gives 82.01% and vSA-SVM equal to 80.01% with five-fold cross-validation training.
Table 5 and Figure 14 show that the training times of vSA-SVM in two groups of datasets are much faster than those of vLSA-SVM. In that case, the LSA-SVM method did not achieve the best performance, whereas time consumption was equal to 69.4 msec for vLSA-SVM and 46.3 msec, for vSA-SVM but it still reached a real time that is less than 300msec as a standard real time measurement [22,23]. Overall, in most cases, the adaptation of the Label function used in LSA-SVM can improve the performance of the classical support vector machine and SA- SVM.

4.2. Experiments on UCI Datasets

LSA-SVM proved that it is capable of classifying four classes of leg movements consisting of a combination of three classes of the unhealthy leg (Mild, Moderate and Severe patients) and the fourth is the healthy class. This section investigates the performance of LSA-SVM on the benchmark dataset that is accessible online on the UCI machine learning website. The experiment depends on the size of the data. We implement 3-fold cross-validation on the larger sized data, while we executed 5-fold cross validation for the small and medium-sized data. Table 6 shows the data specification for a benchmark for 2-Classes dataset types.
This experiment involved six classifiers—v-SVM, c-SVM, c-SA-SVM, v-SA-SVM, c-LSA- SVM, v-LSA-SVM. The optimal parameters are established and noted its effects on the accuracy and time performance of the classifier. Table 7 provides all parameters that were utilized in this experiment.
Table 8 and Figure 15 show that LSA-SVM had performed reasonably through a seven different data sets with 2 classes. The comparison of LSA-SVM and SA-SVM shows that average accuracy are quite similar in some dataset. As for LSA-SVM, the accuracy of v- LSA-SVM is significantly better than c-LSA-SVM only in “Breast Cancer” equal to 92.09 dataset. From the results, recognized that LSA-SVM is the most accurate classifier across seven datasets equal to 96.42 in “Australian Credit (Statelog)” except “Pima Indians diabetes” and “Spambase” datasets that is equal to 83.31% and 62.52% respectively.
The processing time (time consumption) of classifiers also calculated. Table 9 provides the training time. Figure 16 clarifies that the training time of LSA-SVM is one of the slowest classifiers, equated to other classifiers in overall datasets. Its performance worsens when vLSA-SVM works on big data like the “Skin Segmentation” dataset equal to 133.1 msec. The cLSA-SVM is the slowest classifier, taking around 267 ms to learn “Skin Segmentation” datasets while vSA-SVM shows the faster time, equal to 0.537 msec.

5. Conclusions

LSA-SVM has an advantage over SA-SVM as it works on label data instead of data value. In addition to the Myoelectric leg motion classification, LSA-SVM has been applied to a wide collection of classification problems using a UCI machine learning dataset. The experimental results show that LSA-SVM performs on a wide range of dataset sizes (small to large). Overall, LSA-SVM is a promising classifier for several classification applications, particularly for Myoelectric pattern recognition. In the training test, which was executed to compare the results of LSA-SVM with all other algorithms, the p-value correlated to the ANOVA test by a level of impact α = 0.05 was 0.038, which shows good statistical difference between these groups. Therefore, it can be concluded that LSA-SVM achieved better results than these algorithms. In addition, 68.80 percent sensitivity and 76.5 percent specificity was achieved to classify the testing data for the Hospital. Label Self-Advised Support Vector Machine (LSA-SVM) was implemented and projected the Self-Advised Support Vector Machine (SA-SVM) for leg motion recognition using sEMG signals. Overall, LSA-SVM could classify four leg movements with an accuracy of 99.06 percent, deeming it comparable with renowned classifiers such as SA-SVM, SVM. Therefore, LSA-SVM could improve the performance of the advised-based SVM.
This study presented a new Label classification method, called the Label Self-Advised-Support Vector Machine LSA-SVM, to diagnose leg movements for foot drop patients. Data were collected using a surface electromyography (sEMG) device from Foot Drop Patients from the Metro Rehabilitation Hospital in Sydney, Australia using Ethical Approval (UTS HREC NO. ETH15-0152). Also, the experimental results for the sEMG dataset, UCI and OpenSem benchmark datasets prove its assistance.

Author Contributions

Conceptualization, All Authors; methodology, S.A.A.; software, S.A.A.; validation, S.A.A., S.A.-W. and S.H.A.; formal analysis, All Authors; investigation, S.A.A.; resources, S.A.A. and S.A.-W.; data curation, S.A.A.; writing–original draft preparation, S.A.A. and S.A.-W.; writing–review and editing, All Authors; visualization, S.A.A.; supervision, A.A.-J. and F.A.; project administration, A.A.-J. and F.A.; funding acquisition, A.A.-J. and F.A.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
LAWLab Advised Weight
AWAdvised Weight
SVMSupport Vector Machine
FDFoot drop
SA-SVMSelf Advised Support Vector Machine
LMNLower Motor Neuron
FESFunction Electrical Stimulation
M-PRMyoelectric Pattern Recognition
SA-SVMSelf-Advised-Support Vector Machine
LSA-SVMLabel Self-Advised-Support Vector Machine
LCMlabel classification methods
RBFRadial Basis Function
MDMisclassify Datasets
EMGElectromyography
sEMGsurface EMG
TATibias Anterior
RFRectus Femurs
GasGastrocnemius
SolSoule
TATibias Anterior
CGCrouch Gait
LDALinear discriminant analysis
TD-ARTiming Domain-Autoregressive
SCCsingle class classification
SLCSingle Label Classification
MDPIMultidisciplinary Digital Publishing Institute
DOAJDirectory of open access journals
TLAThree letter acronym
LDlinear dichroism

References

  1. Westhout, F.D.; Paré, L.S.; Linskey, M.E. Central causes of foot drop: Rare and underappreciated differential diagnoses. J. Spinal Cord Med. 2007, 30, 62–66. [Google Scholar] [CrossRef] [PubMed]
  2. Hiam, D.S. The Gale Encyclopedia of Neurological Disorders; Gale: Farmington Hills, MI, USA, 2017. [Google Scholar]
  3. Gastounioti, A.; Makrodimitris, S.; Golemati, S.; Kadoglou, N.P.; Liapis, C.D.; Nikita, K.S. A novel computerized tool to stratify risk in carotid atherosclerosis using kinematic features of the arterial wall. IEEE J. Biomed. Health Inform. 2015, 19, 1137–1145. [Google Scholar] [PubMed]
  4. Vapnik, V.; Vapnik, V. Statistical Learning Theory; Wiley: New York, NY, USA, 1998; pp. 156–160. [Google Scholar]
  5. Wang, Z.; Xue, X. Multi-Class Support Vector Machine. In Support Vector Machines Applications; Ma, Y., Guo, G., Eds.; Springer International Publishing: Cham, Switzerland, 2014; pp. 23–48. [Google Scholar]
  6. Masood, A.; Al-Jumaily, A. SA-SVM based automated diagnostic system for skin cancer. Proc. SPIE 2015, 9443, 94432L. [Google Scholar]
  7. Maali, Y.; Al-Jumaily, A. Self-advising support vector machine. Knowl.-Based Syst. 2013, 52, 214–222. [Google Scholar] [CrossRef]
  8. Masood, A.; Al-Jumaily, A.; Anam, K. Texture analysis based automated decision support system for classification of skin cancer using SA-SVM. In Proceedings of the International Conference on Neural Information Processing, Kuching, Malaysia, 3–6 November 2014; Springer: Cham, Switzerland, 2014; pp. 101–109. [Google Scholar]
  9. Mohammed, A.A.; Sajjanhar, A. Robust single-label classification of facial attributes. In Proceedings of the 2017 IEEE International Conference on Multimedia & Expo Workshops (ICMEW), Hong Kong, China, 10–14 July 2017; pp. 651–656. [Google Scholar]
  10. Xu, J.W.; Suzuki, K. Max-AUC feature selection in computer-aided detection of polyps in CT colonography. IEEE J. Biomed. Health Inform. 2014, 18, 585–593. [Google Scholar] [PubMed]
  11. Herrera, F.; Charte, F.; Rivera, A.J.; Del Jesus, M.J. Multilabel classification. In Multilabel Classification; Springer: Cham, Switzerland, 2016; pp. 17–31. [Google Scholar]
  12. Masood, A. Developing Improved Algorithms for Detection and Analysis of Skin Cancer. Ph.D. Thesis, University of Technology Sydney, Ultimo, Australia, July 2016. [Google Scholar]
  13. Anam, K.; Al Jumaily, A.; Maali, Y. Index Finger Motion Recognition using self-advise support vector machine. Int. J. Smart Sens. Intell. Syst. 2014, 7, 644–657. [Google Scholar] [CrossRef]
  14. Masood, A.; Al-Jumaily, A.; Anam, K. Self-supervised learning model for skin cancer diagnosis. In Proceedings of the 7th International IEEE/EMBS Conference on Neural Engineering (NER), Montpellier, France, 22–24 April 2015; pp. 1012–1015. [Google Scholar]
  15. Kolesov, A.; Kamyshenkov, D.; Litovchenko, M.; Smekalova, E.; Golovizin, A.; Zhavoronkov, A. On multilabel classification methods of incompletely labeled biomedical text data. Comput. Math. Methods Med. 2014, 2014, 781807. [Google Scholar] [CrossRef] [PubMed]
  16. Read, J.; Pfahringer, B.; Holmes, G.; Frank, E. Classifier chains for multi-label classification. Mach. Learn. 2011, 85, 333. [Google Scholar] [CrossRef]
  17. Reynolds, J.S.; Goldsmith, W.T.; Day, J.B.; Abaza, A.A.; Mahmoud, A.M.; Afshari, A.A.; Barkley, J.B.; Petsonk, E.L.; Kashon, M.L.; Frazer, D.G. Classification of voluntary cough airflow patterns for prediction of abnormal spirometry. IEEE J. Biomed. Health Inform. 2016, 20, 963–969. [Google Scholar] [CrossRef] [PubMed]
  18. Sakai, H.; Liu, C.; Nakata, M. Information Dilution: Granule-Based Information Hiding in Table Data—A Case of Lenses Data Set in UCI Machine Learning Repository. In Proceedings of the 2016 Third International Conference on Computing Measurement Control and Sensor Network (CMCSN), Matsue, Japan, 20–22 May 2016; pp. 52–55. [Google Scholar]
  19. Dua, D.; Graff, C. UCI Machine Learning Repository; University of California: Irvine, CA, USA, 2017. [Google Scholar]
  20. Ashok, P.; Nawaz, G.K. Detecting outliers on UCI repository datasets by Adaptive Rough Fuzzy clustering method. In Proceedings of the 2016 Online International Conference on Green Engineering and Technologies (IC-GET), Coimbatore, India, 19 November 2016; pp. 1–6. [Google Scholar]
  21. Trejo, R.L.; Vázquez, J.P.G.; Ramirez, M.L.G.; Corral, L.E.V.; Marquez, I.R. Hand goniometric measurements using leap motion. In Proceedings of the 2017 14th IEEE Annual Consumer Communications & Networking Conference (CCNC), Las Vegas, NV, USA, 8–11 January 2017; pp. 137–141. [Google Scholar]
  22. Osojnik, A.; Panov, P.; Džeroski, S. Multi-label classification via multi-target regression on data streams. Mach. Learn. 2017, 106, 745–770. [Google Scholar] [CrossRef]
  23. Ren, D.; Ma, L.; Zhang, Y.; Sunderraman, R.; Fox, P.T.; Laird, A.R.; Turner, J.A.; Turner, M.D. Online biomedical publication classification using multi-instance multi-label algorithms with feature reduction. In Proceedings of the IEEE 14th International Conference on Cognitive Informatics & Cognitive Computing (ICCI*CC), Beijing, China, 6–8 July 2015; pp. 234–241. [Google Scholar]
Figure 1. Stages for an Offline myoelectric pattern recognition system.
Figure 1. Stages for an Offline myoelectric pattern recognition system.
Biosensors 09 00114 g001
Figure 2. The hardware and software needed for the off-line application.
Figure 2. The hardware and software needed for the off-line application.
Biosensors 09 00114 g002
Figure 3. Knee Flexion with Flexion Dorsiflexion.
Figure 3. Knee Flexion with Flexion Dorsiflexion.
Biosensors 09 00114 g003
Figure 4. Collect data signals from the leg using sEMG device.
Figure 4. Collect data signals from the leg using sEMG device.
Biosensors 09 00114 g004
Figure 5. Knee Extension with Flexion Dorsiflexion.
Figure 5. Knee Extension with Flexion Dorsiflexion.
Biosensors 09 00114 g005
Figure 6. Knee Flexion with Flexion Dorsiflexion.
Figure 6. Knee Flexion with Flexion Dorsiflexion.
Biosensors 09 00114 g006
Figure 7. Knee Flexion with Flexion Dorsiflexion.
Figure 7. Knee Flexion with Flexion Dorsiflexion.
Biosensors 09 00114 g007
Figure 8. Knee Flexion with Extension Planter Dorsiflexion.
Figure 8. Knee Flexion with Extension Planter Dorsiflexion.
Biosensors 09 00114 g008
Figure 9. Knee Extension with Extension Plantarflexion.
Figure 9. Knee Extension with Extension Plantarflexion.
Biosensors 09 00114 g009
Figure 10. Measure Dorsal Plantar Flexion range for the leg [21].
Figure 10. Measure Dorsal Plantar Flexion range for the leg [21].
Biosensors 09 00114 g010
Figure 11. Flowchart of the LSA-SVM methods steps.
Figure 11. Flowchart of the LSA-SVM methods steps.
Biosensors 09 00114 g011
Figure 12. Flowcharts for each stage of M-PR.
Figure 12. Flowcharts for each stage of M-PR.
Biosensors 09 00114 g012
Figure 13. The average classification accuracy of LSA-SVM across two datasets using five-fold cross-validation compared with other classifiers.
Figure 13. The average classification accuracy of LSA-SVM across two datasets using five-fold cross-validation compared with other classifiers.
Biosensors 09 00114 g013
Figure 14. The training time consumption for each type of classifier.
Figure 14. The training time consumption for each type of classifier.
Biosensors 09 00114 g014
Figure 15. The accuracy of seven classifiers on different data using 5-fold cross validation for small and medium-size data and 3-fold cross validation for large size data for 2 classes.
Figure 15. The accuracy of seven classifiers on different data using 5-fold cross validation for small and medium-size data and 3-fold cross validation for large size data for 2 classes.
Biosensors 09 00114 g015
Figure 16. The training time consumption for each type of classifier.
Figure 16. The training time consumption for each type of classifier.
Biosensors 09 00114 g016
Table 1. The hardware and software needed for the off-line application.
Table 1. The hardware and software needed for the off-line application.
ComponentDescriptionPicture
HardwarePersonal Computer Intel
HardwareEMG acquisition device, the FlexComp Infiniti™ System from Thought Technology with frequency sampling 2000 Hz Biosensors 09 00114 i001
HardwareFour EMG sensors: MyoScan™ T9503M Sensors from Though technology Biosensors 09 00114 i002
HardwareFour electrodes Biosensors 09 00114 i003
HardwareOneProCom Infiniti USB Adapter—TT-USB Biosensors 09 00114 i004
HardwareOne small piece of Fiber Optic Cable 15ft.—SA9480 Biosensors 09 00114 i005
HardwareOneProCom Infiniti USB Adapter—TT-USB Biosensors 09 00114 i006
SoftwareMatlab R2016b
SoftwareAPI library from Though Technology connecting the Flexcomp to Matlab
Table 2. Characteristics for sick and healthy subjects collected for the hospital.
Table 2. Characteristics for sick and healthy subjects collected for the hospital.
GenderAge (Years)Height (cm)Weight (kg)Min KFA(deg)Speed (m/s)T (Months)BPL300
F45155.754.91548Yes
F5216054.72218Yes
M611761081815No
M641621021536No
M8414778.23000.1751Yes
F6816589.4453No
M8217270383No
M22134.647500.45424Yes
M6815364.75036Yes
F60154.5816530Yes
M601656501.20 No
F451607101.02 No
F181636401.17 No
Table 3. Optimal parameters for all classifiers.
Table 3. Optimal parameters for all classifiers.
DatasetC-SVMv-SVMFeat TypeWin. Size, Win. Inc.
C γ C γ
Hospital1000.0031000.0031450, 15
crouch gait1000.0031000.0031420, 5
Table 4. The average classification accuracy of LSA-SVMacross two datasets using five-fold cross-validation compared with other classifiers.
Table 4. The average classification accuracy of LSA-SVMacross two datasets using five-fold cross-validation compared with other classifiers.
DatasetAccuracy (%)
SVM
C
SVM
v
SA-SVM
C
SA-SVM
v
LSA-SVM
C
LSA-SVM
v
Hospital92.1893.2698.8698.7598.9999.06
crouch gait76.2070.1081.7380.0181.9782.01
Table 5. The training time consumption for each type of classifier.
Table 5. The training time consumption for each type of classifier.
DatasetTraining Time (ms)
SVM
C
SVM
v
SA-SVM
C
SA-SVM
v
LSA-SVM
C
LSA-SVM
v
Hospital55.354.3107.946.370.769.4
crouch gait48.555.136.338.048.948.4
Table 6. Data specification for benchmarking for 2-Classes datasets.
Table 6. Data specification for benchmarking for 2-Classes datasets.
DatasetGroup# Data# Attributes
IrisSmall size1004
Parkinson19522
Australian Credit Approval (Statlog)Medium size6894
Breast Cancer6999
Pima Indians diabetes7688
SpambaseLarge size460157
Skin Segmentation245,0573
Table 7. The optimal parameters used by each classifier in the UCI [19] dataset experiments for 2-classes.
Table 7. The optimal parameters used by each classifier in the UCI [19] dataset experiments for 2-classes.
DatasetC-SVMv-SVMFeat TypeWin. Size, Win. Inc.
C γ C γ
Iris1110.251120, 5
Parkinson3000.12530000.0031450, 15
Australian Credit Approval (Statlog)0.50.1252140.0031420, 5
Breast Cancer82000.0032140.0031350, 7
Spambase30000.00320500.00514500, 120
Pima Indians diabetes82000.007820500.0021450, 15
Skin Segmentation1000.0031214500, 150
Table 8. The accuracy of seven classifiers on various data using 5-fold cross validation for small and medium-size data and 3-fold cross validation for large size data for 2 classes.
Table 8. The accuracy of seven classifiers on various data using 5-fold cross validation for small and medium-size data and 3-fold cross validation for large size data for 2 classes.
DatasetAccuracy (%)
SVM
C
SVM
v
SA-SVM
C
SA-SVM
v
LSA-SVM
C
LSA-SVM
v
Iris92.8590.1285.7185.8492.8592.86
Parkinson72.3270.071.4364.2971.4374.10
Australian Credit (Statelog)81.680.888.0788.5391.2896.42
Breast Cancer89.8888.7690.9690.3392.0991.01
Pima Indians diabetes82.58078.477.580.9578.3
Spambase61.1762.5261.1862.5362.1961.26
Skin Segmentation83.3172.2660.5067.4784.2384.43
Table 9. The training time consumption for each type of classifier.
Table 9. The training time consumption for each type of classifier.
DatasetTraining Time (ms)
SA-SVM
C
SA-SVM
v
LSA-SVM
C
LSA-SVM
v
Iris0.8370.7013.64.1
Australian Credit (Statelog)1.101.604.65.2
Breast Cancer3.203.37.208.70
Pima Indians diabetes0.9010.8365.04.7
Spambase25.816.134.833.4
Skin Segmentation121.486.3267.2133.1

Share and Cite

MDPI and ACS Style

Adil Abboud, S.; Al-Wais, S.; Abdullah, S.H.; Alnajjar, F.; Al-Jumaily, A. Label Self-Advised Support Vector Machine (LSA-SVM)—Automated Classification of Foot Drop Rehabilitation Case Study. Biosensors 2019, 9, 114. https://doi.org/10.3390/bios9040114

AMA Style

Adil Abboud S, Al-Wais S, Abdullah SH, Alnajjar F, Al-Jumaily A. Label Self-Advised Support Vector Machine (LSA-SVM)—Automated Classification of Foot Drop Rehabilitation Case Study. Biosensors. 2019; 9(4):114. https://doi.org/10.3390/bios9040114

Chicago/Turabian Style

Adil Abboud, Sahar, Saba Al-Wais, Salma Hameedi Abdullah, Fady Alnajjar, and Adel Al-Jumaily. 2019. "Label Self-Advised Support Vector Machine (LSA-SVM)—Automated Classification of Foot Drop Rehabilitation Case Study" Biosensors 9, no. 4: 114. https://doi.org/10.3390/bios9040114

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop