Next Article in Journal
Growth Stages Classification of Potato Crop Based on Analysis of Spectral Response and Variables Optimization
Previous Article in Journal
A Versatile Illumination System for Real-Time Terahertz Imaging
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Letter

A Novel Surface Electromyographic Signal-Based Hand Gesture Prediction Using a Recurrent Neural Network

School of Mechatronic Engineering and Automation, Shanghai University, Shanghai 200444, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(14), 3994; https://doi.org/10.3390/s20143994
Submission received: 30 May 2020 / Revised: 14 July 2020 / Accepted: 15 July 2020 / Published: 17 July 2020
(This article belongs to the Section Biomedical Sensors)

Abstract

:
Surface electromyographic signal (sEMG) is a kind of bioelectrical signal, which records the data of muscle activity intensity. Most sEMG-based hand gesture recognition, which uses machine learning as the classifier, depends on feature extraction of sEMG data. Recently, a deep leaning-based approach such as recurrent neural network (RNN) has provided a choice to automatically learn features from raw data. This paper presents a novel hand gesture prediction method by using an RNN model to learn from raw sEMG data and predict gestures. The sEMG signals of 21 short-term hand gestures of 13 subjects were recorded with a Myo armband, which is a non-intrusive, low cost, commercial portable device. At the start of the gesture, the trained model outputs an instantaneous prediction for the sEMG data. Experimental results showed that the more time steps of data that were known, the higher instantaneous prediction accuracy the proposed model gave. The predicted accuracy reached about 89.6% when the data of 40-time steps (200 ms) were used to predict hand gesture. This means that the gesture could be predicted with a delay of 200 ms after the hand starts to perform the gesture, instead of waiting for the end of the gesture.

1. Introduction

Hand gesture recognition is a promising human-computer interaction, which is widely discussed and studied in various areas. The capability of machines to recognize distinctive gesture characteristics can be harnessed in a wide variety of applications including the control of bionic hands [1,2], virtual game control [3], sign language translation [4], smart wheelchair [5], and intelligent robotics [6,7]. To this day, various sensors have been used to recognize hand gesture. Data glove that contains bending sensors and accelerometers are able to capture the rotation and movement of the hand and fingers [8], while it is not convenient and unnatural to wear a cumbersome glove in daily life. Cameras are also used to recognize hand motion [9], while it is sensitive to the use environment such as background texture, color, and lighting. Surface electromyographic signal (sEMG) is a useful non-intrusive technique for recording the electrical activity produced by muscles through surface sensors placed on the skin, which is a promising candidate for motion detection, gesture recognition and even gesture prediction [10,11,12].
In previous studies, a variety of features of sEMG signal were designed and extracted to classify hand gesture [13,14,15,16,17,18,19,20]. These features included mean absolute value, zero crossing, root mean square, power spectrum ratio, discrete wavelet transform and so on. Machine learning methods were used to classify these features, such as k-nearest neighbors, artificial neural network, gaussian mixture model, linear discriminant analysis, hidden Markov model, support vector machine and random forests. Although the promising performance of these feature-based methods have been shown, the complex process of feature extraction may result in the loss of useful information [21].
To solve the problem, researchers started to use deep learning-based approaches, such as convolutional neural network (CNN), to automatically learn features from a large amount of data. Ding [22] used a multiple-scale CNN structure to deduce information loss in feature extraction. Geng [23] proposed a deep CNN to classify instantaneous sEMG data images generated by a high-density electrodes matrix. Allard [6] transformed sEMG data to frequency spectrum images using Fast Fourier Transform, then used CNN to classify them. Wei [24] proposed a multi-channel CNN, and improved recognition accuracy by learning the correction between muscle and gesture. Shen [25] built a complex model that ensembled multiple CNN with a stacking method. Allard [26] combined CNN and transfer learning to decrease the data requirement of the training model for a new subject.
Recurrent neural network (RNN) is also a deep learning-based approach, which has been widely studied in the field of speech recognition and machine translation in recent years. Different from CNN which is good with hierarchical or spatial data and extracting unlabeled features, RNN is good at temporal or otherwise sequential data [27]. Simao [28] compared the performance between forward propagation neural network and RNN on the recognition of eight hand gestures. He [29] combined long short-term memory (LSTM) and multiple layer perceptron to learn features of static hand gesture. Nadia [30] used raw sEMG signal to recognize six gestures, which could adapt to new subjects. Furthermore, some researchers used the combination of CNN and RNN to recognize hand gesture. Wu [21] proposed a method to combine LSTM and CNN to recognize hand gesture. Xie [31] compared the performance of a single CNN model, single LSTM model and hybrid model to recognize different gestures. Hu [27] introduced an attention mechanism into a hybrid model of CNN and RNN to recognize gesture in five sEMG benchmark databases.
However, to our knowledge, most of the existing RNN-based methods [27,28,29,30,31,32] still focus on hand gesture recognition, which recognize hand gestures based on the whole temporal sEMG data, but make a prediction by RNN, which has already been used in other fields [33,34,35,36] and has not been studied. In this paper, we propose a novel method by using RNN to predict hand gesture. The trained RNN model could output instantaneous prediction for the sEMG data from the previous sampling time step. The sEMG signals of 21 hand gestures of 13 subjects were recoded to train the model. In the test part, when the movement of a gesture starts, the model outputs an instantaneous prediction at every sampling time step.
The main contributions of this paper are as follows:
  • A hand gesture dataset containing 21 short-term gestures of 13 subjects is recorded by the Myo armband, which is publicly available on the Github that also includes our code (https://github.com/ChauncyHe/HandGesturePrediction).
  • A novel RNN model to predict hand gesture is proposed, which is able to predict the gesture in the process of the gesture. When sEMG data points of 200 ms are used, which are generated after the motion start of the gesture, the accuracy could be about 89.6%.
The rest of this paper is organized as follows: Section 2 describes the sEMG sensors used in this work, the process of data acquisition, and the proposed method in detail. Section 3 presents the experimental results and analysis. Section 4 summarizes the paper.

2. Materials and Methods

2.1. sEMG Dataset

2.1.1. Recording Device

There are various commercial devices or sensors to record sEMG signals, such as Thalmic Myo armband, Otto Bock 13e200 electrodes, Cometa Wave Plus system, Ag-AgCl Duo-Trode electrodes and Delsys Trigno system. Considering the portability in a real application scenario, we selected one low-cost wireless Myo armband from Thalmic Labs (Figure 1) to record sEMG data, which has been used to recognize hand gestures in many studies [37]. The Myo armband consists of a low-consumption ARM Cortex-M4 120 MHZ microprocessor, 8 dry electrodes and 9-axis inertial measurement unit (IMU). Its 8 non-intrusive dry electrodes can record sEMG signals generated by muscles at a sampling frequency of 200 Hz with 8 bits of resolution for each channel.
As shown in Figure 2, the Myo armband wearing protocol is stipulated as follows: Firstly, the armband is located at a fixed position on the forearm of the right hand. The distance between the edge of the armband and the elbow joint is approximately the width of an index finger. Secondly, the wearing orientation is fixed. The first electrode is placed on the back of the forearm and aligned to the middle finger. Thirdly, the position and orientation of the armband is kept unchanged when recording data. Because the prediction model is subject-specified, the precisely same wearing position and orientation for all subjects is not necessary.

2.1.2. Hand Gestures

We designed 21 common gestures as shown in Figure 3. These gestures are mainly controlled by the traction of several major muscles on the forearm, and high visual similarity exists in some of these gestures. For example, grabbing a cylinder (A19) and grabbing a sphere (A18) are two similar behaviors. In addition, A09, A15, A14, A05 and A04 are five finger gestures corresponding to numbers 1–5, respectively. A20 is finished by bending five fingers into the shape of zero, which is also similar to a fist (A03). Most of these gestures are realized by the motion of the finger joints, while A16 and A17 are finished by the motion of the wrist joint. Relaxation gesture is regarded as the 21st gesture. Different from previous studies, all gestures are asked to be finished in the short term, which occurs within 2 s instead of maintaining more than 5 s.

2.1.3. Acquisition Protocol

We recruited a total of 13 healthy right-handed volunteers (including 8 males and 5 females, aged from 23 to 25). Each subject wore the Myo armband on the right forearm near the elbow joint in accordance with the aforementioned wearing protocol. Before formal recording of data, the subject was instructed on how to perform each kind of hand gesture until the subject was able to perform all gestures by themselves without difficulty. Real-time sEMG data were transmitted from the Myo armband to a personal computer by a wireless receiver. The data of 30 repetitions of 21 kinds of hand gestures was recorded for each subject. Although the sampling time of each repetition lasted 2 s, a complete hand gesture was suggested to be finished in 0.5–1.5 s. Muscle fatigue will be caused by continuous multiple repetitions [38]. Therefore, to alleviate the harmful impact of muscle fatigue, a 2-s break between repetitions and a 5-min break between different gestures were reserved. In Table 1, a detailed data acquisition configuration for all subjects is shown. For each subject, a total of 630 samples were recorded. A sample is comprised of sEMG signal X 400 × 8 and corresponding label Y 400 × 21 , where X is 8-channel sEMG data in 400-time steps and Y is the one-hot coding label of these time steps.

2.2. Methods

In this section, the details of the proposed method are described. It is worth noting that we do not perform any common data preprocessing on input sEMG signals such as filtering and rectification. That is to say, we will use raw sEMG signals to train the RNN model, which could give a category prediction result by using raw sEMG signals in the test stage.
Firstly, before training the RNN model, motion detection which detects the start and end of a hand gesture is used to label the gesture. Then, the structure of the recurrent neural network model is described. Finally, post-processing is performed to obtain the final prediction result for a sample from instantaneous predictions generated at every time step.

2.2.1. Motion Detection

In the process of data recording, since subjects are only required to finish a complete hand gesture within 2 s, the gesture start time step t s and the end time step t e for each sample are both unknown. Motion detection is a key process to label gestures. There are many methods including moving average algorithm [3], standard deviation [5] and spectrogram technology [39]. In this study, we used the standard deviation of multi-channel sEMG signals in the time domain to detect the gesture. A sliding window was used to extract the average standard deviation of 8 channels.
S 1 [ t , c ] = 1 w t w t ( X [ t , c ] 1 w t w t X [ t , c ] ) 2 ,
S 2 [ t ] = 1 N c c = 1 N c S 1 [ t , c ] .
where t = 1 , 2 , 3 , , 400 represents time steps, and c = 1 , 2 , 3 , 8 represents channels. w represents the width of the sliding window. X is the sEMG data of a sample, and N C represents the number of channels.
The sliding window is computing the standard deviation of the front w data points at time step t . To guarantee the consistency of array shape, zero-padding is used when t < w . Considering the sensitivity of each channel to gestures are different, an average operation on channels is performed to get a more robust detection (Figure 4).

2.2.2. Model Structure

The RNN model is established by using gated recurrent units (GRU), a variant of recurrent neural units, as shown in Figure 5. The input layer is 8 channels raw sEMG data X 400 × 8 of a gesture sample. Layer 1 consists of 50 GRU units and uses the tanh activation function, which memorizes the signal change in the time domain. Layer 2 is a fully connected layer with 200 general units, and uses Tanh as the activation function. Layer 3 is also a recurrent layer that is the same as layer 1. Layer 4 has 21 units and uses SoftMax as the activation function to output probabilities of multiple categories. L1 or L2 regularization is not used in this study, and dropout is also not necessary by our experiments. For two recurrent layers, the sequence output is used, which guarantee the model output is also a sequence.
Because of the adding of GRU units in layer 1 and layer 3, the model output contains the time information as well. Model output Y ^ 400 × 21 is the changing process of the instantaneous prediction result with respect to the increasing time steps of sEMG data. In detail, Y ^ [ t , : ] 21 is an instantaneous prediction result at timestep t , which is represented by a probability vector obtained by layer 4. Additionally, according to the time characteristic of a unidirectional recurrent neural network, the instantaneous prediction result Y ^ [ t , : ] is merely determined by sEMG data slice X [ 1 : t , : ] t × 21 .

2.2.3. Post-Processing

When the gesture start is detected at t s by motion detection, the model starts to output an instantaneous prediction result at every time step. In fact, when t = t s + M , a total of M instantaneous prediction results are obtained. To improve the real-time performance of gesture prediction, a small number of m ( m < M ) instantaneous results are used to make a decision in post-processing. To get a final prediction category label from these m instantaneous prediction results, a simple and effective approach is to merely consider the last instantaneous result at t = t s + m , which could be formed as the following:
L ^ = argmax   Y [ t s : t s + m , : ] .
where t s is the gesture start timestep of this sample, and m denotes how many time steps of sEMG that are used to predict. Then L ^ { 1 , 2 , , 21 } represents 21 different gestures. When m equals 40 timesteps, it means that the final prediction category label could be obtained with a delay of 200 ms after the hand starts to perform the gesture, instead of waiting for the end of the gesture.

2.3. Training and Test Details

Considering the relatively small number of samples for training a deep learning-based model with raw sEMG signals, stratified 5-folds cross-validation is used to partition datasets and evaluate model performance. Stratified partition guarantees enough sample numbers to train each category, and prevents resulting in an unbalanced model because of common random partitions. An average test accuracy on these 5 models is regarded as the generalization ability of the current model structure and hyper-parameters. Training details and test details are shown in Figure 6.
In the training stage, gesture start time step t s and gesture end time step t e of all training samples are detected to label the dataset by motion detection, and the RNN model is trained on the labelled dataset. We selected Adam optimizer to update model weights and used the cross entropy loss function to measure model error. The initial learning rate was set to 0.01 and a learning rate reducing mechanism was used.
In the testing stage, when the gesture start is detected by motion detection, the model makes use of m time steps of record data after the gesture starts to predict hand gesture. As shown in Figure 6, record data obtained from motion detection are sent to the trained model after data filling which fills zero into the record data to make the input data X 400 × 8 . Prediction results are obtained from the model output after post processing.
After getting prediction category labels of every sample in the testing set by post-processing, we could obtain the evaluation performance of the model as usual classification tasks. Prediction accuracy of the model is determined by:
A C C = i = 1 N S I ( L ^ i = L i ) N S
where Ns is the number of test samples, final category prediction result is represented by L ^ { 0 , 1 , , 20 } , L i is the true category label of sample i, I ( p ) equals to 1 when p is true otherwise it equals to 0.

3. Results and Analysis

3.1. Accuracy Performance

Prediction accuracy of each subject is shown in Figure 7. The result shows that prediction accuracy increases with m , which conforms to the intuition that more known data points of gesture leads to a more reliable result. The average accuracy of all subjects on some major values of m are recorded in Table 2. When m = 40 time steps ( T m = 200   ms ), the average prediction accuracy on testing set reached 89.6%.
To figure out the prediction performance of the model on each kind of hand gesture, the confusion matrix is plotted in Figure 8, on the condition that m = 40 time steps. In order to more objectively express the model prediction performance on these 21 types of gestures, the confusion matrix is the hybrid of all subjects. From this figure, we found that several similar finger motion gestures representing numbers such as A04 (five), A05 (four), A14 (three) and A15 (two) were more likely to be confused with each other. On the contrary, A16 (wave in) and A17 (wave out) that only used the wrist joint, were relatively easy to predict.

3.2. Real-Time Performance

For a hand gesture, the time T a = t e t s represents the actual time span of a complete gesture. The average T a on all samples of a subject varies between 463 ms and 1582 ms as shown in Figure 9. When m = 40 time steps, corresponding T m = 200   ms . From the figure, we find that T m     T a , which means the model will give prediction output after the gesture has started 200 ms before the gesture ends.

3.3. The Feasibility of Prediction

According to [37], a gesture contains transient state and steady state, and transient state is generated when the gesture is in motion, while steady state is produced when the gesture is maintained. Classification of hand gestures in the transient state has lower accuracy than in the steady state. To explore whether the data in the transient state of a gesture is able to be used to predict gesture, Tp is defined as the time of transient state and a comparison between T a , T p and T m of a subject is illustrated in Figure 10. When T m = 200   ms , the accuracy of the model on the test set reached 89.6%. We can see for some drastic gestures such as A03 (fist) and A16 (wave in), transient time Tp was longer than that of other gestures. Additionally, for these gestures, sEMG data only in Tp were used to predict hand gestures. The results show that it is feasible to obtain a high prediction performance by using our RNN method.

3.4. Comparison with Other Methods

In this study, the proposed method is used to predict hand gestures and the real-time performance, which is more important. Table 3 shows the real-time performance and other conditions of the proposed model and other previous studies which used RNN methods. Most of these models were not real-time models, especially those methods which were a combination of LSTM and CNN. Nasri [30] made a very high validation accuracy on six gestures, but each gesture away repeated about 200 times for a subject, which is an extremely heavy workload. In addition, the use of a sliding window of 940 ms severely decreases the model’s real-time performance. He [21] used a sliding window of 400 ms but the accuracy of gestures was 75.5%. Compared to these studies, our method could obtain a real-time performance with a good accuracy.

3.5. Limitations

We regret that, although the Myo armband is a popular low-cost device widely used in relevant studies in the past, it is unfortunately not commercially available since 2018 [39]. To our knowledge, there are some sEMG devices which use dry electrodes like the Myo armband, including Delsys Trigno system, gForce-100 Armband [40], etc.
In addition, more concerns have to be taken into consideration in the application of real-time hand gesture prediction. In data acquisition, the wearing position and orientation of the Myo armband is fixed for each subject, so the trained subject-specified model is sensitive to the subject and wearing situation of the Myo armband in the test stage. Generally, the robustness of the model needs to be enhanced in further work.

4. Conclusions

This paper proposes a hand gesture prediction method based on raw multi-channel sEMG signals. Firstly, a dataset containing 30 repetitions of 21 short-term hand gestures of 13 subjects was collected with a Myo armband, in which each complete gesture was finished within 0.5–1.5 s. Then, a RNN model was built, which output an instantaneous probability distribution at every time step after the start of the gesture was detected. Finally, the prediction result was acquired by post-processing from these instantaneous outputs. Experimental results show that prediction accuracy by our model could reach about 89.6% using data of 200 ms, which can be collected after the start of the gesture.

Author Contributions

Conceptualization, Z.Z.; methodology, Z.Z.; software, C.H.; validation, C.H., K.Y. and Z.Z.; formal analysis, C.H.; investigation, C.H.; resources, K.Y.; data curation, K.Y.; writing—original draft preparation, C.H.; writing—review and editing, Z.Z.; visualization, Z.Z.; supervision, Z.Z.; project administration, Z.Z.; funding acquisition, Z.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under grant number 91748119 and Shanghai Science and Technology Commission fgrant number 18JC1410402.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Huang, Y.; Englehart, K.; Hudgins, B. A Gaussian Mixture Model Based Classification Scheme for Myoelectric Control of Powered Upper Limb Prostheses. IEEE Trans. Biomed. Eng. 2005, 52, 1801–1811. [Google Scholar] [CrossRef]
  2. Zhang, Z.; Yang, X. Bio-inspired motion planning for reaching movement of a manipulator based on intrinsic tau jerk guidance. Adv. Manuf. 2019, 7, 315–325. [Google Scholar] [CrossRef]
  3. Zhang, X.; Chen, X.; Li, Y. A framework for hand gesture recognition based on accelerometer and emg sensors. IEEE Trans. Syst. Man Cybern. Part A Syst. Humans 2011, 41, 1064–1076. [Google Scholar] [CrossRef]
  4. Yang, X.; Chen, X.; Cao, X.; Wei, S.; Zhang, X. Chinese sign language recognition based on an optimized tree-structure framework. IEEE J. Biomed. Health Inf. 2016, 21, 994–1004. [Google Scholar] [CrossRef] [PubMed]
  5. Kundu, A.S.; Mazumder, O.; Lenka, P.K.; Bhaumik, H. Hand Gesture Recognition Based Omnidirectional Wheelchair Control Using Imu and Emg Sensors. J. Intell. Rob. Syst. 2018, 91, 529–541. [Google Scholar] [CrossRef]
  6. Allard, U.C.; Nougarou, F.; Fall, C.L.; Giguère, P.; Gosselin, C.; Laviolette, F.; Gosselin, B. A convolutional neural network for robotic arm guidance using sEMG based frequency-features. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Daejeon, Korea, 9–14 October 2016; pp. 2464–2470. [Google Scholar]
  7. Xiao, S.; Liu, S.; Wang, H.; Lin, Y.; Song, M.; Zhang, H. Nonlinear dynamics of coupling rub-impact of double translational joints with subsidence considering the flexibility of piston rod. Nonlinear Dyn. 2020, 100, 1203–1229. [Google Scholar] [CrossRef]
  8. Fang, G.; Gao, W.; Zhao, D. Large Vocabulary Sign Language Recognition Based on Fuzzy Decision Trees. IEEE Trans. Syst. Man Cybern. Part A Syst. Humans 2004, 34, 305–314. [Google Scholar] [CrossRef]
  9. Starner, T.; Weaver, J.; Pentland, A. Real-time American Sign Language Recognition Using Desk and Wearable Computer Based Video. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 1371–1375. [Google Scholar] [CrossRef]
  10. Zhang, Z.; Yu, X.; Qian, J. Classification of Finger Movements for Prosthesis Control with Surface Electromyography. Sens. Mater. 2020, 32, 1523–1532. [Google Scholar]
  11. Xu, Z.; Shen, L.; Qian, J.; Zhang, Z. Advanced Hand Gesture Prediction Robust to Electrode Shift with an Arbitrary Angle. Sensors 2020, 20, 1113. [Google Scholar] [CrossRef] [Green Version]
  12. Zhang, Z.; Yang, K. Real-time Surface Emg Pattern Recognition for Hand Gestures Based on an Artificial Neural Network. Sensors 2019, 19, 3170. [Google Scholar] [CrossRef] [Green Version]
  13. Kong, D.; Zhu, J.; Duan, C.; Lu, L.; Chen, D. Bayesian linear regression for surface roughness prediction. Mech. Syst. Sig. Process. 2020, 142, 106770. [Google Scholar] [CrossRef]
  14. Al-Angari, H.; Kanitz, G.; Tarantino, S.; Cipriani, C. Distance and mutual information methods for emg feature and channel subset selection for classification of hand movements. Biomed. Signal Process. Control 2016, 27, 24–31. [Google Scholar] [CrossRef]
  15. Benatti, S.; Milosevic, B.; Farella, E.; Gruppioni, E.; Benini, L. A Prosthetic Hand Body Area Controller Based on Efficient Pattern Recognition Control Strategies. Sensors 2017, 17, 869. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Kakoty, N.M.; Hazarika, S.M.; Gan, J.Q. EMG Feature Set Selection Through Linear Relationship for Grasp Recognition. J. Med. Biol. Eng. 2016, 36, 883–890. [Google Scholar] [CrossRef]
  17. Shi, W.T.; Lyu, Z.J.; Tang, S.T.; Chia, T.; Yang, C. A bionic hand controlled by hand gesture recognition based on surface EMG signals: A preliminary study. Biocybern. Biomed. Eng. 2018, 38, 126–135. [Google Scholar] [CrossRef]
  18. Yang, C.; Long, J.; Urbin, M.A.; Feng, Y.; Song, G.; Weng, J.; Li, Z. Real-Time Myocontrol of a Human–Computer Interface by Paretic Muscles After Stroke. IEEE T. Cogn. Dev. Syst. 2018, 10, 1126–1132. [Google Scholar] [CrossRef]
  19. Dardas, N.H.; Georganas, N.D. Real-time hand gesture detection and recognition using bag-of-features and support vector machine techniques. IEEE Trans. Instrum. Meas. 2011, 60, 3592–3607. [Google Scholar] [CrossRef]
  20. Li, Z.; Guan, X.; Zou, K.; Xu, C. Estimation of Knee Movement From Surface EMG Using Random Forest with Principal Component Analysis. Electronics 2020, 9, 43. [Google Scholar] [CrossRef] [Green Version]
  21. Wu, Y.; Zheng, B.; Zhao, Y. Dynamic Gesture Recognition Based on LSTM-CNN. In Proceedings of the 2018 Chinese Automation Congress, Xi’an, China, 30 November–2 December 2018; pp. 2446–2450. [Google Scholar]
  22. Ding, Z.; Yang, C.; Tian, Z.; Yi, C.; Fu, Y.; Jiang, F. sEMG-based Gesture Recognition with Convolution Neural Networks. Sustainability 2018, 10, 1865. [Google Scholar] [CrossRef] [Green Version]
  23. Geng, W.; Du, Y.; Jin, W.; Wei, W.; Hu, Y.; Li, J. Gesture Recognition By Instantaneous Surface Emg Images. Sci. Rep. 2016, 6, 36571. [Google Scholar] [CrossRef]
  24. Wei, W.; Wong, Y.; Du, Y.; Hu, Y.; Kankanhalli, M.S.; Geng, W. A Multi-stream Convolutional Neural Network for sEMG-based Gesture Recognition in Muscle-computer Interface. Pattern Recognit. Lett. 2019, 119, 131–138. [Google Scholar] [CrossRef]
  25. Shen, S.; Gu, K.; Chen, X.; Yang, M.; Wang, R. Movements Classification of Multi-channel sEMG Based on CNN and Stacking Ensemble Learning. IEEE Access 2019, 7, 137489–137500. [Google Scholar] [CrossRef]
  26. Coteallard, U.; Fall, C.L.; Drouin, A.; Campeaulecours, A.; Gosselin, C.; Glette, K.; Laviolette, F.; Gosselin, B. Deep learning for electromyographic hand gesture signal classification using transfer learning. IEEE Trans. Neural Syst. Rehabil. Eng. 2019, 27, 760–771. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Hu, Y.; Wong, Y.; Wei, W.; Du, Y.; Kankanhalli, M.S.; Geng, W. A novel attention-based hybrid CNN-RNN architecture for sEMG-based gesture recognition. PLoS One 2018, 13, 10. [Google Scholar] [CrossRef] [Green Version]
  28. Simao, M.A.; Neto, P.; Gibaru, O. EMG-based Online Classification of Gestures with Recurrent Neural Networks. Pattern Recognit. Lett. 2019, 128, 45–51. [Google Scholar] [CrossRef]
  29. He, Y.; Fukuda, O.; Bu, N.; Okumura, H.; Yamaguchi, N. Surface EMG Pattern Recognition Using Long Short-Term Memory Combined with Multilayer Perceptron. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 18–21 July 2018; pp. 5636–5639. [Google Scholar]
  30. Nasri, N.; Orts-Escolano, S.; Gomez-Donoso, F.; Cazorla, M. Inferring Static Hand Poses from a Low-Cost Non-Intrusive sEMG Sensor. Sensors 2019, 19, 371. [Google Scholar] [CrossRef] [Green Version]
  31. Xie, B.; Li, B.; Harland, A.R. Movement and Gesture Recognition Using Deep Learning and Wearable-sensor Technology. In Proceedings of the 2018 International Conference on Artificial Intelligence and Pattern Recognition, Beijing, China, 18–20 August 2018; pp. 26–31. [Google Scholar]
  32. Ali, S. Gated Recurrent Neural Networks for EMG-based Hand Gesture Classification. A Comparative Study. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 18–21 July 2018; pp. 1094–1097. [Google Scholar]
  33. Tuncer, T.; Dogan, S.; Subasi, A. Surface EMG signal classification using ternary pattern and discrete wavelet transform based feature extraction for hand movement recognition. Biomed. Signal Process. Control 2020, 58, 101872. [Google Scholar] [CrossRef]
  34. Wang, C.; Guo, W.; Zhang, H.; Guo, L.; Huang, C.; Lin, C. sEMG-based continuous estimation of grasp movements by long-short term memory network. Biomed. Signal Process. Control 2020, 59, 101774. [Google Scholar] [CrossRef]
  35. Zhou, Y.; Huang, Y.; Pang, J.; Wang, K. Remaining useful life prediction for supercapacitor based on long short-term memory neural network. J. Power Sources 2019, 440, 227149. [Google Scholar] [CrossRef]
  36. Jaramillo-Yánez, A.; Benalcázar, M.E.; Mena-Maldonado, E. Real-Time Hand Gesture Recognition Using Surface Electromyography and Machine Learning: A Systematic Literature Review. Sensors 2020, 20, 2467. [Google Scholar] [CrossRef] [PubMed]
  37. Atzori, M.; Gijsberts, A.; Castellini, C.; Caputo, B.; Hager, A.M.; Elsig, S.; Giatsidis, G.; Bassetto, F.; Muller, H. Electromyography Data for Non-invasive Naturally-controlled Robotic Hand Prostheses. Sci. Data 2014, 1, 1–13. [Google Scholar]
  38. Benalcazar, M.E.; Motoche, C.; Zea, J.A.; Jaramillo, A.G.; Anchundia, C.E.; Zambrano, P.; Segura, M.; Benalcazar, P.F.; Perez, M. Real-time hand gesture recognition using the Myo armband and muscle activity detection. In Proceedings of the 2017 IEEE 2nd Ecuador Technical Chapters Meeting, Salinas, Ecuador, 16–20 October 2017; pp. 1–6. [Google Scholar]
  39. Welcome to Myo Support. Available online: https://support.getmyo.com/hc/en-us (accessed on 15 July 2020).
  40. gForce-100 Gesture Armband. Available online: http://oymotion.com/en/product32/150 (accessed on 15 July 2020).
Figure 1. Myo armband.
Figure 1. Myo armband.
Sensors 20 03994 g001
Figure 2. Wearing position and orientation.
Figure 2. Wearing position and orientation.
Sensors 20 03994 g002
Figure 3. A01–A20 are 20 different hand gestures, and R represents the relaxation gesture. All these gestures (except R) are rapidly finished by going through a process of R-A-R within 2 s.
Figure 3. A01–A20 are 20 different hand gestures, and R represents the relaxation gesture. All these gestures (except R) are rapidly finished by going through a process of R-A-R within 2 s.
Sensors 20 03994 g003aSensors 20 03994 g003b
Figure 4. The curves of surface electromyographic (sEMG) signals and the motion detection result.
Figure 4. The curves of surface electromyographic (sEMG) signals and the motion detection result.
Sensors 20 03994 g004
Figure 5. The model structure of recurrent neural network (RNN).
Figure 5. The model structure of recurrent neural network (RNN).
Sensors 20 03994 g005
Figure 6. Flow diagram of the training stage. MD: motion detection. DF: data filling. (a) Training data. (b) Modify label of sample. (c) Test data. (d) Data after motion detection. (e) Prepare input data by DF. (f) The output of the RNN model.
Figure 6. Flow diagram of the training stage. MD: motion detection. DF: data filling. (a) Training data. (b) Modify label of sample. (c) Test data. (d) Data after motion detection. (e) Prepare input data by DF. (f) The output of the RNN model.
Sensors 20 03994 g006
Figure 7. Accuracy with respect to m on 13 subjects. The horizontal axis m denotes how many time steps of sEMG data are used to predict gesture. Additionally, the vertical axis denotes the average accuracy of 5-fold cross validation.
Figure 7. Accuracy with respect to m on 13 subjects. The horizontal axis m denotes how many time steps of sEMG data are used to predict gesture. Additionally, the vertical axis denotes the average accuracy of 5-fold cross validation.
Sensors 20 03994 g007
Figure 8. Hybrid test confusion matrix of all subjects.
Figure 8. Hybrid test confusion matrix of all subjects.
Sensors 20 03994 g008
Figure 9. The motion time T a of each subject and T m = 200   ms .
Figure 9. The motion time T a of each subject and T m = 200   ms .
Sensors 20 03994 g009
Figure 10. The comparison between T a , T p and T m of a subject.
Figure 10. The comparison between T a , T p and T m of a subject.
Sensors 20 03994 g010
Table 1. Acquisition configuration summary for the collection of datasets.
Table 1. Acquisition configuration summary for the collection of datasets.
Acquisition DeviceMyo ArmbandGestures21
Sampling frequency200 HzRepetition times30
Channel number8Sampling time of a repetition2 s
Subject number13Finish time of a gesture0.5–1.5 s
Age range of subjects22–26Repetition interval 2 s
Health stateIntact subjectsGesture interval 5 min
Table 2. Prediction accuracy with different m .
Table 2. Prediction accuracy with different m .
m (Time Steps)20406080
T m (ms)100200300400
Accuracy83.8 ± 7.5%89.6 ± 5.5%91.5 ± 4.7%90.8 ± 5.4%
Table 3. Comparison with state-of-the-art using RNN.
Table 3. Comparison with state-of-the-art using RNN.
WorkChannelsDeviceSampling Rate (Hz)GesturesRepetitionGesture Duration (s)SubjectsClassifierAccuracy (%)RTP (ms)
Hu [27]10Myo(8+2)1005210527LCNN87.0NRT
Xie [31]16Myo(8+8)200176510LCNN83.6NRT
Wu [21]8Myo(8)2005125NILCNN98.0NRT
Simao [28]16Myo(8+8)2008NININILSTM/GRU95.0NRT
Samadani [32]12Myo(8+4)100186540LSTM89.5NRT
Nasri [30]8Myo(8)20061951035GRU99.8940
He [29]12Myo(8+4)1005210527LSTM75.5400
Ours8Myo(8)2002130213GRU89.6200
Note: RTP represents real-time performance, which usually can be denoted as the width of the sliding window. NI denotes the corresponding term is not indicated in the paper clearly. LCNN is the combination of LSTM and CNN. NRT means the method is not a real time approach.

Share and Cite

MDPI and ACS Style

Zhang, Z.; He, C.; Yang, K. A Novel Surface Electromyographic Signal-Based Hand Gesture Prediction Using a Recurrent Neural Network. Sensors 2020, 20, 3994. https://doi.org/10.3390/s20143994

AMA Style

Zhang Z, He C, Yang K. A Novel Surface Electromyographic Signal-Based Hand Gesture Prediction Using a Recurrent Neural Network. Sensors. 2020; 20(14):3994. https://doi.org/10.3390/s20143994

Chicago/Turabian Style

Zhang, Zhen, Changxin He, and Kuo Yang. 2020. "A Novel Surface Electromyographic Signal-Based Hand Gesture Prediction Using a Recurrent Neural Network" Sensors 20, no. 14: 3994. https://doi.org/10.3390/s20143994

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop