Next Article in Journal
Theoretical Analysis and Experiments on the Sound Absorption Properties of Foam Sound Absorbers with Thin Membranes Naturally Present in Foams Using Nano-Computed Tomography Scan Images
Previous Article in Journal
Prediction Method of Available Nitrogen in Red Soil Based on BWO-CNN-LSTM
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Parallel Network for Continuous Motion Estimation of Finger Joint Angles with Surface Electromyographic Signals

School of Information Science and Technology, Dalian Maritime University, Dalian 116026, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(20), 11078; https://doi.org/10.3390/app152011078
Submission received: 4 April 2024 / Revised: 30 April 2024 / Accepted: 7 May 2024 / Published: 16 October 2025

Abstract

The implementation of surface electromyographic (sEMG) signals in the interaction between human beings and machines is an important line of research. In the system of human–machine interaction, continuous-motion-estimation-based control plays an important role because it is more natural and intuitive than pattern recognition-based control. In this paper, we propose a parallel network consisting of a CNN with a multi-head attention mechanism and a BiLSTM (bidirectional long short-term memory) network to improve the accuracy of continuous motion estimation. The proposed network is evaluated in the Ninapro dataset. Six finger movements of 10 subjects were tested in the Ninapro DB2 dataset to evaluate the performance of the neural network and calculate the PCC (Pearson Correlation Coefficient) between the predicted joint angle sequence and the actual joint angle sequence. The experimental results show that the average accuracy (PCC) of the proposed network reaches 0.87 ± 0.02, which is significantly better than that of the BiLSTM network (0.79 ± 0.04, p < 0.05), CNN-Attention (0.80 ± 0.01, p < 0.05), CNN (0.70 ± 0.03, p < 0.05), CNN-BiLSTM (0.83 ± 0.02, p < 0.05), and TCN (0.76 ± 0.05, p < 0.05). It is worth noting that in this work, we extract multiple features from the raw sEMG signals and fuse them. We found that better continuous estimation accuracy can be achieved using multi-feature sEMG data. The model proposed in this paper skillfully integrates the convolutional neural network, multi-head attention mechanism, and bidirectional long short-term memory network, and its performance has good stability and accuracy. The model realizes more natural and accurate human–computer interaction.

1. Introduction

Electromyography (EMG) is used to record the electrical signals produced when muscles contract [1]. The EMG reflects the motor intention of the subject and contains a large amount of temporal and spatial information. Due to its low acquisition cost and easy acquisition, surface electromyography is widely used in gesture recognition [2], robot control [3], clinical diagnosis, sports science [4], and other fields. Surface electromyography is an important direction, providing a more natural method for human–computer collaboration.
Methods commonly used to decode human motion intentions include electromyography, video, inertial measurement unit (IMU), etc. Video-based techniques are vulnerable to environmental variables like lighting, noise, and camera placement, and require substantial computational resources. Inertial measurement unit (IMU)-based approaches are capable of accurately estimating joint angles. However, a large time delay is an unavoidable disadvantage of this method. EMG signals can give information about motor intention 50–100 milliseconds before the movement occurs [5]. Therefore, methods based on EMG signals can predict and estimate joint angles more naturally and fluently.
Overall, surface-EMG-based recognition tasks can be categorized as either classifying or regressing tasks. Classification tasks mainly include gesture recognition [6,7], body movement classification, etc. The regression task revolves around predicting continuous motion using EMG data. With the development of artificial intelligence, pure motion classification can no longer meet the requirements of human–computer interaction. Accurate continuous motion estimation plays a crucial role in realizing more natural human–computer interaction.
For EMG-based continuous motion prediction, both model-based and model-free approaches can be used [8]. Model-based methods estimate human motion parameters by building kinematic, dynamic, or musculoskeletal models. In this class of methods, Liu et al. [9] investigated a nonlinear EMG signal–torque model and an EMG signal amplitude estimation processor to estimate human joint torque. Wang et al. [10] studied and used the Hill-type model to estimate muscle force, and then obtained the moment arm through the musculoskeletal geometric model to achieve the purpose of predicting joint torque. The limitation of this type of method is that it requires more prior knowledge about human prostheses, and the number of parameters is large and difficult to estimate. Methods based on machine learning or deep learning, that is, model-free methods, do not include complex feature extraction or feature selection processes, which alleviate the requirements for kinematic modeling and are often better than model-based methods in terms of accuracy. For example, Chen et al. [11] proposed an LSTM-based upper limb motion model for multi-joint upper limb motion experiments. Gautam et al. [12] presented a deep learning model using a combination of CNNs and LSTMs, which can simultaneously classify the movement of the lower extremities and predict the corresponding knee angles. Chen et al. [13] used CNN, RNN, and traditional machine learning methods to continuously estimate the force of fingers with multiple degrees of freedom. Empirical findings indicate that the approach reliant upon neural networks is superior to the conventional technique. Ma et al. [14] aimed to predict joint angles with ten degrees of freedom and presented a novel method for surface electromyographic signal feature extraction for use in deep learning models. The proposed approach yields highly precise joint angle predictions. Zhang et al. [15] developed an adaptive radial basis function neural network (RBF NN)-based dynamics learning scheme to achieve the accurate identification of dominant dynamics of a robot under different failure modes. And physical experiments were conducted on a soft trunk robot. Jandaghi et al. [16] introduce a novel machine learning method, called deterministic learning, for training soft trunk robot models using radial basis function neural networks. The identification of fault occurrences during the operation of soft trunk robots is also investigated. Bodaghi et al. [17] present a flow-learning-based dimensionality reduction intermediate multimodal fusion network and compare various dimensionality reduction techniques for different variants of unimodal and multimodal networks.
In recent years, the utilization of deep learning techniques has grown in the field of recognizing patterns in biomedical signals. Due to the excellent performance of this method in terms of prediction accuracy, its future application scenarios are very broad. This paper presents a parallel network. The network consists of a convolutional neural network (CNN) with a multi-head attention mechanism and a bidirectional long short-term memory (BiLSTM) network. Six finger movements of 10 subjects were tested in the Ninapro DB2 dataset and benchmarked with five other deep learning methods. In addition, we also tested the network proposed in this paper in Ninapro DB7 for verifying the robustness of the network in different datasets. The test results show that the network has strong robustness.
The main contributions of this paper are the following four aspects: (1) We extract multiple features from the original sEMG signal and then fuse them. This effectively improves the accuracy of continuous estimation. (2) The network proposed in this paper skillfully combines the convolutional neural network, recurrent neural network, and attention mechanism. The special parallel mode ensures accuracy with high processing efficiency. (3) We tested the proposed model on a new database, and the results show that the model proposed in this paper has good robustness. (4) We also investigated the efficiency of the model, and the experimental results show that the model has a high inference speed, which meets the requirement of real-time continuous motion estimation (less than 100 ms).
The remainder of this paper is organized as follows: Section 2 discusses the dataset, the feature extraction method, the construction of the control model, and the PNCB model. Section 3 gives the evaluation metrics for continuous motion estimation and the test results of the PNCB model and the control model. Section 4 and Section 5 provide the discussion and conclusion of this paper.

2. Methods

2.1. Dataset

In order to evaluate the performance of the newly introduced algorithm and to compare it fairly with other traditional deep learning algorithms, this study chose the Ninapro open dataset. The Ninapro dataset records the motion data of multiple intact subjects and amputee subjects in the form of electromyography and is widely used in the development and testing of motion recognition control algorithms [18]. Ninapro is divided into 10 sub-datasets according to the type of experiment. In this study, we used the second one, referred to as DB2. This dataset contains motion data of 40 complete subjects.
During the process of obtaining hand movement data, subjects must follow the prompts to make corresponding actions and perform each movement for five seconds, alternating with rest positions for three seconds. Twenty-two joint angles were accurately measured at a sampling rate of 20 Hz using the CyberGlove II data glove. Surface EMG signals were measured and acquired through two models of dual differential surface EMG signal electrodes. One of the electrode combinations, called Delsys Trigno Wireless System, samples raw surface electromyography signals at a frequency of 2 kHz. It contains 12 wireless surface electromyography signal electrodes and a base station. In order to synchronize the frequency of data collected by the above two devices, the joint angles were resampled to 2 kHz.
In this study, we chose to use the 12-channel EMG to estimate 10 joint angles. As shown in Figure 1A, the ten joint angles selected, including the proximal interphalangeal and metacarpophalangeal joint points, were the primary active joints in the grasping movement. To ensure the generalizability of the algorithm, we selected 10 representative subjects from Ninapro DB2. Their height ranged from 169 to 187 cm and their weight ranged from 58 to 75 kg. Six different grasping actions were chosen for each participant, as shown in Figure 1B, since the focus of this paper is on the continuous estimation of joint angles for grasping actions. In order to verify the robustness of the model proposed in this paper, we also tested the same six grasping actions in Ninapro DB7 with a wide selection of 10 subjects.

2.2. Data Processing

Joint angle continuous motion is motion across multiple degrees of freedoms (DoFs). It represents a higher-dimensional and more complex target space and requires more comprehensive feature extraction methods. However, conventional deep learning feature extraction methods struggle to provide sufficient information for complex motion estimation across multiple degrees of freedom. Therefore, we used and fused multiple surface EMG signal feature extraction methods oriented to deep learning models. The extracted features such as Peak Stress (PS) and Shake Expectation (SE) are closely related to the strength and period of muscle contraction, which determines the accuracy of joint angle estimation.
In order to more completely extract data features, a sliding window of 100 ms, with a step size of 0.5 ms, was segmented from the surface electromyography signal and the resampled hand movement joint angle signal. Then, we extracted the following features from the segmented data intervals.

2.2.1. Root Mean Square (RMS)

Because of its simple calculation form and rich information, RMS is often used as a feature of surface electromyographic signals [19].

2.2.2. Peak Stress (PS)

Peak Stress (PS) is defined as a measurement that quantifies the amount of stress that becomes the peak in each sliding window. This feature is described using zeroth-, second-, and fourth-order moments. The calculation formulas are as follows:
P S = m 4 m 2 m 0
where m 0 represents the intensity of muscle contraction, m 2 can be seen as a formula describing changes in surface electromyographic signals, and m 4 describes the changes in m 2 . The quotient of m 4 and m 2 represents the stress change of the surface electromyographic signal, and then the stress of each unit can be obtained by dividing by m 0 .

2.2.3. Shake Expectation (SE)

The second derivative can explain the curvature of the surface EMG signal. Therefore, the amplitude changes the speed expectation of each sliding window, that is, Shake Expectation (SE) can be described by the average absolute second derivative of the surface electromyographic signal, and its calculation formula is
S E = 1 N i = 0 N 1 Δ 2 x i
where Δ 2 represents the second derivative.

2.2.4. Unbiased Standard Deviation (USTD)

A sliding window of surface EMG signals can be viewed as a sample of movement progression. Unbiased Standard Deviation (USTD) uses unbiased estimators to evaluate the distribution of sample values. We assume that x ¯ is the prediction of the surface electromyographic signal amplitude in the sliding window, with the following calculation formula:
U S T D = 1 N 1 i = 0 N 1 | x i x ¯ | 2

2.3. Bidirectional Long Short-Term Memory Network

The LSTM model was proposed to address the issue of vanishing gradients in conventional recurrent neural networks (RNNs). As a result, it demonstrates exceptional performance in capturing dependencies over longer distances. BiLSTM is an LSTM variant that connects in both directions. BiLSTM can effectively improve the long-term dependence of learning and further improve the model prediction accuracy [20]. Thus, the BiLSTM model finds extensive usage in natural language processing [21] and regression forecasting [22].
The fundamental component of the LSTM network consists of three gate structures, the forget gate, the input gate, and the output gate, as shown in Figure 2. The forget gate usually uses the Sigmoid function to achieve the purpose of retaining or deleting existing information. The input gate is made up of Sigmoid and Tanh layers that determine how much new information is being stored in memory. The output gate passes the information to the next LSTM unit. The specific mathematical formulas of the gates control unit are as follows:
f t = σ W f h t 1 , x t + b f
i t = σ W i h t 1 , x t + b i
C ˜ t = tanh W C h t 1 , x t + b C
C t = f t C t 1 + i t C ˜ t
o t = σ W o h t 1 , x t + b o
h t = o t tanh C t
where σ and tanh are the Sigmoid and Tanh activation functions, respectively. The W item is the weight matrix, and its value range is 0–1. The subscripts f , i , C , and o represent the forget layer, input layer, hidden layer, and output layer, respectively. The b item is the bias matrix of each layer. f t , i t , and o t represent the status of the forget layer, input layer, and output layer at time t , respectively. C t and C ˜ t represent the unit state vector. h t is the final unit state vector at time t .
In this study, we utilized a bidirectional long short-term memory network consisting of two layers, with a predetermined number of 64 hidden units. Our tests showed that this model structure performs well in terms of both training speed and accuracy. The output is generated via a fully connected layer.

2.4. Temporal Convolutional Network

It is well known that recurrent neural networks (RNNs) have excellent performance in processing time series problems, but they can only handle one time step at a time and cannot be processed in large-scale parallelism like convolutional neural networks. A Temporal Convolutional Network (TCN) using a convolutional architecture with extended causal convolutions and residual connections was introduced by Bai et al. [23]. Due to its excellent performance on time series problems, the TCN is widely used in fields such as signal processing [24] and regression prediction [25].
Causal convolution allows the network to only see future data and is a strictly time-constrained model. The size of the convolution kernel limits the time available for modelling by simple causal convolution. This requires a very deep network when processing tasks with a large workload, which will occupy a very large amount of computing resources. The extended causal convolution of the TCN effectively expands the receptive field of the convolutional neural network and solves this problem. The TCN incorporates a residual connection component that allows gradient transmission across layers to effectively prevent gradient disappearance.
In this study, we constructed a 5-layer TCN with 48 convolution channels and a convolution kernel size of 3. Finally, the output utilizes a fully connected layer. The above configuration was experimentally verified to achieve a relatively optimistic accuracy rate.

2.5. Parallel Network of CNN-Attention and BiLSTM

Convolution and recurrent neural networks are gaining ground in surface EMG signal processing and have shown good results in the past. However, when it comes to the regression problem based on surface electromyography signals, the above two networks have problems such as large prediction fluctuations and difficulty in convergence. Solving the above problems according to previous ideas requires expanding the scale of the network, but this will bring challenges to computing resources. To this end, we introduce a multi-head attention mechanism and combine it with a convolutional neural network, which is CNN-Attention. Considering the excellent performance of the BiLSTM network in solving long sequences, we connected it in parallel with the CNN-Attention network and finally obtained the model proposed in this article, called PNCB.

2.5.1. Multi-Scale Convolution

In the CNN module of the PNCB model, we adopt a multi-scale convolution module. The conventional convolution process has a small responsive area with a fixed filter size, which leads to the inadequate extraction of information when dealing with long sequence data like surface electromyographic signals. Therefore, in this study, we adopt a convolution operation with multiple size filters that can extract large-scale and small-scale information simultaneously. As shown in Figure 3, we set up convolution modules with three different convolution kernel sizes (the convolution kernel sizes used in this article are 3, 4, and 5, respectively). The preprocessed surface electromyography signal data pass through three convolution modules in a parallel manner. Each convolution module contains three parts, namely the convolution layer, maximum pooling layer, and ELU activation function. Finally, to obtain the output of the multi-scale convolution module, the results of the three convolution modules are combined based on their corresponding dimensions and then passed through a fully connected layer. Suppose the input signal is X ; then, the output X of the multi-scale convolution module can be expressed by the following equation:
X = σ f n 1 × D ; C X + σ f n 2 × D ; C X + σ f n 3 × D ; C X
where f n 1 × D ; C denotes the convolution operation, C is the number of convolution kernels, D is the number of channels of the input signal, n 1 is the convolution kernel size, and σ is the ELU activation function.

2.5.2. Multi-Head Attention Mechanism

The attention mechanism has found a wide range of applications in disciplines such as image processing [26] and machine translation [27] due to its superior parallelism and efficiency. The model of attention establishes an interactive relationship between a variety of input vectors in a linear space of projections. Multi-head attention aims to create distinct projections in multiple diverse projection spaces, perform different projections on the input matrix, and after obtaining many output matrices, splice them together. As shown in Figure 4B, in the context of single-head attention within multi-head attention, the network first calculates the attention between the key vector and the query vector and then calculates the weight value vector. The mathematical expressions for query vector Q , key vector K , weight value vector V , and attention A t t e n t i o n Q , K , V are as follows:
Q = W q x + b q
K = W k x + b k
V = W v x + b v
A t t e n t i o n Q , K , V = s o f t m a x Q K T d k · V
where W q , W k , and W v are in R C × C , and b q , b k , and b v are in R C . C here refers to the channel dimension of the single-head attention mechanism layer. d k is the dimension of Q and K .
As shown in Figure 4A, we set up a multi-head attention mechanism with 8 heads (n = 8) in parallel. This module divides the input sequence into n groups according to dimensions. Then, it splices the outputs of multiple single-head attention layers according to the corresponding dimensions. Finally, a fully connected layer provides the output of the multi-head attention module. The mathematical expressions of this process are as follows:
h e a d i = A t t e n t i o n Q i , K i , V i
M u l t i H e a d = C o n c a t h e a d 1 , , h e a d n W + b
where W is in R D × D and b is in R D .

2.5.3. PNCB Model Structure

As shown in Figure 5, the original surface EMG signals are sent to the feature extractor, respectively, to extract the four features mentioned in Section 2.2. The data after feature extraction are spliced according to the channel dimension to obtain 48-channel feature signal data. The processed data are then input into the PNCB model, and predictions of 10 joint angles are finally obtained. The PNCB model has two parallel branches. One branch is a CNN-Attention network composed of multi-scale convolution modules and multi-head attention mechanism modules in series. The parameters and specific implementation of this branch are described in Section 2.5.1. And the other branch is composed of BiLSTM and a maximum pooling layer. The parameters of this branch of the BiLSTM remain the same as those mentioned in Section 2.3. The outputs of the two branches are spliced according to the channel dimension, and finally the features are mapped to 10 target angles through two fully connected layers.
It is worth noting that in addition to testing the TCN and BiLSTM network as control experiments, we also tested three sub-module networks in the PNCB model. They are multi-scale convolution modules, namely the CNN, CNN-Attention network, and CNN-BiLSTM network (the network that removes the multi-head attention module in the PNCB model). The purpose of this is not only to enrich the control experiment, but also to test whether each module of the model promotes the entire system.

3. Result

3.1. Evaluation Metrics

In this study, we use the Pearson Correlation Coefficient (PCC) to assess the association between the expected and actual joint angles, and thus the accuracy and effectiveness of the algorithm. The closer the PCC is to 1, the more accurate the algorithm is. Its mathematical expression is as follows:
P C C = i = 1 N θ e s t θ e s t ¯ θ r e a l θ r e a l ¯ i = 1 N θ e s t θ e s t ¯ 2 i = 1 N θ r e a l θ r e a l ¯ 2
where θ e s t , θ e s t ¯ , θ r e a l , and θ r e a l ¯ represent joint angle estimates, average joint angle estimates, actual joint angle values, and actual joint angle averages, respectively.
The deviation between the prediction and the actual value is the root-mean-square error (RMSE). This article uses the RMSE to indicate the amount of error in degrees (°) between the predicted and actual joint angle.
R2 objectively assesses the overall accuracy of the algorithm as a comprehensive evaluation index. The coefficient of determination represents the percentage of variance in the predicted value that can be explained by the prediction. The indicator range lies between 0 and 1. The better the estimation performance of the algorithm, the higher the value of the indicator. The calculation formula is
R 2 = 1 i = 1 N θ r e a l θ e s t 2 i = 1 N θ r e a l θ r e a l ¯ 2

3.2. Experimental Parameters and Statistical Analysis

We maintain consistency in the training parameters, including the type of loss function, the type of optimizer, and the learning rate, when training a neural network in this study. The loss function used is the mean squared error, with Adam optimization as the optimizer, and the learning rate is uniformly set to 0.0002. In regression prediction problems, mean squared error is a commonly used loss function. As an effective deep learning optimizer, Adam optimization has often been used in the fields of signal processing and image processing in recent years. It is based on the Stochastic Gradient Descent Technique [28]. We keep the network structure parameters in the control experimental algorithms CNN, CNN-Attention, CNN-BiLSTM, and BiLSTM consistent with those of the corresponding modules of PNCB. All neural networks are built on the Pytorch framework.
The t-test uses the theory of the t-distribution to infer the probability of a difference occurring, and thus to compare whether the difference between two sets of data is significant or not. It is applicable to normal distributions with a small sample size and unknown overall standard deviation. In this article, we used it to examine the differences between the comparison algorithm used in the experiment and the model proposed in this article. The dependent variables used in this study were PCC, RMSE, and R2 of the evaluation indicators. In this research, the statistical significance threshold was established at p value < 0.05.

3.3. Experimental Results

The general performance of the PNCB model and the other five comparison algorithms for the continuous motion estimation of finger joints is shown in Figure 6. Figure 6A–C show the average PCC, RMSE, and R2 of the six deep learning algorithms on 10 subjects, respectively. The experimental results show that the mean value of the PCC of the PNCB model is 0.87 ± 0.02, and the average PCC values of the other five comparison algorithms are BiLSTM (0.79 ± 0.04, p < 0.05), CNN-Attention (0.80 ± 0.01, p < 0.05), CNN (0.70 ± 0.03, p < 0.05), CNN-BiLSTM (0.83 ± 0.02, p < 0.05), and TCN (0.76 ± 0.05, p < 0.05). The average RMSE of the PNCB model is 9.57 ± 1.01, and the average RMSEs of the other five comparison algorithms are BiLSTM (11.85 ± 1.80, p < 0.05), CNN-Attention (11.65 ± 1.38, p < 0.05), CNN (13.69 ± 1.59, p < 0.05), CNN-BiLSTM (10.78 ± 1.36, p < 0.05), and TCN (12.72 ± 1.87, p < 0.05). The average R2 value of the PNCB model is 0.72 ± 0.04, and the average R2 values of the other five comparison algorithms are BiLSTM (0.58 ± 0.09, p < 0.05), CNN-Attention (0.57 ± 0.07, p < 0.05), CNN (0.47± 0.06, p < 0.05), CNN-BiLSTM (0.65 ± 0.06, p < 0.05), and TCN (0.53 ± 0.10, p < 0.05). Our proposed PNCB model outperforms the other five comparative algorithms on three evaluation indicators, as shown by the experimental results. The p-values are all below 0.05. This means that there is a statistically significant difference in performance between the six methods.
Figure 7A–C show the accuracy performance of the PNCB model and other five comparison algorithms among 10 subjects. The PCC value of the PNCB model remains consistently high across all subjects, exceeding 0.85 and indicating a very strong level of correlation. Additionally, the PNCB model maintains a lower RMSE and higher R2 on each of the 10 subjects. The results above show that the PNCM is superior to the other five algorithms compared and is more widely applicable.
Figure 8A–C show joint angle curves predicted by PNCB, BiLSTM and the TCN alongside the actual joint angle curves. It is not difficult to see from the figure that the predicted joint angle curve of PNCB has the best consistency with the actual joint angle curve, which is better than those of BiLSTM and the TCN. The joint angle curve predicted by BiLSTM agrees well with the actual joint angle curve. However, the prediction curves of some action segments fluctuate greatly, and the prediction curves have more prominent abnormalities. The TCN-predicted joint angle curves fluctuated greatly and did not match the actual joint angle curves well enough.
In order to verify the robustness of PNCB, we selected the same six actions from a wide range of 10 more subjects in Ninapro DB7 and performed a continuous estimation test. The experimental results show that the average PCC value is 0.86 ± 0.03, the average RMSE is 9.38 ± 1.05, and the average R2 value is 0.72 ± 0.06. The experimental results indicate that PNCB is robust enough to be applied to other datasets. In addition, we have also conducted a study on the efficiency of the PNCB model. The size of the model is 0.991 MB, and the inference time of the model on a GeForce RTX3060 graphics card is 10.92 ms. Although the PNCB model has a large number of parameters, its inference time is very short and the model is efficient because it can be computed in parallel.

4. Discussion

This paper introduces the PNCB model, which skillfully integrates multi-scale convolutional neural networks, multi-head attention mechanisms, and bidirectional long short-term memory networks to improve the accuracy and universality of continuous estimation of hand joint angles. And 10 subjects were extensively selected in the Ninapro DB2 dataset to conduct a continuous estimation of hand motion joint angles for six daily grasping movements. Five other algorithms were also tested as control experiments.
To ensure precise and quantitative evaluation of the PNCB model and other control algorithms, we measured three evaluation indicators: PCC, RMSE, and R2. The findings of this study demonstrate that the PNCB model, as presented in this paper, outperforms the control algorithm in terms of both overall and individual subject performance. And it has good stability and universality. It is worth noting that the comparison algorithms CNN, CNN-Attention, and CNN-BiLSTM all come from certain modules of the PNCB model. The purpose of this is to enrich the control experiments and to test whether the stacking method of each module of the PNCB model promotes the overall performance of the model. It is not difficult to see from the experimental results (Figure 6) that the CNN with only one multi-scale convolution module performed worst among the control algorithms. The comparison algorithms CNN-Attention and CNN-BiLSTM stacked by the two modules have better performance than CNN, and their performance has been greatly improved. This shows that each module of PNCB promotes model performance.
To more intuitively see the accuracy of the PNCB model, BiLSTM, and TCN in the continuous estimation of hand motion joints, we drew the actual joint angle curves and the predicted joint angle curves of the above three networks, as shown in Figure 8. The superiority of the PNCB model is further verified. The predicted joint angle curve of the PNCB model is more stable and smoother than those of BiLSTM and the TCN, and it has a high degree of fit with the actual curve. There are several reasons for the excellent performance of the PNCB model. The first is the use of a multi-head attention mechanism, which can obtain global and local relationships in one step in a parallel manner. However, the BiLSTM network acquires global information in a recursive manner and is susceptible to noise, leading to reduced accuracy. In the same way, the TCN obtains global information through layering and the receptive field is not large enough, which has the same problem as BiLSTM. In addition, multi-scale convolution can obtain the characteristics of different scales of surface electromyographic signals, further improving the performance of the PNCB model. Finally, we added BiLSTM to the model in parallel and used its unique loop structure to once again improve the performance of the model. The advantage of using a parallel form of adding BiLSTM is that this can fully utilize the strengths of the different networks, avoid the weaknesses, and also improve the processing efficiency of the model.
The following shortcomings may exist in this study. There are many factors that affect EMG, including electrode movement, noise, and arm position [29]. These factors actually exist and can reduce the accuracy of the continuous estimation of the model. In future research, we will consider modeling these factors by adding noise to the data to test and improve the model to make it more robust. In addition, it is our goal to use the model in real-world scenarios such as smart prosthetics and controlling robotic hands using EMG. In the future, we are considering testing the model on edge devices so that we can apply the model to real-world scenarios.

5. Conclusions

This paper proposes a parallel model that cleverly integrates convolutional neural networks, multi-head attention mechanisms, and bidirectional long short-term memory networks, taking full advantage of the benefits of each neural network. And the continuous estimation of hand motion joints was tested on the Ninapro dataset. Its performance is significantly better than those of the five comparison algorithms mentioned in this article and has good stability and universality. Furthermore, upon comparing the experimental outcomes, it is apparent that the stacking technique for each module in the PNCB model proposed encourages the overall performance of the model and can suggest insights on constructing deep learning networks. It is worth noting that we also innovatively extracted multiple features of the surface electromyographic signal and fused them, which improved the accuracy of continuous estimation of hand motion. We also test it on a wide selection of subjects in different datasets, and the experimental results show that PNCB possesses strong robustness. In brief, it is anticipated that PNCB will have a significant role in the continuous motion assessment area that depends on the surface electromyography method in the future.

Author Contributions

Methodology, C.L.; software, S.Z.; validation, S.Z.; writing—original draft preparation, S.Z.; writing—review and editing, C.L.; project administration, C.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Leading talent project of Dalian Maritime University.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. These data can be found here: https://ninapro.hevs.ch/ (accessed on 15 March 2024).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Xiong, D.; Zhang, D.; Zhao, X.; Zhao, Y. Deep Learning for EMG-based Human-Machine Interaction: A Review. IEEE/CAA J. Autom. Sin. 2021, 8, 512–533. [Google Scholar] [CrossRef]
  2. Benalcázar, M.E.; González, J.; Jaramillo-Yánez, A.; Anchundia, C.E.; Zambrano, P.; Segura, M. A Model for Real-Time Hand Gesture Recognition Using Electromyography (EMG), Covariances and Feed-Forward Artificial Neural Networks. In Proceedings of the 2020 IEEE ANDESCON, Quito, Ecuador, 13–16 October 2020; pp. 1–6. [Google Scholar] [CrossRef]
  3. Kaplan, K.E.; Nichols, K.A.; Okamura, A.M. Toward human-robot collaboration in surgery: Performance assessment of human and robotic agents in an inclusion segmentation task. In Proceedings of the 2016 IEEE International Conference on Robotics and Automation (ICRA), Stockholm, Sweden, 16–21 May 2016; pp. 723–729. [Google Scholar] [CrossRef]
  4. Hakonen, M.; Piitulainen, H.; Visala, A. Current state of digital signal processing in myoelectric interfaces and related applications. Biomed. Signal Process. Control. 2015, 18, 334–359. [Google Scholar] [CrossRef]
  5. Artemiadis, P. EMG-based Robot Control Interfaces: Past, Present and Future. Adv. Robot. Autom. 2012, 1, e107. [Google Scholar] [CrossRef]
  6. Chen, J.; Bi, S.; Zhang, G.; Cao, G. High-Density Surface EMG-Based Gesture Recognition Using a 3D Convolutional Neural Network. Sensors 2020, 20, 1201. [Google Scholar] [CrossRef] [PubMed]
  7. Park, K.H.; Lee, S.W. Movement intention decoding based on deep learning for multiuser myoelectric interfaces. In Proceedings of the 2016 4th International Winter Conference on Brain-Computer Interface (BCI), Gangwon, Republic of Korea, 22–24 February 2016; pp. 1–2. [Google Scholar] [CrossRef]
  8. Bi, L.; Guan, C. A review on EMG-based motor intention prediction of continuous human upper limb motion for human-robot collaboration. Biomed. Signal Process. Control. 2019, 51, 113–127. [Google Scholar] [CrossRef]
  9. Liu, P.; Liu, L.; Clancy, E.A. Influence of Joint Angle on EMG-Torque Model During Constant-Posture, Torque-Varying Contractions. IEEE Trans. Neural Syst. Rehabil. Eng. 2015, 23, 1039–1046. [Google Scholar] [CrossRef] [PubMed]
  10. Wang, L.; Buchanan, T.S. Prediction of joint moments using a neural network model of muscle activations from EMG signals. IEEE Trans. Neural Syst. Rehabil. Eng. 2002, 10, 30–37. [Google Scholar] [CrossRef] [PubMed]
  11. Chen, Y.; Yu, S.; Ma, K.; Huang, S.; Li, G.; Cai, S.; Xie, L. A Continuous Estimation Model of Upper Limb Joint Angles by Using Surface Electromyography and Deep Learning Method. IEEE Access 2019, 7, 174940–174950. [Google Scholar] [CrossRef]
  12. Gautam, A.; Panwar, M.; Biswas, D.; Acharyya, A. MyoNet: A Transfer-Learning-Based LRCN for Lower Limb Movement Recognition and Knee Joint Angle Prediction for Remote Monitoring of Rehabilitation Progress From sEMG. IEEE J. Transl. Eng. Health Med. 2020, 8, 2100310. [Google Scholar] [CrossRef]
  13. Chen, Y.; Dai, C.; Chen, W. Cross-Comparison of EMG-to-Force Methods for Multi-DoF Finger Force Prediction Using One-DoF Training. IEEE Access 2020, 8, 13958–13968. [Google Scholar] [CrossRef]
  14. Ma, C.; Guo, W.; Zhang, H.; Samuel, O.W.; Ji, X.; Xu, L.; Li, G. A Novel and Efficient Feature Extraction Method for Deep Learning Based Continuous Estimation. IEEE Robot. Autom. Lett. 2021, 6, 7341–7348. [Google Scholar] [CrossRef]
  15. Zhang, J.; Chen, X.; Jandaghi, E.; Zeng, W.; Zhou, M.; Yuan, C. Dynamics Learning-Based Fault Isolation for A Soft Trunk Robot. In Proceedings of the 2023 American Control Conference (ACC), San Diego, CA, USA, 31 May–2 June 2023; pp. 40–45. [Google Scholar] [CrossRef]
  16. Jandaghi, E.; Chen, X.; Yuan, C. Motion Dynamics Modeling and Fault Detection of a Soft Trunk Robot. In Proceedings of the 2023 IEEE/ASME International Conference on Advanced Intelligent Mechatronics (AIM), Seattle, WA, USA, 28–30 June 2023; pp. 1324–1329. [Google Scholar] [CrossRef]
  17. Bodaghi, M.; Hosseini, M.; Gottumukkala, R. A Multimodal Intermediate Fusion Network with Manifold Learning for Stress Detection. arXiv 2024. [Google Scholar] [CrossRef]
  18. Atzori, M.; Muller, H. The Ninapro database: A resource for sEMG naturally controlled robotic hand prosthetics. In Proceedings of the 2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Milan, Italy, 25–29 August 2015; pp. 7151–7154. [Google Scholar] [CrossRef]
  19. Jang, G.; Kim, J.-H.; Lee, S.; Choi, Y. EMG-Based Continuous Control Scheme with Simple Classifier for Electric-Powered Wheelchair. IEEE Trans. Ind. Electron. 2016, 63, 3695–3705. [Google Scholar] [CrossRef]
  20. Siami-Namini, S.; Tavakoli, N.; Namin, A.S. The Performance of LSTM and BiLSTM in Forecasting Time Series. In Proceedings of the 2019 IEEE International Conference on Big Data (Big Data), Los Angeles, CA, USA, 9–12 December 2019; pp. 3285–3292. [Google Scholar] [CrossRef]
  21. Xu, G.; Meng, Y.; Qiu, X.; Yu, Z.; Wu, X. Sentiment Analysis of Comment Texts Based on BiLSTM. IEEE Access 2019, 7, 51522–51532. [Google Scholar] [CrossRef]
  22. Ma, C.; Lin, C.; Samuel, O.W.; Guo, W.; Zhang, H.; Greenwald, S.; Xu, L.; Li, G. A Bi-Directional LSTM Network for Estimating Continuous Upper Limb Movement from Surface Electromyography. IEEE Robot. Autom. Lett. 2021, 6, 7217–7224. [Google Scholar] [CrossRef]
  23. Bai, S.; Kolter, J.Z.; Koltun, V. An Empirical Evaluation of Generic Convolutional and Recurrent Networks for Sequence Modeling. arXiv 2018. [Google Scholar] [CrossRef]
  24. Zanghieri, M.; Benatti, S.; Conti, F.; Burrello, A.; Benini, L. Temporal Variability Analysis in sEMG Hand Grasp Recognition using Temporal Convolutional Networks. In Proceedings of the 2020 2nd IEEE International Conference on Artificial Intelligence Circuits and Systems (AICAS), Genova, Italy, 31 August–2 September 2020; pp. 228–232. [Google Scholar] [CrossRef]
  25. Song, Y.; Gao, S.; Li, Y.; Jia, L.; Li, Q.; Pang, F. Distributed Attention-Based Temporal Convolutional Network for Remaining Useful Life Prediction. IEEE Internet Things J. 2020, 8, 9594–9602. [Google Scholar] [CrossRef]
  26. Zhou, Y.; Chen, H.; Li, Y.; Liu, Q.; Xu, X.; Wang, S.; Yap, P.T.; Shen, D. Multi-task learning for segmentation and classification of tumors in 3D automated breast ultrasound images. Med. Image Anal. 2021, 70, 101918. [Google Scholar] [CrossRef]
  27. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. arXiv 2017. [Google Scholar] [CrossRef]
  28. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014. [Google Scholar] [CrossRef]
  29. Patel, G.K.; Castellini, C.; Hahne, J.M.; Farina, D.; Dosen, S. A Classification Method for Myoelectric Control of Hand Prostheses Inspired by Muscle Coordination. IEEE Trans. Neural Syst. Rehabil. Eng. 2018, 26, 1745–1755. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Dataset description. (A) Ten selected joint angles, (B) six grasping actions.
Figure 1. Dataset description. (A) Ten selected joint angles, (B) six grasping actions.
Applsci 15 11078 g001
Figure 2. Structure of LSTMCell.
Figure 2. Structure of LSTMCell.
Applsci 15 11078 g002
Figure 3. Structure of multi-scale convolution.
Figure 3. Structure of multi-scale convolution.
Applsci 15 11078 g003
Figure 4. Multi-head attention module. (A) Structure of multi-head attention. (B) Structure of scaled dot-product attention.
Figure 4. Multi-head attention module. (A) Structure of multi-head attention. (B) Structure of scaled dot-product attention.
Applsci 15 11078 g004
Figure 5. PNCB model architecture and overall processing flow.
Figure 5. PNCB model architecture and overall processing flow.
Applsci 15 11078 g005
Figure 6. General performance of the PNCB model and the other five comparison algorithms in (A) average PCC, (B) average RMSE, and (C) average R2.
Figure 6. General performance of the PNCB model and the other five comparison algorithms in (A) average PCC, (B) average RMSE, and (C) average R2.
Applsci 15 11078 g006
Figure 7. Accuracy performance of the PNCB model and the other five comparison algorithms among 10 subjects in (A) PCC, (B) RMSE, and (C) R2.
Figure 7. Accuracy performance of the PNCB model and the other five comparison algorithms among 10 subjects in (A) PCC, (B) RMSE, and (C) R2.
Applsci 15 11078 g007
Figure 8. Actual joint angle curve and joint angle curve predicted by (A) PNCB, (B) BiLSTM, and (C) TCN. The abscissa in the figure represents the sampling point, and the ordinate represents the joint angle. The orange curve represents the actual joint angle curve and the blue curve represents the predicted joint angle curve.
Figure 8. Actual joint angle curve and joint angle curve predicted by (A) PNCB, (B) BiLSTM, and (C) TCN. The abscissa in the figure represents the sampling point, and the ordinate represents the joint angle. The orange curve represents the actual joint angle curve and the blue curve represents the predicted joint angle curve.
Applsci 15 11078 g008
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lin, C.; Zhou, S. A Parallel Network for Continuous Motion Estimation of Finger Joint Angles with Surface Electromyographic Signals. Appl. Sci. 2025, 15, 11078. https://doi.org/10.3390/app152011078

AMA Style

Lin C, Zhou S. A Parallel Network for Continuous Motion Estimation of Finger Joint Angles with Surface Electromyographic Signals. Applied Sciences. 2025; 15(20):11078. https://doi.org/10.3390/app152011078

Chicago/Turabian Style

Lin, Chuang, and Shengshuo Zhou. 2025. "A Parallel Network for Continuous Motion Estimation of Finger Joint Angles with Surface Electromyographic Signals" Applied Sciences 15, no. 20: 11078. https://doi.org/10.3390/app152011078

APA Style

Lin, C., & Zhou, S. (2025). A Parallel Network for Continuous Motion Estimation of Finger Joint Angles with Surface Electromyographic Signals. Applied Sciences, 15(20), 11078. https://doi.org/10.3390/app152011078

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop