Skip to Content
BiomimeticsBiomimetics
  • Article
  • Open Access

14 June 2023

A 3D Printed, Bionic Hand Powered by EMG Signals and Controlled by an Online Neural Network

,
,
,
and
1
Neuroimaging and Bioengineering Laboratory (LNB), Facultad de Ingeniería en Mecánica y Ciencias de la Producción, Escuela Superior Politécnica del Litoral (ESPOL), Campus Gustavo Galindo km 30.5 Vía Perimetral, Guayaquil 090903, Ecuador
2
SDAS Research Group, Ben Guerir 43150, Morocco
3
Faculty of Law, Administrative and Social Sciences, Universidad UTE, Quito 170147, Ecuador
4
College of Computing, Mohammed VI Polytechnic University, Ben Guerir 47963, Morocco

Abstract

About 8% of the Ecuadorian population suffers some type of amputation of upper or lower limbs. Due to the high cost of a prosthesis and the fact that the salary of an average worker in the country reached 248 USD in August 2021, they experience a great labor disadvantage and only 17% of them are employed. Thanks to advances in 3D printing and the accessibility of bioelectric sensors, it is now possible to create economically accessible proposals. This work proposes the design of a hand prosthesis that uses electromyography (EMG) signals and neural networks for real-time control. The integrated system has a mechanical and electronic design, and the latter integrates artificial intelligence for control. To train the algorithm, an experimental methodology was developed to record muscle activity in upper extremities associated with specific tasks, using three EMG surface sensors. These data were used to train a five-layer neural network. the trained model was compressed and exported using TensorflowLite. The prosthesis consisted of a gripper and a pivot base, which were designed in Fusion 360 considering the movement restrictions and the maximum loads. It was actuated in real time thanks to the design of an electronic circuit that used an ESP32 development board, which was responsible for recording, processing and classifying the EMG signals associated with a motor intention, and to actuate the hand prosthesis. As a result of this work, a database with 60 electromyographic activity records from three tasks was released. The classification algorithm was able to detect the three muscle tasks with an accuracy of 78.67% and a response time of 80 ms. Finally, the 3D printed prosthesis was able to support a weight of 500 g with a safety factor equal to 15.

1. Introduction

The development of prostheses is necessary to improve the quality of life of people who have lost limbs. These artificial devices play a crucial role in replacing missing parts of the body, allowing for the restoration of functionality, mobility, and autonomy of affected individuals [1,2,3]. Additionally, prosthesis provide long-term health benefits, drive technological advancements, and enable adaptation. Continuous research and development in this field are fundamental to further improving the lives of those who depend on these devices.
The hand is one of the most developed organs of the body. It allows humans to interact with their environment through complex movements due to the considerable number of degrees of freedom in its structure (27 in total: four in each finger, five in the thumb and six in the wrist). In addition, it is a fundamental part of physical and social interactions. The upper extremities depend on the hand, so losing one implies a reduction in autonomy, limitations in the development of work and daily activities, and a drastic change in the lifestyle of people [4].
In Ecuador, more than 8% of the population has some need in their upper and/or lower limbs. The cost of a prosthesis, which is around 8000 USD, is excessively high compared to the income of the average Ecuadorian, which, according to Encuesta Nacional de Empleo, Desempleo y Subempleo (Enemdu), reached 248 USD in August 2021 [5]. In addition, of the people with physical disabilities, only 17% are employed and most of them are in vulnerable economic situations. Despite several initiatives dedicated to the design and manufacture of prostheses, the needs of existing patients have not been met. The “Las Manuelas” mission, founded in 2007, acquired machinery in 2012 to produce 300 prostheses per month. Until 2019, the prostheses produced did not exceed 10% of the target.
Considering the aforementioned information, it is vital to develop systems that can adjust and optimize gripping patterns and gripping capacity [1] and hand movements according to the needs and preferences of each individual. This will allow for greater customization and comfort in the use of prostheses. Bionic hands, powered by electromyographic (EMG) signals, can interpret and translate the electrical signals generated by residual muscles into precise commands for finger and hand movement [6]. This provides a higher degree of control and a more natural experience when using the prosthesis.
Furthermore, advances in artificial intelligence (AI) have played a crucial role in the design and creation of bionic hands capable of interpreting EMG signals for precise movement control. These techniques use machine learning algorithms, artificial neural networks (ANNs), and computer vision techniques to process EMG signals and enhance prostheses [3], which play a key role in the design of bionic hands [7]. AI has also improved the bionic hands’ ability to learn and adapt as they interact with the environment. By collecting and analyzing real-time data, bionic hands can adjust their movements and grip strength more precisely and efficiently. These advancements have allowed bionic hands to provide better and more sophisticated functionality.
Several studies have focused on the design and manufacture of hand prosthetics that utilize EMG signals and AI for real-time control. For example, Amsüss et al. [8] propose a pattern recognition system for surface EMG signals to control upper limb prosthetics using a trained ANN. Other researchers [9] present a study based on EMG signals acquired from muscles and motion detection through a Human–Machine Interface, designing an upper limb prosthesis using an AI-based controller. In another work [10], hardware design for hand gesture recognition using EMG is developed and implemented on a Zynq platform, processing the acquired EMG signals with an eight-channel Myo sensor. Furthermore, in [11], various methods are applied to detect and classify muscle activities using sEMG signals, with ANNs showing the highest accuracy in recognizing movements among and within subjects. Additionally, in [12], a woven band sensor is manufactured for myoelectric control of prosthetic hands based on single-channel sEMG signals.
The use of artificial intelligence techniques, especially ANNs, has revolutionized the design of bionic hands by allowing control through EMG signals, providing a higher degree of control, functionality, and adaptability. These advancements have had a significant impact on the quality of life of people with amputations or disabilities in their upper limbs. However, despite the progress and promising applications of artificial intelligence in EMG signal processing and other fields, there are still challenges and considerations that need to be addressed. In this regard, this work proposes the design and manufacture of an accessible hand prosthesis powered by EMG signals and controlled by an online neural network. The system demonstrates accuracy in classification and load capacity.
The rest of this manuscript is structured as follows: Section 2 presents some notable related works and provides general information on the types of prosthesis, techniques, and EMG signals and their applications using artificial intelligence. Section 3 describes the system design, the mechanical and control circuit design, data acquisition and analysis, and system integration for controlling a bionic hand. Section 4 gathers the experimental results and presents the discussion. Finally, Section 5 presents the conclusions of this study and states the future work.

3. System Design

3.1. Mechanical Design

The mechanical design steps are detailed below, starting with the design of the prosthetic fingers whose dimensions are used as the basis for establishing other system parameters.

3.1.1. Finger Design

In order to ensure ideal positions for gripping, the angles between the sections representing the distal, medial, and proximal phalanges of the fingers were defined according to Table 1. These are based on the angular difference established for abduction and adduction at the metacarpophalangeal joint [59] and the angle in functional position established for the proximal and distal interphalangeal joint [30]. Additionally, an empirical analysis of the angles between the phalanges observed in the Variplus Speed prosthesis model offered by Ottobock, which resembles the information collected, is presented in Figure 7.
Table 1. Interphalangeal angles.
Figure 7. Empirical analysis of interphalangeal angles.
The length of the phalanges was established proportionally to comply with the empirically determined length of the corresponding fingers: 85 mm for the index finger and 60 mm for the thumb. To simplify the design, the index and middle fingers were set to be the same length; only the medial and distal phalanges of the thumb were modeled since they are the ones that protrude from the palm and participate in the grip. Table 2 shows the established dimensions.
Table 2. Phalanx length.

3.1.2. Flexion–Extension Mechanism Synthesis

The flexion–extension mechanism was designed using the two-position synthesis method since the abduction and adduction of the fingers can be classified as a coupler output. Based on the desired output, it was known that a four-bar triple rocker would be obtained that could be motor driven by the addition of two intermediate links between the motor coupler and the fingers, resulting on a six-bar Watt’s chain in which the four-bar subchain describes the desired motion [60]. To reduce the unknown variables in the design, lengths for the links that will form part of the fingers were defined from the known length of the selected motor metal coupling, as shown in Table 3.
Table 3. Established dimensions for links.
From these dimensions, the missing links were synthesized graphically using Autodesk Fusion 360. In this program, the separation distance of the links forming part of the index finger and thumb was defined based on the dimensions of the cover designed for the selected engine. Finally, the lengths of the remaining links were adjusted to describe the desired movement and maximum opening, obtaining the lengths shown in Table 4.
Table 4. Dimensions determined through graphic synthesis.
In Figure 8 and Figure 9, the links of the mechanism obtained and their position limits are presented.
Figure 8. Resulting links.
Figure 9. Maximum extension (left) and flexion (right).

3.1.3. Motor Selection

Based on the required torque, the servomotors shown in Table 5 were considered. After analyzing the characteristics of these models and determining that they are in similar ranges in terms of weight, dimensions, and stall current, the DS3218MG model was chosen because the stall torque is double that required for the maximum load, which would provide an additional safety factor to counteract potential losses in the links of the mechanism.
Table 5. Motor characteristics.

3.1.4. Covers and Turning Base

Based on the dimensions of the selected servomotor, three covers were designed, as shown in the Figure 10, capable of housing it and holding fixed the flexion–extension mechanism to which it would be coupled.
Figure 10. From Left to Right: front cover, rear cover, upper cover.
Finally, to allow the prosthesis to rotate, a simple shaft connected to the rear cover was designed. This is attached to the rotation base shown in the Figure 11 by means of guide rails and a bearing placed between the top of it and the motor.
Figure 11. From left to right: Axis, pivot base, perspective pivot base.

3.2. Electronic Design

For the design of the control circuit, consecutive stages were established in order to adequately structure both the data recording step and the control of the system in real time.

3.2.1. Sensor

As a first step, a review of the sensors used in similar projects was made. It was found that the MyoWare sensor was the most widely used, due to its accessibility in the market and its low price; its characteristics are shown in Table 6.
Table 6. Technical characteristics of the MyoWare sensor.

3.2.2. Sampling Frequency

Nyquist’s theorem is shown in Equation (1). This states that the sampling frequency of a signal must be at least twice the highest frequency present in the signal. This is the only way to accurately reconstruct the signal.
f s 2 f m a x
Applying that theorem to electromyography, a sampling frequency of at least 300 Hz should be established, since the relevant information is in the range of 50–150 Hz [27]. However, it has been proven that the application of higher sampling frequencies manages to capture substantial information that allows for improved classification [27]. For this reason, a sampling frequency of 1 kHz was chosen.

3.2.3. Development Board Selection

The following characteristics were considered when choosing the development board.
Communication requirement: Communication between the development board and the computer was crucial, since both times had to be synchronized. For this reason, Bluetooth Low Energy (BLE) was chosen. This communication architecture has been used by several myoelectric signal collection systems [61].
Storage requirement: For the data collection stage, the data had to be stored in the most optimal way without affecting the signal sampling rate. This was high, so the data packets were not small. For this reason, it was decided to store the data locally, using a micro SD card reader module.
The development board must have had Bluetooth Low Energy for communication and GPIO pins, the connection of sensors and motors, SPI communication for the ESP32 card reader, and I2C communication for a gyroscope. Among the options was one offered by one of the most recognized manufacturers of AIoT (Artificial Intelligence of Things) chips, Espressif. We chose the development board Esp32 C3 DevKit-M1, which has the characteristics presented in the Table 7.
Table 7. Features of Esp32 DevKit-M1.

3.2.4. Printed Circuit Board

Due to the multiple existing connections, it was decided to design a PCB board that would allow a more compact electronic system. The first step was to create a schematic diagram representing all the connections of the electronic elements, as shown in the following Figure 12.
Figure 12. Preliminary electronic circuit design.
The board was designed in Altium Designer software, as shown in Figure 13.
Figure 13. Electronic circuit design. (a) Bottom layer (b) Top layer.
The electronic circuit was organized on the PCB board; the result is shown on the Figure 14.
Figure 14. Bakelite printed PCB board.
Finally, a case was designed to encapsulate the circuit. This allowed decreasing of the noise in the recorded signals and at the same time prevented the circuit from being exposed to factors that could damage it such as shocks, water, external particles, etc.

3.3. Data Adcquisition

3.3.1. Experimental Organization

Due to the variability of EMG signals between test subjects (associated with the anatomical and physiological properties of each human being), a single test subject was called for data acquisition. The test subject was received by the personnel in charge of the experiment. This was carried out in a clean, organized space, with natural light and a temperature of around 25 degrees Celsius. We tried to eliminate any type of stimulus that could divert the attention of the test subject, such as intense lights, loud sounds, and people entering/leaving the room. The test subject was asked to sign an informed consent form, detailing that the data recorded would be used strictly for research purposes and would remain anonymous. Then, the procedure of the experiment was explained, as well as how the system worked and that it would not affect the test subject’s health for any reason. The test subject was asked to sit in a reclining chair, which had support for both arms (at 145 degrees) and legs (at 90 degrees). The electrodes were then placed on the patient’s brachioradialis, flexor carpi ulnaris, and common extensor digitorum muscles, as shown in Figure 15. According to the MyoWare sensor manufacturer’s instructions [62], each sensor was placed in the center of the muscle. Finally, the screen showing the experiment was placed at a distance of 20 inches from the test subject.
Figure 15. MyoWare sensors positions.

3.3.2. Experimental Methodology

One run of the experiment lasted 7 min, during which visual signals associated with each of the three muscle tasks were displayed: extension and flexion of the wrist and closing of the hand. Each action appeared 10 times on the screen. Initially, a welcome screen was displayed, in which the test subject was told to start the experiment whenever he was ready by pressing the Space key. Each muscle task lasted 5 s, during which the subject had to contract the muscle and hold it. After the 5 s, 10 rest seconds were given to avoid the risk of muscle fatigue. The process was iterative until the defined time was reached, see Figure 16. The actions were randomly displayed on the screen. Each of the three muscle tasks were recorded at a sampling rate of 1 kHz.
Figure 16. Experimental design.

3.4. Dataset

The test subject is an Ecuadorian national man who is 31 years old. He has amputation in both upper limbs. The dataset contains 20 files for each muscle task. In total it contains 60 .csv files. Each file has three columns, one for each MyoWare sensor, and another six columns that record the reading of a gyroscope, during the experiment. The dataset can be found at the following link: https://ieee-dataport.org/documents/emg1k-dataset.
All recorded .csv files were stored with class 1, 2 and 3 tags, as detailed below.
  • Class 1: wrist flexion
  • Class 2: fist
  • Class 3: wrist extension

3.5. Preprocessing

The data obtained were preprocessed using frequency decreasing techniques, data normalization, and feature extraction. All codes can be found at the following link: The data obtained were preprocessed for use during neural network training, using frequency decreasing techniques, data normalization, and feature extraction. All code can be found at the following link: https://github.com/kaviles22/EMG_SignalClassification.git.

3.5.1. Decrease Frequency

EMG signals are noise sensitive, which is why efforts must be made to smoothen and clean them. For this purpose, the RMS envelope technique was used in Equation (2), which proposes to calculate the quadratic mean in fixed time windows. The objective of this method is to smooth the signal without loosing representative information. This is why the size of the moving window cannot be too large, because a lot of information would be lost, nor too small, because the resulting signal will still present a lot of noise.
n = 1 N x n 2 N
Several tests were performed, and it was observed that with very large or very small time window values, the features were not representative; therefore, the time window that yielded the best classification results was chosen, which was 50 ms, as shown in Figure 17.
Figure 17. Rms envelope.

3.5.2. Normalization

This process was used in order to keep all data within the same range, so that the AI model did not assign a greater importance to some samples rather than others. The peak dynamic method, described in Equation (3), was used, which represents the values of a time window as a division between the value obtained and the maximum value of that time window [63].
X n o r m = X X p e a k
With this equation, all values remained within the range of [0, 1].

3.6. Feature Extraction

At this stage, we kept the features in the time domain, because the established pipeline had to be deployed in real time. Thus, it should not represent a high computational cost for the development board. Two approaches for feature extraction from electromyographic signals were analyzed. Both were tested and the best method was chosen.

3.6.1. Statistical Features

There are different types of features that can be used to represent a biological signal. In this case, six statistical measures were used as features: Mean Absolute Value, Mean Absolute Value Weight I, Mean Absolute Value Weight II, Median Absolute Value, Variance, and Standard Deviation [64].

3.6.2. Root Mean Square

Obtain the root mean square in time windows of predefined duration. In this case, a window of 500 ms was chosen from the original window. That is, four RMS values were obtained for each time window (in Section 4.2, it is explained why a final 2-second duration window was chosen).
To evaluate both methods, the classifier (described in more detail in Section 3.7) was used. Its metrics were evaluated using the method 10-fold cross validation, obtaining the results shown in Table 8.
Table 8. Evaluation of feature extraction methods.
Based on the results obtained, the rms features were chosen, and it was concluded that the speed of computation of this type of features in real time would represent an advantage over the other method described.

3.7. Classification

The capabilities of neural networks have been extensively studied in classification applications using bioelectrical signals, such as visual stimulus detection using EEG-SSVEP signals [65]; detection of imaginative-motor intentions of both hands and both feet, obtaining up to 93.7% of accuracy [66]; analysis of EMG signals to facilitate real-time, off-line monitoring of a prosthetic hand [67]; and EMG signal classification using the Wavelet transform in combination with neural networks, obtaining up to 90.7% accuracy [68].

3.7.1. Model

The following model provides a general representation of the bionic hand control system. This hand is powered by EMG signals and controlled through an online neural network. The model begins by explaining how the EMG signals and hand movements are acquired and their relationship. It then illustrates the signal acquisition process and how they are processed before entering the artificial neural network. Subsequently, the model describes the artificial neural network training process, including adjusting weights and connections to achieve proper control of the bionic hand. In real-time, the artificial neural network receives EMG signals and generates control commands to activate the hand’s motors and actuators, enabling desired movements. Specific implementation details such as activation functions, artificial neural network architecture, and other parameters are discussed in the subsequent subsection (described in more detail in Section 3.7.2). The model’s development involves the following steps:
  • Acquisition and preprocessing of EMG signals.
  • Feature extraction: relevant features describing muscle activation patterns are extracted from preprocessed EMG signals.
  • Training dataset: the dataset consists of pairs of EMG signal features and corresponding movements of the bionic hand. We use 60% of the data for training.
  • Artificial neural network structure: input, hidden, and output layers.
  • Artificial neural network training: the neural network is fed with the features of EMG signals. It is trained to learn the correspondence between EMG signals and desired movements of the bionic hand.
  • Validation and adjustment: after training the neural network, its performance is validated using a test dataset. 20% of the data is allocated for validation, and another 20% is used for testing.
  • Implementing the model in the bionic hand.
Below are the inputs and outputs of the model:
Inputs: EMG signals
X = [ x 1 , x 2 , , x n ]
Outputs: movements of the bionic hand
Y = [ y 1 , y 2 , , y n ]
Weights and biases:
P = [ P 1 , P 2 , , P n ] ; b = [ b 1 , b 2 , , b n ]
Hidden layer:
S = i = 1 n X P i + b i
where
  • x i : represents the EMG signal recorded.
  • y i : represents the corresponding movement of the bionic hand.
  • P i and b i : weights and biases of the layers of the neural network.
Feature extraction:
F = [ f 1 , f 2 , , f n ]
where
  • f i : represents the extracted features from segments of EMG signals.
Training dataset:
T = [ ( F 1 , y 1 ) , ( F 2 , y 2 ) , ( F n , y n ) ]
where
  • F i : is a feature vector.
Output:
Y = i = 1 n S P + b
where
  • P and b: weights and biases of the output layer of the artificial neural network.
The mathematical model establishes a relationship between the input EMG signals and the output joint movements of the bionic hand. The artificial neural network is trained to learn this relationship and is used to generate the necessary control signals for the movement of the bionic hand.

3.7.2. Model Training

The Tensorflow library, which provides multiple facilities for developing, training, and deploying machine learning models, was used to design the model. Since we needed an online system that would fit on a hardware with limited resources, we decided to go with a simple and small model that could reach high performance and would fit in the microchip. For this reason, a multilayer perceptron of five layers was used; as shown in Figure 18, intermediate dropout layers with a 20% ratio were used to avoid overfitting the model.
Figure 18. Neural network architecture.
The model was trained with 60% of the data, while the remaining 40% was divided equally between test and validation data. The training parameters are presented in Table 9. The final activation function and the loss function were chosen, given that the task was a multiclass classification. The Softmax function yields a probabilistic distribution of N different classes. The loss or cost function, categorical cross-entropy, measures the precision of the results with respect to an expected value for categorical variables. We tested them out, and the results were stable, so we decided to proceed doing a random search to find the optimal hyperparameters. We used a range from 0.01 to 0.0001 for learning rate, epochs were bounded by [100, 200, 300], and the batch size was constant.
Table 9. Hyperparameters and functions used in model training.
After analyzing results, it was decided to train a binary classifier using only tasks 1 and 3, which corresponded to flexion and extension of the wrist, respectively. The training parameters are shown in Table 10.
Table 10. Hyperparameters and functions used in binary model training.

3.7.3. System Integration

The trained model was integrated into the electromechanical system, allowing the motors to be controlled as follows (see Figure 19):
Figure 19. Muscular tasks. (a) Class 1 (b) Class 2 (c) Class 3.
  • When class 1, the wrist flexion task, is detected, the microcontroller moves the axis of the gripper motor 20 degrees clockwise. In this way, the prosthesis gripper closes, simulating the gripping movement of a hand.
  • When class 2, the muscle task of closing the hand (making a fist), is detected, the microcontroller moves the axis of the swing motor 30 degrees clockwise. In this way, the base of the prosthesis rotates.
  • When class 3 is detected, the wrist extension task, the microcontroller moves the axis of the gripper motor 20 degrees counterclockwise. In this way, the prosthesis gripper opens.

4. Results and Discussion

4.1. Real Time Classification

The trained model was converted into a C++ source file using the Tensorflow Lite library converter. Finally, a 4756 bytes matrix was obtained. This was stored in the development board in order to be able to carry out the classification.
Once the model was deployed on the development board, the firmware was capable of sampling, storing, processing, and classifying the data in real time was developed, as shown in Figure 20. The code was written in C++ (https://github.com/kaviles22/EMG_SignalClassification.git).
Figure 20. Real time classification flow chart.
  • EMG data register using MyoWare sensors: data logging using three MyoWare sensors and a microcontroller, at 1 kHz in 2-second windows.
  • Signal Filtering: filter the signal noise using the RMS envelope technique.
  • Signal Normalization: normalize the three EMG signals to avoid bias during training.
  • Feature extraction: extract features (RMS value in this case) in 500 ms windows.
  • Three classes classification: make predictions using the compiled and reduced model, previously loaded in the microcontroller.
  • Drive Motors: depending on the result of the classification, activate the motors to perform the three actions, depending on the muscle task:
    Wrist Flexion: closes the prosthesis clamp.
    Wrist extension: opens the prosthesis clamp.
    Making a fist: rotates the prosthesis clamp.

4.2. Temporal Window

The system had to have a response time fast enough to be employed in the user’s daily routine. To find this optimal time, all processing steps were run, changing the size of the time window of the EMG signals. Finally, a graph showing the different temporal window sizes vs. the accuracy of the model was obtained, as shown in Figure 21. After analyzing the graph, it was concluded that the most optimal time window duration was 2 s. This complies with a relatively short response time and a relatively high accuracy.
Figure 21. Window size vs. model accuracy.

4.3. Machine Learning Model Testing

To evaluate the artificial intelligence model, the 10-fold cross validation technique was used. The results are shown in Table 11.
Table 11. Model evaluation metrics (multiclass model).
The model classified approximately 78.67% of the classes correctly. When testing the model in real time on the microcontroller, it was preferred not to have many false positives, since in this case, the hand would perform movements without the client’s consent. For these cases in particular, the model had a high performance, since about 80.21% of the positive predictions were correct (precision), meaning that most of the movements performed were voluntary. Approximately 75.67% of the actions were correctly detected (recall).
Despite having a high accuracy, the loss was high and the model made large errors a few times. As can be seen in Figure 22b, there was both bias and variance in the model. To attack this problem, it was necessary to train the model using more data or to use regularization techniques in the intermediate layers. However, as a first instance, it was decided to analyze the behavior of the model in the different classes.
Figure 22. Model performance graphs (a) accuracy graph (b) loss graph.
From the results obtained in Figure 23, it was concluded that the model failed to accurately identify class 2. On the other hand, the accuracies of classes 1 and 3 were relatively high.
Figure 23. Confusion matrix (multiclass model). Task 1: wrist flexion, Task 2: fist, Task 3: wrist extension.
For this reason, it was decided to analyze the behavior of a binary classification model using only classes 1 and 3. Accuracy and loss during training of the binary model are shown in Figure 24a and Figure 24b, respectively. The loss plot shows that the error during training and testing remained close, thus ruling out any possibility of overfitting or underfitting. In turn, this indicates that the bias and variance of the model were low, resulting in a model capable of generalizing effectively. The accuracy reached 100% in less than 300 epochs, while the loss was approximately 0.07. The results are shown in Table 12.
Figure 24. Model performance graphs (binary model). (a) Accuracy graph (b) Loss graph.
Table 12. Model evaluation metrics (binary model).
Class 2, associated with the action of making a fist, was not classified efficiently. Unlike the other two tasks, which are opposite actions, flexion and extension, they activate two different muscles. However, class 2 could have been confused with one of the other two muscular actions. Although the action was not the same, it was noticed that the activation of the muscles was very similar.

4.4. FirmWare Analysis

The firmware deployed on the ESP32 C3 development board took up 49% of program storage (650 Kb) in memory. The Tensorflow Lite model occupied 4.75 Kb.

4.5. Mechanical Analysis

The designed prototype is shown in Figure 25. This consists of the flexion–extension mechanism formed by the group of links connected to a servomotor and a turning base that houses another servomotor for rotation. The turning base also functions as a support for the aforementioned assembly.
Figure 25. Renderings of the robotic prosthesis: (1) Flexion–extension mechanism (2) Turning base.

4.5.1. Dynamic Analysis

Prior to 3D printing the prototype, a dynamic analysis was performed to observe the behavior of the torque when the prosthesis is subjected to the maximum load during the retraction process. For this, the flexion–extension mechanism was isolated, since this part of the prosthesis is the one that interacts directly with the load. This sub-assembly was imported to Autodesk Inventor 2022 to prepare the joints and forces for the dynamic simulation as can be seen in Figure 26.
Figure 26. Flexion–extension mechanism in Autodesk Inventor 2022 dynamic simulation environment.
Forces 1 and 2 observed in Figure 26 were located at the fingertips, as shown in Figure 27. The joints of interest for the analysis were the coupling joint connected to the motor shaft and the link connected to the coupling directly, defined as Crank #1 in Chapter 2, since these could be used to test whether the required torque would exceed that established during the design stage.
Figure 27. Points of interest: (1) Location of loads (2) Link joints connected to the motor coupling (3) Motor coupling joint. Source: Own elaboration.
As final step prior to simulation, the displacement curve for the motion that would be imposed on the motor coupling joint was defined. Limits were defined according to the angular positions for the maximum opening and closing, while the slope was approximated to the speed specifications presented in the motor data sheet as shown in Figure 28.
Figure 28. Displacement curve for the imposed movement.
The simulation results are presented in Figure 29. As can be seen, in no joints are the starting moments considerably higher than the required moment; it is assumed that the servomotor will be able to overcome the initial conditions and move the mechanism. Consequently, for the calculation of the safety factor, the starting moment was considered since it represents the maximum moment required to maintain the coupling in equilibrium. This is presented in Table 13 together with the safety factor obtained based on the motor torque (20 kgfcm or 1.961 Nm).
Figure 29. Moment curve for Crank #1 (left) and coupling (right).
Table 13. Dynamic simulation results.

4.5.2. FEA

Once the dynamic analysis was completed, the resulting reactions on the joints of Crank #1 at the beginning of the movement were exported, as shown in the figure, since at this point they reached their maximum value and would allow establishing the safety factor of the link for the worst operating scenario.
With these reactions, displacements due to the load were observed in the order of 10 4 mm and a safety factor of 15 along the entire body of the link (see Figure 30), so no deformations or fatigue failures are expected due to the maximum load during operation. However, in the event that the hand is subjected to a load above that expected and beyond that allowed by the safety factor, failures would be expected to occur in the lower right side of the link.
Figure 30. Safety factor (left) and link displacement (right).

4.5.3. Load Tests

The prosthesis was subjected to different load tests—see Figure 31—to observe the functioning of the mechanism and to check that it was capable of holding the established maximum load of 500 g, as well as holding delicate and small objects.
Figure 31. Load tests.

4.5.4. Prototype Weight

To verify that the maximum weight was not exceeded, the printed prototype was placed on a digital scale as shown in the Figure 32, obtaining a total weight of 429 g.
Figure 32. Prototype weight.

5. Conclusions and Future Works

The design of a three-finger prosthesis capable of performing precision and strength grips was presented. This design utilizes a six-bar mechanism to alternate between its position limits and is capable of holding the target load of 500 g, as well as gripping small objects that require greater precision such as pens or screwdrivers. In terms of gripping force, it was estimated that based on the selected motor and the length of the fingers, it is capable of holding the target load of 500 g. It would reach 23 N, which is within the average range of research prototypes.
Additionally, this design was validated by a dynamic simulation of the opening and closing of the gripper and a finite element analysis on the most important link, both carried out using Autodesk Inventor 2022. Based on the resulting starting torque from the dynamic simulation, it was determined that the selected motor was adequate, ensuring that the mechanism will not suffer unforeseen displacements due to the load. On the other hand, the connecting link between the motor coupling and the finger link did not show any significant deformations or fatigue failures, so its design is adequate for the target load. To reduce the weight of the prototype, as well as the cost of printing while maintaining structural integrity, we plan to simplify the covers and turning base in future iterations using the shape optimization tool available in Fusion 360.
The gyroscope, which was used during data collection, did not provide relevant information to the problem. This is due to the fact that during the recording it was ensured that the arm maintained a single position and angle. Therefore, the sensor did not register significant movement changes. Its integration could be useful to explore future topics.
The sampling frequency used in the experiment (1 kHz) stood out compared to commercial EMG signal collection systems, which oscillate at around 128 Hz. This allowed us to work with smaller time windows during preprocessing, to filter the signal without losing representative information, and to characterize the signal efficiently by calculating the RMS values. Additionally, the latter technique allowed us to maintain the temporality of the signals.
Although the overall accuracy of the algorithm was 80.21%, the performance of the model with respect to the action of making a fist was specifically poor compared to the other actions, with an accuracy of 70%. Based on observations, in healthy subjects, this particular action involves both the inner and outer forearm muscle. However, in the test subject, this same action involved only the inner muscle. For example, the wrist flexion and extension actions are opposite actions that activate mostly opposite muscles of the arm: brachioradialis muscle and flexor carpi ulnaris. Therefore, it is easier to discriminate between them. As a solution to this challenge, it is proposed to train the algorithm using more data for it to generalize better across actions.
Since the hardware designed uses the ESP32-C3, its RISC-V based architecture presents advantages in the storage and reading of data in memory, allowing us to obtain a sampling rate of 1kHz for each of the nine variables of interest (three EMG sensors, three acceleration axes, three gyroscope axes). The designed electronic circuit was able to collect data for the training stage and in turn, was used for real-time operation to control the gripper. The ESP32-C3 development board was able to process the EMG signals and drive the gripper servomotors according to the outputs of the AI model.
Unlike EMG systems that may use signal intensity as a trigger, this system is based on an analysis of signal behavior; this provides robustness to the system. If a user presents any symptom of muscle fatigue due to prolonged use, the system would be able to characterize the shape of the signal, regardless of its relative intensity.
Through the analysis of the accuracy in different time windows, it was concluded that using a larger one during data recording does not necessarily represent a significant increase in the accuracy of the model. This is due to the fact that EMG MyoWare sensors have an active filter which prevents the signal from remaining active for a long period of time. Therefore, the most relevant gait is found in the activation of the muscle and not in its deactivation. Thus, reducing the time window could help to reduce the amount of not representative data.
The inference algorithm was able to efficiently classify 95.13% of two of the classes using data collected from one subject. The online classification time of the system was 0.08 s, while a data recording of 2 s of duration was needed, giving a total prosthesis reaction time of 2.18. By using a threshold to consider only predictions with an accuracy greater than 70%, we were able to reduce the number of erroneous predictions between classes. The priority was to reduce false positives as much as possible. In other words, when the subject performed an action, he preferred to keep the gripper still rather than having it move incorrectly. In this case, bias or variance phenomena was not detected in the model as training and testing errors were low.
The project was framed by certain specific objectives and met all the requirements. However, we recognize that the applications are limited to the framework of the project. The hand design could be improved to endure more weight, particularly on the wrist. Regarding the portability, as a future work, it is proposed to design a much more light and compact circuit that can be easily carried on the subject’s arm. Additionally, it is proposed to redesign the connections of the development board in order to reduce the noise generated in the EMG signals, due to physical factors, as much as possible. Regarding the artificial intelligence model used, it is proposed to evaluate the use of deep neural networks capable of generalizing between data from different test subjects, reducing the preprocessing time, and increasing the accuracy of the model. In addition use of recurrent neural networks to explore the capability of expanding the time window analysis can be studied.
For more reality-oriented implementations, it is proposed to analyze the use of the gyroscope. In addition, since the test subject does not remain static, but requires the use of the prosthesis while in motion, it is proposed to use dry electrodes, since adhesive sensors tend to cause skin irritation.

Author Contributions

K.A.-M.: conceptualization, methodology, validation, formal analysis, investigation, resources, data curation, writing—review, editing, and visualization; N.G.G.-L.: conceptualization, methodology, validation, formal analysis, investigation, resources, data curation, writing—review, editing, and visualization; V.A.: conceptualization, methodology, investigation, supervision, writing—original draft preparation, and funding acquisition; L.L.L.-L.: methodology, investigation, supervision, writing—review and editing, and formal analysis; D.H.P.-O.: methodology, investigation, resources, writing—review and editing, supervision, and project administration. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

An application form for ethics approval alongside with the corresponding supporting documents was timely submitted by the authors and subsequently approved by the Institutional Review Board (IRB) of the Smart Data Analysis Systems Group (Meeting minutes IBR-SG-2022-001).

Data Availability Statement

Data regarding images and annotations can be accessed at: repository name: ieee-dataport; data identification number: https://doi.org/10.21227/ppsg-g354; direct URL to data: https://ieee-dataport.org/documents/emg1k-dataset; accessed date: 15 November 2022.

Acknowledgments

The authors also are greatly grateful by the support given by the SDAS Research Group (www.sdas-group.com) as well as the internal editor J. Mejía-Ordóñez for the manuscript reviewing and editing.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cianchetti, M.; Laschi, C.; Menciassi, A.; Dario, P. Biomedical applications of soft robotics. Nat. Rev. Mater. 2018, 3, 143–153. [Google Scholar] [CrossRef]
  2. Amsan, A.N.; Nasution, A.K.; Ramlee, M.H. A short review on the cost, design, materials and challenges of the prosthetics leg development and usage. In Proceedings of the International Conference of CELSciTech 2019—Science and Technology Track (ICCELST-ST 2019), Pekanbaru, Indonesia, 26 September 2019; pp. 59–64. [Google Scholar]
  3. Aman, M.; Sporer, M.E.; Gstoettner, C.; Prahm, C.; Hofer, C.; Mayr, W.; Farina, D.; Aszmann, O.C. Bionic hand as artificial organ: Current status and future perspectives. Artif. Organs 2019, 43, 109–118. [Google Scholar] [CrossRef]
  4. Cordella, F.; Ciancio, A.L.; Sacchetti, R.; Davalli, A.; Cutti, A.G.; Guglielmelli, E.; Zollo, L. Literature Review on Needs of Upper Limb Prosthesis Users. Front. Neurosci. 2016, 10, 209. [Google Scholar] [CrossRef] [PubMed]
  5. Coba, G. El Ingreso Promedio en Ecuador cae y las Horas de Trabajo Suben Ligeramente. Premicias. Available online: https://www.primicias.ec/noticias/economia/ingreso-promedio-disminucion-horas-trabajo-suben/ (accessed on 15 April 2023).
  6. Parajuli, N.; Sreenivasan, N.; Bifulco, P.; Cesarelli, M.; Savino, S.; Niola, V.; Esposito, D.; Hamilton, T.J.; Naik, G.R.; Gunawardana, U.; et al. Real-Time EMG Based Pattern Recognition Control for Hand Prostheses: A Review on Existing Methods, Challenges and Future Implementation. Sensors 2019, 19, 4596. [Google Scholar] [CrossRef] [PubMed]
  7. Saikia, A.; Mazumdar, S.; Sahai, N.; Paul, S.; Bhatia, D.; Verma, S.; Rohilla, P.K. Recent advancements in prosthetic hand technology. J. Med. Eng. Technol. 2016, 40, 255–264. [Google Scholar] [CrossRef] [PubMed]
  8. Amsüss, S.; Goebel, P.M.; Jiang, N.; Graimann, B.; Paredes, L.; Farina, D. Self-correcting pattern recognition system of surface EMG signals for upper limb prosthesis control. IEEE Trans. Biomed. Eng. 2013, 61, 1167–1176. [Google Scholar] [CrossRef] [PubMed]
  9. Elbreki, A.M.; Ramdan, S.; Mohamed, F.; Alshari, K.; Rajab, Z.; Elhub, B. Practical Design of an Upper Prosthetic Limb Using Three Dimensional Printer with an Artificial Intelligence Based Controller. In Proceedings of the 2022 International Conference on Engineering & MIS (ICEMIS), Istanbul, Turkey, 4–6 July 2022; pp. 1–6. [Google Scholar]
  10. Kerdjidj, O.; Amara, K.; Harizi, F.; Boumridja, H. Implementing Hand Gesture Recognition using EMG on the Zynq Circuit. IEEE Sens. J. 2023, 9, 10054–10061. [Google Scholar] [CrossRef]
  11. Hye, N.M.; Hany, U.; Chakravarty, S.; Akter, L.; Ahmed, I. Artificial Intelligence for sEMG-based Muscular Movement Recognition for Hand Prosthesis. IEEE Access 2023, 11, 38850–38863. [Google Scholar] [CrossRef]
  12. Lee, S.; Kim, M.O.; Kang, T.; Park, J.; Choi, Y. Knit band sensor for myoelectric control of surface EMG-based prosthetic hand. IEEE Sens. J. 2018, 18, 8578–8586. [Google Scholar] [CrossRef]
  13. Maat, B.; Smit, G.; Plettenburg, D.; Breedveld, P. Passive prosthetic hands and tools: A literature review. Prosthetics Orthot. Int. 2018, 42, 66–74. [Google Scholar] [CrossRef]
  14. Ottobock. Passive Arm Prostheses: Natural-Looking Arm Prostheses. Available online: https://www.ottobockus.com/prosthetics/upper-limb-prosthetics/solution-overview/passive-arm-prostheses/ (accessed on 24 May 2022).
  15. ArmDynamics. Prosthetic Options. Available online: https://www.armdynamics.com/our-care/prosthetic-options (accessed on 27 May 2022).
  16. Almao, D. Brazos Mecánicos: Tipos, Aplicaciones, Fabricantes y los Brazos Robóticos Más Avanzados del Mercado. Available online: https://www.cinconoticias.com/brazos-mecanicos/ (accessed on 30 May 2022).
  17. Farina, D. Retroalimentación Artificial a los Usuarios de Prótesis. Available online: https://www.catalunyavanguardista.com/retroalimentacion-artificial-a-los-usuarios-de-protesis/ (accessed on 30 May 2022).
  18. De Luca, C. Electromyography. In Encyclopedia of Medical Devices and Instrumentation; John Wiley & Sons Inc.: Hoboken, NJ, USA, 2006. [Google Scholar]
  19. Zayia, L.C.; Tadi, P. Neuroanatomy, motor neuron. In StatPearls; StatPearls Publishing: Tampa, FL, USA, 2021. [Google Scholar]
  20. Reaz, M.; Hussain, M.; Mohd-Yasin, F. Techniques of EMG signal analysis: Detection, processing, classification and applications. Biol. Proced. Online 2006, 8, 11–35. [Google Scholar] [CrossRef] [PubMed]
  21. Ahsan, M.R.; Ibrahimy, M.I.; Khalifa, O.O. Electromygraphy (EMG) signal based hand gesture recognition using artificial neural network (ANN). In Proceedings of the 2011 4th International Conference on Mechatronics (ICOM), Kuala Lumpur, Malaysia, 17–19 May 2011; pp. 1–6. [Google Scholar]
  22. McManus, L.; De Vito, G.; Lowery, M.M. Analysis and biophysics of surface EMG for physiotherapists and kinesiologists: Toward a common language with rehabilitation engineers. Front. Neurol. 2020, 11, 576729. [Google Scholar] [CrossRef] [PubMed]
  23. Litin, S.C.; Nanda, S. Mayo Clinic Family Health Book; Time Incorporated Home Entertainment Des Moines, 2019. [Google Scholar]
  24. Mahmood, D.; Riaz, H.N.; Nisar, H. Introduction to Non-Invasive Biomedical Signals for Healthcare. In Advances in Non-Invasive Biomedical Signal Sensing and Processing with Machine Learning; Springer International Publishing: Cham, Switzerland, 2023. [Google Scholar]
  25. What is EMG Sensor, Myoware and How to Use with Arduino? Available online: https://www.seeedstudio.com/blog/2019/12/29/what-is-emg-sensor-myoware-and-how-to-use-with-arduino/ (accessed on 29 May 2022).
  26. Farnsworth, B. What Is EMG (Electromyography) and How Does It Work? Available online: https://imotions.com/blog/electromyography-101/ (accessed on 29 May 2022).
  27. Phinyomark, A.; Khushaba, R.N.; Scheme, E. Feature Extraction and Selection for Myoelectric Control Based on Wearable EMG Sensors. Sensors 2018, 18, 1615. [Google Scholar] [CrossRef]
  28. Wilson, A.W.; Losier, Y.G.; Parker, P.A.; Lovely, D.F. A bus-based smart myoelectric electrode/amplifier—System requirements. IEEE Trans. Instrum. Meas. 2011, 60, 3290–3299. [Google Scholar] [CrossRef]
  29. Chim, H. Hand and wrist anatomy and biomechanics: A comprehensive guide. Plast. Reconstr. Surg. 2017, 140, 865. [Google Scholar] [CrossRef] [PubMed]
  30. Levangie, P.K.; Norkin, C.C.; Lewek, M.D. Joint Structure and Function: A Comprehensive Analysis; F.A. Davis: Philadelphia, PA, USA, 2019. [Google Scholar]
  31. Artificial Intelligence (AI). IBM Cloud Education. Available online: https://www.ibm.com/cloud/learn/what-is-artificial-intelligence (accessed on 10 October 2022).
  32. Côté-Allard, U.; Fall, C.L.; Drouin, A.; Campeau-Lecours, A.; Gosselin, C.; Glette, K.; Laviolette, F.; Gosselin, B. Deep Learning for Electromyographic Hand Gesture Signal Classification Using Transfer Learning. IEEE Trans. Neural Syst. Rehabil. Eng. 2019, 27, 760–771. [Google Scholar] [CrossRef] [PubMed]
  33. Asanza, V.; Peláez, E.; Loayza, F.; Lorente-Leyva, L.L.; Peluffo-Ordóñez, D.H. Identification of Lower-Limb Motor Tasks via Brain–Computer Interfaces: A Topical Overview. Sensors 2022, 22, 2028. [Google Scholar] [CrossRef]
  34. Mambou, S.J.; Maresova, P.; Krejcar, O.; Selamat, A.; Kuca, K. Breast Cancer Detection Using Infrared Thermal Imaging and a Deep Learning Model. Sensors 2018, 18, 2799. [Google Scholar] [CrossRef]
  35. Daimiel Naranjo, I.; Gibbs, P.; Reiner, J.S.; Lo Gullo, R.; Sooknanan, C.; Thakur, S.B.; Jochelson, M.S.; Sevilimedu, V.; Morris, E.A.; Baltzer, P.A.T.; et al. Radiomics and Machine Learning with Multiparametric Breast MRI for Improved Diagnostic Accuracy in Breast Cancer Diagnosis. Diagnostics 2021, 11, 919. [Google Scholar] [CrossRef]
  36. Gulum, M.A.; Trombley, C.M.; Kantardzic, M. A Review of Explainable Deep Learning Cancer Detection Models in Medical Imaging. Appl. Sci. 2021, 11, 4573. [Google Scholar] [CrossRef]
  37. Zivkovic, M.; Bacanin, N.; Antonijevic, M.; Nikolic, B.; Kvascev, G.; Marjanovic, M.; Savanovic, N. A Hybrid CNN and XGBoost Model Tuned by Modified Arithmetic Optimization Algorithm for COVID-19 Early Diagnostics from X-ray Images. Electronics 2022, 11, 3798. [Google Scholar] [CrossRef]
  38. Toro-Ossaba, A.; Jaramillo-Tigreros, J.; Tejada, J.C.; Peña, A.; López-González, A.; Castanho, R.A. LSTM Recurrent Neural Network for Hand Gesture Recognition Using EMG Signals. Appl. Sci. 2022, 12, 9700. [Google Scholar] [CrossRef]
  39. Valdivieso Caraguay, Á.L.; Vásconez, J.P.; Barona López, L.I.; Benalcázar, M.E. Recognition of Hand Gestures Based on EMG Signals with Deep and Double-Deep Q-Networks. Sensors 2023, 23, 3905. [Google Scholar] [CrossRef] [PubMed]
  40. Xu, Y.; Barbosa, P.C.; da Cunha Neto, J.S.; Zhang, L.; Shanmuganathan, V.; de Albuquerque, V.H.C.; Pasupathi, S. Development of intelligent and integrated technology for pattern recognition in EMG signals for robotic prosthesis command. Expert Syst. 2023, 40, e13109. [Google Scholar] [CrossRef]
  41. Berry, M.W.; Mohamed, A.; Yap, B.W. Supervised and Unsupervised Learning for Data Science. In Unsupervised and Semi-Supervised Learning for Data Science; Emre Celebi, M., Ed.; Springer Nature: Berlin/Heidelberg, Germany, 2019. [Google Scholar]
  42. Dokic, K.; Martinovic, M.; Mandusic, D. Inference speed and quantisation of neural networks with TensorFlow Lite for Microcontrollers framework. In Proceedings of the 2020 5th South-East Europe Design Automation, Computer Engineering, Computer Networks and Social Media Conference (SEEDA-CECNSM), Corfu, Greece, 25–27 September 2020; pp. 1–5. [Google Scholar]
  43. Tensorflow Lite. IBM Cloud Education. Available online: https://www.tensorflow.org/lite/guide?hl=es-419 (accessed on 10 August 2022).
  44. Belter, J.; Segil, J.; Dollar, A.; Weir, R. Mechanical design and performance specifications of anthropomorphic prosthetic hands: A review. J. Rehabil. Res. Dev. 2013, 50, 599–618. [Google Scholar] [CrossRef]
  45. Ottobock. SensorHand Speed. Available online: https://shop.ottobock.us/Prosthetics/Upper-Limb-Prosthetics/Myo-Hands-and-Components/Myo-Terminal-Devices/SensorHand-Speed/p/8E38~58#product-documents-section (accessed on 30 October 2022).
  46. Liow, L.; Clark, A.B.; Rojas, N. OLYMPIC: A Modular, Tendon-Driven Prosthetic Hand with Novel Finger and Wrist Coupling Mechanisms. IEEE Robot. Autom. Lett. 2019, 5, 299–306. [Google Scholar] [CrossRef]
  47. Pancholi, S.; Agarwal, R. Development of low cost EMG data acquisition system for Arm Activities Recognition. In Proceedings of the 2016 International Conference on Advances in Computing, Communications and Informatics (ICACCI), Jaipur, India, 21–24 September 2016; pp. 2465–2469. [Google Scholar]
  48. Pizzolato, S.; Tagliapietra, L.; Cognolato, M.; Reggiani, M.; Müller, H.; Atzori, M. Comparison of six electromyography acquisition setups on hand movement classification tasks. PLoS ONE 2017, 12, e0186132. [Google Scholar] [CrossRef]
  49. Fang, Y.; Yang, J.; Zhou, D.; Ju, Z. Modelling EMG driven wrist movements using a bio-inspired neural network. Neurocomputing 2022, 470, 89–98. [Google Scholar] [CrossRef]
  50. Nazmi, N.; Rahman, M.A.A.; Yamamoto, S.I.; Ahmad, S.A. Walking gait event detection based on electromyography signals using artificial neural network. Biomed. Signal Process. Control. 2019, 47, 334–343. [Google Scholar] [CrossRef]
  51. Toledo-Pérez, D.C.; Rodríguez-Reséndiz, J.; Gómez-Loenzo, R.A.; Jauregui-Correa, J.C. Support vector machine-based EMG signal classification techniques: A review. Appl. Sci. 2019, 9, 4402. [Google Scholar] [CrossRef]
  52. Too, J.; Abdullah, A.R.; Zawawi, T.T.; Saad, N.M.; Musa, H. Classification of EMG signal based on time domain and frequency domain features. Int. J. Hum. Technol. Interact. (IJHaTI) 2017, 1, 25–30. [Google Scholar]
  53. Javaid, H.A.; Rashid, N.; Tiwana, M.I.; Anwar, M.W. Comparative analysis of emg signal features in time-domain and frequency-domain using myo gesture control. In Proceedings of the 2018 4th International Conference on Mechatronics and Robotics Engineering, Valenciennes, France, 7–11 February 2018; pp. 157–162. [Google Scholar]
  54. Tuncer, T.; Dogan, S.; Subasi, A. Surface EMG signal classification using ternary pattern and discrete wavelet transform based feature extraction for hand movement recognition. Biomed. Signal Process. Control 2020, 58, 101872. [Google Scholar] [CrossRef]
  55. Lima, A.A.M.; Araujo, R.M.; dos Santos, F.A.G.; Yoshizumi, V.H.; de Barros, F.K.; Spatti, D.H.; Liboni, L.H.B.; Dajer, M.E. Classification of hand movements from emg signals using optimized mlp. In Proceedings of the 2018 International Joint Conference on Neural Networks (IJCNN), Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–7. [Google Scholar]
  56. Caesarendra, W.; Lekson, S.U.; Mustaqim, K.A.; Winoto, A.R.; Widyotriatmo, A. A classification method of hand EMG signals based on principal component analysis and artificial neural network. In Proceedings of the 2016 International Conference on Instrumentation, Control and Automation (ICA), Bandung, Indonesia, 29–31 August 2016; pp. 22–27. [Google Scholar]
  57. Mukhopadhyay, A.K.; Samui, S. An experimental study on upper limb position invariant EMG signal classification based on deep neural network. Biomed. Signal Process. Control 2020, 55, 101669. [Google Scholar] [CrossRef]
  58. Buongiorno, D.; Cascarano, G.D.; De Feudis, I.; Brunetti, A.; Carnimeo, L.; Dimauro, G.; Bevilacqua, V. Deep learning for processing electromyographic signals: A taxonomy-based survey. Neurocomputing 2021, 452, 549–565. [Google Scholar] [CrossRef]
  59. Hirt, B.; Seyhan, H.; Wagner, M.; Zumhasch, R. Anatomie und Biomechanik der Hand; Georg Thieme: Stuttgart, Alemany, 2014. [Google Scholar]
  60. Norton, R.L. Design of Machinery: An Introduction to the Synthesis and Analysis of Mechanisms and Machines; McGraw Hill: New York, NY, USA, 2019. [Google Scholar]
  61. Visconti, P.; Gaetani, F.; Zappatore, G.A.; Primiceri, P. Technical Features and Functionalities of Myo Armband: An Overview on Related Literature and Advanced Applications of Myoelectric Armbands Mainly Focused on Arm Prostheses. Int. J. Smart Sens. Intell. Syst. 2018, 11, 1–25. [Google Scholar] [CrossRef]
  62. Technologies, A. 3-Lead Muscle/Electromyography Sensor for Microcontroller Applications. 2015. Available online: https://www.mouser.com/datasheet/2/813/MyowareUserManualAT-04-001-1223951.pdf (accessed on 30 October 2022).
  63. Ghazwan, A.; Forrest, S.M.; Holt, C.A.; Whatling, G.M. Can activities of daily living contribute to EMG normalization for gait analysis? PLoS ONE 2017, 12, e0174670. [Google Scholar] [CrossRef]
  64. Mesa, I.; Rubio, A.; Tubia, I.; De No, J.; Diaz, J. Channel and feature selection for a surface electromyographic pattern recognition task. Expert Syst. Appl. 2014, 41, 5190–5200. [Google Scholar] [CrossRef]
  65. Asanza, V.; Avilés-Mendoza, K.; Trivino-Gonzalez, H.; Rosales-Uribe, F.; Torres-Brunes, J.; Loayza, F.L.; Peláez, E.; Cajo, R. SSVEP-EEG Signal Classification based on Emotiv EPOC BCI and Raspberry Pi. IFAC-PapersOnLine 2021, 54, 388–393. [Google Scholar] [CrossRef]
  66. Constantine, A.; Asanza, V.; Loayza, F.L.; Peláez, E.; Peluffo-Ordóñez, D. BCI System using a Novel Processing Technique Based on Electrodes Selection for Hand Prosthesis Control. IFAC-PapersOnLine 2021, 54, 364–369. [Google Scholar] [CrossRef]
  67. Tinoco-Varela, D.; Ferrer-Varela, J.A.; Cruz-Morales, R.D.; Padilla-García, E.A. Design and Implementation of a Prosthesis System Controlled by Electromyographic Signals Means, Characterized with Artificial Neural Networks. Micromachines 2022, 13, 1681. [Google Scholar] [CrossRef]
  68. Subasi, A.; Yilmaz, M.; Ozcalik, H.R. Classification of EMG signals using wavelet neural network. J. Neurosci. Methods 2006, 156, 360367. [Google Scholar] [CrossRef] [PubMed]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.