Next Article in Journal
A Quantitative Model for Optical Coherence Tomography
Next Article in Special Issue
Social Robot Navigation Tasks: Combining Machine Learning Techniques and Social Force Model
Previous Article in Journal
Application of Dual-Frequency Self-Injection Locked DFB Laser for Brillouin Optical Time Domain Analysis
Previous Article in Special Issue
Robust Tightly Coupled Pose Measurement Based on Multi-Sensor Fusion in Mobile Robot System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Biosignal-Based Human–Machine Interfaces for Assistance and Rehabilitation: A Survey

1
Department of Electrical Engineering and Information Technologies, Polytechnic and Basic Sciences School, University of Naples “Federico II”, 80125 Naples, Italy
2
School of Engineering, Design and Built Environment, Western Sydney University, Penrith, NSW 2747, Australia
3
The MARCS Institute, Western Sydney University, Penrith, NSW 2751, Australia
4
The Adelaide Institute for Sleep Health, Flinders University, Bedford Park, SA 5042, Australia
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(20), 6863; https://doi.org/10.3390/s21206863
Submission received: 31 August 2021 / Revised: 30 September 2021 / Accepted: 12 October 2021 / Published: 15 October 2021

Abstract

:
As a definition, Human–Machine Interface (HMI) enables a person to interact with a device. Starting from elementary equipment, the recent development of novel techniques and unobtrusive devices for biosignals monitoring paved the way for a new class of HMIs, which take such biosignals as inputs to control various applications. The current survey aims to review the large literature of the last two decades regarding biosignal-based HMIs for assistance and rehabilitation to outline state-of-the-art and identify emerging technologies and potential future research trends. PubMed and other databases were surveyed by using specific keywords. The found studies were further screened in three levels (title, abstract, full-text), and eventually, 144 journal papers and 37 conference papers were included. Four macrocategories were considered to classify the different biosignals used for HMI control: biopotential, muscle mechanical motion, body motion, and their combinations (hybrid systems). The HMIs were also classified according to their target application by considering six categories: prosthetic control, robotic control, virtual reality control, gesture recognition, communication, and smart environment control. An ever-growing number of publications has been observed over the last years. Most of the studies (about 67%) pertain to the assistive field, while 20% relate to rehabilitation and 13% to assistance and rehabilitation. A moderate increase can be observed in studies focusing on robotic control, prosthetic control, and gesture recognition in the last decade. In contrast, studies on the other targets experienced only a small increase. Biopotentials are no longer the leading control signals, and the use of muscle mechanical motion signals has experienced a considerable rise, especially in prosthetic control. Hybrid technologies are promising, as they could lead to higher performances. However, they also increase HMIs’ complexity, so their usefulness should be carefully evaluated for the specific application.

1. Introduction

A Human–Machine Interface (HMI) establishes a connection between a person and a device: sensors decipher human intentions and control machine actions, usually allowing real-time, bidirectional interactions. In a broad sense, we can consider the PC keyboard and monitor or even a simple switch as HMIs. Modern HMIs are software-based and replace manually activated controls, providing sophisticated interaction with machinery. HMIs are widely used in industrial control systems, automotive, aviation, military, etc. [1]. User movements, or movement intention, are typically used to interpret the person’s will to operate a device. To this end, multiple and varied sensors have been developed to monitor specific user’s activities, such as hand, eyes, joint, limbs movements, etc. Rotational, displacement, force, pressure, strain, acceleration, and inertial sensors can detect a user’s motion when opportunely connected to the body [1,2,3]. Since muscles generate human motion, physiological signal related to muscle contraction such as electromyography (EMG), mechanomyography (MMG), and force myography (FMG) can be successfully used for HMI [3,4]. Recently, electroencephalography (EEG), able to detect motor brain function, has increasingly been proposed as HMI control signals (these techniques are referred to as Brain–Computer Interfaces (BCIs) [5,6,7,8]). Human movement can also be captured by cameras (the so-called image-based HMIs), which do not require any physical contact with the user [9,10]. Different sensors can be combined to obtain greater sensitivity and specificity in recognizing the user’s intention. Therefore, we can define such mix as hybrid HMI control [11,12,13].
Signals from sensors typically need to be processed for robust recognition of the user’s intention. This processing can widely vary in complexity and ranges from simple thresholding to complex machine learning techniques. Today, machine learning has developed into a scientific branch of its own. For this reason, this paper does not explicitly review this topic.
Most HMIs provide feedback to the user offering visual, acoustic, tactile sensations, which help interaction [14,15,16]. Feedback provides information on the controlled system’s performance to the user. It can consist of a simple light or sound indication or graphical representations or create complex visual–acoustic experiences like virtual reality or tactile sensations (e.g., via vibrations or forces delivery) such as in haptic devices.
In the medical field, biosignal-based HMIs have increasingly been used for assistance and rehabilitation. An “assistive technology” is any system or object used to enhance, maintain, or improve the capabilities of a disabled individual and, more generally, any technology that allows accomplishing something that generally cannot be accomplished. “Rehabilitation” is defined as the physical restoration of a sick or disabled person by therapeutic measures and re-education to participate in everyday life activities within the limitations of the person’s physical disability [17]. People with severe disabilities enormously benefit from using these new HMIs for assistance and rehabilitation purposes. Clear examples are subjects with paraplegia or quadriplegia; those affected by neuromuscular disorders [5,18,19,20] such as Muscular Dystrophy (MD), Amyotrophic Lateral Sclerosis (ALS), or Multiple Sclerosis (MS); people with Spinal Cord Injury (SCI) or Cerebral Palsy (CP); or even stroke survivors and amputees. Literature reviews in these fields are currently limited to particular applications (e.g., prosthetic control, BCI, exergaming, etc.) or focused on specific biosignals. Table 1 provides a summary of some recent surveys regarding assistive and rehabilitative HMIs. As an example, Mohebbi et al. [21] proposed a review about human–robot interaction in assistive and rehabilitation robotics, while Frisoli et al. [22] focused on wearable technologies and, in particular, on a robotic exoskeleton for assistance in performing activities of daily living (ADL). Baniqued et al. [7] presented a review study on BCI robotics for motor rehabilitation of hand movements after stroke. Different surveys such as [9,10,23,24,25,26,27,28] focused specifically on the state-of-the-art and control strategies of upper limb prostheses, while further reviews presented exergaming applications for rehabilitation neuromotor functions [29,30,31,32].
To the best of our knowledge, a broad overview of the current research trends on assistive and rehabilitation HMIs is missing. The current survey aims to review the extensive literature of the last two decades regarding biosignal-based HMIs for assistance and rehabilitation to outline state-of-the-art and identify emerging technologies and potential future research trends.

2. Survey Method

Over four months (April 2021 to July 2021), we surveyed Google Scholar, Scopus, PubMed, IEEE Xplore, MDPI, Frontiers, and ScienceDirect to identify HMIs with applications as assistive technology and/or rehabilitation.
Using the following keywords: “Assistive HMI”, “Rehabilitation HMI”, “Prosthetic control”, “Exoskeleton control”, “Assistive robot”, “Rehabilitation robot”, and “Exergaming AND rehabilitation”, we restricted our search to the last two decades, including only papers written in English and fully peer-reviewed.
The initially selected papers were then individually screened in three levels: (1) title, (2) abstract, and (3) full-text, to verify the research presented, the results obtained, and their applications (even those potential). Any paper that did not appear to use biosignals to control HMIs for assistive or rehabilitation was discarded. This process produced 144 journal papers and 37 conference papers.
Four macrocategories were considered to classify the biosignals used as a control for HMIs: (1) biopotential (EMG, EEG, etc.), (2) muscle mechanical motion (gross muscle motion, muscle vibrations, muscle–tendons movements), (3) body motion (limb or joint motion, hand gesture, eye movements), and (4) hybrid (a combination of two or more different biosignals).
The type of sensor(s), the site(s) of its/their application, and the HMI target(s) (prosthetic control, robotic control, virtual reality control, gesture recognition, communication, smart environment control) was reported for each study.
To facilitate reading, it is worth highlighting briefly the definitions of the primary devices for assistive and rehabilitative purposes that have emerged from the survey:
  • Prosthesis: an assistive device designed to replace a missing part of the body or to make a part of the body work better. Prosthetic devices commonly replace the diseased or missing eyes, arms, hands, legs, or joints [9,10,23,24,25,26,27,28].
  • Exoskeleton: an assistive/rehabilitation device that connects with the human body in a wearable way and can control the movement of joints (leg, wrist, hand, fingers), providing the person with the support and energy necessary to perform a movement (hand closing/opening, walking, etc.) [34,35,36].
  • Robotic arm: a type of mechatronic arm with functions similar to a human arm, which can provide assistance to the person in performing tasks (grabbing and moving objects, etc.) and can also be used as a rehabilitation device [37,38].
  • Smart wheelchair: a power wheelchair to which computers, sensors, and assistive technology are attached [39,40,41].
  • Smart environment: environment equipped with intelligent technologies capable of monitoring and assisting the health of people who have difficulty living independently at home [42,43].
  • Exergaming: a specific type of serious game (not designed for pure entertainment) is the so-called exergame: a human-activated video game that tracks the user’s gestures or movements and simulates them into a connected screen. It can be used as a potential rehabilitation tool to increase physical activity and improve health and physical function in patients with neuromuscular diseases [29,30,31,32].
Figure 1 shows a graphical representation of the focus of this survey, highlighting the kind of biosignal-based control and the primary devices that realize HMIs for assistance and rehabilitation.

3. HMI Control Strategies

3.1. HMI Control Based on Biopotentials

A large group of HMIs is based on the acquisition of biopotentials, such as electroencephalogram (EEG), electromyogram (EMG), electroneurogram (ENG), electrooculogram (EOG), and electroretinogram (ERG), which are generated by electrical sources in the human body [44]. Hence, they reflect the function of some organs (e.g., brain, muscles, eyes) in the form of electrical activity, providing relevant information about them [44,45]. For this reason, biopotentials are used as control signals in many biomedical HMI applications [25,42,46,47,48,49,50,51,52,53,54,55,56,57,58,59]. These biosignals have their origin in electrophysiological phenomena associated with biochemical events occurring at a cellular level. In detail, some tissues (e.g., nervous, muscular) are composed of excitable cells. At rest, an excitable cell exhibits a transmembrane potential difference in response to a diffusion gradient. Indeed, electrically, the cell membrane works like a leaky capacitor since it consists of a thin insulating layer that separates charge distributions in two conductive media, the intracellular and extracellular environments.
Furthermore, within the dielectric cellular material, the presence of ion-specific passive channels, i.e., pores, results in different ionic concentrations, which drive leakage currents to flow across the membrane. The charge and discharge current of the membrane capacitance is opposed to these ionic currents. At equilibrium, the net current flow is zero. Therefore, the resting potential is achieved. It is maintained by the action of the well-known sodium–potassium pump that actively transports sodium and potassium ions against the electrochemical gradient under the consumption of energy in the form of ATP. When properly stimulated, the excitable cell produces an action potential representing an all-or-none event. It starts only if the transmembrane potential exceeds a threshold value in a specific time interval and travels without attenuation at constant conduction velocity. Due to the electrical activity of many excitable cells, the equivalent current propagates through the surrounding volume conductor, represented by biological tissues, up to the body surface. Consequently, if electrodes are placed in the correct positions on the skin or within underlying tissues, a biopotential representing the spiking cell/tissue phenomenon can be measured [44,45,60,61].

3.1.1. EEG-Based HMIs

The EEG signal originates from the electrical activity of neurons in the brain cortex. To obtain a perceptible signal on the scalp, large populations of neurons must depolarize and repolarize simultaneously in the brain cortex. This synchronization generates the classic oscillations or waves (delta, theta, alpha, beta, and gamma) found in the EEG. In general, some oscillations (as the μ waves of the motor cortex) correspond to an idle rhythm, that is, a state of inactivity. For example, when a person performs or even thinks of acting (such as clench a fist), the idle rhythm is disrupted (e.g., event-related desynchronization) and replaced by a smaller high-frequency signal. Some EEG electrodes can detect this, and this information can be used to give commands to a device. Event-related potentials (e.g., Steady-State Visually Evoked Potentials (SSVEPs), P300) and slow cortical potentials can also be used to implement HMIs. EEG can be classified into two main types: invasive and noninvasive. Noninvasive EEG is commonly recorded by placing electrodes on the scalp.
On the other hand, invasive EEG is acquired intracranially, and it is generally referred to as intracranial EEG (iEEG). It includes both electrocorticogram (ECoG), which is performed by placing electrodes directly onto the brain surface to record electrical activity from the cerebral cortex, and the EEG signal acquired by means of depth electrodes to record electrical activity from deep brain regions, as the stereo-electroencephalogram (sEEG). Scalp EEG amplitude is much lower than other noninvasive biopotentials, and there are issues for electrode fixation and endurance [44,60,61].
Many HMIs are based on the acquisition of the EEG signal, mostly of noninvasive type. Specifically, these technologies are referred to as Brain–Machine Interfaces (BMIs) or Brain–Computer Interfaces (BCIs), since they provide an alternative interaction pathway with the surrounding environment by capturing brain waves and translating them into control signals or commands for an external device [62]. Generally, after a preprocessing phase, some meaningful features are extracted from the acquired EEG signals, and then a classification step is performed to interpret human intention. Afterward, each class is associated with a specific command. In this context, the user is often required to focus on a cognitive task (e.g., motor imagery), or an external stimulation (e.g., visual, auditory, somatosensory) is provided to induce a brain response. The resultant EEG signals are Event-Related Potentials (ERPs), the most commonly used brain waves in a BMI/BCI system [63]. For this reason, in recent years, EEG-based HMIs have found strong application in the assistance and rehabilitation field. Indeed, they represent a powerful tool for people with severe motor disabilities, who need to assist or restore their muscle function or even replace a missing limb, thus improving their quality of life. Clear examples are subjects affected by neuromuscular disorders, such as MD, ALS, MS, SCI, and CP, or even poststroke patients and amputees [5,18,19,20]. In this scenario, two main targets can be identified: robotic control [14,15,34,49,50,64,65,66,67,68,69] and prosthetic control [46,47,48,69,70,71,72,73,74,75,76,77,78].
Regarding robotic control, Song et al. proposed an efficient EEG-based method to control an upper-limb assist robot to help paralysed people perform practical activities or rehabilitation exercises. After applying a visual stimulus, the related EEG signal is classified to decode the human motor intention [49]. Similarly, Korovesis et al. developed a BCI-based system that allows the user to control the motion of a robot vehicle by converting eyes opening and closing into binary sequences, then associating then with different motor commands [50]. Antoniou et al. presented a BCI adopting an eye movement recognition method for hand-free wheelchair control [64]. Xu et al. proposed a motor imagery-based BCI system for teleoperation with tactile feedback to remotely control a robot arm in grasping and reaching tasks [14]. In this context, different studies that use EEG signals for exoskeleton control should be mentioned. Liang et al. extracted some features from EEG signals related to the shoulder’s joint flexion and extension movement. They demonstrated the existence of a relationship between changes in EMG and EEG signals, thus showing feasible to estimate from EEG the minimum torque for controlling an upper-limb exoskeleton robot [34]. He et al. demonstrated the feasibility of decoding joint kinematics and sEMG patterns from scalp EEG to control a powered exoskeleton for gait rehabilitation after stroke [35]. Moreover, Tang et al. proved the effectiveness of a BMI based on Event-Related Desynchronization/Synchronization (ERD/ERS) for controlling an upper-limb exoskeleton, which can assist people in daily living activities [79]. Randazzo et al. designed “mano”, a wearable hand exoskeleton for both assistance and neurorehabilitation purposes. In particular, it exploits a motor imagery-based BMI, which can decode flexion and extension of each finger [80]. Again, Li et al. proposed a BMI-controlled upper limb exoskeleton. The system comprises three main components: EEG signal acquisition and classification into motion commands in task space, redundant motion planning to transform motion commands in joint space, and adaptative neural control to perform manipulation tasks. In detail, visual stimuli are presented to the subject to induce brain signals (i.e., SSVEPs). Then, they are translated into motion commands for controlling the cursor on a computer screen. Afterward, cursor control in task space is converted into exoskeleton control in joint space [81].
Furthermore, López-Larraz et al. presented a BCI system to control an ambulatory exoskeleton for gait rehabilitation of SCI patients, which can benefit from an assist-as-needed paradigm based on the measurement of the stiffness parameter [67]. Xu et al. [36] proposed a BCI system to control an ankle–foot exoskeleton for stroke rehabilitation, which was proven effective for inducing cortical neuroplasticity. Kwak et al. [82] developed an EEG-controlled lower limb exoskeleton for rehabilitation purposes that is based on SSVEPs. Finally, Araujo et al. presented HERO, Hand Exoskeleton for Rehabilitation Objective, to recover flexion and extension of the fingers in patients following stroke. This novel, low-cost exoskeleton is 3D-printed on textiles and controlled by motor imagery-based BMI [83].
Regarding prosthetic control, Gao et al. developed a motor imagery-based BCI to control a prosthetic leg walking in different terrains, intending to reproduce natural human gait [46]. Gannouni et al. proposed a BCI system that can detect finger movements [47]. Fuentes-Gonzalez et al. designed an EEG-controlled prosthetic arm [48], similarly to Zhan Hong et al. [70]. Murphy et al. presented a case study of BCI for controlling a prosthetic knee in transfemoral amputees with only the EEG signals arising from movement imagination [73].
Although less common, invasive EEG-based HMIs also exist. For instance, Li et al. tested the feasibility of an HMI based on the acquisition of sEEG signals to control a prosthetic hand [74]. Morishita et al. proposed an invasive BCI-based prosthetic arm. In particular, they decoded the motor intention by estimating muscle activity from ECoGs [76]. Yanagisawa et al. presented an invasive BCI system for controlling a prosthetic hand by using ECoG signals of a poststroke patient [78]. Finally, Fatima et al. reviewed intracortical BMIs for controlling upper-limb powered muscle and robotic systems in SCI patients [19].
Furthermore, these technologies can also be used as nonverbal communication tools. For example, Kashihara et al. proposed a BCI that allows interpretation of emotions by decoding facial expressions. In detail, the user is asked to look at some face images on a monitor. The visual stimulation triggers the activation of specific brain areas. Applying a classifier to the acquired EEG signal makes it possible to capture the subject’s reaction, which is correlated to his actual emotional state [84]. In addition, a BCI can be beneficial also in case of blindness, as described in [6]. Finally, a wide range of BMI-based neuroprostheses should be mentioned [23].
Moreover, it is noteworthy that another technique for BCI/BMI applications is functional Near-Infrared Spectroscopy (fNIRS), which acquires a biosignal that measures changes in blood oxygenation under the activation of specific brain areas. For example, this kind of BCI/BMIs can be used as a nonverbal communication tool or for controlling a wearable robotic exoskeleton [26,27].
Table 2 shows a summary of the selected EEG-based HMIs.

3.1.2. EMG-Based HMIs

EMG reflects the electrical activity of skeletal muscles during voluntary contraction. The EMG signal is made of superpositions of small Motor Unit Action Potential waveforms with random phases. Hence, the EMG signal does not have a specific waveform but rather an interference pattern. Concise parameters of EMG are often used to monitor muscle contraction. The EMG time-varying RMS and the EMG linear envelope are largely used to assess the level of muscle contraction and the strength it develops (the stronger is the muscle contraction, the greater is the number and the frequency of active fibres, and, in turn, the higher is the amplitude of the EMG). Some other EMG parameters computed in the frequency domain can provide information about muscle fatigue. Since the EMG signal is specific to a single muscle, through the use of multiple electrodes, many muscles can be monitored simultaneously to decipher complex actions such as hand gestures or even movements of a single finger. EMG can be recorded noninvasively by employing skin-mounted electrodes (the so-called surface EMG (sEMG), which is by far the most used option) or invasively using implanted electrodes into the muscle (enabling more specific information on neural control) [44,60,86].
Currently, sEMG is the most widely used signal in biomedical HMI applications for prosthetic control [26,27,28,33]. It is easy to access and provides an intuitive control strategy to reproduce the function of a biological limb. This biopotential contains information about neural signals transmitted from the brain to muscles to perform a motor task. Hence, it allows for capturing the subject’s movement intention. EMG-controlled artificial limbs are referred to as myoelectric prostheses. Amputees represent the primary users of these assistive technologies. Indeed, people with a missing limb need to replace the lost function, in order not to suffer from their disability and to carry out daily activities autonomously and as naturally as possible. In the case of postamputation patients, the sEMG signal must be acquired from muscle groups of the residual limb, generally on the forearm or on the leg stump. Information associated with movements is captured from the EMG signal and converted into input commands to the prosthesis. To this aim, in recent years, machine learning algorithms, which substantially consist of feature extraction and classification steps, have gained ground in the biosignal analysis field. On this trend, Eisenberg et al. presented a novel segmentation technique that allows separating muscle contraction and rest periods of sEMG signals to perform real-time hand movements classification [53]. Tavakoli et al. proposed a simple, low-cost and efficient approach for hand gesture recognition from a single-channel sEMG. In particular, this method can recognize four hand gestures, hand closing, hand opening, wrist flexion, and double wrist flexion to control a prosthetic hand [54]. Bai et al. proposed a deep learning gesture recognition method from multichannel sEMG [87].
In contrast, Cao et al. presented a sEMG detection and action recognition system based on overlapping sliding window analysis and synchronous command transmission for real-time manipulator control [88]. Besides, Benatti et al. designed a wearable prosthetic hand controller based on gesture recognition that is able to communicate with a smartphone via Bluetooth, providing a customized control strategy [89]. In [90], an EMG-driven upper limb prosthesis for amputees living in developing countries was presented. The EMG signal is acquired via an Electro Resistive Band (ERB) sensor, which is low-cost, cheap, washable, and withstands wear and tear. The EMG signal’s amplitude drives the motor to set hand opening or closing commands. Polisiero et al. [91] designed a low-cost, EMG-controlled prosthetic hand intended for amputees living in developing countries. Two-channel sEMG signal is acquired on the forearm muscles and, after a preprocessing phase, the EMG envelope is computed to detect muscle contraction. Afterward, it is compared with a proper threshold to issue prosthesis hand opening or closing commands. In this way, the motor torque is proportional to EMG amplitude.
Gailey et al. provided a proof of concept of an online EMG-based prosthetic hand control method that can predict different hand gestures and individual finger forces via a Support Vector Machine (SVM) classifier [92]. Bernardino et al. proposed a control method for a 3D-printed prosthetic hand that uses the sEMG signals acquired from abdominal muscles [93]. In [94], pattern recognition from three-channel sEMG was performed to control a five-fingered underactuated prosthetic hand. The control method is divided into two main steps: (i) feature extraction through parametric autoregressive (AR) model, wavelet transform, and integral of EMG signals, and (ii) classification via a variable learning rate (VLR)-based, three-layer feedforward neural network applied to EMG features. It can discriminate between flexion and extension of three fingers: thumb, index, and middle. Carrozza et al. designed a novel EMG-controlled prosthetic hand that allows hand opening and closing tasks [95]. In [96], a novel sEMG sensor using polypyrrole-coated nonwoven fabric sheet was used to capture a subject’s motion intention for controlling a myoelectric prosthetic hand. Brunelli et al. proposed a fully wireless, low-cost, wearable multichannel sEMG acquisition system that can communicate via Bluetooth with a mobile processing device, which in turn can send motor commands to a prosthetic hand [97]. In [98], individual and combined finger movement features were extracted from sEMG signals using time–frequency distribution. The performances of three different classifiers in discriminating these movements were compared to obtain a more dexterous prosthetic hand control. Khushaba et al. presented a pattern recognition method that can classify different individuals and combined finger movements from two-channel sEMG to control a prosthetic hand [99]. In particular, a Bayesian data fusion postprocessing algorithm was applied to enhance the classification performance.
Furthermore, Kamavuako et al. [100] investigated the feasibility of using the intramuscular EMG (imEMG) signal for the control of a myoelectric prosthetic hand by applying a Fitts’ Law test, while Dewald et al. [101] presented a case study of imEMG-based prosthetic control. Al-Timemy et al. proposed a multichannel sEMG-based pattern recognition method to classify different individual and combined finger movements for dexterous hand prosthesis control [102]. In detail, they performed feature extraction through time domain–autoregressive approach, feature reduction by applying orthogonal fuzzy neighbourhood discriminant analysis (OFNDA), and classification via a Linear Discriminant Analysis (LDA). Zhang et al. designed an anthropomorphic EMG-controlled prosthetic hand [103]. Dalley et al. proposed a two-channel sEMG-based method for controlling a multigrasp prosthetic hand [104]. Russo et al. designed an artificial EMG-controlled prosthetic hand. In detail, sEMG signals are acquired via MyoWare muscle sensor, and three different hand movements are recognized through a SVM classifier [105]. In [106], neck and face sEMG signals were used to control a voice rehabilitative device for people with total laryngectomy.
Moreover, the “Myo armband” by Thalmic Lab is worth mentioning, a wearable device for human–machine interaction (discontinued from the market). It detects muscle activity via eight dry sEMG sensors and hand/forearm movements through a nine-axes Inertial Measurement Unit (IMU). The acquired data are transmitted via Bluetooth and can be used to control different devices (PC, robot, prosthetic hand, etc.), as described by Visconti et al. in [107]. Furthermore, Lu and Zhou proposed a hand-free HMI driven by facial EMG signal for controlling electronic devices (e.g., mouse, touch screen, etc.) [108].
In [109], sEMG signals were used to control an electric wheelchair for patients suffering from SCI at lower cervical levels. The classification performances of a SVM and a k-Nearest Neighbour (kNN) were compared to distinguish five different commands: left, right, forward, stop, and rest. Kalani et al. presented an sEMG-based teleoperated masticatory rehabilitation robot [110]. Briefly, three kinematic parameters are predicted using eight-channel sEMG (acquired from four jaw muscles of a teleoperator) and used to control the robot’s masticatory trajectory. Alibhai et al. performed gesture recognition from sEMG signals acquired via a Myo armband to control an intelligent wheelchair [39]. In [37], a three-channel sEMG signal was used to wirelessly control the commercially available JACO robotic arm. This control method was proved to be simple, low-cost, comfortable, easy to use and, above all, suitable for people with severe motor disabilities who cannot make use of the joystick. Song et al. addressed the problem of muscular fatigue by proposing a robust pattern recognition method based on a fuzzy neural network to control a powered wheelchair [40]. Laksono et al. investigated the best model to classify three different upper limb movements, i.e., elbow extension, shoulder extension, and combined elbow and shoulder extension, from a three-channel sEMG signal for the control of a robotic arm by applying machine learning [38]. In detail, they compared 48 different classification models that were obtained with or without data division by time, with or without Teager–Kaiser energy operator (TKEO), and by using conventional features in the feature extraction phase and testing eight different classifiers in the classification step. In [41], facial sEMG signals were used for hands-free control of an intelligent wheelchair. In particular, the muscle movements during chewing, namely single jaw click and double jaw click, are converted into commands to the device, i.e., forward, backward, left, right, and stop. In [111], facial sEMG signal was acquired during forehead single and double click movements and used to issue commands to an intelligent wheelchair (i.e., forward, backward, turn left, turn right, and stop). Hamedi et al. proposed an EMG-based facial expression recognition method suitable for HMI applications [112].
In the field of exoskeleton control, Wege and Zimmermann presented an EMG-controlled hand exoskeleton for rehabilitation purpose. Ten-channel sEMG is acquired on the forearm to detect subject’s motion intention. After blind source separation, rectification, and low pass filtering, the signal is decomposed into different components to capture each finger’s flexion and extension. Then, corresponding force values are computed and compared with appropriate thresholds to generate a trajectory executed by the motor control [113]. Ho et al. designed an EMG-driven exoskeleton for hand closing and opening rehabilitation in chronic patients after stroke. EMG sensors are placed on the forearm to capture voluntary contraction of abductor pollicis brevis and extensor digitorum. Hand closing is triggered by 20% exceeding the maximum voluntary contraction value of the former, whereas hand opening is associated with 20% exceeding the maximum voluntary contraction value of the latter [114]. Besides, Loconsole et al. proposed an EMG-controlled rehabilitation hand exoskeleton for bilateral grasp training. The signal is acquired on three forearm muscles of the healthy upper limb and used to control the grasp function of the impaired one [115]. Hussain et al. presented an EMG interface for the control of a supernumerary robotic finger. In particular, they used an EMG armband for hand gesture recognition to control motion and an sEMG channel to control joints compliance [116].
Furthermore, Abdallah et al. designed a 3D-printed EMG-driven robotic exoskeleton to assist finger flexion and extension movements in patients following stroke [117]. Secciani et al. presented a low-cost, fully wearable EMG-driven hand exoskeleton to assist impaired people in daily life activities [118]. In detail, sEMG signals are acquired on the extensor digitorum superficialis and flexor digitorum superficialis to capture the subject’s motor intention, i.e., hand opening, closing, or resting. Song et al. proposed an EMG-controlled exoskeleton for wrist flexion and extension rehabilitation in patients after stroke [119]. In [120], a myoelectric control method for a lightweight, 3D-printed upper arm exoskeleton, intended to improve muscle strength in home-based rehabilitation sessions, was presented. In detail, the method exploits sEMG signal acquisition through a Myo armband and different training load classifications via a kNN. Cai et al. presented ReRobot, an EMG-controlled upper limb exoskeleton for rehabilitation of poststroke patients by exploiting mirror therapy [121]. Indeed, sEMG signals acquired on the healthy side are classified by a SVM to decode the subject’s motor intention, then executed by the exoskeleton attached to the impaired side. Yin et al. extracted gait cycle durations (GCDs) from an eight-channel sEMG signal using the autocorrelation and the Bayesian fusion algorithm for controlling the motion speed of an exoskeleton–treadmill gait rehabilitation system [122]. In [123], an upper limb EMG-controlled power-assist exoskeleton intended for elbow rehabilitation was developed. Lu et al. proposed a four-channel sEMG-based real-time control for a rehabilitation hand exoskeleton [124]. In [125], a custom lower limb robotic exoskeleton control method was presented by estimating the active joint torque from the sEMG signal.
La Scaleia et al. showed a novel approach for associating the EMG signal from shoulder muscles with leg kinematics, which can be adopted to control an avatar walking in virtual reality or a robotic lower-limb exoskeleton [126]. Finally, Lyu et al. developed an EMG-controlled knee exoskeleton for home rehabilitation of poststroke subjects. In particular, the EMG signal is acquired via Myo thigh band, which is obtained by combining two Myo armbands. In addition, for greater involvement of patients, they are asked to perform training sessions in a visuomotor game context [127].
Table 3 presents a summary of the considered EMG-based HMIs.

3.1.3. ENG-Based HMIs

ENG is the recording of electrical activity directly from a peripheral motor nerve. ENG is an invasive technique that involves electrodes implanted in the neural tissue. ENG provides precise and high-resolved information on a group of neurons or even on a single neuron activity. Differently, EEG can only record general activities of the brain, while sEMG provides aggregated information about neuromuscular control [44,60].
In the last decade, even ENG signal has captured attention in biomedical HMIs, principally for prosthetic hand control. However, in the literature, there are still few studies about ENG-based HMIs [24]. The central idea is to extract user’s motor intention by recording the electrical activity of peripheral nerves. Noce et al. proposed an ENG-based hand prosthesis control. In particular, they acquired neural signals from an amputee’s median and ulnar nerves and computed the ENG envelope. Afterward, by applying typical EMG pattern recognition techniques, they performed hand gesture recognition without needing a feature extraction phase. Furthermore, for comparison purposes, they applied the proposed method also to sEMG signal, thus confirming the validity of this novel approach [52]. Nguyen et al. presented a bioelectric neural interface for hand prosthesis control, which is integrated with implantable microelectrodes to record ENG signal from peripheral nerves in the residual limb, and deep-learning algorithm decode subject’s motor intention [55]. Finally, Noce et al. developed an ENG-based classification algorithm for discriminating two different grasp commands to control a prosthetic hand [59].
Table 4 shows a summary of the selected ENG-based HMIs.

3.1.4. EOG-Based HMIs

EOG is the measurement of a steady potential generated by the eyeball’s movements due to separated positive and negative electric charges. Therefore, a current flows from the cornea to the retina. This biopotential can be recorded by placing surface electrodes on both sides of the eye, either horizontally or vertically. Indeed, when the gaze is straight ahead, the EOG signal is zero because of the eyeball’s symmetric horizontal position for the two electrodes. On the contrary, when the gaze is shifted to one side, a nonzero EOG signal is detectable. The same principle is applied to measurement in a vertical configuration. Therefore, there is a relationship between the horizontal or vertical gaze angle and the EOG amplitude [44,45,60].
Various biomedical HMIs exploit eye tracking from EOG signals, enabling disabled people to interact with the outside world through eye motions. EOG-based assistive technologies can be used as nonverbal communication tools or for controlling a robot, a prosthesis, a wheelchair, a device in a smart environment, the cursor on a computer screen, and so on. Golparvar and Yapici designed a graphene textile-based wearable assistive device that allows eye tracking from EOG for remote control objects [56]. Zhang et al. developed an EOG-based HMI for smart home environment control (e.g., TV control, air conditioner control, wheelchair control) in the case of SCI patients. In particular, a visual stimulus, in the form of flashing buttons on a Graphical User Interface (GUI), is presented to the subject. Each button corresponds to a specific command. The user selects the input command by blinking his eyes synchronously to the flashes. The acquired EOG signal is submitted to feature extraction and classification operations to decode the subject’s intention [42].
Similarly, Huang et al. presented an EOG-based HMI for wheelchair control [57]. Martínez-Cerveró et al. developed an open-source hardware/software platform for eye movements recognition from EOG signal. In detail, this system is able to classify four different eye movement directions and can be used in HMI applications for assisted communication in paralysed people [128]. Moreover, Perez Reynoso et al. proposed an EOG-based HMI for real-time trajectory tracking of a manipulator robot by using neural network modelling [129]. Again, Choudhari et al. presented an EOG-based HMI for wheelchair control by converting voluntary single, double, and triple eye blinks into control actions, i.e., forward, right turn, left turn, and stop [130].
In contrast, Heo et al. developed a headband-type forehead EOG measurement system for HMI applications, consisting of wireless wearable sensors for detecting six class-eye movements [131]. Guo et al. developed a wearable HMI based on single-channel EOG recorded with a patchable sensor [132]. Finally, Wu et al. designed a single-channel EOG-based HMI integrated with encoding/decoding paradigms of eye blinking and looking up to interpret the user’s intention [133].
Table 5 shows a summary of the selected ENG-based HMIs.

3.1.5. Hybrid Biopotential-Based HMIs

Hybrid or multimodal biopotential-based HMI is the result of combining different types of bioelectric signals to increase their performance and reliability. Moreover, by exploiting the advantages of each category and associating many functions, these HMIs improve their effectiveness as assistive and rehabilitation technologies. Indeed, they allow people with disabilities to perform multiple tasks or implement better control. Furthermore, various combinations of biopotentials exist.
In the field of robotic control, Gordleeva et al. presented a real-time hybrid HMI for controlling a lower-limb exoskeleton. In particular, they combined a foot motor imagery-based BCI and multichannel EMG signals acquired from leg muscles to extract user’s motor intention, then converted them into input commands to the exoskeleton [51]. Ferreira et al. proposed a hybrid HMI based on EMG signal caused by eye blinks and EEG signal from the cerebral area responsible for processing the visual information to control a robotic wheelchair [134]. Zhang et al. developed a multimodal HMI combining EOG, EMG, and EEG signals to control a soft robot hand. In detail, this HMI exploits hand motor imagery from EEG, looking-left and looking-right eye movements recognition from EOG, and hand gesture recognition from EMG [135]. Huang et al. proposed a novel hybrid BCI that acquires EEG and EOG signals and translates them into commands for controlling an integrated wheelchair robotic arm system. By interacting with dedicated panels on a GUI, the user can control turning left/right of the wheelchair by performing hand motor imagery and issue other commands to the wheelchair and the robotic arm through eye blinking and eyebrow raising [136]. Finally, Ma et al. presented an EOG/EEG hybrid HMI for single and multiple robot control, which is based on four class-eye movements recognition from EOG and ERPs classification after the application of visual stimuli [12,137].
Regarding prosthetic control, Arrow et al. demonstrated the feasibility of using ERG signals for improved myoelectric hand prosthesis control. ERG is recording a transient potential that develops on the retina’s surface or the cornea in response to a light stimulus [44,61]. In particular, they used a threshold-based control strategy. Indeed, the number of naturally occurring neuronal action potential spike trains, generated by the retina after a visual stimulation, is counted and then compared with a predefined threshold to set the command to be sent to the prosthetic hand in order to improve the response time and the desired grip [58]. Rezazadeh et al. designed a coadaptive and affective HMI, which is able to assess the subject’s mental workload from EEG signal, for a better myoelectric forearm prosthesis control [138].
Furthermore, again Rezazadeh et al. proposed a multimodal HMI based on the acquisition of facial EEG, EOG, and EMG to perform face gesture recognition. In detail, features related to facial movements, eye movement direction, and mental states are extracted from fEMG, fEOG, and fEEG, respectively, thus enabling nonverbal communication [139]. Iáñez et al. presented a hybrid HMI that combines EEG and EOG signals to move a dot on a GUI [140]. Finally, Laport et al. compared two different HMI systems based on single-channel EEG and EOG to control smart home devices [141].
Table 6 shows a summary of the considered hybrid biopotential-based HMIs.

3.2. HMI Control Based on Muscle Mechanical Motion

Information about muscle mechanical motion is of great interest to many biomedical HMI applications, particularly for assistance and rehabilitation purposes, providing alternative strategies for controlling prostheses, robots, virtual reality, or even devices in smart environment. The term “muscle mechanical motion” refers to all mechanical events occurring during voluntary contraction, of which the EMG represents the electrical counterpart. Their monitoring allows the measurement of mechanical-induced muscle morphological changes. Indeed, it is well known that a contracting muscle changes its shape or dimensions (e.g., its cross-section), and small vibrations due to the progressive recruitment of motor units can be perceived (i.e., the mechanomyogram). Furthermore, muscle contraction also involves an increased blood afflux and displacement of muscle–tendon groups or even bones [4,143,144,145,146,147]. Therefore, these features reflect three main mechanical components: gross motion of specific muscle groups with associated muscle swelling; muscle vibrations; and movement of musculotendinous groups. To the aim of muscle contraction detection, multiple sensor technologies and different acquisition techniques have been proposed.

3.2.1. Muscle Gross Motion-Based HMIs

As mentioned before, muscle contraction is generally associated with muscle volume, cross-sectional area, and stiffness changes. Therefore, muscle contraction detection can be performed by measuring specific muscle or muscle group changes due to the so-called muscle gross motion. To date, several noninvasive sensor technologies have been developed that measure different physical quantities, which exhibit a direct or indirect relationship with contractile force, such as force/pressure sensors and triboelectric sensors. They are placed on the skin in order to record mechanical signals from the underlying muscles.
First, force sensors are sensitive to the force/pressure exerted by muscle contraction. A wide range of force sensors is available, such as resistive, piezoresistive, pressure, piezoelectric, and even optical fibre sensors. Many of these enable the recording of force myography (FMG) signals. Force myography is a noninvasive technique that uses force sensing elements to detect “changes in stiffness of corresponding musculotendinous complex against a default state” [4], providing information about muscle contraction in the low-frequency range (i.e., <10 Hz) [148]. Force-sensitive resistors (FSRs) change their electrical resistance as a function of the applied force and are widely used in FMG recordings. Several studies show the potential of FMG as an alternative control strategy for HMI applications, particularly for upper-limb prosthetic control [9]. In this context, Sakr et al. demonstrated the feasibility of estimating hand isometric force/torque from FMG signals recorded via 60 FSRs, which were embedded into four bands and placed in different locations around the arm [149]. Again, Sakr et al. showed the possibility of predicting force in dynamic conditions by using FSRs worn around the arm [150]. Ahmadizadeh et al. explored the application of feature selection to three high-density FMG datasets in order to reduce its dimensionality and, at the same time, achieve the same performance but with lower cost and complexity [151]. Xiao et al. proposed a novel FMG system, consisting of a strap embedded with eight FSRs, to detect different forearm positions for controlling a custom-made forearm pronation/supination exoskeleton [152]. Ferigo et al. presented a case study of an FMG-controlled bionic hand prosthesis for a transradial amputee [153].
Furthermore, Esposito et al. developed a piezoresistive array armband for hand gesture recognition equipped with only three FSR-based sensors. It is able to detect eight different hand gestures and can be used for prosthetic control [154]. A study by Prakash et al. proposed an FMG-controlled prosthetic hand for upper limb amputees. An FMG sensor consisting of two FSRs was applied on the residual forearm of an amputee to detect muscle contractile force, then converted into input commands to the prosthesis [144]. Esposito et al. also presented an alternative approach to EMG, with comparable performances, for improved hand prosthesis control based on muscle contraction detection via an FSR-based sensor [143]. The sensor was placed on a forearm muscle, proving to be as effective as the EMG envelope to control a hand prosthesis prototype [155,156]. Ha et al. explored the prediction of hand gestures by applying piezoelectric sensors around the forearm to map muscle contraction [157,158]. A piezoelectric sensor converts its mechanical deformation due to the applied force into an electrical signal. Ahmadizadeh et al. showed the feasibility of using selected locations of FSRs for FMG-controlled prosthesis with performances comparable to high-density FMG [151].
Furthermore, Fujiwara et al. proposed a low-cost optical FMG sensor for hand gesture recognition based on modulation of light intensity due to the applied force by the microblading effect, which can be applied for hand prosthesis control [159]. Moreover, Bifulco et al. tested the feasibility of a conductive rubber sensor, that changes its electrical resistance when stretched, to detect muscle contraction on the forearm, and also to control a hand prosthesis prototype [160]. Radmand et al. explored high-density FMG for prosthetic control. An array with a high number of FSRs, mounted into a prosthetic socket, was used to measure changes in surface pressure on the forearm [161]. Again, Cho et al. investigated the feasibility of FMG-controlled upper extremity prostheses [162].
Muscle contraction detection can also be performed via triboelectric sensors, which convert their mechanical deformation into electrical output, thus acquiring contractile force. The working principle of such a sensor technology is based on electrification and electrostatic induction phenomena occurring when materials with different electronegativities come into contact. Dong et al. developed a wearable triboelectric HMI, in the form of a smart glove, using a nanophotonic readout for hand robotic control and virtual/augmented reality applications [163]. Similarly, Zhu et al. presented a smart glove equipped with elastomer-based triboelectric nanogenerators and piezoelectric mechanical stimulator for robotic control and virtual/augmented reality applications [16]. Finally, An et al. proposed a tattoo-like triboelectric self-powered wearable sensor for controlling robots or devices in a smart environment [164].
Moreover, some researchers proposed a novel, invasive approach for hand prosthesis control by tracking the position of permanent magnets directly implanted into the upper residual muscles of amputees. By externally measuring changes in the magnetic field due to magnets displacement, it is possible to detect the force exerted by muscle contraction and thus to capture valuable information to issue commands to the prosthesis. This innovative HMI was called myokinetic controller [165,166].
Table 7 shows a summary of the considered muscle gross motion-based HMIs.

3.2.2. Muscle Vibrations-Based HMIs

During muscle contraction, muscle mechanical vibrations occur due to three main processes: (1) internal muscle vibrations, (2) oscillations of the human motor system (e.g., tremor and clonus), and (3) artifacts [172]. Mechanomyography (MMG) is a noninvasive technique that allows capturing high-frequency information (i.e., from 1–2 Hz to 100 Hz [3]) related to muscle vibrations. Islam et al. reported that MMG frequency content is closely related to the resonant frequency of muscle, which is affected by muscle stiffness [173]. However, the origin of the MMG signal is not yet fully understood. In Beck et al. [3], it is suggested that the MMG signal reflects three main physiological phenomena: (1) gross muscle movement during contraction, (2) muscle lateral oscillations at its resonant frequency, and (3) muscle fibres dimensional changes. Indeed, the recruitment of motor units during contraction results in dimensional changes of muscle fibres (e.g., their shortening and increase in diameter), which produce oscillations, i.e., pressure waves, that propagate from muscle up to the skin where they are detectable [174,175]. Different technological solutions have been proposed for MMG recordings, such as microphones, accelerometers, piezoelectric, or laser distance sensors [3,173,174,175]. Asheghabadi et al. presented a single-site MMG sensor consisting of a piezo plate and a strain gauge, which can capture both electrical and acoustic features from vibrations of a single muscle to perform multichannel finger pattern recognition [146]. Castillo et al. designed a wearable armband equipped with four MMG microphones and placed around the forearm to study the relationship between distributed normal force and MMG informational content. In detail, they showed that, as average force increases, tissue viscoelasticity changes, resulting in increased mechanical conductivity. In this way, the sensor is able to detect vibrations of deeper muscles with improved discriminative power, which is very useful in HMI applications for prosthetic control [176]. Moreover, Wicaksono et al. proposed a wireless synchronous carbon nanotube piezoresistive patch sensor to record MMG signal from leg muscles, which can be used for prosthetic or robotic control [177]. Finally, Xie et al. performed MMG recordings via two accelerometers for hand and wrist gesture recognition to perform multifunctional prosthetic control [178].
Table 8 shows a summary of the considered muscle vibrations-based HMIs.

3.2.3. Muscle–Tendons Movement-Based HMIs

It is worth mentioning other acquisition techniques that detect movement of musculotendinous groups resulting in morphological changes. In this context, Wu et al. presented an HMI using Electrical Impedance Tomography (EIT) for hand prosthesis control. The system performs electrical bioimpedance imaging, thus capturing changes in electrical conductivity under the movement of muscles and bones. In particular, an array of bioimpedance electrodes in a wristband was placed around the forearm to recognize nine hand gestures [145]. Furthermore, there are several HMI applications based on ultrasounds (US). They are safe, provide high temporal/spatial resolution, and can be acquired noninvasively. However, the major drawback is represented by the cumbersome US probe. Huang et al. compared simultaneous EMG recording and US imaging acquired on the forearm in terms of finger gesture recognition accuracies. They concluded that the US allows more dexterous and accurate control, thus showing its feasibility for prosthetic or robotic control [179]. Li et al. proposed a multichannel HMI based on an armband with US transducers for finger gesture recognition. It is designed to be applied in rehabilitation robotics [180]. Furthermore, Ortenzi et al. presented a comparative study of features and classification methods in US-based HMI for hand prosthetic control [181]. Sikdar et al. developed a novel method for predicting dexterous individual finger movements by imaging muscle activity using a wearable ultrasonic system [182].
Moreover, Sierra González and Castellini tested the feasibility of US-based HMI in a realistic scenario. They showed a linear relationship between the spatial first-order US features of the forearm and hand kinematic. In this way, it is possible to predict forces at the fingertips, thus revealing very useful for controlling a prosthesis [183,184]. Finally, two studies presented HMIs based on Sonomyography (SMG), a novel technique based on US acquisition that is able to detect muscle architectural changes and can be used to control a hand prosthesis. In particular, they investigated changes in muscle thickness during wrist flexion/extension. Maximum values of SMG are associated with wrist extension and, therefore, fingers opening, whereas minimum values of SMG imply wrist flexion and, thus, fingers closing [147,185].
Table 9 shows a summary of the considered HMIs based on muscle–tendon movements.

3.2.4. Hybrid Muscle Mechanical Motion-Based HMIs

Finally, HMIs combining different techniques of muscle mechanical motion detection also exist. Esposito et al. presented a piezoresistive FSR sensor that is able to measure changes in muscle cross-sectional area and MMG signal, proving very useful for prosthetic control [143]. Booth and Goldsmith developed a wrist-worn piezoelectric sensor for finger gesture recognition. The sensor is able to capture both MMG signal and changes in muscle shape, providing information that can be used for prosthetic, robotic, virtual reality, or smart environment control [186].
Table 10 shows a summary of the selected hybrid muscle mechanical motion-based HMIs.

3.3. Body Motion-Based HMIs

3.3.1. Image-Based Body Motion HMIs

A further class of HMIs is aimed at tracking the motion of different body parts, such as the eyes [43,187,188], the upper and lower limbs [189], and the head [190,191,192], via different kinds of vision devices. Eye-tracking technologies are usually based on infrared (IR) illuminators pointed at the eyes. IR sensors or cameras capture the reflected IR light to determine the gazing point on a screen and control a cursor to interact with various applications. The other vision-based HMIs mainly rely on standard cameras and sometimes also on depth cameras and involve a processing stage to recognize the gestures performed by the tracked body parts, e.g., translations and rotations of the head and various kinds of hand gestures. Moreover, a branch of vision-based HMIs comprises devices for exergames, which require active body movements to control the gaming experience and have potential applications for disease prevention, health promotion, and rehabilitation [18,20,29,30,31,32].
Maule et al. proposed an eye-tracking-based HMI called RoboEye [187]. RoboEye is composed of a standard power wheelchair integrated with an innovative, cost-effective, and user-friendly control system based on an eye-tracking system, a 3D camera, and a computer monitor. The system provides the users with comfortable navigation and two driving options, namely “direct” and “semiautonomous”, which allow them to move easily and autonomously within their homes. The natural modality uses eye-tracking to detect gazing at different areas of the monitor and provide continuous control of frontal and angular wheelchair velocities. This modality also enables efficient control of the wheelchair for users who cannot use standard interfaces (e.g., joystick, keyboard). The semiautonomous modality allows navigation toward a selected point in the environment by just pointing and activating the wished destination. At the same time, the system autonomously plans and follows the trajectory that brings the wheelchair to that point with the support of the 3D camera.
Bissoli et al. presented an intelligent home system that provides eye-tracking-based control of four devices, namely a television, a radio, a lamp, and a fan, as well as remote monitoring via the Internet of Things (IoT) protocols [43]. The control system and remote monitoring are intended for users with severe disabilities and related caregivers. The eye-tracking interface replaces the mouse control of a personal computer (PC): the changes of gazing points are translated into mouse movements, while the persistence in the same point for few seconds is translated into a mouse click. The system has been first tested on 29 healthy participants and then on a woman with severe disabilities in her own home for seven days. The efficacy of the smart home system was evaluated via a System Usability Scale (SUS) questionnaire, which was administered to both the healthy subjects and the person with severe disabilities, reporting very high scores.
Lin et al. proposed an HMI to control a computer mouse [188]. An IR camera captures images of the user’s eye, illuminated by an auxiliary infrared LED source. Features of the pupil are extracted to determine the gazing point on the screen, thus allowing the user to control a PC mouse. An autocorrection method is implemented to correct the gazing point estimate, leading to improved eye-tracking-based mouse control in the 90% of involved subjects, who achieved an overall accuracy of 97%.
An HMI for smart home control based on a Virtual Interactive Blackboard (VIB) has been designed by Conci et al. to remove any physical connection between the user and the domotic system [189]. VIB is based on the low-cost hardware architecture: the visual interface is projected through a beamer on a flat surface. A fixed webcam achieves the user’s gesture, which captures the scene at 10 fps. To achieve real-time processing, unnecessary information is first removed. The illumination is assumed to be slowly varying, so a background suppression procedure is performed by discarding illumination variations and focusing on areas corresponding to human skin colour to simplify hand tracking. An AdaBoost classifier is trained to provide real-time hand gesture recognition. Three hand gestures, namely opening, closing, and finger-pointing, are used to control the position of a cursor, trigger clicking actions, and trace curves for writing and drawing. A central processing unit connects the VIB interface to actuators taking charge of user commands, which can access several services (e.g., TV, phone) and control the related parameters (tuning, switching, opening–closing, moving).
A HMI based on head and mouth movement analysis has been proposed by Baklouti et al. as a control system for an exoskeleton, designed to compensate for the loss of upper limbs mobility in patients suffering from myopathy [190]. The system is based on a monocular camera to capture the user’s head and provides two control modalities, namely head control and mouth control. The head control uses an attention region approach to minimize the computational burden of face detection, then performs, in real time, a face detection and global feature extraction via Adaboost, and finally a pose estimation that provides the translation and rotation matrix of the head, which is used to generate an exoskeleton command. The mouth control performs a mouth extraction by combining a colorimetric method (thresholding on Q component of the YIQ colour space) with edge detection and active contour to locate extremum points for mean square fitting of lips profile with second-order polynomials. It then classifies the mouth expression from the modelled lips profiles to generate a command for the exoskeleton.
Chang et al. proposed an HMI for people who cannot use their hands to support using a computer or speaking [191]. A camera captures images of user’s head movements. An initial skin colour adjustment is performed to optimize the tracking of the user’s face within the frames. A binary mask is obtained with a group of pixels that roughly covers the user’s face in each frame. The barycentre of this group of pixels is computed to determine head movement direction among the eight directions considered by the system (two along the vertical axis, two along the horizontal axis, and two along each of the two 45° diagonals). The head tracking system is used to compose two-digit or three-digit codes corresponding to letters or sentences in predefined tables, which are then reproduced via a sound-generating system to allow disabled users to communicate.
A HMI based on head movement recognition has been presented by Gautam et al. to control a small robotic car [192]. A camera captures the user’s head images, which are processed to extract binary maps of face pixels. The image field is divided into three areas: left, centre, and right, and the head movement is classified based on the area of the binary map, including the most significant number of face pixels. The result of the classification is translated into a command for the robotic car, namely “turn left”, “go ahead”, and “turn right”. When user’s head falls out of the image field, i.e., no face pixels are detected in all areas, a “stop” command is sent to the robotic car.
Furthermore, in the context of the vision-based HMIs, it is worth highlighting the impact of the “exergames” on physical and cognitive functions, as demonstrated by different studies published over the last decade. Rosly et al. [31] presented a review on exergaming applications for subjects with neurological disabilities. They underlined that exergaming can provide outcomes with equivalent dose-potency as traditional physical exercise in clinic or home environments. Another survey by Reis et al. [32], showed that exergames could be used as a complement to traditional forms of motor rehabilitation in older subjects, improving their balance, gait, muscle strength, upper limb function, and dexterity. A combined intervention in which traditional physiotherapy is integrated with exergames appears to be more efficient than each type separately. Exergames were also used as a rehabilitation treatment for people affected by Parkinson’s disease [20] and MS [18]. Furthermore, the commercial device “Microsoft Kinect” (an RGB–D camera) was widely tested for motor rehabilitation purposes, as reported in the review by De Gama et al. [29]: many studies presented Kinect-based systems in which the subject had to simulate in a virtual environment a functional activity and achieve some objects to complete a task. A Cochrane review by Laver et al. [30] found evidence that virtual reality and interactive video gaming for stroke rehabilitation may be beneficial in improving global motor, upper limb function, and activities of daily living (ADL) when used as an addition to usual care or when compared with the same dose of conventional therapy. In summary, these reviews indicate that exergames appear to be a viable and effective rehabilitation tool for people with neuromuscular diseases.
Below are some articles that feature different technologies for exergaming purposes. A mapping study by Gmez-Portes et al. [193] showed how exergames could address home rehabilitation for children and teenagers. In particular, the study presents a home rehabilitation software prototype based on “Microsoft Azure Kinect DK”. A virtual avatar, showed in a connected screen, mimics the patient’s movements, and it is necessary to perform certain tasks to achieve the predetermined goals. Palaniappan et al. [194] tested the “HTC Vive” for virtual reality exergames on patients affected by spinal cord injury. The authors used the “Vive tracker” to track the movements of the patient’s arm by attaching it to a Velcro hand strap through a 3D-printed interface plate (the standard controller requires fine motor control to grip and use buttons). A further study by Nguyen et al. [195] showed an exergames room based on the “Jintronix Rehabilitation System (JRB)” (a virtual reality software compatible with the Microsoft Kinect) and the “Meditouch HandTutor (MHT)” (a set of electro goniometers measuring wrist and elbow movements) to interface the patient in stroke rehabilitation with the virtual reality and guide him in performing specific movements.
Table 11 outlines the image-based body motion HMIs studies included in the current survey.

3.3.2. Nonimage-Based Body Motion HMIs

Body motions, such as hand or finger motions, are generally tracked by vision-based systems. An alternative approach is based on wearable devices, such as inertial sensors (the most widely employed for human motion recognition), touch sensors, strain gauges, flex sensors, and ultrasonic sensors. Some studies based on these technologies with assistive and rehabilitative purposes are reported below.
A study by Chuang et al. [196] focused on a smart glove equipped with flex sensors (variable resistors with the degree of deflection) for finger gesture recognition purpose. The study aimed to detect different thumb, index finger, and middle finger movements using neural network algorithms. A further paper by Dong et al. [197] proposed a smart glove embedded with piezoresistive strain sensors based on stretchable polydimethylsiloxane–carbon black to recognize finger gestures and control robot fingers. Zhu et al. [198] presented a different smart glove based on stretchable conductive metal-coated yarns (a nylon yarn covered by a metal film) to remotely control a robotic hand and manipulate the colour switching of light by using gesture recognition. Hang et al. [199] detected various human body motions using a poly(acrylamide) hydrogel-based strain sensor. Thanks to the high extensibility of these sensors, the authors measured the bending of fingers, wrist, elbow, and knee and the cheek bulging. Gesture recognition and a robotic hand’s gesture control were experimented with by embedding five hydrogel sensors in a glove.
In the context of exoskeleton control, Ueki et al. developed a hand exoskeleton for self-controlled rehabilitation therapy. It uses a closed-loop master and slave system to assist flexion/extension, abduction/adduction of hand joints, and thumb opposability. In detail, the subject performs movements in virtual reality with his healthy hand wearing a glove (i.e., master level), which is equipped with 3D motion sensor. The movement is reproduced by the exoskeleton attached to the impaired hand (i.e., slave level), which uses three-axis force sensors [199]. Similarly, a further study by Rahman and Al-Jumaily [200] proposed a hand rehabilitation exoskeleton consisting of a master and slave system. The healthy hand wears the glove acting as master and is equipped with flex sensors, whereas the impaired one wears the exoskeleton acting as a slave. The glove wirelessly transmits data to the exoskeleton that executes commands for fingers flexion and extension. Again, Cortese et al. proposed a hand rehabilitation exoskeleton in the form of a master and slave system. The therapist wears a glove (i.e., master unit), while the patient wears the exoskeleton (i.e., slave unit). These two devices exchange data between them. In particular, the therapist guides the exercises to be performed by the patient. The glove presents six three-axis MEMS accelerometers, which obtain position vectors classified and compared with proper thresholds to understand the specific command (i.e., rest, grasp, pinch) and, eventually, its percentage [201].
Moreover, some approaches use only wearable inertial sensors. An example is the research by Han and Yoon [202], who experimented with a three-axis gyroscope fixed on the hand back intending to discriminate six hand gestures (up, down, left, right, clockwise rotation, anticlockwise rotation) and track the hand trajectories. Then the HMI was tested as a PC input device for controlling various applications (presentation, video player, web browser).
Table 12 presents a summary of the considered non-image-based body motion HMIs.

3.4. Hybrid HMIs

Hybrid HMIs are characterized by two or more technologies that work together to achieve a common goal [11]. Various combinations of biopotentials with vision systems, force sensors, accelerometers, US probes, microphones, IMU sensors, etc. have been presented in the last two decades. Compared with single modality HMIs, multimodality HMIs could enrich the controllability of the interface, optimize the communication methods, and improve the overall performance of the interface [204]. Many studies combined biopotentials with morphological sensors (force sensors, US probes, etc.) to obtain electrophysiological and morphological information of the same muscle simultaneously [13].

3.4.1. Biopotentials and Image-Based Systems

The current subsection shows biomedical HMIs with hybrid controls based on biopotentials detection and vision systems. A study by Wei et al. [204] presented a hybrid HMI designed for hands-free control of an electric-powered wheelchair. Forehead EMG signals and colour face image information were used to identify movements of the human face. In detail, five winking and jaw clenching movement patterns were selected and classified, mapping into six control commands to drive an electric-powered wheelchair in an indoor environment.
Haung et al. [205] presented a multimodal emotion recognition system by combining facial images and EEG. The acquired images were classified by a convolutional neural network (CNN) architecture for facial expression detection. Different SVM classifiers classified the EEG signals. Finally, the facial expressions and EEG signals were combined to recognize the facial emotions.
Downey et al. [206] experimented with a combined control system for a robotic arm/hand on patients with spinal cord injury. Microelectrodes arrays were implanted on the motor cortex of the patients for realizing a BMI. Moreover, a computer vision system composed of an RGB–Depth camera was mounted above the arm base to identify objects by matching depth image templates from a model library. During the experiment, the subjects controlled the robotic arm/hand to perform grasping tasks.
A conference paper by Bu et al. [207] proposed a hybrid control method for prosthetic hand, combining EMG signals with a vision-based object classifier. Information of target objects (shape, dimension) are obtained from images and then utilized to generate control commands for motors in combination with the EMG signals.
A study by Malechka et al. [208] presented the “sBCI eye tracking system” composed of a headset that integrates multichannel EEG equipment, an eye tracker system (two cameras for tracking of left and right eye) based on video-oculography (VOG), an environmental observation camera, and an integrated visual stimulator for a SSVEPs. The system detects the user’s intention to interact with a specific device in its environment. McMullen et al. [209] presented a hybrid HMI named “HARMONIE system”. This system utilizes ECoG and depth electrodes within the motor cortex for realizing an iEEG-BMI, with the aim to identify the movement intention. A Microsoft Kinect sensor is used to record both the depth information and RGB of the experimental workspace. The patient’s gaze is captured by a monitor-mounted eye-tracking system. A patient implanted with iEEG electrodes can use the eye tracking system to visually select an object displayed in real-time on the monitor and provide control signals to an upper prosthetic limb.
Frisoli et al. proposed architecture for controlling a rehabilitation arm exoskeleton in reaching and grasping functions. In summary, the system is composed of a motor imagery-based BCI to decode the subject’s motion intention, an eye-tracking to detect subject’s gaze direction while selecting the target object in the task space, and a Kinect sensor to track the object in the 3D space and communicate its position to the robotic device [210].
Table 13 outlines the hybrid controls for HMIs based on biopotentials and image-based systems included in the current survey.

3.4.2. Biopotentials and Mechanical Motion Detection

This subsection shows some biomedical HMIs with hybrid controls based on biopotentials and mechanical motion detection.
The human anatomy inspires a prosthetic hand presented by Dunai et al. [211]. It works by using sEMG sensors to allow the user to activate the prosthesis and FSR sensors to simulate the touch pressure of the fingers. The grasping is completed when the FSR in the thumb signal is above a given threshold.
Krasoulis et al. [211] proposed a multimodal control for prosthetic hand, which combines sEMG sensors with IMU systems. In detail, each EMG electrode incorporated a nine-degree-of-freedom (DOF) IMU, i.e., a triaxial accelerometer; gyroscope; and magnetometer measuring acceleration, angular velocity, and magnetic field, respectively. A total of 12 hybrid sensors were used for monitoring the activation of the arm and forearm muscles and provided control signals to the prosthetic hand. A further example of a hybrid sEMG-IMU interface is the study presented by Shahzad et al. [212] to evaluate the variation of the sEMG signals with the arm position. The system consists of multiple sEMG sensors and nine-DOF IMUs to measure the electrical activity of the muscles and the position of the forearm relative to the shoulder while performing six classes of hand motions. A conference paper by Kyranou et al. [213] also experimented with a hybrid control for robotic prosthetic hand, based on 12 sEMG sensors, each integrating a nine-DOF IMU. A multiclass LDA classifier was used to discriminate six hand grip patterns and use the predictions to control a robotic prosthetic hand in real-time.
Moreover, Jaquier et al. [214] experimented with a hybrid device composed of sEMG and pressure sensors to predict the activation of the wrist, hand, and single-finger for future prosthetic implementations. A proximal forearm device is a tactile bracelet with 10 pressure sensors (made by resistive elastomers); a distal forearm device is composed of 10 sEMG electrodes (Ottobock Myobock).
A different combined control for upper-limb prosthesis, presented by Guo et al. [215], was based on sEMG and near-infrared spectroscopy (NIRS) sensors (light sources at three wavelength: 730, 805, and 850 nm). The study investigated the classification accuracy in discriminating 13 wrist and hand motions. It tested the performance in controlling a virtual prosthetic hand by using four hybrid sEMG/NIRS to the forearm muscles of amputees.
A further type of hybrid control is the combined sensor sEMG/US presented by Xia et al. [13]. The paper proposed a portable hybrid sensor system: the sEMG electrodes and an US transducer (for A-Mode signals) are fixed into a module to detect electrophysiological and morphological information of the same muscle. An armband equipped with four hybrid modules is mounted on the subject’s forearm to recognize 20 different gestures (6 wrist motions, 13 finger motions, and a rest state motion).
Furthermore, Dwivedi et al. proposed a soft assistive exoskeleton that combines a glove and a muscle–machine interface in the form of a sensorised sleeve. The interface is equipped with EMG and FMG sensors to decode the subject’s motion intention. The acquired data are classified into different grasp types and used to trigger the motors [216].
Table 14 summarizes the selected hybrid controls for HMIs based on biopotentials and mechanical motion detection.

3.4.3. Other Various Hybrid Controls

The following subsection shows some examples of biomedical HMIs not classifiable in the previous subsections whose controls are based on hybrid combinations of sensors (biopotentials, IMU, US, vision system, voice recognition, microphones, Radio-Frequency Identification (RFID), etc.) and whose application fields are principally involved in assistance or rehabilitation.
Ubeda et al. [11] presented a shared control combining a BMI with RFID technology to control a robot arm. RFID tags were placed in the experimental setup to give information about the position of the objects in the scene. An antenna, placed on the robot arm’s end effector, is used to read/write the tags. Information about objects’ positions is stored in the tags. The shared control system uses this information and supports the user to make high-level BMI decisions via mental tasks (move left, move right, pick, or place). The authors showed that the total amount of BMI commands can be significantly reduced thanks to the RFID tags.
A further study by Perez et al. [218] proposed a hybrid control for a robotic wheelchair to be used by people with severe motor disability. The system is based on a Visual-Based Interface (VBI) and an IMU sensor and combines the information of the user’s head orientation (obtained by both the VBI and the IMU sensor) for providing control signals to the wheelchair. In [219], another hybrid HMI for controlling a robotic wheelchair was presented, which combines six different control strategies, i.e., eye blinks from sEMG, eyes and face movements acquired using video cameras, head movements detected via an IMU sensor, sip-and-puff captured by a pressure sensor, and brain waves from EEG.
Voice recognition systems have also been experimented with for controlling wheelchairs. An example is the study by Anwer et al. [220] that presented an eye- and voice-controlled wheelchair. The device carries out both voice commands (on, right, left, stop), which are processed through an Arduino platform, and commands given by the movement of the eyes (to the right or left), captured with a front camera and processed with a Raspberry PI platform.
Another low-cost multimodal sensor was presented by Gardner et al. [221]. It consists of an acoustic MMG sensor, a nine-DOF IMU sensor, and a video camera. On an amputee participant, a compression sleeve containing both the MMG and the IMU sensors were positioned on the biceps. The participant wore the video camera on a pair of glasses. The study predicts different grasp strategies and the successive control of a commercial prosthetic hand (BeBionic Hand).
Moreover, in order to allow people with severe disabilities (ALS, CP, stroke, etc.) to communicate, a study presented a wireless home assistive system [222], consisting of various types of sensors: EOG (for detecting eyes movements) and switches (based on: push button, InfraRed, mercury, long–short tone, and pacifier) were used for generating a Morse Code. Afterward, a fuzzy algorithm-based-Morse Code recognition allowed it to provide input commands to a connected PC.
Table 15 presents a summary of the considered hybrid controls for HMIs based on various combinations of sensors.

4. Discussion

This survey analysed the literature on biosignal-based HMIs that use biopotentials, mechanical muscle movements, body motion, or their combinations as control signals, with applications in the field of assistance and/or rehabilitation. Journal and conference papers presented in the past 20 years were considered.

4.1. Statistical Analysis

Statistical data extracted from the selected papers (144 journals papers and 37 conference papers) are reported below. Figure 2 shows the number of publications over the last years: the trend is growing. This proves the ever-growing interest to develop HMI technologies for assistive and/or rehabilitative purposes. Figure 2B shows how the studies are related to the considered applications. The highest percentage (about 67%) pertains to the assistive field, while rehabilitation and assistance rehabilitation represent about the 20% and 13% of all studies, respectively.
Figure 3A shows the percentages of use of the biosignals for HMI control. Biopotentials constitute the highest percentage (about 57%), while muscle mechanical motion represents 21%. Lower percentages are associated with body motion controls (about 10%) and hybrid ones (about 12%). The pie charts in Figure 3B–E represent the distributions of the subcategories of biosignals. The most used biopotentials are still the EMG and the EEG (>75%). Among the muscle mechanical motion signals, the muscle gross motions are the most used ones and, together with the related tendon motions, are involved in about 85% of the studies. The studies on hybrid HMIs are well-balanced; however, it is worth highlighting that biopotentials are involved in 75% of these studies.
Figure 4A shows the distribution of studies focusing on the different targets, while Figure 4B shows their timeline. About 70% of these studies addressed robotic and prosthetic control applications. The number of studies published before 2010 only represents about 6% of the total. Starting from 2010, a moderate increase can be observed in studies focusing on robotic control, prosthetic control, and gesture recognition, while studies on the other targets experienced only a minor increase.

4.2. Advantages and Disadvantages of Biosignal Categories

A summary of the principal technologies/sensors/systems associated with each different macrocategory of biosignal is displayed in Figure 5. For the sake of simplicity, hybrid systems have been omitted.
It is interesting to briefly report the pros and cons of each macrocategory of biosignal-based HMIs.
Biopotential-based HMIs allow capturing the subject’s intention from different body regions (i.e., brain, muscles, peripheral nerves, or eyes). They are widely used for controlling all kinds of assistive and rehabilitation devices. BCI systems are promising technologies, particularly for people with severe motor disabilities, such as quadriplegia or paraplegia. Most biopotentials are acquired noninvasively (e.g., scalp EEG, sEMG, EOG). In this case, stable electrode placement and skin preparation are required. Wet electrodes need the application of a conductive gel to ensure better contact; on the contrary, dry electrodes are very useful for long-term recordings [47]. Amplification and real-time processing are needed because biosignals have small amplitude and are affected by noise, such as electromagnetic interference, motion artifacts, and crosstalk with other biosignals, and modifications over time (e.g., EMG is susceptible to muscle fatigue). Increasing use of machine learning techniques has been documented to decode the subject’s intention. Some acquisition techniques are invasive (e.g., ECoG electrodes are surgically implanted and may involve a risk of infection) and therefore more complex and expensive [65]. However, they allow acquiring signals from otherwise inaccessible regions with much higher specificity (e.g., ENG allows more intuitive prosthetic control). The use of this type of HMI is not intuitive. Therefore subjects must first be trained [52].
Muscle mechanical motion-based HMIs exploit the morphological changes of muscles during contraction. The FMG technique (force, pressure, piezoelectric, triboelectric sensors, etc.) detects muscle gross variation, allows performance comparable to EMG without using electrodes (no electrical risk), requires much simpler processing, has a lower cost and is less affected by electromagnetic and environmental interferences. On the other hand, FMG is susceptible to muscle fatigue and crosstalk of adjacent muscles and requires a suitable capacity for muscle contraction (technique not usable in subjects with severe motor disabilities) [4]. The MMG technique (microphone, accelerometer, piezoelectric sensor, etc.), which allows detection of muscle vibrations during contraction, has the advantage of being more sensitive than FMG in interpreting movement intentions, even in subjects with poor muscle contraction ability but is more prone to noise and motion artifacts [3]. As for the detection of muscle–tendon movements, the US technique guarantees good resolution for both superficial and deep muscles, but the US probe is bulky and expensive; it is not wearable and portable, so it is suitable only for clinical or research environments [183].
Finally, concerning the body motion-based HMIs, the image-based ones have the advantage of being contactless and easy to use in a home environment. They are able to detect the movement of body parts (e.g., head, hands, eyes) but are susceptible to illumination conditions, camera field of view, and overlapping objects. They require complex image processing algorithms to extract useful information for controlling devices, and real-time implementation is difficult. This type of interface is essentially used for smart environment and virtual reality control [30]. On the other hand, nonimage-based HMIs exploits wearable, low-cost sensors which can be embedded into electronic devices, e.g., smartphone, or in gloves that can turn out to be uncomfortable for the user [197]. They are mainly used for robotic control.

4.3. Latest Trends

The pie charts in Figure 6 outline the percentages of use of the considered biosignals in each specific target’s state-of-the-art by considering the studies published only in the last five years (i.e., 2015–2020).
As opposed to the beginning of the 20 years considered in this survey, biopotentials are no longer the leading control signals of biosignal-based HMIs. Indeed, the use of muscle mechanical motion signals has reached or overcome biopotentials in all targets (more than 20% of studies), except for the robotic control, where biopotentials are still used in about the 65% of studies. In prosthetic control, muscle mechanical motion signals have been used in about 68% of studies, so their use has primarily overcome biopotentials (about 10% of the studies). In particular, the measurement of muscle gross motion (about 61% of all muscle mechanical motion signals) has experienced impressive growth as an alternative to EMG. Undoubtedly, FMG overcomes many well-known EMG limitations (e.g., artifacts due to unstable electrical contact, drying of pregelled electrodes over long periods, susceptibility to electromagnetic interferences), thus standing as a good, robust control signal for long-term HMI applications. On the other hand, while reliable standards are available for EMG measurement (e.g., SENIAM project), the properties of FMG sensors as well as their number and positioning for accurate measurements of muscle mechanical motion have not been standardized yet. Future research on FMG should address this issue to provide clear guidelines that ensure good quality measurements and increase the transferability of results obtained in different studies. Obviously, HMIs based on muscle mechanical motion are unfeasible for subjects with limited to no muscle activity, while those based on brain-related biopotentials, such as the BCIs, enable these impaired subjects to interact with machines that provide them with assistance or support their rehabilitation.
Body motion is prevalent in HMIs dedicated to controlling virtual realities and smart environments, as it has been used in about 30% of studies focusing on these targets. It has also been used in more than 20% of studies on communication and gesture recognition, and in few studies focusing on robotic control, its use is still limited. However, it has never been used for prosthetic control HMIs, probably because of the unsuitability of body motion capture technologies in the typical prosthetic control scenarios. Indeed, a great part of these technologies are image-based, so they do not fit very well to the inherent wearability needs of prosthetic applications and, in addition, they require a considerably higher computational power/cost; the nonimage-based technologies are based on inertial measurement units, which usually require specific limbs motion, thus interfering with the final limb actions to be performed or requiring higher coordination efforts even to perform simple tasks.
The use of hybrid approaches has also experienced considerable growth. Undoubtedly, the combined use of different biosignals could help compensate for the weaknesses of each single biosignal, thus leading to improved performances. On the other hand, however, this is achieved at higher costs in terms of both hardware and software complexities. Therefore, the pros and cons of hybrid approaches should be carefully evaluated and may vary among different targets. Indeed, their use in the control of robots, smart environments, and virtual realities is still limited. In contrast, in prosthetic control, gesture recognition, and communication, their use has reached or overcome the use of other biosignals. In particular, hybrid technologies have been used in 40% of all studies focused on communication HMIs, thus overcoming all other biosignals. This result suggests that the considerable complexity of communication tasks benefited from combining different biosignals much more than other target applications.

5. Conclusions

This survey outlines the developments and research trends in biosignal-based Human–Machine Interfaces designed for assistive and rehabilitation purposes. Recently, several patents have filed for assistance and rehabilitation devices, demonstrating the ever-growing interest in developing new technologies by companies, universities, and research institutes. Some examples are: an ambulation exoskeleton for limb rehabilitation in patients suffering from poststroke motor disabilities; a lower limb wearable robotic system able to support the flexion/extension movements of the hip and knee joints for rehabilitation applications; a hand exoskeleton for use in robot-assisted rehabilitation; a vision rehabilitation system based on the remote control of the physician; a grip assistance device based on functional electrical stimulation (FES), which perceives the movement intentions by the EMG signal from a paretic limb and provides electrical stimuli in order to activate the muscles involved in grasp actions; etc. [223].
In conclusion, biopotential-based HMIs are involved in all considered targets. Their major drawback is the need for stable electrical contact over time. Among these HMIs, BCI is a promising exploration area and could be crucial for people with severe disabilities. However, this technology is not mature enough to be reliably used in daily life environments.
In the last several years, many alternatives to biopotentials have been available in the field of assistive and rehabilitative HMIs. Recently, research on prosthetic control has been strongly directed towards using muscle mechanical motion as an EMG alternative; it has also been widely tested for all the other targets considered. HMIs that exploit this kind of biosignals achieve comparable performances to EMG, overcoming its well-known limitations. Nonetheless, they do require the user to have adequate muscle contraction capacity. Body motion-based HMIs are mainly used for smart environment and virtual reality control. The use of image-based and nonimage-based techniques turned out to be well balanced. Hybrid HMIs represent an emergent trend and have already provided a remarkable contribution to the communication target. They could help compensate for single biosignals’ weaknesses but must be adapted to specific applications.
The future will probably be characterized by the development of interactive/intelligent systems that allow any person, even someone affected by severe motor disabilities, to provide commands to a wide range of machines. Shared control, predictable machine learning techniques, closed-loop control based on sensory feedback, and human-oriented design requirements will be the critical challenges for future research. In addition, the HMIs design, especially in the field of assistive technologies, would probably be directed towards the development of “universal interfaces” [224], capable of recognizing and executing any user command (e.g., home automation devices, smart wheelchairs, personal communicators, etc.) in order to let these systems to be easily used by all kinds of users, regardless of age, languages, and degree of disability.

Author Contributions

Conceptualization, D.E. and P.B.; methodology and formal analysis, D.E. and J.C.; investigation, D.E., J.C., E.A., G.D.G., G.R.N. and P.B.; writing—original draft preparation, D.E. and J.C.; writing—review and editing, D.E., J.C., E.A., G.D.G., G.R.N. and P.B.; visualization, D.E. and J.C.; supervision, P.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Singh, H.P.; Kumar, P. Developments in the Human Machine Interface Technologies and Their Applications: A Review. J. Med. Eng. Technol. 2021, 45, 552–573. [Google Scholar] [CrossRef]
  2. Kakkos, I.; Miloulis, S.-T.; Gkiatis, K.; Dimitrakopoulos, G.N.; Matsopoulos, G.K. Human–Machine Interfaces for Motor Rehabilitation. In Advanced Computational Intelligence in Healthcare-7: Biomedical Informatics; Maglogiannis, I., Brahnam, S., Jain, L.C., Eds.; Studies in Computational Intelligence; Springer: Berlin/Heidelberg, Germany, 2020; pp. 1–16. ISBN 978-3-662-61114-2. [Google Scholar]
  3. Beck, T.W.; Housh, T.J.; Cramer, J.T.; Weir, J.P.; Johnson, G.O.; Coburn, J.W.; Malek, M.H.; Mielke, M. Mechanomyographic Amplitude and Frequency Responses during Dynamic Muscle Actions: A Comprehensive Review. Biomed. Eng. Online 2005, 4, 67. [Google Scholar] [CrossRef] [Green Version]
  4. Xiao, Z.G.; Menon, C. A Review of Force Myography Research and Development. Sensors 2019, 19, 4557. [Google Scholar] [CrossRef] [Green Version]
  5. Lazarou, I.; Nikolopoulos, S.; Petrantonakis, P.C.; Kompatsiaris, I.; Tsolaki, M. EEG-Based Brain–Computer Interfaces for Communication and Rehabilitation of People with Motor Impairment: A Novel Approach of the 21st Century. Front. Hum. Neurosci. 2018, 12, 14. [Google Scholar] [CrossRef] [Green Version]
  6. Ptito, M.; Bleau, M.; Djerourou, I.; Paré, S.; Schneider, F.C.; Chebat, D.-R. Brain-Machine Interfaces to Assist the Blind. Front. Hum. Neurosci. 2021, 15, 638887. [Google Scholar] [CrossRef]
  7. Baniqued, P.D.E.; Stanyer, E.C.; Awais, M.; Alazmani, A.; Jackson, A.E.; Mon-Williams, M.A.; Mushtaq, F.; Holt, R.J. Brain–Computer Interface Robotics for Hand Rehabilitation after Stroke: A Systematic Review. J. Neuroeng. Rehabil. 2021, 18, 15. [Google Scholar] [CrossRef]
  8. Mrachacz-Kersting, N.; Jiang, N.; Stevenson, A.J.T.; Niazi, I.K.; Kostic, V.; Pavlovic, A.; Radovanovic, S.; Djuric-Jovicic, M.; Agosta, F.; Dremstrup, K.; et al. Efficient Neuroplasticity Induction in Chronic Stroke Patients by an Associative Brain-Computer Interface. J. Neurophysiol. 2016, 115, 1410–1421. [Google Scholar] [CrossRef] [PubMed]
  9. Ahmadizadeh, C.; Khoshnam, M.; Menon, C. Human Machine Interfaces in Upper-Limb Prosthesis Control: A Survey of Techniques for Preprocessing and Processing of Biosignals. IEEE Signal Process. Mag. 2021, 38, 12–22. [Google Scholar] [CrossRef]
  10. Grushko, S.; Spurný, T.; Černý, M. Control Methods for Transradial Prostheses Based on Remnant Muscle Activity and Its Relationship with Proprioceptive Feedback. Sensors 2020, 20, 4883. [Google Scholar] [CrossRef] [PubMed]
  11. Úbeda, A.; Iáñez, E.; Azorín, J.M. Shared Control Architecture Based on RFID to Control a Robot Arm Using a Spontaneous Brain–Machine Interface. Robot. Auton. Syst. 2013, 61, 768–774. [Google Scholar] [CrossRef]
  12. Ma, J.; Zhang, Y.; Cichocki, A.; Matsuno, F. A Novel EOG/EEG Hybrid Human-Machine Interface Adopting Eye Movements and ERPs: Application to Robot Control. IEEE Trans. Biomed. Eng. 2015, 62, 876–889. [Google Scholar] [CrossRef]
  13. Xia, W.; Zhou, Y.; Yang, X.; He, K.; Liu, H. Toward Portable Hybrid Surface Electromyography/A-Mode Ultrasound Sensing for Human–Machine Interface. IEEE Sens. J. 2019, 19, 5219–5228. [Google Scholar] [CrossRef]
  14. Xu, B.; Li, W.; He, X.; Wei, Z.; Zhang, D.; Wu, C.; Song, A. Motor Imagery Based Continuous Teleoperation Robot Control with Tactile Feedback. Electronics 2020, 9, 174. [Google Scholar] [CrossRef] [Green Version]
  15. Varada, V.; Moolchandani, D.; Rohit, A. Measuring and Processing the Brain’s EEG Signals with Visual Feedback for Human Machine Interface. Int. J. Sci. Eng. Res. 2013, 4, 1–4. [Google Scholar]
  16. Zhu, M.; Sun, Z.; Zhang, Z.; Shi, Q.; He, T.; Liu, H.; Chen, T.; Lee, C. Haptic-Feedback Smart Glove as a Creative Human-Machine Interface (HMI) for Virtual/Augmented Reality Applications. Sci. Adv. 2020, 6, eaaz8693. [Google Scholar] [CrossRef] [PubMed]
  17. National Library of Medicine—National Institutes of Health. Available online: https://www.nlm.nih.gov/ (accessed on 9 August 2021).
  18. Taylor, M.; Griffin, M. The Use of Gaming Technology for Rehabilitation in People with Multiple Sclerosis. Mult. Scler. 2015, 21, 355–371. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Fatima, N.; Shuaib, A.; Saqqur, M. Intra-Cortical Brain-Machine Interfaces for Controlling Upper-Limb Powered Muscle and Robotic Systems in Spinal Cord Injury. Clin. Neurol. Neurosurg. 2020, 196, 106069. [Google Scholar] [CrossRef]
  20. Garcia-Agundez, A.; Folkerts, A.-K.; Konrad, R.; Caserman, P.; Tregel, T.; Goosses, M.; Göbel, S.; Kalbe, E. Recent Advances in Rehabilitation for Parkinson’s Disease with Exergames: A Systematic Review. J. Neuroeng. Rehabil. 2019, 16, 17. [Google Scholar] [CrossRef]
  21. Mohebbi, A. Human-Robot Interaction in Rehabilitation and Assistance: A Review. Curr. Robot. Rep. 2020, 1, 131–144. [Google Scholar] [CrossRef]
  22. Frisoli, A.; Solazzi, M.; Loconsole, C.; Barsotti, M. New Generation Emerging Technologies for Neurorehabilitation and Motor Assistance. Acta Myol. 2016, 35, 141–144. [Google Scholar] [PubMed]
  23. Wright, J.; Macefield, V.G.; van Schaik, A.; Tapson, J.C. A Review of Control Strategies in Closed-Loop Neuroprosthetic Systems. Front. Neurosci. 2016, 10, 312. [Google Scholar] [CrossRef] [PubMed]
  24. Ciancio, A.L.; Cordella, F.; Barone, R.; Romeo, R.A.; Bellingegni, A.D.; Sacchetti, R.; Davalli, A.; Di Pino, G.; Ranieri, F.; Di Lazzaro, V.; et al. Control of Prosthetic Hands via the Peripheral Nervous System. Front. Neurosci. 2016, 10, 116. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Ngan, C.G.Y.; Kapsa, R.M.I.; Choong, P.F.M. Strategies for Neural Control of Prosthetic Limbs: From Electrode Interfacing to 3D Printing. Materials 2019, 12, 1927. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Parajuli, N.; Sreenivasan, N.; Bifulco, P.; Cesarelli, M.; Savino, S.; Niola, V.; Esposito, D.; Hamilton, T.J.; Naik, G.R.; Gunawardana, U.; et al. Real-Time EMG Based Pattern Recognition Control for Hand Prostheses: A Review on Existing Methods, Challenges and Future Implementation. Sensors 2019, 19, 4596. [Google Scholar] [CrossRef] [Green Version]
  27. Igual, C.; Pardo, L.A.; Hahne, J.M.; Igual, J. Myoelectric Control for Upper Limb Prostheses. Electronics 2019, 8, 1244. [Google Scholar] [CrossRef]
  28. Kumar, D.K.; Jelfs, B.; Sui, X.; Arjunan, S.P. Prosthetic Hand Control: A Multidisciplinary Review to Identify Strengths, Shortcomings, and the Future. Biomed. Signal Process. Control 2019, 53, 101588. [Google Scholar] [CrossRef]
  29. Da Gama, A.; Fallavollita, P.; Teichrieb, V.; Navab, N. Motor Rehabilitation Using Kinect: A Systematic Review. Games Health J. 2015, 4, 123–135. [Google Scholar] [CrossRef]
  30. Laver, K.E.; Lange, B.; George, S.; Deutsch, J.E.; Saposnik, G.; Crotty, M. Virtual Reality for Stroke Rehabilitation. Cochrane Database Syst. Rev. 2017, 2017, CD008349. [Google Scholar] [CrossRef] [Green Version]
  31. Mat Rosly, M.; Mat Rosly, H.; Davis Oam, G.M.; Husain, R.; Hasnan, N. Exergaming for Individuals with Neurological Disability: A Systematic Review. Disabil. Rehabil. 2017, 39, 727–735. [Google Scholar] [CrossRef]
  32. Reis, E.; Postolache, G.; Teixeira, L.; Arriaga, P.; Lima, M.L.; Postolache, O. Exergames for Motor Rehabilitation in Older Adults: An Umbrella Review. Phys. Ther. Rev. 2019, 24, 84–99. [Google Scholar] [CrossRef]
  33. Li, W.; Shi, P.; Yu, H. Gesture Recognition Using Surface Electromyography and Deep Learning for Prostheses Hand: State-of-the-Art, Challenges, and Future. Front. Neurosci. 2021, 15, 621885. [Google Scholar] [CrossRef]
  34. Liang, H.; Zhu, C.; Iwata, Y.; Maedono, S.; Mochita, M.; Liu, C.; Ueda, N.; Li, P.; Yu, H.; Yan, Y.; et al. Feature Extraction of Shoulder Joint’s Voluntary Flexion-Extension Movement Based on Electroencephalography Signals for Power Assistance. Bioengineering 2018, 6, 2. [Google Scholar] [CrossRef] [Green Version]
  35. He, Y.; Nathan, K.; Venkatakrishnan, A.; Rovekamp, R.; Beck, C.; Ozdemir, R.; Francisco, G.E.; Contreras-Vidal, J.L. An Integrated Neuro-Robotic Interface for Stroke Rehabilitation Using the NASA X1 Powered Lower Limb Exoskeleton. In Proceedings of the 2014 36th Annual International Conference of the IEEE Engineering in Medicine and Biology, Chicago, IL, USA, 26–30 August 2014; pp. 3985–3988. [Google Scholar] [CrossRef]
  36. Xu, R.; Jiang, N.; Mrachacz-Kersting, N.; Lin, C.; Asín Prieto, G.; Moreno, J.C.; Pons, J.L.; Dremstrup, K.; Farina, D. A Closed-Loop Brain-Computer Interface Triggering an Active Ankle-Foot Orthosis for Inducing Cortical Neural Plasticity. IEEE Trans. Biomed. Eng. 2014, 61, 2092–2101. [Google Scholar] [CrossRef]
  37. Fall, C.L.; Gagnon-Turcotte, G.; Dube, J.-F.; Gagne, J.S.; Delisle, Y.; Campeau-Lecours, A.; Gosselin, C.; Gosselin, B. Wireless SEMG-Based Body-Machine Interface for Assistive Technology Devices. IEEE J. Biomed. Health Inform. 2017, 21, 967–977. [Google Scholar] [CrossRef]
  38. Laksono, P.W.; Kitamura, T.; Muguro, J.; Matsushita, K.; Sasaki, M.; Amri bin Suhaimi, M.S. Minimum Mapping from EMG Signals at Human Elbow and Shoulder Movements into Two DoF Upper-Limb Robot with Machine Learning. Machines 2021, 9, 56. [Google Scholar] [CrossRef]
  39. Alibhai, Z.; Burreson, T.; Stiller, M.; Ahmad, I.; Huber, M.; Clark, A. A Human-Computer Interface For Smart Wheelchair Control Using Forearm EMG Signals. In Proceedings of the 2020 3rd International Conference on Data Intelligence and Security (ICDIS), South Padre Island, TX, USA, 24–26 June 2020; pp. 34–39. [Google Scholar]
  40. Song, J.-H.; Jung, J.-W.; Lee, S.-W.; Bien, Z. Robust EMG Pattern Recognition to Muscular Fatigue Effect for Powered Wheelchair Control. J. Intell. Fuzzy Syst. 2009, 20, 3–12. [Google Scholar] [CrossRef]
  41. Xu, X.; Zhang, Y.; Luo, Y.; Chen, D. Robust Bio-Signal Based Control of an Intelligent Wheelchair. Robotics 2013, 2, 187–197. [Google Scholar] [CrossRef] [Green Version]
  42. Zhang, R.; He, S.; Yang, X.; Wang, X.; Li, K.; Huang, Q.; Yu, Z.; Zhang, X.; Tang, D.; Li, Y. An EOG-Based Human–Machine Interface to Control a Smart Home Environment for Patients with Severe Spinal Cord Injuries. IEEE Trans. Biomed. Eng. 2019, 66, 89–100. [Google Scholar] [CrossRef] [PubMed]
  43. Bissoli, A.; Lavino-Junior, D.; Sime, M.; Encarnação, L.; Bastos-Filho, T. A Human–Machine Interface Based on Eye Tracking for Controlling and Monitoring a Smart Home Using the Internet of Things. Sensors 2019, 19, 859. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Clark, J.W.J. The origin of biopotentials. In Medical Instrumentation: Application and Design, 4th ed.; Webster, J.G., Ed.; John Wiley & Sons: Hoboken, NJ, USA, 2010; pp. 126–181. [Google Scholar]
  45. Thakor, N.V. Biopotentials and Electrophysiology Measurements. In Measurement, Instrumentation, and Sensors Handbook: Electromagnetic, Optical, Radiation, Chemical, and Biomedical Measurement, 2nd ed.; Webster, J.G., Eren, H., Eds.; CRC Press: Boca Raton, FL, USA, 2017; pp. 1–7. [Google Scholar]
  46. Gao, H.; Luo, L.; Pi, M.; Li, Z.; Li, Q.; Zhao, K.; Huang, J. EEG-Based Volitional Control of Prosthetic Legs for Walking in Different Terrains. IEEE Trans. Autom. Sci. Eng. 2021, 18, 530–540. [Google Scholar] [CrossRef]
  47. Gannouni, S.; Belwafi, K.; Aboalsamh, H.; AlSamhan, Z.; Alebdi, B.; Almassad, Y.; Alobaedallah, H. EEG-Based BCI System to Detect Fingers Movements. Brain Sci. 2020, 10, 965. [Google Scholar] [CrossRef]
  48. Fuentes-Gonzalez, J.; Infante-Alarcón, A.; Asanza, V.; Loayza, F.R. A 3D-Printed EEG Based Prosthetic Arm. In Proceedings of the 2020 IEEE International Conference on E-health Networking, Application Services (HEALTHCOM), Shenzhen, China, 1–2 March 2021; pp. 1–5. [Google Scholar]
  49. Song, Y.; Cai, S.; Yang, L.; Li, G.; Wu, W.; Xie, L. A Practical EEG-Based Human-Machine Interface to Online Control an Upper-Limb Assist Robot. Front. Neurorobot. 2020, 14, 32. [Google Scholar] [CrossRef]
  50. Korovesis, N.; Kandris, D.; Koulouras, G.; Alexandridis, A. Robot Motion Control via an EEG-Based Brain–Computer Interface by Using Neural Networks and Alpha Brainwaves. Electronics 2019, 8, 1387. [Google Scholar] [CrossRef] [Green Version]
  51. Gordleeva, S.Y.; Lobov, S.A.; Grigorev, N.A.; Savosenkov, A.O.; Shamshin, M.O.; Lukoyanov, M.V.; Khoruzhko, M.A.; Kazantsev, V.B. Real-Time EEG–EMG Human–Machine Interface-Based Control System for a Lower-Limb Exoskeleton. IEEE Access 2020, 8, 84070–84081. [Google Scholar] [CrossRef]
  52. Noce, E.; Dellacasa Bellingegni, A.; Ciancio, A.L.; Sacchetti, R.; Davalli, A.; Guglielmelli, E.; Zollo, L. EMG and ENG-Envelope Pattern Recognition for Prosthetic Hand Control. J. Neurosci. Methods 2019, 311, 38–46. [Google Scholar] [CrossRef]
  53. Eisenberg, G.D.; Fyvie, K.G.H.M.; Mohamed, A.-K. Real-Time Segmentation and Feature Extraction of Electromyography: Towards Control of a Prosthetic Hand. IFAC-PapersOnLine 2017, 50, 151–156. [Google Scholar] [CrossRef]
  54. Tavakoli, M.; Benussi, C.; Lourenco, J.L. Single Channel Surface EMG Control of Advanced Prosthetic Hands: A Simple, Low Cost and Efficient Approach. Expert Syst. Appl. 2017, 79, 322–332. [Google Scholar] [CrossRef]
  55. Nguyen, A.T.; Xu, J.; Jiang, M.; Luu, D.K.; Wu, T.; Tam, W.-K.; Zhao, W.; Drealan, M.W.; Overstreet, C.K.; Zhao, Q.; et al. A Bioelectric Neural Interface towards Intuitive Prosthetic Control for Amputees. J. Neural Eng. 2020, 17, 066001. [Google Scholar] [CrossRef] [PubMed]
  56. Golparvar, A.J.; Yapici, M.K. Toward Graphene Textiles in Wearable Eye Tracking Systems for Human–Machine Interaction. Beilstein J. Nanotechnol. 2021, 12, 180–189. [Google Scholar] [CrossRef] [PubMed]
  57. Huang, Q.; He, S.; Wang, Q.; Gu, Z.; Peng, N.; Li, K.; Zhang, Y.; Shao, M.; Li, Y. An EOG-Based Human–Machine Interface for Wheelchair Control. IEEE Trans. Biomed. Eng. 2018, 65, 2023–2032. [Google Scholar] [CrossRef] [PubMed]
  58. Arrow, C.; Wu, H.; Baek, S.; Iu, H.H.C.; Nazarpour, K.; Eshraghian, J.K. Prosthesis Control Using Spike Rate Coding in the Retina Photoreceptor Cells. In Proceedings of the 2021 IEEE International Symposium on Circuits and Systems (ISCAS), Daegu, Korea, 22–28 May 2021; pp. 1–5. [Google Scholar]
  59. Noce, E.; Gentile, C.; Cordella, F.; Ciancio, A.L.; Piemonte, V.; Zollo, L. Grasp Control of a Prosthetic Hand through Peripheral Neural Signals. J. Phys. Conf. Ser. 2018, 1026, 012006. [Google Scholar] [CrossRef]
  60. Towe, B.C. Bioelectricity and its measurement. In Biomedical Engineering and Design Handbook, 2nd ed; Kutz, M., Ed.; McGraw-Hill Education: New York, NY, USA, 2009; pp. 481–527. [Google Scholar]
  61. Barr, R.C. Basic Electrophysiology. In Biomedical Engineering Handbook; Bronzino, J.D., Ed.; CRC Press: Boca Raton, FL, USA, 1999; pp. 146–162. [Google Scholar]
  62. Miller, L.E.; Hatsopoulos, N. Neural activity in motor cortex and related areas. In Brain–Computer Interfaces: Principles and Practice; Wolpaw, J., Wolpaw, E.W., Eds.; Oxford University Press: New York, NY, USA, 2012; pp. 15–43. [Google Scholar]
  63. Pfurtscheller, G.; Lopes da Silva, F.H. Event-Related EEG/MEG Synchronization and Desynchronization: Basic Principles. Clin. Neurophysiol. 1999, 110, 1842–1857. [Google Scholar] [CrossRef]
  64. Antoniou, E.; Bozios, P.; Christou, V.; Tzimourta, K.D.; Kalafatakis, K.; Tsipouras, M.G.; Giannakeas, N.; Tzallas, A.T. EEG-Based Eye Movement Recognition Using Brain–Computer Interface and Random Forests. Sensors 2021, 21, 2339. [Google Scholar] [CrossRef] [PubMed]
  65. Matsushita, K.; Hirata, M.; Suzuki, T.; Ando, H.; Yoshida, T.; Ota, Y.; Sato, F.; Morris, S.; Sugata, H.; Goto, T.; et al. A Fully Implantable Wireless ECoG 128-Channel Recording Device for Human Brain–Machine Interfaces: W-HERBS. Front. Neurosci. 2018, 12, 511. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  66. Spataro, R.; Chella, A.; Allison, B.; Giardina, M.; Sorbello, R.; Tramonte, S.; Guger, C.; La Bella, V. Reaching and Grasping a Glass of Water by Locked-In ALS Patients through a BCI-Controlled Humanoid Robot. Front. Hum. Neurosci. 2017, 11, 68. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  67. López-Larraz, E.; Trincado-Alonso, F.; Rajasekaran, V.; Pérez-Nombela, S.; del-Ama, A.J.; Aranda, J.; Minguez, J.; Gil-Agudo, A.; Montesano, L. Control of an Ambulatory Exoskeleton with a Brain–Machine Interface for Spinal Cord Injury Gait Rehabilitation. Front. Neurosci. 2016, 10, 359. [Google Scholar] [CrossRef]
  68. Hortal, E.; Planelles, D.; Costa, A.; Iáñez, E.; Úbeda, A.; Azorín, J.M.; Fernández, E. SVM-Based Brain–Machine Interface for Controlling a Robot Arm through Four Mental Tasks. Neurocomputing 2015, 151, 116–121. [Google Scholar] [CrossRef]
  69. Wang, H.; Su, Q.; Yan, Z.; Lu, F.; Zhao, Q.; Liu, Z.; Zhou, F. Rehabilitation Treatment of Motor Dysfunction Patients Based on Deep Learning Brain–Computer Interface Technology. Front. Neurosci. 2020, 14, 595084. [Google Scholar] [CrossRef]
  70. Hong, L.Z.; Zourmand, A.; Victor Patricks, J.; Thing, G.T. EEG-Based Brain Wave Controlled Intelligent Prosthetic Arm. In Proceedings of the 2020 IEEE 8th Conference on Systems, Process and Control (ICSPC), Melaka, Malaysia, 11–12 December 2020; pp. 52–57. [Google Scholar]
  71. Ortiz, M.; Ferrero, L.; Iáñez, E.; Azorín, J.M.; Contreras-Vidal, J.L. Sensory Integration in Human Movement: A New Brain-Machine Interface Based on Gamma Band and Attention Level for Controlling a Lower-Limb Exoskeleton. Front. Bioeng. Biotechnol. 2020, 8, 735. [Google Scholar] [CrossRef]
  72. Kasim, M.A.A.; Low, C.Y.; Ayub, M.A.; Zakaria, N.A.C.; Salleh, M.H.M.; Johar, K.; Hamli, H. User-Friendly LabVIEW GUI for Prosthetic Hand Control Using Emotiv EEG Headset. Procedia Comput. Sci. 2017, 105, 276–281. [Google Scholar] [CrossRef]
  73. Murphy, D.P.; Bai, O.; Gorgey, A.S.; Fox, J.; Lovegreen, W.T.; Burkhardt, B.W.; Atri, R.; Marquez, J.S.; Li, Q.; Fei, D.-Y. Electroencephalogram-Based Brain–Computer Interface and Lower-Limb Prosthesis Control: A Case Study. Front. Neurol. 2017, 8, 696. [Google Scholar] [CrossRef] [Green Version]
  74. Li, G.; Jiang, S.; Xu, Y.; Wu, Z.; Chen, L.; Zhang, D. A Preliminary Study towards Prosthetic Hand Control Using Human Stereo-Electroencephalography (SEEG) Signals. In Proceedings of the 2017 8th International IEEE/EMBS Conference on Neural Engineering (NER), Shanghai, China, 25–28 May 2017; pp. 375–378. [Google Scholar]
  75. Bhagat, N.A.; Venkatakrishnan, A.; Abibullaev, B.; Artz, E.J.; Yozbatiran, N.; Blank, A.A.; French, J.; Karmonik, C.; Grossman, R.G.; O’Malley, M.K.; et al. Design and Optimization of an EEG-Based Brain Machine Interface (BMI) to an Upper-Limb Exoskeleton for Stroke Survivors. Front. Neurosci. 2016, 10, 122. [Google Scholar] [CrossRef]
  76. Morishita, S.; Sato, K.; Watanabe, H.; Nishimura, Y.; Isa, T.; Kato, R.; Nakamura, T.; Yokoi, H. Brain-Machine Interface to Control a Prosthetic Arm with Monkey ECoGs during Periodic Movements. Front. Neurosci. 2014, 8, 417. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  77. Zhang, X.; Li, R.; Li, Y. Research on Brain Control Prosthetic Hand. In Proceedings of the 2014 11th International Conference on Ubiquitous Robots and Ambient Intelligence (URAI), Kuala Lumpur, Malaysia, 12–15 November 2014; pp. 554–557. [Google Scholar]
  78. Yanagisawa, T.; Hirata, M.; Saitoh, Y.; Goto, T.; Kishima, H.; Fukuma, R.; Yokoi, H.; Kamitani, Y.; Yoshimine, T. Real-Time Control of a Prosthetic Hand Using Human Electrocorticography Signals. J. Neurosurg. 2011, 114, 1715–1722. [Google Scholar] [CrossRef] [PubMed]
  79. Tang, Z.; Sun, S.; Zhang, S.; Chen, Y.; Li, C.; Chen, S. A Brain-Machine Interface Based on ERD/ERS for an Upper-Limb Exoskeleton Control. Sensors 2016, 16, 2050. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  80. Randazzo, L.; Iturrate, I.; Perdikis, S.; Millán, J.D. Mano: A Wearable Hand Exoskeleton for Activities of Daily Living and Neurorehabilitation. IEEE Robot. Autom. Lett. 2018, 3, 500–507. [Google Scholar] [CrossRef] [Green Version]
  81. Li, Z.; Li, J.; Zhao, S.; Yuan, Y.; Kang, Y.; Chen, C.L.P. Adaptive Neural Control of a Kinematically Redundant Exoskeleton Robot Using Brain–Machine Interfaces. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3558–3571. [Google Scholar] [CrossRef]
  82. Kwak, N.-S.; Müller, K.-R.; Lee, S.-W. A Lower Limb Exoskeleton Control System Based on Steady State Visual Evoked Potentials. J. Neural Eng. 2015, 12, 056009. [Google Scholar] [CrossRef]
  83. Araujo, R.S.; Silva, C.R.; Netto, S.P.N.; Morya, E.; Brasil, F.L. Development of a Low-Cost EEG-Controlled Hand Exoskeleton 3D Printed on Textiles. Front. Neurosci. 2021, 15, 661569. [Google Scholar] [CrossRef]
  84. Kashihara, K. A Brain-Computer Interface for Potential Non-Verbal Facial Communication Based on EEG Signals Related to Specific Emotions. Front. Neurosci. 2014, 8, 244. [Google Scholar] [CrossRef] [Green Version]
  85. Mahmoudi, B.; Erfanian, A. Single-Channel EEG-Based Prosthetic Hand Grasp Control for Amputee Subjects. In Proceedings of the Second Joint 24th Annual Conference and the Annual Fall Meeting of the Biomedical Engineering Society, Engineering in Medicine and Biology, Houston, TX, USA, 23–26 October 2002; Volume 3, pp. 2406–2407. [Google Scholar]
  86. De Luca, C.J. Electromyography. In Encyclopedia of Medical Devices and Instrumentation, 2nd ed.; Webster, J.G., Ed.; John Wiley & Sons: Hoboken, NJ, USA, 2006; pp. 98–109. [Google Scholar]
  87. Bai, D.; Liu, T.; Han, X.; Chen, G.; Jiang, Y.; Hiroshi, Y. Multi-Channel SEMG Signal Gesture Recognition Based on Improved CNN-LSTM Hybrid Models. In Proceedings of the 2021 IEEE International Conference on Intelligence and Safety for Robotics (ISR), Tokoname, Japan, 4–6 March 2021; pp. 111–116. [Google Scholar]
  88. Cao, T.; Liu, D.; Wang, Q.; Bai, O.; Sun, J. Surface Electromyography-Based Action Recognition and Manipulator Control. Appl. Sci. 2020, 10, 5823. [Google Scholar] [CrossRef]
  89. Benatti, S.; Milosevic, B.; Farella, E.; Gruppioni, E.; Benini, L. A Prosthetic Hand Body Area Controller Based on Efficient Pattern Recognition Control Strategies. Sensors 2017, 17, 869. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  90. Ulloa, G.D.F.; Sreenivasan, N.; Bifulco, P.; Cesarelli, M.; Gargiulo, G.; Gunawardana, U. Cost Effective Electro—Resistive Band Based Myo Activated Prosthetic Upper Limb for Amputees in the Developing World. In Proceedings of the 2017 IEEE Life Sciences Conference (LSC), Sydney, NSW, Australia, 13–15 December 2017; pp. 250–253. [Google Scholar]
  91. Polisiero, M.; Bifulco, P.; Liccardo, A.; Cesarelli, M.; Romano, M.; Gargiulo, G.D.; McEwan, A.L.; D’Apuzzo, M. Design and Assessment of a Low-Cost, Electromyographically Controlled, Prosthetic Hand. Med. Devices 2013, 6, 97–104. [Google Scholar] [CrossRef] [Green Version]
  92. Gailey, A.; Artemiadis, P.; Santello, M. Proof of Concept of an Online EMG-Based Decoding of Hand Postures and Individual Digit Forces for Prosthetic Hand Control. Front. Neurol. 2017, 8, 7. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  93. Bernardino, A.; Rybarczyk, Y.; Barata, J. Versatility of Human Body Control through Low-Cost Electromyographic Interface. In Proceedings of the International Conference on Applications of Computer Engineering, San Francisco, CA, USA, 22–24 October 2014. [Google Scholar]
  94. Zhao, J.; Jiang, L.; Shi, S.; Cai, H.; Liu, H.; Hirzinger, G. A Five-Fingered Underactuated Prosthetic Hand System. In Proceedings of the 2006 International Conference on Mechatronics and Automation, Luoyang, China, 25–28 June 2006; pp. 1453–1458. [Google Scholar]
  95. Carozza, M.C.; Cappiello, G.; Stellin, G.; Zaccone, F.; Vecchi, F.; Micera, S.; Dario, P. On the Development of a Novel Adaptive Prosthetic Hand with Compliant Joints: Experimental Platform and EMG Control. In Proceedings of the 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems, Edmonton, AB, Canada, 2–6 August 2005; pp. 1271–1276. [Google Scholar]
  96. Jiang, Y.; Togane, M.; Lu, B.; Yokoi, H. SEMG Sensor Using Polypyrrole-Coated Nonwoven Fabric Sheet for Practical Control of Prosthetic Hand. Front. Neurosci. 2017, 11, 33. [Google Scholar] [CrossRef] [PubMed]
  97. Brunelli, D.; Tadesse, A.M.; Vodermayer, B.; Nowak, M.; Castellini, C. Low-Cost Wearable Multichannel Surface EMG Acquisition for Prosthetic Hand Control. In Proceedings of the 2015 6th International Workshop on Advances in Sensors and Interfaces (IWASI), Gallipoli, Italy, 18–19 June 2015; pp. 94–99. [Google Scholar]
  98. Shair, E.F.; Jamaluddin, N.A.; Abdullah, A.R. Finger Movement Discrimination of EMG Signals Towards Improved Prosthetic Control Using TFD. Int. J. Adv. Comput. Sci. Appl. (IJACSA) 2020, 11, 244–251. [Google Scholar] [CrossRef]
  99. Khushaba, R.N.; Kodagoda, S.; Takruri, M.; Dissanayake, G. Toward Improved Control of Prosthetic Fingers Using Surface Electromyogram (EMG) Signals. Expert Syst. Appl. 2012, 39, 10731–10738. [Google Scholar] [CrossRef]
  100. Kamavuako, E.N.; Scheme, E.J.; Englehart, K.B. On the Usability of Intramuscular EMG for Prosthetic Control: A Fitts’ Law Approach. J. Electromyogr. Kinesiol. 2014, 24, 770–777. [Google Scholar] [CrossRef]
  101. Dewald, H.A.; Lukyanenko, P.; Lambrecht, J.M.; Anderson, J.R.; Tyler, D.J.; Kirsch, R.F.; Williams, M.R. Stable, Three Degree-of-Freedom Myoelectric Prosthetic Control via Chronic Bipolar Intramuscular Electrodes: A Case Study. J. Neuroeng. Rehabil. 2019, 16, 147. [Google Scholar] [CrossRef] [Green Version]
  102. Al-Timemy, A.H.; Bugmann, G.; Escudero, J.; Outram, N. Classification of Finger Movements for the Dexterous Hand Prosthesis Control with Surface Electromyography. IEEE J. Biomed. Health Inform. 2013, 17, 608–618. [Google Scholar] [CrossRef]
  103. Zhang, T.; Wang, X.Q.; Jiang, L.; Wu, X.; Feng, W.; Zhou, D.; Liu, H. Biomechatronic Design and Control of an Anthropomorphic Artificial Hand for Prosthetic Applications. Robotica 2016, 34, 2291–2308. [Google Scholar] [CrossRef]
  104. Dalley, S.A.; Varol, H.A.; Goldfarb, M. A Method for the Control of Multigrasp Myoelectric Prosthetic Hands. IEEE Trans. Neural Syst. Rehabil. Eng. 2012, 20, 58–67. [Google Scholar] [CrossRef] [Green Version]
  105. Russo, R.E.; Fernández, J.; Rivera, R.; Kuzman, M.G.; López, J.; Gemin, W.; Revuelta, M.Á. Algorithm of Myoelectric Signals Processing for the Control of Prosthetic Robotic Hands. J. Comput. Sci. Technol. 2018, 18, 28–34. [Google Scholar] [CrossRef]
  106. Stepp, C.E.; Heaton, J.T.; Rolland, R.G.; Hillman, R.E. Neck and Face Surface Electromyography for Prosthetic Voice Control after Total Laryngectomy. IEEE Trans. Neural Syst. Rehabil. Eng. 2009, 17, 146–155. [Google Scholar] [CrossRef] [PubMed]
  107. Visconti, P.; Gaetani, F.; Zappatore, G.A.; Primiceri, P. Technical Features and Functionalities of Myo Armband: An Overview on Related Literature and Advanced Applications of Myoelectric Armbands Mainly Focused on Arm Prostheses. Int. J. Smart Sens. Intell. Syst. 2018, 11, 1–25. [Google Scholar] [CrossRef] [Green Version]
  108. Lu, Z.; Zhou, P. Hands-Free Human-Computer Interface Based on Facial Myoelectric Pattern Recognition. Front. Neurol. 2019, 10, 444. [Google Scholar] [CrossRef] [PubMed]
  109. Kumar, B.; Paul, Y.; Jaswal, R.A. Development of EMG Controlled Electric Wheelchair Using SVM and KNN Classifier for SCI Patients. In Proceedings of the Advanced Informatics for Computing Research, Shimla, India, 15–16 June 2019; pp. 75–83. [Google Scholar]
  110. Kalani, H.; Moghimi, S.; Akbarzadeh, A. Towards an SEMG-Based Tele-Operated Robot for Masticatory Rehabilitation. Comput. Biol. Med. 2016, 75, 243–256. [Google Scholar] [CrossRef] [PubMed]
  111. Zhang, Y.; Zhu, X.; Dai, L.; Luo, Y. Forehead SEMG Signal Based HMI for Hands-Free Control. J. China Univ. Posts Telecommun. 2014, 21, 98–105. [Google Scholar] [CrossRef]
  112. Hamedi, M.; Salleh, S.-H.; Swee, T.T. Kamarulafizam Surface Electromyography-Based Facial Expression Recognition in Bi-Polar Configuration. J. Comput. Sci. 2011, 7, 1407–1415. [Google Scholar] [CrossRef] [Green Version]
  113. Wege, A.; Zimmermann, A. Electromyography Sensor Based Control for a Hand Exoskeleton. In Proceedings of the 2007 IEEE International Conference on Robotics and Biomimetics (ROBIO), Sanya, China, 15–18 December 2007. [Google Scholar] [CrossRef]
  114. Ho, N.S.K.; Tong, K.Y.; Hu, X.L.; Fung, K.L.; Wei, X.J.; Rong, W.; Susanto, E.A. An EMG-Driven Exoskeleton Hand Robotic Training Device on Chronic Stroke Subjects: Task Training System for Stroke Rehabilitation. In Proceedings of the 2011 IEEE International Conference on Rehabilitation Robotics, Zurich, Switzerland, 29 June–1 July 2011; pp. 1–5. [Google Scholar]
  115. Loconsole, C.; Leonardis, D.; Barsotti, M.; Solazzi, M.; Frisoli, A.; Bergamasco, M.; Troncossi, M.; Foumashi, M.M.; Mazzotti, C.; Castelli, V.P. An Emg-Based Robotic Hand Exoskeleton for Bilateral Training of Grasp. In Proceedings of the 2013 World Haptics Conference (WHC), Daejeon, Korea, 14–17 April 2013; pp. 537–542. [Google Scholar]
  116. Hussain, I.; Spagnoletti, G.; Salvietti, G.; Prattichizzo, D. An EMG Interface for the Control of Motion and Compliance of a Supernumerary Robotic Finger. Front. Neurorobot. 2016, 10, 18. [Google Scholar] [CrossRef] [Green Version]
  117. Abdallah, I.B.; Bouteraa, Y.; Rekik, C. Design and Development of 3d Printed Myoelectric Robotic Exoskeleton for Hand Rehabilitation. Int. J. Smart Sens. Intell. Syst. 2017, 10, 341–366. [Google Scholar] [CrossRef] [Green Version]
  118. Secciani, N.; Bianchi, M.; Meli, E.; Volpe, Y.; Ridolfi, A. A Novel Application of a Surface ElectroMyoGraphy-Based Control Strategy for a Hand Exoskeleton System: A Single-Case Study. Int. J. Adv. Robot. Syst. 2019, 16, 1729881419828197. [Google Scholar] [CrossRef] [Green Version]
  119. Song, R.; Tong, K.; Hu, X.; Zhou, W. Myoelectrically Controlled Wrist Robot for Stroke Rehabilitation. J. Neuroeng. Rehabil. 2013, 10, 52. [Google Scholar] [CrossRef] [Green Version]
  120. Liu, Y.; Li, X.; Zhu, A.; Zheng, Z.; Zhu, H. Design and Evaluation of a Surface Electromyography-Controlled Lightweight Upper Arm Exoskeleton Rehabilitation Robot. Int. J. Adv. Robot. Syst. 2021, 18, 17298814211003460. [Google Scholar] [CrossRef]
  121. Cai, S.; Chen, Y.; Huang, S.; Wu, Y.; Zheng, H.; Li, X.; Xie, L. SVM-Based Classification of SEMG Signals for Upper-Limb Self-Rehabilitation Training. Front. Neurorobot. 2019, 13, 31. [Google Scholar] [CrossRef] [Green Version]
  122. Yin, G.; Zhang, X.; Chen, D.; Li, H.; Chen, J.; Chen, C.; Lemos, S. Processing Surface EMG Signals for Exoskeleton Motion Control. Front. Neurorobot. 2020, 14, 40. [Google Scholar] [CrossRef]
  123. Tang, Z.; Zhang, K.; Sun, S.; Gao, Z.; Zhang, L.; Yang, Z. An Upper-Limb Power-Assist Exoskeleton Using Proportional Myoelectric Control. Sensors 2014, 14, 6677–6694. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  124. Lu, Z.; Chen, X.; Zhang, X.; Tong, K.-Y.; Zhou, P. Real-Time Control of an Exoskeleton Hand Robot with Myoelectric Pattern Recognition. Int. J. Neural Syst. 2017, 27, 1750009. [Google Scholar] [CrossRef]
  125. Gui, K.; Liu, H.; Zhang, D. A Practical and Adaptive Method to Achieve EMG-Based Torque Estimation for a Robotic Exoskeleton. IEEE/ASME Trans. Mechatron. 2019, 24, 483–494. [Google Scholar] [CrossRef]
  126. La Scaleia, V.; Sylos-Labini, F.; Hoellinger, T.; Wang, L.; Cheron, G.; Lacquaniti, F.; Ivanenko, Y.P. Control of Leg Movements Driven by EMG Activity of Shoulder Muscles. Front. Hum. Neurosci. 2014, 8, 838. [Google Scholar] [CrossRef] [PubMed]
  127. Lyu, M.; Chen, W.-H.; Ding, X.; Wang, J.; Pei, Z.; Zhang, B. Development of an EMG-Controlled Knee Exoskeleton to Assist Home Rehabilitation in a Game Context. Front. Neurorobot. 2019, 13, 67. [Google Scholar] [CrossRef] [Green Version]
  128. Martínez-Cerveró, J.; Ardali, M.K.; Jaramillo-Gonzalez, A.; Wu, S.; Tonin, A.; Birbaumer, N.; Chaudhary, U. Open Software/Hardware Platform for Human-Computer Interface Based on Electrooculography (EOG) Signal Classification. Sensors 2020, 20, 2443. [Google Scholar] [CrossRef]
  129. Perez Reynoso, F.D.; Niño Suarez, P.A.; Aviles Sanchez, O.F.; Calva Yañez, M.B.; Vega Alvarado, E.; Portilla Flores, E.A. A Custom EOG-Based HMI Using Neural Network Modeling to Real-Time for the Trajectory Tracking of a Manipulator Robot. Front. Neurorobot. 2020, 14, 67. [Google Scholar] [CrossRef]
  130. Choudhari, A.M.; Porwal, P.; Jonnalagedda, V.; Mériaudeau, F. An Electrooculography Based Human Machine Interface for Wheelchair Control. Biocybern. Biomed. Eng. 2019, 39, 673–685. [Google Scholar] [CrossRef]
  131. Heo, J.; Yoon, H.; Park, K.S. A Novel Wearable Forehead EOG Measurement System for Human Computer Interfaces. Sensors 2017, 17, 1485. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  132. Guo, X.; Pei, W.; Wang, Y.; Chen, Y.; Zhang, H.; Wu, X.; Yang, X.; Chen, H.; Liu, Y.; Liu, R. A Human-Machine Interface Based on Single Channel EOG and Patchable Sensor. Biomed. Signal Process. Control 2016, 30, 98–105. [Google Scholar] [CrossRef]
  133. Wu, J.F.; Ang, A.M.S.; Tsui, K.M.; Wu, H.C.; Hung, Y.S.; Hu, Y.; Mak, J.N.F.; Chan, S.C.; Zhang, Z.G. Efficient Implementation and Design of a New Single-Channel Electrooculography-Based Human–Machine Interface System. IEEE Trans. Circuits Syst. II Express Briefs 2015, 62, 179–183. [Google Scholar] [CrossRef] [Green Version]
  134. Ferreira, A.; Silva, R.L.; Celeste, W.C.; Filho, T.F.B.; Filho, M.S. Human–Machine Interface Based on Muscular and Brain Signals Applied to a Robotic Wheelchair. J. Phys. Conf. Ser. 2007, 90, 012094. [Google Scholar] [CrossRef]
  135. Zhang, J.; Wang, B.; Zhang, C.; Xiao, Y.; Wang, M.Y. An EEG/EMG/EOG-Based Multimodal Human-Machine Interface to Real-Time Control of a Soft Robot Hand. Front. Neurorobot. 2019, 13, 7. [Google Scholar] [CrossRef] [Green Version]
  136. Huang, Q.; Zhang, Z.; Yu, T.; He, S.; Li, Y. An EEG-/EOG-Based Hybrid Brain-Computer Interface: Application on Controlling an Integrated Wheelchair Robotic Arm System. Front. Neurosci. 2019, 13, 1243. [Google Scholar] [CrossRef] [Green Version]
  137. Ma, J.; Zhang, Y.; Nam, Y.; Cichocki, A.; Matsuno, F. EOG/ERP Hybrid Human-Machine Interface for Robot Control. In Proceedings of the 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, Tokyo, Japan, 3–7 November 2013; pp. 859–864. [Google Scholar]
  138. Rezazadeh, I.M.; Firoozabadi, M.; Hu, H.; Golpayegani, S.M.R.H. Co-Adaptive and Affective Human-Machine Interface for Improving Training Performances of Virtual Myoelectric Forearm Prosthesis. IEEE Trans. Affect. Comput. 2012, 3, 285–297. [Google Scholar] [CrossRef]
  139. Rezazadeh, I.M.; Firoozabadi, S.M.; Hu, H.; Hashemi Golpayegani, S.M.R. A Novel Human–Machine Interface Based on Recognition of Multi-Channel Facial Bioelectric Signals. Australas. Phys. Eng. Sci. Med. 2011, 34, 497–513. [Google Scholar] [CrossRef] [PubMed]
  140. Iáñez, E.; Ùbeda, A.; Azorín, J.M. Multimodal Human-Machine Interface Based on a Brain-Computer Interface and an Electrooculography Interface. In Proceedings of the 2011 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Boston, MA, USA, 30 August–3 September 2011; pp. 4572–4575. [Google Scholar] [CrossRef]
  141. Laport, F.; Iglesia, D.; Dapena, A.; Castro, P.M.; Vazquez-Araujo, F.J. Proposals and Comparisons from One-Sensor EEG and EOG Human-Machine Interfaces. Sensors 2021, 21, 2220. [Google Scholar] [CrossRef] [PubMed]
  142. Neto, A.F.; Celeste, W.C.; Martins, V.R.; Filho, T.; Filho, M.S. Human-Machine Interface Based on Electro-Biological Signals for Mobile Vehicles. In Proceedings of the 2006 IEEE International Symposium on Industrial Electronics, Montreal, QC, Canada, 9–13 July 2006. [Google Scholar] [CrossRef]
  143. Esposito, D. A Piezoresistive Sensor to Measure Muscle Contraction and Mechanomyography. Sensors 2018, 18, 2553. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  144. Prakash, A.; Sahi, A.K.; Sharma, N.; Sharma, S. Force Myography Controlled Multifunctional Hand Prosthesis for Upper-Limb Amputees. Biomed. Signal Process. Control 2020, 62, 102122. [Google Scholar] [CrossRef]
  145. Wu, Y.; Jiang, D.; Liu, X.; Bayford, R.; Demosthenous, A. A Human-Machine Interface Using Electrical Impedance Tomography for Hand Prosthesis Control. IEEE Trans. Biomed. Circuits Syst. 2018, 12, 1322–1333. [Google Scholar] [CrossRef] [Green Version]
  146. Asheghabadi, A.S.; Moqadam, S.B.; Xu, J. Multichannel Finger Pattern Recognition Using Single-Site Mechanomyography. IEEE Sens. J. 2021, 21, 8184–8193. [Google Scholar] [CrossRef]
  147. Chen, X.; Zheng, Y.-P.; Guo, J.-Y.; Shi, J. Sonomyography (SMG) Control for Powered Prosthetic Hand: A Study with Normal Subjects. Ultrasound Med. Biol. 2010, 36, 1076–1088. [Google Scholar] [CrossRef]
  148. Xiao, Z.; Menon, C. Performance of Forearm FMG and SEMG for Estimating Elbow, Forearm and Wrist Positions. J. Bionic Eng. 2017, 14, 284–295. [Google Scholar] [CrossRef]
  149. Sakr, M.; Jiang, X.; Menon, C. Estimation of User-Applied Isometric Force/Torque Using Upper Extremity Force Myography. Front. Robot. AI 2019, 6, 120. [Google Scholar] [CrossRef] [Green Version]
  150. Sakr, M.; Menon, C. Exploratory Evaluation of the Force Myography (FMG) Signals Usage for Admittance Control of a Linear Actuator. In Proceedings of the 2018 7th IEEE International Conference on Biomedical Robotics and Biomechatronics (Biorob), Enschede, The Netherlands, 26–29 August 2018. [Google Scholar] [CrossRef]
  151. Ahmadizadeh, C.; Pousett, B.; Menon, C. Investigation of Channel Selection for Gesture Classification for Prosthesis Control Using Force Myography: A Case Study. Front. Bioeng. Biotechnol. 2019, 7, 331. [Google Scholar] [CrossRef] [PubMed]
  152. Xiao, Z.; Elnady, A.M.; Menon, C. Control an Exoskeleton for Forearm Rotation Using FMG. In Proceedings of the 5th IEEE RAS/EMBS International Conference on Biomedical Robotics and Biomechatronics, Sao Paulo, Brazil, 12–15 August 2014. [Google Scholar] [CrossRef]
  153. Ferigo, D.; Merhi, L.-K.; Pousett, B.; Xiao, Z.; Menon, C. A Case Study of a Force-Myography Controlled Bionic Hand Mitigating Limb Position Effect. J. Bionic Eng. 2017, 14, 692–705. [Google Scholar] [CrossRef]
  154. Esposito, D. A Piezoresistive Array Armband with Reduced Number of Sensors for Hand Gesture Recognition. Front. Neurorobot. 2020, 13, 114. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  155. Esposito, D.; Savino, S.; Andreozzi, E.; Cosenza, C.; Niola, V.; Bifulco, P. The “Federica” Hand. Bioengineering 2021, 8, 128. [Google Scholar] [CrossRef] [PubMed]
  156. Esposito, D.; Cosenza, C.; Gargiulo, G.D.; Andreozzi, E.; Niola, V.; Fratini, A.; D’Addio, G.; Bifulco, P. Experimental Study to Improve “Federica” Prosthetic Hand and Its Control System. In Proceedings of the XV Mediterranean Conference on Medical and Biological Engineering and Computing—MEDICON 2019, Coimbra, Portugal, 26–28 September 2019; pp. 586–593. [Google Scholar]
  157. Ha, N.; Withanachchi, G.P.; Yihun, Y. Force Myography Signal-Based Hand Gesture Classification for the Implementation of Real-Time Control System to a Prosthetic Hand. In Proceedings of the 2018 Design of Medical Devices Conference, Minneapolis, MN, USA, 9–12 April 2018. [Google Scholar]
  158. Ha, N.; Withanachchi, G.P.; Yihun, Y. Performance of Forearm FMG for Estimating Hand Gestures and Prosthetic Hand Control. J. Bionic Eng. 2019, 16, 88–98. [Google Scholar] [CrossRef]
  159. Fujiwara, E.; Wu, Y.T.; Suzuki, C.K.; de Andrade, D.T.G.; Neto, A.R.; Rohmer, E. Optical Fiber Force Myography Sensor for Applications in Prosthetic Hand Control. In Proceedings of the 2018 IEEE 15th International Workshop on Advanced Motion Control (AMC), Tokyo, Japan, 9–11 March 2018; pp. 342–347. [Google Scholar]
  160. Bifulco, P.; Esposito, D.; Gargiulo, G.D.; Savino, S.; Niola, V.; Iuppariello, L.; Cesarelli, M. A Stretchable, Conductive Rubber Sensor to Detect Muscle Contraction for Prosthetic Hand Control. In Proceedings of the 2017 E-Health and Bioengineering Conference (EHB), Sinaia, Romania, 22–24 June 2017; pp. 173–176. [Google Scholar]
  161. Radmand, A.; Scheme, E.; Englehart, K. High-Density Force Myography: A Possible Alternative for Upper-Limb Prosthetic Control. J. Rehabil. Res. Dev. 2016, 53, 443–456. [Google Scholar] [CrossRef]
  162. Cho, E.; Chen, R.; Merhi, L.-K.; Xiao, Z.; Pousett, B.; Menon, C. Force Myography to Control Robotic Upper Extremity Prostheses: A Feasibility Study. Front. Bioeng. Biotechnol. 2016, 4, 18. [Google Scholar] [CrossRef] [Green Version]
  163. Dong, B.; Yang, Y.; Shi, Q.; Xu, S.; Sun, Z.; Zhu, S.; Zhang, Z.; Kwong, D.-L.; Zhou, G.; Ang, K.-W.; et al. Wearable Triboelectric-Human-Machine Interface (THMI) Using Robust Nanophotonic Readout. ACS Nano 2020, 14, 8915–8930. [Google Scholar] [CrossRef]
  164. An, T.; Anaya, D.V.; Gong, S.; Yap, L.W.; Lin, F.; Wang, R.; Yuce, M.R.; Cheng, W. Self-Powered Gold Nanowire Tattoo Triboelectric Sensors for Soft Wearable Human-Machine Interface. Nano Energy 2020, 77, 105295. [Google Scholar] [CrossRef]
  165. Clemente, F.; Ianniciello, V.; Gherardini, M.; Cipriani, C. Development of an Embedded Myokinetic Prosthetic Hand Controller. Sensors 2019, 19, 3137. [Google Scholar] [CrossRef] [Green Version]
  166. Tarantino, S.; Clemente, F.; Barone, D.; Controzzi, M.; Cipriani, C. The Myokinetic Control Interface: Tracking Implanted Magnets as a Means for Prosthetic Control. Sci. Rep. 2017, 7, 17149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  167. Kumar, S.; Sultan, M.J.; Ullah, A.; Zameer, S.; Siddiqui, S.; Sami, S.K. Human Machine Interface Glove Using Piezoresistive Textile Based Sensors. IOP Conf. Ser. Mater. Sci. Eng. 2018, 414, 012041. [Google Scholar] [CrossRef]
  168. Castellini, C.; Kõiva, R.; Pasluosta, C.; Viegas, C.; Eskofier, B.M. Tactile Myography: An Off-Line Assessment of Able-Bodied Subjects and One Upper-Limb Amputee. Technologies 2018, 6, 38. [Google Scholar] [CrossRef] [Green Version]
  169. Dong, W.; Xiao, L.; Hu, W.; Zhu, C.; Huang, Y.; Yin, Z. Wearable Human–Machine Interface Based on PVDF Piezoelectric Sensor. Trans. Inst. Meas. Control 2017, 39, 398–403. [Google Scholar] [CrossRef]
  170. Lim, S.; Son, D.; Kim, J.; Lee, Y.B.; Song, J.-K.; Choi, S.; Lee, D.J.; Kim, J.H.; Lee, M.; Hyeon, T.; et al. Transparent and Stretchable Interactive Human Machine Interface Based on Patterned Graphene Heterostructures. Adv. Funct. Mater. 2015, 25, 375–383. [Google Scholar] [CrossRef]
  171. Rasouli, M.; Ghosh, R.; Lee, W.W.; Thakor, N.V.; Kukreja, S. Stable Force-Myographic Control of a Prosthetic Hand Using Incremental Learning. In Proceedings of the 2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Milan, Italy, 25–29 August 2015; pp. 4828–4831. [Google Scholar]
  172. Islam, M.A.; Sundaraj, K.; Ahmad, R.B.; Sundaraj, S.; Ahamed, N.U.; Ali, M.A. Cross-Talk in Mechanomyographic Signals from the Forearm Muscles during Sub-Maximal to Maximal Isometric Grip Force. PLoS ONE 2014, 9, e96628. [Google Scholar] [CrossRef] [Green Version]
  173. Islam, M.A.; Sundaraj, K.; Ahmad, R.; Ahamed, N.; Ali, M. Mechanomyography Sensor Development, Related Signal Processing, and Applications: A Systematic Review. IEEE Sens. J. 2013, 13, 2499–2516. [Google Scholar] [CrossRef]
  174. Orizio, C.; Gobbo, M. Mechanomyography. In Encyclopedia of Medical Devices and Instrumentation, 2nd ed.; John Wiley & Sons: Hoboken, NJ, USA, 2006. [Google Scholar]
  175. Ibitoye, M.O.; Hamzaid, N.A.; Zuniga, J.M.; Abdul Wahab, A.K. Mechanomyography and Muscle Function Assessment: A Review of Current State and Prospects. Clin. Biomech. 2014, 29, 691–704. [Google Scholar] [CrossRef]
  176. Castillo, C.S.M.; Wilson, S.; Vaidyanathan, R.; Atashzar, S.F. Wearable MMG-Plus-One Armband: Evaluation of Normal Force on Mechanomyography (MMG) to Enhance Human-Machine Interfacing. IEEE Trans. Neural Syst. Rehabil. Eng. 2021, 29, 196–205. [Google Scholar] [CrossRef]
  177. Wicaksono, D.H.B.; Soetjipto, J.; Ughi, F.; Iskandar, A.A.; Santi, F.A.; Biben, V. Wireless Synchronous Carbon Nanotube-Patch Mechanomyography of Leg Muscles. In Proceedings of the 2020 IEEE SENSORS, Rotterdam, The Netherlands, 25–28 October 2020; pp. 1–4. [Google Scholar]
  178. Xie, H.-B.; Zheng, Y.-P.; Guo, J.-Y. Classification of the Mechanomyogram Signal Using a Wavelet Packet Transform and Singular Value Decomposition for Multifunction Prosthesis Control. Physiol. Meas. 2009, 30, 441–457. [Google Scholar] [CrossRef] [PubMed]
  179. Huang, Y.; Yang, X.; Li, Y.; Zhou, D.; He, K.; Liu, H. Ultrasound-Based Sensing Models for Finger Motion Classification. IEEE J. Biomed. Health Inform. 2018, 22, 1395–1405. [Google Scholar] [CrossRef] [Green Version]
  180. Li, Y.; He, K.; Sun, X.; Liu, H. Human-Machine Interface Based on Multi-Channel Single-Element Ultrasound Transducers: A Preliminary Study. In Proceedings of the 2016 IEEE 18th International Conference on e-Health Networking, Applications and Services (Healthcom), Munich, Germany, 14–16 September 2016; pp. 1–6. [Google Scholar]
  181. Ortenzi, V.; Tarantino, S.; Castellini, C.; Cipriani, C. Ultrasound Imaging for Hand Prosthesis Control: A Comparative Study of Features and Classification Methods. In Proceedings of the 2015 IEEE International Conference on Rehabilitation Robotics (ICORR), Singapore, 11–14 August 2015; pp. 1–6. [Google Scholar]
  182. Sikdar, S.; Rangwala, H.; Eastlake, E.B.; Hunt, I.A.; Nelson, A.J.; Devanathan, J.; Shin, A.; Pancrazio, J.J. Novel Method for Predicting Dexterous Individual Finger Movements by Imaging Muscle Activity Using a Wearable Ultrasonic System. IEEE Trans. Neural Syst. Rehabil. Eng. 2014, 22, 69–76. [Google Scholar] [CrossRef]
  183. Sierra González, D.; Castellini, C. A Realistic Implementation of Ultrasound Imaging as a Human-Machine Interface for Upper-Limb Amputees. Front. Neurorobot. 2013, 7, 17. [Google Scholar] [CrossRef] [Green Version]
  184. Castellini, C.; Gonzalez, D.S. Ultrasound Imaging as a Human-Machine Interface in a Realistic Scenario. In Proceedings of the 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, Tokyo, Japan, 3–7 November 2013; pp. 1486–1492. [Google Scholar]
  185. Shi, J.; Chang, Q.; Zheng, Y.-P. Feasibility of Controlling Prosthetic Hand Using Sonomyography Signal in Real Time: Preliminary Study. J. Rehabil. Res. Dev. 2010, 47, 87–98. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  186. Booth, R.; Goldsmith, P. A Wrist-Worn Piezoelectric Sensor Array for Gesture Input. J. Med. Biol. Eng. 2018, 38, 284–295. [Google Scholar] [CrossRef]
  187. Maule, L.; Luchetti, A.; Zanetti, M.; Tomasin, P.; Pertile, M.; Tavernini, M.; Guandalini, G.M.A.; De Cecco, M. RoboEye, an Efficient, Reliable and Safe Semi-Autonomous Gaze Driven Wheelchair for Domestic Use. Technologies 2021, 9, 16. [Google Scholar] [CrossRef]
  188. Lin, C.-S.; Chan, C.-N.; Lay, Y.-L.; Lee, J.-F.; Yeh, M.-S. An Eye-Tracking Human-Machine Interface Using an Auto Correction Method. J. Med. Biol. Eng. 2007, 27, 105–109. [Google Scholar]
  189. Conci, N.; Ceresato, P.; De Natale, F.G.B. Natural Human-Machine Interface Using an Interactive Virtual Blackboard. In Proceedings of the 2007 IEEE International Conference on Image Processing, San Antonio, TX, USA, 16 September–19 October 2007; Volume 5, pp. V-181–V-184. [Google Scholar]
  190. Baklouti, M.; Bruin, M.; Guitteny, V.; Monacelli, E. A Human-Machine Interface for Assistive Exoskeleton Based on Face Analysis. In Proceedings of the 2008 2nd IEEE RAS EMBS International Conference on Biomedical Robotics and Biomechatronics, Scottsdale, AZ, USA, 19–22 October 2008; pp. 913–918. [Google Scholar]
  191. Chang, C.-M.; Lin, C.-S.; Chen, W.-C.; Chen, C.-T.; Hsu, Y.-L. Development and Application of a Human–Machine Interface Using Head Control and Flexible Numeric Tables for the Severely Disabled. Appl. Sci. 2020, 10, 7005. [Google Scholar] [CrossRef]
  192. Gautam, A.K.; Vasu, V.; Raju, U.S.N. Human Machine Interface for Controlling a Robot Using Image Processing. Procedia Eng. 2014, 97, 291–298. [Google Scholar] [CrossRef] [Green Version]
  193. Gmez-Portes, C.; Lacave, C.; Molina, A.I.; Vallejo, D. Home Rehabilitation Based on Gamification and Serious Games for Young People: A Systematic Mapping Study. Appl. Sci. 2020, 10, 8849. [Google Scholar] [CrossRef]
  194. Palaniappan, S.M.; Duerstock, B.S. Developing Rehabilitation Practices Using Virtual Reality Exergaming. In Proceedings of the 2018 IEEE International Symposium on Signal Processing and Information Technology (ISSPIT), Louisville, KY, USA, 6–8 December 2018; pp. 090–094. [Google Scholar]
  195. Nguyen, A.-V.; Ong, Y.-L.A.; Luo, C.X.; Thuraisingam, T.; Rubino, M.; Levin, M.F.; Kaizer, F.; Archambault, P.S. Virtual Reality Exergaming as Adjunctive Therapy in a Sub-Acute Stroke Rehabilitation Setting: Facilitators and Barriers. Disabil. Rehabil. Assist. Technol. 2019, 14, 317–324. [Google Scholar] [CrossRef]
  196. Chuang, W.-C.; Hwang, W.-J.; Tai, T.-M.; Huang, D.-R.; Jhang, Y.-J. Continuous Finger Gesture Recognition Based on Flex Sensors. Sensors 2019, 19, 3986. [Google Scholar] [CrossRef] [Green Version]
  197. Dong, W.; Yang, L.; Fortino, G. Stretchable Human Machine Interface Based on Smart Glove Embedded with PDMS-CB Strain Sensors. IEEE Sens. J. 2020, 20, 8073–8081. [Google Scholar] [CrossRef]
  198. Zhu, C.; Li, R.; Chen, X.; Chalmers, E.; Liu, X.; Wang, Y.; Xu, B.B.; Liu, X. Ultraelastic Yarns from Curcumin-Assisted ELD toward Wearable Human–Machine Interface Textiles. Adv. Sci. 2020, 7, 2002009. [Google Scholar] [CrossRef] [PubMed]
  199. Hang, C.-Z.; Zhao, X.-F.; Xi, S.-Y.; Shang, Y.-H.; Yuan, K.-P.; Yang, F.; Wang, Q.-G.; Wang, J.-C.; Zhang, D.W.; Lu, H.-L. Highly Stretchable and Self-Healing Strain Sensors for Motion Detection in Wireless Human-Machine Interface. Nano Energy 2020, 76, 105064. [Google Scholar] [CrossRef]
  200. Ueki, S.; Kawasaki, H.; Ito, S.; Nishimoto, Y.; Abe, M.; Aoki, T.; Ishigure, Y.; Ojika, T.; Mouri, T. Development of a Hand-Assist Robot With Multi-Degrees-of-Freedom for Rehabilitation Therapy. IEEE/ASME Trans. Mechatron. 2012, 17, 136–146. [Google Scholar] [CrossRef]
  201. Rahman, M.A.; Al-Jumaily, A. Design and Development of a Hand Exoskeleton for Rehabilitation Following Stroke. Procedia Eng. 2012, 41, 1028–1034. [Google Scholar] [CrossRef] [Green Version]
  202. Cortese, M.; Cempini, M.; de Almeida Ribeiro, P.R.; Soekadar, S.R.; Carrozza, M.C.; Vitiello, N. A Mechatronic System for Robot-Mediated Hand Telerehabilitation. IEEE/ASME Trans. Mechatron. 2015, 20, 1753–1764. [Google Scholar] [CrossRef]
  203. Han, H.; Yoon, S.W. Gyroscope-Based Continuous Human Hand Gesture Recognition for Multi-Modal Wearable Input Device for Human Machine Interaction. Sensors 2019, 19, 2562. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  204. Wei, L.; Hu, H. A Hybrid Human-Machine Interface for Hands-Free Control of an Intelligent Wheelchair. Int. J. Mechatron. Autom. 2011, 1, 97–111. [Google Scholar] [CrossRef] [Green Version]
  205. Huang, Y.; Yang, J.; Liu, S.; Pan, J. Combining Facial Expressions and Electroencephalography to Enhance Emotion Recognition. Future Internet 2019, 11, 105. [Google Scholar] [CrossRef] [Green Version]
  206. Downey, J.E.; Weiss, J.M.; Muelling, K.; Venkatraman, A.; Valois, J.-S.; Hebert, M.; Bagnell, J.A.; Schwartz, A.B.; Collinger, J.L. Blending of Brain-Machine Interface and Vision-Guided Autonomous Robotics Improves Neuroprosthetic Arm Performance during Grasping. J. Neuroeng. Rehabil. 2016, 13, 28. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  207. Bu, N.; Bandou, Y.; Fukuda, O.; Okumura, H.; Arai, K. A Semi-Automatic Control Method for Myoelectric Prosthetic Hand Based on Image Information of Objects. In Proceedings of the 2017 International Conference on Intelligent Informatics and Biomedical Sciences (ICIIBMS), Okinawa, Japan, 24–26 November 2017; pp. 23–28. [Google Scholar]
  208. Malechka, T.; Tetzel, T.; Krebs, U.; Feuser, D.; Graeser, A. SBCI-Headset—Wearable and Modular Device for Hybrid Brain-Computer Interface. Micromachines 2015, 6, 291–311. [Google Scholar] [CrossRef] [Green Version]
  209. McMullen, D.P.; Hotson, G.; Katyal, K.D.; Wester, B.A.; Fifer, M.S.; McGee, T.G.; Harris, A.; Johannes, M.S.; Vogelstein, R.J.; Ravitz, A.D.; et al. Demonstration of a Semi-Autonomous Hybrid Brain-Machine Interface Using Human Intracranial EEG, Eye Tracking, and Computer Vision to Control a Robotic Upper Limb Prosthetic. IEEE Trans. Neural Syst. Rehabil. Eng. 2014, 22, 784–796. [Google Scholar] [CrossRef] [Green Version]
  210. Frisoli, A.; Loconsole, C.; Leonardis, D.; Banno, F.; Barsotti, M.; Chisari, C.; Bergamasco, M. A New Gaze-BCI-Driven Control of an Upper Limb Exoskeleton for Rehabilitation in Real-World Tasks. IEEE Trans. Syst. Man Cybern. Part C (Appl. Rev.) 2012, 42, 1169–1179. [Google Scholar] [CrossRef]
  211. Dunai, L.; Novak, M.; García Espert, C. Human Hand Anatomy-Based Prosthetic Hand. Sensors 2021, 21, 137. [Google Scholar] [CrossRef]
  212. Krasoulis, A.; Kyranou, I.; Erden, M.S.; Nazarpour, K.; Vijayakumar, S. Improved Prosthetic Hand Control with Concurrent Use of Myoelectric and Inertial Measurements. J. Neuroeng. Rehabil. 2017, 14, 71. [Google Scholar] [CrossRef]
  213. Shahzad, W.; Ayaz, Y.; Khan, M.J.; Naseer, N.; Khan, M. Enhanced Performance for Multi-Forearm Movement Decoding Using Hybrid IMU–SEMG Interface. Front. Neurorobot. 2019, 13, 43. [Google Scholar] [CrossRef] [Green Version]
  214. Kyranou, I.; Krasoulis, A.; Erden, M.S.; Nazarpour, K.; Vijayakumar, S. Real-Time Classification of Multi-Modal Sensory Data for Prosthetic Hand Control. In Proceedings of the 2016 6th IEEE International Conference on Biomedical Robotics and Biomechatronics (BioRob), Singapore, 26–29 June 2016; pp. 536–541. [Google Scholar]
  215. Jaquier, N.; Connan, M.; Castellini, C.; Calinon, S. Combining Electromyography and Tactile Myography to Improve Hand and Wrist Activity Detection in Prostheses. Technologies 2017, 5, 64. [Google Scholar] [CrossRef] [Green Version]
  216. Guo, W.; Sheng, X.; Liu, H.; Zhu, X. Toward an Enhanced Human–Machine Interface for Upper-Limb Prosthesis Control with Combined EMG and NIRS Signals. IEEE Trans. Hum.-Mach. Syst. 2017, 47, 564–575. [Google Scholar] [CrossRef]
  217. Dwivedi, A.; Gerez, L.; Hasan, W.; Yang, C.-H.; Liarokapis, M. A Soft Exoglove Equipped with a Wearable Muscle-Machine Interface Based on Forcemyography and Electromyography. IEEE Robot. Autom. Lett. 2019, 4, 3240–3246. [Google Scholar] [CrossRef]
  218. Perez, E.; López, N.; Orosco, E.; Soria, C.; Mut, V.; Freire-Bastos, T. Robust Human Machine Interface Based on Head Movements Applied to Assistive Robotics. Sci. World J. 2013, 2013, e589636. [Google Scholar] [CrossRef] [Green Version]
  219. Bastos-Filho, T.F.; Cheein, F.A.; Müller, S.M.T.; Celeste, W.C.; de la Cruz, C.; Cavalieri, D.C.; Sarcinelli-Filho, M.; Amaral, P.F.S.; Perez, E.; Soria, C.M.; et al. Towards a New Modality-Independent Interface for a Robotic Wheelchair. IEEE Trans. Neural Syst. Rehabil. Eng. 2014, 22, 567–584. [Google Scholar] [CrossRef] [PubMed]
  220. Anwer, S.; Waris, A.; Sultan, H.; Butt, S.I.; Zafar, M.H.; Sarwar, M.; Niazi, I.K.; Shafique, M.; Pujari, A.N. Eye and Voice-Controlled Human Machine Interface System for Wheelchairs Using Image Gradient Approach. Sensors 2020, 20, 5510. [Google Scholar] [CrossRef] [PubMed]
  221. Gardner, M.; Mancero Castillo, C.S.; Wilson, S.; Farina, D.; Burdet, E.; Khoo, B.C.; Atashzar, S.F.; Vaidyanathan, R. A Multimodal Intention Detection Sensor Suite for Shared Autonomy of Upper-Limb Robotic Prostheses. Sensors 2020, 20, 6097. [Google Scholar] [CrossRef] [PubMed]
  222. Wu, C.-M.; Chen, Y.-J.; Chen, S.-C.; Yeng, C.-H. Wireless Home Assistive System for Severely Disabled People. Appl. Sci. 2020, 10, 5226. [Google Scholar] [CrossRef]
  223. Assistive Robotics and Rehabilitation. Available online: https://www.knowledge-share.eu/en/sector/assistive-robotics-and-rehabilitation/ (accessed on 26 August 2021).
  224. Andreoni, G.; Parini, S.; Maggi, L.; Piccini, L.; Panfili, G.; Torricelli, A. Human Machine Interface for Healthcare and Rehabilitation. In Advanced Computational Intelligence Paradigms in Healthcare-2; Vaidya, S., Jain, L.C., Yoshida, H., Eds.; Studies in Computational Intelligence; Springer: Berlin/Heidelberg, Germany, 2007; pp. 131–150. ISBN 978-3-540-72375-2. [Google Scholar]
Figure 1. Focus of the survey: biosignal-based HMIs with assistive and rehabilitation purposes.
Figure 1. Focus of the survey: biosignal-based HMIs with assistive and rehabilitation purposes.
Sensors 21 06863 g001
Figure 2. (A) Histogram illustrating the trend in the last two decades in the number of studies selected for the current survey. (B) Pie chart presenting the percentage distribution of the selected studies, according to the field of application.
Figure 2. (A) Histogram illustrating the trend in the last two decades in the number of studies selected for the current survey. (B) Pie chart presenting the percentage distribution of the selected studies, according to the field of application.
Sensors 21 06863 g002
Figure 3. (A) Pie chart showing the percentage distribution of the biosignals implemented by the HMIs considered in this survey. (BE) Pie charts representing the percentage distributions of the biosignals subcategories: (B) Biopotentials-based HMIs; (C) Muscle mechanical motion-based HMIs; (D) Body motion-based HMIs; (E) Hybrid HMIs.
Figure 3. (A) Pie chart showing the percentage distribution of the biosignals implemented by the HMIs considered in this survey. (BE) Pie charts representing the percentage distributions of the biosignals subcategories: (B) Biopotentials-based HMIs; (C) Muscle mechanical motion-based HMIs; (D) Body motion-based HMIs; (E) Hybrid HMIs.
Sensors 21 06863 g003
Figure 4. (A) Pie diagram illustrating the percentage distribution of the considered targets by the selected HMIs. (B) Graph illustrating the trend in the last two decades of each target (depicted in a different colour) of the considered HMIs with assistive and/or rehabilitative applications. Linear regressions (starting from 2010) are also superimposed as dashed lines.
Figure 4. (A) Pie diagram illustrating the percentage distribution of the considered targets by the selected HMIs. (B) Graph illustrating the trend in the last two decades of each target (depicted in a different colour) of the considered HMIs with assistive and/or rehabilitative applications. Linear regressions (starting from 2010) are also superimposed as dashed lines.
Sensors 21 06863 g004
Figure 5. Types of biosignal and main technologies/sensors/systems used by assistive and/or rehabilitative HMIs.
Figure 5. Types of biosignal and main technologies/sensors/systems used by assistive and/or rehabilitative HMIs.
Sensors 21 06863 g005
Figure 6. Pie charts illustrating the percentage distributions of the HMIs targets in the period 2015–2020: (A) Communication; (B) Gesture recognition; (C) Prosthetic control; (D) Robotic control; (E) Smart environment control; (F) Virtual reality control.
Figure 6. Pie charts illustrating the percentage distributions of the HMIs targets in the period 2015–2020: (A) Communication; (B) Gesture recognition; (C) Prosthetic control; (D) Robotic control; (E) Smart environment control; (F) Virtual reality control.
Sensors 21 06863 g006
Table 1. Review studies about HMIs for assistive and rehabilitation purposes (starting from 2015 to 2021).
Table 1. Review studies about HMIs for assistive and rehabilitation purposes (starting from 2015 to 2021).
Authors [Reference]TitleTopic
Taylor et al. [18]The use of gaming technology for rehabilitation in people with multiple sclerosisExergaming
De Gama et al. [29]Motor Rehabilitation Using Kinect: A Systematic ReviewExergaming
Laver et al. [30]Virtual reality for stroke rehabilitationExergaming
Wright et al. [23]A Review of Control Strategies in Closed-Loop Neuroprosthetic SystemsProsthetic control
Ciancio et al. [24]Control of Prosthetic Hands via the Peripheral Nervous SystemProsthetic control
Frisoli et al. [22]New generation emerging technologies for neurorehabilitation and motor assistanceWearable devices (exoskeletons)
Rosly et al. [31]Exergaming for individuals with neurological disability: A systematic reviewExergaming
Lazarou et al. [5]EEG-Based Brain–Computer Interfaces for Communication and Rehabilitation of People with Motor Impairment: A Novel Approach of the 21st CenturyBCI
Ngan et al. [25]Strategies for neural control of prosthetic limbs: From electrode interfacing to 3D printingProsthetic control
Parajuli et al. [26]Real-Time EMG Based Pattern Recognition Control for Hand Prostheses: A Review on Existing Methods, Challenges, and Future ImplementationProsthetic control
Igual et al. [27]Myoelectric Control for Upper Limb ProsthesesProsthetic control
Kumar et al. [28]Prosthetic hand control: A multidisciplinary review to identify strengths, shortcomings, and the futureProsthetic control
Reis et al. [32]Exergames for motor rehabilitation in older adults: An umbrella reviewExergaming
Garcia-Agundez et al. [20]Recent advances in rehabilitation for Parkinson’s Disease with exergames: A Systematic ReviewExergaming
Fatima et al. [19]Intracortical brain–machine interfaces for controlling
upper-limb-powered muscle and robotic systems in
spinal cord injury
Prosthetic control
Grushko et al. [10]Control Methods for Transradial Prostheses Based on Remnant Muscle Activity and Its Relationship with
Proprioceptive Feedback
Prosthetic control
Mohebbi et al. [21]Human–Robot Interaction in Rehabilitation and
Assistance: A Review
Robotic control
Ptito et al. [6]Brain–Machine Interfaces to Assist the BlindBCI
Li et al. [33]Gesture Recognition Using Surface Electromyography and Deep Learning for Prostheses Hand: State-of-the-Art, Challenges, and FutureProsthetic control
Ahmadizadeh et al. [9]Human Machine Interfaces in Upper-Limb Prosthesis Control: A Survey of Techniques for Preprocessing and Processing of BiosignalsProsthetic control
Baniqued et al. [7]Brain–computer interface robotics for hand rehabilitation after stroke: A systematic reviewBCI
Table 2. EEG-based HMIs.
Table 2. EEG-based HMIs.
Authors [Reference]Kind of BiopotentialTargetField
Gao et al. [46]Scalp EEGProsthetic ControlAssistance
Gannouni et al. [47]Scalp EEGProsthetic ControlAssistance, Rehabilitation
Fuentes-Gonzalez et al. [48]Scalp EEGProsthetic ControlAssistance
Song et al. [49]Scalp EEGRobotic ControlAssistance, Rehabilitation
Korovesis et al. [50]Scalp EEGRobotic ControlAssistance
Antoniou et al. [64]Scalp EEGRobotic ControlRehabilitation
Xu et al. [14]Scalp EEGRobotic ControlAssistance, Rehabilitation
Liang et al. [34]Scalp EEGRobotic ControlAssistance, Rehabilitation
Matsushita et al. [65]ECoGRobotic ControlAssistance
Spataro et al. [66]Scalp EEGRobotic ControlAssistance
López-Larraz et al. [67]Scalp EEGRobotic ControlRehabilitation
Xu et al. [36]Scalp EEGRobotic ControlRehabilitation
Kwak et al. [82]Scalp EEGRobotic ControlRehabilitation
Hortal et al. [68]Scalp EEGRobotic ControlAssistance, Rehabilitation
Varada et al. [15]Scalp EEGRobotic Control, Smart Environment ControlAssistance, Rehabilitation
Wang et al. [69]Scalp EEGRobotic Control, Prosthetic ControlRehabilitation
Zhan Hong et al. [70]Scalp EEGProsthetic ControlAssistance
Ortiz et al. [71]Scalp EEGRobotic ControlAssistance, Rehabilitation
Kasim et al. [72]Scalp EEGProsthetic ControlAssistance
Murphy et al. [73]Scalp EEGProsthetic ControlAssistance
Li et al. [74]sEEGProsthetic ControlAssistance
Bhagat et al. [75]Scalp EEGRobotic ControlRehabilitation
Morishita et al. [76]ECoGProsthetic ControlRehabilitation
Zhang et al. [77]Scalp EEGProsthetic ControlAssistance, Rehabilitation
Yanagisawa et al. [78]ECoGProsthetic ControlAssistance, Rehabilitation
He et al. [35]Scalp EEGRobotic ControlRehabilitation
Tang et al. [79]Scalp EEGRobotic ControlAssistance
Randazzo et al. [80]Scalp EEGRobotic ControlAssistance, Rehabilitation
Li et al. [81]Scalp EEGRobotic ControlAssistance, Rehabilitation
Araujo et al. [83]Scalp EEGRobotic ControlRehabilitation
Kashihara et al. [84]Scalp EEGCommunicationAssistance
Mahmoudi and Erfanian [85]Scalp EEGProsthetic ControlAssistance
Table 3. EMG-based HMIs.
Table 3. EMG-based HMIs.
Authors [Reference]Kind of BiopotentialTargetField
Eisenberg et al. [53]sEMGGesture Recognition, Prosthetic ControlAssistance
Tavakoli et al. [54]sEMGGesture Recognition, Prosthetic ControlAssistance
Bai et al. [87]sEMGGesture Recognition, Prosthetic ControlAssistance
Cao et al. [88]sEMGProsthetic ControlAssistance
Benatti et al. [89]sEMGGesture Recognition, Prosthetic ControlAssistance
Ulloa et al. [90]sEMGProsthetic ControlAssistance
Polisiero et al. [91]sEMGProsthetic ControlAssistance
Gailey et al. [92]sEMGGesture Recognition, Prosthetic ControlAssistance
Bernardino et al. [93]sEMGGesture Recognition, Prosthetic ControlAssistance
Zhao et al. [94]sEMGGesture Recognition, Prosthetic ControlAssistance
Carrozza et al. [95]sEMGProsthetic ControlAssistance
Jiang et al. [96]sEMGGesture Recognition, Prosthetic ControlAssistance
Brunelli et al. [97]sEMGGesture Recognition, Prosthetic ControlAssistance
Shair et al. [98]sEMGProsthetic ControlAssistance
Khushaba et al. [99]sEMGGesture Recognition, Prosthetic ControlAssistance
Kamavuako et al. [100]imEMGProsthetic ControlAssistance
Dewald et al. [101]imEMGGesture Recognition, Prosthetic Control, Virtual Reality ControlAssistance
Al-Timemy et al. [102]sEMGGesture Recognition, Prosthetic ControlAssistance
Zhang et al. [103]sEMGProsthetic ControlAssistance
Dalley et al. [104]sEMGProsthetic ControlAssistance
Russo et al. [105]sEMGGesture Recognition, Prosthetic ControlAssistance
Stepp et al. [106]sEMGProsthetic ControlRehabilitation
Visconti et al. [107]sEMGGesture Recognition, Prosthetic Control, Robotic Control, Smart Environment Control, Virtual
Reality Control
Assistance, Rehabilitation
Lu and Zhou [108]sEMGSmart Environment ControlAssistance
Kumar et al. [109]sEMGRobotic ControlAssistance
Kalani et al. [110]sEMGRobotic ControlRehabilitation
Alibhai et al. [39]sEMGGesture Recognition, Robotic ControlAssistance
Fall et al. [37]sEMGRobotic ControlAssistance
Song et al. [40]sEMGGesture Recognition, Robotic ControlAssistance
Laksono et al. [38]sEMGRobotic ControlAssistance
Xu et al. [41]sEMGRobotic ControlAssistance
Zhang et al. [111]sEMGRobotic ControlAssistance
Hamedi et al. [112]sEMGGesture RecognitionAssistance, Rehabilitation
Wege and Zimmermann [113]sEMGRobotic ControlRehabilitation
Ho et al. [114]sEMGRobotic ControlRehabilitation
Loconsole et al. [115]sEMGRobotic ControlRehabilitation
Hussain et al. [116]sEMGGesture Recognition, Robotic ControlAssistance
Abdallah et al. [117]sEMGRobotic ControlRehabilitation
Secciani et al. [118]sEMGRobotic ControlAssistance
Song et al. [119]sEMGRobotic ControlRehabilitation
Liu et al. [120]sEMGRobotic ControlRehabilitation
Cai et al. [121]sEMGRobotic ControlRehabilitation
Yin et al. [122]sEMGRobotic ControlRehabilitation
Tang et al. [123]sEMGRobotic ControlRehabilitation
Lu et al. [124]sEMGRobotic ControlRehabilitation
Gui et al. [125]sEMGRobotic ControlRehabilitation
La Scaleia et al. [126]sEMGRobotic Control, Virtual Reality ControlAssistance, Rehabilitation
Lyu et al. [127]sEMGRobotic ControlRehabilitation
Table 4. ENG-based HMIs.
Table 4. ENG-based HMIs.
Authors [Reference]TargetField
Noce et al. [52]Gesture Recognition, Prosthetic ControlAssistance
Nguyen et al. [55]Prosthetic ControlAssistance
Noce et al. [59]Gesture Recognition, Prosthetic ControlAssistance
Table 5. EOG-based HMIs.
Table 5. EOG-based HMIs.
Authors [Reference]TargetField
Golparvar and Yapici [56]Robotic Control, Smart Environment ControlAssistance
Zhang et al. [42]Smart Environment ControlAssistance
Huang et al. [57]Robotic ControlAssistance
Martínez-Cerveró et al. [128]CommunicationAssistance
Perez Reynoso et al. [129]Robotic ControlAssistance
Choudhari et al. [130]Robotic ControlAssistance
Heo et al. [131]Communication, Robotic ControlAssistance
Guo et al. [132]Smart Environment ControlAssistance
Wu et al. [133]Robotic Control, Smart Environment ControlAssistance,
Rehabilitation
Table 6. Hybrid biopotential-based HMIs.
Table 6. Hybrid biopotential-based HMIs.
Authors [Reference]Kind of BiopotentialTargetField
Gordleeva et al. [51]EEG + EMGRobotic ControlRehabilitation
Ferreira et al. [134]EEG + EMGRobotic ControlAssistance
Zhang et al. [135]EEG + EMG + EOGGesture Recognition, Robotic ControlAssistance, Rehabilitation
Huang et al. [136]EEG + EOGRobotic ControlAssistance
Ma et al. [12]EEG + EOGRobotic ControlAssistance
Ma et al. [137]EEG + EOGRobotic ControlAssistance
Arrow et al. [58]EMG + ERGProsthetic ControlAssistance
Rezazadeh et al. [138]EEG + EMGVirtual Reality ControlAssistance
Rezazadeh et al. [139]EEG + EMG + EOGCommunication, Gesture RecognitionAssistance, Rehabilitation
Iáñez et al. [140]EEG + EOGSmart Environment ControlAssistance
Laport et al. [141]EEG + EOGSmart Environment ControlAssistance
Neto et al. [142]EEG + EMG + EOGRobotic ControlAssistance
Table 7. Muscle gross motion-based HMIs.
Table 7. Muscle gross motion-based HMIs.
Authors [Reference]Kind of sensorApplication SiteTargetField
Prakash et al. [144]FSRForearmProsthetic ControlAssistance
Clemente et al. [165]Magnetic FieldForearmProsthetic ControlAssistance
Xiao et al. [152]FSRForearmRobotic ControlRehabilitation
Ferigo et al. [153]FSRForearmProsthetic ControlAssistance
Esposito et al. [154]FSRForearmGesture Recognition, Prosthetic ControlAssistance
Esposito et al. [155]FSRForearmProsthetic ControlAssistance
Esposito et al. [156]FSRForearmProsthetic ControlAssistance
Ha et al. [157]PiezoelectricForearmProsthetic ControlAssistance
Ha et al. [158]PiezoelectricForearmProsthetic ControlAssistance
Ahmadizadeh et al. [151]FSRForearmProsthetic ControlAssistance
Fujiwara et al. [159]Optical FibreForearmGesture Recognition, Prosthetic Control, Virtual Reality ControlAssistance,
Rehabilitation
Bifulco et al. [160]ResistiveForearmProsthetic ControlAssistance
Radmand et al. [161]FSRForearmProsthetic ControlAssistance
Cho et al. [162]FSRForearmProsthetic ControlAssistance
Dong et al. [163]TriboelectricHandRobotic Control, Virtual Reality ControlAssistance,
Rehabilitation
Zhu et al. [16]TriboelectricHandRobotic Control, Virtual Reality ControlAssistance,
Rehabilitation
An et al. [164]TriboelectricArmRobotic Control, Smart Environment ControlAssistance
Tarantino et al. [166]Magnetic fieldForearmProsthetic ControlAssistance
Kumar et al. [167]PiezoresistiveHandCommunication, Smart Environment ControlAssistance
Castellini et al. [168]ResistiveForearmProsthetic ControlAssistance
Dong et al. [169]PiezoelectricWristProsthetic ControlAssistance
Lim et al. [170]PiezoelectricForearm, WristRobotic ControlAssistance
Rasouli et al. [171]PiezoelectricForearmProsthetic ControlAssistance
Table 8. Muscle vibrations-based HMIs.
Table 8. Muscle vibrations-based HMIs.
Authors [Reference]Kind of SensorApplication SiteTargetField
Asheghabadi et al. [146]Piezoelectric + Strain GaugeForearmProsthetic ControlAssistance
Castillo et al. [176]MicrophoneForearmProsthetic ControlAssistance
Wicaksono et al. [177]PiezoresistiveLower limbProsthetic Control, Robotic ControlAssistance,
Rehabilitation
Xie et al. [178]AccelerometerForearmGesture Recognition, Prosthetic ControlAssistance
Table 9. Muscle–tendons movement-based HMIs.
Table 9. Muscle–tendons movement-based HMIs.
Authors [Reference]Kind of SensorApplication SiteTargetField
Wu et al. [145]BioamplifierForearmGesture Recognition, Prosthetic ControlAssistance
Chen et al. [147]US probeForearmProsthetic ControlAssistance
Huang et al. [179]US probeForearmGesture Recognition, Prosthetic
Control, Robotic Control
Assistance
Li et al. [180]US transducerForearmGesture Recognition, Robotic ControlRehabilitation
Ortenzi et al. [181]US probeForearmProsthetic ControlAssistance
Sikdar et al. [182]US probeForearmProsthetic ControlAssistance
Sierra González et al. [183]US probeForearmRobotic ControlRehabilitation
Castellini et al. [184]US probeForearmRobotic ControlRehabilitation
Shi et al. [185]US probeForearmProsthetic ControlAssistance
Table 10. Hybrid muscle mechanical motion-based HMIs.
Table 10. Hybrid muscle mechanical motion-based HMIs.
Authors [Reference]Kind of SensorApplication SiteTargetField
Esposito et al. [143]FSRForearmProsthetic ControlAssistance
Booth et al. [186]PiezoelectricWristGesture Recognition, Prosthetic Control,
Robotic Control, Smart Environment Control, Virtual Reality Control
Assistance,
Rehabilitation
Table 11. Image-based body motion HMIs.
Table 11. Image-based body motion HMIs.
Authors [Reference]Tracked Body PartTargetField
Maule et al. [187]EyesRobotic ControlAssistance
Bissoli et al. [43]EyesSmart Environment ControlAssistance
Lin et al. [188]EyesSmart Environment ControlAssistance
Conci et al. [189]HandsGesture Recognition, Smart Environment ControlAssistance
Baklouti et al. [190]Head/MouthRobotic controlRehabilitation
Chang et al. [191]HeadCommunicationAssistance
Gautam et al. [192]HeadRobotic ControlAssistance
Gmez-Portes et al. [193]Whole bodyVirtual Reality ControlRehabilitation
Palaniappan et al. [194]Upper limbVirtual Reality ControlRehabilitation
Nguyen et al. [195]Whole body with “JRS”; wrist and elbow with “MHT”Virtual Reality ControlRehabilitation
Table 12. Nonimage-based body motion HMIs.
Table 12. Nonimage-based body motion HMIs.
Authors [Reference]Kind of SensorsApplication Sites of SensorsTargetField
Chuang et al. [196]Resistive flex sensorsEmbedded in a gloveGesture RecognitionAssistance
Dong et al. [197]Piezoresistive strain sensors (based on PDMS-CB)Embedded in a gloveGesture Recognition,
Robotic Control
Assistance,
Rehabilitation
Zhu et al. [198]Stretchable conductive yarnsEmbedded in a gloveRobotic Control, Smart
Environment Control
Assistance
Hang et al. [203]PAAm hydrogel-based strain sensorVarious body positionsGesture Recognition,
Robotic Control
Assistance
Ueki et al. [199]Force/torque sensors and 3D motion sensorEmbedded in a glove, hand and forearmRobotic Control,
Virtual Reality Control
Rehabilitation
Rahman et al. [200]Flex sensorsEmbedded in a glove, handRobotic ControlRehabilitation
Cortese et al. [201]MEMS accelerometersEmbedded in a glove, handRobotic ControlRehabilitation
Han et al. [202]Three-axis gyroscopeHand backGesture Recognition, Smart Environment ControlAssistance
Table 13. Hybrid controls for HMIs based on Biopotentials and Image-based systems.
Table 13. Hybrid controls for HMIs based on Biopotentials and Image-based systems.
Authors [Reference]Kind of SensorsApplication Site of ElectrodesLocation of Video
System/s
TargetField
Wei and Hu [204]EMG electrodes + Video cameraForeheadtowards the subject’s faceRobotic ControlAssistance
Haung et al. [205]Video camera +
EEG electrodes
10–20 EEG international systemtowards the subject’s faceCommunicationAssistance
Downey et al. [206]Intracortical microelectrode arrays + RGB–D cameraMotor cortexon the arm of the robotRobotic ControlAssistance
Bu et al. [207]EMG electrodes + Video cameraForearmtowards the target
objects
Prosthetic
Control
Assistance
Malechka et al. [208]EEG electrodes +
3 video cameras
10–10 EEG international systemtwo video cameras towards subject’s face (one for each eye
tracking); one video camera towards the
target objects
Smart
Environment
Control
Assistance
McMullen et al. [209]ECoG and depth electrodes + Microsoft
Kinect + video camera
Motor cortexKinect sensor towards the target objects; video camera towards the
subject’s face
Prosthetic
Control
Assistance
Frisoli et al. [210]EEG electrodes + scene camera (i.e., 2 infrared cameras + 2 infrared LEDs + 1 wide-angle camera) + Microsoft KinectOver sensorimotor cortexScene camera mounted on glasses; Kinect sensor towards the target
objects
Robotic ControlRehabilitation
Table 14. Hybrid controls for HMIs based on Biopotentials and Mechanical Motion Detection.
Table 14. Hybrid controls for HMIs based on Biopotentials and Mechanical Motion Detection.
Authors [Reference]Kind of Hybrid
Sensors
Application Sites of Hybrid
Sensors
TargetField
Dunai et al. [217]sEMG electrodes + FSR sensorssEMG electrodes on Extensor digitorum (forearm).
FSR sensors on prosthetic fingertips.
Prosthetic ControlAssistance
Krasoulis et al. [211]Hybrid sEMG/IMU
sensors
Eight hybrid sensors are equally spaced around the forearm (3 cm
below the elbow); two are placed on the extrinsic hand muscles
superficialis; two are placed on the biceps and triceps brachii muscles.
Prosthetic ControlAssistance
Shahzad et al. [212]sEMG electrodes + IMUTwo sEMG sensors are placed on the forearm flexors, and other two are placed at the forearm extensors.
The forearm IMU was placed
proximal to the wrist, and the upper arm IMU was paced over the biceps brachii muscle.
Gesture Recognition, Prosthetic controlAssistance
Kyranou et al. [213]Hybrid sEMG/IMUTwelve hybrid sensors are placed on the proximal forearm via an elastic bandage.Gesture Recognition, Prosthetic controlAssistance
Jaquier et al. [214]sEMG electrodes + pressure sensors
(resistive elastomers)
Ten sEMG sensors are placed on the proximal forearm.
Ten pressure sensors (via a bracelet) are placed on the proximal forearm.
Gesture Recognition, Prosthetic controlAssistance
Guo et al. [215]Hybrid sEMG/NIRS
sensors
Four hybrid sensors are attached above flexor carpi ulnaris, flexor carpi radialis, extensor carpi radialis longus, and extensor digitorum.Gesture Recognition, Virtual Reality ControlAssistance
Xia et al. [13]Hybrid sEMG/US
sensors
Four hybrid sensors are mounted on the forearm by means of an
armband.
Gesture Recognition, Prosthetic ControlAssistance
Dwivedi et al. [216]sEMG + FSR
sensors
Three EMG sensors are embedded in a sleeve.
Five FSR sensors are embedded in a sleeve.
Robotic ControlAssistance
Table 15. Hybrid controls for HMIs based on various combinations of sensors.
Table 15. Hybrid controls for HMIs based on various combinations of sensors.
Authors
[Reference]
Kind of Hybrid SensorsApplication Sites of Hybrid SensorsTargetField
Ubeda et al. [11]EEG electrodes + RFID tags10–20 EEG international system;
RFID tags near by the target objects.
Robotic ControlAssistance, Rehabilitation
Perez et al. [218]Video camera + IMU sensorVideo camera (webcam) towards the
patient’s face;
IMU sensor mounted on a cap or headband.
Robotic ControlAssistance
Bastos-Filho et al. [219]EMG electrodes + video cameras + IMU sensor +
pressure sensor + EEG electrodes
EMG electrodes on temporal muscles;
Video camera on a pair of glasses worn by the user;
IMU sensor mounted on a cap;
Pressure sensor into a straw;
10–20 EEG international system;
Video camera towards the user’s face.
Robotic ControlAssistance
Anwer et al. [220]Microphone + video cameraMicrophone embedded in the wheelchair;
Video camera towards the user’s face.
Robotic ControlAssistance
Gardner et al. [221]Acoustic MMG
sensor + IMU sensor + video camera
MMG and IMU (embedded in a compression sleeve) on the biceps;
Video camera on a pair of glasses worn by the user.
Prosthetic ControlAssistance
Wu et al. [222]EOG electrodes + switches (push
button, InfraRed, mercury, long–short tone, and pacifier)
EOG electrodes on eyebrow arch;
Various switches are positioned to be
activated by the user.
CommunicationAssistance
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Esposito, D.; Centracchio, J.; Andreozzi, E.; Gargiulo, G.D.; Naik, G.R.; Bifulco, P. Biosignal-Based Human–Machine Interfaces for Assistance and Rehabilitation: A Survey. Sensors 2021, 21, 6863. https://doi.org/10.3390/s21206863

AMA Style

Esposito D, Centracchio J, Andreozzi E, Gargiulo GD, Naik GR, Bifulco P. Biosignal-Based Human–Machine Interfaces for Assistance and Rehabilitation: A Survey. Sensors. 2021; 21(20):6863. https://doi.org/10.3390/s21206863

Chicago/Turabian Style

Esposito, Daniele, Jessica Centracchio, Emilio Andreozzi, Gaetano D. Gargiulo, Ganesh R. Naik, and Paolo Bifulco. 2021. "Biosignal-Based Human–Machine Interfaces for Assistance and Rehabilitation: A Survey" Sensors 21, no. 20: 6863. https://doi.org/10.3390/s21206863

APA Style

Esposito, D., Centracchio, J., Andreozzi, E., Gargiulo, G. D., Naik, G. R., & Bifulco, P. (2021). Biosignal-Based Human–Machine Interfaces for Assistance and Rehabilitation: A Survey. Sensors, 21(20), 6863. https://doi.org/10.3390/s21206863

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop