Next Article in Journal
Risk Calculators in Bipolar Disorder: A Systematic Review
Previous Article in Journal
Analgesic Effects and Impairment in Locomotor Activity Induced by Cannabinoid/Opioid Combinations in Rat Models of Chronic Pain
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A 36-Class Bimodal ERP Brain-Computer Interface Using Location-Congruent Auditory-Tactile Stimuli

1
College of Intelligence Science and Technology, National University of Defense Technology, Changsha 410073, China
2
National Key Laboratory of Human Factors Engineering, China Astronaut Research and Training Center, Beijing 100094, China
*
Author to whom correspondence should be addressed.
Brain Sci. 2020, 10(8), 524; https://doi.org/10.3390/brainsci10080524
Submission received: 23 June 2020 / Revised: 3 August 2020 / Accepted: 4 August 2020 / Published: 6 August 2020

Abstract

:
To date, traditional visual-based event-related potential brain-computer interface (ERP-BCI) systems continue to dominate the mainstream BCI research. However, these conventional BCIs are unsuitable for the individuals who have partly or completely lost their vision. Considering the poor performance of gaze independent ERP-BCIs, it is necessary to study techniques to improve the performance of these BCI systems. In this paper, we developed a novel 36-class bimodal ERP-BCI system based on tactile and auditory stimuli, in which six-virtual-direction audio files produced via head related transfer functions (HRTF) were delivered through headphones and location-congruent electro-tactile stimuli were simultaneously delivered to the corresponding position using electrodes placed on the abdomen and waist. We selected the eight best channels, trained a Bayesian linear discriminant analysis (BLDA) classifier and acquired the optimal trial number for target selection in online process. The average online information transfer rate (ITR) of the bimodal ERP-BCI reached 11.66 bit/min, improvements of 35.11% and 36.69% compared to the auditory (8.63 bit/min) and tactile approaches (8.53 bit/min), respectively. The results demonstrate the performance of the bimodal system is superior to each unimodal system. These facts indicate that the proposed bimodal system has potential utility as a gaze-independent BCI in future real-world applications.

1. Introduction

Brain-computer interface (BCI) systems are novel human-computer interaction technology, which bypasses peripheral nerves and muscles and instead uses human brain activity to directly communicate with a computer or external devices [1,2,3]. BCI technology has developed rapidly since the 1970s [4,5,6]. In terms of signal acquisition methods, BCIs can be divided into two categories: invasive BCIs and non-invasive BCIs [7]. Although signals acquired using invasive BCIs end to be more accurate, scientific research is difficult to perform in general laboratory environments because of the need for brain surgery [8,9]. As such, non-invasive BCIs, most commonly using electroencephalography (EEG), have been widely applied in neural system and rehabilitation engineering [10,11]. In terms of EEG signal patterns, BCIs can be divided into spontaneous and evoked potential-based BCIs [12,13]. Specifically, event-related potential (ERP)-based BCIs are an important subset of evoked potential-based BCIs, in which, subjects are required to concentrate on infrequent target stimuli and ignore other non-target stimuli. Once the target event occurs, the EEG signals in the corresponding channels will change from a neutral ground state to an excited state [14]. Examples of ERPs include early components, such as the N200, and late components, such as the P300 and N400, etc. [15], the latency of which vary across individuals and ranges from 200 ms to 700 ms after target event onset [16].
Recently, researches on ERP-BCI have attracted extensive attention, and a number of practical application paradigms have been developed based on ERP-BCIs [17], which benefit from the following advantages: (1) target detection accuracy; (2) training time; (3) population adaptability; (4) information transfer rate (ITR); (5) paradigm flexibility, etc. [18,19]. Among the most of these researches, target events are usually based on visual stimuli, and require subjects to fix their eyesight on the stimulus interface at all times [20,21]. Nevertheless, for the individuals who have partly or completely lost their vision, traditional visual-based ERP-BCIs are ill-suited [22,23]. For instance, when a patient with paralysis manipulates an intelligent wheelchair with a visual-based ERP-BCI, he or she is required to fixate their attention onto a stimulus presentation screen causing them to ignore changes in their surroundings environment. To overcome these current limitations, auditory and tactile-based ERP-BCIs have emerged as a promising substitute for gaze-dependent systems [24]. Despite the usually lower ITR and classification accuracies, auditory- and tactile-based ERP-BCIs require few gaze shifts, and so visual feedback can be used to convey external scene instead of just being used as a simple stimulus source [25].
In respect to auditory-based ERP-BCIs, the first relevant paradigm was reported by Hill et al. in 2005, in which a binary target BCI was proposed by recognizing standard versus deviant voices [26]. Since 2010, Schreuder et al. realized a new auditory multi-class BCI paradigm and found that the performance of this auditory-based ERP-BCI improved by stimulating the subjects from different spatial orientations [27]. Guo et al. and Xu et al. further certified that active mental tasks increased target and non-target discrimination accuracy in an ERP-BCI paradigm, and thus improved the performance [28,29]. Building on these foundational researches, a number of studies have been expanded in the field in the last three years. Baykara et al. developed an auditory ERP-BCI system to spell words by using a 5 × 5 matrix of animal sounds and the corresponding ITR increased from 3.72 bit/min to 5.63 bit/min [30]. Moreover, Halder et al. optimized the aforementioned system and further improved the performance of the BCI when operated by subjects with motor impairments (i.e., the ITR to 5.78 bit/min, and accuracy to 92%) [31]. In addition, Miho et al. developed an ERP-BCI paradigm based on detection of six-direction virtual sound sources and found that the performance could be improved for both neurophysiological and behavioral responses, by means of shortening the stimulus onset asynchrony (SOA) [32].
More recently, tactile-based ERP-BCIs were initially explored by Brouwer et al. in 2010 [33]. In this paradigm, tactile stimuli were delivered via vibrating motors placed at different positions around the waist of the subject. Afterwards, moving towards the functional applications, Waal et al. developed a tactile ERP speller by classifying tactile stimuli delivered to different fingers and showed that a tactile speller provided a useful alternative to existing visual-based systems [34]. Subsequently, Kaufmann et al. designed another tactile ERP speller, in which the 26 letters in the English alphabet were divided into four groups and the tactile stimuli were delivered to four different areas of the subject’s left arm [35]. Furthermore, Kaufmann et al. also developed a wheelchair control system based on tactile ERP-BCI and investigated the use of a dynamic stopping method to improve the speed of the system [36]. Another wheelchair control system based on a tactile ERP-BCI was achieved by Herweg et al., in which elderly subjects participated in five sessions and tactors were placed on the legs, abdomen and back, and the system was validated on the elderly [37]. Most recently, Liu et al. proposed a tactile-based ERP-BCI paradigm for communication in 2018. In their study, the vibrators were positioned on the arm of the subjects to simulate a robotic arm. Subjects were able to successfully control a simulated mechanical arm with six degrees of freedom [38]. To date, vibration stimuli have been the most widely applied stimulus type in tactile-based ERP-BCI.
Different stimuli modalities are inherently useful as they can be used to evoke different potentials [39]. BCIs exist to leverage most of our five fundamental senses; however, the olfactory and gustatory perception channels are less developed from evolutionary standpoint, and it is difficult to reliably deliver stimuli. Researchers have already conducted studies on multisensory BCIs based on visual, auditory and tactile stimuli (though mainly visual-auditory and visual-tactile) [40,41]. Due to multisensory integration, the human brain synthesizes the neural information of each sensory channel independently, thus evoking a stronger ERP under multi-modal stimuli, leading to enhancement of the classification accuracy in ERP-BCIs [42,43,44,45]. For example, An et al. presented a visual-auditory ERPBCI speller for two different stimuli types (i.e., simultaneous redundant streams and interleaved independent streams), and explicitly explored the various combinations of these visual and auditory stimuli in a gaze independent BCI [46]. However, compared with each unimodal BCI, the performance of the bimodal BCI system did not reach a higher level. Recently, Sun et al. designed a novel improved visual-tactile bimodal stimulus ERP-BCI paradigm [47]. The average online ITR of the proposed pattern reached 12.49 bit/min, while the online ITR of the color-change pattern reached 8.87 bit/min on average. These results demonstrate that the picture-vibrate pattern achieved higher classification accuracy and online ITR than the color-change pattern. The limitations of the prior researches were mainly focused on three points: (1) the target for selection was too scarce; (2) using loudspeakers, the hardware load of the system increased significantly, which may give rise to mutual interference; (3) the time delay of the vibration tactile stimuli likely affects the instantaneous performance of the system. Consequently, our primary objective in this study is to overcome these three limitations and to further increase the ITR.
We proposed and developed a novel 36-class bimodal ERP-BCI system in this paper, in which the auditory and tactile stimuli were simultaneously generated from the same directions in a randomized order. Specifically, six-virtual-direction audio files were delivered from “1” to “6” through headphones, and the male voices were used for row selection as well as the female voices for column selection. In addition, electro-tactile stimuli were delivered by six pairs of electrodes placed around the abdomen and waist of the subjects corresponding with the angle of each virtual direction generated by the head related transfer functions (HRTF). Moreover, by offline data training, we selected the eight best channels, trained the Bayesian linear discriminant analysis (BLDA) classifier and obtained the optimal trial number and used these parameters in the online signal processing.

2. Materials and Methods

2.1. Subjects

There were 12 healthy subjects (age 18–25, mean age 20.83, denoted as S1–S12) in the experiments. All subjects had no history of neurological or psychological disorders. In addition, the subjects had normal hearing and normal touch sense in the waist and abdomen areas, and had no difficulties localizing sounds in space or history of trauma. None of them had participated in the experiments based on auditory or tactile BCI. Before the experiments, all the subjects were familiar with the purpose and tasks and signed the informed consent form. The study was approved by the ethical approval committee of Xiangya Hospital, Central South University, with the ethical approval code of “2018101045”, and the approval date was 10 October 2018.

2.2. The BCI Design

In our study, the BCI systems were composed of two portions: (1) user interface and (2) bimodal stimulus presentation, based on the MATLAB (MathWorks, Inc., Natick, MA, USA) platform. The user interface was used to present the characters on a 23.6-LCD monitor with a 1920 × 1080 resolution and a 60 Hz refresh rate. The distance between the monitor and the subject’s eyes was set to 70 cm. The character presentation was controlled by PTB-3 in MATLAB [48]. As shown in Figure 1a, a total of 36 (6 × 6) characters (26 letters, six digits, three punctuations and an operator) were presented on the screen. Each character was presented within a 40 × 40-pixel square. The vertical and horizontal distances between two neighboring characters were both 50 pixels. We employed the classical standard BCI speller with a 6 × 6 matrix, in which the target character was selected by the row/column (RC) mode. In the RC mode, the selection of each item in the matrix was grouped into row and column. The bimodal stimuli were presented in a random order to select the row and column indexes of the target character in succession. The intersection of the row and column determined by the ERP detection algorithms was identified as the target character. For the individuals who have partly or completely lost their vision, if their auditory and tactile functions are normal, they can also use the BCI system by memorizing the coordinates of each character, instead of by vision information. For example, when a user wishes to spell “J”, they will perform the following steps:
  • Recognize “J” is in the 3rd row. Focus on the stimulus “3” (at 135°) and count the number of randomly delivered repetitions of the desired stimulus;
  • The computer determines the user is focusing on the third row after the preset number of repetitions of stimuli is reached;
  • Recognize “J” is in the 4th column. Focus on the stimulus “4” (at 180°) and count the number of randomly delivered repetitions of the desired stimulus;
  • Lastly, computer determines the user was focusing on “4”. Consequently, the character “J” is presented on the screen if both the row and column indexes are correctly selected.
The bimodal stimulus presentation portion was composed of two major components: a stereo headset for playing the directional audio files, and an electrical stimulation controller to deliver the electro-tactile stimulus. The stimulus duration was set to 260 ms, and the inter-stimulus interval was set to 250 ms. A trial was defined as one complete stimulus cycle, in which each of the six stimuli were delivered once. At least two cycles were used for the selection of the row and column index. The direction of the stimulus was coded from “1” to “6”. The commands for controlling the onset of the stimuli were simultaneously sent from MATLAB to the electrical stimulation controller and stereo headset, which ensured the time synchronization of the auditory and tactile stimuli. A time break (1 s) was given after each selection, during which time the next target character turned red to prompt subjects to the next selection task. In addition, there was a short pause (0.75 s) when switching between the selection of the row and column index, so that subjects could pay attention to the column index in which the target character was located. We developed auditory, tactile and bimodal BCI to compare the unimodal system with the bimodal system. Additional details of the stimulation mechanism are shown in Figure 1b.

2.3. The Stimulus Design

2.3.1. Auditory Stimulus Design

The headphones delivered the auditory stimuli in six different directions (“1” at 0° relative to where the subjects faced, “2” at 45°, “3” at 135°, “4” at 180°, “5” at 225° and “6” at 315° in counterclockwise order) (see Figure 1b). To generate the sounds in six directions, we employed the HRTF [49] to turn two-channel audio files into the virtual six-direction audio files. HRTF is defined as:
{ H L = P L ( r , θ , φ , ω , α ) / P O ( r , ω ) H R = P R ( r , θ , φ , ω , α ) / P O ( r , ω )
where PL and PR are the complex sound pressure generated by the simple harmonic point sound source in the left and right ears of the listener, respectively. PO is the plural sound pressure at the center of the head. HL, HR is the function of the horizontal azimuth angle of the sound source θ, the elevation angle of the sound source φ, the distance from the sound source to the center of the head r, the angular frequency of the sound wave ω, and the head size α.
In our approach, the six ordinary audio files used for stimuli were “1” to “6” in the Chinese language and each was of equal duration. Then, the audio files were transformed to a virtual direction using the HRTF. After transformation, each stimulus had duration of 260 ms. The generated audio files were divided into male and female voice files for the selection of the row index and the column index, respectively. The male voice was used to select row index and the female voice was used to select column index, which made it convenience for the subject to distinguish between row selection and column selection. Consequently, a total of 12 audio files were generated. The volume of the headphone was set to approximately 80 db. Headphones were employed instead of loudspeakers to reduce the system setup time and the effects of disruptive surrounding noise. Since active mental task has been proved to enhance the performance of auditory-based BCI, during the experiments, the subjects were asked to mentally count the number of repetitions of the desired target [28].

2.3.2. Electro-Tactile Stimulus Design

The electro-tactile stimuli were delivered through a six-channel current controlled electrical stimulation generator (Neuracle Tech., Changzhou, China). Compared to mechanical approaches, such as vibration motors, electrical approaches have numerous advantages including easier adjustability of stimulus intensity and lower time delay. Six pairs of two standard self-adhesive disposable electrocardiogram (ECG) stimulation electrodes (Ch50rb, HealForce Bio-Meditech, Shanghai, China) were around the subject’s abdomen and waist with the same angle of each virtual direction of auditory stimulus. The electro-tactile stimuli were generated by the local current in each pair of electrodes on the skin surface. The electrodes of a given pair were 5 cm apart and interfaced to the computer. The instructions were delivered through a MATLAB program.
The generators provided electro-tactile stimuli by means of discrete electrical impulses. The current intensity ranged from 0.1 to 2 mA, the impulse width ranged from 0.1 to 6553.5 ms, and the time interval ranges from 10 to 65,535 ms. These types of generators are safe, portable and robust. Before placing the electrodes at corresponding locations on the subject’s waist, the candidate electrode locations were wiped with exfoliating cream twice to reduce the skin impedance so that subjects could more easily to perceive the electro-tactile stimuli. For the safety and comfort of the subjects, we set the current intensity (<2 mA) in each direction separately. Moreover, the impulse width was set to 260 ms (the same duration of each audio file), while the time interval was fixed to 250 ms. The mechanic of electro-tactile stimuli is shown in Figure 2.

2.3.3. Bimodal Stimulus Design

In the bimodal stimulus paradigm, ERPs were evoked via location-congruent auditory and electro-tactile stimuli simultaneously. As shown in Figure 1b, the bimodal stimuli were delivered from six directions. Furthermore, we delivered the virtual direction auditory and electro-tactile stimuli at corresponding positions. This is a location-congruent paradigm, and the coded number “1” to “6” represents the directional information of the bimodal stimulus system. When prompted in a certain direction, the subjects need to focus on the auditory and tactile stimuli in that direction, while ignoring the stimuli from the other directions.

2.4. Experiment Procedure

In our approach, the experiments were conducted in a quiet laboratory. The subjects were instructed to sit on the chair and face the LCD monitor place in front of them. The subjects were asked kept their eyes focused on the red prompted character so that blinking and eye movements were reduced. The subjects performed an exercise to familiarize themselves with the experimental tasks before each session. Each subject performed auditory, tactile and bimodal paradigm in three different sessions. We set the sessions in a random order to avert the sequence effect.
In each session, the subjects conducted 15 runs. As shown in Figure 3, each run was comprised of three character selections. Each character selection consisted of 10 row and column repetitions for each stimulus in a randomized order. In each repetition, the subjects were asked to pay attention to one number at a time for each row/column selection. An example sequence might be “4-2-5-1-3-6”. To avert prejudgment, the order of these coded numbers was randomized. The ERP-BCI system detected the row and column numbers focused by subjects. The intersection of the row and column was identified as the target character. The first ten runs were in the offline phase, which was used for channel selection, the ERP classifier training and the optimal trial selection; in the online phase, we applied the next five runs. Based on the offline and online phase, we compared the performance between bimodal and unimodal BCI. Between two adjacent runs, we arranged a five-minute interval for each subject to prevent fatigue.

3. Signal Processing

Offline training procedure:
  • Signal acquisition and preprocessing;
  • Channel selection;
  • BLDA training;
  • Trial number optimization.
Online selection procedure:
  • Signal acquisition and preprocessing;
  • ERP feature extraction;
  • Target selection.
More details on the signal processing are as below.

3.1. Signal Acquisition and Preprocessing

The raw EEG signals were acquired utilizing a standard EEG cap (Neuracle Tech., Changzhou, China) with 32 active Ag/AgCl electrodes following the international 10−20 system. The signals were sampled at a sampling rate of 250 Hz and amplified with a wireless amplifier (Neuracle, Changzhou, China). The configuration of electrode locations was in Figure 4. The impedance of each electrode was kept below 10 kΩ.
Both offline and online, signal preprocessing was the required procedure. Initially, the raw EEG data were treated with a 50 Hz notch filter and then with a 0.05−45 Hz band-pass filter. Then, the 1000 ms segment after each stimulus onset was extracted for the ERP feature analysis. In addition, the 400 ms pre-stimulus portion was subtracted from the post-stimulus segment to estimate a mean baseline amplitude. Finally, the data segments were down sampled from 250 to 25 Hz to acquire the major ERP information by an anti-aliasing moving average filter.

3.2. Channel Selection

In practical applications, the number of EEG channels used in the classification should be as small as possible to reduce the computational complexity and hardware requirements. However, it is inherently difficult to select the smallest channel set while maximizing detection accuracy because EEG feature spaces vary across individuals. The bilateral temporal lobes dominate the auditory function, and the posterior central gyrus dominates the tactile function [50,51]. Therefore, it is preferable to perform channel selection on an individual basis using detection accuracy to optimize the channel subset. Moreover, an eight-channel subset has produced tantamount accuracies to the traditional 32-channel complete set [52].
As shown in Figure 5, we adopted the Jumpwise regression algorithm to obtain optimal eight-channel subset [53]. Initially, the selected channel set (SC) and the unselected channel set (UC) are defined as an empty set and a complete set with 30 elements (30 in 32 channels, except REF and GND), respectively. Then, the channel set SC is checked to determine whether it contains eight elements; if so, the results are output; otherwise, channels are added or removed to SC until the target channel number (i.e., eight) is attained. Additional details are shown in Figure 5.

3.3. BLDA Classifier

In our approach, we applied Bayesian linear discriminant analysis (BLDA) algorithm [54,55,56] to train the feasible classifier. BLDA algorithm is derived as follows:
Initially, we assume that the linear regression target y is related to the feature vector x, and we define the linear discriminant function as:
y = w T x + n
where w is the weight vector and n denotes Gaussian noise. The likelihood function of w can be defined as:
p ( D | β , w ) = ( β 2 π ) N 2 e x p ( β 2 || X T w y || )
where D represents the dimension of the feature space, β is the inverse of the noise variance, N is sample size and X denotes the horizontal superposition matrix of the feature vectors. In Bayesian method, the prior probability distribution of w can be defined as:
p ( w | α ) = ( α 2 π ) D 2 ( ε 2 π ) 1 2 e x p ( 1 2 w T I ( α ) w )
where I’(α) is (D + 1) dimensional diagonal matrix defined as:
I ( α ) = [ α 0 0 0 α 0 0 0 ε ]
Therefore, the mean value m and the variance C of the posterior probability distribution of w can be respectively defined as:
{ m = β ( β X X T + I ( a ) ) 1 X y C = ( β X X T + I ( a ) ) 1
In conclusion, we obtain the estimated value of the mean value μ and variance σ2 of the linear regression target y:
{ μ = m T x ^ σ 2 = 1 β + x ^ T C x ^

3.4. ERP Detection

In the online signal processing procedure, the BLDA classifier was used to calculate the scores of ERP responses based on the following equation:
{ s c o r e r k = W T X r k s c o r e c k = W T X c k
where r , c { 1 , 2 , 3 , 4 , 5 , 6 } are the indexes of row and column stimulus codes, respectively, k { 1 , 2 , 3 , , K } is the trial number, K is the total trial number for each row/column index selection and W is the weight column vector of the BLDA classifier with 200 elements (8 channels by 25 time points). Before calculating the ERP scores, each preprocessed data segment was reshaped into a column vector X with the same elements as W. In addition, the scores of each row/column code were obtained by averaging all the scores with the same code number.
{ S c o r e r = 1 K k = 1 K s c o r e r k S c o r e c = 1 K k = 1 K s c o r e c k
where K is the total trial number for the current selection of row/column index. Then, the row and column indexes were determined by the maximum scores of the ERP response, which are denoted as:
{ I n d e x r = a r g m a x r { 1 , 2 , 3 , 4 , 5 , 6 } ( S c o r e r ) I n d e x c = a r g m a x c { 1 , 2 , 3 , 4 , 5 , 6 } ( S c o r e c )
Eventually, the target character was selected based on the intersection of the row and the column index in the given character matrix (see Figure 1a):
T A R G E T = ( I n d e x r ,   I n d e x c )

3.5. Selection of Optimal Trials

ITR is the most common standard to evaluate the system performance in BCI literature [1]. It is defined as:
ITR = { l o g 2 N + A l o g 2 A + ( 1 A ) l o g 2 1 A N 1 } / T
where N is the total number of all characters (i.e., 36), A represents the selection accuracy rate and T denotes the time interval (including 1 s for the break) per selection. In this study, the optimal trial number was selected by maximizing the ITR of each subject. To avoid overestimation, we adopted leave-one-block-out cross validation (LOOCV) to obtain the ITR for each subject. Specifically, we utilized nine runs for training and the last run for testing to calculate the ITR. To reduce fatigue and avoid extensive waiting periods, the maximum number of trials was set to ten. Then, the optimal number of trials per subject was determined as the number of repetitions which the ITR was at a local maximum. Finally, the optimal number of trials was used in the online signal processing.

3.6. Statistical Analysis

In this study, SPSS software (IBM SPSS statistics, IBM Corporation, Armonk, NY, USA) was used for statistical analysis. One-way repeated-measures analysis of variance (ANOVA) was applied to test the difference for classification accuracy, the number of trials and ITR between different methods (bimodal, auditory and tactile BCI), respectively. If the data did not conform to the sphericity assumption obtained by Mauchly’s test of sphericity, the Greenhouse–Geisser correction was performed. All post hoc pairwise comparisons were Bonferroni corrected. The alpha level was set at 0.05.

4. Results

4.1. Online Performance

In Table 1, the online performances of the ERP-BCIs using auditory, tactile and bimodal stimuli are presented. The experimental results demonstrated that the online performance of the bimodal approach was significantly better than that of each unimodal (i.e., auditory and tactile). One-way repeated measures ANOVA showed there was significant difference of ITR (F(2,33) = 5.2, p < 0.05) between three methods (bimodal, auditory and tactile). However, one-way repeated measures ANOVA showed there was no significant difference for the classification accuracy and the number of trials between three methods. Specifically, compared with the auditory (8.63 bit/min) and tactile approach (8.53 bit/min), the average ITR of the bimodal BCI reached 11.66 bit/min, an improvement of 35.11% and 36.69%, respectively. Moreover, the bimodal BCI resulted in a higher mean classification accuracy (72.78%) than that of each unimodal approach (auditory: 63.89%; tactile: 70.00%) with a fewer average number of trials (bimodal: 2.33; auditory: 2.58; tactile: 3.00). For S1 and S2, the ITR of the tactile BCI was superior to those produced by the auditory and bimodal BCIs, which suggest that tactile stimulus is preferable for these subjects.

4.2. Offline Analyses

We performed the offline training performance comparisons among the auditory, tactile and bimodal in terms of ITR and classification accuracy. We applied a standard leave-one-block-out cross validation (LOOCV) to evaluate the offline performance. Specifically, there were 10 blocks (here, a block is a run: see Section 2.4) for each subject in the offline training data, and 10 estimations of the classifier performance were conducted. In each calculation, nine blocks were used for training, and the remaining one was tested. Therefore, the leave one block out cross validation here was equivalent to 10-fold cross validation. As shown in Figure 6, both the average ITR and the classification accuracy of the bimodal BCI were consistently higher than those of each unimodal BCI across all repetitions. One-way repeated measures ANOVA showed there was a significant difference for ITR (p < 0.05) and the classification accuracy (p < 0.05) in each repetition of trial between three methods (bimodal, auditory and tactile). In addition, with increasing trial repetition number (i.e., an increase in T), the classification accuracy increased rapidly before converging to an asymptote. Hence, we are able to select the optimal number of trials for each subject by selecting the point at which the ITR exhibits a local maximum by assessing the offline performance of each subject. These results indicate that the overall system performance of the bimodal BCI is superior to each unimodal BCI.

5. Discussions

5.1. Classification Accuracy of Each Direction

The classification accuracy comparisons among the six directions of bimodal, auditory and tactile approaches are illustrated in Figure 7. To further discuss whether the classification performance differed across the six directions, we obtained the mean three-trial classification accuracy of each direction for all the subjects. A ten-fold cross-validation procedure was applied to evaluate the accuracy using all runs in each session. We selected the three-trial accuracy as a benchmark measure since all of the modalities achieved minimum accuracies required for effective communication (i.e., each was higher than 60%) when using three trials at least (See Figure 6b).
As shown in Figure 7, the classification accuracies of the bimodal BCI were higher than the corresponding accuracies of each unimodal BCI for all directions, which further validates the effectiveness and utility of the proposed bimodal BCI. In addition, although the same stimulus intensity (i.e., the playing volume and the current intensity) was delivered to the subjects in each direction, the mathematical statistical results indicate that the average accuracy of the right-posterior area (“4”, “5” and “6”) was significantly higher than that of the left-anterior area (“1”, “2” and “3”) for each modality (p < 0.05 for auditory; p < 0.05 for tactile; p < 0.01 for the bimodal) by using an one-way repeated measures ANOVA. We attempt to further explore the reasons for the superior performance observed using the bimodal BCI system as follows:
  • In total, 80% of the information received by human beings is from the visual system and the rest is collectively referred to as non-visual information. Because our eyes are in the anterior area, we cannot focus on the posterior area in a visual way. As previous research indicates, auditory and tactile stimuli play a quite important role in enhancing human spatial perception of the posterior area, whereas auditory and tactile are not crucial for target identification within the range of human vision [57,58]. Hence, the posterior area was superior to the anterior area for non-visual ERP-BCIs, especially auditory and tactile ERP-BCIs.
  • Tommasi and Marzoli demonstrated that the human brain processes sounds heard from our two ears differently. Signals received by the right ear are processed first and generally commands received by our right ear are easier to execute. This is defined as right-ear advantage [59] or multisensory right-side-of-space advantage [60]. To be specific, there might be a right-ear advantage in dichotic listening compared to the left in most individuals, which leads to the phenomenon that the left hemisphere of the brain is used for language processing and may be more sensitive to right-sided stimuli [61]. Consequently, the detection accuracy of right-sided ERP targets was superior to targets presentation on the left side.

5.2. Results of Channel Selection

We illustrated the eight optimal channel sets for each subject in the form of scatter diagrams and brain topographic maps. As shown in Figure 8a–c, the selected channels did vary from individual to individual across all the three stimulus modalities, thus validating the need to implement channel selection on a per subject basis.
In particular, the four most selected channels for the auditory, tactile and bimodal stimulus approaches were (Cz, Fp2, Fz and T3), (Cz, Fp1, Fp2 and F8) and (Fp2, Fz, F3 and T3), respectively. Fp2, Fz, Cz and T3 were the four most frequently selected channels among the three stimulus modalities. In Figure 8d–f, we found that Fp2 (located at right frontal pole) remained crucial for most subjects regardless of modality. Then, this finding corroborates the research on spatiotemporal analysis of ERPs by Cao et al., suggesting that the preparation effects appear in the right frontal pole first followed by the left pole [62]. We also noted that the Fz position (frontal midline) was often selected in the auditory and bimodal stimulus conditions. This selection preference to Fz could be explained through prior research, in which Gill et al. and Choi et al. determined the signals in the medial frontal lobe area can be applied to detect the surprisal response [63,64]. In addition, the channel Cz was selected for most of the subjects in the auditory and tactile ERP-BCIs (especially in the tactile case where it was chosen in 10 of 12 subjects). This observation is remarkably similar to previous findings that the tactile function area is primarily located in central area of the brain [34]. Another fact was that the channel T3 (left hemisphere) was selected more frequent than T4 (right hemisphere) for both the auditory and bimodal approaches. This finding is consistent with other previous studies [65,66,67,68], suggesting that the auditory functional areas are located in left temporal lobe areas in 95% of humans. Furthermore, the channels selected in the bimodal stimulus approach tended to consist of a union of channel used in the auditory and the tactile approaches. Hence, the performance improvement observed in the bimodal ERP-based BCI may be attributable to simultaneous activation of both the sensorimotor and the auditory cortices. In conclusion, the procedure of channel selection effectively reduced the computational complexity and the classification performance of the BCI was improved by means of combining the auditory and tactile stimuli.

5.3. Limitations and Optimized Orientations

Our major objective is to establish approaches to further improve the performance of an auditory-tactile bimodal ERP-BCI system. The limitations of current study and the corresponding optimized orientations are discussed as below:
  • The current optimal number of trials is estimated using offline calibration data and is fixed prior to the online use. During online signal processing, the target is determined once the optimal number of trials is met. However, the brain state as well as individual electrode conditions change over time. The optimal number of trials determined offline may not be optimal for the subsequent online experiment. Therefore, the number of trials can be further optimized by applying dynamic stopping strategy, which adaptively determines a selection time in each single character selection [69].
  • The performance of the bimodal BCI system proposed in this paper is superior to the auditory or tactile BCI systems alone. However, we still do not know the underlying basis for this superior performance. As previous researches indicated, Xu et al. optimized the performance of visual-based BCI systems by combing the P300 and SSVEP features [70]. Therefore, in our future work, we will attempt to extract more potentials, such as steady state somatosensory evoked potential (SSSEP) and steady state auditory evoked potential (SSAEP), to analyze the dynamic information of human brain neural activity under the condition of delivering the auditory-tactile bimodal stimuli.
  • Recently, BCI technology has been widely used in neural system and rehabilitation engineering especially as a potential communication solution for persons with severe motor impairments. Traditional visual-based BCI systems are dependent on the user’s vision and it is thus difficult to apply these solutions to the patients with visual impairments. The proposed BCI approach provides an alternative way to establish a visual-saccade-independent online brain-computer cooperative control system based on multisensory information. While the proposed approach is currently still in the laboratory phase, it is hoped that the system can be easily adapted to real-world applications.

6. Conclusions

A novel 36-class auditory-tactile bimodal ERP-BCI system was proposed in this paper, in which six-virtual-direction audio files were delivered through the headphones and six-position electro-tactile stimuli were delivered from the corresponding directions simultaneously. Moreover, we used Jumpwise regression to select the optimum subset of eight channels, and trained a BLDA classifier for automated target detection. The optimal number of trials for each subject was also tailored and used in the online process. The mean online ITR of the bimodal stimulus BCI reached 11.66 bit/min, increases of 35.11% and 36.69% compared with the auditory (8.63 bit/min) and tactile (8.53 bit/min) BCIs, respectively. The results demonstrate the performance of bimodal system was superior to each unimodal system (i.e., auditory and tactile). As far as we know, the bimodal stimulus BCI has achieved the highest number of possible characters (i.e., 36-class) and the highest reported ITR (11.66 bit/min) for a non-visual ERP-BCI to date. We believe that the proposed paradigm should be feasible for the patients with severe motor impairments who are currently ill-served by existing gaze-dependent BCI systems.

Author Contributions

Conceptualization, B.Z. and Z.Z.; methodology, B.Z.; software, B.Z. and J.J.; writing—original draft, B.Z.; writing—review and editing, B.Z. All authors have read and agreed to the published version of the manuscript.

Funding

National Key Research and Development Program: 2017YFB1002505.

Acknowledgments

This work was supported in part by the National Key Research and Development Program under Grant 2017YFB1002505.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wolpaw, J.R.; Birbaumer, N.; Heetderks, W.; Mcfarland, D.; Peckham, P.; Schalk, G.; Donchin, E.; Quatrano, L.; Robinson, C.; Vaughan, T. Brain–computer interface technology: A review of the first international meeting. IEEE Trans. Rehabil. Eng. 2000, 8, 164–173. [Google Scholar] [CrossRef] [PubMed]
  2. Allison, B.Z.; Wolpaw, E.W.; Wolpaw, J.R. Brain–computer interface systems: Progress and prospects. Expert Rev. Med. Devices 2007, 4, 463–474. [Google Scholar] [CrossRef] [PubMed]
  3. Xu, M.; Xiao, X.; Wang, Y.; Qi, H.; Jung, T.P.; Ming, D. A brain computer interface based on miniature event-related potentials induced by very small lateral visual stimuli. IEEE Trans. Biomed. Eng. 2018, 65, 1166–1175. [Google Scholar]
  4. Coyle, S.; Ward, T.; Markham, C. Brain computer interfaces, a review. Int. Sci. Rev. 2003, 28, 112–118. [Google Scholar] [CrossRef] [Green Version]
  5. Lance, B.J.; Kerick, S.E.; Ries, A.J.; Oie, K.S.; Mcdowell, K. Brain-computer interface technologies in the coming decades. Proc. IEEE 2012, 100, 1585–1599. [Google Scholar] [CrossRef]
  6. Burns, A.; Adeli, H.; Buford, J.A. Brain-computer interface after nervous system injury. Neuroscientist 2014, 20, 639–651. [Google Scholar] [CrossRef]
  7. Ortiz-Rosario, A.; Adeli, H. Brain-computer interface technologies: From signal to action. Rev. Neurosci. 2013, 24, 537–552. [Google Scholar] [CrossRef]
  8. Ortiz-Rosario, A.; Berrios-Torres, I.; Adeli, H.; Buford, J.A. Combined corticospinal and reticulospinal effects on upper limb muscles. Neurosci. Lett. 2014, 561, 30–34. [Google Scholar] [CrossRef] [Green Version]
  9. Spueler, M.; Walter, A.; Ramos-Murguialday, A.; Naros, G.; Birbaumer, N.; Gharabaghi, A. Decoding of motor intentions from epidural ECoG recordings in severely paralyzed chronic stroke patients. J. Neural Eng. 2014, 11, 066008. [Google Scholar] [CrossRef] [Green Version]
  10. Moghimi, S.; Kushki, A.; Guerguerian, A.M.; Chau, T. A review of EEG-based brain–computer interfaces as access pathways for individuals with severe disabilities. Assist. Technol. 2013, 25, 99–110. [Google Scholar] [CrossRef]
  11. Chen, X.; Xu, X.; Liu, A.; Mckeown, M.J.; Wang, Z.J. The use of multivariate EMD and CCA for denoising muscle artifacts from few-channel EEG recordings. IEEE Trans. Instrum. Meas. 2018, 67, 359–370. [Google Scholar] [CrossRef]
  12. Jiang, J.; Zhou, Z.T.; Yin, E.W.; Yu, Y.; Liu, Y.D.; Hu, D.W. A novel Morse code-inspired method for multiclass motor imagery brain–computer interface (BCI) design. Comput. Biol. Med. 2015, 66, 11–19. [Google Scholar] [CrossRef]
  13. Feng, J.; Yin, E.; Jin, J.; Saab, R.; Daly, I.; Wang, X. Towards correlation-based time window selection method for motor imagery BCIs. Neural Netw. 2018, 102, 87–95. [Google Scholar] [CrossRef] [Green Version]
  14. Farwell, L.A. Talking off the top of your head: Toward a mental prosthesis utilizing event-related brain potentials. Electroencephalogr. Clin. Neurophysiol. 1988, 70, 510–523. [Google Scholar] [CrossRef]
  15. Jin, J.; Sellers, E.W.; Zhou, S.; Zhang, Y.; Wang, X.; Cichocki, A. A P300 brain computer interface based on a modification of the mismatch negative paradigm. Int. J. Neural Syst. 2015, 25, 150011. [Google Scholar] [CrossRef] [PubMed]
  16. Sutton, S.; Braren, M.; Zubin, J.; John, E.R. Evoked-potential correlates of stimulus uncertainty. Science 1965, 150, 1187–1188. [Google Scholar] [CrossRef] [PubMed]
  17. Jin, J.; Zhang, H.; Daly, I.; Wang, X.; Cichocki, A. An improved P300 pattern in BCI to catch user’s attention. J. Neural Eng. 2017, 14, 036001. [Google Scholar] [CrossRef] [PubMed]
  18. Mak, J.N.; Arbel, Y.; Minett, J.W.; Mccane, L.M.; Yuksel, B.; Ryan, D. Optimizing the p300-based brain–computer interface: Current status and limitations and future directions. J. Neural Eng. 2011, 8, 025003. [Google Scholar] [CrossRef]
  19. Fazel-Rezai, R.; Allison, B.Z.; Guger, C.; Sellers, E.W.; Kleih, S.C.; Kübler, A. P300 brain computer interface: Current challenges and emerging trends. Front. Neuroeng. 2012, 5, 00014. [Google Scholar] [CrossRef] [Green Version]
  20. Riccio, A.; Mattia, D.; Simione, L.; Olivetti, M.; Cincotti, F. Eye-gaze independent EEG-based brain-computer interfaces for communication. J. Neural Eng. 2012, 9, 045001. [Google Scholar] [CrossRef]
  21. Acqualagna, L.; Blankertz, B. Gaze-independent BCI-spelling using rapid serial visual presentation (RSVP). Clin. Neurophysiol. 2013, 124, 901–908. [Google Scholar] [CrossRef] [PubMed]
  22. Liu, Y.; Zhou, Z.; Hu, D. Gaze independent brain-computer speller with covert visual search tasks. Clin. Neurophysiol. 2011, 122, 1127–1136. [Google Scholar] [CrossRef]
  23. Treder, M.S.; Schmidt, N.M.; Blankertz, B. Gaze-independent brain-computer interfaces based on covert attention and feature attention. J. Neural Eng. 2011, 8, 066003. [Google Scholar] [CrossRef]
  24. Barbosa, S.; Pires, G.; Nunes, U. Toward a reliable gaze-independent hybrid BCI combining visual and natural auditory stimuli. J. Neurosci. Methods 2011, 261, 47–61. [Google Scholar] [CrossRef]
  25. Xie, Q.; Pan, J.; Chen, Y.; He, Y.; Ni, X.; Zhang, J. A gaze-independent audiovisual brain-computer Interface for detecting awareness of patients with disorders of consciousness. BMC Neurol. 2018, 18, 144. [Google Scholar] [CrossRef]
  26. Hill, N.J.; Lal, T.N.; Bierig, K.; Birbaumer, N.; Schlkopf, B. An Auditory Paradigm for Brain–computer Interfaces. In Advances in Neural Information Processing Systems; MIT: Cambridge, MA, USA, 2005; pp. 569–576. [Google Scholar]
  27. Schreuder, M.; Blankertz, B.; Tangermann, M. A new auditory multi-class brain–computer interface paradigm: Spatial hearing as an informative cue. PLoS ONE 2010, 5, E9813. [Google Scholar] [CrossRef]
  28. Guo, J.; Gao, S.; Hong, B. An auditory brain– computer interface using active mental response. IEEE Trans. Neural Syst. Rehabil. Eng. 2010, 18, 230–235. [Google Scholar]
  29. Xu, H.; Zhang, D.; Ouyang, M.; Hong, B. Employing an active mental task to enhance the performance of auditory attention-based brain–computer interfaces. Clin. Neurophysiol. 2013, 124, 83–90. [Google Scholar] [CrossRef]
  30. Baykara, E.; Ruf, C.A.; Fioravanti, C.; Kthner, I.; Halder, S. Effects of training and motivation on auditory P300 brain-computer interface performance. Clin. Neurophysiol. 2016, 127, 379–387. [Google Scholar] [CrossRef]
  31. Halder, S.; KaThner, I.; Kübler, A. Training leads to increased auditory brain–computer interface performance of end-users with motor impairments. Clin. Neurophysiol. 2016. [Google Scholar] [CrossRef]
  32. Miho, S.; Yutaka, H.; Isao, N.; Alejandro, G.; Yoshinori, T.; Shohei, Y. Improving the performance of an auditory brain-computer interface using virtual sound sources by shortening stimulus onset asynchrony. Front. Neurosci. 2018, 12, 108–118. [Google Scholar]
  33. Brouwer, A.M. A tactile p300 brain–computer interface. Front. Neurosci. 2010, 4, 19. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Waal, M.V.D.; Severens, M.; Geuze, J.; Desain, P. Introducing the tactile speller: An ERP-based brain-computer interface for communication. J. Neural Eng. 2012, 9, 045002. [Google Scholar] [CrossRef]
  35. Kaufmann, T.; Holz, E.M.; Kübler, A. Comparison of tactile and auditory and visual modality for brain–computer interface use: A case study with a patient in the locked-in state. Front. Neurosci. 2013, 7, 129. [Google Scholar]
  36. Kaufmann, T.; Herweg, A.; Andrea, K. Toward brain-computer interface based wheelchair control utilizing tactually-evoked event-related potentials. J. Neural Eng. Rehabil. 2014, 11, 7. [Google Scholar] [CrossRef] [Green Version]
  37. Herweg, A.; Gutzeit, J.; Kleih, S.; Kübler, A. Wheelchair control by elderly participants in a virtual environment with a brain-computer interface (BCI) and tactile stimulation. Biol. Psychol. 2016, 121, 117–124. [Google Scholar] [CrossRef]
  38. Liu, Y.; Wang, J.; Yin, E.; Yu, Y.; Zhou, Z.; Hu, D. A tactile ERP-based brain–computer interface for communication. Int. J. HCI 2018, 35, 1–9. [Google Scholar] [CrossRef]
  39. Bernasconi, F.; Noel, J.-P.; Park, H.D.; Faivre, N.; Seeck, M.; Laurent, S.; Schaller, K.; Blanke, O.; Serino, A. Audio-tactile and peripersonal space processing around the trunk in human parietal and temporal cortex: An intracranial EEG study. Cereb. Cortex 2018, 28, 3385–3397. [Google Scholar] [CrossRef]
  40. Gao, S.; Wang, Y.; Gao, X.; Hong, B. Visual and auditory brain-computer interfaces. IEEE Trans. Bio Med. Eng. 2014, 61, 1436–1447. [Google Scholar]
  41. Thurlings, M.E.; Anne-Marie, B.; Van, E.J.B.F.; Peter, W. Gaze-independent ERP-BCIs: Augmenting performance through location-congruent bimodal stimuli. Front. Neurosci. 2014, 8. [Google Scholar] [CrossRef] [PubMed]
  42. Stein, B.E.; Stanford, T.R. Multisensory integration: Current issues from the perspective of the single neuron. Nat. Rev. Neurosci. 2008, 9, 255–266. [Google Scholar] [CrossRef] [PubMed]
  43. Ito1, T.; Gracco, V.L.; Ostry, D.J. Temporal factors affecting somatosensory-auditory interactions in speech processing. Front. Psychol. 2014, 5, 1198. [Google Scholar]
  44. Tidoni, E.; Gergondet, P.; Fusco, G. The role of audio-visual feedback in a thought-based control of a humanoid robot: A BCI study in healthy and spinal cord injured people. IEEE Trans. Neural Syst. Rehabil. Eng. 2017, 25, 772–781. [Google Scholar] [CrossRef]
  45. Tonelli, A.; Campus, C.; Serino, A.; Gori, M. Enhanced audio-tactile multisensory interaction in a peripersonal task after echolocation. Exp. Brain Res. 2019, 237, 3–4. [Google Scholar] [CrossRef] [Green Version]
  46. An, X.; Johannes, H.; Dong, M.; Benjamin, B.; Virginie, V.W. Exploring combinations of auditory and visual stimuli for gaze-independent brain-computer interfaces. PLoS ONE 2014, 9, E111070. [Google Scholar] [CrossRef] [Green Version]
  47. Sun, H.; Jin, J.; Zhang, Y.; Wang, B.; Wang, X. An Improved Visual-Tactile P300 Brain Computer Interface. In International Conference on Neural Information Processing; Springer: Cham, Switzerland, 2017. [Google Scholar]
  48. Wilson, A.D.; Tresilian, J.; Schlaghecken, F. The masked priming toolbox: An open-source MATLAB toolbox for masked priming researchers. Behav. Res. Methods 2011, 43, 201–214. [Google Scholar] [CrossRef]
  49. Gardner, B. HRTF Measurements of a KEMAR Dummy Head Microphone. In MIT Media Lab Perceptual Computing Technical Report; MIT Media Laboratory: Cambridge, MA, USA, 1994. [Google Scholar]
  50. Bruns, P.; Liebnau, R.; Röder, B. Cross-modal training induces changes in spatial representations early in the auditory processing pathway. Psychol. Sci. 2011, 22, 1120–1126. [Google Scholar] [CrossRef]
  51. Yao, L.; Sheng, X.; Mrachacz-Kersting, N.; Zhu, X.; Farina, D.; Jiang, N. Sensory stimulation training for BCI system based on somatosensory attentional orientation. IEEE Trans. Biomed. Eng. 2018, 66, 640–646. [Google Scholar] [CrossRef] [Green Version]
  52. Cecotti, H.; Rivet, B.; Congedo, M.; Jutten, C.; Bertrand, O.; Maby, E. A robust sensor-selection method for P300 braincomputer interfaces. J. Neural Eng. 2011, 8, 016001. [Google Scholar] [CrossRef] [Green Version]
  53. Colwell, K.A.; Ryan, D.B.; Throckmorton, C.S.; Sellers, E.W.; Collins, L.M. Channel selection methods for the P300 speller. J. Neurosci. Methods 2014, 232, 6–15. [Google Scholar] [CrossRef] [Green Version]
  54. David, J.C. Bayesian interpolation. Neural Comput. 1992, 4, 415–447. [Google Scholar]
  55. Hoffmann, U.; Vesin, J.M.; Ebrahimi, T.; Diserens, K. An efficient P300-based brain–computer interface for disabled subjects. J. Neurosci. Methods 2008, 167, 115–125. [Google Scholar] [CrossRef] [Green Version]
  56. Zhou, W.; Liu, Y.; Yuan, Q.; Li, X. Epileptic seizure detection using lacunartiy and Bayesian linear discriminant analysis in intracranial EEG. IEEE Trans. Biomed. Eng. 2013, 60, 3375–3381. [Google Scholar] [CrossRef]
  57. Occelli, V.; Spence, C.; Zampini, M. Audiotactile interactions in front and rear space. Neurosci. Biobehav. Rev. 2011, 35, 589–598. [Google Scholar] [CrossRef]
  58. Farne, A.; Làdavas, E. Auditory peripersonal space in humans. J. Cognit. Neurosci. 2002, 14, 1030–1043. [Google Scholar] [CrossRef]
  59. Marzoli, D.; Tommasi, L. Side biases in humans (Homo sapiens): Three ecological studies on hemispheric asymmetries. Naturwissenschaften 2009, 96, 1099–1106. [Google Scholar] [CrossRef]
  60. Hiscock, M.; Kinsbourne, M. Attention and the right-ear advantage: What is the connection? Brain Cognit. 2011, 76, 263–275. [Google Scholar] [CrossRef]
  61. Satrevik, B. The right ear advantage revisited: Speech lateralization in dichotic listening using consonant-vowel and vowel-consonant syllables. Laterality 2012, 17, 119–127. [Google Scholar] [CrossRef]
  62. Cao, Y.; Zhou, S.; Wang, Y. Neural dynamics of cognitive flexibility: Spatiotemporal analysis of event-related potentials. J. South. Med. Uni. 2017, 37, 755–760. [Google Scholar]
  63. Gill, P.; Woolley, S.M.N.; Fremouw, T.; Theunissen, F.E. What’s that sound? Auditory area CLM encodes stimulus surprise and not intensity or intensity changes. J. Neurophysiol. 2008, 99, 2809–2820. [Google Scholar] [CrossRef] [Green Version]
  64. Choi, D.; Nishimura, T.; Midori, M. Effect of empathy trait on attention to various facial expressions: Evidence from n170 and late positive potential (LPP). J. Physiol. Anthropol. 2014, 33, 18. [Google Scholar] [CrossRef] [Green Version]
  65. Catani, M.; Jones, D.K.; Ffytche, D.H. Perisylvian language networks of the human brain. Ann. Neurol. 2005, 57, 8–16. [Google Scholar] [CrossRef]
  66. Apostolova, L.G.; Lu, P.; Rogers, S.; Dutton, R.A.; Hayashi, K.M.; Toga, A.W. 3D mapping of language networks in clinical and pre-clinical Alzheimer’s disease. Brain Lang. 2008, 104, 33–41. [Google Scholar] [CrossRef] [Green Version]
  67. Leavitt Victoria, M.; Molholm, S.; Gomez-Ramirez, M. “What” and “Where” in auditory sensory processing: A high-density electrical mapping study of distinct neural processes underlying sound object recognition and sound localization. Front. Integr. Neurosci. 2011, 5, 23. [Google Scholar]
  68. Bizley, J.K.; Cohen, Y.E. The what, where and how of auditory-object perception. Nat. Rev. Neurosci. 2013, 14, 693–707. [Google Scholar] [CrossRef]
  69. Jiang, J.; Yin, E.; Wang, C.; Xu, M.; Ming, D. Incorporation of dynamic stopping strategy into the highspeed SSVEP-based BCIs. J. Neural Eng. 2018, 15, 046025. [Google Scholar] [CrossRef]
  70. Xu, M.; Qi, H.; Wan, B.; Yin, T.; Liu, Z.; Ming, D. A hybrid BCI speller paradigm combining P300 potential and the SSVEP blocking feature. J. Neural Eng. 2013, 10, 026001. [Google Scholar] [CrossRef]
Figure 1. Illustration of the character distribution of the 36-class brain-computer interface (BCI) system and the bimodal stimulus paradigm. In subfigure (a), the prompted character “J”, as an example, is marked red. In subfigure (b), the stimuli were delivered from corresponding location. Each yellow broken line represents a pair of electrical tactile electrodes placed at waist level. The arrows represent the directions of auditory stimuli delivered from headphones. The numbers identify the stimulus codes (location-congruent bimodal stimuli from “1” to “6”).
Figure 1. Illustration of the character distribution of the 36-class brain-computer interface (BCI) system and the bimodal stimulus paradigm. In subfigure (a), the prompted character “J”, as an example, is marked red. In subfigure (b), the stimuli were delivered from corresponding location. Each yellow broken line represents a pair of electrical tactile electrodes placed at waist level. The arrows represent the directions of auditory stimuli delivered from headphones. The numbers identify the stimulus codes (location-congruent bimodal stimuli from “1” to “6”).
Brainsci 10 00524 g001
Figure 2. Illustration of electro-tactile impulse stimuli.
Figure 2. Illustration of electro-tactile impulse stimuli.
Brainsci 10 00524 g002
Figure 3. Stimulus sequence generation.
Figure 3. Stimulus sequence generation.
Brainsci 10 00524 g003
Figure 4. Configuration of electrode locations.
Figure 4. Configuration of electrode locations.
Brainsci 10 00524 g004
Figure 5. Flow diagram of Jumpwise regression algorithm. The value pU is the p-value of the partial F-test statistics of the model containing all features of the selected channel set (SC) against the features in SC except those belonging to the unselected channel set (UC), and the value pS is the p-value of the partial F-test statistics of the model containing all features of SC against the features in SC except those belonging to SC. The p-value thresholds for adding and removing channels were set to 0.1 and 0.15, respectively.
Figure 5. Flow diagram of Jumpwise regression algorithm. The value pU is the p-value of the partial F-test statistics of the model containing all features of the selected channel set (SC) against the features in SC except those belonging to the unselected channel set (UC), and the value pS is the p-value of the partial F-test statistics of the model containing all features of SC against the features in SC except those belonging to SC. The p-value thresholds for adding and removing channels were set to 0.1 and 0.15, respectively.
Brainsci 10 00524 g005
Figure 6. Offline ITR (a) and classification accuracy (b) of auditory, tactile and bimodal BCI using the first ten runs of each session. Error bars denote the standard deviation.
Figure 6. Offline ITR (a) and classification accuracy (b) of auditory, tactile and bimodal BCI using the first ten runs of each session. Error bars denote the standard deviation.
Brainsci 10 00524 g006
Figure 7. Classification accuracy comparisons among the six directions of auditory (a), tactile (b), and bimodal (c) approaches. The green line divided the stimulus area into the right-posterior part (“4”, “5” and “6”) and the left-anterior part (“1”, “2” and “3”) for the ease of analyzing significance.
Figure 7. Classification accuracy comparisons among the six directions of auditory (a), tactile (b), and bimodal (c) approaches. The green line divided the stimulus area into the right-posterior part (“4”, “5” and “6”) and the left-anterior part (“1”, “2” and “3”) for the ease of analyzing significance.
Brainsci 10 00524 g007
Figure 8. The results of channel selection for each subject. (ac) denotes the eight optimal channels of each subject for the three modalities; (df) represents the total number of selected channels on the scalp for the twelve subjects. The color bar from dark blue to dark red indicates the summed number of channel selections for each electrode from zero to eight times, respectively.
Figure 8. The results of channel selection for each subject. (ac) denotes the eight optimal channels of each subject for the three modalities; (df) represents the total number of selected channels on the scalp for the twelve subjects. The color bar from dark blue to dark red indicates the summed number of channel selections for each electrode from zero to eight times, respectively.
Brainsci 10 00524 g008
Table 1. Online performance comparisons (Acc. (%), information transfer rate (ITR) (bit/min)).
Table 1. Online performance comparisons (Acc. (%), information transfer rate (ITR) (bit/min)).
SubjectAuditoryTactileBimodal
TrialAcc.ITRTrialAcc.ITRTrialAcc.ITR
S1266.6710.903100.0015.43280.0014.68
S2273.3312.723100.0015.43280.0014.68
S3133.336.34366.677.58146.6710.96
S4253.337.63373.338.85280.0014.68
S5266.6710.90360.006.41280.0014.68
S6260.009.21366.677.58266.6710.90
S7133.336.34246.676.17266.6710.90
S8473.336.78473.336.78373.338.85
S9473.336.78373.338.855100.009.59
S10586.677.27353.335.31380.0010.21
S11480.007.83246.676.17373.338.85
S12266.6710.90480.007.83146.6710.96
Mean2.5863.898.633.0070.008.532.3372.7811.66
Std1.2615.982.110.5816.893.251.0314.272.25

Share and Cite

MDPI and ACS Style

Zhang, B.; Zhou, Z.; Jiang, J. A 36-Class Bimodal ERP Brain-Computer Interface Using Location-Congruent Auditory-Tactile Stimuli. Brain Sci. 2020, 10, 524. https://doi.org/10.3390/brainsci10080524

AMA Style

Zhang B, Zhou Z, Jiang J. A 36-Class Bimodal ERP Brain-Computer Interface Using Location-Congruent Auditory-Tactile Stimuli. Brain Sciences. 2020; 10(8):524. https://doi.org/10.3390/brainsci10080524

Chicago/Turabian Style

Zhang, Boyang, Zongtan Zhou, and Jing Jiang. 2020. "A 36-Class Bimodal ERP Brain-Computer Interface Using Location-Congruent Auditory-Tactile Stimuli" Brain Sciences 10, no. 8: 524. https://doi.org/10.3390/brainsci10080524

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop