Next Article in Journal
Interdisciplinary Approach to Biological and Health Implications in Selected Professional Competences
Previous Article in Journal
Deep Learning Architecture Reduction for fMRI Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

cVEP Training Data Validation—Towards Optimal Training Set Composition from Multi-Day Data

Faculty of Technology and Bionics, Rhine-Waal University of Applied Sciences, 47533 Kleve, Germany
*
Author to whom correspondence should be addressed.
Brain Sci. 2022, 12(2), 234; https://doi.org/10.3390/brainsci12020234
Submission received: 31 December 2021 / Revised: 1 February 2022 / Accepted: 4 February 2022 / Published: 8 February 2022
(This article belongs to the Section Computational Neuroscience and Neuroinformatics)

Abstract

:
This paper investigates the effects of the repetitive block-wise training process on the classification accuracy for a code-modulated visual evoked potentials (cVEP)-based brain–computer interface (BCI). The cVEP-based BCIs are popular thanks to their autocorrelation feature. The cVEP-based stimuli are generated by a specific code pattern, usually the m-sequence, which is phase-shifted between the individual targets. Typically, the cVEP classification requires a subject-specific template (individually created from the user’s own pre-recorded EEG responses to the same stimulus target), which is compared to the incoming electroencephalography (EEG) data, using the correlation algorithms. The amount of the collected user training data determines the accuracy of the system. In this offline study, previously recorded EEG data collected during an online experiment with 10 participants from multiple sessions were used. A template matching target identification, with similar models as the task-related component analysis (TRCA), was used for target classification. The spatial filter was generated by the canonical correlation analysis (CCA). When comparing the training models from one session with the same session’s data (intra-session) and the model from one session with the data from the other session (inter-session), the accuracies were (94.84%, 94.53%) and (76.67%, 77.34%) for intra-sessions and inter-sessions, respectively. In order to investigate the most reliable configuration for accurate classification, the training data blocks from different sessions (days) were compared interchangeably. In the best training set composition, the participants achieved an average accuracy of 82.66% for models based only on two training blocks from two different sessions. Similarly, at least five blocks were necessary for the average accuracy to exceed 90%. The presented method can further improve cVEP-based BCI performance by reusing previously recorded training data.

1. Introduction

The electroencephalography (EEG)-based brain–computer interface (BCI) is a technology that has the potential of replacing, enhancing, and improving human interaction with the surrounding/environment as well as enhancing digital life [1,2,3,4].
While the most popular and easy to implement paradigms are based on the visually evoked potentials (e.g., SSVEP-based BCI [5,6]), the code-modulated VEP (cVEP) has been highly researched in the recent past with very promising results in terms of accuracy and information transfer rate (ITR) [7,8]. Spüler et al. used a one class support vector machine and error-related potentials for an online adaptation of a cVEP-based BCI system [9]. In 2016, Nakanishi et al. utilised canonical correlation analysis (CCA) and datasets recorded on two different days to evaluate the performance of different session-to-session transfer learning approaches, in order to optimise the SSVEP classification [10]. Nakanishi et al. presented an enhanced ensemble task-related component analysis (TRCA) method, where the spatial filter created from different targets were combined in order to increase the overall SSVEP-based BCI system performance [11]. In an online cue-guided experiment with 20 subjects, a mean accuracy of 89.8% was achieved. In another relevant paper published in 2015, Yuan et al. improved the performances of an SSVEP-based BCI while exploiting inter-subject template transformation [12]. Compared to the standard CCA, when using the proposed method, readings of 7.4% and 18.8% increases in accuracy for data lengths (time-window lengths) of 2.0 s and 1.5 s, respectively, were obtained for 12 participants in a simulated online experiment. Wong et al. presented a subject transfer-based CCA method to combine the information between the subjects and within the subject, reaching an average ITR of 198.18 bits/min [13]. Wang et al. proposed an inter- and intra-subject template-based multivariate synchronization index with an adaptive threshold for a 12-class SSVEP-based BCI dataset [14]. The results from 10 subjects showed an average accuracy of 99.2% for the new method compared to the standard CCA, which reached 93.6%. In 2020, Gembler et al. presented a session-to-session training transfer for a cVEP-based BCI system [15]. Eight out of 10 participants achieved an average accuracy of 97.1% when utilising the training data recorded 2 weeks earlier in an online copy-spelling task.
Exploring inter-subject variability is recently a popular research interest in the BCI field across different paradigms: in the sensorimotor rhythms (SMR)-based BCI system e.g., Saha et al. [16], Saha and Baumert [17]; in the simultaneous P300 event-related potential (ERP) and functional near-infrared spectroscopy (fNIRS) by Li et al. [18]; and in SSVEP-based BCI, Wei et al. [19], Tanaka [20]. In a recent review, Zarefa et al. compared different approaches to the users’ training and feature extraction algorithms of SSVEP-based BCI systems [21].
In 2021, Stawicki et al. [22] tested similarities in SSVEP and steady-state motion visual evoked potentials (SSMVEP) training by cross-analysing the training sessions for SSVEP and SSMVEP stimulus designs. The analysed data came from a previously performed comprehensive study with 86 participants [8].
The common necessity is that long repetitive training sessions ensure good signal-to-noise ratios for VEP-based BCI analysis. However, with respect to user-friendliness, these commonly-used repetitive training sessions may be tedious, as subjects are required to focus their gaze for a relatively long time on the stimulus, usually, multiple times. This is a time-consuming approach that commonly causes visual fatigue [8]. While the BCI system was developed for everyday use, alternative approaches were beneficial, such as training-free algorithms or subject-independent methods. Recent popular approaches in the SSVEP-based BCI systems are the transfer of subject-specific learning across stimulus frequencies [23], by transferring the learning knowledge within the subject and between subjects [24], or transferring training data between sessions [15].
Recently developed novel cVEP-based BCI systems focus on skipping the mostly required training session [25,26]. In 2018, Nagel and Spüler developed a model that predicts the arbitrary visual stimuli from EEG and the triggered brain response [25]. In an online experiment, nine participants achieved an average accuracy and ITR of 98.0% and 108.1 bits/min, respectively, for an optimised EEG2Code model using 32 EEG electrodes. In the recent paper, Thielen et al. developed a calibration-free method (0-training approach), where in an online copy-spelling task, nine participants achieved an average accuracy and ITR of 99.7% and 67.9 bits/min, respectively, using eight EEG electrodes and gold codes-based stimuli [26].
The herein paper utilises previously recorded EEG data from a cVEP study with 10 subjects [15]. The recordings were analysed with transfer data from one session to another session (both directions) of the same participant, and transfer data between different participants and sessions were included. In other words, the models (training data) from one participant were used to classify the data (testing data) from all participants from the same session and for the testing data of all participants of the other session. During this study (further referred to as the original study), two recording sessions took place per participant, which were spread apart by about 2 weeks. Here, we combine these recorded data into one pool, which is further analysed.

2. Materials and Methods

2.1. Participants

The data of the 10 participants of the original experiment [15] were utilised in this offline analysis. All subjects were recruited from Rhine-Waal University of Applied Sciences (two female), with a mean (SD) age of 25.4 years (4.1). A written informed consent was signed before the experiment; the experiment was approved by the ethical committee of the medical faculty of the University Duisburg-Essen, Germany. Participants had normal or corrected-to-normal vision and little to no prior BCI experience. All participants received a financial reward for their participation.

2.2. Stimulus Presentation

The stimuli were presented on a 24.5-inch monitor (Acer predator XB252Q) with a refresh rate of 240 Hz and a resolution of 1920 × 1080 pixels, which was positioned approximately 60 cm from the subject’s eyes [15]. The stimuli consisted of 32 white-squared targets with sizes of 150 × 150 pixels or 4.3 × 4.3 cm, which correspond to a visual angle of 4.1°, presented over a black background; see Figure 1. The distance between the targets was 75 pixels or 2.1 cm (vertical and horizontal).
The cVEP stimuli used the 63-bit m-sequence c 1 = 101011001101110110100100111000101111001010001100001000001111110 [27], where ‘0’ represented ‘black’ and ‘1’ represented ‘white’, thus presenting at full contrast. The remaining stimuli c k , k = 2 , , K , where K = 32, were generated by circularly left shifting c 1 by k · 2 bits. The update rate was set to 60 Hz; thus, the colour of the stimulus changed in accordance with the bit sequence: every fourth frame with the used refresh rate of 240 Hz. This m-sequence was displayed for a duration of 1.05 s, which emulates to the historical vertical refresh rate of 60 Hz, which was popular in many previous studies.
In order to achieve the most reliable stimulus presentation, the G-Sync® technology was disabled. Instead, the fixed refresh rate (240 Hz) was used, where a single frame is drawn within 4.17 milliseconds. Additionally, with the graphic card manufacturer tool, the number of pre-rendered frames (in the graphics card memory) was set to 1 (minimal available value). This reduced the internal drawing delays usually present in graphics card hardware to the actual drawing on the screen during the vertical screen refresh. Further details regarding the stimulus design can be found in [15].

2.3. Recordings

The recordings used for this study were the training phase of the original study, which consisted of a series of six repetitions (blocks). In every block, the participants had to focus their gaze at each target, which was cued sequentially through all 32 targets, for 2.10 s with a gaze shift pause of 1.0 s between every target. Between the blocks, the participants had an opportunity for a short break (while the EEG was still connected and the participant was still seated), after which they continued by pressing the space key.

2.4. Hardware and Software for Data Analysis

The hardware used for the offline analysis was a MSI GT73 notebook equipped with an Intel processor (Intel Core i7-6820HK CPU @2.70 GHz) and 16 GB of RAM, running on Windows 10 Education. The software utilised for the offline analysis was MATLAB® 2021b, and the ensemble TRCA methods from [11] were adopted.

2.5. CCA-Based Spatial Filter Design and Template Generation

Canonical correlation analysis is the most popular and widely used method for generating the spatial filters, which finds a linear transformation that maximizes the correlation between the recorded signal and a template signal, e.g., sine–cosine signals or averaged EEG template signals [28,29]. Typically, only the first canonical correlation and corresponding weights (w) are used for the classification and construction of filters [11].
Similar to our previous study, we combined the CCA method with the TRCA template construction [15,22,30]. Each training trial was stored in an m × n matrix, where m denotes the number of electrode channels (here, m = 16 ) and n denotes the number of sample points (here, there are two 1.05 s stimulus cycles, n = 1.05 · F S · 2 = 1260 ).
Given two multi-dimensional variables X R m 1 × n and Y R m 2 × n , CCA identifies weights w X R m 1 and w Y R m 2 that maximize the correlation, ρ , between the so-called canonical variates x = X T w X and y = Y T w Y by solving
ρ = max w X , w Y w X T X Y T w Y w X T X X T w X w Y T Y Y T w Y .
The correlation value ρ that solves (1) is the canonical correlation. For most VEP-based BCI realizations, only the first canonical correlation weights ( w X , w Y ) are used for classification or for the spatial filter design.
The recorded data were segmented into single trials T i R m × n , and used to generate a CCA-based spatial filter w R m . In total, 32 × 6 = 192 of such trials T i , j R m × n , i = 1 , , K , j = 1 , , n b , where n b = 6 , were recorded.
For each target, individual templates X i R m × n and filters w i were determined ( i = 1 , , K ). In order to generate the spatial filters, the two matrices were constructed,
T ^ i = [ T i , 1 T i , 2 T i , n b ] and X i = [ X ^ i X ^ i X ^ i n b ] ,
where X ^ i represents the average of the n b trials corresponding to the i-th class. When inserted into Equation (1), these two matrices yield the filter vectors: w i = w T ^ i , i = 1 , , K .

2.6. Classification

In order to add further improvements to the target classification, the ensemble spatial filter w, introduced in the TRCA method [11], which concatenates the spatial filters of all targets (classes), and the filter bank-based analysis [31], were implemented. For the filter bank design, a k was calculated as
a k = α k k = 1 K α k , where α k = k 1.25 + 0.25 ,
which gives the following normalised results: 0.386, 0.207, 0.156, 0.132, and 0.119. The lower cut-off frequencies for the k-th sub-band were 6, 14, 22, 30, and 38 Hz, and the upper cut-off frequency was 60 Hz for all-sub, which was filtered with an 4th-order Butterworth filter. In order to cancel out any phase response, forward and reverse filtering methods were applied.
The output command (C) was calculated based on the weighted linear combinations of the Pearson’s correlation coefficient ( ρ k ( X i ^ T w i , T i ^ T w i ) ) and the filter bank sub-bands (K) and amplitude ( a k ),
C = arg max k = 1 , , K λ k , where λ K = k = 1 K a k ρ k
for every trial in each block.
The information transfer rate was calculated utilising the regular ITR formula:
I T R = ( l o g 2 ( N ) + P l o g 2 ( P ) + ( 1 P ) l o g 2 ( 1 P N 1 ) ) * 60 / T ,
where N is the number of targets, P is accuracy, and T is the total time, including the gaze shift (inter-trial) pause. An online ITR calculator can be found at (https://bci-lab.hochschule-rhein-waal.de/en/itr.html, accessed on 31 December 2021).

2.7. Procedure

Both training sessions from the original study [15] were recorded on different days (mean distance 14.4 days), where each training session contained 6 training blocks. All the blocks from both sessions were used as follows: the testing data from 1st session ( D 1 ) blocks were numbered from 1 to 6, and the testing data from 2nd session ( D 2 ) blocks were numbered 7 to 12; see Figure 2.
For this study, the blocks were arranged in the following sequence ( s e q ) = 1, 7, 2, 8, 3, 9, 4, 10, 5, 11, 6, and 12 for the leave-one-out, stepwise correlation and for 5 different variant combinations, which are described as follows (see Figure 2):
  • Variant 1 ( V 1 ), where block b l 12 was used for testing and blocks b l 1 b l 11 in s e q order were used for model building.
  • Variant 2 ( V 2 ), where blocks b l 12 and b l 6 were used for testing and blocks b l 1 b l 11 in s e q order without b l 6 were used for model building.
  • Variant 3 ( V 3 ), where blocks b l 12 , b l 6 , and b l 11 were used for testing and blocks b l 1 b l 10 in s e q order without b l 6 were used for model building.
  • Variant 4 ( V 4 ), where blocks b l 12 , b l 6 , b l 11 , and b l 5 were used for testing and blocks b l 1 b l 10 in s e q order without b l 5 and b l 6 were used for model building.
  • Variant 5 ( V 5 ), where blocks b l 12 , b l 6 , b l 11 , b l 5 , and b l 10 were used for testing and blocks b l 1 b l 9 in s e q order without b l 5 and b l 6 were used for model building.
Additionally, a cross testing between session 1 ( D 1 ) containing b l 1 b l 6 and session2 ( D 2 ) containing b l 7 b l 12 was performed, where models built from one session were used for testing the data from the other session, e.g., model M (training data, x-axis) from subject 1 session 1 was used to classify the testing data (y-axis) of all subjects from the same session (session 1, Figure 3, left side) and all subjects from the other session (session 2, Figure 4, left side). The other direction was also analysed e.g., model M (training data, x-axis) from subject 1 session 2 was used to classify the testing data (y-axis) of all subjects from the same session (session 2, Figure 3, right side) and all subjects from the other session (session 1, Figure 4, right side). The accuracy of these results is presented in Figure 3 and Figure 4, and the detailed diagonals are presented in Table 1.
The stepwise increasing concatenated models were built as follows (e.g., for testing b l 12 ):
  • Step 1—model b l 1
  • Step 2—model b l 1 + b l 7
  • Step 3—model b l 1 + b l 7 + b l 2
  • Step 4—model b l 1 + b l 7 + b l 2 + b l 8
  • Step 5—model b l 1 + b l 7 + b l 2 + b l 8 + b l 3
  • Step 6—model b l 1 + b l 7 + b l 2 + b l 8 + b l 3 + b l 9
  • Step 7—model b l 1 + b l 7 + b l 2 + b l 8 + b l 3 + b l 9 + b l 4
  • Step 8—model b l 1 + b l 7 + b l 2 + b l 8 + b l 3 + b l 9 + b l 4 + b l 10
  • Step 9—model b l 1 + b l 7 + b l 2 + b l 8 + b l 3 + b l 9 + b l 4 + b l 10 + b l 5
  • Step 10—model b l 1 + b l 7 + b l 2 + b l 8 + b l 3 + b l 9 + b l 4 + b l 10 + b l 5 + b l 11
  • Step 11—model b l 1 + b l 7 + b l 2 + b l 8 + b l 3 + b l 9 + b l 4 + b l 10 + b l 5 + b l 11 + b l 6
For the leave-one-out cross-validation, if one block was tested, the steps would be built accordingly without these blocks.
The accuracy was calculated for every trial out of 32 in one block.

3. Results

Table 1 shows the results for session 1 ( b l 1 b l 6 ) vs. session 2 ( b l 7 b l 12 ) cross analysis, where the training data of one session were used to create the reference models M and the data from both were used for testing D. The training data of session 1 was used to build the model session 1 ( M s e s s i o n 1 ), and this model was used to classify the testing data from session 1 ( D s e s s i o n 1 ) and the testing data of session 2 ( D s e s s i o n 2 ). Similar, the training data of session 2 was used to build the model session 2 ( M s e s s i o n 2 ), and this model was used to classify the testing data of sessions 1 ( D s e s s i o n 1 ) and session 2 ( D s e s s i o n 2 ). A pairwise Mann–Whitney U-test between the M s e s s i o n 1 D s e s s i o n 1 vs. M s e s s i o n 1 D s e s s i o n 2 showed a statistical significance with U = 21.5, Z = −2.17, p < 0.05; between the M s e s s i o n 1 D s e s s i o n 1 vs. M s e s s i o n 2 D s e s s i o n 1 showed a statistical significance with U = 23.0, Z = −2.05, p < 0.05; between the M s e s s i o n 2 D s e s s i o n 2 vs. M s e s s i o n 1 D s e s s i o n 2 showed no statistical significance with U = 30.5, Z = −1.48, p = 0.14; between the M s e s s i o n 2 D s e s s i o n 2 vs. M s e s s i o n 2 D s e s s i o n 1 showed no statistical significance with U = 30.5, Z = −1.48, p = 0.14.
The bottom part of Table 1 shows the corresponding information transfer rates (ITRs), as shown in Equation (5).
The values in Table 1 also represent the corresponding diagonals of Figure 3 and Figure 4, where the additional cross-subject validation was performed. In Figure 3, the models and data from the same session (intra session) were cross-validated between the subjects and in Figure 4 the models and data from the other session (inter session) were cross-validated between the subjects.
In order to test the interchangeability of single training blocks from different multi-day sessions, the testing sequence arrangement (see Figure 2) was constructed and evaluated in Figure 5 with a single block bl12 against a concatenating increasing model from both sessions. This approach was also used to compare all other block from the multi-day pool in a leave-one-out cross-validation with a stepwise increasing (concatenating) models, including all other blocks in Table 2. The eleven bars in Figure 5 corresponds to the column labelled 12 in Table 2.
The different variants constructed in Figure 2 and evaluated in Table 3 test the different multi-day training data in order to build a minimal training model for improving the accuracy. Here, a Kruskal–Wallis Test showed no statistical significance between the different variants V 1 V 5 with H(4) = 0.271, p = 0.992.

4. Discussion

This paper focuses on the offline analysis of the cVEP training data only of the original study [15]; the training sessions were unified for all participants of the original study, which was performed on 2 separate days, spread apart for around 14 days.
Here, the focus lies on finding the suitable training process for the used cVEP system by cross-analysing the training data of both sessions and finding the best training strategy. First, the training data of each session were used for TRCA model building and testing (see Figure 3). When analysing this training intra-session’s classification accuracies, averaged over all six blocks of a leave-one-out cross-validation, session 1 and session 2 reached 94.84% and 94.53%, respectively, which correspond to an ITR of 88.0 bits/min and 87.1 bits/min, respectively (see Table 1).
Both training sessions’ data from the 1st and the 2nd day, session 1 and session 2, respectively, were cross-validated against the models from the other session (inter-session); see Table 1 and Figure 4. Interestingly, this cross-session analysis (presented in Table 1) shows that the participants could still achieve reliable control over the system with accuracies at around 76.67% and 77.34%, which shows that the fresh model can classify old data.
In order to avoid favouring the second session and the dependencies arising from the session continuity, the blocks data of both sessions were interchangeably (alternately) stacked, creating a continuous testing sequence s e q . This sequence was also used for stepwise increasing models. These models were analysed with a leave-one-out cross-validation of every block and averaged across participants (Table 2), yielding a maximum accuracy of 96.88%, which is higher than the intra-session average of both of the sessions (Table 1). This result was achieved for testing the block b l 8 with the model at the step 6. The results of the testing sequence show a relevant increase in accuracy when additional blocks are added (increasing step number) to the model build. Interestingly, in step 2, where just two blocks were used for the model, the averaged accuracy was 82.66%. This accuracy increased with every additional step until it reached 94.44%.
Lastly, to find the minimalistic training procedure, the tested sequence s e q data were arranged into five variants. In every variant, the accuracy was calculated by averaging the results for every testing block (see Table 3, Figure 2) with a fixed model. When analysing these results, the average accuracy had almost no difference between the variants V 1 V 5 . Interestingly, when more testing data were used in a variant (e.g., V 5 ), there was only a slight increase in the classification accuracies, reaching 93.56% when compared to e.g., V 2 , which achieved 92.5%, the same as V 1 .
These results show that from at least two sessions a minimum of two training blocks are required, in order to achieve sufficient cVEP-based BCI system control, so that the accuracy exceeds 80%, as mentioned in Renton et al. [32]. In this case, the average classification accuracies achieved 82%, and if at least five training blocks are used, the average accuracy exceeds 90% (see Table 2). The 2nd step of the tested sequence included the 1st blocks of both sessions, thus, one old (previous session) and one new (current session) block of training data. The findings of this study suggest that the already collected training data can be efficiently reused, e.g., as transferred training data, to increase the user-friendliness of the BCI system by reducing repetitive training and thus minimising the potential visual fatigue caused by the stimuli.
Compared to the system developed by Thielen et al. [26], which only requires an initial warm-up period of 12 s, the here-presented approach requires a minimum of one new training block, which corresponds to 98.2 s (including the inter-trial time of 1.0 s). On the other hand, some minimal training can be beneficial (easier adaptation) for different stimulation hardware (higher refresh rates which corresponds to higher cVEP carrier frequency), where the number of gold codes can be limited; instead, longer m-sequences can be implemented: one example is a 124-bit quintary m-sequence [33]).

5. Conclusions

Our findings show that at least two training blocks from two sessions form a sufficient starting point for further improvements in a modern cVEP-based BCI system, which in future research can further be improved with e.g., an online adaptation process. While the novel developed/emerging systems can work without a user-specific training (e.g., Spüler et al. [9] or Thielen et al. [26]), most of the currently used and popular cVEP-based BCI systems require some training phase that could benefit from previously collected training data, also from other subjects. This cross-subject feature extraction requires further research. While nothing can replace a “fresh” EEG data from the training (a necessity for an optimal model), the need for new re-training data can be reduced to the bare minimum. This can help some potential users in re-using their own previously collected data for the BCI system for everyday use and thus boost the setup times and further increase the initial daily performance of a cVEP system. This is also valid for many other BCI paradigms.

Author Contributions

Conceptualization, P.S.; methodology, P.S.; software, P.S.; validation, P.S.; formal analysis, P.S.; investigation, P.S.; data curation, P.S.; writing—original draft preparation, P.S.; writing—review and editing, P.S. and I.V.; visualization, P.S.; supervision, I.V.; project administration, I.V. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Ethic Committee Name: University Duisburg-Essen, Germany Approval Code: 17-7693-BO Approval Date: 12 September 2017.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Acknowledgments

The authors gratefully acknowledge the financial support by the association “The Friends of the University Rhein-Waal–Campus Cleve”.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Brunner, C.; Birbaumer, N.; Blankertz, B.; Guger, C.; Kübler, A.; Mattia, D.; Millán, J.d.R.; Miralles, F.; Nijholt, A.; Opisso, E.; et al. BNCI Horizon 2020: Towards a roadmap for the BCI community. Brain-Comput. Interfaces 2015, 2, 1–10. [Google Scholar] [CrossRef] [Green Version]
  2. Abiri, R.; Borhani, S.; Sellers, E.W.; Jiang, Y.; Zhao, X. A comprehensive review of EEG-based brain–computer interface paradigms. J. Neural Eng. 2019, 16, 011001. [Google Scholar] [CrossRef] [PubMed]
  3. Stegman, P.; Crawford, C.S.; Andujar, M.; Nijholt, A.; Gilbert, J.E. Brain–computer interface software: A review and discussion. IEEE Trans. Hum.-Mach. Syst. 2020, 50, 101–115. [Google Scholar] [CrossRef]
  4. Rashid, M.; Sulaiman, N.; PP Abdul Majeed, A.; Musa, R.M.; Bari, B.S.; Khatun, S. Current status, challenges, and possible solutions of EEG-based brain-computer interface: A comprehensive review. Front. Neurorobot. 2020, 14, 25. [Google Scholar] [CrossRef]
  5. Rezeika, A.; Benda, M.; Stawicki, P.; Gembler, F.; Saboor, A.; Volosyak, I. Brain-Computer Interface Spellers: A Review. Brain Sci. 2018, 8, 57. [Google Scholar] [CrossRef] [Green Version]
  6. Li, M.; He, D.; Li, C.; Qi, S. Brain–Computer Interface Speller Based on Steady-State Visual Evoked Potential: A Review Focusing on the Stimulus Paradigm and Performance. Brain Sci. 2021, 11, 450. [Google Scholar] [CrossRef] [PubMed]
  7. Martínez-Cagigal, V.; Thielen, J.; Santamaría-Vázquez, E.; Pérez-Velasco, S.; Desain, P.; Hornero, R. Brain–computer interfaces based on code-modulated visual evoked potentials (c-VEP): A literature review. J. Neural Eng. 2021, 18, 061002. [Google Scholar] [CrossRef] [PubMed]
  8. Volosyak, I.; Rezeika, A.; Benda, M.; Gembler, F.; Stawicki, P. Towards solving of the Illiteracy phenomenon for VEP-based brain-computer interfaces. Biomed. Phys. Eng. Express 2020, 6, 035034. [Google Scholar] [CrossRef]
  9. Spüler, M.; Rosenstiel, W.; Bogdan, M. Online Adaptation of a C-VEP Brain-Computer Interface(BCI) Based on Error-Related Potentials and Unsupervised Learning. PLoS ONE 2012, 7, e51077. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Nakanishi, M.; Wang, Y.; Jung, T.P. Session-to-Session Transfer in Detecting Steady-State Visual Evoked Potentials with Individual Training Data. In Foundations of Augmented Cognition: Neuroergonomics and Operational Neuroscience; Schmorrow, D.D., Fidopiastis, C.M., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 253–260. [Google Scholar]
  11. Nakanishi, M.; Wang, Y.; Chen, X.; Wang, Y.T.; Gao, X.; Jung, T.P. Enhancing Detection of SSVEPs for a High-Speed Brain Speller Using Task-Related Component Analysis. IEEE Trans. Biomed. Eng. 2017, 65, 104–112. [Google Scholar] [CrossRef]
  12. Yuan, P.; Chen, X.; Wang, Y.; Gao, X.; Gao, S. Enhancing performances of SSVEP-based brain–computer interfaces via exploiting inter-subject information. J. Neural Eng. 2015, 12, 046006. [Google Scholar] [CrossRef] [PubMed]
  13. Wong, C.M.; Wang, Z.; Wang, B.; Lao, K.F.; Rosa, A.; Xu, P.; Jung, T.P.; Chen, C.P.; Wan, F. Inter-and Intra-Subject Transfer Reduces Calibration Effort for High-Speed SSVEP-Based BCIs. IEEE Trans. Neural Syst. Rehabil. Eng. 2020, 28, 2123–2135. [Google Scholar] [CrossRef] [PubMed]
  14. Wang, H.; Sun, Y.; Li, Y.; Chen, S.; Zhou, W. Inter-and intra-subject template-based multivariate synchronization index using an adaptive threshold for SSVEP-based BCIs. Front. Neurosci. 2020, 14, 717. [Google Scholar] [CrossRef] [PubMed]
  15. Gembler, F.; Stawicki, P.; Rezeika, A.; Benda, M.; Volosyak, I. Exploring Session-to-Session Transfer for Brain-Computer Interfaces based on Code-Modulated Visual Evoked Potentials. In Proceedings of the 2020 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Toronto, ON, Canada, 11–14 October 2020; pp. 1505–1510. [Google Scholar]
  16. Saha, S.; Ahmed, K.I.U.; Mostafa, R.; Hadjileontiadis, L.; Khandoker, A. Evidence of variabilities in EEG dynamics during motor imagery-based multiclass brain–computer interface. IEEE Trans. Neural Syst. Rehabil. Eng. 2017, 26, 371–382. [Google Scholar] [CrossRef]
  17. Saha, S.; Baumert, M. Intra-and inter-subject variability in EEG-based sensorimotor brain computer interface: A review. Front. Comput. Neurosci. 2020, 13, 87. [Google Scholar] [CrossRef] [Green Version]
  18. Li, F.; Tao, Q.; Peng, W.; Zhang, T.; Si, Y.; Zhang, Y.; Yi, C.; Biswal, B.; Yao, D.; Xu, P. Inter-subject P300 variability relates to the efficiency of brain networks reconfigured from resting-to task-state: Evidence from a simultaneous event-related EEG-fMRI study. NeuroImage 2020, 205, 116285. [Google Scholar] [CrossRef]
  19. Wei, C.S.; Nakanishi, M.; Chiang, K.J.; Jung, T.P. Exploring human variability in steady-state visual evoked potentials. In Proceedings of the 2018 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Miyazaki, Japan, 7–10 October 2018; pp. 474–479. [Google Scholar]
  20. Tanaka, H. Group task-related component analysis (gTRCA): A multivariate method for inter-trial reproducibility and inter-subject similarity maximization for EEG data analysis. Sci. Rep. 2020, 10, 1–17. [Google Scholar] [CrossRef]
  21. Zerafa, R.; Camilleri, T.; Falzon, O.; Camilleri, K.P. To Train or Not to Train? A Survey on Training of Feature Extraction Methods for SSVEP-Based BCIs. J. Neural Eng. 2018, 15, 051001. [Google Scholar] [CrossRef]
  22. Stawicki, P.; Rezeika, A.; Volosyak, I. Effects of Training on BCI. In International Work-Conference on Artificial Neural Networks; Springer: Berlin/Heidelberg, Germany, 2021; pp. 69–80. [Google Scholar]
  23. Wong, C.M.; Wang, Z.; Rosa, A.C.; Chen, C.P.; Jung, T.P.; Hu, Y.; Wan, F. Transferring Subject-Specific Knowledge Across Stimulus Frequencies in SSVEP-Based BCIs. IEEE Trans. Autom. Sci. Eng. 2021, 18, 552–563. [Google Scholar] [CrossRef]
  24. Wang, H.; Sun, Y.; Wang, F.; Cao, L.; Zhou, W.; Wang, Z.; Chen, S. Cross-Subject Assistance: Inter-and Intra-Subject Maximal Correlation for Enhancing the Performance of SSVEP-Based BCIs. IEEE Trans. Neural Syst. Rehabil. Eng. 2021, 29, 517–526. [Google Scholar] [CrossRef]
  25. Nagel, S.; Spüler, M. Modelling the brain response to arbitrary visual stimulation patterns for a flexible high-speed brain-computer interface. PLoS ONE 2018, 13, e0206107. [Google Scholar] [CrossRef] [Green Version]
  26. Thielen, J.; Marsman, P.; Farquhar, J.; Desain, P. From full calibration to zero training for a code-modulated visual evoked potentials for brain–computer interface. J. Neural Eng. 2021, 18, 056007. [Google Scholar] [CrossRef]
  27. Bin, G.; Gao, X.; Wang, Y.; Li, Y.; Hong, B.; Gao, S. A High-Speed BCI Based on Code Modulation VEP. J. Neural Eng. 2011, 8, 025015. [Google Scholar] [CrossRef] [PubMed]
  28. Zhang, Y.; Xie, S.Q.; Wang, H.; Zhang, Z. Data Analytics in Steady-State Visual Evoked Potential-Based Brain–Computer Interface: A Review. IEEE Sens. J. 2020, 21, 1124–1138. [Google Scholar] [CrossRef]
  29. Zhuang, X.; Yang, Z.; Cordes, D. A Technical Review of Canonical Correlation Analysis for Neuroscience Applications. Hum. Brain Mapp. 2020, 41, 3807–3833. [Google Scholar] [CrossRef]
  30. Gembler, F.; Stawicki, P.; Saboor, A.; Volosyak, I. Dynamic Time Window Mechanism for Time Synchronous VEP-Based BCIs—Performance Evaluation with a Dictionary-Supported BCI Speller Employing SSVEP and c-VEP. PLoS ONE 2019, 14, e0218177. [Google Scholar] [CrossRef] [Green Version]
  31. Chen, X.; Wang, Y.; Gao, S.; Jung, T.P.; Gao, X. Filter Bank Canonical Correlation Analysis for Implementing a High-Speed SSVEP-Based Brain-Computer Interface. J. Neural Eng. 2015, 12, 046008. [Google Scholar] [CrossRef] [PubMed]
  32. Renton, A.I.; Mattingley, J.B.; Painter, D.R. Optimising non-invasive brain-computer interface systems for free communication between naïve human participants. Sci. Rep. 2019, 9, 1–18. [Google Scholar] [CrossRef]
  33. Gembler, F.W.; Rezeika, A.; Benda, M.; Volosyak, I. Five shades of grey: Exploring quintary m-sequences for more user-friendly c-vep-based bcis. Comput. Intell. Neurosci. 2020, 2020, 7985010. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The graphical user interface (GUI) used in the training phase of online experiment [15] with 32 targets. During the training session, the current target on which the participants needed to fix their gaze was marked with a green frame.
Figure 1. The graphical user interface (GUI) used in the training phase of online experiment [15] with 32 targets. During the training session, the current target on which the participants needed to fix their gaze was marked with a green frame.
Brainsci 12 00234 g001
Figure 2. Different testing arrangements used in this study. The training data from the original experiment [15] were divided into 12 blocks and rearranged into a testing sequence, which was divided into model building and testing datasets.
Figure 2. Different testing arrangements used in this study. The training data from the original experiment [15] were divided into 12 blocks and rearranged into a testing sequence, which was divided into model building and testing datasets.
Brainsci 12 00234 g002
Figure 3. Cross-validation of the training EEG data (2.1 s). This figure shows the accuracy of M session 1 Data session 1 and M session 2 Data session 2 tested across all participants, where e.g., the model (M) from session 1 was used to classify the session 1 training EEG data (left side), and model (M) from session 2 was used to classify the session 2 training EEG data (right side).
Figure 3. Cross-validation of the training EEG data (2.1 s). This figure shows the accuracy of M session 1 Data session 1 and M session 2 Data session 2 tested across all participants, where e.g., the model (M) from session 1 was used to classify the session 1 training EEG data (left side), and model (M) from session 2 was used to classify the session 2 training EEG data (right side).
Brainsci 12 00234 g003
Figure 4. Cross-validation of the training EEG data across paradigms for 2.1 s time window. This figure shows the accuracy of M session 1 Data session 2 and M session 2 Data session 1 tested across all participants, where e.g., the model (M) from session 1 was used to classify the session 2 training EEG data (left side), and model (M) from session 2 was used to classify the session 1 training EEG data (right side).
Figure 4. Cross-validation of the training EEG data across paradigms for 2.1 s time window. This figure shows the accuracy of M session 1 Data session 2 and M session 2 Data session 1 tested across all participants, where e.g., the model (M) from session 1 was used to classify the session 2 training EEG data (left side), and model (M) from session 2 was used to classify the session 1 training EEG data (right side).
Brainsci 12 00234 g004
Figure 5. This figure shows the classification accuracy of the b l 12 against models built from all other blocks arranged in the testing sequence ( s e q ) = 1, 7, 2, 8, 3, 9, 4, 10, 5, 11, 6. The models were built stepwise added together; the blocks in the s e q order, e.g., “1”, “7”, “2”, “8”, etc. represent the models built from b l 1 , b l 1 + b l 7 , b l 1 + b l 7 + b l 2 , and b l 1 + b l 7 + b l 2 + b l 8 , respectively. The whiskers mark the standard error.
Figure 5. This figure shows the classification accuracy of the b l 12 against models built from all other blocks arranged in the testing sequence ( s e q ) = 1, 7, 2, 8, 3, 9, 4, 10, 5, 11, 6. The models were built stepwise added together; the blocks in the s e q order, e.g., “1”, “7”, “2”, “8”, etc. represent the models built from b l 1 , b l 1 + b l 7 , b l 1 + b l 7 + b l 2 , and b l 1 + b l 7 + b l 2 + b l 8 , respectively. The whiskers mark the standard error.
Brainsci 12 00234 g005
Table 1. Cross sessions accuracies (%) and ITR (bits/min) results. The data of session 1 ( b l 1 b l 6 ) and session 2 ( D b l 7 b l 12 ) were used to create the models M s e s s i o n 1 and M s e s s i o n 2 , respectively.
Table 1. Cross sessions accuracies (%) and ITR (bits/min) results. The data of session 1 ( b l 1 b l 6 ) and session 2 ( D b l 7 b l 12 ) were used to create the models M s e s s i o n 1 and M s e s s i o n 2 , respectively.
AccuracyModel Session 1Model Session 2
SubjectData Session 1Data Session 2Data Session 1Data Session 2
197.4098.9696.8899.48
210068.7584.9094.27
310094.2798.9698.96
499.4899.4898.4499.48
598.4482.2996.3596.88
675.0025.5202.0870.83
799.4898.9699.48100
899.4896.8892.71100
979.1704.6911.9889.06
1010096.8891.6796.35
Mean94.8476.6777.3494.53
SD09.4534.2137.3909.00
ITRModel Session 1Model Session 2
SubjectData Session 1Data Session 2Data Session 1Data Session 2
190.9194.1689.995.37
296.7749.4770.4485.15
396.7785.1594.1694.16
495.3795.3793.0395.37
593.0366.7588.989.90
657.1009.500051.95
795.3794.1695.3796.77
895.3789.9082.4996.77
962.5100.1002.1476.64
1096.7789.9080.7888.90
Mean88.0067.4569.7287.10
SD15.0336.1336.9513.88
Table 2. Averaged accuracy (Acc.) results of the leave-one-out cross-validation of every single block of the testing sequence with the stepwise increasing model (step). Presented are means and SD (bottom table) of all 10 participants. The mean (SD) in the top part are average values for every step.
Table 2. Averaged accuracy (Acc.) results of the leave-one-out cross-validation of every single block of the testing sequence with the stepwise increasing model (step). Presented are means and SD (bottom table) of all 10 participants. The mean (SD) in the top part are average values for every step.
Acc.Single Blocks of Testing Sequence
Step172839410511612MeanSD
113.4413.1319.6915.6321.2511.8818.4412.5022.8110.6316.5612.8115.733.35
283.1375.0083.7589.6983.4482.8182.8184.0684.6979.6981.5681.2582.662.19
385.6386.2586.2590.3186.2585.9487.1989.3885.9482.8187.5083.7586.431.44
488.1389.0688.1390.9488.4489.3889.3891.5687.8187.5085.9488.7588.751.09
589.0690.3188.4495.3188.1390.0092.8190.9490.6388.4488.4489.6990.181.51
690.3192.1990.0096.8890.3190.0092.5092.8192.5091.5688.1391.8891.591.54
790.3193.7590.6395.9490.9492.8193.1394.6994.3892.5090.6392.5092.681.43
893.1393.7592.8195.6393.1392.8193.7594.6993.4492.8190.9494.3893.440.83
993.1394.0691.8895.9493.7593.7595.0095.3194.0692.8191.2594.3893.781.02
1093.4494.0694.3896.2593.1393.4496.2594.0695.6392.5091.5695.0094.141.13
1193.1394.3893.1396.5692.8194.0696.2595.3196.2593.4492.5095.3194.431.26
SDSingle Blocks of Testing Sequence
Step172839410511612
16.928.9423.2511.7922.4312.8323.9110.2126.2310.5422.8710.46
233.7934.8632.9318.8131.8429.429.427.0228.8134.3731.0628.41
328.6525.4828.6519.5126.8526.6925.6621.6128.0829.9822.6325.55
424.1519.7824.7718.7222.4419.7222.5918.5826.4922.9225.118.41
523.9520.7524.710.6523.2418.8512.8517.2720.6221.4521.9516.41
621.5217.2619.76.9119.6219.0813.615.817.0716.0121.1312.6
720.9613.3418.169.7818.3713.5912.6611.611.6712.9417.3710.44
815.2213.3414.4410.7414.4913.7411.4113.5913.313.0116.768.18
915.9212.3616.3511.7913.2611.698.9811.6211.5513.0116.467.48
1013.7812.3611.110.8114.4912.456.5612.87.6812.9415.456.94
1114.6411.3913.728.7715.1710.366.5610.656.5610.9713.916.29
Table 3. Accuracies of the different variants for the tested 10 participants (P1–P10).
Table 3. Accuracies of the different variants for the tested 10 participants (P1–P10).
Acc V 1 V 2 V 3 V 4 V 5
P110098.4497.9298.4498.75
P293.7598.4496.8897.6698.13
P396.88100100100100
P410098.4498.9698.4498.75
P593.7596.8896.8898.4496.25
P671.8870.3168.7568.7566.25
P7100100100100100
P810098.4498.9699.2299.38
P975.067.1970.8375.080.0
P1093.7596.8896.8897.6698.13
Mean92.5092.50092.6193.3693.56
SD10.4412.5812.0911.4511.30
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Stawicki, P.; Volosyak, I. cVEP Training Data Validation—Towards Optimal Training Set Composition from Multi-Day Data. Brain Sci. 2022, 12, 234. https://doi.org/10.3390/brainsci12020234

AMA Style

Stawicki P, Volosyak I. cVEP Training Data Validation—Towards Optimal Training Set Composition from Multi-Day Data. Brain Sciences. 2022; 12(2):234. https://doi.org/10.3390/brainsci12020234

Chicago/Turabian Style

Stawicki, Piotr, and Ivan Volosyak. 2022. "cVEP Training Data Validation—Towards Optimal Training Set Composition from Multi-Day Data" Brain Sciences 12, no. 2: 234. https://doi.org/10.3390/brainsci12020234

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop