Next Article in Journal
Steady-State Visual Evoked Potential Classification Using Complex Valued Convolutional Neural Networks
Previous Article in Journal
Third Party Certification of Agri-Food Supply Chain Using Smart Contracts and Blockchain Tokens
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Solving the SSVEP Paradigm Using the Nonlinear Canonical Correlation Analysis Approach

by
Danni Rodrigo De la Cruz-Guevara
1,2,*,
Wilfredo Alfonso-Morales
2 and
Eduardo Caicedo-Bravo
2
1
Department of Electrical, Electronics and Telecommunications Engineering, Universidad de las Fuerzas Armadas, Sangolqui 171103, Ecuador
2
School of Electrical and Electronics Engineering, Faculty of Engineering, Universidad del Valle, Calle 13 #100-00, Cali 760032, Colombia
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(16), 5308; https://doi.org/10.3390/s21165308
Submission received: 7 June 2021 / Revised: 31 July 2021 / Accepted: 1 August 2021 / Published: 6 August 2021
(This article belongs to the Section Biomedical Sensors)

Abstract

:
This paper presents the implementation of nonlinear canonical correlation analysis (NLCCA) approach to detect steady-state visual evoked potentials (SSVEP) quickly. The need for the fast recognition of proper stimulus to help end an SSVEP task in a BCI system is justified due to the flickering external stimulus exposure that causes users to start to feel fatigued. Measuring the accuracy and exposure time can be carried out through the information transfer rate—ITR, which is defined as a relationship between the precision, the number of stimuli, and the required time to obtain a result. NLCCA performance was evaluated by comparing it with two other approaches—the well-known canonical correlation analysis (CCA) and the least absolute reduction and selection operator (LASSO), both commonly used to solve the SSVEP paradigm. First, the best average ITR value was found from a dataset comprising ten healthy users with an average age of 28, where an exposure time of one second was obtained. In addition, the time sliding window responses were observed immediately after and around 200 ms after the flickering exposure to obtain the phase effects through the coefficient of variation (CV), where NLCCA obtained the lowest value. Finally, in order to obtain statistical significance to demonstrate that all approaches differ, the accuracy and ITR from the time sliding window responses was compared using a statistical analysis of variance per approach to identify differences between them using Tukey’s test.

1. Introduction

Electroencephalography (EEG) signals are nonlinear, i.e., they present unstable characteristics and often vary in quality between user-to-user trials or even trials of the same user, posing significant challenges to build brain–computer interfaces (BCIs). The human brain is a complex nonlinear system, where the biosignals among these brain activities are ‘3N’: nonstationary, nonlinear, and noisy [1]; although, at small time windows of less than 4 s, the signal can be considered quasi-stationary [2], and the registering or discarding of these 3N characteristics can significantly influence the results. This is why a better analysis of EEG signals for steady-state visual evoked potentials (SSVEP) requires unconventional analyses and methods to address the nonlinear and nonstationary nature of these [3]. Additionally, a better analysis is also justified as a blinking visual stimulus and becomes weaker in EEG signals as the frequency of the stimuli increases [4], which results in a range of limited frequencies.
For almost a century, EEG signal acquisitions have demanded strict experimental conditions that include rigorous experimental protocols [5]. Improvements through the processing of signals, data analysis, and statistical modeling of the collected information from multiple electrodes to remove artifacts in time, frequency, or phase domains by proposing filters or pipelines as PREP [6], HAPPE [7], Automagic [8], or MADE [9] are being studied. However, for the SSVEP paradigm, the artifact-removal techniques are considered unnecessary due to the short period required to collect a trial.
The SSVEP paradigm is a natural response to visual stimuli with specific frequencies. The brain generates a response with the same spectrum coming from the retina stimulated by a flickering exogenous input [10]. The response generally occurs in the occipital and parietal lobes of the brain, where it is possible to collect multiple trials quickly. These stimuli produce steady-stable and small-amplitude visual evoked potentials if the eyes face the same exogenous input [11,12] and its frequency is greater than 2 Hz; otherwise, it produces a transient VEP. However, the stimulus achieves a periodic response called SSVEP when the frequency is higher (>6 Hz) [13]. The VEP responses also undergo an early transitional stage before entering SSVEP (see Figure 1, figure adapted from [14]). The VEP response initially has three components or waves: the P100 is a positive wave that appears 100 ms after the stimulus and is rounded by the other two minor, harmful waves N75 and N135, which appear 75-ms and 135-ms later than the stimulus, respectively. N135’s waveform depends on the contour process and changes in contrast when binocular interactions occur [15].
BCI systems based on SSVEP will achieve a better performance only when the responses overcome the transitional stage. Taking Figure 1, a displacement longer than 195 ms after the stimulus (P195) was observed as an effective starting point since the stability of the response for feature extraction is guaranteed. The time-frequency decomposition analysis showed a stable SSVEP result that started 250 ms after stimulus presentation at each fundamental frequency and its harmonic components [16]. Another relevant task is defining the minimal time windows to be analyzed. The time window is chosen in the first instance by the system developer and represents a trade-off between accuracy and classification speed [17]. A frequency greater than 6 Hz is necessary to define a steady-state into VEP responses. Therefore, the studies should collect at least three wavelengths to analyze the EEG signals more effectively [18], e.g., a 6 Hz stimulus requires a time window of 500 ms.
On the other hand, the amplitude and phase of SSVEP responses are highly sensitive to stimulus parameters such as repetition rate, color, luminance or modulation depth, and spatial frequency [19,20]. However, the SSVEP paradigm stands out for its minimal training capacity; robustness, high signal-to-noise ratio (SNR); and high information transfer rate (ITR) [21,22]. Furthermore, for the analysis and design of BCI systems, it is necessary to consider three fundamental aspects: the short processing time, the high information transfer rate (ITR), and the fact that these should be as noninvasive as possible [23,24]. SSVEP complies with all of these aspects since its records are minimally invasive, collects multiple trials in a short period, and provides a high information transfer rate [25]. A well-designed SSVEP-based BCI system would decrease ocular fatigue by reducing the user’s exposure to the flickering exogenous inputs.
Following the above idea, many researchers have developed different feature extraction methods for the SSVEP paradigm, mainly focused on the well-known CCA approach, since this has been proven to be efficient, stable, and simple to apply [26]. The most current review about BCI mainly shows CCA-based methods and others for feature extraction [27] such as the filter bank CCA, the multiset CCA, the individual template CCA, or the multilayer correlation maximization. The study also includes other linear approaches such as task-related component analysis (TRCA), correlated component analysis (CORCA), and least absolute shrinkage and selection operator (LASSO). However, as we mentioned before, the EEG signals are nonlinear; so, approaches such as empirical and variational mode decomposition (i.e., EMD and VMD as nonlinear) have had outstanding performances, as mentioned in Labecki et al. [28], De la Cruz et al. [29].
Based on the above statement, this paper presents the nonlinear CCA (NLCCA) implementation to classify EEG signals using the SSVEP paradigm. No reference to NLCCA as a solution for SSVEP-based BCI systems was established, although the method has been used since its beginnings in other applications, especially climatic ones, such as the work of [30], and in more current ones such as [31]. Besides, we also compared its performance with other approaches such as CCA and LASSO. The first method represents the basis to show the treatment effects of the signals; the second is a simple approach that has served to obtain better performance than CCA [32]. Thus, this paper presents the average accuracy and ITR as comparison metrics using different time windows in two scenarios: full-users and removing the worst user. First, the best time window base on the average ITR value was defined. Next, an inspection of the phase effects during the time sliding window was carried out by taking into account and discarding the transitional stage issue, i.e., using EEG data when the stimuli appeared without shifting and 200 ms-later; here, the coefficient of variation metric was used. Finally, a statistical analysis was conducted to identify differences between the approaches using analysis of variance (ANOVA) and post-Tukey’s test [33,34].

2. Materials and Methods

In this study, an externally provided database was used. This dataset corresponds to the experiments developed at the Swartz Center for Computational Neuroscience, Institute for Neural Computation, from the University of San Diego [35] (Dataset available in https://www.kaggle.com/lzyuuu/ssvep-sandiego (accessed on 7 June 2021)). The dataset contains the records of ten volunteers without any reported illnesses who observed twelve visual stimuli while an EEG system captured their signals. The following subsections present a description of the main characteristics of the dataset that are relevant to the study (Section 2.1, Section 2.2 and Section 2.3), the implemented approaches (Section 2.4), and the used performance metrics (Section 2.5).

2.1. Experimental Setup and Volunteers

Ten healthy, right-handed volunteers (S1–S10) with standard or normal-corrected vision participated in the experiments (nine men and one woman with an average age of 28 years). They sat in a comfortable chair positioned 60 cm in front of a 27 in LCD monitor (60 Hz refresh rate, 1280 × 800 screen resolution) in a low-illuminated room.

2.2. Experimental Protocol

Each volunteer completed 15 blocks, and each block consisted of 12 trials corresponding to each stimulus (from 9.25 Hz to 14.75 Hz at 0.5 Hz intervals). In each trial, the user had one second to fix his/her gaze to the target stimulus position, which was indicated by a red square marker. Subsequently, all stimuli started to flicker simultaneously for 4 s on the monitor. It is important to note that the marker was randomly located for each new trial until the volunteer ended a block [35]. Figure 2 shows the time of each trial.

2.3. EEG Signal Recordings and Preprocessing

The sampling frequency of EEG signal recordings was found to be 256 Hz. Eight Ag/AgCl electrodes covered the occipital area using a BioSemi ActiveTwo EEG system (Biosemi, Inc. Amsterdam, The Netherlands), as shown in Figure 3, corresponding to electrodes O1, O2, Oz, PO7, PO3, POz, PO4, and PO8.
The raw recordings were then filtered using a bandpass between 6 Hz and 80 Hz, with an infinite impulse response (IIR) filter. Here, zero-phase forward and reverse IIR filtering was implemented as suggested in [35].

2.4. SSVEP Pattern Recognition Approaches

2.4.1. Least Absolute Shrinkage and Selection Operator (LASSO)

LASSO is a scattered regression model that offers high interpretable regression coefficients. Zhang et al. [36] proposed the implementation of the LASSO method for the detection of SSVEP signals [36], and it did in fact prove to be useful and robust in the feature extraction and selection of EEG signals into the SSVEP paradigm. LASSO’s main purpose is to solve a standard linear regression model [37,38].
LASSO obtains an approximation of the g n term in the SSVEP model denoted in
y n = X f g n + e n
where y n represents the EEG signal, X f the SSVEP reference signal, and e n is an additive noise vector [39]. β ^ is defined as the LASSO estimator and is calculated using quadratic programming, as given below [40]:
β ^ = a r g min β y X β 2 2 + λ β 1 ,
where · 1 , · 2 denote the norm l 1 and the norm l 2 , respectively; λ is a penalty parameter that provides a sparse solution (i.e., it forces many entries to zero). Thus, the optimization problem is to find the optimal β ^ vector—here, β ^ = [ β 1 , 1 , , β 1 , H , , β F , 1 , , β F , H ] T , where H is the representation of the number of harmonics. Each element indicates the level of contribution of the stimulus frequency f to the EEG signal. The largest LASSO estimator is established by the target stimulus, which is the one that provides the highest contribution and is defined as
f t = m a x f i = 1 N h = 1 H β f , h i N , f f 1 , , f F ,
where N is the number of channels of the EEG signal. LASSO is considered a subject-independent technique since the penalty parameter that influences the LASSO performance is calculated offline by taking advantage of the SSVEP data from multiple subjects [41]. There has been recent research where the LASSO method was applied because of the ability to perform multichannel analysis for identifying SSVEP signal frequencies (e.g., [32,42]).

2.4.2. Canonical Correlation Analysis (CCA)

CCA is a feature extraction method that is widely used to detect the frequency of EEG signals through the underlying correlation between two multidimensional variables. The idea of CCA for detection of the SSVEP paradigm is to take the EEG signal and relate it to a set of frequencies that coincide with the frequencies of the target stimuli and establish the highest multidimensional correlation [43].
The first multidimensional variable (x) indicates the multichannel EEG signals. The second (y) refers to sinusoidal reference signals, which correspond to the primary and other harmonics (usually the second and third) from the fundamental frequency of the observed stimulus [44]. This is given by
y = sin ( 2 π f k t ) cos ( 2 π f k t ) sin ( 2 π N h f k t ) cos ( 2 π N h f k t ) ;       t = 1 S , 2 S T S ,
where f k is the stimulus frequency, N h is the number of harmonics, T is the number of sampling points, and S is the sampling rate.
Then, a pair of linear combinations u = x T W x and v = y T W y , called canonical variates, are calculated by tuning the weight vectors W x and W y in order to ensure the correlation between them achieves a maximum value. Notice that each stimulus will have its weight vectors for frequency, so there is a need to calculate the canonical variates from the multidimensional variables (i.e., EEG and reference signals). Subsequently, the correlations between canonical variates (CCA coefficients) were calculated. The maximum coefficient argument of these canonical variates per frequency should correspond to the stimulus [45]. A summary of the above explanation is given by
k ^ = a r g max k ρ k ,         k = 1 , 2 , , K ,
where ρ are the CCA coefficients, K is the stimulus frequency number, and k ^ 1 ,   K is the selected stimulus.

2.4.3. Nonlinear Canonical Correlation Analysis (NLCCA)

The NLCCA method follows the same approach as the CCA base technique. The modification is found in the linear mappings of Equation (6), by nonlinear mapping functions with the use of neural networks (NNs). In Figure 4 found on the left, the mappings from x to u and y to v are represented by the double-barreled NNs [46].
Consider one dataset x i ( t ) with variables i and another dataset y j ( t ) with j variables, where each dataset has t = 1 , , n samples. Variables x i ( t ) can be grouped together to form vector x ( t ) and variables y j ( t ) can be grouped together to form vector y ( t ) . As shown previously, CCA obtains the linear combinations according to
u t = x t T · W x ,       v t = y t T · W y
where W x and W y are weight vectors that represent correlated spatial patterns corresponding to the fields of the EEG data. However, the NLCCA corresponds to u and v, and u = f ( ω x , x ) and v = f ( ω y , y ) are functions where ω x and ω y are all neural network parameters that require establishing. Therefore, a learning procedure such as the CCA approach is required to maximize the correlation between u and v.
Based on Figure 4, the networks on the left assign nonlinearly x u , y v depending on
h ( x ) = tanh [ x T W ( x ) + b ( x ) T ] ,   u = w ˜ ( x ) · h ( x ) T + b ˜ ( x ) T h ( y ) = tanh [ y T W ( y ) + b ( y ) T ] ,   v = w ˜ ( y ) · h ( y ) T + b ˜ ( y ) T
where h ( x ) and h ( y ) are the output nodes of the hidden layer and tanh [ · ] is the hyperbolic tangent function. The neural network parameters are W ( x ) and W ( y ) as the hidden layer weight matrices, b ( x ) and b ( y ) as the hidden layer bias vectors, w ˜ ( x ) and w ˜ ( y ) as the weight vectors of the output layer, and b ˜ ( x ) and b ˜ ( y ) as the bias vectors of the output layer.
Next, the Pearson correlation in the first stage of the architecture, cor u , v , between the canonical variables u and v should be maximized. The Pearson correlation is given by
r = cor u , v = n u v u v n u 2 ( u ) 2 n v 2 ( v ) 2
where n is the length of any vector u or v.
The neural network parameters were obtained by minimizing the cost function
C 1 = cor ( u , v ) + u 2 + v 2 + u 2 1 2 1 2 + v 2 1 2 1 2 + P 1 [ k i ( W k i ( x ) ) 2 + l j ( W l j ( y ) ) 2 ]
In the above expression, g denotes the sample or temporal mean, i.e., q n . The first term maximizes the correlation between the canonical variables u and v; the second and third terms represent the sum of the squared samples; the fourth and fifth terms are normalization constraints that force u and v to have a zero mean and unit variance. The sixth term is a regularization term ( L 2 -norm type) to avoid overfitting in the neural networks, and its relative magnitude is controlled by P 1 . Larger P 1 values lead to smaller weights (fewer effective model parameters), resulting in a more linear model [46].
The right-hand side of the two neural networks, shown in Figure 4, represents the inverse mappings for x ^ and y ^ with the computed canonical variables u and v.
h k ( u ) = tanh [ ( w ( u ) u + b ( u ) ) k ] , x ^ = W ˜ ( u ) h ( u ) + b ˜ ( u ) h l ( v ) = tanh [ ( w ( v ) v + b ( v ) ) l ] , y ^ = W ˜ ( v ) h ( v ) + b ˜ ( v )
To find the bias and weight parameters of the two networks on the right, the cost functions were minimized, as given by
C 2 = x ^ x 2 2 + P 2 k w k ( u ) 2 C 3 = y ^ y 2 2 + P 3 l w l ( v ) 2
where · 2 2 is the square of L 2 -norm, with L p -norm as in
L p ( e ) = ( e p ) 1 / p = i e i p 1 / p
The mean square error (MSE) is given by the first term of C 2 and C 3 in Equation (11). P 2 and P 3 correspond to penalty terms of the synaptic weights for the neural network.
However, in this study, the robust version of NLCCA was used, as presented by Cannon and Hsieh in 2008, since the previous architecture presented by Hsieh in 2001 is sensitive to overfitting. This modified robust version involved two changes. First, the section that calculates the MSE within the cost functions C 2 and C 3 are replaced by the L 1 -norm in order to calculate the mean absolute error (MAE). Second, the biweight midcorrelation (bicor) is used instead of the Pearson correlation on the C 1 cost function to measure the similarity between the canonical variables u and v, i.e., the cor u , v is replaced by bicor u , v .
The literature has shown the advantages of nonlinear CCA methods over linear CCA methods in several applications (e.g., [47,48]). However, the nonlinear methods have not been explored much, and therefore have few recent studies related to, most notably, BCI systems (e.g., [49]), where the application areas are primarily found in forecasting and climate studies [50]. In this study, the 2008 version mentioned above was used to apply NLCCA in the feature extraction of EEG–SSVEP signals.
The procedure for calculating the biweight midcorrelation and Pearson’s correlation coefficient is similar, with the exception that robust parameters replace the nonrobust parameters (i.e., covariance, mean, and expected deviation). The bicor function forecasts u from v and vice-versa, as it was calculated in the standard NLCCA model presented in Equation (17).
The biweight midcorrelation function for two vectors x and y ( bicor x , y ) is calculated, first defined as the median of the vector x ( m e d ( x ) ) and the median absolute deviation ( m a d ( x ) ); then, u i and v i are determined as [51], where
u i = x i m e d ( x ) 9 m a d ( x ) , v i = y i m e d ( y ) 9 m a d ( y )
where i is the number of items for each vector x and y.
The weights w i ( x ) and w i ( y ) are defined as
w i ( x ) = 1 u i 2 2 I 1 u i w i ( y ) = 1 v i 2 2 I 1 v i
where I represents the identity function, where
I x = 1 , if   x > 0 0 , otherwise .
Then, it is normalized to get a sum of the weights equal to one as follows:
x ˜ i = ( x i m e d ( x ) ) w i ( x ) j = 1 m x j m e d ( x ) w j ( x ) 2 y ˜ i = ( y i m e d ( y ) ) w i ( y ) j = 1 m y j m e d ( y ) w j ( y ) 2
The biweight midcorrelation function finally is determined as
δ = bicor ( x , y ) = i = 1 m x ˜ i y ˜ i
The use of bicor and MAE functions in the robust NLCCA model leads to a more stable algorithm, with improved performance and a decreasing sensitivity to overfitting [43]. These characteristics make robust NLCCA an excellent approach to analyze databases with low signal-to-noise ratios such as electroencephalographic (EEG) signals.
The cost function C 1 must be derived to implement the backpropagation algorithm; however, it is possible to use numerical approximation methods to achieve it. In order to obtain an effective tuning, the next stage is transitioned to by following an interactive process where the goal is to guarantee that all functions achieve minimum values. By inspection, C 1 should see a value near 1 , and C 2 and C 3 values near 0 are expected. The derivative of bicor x , y is represented as follows:
x i ( bicor ( x i , y i ) ) = y ˜ i ( 1 x ˜ i 2 ) w i ( x ) 4 u i 2 ( 1 u i 2 ) I ( 1 u i ) j = 1 m ( x j m e d ( x ) ) w j ( x ) 2
Depending on which path (u or v), the first part of the network involves taking, from which data is derived. Therefore, if the output v is constant, u is derived, and vice versa. Thus, the corresponding changes of the left-hand neural network parameters are taken to later verify that the cost function achieves the minimal values.
Subsequently, each stimulus passes through the trained neural networks to obtain both u and v vectors, where the bicor function is expected to obtain the highest correlation. From the nonlinear point of view, one advantage is that by maximizing the correlation, there is a more discriminating factor between those that belong to a specific stimulus and those that do not. In CCA, the linear representation that is relatively coarse brings with it differences between the incoming stimulus, resulting in little effective discrimination concerning those found in the templates.
For each template, a neural representation is necessary as it becomes the element of comparison concerning a new arrival of a stimulus, to then verify the bicor u , v through its transformation, (N.B. an approximation given by u ^ is provided). All templates have a specific v that is to be compared with a new input signal that will generate a vector u ^ , where the highest correlation indicates the possible observed stimulus, as shown in the following expression:
k ^ = a r g max k δ k ,         k = 1 , 2 , , K ,
where δ k is the correlation coefficient between u ^ and v k , as obtained from the NLCCA approach, with k = 1 , , K trained templates.

2.5. Performance Metrics

The information transfer rate introduced by [52] as a performance measure for BCIs has been preferred by researchers. The measures for calculating the ITR are related to how the user performs the task quickly and effectively, the task speed time, the performance time, the possibility of completing the task, and the selected time being involved in efficiency measures.
For a BCI system with N targets and a general classification precision P, the ITR can be defined as the amount of information reported per unit of time, which depends on both the transmission speed and the precision. In BCIs, the bit rate per trial, B (bits/trials), is expressed as
B = log 2 N + P log 2 P + ( 1 P ) log 2 1 P N 1 ,
where the transfer rate B can be translated into bits/minute, once it is multiplied by the frequency of decisions per minute T, as permitted in the BCI system shown in expression (21).
ITR = B T
The performance of the BCI system can be determined by calculating the precision ( A c c —Accuracy), and is defined as
A c c = T P + T N T P + T N + F P + F N × 100 ,
where T P and T N represent the total number of correctly detected true positive events and true negative events, respectively. F P and F N represent the total number of incorrect positive events and incorrect negative events, respectively [53].

3. Analysis of Results

This section assesses the potential of the NLCCA approach. First, the accuracy and ITR performance per user was verified by using different time windows in order to identify the most appropriate. Later, all users were grouped together, and the worst was removed (User 3) to show the average accuracy and ITR performance. This first analysis shows the general performance of each approach. The following analysis seeks to verify the time sliding window response using the best established time window. Following the flickering exposure, here, two scenarios immediately after and around 200 ms were presented in order show the transient response in VEPs. The stationary responses to measure each coefficient of variation were taken to define the robustness associated with phase shifting. Finally, an analysis of variance (ANOVA) with a post- Tukey’s test was performed, where the time sliding window for one second per approach was used to define the most suitable.

3.1. Accuracy and ITR Inspection per User

The recognition accuracy of each subject for 12 stimuli frequencies and the corresponding ITR derived from the three approaches is shown in Figure 5 and Figure 6. The red, blue, and green lines represent the CCA, Lasso, and NLCCA approaches.
The results show a clear tendency for 9 of 10 subjects to show a higher performance for the NLCCA method, followed by the CCA, and finally the LASSO method. User 3 shows inferior performance for all approaches, which is not consistent with the other results, suggesting that the user’s data contains possible errors. As a result, a decision was made to exclude this user from the analysis.
Figure 7a,b present the average accuracy and ITR for 10 subjects, while Figure 8a,b exclude user 3. These last two graphs show that the NLCCA method performs better than the other two, especially within the first 2 s; further, later, it converges with the CCA method. Based on these results, a good performance was evidenced to have been obtained with smaller time windows using NLCCA. In addition, Figure 8a,b show that in the second time window, the NLCCA approach achieves a recognition accuracy of 73 % , which is considerably higher than the 48.4 % and 25.6 % of CCA and LASSO, respectively. Here, one second is determined to be sufficient for producing a suitable performance, since the difference between the ITR results for one second and one and a half seconds are undistinguished, as checked by their variances. Besides, a short time window to reduce the fatigue is part of the objective, so one second as a reference value was chosen.
Table 1 illustrates the maximum ITR value per user, and the corresponding recognition accuracy and time window. The last two rows show the average of all the results (excluding user 3). In general, Table 1 reflects a better performance for the NLCCA approach against the others, where NLCCA obtained the highest ITR values for 8 of the 9 users (1, 2, 4–7, 9, and 10)—user 8 being the exception with only 1.1 (bit/min) below the CCA. In addition, the time taken to obtain these highest ITR values was shorter than with the other approaches, with an average of 1.28 s for NLCCA. In contrast, the time taken to obtain the highest ITR values for the CCA and LASSO corresponded to 2.39 and 2.89 s, exceeding the NLCCA by more than one second. The LASSO method obtained the lowest performance in all cases.

3.2. Transient and Steady-State Responses by the Time Sliding Windows (Phase Effects)

The Accuracy and ITR data shown below were analyzed to determine any effect when the phase is modified. Here, a one-second sliding window in steps of 3.9 ms was taken (one for each sample), and its effects with and without transient response for the average performance was taken for all users except User 3. Figure 9 and Figure 10 show the results obtained by including the transient response, i.e., taking data immediately after the flickering exposure. Figure 11 and Figure 12 show the results around 200 ms after the flickering, when the steady-state response for the average performance with all users excluding User 3 was taken, respectively.
From Figure 9, Figure 10, Figure 11 and Figure 12, the phase-shifting was observed to have produced nocive effects. It can be observed that when using the full signal (Figure 9 and Figure 10), the three methods show a drop in performance during the first 500 ms when compared to the same performance after 500 ms. However, the most affected method is identified as the CCA, since the drop in its performance is considerably greater. However, when using the signal without the first 200 ms, i.e., the transient (Figure 11 and Figure 12), the initial performance drop in all three methods is completely eliminated. Here, the initial performance is seen to remain in the same range of variation as the rest of the signal.
When comparing this behavior to previous studies, it is important to highlight the need for achieving a steady-state in order to obtain the best performance in each approach. However, the study also revealed that LASSO and NLCCA still offer some resistance to the transient response, while CCA does not. Although both ITR results evidence no significant change, NLCCA is much more suitable when the shifting immediately after the flickering exposure is ignored.
On the other hand, considering the steady-state (around 200 ms later), the phase-shifting effects are still observed. These effects produce variations in the accuracy and ITR performances, which were analyzed using the coefficient of variation. In addition, some slight differences that appear between NLCCA with and without transient responses were observed, which other approaches do not have. These differences are due to different neural networks being trained for each procedure, although the data quality facilitates a better performance. Table 2 shows the mean, standard deviation, and coefficient of variation data for each of the signals presented in Figure 11 and Figure 12. It can be seen that the mean of NLCCA is higher than the other two methods, both in precision (80–86) and ITR (72–79), and the standard deviation is similar in all three methods. This results in the coefficient of variation of NLCCA being the lowest in all cases (2.37 to 4.22), thus indicating that this method is more robust.

3.3. Identifying the Best Approach

Table 3 shows the analysis of variance performed on the obtained values, as shown in Figure 11 and Figure 12. The calculated F-value for all cases (from 44,000 onwards) was observed and, concerning the reliability of 95 % , this was found to consistently be more significant than the F-value of 3. This result indicates that there are significant differences between the three analyzed approaches.
Finally, to determine which approach these differences can be found in, and to identify the best performance, the post-Tukey’s test was used, and Table 4 shows the results. According to the table, all differences between each case exceed the DVS value, indicating significant differences between all the methods. Therefore, the NLCCA ( U 1 ) is shown to have the best performance, followed by CCA ( U 2 ), and finally LASSO ( U 3 ). In addition, these results correspond to a one-second sliding window, where a suitable enough flickering exposure to avoid fatigue was defined.
This paper presents a first analysis of the NLCCA performance compared to the CCA and LASSO approaches. The NLCCA approach was found to be more effective than the other two, especially in the first 2 s, where it reached higher ITR values, thus implying that NLCCA can obtain an acceptable result in less time. On the other hand, due to the nonlinear nature of brain signals, nonlinear methods would respond more appropriately than linear methods, which is typical of EEG signals [28]. Therefore, this paper shows the implementation of the NLCCA approach and the CCA approach, the latter being nonlinear and the most widely used in this field.
The NLCCA, based on neural networks, is characterized by being adaptive and representative of the CCA model in a nonlinear mode [46]. Its complex structure and the use of nonrobust functions aid in the performing of an overfitting process, mainly when manipulating noisy and short databases. The robust NLCCA solved these issues by using robust cost functions, explicitly replacing the inverse mapping MSE network with MAE, and changing the Pearson correlation by the biweight midcorrelation in the double barrier network. The robust NLCCA is expected to be ideal for EEG signals due to the biosignals generally being nonstationary, nonlinear, and noisy [1]. Indeed, the implementation showed that NLCCA increases their stability when the data quality exclusively includes the steady-state responses, showing an outstanding performance when the database is limited in size.

4. Conclusions

The NLCCA approach for EEG signal treatment into the SSVEP paradigm shows its potential as it offers outstanding performance, as evaluated using statistical analysis. All its results showed a better performance in terms of accuracy and ITR. Here, the average behavior for the 12 stimuli per user was evaluated. Later, the average and standard deviation in both scenarios (with and without User 3) was calculated, where NLCCA obtained the shortest time windows in almost all subjects with the best ITR; the results also showed that a one-second window was enough to obtain suitable performance. It is crucial to reduce the time window with a high ITR since this reduces user fatigue during the experimental stage.
A second experiment determined the phase effects using the time sliding window; this experiment combined an additional interaction to include the transient response into the analysis. Several studies have reported that the steady-state occurs approximately after N135, while others indicated it occurred after 200 ms [54]. Figure 9 and Figure 10 showed an initial impairment for each one of the responses. The findings of this study showed that NLCCA tries to keep a low variability as LASSO did also, but these findings are incomparable due to the significant difference in performances. The CCA was also observed to not be significantly affected by the transient response. In addition, in being able to identify the time when the flickering exposure appears in a BCI development by taking the attendance user, the different methods will show a good performance. However, with a lack of attendance control, the NLCCA seems more appropriate based on ITR performance.
On the other hand, the NLCCA performance showed that the signals in steady-state slightly improved due to the neural networks being newly trained. LASSO and CCA approaches do not have this advantage since they are deterministic. Despite this, the variations due to the phase effects are still evident with all approaches. Nonetheless, the coefficients of variation to determine the best approach were verified, identifying the NLCCA to have the lowest values when compared to the others.
Finally, the third experiment involved the one-second sliding window data to design the analysis of variance to identify differences between the approaches. The scenarios showed differences between them, so a post-Tukey’s test to define the statistical differences was used. The results helped to state that the NLCCA approach presented the best performance.
Future studies could aim to investigate the behavior of the NLCCA approach when under other operating scenarios, always focused on the objective frequency discrimination stage. The analysis of new research studies will improve validation parameters such as ITR, accuracy, and reduced time window to mitigate user fatigue as much as possible. One way to explore these issues is to use preprocessing methods to improve the signal quality before implementing NLCCA as the empirical or the variational mode decomposition.

Author Contributions

Conceptualization, D.R.D.l.C.-G., W.A.-M. and E.C.-B.; methodology, D.R.D.l.C.-G. and W.A.-M.; software, D.R.D.l.C.-G. and W.A.-M.; validation, D.R.D.l.C.-G. and W.A.-M.; formal analysis, D.R.D.l.C.-G. and W.A.-M.; investigation, D.R.D.l.C.-G. and W.A.-M.; resources, D.R.D.l.C.-G.; data curation, D.R.D.l.C.-G.; writing—original draft preparation, D.R.D.l.C.-G. and W.A.-M.; writing—review & editing, D.R.D.l.C.-G. and W.A.-M.; visualization, D.R.D.l.C.-G. and W.A.-M.; supervision, W.A.-M. and E.C.-B.; project administration, W.A.-M.; funding acquisition, D.R.D.l.C.-G. and W.A.-M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Office of the Vice-President of Research at Universidad del Valle (CI-21138) and the Department of Energy Sciences and Mechanics—DECEM of the ESPE.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank the WiCOM research group from ESPE and the Perception and Intelligent Systems research group at Universidad del Valle, who gave the financial support and the installations to perform this study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Klonowski, W. Everything you wanted to ask about EEG but were afraid to get the right answer. Nonlinear Biomed. Phys. 2009, 3, 2. [Google Scholar] [CrossRef] [Green Version]
  2. Borghini, G.; Aricò, P.; Di, F.G.; Sciaraffa, N.; Babiloni, F. Correlation and Similarity between Cerebral and Non-Cerebral Electrical Activity for User’s States Assessment. Sensors 2019, 3, 704. [Google Scholar] [CrossRef] [Green Version]
  3. Chen, Y.F.; Atal, K.; Xie, S.Q.; Liu, Q. A new multivariate empirical mode decomposition method for improving the performance of SSVEP-based brain–computer interface. J. Neural Eng. 2017, 14, 046028. [Google Scholar] [CrossRef] [Green Version]
  4. Demir, A.F.; Arslan, H.; Uysal, I. Bio-inspired filter banks for SSVEP-based brain-computer interfaces. In Proceedings of the 2016 IEEE-EMBS International Conference on Biomedical and Health Informatics (BHI), Las Vegas, NV, USA, 24–27 February 2016; pp. 144–147. [Google Scholar] [CrossRef] [Green Version]
  5. Choi, I.; Rhiu, I.; Lee, Y.; Yun, M.; Nam, C. A systematic review of hybrid brain-computer interfaces: Taxonomy and usability perspectives. PLoS ONE 2017, 12, e0176674. [Google Scholar] [CrossRef] [Green Version]
  6. Bigdely-Shamlo, N.; Mullen, T.; Kothe, C.; Su, K.; Robbins, K. The PREP Pipeline: Standardized preprocessing for large-scale EEG analysis. Front. Neuroinform. 2015, 9, 16. [Google Scholar] [CrossRef]
  7. Gabard-Durnam, L.; Méndez Leal, A.; Wilkinson, C.; Levin, A. The Harvard Automated Processing Pipeline for Electroencephalography (HAPPE): Standardized Processing Software for Developmental and High-Artifact Data. Front. Neurosci. 2018, 12, 97. [Google Scholar] [CrossRef] [PubMed]
  8. Pedroni, A.; Bahreini, A.; Langer, N. Automagic: Standardized Preprocessing of Big EEG Data. bioRxiv 2019, 200, 460–473. [Google Scholar] [CrossRef] [Green Version]
  9. Debnath, R.; Buzzell, G.; Morales, S.; Bowers, M.; Leach, S.; Fox, N. The Maryland analysis of developmental EEG (MADE) pipeline. Psychophysiology 2020, 57, e13580. [Google Scholar] [CrossRef] [PubMed]
  10. Vialatte, F.; Maurice, M.; Tanaka, T.; Yamaguchi, Y.; Cichocki, A. Analyzing steady state visual evoked potentials using blind source separation. In Proceedings of the APSIPA ASC 2010—Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, Lanzhou, China, 18–21 November 2010; pp. 578–582. [Google Scholar]
  11. Han, C.; Xu, G.; Xie, J.; Chen, R.; Zhang, S. Highly Interactive Brain–Computer Interface Based on Flicker-Free Steady-State Motion Visual Evoked Potential. Sci. Rep. 2018, 8, 5835. [Google Scholar] [CrossRef] [PubMed]
  12. Li, M.; He, D.; Li, C.; Qi, S. Brain—Computer Interface Speller Based on Steady-State Visual Evoked Potential: A Review Focusing on the Stimulus Paradigm and Performance. Brain Sci. 2021, 11, 450. [Google Scholar] [CrossRef]
  13. Kaya, I.; Bohorquez, J.; Ozdamar, O. A BCI Gaze Sensing Method Using Low Jitter Code Modulated VEP. Sensors 2019, 19, 3797. [Google Scholar] [CrossRef] [Green Version]
  14. Salomao, S.R. Desenvolvimento da acuidade visual de grades. Psicol. USP 2007, 18, 63–81. [Google Scholar] [CrossRef] [Green Version]
  15. Creel, D.J. Chapter 34—Visually evoked potentials. In Clinical Neurophysiology: Basis and Technical Aspects; Levin, K.H., Chauvel, P., Eds.; Elsevier: Amsterdam, The Netherlands, 2019; Volume 160, pp. 501–522. [Google Scholar] [CrossRef]
  16. Tsoneva, T.; Garcia-Molina, G.; Desain, P. SSVEP phase synchronies and propagation during repetitive visual stimulation at high frequencies. Sci. Rep. 2021, 11, 4975. [Google Scholar] [CrossRef]
  17. Johnson, E.; Norton, J.; Jun, D.; Bretl, T.; Jones, D. Sequential selection of window length for improved SSVEP-based BCI classification. In Proceedings of the 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Osaka, Japan, 3–7 July 2013; pp. 7060–7063. [Google Scholar] [CrossRef]
  18. Martišius, I.; Damaševičius, R. A prototype SSVEP based real time BCI gaming system. Comput. Intell. Neurosci. 2016, 2016, 3861425. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Floriano, A.; Diez, P.; Freire, T. Evaluating the Influence of Chromatic and Luminance Stimuli on SSVEPs from Behind-the-Ears and Occipital Areas. Sensors 2018, 18, 615. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Ding, J.; Sperling, G.; Srinivasan, R. Attentional modulation of SSVEP power depends on the network tagged by the flicker frequency. Cereb. Cortex 2006, 16, 1016–1029. [Google Scholar] [CrossRef] [PubMed]
  21. Salelkar, S.; Ray, S. Interaction between steady-state visually evoked potentials at nearby flicker frequencies. Sci. Rep. 2020, 10, 5344. [Google Scholar] [CrossRef]
  22. Xu, M.; Han, J.; Wang, Y.; Jung, T.P.; Ming, D. Implementing over 100 command codes for a high-speed hybrid brain-computer interface using concurrent P300 and SSVEP features. IEEE Trans. Biomed. Eng. 2020, 67, 3073–3082. [Google Scholar] [CrossRef]
  23. Muller, S.; Bastos, T.; Sarcinelli-Filho, M. Incremental SSVEP analysis for BCI implementation. In Proceedings of the 2010 Annual International Conference of the IEEE Engineering in Medicine and Biology, Buenos Aires, Argentina, 31 August–4 September 2010; pp. 3333–3336. [Google Scholar] [CrossRef]
  24. Safi, S.M.M.; Pooyan, M.; Motie Nasrabadi, A. Improving the performance of the SSVEP-based BCI system using optimized singular spectrum analysis (OSSA). Biomed. Signal Process. Control 2018, 46, 46–58. [Google Scholar] [CrossRef]
  25. Zhang, Y.; Zhang, E. Data Analytics in Steady-State Visual Evoked Potential-based Brain-Computer Interface: A Review. IEEE Sens. J. 2020, 21, 1124–1138. [Google Scholar] [CrossRef]
  26. Behroozi, M.; Daliri, M.R. A high performance steady state visual evoked potential BCI system based on variational mode decomposition. In Proceedings of the 2018 Electric Electronics, Computer Science, Biomedical Engineerings’ Meeting (EBBT), Istanbul, Turkey, 18–19 April 2018; pp. 1–4. [Google Scholar] [CrossRef]
  27. Kawala-Sterniuk, A.; Browarska, N.; Al-Bakri, A. Summary of over fifty years with brain-computer interfaces—A review. Brain Sci. 2021, 11, 43. [Google Scholar] [CrossRef]
  28. Labecki, M.; Kuś, R.; Brzozowska, A.; Stacewicz, T.; Bhattacharya, B.; Suffczynski, P. Nonlinear origin of SSVEP spectra—A combined experimental and modeling study. Front. Comput. Neurosci. 2016, 10, 129. [Google Scholar] [CrossRef] [Green Version]
  29. De la Cruz, D.R.; Alfonso-Morales, W.; Caicedo-Bravo, E. Analysis of the information transfer rate-ITR in linear and non-linear feature extraction methods for SSVEP signals. Int. J. Adv. Sci. Eng. Inf. Technol. 2021, 11, 761–768. [Google Scholar] [CrossRef]
  30. Wu, A.; Hsieh, W. Nonlinear canonical correlation analysis of the tropical Pacific wind stress and sea surface temperature. Clim. Dyn. 2002, 19, 713–722. [Google Scholar] [CrossRef]
  31. Ouali, D.; Chebana, F.; Ouarda, T. Non-linear canonical correlation analysis in regional frequency analysis. Stoch. Environ. Res. Risk Assess. 2016, 30, 449–462. [Google Scholar] [CrossRef]
  32. Yousefiankalareh, A.; Manoochehri, M.A.; Alireza Khoshnevis, S.; Surakanti, S.R. Improve the Frequency Identification in SSVEP based BCI Systems with Moving Windows Algorithm. In Proceedings of the 2019 6th International Conference on Control, Instrumentation and Automation (ICCIA), Sanandaj, Iran, 30–31 October 2019; pp. 1–5. [Google Scholar] [CrossRef]
  33. Cao, Z.; Wang, Y.K.; Hussain, F.; Al-Jumaily, A.; Lin, C.T. Effects of Repetitive SSVEPs on EEG Complexity using Multiscale Inherent Fuzzy Entropy. Neurocomputing 2019, 389, 198–206. [Google Scholar] [CrossRef] [Green Version]
  34. Çetin, V.; Ozekes, S.; Varol, H. Harmonic analysis of steady-state visual evoked potentials in brain computer interfaces. Biomed. Signal Process. Control 2020, 60, 101999. [Google Scholar] [CrossRef]
  35. Nakanishi, M.; Wang, Y.; Wang, Y.T.; Jung, T.P. A comparison study of canonical correlation analysis based methods for detecting steady-state visual evoked potentials. PLoS ONE 2015, 10, e0140703. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Zhang, Y.; Jin, J.; Qing, X.; Wang, B.; Wang, X. LASSO based stimulus frequency recognition model for SSVEP BCIs. Biomed. Signal Process. Control 2012, 7, 104–111. [Google Scholar] [CrossRef]
  37. Meier, L.; van de Geer, S.; Bühlmann, P. The group lasso for logistic regression. J. R. Stat. Soc. Ser. Stat. Methodol. 2008, 70, 53–71. [Google Scholar] [CrossRef] [Green Version]
  38. Nigham, A.; Aggarwal, V. The LPASSO Method for Regression Regularization; Techreport; Massachusetts Institute Technology: Boston, MA, USA, 2005. [Google Scholar]
  39. Tibshirani, R. Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Ser. B Methodol. 1996, 58, 267–288. [Google Scholar] [CrossRef]
  40. Schittowski, K. NLPQL: A fortran subroutine solving constrained nonlinear programming problems. Ann. Oper. Res. 1986, 5, 485–500. [Google Scholar] [CrossRef]
  41. Zerafa, R.; Camilleri, T.; Falzon, O.; Camilleri, K. To train or not to train? A survey on training of feature extraction methods for SSVEP-based BCIs. J. Neural Eng. 2018, 15, 051001. [Google Scholar] [CrossRef]
  42. Neghabi, M.; Marateb, H.; Mahnam, A. A Comprehensive Comparison Between Steady-State Visual Evoked Potentials Frequency Estimation Methods in Brain-Computer Interface with the Minimum Number of EEG Channels. Basic Clin. Neurosci. J. 2019, 10, 245. [Google Scholar] [CrossRef]
  43. Cannon, A.; Hsieh, W. Robust nonlinear canonical correlation analysis: Application to seasonal climate forecasting. Nonlinear Process. Geophys. 2008, 15, 221–232. [Google Scholar] [CrossRef] [Green Version]
  44. Nan, W.; Wong, C.; Wang, B.; Wan, F.; Mak, P.; Mak, P.I.; Vai, M. A comparison of minimum energy combination and canonical correlation analysis for SSVEP detection. In Proceedings of the 2011 5th International IEEE/EMBS Conference on Neural Engineering, Cancun, Mexico, 27 April–1 May 2011; pp. 469–472. [Google Scholar] [CrossRef]
  45. Wei, Q.; Xiao, M.; Lu, Z. A comparative study of canonical correlation analysis and power spectral density analysis for SSVEP detection. In Proceedings of the 2011 Third International Conference on Intelligent Human-Machine Systems and Cybernetics, Hangzhou, China, 26–27 August 2011; Volume 2, pp. 7–10. [Google Scholar] [CrossRef]
  46. Hsieh, W. Nonlinear canonical correlation analysis by neural networks. Neural Netw. Off. J. Int. Netw. Soc. 2001, 13, 1095–1105. [Google Scholar] [CrossRef]
  47. Hardoon, D.; Szedmak, S.; Shawe-Taylor, J. Canonical Correlation Analysis: An Overview with Application to Learning Methods. Neural Comput. 2005, 16, 2639–2664. [Google Scholar] [CrossRef] [Green Version]
  48. Wang, W.; Arora, R.; Livescu, K.; Bilmes, J. On Deep Multi-View Representation Learning. In Proceedings of the 32nd International Conference on International Conference on Machine Learning, Lille, France, 6–11 July 2015; Volume 37, pp. 1083–1092. [Google Scholar]
  49. Liu, Q.; Jiao, Y.; Miao, Y.; Zuo, C.; Wang, X.; Cichocki, A.; Jin, J. Efficient representations of EEG signals for SSVEP frequency recognition based on deep multiset CCA. Neurocomputing 2019, 378, 36–44. [Google Scholar] [CrossRef]
  50. Woldesellasse, H.; Marpu, P.; Ouarda, T. Long-term forecasting of wind speed in the UAE using nonlinear canonical correlation analysis (NLCCA). Arab. J. Geosci. 2020, 13, 962. [Google Scholar] [CrossRef]
  51. Wilcox, R.R. Introduction to Robust Estimation and Hypothesis Testing; Academic Press: Cambridge, MA, USA, 2011. [Google Scholar]
  52. Wolpaw, J.R.; Birbaumer, N.; Heetderks, W.J.; Mcfarland, D.; Peckham, P.; Schalk, G.; Donchin, E.; Quatrano, L.A.; Robinson, C.; Vaughan, T. Brain-computer interface technology: A review of the first international meeting. IEEE Trans. Rehabil. Eng. 2000, 8, 164–173. [Google Scholar] [CrossRef] [PubMed]
  53. Bajaj, V.; Pachori, R. Classification of seizure and nonseizure EEG signals using empirical mode decomposition. IEEE Trans. Inf. Technol. Biomed. 2012, 16, 1135–1142. [Google Scholar] [CrossRef] [PubMed]
  54. Yoshimura, N.; Itakura, N. Usability of transient VEPs in BCIs. In Recent Advances in Brain-Computer Interface Systems; Fazel-Rezai, R., Ed.; IntechOpen: Rijeka, Croatia, 2011; Chapter 6. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Stimulation of the inverted pattern of the SSVEP signal [14].
Figure 1. Stimulation of the inverted pattern of the SSVEP signal [14].
Sensors 21 05308 g001
Figure 2. Distribution of the capture time of the SSVEP signal.
Figure 2. Distribution of the capture time of the SSVEP signal.
Sensors 21 05308 g002
Figure 3. Electrode placement during experimentation.
Figure 3. Electrode placement during experimentation.
Sensors 21 05308 g003
Figure 4. Neural Network Architecture for NLCCA [43].
Figure 4. Neural Network Architecture for NLCCA [43].
Sensors 21 05308 g004
Figure 5. Comparison of accuracy for each subject between CCA, LASSO, and NLCCA.
Figure 5. Comparison of accuracy for each subject between CCA, LASSO, and NLCCA.
Sensors 21 05308 g005
Figure 6. Comparison of ITR (information transfer rate) for each subject between CCA, LASSO, and NLCCA.
Figure 6. Comparison of ITR (information transfer rate) for each subject between CCA, LASSO, and NLCCA.
Sensors 21 05308 g006
Figure 7. Comparison of (a) average accuracy and (b) average ITR between CCA, LASSO, and NLCCA.
Figure 7. Comparison of (a) average accuracy and (b) average ITR between CCA, LASSO, and NLCCA.
Sensors 21 05308 g007
Figure 8. Comparison of (a) average accuracy and (b) average ITR (excluding User 3) between CCA, LASSO, and NLCCA.
Figure 8. Comparison of (a) average accuracy and (b) average ITR (excluding User 3) between CCA, LASSO, and NLCCA.
Sensors 21 05308 g008
Figure 9. One-second sliding window for (a) Accuracy and (b) ITR.
Figure 9. One-second sliding window for (a) Accuracy and (b) ITR.
Sensors 21 05308 g009
Figure 10. One-second sliding window for (a) Accuracy and (b) ITR removing User 3.
Figure 10. One-second sliding window for (a) Accuracy and (b) ITR removing User 3.
Sensors 21 05308 g010
Figure 11. One–second sliding window (trimmed 50 first data—200 ms) of (a) Accuracy and (b) ITR.
Figure 11. One–second sliding window (trimmed 50 first data—200 ms) of (a) Accuracy and (b) ITR.
Sensors 21 05308 g011
Figure 12. One–second sliding window (trimmed 50 first data–around 200 ms) of (a) Accuracy and (b) ITR removing User 3.
Figure 12. One–second sliding window (trimmed 50 first data–around 200 ms) of (a) Accuracy and (b) ITR removing User 3.
Sensors 21 05308 g012
Table 1. Maximum ITR value for each user related to each method, and the corresponding accuracy and time window values.
Table 1. Maximum ITR value for each user related to each method, and the corresponding accuracy and time window values.
LASSOCCANLCCA
SubjectTWAccITRTWAccITRTWAccITR
(s)(%)(bit/min)(s)(%)(bit/min)(s)(%)(bit/min)
S1334.175.7246014.761.557.527.14
S2479.1725.512.588.3345.63172.553.55
S30.550.48477.524.453.5252.38
S42.571.6729.91393.3345.01281.6745.27
S5390.8342.39297.566.619083.1
S6468.3319.07292.558.82195.8395.73
S7281.6745.27298.3368.1186.6776.72
S8283.3347.1729055.41.581.6754.32
S9299.1769.731.599.16783.68195.8395.73
S103.571.6723.262.593.3351.451.59066.48
Avg.2.6568.5030.852.5588.9951.391.577.6760.04
Avg. (w/o S3)2.8975.5634.232.3990.2854.381.2883.5266.45
Table 2. Mean and standard deviation of the offset plot.
Table 2. Mean and standard deviation of the offset plot.
LASSOCCANLCCA
MeanSTDCV (%)MeanSTDCV (%)MeanSTDCV (%)
Accuracy10 user37.201.945.2164.443.004.6680.542.372.95
9 user40.192.125.2868.413.154.6186.142.372.75
ITR10 user17.951.9110.6447.544.168.7472.583.945.43
9 user19.912.1210.6551.954.508.6779.654.225.30
Table 3. Analysis of Variance (ANOVA).
Table 3. Analysis of Variance (ANOVA).
CaseSource
Variability
GDLSum
Squares
Mean
Squares
FcF (95%)
Accuracy (10 users)Treatments2690,084.705345,042.3556,214.053
Error215413,221.276.138
Total2156703,305.98
56,214.05 > 3
Accuracy (9 users)Treatments2772,312.22386,156.1157,822.383
Error215414,385.16.68
Total2156786,697.31
57,822.38 > 3
ITR (10 users)Treatments21,075,514.163537,757.0844,268.263
Error215426,166.1212.15
Total21561,101,680.28
44,268.26 > 3
ITR (9 users)Treatments21,285,243.27642,621.63545,287.1863
Error215430,565.0914.19
Total21561,315,808.36
45,287.186 > 3
Table 4. Post-Tukey’s test for Accuracy performance.
Table 4. Post-Tukey’s test for Accuracy performance.
CaseHypothesis | X i X j | Coefficient DVS c Comparison
q (5%) 3,2154 q CM d n
Accuracy (10 users) H 0 : U 1 = U 2 27.243.320.30627.24 > 0.306
H 0 : U 1 = U 3 43.3443.34 > 0.306
H 0 : U 2 = U 3 16.116.1 > 0.306
Accuracy (9 users) H 0 : U 1 = U 2 28.223.320.3228.22 > 0.32
H 0 : U 1 = U 3 45.9545.95 > 0.32
H 0 : U 2 = U 3 17.7317.73 > 0.32
ITR (10 users) H 0 : U 1 = U 2 29.593.320.43129.59 > 0.431
H 0 : U 1 = U 3 54.6354.63 > 0.431
H 0 : U 2 = U 3 25.0425.04 > 0.431
ITR (9 users) H 0 : U 1 = U 2 32.043.320.46632.04 > 0.466
H 0 : U 1 = U 3 59.7459.74 > 0.466
H 0 : U 2 = U 3 27.727.7 > 0.466
The mean difference is greater than the DVSc value in all cases. Therefore, all null hypotheses are rejected.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

De la Cruz-Guevara, D.R.; Alfonso-Morales, W.; Caicedo-Bravo, E. Solving the SSVEP Paradigm Using the Nonlinear Canonical Correlation Analysis Approach. Sensors 2021, 21, 5308. https://doi.org/10.3390/s21165308

AMA Style

De la Cruz-Guevara DR, Alfonso-Morales W, Caicedo-Bravo E. Solving the SSVEP Paradigm Using the Nonlinear Canonical Correlation Analysis Approach. Sensors. 2021; 21(16):5308. https://doi.org/10.3390/s21165308

Chicago/Turabian Style

De la Cruz-Guevara, Danni Rodrigo, Wilfredo Alfonso-Morales, and Eduardo Caicedo-Bravo. 2021. "Solving the SSVEP Paradigm Using the Nonlinear Canonical Correlation Analysis Approach" Sensors 21, no. 16: 5308. https://doi.org/10.3390/s21165308

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop