Next Article in Journal
Optimization of Kernel Type and Sharpness Level Improves Objective and Subjective Image Quality for High-Pitch Photon Counting Coronary CT Angiography
Previous Article in Journal
Impact of the FilmArray Rapid Multiplex PCR Assay on Clinical Outcomes of Patients with Bacteremia
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hybrid Deep Learning Approach for Stress Detection Using Decomposed EEG Signals

1
Department of Computer Science Engineering-AI & ML, Siliguri Institute of Technology, Siliguri 734009, India
2
School of Computing Science and Engineering, Vellore Institute of Technology Bhopal University, Bhopal 466114, India
3
Department of Computer Science and Engineering, National Institute of Technology, Patna 800001, India
4
Department of Civil Engineering, Roorkee Institute of Technology, Roorkee 247667, India
5
Department of Computer Science and Engineering, Pandit Deendayal Energy University, Gandhinagar 382426, India
6
Department of Civil and Environmental Engineering, Incheon National University, Incheon 22022, Republic of Korea
7
Incheon Disaster Prevention Research Center, Incheon National University, Incheon 22022, Republic of Korea
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(11), 1936; https://doi.org/10.3390/diagnostics13111936
Submission received: 16 April 2023 / Revised: 14 May 2023 / Accepted: 26 May 2023 / Published: 1 June 2023
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
Stress has an impact, not only on a person’s physical health, but also on the ability to perform at the workplace in daily life. The well-established relation between psychological stress and its pathogeneses highlights the need for detecting psychological stress early, in order to prevent disease advancement and to save human lives. Electroencephalography (EEG) signal recording tools are widely used to collect these psychological signals/brain rhythms in the form of electric waves. The aim of the current research was to apply automatic feature extraction to decomposed multichannel EEG recordings, in order to efficiently detect psychological stress. The traditional deep learning techniques, namely the convolution neural network (CNN), long short-term memory (LSTM), bidirectional long short-term memory (BiLSTM), gated recurrent unit (GRU) and recurrent neural network (RNN) models, have been frequently used for stress detection. A hybrid combination of these techniques may provide improved performance, and can handle long-term dependencies in non-linear brain signals. Therefore, this study proposed an integration of deep learning models, called DWT-based CNN, BiLSTM, and two layers of a GRU network, to extract features and classify stress levels. Discrete wavelet transform (DWT) analysis was used to remove the non-linearity and non-stationarity from multi-channel (14 channel) EEG recordings, and to decompose them into different frequency bands. The decomposed signals were utilized for automatic feature extraction using the CNN, and the stress levels were classified using BiLSTM and two layers of GRU. This study compared five combinations of the CNN, LSTM, BiLSTM, GRU and RNN models with the proposed model. The proposed hybrid model performed better in classification accuracy compared to the other models. Therefore, hybrid combinations are appropriate for the clinical intervention and prevention of mental and physical problems.
Keywords:
EEG; DWT; CNN; LSTM; BiLSTM; GRU

1. Introduction

Human life today is not as simple as it once was. According to a recent study, the greatest impact on routine life is on working professionals aged 25 to 40 [1]. Stress has been revealed to be a silent killer for the human brain. Stress is the root cause of every mental problem, and is due to various physical or emotional states in the human body. Work pressures to meet deadlines, financial crunches, job dismissals, unemployment and various corporate demands contribute to increasing stress levels. The human brain progresses to stressors in order to maintain balance in the nervous system [1]. If professionals are under stress for an extended period, their performance suffers, and they become distressed. This stress seems to have negative impacts on the human body, causing diseases such as insomnia, decreased immunity, infections, cervical impairments and migraines, and so on [1]. A person is unaware of the stress that is silently killing his/her mind. Scenario and strategic management from health care institutions, digital technologies for stress prediction and entrepreneurship development on the national medical services market could boost a population’s quality of life and its nation’s human potential [2,3,4].
The main entities that reflect stress are the following: human body temperatures, brain activity and eye blinks [1]. As stated by a World Health Organization (WHO) survey, depression is one of the leading causes of disability worldwide [5]. In accordance with India’s national mental health survey (2015–2016), held by the National Institute of Mental Health and Neuro Sciences (NIMHANS) [2], the population diagnosed with mental illness has increased from 7.5 percent in 2014 to 10.6 percent in 2016. The ratio of patients with doctors in the low- and middle-income classes is being threatened [5]. In India, more than 150 million people are suffering from different mental illnesses such as anxiety, depression and other personality disorders. Stress has also been made worse by the COVID-19 pandemic’s effects on people’s lives [6]. These mental illnesses are in desperate need of mental health care; however, there is a treatment gap ranging from 74% to 90% for such services.
There are several approaches to record/collect the human stress levels. A phonocardiography (PCG) signal can be obtained using an electronic stethoscope, which can be utilized as a valuable diagnostic tool in rural locations, with babies, and for homecare purposes [5]. Electroencephalography (EEG), electrooculography (EOG), electromyography (EMG) and electrocardiography (ECG) are the four methods that are utilized most frequently for the purpose of recording physiological signals in response to generated approaches. Photoplethysmography (PPG) also plays a critical role in the collection of physiological signals [7]. Conferring with the deferent literature, bio-chemical and bio-logical-based methodologies have produced contradictory results; these were attributed to hormone instability.
Mental stress is an important belief that is slowly gaining attention in different research fields related to neuroscience, psychology, medicine and other fields such as sentiment computing. As a result, it is critical to investigate stress using a variety of methods, such as EEG data. Signals such as EEGs are extremely effective at revealing correlations between various rhythmic signals [8]. Due to the scarcity of professional automation and semi-automation, the study of different multimodal signals such as EEGs/ECGs is critical [8]. Big data and the Internet of Medical Things have made it more important than ever to diagnose, detect, and treat mental illnesses [9,10]. Consequently, more than 45 percent of high school scholars are stressed, which has an undesirable impact on their learning performance [11,12]. Stress, on the other hand, manifests itself in a variety of ways [13]. As a consequence of this, it is absolutely necessary to research the effects of stress, utilizing individual EEG signals and the data on the human brain’s bioelectrical transmissions [14,15]. Based on the EEG data obtained from the scalp, it is possible to examine the electrical activities of the brain [16,17]. Additionally, electroencephalography has developed into a crucial non-invasive method for gauging brain activity that can detect anomalies, abnormalities, and mental illnesses [18,19]. The EEG signal has been widely utilized to identify and analyze human stress [20], particularly in the frontal lobe [21]. Many recent studies have focused on using EEG signals to detect and diagnose mental stress, as well as the link between frontal lobe EEG alpha-bands and emotional states activity [22,23].
Generally, EEG signals are typically electrical recordings that are random, non-stationary, non-correlated and non-linear in character. Therefore, the proper diagnosis of disease from EEG signals requires advanced signal processing tools to identify the brain’s rhythms [24]. There are four types of features of extraction techniques for raw EEG signals: (i) time-domain-based, (ii) frequency domain-based, (iii) time–frequency domain-based, and (iv) spatial-time–frequency domain-based [24]. Table 1 demonstrates the feature extraction methods for classification procedures. Frequency and temporal investigations can improve EEG studies [4]. This research applied a time–frequency analysis technique for EEG signals called discrete wavelet transform (DWT) analysis. The DWT technique is useful because it can convert fluctuating EEG signals into more stable linear ones [25,26,27].

1.1. Machine Learning/Deep Learning for Classification of EEG Signal

After successful extraction of the features from an EEG signal, machine learning (ML)/deep learning (DL) models are used to classify the EEG Signals. ML/DL models can be able to learn automatically from data without any human interaction [28,29,30,31,32,33,34,35,36].
Recently, DL techniques have increasingly been used for the analysis of EEG signals, due to their remarkable characteristics [34]. A DL model can be an artificial neural network (ANN) with multiple hidden layers, such as CNN, RNN, etc.. DL models use multiple layers of neural connections for extracting different features from the data and progressively improving the accuracy of the results. These models showed efficient results in EEG signals classification for several disease diagnoses [27,34,36,37]. In addition, DL algorithms perform better than typical ML models when it comes to the classification of EEG data (brain signals) [24]. Table 1 shows the approaches of previous research in the analysis of EEG signals.
The previous research used various ML approaches, such as naïve Bayes (NB) and support vector machine (SVM), to detect and classify stress conditions using EEG signals [38], and proposed different stress classification techniques. The improved Elman neural network (IENN) was used to develop a stress detection system in [39]. A cognitive design of an autonomously intelligent agent implemented an ANN [40].
Studies of EEG signals routinely employ a wide variety of ML models, including SVM, neural networks (NN), k-nearest neighbor (KNN), stochastic gradient decent (SGD) and linear regression (LR), multilayer perceptron (MLP), random forest (RF) and fuzzy logic (FL) [41,42,43].
Table 1. Research analysis of previous models using ML and DL.
Table 1. Research analysis of previous models using ML and DL.
ClassifierVolunteers/SubjectsFeature Engineering (Domain)Pros. of ClassifierCons. of ClassifierAccuracy
SVM15 volunteers [41]Correlation analysis
(Time)
Works effectively when classes are well-separatedUnsuitable for large data sets86.94%
MLP33 subjects,
eyes open and closed conditions [44]
Neuro-physiological
Features
(Time)
More efficient on non-linear dataClassification task computation are complex and consuming time85.20%
SVM6 subjects’ EEG dataset [45]Hilbert Huang Transform
(Time-Frequency)
-Performs poorly when target classes overlap due to noise.89.07%
LR4 EEG channels features of 27 subjects [38]Band power (Frequency)Works well when data are linearly separable.Requires average or no independent variable multi-collinearity98.76%
SVM17 patients were taken from subjects [46]NIL--90.0%
SVM34 patients were taken from subjects [47]Band Power (Frequency)-Needs extensive testing such as cross validation85.0%
NB48 practice patterns were taken from subjects [48]DWT
(Time and frequency)
Process high-dimensional data efficientlyNB struggles to predict minorities class data91.60%
DL Network32 samples were taken from the subjects [49]Power spectrum density
(Frequency)
The model learns relevant features without manual feature engineering.The training data, and model’s performance can decline in diverse hardware resources.53.42%
RF17 scalp patients and 10 intracranial were taken from subjects [50]DWT
(Time and frequency)
It automatically selects a subset of characteristics at each split, reducing the causes of dimensionality and irrelevant features.Low-cardinality features may be less important or require preprocessing to avoid bias.62.00%
FL19 patients were taken from subjects [51]Band Power
(Frequency)
It permits the formulation of rules that account for varying degrees of uncertainty and exceptionsComplex fuzzy systems demand more processing and memory, making them unsuitable for real-time or resource-constrained applications.91.80%
KNN32 healthy subjects only [52]DWT
(Time and frequency)
It detects linear and non-linear data relationshipsStruggles with class imbalances95.69%
Long Short-Term Memory (LSTM)32 EEG channels of 32 subjects [53]Band power
(Frequency)
Successfully captures long-term relationships and can alleviate the vanishing gradient issue that is typical in standard RNNs, making training and optimization simpler.It is susceptible to overfitting, especially when the model has a high number of parameters and the training data are restricted.94.69%
(Bidirectional Long Short-Term Memory) BiLSTM-LSTM14 EEG channels of 48 subjects [54]Power spectral density
(Frequency)
Processes data in both directions, and is able to properly capture past and future contexts of the dataRequires an extensive amount of data to train successfully, this may pose a problem if there are not enough labelled data.97.80%
LR, NN, RNN1488 abnormal, 1529 normal patients were taken from subjects [55]Raw EEGRNNs can learn context from previous inputsRNNs face vanishing gradient problemRNN achieve3.47% more
VGG16-CNN 16 and 19 channels were taken from 45 and 28 subjects, respectively [35]Continuous Wavelet Transform
(Time-Frequency)
VGG16 can be used as a feature extractor or as a starting point for transfer learningVGG16 requires more computation compared other more streamlined CNN architectures such as ResNet or Inception.98%
2D-CNN-LSTM5 channels are taken from 60 subjects [36]Raw EEGIn hybrid model, CNNs are powerful in automatically learning hierarchical features from input data and LSTM networks; on the other hand, can handle temporal variations and long-term dependencies in sequential data.It is susceptible to overfitting, especially when the model has a high number of parameters and the training data are restricted.72.55%
CNN + LSTM60 channels features of brain EEGs were taken from 54 subjects [27]Fuzzy Entropy and fast Fourier transform (Frequency)Hybrid DL model performed well on high dimension dataHybrid DL Model takes long computation time for training 99.22%
Figure 1 shows how researchers collect, analyze, and classify brain signals using EEG signal analysis. The steps are as follows: data capture, preprocessing, feature extraction and categorization [56].

1.2. Research Contributions

This study presents a novel hybrid deep learning approach for stress detection. The simultaneous task EEG workload (STEW) dataset was used [57], and an effective technique called DWT for frequency band decompression and noise removal from raw EEG signals was utilized. DWT delivers reliable frequency and timing information at low and high frequencies. Hence, the DWT is ideal for asymmetrical data analysis [58,59]. Decomposed EEG signals are taken as the input to a CNN-based automatic feature selection technique. For the classification of stress levels, a hybrid combination of deep learning models called LSTM, BiLSTM, two layers of Gated Recurrent Unit (GRU) and RNN were applied to the classification of human stress. The hybrid combination of these techniques provided improved performance, and could efficiently handle long-term dependencies in non-linear brain signals. The proposed hybrid DL model, in contrast to more traditional methods of anomaly identification, attained efficient accuracy.

2. Approaches and Data Description

This section describes the EEG dataset used. It introduces DWT, which was used for signal de-noising and decomposing; CNN, which was used for automatic feature extraction; and RNN, LSTM, BILSTM and GRU, which are briefly introduced.

2.1. Dataset Descriptions

In this research, a STEW, simultaneous task EEG workload [57] dataset, was used. This dataset consisted of total 48 subjects. A commercial psychological test single-session simultaneous capacity (SIMKAP) experiment was performed, and the EEG signal activity was evaluated with MATLAB EEGLAB toolbox [60]. An emotive EPOC (high resolution, multi-channel, wireless neuroheadset) EEG device was used for EEG data collection, with 128 Hz as the sampling frequency. According to the 10–20 international system, the device had fourteen electrodes located at AF3, F3, F7, FC5, T7, P7, O2, O1, T8, P8, FC6, AF4, F4 and F8. This research considered only a SIMKAP experiment based on subjects’ ratings on a scale of 1–9. In reality, the inspection process as a whole was a form of validation for the participant, who faced a greater burden while performing the test. A 1 to 3 rating was a low burden, 4 to 6 was a moderate burden and 7 to 9 was a high burden. The EEG recordings consisted of a total of fourteen channels. Figure 2 depicts the positions of the electrodes, according to the 10–20 international system.
Before starting any analysis, it was critical to process the raw EEG signal data to remove the artefacts that were caused by muscle movement, and to clear the data of noise. More details of the dataset description are presented in [57].
The general steps were as follows:
(a)
Apply a 1 Hz high-pass filter on the raw data.
(b)
Eliminate the line noise.
(c)
Carry out artifact subspace reconstruction (ASR).
(d)
Re-assign data to the average.

2.2. Discrete Wavelet Transform

The wavelength techniques were already fulfilled the signal decomposing along with de-noising significantly. The transformation coefficients could be estimated to the initial signal [58]. Wavelets can be used to classify the neighborhood features of the signals in both the frequency and time domains. The frequency domain generates low-frequency wavelets to compare to the large-scale time domain [61]. The continuous wavelet transform (CWT) of signal x(n) is enumerated as follows:
WT x ( a , τ ) = 1 a   x ( n ) ψ n τ a δ n
where a is the scale displacement, τ is the time displacement, and ψ(i) indicates a wavelet basis function. EEG signals are discrete signals; for this reason, DWTs are essential requirements of discrete wavelets. In comparison to the CWT, the DWT restricts the a and τ from the wavelet basis function ψ(a,τ) to different discrete points, i.e., the scale and displacement are discretized. The discrete wavelet basis function is expressed as ψ(j,l) (n) = 2^(−i/(2)) ψ(2^(−j) n−l) where, j ∈ Z, l ∈ Z indicates DWT.
WT x j , l = x n ψ * n δ t
In DWT, the scaling function brings off both the low- as well as the high-pass filter. The procedure of the DWT workflow is shown in Figure 3, where approximation coefficients have a low-pass frequency resolution but a high time resolution, whereas the detail coefficient has the reverse condition [61]. The wavelets of the four-level EEG signals are decomposed using low-pass (LP) and high-pass (HP) filter coefficients (Figure 3), which are detail coefficients (D1, 30–65 Hz; D2, 14–30 Hz; D3; 8–14 Hz; D4; 4–8 Hz, Theta) and approximation coefficients (A1: 0–32 Hz; A2: 0–16 Hz; A3: 0–8 Hz; A4: 0–4 Hz, Delta) [9].

2.3. Convolutional Neural Network (CNN)

The CNN structure can mimic the activity of the human brain’s composite cerebral cortex. To train a multiplex model, it predicts based on a large training dataset that uses many algorithms, such as back propagation and gradient descent optimization, to find out effective features. It utilizes a multiple number of filtering techniques, non-linear activation and normalization methods to extract different important features [62,63].
The recommended simple CNN input layer is involved by the convolutional layer, which then passes the result to the next layer. The filter application and the feature extraction properties in the convolutional layers act as an input signal [64]. Each sub-sample input layer minimizes its dimension to reduce different number of parameters. It learns how to reduce calculation costs by utilizing standard discretization max-pooling-1D blocks. The flattened layer is utilized for traditional multidimensional data to flatten out. In the classification process, dropout layers prevent the loss of validity by normalizing and improving the neural network over-fitting problem.

2.4. Recurrent Neural Networks (RNN)

RNN are powerful and robust in nature, and consist of recurrent networks along with internal memory. Since RNN weights are considered for both the input and looping back output signals, these types of weights are adjusted with the help of gradient descent or back propagation [64] algorithm. The deficit in RNN is long-term dependencies [65], while the LSTM can solve this problem due to the design of its repeating module. RNN are time delay networks with training complexity issues, which is a key problem, because at each back propagation step of computation there is gradient loss. Therefore, the LSTM model may be utilized in place of the RNN model without having these side effects [66].

2.5. Long Short-Term Memory (LSTM)

The LSTM network presents a unique configuration known as a memory cell [64,67]. This memory cell consists of four major components: a neuron, a forget gate, an input gate, and an output gate with a self-recurrent structure (Figure 4). The ability of cells to store and access information for longer durations is supported by these gates.
The hidden states are calculated by the LSTM network using the following equations:
i = σ x n U i + s n 1 W i
f = σ x n U f + s n 1 W f
o = σ x n U + s n 1 W
g = tanh x n U g + s n 1 W g
c n = c n 1 f + g i
s n = tanh c n o
y = softmax Vs n
Traditionally, neural networks are used, such as feedforward neural networks as well as recurrent neural networks. In essence, a feedforward neural network is an ANN in which the output of any layer does not affect the overall performance of that same layer, i.e., there is no cycle formed by the connections between the two units. However, feedforward networks are processed to the network by both the input and output layers.

2.6. Bidirectional LSTM

The BiLSTM learning technique is a series of processing models that includes two LSTM networks: the first one acts in a forward direction, and second one in a backwards direction [68,69]. BiLSTMs effectively increase the amount of information available to the network, giving the algorithm more context. Figure 4 shows a BiLSTM model architecture which consists of forward, backward and hidden layers. In the BiLSTM model (Figure 5) architecture, σ is the activation function for the layers, and x and y are the input and output, respectively.

2.7. Gated Recurrent Units (GRU)

Traditional RNN architectures suffer from vanishing and rising gradients [70]; this makes optimization challenging, and prevents the networks from learning long-term dependencies. To address this issue, several RNN modifications have been proposed, the most popular of which are long short-term memory units (LSTMs) [71,72]. In comparison to LSTMs, GRUs are easier to implement, require fewer parameters, and have better performance in a number of scenarios [73]. Figure 6 shows the structure of a GRU cell
Where reset α t , update ( β t ) and output h t are gates at time t. outputs of gates   ĥ t   and ĥ t 1 are the output at times t and t − 1, respectively, Input   ( X t )   is at time t, and the activation functions ( σ ,   tanh ) . W α ,   W β ,   W ĥ   and   W o are the weights for the input and output of the models. y ^ t   shows the output of the training sample at time t. The calculation process of the memory unit is expressed by Equations (10)–(14).
α t = σ W α ĥ t 1 , X t
β t = σ W β ĥ t 1 , X t
h t = tan h W ĥ α t * ĥ t 1 , X t
ĥ t = 1 β t * ĥ t 1 + z t * h t
ŷ t = σ W o ĥ t

3. The DWT-Based Hybrid DL Models

This section describes the proposed methods with a flow diagram and the parameter settings. These experiments were performed using a Python rich library tool [74], using a hardware configuration of RAM-32GB, i9 and CUDA-GPU. The following describes the steps of the proposed methodology.
Step 1: Feature extraction from EEG signals
The DWT is a familiar tool used to remove noise, as well as for feature extraction from the signals. The wavelet decomposition of a noisy signal emphasizes essential signal information in a few large absolute-valued wavelet coefficients, without changing the random noise distribution [75]. In this research, each subject has fourteen EEG channels. The signal is extracted using DWT from the EEG dataset, and signals are decomposed in four levels with Daubechies (dB4) wavelet function. Order 4 of the Daubechies wavelet function performs better than the other orders [76]. The de-noising method is used with a free distributed hypothesis test threshold (FDR), tunable oscillatory behavior (Q-factor) with a value of 0.05; the threshold rule is hard and independent of the noise level. Figure 7 shows a decomposed signal in frequencies called Alpha, Beta, Gamma, Delta and Theta.
Step 2: CNN and max pooling layer description
The CNN selects features from decomposed EEG signals in the model. In signal processing, the CNN, a deep learning subset, has gained attention [77]. In the model, one CNN-1d layer is continuously used with filter size ‘128′, kernel size 1, padding set as ‘valid’, and activation set as ‘Softmax’. These filters were used as inputs to the next layer. For the purpose of feature extraction from input signals, the ‘Softmax’ activation function was employed to describe a probability distribution over an n-valued discrete signal with kernel size 1. For the input signals to be entirely covered by the filter, it was assumed that all dimensions are accurate. The CNN layer receives pre-processed EEG signals as the input, and the max-pooling filter acts as a window through which only the highest score is selected for the output. The hit-and-trial method was used to select all of the learning parameters.
Step 3: Different hybrid combinations for stress classification
In this research, six hybrid combinations with the BiLSTM, RNN, LSTM and GRU models were developed for classifying human stress levels. The hybrid combination of DL models may provide improved performance, and can handle long-term dependencies in non-linear brain signals. The hybridization of DLs leverages the strengths of different DL models, and removes the limitations of single DL models. It enhances the performance efficiency for EEG signal classification tasks [27,35,36]. Figure 8 shows the six hybrid model combinations. The described CNN layers functioning in step 2 were same for all combinations. In each of the first three combinations (Figure 8), the BiLSTM layer was followed by two GRU layers (CBGG) or two LSTM layers (CBLL), or two RNN layers (CBRR). Later, three were CNN–RNN, CNN–LSTM and CNN–GRU. Table 2 represents each hybrid model’s parameter description. In order to solve the issue of over-fitting that arises during the learning process, a dropout unit rate was set as 0.2. A single neuron with a sigmoid activation function was employed to feed information into the last dense layer. The hit-and-trial method was utilized throughout the entire process of configuring the parameters. The model’s objective function was set as “binary_crossentropy” and the optimizer was “Adam”. The other fitting parameters are epoch (200) and batch_size (50) were chosen based on the hit-and-trial approach.

4. Results and Analysis

This section presents findings based on the matrices, convergence and receiver operating characteristic (ROC) curves, which are provided for a visual comparison of the hybrid models.

4.1. Metrics-Based Performance

In this study, the data were classified into two classes, the stress state and the relaxed state. The binary classification was evaluated on the basis of confusion matrix parameters. The parameters’ names and formulations are specified in Table 3.

4.2. Performance Evaluation

The performance of the proposed hybrid DWT-based CBGG model was compared with the CBRR, CBLL, CNN–RNN, CNN–LSTM and CNN–GRU models to prove its better efficiency. The STEW [48] dataset was used to prove the usability of the hybrid models. The dataset had 14 channels and 921,600 (128 sampling frequency × 150 s recoding time × 48 participants) data points per channel, with 70 percent used for instruction and 30 percent used for testing. The training data were likewise divided 70:30 for the purposes of validating and testing the performance of the models.
The proposed CBGG model’s outcomes were evaluated using the following metrics: accuracy, sensitivity, F-score, specificity, precision, +LR and −LR. Table 4 represents a comparison of the models based on stated parameters. The proposed model outperformed the existing models, and achieved highest accuracy of 98.10%. The CBRR and CBLL models showed better efficiency compared to the CNN–RNN, CNN–LSTM and CNN–GRU hybrid models. Furthermore, the performance metrics scores for the sensitivity (98.08%), F-score (98.20%), specificity (97.76%), precision (98.27%), +LR (44) and −LR (0.02) of the proposed CBGG model were higher compared to those of the other models. Figure 9 shows a graphical representation of confusion matrix parameters over the proposed hybrid model in comparison with the other models.

4.3. Convergence Curve Analysis

A convergence curve for both the training and validation phases can be thought of as a representation of the optimum possible value for a learning parameter in relation to the accuracy, with regard to the loss function. Figure 10a–f show the training accuracy curve and validation accuracy curve traces for the hybrid CNN–RNN, CNN–LSTM, CNN–GRU, CBRR, CBLL and CBGG models, respectively (200 epochs). The proposed model’s (CBGG) (Figure 10f) train and validation accuracy showed a faster convergence rate compared to that of the other developed models, with a higher accuracy of 98.10%.

4.4. Receiver Operating Characteristic (ROC) Curve Analysis of Models

The ROC curve is a measurement of how well a model distinguishes between the stressed and relaxed states. Since there were so many inconsistent data points in the collection, the performance indicator based on confusion was insufficient. It was possible to determine the ratio of accurate positives to erroneous ones using an ROC curve, which was applied to the test data. On the X-axis, the false positive rate was plotted, while the real positive rate was plotted on the Y-axis. Figure 11a shows ROC curves of the CNN–RNN, CNN–LSTM and CNN–GRU models, and Figure 11b shows the ROC curves for the CBRR, CBLL and CBGG models. In Figure 11b, the ROC show better efficiencies for the CBRR, CBLL and CBGG models compared to the other three hybrid models in Figure 11a. The proposed model (CBGG), which includes a larger ROC covered area (area under curve), indicates that it performed better than the CBRR and CBLL models.

4.5. Comparison of Proposed and Existing Works

This section compares the proposed research with existing machine learning models built using EEG channels and feature selection methods on the same dataset. The detection of relaxation and tension from EEG signals has also been the subject of numerous studies in the literature. Table 5 shows a comparison of the CBGG’s performance with those of existing models.

4.6. Validation of Proposed Model

To provide quantitative statistical results, predictive models were constructed on a dataset and then evaluated using resampling techniques. Lastly, statistical analysis was carried out to select the ideal model. The average classification accuracy in this study, based on training and testing the CBGG model using a stratified 10-fold cross-validation method, was 97.60%.

4.7. Limitations of the Proposed Hybrid DL Models

The main limitation of DL-based hybrid models is the complexity due to significant parameter tuning, which required higher running time. Furthermore, the selection of different hybrid combinations was a tedious task in proving the validity of the models on the EEG dataset.

5. Conclusions

This study developed DWT-based hybrid deep learning models used for the classification of stress using a STEW dataset that consisted of a total of 48 subjects. The occurrence of stress is quite common, and it causes many health-related issues, such as insomnia, decreased immunity, infections, cervical impairments and migraines. EEG signals are one of the reliable tools for stress detection. Therefore, stress can be cured before it becomes worse. In this research, an EEG signal is used as the input, and five different frequency bands were extracted using a DWT. After frequency extraction, a CNN was used for automatic feature extraction to achieve better prediction performance. Finally, a deep hybrid model named as CBGG showed significant performance for stress level detection. The attained accuracy using CBGG was 98.10%. The attained results demonstrate the feasibility of using EEG signals for the detection of stress. Therefore, this method is appropriate for the clinical intervention and prevention of mental and physical problems. In future research, the proposed automatic feature extraction-based hybrid deep learning model will be tested in prediction tasks that involve a greater number of EEG datasets. Furthermore, the large number of parameters of these hybrid combinations can be reduced, in order to run them in edge devices.

Author Contributions

Supervision, B.R.; conceptualization, B.R., LM., S.M. and R.K.; methodology, B.R. and L.M.; software, L.M. and R.K.; validation, A.K., T.B., R.K. and J.W.H.; formal analysis, B.R. and L.M.; investigation, B.R. and L.M.; resources, A.K., T.B., R.K. and J.W.H.; data curation, L.M., S.M., R.K. and A.K.; visualization, B.R., L.M., S.M., R.K., A.K. and J.W.H.; writing—original draft preparation, B.R., L.M., S.M., R.K., A.K., T.B. and J.W.H.; writing—review and editing, B.R., L.M., S.M., R.K., A.K., T.B. and J.W.H.; funding acquisition, J.W.H. All authors discussed the results and approved the final study. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by the Korea Agency for Infrastructure Technology Advancement (KAIA) grant funded by the Ministry of Land, Infrastructure and Transport (Grant RS-2022-00143541).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available dataset [49].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sharma, R.; Chopra, K. EEG signal analysis and detection of stress using classification techniques. J. Inf. Optim. Sci. 2020, 41, 229–238. [Google Scholar] [CrossRef]
  2. Mikhno, I.; Koval, V.; Ternavskyi, A. Strategic management of healthcare institution development of the national medical services market. ACCESS Access Sci. Bus. Innov. Digit. Econ. 2020, 1, 157–170. [Google Scholar] [CrossRef] [PubMed]
  3. Qadri, A.; Yan, H. To promote entrepreneurship: Factors that influence the success of women entrepreneurs in Pakistan. Access J. 2023, 4, 155–167. [Google Scholar] [CrossRef] [PubMed]
  4. Singh, S.K.; Singh, S.S.; Singh, V.L. Predicting adoption of next generation digital technology utilizing the adoption-diffusion model fit: The case of mobile payments interface in an emerging economy. Access J. 2023, 4, 130–148. [Google Scholar] [CrossRef]
  5. Cheema, A.; Singh, M. Psychological stress detection using phonocardiography signal: An empirical mode decomposition approach. Biomed. Signal Process. Control. 2019, 49, 493–505. [Google Scholar] [CrossRef]
  6. Petrova, M.; Tairov, I. Solutions to Manage Smart Cities’ Risks in Times of Pandemic Crisis. Risks 2022, 10, 240. [Google Scholar] [CrossRef]
  7. Salankar, N.; Koundal, D.; Qaisar, S.M. Stress classification by multimodal physiological signals using variational mode decomposition and machine learning. J. Health Eng. 2021, 2021, 2146369. [Google Scholar] [CrossRef]
  8. AlShorman, O.; Masadeh, M.; Bin Heyat, B.; Akhtar, F.; Almahasneh, H.; Ashraf, G.; Alexiou, A. Frontal lobe real-time EEG analysis using machine learning techniques for mental stress detection. J. Integr. Neurosci. 2022, 21, 20. [Google Scholar] [CrossRef]
  9. Hasan, M.J.; Kim, J.M. A hybrid feature pool-based emotional stress state detection algorithm using EEG signals. Brain Sci. 2019, 9, 376. [Google Scholar] [CrossRef]
  10. AlShorman, O.; Alshorman, B.; Alkahtani, F. A review of wearable sensors based monitoring with daily physical activity to manage type 2 diabetes. Int. J. Electr. Comput. Eng. 2021, 11, 646–653. [Google Scholar] [CrossRef]
  11. Dushanova, J.; Christov, M. The effect of aging on EEG brain oscillations related to sensory and sensorimotor functions. Adv. Med. Sci. 2014, 59, 61–67. [Google Scholar] [CrossRef]
  12. Mason, A.E.; Adler, J.M.; Puterman, E.; Lakmazaheri, A.; Brucker, M.; Aschbacher, K.; Epel, E.S. Stress resilience: Narrative identity may buffer the longitudinal effects of chronic caregiving stress on mental health and telomere shortening. Brain Behav. Immun. 2019, 77, 101–109. [Google Scholar] [CrossRef]
  13. Belleau, E.L.; Treadway, M.T.; Pizzagalli, D.A. The Impact of Stress and Major Depressive Disorder on Hippocampal and Medial Prefrontal Cortex Morphology. Biol. Psychiatry 2019, 85, 443–453. [Google Scholar] [CrossRef]
  14. Fernández, J.R.; Anishchenko, L. Mental stress detection using bioradar respiratory signals. Biomed. Signal Process. Control. 2018, 43, 244–249. [Google Scholar] [CrossRef]
  15. Heyat, M.B.; Hasan, Y.M.; Siddiqui, M.M. EEG signals and wireless transfer of EEG Signals. Int. J. Adv. Res. Comput. Commun. Eng. 2015, 4, 10–12. [Google Scholar]
  16. Bakhshayesh, H.; Fitzgibbon, S.; Janani, A.S.; Grummett, T.S.; Pope, K. Detecting synchrony in EEG: A comparative study of functional connectivity measures. Comput. Biol. Med. 2019, 105, 1–15. [Google Scholar] [CrossRef]
  17. Heyat, M.B.; Siddiqui, M.M. Recording of eegecgemg signal. Int. J. Adv. Res. Comput. Sci. Softw. Eng. 2015, 5, 813–815. [Google Scholar]
  18. Pal, R.; Heyat, M.B.; You, Z.; Pardhan, B.; Akhtar, F.; Abbas, S.J.; Guragai, B.; Acharya, K. Effect of Maha Mrityunjaya HYMN recitation on human brain for the analysis of single EEG channel C4-A1 using machine learning classifiers on yoga practitioner. In Proceedings of the 2020 17th International Computer Conference on Wavelet Active Media Technology and Information Processing (ICCWAMTIP), Chengdu, China, 18–20 December 2020; pp. 89–92. [Google Scholar]
  19. Cea-Canas, B.; Gomez-Pilar, J.; Nunez, P.; Rodriguez-Vazquez, E.; de Uribe, N.; Diez, A.; Perez-Escudero, A.; Molina, V. Connectivity strength of the EEG functional network in schizophrenia and bipolar disorder. Prog. Neuro-Psychopharmacol. Biol. Psychiatry 2020, 98, 109801. [Google Scholar] [CrossRef]
  20. Dushanova, J.A.; Tsokov, S.A. Small-world EEG network analysis of functional connectivity in developmental dyslexia after visual training intervention. J. Integr. Neurosci. 2020, 19, 601–618. [Google Scholar] [CrossRef]
  21. Olson, E.A.; Cui, J.; Fukunaga, R.; Nickerson, L.D.; Rauch, S.L.; Rosso, I.M. Disruption of white matter structural integrity and connectivity in posttraumatic stress disorder: A TBSS and tractography study. Depress. Anxiety 2017, 34, 437–445. [Google Scholar] [CrossRef]
  22. Zubair, M.; Yoon, C. Multilevel mental stress detection using ultra-short pulse rate variability series. Biomed. Signal Process. Control. 2020, 57, 101736. [Google Scholar] [CrossRef]
  23. Goodman, R.N.; Rietschel, J.C.; Lo, L.-C.; Costanzo, M.E.; Hatfield, B.D. Stress, emotion regulation and cognitive performance: The predictive contributions of trait and state relative frontal EEG alpha asymmetry. Int. J. Psychophysiol. 2013, 87, 115–123. [Google Scholar] [CrossRef] [PubMed]
  24. Luján, M..; Jimeno, M.V.; Sotos, J.M.; Ricarte, J.J.; Borja, A.L. A Survey on EEG Signal Processing Techniques and Machine Learning: Applications to the Neurofeedback of Autobiographical Memory Deficits in Schizophrenia. Electronics 2021, 10, 3037. [Google Scholar] [CrossRef]
  25. Hosseini, M.-P.; Hosseini, A.; Ahi, K. A Review on Machine Learning for EEG Signal Processing in Bioengineering. IEEE Rev. Biomed. Eng. 2020, 14, 204–218. [Google Scholar] [CrossRef] [PubMed]
  26. Aggarwal, S.; Chugh, N. Review of machine learning techniques for EEG based brain computer interface. Arch. Comput. Methods Eng. 2022, 29, 3001–3020. [Google Scholar] [CrossRef]
  27. Sun, J.; Cao, R.; Zhou, M.; Hussain, W.; Bin Wang, B.; Xue, J.; Xiang, J. A hybrid deep neural network for classification of schizophrenia using EEG Data. Sci. Rep. 2021, 11, 4706. [Google Scholar] [CrossRef]
  28. Zuo, X.-N. A machine learning window into brain waves. Neuroscience 2020, 436, 167–169. [Google Scholar] [CrossRef]
  29. Najafzadeh, H.; Esmaeili, M.; Farhang, S.; Sarbaz, Y.; Rasta, S.H. Automatic classification of schizophrenia patients using resting-state EEG signals. Phys. Eng. Sci. Med. 2021, 44, 855–870. [Google Scholar] [CrossRef]
  30. Barros, C.; Silva, C.A.; Pinheiro, A.P. Advanced EEG-based learning approaches to predict schizophrenia: Promises and pitfalls. Artif. Intell. Med. 2021, 114, 102039. [Google Scholar] [CrossRef]
  31. Vázquez, M.A.; Maghsoudi, A.; Mariño, I.P. An Interpretable Machine Learning Method for the Detection of Schizophrenia Using EEG Signals. Front. Syst. Neurosci. 2021, 15, 652662. [Google Scholar] [CrossRef]
  32. Mortaga, M.; Brenner, A.; Kutafina, E. Towards interpretable machine learning in EEG analysis. In German Medical Data Sciences 2021: Digital Medicine: Recognize–Understand–Heal 2021; IOS Press: Amsterdam, The Netherlands, 2021; pp. 32–38. [Google Scholar]
  33. da Silva Lourenço, C.; Tjepkema-Cloostermans, M.C.; van Putten, M.J. Machine learning for detection of interictal epileptiform discharges. Clin. Neurophysiol. 2021, 132, 1433–1443. [Google Scholar] [CrossRef]
  34. Gao, Z.; Dang, W.; Wang, X.; Hong, X.; Hou, L.; Ma, K.; Perc, M. Complex networks and deep learning for EEG signal analysis. Cogn. Neurodyn. 2021, 15, 369–388. [Google Scholar] [CrossRef]
  35. Aslan, Z.; Akin, M. A deep learning approach in automated detection of schizophrenia using scalogram images of EEG signals. Phys. Eng. Sci. Med. 2022, 45, 83–96. [Google Scholar] [CrossRef]
  36. Ahmedt-Aristizabal, D.; Fernando, T.; Denman, S.; Robinson, J.E.; Sridharan, S.; Johnston, P.J.; Laurens, K.R.; Fookes, C. Identification of Children at Risk of Schizophrenia via Deep Learning and EEG Responses. IEEE J. Biomed. Health Inform. 2020, 25, 69–76. [Google Scholar] [CrossRef]
  37. Nikolaev, D.; Petrova, M. Application of Simple Convolutional Neural Networks in Equity Price Estimation. In Proceedings of the 2021 IEEE 8th International Conference on Problems of Infocommunications, Science and Technology (PIC S&T), Kharkiv, Ukraine, 5–7 October 2021; pp. 147–150. [Google Scholar] [CrossRef]
  38. Asif, A.; Majid, M.; Anwar, S.M. Human stress classification using EEG signals in response to music tracks. Comput. Biol. Med. 2019, 107, 182–196. [Google Scholar] [CrossRef]
  39. Ranjith, C.; Arunkumar, B. An improved elman neural network based stress detection from EEG signals and reduction of stress using music. Int. J. Eng. Res. Technol. 2019, 12, 16–23. [Google Scholar]
  40. Dyachenko, Y.; Nenkov, N.; Petrova, M.; Skarga-Bandurova, I.; Soloviov, O. Approaches to cognitive architecture of autonomous intelligent agent. Biol. Inspired Cogn. Arch. 2018, 26, 130–135. [Google Scholar] [CrossRef]
  41. Betti, S.; Lova, R.M.; Rovini, E.; Acerbi, G.; Santarelli, L.; Cabiati, M.; Del Ry, S.; Cavallo, F. Evaluation of an Integrated System of Wearable Physiological Sensors for Stress Monitoring in Working Environments by Using Biological Markers. IEEE Trans. Biomed. Eng. 2017, 65, 1748–1758. [Google Scholar] [CrossRef]
  42. Attallah, O. An Effective Mental Stress State Detection and Evaluation System Using Minimum Number of Frontal Brain Electrodes. Diagnostics 2020, 10, 292. [Google Scholar] [CrossRef]
  43. Ahani, A.; Wahbeh, H.; Nezamfar, H.; Miller, M.; Erdogmus, D.; Oken, B. Quantitative change of EEG and respiration signals during mindfulness meditation. J. Neuroeng. Rehabil. 2014, 11, 87. [Google Scholar] [CrossRef]
  44. Şeker, M.; Özbek, Y.; Yener, G.; Özerdem, M.S. Complexity of EEG dynamics for early diagnosis of Alzheimer’s disease using permutation entropy neuromarker. Comput. Methods Programs Biomed. 2021, 206, 106116. [Google Scholar] [CrossRef] [PubMed]
  45. Vanitha, V.; Krishnan, P. Real time stress detection system based on EEG signals. Biomed. Res. 2016, 2017 (Suppl. S1), S271–S275. [Google Scholar]
  46. Aghajani, H.; Garbey, M.; Omurtag, A. Measuring mental workload with EEG+ fNIRS. Front. Hum. Neurosci. 2017, 11, 359. [Google Scholar] [CrossRef] [PubMed]
  47. Aydin, S.; Arica, N.; Ergul, E.; Tan, O. Classification of obsessive compulsive disorder by EEG complexity and hemispheric dependency measurements. Int. J. Neural Systems 2015, 25, 1550010. [Google Scholar] [CrossRef]
  48. Amin, H.U.; Mumtaz, W.; Subhani, A.R.; Saad, M.N.M.; Malik, A.S. Classification of EEG Signals Based on Pattern Recognition Approach. Front. Comput. Neurosci. 2017, 11, 103. [Google Scholar] [CrossRef]
  49. Jirayucharoensak, S.; Pan-Ngum, S.; Israsena, P. EEG-Based Emotion Recognition Using Deep Learning Network with Principal Component Based Covariate Shift Adaptation. Sci. World J. 2014, 2014, 627892. [Google Scholar] [CrossRef]
  50. Le Douget, J.E.; Fouad, A.; Filali, M.M.; Pyrzowski, J.; Le Van Quyen, M. Surface and intracranial EEG spike detection based on discrete wavelet decomposition and random forest classification. In Proceedings of the 2017 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Jeju, Republic of Korea, 11–15 July 2017; pp. 475–478. [Google Scholar]
  51. Amezquita-Sanchez, J.P.; Mammone, N.; Morabito, F.C.; Adeli, H. A New dispersion entropy and fuzzy logic system methodology for automated classification of dementia stages using electroencephalograms. Clin. Neurol. Neurosurg. 2021, 201, 106446. [Google Scholar] [CrossRef]
  52. Li, M.; Xu, H.; Liu, X.; Lu, S. Emotion recognition from multichannel EEG signals using K-nearest neighbor classification. Technol. Health Care 2018, 26, 509–519. [Google Scholar] [CrossRef]
  53. Nath, D.; Singh, M.; Sethia, D.; Kalra, D.; Indu, S. An efficient approach to eeg-based emotion recognition using lstm network. In Proceedings of the 2020 16th IEEE international colloquium on signal processing & its applications (CSPA), Langkawi, Malaysia, 28–29 February 2020; pp. 88–92. [Google Scholar]
  54. Das Chakladar, D.; Dey, S.; Roy, P.P.; Dogra, D.P. EEG-based mental workload estimation using deep BLSTM-LSTM network and evolutionary algorithm. Biomed. Signal Process. Control. 2020, 60, 101989. [Google Scholar] [CrossRef]
  55. Roy, S.; Kiral-Kornek, I.; Harrer, S. ChronoNet: A deep recurrent neural network for abnormal EEG identification. In Proceedings of the Artificial Intelligence in Medicine: 17th Conference on Artificial Intelligence in Medicine, AIME 2019, Poznan, Poland, June 26–29; 2019; pp. 47–56. [Google Scholar]
  56. Nicolas-Alonso, L.F.; Gomez-Gil, J. Brain computer interfaces, a review. Sensors 2012, 12, 1211–1279. [Google Scholar] [CrossRef]
  57. Lim, W.L.; Sourina, O.; Wang, L.P. STEW: Simultaneous task EEG workload data set. IEEE Trans. Neural Syst. Rehabil. Eng. 2018, 26, 2106–2114. [Google Scholar] [CrossRef]
  58. Lakshmi, M.R.; Prasad, T.V.; Prakash, D.V. Survey on EEG signal processing methods. Int. J. Adv. Res. Comput. Sci. Softw. Eng. 2014, 4, 84–91. [Google Scholar]
  59. Malviya, L.; Mal, S. CIS feature selection based dynamic ensemble selection model for human stress detection from EEG signals. Clust. Comput. 2023, 1–15. [Google Scholar] [CrossRef]
  60. Ji, N.; Ma, L.; Dong, H.; Zhang, X. EEG Signals Feature Extraction Based on DWT and EMD Combined with Approximate Entropy. Brain Sci. 2019, 9, 201. [Google Scholar] [CrossRef]
  61. Aamir, M.; Pu, Y.-F.; Rahman, Z.; Tahir, M.; Naeem, H.; Dai, Q. A Framework for Automatic Building Detection from Low-Contrast Satellite Images. Symmetry 2018, 11, 3. [Google Scholar] [CrossRef]
  62. Wu, H.; Niu, Y.; Li, F.; Li, Y.; Fu, B.; Shi, G.; Dong, M. A Parallel Multiscale Filter Bank Convolutional Neural Networks for Motor Imagery EEG Classification. Front. Neurosci. 2019, 13, 1275. [Google Scholar] [CrossRef]
  63. Mohseni, M.; Shalchyan, V.; Jochumsen, M.; Niazi, I.K. Upper limb complex movements decoding from pre-movement EEG signals using wavelet common spatial patterns. Comput. Methods Programs Biomed. 2020, 183, 105076. [Google Scholar] [CrossRef]
  64. Smagulova, K.; James, A.P. Overview of long short-term memory neural networks. In Deep Learning Classifiers with Memristive Networks: Theory and Applications; Springer: Cham, Switzerland, 2020; pp. 139–153. [Google Scholar]
  65. Bengio, Y.; Simard, P.; Frasconi, P. Learning long-term dependencies with gradient descent is difficult. IEEE Trans. Neural Netw. 1994, 5, 157–166. [Google Scholar] [CrossRef]
  66. Kumar, D.; Singh, A.; Samui, P.; Jha, R.K. Forecasting monthly precipitation using sequential modelling. Hydrol. Sci. J. 2019, 64, 690–700. [Google Scholar] [CrossRef]
  67. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  68. Yang, J.; Huang, X.; Wu, H.; Yang, X. EEG-based emotion classification based on Bidirectional Long Short-Term Memory Network. Procedia Comput. Sci. 2020, 174, 491–504. [Google Scholar] [CrossRef]
  69. Malviya, L.; Mal, S. A novel technique for stress detection from EEG signal using hybrid deep learning model. Neural Comput. Appl. 2022, 34, 19819–19830. [Google Scholar] [CrossRef]
  70. Abuqaddom, I.; Mahafzah, B.A.; Faris, H. Oriented stochastic loss descent algorithm to train very deep multi-layer neural networks without vanishing gradients. Knowl. Based Systems 2021, 230, 107391. [Google Scholar] [CrossRef]
  71. Graves, A. Supervised Sequence Labelling. In Supervised Sequence Labelling with Recurrent Neural Networks; Studies in Computational Intelligence; Springer: Berlin/Heidelberg, Germany, 2012; pp. 37–45. [Google Scholar] [CrossRef]
  72. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  73. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar]
  74. Geldiev, E.M.; Nenkov, N.V.; Petrova, M.M. Exercise of machine learning using some python tools and techniques. CBU Int. Conf. Proc. 2018, 6, 1062–1070. [Google Scholar] [CrossRef]
  75. Faust, O.; Acharya, U.R.; Adeli, H.; Adeli, A. Wavelet-based EEG processing for computer-aided seizure detection and epilepsy diagnosis. Seizure 2015, 26, 56–64. [Google Scholar] [CrossRef]
  76. Ismail, A.R.; Asfour, S.S. Discrete wavelet transform: A tool in smoothing kinematic data. J. Biomech. 1999, 32, 317–321. [Google Scholar] [CrossRef]
  77. Acharya, U.R.; Oh, S.L.; Hagiwara, Y.; Tan, J.H.; Adeli, H. Deep convolutional neural network for the automated detection and diagnosis of seizure using EEG signals. Comput. Biol. Med. 2018, 100, 270–278. [Google Scholar] [CrossRef]
Figure 1. EEG signal analysis general steps.
Figure 1. EEG signal analysis general steps.
Diagnostics 13 01936 g001
Figure 2. Positions of electrodes according to the 10–20 international system.
Figure 2. Positions of electrodes according to the 10–20 international system.
Diagnostics 13 01936 g002
Figure 3. Discrete wavelet transforms analysis.
Figure 3. Discrete wavelet transforms analysis.
Diagnostics 13 01936 g003
Figure 4. Structure of an LSTM memory cell.
Figure 4. Structure of an LSTM memory cell.
Diagnostics 13 01936 g004
Figure 5. BiLSTM model architecture.
Figure 5. BiLSTM model architecture.
Diagnostics 13 01936 g005
Figure 6. Structure of a GRU cell.
Figure 6. Structure of a GRU cell.
Diagnostics 13 01936 g006
Figure 7. Decomposed EEG signal.
Figure 7. Decomposed EEG signal.
Diagnostics 13 01936 g007
Figure 8. Combination DWT-based hybrid DL models.
Figure 8. Combination DWT-based hybrid DL models.
Diagnostics 13 01936 g008
Figure 9. Graphical comparison of the proposed model with other models.
Figure 9. Graphical comparison of the proposed model with other models.
Diagnostics 13 01936 g009
Figure 10. Models train vs. validation accuracy convergence curve (a) CNN–RNN (b) CNN–LSTM, (c) CNN–GRU, (d) CBRR, (e) CBLL, (f) CBGG.
Figure 10. Models train vs. validation accuracy convergence curve (a) CNN–RNN (b) CNN–LSTM, (c) CNN–GRU, (d) CBRR, (e) CBLL, (f) CBGG.
Diagnostics 13 01936 g010
Figure 11. (a) ROC curves of the CNN–RNN, CNN–LSTM, CNN–GRU and (b) CBRR, CBLL and CBGG models.
Figure 11. (a) ROC curves of the CNN–RNN, CNN–LSTM, CNN–GRU and (b) CBRR, CBLL and CBGG models.
Diagnostics 13 01936 g011
Table 2. Model parameter descriptions.
Table 2. Model parameter descriptions.
Model Sequential LayersLayerParameters and Values
CBRRCNN-ID
MaxPooling1D
BiLSTM Layer
RNN Layer
RNN Layer
Dropout
Total Parameters
Filter = 128, kernel size = 1, padding = valid, activation = softmax
Pool_size = 1
Filter =64
Filter =32
Filter =16
0.2
112,927
CBLLCNN-ID Layer
MaxPooling1D
BiLSTM Layer
LSTM Layer
LSTM Layer
Dropout
Total Parameters
Filter = 128, kernel size = 1, padding = valid, activation =softmax
Pool_size = 1
Filter = 64
Filter =32
Fiter =16
0.2
153,681
CBGGCNN-ID Layer
MaxPooling1D
BiLSTM Layer
GRULayer
GRU Layer
Dropout
Total Parameters
Filter = 128, kernel size = 1, padding = valid, activation =softmax
Pool_size = 1
Filter = 64
Filter =32
Fiter =16
0.2
117,657
CNN-RNNCNN-ID Layer
MaxPooling1D
RNN
Dropout
Total Parameters
Filter = 128, kernel size = 1, padding = valid, activation =softmax
Pool_size = 1
Filter = 64
0.2
12,673
CNN-LSTMCNN-ID Layer
MaxPooling1D
LSTM
Dropout
Total Parameters
Filter = 128, kernel size = 1, padding = valid, activation =softmax
Pool_size = 1
Filter = 64
0.2
49,729
CNN-GRUCNN-ID Layer
MaxPooling1D
GRU
Dropout
Total Parameters
Filter = 128, kernel size = 1, padding = valid, activation =softmax
Pool_size = 1
Filter = 64
0.237,569
Table 3. Metrics formulation of parameters.
Table 3. Metrics formulation of parameters.
ParameterFormula
Precision T P T P + F P * 100
Sensitivity T P T P + F N * 100
Specificity T N F P + T N * 100
F1-Score 2     ( P P V     Sensitivity )   P P V +   Sensitivity  
Accuracy T P + T N T P + F N + F N + T N     100
Positive Likelihood Ratio (+LR)   Sensitivity   100   Specificity  
Negative Likelihood Ratio (−LR) 100   Sensitivity     Specificity  
Table 4. Model comparison results.
Table 4. Model comparison results.
ModelAccuracy (%) Precision (%)Sensitivity (%)F1-Score (%)Specificity (%)+LR−LR
CNN–RNN89.9189.7089.8389.8089.8080.12
CNN–LSTM95.6095.9095.6495.9094.90190.05
CNN–GRU95.2096.0995.7795.5094.70180.04
CBRR97.1096.5397.7497.1496.45280.02
CBLL97.5497.3297.8097.5697.27350.02
CBGG98.1098.2798.0898.2097.76440.02
Table 5. Comparison with the most relevant models in dealing with the identification of stress in STEW EEG signals.
Table 5. Comparison with the most relevant models in dealing with the identification of stress in STEW EEG signals.
Feature Extraction/
Selection
ClassifierAccuracyCross-Validation
Particle Swarm Optimization (PSO) [46]BiLSTM, LSTM86.33%-
PSD features via FFT [48])KNN, SVM69.00%-
CNN-based features from DWT signalsCBGG98.10%97.60% (Stratified 10-Fold)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Roy, B.; Malviya, L.; Kumar, R.; Mal, S.; Kumar, A.; Bhowmik, T.; Hu, J.W. Hybrid Deep Learning Approach for Stress Detection Using Decomposed EEG Signals. Diagnostics 2023, 13, 1936. https://doi.org/10.3390/diagnostics13111936

AMA Style

Roy B, Malviya L, Kumar R, Mal S, Kumar A, Bhowmik T, Hu JW. Hybrid Deep Learning Approach for Stress Detection Using Decomposed EEG Signals. Diagnostics. 2023; 13(11):1936. https://doi.org/10.3390/diagnostics13111936

Chicago/Turabian Style

Roy, Bishwajit, Lokesh Malviya, Radhikesh Kumar, Sandip Mal, Amrendra Kumar, Tanmay Bhowmik, and Jong Wan Hu. 2023. "Hybrid Deep Learning Approach for Stress Detection Using Decomposed EEG Signals" Diagnostics 13, no. 11: 1936. https://doi.org/10.3390/diagnostics13111936

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop