You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

15 November 2025

A Bioinspired Multimodal CNN-LSTM Network for EEG Analysis of Patients in Coma

,
,
,
,
,
and
1
Department of Computing and Mathematics, University of São Paulo, Ribeirão Preto 14040-901, Brazil
2
Faculty of Computing, Federal University of Uberlândia, Uberlândia 38400-902, Brazil
3
Faculty of Electrical Engineering, Federal University of Uberlândia, Uberlândia 38400-902, Brazil
4
Uberlândia Clinical Hospital, Federal University of Uberlândia, Uberlândia 38405-320, Brazil
Sensors2025, 25(22), 6981;https://doi.org/10.3390/s25226981 
(registering DOI)
This article belongs to the Special Issue AI and Neural Networks for Advanced Biomedical Sensor Applications

Abstract

Electroencephalography (EEG) is widely used for diagnosis and evaluation of neurological diseases, despite challenges from its high-dimensional and noisy temporal data, which complicate accurate brain signal classification. This study proposes a multimodal deep learning model combining Convolutional Neural Network (CNN) and Long Short-Term Memory (LSTM) layers to classify EEG signals, integrating patient information as additional modalities. CNN layers effectively extract spatial features and reduce EEG data dimensionality, while LSTM layers capture long-term temporal dependencies. A Genetic Algorithm (GA) selects relevant multimodal features and optimizes CNN-LSTM hyperparameters. The model was applied to outcome classification in comatose patients, achieving improved classifier performance compared to unimodal approaches. Experimental results demonstrate that multimodal integration and GA optimization significantly enhance accuracy, robustness, and generalization. The architecture shows promise for broader EEG classification tasks, potentially advancing clinical decision support based on EEG signals.

1. Introduction

Electroencephalography (EEG) is typically a cheap and noninvasive technique that measures electrical activity in the brain using electrodes attached to the scalp. The electrodes capture the summed electrical activity produced by groups of neurons in the cerebral cortex []. An EEG recording provides valuable insights for monitoring and diagnosing various neurological diseases and conditions, such as Alzheimer’s disease [], Parkinson’s disease [], epilepsy [], sleep disorders [], and coma []. Through analysis of EEG patterns, both normal and abnormal brain activities can be identified, aiding doctors in identifying neurological disorders.
Despite the satisfactory performance achieved by traditional machine learning methods in classifying EEG signals [,,], a significant challenge lies in extracting relevant features to be used as inputs for these algorithms. Understanding the essential features that enable machine learning algorithms to classify EEG signals is crucial. Manual feature extraction methods may lack the necessary accuracy due to the inherent variability of EEG signal recordings.
Artificial Neural Network models (ANNs), particularly those based on Convolutional Neural Networks (CNNs) and/or Recurrent Neural Networks (RNNs), have been used for classifying EEG signals [,,]. In contrast to traditional machine learning approaches based on predefined features, deep learning algorithms can automatically extract features from raw EEG data. RNNs have demonstrated their efficacy in capturing temporal information, making them useful in various medical applications. Unlike conventional RNNs, Long Short-Term Memory (LSTM) networks can handle long time-series data [].
In this paper, we propose a new multimodal hybrid deep learning model to classify EEG signals named CNN-LSTM-GA. The proposed model is used to aid the outcome prediction of patients in coma. Coma can be caused by various health conditions []. Accurately predicting patient outcomes is vital for clinical decision-making and providing timely interventions [,]. In addition, outcome prediction enables optimizing treatment at the Intensive Care Unit (ICU), leading to a better work division within the health staff, as well as improving the preparation of possible organ donations.
The proposed hybrid architecture is defined by the combination of CNN and LSTM networks. The CNN layers provide more efficient representations of the EEG signals while reducing their high-dimensional space. On the other hand, the LSTM layers capture longer temporal dependence over sequential data. Additionally, the expressiveness of our hybrid architecture is increased by its ability to learn from multimodalities, such as those related to patient information or extracted from EEG signals by using predefined filters. Our hypothesis is that the predictive performance of CNN-LSTM-GA models is improved when using multimodal information, such as raw EEG signals and patient features obtained from clinical records.
Using CNN combined with LSTM allows the network to perform well on both spatial and temporal data. On the other hand, it represents a more complex technique in relation to the configuration of hyperparameters and architecture. In addition, the adoption of different sources of information increases the number of features, which does not necessarily mean an improvement in predictive performance. To handle this problem, we propose a Genetic Algorithm (GA) to optimize hyperparameters for the hybrid CNN-LSTM, and also to select the most important features. Experiments with hundreds of EEG data collected from 60 comatose patients at a public hospital were conducted in order to evaluate the CNN-LSTM-GA architecture, the strategies to consider additional modalities of information, and the optimization process conducted by GA. In summary, the main contributions of this manuscript are listed as follows: 
  • A multimodal CNN-LSTM-GA model is proposed to predict the outcome of patients in coma from their EEG records, while CNN layers extract features from the EEG records, the LSTM layers capture relationships over temporal sequences.
  • A GA is proposed for the selection of the most important features and to optimize the hyperparameters of the proposed model.
  • The proposed model is able to learn from other modalities of information besides the EEG signals, such as those related to patient information (e.g., sex and age). A fully connected layer fuses the outputs of the LSTM layer and the selected features from other modalities.
The multimodal CNN-LSTM-GA architecture proposed is a novel approach aimed at significantly improving the outcome prediction of comatose patients. By enhancing the accuracy of outcome predictions, more informed clinical decisions and timely interventions can be taken for better patient care, also including the improvement of work organization within the ICU and more efficient possible organ donations.
Furthermore, the proposed GA extends beyond conventional hyperparameter tuning or feature selection approaches. It performs a joint optimization of the CNN–LSTM architecture and multimodal feature subsets within a framework that balances predictive performance and model complexity. Unlike regularization-based or cross-validation strategies, which primarily adjust model parameters or assess robustness, the GA explores the global search space by evolving both the structure and parameters of the network. Although this process introduces some computational cost, it enables a broader and more effective exploration of the multimodal feature space, leading to improved predictive performance and more insights regarding interpretability.
The remainder of this paper is organized as follows. Section 2 presents the main related works in the literature. Section 3 introduces the problem addressed by our investigation as well as the multimodal dataset considered. Section 4 describes the proposed CNN-LSTM-GA architecture and the optimization strategy developed to optimize hyperparameters and feature selection. Section 5 discusses the results of the proposed technique in comparison to other techniques, and presents the main insights obtained from the experiments. Section 6 concludes the paper.

3. EEG Dataset and Feature Description

Predicting patient outcomes is essential for making clinical decisions and timely interventions. In this way, we consider the problem of outcome prediction of patients in coma, in which the model’s output corresponds to one of two categories: favorable or unfavorable outcome.

3.1. Dataset Description

The dataset is composed of information from 60 comatose patients who were admitted to the Clinical Hospital of the Federal University of Uberlândia (HC-UFU) between 2010 and 2014. The data were collected by the hospital’s technical team, responsible for comatose patients in the adult ICU, with the approval of the Ethics Committee of the Federal University of Uberlândia under protocol number 369/11.
The patients underwent an EEG recording session approximately 14 days after being admitted to the hospital. The EEG recording lasted approximately 20 min and was conducted using a 20-channel EEG cap with active electrodes and a BrainNet BNT-EEG device (Linx BNT®). The acquisition process involved placing 20 electrodes in specific regions of the cerebral cortex, following the international 10-20 system: FP1, FP2, F7, F3, FZ, F4, F8, T3, C3, CZ, C4, T4, T5, P3, PZ, P4, T6, O1, OZ, and O2. It is important to note that no external stimuli were presented to the patients during the recording sessions.
Additional information, including age, sex, coma etiology, and outcome, was collected from the clinical records of each individual. The average age of the patients was 49.16 ± 19.18, with 44 males and 16 females. Table 2 presents an overview of the number of patients for each coma etiology and their corresponding outcomes. A favorable outcome was defined as patient survival, while an unfavorable outcome was defined as patient death.
Table 2. Description of our EEG dataset based on the respective number of segments for each coma etiology and outcome.
Comatose patient’s EEG data are characterized by signal power concentrated at low frequencies (1–7 Hz, or delta and theta bands), throughout all cerebral regions, so that there is almost no signal power at high frequencies (above 7 Hz, like alpha and beta waves) []. Otherwise, normal individual EEG recordings present most of the signal power at high frequencies, with alpha waves lying at the posterior regions. This can be observed in Figure 1, which presents a comparison of EEG data collected from a comatose patient and a healthy control patient.
Figure 1. EEG signals of a comatose and a healthy patient.
For each 20 min EEG session, 10 segments, each lasting 2 s, were extracted, resulting in a total of 10 segments’ samples (examples of the dataset) for each patient, which resulted in a dataset with a total of 600 segments’ samples. Each segment corresponds to a fraction of the EEG data captured by the 20 channels. The frequency range of the recordings varied between 100 and 600 Hz. As a result, the dataset comprises segments’ samples from different patients, each with its frequency range. No downsampling was performed to establish a standardized sampling rate. A team of clinical neurologists with extensive expertise conducted a detailed visual analysis of the EEG recordings and selected each one of the segments, ensuring minimal noise presence.

3.2. Feature Description

In the following, we describe both patient and time-domain features considered in this study: 
  • Patient Features. The patient features, another modality of information, were extracted from the patient’s electronic medical records. They are (i) patient’s age; (ii) patient’s sex; (iii) coma etiology; and (iv) coma outcome.
  • Time-Domain (TD) Features. We extracted time-domain features from each 2 s segment of the EEG data. As CNNs have the ability to automatically extract features from inputs, we opted to extract basic features directly from the signals. The selected TD features are (i) mean; (ii) minimum; (iii) maximum; (iv) standard deviation; and (v) variance. As every segment had 20 channels, and each channel captures its own signal, the total of TD features extracted from each segment is 100.

4. The Proposed Multimodal CNN-LSTM-GA

In this section, we present our multimodal CNN-LSTM architecture optimized by GA to aid medical experts in analyzing EEG signals. The model takes EEG signals as inputs and incorporates multimodal features, such as patient information from clinical records and features extracted from EEG signals, as additional inputs to the latent representation learned by the CNN-LSTM. In addition, we also designed a GA to optimize the subset of multimodal features from EEG signals and patient features, as well as to find the optimal hyperparameters of the CNN-LSTM model. We refer to our proposed architecture as CNN-LSTM-GA.
A general overview of our multimodal CNN-LSTM-GA architecture is illustrated by Figure 2. CNN-LSTM-GA is able to process different modalities of information through a CNN-LSTM architecture. On the one hand, multimodal features aim to increase the expressiveness of our hybrid architecture by providing complementary information. On the other hand, CNN-LSTM architecture allows the model to capture both spatial and temporal aspects of the input data.
Figure 2. Overview of the proposed multimodal CNN-LSTM-GA architecture. Electroencephalogram (EEG) segments are processed through Convolutional Neural Network (CNN) layers followed by Long Short-Term Memory (LSTM) layers to extract temporal and spatial features, producing a latent representation. Patient features and time-domain EEG descriptors are then concatenated with this latent representation. A Genetic Algorithm (GA) is employed to select the optimal subset of multimodal features and to optimize the model architecture and hyperparameters.
According to Figure 2, the multimodal information is processed as follows. On the one hand, EEG segments are input to the initial CNN layer and pass through a series of convolutional and pooling layers to automatically extract informative features for classification. These features are then fed into LSTM layers to capture temporal dependencies, and the LSTM output forms the latent representation. On the other hand, patient features from clinical records and TD features from EEG segments are fused with this latent representation to provide richer input to the fully connected layers.
To formally describe this fusion process, we now present the mathematical formulation, which constitutes a key component of our model.
  • Let X E E G R B × T × C denote the EEG input, where B is the batch size, T is the number of time samples, and C is the number of EEG channels. Each EEG segment corresponds to a multichannel time series captured by 20 electrodes ( C = 20 ) with approximately T = 2000 samples per segment.
The CNN-LSTM subnetwork first applies a sequence of one-dimensional convolutions and pooling operations to extract spatial patterns from the EEG input: 
F C N N = CNN ( X E E G ) R B × T × D C N N ,
where T is the reduced temporal dimension after pooling operations, and D C N N is the number of convolutional filters (i.e., the depth of the CNN feature maps).
The feature maps F C N N are then processed by the LSTM layers to capture long-term temporal dependencies and produce a latent temporal representation: 
H L S T M = LSTM ( F C N N ) R B × D L S T M .
where D L S T M is the dimensionality of the hidden state (i.e., the number of LSTM units).
Patient and time-domain (TD) features are represented as X P R B × D P and X T D R B × D T D , respectively, where D P and D T D , correspond to the patient variables and the time-domain descriptors extracted from each EEG segment selected in the GA. These two feature sets are first concatenated to form a single multimodal vector: 
Z P T D = [ X P X T D ] R B × ( D P + D T D ) .
The resulting vector Z P T D is then concatenated with the latent EEG representation obtained from the LSTM: 
Z f u s i o n = [ H L S T M Z P T D ] R B × ( D L S T M + D P + D T D ) ,
where ∥ denotes concatenation along the feature dimension. The fused multimodal representation is finally passed through fully connected layers to produce the classification output: 
y ^ = Softmax ( f F C ( Z f u s i o n ) ) .
This corresponds to a hierarchical fusion strategy, where the EEG signal is first encoded through CNN and LSTM layers to obtain a latent temporal–spatial representation H L S T M , which is then combined with patient and TD embeddings after modality-specific processing. For the dataset used in this study, the resulting latent representation has dimension H L S T M R B × 128 before fusion.
We also developed a GA to optimize the selection of multimodal information related to patient and TD features, as well as to find the optimal hyperparameters of the CNN-LSTM model. In a GA, a population of individuals or chromosomes, which represent potential solutions to a problem, undergoes selection, crossover, and mutation inspired by natural selection and genetics [], in which each individual is evaluated by a fitness function able to define how good that corresponding solution is.
We designed an encoding in which each GA individual consists of two major parts: one for the feature selection (part 1) and one for optimization of the hyperparameters that define the CNN-LSTM architecture (part 2). Each GA individual encodes a subset of features (binary representation) and the hyperparameters of the CNN-LSTM (integer, real, or categorical values), forming a mixed-type optimization problem. Figure 3 shows how the feature selection strategy and the hyperparameter optimization of the CNN-LSTM are codified in the chromosome of a GA individual. In the part 1 (left side of an individual denoted by red color in the figure), each individual represents a specific subset of features, in which the i-th binary element of the chromosome indicates whether the i-th feature is included or excluded. In the part 2 (right side of an individual denoted by blue color in the figure), the hyperparameters to be optimized are divided in four groups related to CNN layers (part 2a), LSTM layers (part 2b), fully connected (dense) layers (part 2c), and general training parameters (part 2d). In addition, we also defined empirically that the GA-coded architectures have up to five convolution layers, up to three LSTM layers, and up to two dense layers, which is often enough to obtain good results. Table A1 of the Appendix A provides the range of hyperparameter values that can be selected by the GA at each position of the encoding.
Figure 3. Representation of the solution (values for the hyperparameters of the CNN-LSTM and features) in the chromosome of the individual.
The simplified flowchart of the proposed GA is presented in Figure 4. Firstly, an initial population of candidate solutions is randomly generated following the encoding presented in Figure 3. Each individual is then evaluated according to a fitness function in which its corresponding features and hyperparameters are considered. Genetic operators of selection and replacement (crossover and mutation) are applied in order to produce better configurations of candidate solutions. Such a procedure is repeated by a fixed number of generations (iterations). At the end, we expect to achieve an optimal configuration of multimodal features and hyperparameters able to provide good results for a given EEG classification task.
Figure 4. Flowchart of the GA utilized for feature selection and the selection of hyperparameters for the hybrid CNN-LSTM model.
We defined the F1 score as the fitness function for each candidate solution. To obtain that, the dataset was divided into three distinct parts: 40% for training, 40% for validation, and 20% for testing. The stratified method was used, which ensures that the proportions of samples for each category are maintained when randomly partitioning the dataset. During the GA execution process, the training and validation parts were utilized. Each GA individual was evaluated by training the corresponding model using the training data, and the F1 score was calculated using the validation examples. At the end of the GA execution, the best individual was evaluated by retraining the model with both the training and validation sets and testing it using the test set.
Regarding the genetic operators, we designed selection and crossover as follows: selection adopts a comparison (tournament) between two individuals randomly selected from the population, in which the individuals with better fitness are considered for crossover; and crossover considers a slice of two points randomly defined to produce new individuals by swapping such chromossomes’ values between a given pair of selected individuals. Given that the individuals consist of two major parts, their chromosomes contain a variety of data types, including Boolean, integer, or string. Consequently, different types of mutations are implemented based on the specific data type found in the randomly chosen chromosome position undergoing mutation: 
  • Boolean mutation occurs when a mutation affects a position on the chromosome where the data type is Boolean. This mutation involves flipping the value present in the original chromosome.
  • Window mutation is applied when the mutation occurs at a chromosome position where the data type is integer or real. In this case, the values are ordered, and the mutation involves creating a subset considering the two values before (smaller) and after (bigger) the value present in the chromosome, if any. A new value is then randomly chosen in this subset of values.
  • Nominal mutation is used when a mutation occurs in a chromosome position that represents a nominal categorical variable (string), meaning a variable with unordered values. In this case, a new value is randomly selected, from all possible values, to replace the value present in the chromosome. Notice that, in the case of nominal categorical variables with only two values, the nominal mutation is similar to the Boolean mutation.
The pseudocode describing the optimization process is presented in Algorithm 1. The complete implementation of the proposed CNN–LSTM–GA model, including data preprocessing, training, and evaluation procedures, is publicly available at our GitHubrepository (https://github.com/SergioBaldo/CNN-LSTM-GA, accessed on 22 September 2025).
Algorithm 1 CNN-LSTM-GA: Feature selection and hyperparameter optimization
  • Input: D—Dataset, N p o p —population size, N g e n —number of generations, p c —crossover probability, p m —mutation probability  
  • Output:  Best individual I * and corresponding CNN-LSTM model M *  
  1:
Split D into train (40%), validation (40%) and test (20%) subsets;
// Initialize population P with N p o p individuals
  2:
for each individual i P  do
  3:
      Randomly initialize feature subset F i and model hyperparameters H i ;
  4:
end for
  5:
for generation = 1 to N g e n  do
  6:
      for each individual i P  do
  7:
            Decode chromosome → feature subset F i and model hyperparameters H i ;
  8:
            Build CNN-LSTM( H i );
  9:
            Construct T t r a i n and T t e s t datasets from EEG data using feature subset F i ;
10:
            Train CNN-LSTM( H i ) model with T t r a i n ;
11:
            Compute fitness f ( i ) as the F1-score of CNN_LSTM( H i ) evaluated on T v a l i d a t i o n ;
12:
      end for
13:
      Apply tournament selection to choose parents based on f ( i ) ;
14:
      For each parent pair, generate two offspring using crossover with probability p c .
15:
      for each offspring gene g with probability p m  do
16:
            if g is Boolean then
17:
                  Flip gene g;
18:
            else if g is Integer or Real then
19:
                  Apply window mutation;
20:
            else if g is Nominal then
21:
                  Replace with random category;
22:
            end if
23:
      end for
24:
      Form new population P combining best individuals and offspring;
25:
end for
26:
Select best individual I * = arg max i f ( i ) ;
27:
Decode individual I * → feature subset F i and model hyperparameters H i ;
28:
Build CNN-LSTM( H i );
29:
Construct T t r a i n and T t e s t datasets from EEG data using feature subset F i ;
30:
Train CNN-LSTM( H i ) model with T t r a i n ;
31:
Evaluate final model M * on T t e s t ;
32:
return  I * , M *

5. Results and Discussion

The algorithms were implemented in Python (version 3.9.15) using the TensorFlow library (version 2.10.0). A workstation with an Intel@ Core i7-12700 Alder Lake processor (12 cores, 20 threads, 25 MB cache), 128 GB of RAM, and a GF AMPERE RTX 3060 12 GB 3584 CUDA GPU was used for running the codes. We compared the proposed model against several other models, considering monomodal and multimodal simulations. By comparing different models and modality variations, it is possible to evaluate the individual contribution of different mechanisms, e.g., to assess whether adding patient features is beneficial or not. The compared models with only a simple modality are: 
  • EEGTransformer []: A TransformerEncoder based model designed to capture complex spatio-temporal patterns in raw EEG data, thereby enhancing class separability and prediction reliability.
  • MLP w/Patient Feats: an MLP that uses only patient features as inputs. All MLP models were trained for 30 epochs, using the Adam optimizer with a learning rate of 0.0001, and the features were normalized using standard scaling. The number of neurons in the hidden layer was defined as the number of input features. The output layer adopts a sigmoid function with a single neuron, and the model was trained using the binary accuracy metric and the binary crossentropy loss;
  • MLP w/TD Feats: an MLP that uses only TD features as inputs;
  • CNN-LSTM-baseline: a CNN-LSTM model that uses only raw EEG signals as inputs to the CNN;
  • CNN-LSTM w/Patient Feats: a CNN-LSTM model that uses EEG signals as inputs to the CNN and patient features as additional inputs to the first dense layer;
  • CNN-LSTM w/TD Feats: a CNN-LSTM model that uses raw EEG signals as inputs to the CNN and time-domain features extracted from EEG signals as additional inputs to the first dense layer.
The multimodal models under comparison are: 
  • MLP w/Patient + TD Feats: an MLP that uses patient and time-domain features extracted from EEG as inputs;
  • CNN-LSTM w/Patient + TD Feats: a CNN-LSTM model that uses raw EEG signals as inputs to the CNN and patient and time-domain features extracted from EEG signal as additional inputs to the first dense layer;
  • CNN-LSTM-GA (Proposed): a CNN-LSTM model with the hyperparameters optimized by GA that uses EEG signals as inputs to the CNN and multimodal features selected by GA as additional inputs to the first dense layer (Figure 2). Table A2 of the Appendix A presents the architecture and the configuration of hyperparameters for CNN-LSTM-GA.
The number of runs of the GA was 10; in each run, a different pseudo-random seed was used. The population size was 30 individuals, and the number of generations was 30. The crossover rate was 0.6 and the mutation rate was 2 n , where n is the size of the chromosome. F1 score (macro) was used as the fitness measure for evaluating the individuals.

5.1. Evaluation Procedures

Two evaluation procedures have been defined to our experiments as follows: 
  • Hold-out. In order to simulate a real-world scenario, we randomly divided the dataset into 40% for training, 40% for validation, and 20% for testing.
  • Cross-validation (CV). We also employed a stratified five-fold cross-validation approach where the dataset was randomly partitioned based on patients. The patients were divided into five equal-sized groups. During each iteration, one group was designated as the test set, while the remaining groups were used for training.
Since there are ten segments’ samples per patient, in both hold-out and CV experiments we only utilized the first nine segments for our tests, i.e., the classification was performed on these nine segments’ samples, and the algorithm’s final prediction for each patient was determined by selecting the most frequent classified category.
For the evaluation of EEG classification models, metrics that capture both overall accuracy and class-specific performance are essential, particularly due to the common presence of imbalanced class distributions. Accuracy provides a general overview of correct predictions; however, it may be misleading when classes are not evenly represented. Therefore, precision, also known as positive predictive value, measures the proportion of true positive predictions within all positive predictions for a class, while recall (sensitivity) reflects the proportion of actual positive cases correctly identified by the model. Additionally, the F1-score, as the harmonic mean of precision and recall, offers a balanced performance metric at the class level, especially valuable in imbalanced scenarios. These metrics derive from the confusion matrix, which displays counts of true positives, true negatives, false positives, and false negatives, as illustrated in Table 3. The corresponding equations are detailed in Table 4.
Table 3. Confusion matrix example for binary classification.
Table 4. Performance metrics and equations.

5.2. Prediction of Coma Outcome

In this experiment, the model’s output corresponds to either a “favorable” or an “unfavorable” outcome. The patient’s features were age, sex, and etiology of the coma. As sex and coma etiology are categorical attributes, we transformed them into numerical variables using the One-Hot-Encoding technique. This transformation resulted in a total of 14 patient features, besides the 100 features extracted from the EEG signals.
Table 5 presents the results obtained for the eight models, including the proposed CNN-LSTM-GA model. The mean and standard deviation of the results obtained in the 10 runs of the GA are also presented. The evaluation is based on the test set, using a dataset division with 80% of the data in the training set (40% training and 40% validation) and 20% in the test set (hold-out). Additionally, the table provides the mean and standard deviation obtained in stratified five-fold cross-validation (CV). These metrics were calculated using the results obtained from the test set of each fold. Results in terms of other complementary metrics, like Geometric Mean, can be seen in Table A4 of Appendix A.
Table 5. The results obtained by the models for predicting outcome of the coma, including the best model identified by the GA, as well as the average results for all 10 runs. The best results for each metric are in bold.
According to Table 5, MLP w/Patient + TD Feats, CNN-LSTM w/Patient Feats and CNN-LSTM-GA (Proposed) obtained the best results for the hold-out experiments. On the other hand, the CNN-LSTM-GA (Proposed) obtained the best predictive performance for CV, with an F1(macro) of 0.922 ± 0.046 for the best GA run. Figure 5 shows the results of each model, considering each evaluation metric in each fold of CV, which shows the robust predictive performance of our CNN-LSTM-GA model.
Figure 5. Evaluation for each fold in the coma outcome classification experiment. (a) Accuracy; (b) F1 score (macro); (c) precision; (d) recall.
By analyzing the results of the techniques that learn from multimodalities, opposite behavior can be noticed. In the hold-out validation, the predictive performance of CNN-LSTM w/Patient Feats was similar to the results obtained by the MLP w/Patient + TD Feats model, which incorporates both the patient and TD features as input, with a slight advantage for MLP in terms of F1(macro). However, the results obtained in CV show that the CNN-LSTM w/Patient Feats model outperformed the MLP w/Patient + TD Feats considerably in terms of all predictive metrics. The same occurred in relation to other techniques like Random Forest and Support Vector Machine, the results of which are presented in Table A5 of Appendix A.
The performance obtained by the CNN-LSTM w/Patient Feats model suggests that incorporating patient features as a new modality for the first dense layer enhances the performance of the CNN-LSTM-baseline model. However, when all available features, including those of the patient (age, sex, and etiology of the coma) and those extracted from the EEG signals, were used as additional inputs for the first dense layer in the CNN-LSTM w/Patient + TD Feats model, the classifier performance deteriorated. The worse results can be mainly attributed to the significantly higher number of features (114 features) used by the CNN-LSTM w/Patient + TD Feats model compared to only 14 features utilized in the CNN-LSTM w/Patient Feats model. Additionally, the features extracted from the EEG signals are relatively simple and most of them contribute little to the classification. These facts also help to explain the better results of the model optimized by the GA, which are discussed next.
In Table 5, one can observe that our model, CNN-LSTM-GA, obtained the same performance as the CNN-LSTM w/Patient Feats model in the hold-out experiments. However, in the CV experiments, CNN-LSTM-GA obtained the best results, with an F1(macro) of 0.922 ± 0.046. The model optimized by GA utilized only 38 features as additional inputs; five of them are patient features and thirty-three are features extracted from the EEG signals. This number of features is smaller than that used in CNN-LSTM w/Patient + TD Feats. These results suggest that the inclusion of specific multimodal features extracted from the EEG signals, along with patient features, contributed to improving the classifier’s performance. This improvement becomes evident when comparing the mean result of CNN-LSTM-GA presented in Table A3 of Appendix A, against the results obtained by other CNN-LSTM variants. Taking the 10 GA runs, CNN-LSTM-GA achieved better averaged results of accuracy, F1(macro) and recall than all other models.
As one of the state-of-the-art methods, the EEGTransformer [] model has demonstrated good performance in EEG classification tasks. Our proposed CNN-LSTM-GA architecture surpasses this model in both hold-out and cross-validation evaluations. This improvement underscores the effectiveness of integrating hybrid deep learning with multimodal inputs and Genetic Algorithm-based optimization techniques.
Analyzing Table 5 reveals a significant difference between the two model performance evaluation methods: hold-out and cross-validation. This difference arises from how each technique operates. In the hold-out approach, the dataset is initially split into three distinct sets: 40% for training, 40% for validation, and 20% for testing. These first two sets are combined into a single training set on which the model is both trained and validated, and afterwards, the model is evaluated on the separate test set. Conversely, in cross-validation, all available data are combined into a single dataset, which is randomly divided into five folds for cross-validation. The model is iteratively trained on four folds and tested on the remaining fold, cycling through all folds to provide a more stable and reliable performance estimate by reducing variability caused by any single data split.
The experimental results indicate that the GA can select an optimized set of features for the first dense layer of the CNN-LSTM model. In addition, the GA can find a model with superior performance and optimized hyperparameters for solving the problem at hand. Additionally, as shown in Table A3 of Appendix A, the CNN-LSTM-GA average results obtained from all GA runs are better than those of almost all other models under comparison listed in Table 5. This suggests that most of the GA runs were able to find a model that outperformed the reference models.
Another interesting analysis is presented in Figure 6, which depicts the most frequent features selected by our multimodal CNN-LSTM-GA architecture, i.e., those ones whose frequency attains at least in 50% of the simulations. Figure 6a,b present results regarding TD features obtained directly from the signals of the electrodes, pointing out that right-side and central electrodes (T4, T6, O2, F4, Fz, Cz) play a relevant role in terms of outcome prediction. Particularly, electrodes T6 and F4 achieve a frequency of 70%, thus pointing out that information tied to the electrical activity related to the cerebral right anterior region is very important for outcome prediction. This result deserves further neurological interpretations and discussions. On the other hand, Figure 6c depicts results regarding another modality related to patient clinical data. One can see that age presents a frequency of 60%, whereas, for hypoxic etiology, frequency attains 50%. This is already expected, since from a medical viewpoint, elderly patients are tied to a minimal recovery ability, and hypoxic phenomena lead to a high rate of patient mortality [].
Figure 6. Features selected by the best solution found in each of the 10 runs of the GA in the coma outcome classification experiment. The total of features is 114.

5.3. Applicability

This investigation has been conducted through a partnership with a Brazilian public hospital. Therefore, we assume that the health units have limited resources, where electroencephalogram equipment is available, but other more sophisticated (and expensive) equipment is not. This assumption is in conformation with most Brazilian health units of the public health system. Specifically, the application of the multimodal architectures proposed in this paper considers the prognosis of patients in coma.
The prognosis of comatose patients admitted to the ICU involves several factors related to coma severity, patient history, demographics, etc. Despite the big challenge behind such a task, the prognosis is crucial to select the most appropriate treatments, as well as to verify their effectiveness. In addition, it can also be helpful in specific cases like brain death, in which clinical protocols aimed at organ donation may be started earlier. In this paper, we found that our CNN-LSTM-GA architecture was able to effectively differentiate patients of favorable and unfavorable outcomes. The results also suggest that other EEG analysis tasks may benefit from our innovative model.

6. Conclusions

In this study, we propose a multimodal model defined by the combination of CNN and LSTM networks for the classification of EEG signals related to the prognosis of patients in coma. Our multimodal design allows the combination of EEG signals processed by CNN-LSTM layers with additional modalities, such as those related to patient information. To improve the performance of the classifier, we employ a GA to select the most relevant patient features and extracted EEG features, as well as to optimize the hyperparameters of the hybrid model.
The results obtained from the experiments indicate that incorporating patient features from their clinical records and/or TD features extracted from the EEG signals as additional information in the first dense layer of the CNN-LSTM model enhances the classifier’s performance. However, not all features contribute equally to this improvement. Therefore, utilizing the GA for feature selection enables the creation of a classifier with superior results compared to the several other mono- and multimodal models. Additionally, the GA is effective in finding the optimal CNN-LSTM hybrid architecture for classifying the EEG signals.
Another important contribution is the use of hundreds of EEG data collected from 60 Brazilian patients, which were visually analyzed and selected by expert neurologists to minimize noise in the signals. These procedures are crucial for achieving the statistical power and confidence needed in neurological studies []. From a medical standpoint, the results are statistically consistent and promising, warranting further investigation, as noted by the experts who contributed to this article.
The limitations of our study include: (i) the very simple TD features extracted from the EEG signals; and (ii) the dataset size from a deep learning perspective. As our experiments revealed, the extracted features do not contribute significantly to the outcome prediction of patients in coma. To address such a limitation, future works should explore more sophisticated features obtained through signal processing to broaden the scope of analysis. Although our dataset comprises more patients than most related EEG datasets in the literature, expanding its size could further enhance the training of deep learning algorithms. Therefore, we will persist in our efforts to collect additional data for future research.
Regarding future works, the proposed model is general and can be applied to other problems involving the classification of EEG signals, with plans to perform a more detailed statistical analysis to rigorously compare the classification performance across multiple diverse datasets. Therefore, its application to medical problems remains a highly relevant research topic. Additionally, forthcoming studies will explore explainable artificial intelligence (XAI) methods to provide interpretability for the CNN-LSTM model’s decisions. Although some efficient machine learning techniques, such as CNNs, lack intrinsic mechanisms for result explanation, interpretability is essential when applying these models in sensitive domains like medicine and healthcare.
Future efforts will also evaluate advanced optimization algorithms, such as the Coyote Optimization Algorithm, which has demonstrated promising results in various optimization applications. In addition, the effect of demographic variables including sex, age, and coma etiology on the training and validation processes will be evaluated to enhance the robustness of the model. A detailed neurophysiological interpretation and discussion of the electrodes that showed greater relevance in the classification will also be conducted to better understand their significance.

Author Contributions

Conceptualization, L.Z. and R.T.; methodology, M.G.C., L.Z. and R.T.; software, S.B.-J.; validation, S.B.-J., M.G.C. and R.T.; investigation, S.B.-J., M.G.C. and J.L.M.B.; data curation, J.B.D.-F. and M.C.; writing—original draft preparation, S.B.-J.; writing—review and editing, M.G.C., J.L.M.B., L.Z., J.B.D.-F. and R.T.; visualization, S.B.-J., M.G.C. and R.T.; supervision, R.T.; project administration, R.T.; funding acquisition, M.G.C., L.Z., J.B.D.-F. and R.T. All authors have read and agreed to the published version of the manuscript.

Funding

This work was carried out at the Center for Artificial Intelligence (C4AI-USP), with support from the São Paulo Research Foundation (FAPESP grants 2019/07665-4, 2021/09720-2, 2024/02783-7 and 2024/15430-5) and from the IBM Corporation. Authors also thank the financial support given by the Brazilian National Council for Scientific and Technological Development—CNPq (grant number 420212/2023-0), and by the National Institute of Artificial Intelligence (IAIA) of CNPq (grant number 406417/2022-9).

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by Ethics Committee of the Federal University of Uberlândia (protocol number 369/11).

Data Availability Statement

Data may be made available upon request due to privacy restrictions.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

This is the Appendix of the paper entitled “A Bioinspired Multimodal CNN-LSTM Network for EEG Analysis of Patients in Coma”, submitted to Sensors. The document is organized as follows: Appendix A.1 presents an overview about the hyperparameters considered by our CNN-LSTM-GA encoding; Appendix A.2 shows the optimal configuration of hyperparameters found by model; Appendix A.3 presents the results obtained by CNN-LSTM-GA in each one of the 10 GA runs; Appendix A.4 presents experimental results in terms of complementary metrics; and Appendix A.5 shows results obtained by other mono- and multimodal techniques.

Appendix A.1. Hyperparameters Encoding in CNN-LSTM-GA

Table A1 shows the parameters that can be selected by the GA at each position of the designed encoding.
Table A1. Range of values that can be adopted in each encoding position of a GA individual.
Table A1. Range of values that can be adopted in each encoding position of a GA individual.
DescriptionParameterValuesType
Activation function of convolutional layersactivation_CNN_layers [ relu , tanh ] string
Activation function of dense layersactivation_Dense_layers [ relu , tanh ] string
Dropout ratedropout [ 0.0 , 0.2 , 0.3 , 0.4 , 0.5 ] float
Number of neurons in lstm layerlstm_units [ 0 , 8 , 16 , 32 , 64 ] integer
Number of neurons in dense layerdense_units [ 0 , 32 , 64 , 128 , 256 ] integer
The number of output filters in the convolutionfilter [ 0 , 32 , 64 , 128 , 256 ] integer
The length of the 1D convolution windowkernel_size [ 2 , 6 , 9 , 13 ] integer
Number of epochs to train the modelepochs [ 5 , 10 , 15 , 20 ] integer
Optimizer used for training the modeloptimizer [ adam , sgd ] string
Learning ratelearning_rate [ 0.0001 , 0.001 , 0.01 , 0.05 , 0.1 ] float
Scaling method for the featuresscaler[‘standard’, ‘minmax’]string
Momentummomentum [ 0.3 , 0.5 , 0.7 , 0.9 ] float
Refers to adding additional values at the beginning and/or end of a 1D input sequence to preserve its length and/or prevent information losspadding[‘same’, ‘causal’]string
Size of the max pooling windowpool_size [ 0 , 2 , 3 , 4 , 5 ] integer
The corresponding feature is selected (True) or not (False). There is an element of the chromosome for each feature.feature[’true’, ’false’]boolean
A value of 0 indicates the absence of the respective layer.

Appendix A.2. Optimal Configuration of Hyperparameters for CNN-LSTM-GA

Table A2 presents the optimized architecture and hyperparameters of the CNN-LSTM-GA model in the coma outcome classification experiment. The model was trained using the binary_accuracy metric function and the loss binary_crossentropy function.
Table A2. Architecture and parameters of the proposed CNN-LSTM-GA model in the coma outcome classification experiment. The model was trained for 20 epochs, using the Adam optimizer with lr = 0.001 and the external features data was normalized using standard scaling. The learning and architecture parameters were found by the GA.
Table A2. Architecture and parameters of the proposed CNN-LSTM-GA model in the coma outcome classification experiment. The model was trained for 20 epochs, using the Adam optimizer with lr = 0.001 and the external features data was normalized using standard scaling. The learning and architecture parameters were found by the GA.
No.LayerHyperparameters
0Input[signal, 20]
1Conv1Dfilters = 32, kernel_size = 6, padding = same, activation = TanH
2Dropoutrate = 0.5
3Conv1Dfilters = 256, kernel_size = 2, padding = same, activation = TanH
4MaxPooling1Dpool_size = 5, strides = 5
5Dropoutrate = 0.2
6MaxPooling1Dpool_size = 5, strides = 5
7Dropoutrate = 0.5
8Conv1Dfilters = 128, kernel_size = 2, padding = causal, activation = TanH
9MaxPooling1Dpool_size = 2, strides = 2
10Dropoutrate = 0.5
11Dropoutrate = 0.5
12LSTMunits = 16, return_sequences = True
13Dropoutrate = 0.3
14LSTMunits = 16, return_sequences = True
15GlobalAveragePooling1D-
16Flatten-
17Concatenate(Flatten output + 38 External features)
18Denseunits = 32, activation = TanH
19Dropoutrate = 0.2
20Denseunits = 1, activation = Sigmoid

Appendix A.3. CNN-LSTM-GA Results

Table A3 shows the results obtained by the CNN-LSTM-GA in each of the 10 runs of the GA for the coma outcome classification task.
Table A3. Results obtained by the CNN-LSTM-GA in each of the 10 runs of the GA in the coma outcome classification experiment. The results from the best model are in bold.
Table A3. Results obtained by the CNN-LSTM-GA in each of the 10 runs of the GA in the coma outcome classification experiment. The results from the best model are in bold.
RunDatasetAccuracyF1 Score (Macro)PrecisionRecall
180% Train/20% Test0.8330.8330.8330.833
Cross-Validation0.917 ± 0.0750.913 ± 0.0790.950 ± 0.1000.900 ± 0.133
280% Train/20% Test0.7500.7270.8000.677
Cross-Validation0.867 ± 0.0410.855 ± 0.0490.950 ± 0.1000.800 ± 0.125
380% Train/20% Test0.7500.7270.8000.667
Cross-Validation0.867 ± 0.0410.855 ± 0.0490.950 ± 0.1000.800 ± 0.125
480% Train/20% Test0.7500.7270.8000.667
Cross-Validation0.833 ± 0.0850.886 ± 0.0770.905 ± 0.1310.900 ± 0.133
580% Train/20% Test0.7500.7270.8000.667
Cross-Validation0.833 ± 0.0410.868 ± 0.0560.971 ± 0.0570.800 ± 0.125
680% Train/20% Test0.8330.8330.8330.833
Cross-Validation0.867 ± 0.0670.865 ± 0.0730.881 ± 0.1030.867 ± 0.125
780% Train/20% Test0.7500.7270.8000.667
Cross-Validation0.833 ± 0.0410.868 ± 0.0560.971 ± 0.0570.800 ± 0.125
880% Train/20% Test0.8330.8330.8330.833
Cross-Validation0.833 ± 0.0670.876 ± 0.0770.921 ± 0.1020.867 ± 0.163
980% Train/20% Test0.8330.8330.8330.833
Cross-Validation0.917 ± 0.0530.922 ± 0.0460.893 ± 0.0960.967 ± 0.067
1080% Train/20% Test0.7500.7270.8000.667
Cross-Validation0.850 ± 0.0820.857 ± 0.0790.843 ± 0.1290.900 ± 0.133
Mean80% Train/20% Test0.783 ± 0.0430.769 ± 0.0550.813 ± 0.0170.733 ± 0.086
Cross-Validation0.877 ± 0.0260.877 ± 0.0240.924 ± 0.0420.860 ± 0.058

Appendix A.4. Stratified 5-Fold Cross-Validation Results with Complementary Metrics

Table A4 presents the results obtained in the coma outcome classification experiment for mono- and multimodal models in stratified five-fold cross-validation, including the CNN-LSTM-GA. The table provides the following metrics based on stratified five-fold cross-validation: mean, standard deviation, Geometric Mean (G-mean), maximum, and minimum values. These metrics were calculated using the results obtained from the test set of each fold.
Table A4. Results obtained by the mono- and multimodal models for coma outcome classification considering a stratified 5-fold cross-validation.
Table A4. Results obtained by the mono- and multimodal models for coma outcome classification considering a stratified 5-fold cross-validation.
ModelCross
Validation
AccuracyF1 Score (Macro)PrecisionRecall
EEGTransformer []Mean and std0.88±0.090.88±0.090.97±0.060.80±0.19
Max1.001.001.001.00
G-Mean0.880.870.970.77
Min0.750.730.860.50
MLP w/Patient FeatsMean and std0.67 ± 0.090.65 ± 0.130.68 ± 0.090.63 ± 0.16
Max0.830.830.830.83
G-Mean0.660.630.680.61
Min0.580.440.570.33
MLP w/TD FeatsMean and std0.52 ± 0.130.49 ± 0.190.49 ± 0.140.50 ± 0.24
Max0.670.710.670.83
G-Mean0.500.450.470.43
Min0.330.220.330.17
CNN-LSTM-baselineMean and std0.83 ± 0.050.82 ± 0.050.91 ± 0.120.77 ± 0.08
Max0.920.911.000.83
G-Mean0.830.820.900.76
Min0.750.770.710.67
CNN-LSTM w/Patient FeatsMean and std0.83 ± 0.090.83 ± 0.090.87 ± 0.120.80 ± 0.13
Max0.920.921.001.00
G-Mean0.830.820.860.79
Min0.670.680.680.68
CNN-LSTM w/TD FeatsMean and std0.80 ± 0.070.80 ± 0.060.85 ± 0.130.77 ± 0.08
Max0.920.911.000.83
G-Mean0.800.790.840.76
Min0.750.730.710.67
MLP w/Patient + TD FeatsMean and std0.65 ± 0.130.64 ± 0.160.65 ± 0.140.63 ± 0.19
Max0.830.830.830.83
G-Mean0.640.610.640.60
Min0.500.400.500.33
CNN-LSTM w/Patient + TD FeatsMean and std0.82 ± 0.120.81 ± 0.130.85 ± 0.130.77 ± 0.13
Max1.001.001.001.00
G-Mean0.810.800.840.76
Min0.670.670.670.67
CNN-LSTM-GA (Proposed)Mean and std0.92 ± 0.050.92 ± 0.050.89 ± 0.100.97 ± 0.07
Max1.001.001.001.00
G-Mean0.920.920.890.96
Min0.830.860.750.83

Appendix A.5. Results of Other Mono- and Multimodal Models

Table A5 presents the results obtained in the coma outcome classification experiment for the Random Forest and Support Vector Machine (SVM) in stratified five-fold cross-validation. The table provides the following metrics based on stratified five-fold cross-validation: mean, standard deviation, Geometric Mean (G-mean), maximum, and minimum values. These metrics were calculated using the results obtained from the test set of each fold.
Table A5. The results obtained by other mono and multimodal models for predicting coma outcome in stratified 5-fold cross-validation.
Table A5. The results obtained by other mono and multimodal models for predicting coma outcome in stratified 5-fold cross-validation.
ModelCross
Validation
AccuracyF1 Score
(Macro)
PrecisionRecall
Random Forest w/Patient FeatsMean and std0.72 ± 0.090.68 ± 0.140.74 ± 0.060.67 ± 0.21
Max0.830.860.801.00
G-Mean0.710.670.730.63
Min0.580.440.670.33
Random Forest w/TD FeatsMean and std0.48 ± 0.080.50 ± 0.120.48 ± 0.080.53 ± 0.19
Max0.580.630.570.83
G-Mean0.480.480.470.50
Min0.330.330.330.33
Random Forest w/Patient + TD FeatsMean and std0.60 ± 0.100.55 ± 0.130.62 ± 0.130.50 ± 0.15
Max0.750.730.800.67
G-Mean0.590.530.610.48
Min0.500.400.500.33
SVM w/Patient FeatsMean and std0.65 ± 0.100.65 ± 0.090.67 ± 0.120.63 ± 0.07
Max0.750.730.800.67
G-Mean0.640.640.660.63
Min0.500.500.500.50
SVM w/TD FeatsMean and std0.42 ± 0.120.45 ± 0.180.40 ± 0.120.53 ± 0.29
Max0.580.710.551.00
G-Mean0.400.410.380.45
Min0.250.180.200.17
SVM w/Patient + TD FeatsMean and std0.62 ± 0.100.61 ± 0.080.64 ± 0.130.60 ± 0.08
Max0.750.730.800.67
G-Mean0.610.610.630.59
Min0.500.500.500.50

References

  1. Müller-Putz, G.R. Electroencephalography. In Handbook of Clinical Neurology; Elsevier: Amsterdam, The Netherlands, 2020; Volume 168, pp. 249–262. [Google Scholar]
  2. Melissant, C.; Ypma, A.; Frietman, E.; Stam, C. A method for detection of Alzheimer’s disease using ICA-enhanced EEG measurements. Artif. Intell. Med. 2005, 33, 209–222. [Google Scholar] [CrossRef]
  3. Klassen, B.T.; Hentz, J.G.; Shill, H.A.; Driver-Dunckley, E.; Evidente, V.G.; Sabbagh, M.N.; Adler, C.H.; Caviness, J.N. Quantitative EEG as a predictive biomarker for Parkinson disease dementia. Neurology 2011, 77, 118–124. [Google Scholar] [CrossRef]
  4. Acharya, U.R.; Subbhuraam, V.S.; Goutham, S.; Martis, R.; Suri, J. Automated EEG analysis of epilepsy: A review. Knowl.-Based Syst. 2013, 45, 147–165. [Google Scholar] [CrossRef]
  5. Siddiqui, M.M.; Srivastava, G.; Saeed, S.H. Diagnosis of insomnia sleep disorder using short time frequency analysis of PSD approach applied on EEG signal using channel ROC-LOC. Sleep Sci. 2016, 9, 186–191. [Google Scholar] [CrossRef]
  6. Ardeshna, N.I. EEG and Coma. Neurodiagn. J. 2016, 56, 1–16. [Google Scholar] [CrossRef]
  7. Lehmann, C.; Koenig, T.; Jelic, V.; Prichep, L.; John, R.E.; Wahlund, L.O.; Dodge, Y.; Dierks, T. Application and comparison of classification algorithms for recognition of Alzheimer’s disease in electrical brain activity (EEG). J. Neurosci. Methods 2007, 161, 342–350. [Google Scholar] [CrossRef]
  8. Shoeb, A.H.; Guttag, J.V. Application of machine learning to epileptic seizure detection. In Proceedings of the International Conference on Machine Learning (ICML 2010), Haifa, Israel, 21–24 June 2010. [Google Scholar]
  9. Qureshi, A.A.; Zhang, C.; Zheng, R.; Elmeligi, A. Ischemic Stroke Detection Using EEG Signals. In Proceedings of the 28th Annual International Conference on Computer Science and Software Engineering, Markham, ON, Canada, 29–31 October 2018; pp. 301–308. [Google Scholar]
  10. Ay, B.; Yildirim, O.; Talo, M.; Baloglu, U.B.; Aydin, G.; Puthankattil, S.D.; Acharya, U.R. Automated Depression Detection Using Deep Representation and Sequence Learning with EEG Signals. J. Med. Syst. 2019, 43, 205. [Google Scholar] [CrossRef]
  11. Lee, S.; Hussein, R.; McKeown, M.J. A Deep Convolutional-Recurrent Neural Network Architecture for Parkinson’s Disease EEG Classification. In Proceedings of the 2019 IEEE Global Conference on Signal and Information Processing (GlobalSIP), Ottawa, ON, Canada, 11–14 November 2019; pp. 1–4. [Google Scholar]
  12. Xu, G.; Ren, T.; Chen, Y.; Che, W. A One-Dimensional CNN-LSTM Model for Epileptic Seizure Recognition Using EEG Signal Analysis. Front. Neurosci. 2020, 14, 578126. [Google Scholar] [CrossRef] [PubMed]
  13. Hochreiter, S. The Vanishing Gradient Problem During Learning Recurrent Neural Nets and Problem Solutions. Int. J. Uncertain. Fuzziness-Knowl.-Based Syst. 1998, 6, 107–116. [Google Scholar] [CrossRef]
  14. Young, G.B. Coma. Ann. N. Y. Acad. Sci. 2009, 1, 32–47. [Google Scholar] [CrossRef] [PubMed]
  15. Sandroni, C.; Cariou, A.; Cavallaro, F.; Cronberg, T.; Friberg, H.; Hoedemaekers, C.; Horn, J.; Nolan, J.P.; Rossetti, A.O.; Soar, J. Prognostication in comatose survivors of cardiac arrest: An advisory statement from the European Resuscitation Council and the European Society of Intensive Care Medicine. Intensive Care Med. 2014, 40, 1816–1831. [Google Scholar] [CrossRef]
  16. Bissaro, L.Z.; Junior, O.O.N.; Destro Filho, J.B.; Jin, Y.; Carneiro, M.G. Towards the Prognosis of Patients in Coma using Echo State Networks for EEG Analysis. Procedia Comput. Sci. 2023, 222, 509–518. [Google Scholar] [CrossRef]
  17. Kachare, P.H.; Sangle, S.B.; Puri, D.V.; Khubrani, M.M.; Al-Shourbaji, I. STEADYNet: Spatiotemporal EEG analysis for dementia detection using convolutional neural network. Cogn. Neurodyn. 2024, 18, 3195–3208. [Google Scholar] [CrossRef]
  18. Pandey, S.K.; Janghel, R.R.; Mishra, P.K.; Ahirwal, M.K. Automated epilepsy seizure detection from EEG signal based on hybrid CNN and LSTM model. Signal Image Video Process. 2022, 17, 1113–1122. [Google Scholar] [CrossRef]
  19. Aszemi, N.M.; Aszemi, P. Hyperparameter Optimization in Convolutional Neural Network using Genetic Algorithms. Int. J. Adv. Comput. Sci. Appl. 2019, 10, 269–278. [Google Scholar] [CrossRef]
  20. Nikbakht, S.; Anitescu, C.; Rabczuk, T. Optimizing the neural network hyperparameters utilizing genetic algorithm. J. Zhejiang Univ.-Sci. A 2021, 22, 407–426. [Google Scholar] [CrossRef]
  21. Del-Lama, R.S.; Cândido, R.M.; Chiari-Correa, N.S.; Nogueira-Barbosa, M.H.; Azevedo-Marques, P.M.; Tinós, R. Computer-Aided Diagnosis of Vertebral Compression Fractures Using Convolutional Neural Networks and Radiomics. J. Digit. Imaging 2022, 35, 446–458. [Google Scholar] [CrossRef]
  22. Baldo-Júnior, S.; Carneiro, M.G.; Destro-Filho, J.B.; Zhao, L.; Tinós, R. Classification of coma etiology using convolutional neural networks and long-short term memory networks. In Proceedings of the 2023 International Joint Conference on Neural Networks (IJCNN), Gold Coast, Australia, 18–23 June 2023; pp. 1–8. [Google Scholar]
  23. Xue, B.; Zhang, M.; Browne, W.N.; Yao, X. A Survey on Evolutionary Computation Approaches to Feature Selection. IEEE Trans. Evol. Comput. 2016, 20, 606–626. [Google Scholar] [CrossRef]
  24. Hu, F.; He, K.; Wang, C.; Zheng, Q.; Zhou, B.; Li, G.; Sun, Y. STRFLNet: Spatio-Temporal Representation Fusion Learning Network for EEG-Based Emotion Recognition. IEEE Trans. Affect. Comput. 2025, 16, 1–16. [Google Scholar] [CrossRef]
  25. Barbosa, J.L.M.; Carneiro, M.G.; Baldo-Júnior, S.; Tinos, R.; Ji, D.; Zhao, L.; Destro-Filho, J.B. Attention to EEG signals: A transformer-based architecture for the Prognosis of Patients in Coma. In Proceedings of the International Joint Conference on Neural Networks-IJCNN, Rome, Italy, 30 June–5 July 2025. [Google Scholar]
  26. Ramos, C.D.; Destro-Filho, J.B. Measurements of coherence in EEG signal in brazilian people: A comparison of different consciousness states. Biosci. J. 2021, 37, 1981–3163. [Google Scholar] [CrossRef]
  27. Carneiro, M.G.; Ramos, C.D.; Destro-Filho, J.B.; Zhu, Y.; Ji, D.; Zhao, L. High-Level Classification for EEG Analysis. In Proceedings of the 2023 International Joint Conference on Neural Networks (IJCNN), Gold Coast, Australia, 18–23 June 2023; pp. 1–8. [Google Scholar]
  28. Schomer, D.L.; Lopes da Silva, F. Niedermeyer’s Electroencephalography: Basic Principles, Clinical Applications, and Related Fields, 6th ed.; Lippincott Williams & Wilkins: Philadelphia, PA, USA, 2012; ISBN 9781451122435. [Google Scholar]
  29. Chen, C.C.; Massey, S.L.; Kirschen, M.P.; Yuan, I.; Padiyath, A.; Simpao, A.F.; Tsui, F.R. Electroencephalogram-based machine learning models to predict neurologic outcome after cardiac arrest: A systematic review. Resuscitation 2023, 194, 110049. [Google Scholar] [CrossRef] [PubMed]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Article metric data becomes available approximately 24 hours after publication online.