Next Article in Journal
Ammonium Pyrrolidine Dithiocarbamate-Modified CdTe/CdS Quantum Dots as a Turn-on Fluorescent Sensor for Detection of Trace Cadmium Ions
Next Article in Special Issue
Enhancing BCI-Based Emotion Recognition Using an Improved Particle Swarm Optimization for Feature Selection
Previous Article in Journal
Improvement of Flow Velocity Measurement Algorithms Based on Correlation Function and Twin Plane Electrical Capacitance Tomography
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Wrapper Feature Selection Algorithm: An Emotional Assessment Using Physiological Recordings from Wearable Sensors

by
Inma Mohino-Herranz
1,*,
Roberto Gil-Pita
1,
Joaquín García-Gómez
1,
Manuel Rosa-Zurera
1 and
Fernando Seoane
2,3,4
1
Department of Signal Theory and Communications, University of Alcalá, Alcalá de Henares, 28805 Madrid, Spain
2
Institute for Clinical Science, Intervention and Technology, Karolinska Institutet, 17177 Solna Stockholm, Sweden
3
Department of Medical Care Technology, Karolinska University Hospital, 14157 Huddinge, Sweden
4
Textile Materials Technology, Department of Textile Technology, Faculty of Textiles, Engineering and Businees Swedish School of Textiles, University of Boras, 50190 Boras, Sweden
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(1), 309; https://doi.org/10.3390/s20010309
Submission received: 28 November 2019 / Revised: 29 December 2019 / Accepted: 3 January 2020 / Published: 6 January 2020
(This article belongs to the Special Issue Emotion Monitoring System Based on Sensors and Data Analysis)

Abstract

:
Assessing emotional state is an emerging application field boosting research activities on the topic of analysis of non-invasive biosignals to find effective markers to accurately determine the emotional state in real-time. Nowadays using wearable sensors, electrocardiogram and thoracic impedance measurements can be recorded, facilitating analyzing cardiac and respiratory functions directly and autonomic nervous system function indirectly. Such analysis allows distinguishing between different emotional states: neutral, sadness, and disgust. This work was specifically focused on the proposal of a k-fold approach for selecting features while training the classifier that reduces the loss of generalization. The performance of the proposed algorithm used as the selection criterion was compared to the commonly used standard error function. The proposed k-fold approach outperforms the conventional method with 4% hit success rate improvement, reaching an accuracy near to 78%. Moreover, the proposed selection criterion method allows the classifier to produce the best performance using a lower number of features at lower computational cost. A reduced number of features reduces the risk of overfitting while a lower computational cost contributes to implementing real-time systems using wearable electronics.

1. Introduction

In the literature, numerous studies show that the human body generates information regarding emotional state continuously. Often, people express their feelings verbally and through body gestures. However, emotional state and mental and physical load influence the individual at a physiological level, and such influences can be reflected in the parasympathetic activity of the autonomous nervous system (ANS). Its parts are the sympathetic nervous system (SNS) and the parasympathetic nervous system (PSNS). The main function of the SNS is to mobilize the body’s nervous system in what is known as the fight-or-flight response [1,2]. It is a physiological reaction produced by a perceived harmful event, attack, or threat to survival. The PSNS has its roots in the brain stern and spinal cord in the lower back. Its function is to bring back sanity after an emergency situation that led to the SNS taking control. Certain reactions are provided by SNS and PSNS in such organs as the eyes, heart, lungs, blood vessels, sweat glands, and digestive tract. When communicating human emotions, there are several tools involved, which humans use willfully and unconsciously to communicate their own emotions or feelings. Apart from gestures and speech, emotions modify some uncontrollable physiological signals produced in the human body; some can be read by the electrocardiogram (ECG), thoracic electrical bioimpedance (TEB), and skin conductance.
ECG signal analysis is widely used in medicine, since it offers information regarding diseases and disorders affecting the cardiac function. Several studies [3,4,5,6,7] conclude that emotions clearly affect the cardiac activity.
The TEB measurement is a non-invasive technique used to monitor the dynamics of the thoracic activity. This is directly related to the breathing. In the literature, it is possible to find multiples views about the relationship between emotions and breathing activity. A long time ago, in [8], the relationships between emotions and breathing and the possibility of controlling a anxiety and panic attacks through breathing were studied. Across many years it is possible to find numerous articles that study such relationship [9,10].
TEB has also been related to emotions in [11], wherein it was used to study the hemodynamic reactions to road traffic noise in young adults. The thoracic bioelectrical impedance is related to stress induced by traffic noise. Moreover, several studies have demonstrated that the analysis of ECG and TEB recordings can indeed provide information about the emotional status of individuals [12,13].
To register the red cited physiological signals, there are several types of devices. The proliferation of wearable sensing devices [14,15,16], such as smart watches and wrist and chest-bands, combined with the communication and data processing capabilities of smartphones, is boosting the potential for pervasive monitoring.
Emotional analysis with recorded ECG or TEB signals is understood as a classification problem, implemented with several steps: (a) meaningful features are calculated from the signals; (b) the vector of features is applied as input to a classifier, whose outputs indicate a person’s emotions. Meaningful feature selection is one the main tasks in classification for several reasons: (a) models can be simplified to make them easier to be interpreted; (b) the models can be trained in shorter times; (c) it helps to avoid the curse of dimensionality, enhancing generalization by reducing overfitting.
The main objective of this paper is to select the most valuable and efficient features to distinguish three emotional states—sadness, disgust, and neutral—using the physiological signals aforementioned.
The problems of generalization are given by several factors in emotional recognition. Generalization is a measure of how accurately an algorithm is able to classify previously unseen data. The capacity of generalization is related to the number of parameters to be tuned during training and the number of samples available for training the classifier. Generalization error can be minimized by avoiding overfitting in the learning algorithm. A common strategy to reduce the number of parameters to be tuned, so as to reduce the possibility of overfitting, is feature selection. It allows reducing the dimension of the input vector, and consequently, the complexity of the system.
In the literature, numerous algorithms to improve generalization can be found; for example, using neural networks [17] or combining genetic algorithms (GA) with the hybrid learning algorithm (HLA) [18]. A study of generalization capabilities of SVMs trained with k-fold cross validation was carried out in [19].
In this paper, an algorithm is proposed whose main objective is to improve the selection of features to minimize these undesirable generalization problems. We compared two different fitting functions aimed to reduce the loss of generalization: one of them is based on the use of a standard error function as the selection criterion and the other is the proposed k-fold-based fitting function.

2. System Overview

2.1. System Architecture

A functional work-flow overview of the system is depicted in Figure 1. The system can be included in the known passive brain-computer interface (pBCI). These systems register biosignals for assessment of emotional and mental workload. In [20], a review is presented about the pBCI with the current trends and future directions. These systems are the interface between humans and computers through signal acquisition, feature extraction, and finally, feature selection.
The first block represents the feature extraction task, where the features used in this study were calculated in order to extract the information from biosignals. The second block, the feature selection block, aims at selecting the most efficient features, which in this paper was performed using an evolutionary algorithm and a quadratic classifier. Finally, the classification block—where the classifier produces outputs to classify the input signals minimizing an objective function, such as the mean squared error. The classification task is implemented in both the feature selection block and the classification block. In both cases, the classifier selected was a quadratic classifier; concretely, the least square diagonal quadratic classifier (LSDQC), which is explained in detail in [21].
The Feature Selection block was based on wrapper methods, which were proposed in [22]. These models require a predetermined learning algorithm to select features and use their performance to evaluate and determine which features are selected. Usually, wrapper methods can generate potential feature subsets with high accuracy because of the good matching of that subset of features with the learning algorithms. The disadvantage comes from the large computational cost, due to the use of the classifier for each subset of features. The concept of wrapper approach is presented in Figure 2. A deeply studied variant of the wrapper model is the randomized one, which relies on search strategies such as genetic algorithms (GAs) [23], hill climbing [24], and simulated annealing [25]. In some cases, part of the classifier design process (tuning of some parameters) is embedded within feature selection in the optimization process [26].
The training and validation processes are carried out offline. Once the best features are obtained, which are selected based on performance and computational cost, the real-time system only calculates the features that were selected. With these features and the optimized model, it is possible to assign the pattern to a class.

2.2. Feature Extraction

The ECG and TEB signals are analyzed to generate several characterization features, such as time analysis, QRS complex analysis or heart rate variability (HRV) [27,28], and the power ratios in different frequency bands [29,30].
In order to include information about time variation, several temporal averages of calculated features are included in the features set. In the present work, the features are calculated similarly to [31], as it is shown in Figure 3. The scheme shows a block diagram, where it is possible to distinguish the different functional operations implemented on the biosignal recordings to extract the features.
In the present work, the features were determined every 10 s. The choice of this value is a compromise between the minimum time to calculate the features and the implementation in real time. Since the feature extraction is expected to be implemented in a smartphone, it is necessary to calculate the computational complexity associated with each feature.
The blocks Low Frequency (LF), Respiration Frequency (RF), Medium Frequency (MF), and Electrocardiogram Frequency (EF) are filters.
LF and MF are anti-aliasing filters, which allow the use of interpolated finite impulse response (IFIR) filters [32]. The outputs of these anti-aliasing filters are applied to two different IFIR filters, RF and EF, with a stretch factor SF = 25 for the ECG measurement and SF = 10 for the TEB measurement.
For calculating the computational cost for each filter, we used the following reasoning: The cost of the anti-aliasing filter was obtained as C A N T = N · F , N being the filter order and F the sampling frequency of the measurement used (ECG or TEB). Analogously, the cost associated with the IFIR filter is calculated with C I F I R = N · F S F .
In Table 1 it is possible to observe the kind of filters used, the cut-off frequencies, the order, and computational cost.
The outputs of the filters are applied to different blocks denominated BPM (breaths per minute) and PPM (pulsations per minute). The block BPM considers the signal to be sine-shaped, and it periodically determines the minimum and the maximum values of the signal, allowing estimation of the number of breaths per minute and its amplitude.
The block PPM obtains the pulsations per minute. The outputs of BPM and PPM blocks are used by the next block, denominated Inter, to generate interpolated signals, resulting from piece-wise constant interpolation to the last known values of the number of breaths per minute, its amplitude, or the number of beats per minute. The sampling frequency of the generated signals after interpolation is 50 Hz. The associated computational cost analysis is included in Section 4.3.
The block parameters represent the calculations of 14 statistical parameters (statistics), which are: trimmed mean of 25%, median, percentile 25%, percentile 75%, kurtosis, skewness, standard deviation, mean absolute deviation, mean absolute deviation, geometric mean, harmonic mean, baseline, maximum, minimum, and mean. Finally, we obtain three signals from each measurement. Since the number of parameters used is 14, the total number of features is 84, 42 from each measurement. The computational cost calculated for these blocks is shown in the results section.

2.3. Emotion Classifier

In feature selection using GA, the classifier used to test the performance with the selected features at each stage must be simple for the algorithm to be feasible. The simplest classifier is a linear classifier, but it only implements linear boundaries to separate the classes. In this study, we relied on the least square diagonal quadratic classifier [21], which can be used to implement closed boundaries, and at the same time is quite simple.
Let us consider a set of training patterns x = [ x 1 , x 2 , , x L ] T , where each of these patterns is assigned to one of the possible classes denoted as C m m = 1 , , M . , M being the number of classes. In a quadratic classifier, the decision rule can be obtained using a set of k combinations, as shown in Equation (1).
y m = w m 0 + n = 1 L w m n x n + n = 1 L v n m x n 2 .
The pattern matrix Q , which contains the input features for classification and their quadratic values, is expressed in Equation (2).
Q = 1 1 1 1 x 11 x 12 x 13 x 1 N x L 1 x L 2 x L 3 x L N x 11 2 x 12 2 x 13 2 x 1 N 2 x L 1 2 x L 2 2 x L 3 2 x L N 2 .
Defining V , the weights matrix, with Equation (3).
V = w 10 w 11 w 1 L v 11 v 1 L w M 0 w M 1 w M L v M 1 v M L .
The output of the quadratic classifier is obtained in Equation (4).
Y = V · Q .
The target matrix, which contains the labels of each patterns is defined as:
T = t 11 t 12 t 13 t 1 N t M 1 t M 2 t M 3 t M N ,
where N is the number of data samples, and t m n = 1 if the n-th pattern belongs to class C m , and 0 otherwise. Then, the error is the difference between the outputs of the classifier and the correct values, which are contained in the target vector:
Consequently, the mean squared error is defined with Equation (6).
M S E = 1 N Y - T 2 = 1 N V · Q - T 2 .
In the least squares approach, the weights are adjusted in order to minimize the mean squared value of this error (MSE). The matrix V which minimizes Equation (6) is obtained with the Wiener–Hopf equation:
V = T · Q T · Q · Q T 1 .
This expression allows us to determine the values of the coefficients that minimize the mean squared error for a given set of features.
To avoid the loss of generalization in the results while maximizing the accuracy in the estimation of the error rate, k-fold cross-validation was used in the experiments, with k set equal to S, the number of subjects available in the design database. Thus, the data were divided into k folds or subsets containing data from each subject, and each time, the registers from one given subject were used as a test set, with the data from the k 1 subject used for the training task.
Therefore, Q | k = s stands for the sub-matrix of Q containing the data of the s-th fold; that is, data from the s-th subject of the database. Thus, in the k-fold validation process, the output of the classifier for the s-th individual is obtained using data from the remaining individuals; that is, Q | k s . Thus, taking into account Equations (4) and (7), we obtain the output of the classifier for the s-th subject of the database as follows:
Y | k = s = V | k s · Q | k = s = T | k s Q | k s T ( Q | k s · Q | k s T ) 1 · Q | k = s .
Once this output is obtained, the classification decision is made, determining which term of the output is maximum. The classification error rate is then estimated as the average of the k-folds classification errors.

3. Proposed Feature Selection Method

For the feature selection task, the algorithm selected was a tailored version of a genetic algorithm (GA) [33]. The performance of GAs is based on the application of evolutionary laws (crossover, mutation, and selection of the fittest or elitism) to an initial generic population of possible solutions to meta-heuristically find a good solution to the problem. With these algorithms, a set of features can be found out as the solution of an optimization problem with the objective of minimizing the classification error with a constraint related to the number of operations, N o p . Since the application of GA requires the calculation of the classifier error for each combination of features in the current population of feature sets, the classifier should not be complex in order to reduce the computational complexity of the solution. For this reason, quadratic classifiers were used; they combine reduced computational complexity and reasonable performance.
To understand the procedure used, it is important to note that to determine the set of features, we took into account both the performance of the classifier (in terms of error rate) and the computational complexity. For simplicity, we considered the number of simple operations per second ( N o p ) as an indicator of the computational complexity of the real-time implementation of a given set of features, since this value is proportional to the CPU load of the final implementation.
As it was stated above, the selection of the fitting function in the GA-based feature selection process is a critical issue, since its choice may cause generalization problems, due to the small number of available data. To avoid generalization problems risen from the limited size of the dataset, in this paper, two fitting functions were compared: a standard MSE-based function, and a novel k-fold-based fitting function, which aims at increasing the generalization capability of the feature selection process.

3.1. Standard Design Error Optimization (SDEO)

In a first approach, we select the set of features that minimize the mean squared error over the design set. At this point we have to consider that, as it was stated above, we use a k-fold validation using the different individuals of the database as folds (k must match S, the number of individuals in the database). In this validation method, the database is divided into k folds, one for each individual in the dataset. The design set for the s-th iteration of the k-fold validation process is denominated, at this point, Q | k s (since it contains all data except those patterns included in the s-th fold), and the test set for the s-th iteration of the k-fold validation process is denominated Q | k = s (containing only those patterns included in the s-th fold). The MSE of the k-fold validation process can be calculated using Equation (9).
M S E = 1 S s = 1 S 1 N s V | k s · Q | k s T | k = s 2 .
It is worth mentioning that feature selection is part of training; therefore, in order to guarantee generalization of the results, the test set cannot be used in order to select the best features. Thus, the SDEO for the s-th iteration of the k-fold validation process is exclusively obtained using the design data of that iteration; that is, the mean squared error over the design set is expressed using Equation (10).
S D E O s = 1 i s N i V | k s · Q | k s T | k s 2 .

3.2. k-Fold-Based Error Optimization (KFBEO)

Taking into account that the database does not contain a large number of data from a large number of individuals, the SDEO might present loss of generalization in the selection of the features. That is, the features selected in each iteration of the k-fold process may be especially good for the given design set. In order to overcome generalization loss, we propose the use of an additional k-fold optimization process for feature selection. In the design test, to find out the select features, the output is obtained for systems trained with the folds in the design set, but one is left out which has input equal to that fold that was left out. The KFBEO proposed to be minimized in order to determine the selected features for the s-th iteration of the k-fold validation process will be consequently given by Equation (11), by averaging over the S 1 folds (being S the number of subjects in the database).
K F B E O s = 1 n = 1 n s S N n n = 1 n s S | | V | k ( n , s ) · Q | k = n T | k = n | | 2 ,
where V | k ( n , s ) are the weights of the quadratic classifier, determined using Equation (7) over the database excluding both the s-th subject (the test subject of the main k-fold validation process), and the n-th subject (the design subject of the k-fold process related to the evaluation of the KFBEO fitting function).

4. Experiments and Results

4.1. Database Description

The database consists of the ECG and TEB signals recorded from 40 subjects, while the subjects watched certain sequences of films whose objective was to elicit the different emotions under study, neutral emotion, sadness, and disgust. The sample frequency was 100 Hz for TEB and 250 Hz for ECG. The database is described in detail in [31] and it can be downloaded from https://www.mdpi.com/1424-8220/19/24/5524 associated with [34]. A brief description of the database can be found in Appendix A.

4.2. Parameters Used for Evaluation

To quantify the error and improvement obtained using the proposed algorithm, several parameters are used:
  • Computational cost.
  • Performance of the classifier measured with the MSE. This error can be calculated when features are selected using any one of the proposed methods. In the present case, the database is balanced, which means that the number of patterns of each class is the same. For this reason, it is enough to analyze the accuracy, since there is no possibility to fall into the accuracy paradox [35].
  • Ranking of features. It is very important to know which features provide more valuable information so as to reduce calculation time and computational cost.
  • Power consumption in a specific smartphone. The system has been implemented in a smartphone, in order to compare the required power consumption with each possible solution, when it is implemented in a embedded device.

4.3. Results

4.3.1. Computational Cost

The computational cost obtained for the execution of algorithms and the extraction of features is reported as follows. Table 2 lists the computational cost obtained for each of the implemented filters (LF, MF, RF, and EF), while for the algorithms PPM and BPM the estimated computational costs are 9050 and 8800 operations per second, respectively (the explanation about the way to estimate the computational cost is part of Section 2.2).
Some of the features are obtained by calculating statistics over some primary features. The number of operations required to calculate these statistics must also be considered to evaluate the complexity of our solution. The computational cost for the calculation of these statistics is shown in Table 3.

4.3.2. Classification Performance

Figure 4 contains the violin plots for the classification performance obtained by each of the algorithms. The obtained classification error is plotted in function of the computational cost expressed as the maximum number of operations per second required using the features selected by each of the algorithms. The classification errors are presented as means of 100 estimations, and for the sake of clarity, they were normalized for the figure (the most valuable information is the mean value and how the results spread around the mean value). The minimum error probability obtained using the SDEO is approximately 26.5% (mean value), for 50,000 operations per second, using 35 features. The minimum mean error probability with the proposed KFBEO method for feature selection is around 22.5% for 50,000 operations per second with 10 features.

4.3.3. Ranking of Selected Features

The label of each feature was constructed from the name of the biosignal and the algorithm used: ECG_PPM, ECG_RD, ECG_RT, TEB_BPM, TEB_RD, and TEB_RT. Said names are followed by the names of the parameter or statistic used, respectively. Feature selection has been repeated 100 times, being 40-fold. The number of attempts to estimate the percentage of positive feature selections is the product of k and the number of the repetitions (100).
Table 4 presents the features that have been selected in more than 20% of the experiments. In this case, 60 of the 84 features were selected at least once.
The number of features selected at least once using the KFBEO is 36. And number of features selected more than 20% of the attempts is five; they are shown in Table 5.
The number of features selected for obtaining the best result for the SDEO is 35 features, meanwhile for the KFBEO, that number is 10 features. That is the reason for which the computational cost of KFBEO is lower than the computational cost of SDEO.

4.3.4. Power Consumption

Feature extraction was implemented in a Galaxy Pocket smartphone, manufactured by Samsung Electronics, Co., running with an 832 MHz CPU, 512 MB, and a 1200 mAh battery. In order to assess whether the system was suitable for real-time applications using the given smartphone, the application was executed for four hours of full performance, during which the battery level decreased an additional 10% when compared with the standard battery usage with the smartphone operating in sleep mode. The battery consumption was mainly due to Bluetooth. Regarding resource management, the app running the feature extraction used less than 6% of the CPU resources.

5. Discussion

Nowadays, the analysis of biological signals to determine the emotional state in real-time is a promising research field [36,37,38]. Through the analysis of electrocardiogram and thoracic impedance recordings obtained with wearable sensors to monitor cardiac and breathing activity, respectively, it is possible to distinguish among three different emotions: neutral emotion, sadness, and disgust. In the study of the ECG and TEB signals, numerous characterizing features can be obtained, carrying information useful for different purposes. Working with a large number of features in classification applications might impact in the generalization ability of the obtained classifiers, resulting in systems with sub-optimal performances when applied in real scenarios [39].
The classifier which uses the features selected with the k-fold-based approach achieves a lower probability of error than the classifier which uses the features selected with the standard fitting function. This is due to an improvement in generalization during training, because the database for training does not contain a large amount of data from many individuals. The use of the SDEO produces loss of generalization.
In this work, we aimed at selecting the best features targeting both low error rate and reduced computational complexity. A wrapper feature selection algorithm based on GA has been used [40,41]. In these cases, the selection of the fitting function in the GA-based wrapper feature selection process is of paramount importance, since a wrong choice may lead to generalization issues. In cases with a small amount of data available, the selection of the fitting function becomes critical.
As an alternative approach, we propose applying a k-fold technique to the evaluation of the fitting function, each fold corresponding to one subject of the design set. The results show that the KFBEO renders a lower error probability than the SDEO. The KFBEO renders an improvement, which is around 3% in error probability using the same number of operations per second. Moreover, the error obtained is in all cases lower using the KFBEO.
Besides, the number of features selected for obtaining the best result for the SDEO is near to 35, but for the proposed KFBEO is only 10 features. The number of features is likely to affect the performance of the classifier, since increasing the number of features also increases the possibility of over-fitting [42]. Concerning the operations per second required to obtain the best result, the number is remarkably lower for KFBEO than SDEO, which implies that the computational cost is smaller with KFBEO than with SDEO.

6. Conclusions

Comparing the results rendered by the emotions classification system using the features selected with SDEO an KFBEO, we can conclude that the error probability provided by the classifier, which uses the features selected with the KFBEO, is notably better than the error probability provided by the classifier, which uses the features selected with SDEO. In addition, as the number of features selected with KFBEO-based approach is lower, the resulting computational complexity is lower too.
Comparing classifiers with a similar number of operations fed with features selected with both methods, an improvement of 3% in error classification is obtained.
According to the number of features needed for obtaining the best results in each algorithm, the SDEO selects 35 features; meanwhile, the KFBEO only requires the selection of 10. This implies that the classifier trained with the KFBEO presents a higher tolerance to overfitting, since less features are used than when using SDEO fitting function for feature selection. The number of features selected is related to the computational cost, and therefore, to the battery consumption. The implementation of the proposed classification solution in a smartphone is feasible, since the computational cost is low in comparison with the number of operations per second that a smartphone can perform. Therefore, a future work can be to test the features in other realistic mobile computing environments with Android or i-OS.

Author Contributions

Conceptualization, I.M.-H. and R.G.-P.; methodology, I.M.-H. and R.G.-P.; software, I.M.-H. and J.G.-G. validation, I.M.-H., R.G.-P., F.S., and J.G.-G.; formal analysis, I.M.-H., R.G.-P., and F.S.; investigation, I.M.-H.; resources, I.M.-H., M.R.-Z., and F.S.; data curation, I.M.-H. and J.G.-G.; writing—original draft preparation, I.M.-H. and F.S.; writing—review and editing, I.M.-H., R.G.-P., M.R-Z., and F.S.; visualization, I.M.-H., J.G.-G., and F.S.; supervision, I.M.-H., R.G.-P., M.R.-Z., and F.S.; project administration, M.R.-Z. and R.G.-P.; funding acquisition, M.R.-Z. and R.G.-P. All authors have read and agreed to the published version of the manuscript.

Funding

This work has been funded by the Spanish Ministry of Economy and Competitiveness FEDER under Project RTI2018-098085-B-C42.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Database

The description of the database in detail can be found in [31] and https://www.mdpi.com/1424-8220/19/24/5524 associated with [34]. The database consists of the ECG and TEB signals recorded from 40 subjects (students and climbers) aged between 20 and 49 years, of which 12 were females and 28 males. The sampling frequency is 100 Hz for TEB and 250 Hz for ECG. The total experiment lasts approximately 90 min. All of the experiments were performed under the conditions of respect for individual rights and ethical principles that govern biomedical research involving humans, and written informed consent was obtained from all participants. The experiments were approved by the Research Ethics Committee at the University of Alcala, and the protocol number assigned to this experiment is CEI: 2013/025/201130624.
The part used in the present paper studied the emotional response, in which the main objective is to distinguish among three different emotions:
  • Neutral emotional response
  • Sadness emotion
  • Disgust emotion
For this purpose, the subject under study watched three sequences of several films. In Figure A1 it is possible to observe the films used to elicit the three emotions under study. The sequences selected were watched twice, and we used the recordings from the last 200 s of each movie, generating 400 s for the neutral emotion, 400 s for sadness and 400 s for disgust for each subject. The first film was Earth, the second one was elected by the user among American History X, I am a legend, or La vita é bella and finally Cannibal Holocaust.
Figure A1. Films used to elicit the three emotions: neutral, sadness and disgust.
Figure A1. Films used to elicit the three emotions: neutral, sadness and disgust.
Sensors 20 00309 g0a1

References

  1. Jansen, A.S.; Van Nguyen, X.; Karpitskiy, V.; Mettenleiter, T.C.; Loewy, A.D. Central command neurons of the sympathetic nervous system: Basis of the fight-or-flight response. Science 1995, 270, 644. [Google Scholar] [CrossRef] [PubMed]
  2. Taylor, S.E.; Klein, L.C.; Lewis, B.P.; Gruenewald, T.L.; Gurung, R.A.; Updegraff, J.A. Biobehavioral responses to stress in females: Tend-and-befriend, not fight-or-flight. Psychol. Rev. 2000, 107, 411. [Google Scholar] [CrossRef] [PubMed]
  3. Cai, J.; Liu, G.; Hao, M. The research on emotion recognition from ECG signal. In Proceedings of the International Conference on Information Technology and Computer Science, Kiev, Ukraine, 25–26 July 2009; pp. 497–500. [Google Scholar]
  4. Dhillon, H.S.; Rekhi, N.S. The effect of emotions on electrocardiogram. Acad. Res. Int. 2011, 1, 280. [Google Scholar]
  5. Lee, C.; Yoo, S.K.; Park, Y.; Kim, N.; Jeong, K.; Lee, B. Using neural network to recognize human emotions from heart rate variability and skin resistance. In Proceedings of the 2005 IEEE Engineering in Medicine and Biology 27th Annual Conference, Shanghai, China, 17–18 January 2006; pp. 5523–5525. [Google Scholar]
  6. Ma, C.-W.; Liu, G.-Y. Feature extraction, feature selection and classification from electrocardiography to emotions. In Proceedings of the 2009 International Conference on Computational Intelligence and Natural Computing, Wuhan, China, 6–7 June 2009; pp. 190–193. [Google Scholar]
  7. Wiens, S.; Mezzacappa, E.S.; Katkin, E.S. Heartbeat detection and the experience of emotions. Cogn. Emot. 2000, 14, 417–427. [Google Scholar] [CrossRef]
  8. Calhoun, C.; Solomon, R.C. What is an Emotion: Classic Readings in Philosophical Psychology; Oxford University Press: Oxford, UK, 1984. [Google Scholar]
  9. Doll, A.; Hölzel, B.K.; Bratec, S.M.; Boucard, C.C.; Xie, X.; Wohlschläger, A.M.; Sorg, C. Mindful attention to breath regulates emotions via increased amygdala–prefrontal cortex connectivity. Neuroimage 2016, 134, 305–313. [Google Scholar] [CrossRef]
  10. Homma, I.; Masaoka, Y. Breathing rhythms and emotions Experimental physiology. Wiley Online Lib. 2008, 93, 1011–1021. [Google Scholar]
  11. Paunovic, K.; Stojanov, V.; Jakovljevic, B.; Belojevic, G. Thoracic Bioelectrical Impedance Assessment of the Hemodynamic Reactions to Recorded Road-Traffic Noise in Young Adults. Environ. Res. 2014, 129, 52–58. [Google Scholar] [CrossRef]
  12. Seoane, F.; Mohino-Herranz, I.; Ferreira, J.; Alvarez, L.; Buendia, R.; Ayllón, D.; Llerena, C.; Gil-Pita, R. Wearable biomedical measurement systems for assessment of mental stress of combatants in real time. Sensors 2014, 14, 7120–7141. [Google Scholar] [CrossRef] [Green Version]
  13. Sloan, R.; Shapiro, P.; Bagiella, E.; Boni, S.; Paik, M.; Bigger, J.; Steinman, R.C.; Gorman, J.M. Effect of mental stress throughout the day on cardiac autonomic control. Biol. Psychol. 1994, 37, 89–99. [Google Scholar] [CrossRef]
  14. Hataji, O.; Nishii, Y.; Ito, K.; Sakaguchi, T.; Saiki, H.; Suzuki, Y.; D’Alessandro-Gabazza, C.; Fujimoto, H.; Kobayashi, T.; Gabazza, E.C.; et al. Smart watch-based coaching with tiotropium and olodaterol ameliorates physical activity in patients with chronic obstructive pulmonary disease. Exp. Ther. Med. 2017, 14, 4061–4064. [Google Scholar] [PubMed]
  15. Zheng, Y.L.; Ding, X.R.; Poon, C.C.Y.; Lo, B.P.L.; Zhang, H.; Zhou, X.L.; Yang, G.-Z.; Zhao, N.; Zhang, Y.-T. Unobtrusive sensing and wearable devices for health informatics. IEEE Trans. Biomed. Eng. 2014, 61, 1538–1554. [Google Scholar] [CrossRef] [PubMed]
  16. Yang, T.; Xie, D.; Li, Z.; Zhu, H. Recent advances in wearable tactile sensors: Materials, sensing mechanisms, and device performance. Mater. Sci. Eng. R Rep. 2017, 115, 1–37. [Google Scholar] [CrossRef]
  17. Jean, J.S.N.; Wang, J. Weight smoothing to improve network generalization. IEEE Trans. Neural Netw. 1994, 5, 752–763. [Google Scholar] [CrossRef] [PubMed]
  18. Zhao, Z.-Q.; Huang, D.-S. A mended hybrid learning algorithm for radial basis function neural networks to improve generalization capability. Appl. Math. Model. 2007, 31, 1271–1281. [Google Scholar] [CrossRef]
  19. Anguita, D.; Ridella, S.; Rivieccio, F. K-fold generalization capability assessment for support vector classifiers. In Proceedings of the 2005 IEEE International Joint Conference on Neural Networks, Montreal, QC, Canada, 31 July–4 August 2005; pp. 1271–1281. [Google Scholar]
  20. Aricò, P.; Borghini, G.; Di Flumeri, G.; Sciaraffa, N.; Babiloni, F. Passive BCI beyond the lab: Current trends and future directions. Physiol. Meas. 2018, 39, 08TR02. [Google Scholar] [CrossRef] [PubMed]
  21. Mohino-Herranz, I.; Gil-Pita, R.; Alonso-Diaz, S.; Rosa-Zurera, M. Synthetical enlargement of mfcc based training sets for emotion recognition. Int. J. Comput. Sci. Inf. Technol. 2014, 6, 249–259. [Google Scholar]
  22. Kohavi, R.; John, G.H. Wrappers for feature subset selection. Artif. Intell. 1997, 97, 273–324. [Google Scholar] [CrossRef] [Green Version]
  23. Shah, S.C.; Kusiak, A. Data mining and genetic algorithm based gene/SNP selection. Artif. Intell. Med. 2004, 31, 183–196. [Google Scholar] [CrossRef]
  24. Long, N.; Gianola, D.; Rosa, G.; Weigel, K. Dimension reduction and variable selection for genomic selection: Application to predicting milk yield in Holsteins. J. Anim. Breed. Genet. 2011, 128, 247–257. [Google Scholar] [CrossRef]
  25. Meiri, R.; Zahavi, J. Using simulated annealing to optimize the feature selection problem in marketing applications. Eur. J. Oper. Res. 2006, 171, 842–858. [Google Scholar] [CrossRef]
  26. Hernandez, J.C.H.; Duval, B.; Hao, J.K. A genetic embedded approach for gene selection and classification of microarray data. EvoBIO 2007, 7, 90–101. [Google Scholar]
  27. Healey, J.A.; Picard, R.W. Detecting stress during real-world driving tasks using physiological sensors. Intell. Transp. Syst. IEEE Trans. 2005, 6, 156–166. [Google Scholar] [CrossRef] [Green Version]
  28. McCraty, R.; Atkinson, M.; Tiller, W.A.; Rein, G.; Watkins, A.D. The effects of emotions on short-term power spectrum analysis of heart rate variability. Am. J. Cardiol. 1995, 76, 1089–1093. [Google Scholar] [CrossRef]
  29. Piccirillo, G.; Vetta, F.; Fimognari, F.; Ronzoni, S.; Lama, J.; Cacciafesta, M.; Marigliano, V. Power spectral analysis of heart rate variability in obese subjects: Evidence of decreased cardiac sympathetic responsiveness. Int. J. Obes. Relat. Metab. Disord. J. Int. Assoc. Study Obes. 1996, 20, 825–829. [Google Scholar]
  30. Billman, G.E. The LF/HF ratio does not accurately measure cardiac sympatho-vagal balance. Front. Physiol. 2013, 4, 26. [Google Scholar] [CrossRef] [Green Version]
  31. Mohino-Herranz, I.; Gil-Pita, R.; Ferreira, J.; Rosa-Zurera, M.; Seoane, F. Assessment of mental, emotional and physical stress through analysis of physiological signals using smartphones. Sensors 2015, 15, 25607–25627. [Google Scholar] [CrossRef]
  32. Mehrnia, A.; Willson, A.N., Jr. On optimal IFIR filter design. In Proceedings of the 2004 IEEE International Symposium on Circuits and Systems (IEEE Cat. No.04CH37512), Vancouver, BC, Canada, 23–26 May 2004; Volume 3. [Google Scholar]
  33. Davis, L. Handbook of Genetic Algorithms; Van Nostrand Reinhold: New York, NY, USA.
  34. Mohino-Herranz, I.; Gil-Pita, R.; Rosa-Zurera, M.; Seoane, F. Activity Recognition Using Wearable Physiological Measurements: Selection of Features from a Comprehensive Literature Study. Sensors 2019, 19, 5524. [Google Scholar] [CrossRef] [Green Version]
  35. Prabhakar, V.; Varde, P.; Michael, G. Risk-Based Engineering: An Integrated Approach to Complex Systems-special Reference to Nuclear Plants; Springer: Basel, Switzerland, 2019. [Google Scholar]
  36. Liu, S.; Zhu, M.; Yu, D.J.; Rasin, A.; Young, S.D. Using Real-Time Social Media Technologies to Monitor Levels of Perceived Stress and Emotional State in College Students: A Web-Based Questionnaire Study. JMIR Ment. Health 2017, 4, e2. [Google Scholar] [CrossRef]
  37. Liu, Y.J.; Yu, M.; Zhao, G.; Song, J.; Ge, Y.; Shi, Y. Real-time movie-induced discrete emotion recognition from EEG signals. IEEE Trans. Affect. Comput. 2018, 9, 550–562. [Google Scholar] [CrossRef]
  38. Turabzadeh, S.; Meng, H.; Swash, R.M.; Pleva, M.; Juhar, J. Real-time emotional state detection from facial expression on embedded devices. In Proceedings of the 2017 Seventh International Conference on Innovative Computing Technology (INTECH), Luton, UK, 16–18 August 2017; pp. 46–51. [Google Scholar]
  39. Keogh, E.; Mueen, A. Curse of dimensionality. In Encyclopedia of Machine Learning; Springer: Basel, Switzerland, 2011; pp. 257–258. [Google Scholar]
  40. Zhuo, L.; Zheng, J.; Li, X.; Wang, F.; Ai, B.; Qian, J. A genetic algorithm based wrapper feature selection method for classification of hyperspectral images using support vector machine. In Proceedings of the Geoinformatics 2008 and Joint Conference on GIS and Built Environment: Geo-Simulation and Virtual GIS Environments, Guangzhou, China, 28–29 June 2008. [Google Scholar]
  41. Huang, J.; Cai, Y.; Xu, X. A hybrid genetic algorithm for feature selection wrapper based on mutual information. Pattern Recognit. Lett. 2007, 28, 1825–1844. [Google Scholar] [CrossRef]
  42. Kohavi, R.; Sommerfield, D. Feature Subset Selection Using theWrapper Method: Overfitting and Dynamic Search Space Topology. In Proceedings of the First International Conference on Knowledge Discovery and Data Mining, Montreal, QC, Canada, 20–21 August 1995; pp. 192–197. [Google Scholar]
Figure 1. Complete system.
Figure 1. Complete system.
Sensors 20 00309 g001
Figure 2. Wrapper approach scheme.
Figure 2. Wrapper approach scheme.
Sensors 20 00309 g002
Figure 3. Feature extraction scheme.
Figure 3. Feature extraction scheme.
Sensors 20 00309 g003
Figure 4. Error probabilities obtained using standard design error optimization (SDEO) and fold-based error optimization (KFBEO).
Figure 4. Error probabilities obtained using standard design error optimization (SDEO) and fold-based error optimization (KFBEO).
Sensors 20 00309 g004
Table 1. Types of filters and computational cost.
Table 1. Types of filters and computational cost.
NameTypeOrder (N)Cutoff (Hz)
L F E C G Low-pass (anti-aliasing)1003
M F E C G Low-pass (anti-aliasing)10030
R F E C G Band-pass (IFIR)11500.1–0.5
L F T E B Low-pass (anti-aliasing)1003
R F T E B Band-pass (IFIR)4000.1–0.5
E F T E B Band-pass (IFIR)4000.8–2.9
Table 2. Computational cost for each filter in the real-time implementation.
Table 2. Computational cost for each filter in the real-time implementation.
Name LF ECG MF ECG RF ECG LF TEB RF TEB EF TEB
C o s t ( N o p ) 25,00025,00011,50010,00040004000
Table 3. List of statistics and number of operations per second required.
Table 3. List of statistics and number of operations per second required.
ParameterOperations per Second
Trimmed mean of 25%27,805
Median27,580
Percentile 25%27,580
Percentile 75%27,580
Kurtosis2701
Skewness2101
Standard deviation1201
Mean absolute deviation1200
Geometric mean3901
Harmonic mean3301
Baseline550
Maximum300
Minimum300
Mean300
Table 4. Features selected more than 20% in all experiments for SDEO.
Table 4. Features selected more than 20% in all experiments for SDEO.
FeaturePercentage (%)
TEB_PPM_Baseline100
TEB_PPM_Max99.9
TEB_PPM_Mean99.9
TEB_PPM_Min99.3
TEB_PPM_Std97.6
TEB_PPM_Mad97.6
TEB_PPM_Harmean77.3
TEB_RT_Baseline73.3
TEB_RT_Mean73.3
TEB_RT_Min72.8
TEB_RT_Max44.3
E_PPM_Mean44.3
E_PPM_Min44.2
E_PPM_Baseline43.9
TEB_PPM_Skewness43.6
E_PPM_Max42.7
E_PPM_Std40.5
E_PPM_Mad39.7
TEB_RD_Baseline30.0
TEB_RD_Mean30.0
E_PPM_Harmean27.9
TEB_RT_Mad25.1
TEB_RD_Mad25.1
TEB_RD_Max20.6
Table 5. Features selected more than 20% in all experiments for KFBEO.
Table 5. Features selected more than 20% in all experiments for KFBEO.
FeaturePercentage (%)
TEB_PPM_Baseline100
TEB_RT_Max100
TEB_PPM_Mean97
E_PPM_Mad48.73
E_PPM_Mean48.58

Share and Cite

MDPI and ACS Style

Mohino-Herranz, I.; Gil-Pita, R.; García-Gómez, J.; Rosa-Zurera, M.; Seoane, F. A Wrapper Feature Selection Algorithm: An Emotional Assessment Using Physiological Recordings from Wearable Sensors. Sensors 2020, 20, 309. https://doi.org/10.3390/s20010309

AMA Style

Mohino-Herranz I, Gil-Pita R, García-Gómez J, Rosa-Zurera M, Seoane F. A Wrapper Feature Selection Algorithm: An Emotional Assessment Using Physiological Recordings from Wearable Sensors. Sensors. 2020; 20(1):309. https://doi.org/10.3390/s20010309

Chicago/Turabian Style

Mohino-Herranz, Inma, Roberto Gil-Pita, Joaquín García-Gómez, Manuel Rosa-Zurera, and Fernando Seoane. 2020. "A Wrapper Feature Selection Algorithm: An Emotional Assessment Using Physiological Recordings from Wearable Sensors" Sensors 20, no. 1: 309. https://doi.org/10.3390/s20010309

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop