1. Introduction
Empathy is a construct that has been addressed from multidimensional perspectives for decades [
1,
2]. One of the most popular distinctions has been pointed out between cognitive and affective empathy. The first refers to the ability to understand emotions, while the latter refers to the ability to share those emotions [
3]. This distinction has found anatomical support, with different structures being differentiated for each dimension [
4,
5]. While the left anterior cingulate cortex has been related to cognitive empathy, the right anterior insula and the right inferior frontal gyrus seems to be more related to the affective, according to a meta-analysis [
6].
As both definitions of empathy are considered latent constructs that cannot be directly assessed, this study is focused on behavioral response. While facial emotion recognition cannot be considered as a direct reflection of empathy, the literature has considered an overlap with emotional empathy in perceptual processing, as well as of cognitive empathy with inferential processes [
7,
8]. The literature also seems to indicate that highly empathetic individuals are more sensitive to human expressions, hypothesizing that highly-empathetic individuals might pay more attention than individuals with lower empathetic scores when discriminating facial expressions [
9]. Some studies have questioned if trait empathy correlates with attention elicited by discriminating facial expressions, while others have indicated evidence of relationship between trait empathy and brain activation elicited by facial expressions [
10].
According to the mirror neuron activation approach, for a proper empathic response, perceptual ability is necessary to process socially relevant stimuli, e.g., facial expressions [
11]. Not surprisingly, individuals with higher affective empathy also seem to be particularly sensitive in reacting to facial expressions in an accurate way [
12]. However, it has also been hypothesized that high affective empathy may interfere with the efficiency of visual recognition, causing personal distress [
13]. In that case, cognitive empathy might be a more favorable resource. Cognitive empathy has sometimes been related to theory-of-mind or perspective-taking abilities [
14]. However, when focusing on popular tools such as the eyes test [
15] in the field, a relationship of this tool with affective measures has been described [
16,
17].
The indirect and probably most common way of measuring affective and cognitive empathy is through questionnaires. The revised
Reading the Mind in the Eyes test is one of the most widely used tools. It is employed to assess emotion recognition, providing insight into the ability to recognize emotional expressions by determining an individual’s complex cognitive mental state from a partial facial expression. On this front, the literature has pointed out that facial emotion recognition partially mediated the relationship between reasoning by analogy and social cognition by employing the Reading the Mind in the Eyes Test [
18]. With regards to the distinction between cognitive and affective empathy, an optimal strategy is to use the Cognitive and Affective Empathy Test [
19], also known as TECA (
Test de Empatía Cognitiva y Afectiva) [
20] or the Interpersonal Reactivity Index [
1]. Both tools are divided into four subscales, two of them assessing cognitive empathy (Perspective Taking and Emotional Understanding), the other two assessing affective empathy (Empathic Distress and Empathic Happiness).
By combining a questionnaire methodology with an experimental task, this study aims to examine the relationship between affective and cognitive empathy scores and perceptual face recognition skills under the cognitive and affective model proposed by Davis (1980;1983). Considering previous literature linking cognitive empathy to theory-of-mind [
14], it is hypothesized that cognitive empathy scores are directly related to hit rates in a simple emotional face recognition task (hypothesis 1). Individuals with high cognitive empathy scores would show better perceptual and abstractive processing. Finally, it is hypothesized that affective empathy scores are directly related to reaction times (hypothesis 2). This would explain a higher cognitive cost of emotion processing in individuals with high affective empathy scores, as described in the previous literature [
13].
2. Materials and Methods
2.1. Procedure and Participants
The study was conducted after obtaining ethical approval from the Universidad Católica San Vicente Mártir Ethics Committee (UCV2017-2018-31) and receiving participants’ written informed consent. Anonymous data was collected from young adults on the Southeastern coast of Spain in March 2022.
Before conducting the experiment, participants were informed about the implications of the study and were asked to sign the written consent form. Participants were assessed in a silent room with a computer. For the first part, the face recognition task, on the screen they could observe two pictures of two people: a man and a woman with neutral expression, and six emotional facial expressions: anger, surprise, disgust, enjoyment, fear, and sadness [
21]. Participants had to memorize the two faces in order to be able to recognize them afterwards. Furthermore, these figures could be expressing different emotions. As depicted in
Figure 1, each time they recognized one on the picture, they had to tap the M keyboard key, and each time they were presented with a picture of someone else, they had to press the Z. Due to the copyright of the KDEF images, they have been simulated in
Figure 1. The DMDX software [
22] was employed for image presentation.
A total of 18 young adults participated in the study. Sex was controlled across the sample (50% were women and 50% men) with ages ranging from 25 to 30 years. In terms of inclusion criteria, all the participants had normal or corrected to normal vision, were native Spanish speakers and did not report cognitive or neurological disorders. The participants were chosen after a personal interview.
2.2. Measures
After a sociodemographic battery, individual differences in empathic tendencies were evaluated using the Cognitive and Affective Empathy Test (TECA), which consists of 33 questions, with Likert-type responses ranging from 1 (strongly disagree) to 5 (strongly agree). It allows the measurement of both cognitive empathy and emotional empathy. Regarding the validity of the TECA, it has shown strong convergent validity, with high correlations with the Spanish adaptation of the Interpersonal Reactivity Index [
23].
The eyes test (Baron-Cohen et al., 2001) was used to assess cognitive empathy. It consists of 36 slides with pictures of facial expressions on the eye line only. Each sheet contains four adjectives from which the participant must choose that that best describes the expression in the image. The adjectives correspond to complex emotional states such as bored or arrogant, rather than simple emotional states such as sad or happy. The test is corrected by assigning one point for each correct answer, so the score varies from 0 to 36.
Lastly, the Karolinska battery of images expressing different emotions was employed for the experimental task [
24]. The task has two blocks, a presentation, and a recognition phase. Thus, there was a total of 392 items per participant (or 28 items per 7 conditions, where 14 were a man a 14 a woman); 192 in the presentation and 392 in the recognition block (192 targets and 192 distractors divided into 7 face emotions). The emotions selected as stimuli, as described in previous literature, were fear, sadness, happiness, anger, surprise, and disgust, besides another neutral image. The DMDX software was employed [
22]. This is a display system that carries out the randomization of the images’ presentation and records reaction time of the participants to respond to the stimuli presented.
2.3. Design and Statistical Analysis
This study uses a combination of experimental and survey method. G*Power [
25] was employed to calculate the sample size. A medium effect for an experimental paradigm suggested a minimum of 18 participants. Stimuli in the experiment were counterbalanced to diminish progressive error. This allowed us to examine any possible biases related to the stimuli characteristics. For example, one of the images has characteristics which people remember better than the others. It should be noted that we did not find statistically significant differences in terms of list of presentation (
p > 0.05). A cut-off technique was applied for reaction time lower than 250 ms and higher than 1500, as described in previous literature [
26,
27,
28,
29], trimming 3% from the whole data set. Data analysis was performed using JASP (Version 0.12.2).
3. Results
Results were examined in terms of accuracy (hypothesis 1) and response latency (hypothesis 2).
Table 1 describes the whole scores on the variables under study. Differences across participant’s sex were examined using Mann-Whitney U test. No statistically significant difference was found (
p > 0.05), though Reaction Time (RT) across distraction and target images related to each other in a positive way, and Empathic Distress (ED) and Empathic Happiness (EH) appeared to be in the same direction. When Spearman’s rho coefficients were carried out to check for accuracy of each emotion (hits) and the other questionnaire variables, a positive correlation between Emotional Understanding (EU) and surprise for target images was found (rho = 0.528;
p < 0.05). Moreover, a positive correlation between Perspective Taking (PT) and disgust for distracting images was found (rho = 0.494;
p < 0.05).
Secondly, when the same analysis was carried out for RT across each emotion and the other questionnaire variables, a positive correlation between PT and disgust for target images was found (rho = 0.473; p < 0.05), as well as sadness for target images (rho = 0.570; p < 0.05). PT showed a positive relationship with surprise for target images, but this result did not reach statistically significant level (rho = 0.469; p = 0.05). With regards to the distracting stimuli, there was a positive correlation between PT and disgust (rho = 0.526; p < 0.05).
Lastly, linear multiple regression models, were carried out. Thus, the scores for eyes test and TECA were entered as the predictors and the outcome variables were the seven stimuli conditions for emotional faces (neutral, fear, sadness, happiness, anger, surprise and disgust) in the target and distracting condition across accuracy and RT dependent variables. None of these models reached statistically significant levels except for the model on the prediction of surprise emotion for the accuracy on target images. In this case the EU depicted the following values: β = 0.66; p < 0.05.
4. Conclusions and Discussion
The aim of this study was to examine the relationship between affective and cognitive empathy scores with emotional face recognition in university students. For this purpose, an experimental task of emotional face recognition, combined with questionnaires measuring cognitive and affective empathy (TECA), and the eyes test, which measures individual differences in theory of mind capabilities and more precisely emotion perception [
16,
17], were used in a university sample of 18 participants. Cognitive empathy sub-factors were found to be more related to the hit rate on recognition of surprise faces and the discarding of disgust faces. The hit rate measure for distracting disgust faces was positively related to perspective taking. Reaction time to Cognitive empathy subfactors were found to be related to recognition of disgust, surprise and sadness. Finally, Perspective taking was again related to the discarding of disgust reaction time. The relationships between affective empathy and other measures for emotional face recognition were not statistically significant.
According to the first hypothesis, current results support that cognitive empathy scores are directly related to hit rates for recognition of surprise and disgust facial expressions. Individuals with high cognitive empathy scores show higher accuracy for disgust and surprise faces. However, this was not the case for affective empathy. The findings of this study aligns with the definition of empathy as a cognitive dimension in which one person attempts to comprehend the feelings, thoughts, or intentions of another person [
30]. However, when individuals with high emotional reactions are tested with TECA and eyes test, affective measure does not seem to benefit from this strategy.
The results of the second hypothesis does not show any interference in reaction times or cognitive cost for individuals with higher affective empathy scores. Thus, current results seem to be inconclusive for the second hypothesis. One of the limitations that might explain this result could be related to the use of instruments under study. Although TECA and IRI measures the same constructs, there are differences between the sub-factors of both scales. Previous literature using IRI found affective empathy to have opposite relationships with the recognition of facial expressions of emotions. Specifically, the authors have described that empathic concern was positively related, while personal distress was negatively related, to accurate emotion recognition. However, the present results do not support this finding with the TECA and eyes test.
Moreover, the sample size is one of the limitations of the of the present study. While experimental studies with repeated measures can be used with smaller numbers of participants, future studies should consider larger a samplesl using survey techniques.
The novelty of this study lies in the analysis of emotional face recognition and the study of the empathy relationship under the Davis model (1980, 1983). These results might be of interest for interventions to help recognize emotions and improve social interactions. Knowing that there are individual differences in cognitive and affective empathy, as well as in their relationship with behavioral responses such as the recognition or dismissal of emotional faces, the results could be connected to aspects of social interaction and psychotherapy.
Author Contributions
Conceptualization, C.M.-T. and T.Q.I.; methodology and software, C.M.-T.; resources, C.M.-T. and P.M.-R.; data curation, P.M.-R.; writing—original draft preparation, C.M.-T.; writing—review and editing, as well as funding acquisition, all authors. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by Ayudas TFG−TFM para proyectos de Desarrollo Sostenible 2022/Vicerrectorado de Alumnos y Acción Social (Universidad Católica de Valencia San Vicente Mártir).
Institutional Review Board Statement
The study was conducted in accordance with the Declaration of Helsinki, and approved by the Universidad Católica San Vicente Mártir Ethics Committee (UCV2017-2018-31).
Informed Consent Statement
Informed consent was obtained from all subjects involved in the study.
Data Availability Statement
The data presented in this study are available on request from the corresponding and second author.
Acknowledgments
We would like to thank all the participants involved in this study for their invaluable time. We would like to thank the Universidad Católica de Valencia San Vicente Mártir, and the Vicerrectorado de Alumnos y Acción Social (UCV) for their support. We also thank the Master de Psicología Jurídica from this institution, where the research takes place.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Davis, M.H. Measuring Individual Differences in Empathy: Evidence for a Multidimensional Approach. J. Pers. Soc. Psychol. 1983, 44, 113–126. [Google Scholar] [CrossRef]
- Powell, P.A. Individual Differences in Emotion Regulation Moderate the Associations between Empathy and Affective Distress. Motiv. Emot. 2018, 42, 602–613. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Thompson, N.M.; van Reekum, C.M.; Chakrabarti, B. Cognitive and Affective Empathy Relate Differentially to Emotion Regulation. Affect. Sci. 2022, 3, 118–134. [Google Scholar] [CrossRef] [PubMed]
- Schnell, K.; Bluschke, S.; Konradt, B.; Walter, H. Functional Relations of Empathy and Mentalizing: An FMRI Study on the Neural Basis of Cognitive Empathy. NeuroImage 2011, 54, 1743–1754. [Google Scholar] [CrossRef] [PubMed]
- Uribe, C.; Puig-Davi, A.; Abos, A.; Baggio, H.C.; Junque, C.; Segura, B. Neuroanatomical and Functional Correlates of Cognitive and Affective Empathy in Young Healthy Adults. Front. Behav. Neurosci. 2019, 13, 85. [Google Scholar] [CrossRef]
- Fan, Y.; Duncan, N.W.; de Greck, M.; Northoff, G. Is There a Core Neural Network in Empathy? An FMRI Based Quantitative Meta-Analysis. Neurosci. Biobehav. Rev. 2011, 35, 903–911. [Google Scholar] [CrossRef]
- Holland, A.C.; O’Connell, G.; Dziobek, I. Facial Mimicry, Empathy, and Emotion Recognition: A Meta-Analysis of Correlations. Cogn. Emot. 2021, 35, 150–168. [Google Scholar] [CrossRef]
- Schirmer, A.; Adolphs, R. Emotion Perception from Face, Voice, and Touch: Comparisons and Convergence. Trends Cogn. Sci. 2017, 21, 216–228. [Google Scholar] [CrossRef] [Green Version]
- Choi, D.; Watanuki, S. Effect of Empathy Trait on Attention to Faces: An Event-Related Potential (ERP) Study. J. Physiol. Anthropol. 2014, 33, 4. [Google Scholar] [CrossRef] [Green Version]
- Jabbi, M.; Swart, M.; Keysers, C. Empathy for Positive and Negative Emotions in the Gustatory Cortex. NeuroImage 2007, 34, 1744–1753. [Google Scholar] [CrossRef]
- Enticott, P.G.; Johnston, P.J.; Herring, S.E.; Hoy, K.E.; Fitzgerald, P.B. Mirror Neuron Activation Is Associated with Facial Emotion Processing. Neuropsychologia 2008, 46, 2851–2854. [Google Scholar] [CrossRef] [PubMed]
- Dimberg, U.; Andréasson, P.; Thunberg, M. Emotional Empathy and Facial Reactions to Facial Expressions. J. Psychophysiol. 2011, 25, 26–31. [Google Scholar] [CrossRef]
- Israelashvili, J.; Sauter, D.; Fischer, A. Two Facets of Affective Empathy: Concern and Distress Have Opposite Relationships to Emotion Recognition. Cogn. Emot. 2020, 34, 1112–1122. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Baron-Cohen, S.; Wheelwright, S. The Empathy Quotient: An Investigation of Adults with Asperger Syndrome or High Functioning Autism, and Normal Sex Differences. J. Autism Dev. Disord. 2004, 34, 163–175. [Google Scholar] [CrossRef]
- Baron-Cohen, S.; Wheelwright, S.; Hill, J.; Raste, Y.; Plumb, I. The “Reading the Mind in the Eyes” Test Revised Version: A Study with Normal Adults, and Adults with Asperger Syndrome or High-functioning Autism. J. Child Psychol. Psychiatry 2001, 42, 241–251. [Google Scholar] [CrossRef] [PubMed]
- Guastella, A.J.; Einfeld, S.L.; Gray, K.M.; Rinehart, N.J.; Tonge, B.J.; Lambert, T.J.; Hickie, I.B. Intranasal Oxytocin Improves Emotion Recognition for Youth with Autism Spectrum Disorders. Biol. Psychiatry 2010, 67, 692–694. [Google Scholar] [CrossRef] [PubMed]
- Olderbak, S.; Wilhelm, O.; Olaru, G.; Geiger, M.; Brenneman, M.W.; Roberts, R.D. A Psychometric Analysis of the Reading the Mind in the Eyes Test: Toward a Brief Form for Research and Applied Settings. Front. Psychol. 2015, 6, 1503. [Google Scholar] [CrossRef] [Green Version]
- Seo, E.; Koo, S.J.; Kim, Y.J.; Min, J.E.; Park, H.Y.; Bang, M.; Lee, E.; An, S.K. Reading the Mind in the Eyes Test: Relationship with Neurocognition and Facial Emotion Recognition in Non-Clinical Youths. Psychiatry Investig. 2020, 17, 835–839. [Google Scholar] [CrossRef]
- Van Hoey, J.; Moret-Tatay, C.; Santolaya Prego de Oliver, J.A.; Beneyto-Arrojo, M.J. Profile Changes in Male Partner Abuser After an Intervention Program in Gender-Based Violence. Int. J. Offender Ther. Comp. Criminol. 2019, 65, 1411–1422. [Google Scholar] [CrossRef]
- López-Pérez, B.; Fernández-Pinto, I.; García, F.J.A. Test de Empatía Cognitiva y Afectiva. Madr. Spain Tea 2008. [Google Scholar]
- Ekman, P.; Friesen, W. Pictures of Facial Affect; Consulting Psychologists Press: Palo Alto, CA, USA, 1976. [Google Scholar]
- Forster, K.I.; Forster, J.C. DMDX: A Windows Display Program with Millisecond Accuracy. Behav. Res. Methods Instrum. Comput. 2003, 35, 116–124. [Google Scholar] [CrossRef] [PubMed]
- Pérez-Albéniz, A.; De Paúl, J.; Etxeberría, J.; Paz Montes, M.; Torres, E. Adaptación de Interpersonal Reactivity Index (IRI) al Español. Psicothema 2003, 15, 267–272. [Google Scholar]
- Lundqvist, D.; Flykt, A.; Öhman, A. Karolinska Directed Emotional Faces. 2015. [Google Scholar]
- Faul, F.; Erdfelder, E.; Lang, A.-G.; Buchner, A. G*Power 3: A Flexible Statistical Power Analysis Program for the Social, Behavioral, and Biomedical Sciences. Behav. Res. Methods 2007, 39, 175–191. [Google Scholar] [CrossRef]
- Moret-Tatay, C.; Gamermann, D.; Navarro-Pardo, E.; Fernández de Córdoba Castellá, P. ExGUtils: A Python Package for Statistical Analysis With the Ex-Gaussian Probability Density. Front. Psychol. 2018, 9, 612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Moret-Tatay, C.; Baixauli-Fortea, I.; Sevilla, M.D.G.; Irigaray, T.Q. Can You Identify These Celebrities? A Network Analysis on Differences between Word and Face Recognition. Mathematics 2020, 8, 699. [Google Scholar] [CrossRef]
- Moret-Tatay, C.; Baixauli-Fortea, I.; Grau-Sevilla, M.D. Profiles on the Orientation Discrimination Processing of Human Faces. Int. J. Environ. Res. Public Health 2020, 17, 5772. [Google Scholar] [CrossRef] [PubMed]
- Moret-Tatay, C.; Wester, A.G.; Gamermann, D. To Google or Not: Differences on How Online Searches Predict Names and Faces. Mathematics 2020, 8, 1964. [Google Scholar] [CrossRef]
- Herrera, F.; Bailenson, J.; Weisz, E.; Ogle, E.; Zaki, J. Building Long-Term Empathy: A Large-Scale Comparison of Traditional and Virtual Reality Perspective-Taking. PLoS ONE 2018, 13, e0204494. [Google Scholar] [CrossRef]
| Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).